Changes to device_t struct to accomodate the upcoming PCI IRQ arbitration rewrite; Added device.c/h API to obtain name from the device_t struct; Significant changes to win/win_settings.c to clean up the code a bit and fix bugs; Ported all the CPU and AudioPCI commits from PCem; Added an API call to allow ACPI soft power off to gracefully stop the emulator; Removed the Siemens PCD-2L from the Dev branch because it now works; Removed the Socket 5 HP Vectra from the Dev branch because it now works; Fixed the Compaq Presario and the Micronics Spitfire; Give the IBM PC330 its own list of 486 CPU so it can have DX2's with CPUID 0x470; SMM fixes; Rewrote the SYSENTER, SYSEXIT, SYSCALL, and SYSRET instructions; Changed IDE reset period to match the specification, fixes #929; The keyboard input and output ports are now forced in front of the queue when read, fixes a number of bugs, including the AMI Apollo hanging on soft reset; Added the Intel AN430TX but Dev branched because it does not work; The network code no longer drops packets if the emulated network card has failed to receive them (eg. when the buffer is full); Changes to PCI card adding and renamed some PCI slot types, also added proper AGP bridge slot types; USB UHCI emulation is no longer a stub (still doesn't fully work, but at least Windows XP chk with Debug no longer ASSERT's on it); Fixed NVR on the the SMC FDC37C932QF and APM variants; A number of fixes to Intel 4x0 chipsets, including fixing every register of the 440LX and 440EX; Some ACPI changes.
287 lines
8.2 KiB
C
287 lines
8.2 KiB
C
static int opCMC(uint32_t fetchdat)
|
|
{
|
|
flags_rebuild();
|
|
cpu_state.flags ^= C_FLAG;
|
|
CLOCK_CYCLES(2);
|
|
PREFETCH_RUN(2, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int opCLC(uint32_t fetchdat)
|
|
{
|
|
flags_rebuild();
|
|
cpu_state.flags &= ~C_FLAG;
|
|
CLOCK_CYCLES(2);
|
|
PREFETCH_RUN(2, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
static int opCLD(uint32_t fetchdat)
|
|
{
|
|
cpu_state.flags &= ~D_FLAG;
|
|
CLOCK_CYCLES(2);
|
|
PREFETCH_RUN(2, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
static int opCLI(uint32_t fetchdat)
|
|
{
|
|
if (!IOPLp)
|
|
{
|
|
if ((!(cpu_state.eflags & VM_FLAG) && (cr4 & CR4_PVI)) ||
|
|
((cpu_state.eflags & VM_FLAG) && (cr4 & CR4_VME)))
|
|
{
|
|
cpu_state.eflags &= ~VIF_FLAG;
|
|
}
|
|
else
|
|
{
|
|
x86gpf(NULL,0);
|
|
return 1;
|
|
}
|
|
}
|
|
else
|
|
cpu_state.flags &= ~I_FLAG;
|
|
|
|
CLOCK_CYCLES(3);
|
|
PREFETCH_RUN(3, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
|
|
static int opSTC(uint32_t fetchdat)
|
|
{
|
|
flags_rebuild();
|
|
cpu_state.flags |= C_FLAG;
|
|
CLOCK_CYCLES(2);
|
|
PREFETCH_RUN(2, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
static int opSTD(uint32_t fetchdat)
|
|
{
|
|
cpu_state.flags |= D_FLAG;
|
|
CLOCK_CYCLES(2);
|
|
PREFETCH_RUN(2, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
static int opSTI(uint32_t fetchdat)
|
|
{
|
|
if (!IOPLp)
|
|
{
|
|
if ((!(cpu_state.eflags & VM_FLAG) && (cr4 & CR4_PVI)) ||
|
|
((cpu_state.eflags & VM_FLAG) && (cr4 & CR4_VME)))
|
|
{
|
|
if (cpu_state.eflags & VIP_FLAG)
|
|
{
|
|
x86gpf(NULL,0);
|
|
return 1;
|
|
}
|
|
else
|
|
cpu_state.eflags |= VIF_FLAG;
|
|
}
|
|
else
|
|
{
|
|
x86gpf(NULL,0);
|
|
return 1;
|
|
}
|
|
}
|
|
else
|
|
cpu_state.flags |= I_FLAG;
|
|
|
|
/*First instruction after STI will always execute, regardless of whether
|
|
there is a pending interrupt*/
|
|
cpu_end_block_after_ins = 2;
|
|
|
|
CLOCK_CYCLES(2);
|
|
PREFETCH_RUN(2, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
|
|
static int opSAHF(uint32_t fetchdat)
|
|
{
|
|
flags_rebuild();
|
|
cpu_state.flags = (cpu_state.flags & 0xff00) | (AH & 0xd5) | 2;
|
|
CLOCK_CYCLES(3);
|
|
PREFETCH_RUN(3, 1, -1, 0,0,0,0, 0);
|
|
|
|
#if (defined(USE_DYNAREC) && defined(USE_NEW_DYNAREC))
|
|
codegen_flags_changed = 0;
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
static int opLAHF(uint32_t fetchdat)
|
|
{
|
|
flags_rebuild();
|
|
AH = cpu_state.flags & 0xff;
|
|
CLOCK_CYCLES(3);
|
|
PREFETCH_RUN(3, 1, -1, 0,0,0,0, 0);
|
|
return 0;
|
|
}
|
|
|
|
static int opPUSHF(uint32_t fetchdat)
|
|
{
|
|
if ((cpu_state.eflags & VM_FLAG) && (IOPL < 3))
|
|
{
|
|
if (cr4 & CR4_VME)
|
|
{
|
|
uint16_t temp;
|
|
|
|
flags_rebuild();
|
|
temp = (cpu_state.flags & ~I_FLAG) | 0x3000;
|
|
if (cpu_state.eflags & VIF_FLAG)
|
|
temp |= I_FLAG;
|
|
PUSH_W(temp);
|
|
}
|
|
else
|
|
{
|
|
x86gpf(NULL,0);
|
|
return 1;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
flags_rebuild();
|
|
PUSH_W(cpu_state.flags);
|
|
}
|
|
CLOCK_CYCLES(4);
|
|
PREFETCH_RUN(4, 1, -1, 0,0,1,0, 0);
|
|
return cpu_state.abrt;
|
|
}
|
|
static int opPUSHFD(uint32_t fetchdat)
|
|
{
|
|
uint16_t tempw;
|
|
if ((cpu_state.eflags & VM_FLAG) && (IOPL < 3))
|
|
{
|
|
x86gpf(NULL, 0);
|
|
return 1;
|
|
}
|
|
if (cpu_CR4_mask & CR4_VME) tempw = cpu_state.eflags & 0x3c;
|
|
else if (CPUID) tempw = cpu_state.eflags & 0x24;
|
|
else tempw = cpu_state.eflags & 4;
|
|
flags_rebuild();
|
|
PUSH_L(cpu_state.flags | (tempw << 16));
|
|
CLOCK_CYCLES(4);
|
|
PREFETCH_RUN(4, 1, -1, 0,0,0,1, 0);
|
|
return cpu_state.abrt;
|
|
}
|
|
|
|
static int opPOPF_286(uint32_t fetchdat)
|
|
{
|
|
uint16_t tempw;
|
|
|
|
if ((cpu_state.eflags & VM_FLAG) && (IOPL < 3))
|
|
{
|
|
x86gpf(NULL, 0);
|
|
return 1;
|
|
}
|
|
|
|
tempw = POP_W(); if (cpu_state.abrt) return 1;
|
|
|
|
if (!(msw & 1)) cpu_state.flags = (cpu_state.flags & 0x7000) | (tempw & 0x0fd5) | 2;
|
|
else if (!(CPL)) cpu_state.flags = (tempw & 0x7fd5) | 2;
|
|
else if (IOPLp) cpu_state.flags = (cpu_state.flags & 0x3000) | (tempw & 0x4fd5) | 2;
|
|
else cpu_state.flags = (cpu_state.flags & 0x3200) | (tempw & 0x4dd5) | 2;
|
|
flags_extract();
|
|
|
|
CLOCK_CYCLES(5);
|
|
PREFETCH_RUN(5, 1, -1, 1,0,0,0, 0);
|
|
|
|
#if (defined(USE_DYNAREC) && defined(USE_NEW_DYNAREC))
|
|
codegen_flags_changed = 0;
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
static int opPOPF(uint32_t fetchdat)
|
|
{
|
|
uint16_t tempw;
|
|
|
|
if ((cpu_state.eflags & VM_FLAG) && (IOPL < 3))
|
|
{
|
|
if (cr4 & CR4_VME)
|
|
{
|
|
uint32_t old_esp = ESP;
|
|
|
|
tempw = POP_W();
|
|
if (cpu_state.abrt)
|
|
{
|
|
|
|
ESP = old_esp;
|
|
return 1;
|
|
}
|
|
|
|
if ((tempw & T_FLAG) || ((tempw & I_FLAG) && (cpu_state.eflags & VIP_FLAG)))
|
|
{
|
|
ESP = old_esp;
|
|
x86gpf(NULL, 0);
|
|
return 1;
|
|
}
|
|
if (tempw & I_FLAG)
|
|
cpu_state.eflags |= VIF_FLAG;
|
|
else
|
|
cpu_state.eflags &= ~VIF_FLAG;
|
|
cpu_state.flags = (cpu_state.flags & 0x3200) | (tempw & 0x4dd5) | 2;
|
|
}
|
|
else
|
|
{
|
|
x86gpf(NULL, 0);
|
|
return 1;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
tempw = POP_W();
|
|
if (cpu_state.abrt)
|
|
return 1;
|
|
|
|
if (!(CPL) || !(msw & 1))
|
|
cpu_state.flags = (tempw & 0x7fd5) | 2;
|
|
else if (IOPLp)
|
|
cpu_state.flags = (cpu_state.flags & 0x3000) | (tempw & 0x4fd5) | 2;
|
|
else
|
|
cpu_state.flags = (cpu_state.flags & 0x3200) | (tempw & 0x4dd5) | 2;
|
|
}
|
|
flags_extract();
|
|
|
|
CLOCK_CYCLES(5);
|
|
PREFETCH_RUN(5, 1, -1, 1,0,0,0, 0);
|
|
|
|
#if (defined(USE_DYNAREC) && defined(USE_NEW_DYNAREC))
|
|
codegen_flags_changed = 0;
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
static int opPOPFD(uint32_t fetchdat)
|
|
{
|
|
uint32_t templ;
|
|
|
|
if ((cpu_state.eflags & VM_FLAG) && (IOPL < 3))
|
|
{
|
|
x86gpf(NULL, 0);
|
|
return 1;
|
|
}
|
|
|
|
templ = POP_L(); if (cpu_state.abrt) return 1;
|
|
|
|
if (!(CPL) || !(msw & 1)) cpu_state.flags = (templ & 0x7fd5) | 2;
|
|
else if (IOPLp) cpu_state.flags = (cpu_state.flags & 0x3000) | (templ & 0x4fd5) | 2;
|
|
else cpu_state.flags = (cpu_state.flags & 0x3200) | (templ & 0x4dd5) | 2;
|
|
|
|
templ &= (is486 || isibm486) ? 0x3c0000 : 0;
|
|
templ |= ((cpu_state.eflags&3) << 16);
|
|
if (cpu_CR4_mask & CR4_VME) cpu_state.eflags = (templ >> 16) & 0x3f;
|
|
else if (CPUID) cpu_state.eflags = (templ >> 16) & 0x27;
|
|
else if (is486 || isibm486) cpu_state.eflags = (templ >> 16) & 7;
|
|
else cpu_state.eflags = (templ >> 16) & 3;
|
|
|
|
flags_extract();
|
|
|
|
CLOCK_CYCLES(5);
|
|
PREFETCH_RUN(5, 1, -1, 0,1,0,0, 0);
|
|
|
|
#if (defined(USE_DYNAREC) && defined(USE_NEW_DYNAREC))
|
|
codegen_flags_changed = 0;
|
|
#endif
|
|
|
|
return 0;
|
|
}
|