This rewrites the way that CPU alerts work, making them a bitmap (since multiple alerts can happen simultaneously, like SMC and IRQ). This doesn't really fix many games but improves accuracy overall and improves performance on some I/O writes (the ones without side effects). The IRQ raising is now decoupled and explicitely called via a new function (check_and_raise_interrupts) to avoid issues such as invalid CPSR values (doesn't seem to bother most games!). There's more side effects missing, so this just lays the ground for more fixes.
956 lines
44 KiB
ArmAsm
956 lines
44 KiB
ArmAsm
|
|
#include "../gpsp_config.h"
|
|
|
|
#define defsymbl(symbol) \
|
|
.align 2; \
|
|
.type symbol, %function ;\
|
|
.global symbol ; \
|
|
.global _##symbol ; \
|
|
symbol: \
|
|
_##symbol:
|
|
|
|
.text
|
|
.align 2
|
|
|
|
#define REG_R0 (0 * 4)
|
|
#define REG_R1 (1 * 4)
|
|
#define REG_R2 (2 * 4)
|
|
#define REG_R3 (3 * 4)
|
|
#define REG_R4 (4 * 4)
|
|
#define REG_R5 (5 * 4)
|
|
#define REG_R6 (6 * 4)
|
|
#define REG_R7 (7 * 4)
|
|
#define REG_R8 (8 * 4)
|
|
#define REG_R9 (9 * 4)
|
|
#define REG_R10 (10 * 4)
|
|
#define REG_R11 (11 * 4)
|
|
#define REG_R12 (12 * 4)
|
|
#define REG_R13 (13 * 4)
|
|
#define REG_R14 (14 * 4)
|
|
#define REG_SP (13 * 4)
|
|
#define REG_LR (14 * 4)
|
|
#define REG_PC (15 * 4)
|
|
#define REG_CPSR (16 * 4)
|
|
#define CPU_MODE (17 * 4)
|
|
#define CPU_HALT_STATE (18 * 4)
|
|
|
|
#define REG_BUS_VALUE (19 * 4)
|
|
#define REG_N_FLAG (20 * 4)
|
|
#define REG_Z_FLAG (21 * 4)
|
|
#define REG_C_FLAG (22 * 4)
|
|
#define REG_V_FLAG (23 * 4)
|
|
#define CHANGED_PC_STATUS (24 * 4)
|
|
#define COMPLETED_FRAME (25 * 4)
|
|
#define OAM_UPDATED (26 * 4)
|
|
#define REG_SAVE (27 * 4)
|
|
#define REG_SAVE2 (28 * 4)
|
|
#define REG_SAVE3 (29 * 4)
|
|
#define REG_SAVE4 (30 * 4)
|
|
#define REG_SAVE5 (31 * 4)
|
|
|
|
#define CPU_ALERT_HALT (1 << 0)
|
|
#define CPU_ALERT_SMC (1 << 1)
|
|
#define CPU_ALERT_IRQ (1 << 2)
|
|
|
|
#define reg_a0 r0
|
|
#define reg_a1 r1
|
|
#define reg_a2 r2
|
|
|
|
#define reg_base r11
|
|
#define reg_flags r9
|
|
|
|
#define reg_cycles r12
|
|
|
|
#define reg_x0 r3
|
|
#define reg_x1 r4
|
|
#define reg_x2 r5
|
|
#define reg_x3 r6
|
|
#define reg_x4 r7
|
|
#define reg_x5 r8
|
|
|
|
|
|
#define MODE_SUPERVISOR 0x13
|
|
#define SUPERVISOR_OFFSET 0x03
|
|
|
|
@ Memory offsets from reg_base to the different buffers
|
|
#define IWRAM_OFF -0xA8000
|
|
#define VRAM_OFF -0x98000
|
|
#define EWRAM_OFF -0x80000
|
|
#define SPSR_RAM_OFF 0x100
|
|
#define STORE_TBL_OFF 0x118
|
|
#define REGMODE_RAM_OFF 0x400
|
|
#define OAM_RAM_OFF 0x500
|
|
#define PAL_RAM_OFF 0x900
|
|
#define RDMAP_OFF 0xD00
|
|
#define IOREG_OFF 0x8D00
|
|
#define PAL_CONV_OFF 0x9100
|
|
|
|
|
|
#if __ARM_ARCH >= 6
|
|
#define extract_u16(rd, rs) \
|
|
uxth rd, rs
|
|
#else
|
|
#define extract_u16(rd, rs) \
|
|
bic rd, rs, #0xff000000; \
|
|
bic rd, rd, #0x00ff0000
|
|
#endif
|
|
|
|
@ Will load the register set from memory into the appropriate cached registers.
|
|
@ See arm_emit.h for listing explanation.
|
|
|
|
#define load_registers_arm() ;\
|
|
ldr reg_x0, [reg_base, #REG_R0] ;\
|
|
ldr reg_x1, [reg_base, #REG_R1] ;\
|
|
ldr reg_x2, [reg_base, #REG_R6] ;\
|
|
ldr reg_x3, [reg_base, #REG_R9] ;\
|
|
ldr reg_x4, [reg_base, #REG_R12] ;\
|
|
ldr reg_x5, [reg_base, #REG_R14] ;\
|
|
|
|
#define load_registers_thumb() ;\
|
|
ldm reg_base, {reg_x0, reg_x1, reg_x2, reg_x3, reg_x4, reg_x5}
|
|
|
|
|
|
@ Will store the register set from cached registers back to memory.
|
|
|
|
#define store_registers_arm() ;\
|
|
str reg_x0, [reg_base, #REG_R0] ;\
|
|
str reg_x1, [reg_base, #REG_R1] ;\
|
|
str reg_x2, [reg_base, #REG_R6] ;\
|
|
str reg_x3, [reg_base, #REG_R9] ;\
|
|
str reg_x4, [reg_base, #REG_R12] ;\
|
|
str reg_x5, [reg_base, #REG_R14] ;\
|
|
|
|
#define store_registers_thumb() ;\
|
|
stm reg_base, {reg_x0, reg_x1, reg_x2, reg_x3, reg_x4, reg_x5}
|
|
|
|
#define store_registers_cond() ;\
|
|
stmne reg_base, {reg_x0, reg_x1, reg_x2, reg_x3, reg_x4, reg_x5} ;\
|
|
bne 80f ;\
|
|
str reg_x0, [reg_base, #REG_R0] ;\
|
|
str reg_x1, [reg_base, #REG_R1] ;\
|
|
str reg_x2, [reg_base, #REG_R6] ;\
|
|
str reg_x3, [reg_base, #REG_R9] ;\
|
|
str reg_x4, [reg_base, #REG_R12] ;\
|
|
str reg_x5, [reg_base, #REG_R14] ;\
|
|
80:
|
|
|
|
@ Returns an updated persistent cpsr with the cached flags register.
|
|
@ Uses reg as a temporary register and returns the CPSR here.
|
|
|
|
#define collapse_flags_no_update(reg) ;\
|
|
ldr reg, [reg_base, #REG_CPSR] /* reg = cpsr */;\
|
|
bic reg, reg, #0xF0000000 /* clear ALU flags in cpsr */;\
|
|
and reg_flags, reg_flags, #0xF0000000 /* clear non-ALU flags */;\
|
|
orr reg, reg, reg_flags /* update cpsr with ALU flags */;\
|
|
|
|
@ Updates cpsr using the above macro.
|
|
|
|
#define collapse_flags(reg) ;\
|
|
collapse_flags_no_update(reg) ;\
|
|
str reg, [reg_base, #REG_CPSR] ;\
|
|
|
|
@ Loads the saved flags register from the persistent cpsr.
|
|
|
|
#define extract_flags() ;\
|
|
ldr reg_flags, [reg_base, #REG_CPSR] ;\
|
|
msr cpsr_f, reg_flags ;\
|
|
|
|
|
|
#define save_flags() ;\
|
|
mrs reg_flags, cpsr ;\
|
|
|
|
#define restore_flags() ;\
|
|
msr cpsr_f, reg_flags ;\
|
|
|
|
@ Align the stack to 64 bits (ABIs that don't require it, still recommend so)
|
|
#define call_c_saved_regs r2, r3, r12, lr
|
|
|
|
@ Calls a C function - reloads the stack pointer and saves all caller save
|
|
@ registers which are important to the dynarec.
|
|
|
|
#define call_c_function(function) ;\
|
|
stmdb sp!, { call_c_saved_regs } ;\
|
|
bl function ;\
|
|
ldmia sp!, { call_c_saved_regs } ;\
|
|
|
|
|
|
@ Update the GBA hardware (video, sound, input, etc)
|
|
|
|
@ Input:
|
|
@ r0: current PC
|
|
|
|
#define return_straight() bx lr
|
|
#define return_add() add pc, lr, #4
|
|
|
|
#define load_pc_straight() ldr r0, [lr, #-12]
|
|
#define load_pc_add() ldr r0, [lr]
|
|
|
|
#define cycles_straight() mvn r0, reg_cycles
|
|
#define cycles_add() mov r0, #0
|
|
|
|
#define arm_update_gba_builder(name, mode, return_op) ;\
|
|
;\
|
|
defsymbl(arm_update_gba_##name) ;\
|
|
load_pc_##return_op() ;\
|
|
str r0, [reg_base, #REG_PC] /* write out the PC */;\
|
|
;\
|
|
save_flags() ;\
|
|
collapse_flags(r0) /* update the flags */;\
|
|
;\
|
|
store_registers_##mode() /* save out registers */;\
|
|
cycles_##return_op() /* remaining cycles in arg0 */;\
|
|
call_c_function(update_gba) /* update GBA state */;\
|
|
;\
|
|
ldr r1, [reg_base, #COMPLETED_FRAME] /* return if new frame */;\
|
|
cmp r1, #0 ;\
|
|
bne return_to_main ;\
|
|
;\
|
|
mvn reg_cycles, r0 /* load new cycle count */;\
|
|
;\
|
|
ldr r0, [reg_base, #CHANGED_PC_STATUS] /* load PC changed status */;\
|
|
cmp r0, #0 /* see if PC has changed */;\
|
|
bne 1f /* go jump/translate */;\
|
|
;\
|
|
load_registers_##mode() /* reload registers */;\
|
|
restore_flags() ;\
|
|
return_##return_op() /* continue, no PC change */;\
|
|
;\
|
|
1: ;\
|
|
ldr r1, [reg_base, #REG_CPSR] /* r1 = flags */;\
|
|
ldr r0, [reg_base, #REG_PC] /* load new PC */;\
|
|
tst r1, #0x20 /* see if Thumb bit is set */;\
|
|
bne 2f /* if so load Thumb PC */;\
|
|
;\
|
|
load_registers_arm() /* load ARM regs */;\
|
|
call_c_function(block_lookup_address_arm) ;\
|
|
restore_flags() ;\
|
|
bx r0 /* jump to new ARM block */;\
|
|
2: ;\
|
|
load_registers_thumb() /* load Thumb regs */;\
|
|
call_c_function(block_lookup_address_thumb) ;\
|
|
restore_flags() ;\
|
|
bx r0 /* jump to new ARM block */;\
|
|
.size arm_update_gba_##name, .-arm_update_gba_##name
|
|
|
|
arm_update_gba_builder(arm, arm, straight)
|
|
arm_update_gba_builder(thumb, thumb, straight)
|
|
|
|
arm_update_gba_builder(idle_arm, arm, add)
|
|
arm_update_gba_builder(idle_thumb, thumb, add)
|
|
|
|
|
|
@ Cheat hooks for master function
|
|
@ This is called whenever PC == cheats-master-function
|
|
@ Just calls the C function to process cheats
|
|
|
|
#define cheat_hook_builder(mode) ;\
|
|
defsymbl(mode##_cheat_hook) ;\
|
|
save_flags() ;\
|
|
store_registers_##mode() ;\
|
|
call_c_function(process_cheats) ;\
|
|
load_registers_##mode() ;\
|
|
restore_flags() ;\
|
|
bx lr ;\
|
|
|
|
cheat_hook_builder(arm)
|
|
cheat_hook_builder(thumb)
|
|
|
|
|
|
@ These are b stubs for performing indirect branches. They are not
|
|
@ linked to and don't return, instead they link elsewhere.
|
|
|
|
@ Input:
|
|
@ r0: PC to branch to
|
|
|
|
defsymbl(arm_indirect_branch_arm)
|
|
save_flags()
|
|
call_c_function(block_lookup_address_arm)
|
|
restore_flags()
|
|
bx r0
|
|
|
|
defsymbl(arm_indirect_branch_thumb)
|
|
save_flags()
|
|
call_c_function(block_lookup_address_thumb)
|
|
restore_flags()
|
|
bx r0
|
|
|
|
defsymbl(arm_indirect_branch_dual_arm)
|
|
save_flags()
|
|
tst r0, #0x01 @ check lower bit
|
|
bne 1f @ if set going to Thumb mode
|
|
add r0, #2 @ two LSB are cleared after
|
|
call_c_function(block_lookup_address_arm)
|
|
restore_flags()
|
|
bx r0 @ keep executing arm code
|
|
1:
|
|
store_registers_arm() @ save out ARM registers
|
|
ldr r1, [reg_base, #REG_CPSR] @ load cpsr
|
|
load_registers_thumb() @ load in Thumb registers
|
|
orr r1, r1, #0x20 @ set Thumb mode
|
|
str r1, [reg_base, #REG_CPSR] @ store flags
|
|
call_c_function(block_lookup_address_thumb)
|
|
restore_flags()
|
|
bx r0
|
|
.size arm_indirect_branch_dual_arm, .-arm_indirect_branch_dual_arm
|
|
|
|
defsymbl(arm_indirect_branch_dual_thumb)
|
|
save_flags()
|
|
tst r0, #0x01 @ check lower bit
|
|
beq 1f @ if set going to ARM mode
|
|
call_c_function(block_lookup_address_thumb)
|
|
restore_flags()
|
|
bx r0 @ keep executing thumb code
|
|
1:
|
|
store_registers_thumb() @ save out Thumb registers
|
|
ldr r1, [reg_base, #REG_CPSR] @ load cpsr
|
|
load_registers_arm() @ load in ARM registers
|
|
bic r1, r1, #0x20 @ clear Thumb mode
|
|
str r1, [reg_base, #REG_CPSR] @ store flags
|
|
add r0, #2 @ two LSB are cleared after
|
|
call_c_function(block_lookup_address_arm)
|
|
restore_flags()
|
|
bx r0
|
|
.size arm_indirect_branch_dual_thumb, .-arm_indirect_branch_dual_thumb
|
|
|
|
@ Update the cpsr.
|
|
|
|
@ Input:
|
|
@ r0: new cpsr value
|
|
@ [lr]: bitmask (user mode)
|
|
@ [lr+4]: bitmask (privileged mode)
|
|
@ [lr+8]: current PC
|
|
|
|
defsymbl(execute_store_cpsr)
|
|
save_flags()
|
|
ldr r1, [reg_base, #CPU_MODE] @ r1 = cpu_mode
|
|
lsr r1, r1, #4 @ Load privilege bit
|
|
ldr r1, [lr, r1, lsl #2] @ Load mask
|
|
|
|
and reg_flags, r0, r1 @ reg_flags = new_cpsr & store_mask
|
|
ldr r0, [reg_base, #REG_CPSR] @ r0 = cpsr
|
|
bic r0, r0, r1 @ r0 = cpsr & ~store_mask
|
|
orr reg_flags, reg_flags, r0 @ reg_flags = new_cpsr | cpsr
|
|
|
|
mov r0, reg_flags @ also put new cpsr in r0
|
|
|
|
store_registers_arm() @ save ARM registers
|
|
ldr r2, [lr, #8] @ r2 = pc
|
|
call_c_function(execute_store_cpsr_body)
|
|
load_registers_arm() @ restore ARM registers
|
|
|
|
cmp r0, #0 @ check new PC
|
|
bne 1f @ if it's zero, resume
|
|
|
|
restore_flags()
|
|
add pc, lr, #12 @ return (skip data)
|
|
|
|
1:
|
|
call_c_function(block_lookup_address_arm)
|
|
restore_flags()
|
|
bx r0 @ return to PC ARM address
|
|
.size execute_store_cpsr, .-execute_store_cpsr
|
|
|
|
@ Restore the cpsr from the mode spsr and mode shift.
|
|
|
|
@ Input:
|
|
@ r0: current pc
|
|
|
|
defsymbl(execute_spsr_restore)
|
|
save_flags()
|
|
ldr r2, [reg_base, #CPU_MODE] @ r2 = cpu_mode
|
|
ands r2, r2, #0xF @ Ignore privilege bits
|
|
beq 2f @ if user/system mode no side effects
|
|
|
|
add r1, reg_base, #SPSR_RAM_OFF @ r1 = spsr
|
|
ldr r1, [r1, r2, lsl #2] @ r1 = spsr[cpu_mode] (new cpsr)
|
|
str r1, [reg_base, #REG_CPSR] @ update cpsr
|
|
mov reg_flags, r1 @ also, update shadow flags
|
|
|
|
@ This function call will pass r0 (address) and return it.
|
|
store_registers_arm() @ save ARM registers
|
|
call_c_function(execute_spsr_restore_body)
|
|
|
|
ldr r1, [reg_base, #REG_CPSR] @ r1 = cpsr
|
|
tst r1, #0x20 @ see if Thumb mode is set
|
|
bne 1f @ if so handle it
|
|
|
|
load_registers_arm() @ restore ARM registers
|
|
2:
|
|
call_c_function(block_lookup_address_arm)
|
|
restore_flags()
|
|
bx r0
|
|
|
|
1:
|
|
load_registers_thumb() @ load Thumb registers
|
|
call_c_function(block_lookup_address_thumb)
|
|
restore_flags()
|
|
bx r0
|
|
|
|
|
|
@ Setup the mode transition work for calling an SWI.
|
|
|
|
@ Input:
|
|
@ r0: current pc
|
|
|
|
#define execute_swi_builder(mode) ;\
|
|
;\
|
|
defsymbl(execute_swi_##mode) ;\
|
|
save_flags() ;\
|
|
add r1, reg_base, #REGMODE_RAM_OFF /* r1 = reg_mode */;\
|
|
/* REG_MODE(MODE_SUPERVISOR)[6] = pc */;\
|
|
ldr r0, [lr] /* load PC */;\
|
|
str r0, [r1, #((SUPERVISOR_OFFSET * (7 * 4)) + (6 * 4))] ;\
|
|
collapse_flags_no_update(r0) /* r0 = cpsr */;\
|
|
add r1, reg_base, #SPSR_RAM_OFF /* r1 = spsr */;\
|
|
str r0, [r1, #(SUPERVISOR_OFFSET * 4)] /* spsr[MODE_SUPERVISOR] = cpsr */;\
|
|
bic r0, r0, #0x3F /* clear mode flag in r0 */;\
|
|
orr r0, r0, #(0x13 | 0x80) /* supervisor mode + disable IRQ */;\
|
|
str r0, [reg_base, #REG_CPSR] /* update cpsr */;\
|
|
;\
|
|
mov r0, #MODE_SUPERVISOR ;\
|
|
;\
|
|
store_registers_##mode() /* store regs for mode */;\
|
|
call_c_function(set_cpu_mode) /* set the CPU mode to svsr */;\
|
|
load_registers_arm() /* load ARM regs */;\
|
|
ldr r0, =0xe3a02004 /* Update open BUS value */;\
|
|
str r0, [reg_base, #REG_BUS_VALUE] ;\
|
|
;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 /* return */;\
|
|
|
|
execute_swi_builder(arm)
|
|
execute_swi_builder(thumb)
|
|
|
|
|
|
@ Start program execution. Normally the mode should be Thumb and the
|
|
@ PC should be 0x8000000, however if a save state is preloaded this
|
|
@ will be different.
|
|
|
|
@ Input:
|
|
@ r0: initial value for cycle counter
|
|
|
|
@ Uses sp as reg_base; must hold consistently true.
|
|
|
|
defsymbl(execute_arm_translate_internal)
|
|
|
|
@ save the registers to be able to return later
|
|
stmdb sp!, { r4, r5, r6, r7, r8, r9, r10, r11, r12, lr }
|
|
|
|
mov reg_base, r1 @ init base_reg
|
|
|
|
mvn reg_cycles, r0 @ load cycle counter
|
|
|
|
@ Check whether the CPU is sleeping already, we should just wait for IRQs
|
|
ldr r1, [reg_base, #CPU_HALT_STATE]
|
|
cmp r1, #0
|
|
bne alert_loop
|
|
b lookup_pc
|
|
|
|
|
|
@ Epilogue to return to the main thread (whatever called execute_arm_translate)
|
|
|
|
return_to_main:
|
|
@ restore the saved regs and return
|
|
ldmia sp!, { r4, r5, r6, r7, r8, r9, r10, r11, r12, lr }
|
|
bx lr
|
|
|
|
@ Prepares stores for external write function (align + zero extend value)
|
|
#define store_align_8() and r1, r1, #0xff
|
|
#define store_align_16() bic r0, r0, #0x01; extract_u16(r1, r1)
|
|
#define store_align_32() bic r0, r0, #0x03
|
|
|
|
#define mask_addr_8(nbits) ;\
|
|
mov r0, r0, lsl #(32 - nbits) /* isolate bottom n bits in top */;\
|
|
mov r0, r0, lsr #(32 - nbits) /* high bits are now clear */;\
|
|
|
|
#define mask_addr_16(nbits) ;\
|
|
mov r0, r0, lsl #(32 - nbits) /* isolate bottom n bits in top */;\
|
|
mov r0, r0, lsr #(32 - nbits + 1) /* high bits are now clear */;\
|
|
mov r0, r0, lsl #1 /* LSB is also zero */;\
|
|
|
|
#define mask_addr_32(nbits) ;\
|
|
mov r0, r0, lsl #(32 - nbits) /* isolate bottom n bits in top */;\
|
|
mov r0, r0, lsr #(32 - nbits + 2) /* high bits are now clear */;\
|
|
mov r0, r0, lsl #2 /* 2 LSB are also zero */;\
|
|
|
|
@ Vram, OAM and palette memories can only be accessed at a 16 bit boundary
|
|
#define mask_addr_bus16_32(nbits) mask_addr_32(nbits)
|
|
#define mask_addr_bus16_16(nbits) mask_addr_16(nbits)
|
|
#define mask_addr_bus16_8(nbits) mask_addr_16(nbits)
|
|
|
|
@ Write out to memory.
|
|
|
|
@ Input:
|
|
@ r0: address
|
|
@ r1: value
|
|
@
|
|
@ The instruction at LR is not an inst but a u32 data that contains the PC
|
|
@ Used for SMC. That's why return is essentially `pc = lr + 4`
|
|
|
|
#define execute_store_builder(store_type, str_op, str_op16, load_op) ;\
|
|
;\
|
|
ext_store_u##store_type: ;\
|
|
save_flags() ;\
|
|
ldr r2, [lr] /* load PC */;\
|
|
str r2, [reg_base, #REG_PC] /* write out PC */;\
|
|
store_align_##store_type() ;\
|
|
call_c_function(write_memory##store_type) ;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 /* return */;\
|
|
;\
|
|
ext_io_store_u##store_type: ;\
|
|
save_flags() ;\
|
|
ldr r2, [lr] /* load PC */;\
|
|
str r2, [reg_base, #REG_PC] /* write out PC */;\
|
|
;\
|
|
mask_addr_##store_type(10) /* Mask to IO memory (+align) */;\
|
|
call_c_function(write_io_register##store_type) ;\
|
|
;\
|
|
cmp r0, #0 ;\
|
|
bne write_epilogue /* handle additional write stuff */;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 /* return */;\
|
|
;\
|
|
ext_store_iwram_u##store_type: ;\
|
|
save_flags() ;\
|
|
mask_addr_##store_type(15) /* Mask to mirror memory (+align)*/;\
|
|
add r2, reg_base, #(IWRAM_OFF+0x8000) /* r2 = iwram base */;\
|
|
str_op r1, [r0, r2] /* store data */;\
|
|
sub r2, r2, #0x8000 /* r2 = iwram smc base */;\
|
|
load_op r1, [r0, r2] /* r1 = SMC sentinel */;\
|
|
cmp r1, #0 /* Check value, should be zero */;\
|
|
bne 3f /* if so perform smc write */;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 /* return */;\
|
|
;\
|
|
ext_store_ewram_u##store_type: ;\
|
|
save_flags() ;\
|
|
mask_addr_##store_type(18) /* Mask to mirror memory (+align)*/;\
|
|
add r2, reg_base, #EWRAM_OFF /* r2 = ewram base */;\
|
|
str_op r1, [r0, r2] /* store data */;\
|
|
add r2, r2, #0x40000 /* r2 = ewram smc base */;\
|
|
load_op r1, [r0, r2] /* r1 = SMC sentinel */;\
|
|
cmp r1, #0 /* Check value, should be zero */;\
|
|
bne 3f /* if so perform smc write */;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 /* return */;\
|
|
;\
|
|
ext_store_vram_u##store_type: ;\
|
|
save_flags() ;\
|
|
mask_addr_bus16_##store_type(17) /* Mask to mirror memory (+align)*/;\
|
|
cmp r0, #0x18000 /* Check if exceeds 96KB */;\
|
|
subcs r0, r0, #0x8000 /* Mirror to the last bank */;\
|
|
add r2, reg_base, #VRAM_OFF /* r2 = vram base */;\
|
|
restore_flags() ;\
|
|
str_op16 r1, [r0, r2] /* store data */;\
|
|
add pc, lr, #4 /* return */;\
|
|
;\
|
|
ext_store_oam_ram_u##store_type: ;\
|
|
mask_addr_bus16_##store_type(10) /* Mask to mirror memory (+align)*/;\
|
|
add r2, reg_base, #OAM_RAM_OFF /* r2 = oam ram base */;\
|
|
str_op16 r1, [r0, r2] /* store data */;\
|
|
str r2, [reg_base, #OAM_UPDATED] /* write non zero to signal */;\
|
|
add pc, lr, #4 /* return */;\
|
|
;\
|
|
3: /* Flush RAM cache and "resume" execution via re-compile */ ;\
|
|
ldr r0, [lr] /* load PC */;\
|
|
str r0, [reg_base, #REG_PC] /* write out PC */;\
|
|
call_c_function(flush_translation_cache_ram) ;\
|
|
b resume_pc /* continue execution */;\
|
|
|
|
@ for ignored areas, just return
|
|
ext_store_ignore:
|
|
add pc, lr, #4 @ return
|
|
|
|
#define store_lookup_table(store_type) ;\
|
|
.word ext_store_ignore /* -1: ignore, for > 0x0F */;\
|
|
.word ext_store_ignore /* 0x00: BIOS, ignore */;\
|
|
.word ext_store_ignore /* 0x01: ignore */;\
|
|
.word ext_store_ewram_u##store_type /* 0x02: ewram */;\
|
|
.word ext_store_iwram_u##store_type /* 0x03: iwram */;\
|
|
.word ext_io_store_u##store_type /* 0x04: I/O regs */;\
|
|
.word ext_store_palette_u##store_type /* 0x05: palette RAM */;\
|
|
.word ext_store_vram_u##store_type /* 0x06: vram */;\
|
|
.word ext_store_oam_ram_u##store_type /* 0x07: oam ram */;\
|
|
.word ext_store_u##store_type /* 0x08: gamepak: ignore */;\
|
|
.word ext_store_ignore /* 0x09: gamepak: ignore */;\
|
|
.word ext_store_ignore /* 0x0A: gamepak: ignore */;\
|
|
.word ext_store_ignore /* 0x0B: gamepak: ignore */;\
|
|
.word ext_store_ignore /* 0x0C: gamepak: ignore */;\
|
|
.word ext_store_u##store_type /* 0x0D: EEPROM */;\
|
|
.word ext_store_u##store_type /* 0x0E: backup */;\
|
|
.word ext_store_ignore /* 0x0F: ignore */;\
|
|
|
|
execute_store_builder(8, strb, strh, ldrb)
|
|
execute_store_builder(16, strh, strh, ldrh)
|
|
execute_store_builder(32, str, str, ldr )
|
|
|
|
@ Palette writes are special since they are converted on the fly for speed
|
|
|
|
ext_store_palette_u8:
|
|
bic r1, r1, #0xff00 // Duplicate the byte
|
|
orr r1, r1, lsl #8
|
|
ext_store_palette_u16:
|
|
mask_addr_16(10) // Accesses are always 16 bit
|
|
add r2, reg_base, #PAL_RAM_OFF // r2 = palette base
|
|
strh r1, [r0, r2] // store data
|
|
|
|
and r2, r1, #0x3E0 // Convert color point
|
|
lsl r2, r2, #1
|
|
orr r2, r1, lsl #11
|
|
and r1, r1, #0x7C00
|
|
orr r2, r1, lsr #10
|
|
|
|
add r1, reg_base, #PAL_CONV_OFF // r1 = converted palette ram
|
|
strh r2, [r0, r1] // Converted value write (r2)
|
|
add pc, lr, #4 // return
|
|
|
|
ext_store_palette_u32_safe:
|
|
sub lr, lr, #4
|
|
ext_store_palette_u32:
|
|
mask_addr_32(10) // Accesses are always 16 bit
|
|
add r2, reg_base, #PAL_RAM_OFF // r2 = palette base
|
|
str r1, [r0, r2] // store data
|
|
add r2, reg_base, #PAL_CONV_OFF // r2 = converted palette ram
|
|
|
|
lsr r9, r1, #10
|
|
and r9, r9, #0x1F
|
|
orr r9, r1, lsl #11
|
|
bic r1, r1, #0x1F
|
|
bic r1, r1, #0xFC00
|
|
orr r9, r1, lsl #1
|
|
strh r9, [r0, r2] // Write first halfword
|
|
add r0, r0, #2
|
|
|
|
lsr r1, r1, #16
|
|
and r9, r1, #0x3E0
|
|
lsl r9, r9, #1
|
|
orr r9, r1, lsl #11
|
|
and r1, r1, #0x7C00
|
|
orr r9, r1, lsr #10
|
|
strh r9, [r0, r2] // Write second halfword
|
|
|
|
add pc, lr, #4 // return
|
|
|
|
|
|
@ This is a store that is executed in a strm case (so no SMC checks in-between)
|
|
|
|
ext_store_u32_safe:
|
|
ext_io_store_u32_safe:
|
|
save_flags()
|
|
call_c_function(write_memory32) @ Perform 32bit store
|
|
restore_flags()
|
|
bx lr @ Return
|
|
|
|
ext_store_iwram_u32_safe:
|
|
mask_addr_8(15) @ Mask to mirror memory (no need to align!)
|
|
add r2, reg_base, #(IWRAM_OFF+0x8000) @ r2 = iwram base
|
|
str r1, [r0, r2] @ store data
|
|
bx lr @ Return
|
|
|
|
ext_store_ewram_u32_safe:
|
|
mask_addr_8(18) @ Mask to mirror memory (no need to align!)
|
|
add r2, reg_base, #EWRAM_OFF @ r2 = ewram base
|
|
str r1, [r0, r2] @ store data
|
|
bx lr @ Return
|
|
|
|
ext_store_vram_u32_safe:
|
|
mask_addr_8(17) @ Mask to mirror memory (no need to align!)
|
|
save_flags()
|
|
add r2, reg_base, #VRAM_OFF @ r2 = vram base
|
|
cmp r0, #0x18000 @ Check if exceeds 96KB
|
|
subcs r0, r0, #0x8000 @ Mirror to the last bank
|
|
str r1, [r0, r2] @ store data
|
|
restore_flags()
|
|
bx lr @ Return
|
|
|
|
ext_store_oam_ram_u32_safe:
|
|
mask_addr_8(10) @ Mask to mirror memory (no need to align!)
|
|
add r2, reg_base, #OAM_RAM_OFF @ r2 = oam ram base
|
|
str r1, [r0, r2] @ store data
|
|
str r2, [reg_base, #OAM_UPDATED] @ store anything non zero here
|
|
bx lr @ Return
|
|
|
|
|
|
write_epilogue:
|
|
ldr r2, [reg_base, #REG_CPSR] @ Save all register and CPSR
|
|
tst r2, #0x20 @ Check thumb bit
|
|
store_registers_cond() @ Store ARM/Thumb regs
|
|
collapse_flags(r1) @ interrupt needs current flags
|
|
|
|
mov r2, r0 @ r2 is stored across C calls
|
|
tst r2, #CPU_ALERT_SMC @ check for SMC code
|
|
beq 1f
|
|
call_c_function(flush_translation_cache_ram) @ Flush RAM if bit is set
|
|
|
|
1:
|
|
tst r2, #CPU_ALERT_IRQ @ check for IRQs
|
|
beq 2f
|
|
call_c_function(check_and_raise_interrupts) @ Update CPU state to raise IRQ
|
|
|
|
2:
|
|
tst r2, #CPU_ALERT_HALT @ check for CPU halt bit
|
|
beq lookup_pc @ Resume execution if not halted
|
|
|
|
@ Fallthrough to alert_loop on purpose (CPU is now halted)
|
|
|
|
mvn r0, reg_cycles @ setup for update_gba
|
|
|
|
alert_loop:
|
|
call_c_function(update_gba) @ update GBA until CPU isn't halted
|
|
|
|
ldr r1, [reg_base, #COMPLETED_FRAME] @ Check whether a frame was completed
|
|
cmp r1, #0
|
|
bne return_to_main
|
|
|
|
ldr r1, [reg_base, #CPU_HALT_STATE] @ Check whether the CPU is halted
|
|
cmp r1, #0
|
|
bne alert_loop @ Keep looping until it is
|
|
|
|
mvn reg_cycles, r0 @ load new cycle count
|
|
ldr r0, [reg_base, #REG_PC] @ load new PC
|
|
ldr r1, [reg_base, #REG_CPSR] @ r1 = flags
|
|
tst r1, #0x20 @ see if Thumb bit is set
|
|
bne 2f
|
|
|
|
call_c_function(block_lookup_address_arm)
|
|
load_registers_arm()
|
|
extract_flags()
|
|
bx r0 @ jump to new ARM block
|
|
|
|
2:
|
|
call_c_function(block_lookup_address_thumb)
|
|
load_registers_thumb()
|
|
extract_flags()
|
|
bx r0 @ jump to new Thumb block
|
|
|
|
|
|
resume_pc:
|
|
@ Resume regular execution (except we might need to recompile due to flush)
|
|
@ assume flags are spilled to reg_flags
|
|
ldr r0, [reg_base, #REG_PC] @ r0 = new pc
|
|
ldr r1, [reg_base, #REG_CPSR] @ r1 = partial flags valid
|
|
tst r1, #0x20 @ see if Thumb bit is set
|
|
beq 1f @ if not lookup ARM
|
|
|
|
call_c_function(block_lookup_address_thumb)
|
|
restore_flags()
|
|
bx r0 @ jump to new Thumb block
|
|
1:
|
|
call_c_function(block_lookup_address_arm)
|
|
restore_flags()
|
|
bx r0 @ jump to new ARM block
|
|
|
|
lookup_pc:
|
|
@ Restart CPU execution, assumes CPU mode might have changed
|
|
ldr r0, [reg_base, #REG_PC] @ r0 = new pc
|
|
ldr r1, [reg_base, #REG_CPSR] @ r1 = flags
|
|
tst r1, #0x20 @ see if Thumb bit is set
|
|
beq 2f @ if not lookup ARM
|
|
|
|
call_c_function(block_lookup_address_thumb)
|
|
load_registers_thumb()
|
|
extract_flags()
|
|
bx r0 @ jump to new Thumb block
|
|
2:
|
|
call_c_function(block_lookup_address_arm)
|
|
load_registers_arm()
|
|
extract_flags()
|
|
bx r0 @ jump to new ARM block
|
|
|
|
|
|
#define exec_ld_op_s8(mirrorbits) ;\
|
|
mov r0, r0, lsl #(32 - mirrorbits) ;\
|
|
mov r0, r0, lsr #(32 - mirrorbits) ;\
|
|
ldrsb r0, [r2, r0]
|
|
|
|
#define exec_ld_op_u8(mirrorbits) ;\
|
|
mov r0, r0, lsl #(32 - mirrorbits) ;\
|
|
mov r0, r0, lsr #(32 - mirrorbits) ;\
|
|
ldrb r0, [r2, r0]
|
|
|
|
#define exec_ld_op_s16(mirrorbits) ;\
|
|
mov r0, r0, lsl #(32 - mirrorbits) ;\
|
|
mov r0, r0, lsr #(32 - mirrorbits) ;\
|
|
ldrsh r0, [r2, r0]
|
|
|
|
#define exec_ld_op_u16(mirrorbits) ;\
|
|
mov r0, r0, lsl #(32 - mirrorbits) ;\
|
|
mov r0, r0, lsr #(32 - mirrorbits) ;\
|
|
ldrh r0, [r2, r0]
|
|
|
|
#define exec_ld_op_u32(mirrorbits) ;\
|
|
mov r0, r0, lsl #(32 - mirrorbits) ;\
|
|
ldr r0, [r2, r0, lsr #(32 - mirrorbits)] ;\
|
|
|
|
|
|
#define execute_load_builder(load_type, albits, load_function) ;\
|
|
;\
|
|
ld_bios_##load_type: /* BIOS area, need to verify PC */;\
|
|
save_flags() ;\
|
|
ldr r1, [lr] /* r1 = PC */;\
|
|
mov r2, r1, lsr #15 /* r2 = High addr bits from PC */;\
|
|
cmp r2, #0 ;\
|
|
bne 10f /* Jump to slow handler */;\
|
|
ldr r2, [reg_base, #RDMAP_OFF] /* r2 = read_mem[0] */;\
|
|
exec_ld_op_##load_type(15) /* Clear upper bits (15 LSB) */;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
ld_ewram_##load_type: /* EWRAM area */;\
|
|
add r2, reg_base, #EWRAM_OFF ;\
|
|
exec_ld_op_##load_type(18) /* Clear upper bits (18 LSB) */;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
ld_iwram_##load_type: /* IWRAM area */;\
|
|
add r2, reg_base, #(IWRAM_OFF+0x8000) ;\
|
|
exec_ld_op_##load_type(15) /* Clear upper bits (15 LSB) */;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
ld_ioram_##load_type: /* I/O RAM area */;\
|
|
add r2, reg_base, #IOREG_OFF ;\
|
|
exec_ld_op_##load_type(10) /* Clear upper bits (10 LSB) */;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
ld_palram_##load_type: /* Palette RAM area */;\
|
|
add r2, reg_base, #PAL_RAM_OFF ;\
|
|
exec_ld_op_##load_type(10) /* Clear upper bits (10 LSB) */;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
ld_oamram_##load_type: /* OAM RAM area */;\
|
|
add r2, reg_base, #OAM_RAM_OFF ;\
|
|
exec_ld_op_##load_type(10) /* Clear upper bits (10 LSB) */;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
/* ROM area: uses generic memory handlers to handle swapping */ ;\
|
|
ld_rdmap_slow_##load_type: ;\
|
|
save_flags() ;\
|
|
add r2, reg_base, #RDMAP_OFF /* r2 = memory_map_read */;\
|
|
mov r1, r0, lsr #15 /* r1 = page index of address */;\
|
|
ldr r2, [r2, r1, lsl #2] /* r2 = base addr */;\
|
|
cmp r2, #0 ;\
|
|
beq 9f /* Page miss, go slow */;\
|
|
;\
|
|
exec_ld_op_##load_type(15) /* Pages are 32KB big */;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
/* ROM/VRAM area: uses generic memory handlers, assumes is mapped */ ;\
|
|
ld_rdmap_##load_type: ;\
|
|
add r2, reg_base, #RDMAP_OFF /* r2 = memory_map_read */;\
|
|
mov r1, r0, lsr #15 /* r1 = page index of address */;\
|
|
ldr r2, [r2, r1, lsl #2] /* r2 = base addr */;\
|
|
;\
|
|
exec_ld_op_##load_type(15) /* Pages are 32KB big */;\
|
|
add pc, lr, #4 ;\
|
|
;\
|
|
/* Slow load path, for open/unmapped loads */;\
|
|
ld_slow_##load_type: ;\
|
|
save_flags() ;\
|
|
9: ;\
|
|
ldr r1, [lr] /* r1 = PC */;\
|
|
10: ;\
|
|
str r1, [reg_base, #REG_PC] /* update PC */;\
|
|
call_c_function(load_function) ;\
|
|
restore_flags() ;\
|
|
add pc, lr, #4 /* return */;\
|
|
|
|
|
|
#define load_table_gen(load_type, rdmapfn) ;\
|
|
.long ld_slow_##load_type /* -1 (for regions above F) */;\
|
|
.long ld_bios_##load_type /* 0 BIOS */;\
|
|
.long ld_slow_##load_type /* 1 Bad region */;\
|
|
.long ld_ewram_##load_type /* 2 EWRAM */;\
|
|
.long ld_iwram_##load_type /* 3 IWRAM */;\
|
|
.long ld_ioram_##load_type /* 4 I/O */;\
|
|
.long ld_palram_##load_type /* 5 Palette RAM, via map */;\
|
|
.long ld_rdmap_##load_type /* 6 VRAM area */;\
|
|
.long ld_oamram_##load_type /* 7 OAM RAM */;\
|
|
.long ld_##rdmapfn##_##load_type /* 8 ROM, via map */;\
|
|
.long ld_##rdmapfn##_##load_type /* 9 ROM, via map */;\
|
|
.long ld_##rdmapfn##_##load_type /* A ROM, via map */;\
|
|
.long ld_##rdmapfn##_##load_type /* B ROM, via map */;\
|
|
.long ld_##rdmapfn##_##load_type /* C ROM, via map */;\
|
|
.long ld_slow_##load_type /* D ROM or EEPROM/FLASH */;\
|
|
.long ld_slow_##load_type /* E EEPROM/FLASH */;\
|
|
.long ld_slow_##load_type /* F Bad region */;\
|
|
|
|
.pool
|
|
|
|
execute_load_builder(u8, 0, read_memory8 )
|
|
execute_load_builder(s8, 0, read_memory8s )
|
|
execute_load_builder(u16, 1, read_memory16 )
|
|
execute_load_builder(s16, 1, read_memory16s)
|
|
execute_load_builder(u32, 2, read_memory32 )
|
|
|
|
.data
|
|
.align 4
|
|
|
|
defsymbl(st_handler_functions)
|
|
store_lookup_table(8)
|
|
store_lookup_table(16)
|
|
store_lookup_table(32)
|
|
store_lookup_table(32_safe)
|
|
defsymbl(ld_handler_functions)
|
|
load_table_gen(u8, rdmap)
|
|
load_table_gen(s8, rdmap)
|
|
load_table_gen(u16, rdmap)
|
|
load_table_gen(s16, rdmap)
|
|
load_table_gen(u32, rdmap)
|
|
defsymbl(ld_swap_handler_functions)
|
|
load_table_gen(u8, rdmap_slow)
|
|
load_table_gen(s8, rdmap_slow)
|
|
load_table_gen(u16, rdmap_slow)
|
|
load_table_gen(s16, rdmap_slow)
|
|
load_table_gen(u32, rdmap_slow)
|
|
|
|
.bss
|
|
.align 4
|
|
|
|
defsymbl(iwram)
|
|
.space 0x10000
|
|
defsymbl(vram)
|
|
.space 0x18000
|
|
defsymbl(ewram)
|
|
.space 0x80000
|
|
defsymbl(reg)
|
|
.space 0x100
|
|
defsymbl(spsr)
|
|
.space 24
|
|
@ Place lookup tables here for easy access via base_reg too
|
|
defsymbl(st_lookup_tables)
|
|
.space 4*17*4 @ store
|
|
defsymbl(ld_lookup_tables)
|
|
.space 5*17*4 @ loads
|
|
.space 132 @ Padding for alignment
|
|
defsymbl(reg_mode)
|
|
.space 196
|
|
.space 60 @ More padding!
|
|
defsymbl(oam_ram)
|
|
.space 0x400
|
|
defsymbl(palette_ram)
|
|
.space 0x400
|
|
defsymbl(memory_map_read)
|
|
.space 0x8000
|
|
defsymbl(io_registers)
|
|
.space 0x400
|
|
defsymbl(palette_ram_converted)
|
|
.space 0x400
|
|
|
|
@ Vita and 3DS (and of course mmap) map their own cache sections through some
|
|
@ platform-speficic mechanisms.
|
|
#if !defined(MMAP_JIT_CACHE) && !defined(VITA) && !defined(_3DS)
|
|
|
|
@ Make this section executable!
|
|
.text
|
|
.section .jit,"awx",%nobits
|
|
.align 4
|
|
defsymbl(rom_translation_cache)
|
|
.space ROM_TRANSLATION_CACHE_SIZE
|
|
.size rom_translation_cache, .-rom_translation_cache
|
|
defsymbl(ram_translation_cache)
|
|
.space RAM_TRANSLATION_CACHE_SIZE
|
|
.size ram_translation_cache, .-ram_translation_cache
|
|
|
|
#endif
|
|
|