#include "assem_arm64.h"
#include "linkage_offsets.h"
+#if (LO_mem_wtab & 7)
+#error misligned pointers
+#endif
+
.bss
.align 4
.global dynarec_local
DRC_VAR(last_count, 4)
DRC_VAR(pending_exception, 4)
DRC_VAR(stop, 4)
-DRC_VAR(invc_ptr, 4)
+DRC_VAR(branch_target, 4)
DRC_VAR(address, 4)
+DRC_VAR(hack_addr, 4)
DRC_VAR(psxRegs, LO_psxRegs_end - LO_psxRegs)
/* psxRegs */
-DRC_VAR(reg, 128)
+#DRC_VAR(reg, 128)
DRC_VAR(lo, 4)
DRC_VAR(hi, 4)
DRC_VAR(reg_cop0, 128)
#DRC_VAR(intCycle, 256)
DRC_VAR(rcnts, 7*4*4)
-DRC_VAR(mem_rtab, 4)
-DRC_VAR(mem_wtab, 4)
-DRC_VAR(psxH_ptr, 4)
-DRC_VAR(zeromem_ptr, 4)
DRC_VAR(inv_code_start, 4)
DRC_VAR(inv_code_end, 4)
-DRC_VAR(branch_target, 4)
-DRC_VAR(scratch_buf_ptr, 4)
-#DRC_VAR(align0, 12) /* unused/alignment */
+DRC_VAR(mem_rtab, 8)
+DRC_VAR(mem_wtab, 8)
+DRC_VAR(psxH_ptr, 8)
+DRC_VAR(invc_ptr, 8)
+DRC_VAR(zeromem_ptr, 8)
+DRC_VAR(scratch_buf_ptr, 8)
+DRC_VAR(ram_offset, 8)
DRC_VAR(mini_ht, 256)
DRC_VAR(restore_candidate, 512)
/* r0 = virtual target address */
/* r1 = instruction to patch */
.macro dyna_linker_main
- /* XXX: should be able to do better than this... */
+ /* XXX TODO: should be able to do better than this... */
bl get_addr_ht
br x0
.endm
dyna_linker_main
.size dyna_linker_ds, .-dyna_linker_ds
- .align 2
-
-FUNCTION(jump_vaddr):
- bl abort
- .size jump_vaddr, .-jump_vaddr
-
- .align 2
-
-FUNCTION(verify_code_ds):
- bl abort
-FUNCTION(verify_code_vm):
-FUNCTION(verify_code):
- /* r1 = source */
- /* r2 = target */
- /* r3 = length */
- bl abort
- .size verify_code, .-verify_code
- .size verify_code_vm, .-verify_code_vm
-
.align 2
FUNCTION(cc_interrupt):
- bl abort
+ ldr w0, [rFP, #LO_last_count]
+ mov w2, #0x1fc
+ add rCC, w0, rCC
+ str wzr, [rFP, #LO_pending_exception]
+ and w2, w2, rCC, lsr #17
+ add x3, rFP, #LO_restore_candidate
+ str rCC, [rFP, #LO_cycle] /* PCSX cycles */
+# str rCC, [rFP, #LO_reg_cop0+36] /* Count */
+ ldr w19, [x3, w2, uxtw]
+ mov x21, lr
+ cbnz w19, 4f
+1:
+ bl gen_interupt
+ mov lr, x21
+ ldr rCC, [rFP, #LO_cycle]
+ ldr w0, [rFP, #LO_next_interupt]
+ ldr w1, [rFP, #LO_pending_exception]
+ ldr w2, [rFP, #LO_stop]
+ str w0, [rFP, #LO_last_count]
+ sub rCC, rCC, w0
+ cbnz w2, new_dyna_leave
+ cbnz w1, 2f
+ ret
+2:
+ ldr w0, [rFP, #LO_pcaddr]
+ bl get_addr_ht
+ br x0
+4:
+ /* Move 'dirty' blocks to the 'clean' list */
+ lsl w20, w2, #3
+ str wzr, [x3, w2, uxtw]
+5:
+ mov w0, w20
+ add w20, w20, #1
+ tbz w19, #0, 6f
+ bl clean_blocks
+6:
+ lsr w19, w19, #1
+ tst w20, #31
+ bne 5b
+ b 1b
.size cc_interrupt, .-cc_interrupt
- .align 2
-FUNCTION(do_interrupt):
- bl abort
- .size do_interrupt, .-do_interrupt
-
.align 2
FUNCTION(fp_exception):
mov w2, #0x10000000
0:
- ldr w1, [fp, #LO_reg_cop0+48] /* Status */
+ ldr w1, [rFP, #LO_reg_cop0+48] /* Status */
mov w3, #0x80000000
- str w0, [fp, #LO_reg_cop0+56] /* EPC */
+ str w0, [rFP, #LO_reg_cop0+56] /* EPC */
orr w1, w1, #2
add w2, w2, #0x2c
- str w1, [fp, #LO_reg_cop0+48] /* Status */
- str w2, [fp, #LO_reg_cop0+52] /* Cause */
+ str w1, [rFP, #LO_reg_cop0+48] /* Status */
+ str w2, [rFP, #LO_reg_cop0+52] /* Cause */
add w0, w3, #0x80
bl get_addr_ht
br x0
.size fp_exception_ds, .-fp_exception_ds
.align 2
+FUNCTION(jump_break_ds):
+ mov w0, #0x24
+ mov w1, #1
+ b call_psxException
+FUNCTION(jump_break):
+ mov w0, #0x24
+ mov w1, #0
+ b call_psxException
+FUNCTION(jump_syscall_ds):
+ mov w0, #0x20
+ mov w1, #1
+ b call_psxException
FUNCTION(jump_syscall):
- ldr w1, [fp, #LO_reg_cop0+48] /* Status */
- mov w3, #0x80000000
- str w0, [fp, #LO_reg_cop0+56] /* EPC */
- orr w1, w1, #2
- mov w2, #0x20
- str w1, [fp, #LO_reg_cop0+48] /* Status */
- str w2, [fp, #LO_reg_cop0+52] /* Cause */
- add w0, w3, #0x80
- bl get_addr_ht
- br x0
- .size jump_syscall, .-jump_syscall
- .align 2
+ mov w0, #0x20
+ mov w1, #0
- .align 2
-FUNCTION(jump_syscall_hle):
- bl abort
+call_psxException:
+ ldr w3, [rFP, #LO_last_count]
+ str w2, [rFP, #LO_pcaddr]
+ add rCC, w3, rCC
+ str rCC, [rFP, #LO_cycle] /* PCSX cycles */
+ bl psxException
/* note: psxException might do recursive recompiler call from it's HLE code,
* so be ready for this */
-pcsx_return:
- bl abort // w10
- ldr w1, [fp, #LO_next_interupt]
- ldr w10, [fp, #LO_cycle]
- ldr w0, [fp, #LO_pcaddr]
- sub w10, w10, w1
- str w1, [fp, #LO_last_count]
+FUNCTION(jump_to_new_pc):
+ ldr w1, [rFP, #LO_next_interupt]
+ ldr rCC, [rFP, #LO_cycle]
+ ldr w0, [rFP, #LO_pcaddr]
+ sub rCC, rCC, w1
+ str w1, [rFP, #LO_last_count]
bl get_addr_ht
br x0
- .size jump_syscall_hle, .-jump_syscall_hle
-
- .align 2
-FUNCTION(jump_hlecall):
- bl abort
- .size jump_hlecall, .-jump_hlecall
-
- .align 2
-FUNCTION(jump_intcall):
- bl abort
- .size jump_intcall, .-jump_intcall
+ .size jump_to_new_pc, .-jump_to_new_pc
+ /* stack must be aligned by 16, and include space for save_regs() use */
.align 2
FUNCTION(new_dyna_start):
- stp x29, x30, [sp, #-96]! // must be aligned by 16
+ stp x29, x30, [sp, #-SSP_ALL]!
ldr w1, [x0, #LO_next_interupt]
ldr w2, [x0, #LO_cycle]
stp x19, x20, [sp, #16*1]
ldp x23, x24, [sp, #16*3]
ldp x25, x26, [sp, #16*4]
ldp x27, x28, [sp, #16*5]
- ldp x29, x30, [sp], #96
+ ldp x29, x30, [sp], #SSP_ALL
ret
.size new_dyna_leave, .-new_dyna_leave
.align 2
+.macro memhandler_pre
+ /* w0 = adddr/data, x1 = rhandler, w2 = cycles, x3 = whandler */
+ ldr w4, [rFP, #LO_last_count]
+ add w4, w4, w2
+ str w4, [rFP, #LO_cycle]
+.endm
+
+.macro memhandler_post
+ ldr w0, [rFP, #LO_next_interupt]
+ ldr w2, [rFP, #LO_cycle] // memhandlers can modify cc, like dma
+ str w0, [rFP, #LO_last_count]
+ sub w0, w2, w0
+.endm
+
+FUNCTION(do_memhandler_pre):
+ memhandler_pre
+ ret
+
+FUNCTION(do_memhandler_post):
+ memhandler_post
+ ret
+
+.macro pcsx_read_mem readop tab_shift
+ /* w0 = address, x1 = handler_tab, w2 = cycles */
+ ubfm w4, w0, #\tab_shift, #11
+ ldr x3, [x1, w4, uxtw #3]
+ adds x3, x3, x3
+ bcs 0f
+ \readop w0, [x3, w4, uxtw #\tab_shift]
+ ret
+0:
+ stp xzr, x30, [sp, #-16]!
+ memhandler_pre
+ blr x3
+.endm
+
FUNCTION(jump_handler_read8):
- bl abort
+ add x1, x1, #0x1000/4*8 + 0x1000/2*8 /* shift to r8 part */
+ pcsx_read_mem ldrb, 0
+ b handler_read_end
FUNCTION(jump_handler_read16):
- bl abort
+ add x1, x1, #0x1000/4*8 /* shift to r16 part */
+ pcsx_read_mem ldrh, 1
+ b handler_read_end
FUNCTION(jump_handler_read32):
- bl abort
+ pcsx_read_mem ldr, 2
+
+handler_read_end:
+ ldp xzr, x30, [sp], #16
+ ret
+
+.macro pcsx_write_mem wrtop movop tab_shift
+ /* w0 = address, w1 = data, w2 = cycles, x3 = handler_tab */
+ ubfm w4, w0, #\tab_shift, #11
+ ldr x3, [x3, w4, uxtw #3]
+ adds x3, x3, x3
+ bcs 0f
+ mov w0, w2 /* cycle return */
+ \wrtop w1, [x3, w4, uxtw #\tab_shift]
+ ret
+0:
+ stp xzr, x30, [sp, #-16]!
+ str w0, [rFP, #LO_address] /* some handlers still need it... */
+ \movop w0, w1
+ memhandler_pre
+ blr x3
+.endm
FUNCTION(jump_handler_write8):
- bl abort
+ add x3, x3, #0x1000/4*8 + 0x1000/2*8 /* shift to r8 part */
+ pcsx_write_mem strb uxtb 0
+ b handler_write_end
FUNCTION(jump_handler_write16):
- bl abort
+ add x3, x3, #0x1000/4*8 /* shift to r16 part */
+ pcsx_write_mem strh uxth 1
+ b handler_write_end
FUNCTION(jump_handler_write32):
- bl abort
+ pcsx_write_mem str mov 2
-FUNCTION(jump_handler_write_h):
- bl abort
+handler_write_end:
+ memhandler_post
+ ldp xzr, x30, [sp], #16
+ ret
FUNCTION(jump_handle_swl):
+ /* w0 = address, w1 = data, w2 = cycles */
+ ldr x3, [rFP, #LO_mem_wtab]
+ orr w4, wzr, w0, lsr #12
+ ldr x3, [x3, w4, uxtw #3]
+ adds x3, x3, x3
+ bcs 4f
+ add x3, x0, x3
+ mov w0, w2
+ tbz x3, #1, 10f // & 2
+ tbz x3, #0, 2f // & 1
+3:
+ stur w1, [x3, #-3]
+ ret
+2:
+ lsr w2, w1, #8
+ lsr w1, w1, #24
+ sturh w2, [x3, #-2]
+ strb w1, [x3]
+ ret
+10:
+ tbz x3, #0, 0f // & 1
+1:
+ lsr w1, w1, #16
+ sturh w1, [x3, #-1]
+ ret
+0:
+ lsr w2, w1, #24
+ strb w2, [x3]
+ ret
+4:
+ mov w0, w2 // todo
bl abort
+ ret
FUNCTION(jump_handle_swr):
+ /* w0 = address, w1 = data, w2 = cycles */
+ ldr x3, [rFP, #LO_mem_wtab]
+ orr w4, wzr, w0, lsr #12
+ ldr x3, [x3, w4, uxtw #3]
+ adds x3, x3, x3
+ bcs 4f
+ add x3, x0, x3
+ mov w0, w2
+ tbz x3, #1, 10f // & 2
+ tbz x3, #0, 2f // & 1
+3:
+ strb w1, [x3]
+ ret
+2:
+ strh w1, [x3]
+ ret
+10:
+ tbz x3, #0, 0f // & 1
+1:
+ lsr w2, w1, #8
+ strb w1, [x3]
+ sturh w2, [x3, #1]
+ ret
+0:
+ str w1, [x3]
+ ret
+4:
+ mov w0, w2 // todo
bl abort
+ ret
+
+FUNCTION(call_gteStall):
+ /* w0 = op_cycles, w1 = cycles */
+ ldr w2, [rFP, #LO_last_count]
+ str lr, [rFP, #LO_saved_lr]
+ add w1, w1, w2
+ str w1, [rFP, #LO_cycle]
+ add x1, rFP, #LO_psxRegs
+ bl gteCheckStallRaw
+ ldr lr, [rFP, #LO_saved_lr]
+ add rCC, rCC, w0
+ ret