1 /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
2 * linkage_arm.s for PCSX *
3 * Copyright (C) 2009-2011 Ari64 *
4 * Copyright (C) 2010-2011 GraÅžvydas "notaz" Ignotas *
6 * This program is free software; you can redistribute it and/or modify *
7 * it under the terms of the GNU General Public License as published by *
8 * the Free Software Foundation; either version 2 of the License, or *
9 * (at your option) any later version. *
11 * This program is distributed in the hope that it will be useful, *
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of *
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
14 * GNU General Public License for more details. *
16 * You should have received a copy of the GNU General Public License *
17 * along with this program; if not, write to the *
18 * Free Software Foundation, Inc., *
19 * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. *
20 * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
22 #include "arm_features.h"
37 .global pending_exception
45 .global restore_candidate
52 .global inv_code_start
58 .type dynarec_local, %object
59 .size dynarec_local, dynarec_local_end-dynarec_local
61 .space dynarec_local_end-dynarec_local
62 next_interupt = dynarec_local + 64
63 .type next_interupt, %object
64 .size next_interupt, 4
65 cycle_count = next_interupt + 4
66 .type cycle_count, %object
68 last_count = cycle_count + 4
69 .type last_count, %object
71 pending_exception = last_count + 4
72 .type pending_exception, %object
73 .size pending_exception, 4
74 stop = pending_exception + 4
78 .type invc_ptr, %object
80 address = invc_ptr + 4
81 .type address, %object
86 .type psxRegs, %object
87 .size psxRegs, psxRegs_end-psxRegs
98 .type reg_cop0, %object
100 reg_cop2d = reg_cop0 + 128
101 .type reg_cop2d, %object
103 reg_cop2c = reg_cop2d + 128
104 .type reg_cop2c, %object
116 interrupt = cycle + 4
117 .type interrupt, %object
119 intCycle = interrupt + 4
120 .type intCycle, %object
122 psxRegs_end = intCycle + 256
127 rcnts_end = rcnts + 7*4*4
130 .type mem_rtab, %object
132 mem_wtab = mem_rtab + 4
133 .type mem_wtab, %object
135 psxH_ptr = mem_wtab + 4
136 .type psxH_ptr, %object
138 zeromem_ptr = psxH_ptr + 4
139 .type zeromem_ptr, %object
141 inv_code_start = zeromem_ptr + 4
142 .type inv_code_start, %object
143 .size inv_code_start, 4
144 inv_code_end = inv_code_start + 4
145 .type inv_code_end, %object
146 .size inv_code_end, 4
147 branch_target = inv_code_end + 4
148 .type branch_target, %object
149 .size branch_target, 4
150 align0 = branch_target + 4 /* unused/alignment */
151 .type align0, %object
153 mini_ht = align0 + 16
154 .type mini_ht, %object
156 restore_candidate = mini_ht + 256
157 .type restore_candidate, %object
158 .size restore_candidate, 512
159 dynarec_local_end = restore_candidate + 512
176 .macro load_varadr reg var
177 #if defined(__ARM_ARCH_7A__) && !defined(__PIC__)
178 movw \reg, #:lower16:\var
179 movt \reg, #:upper16:\var
185 .macro mov_16 reg imm
186 #ifdef __ARM_ARCH_7A__
189 mov \reg, #(\imm & 0x00ff)
190 orr \reg, #(\imm & 0xff00)
194 .macro mov_24 reg imm
195 #ifdef __ARM_ARCH_7A__
196 movw \reg, #(\imm & 0xffff)
197 movt \reg, #(\imm >> 16)
199 mov \reg, #(\imm & 0x0000ff)
200 orr \reg, #(\imm & 0x00ff00)
201 orr \reg, #(\imm & 0xff0000)
205 .macro dyna_linker_main
206 /* r0 = virtual target address */
207 /* r1 = instruction to patch */
221 ldr r5, [r3, r2, lsl #2]
223 add r6, r1, r12, asr #6
238 moveq pc, r4 /* Stale i-cache */
240 b 1b /* jump_in may have dupes, continue search */
243 beq 3f /* r0 not in jump_in */
249 and r1, r7, #0xff000000
252 add r1, r1, r2, lsr #8
256 /* hash_table lookup */
259 eor r4, r0, r0, lsl #16
265 ldr r5, [r3, r2, lsl #2]
272 /* jump_dirty lookup */
282 /* hash_table insert */
296 FUNCTION(dyna_linker):
297 /* r0 = virtual target address */
298 /* r1 = instruction to patch */
303 bl new_recompile_block
311 .size dyna_linker, .-dyna_linker
313 FUNCTION(exec_pagefault):
314 /* r0 = instruction pointer */
315 /* r1 = fault address */
317 ldr r3, [fp, #reg_cop0+48-dynarec_local] /* Status */
319 ldr r4, [fp, #reg_cop0+16-dynarec_local] /* Context */
320 bic r6, r6, #0x0F800000
321 str r0, [fp, #reg_cop0+56-dynarec_local] /* EPC */
323 str r1, [fp, #reg_cop0+32-dynarec_local] /* BadVAddr */
325 str r3, [fp, #reg_cop0+48-dynarec_local] /* Status */
326 and r5, r6, r1, lsr #9
327 str r2, [fp, #reg_cop0+52-dynarec_local] /* Cause */
328 and r1, r1, r6, lsl #9
329 str r1, [fp, #reg_cop0+40-dynarec_local] /* EntryHi */
331 str r4, [fp, #reg_cop0+16-dynarec_local] /* Context */
335 .size exec_pagefault, .-exec_pagefault
337 /* Special dynamic linker for the case where a page fault
338 may occur in a branch delay slot */
339 FUNCTION(dyna_linker_ds):
340 /* r0 = virtual target address */
341 /* r1 = instruction to patch */
348 bl new_recompile_block
355 mov r2, #0x80000008 /* High bit set indicates pagefault in delay slot */
358 .size dyna_linker_ds, .-dyna_linker_ds
368 FUNCTION(jump_vaddr_r0):
369 eor r2, r0, r0, lsl #16
371 .size jump_vaddr_r0, .-jump_vaddr_r0
372 FUNCTION(jump_vaddr_r1):
373 eor r2, r1, r1, lsl #16
376 .size jump_vaddr_r1, .-jump_vaddr_r1
377 FUNCTION(jump_vaddr_r2):
379 eor r2, r2, r2, lsl #16
381 .size jump_vaddr_r2, .-jump_vaddr_r2
382 FUNCTION(jump_vaddr_r3):
383 eor r2, r3, r3, lsl #16
386 .size jump_vaddr_r3, .-jump_vaddr_r3
387 FUNCTION(jump_vaddr_r4):
388 eor r2, r4, r4, lsl #16
391 .size jump_vaddr_r4, .-jump_vaddr_r4
392 FUNCTION(jump_vaddr_r5):
393 eor r2, r5, r5, lsl #16
396 .size jump_vaddr_r5, .-jump_vaddr_r5
397 FUNCTION(jump_vaddr_r6):
398 eor r2, r6, r6, lsl #16
401 .size jump_vaddr_r6, .-jump_vaddr_r6
402 FUNCTION(jump_vaddr_r8):
403 eor r2, r8, r8, lsl #16
406 .size jump_vaddr_r8, .-jump_vaddr_r8
407 FUNCTION(jump_vaddr_r9):
408 eor r2, r9, r9, lsl #16
411 .size jump_vaddr_r9, .-jump_vaddr_r9
412 FUNCTION(jump_vaddr_r10):
413 eor r2, r10, r10, lsl #16
416 .size jump_vaddr_r10, .-jump_vaddr_r10
417 FUNCTION(jump_vaddr_r12):
418 eor r2, r12, r12, lsl #16
421 .size jump_vaddr_r12, .-jump_vaddr_r12
422 FUNCTION(jump_vaddr_r7):
423 eor r2, r7, r7, lsl #16
425 .size jump_vaddr_r7, .-jump_vaddr_r7
426 FUNCTION(jump_vaddr):
429 and r2, r3, r2, lsr #12
436 str r10, [fp, #cycle_count-dynarec_local]
438 ldr r10, [fp, #cycle_count-dynarec_local]
440 .size jump_vaddr, .-jump_vaddr
444 FUNCTION(verify_code_ds):
445 str r8, [fp, #branch_target-dynarec_local]
446 FUNCTION(verify_code_vm):
447 FUNCTION(verify_code):
475 ldr r8, [fp, #branch_target-dynarec_local]
480 .size verify_code, .-verify_code
481 .size verify_code_vm, .-verify_code_vm
484 FUNCTION(cc_interrupt):
485 ldr r0, [fp, #last_count-dynarec_local]
489 str r1, [fp, #pending_exception-dynarec_local]
490 and r2, r2, r10, lsr #17
491 add r3, fp, #restore_candidate-dynarec_local
492 str r10, [fp, #cycle-dynarec_local] /* PCSX cycles */
493 @@ str r10, [fp, #reg_cop0+36-dynarec_local] /* Count */
501 ldr r10, [fp, #cycle-dynarec_local]
502 ldr r0, [fp, #next_interupt-dynarec_local]
503 ldr r1, [fp, #pending_exception-dynarec_local]
504 ldr r2, [fp, #stop-dynarec_local]
505 str r0, [fp, #last_count-dynarec_local]
508 ldmnefd sp!, {r4, r5, r6, r7, r8, r9, sl, fp, ip, pc}
512 ldr r0, [fp, #pcaddr-dynarec_local]
516 /* Move 'dirty' blocks to the 'clean' list */
527 .size cc_interrupt, .-cc_interrupt
530 FUNCTION(do_interrupt):
531 ldr r0, [fp, #pcaddr-dynarec_local]
535 .size do_interrupt, .-do_interrupt
538 FUNCTION(fp_exception):
541 ldr r1, [fp, #reg_cop0+48-dynarec_local] /* Status */
543 str r0, [fp, #reg_cop0+56-dynarec_local] /* EPC */
546 str r1, [fp, #reg_cop0+48-dynarec_local] /* Status */
547 str r2, [fp, #reg_cop0+52-dynarec_local] /* Cause */
551 .size fp_exception, .-fp_exception
553 FUNCTION(fp_exception_ds):
554 mov r2, #0x90000000 /* Set high bit if delay slot */
556 .size fp_exception_ds, .-fp_exception_ds
559 FUNCTION(jump_syscall):
560 ldr r1, [fp, #reg_cop0+48-dynarec_local] /* Status */
562 str r0, [fp, #reg_cop0+56-dynarec_local] /* EPC */
565 str r1, [fp, #reg_cop0+48-dynarec_local] /* Status */
566 str r2, [fp, #reg_cop0+52-dynarec_local] /* Cause */
570 .size jump_syscall, .-jump_syscall
574 FUNCTION(jump_syscall_hle):
575 str r0, [fp, #pcaddr-dynarec_local] /* PC must be set to EPC for psxException */
576 ldr r2, [fp, #last_count-dynarec_local]
577 mov r1, #0 /* in delay slot */
579 mov r0, #0x20 /* cause */
580 str r2, [fp, #cycle-dynarec_local] /* PCSX cycle counter */
583 /* note: psxException might do recorsive recompiler call from it's HLE code,
584 * so be ready for this */
586 ldr r1, [fp, #next_interupt-dynarec_local]
587 ldr r10, [fp, #cycle-dynarec_local]
588 ldr r0, [fp, #pcaddr-dynarec_local]
590 str r1, [fp, #last_count-dynarec_local]
593 .size jump_syscall_hle, .-jump_syscall_hle
596 FUNCTION(jump_hlecall):
597 ldr r2, [fp, #last_count-dynarec_local]
598 str r0, [fp, #pcaddr-dynarec_local]
601 str r2, [fp, #cycle-dynarec_local] /* PCSX cycle counter */
603 .size jump_hlecall, .-jump_hlecall
606 FUNCTION(jump_intcall):
607 ldr r2, [fp, #last_count-dynarec_local]
608 str r0, [fp, #pcaddr-dynarec_local]
611 str r2, [fp, #cycle-dynarec_local] /* PCSX cycle counter */
613 .size jump_hlecall, .-jump_hlecall
616 FUNCTION(new_dyna_leave):
617 ldr r0, [fp, #last_count-dynarec_local]
620 str r10, [fp, #cycle-dynarec_local]
621 ldmfd sp!, {r4, r5, r6, r7, r8, r9, sl, fp, ip, pc}
622 .size new_dyna_leave, .-new_dyna_leave
625 FUNCTION(invalidate_addr_r0):
626 stmia fp, {r0, r1, r2, r3, r12, lr}
627 b invalidate_addr_call
628 .size invalidate_addr_r0, .-invalidate_addr_r0
630 FUNCTION(invalidate_addr_r1):
631 stmia fp, {r0, r1, r2, r3, r12, lr}
633 b invalidate_addr_call
634 .size invalidate_addr_r1, .-invalidate_addr_r1
636 FUNCTION(invalidate_addr_r2):
637 stmia fp, {r0, r1, r2, r3, r12, lr}
639 b invalidate_addr_call
640 .size invalidate_addr_r2, .-invalidate_addr_r2
642 FUNCTION(invalidate_addr_r3):
643 stmia fp, {r0, r1, r2, r3, r12, lr}
645 b invalidate_addr_call
646 .size invalidate_addr_r3, .-invalidate_addr_r3
648 FUNCTION(invalidate_addr_r4):
649 stmia fp, {r0, r1, r2, r3, r12, lr}
651 b invalidate_addr_call
652 .size invalidate_addr_r4, .-invalidate_addr_r4
654 FUNCTION(invalidate_addr_r5):
655 stmia fp, {r0, r1, r2, r3, r12, lr}
657 b invalidate_addr_call
658 .size invalidate_addr_r5, .-invalidate_addr_r5
660 FUNCTION(invalidate_addr_r6):
661 stmia fp, {r0, r1, r2, r3, r12, lr}
663 b invalidate_addr_call
664 .size invalidate_addr_r6, .-invalidate_addr_r6
666 FUNCTION(invalidate_addr_r7):
667 stmia fp, {r0, r1, r2, r3, r12, lr}
669 b invalidate_addr_call
670 .size invalidate_addr_r7, .-invalidate_addr_r7
672 FUNCTION(invalidate_addr_r8):
673 stmia fp, {r0, r1, r2, r3, r12, lr}
675 b invalidate_addr_call
676 .size invalidate_addr_r8, .-invalidate_addr_r8
678 FUNCTION(invalidate_addr_r9):
679 stmia fp, {r0, r1, r2, r3, r12, lr}
681 b invalidate_addr_call
682 .size invalidate_addr_r9, .-invalidate_addr_r9
684 FUNCTION(invalidate_addr_r10):
685 stmia fp, {r0, r1, r2, r3, r12, lr}
687 b invalidate_addr_call
688 .size invalidate_addr_r10, .-invalidate_addr_r10
690 FUNCTION(invalidate_addr_r12):
691 stmia fp, {r0, r1, r2, r3, r12, lr}
693 .size invalidate_addr_r12, .-invalidate_addr_r12
695 FUNCTION(invalidate_addr_call):
696 ldr r12, [fp, #inv_code_start-dynarec_local]
697 ldr lr, [fp, #inv_code_end-dynarec_local]
701 ldmia fp, {r0, r1, r2, r3, r12, pc}
702 .size invalidate_addr_call, .-invalidate_addr_call
705 FUNCTION(new_dyna_start):
706 /* ip is stored to conform EABI alignment */
707 stmfd sp!, {r4, r5, r6, r7, r8, r9, sl, fp, ip, lr}
708 load_varadr fp, dynarec_local
709 ldr r0, [fp, #pcaddr-dynarec_local]
711 ldr r1, [fp, #next_interupt-dynarec_local]
712 ldr r10, [fp, #cycle-dynarec_local]
713 str r1, [fp, #last_count-dynarec_local]
716 .size new_dyna_start, .-new_dyna_start
718 /* --------------------------------------- */
722 .macro pcsx_read_mem readop tab_shift
723 /* r0 = address, r1 = handler_tab, r2 = cycles */
725 lsr r3, #(20+\tab_shift)
726 ldr r12, [fp, #last_count-dynarec_local]
727 ldr r1, [r1, r3, lsl #2]
734 \readop r0, [r1, r3, lsl #\tab_shift]
737 str r2, [fp, #cycle-dynarec_local]
741 FUNCTION(jump_handler_read8):
742 add r1, #0x1000/4*4 + 0x1000/2*4 @ shift to r8 part
743 pcsx_read_mem ldrccb, 0
745 FUNCTION(jump_handler_read16):
746 add r1, #0x1000/4*4 @ shift to r16 part
747 pcsx_read_mem ldrcch, 1
749 FUNCTION(jump_handler_read32):
750 pcsx_read_mem ldrcc, 2
753 .macro pcsx_write_mem wrtop tab_shift
754 /* r0 = address, r1 = data, r2 = cycles, r3 = handler_tab */
756 lsr r12, #(20+\tab_shift)
757 ldr r3, [r3, r12, lsl #2]
758 str r0, [fp, #address-dynarec_local] @ some handlers still need it..
760 mov r0, r2 @ cycle return in case of direct store
765 \wrtop r1, [r3, r12, lsl #\tab_shift]
768 ldr r12, [fp, #last_count-dynarec_local]
772 str r2, [fp, #cycle-dynarec_local]
775 ldr r0, [fp, #next_interupt-dynarec_local]
777 str r0, [fp, #last_count-dynarec_local]
782 FUNCTION(jump_handler_write8):
783 add r3, #0x1000/4*4 + 0x1000/2*4 @ shift to r8 part
784 pcsx_write_mem strccb, 0
786 FUNCTION(jump_handler_write16):
787 add r3, #0x1000/4*4 @ shift to r16 part
788 pcsx_write_mem strcch, 1
790 FUNCTION(jump_handler_write32):
791 pcsx_write_mem strcc, 2
793 FUNCTION(jump_handler_write_h):
794 /* r0 = address, r1 = data, r2 = cycles, r3 = handler */
795 ldr r12, [fp, #last_count-dynarec_local]
796 str r0, [fp, #address-dynarec_local] @ some handlers still need it..
800 str r2, [fp, #cycle-dynarec_local]
803 ldr r0, [fp, #next_interupt-dynarec_local]
805 str r0, [fp, #last_count-dynarec_local]
809 FUNCTION(jump_handle_swl):
810 /* r0 = address, r1 = data, r2 = cycles */
811 ldr r3, [fp, #mem_wtab-dynarec_local]
813 ldr r3, [r3, r12, lsl #2]
834 lsreq r12, r1, #24 @ 0
844 FUNCTION(jump_handle_swr):
845 /* r0 = address, r1 = data, r2 = cycles */
846 ldr r3, [fp, #mem_wtab-dynarec_local]
848 ldr r3, [r3, r12, lsl #2]
870 .macro rcntx_read_mode0 num
871 /* r0 = address, r2 = cycles */
872 ldr r3, [fp, #rcnts-dynarec_local+6*4+7*4*\num] @ cycleStart
879 FUNCTION(rcnt0_read_count_m0):
882 FUNCTION(rcnt1_read_count_m0):
885 FUNCTION(rcnt2_read_count_m0):
888 FUNCTION(rcnt0_read_count_m1):
889 /* r0 = address, r2 = cycles */
890 ldr r3, [fp, #rcnts-dynarec_local+6*4+7*4*0] @ cycleStart
893 mul r0, r1, r2 @ /= 5
897 FUNCTION(rcnt1_read_count_m1):
898 /* r0 = address, r2 = cycles */
899 ldr r3, [fp, #rcnts-dynarec_local+6*4+7*4*1]
902 umull r3, r0, r1, r2 @ ~ /= hsync_cycles, max ~0x1e6cdd
905 FUNCTION(rcnt2_read_count_m1):
906 /* r0 = address, r2 = cycles */
907 ldr r3, [fp, #rcnts-dynarec_local+6*4+7*4*2]
908 mov r0, r2, lsl #16-3
909 sub r0, r3, lsl #16-3
913 @ vim:filetype=armasm