+// SPDX-License-Identifier: LGPL-2.1-or-later
/*
- * Copyright (C) 2014-2020 Paul Cercueil <paul@crapouillou.net>
- *
- * This library is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * This library is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
+ * Copyright (C) 2014-2021 Paul Cercueil <paul@crapouillou.net>
*/
#include "blockcache.h"
-#include "config.h"
#include "debug.h"
#include "disassembler.h"
#include "emitter.h"
#include "interpreter.h"
+#include "lightrec-config.h"
+#include "lightning-wrapper.h"
#include "lightrec.h"
#include "memmanager.h"
#include "reaper.h"
#include "recompiler.h"
#include "regcache.h"
#include "optimizer.h"
+#include "tlsf/tlsf.h"
#include <errno.h>
-#include <lightning.h>
+#include <inttypes.h>
#include <limits.h>
#if ENABLE_THREADED_COMPILER
#include <stdatomic.h>
#include <stdbool.h>
#include <stddef.h>
#include <string.h>
-#if ENABLE_TINYMM
-#include <tinymm.h>
-#endif
-
-#define GENMASK(h, l) \
- (((uintptr_t)-1 << (l)) & ((uintptr_t)-1 >> (__WORDSIZE - 1 - (h))))
static struct block * lightrec_precompile_block(struct lightrec_state *state,
u32 pc);
+static bool lightrec_block_is_fully_tagged(const struct block *block);
+
+static void lightrec_mtc2(struct lightrec_state *state, u8 reg, u32 data);
+static u32 lightrec_mfc2(struct lightrec_state *state, u8 reg);
static void lightrec_default_sb(struct lightrec_state *state, u32 opcode,
void *host, u32 addr, u8 data)
{
*(u8 *)host = data;
- if (!state->invalidate_from_dma_only)
+ if (!(state->opt_flags & LIGHTREC_OPT_INV_DMA_ONLY))
lightrec_invalidate(state, addr, 1);
}
{
*(u16 *)host = HTOLE16(data);
- if (!state->invalidate_from_dma_only)
+ if (!(state->opt_flags & LIGHTREC_OPT_INV_DMA_ONLY))
lightrec_invalidate(state, addr, 2);
}
{
*(u32 *)host = HTOLE32(data);
- if (!state->invalidate_from_dma_only)
+ if (!(state->opt_flags & LIGHTREC_OPT_INV_DMA_ONLY))
lightrec_invalidate(state, addr, 4);
}
.lw = lightrec_default_lw,
};
-static void __segfault_cb(struct lightrec_state *state, u32 addr)
+static void __segfault_cb(struct lightrec_state *state, u32 addr,
+ const struct block *block)
{
lightrec_set_exit_flags(state, LIGHTREC_EXIT_SEGFAULT);
pr_err("Segmentation fault in recompiled code: invalid "
"load/store at address 0x%08x\n", addr);
+ if (block)
+ pr_err("Was executing block PC 0x%08x\n", block->pc);
}
static void lightrec_swl(struct lightrec_state *state,
u32 opcode, void *host, u32 addr, u32 data)
{
unsigned int shift = addr & 0x3;
- unsigned int mask = GENMASK(31, (shift + 1) * 8);
+ unsigned int mask = shift < 3 ? GENMASK(31, (shift + 1) * 8) : 0;
u32 old_data;
/* Align to 32 bits */
const struct lightrec_mem_map_ops *ops,
void *host, u32 addr)
{
- u32 data = state->ops.cop2_ops.mfc(state, op.opcode, op.i.rt);
+ u32 data = lightrec_mfc2(state, op.i.rt);
ops->sw(state, op.opcode, host, addr, data);
}
u32 opcode, void *host, u32 addr, u32 data)
{
unsigned int shift = addr & 0x3;
- unsigned int mask = GENMASK(31, 32 - shift * 8);
+ unsigned int mask = shift ? GENMASK(31, 32 - shift * 8) : 0;
u32 old_data;
/* Align to 32 bits */
{
u32 data = ops->lw(state, op.opcode, host, addr);
- state->ops.cop2_ops.mtc(state, op.opcode, op.i.rt, data);
+ lightrec_mtc2(state, op.i.rt, data);
}
static void lightrec_invalidate_map(struct lightrec_state *state,
- const struct lightrec_mem_map *map, u32 addr)
+ const struct lightrec_mem_map *map, u32 addr, u32 len)
{
- if (map == &state->maps[PSX_MAP_KERNEL_USER_RAM])
- state->code_lut[lut_offset(addr)] = NULL;
+ if (map == &state->maps[PSX_MAP_KERNEL_USER_RAM]) {
+ memset(lut_address(state, lut_offset(addr)), 0,
+ ((len + 3) / 4) * lut_elm_size(state));
+ }
}
-static const struct lightrec_mem_map *
-lightrec_get_map(struct lightrec_state *state, u32 kaddr)
+static enum psx_map
+lightrec_get_map_idx(struct lightrec_state *state, u32 kaddr)
{
+ const struct lightrec_mem_map *map;
unsigned int i;
for (i = 0; i < state->nb_maps; i++) {
- const struct lightrec_mem_map *map = &state->maps[i];
+ map = &state->maps[i];
if (kaddr >= map->pc && kaddr < map->pc + map->length)
- return map;
+ return (enum psx_map) i;
}
- return NULL;
+ return PSX_MAP_UNKNOWN;
}
-u32 lightrec_rw(struct lightrec_state *state, union code op,
- u32 addr, u32 data, u16 *flags)
+const struct lightrec_mem_map *
+lightrec_get_map(struct lightrec_state *state, void **host, u32 kaddr)
+{
+ const struct lightrec_mem_map *map;
+ enum psx_map idx;
+ u32 addr;
+
+ idx = lightrec_get_map_idx(state, kaddr);
+ if (idx == PSX_MAP_UNKNOWN)
+ return NULL;
+
+ map = &state->maps[idx];
+ addr = kaddr - map->pc;
+
+ while (map->mirror_of)
+ map = map->mirror_of;
+
+ if (host)
+ *host = map->address + addr;
+
+ return map;
+}
+
+u32 lightrec_rw(struct lightrec_state *state, union code op, u32 base,
+ u32 data, u32 *flags, struct block *block, u16 offset)
{
const struct lightrec_mem_map *map;
const struct lightrec_mem_map_ops *ops;
- u32 kaddr, pc, opcode = op.opcode;
+ u32 opcode = op.opcode;
+ bool was_tagged = true;
+ u16 old_flags;
+ u32 addr;
void *host;
- addr += (s16) op.i.imm;
- kaddr = kunseg(addr);
+ addr = kunseg(base + (s16) op.i.imm);
- map = lightrec_get_map(state, kaddr);
+ map = lightrec_get_map(state, &host, addr);
if (!map) {
- __segfault_cb(state, addr);
+ __segfault_cb(state, addr, block);
return 0;
}
- pc = map->pc;
+ if (flags)
+ was_tagged = LIGHTREC_FLAGS_GET_IO_MODE(*flags);
+
+ if (likely(!map->ops)) {
+ if (flags && !LIGHTREC_FLAGS_GET_IO_MODE(*flags)) {
+ /* Force parallel port accesses as HW accesses, because
+ * the direct-I/O emitters can't differenciate it. */
+ if (unlikely(map == &state->maps[PSX_MAP_PARALLEL_PORT]))
+ *flags |= LIGHTREC_IO_MODE(LIGHTREC_IO_HW);
+ /* If the base register is 0x0, be extra suspicious.
+ * Some games (e.g. Sled Storm) actually do segmentation
+ * faults by using uninitialized pointers, which are
+ * later initialized to point to hardware registers. */
+ else if (op.i.rs && base == 0x0)
+ *flags |= LIGHTREC_IO_MODE(LIGHTREC_IO_HW);
+ else
+ *flags |= LIGHTREC_IO_MODE(LIGHTREC_IO_DIRECT);
+ }
- while (map->mirror_of)
- map = map->mirror_of;
+ ops = &lightrec_default_ops;
+ } else if (flags &&
+ LIGHTREC_FLAGS_GET_IO_MODE(*flags) == LIGHTREC_IO_DIRECT_HW) {
+ ops = &lightrec_default_ops;
+ } else {
+ if (flags && !LIGHTREC_FLAGS_GET_IO_MODE(*flags))
+ *flags |= LIGHTREC_IO_MODE(LIGHTREC_IO_HW);
- host = (void *)((uintptr_t)map->address + kaddr - pc);
+ ops = map->ops;
+ }
- if (unlikely(map->ops)) {
- if (flags)
- *flags |= LIGHTREC_HW_IO;
+ if (!was_tagged) {
+ old_flags = block_set_flags(block, BLOCK_SHOULD_RECOMPILE);
- ops = map->ops;
- } else {
- if (flags)
- *flags |= LIGHTREC_DIRECT_IO;
+ if (!(old_flags & BLOCK_SHOULD_RECOMPILE)) {
+ pr_debug("Opcode of block at PC 0x%08x has been tagged"
+ " - flag for recompilation\n", block->pc);
- ops = &lightrec_default_ops;
+ lut_write(state, lut_offset(block->pc), NULL);
+ }
}
switch (op.i.op) {
}
static void lightrec_rw_helper(struct lightrec_state *state,
- union code op, u16 *flags)
+ union code op, u32 *flags,
+ struct block *block, u16 offset)
{
- u32 ret = lightrec_rw(state, op,
- state->native_reg_cache[op.i.rs],
- state->native_reg_cache[op.i.rt], flags);
+ u32 ret = lightrec_rw(state, op, state->regs.gpr[op.i.rs],
+ state->regs.gpr[op.i.rt], flags, block, offset);
switch (op.i.op) {
case OP_LB:
case OP_LWL:
case OP_LWR:
case OP_LW:
- if (op.i.rt)
- state->native_reg_cache[op.i.rt] = ret;
- default: /* fall-through */
+ if (OPT_HANDLE_LOAD_DELAYS && unlikely(!state->in_delay_slot_n)) {
+ state->temp_reg = ret;
+ state->in_delay_slot_n = 0xff;
+ } else if (op.i.rt) {
+ state->regs.gpr[op.i.rt] = ret;
+ }
+ fallthrough;
+ default:
break;
}
}
-static void lightrec_rw_cb(struct lightrec_state *state, union code op)
+static void lightrec_rw_cb(struct lightrec_state *state, u32 arg)
{
- lightrec_rw_helper(state, op, NULL);
+ lightrec_rw_helper(state, (union code) arg, NULL, NULL, 0);
}
-static void lightrec_rw_generic_cb(struct lightrec_state *state,
- struct opcode *op, struct block *block)
+static void lightrec_rw_generic_cb(struct lightrec_state *state, u32 arg)
{
- bool was_tagged = op->flags & (LIGHTREC_HW_IO | LIGHTREC_DIRECT_IO);
+ struct block *block;
+ struct opcode *op;
+ u16 offset = (u16)arg;
+
+ block = lightrec_find_block_from_lut(state->block_cache,
+ arg >> 16, state->curr_pc);
+ if (unlikely(!block)) {
+ pr_err("rw_generic: No block found in LUT for PC 0x%x offset 0x%x\n",
+ state->curr_pc, offset);
+ lightrec_set_exit_flags(state, LIGHTREC_EXIT_SEGFAULT);
+ return;
+ }
- lightrec_rw_helper(state, op->c, &op->flags);
+ op = &block->opcode_list[offset];
+ lightrec_rw_helper(state, op->c, &op->flags, block, offset);
+}
- if (!was_tagged) {
- pr_debug("Opcode of block at PC 0x%08x offset 0x%x has been "
- "tagged - flag for recompilation\n",
- block->pc, op->offset << 2);
+static u32 clamp_s32(s32 val, s32 min, s32 max)
+{
+ return val < min ? min : val > max ? max : val;
+}
+
+static u16 load_u16(u32 *ptr)
+{
+ return ((struct u16x2 *) ptr)->l;
+}
- block->flags |= BLOCK_SHOULD_RECOMPILE;
+static void store_u16(u32 *ptr, u16 value)
+{
+ ((struct u16x2 *) ptr)->l = value;
+}
+
+static u32 lightrec_mfc2(struct lightrec_state *state, u8 reg)
+{
+ s16 gteir1, gteir2, gteir3;
+
+ switch (reg) {
+ case 1:
+ case 3:
+ case 5:
+ case 8:
+ case 9:
+ case 10:
+ case 11:
+ return (s32)(s16) load_u16(&state->regs.cp2d[reg]);
+ case 7:
+ case 16:
+ case 17:
+ case 18:
+ case 19:
+ return load_u16(&state->regs.cp2d[reg]);
+ case 28:
+ case 29:
+ gteir1 = (s16) load_u16(&state->regs.cp2d[9]);
+ gteir2 = (s16) load_u16(&state->regs.cp2d[10]);
+ gteir3 = (s16) load_u16(&state->regs.cp2d[11]);
+
+ return clamp_s32(gteir1 >> 7, 0, 0x1f) << 0 |
+ clamp_s32(gteir2 >> 7, 0, 0x1f) << 5 |
+ clamp_s32(gteir3 >> 7, 0, 0x1f) << 10;
+ case 15:
+ reg = 14;
+ fallthrough;
+ default:
+ return state->regs.cp2d[reg];
}
}
u32 lightrec_mfc(struct lightrec_state *state, union code op)
{
- bool is_cfc = (op.i.op == OP_CP0 && op.r.rs == OP_CP0_CFC0) ||
- (op.i.op == OP_CP2 && op.r.rs == OP_CP2_BASIC_CFC2);
- u32 (*func)(struct lightrec_state *, u32, u8);
- const struct lightrec_cop_ops *ops;
+ u32 val;
if (op.i.op == OP_CP0)
- ops = &state->ops.cop0_ops;
- else
- ops = &state->ops.cop2_ops;
+ return state->regs.cp0[op.r.rd];
+
+ if (op.i.op == OP_SWC2) {
+ val = lightrec_mfc2(state, op.i.rt);
+ } else if (op.r.rs == OP_CP2_BASIC_MFC2)
+ val = lightrec_mfc2(state, op.r.rd);
+ else {
+ val = state->regs.cp2c[op.r.rd];
+
+ switch (op.r.rd) {
+ case 4:
+ case 12:
+ case 20:
+ case 26:
+ case 27:
+ case 29:
+ case 30:
+ val = (u32)(s16)val;
+ fallthrough;
+ default:
+ break;
+ }
+ }
- if (is_cfc)
- func = ops->cfc;
- else
- func = ops->mfc;
+ if (state->ops.cop2_notify)
+ (*state->ops.cop2_notify)(state, op.opcode, val);
- return (*func)(state, op.opcode, op.r.rd);
+ return val;
}
static void lightrec_mfc_cb(struct lightrec_state *state, union code op)
{
u32 rt = lightrec_mfc(state, op);
- if (op.r.rt)
- state->native_reg_cache[op.r.rt] = rt;
+ if (op.i.op == OP_SWC2)
+ state->temp_reg = rt;
+ else if (op.r.rt)
+ state->regs.gpr[op.r.rt] = rt;
}
-void lightrec_mtc(struct lightrec_state *state, union code op, u32 data)
+static void lightrec_mtc0(struct lightrec_state *state, u8 reg, u32 data)
{
- bool is_ctc = (op.i.op == OP_CP0 && op.r.rs == OP_CP0_CTC0) ||
- (op.i.op == OP_CP2 && op.r.rs == OP_CP2_BASIC_CTC2);
- void (*func)(struct lightrec_state *, u32, u8, u32);
- const struct lightrec_cop_ops *ops;
+ u32 status, oldstatus, cause;
+
+ switch (reg) {
+ case 1:
+ case 4:
+ case 8:
+ case 14:
+ case 15:
+ /* Those registers are read-only */
+ return;
+ default:
+ break;
+ }
- if (op.i.op == OP_CP0)
- ops = &state->ops.cop0_ops;
- else
- ops = &state->ops.cop2_ops;
+ if (reg == 12) {
+ status = state->regs.cp0[12];
+ oldstatus = status;
- if (is_ctc)
- func = ops->ctc;
- else
- func = ops->mtc;
+ if (status & ~data & BIT(16)) {
+ state->ops.enable_ram(state, true);
+ lightrec_invalidate_all(state);
+ } else if (~status & data & BIT(16)) {
+ state->ops.enable_ram(state, false);
+ }
+ }
+
+ if (reg == 13) {
+ state->regs.cp0[13] &= ~0x300;
+ state->regs.cp0[13] |= data & 0x300;
+ } else {
+ state->regs.cp0[reg] = data;
+ }
+
+ if (reg == 12 || reg == 13) {
+ cause = state->regs.cp0[13];
+ status = state->regs.cp0[12];
+
+ /* Handle software interrupts */
+ if ((!!(status & cause & 0x300)) & status)
+ lightrec_set_exit_flags(state, LIGHTREC_EXIT_CHECK_INTERRUPT);
- (*func)(state, op.opcode, op.r.rd, data);
+ /* Handle hardware interrupts */
+ if (reg == 12 && !(~status & 0x401) && (~oldstatus & 0x401))
+ lightrec_set_exit_flags(state, LIGHTREC_EXIT_CHECK_INTERRUPT);
+ }
}
-static void lightrec_mtc_cb(struct lightrec_state *state, union code op)
+static u32 count_leading_bits(s32 data)
{
- lightrec_mtc(state, op, state->native_reg_cache[op.r.rt]);
+ u32 cnt = 33;
+
+#ifdef __has_builtin
+#if __has_builtin(__builtin_clrsb)
+ return 1 + __builtin_clrsb(data);
+#endif
+#endif
+
+ data = (data ^ (data >> 31)) << 1;
+
+ do {
+ cnt -= 1;
+ data >>= 1;
+ } while (data);
+
+ return cnt;
}
-static void lightrec_rfe_cb(struct lightrec_state *state, union code op)
+static void lightrec_mtc2(struct lightrec_state *state, u8 reg, u32 data)
+{
+ switch (reg) {
+ case 15:
+ state->regs.cp2d[12] = state->regs.cp2d[13];
+ state->regs.cp2d[13] = state->regs.cp2d[14];
+ state->regs.cp2d[14] = data;
+ break;
+ case 28:
+ state->regs.cp2d[9] = (data << 7) & 0xf80;
+ state->regs.cp2d[10] = (data << 2) & 0xf80;
+ state->regs.cp2d[11] = (data >> 3) & 0xf80;
+ break;
+ case 31:
+ return;
+ case 30:
+ state->regs.cp2d[31] = count_leading_bits((s32) data);
+ fallthrough;
+ default:
+ state->regs.cp2d[reg] = data;
+ break;
+ }
+}
+
+static void lightrec_ctc2(struct lightrec_state *state, u8 reg, u32 data)
+{
+ switch (reg) {
+ case 4:
+ case 12:
+ case 20:
+ case 26:
+ case 27:
+ case 29:
+ case 30:
+ store_u16(&state->regs.cp2c[reg], data);
+ break;
+ case 31:
+ data = (data & 0x7ffff000) | !!(data & 0x7f87e000) << 31;
+ fallthrough;
+ default:
+ state->regs.cp2c[reg] = data;
+ break;
+ }
+}
+
+void lightrec_mtc(struct lightrec_state *state, union code op, u8 reg, u32 data)
+{
+ if (op.i.op == OP_CP0) {
+ lightrec_mtc0(state, reg, data);
+ } else {
+ if (op.i.op == OP_LWC2 || op.r.rs != OP_CP2_BASIC_CTC2)
+ lightrec_mtc2(state, reg, data);
+ else
+ lightrec_ctc2(state, reg, data);
+
+ if (state->ops.cop2_notify)
+ (*state->ops.cop2_notify)(state, op.opcode, data);
+ }
+}
+
+static void lightrec_mtc_cb(struct lightrec_state *state, u32 arg)
+{
+ union code op = (union code) arg;
+ u32 data;
+ u8 reg;
+
+ if (op.i.op == OP_LWC2) {
+ data = state->temp_reg;
+ reg = op.i.rt;
+ } else {
+ data = state->regs.gpr[op.r.rt];
+ reg = op.r.rd;
+ }
+
+ lightrec_mtc(state, op, reg, data);
+}
+
+void lightrec_rfe(struct lightrec_state *state)
{
u32 status;
/* Read CP0 Status register (r12) */
- status = state->ops.cop0_ops.mfc(state, op.opcode, 12);
+ status = state->regs.cp0[12];
/* Switch the bits */
status = ((status & 0x3c) >> 2) | (status & ~0xf);
/* Write it back */
- state->ops.cop0_ops.ctc(state, op.opcode, 12, status);
+ lightrec_mtc0(state, 12, status);
}
-static void lightrec_cp_cb(struct lightrec_state *state, union code op)
+void lightrec_cp(struct lightrec_state *state, union code op)
{
- void (*func)(struct lightrec_state *, u32);
-
- if ((op.opcode >> 25) & 1)
- func = state->ops.cop2_ops.op;
- else
- func = state->ops.cop0_ops.op;
-
- (*func)(state, op.opcode);
-}
+ if (op.i.op == OP_CP0) {
+ pr_err("Invalid CP opcode to coprocessor #0\n");
+ return;
+ }
-static void lightrec_syscall_cb(struct lightrec_state *state, union code op)
-{
- lightrec_set_exit_flags(state, LIGHTREC_EXIT_SYSCALL);
+ (*state->ops.cop2_op)(state, op.opcode);
}
-static void lightrec_break_cb(struct lightrec_state *state, union code op)
+static void lightrec_cp_cb(struct lightrec_state *state, u32 arg)
{
- lightrec_set_exit_flags(state, LIGHTREC_EXIT_BREAK);
+ lightrec_cp(state, (union code) arg);
}
-struct block * lightrec_get_block(struct lightrec_state *state, u32 pc)
+static struct block * lightrec_get_block(struct lightrec_state *state, u32 pc)
{
struct block *block = lightrec_find_block(state->block_cache, pc);
+ u8 old_flags;
- if (block && lightrec_block_is_outdated(block)) {
+ if (block && lightrec_block_is_outdated(state, block)) {
pr_debug("Block at PC 0x%08x is outdated!\n", block->pc);
- /* Make sure the recompiler isn't processing the block we'll
- * destroy */
- if (ENABLE_THREADED_COMPILER)
- lightrec_recompiler_remove(state->rec, block);
+ old_flags = block_set_flags(block, BLOCK_IS_DEAD);
+ if (!(old_flags & BLOCK_IS_DEAD)) {
+ /* Make sure the recompiler isn't processing the block
+ * we'll destroy */
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_recompiler_remove(state->rec, block);
+
+ lightrec_unregister_block(state->block_cache, block);
+ remove_from_code_lut(state->block_cache, block);
+ lightrec_free_block(state, block);
+ }
- lightrec_unregister_block(state->block_cache, block);
- remove_from_code_lut(state->block_cache, block);
- lightrec_free_block(block);
block = NULL;
}
struct block *block;
bool should_recompile;
void *func;
+ int err;
- for (;;) {
- func = state->code_lut[lut_offset(pc)];
+ do {
+ func = lut_read(state, lut_offset(pc));
if (func && func != state->get_next_block)
- return func;
+ break;
block = lightrec_get_block(state, pc);
if (unlikely(!block))
- return NULL;
+ break;
+
+ if (OPT_REPLACE_MEMSET &&
+ block_has_flag(block, BLOCK_IS_MEMSET)) {
+ func = state->memset_func;
+ break;
+ }
- should_recompile = block->flags & BLOCK_SHOULD_RECOMPILE &&
- !(block->flags & BLOCK_IS_DEAD);
+ should_recompile = block_has_flag(block, BLOCK_SHOULD_RECOMPILE) &&
+ !block_has_flag(block, BLOCK_NEVER_COMPILE) &&
+ !block_has_flag(block, BLOCK_IS_DEAD);
if (unlikely(should_recompile)) {
pr_debug("Block at PC 0x%08x should recompile\n", pc);
- lightrec_unregister(MEM_FOR_CODE, block->code_size);
-
- if (ENABLE_THREADED_COMPILER)
+ if (ENABLE_THREADED_COMPILER) {
lightrec_recompiler_add(state->rec, block);
- else
- lightrec_compile_block(block);
+ } else {
+ err = lightrec_compile_block(state->cstate, block);
+ if (err) {
+ state->exit_flags = LIGHTREC_EXIT_NOMEM;
+ return NULL;
+ }
+ }
}
if (ENABLE_THREADED_COMPILER && likely(!should_recompile))
- func = lightrec_recompiler_run_first_pass(block, &pc);
+ func = lightrec_recompiler_run_first_pass(state, block, &pc);
else
func = block->function;
if (likely(func))
- return func;
+ break;
- /* Block wasn't compiled yet - run the interpreter */
- if (!ENABLE_THREADED_COMPILER &&
- ((ENABLE_FIRST_PASS && likely(!should_recompile)) ||
- unlikely(block->flags & BLOCK_NEVER_COMPILE)))
- pc = lightrec_emulate_block(block, pc);
+ if (unlikely(block_has_flag(block, BLOCK_NEVER_COMPILE))) {
+ pc = lightrec_emulate_block(state, block, pc);
+
+ } else if (!ENABLE_THREADED_COMPILER) {
+ /* Block wasn't compiled yet - run the interpreter */
+ if (block_has_flag(block, BLOCK_FULLY_TAGGED))
+ pr_debug("Block fully tagged, skipping first pass\n");
+ else if (ENABLE_FIRST_PASS && likely(!should_recompile))
+ pc = lightrec_emulate_block(state, block, pc);
- if (likely(!(block->flags & BLOCK_NEVER_COMPILE))) {
/* Then compile it using the profiled data */
- if (ENABLE_THREADED_COMPILER)
- lightrec_recompiler_add(state->rec, block);
- else
- lightrec_compile_block(block);
+ err = lightrec_compile_block(state->cstate, block);
+ if (err) {
+ state->exit_flags = LIGHTREC_EXIT_NOMEM;
+ return NULL;
+ }
+ } else if (unlikely(block_has_flag(block, BLOCK_IS_DEAD))) {
+ /*
+ * If the block is dead but has never been compiled,
+ * then its function pointer is NULL and we cannot
+ * execute the block. In that case, reap all the dead
+ * blocks now, and in the next loop we will create a
+ * new block.
+ */
+ lightrec_reaper_reap(state->reaper);
+ } else {
+ lightrec_recompiler_add(state->rec, block);
}
+ } while (state->exit_flags == LIGHTREC_EXIT_NORMAL
+ && state->current_cycle < state->target_cycle);
- if (state->exit_flags != LIGHTREC_EXIT_NORMAL ||
- state->current_cycle >= state->target_cycle) {
- state->next_pc = pc;
- return NULL;
- }
- }
+ state->curr_pc = pc;
+ return func;
}
-static s32 c_generic_function_wrapper(struct lightrec_state *state,
- s32 cycles_delta,
- void (*f)(struct lightrec_state *,
- struct opcode *,
- struct block *),
- struct opcode *op, struct block *block)
+static void * lightrec_alloc_code(struct lightrec_state *state, size_t size)
{
- state->current_cycle = state->target_cycle - cycles_delta;
+ void *code;
- (*f)(state, op, block);
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_code_alloc_lock(state);
- return state->target_cycle - state->current_cycle;
+ code = tlsf_malloc(state->tlsf, size);
+
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_code_alloc_unlock(state);
+
+ return code;
+}
+
+static void lightrec_realloc_code(struct lightrec_state *state,
+ void *ptr, size_t size)
+{
+ /* NOTE: 'size' MUST be smaller than the size specified during
+ * the allocation. */
+
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_code_alloc_lock(state);
+
+ tlsf_realloc(state->tlsf, ptr, size);
+
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_code_alloc_unlock(state);
}
-static s32 c_function_wrapper(struct lightrec_state *state, s32 cycles_delta,
- void (*f)(struct lightrec_state *, union code),
- union code op)
+static void lightrec_free_code(struct lightrec_state *state, void *ptr)
{
- state->current_cycle = state->target_cycle - cycles_delta;
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_code_alloc_lock(state);
- (*f)(state, op);
+ tlsf_free(state->tlsf, ptr);
- return state->target_cycle - state->current_cycle;
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_code_alloc_unlock(state);
}
-static struct block * generate_wrapper(struct lightrec_state *state,
- void *f, bool generic)
+static char lightning_code_data[0x80000];
+
+static void * lightrec_emit_code(struct lightrec_state *state,
+ const struct block *block,
+ jit_state_t *_jit, unsigned int *size)
+{
+ bool has_code_buffer = ENABLE_CODE_BUFFER && state->tlsf;
+ jit_word_t code_size, new_code_size;
+ void *code;
+
+ jit_realize();
+
+ if (ENABLE_DISASSEMBLER)
+ jit_set_data(lightning_code_data, sizeof(lightning_code_data), 0);
+ else
+ jit_set_data(NULL, 0, JIT_DISABLE_DATA | JIT_DISABLE_NOTE);
+
+ if (has_code_buffer) {
+ jit_get_code(&code_size);
+ code = lightrec_alloc_code(state, (size_t) code_size);
+
+ if (!code) {
+ if (ENABLE_THREADED_COMPILER) {
+ /* If we're using the threaded compiler, return
+ * an allocation error here. The threaded
+ * compiler will then empty its job queue and
+ * request a code flush using the reaper. */
+ return NULL;
+ }
+
+ /* Remove outdated blocks, and try again */
+ lightrec_remove_outdated_blocks(state->block_cache, block);
+
+ pr_debug("Re-try to alloc %zu bytes...\n", code_size);
+
+ code = lightrec_alloc_code(state, code_size);
+ if (!code) {
+ pr_err("Could not alloc even after removing old blocks!\n");
+ return NULL;
+ }
+ }
+
+ jit_set_code(code, code_size);
+ }
+
+ code = jit_emit();
+
+ jit_get_code(&new_code_size);
+ lightrec_register(MEM_FOR_CODE, new_code_size);
+
+ if (has_code_buffer) {
+ lightrec_realloc_code(state, code, (size_t) new_code_size);
+
+ pr_debug("Creating code block at address 0x%" PRIxPTR ", "
+ "code size: %" PRIuPTR " new: %" PRIuPTR "\n",
+ (uintptr_t) code, code_size, new_code_size);
+ }
+
+ *size = (unsigned int) new_code_size;
+
+ if (state->ops.code_inv)
+ state->ops.code_inv(code, new_code_size);
+
+ return code;
+}
+
+static struct block * generate_wrapper(struct lightrec_state *state)
{
struct block *block;
jit_state_t *_jit;
unsigned int i;
- int stack_ptr;
- jit_word_t code_size;
- jit_node_t *to_tramp, *to_fn_epilog;
+ jit_node_t *addr[C_WRAPPERS_COUNT - 1];
+ jit_node_t *to_end[C_WRAPPERS_COUNT - 1];
+ u8 tmp = JIT_R1;
+
+#ifdef __sh__
+ /* On SH, GBR-relative loads target the r0 register.
+ * Use it as the temporary register to factorize the move to
+ * JIT_R1. */
+ if (LIGHTREC_REG_STATE == _GBR)
+ tmp = _R0;
+#endif
block = lightrec_malloc(state, MEM_FOR_IR, sizeof(*block));
if (!block)
/* Wrapper entry point */
jit_prolog();
+ jit_tramp(256);
- stack_ptr = jit_allocai(sizeof(uintptr_t) * NUM_TEMPS);
-
- for (i = 0; i < NUM_TEMPS; i++)
- jit_stxi(stack_ptr + i * sizeof(uintptr_t), JIT_FP, JIT_R(i));
-
- /* Jump to the trampoline */
- to_tramp = jit_jmpi();
+ /* Add entry points */
+ for (i = C_WRAPPERS_COUNT - 1; i > 0; i--) {
+ jit_ldxi(tmp, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, c_wrappers[i]));
+ to_end[i - 1] = jit_b();
+ addr[i - 1] = jit_indirect();
+ }
- /* The trampoline will jump back here */
- to_fn_epilog = jit_label();
+ jit_ldxi(tmp, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, c_wrappers[0]));
- for (i = 0; i < NUM_TEMPS; i++)
- jit_ldxi(JIT_R(i), JIT_FP, stack_ptr + i * sizeof(uintptr_t));
+ for (i = 0; i < C_WRAPPERS_COUNT - 1; i++)
+ jit_patch(to_end[i]);
+ jit_movr(JIT_R1, tmp);
- jit_ret();
jit_epilog();
-
- /* Trampoline entry point.
- * The sole purpose of the trampoline is to cheese Lightning not to
- * save/restore the callee-saved register LIGHTREC_REG_CYCLE, since we
- * do want to return to the caller with this register modified. */
jit_prolog();
- jit_tramp(256);
- jit_patch(to_tramp);
+
+ /* Save all temporaries on stack */
+ for (i = 0; i < NUM_TEMPS; i++) {
+ if (i + FIRST_TEMP != 1) {
+ jit_stxi(offsetof(struct lightrec_state, wrapper_regs[i]),
+ LIGHTREC_REG_STATE, JIT_R(i + FIRST_TEMP));
+ }
+ }
+
+ jit_getarg(JIT_R2, jit_arg());
jit_prepare();
jit_pushargr(LIGHTREC_REG_STATE);
- jit_pushargr(LIGHTREC_REG_CYCLE);
- jit_pushargi((uintptr_t)f);
- jit_pushargr(JIT_R0);
- if (generic) {
- jit_pushargr(JIT_R1);
- jit_finishi(c_generic_function_wrapper);
- } else {
- jit_finishi(c_function_wrapper);
+ jit_pushargr(JIT_R2);
+
+ jit_ldxi_ui(JIT_R2, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, target_cycle));
+
+ /* state->current_cycle = state->target_cycle - delta; */
+ jit_subr(LIGHTREC_REG_CYCLE, JIT_R2, LIGHTREC_REG_CYCLE);
+ jit_stxi_i(offsetof(struct lightrec_state, current_cycle),
+ LIGHTREC_REG_STATE, LIGHTREC_REG_CYCLE);
+
+ /* Call the wrapper function */
+ jit_finishr(JIT_R1);
+
+ /* delta = state->target_cycle - state->current_cycle */;
+ jit_ldxi_ui(LIGHTREC_REG_CYCLE, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, current_cycle));
+ jit_ldxi_ui(JIT_R1, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, target_cycle));
+ jit_subr(LIGHTREC_REG_CYCLE, JIT_R1, LIGHTREC_REG_CYCLE);
+
+ /* Restore temporaries from stack */
+ for (i = 0; i < NUM_TEMPS; i++) {
+ if (i + FIRST_TEMP != 1) {
+ jit_ldxi(JIT_R(i + FIRST_TEMP), LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, wrapper_regs[i]));
+ }
}
-#if __WORDSIZE == 64
- jit_retval_i(LIGHTREC_REG_CYCLE);
-#else
- jit_retval(LIGHTREC_REG_CYCLE);
-#endif
-
- jit_patch_at(jit_jmpi(), to_fn_epilog);
+ jit_ret();
jit_epilog();
- block->state = state;
block->_jit = _jit;
- block->function = jit_emit();
block->opcode_list = NULL;
- block->flags = 0;
+ block->flags = BLOCK_NO_OPCODE_LIST;
block->nb_ops = 0;
- jit_get_code(&code_size);
- lightrec_register(MEM_FOR_CODE, code_size);
+ block->function = lightrec_emit_code(state, block, _jit,
+ &block->code_size);
+ if (!block->function)
+ goto err_free_block;
+
+ state->wrappers_eps[C_WRAPPERS_COUNT - 1] = block->function;
- block->code_size = code_size;
+ for (i = 0; i < C_WRAPPERS_COUNT - 1; i++)
+ state->wrappers_eps[i] = jit_address(addr[i]);
if (ENABLE_DISASSEMBLER) {
pr_debug("Wrapper block:\n");
return NULL;
}
+static u32 lightrec_memset(struct lightrec_state *state)
+{
+ u32 kunseg_pc = kunseg(state->regs.gpr[4]);
+ void *host;
+ const struct lightrec_mem_map *map = lightrec_get_map(state, &host, kunseg_pc);
+ u32 length = state->regs.gpr[5] * 4;
+
+ if (!map) {
+ pr_err("Unable to find memory map for memset target address "
+ "0x%x\n", kunseg_pc);
+ return 0;
+ }
+
+ pr_debug("Calling host memset, PC 0x%x (host address 0x%" PRIxPTR ") for %u bytes\n",
+ kunseg_pc, (uintptr_t)host, length);
+ memset(host, 0, length);
+
+ if (!(state->opt_flags & LIGHTREC_OPT_INV_DMA_ONLY))
+ lightrec_invalidate_map(state, map, kunseg_pc, length);
+
+ /* Rough estimation of the number of cycles consumed */
+ return 8 + 5 * (length + 3 / 4);
+}
+
+static u32 lightrec_check_load_delay(struct lightrec_state *state, u32 pc, u8 reg)
+{
+ struct block *block;
+ union code first_op;
+
+ first_op = lightrec_read_opcode(state, pc);
+
+ if (likely(!opcode_reads_register(first_op, reg))) {
+ state->regs.gpr[reg] = state->temp_reg;
+ } else {
+ block = lightrec_get_block(state, pc);
+ if (unlikely(!block)) {
+ pr_err("Unable to get block at PC 0x%08x\n", pc);
+ lightrec_set_exit_flags(state, LIGHTREC_EXIT_SEGFAULT);
+ pc = 0;
+ } else {
+ pc = lightrec_handle_load_delay(state, block, pc, reg);
+ }
+ }
+
+ return pc;
+}
+
+static void update_cycle_counter_before_c(jit_state_t *_jit)
+{
+ /* update state->current_cycle */
+ jit_ldxi_i(JIT_R2, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, target_cycle));
+ jit_subr(JIT_R1, JIT_R2, LIGHTREC_REG_CYCLE);
+ jit_stxi_i(offsetof(struct lightrec_state, current_cycle),
+ LIGHTREC_REG_STATE, JIT_R1);
+}
+
+static void update_cycle_counter_after_c(jit_state_t *_jit)
+{
+ /* Recalc the delta */
+ jit_ldxi_i(JIT_R1, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, current_cycle));
+ jit_ldxi_i(JIT_R2, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, target_cycle));
+ jit_subr(LIGHTREC_REG_CYCLE, JIT_R2, JIT_R1);
+}
+
+static void sync_next_pc(jit_state_t *_jit)
+{
+ if (lightrec_store_next_pc()) {
+ jit_ldxi_ui(JIT_V0, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, next_pc));
+ }
+}
+
static struct block * generate_dispatcher(struct lightrec_state *state)
{
struct block *block;
jit_state_t *_jit;
- jit_node_t *to_end, *to_end2, *to_c, *loop, *addr, *addr2;
+ jit_node_t *to_end, *loop, *addr, *addr2, *addr3, *addr4, *addr5, *jmp, *jmp2;
unsigned int i;
- u32 offset, ram_len;
- jit_word_t code_size;
+ u32 offset;
block = lightrec_malloc(state, MEM_FOR_IR, sizeof(*block));
if (!block)
jit_prolog();
jit_frame(256);
- jit_getarg(JIT_R0, jit_arg());
-#if __WORDSIZE == 64
+ jit_getarg(LIGHTREC_REG_STATE, jit_arg());
+ jit_getarg(JIT_V0, jit_arg());
+ jit_getarg(JIT_V1, jit_arg());
jit_getarg_i(LIGHTREC_REG_CYCLE, jit_arg());
-#else
- jit_getarg(LIGHTREC_REG_CYCLE, jit_arg());
-#endif
/* Force all callee-saved registers to be pushed on the stack */
for (i = 0; i < NUM_REGS; i++)
- jit_movr(JIT_V(i), JIT_V(i));
-
- /* Pass lightrec_state structure to blocks, using the last callee-saved
- * register that Lightning provides */
- jit_movi(LIGHTREC_REG_STATE, (intptr_t) state);
+ jit_movr(JIT_V(i + FIRST_REG), JIT_V(i + FIRST_REG));
loop = jit_label();
/* Call the block's code */
- jit_jmpr(JIT_R0);
+ jit_jmpr(JIT_V1);
+
+ if (OPT_REPLACE_MEMSET) {
+ /* Blocks will jump here when they need to call
+ * lightrec_memset() */
+ addr3 = jit_indirect();
+
+ jit_movr(JIT_V1, LIGHTREC_REG_CYCLE);
+
+ jit_prepare();
+ jit_pushargr(LIGHTREC_REG_STATE);
+
+ jit_finishi(lightrec_memset);
+ jit_retval(LIGHTREC_REG_CYCLE);
+
+ jit_ldxi_ui(JIT_V0, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, regs.gpr[31]));
+ jit_subr(LIGHTREC_REG_CYCLE, JIT_V1, LIGHTREC_REG_CYCLE);
+
+ if (OPT_DETECT_IMPOSSIBLE_BRANCHES || OPT_HANDLE_LOAD_DELAYS)
+ jmp = jit_b();
+ }
+
+ if (OPT_DETECT_IMPOSSIBLE_BRANCHES) {
+ /* Blocks will jump here when they reach a branch that should
+ * be executed with the interpreter, passing the branch's PC
+ * in JIT_V0 and the address of the block in JIT_V1. */
+ addr4 = jit_indirect();
+
+ sync_next_pc(_jit);
+ update_cycle_counter_before_c(_jit);
+
+ jit_prepare();
+ jit_pushargr(LIGHTREC_REG_STATE);
+ jit_pushargr(JIT_V1);
+ jit_pushargr(JIT_V0);
+ jit_finishi(lightrec_emulate_block);
+
+ jit_retval(JIT_V0);
+
+ update_cycle_counter_after_c(_jit);
+
+ if (OPT_HANDLE_LOAD_DELAYS)
+ jmp2 = jit_b();
+
+ }
+
+ if (OPT_HANDLE_LOAD_DELAYS) {
+ /* Blocks will jump here when they reach a branch with a load
+ * opcode in its delay slot. The delay slot has already been
+ * executed; the load value is in (state->temp_reg), and the
+ * register number is in JIT_V1.
+ * Jump to a C function which will evaluate the branch target's
+ * first opcode, to make sure that it does not read the register
+ * in question; and if it does, handle it accordingly. */
+ addr5 = jit_indirect();
+
+ sync_next_pc(_jit);
+ update_cycle_counter_before_c(_jit);
+
+ jit_prepare();
+ jit_pushargr(LIGHTREC_REG_STATE);
+ jit_pushargr(JIT_V0);
+ jit_pushargr(JIT_V1);
+ jit_finishi(lightrec_check_load_delay);
+
+ jit_retval(JIT_V0);
+
+ update_cycle_counter_after_c(_jit);
+ }
/* The block will jump here, with the number of cycles remaining in
* LIGHTREC_REG_CYCLE */
addr2 = jit_indirect();
+ sync_next_pc(_jit);
+
+ if (OPT_HANDLE_LOAD_DELAYS && OPT_DETECT_IMPOSSIBLE_BRANCHES)
+ jit_patch(jmp2);
+
+ if (OPT_REPLACE_MEMSET
+ && (OPT_DETECT_IMPOSSIBLE_BRANCHES || OPT_HANDLE_LOAD_DELAYS)) {
+ jit_patch(jmp);
+ }
+
+ /* Store back the next PC to the lightrec_state structure */
+ offset = offsetof(struct lightrec_state, curr_pc);
+ jit_stxi_i(offset, LIGHTREC_REG_STATE, JIT_V0);
+
/* Jump to end if state->target_cycle < state->current_cycle */
to_end = jit_blei(LIGHTREC_REG_CYCLE, 0);
/* Convert next PC to KUNSEG and avoid mirrors */
- ram_len = state->maps[PSX_MAP_KERNEL_USER_RAM].length;
- jit_andi(JIT_R0, JIT_V0, 0x10000000 | (ram_len - 1));
- to_c = jit_bgei(JIT_R0, ram_len);
-
- /* Fast path: code is running from RAM, use the code LUT */
-#if __WORDSIZE == 64
- jit_lshi(JIT_R0, JIT_R0, 1);
-#endif
- jit_addr(JIT_R0, JIT_R0, LIGHTREC_REG_STATE);
- jit_ldxi(JIT_R0, JIT_R0, offsetof(struct lightrec_state, code_lut));
+ jit_andi(JIT_V1, JIT_V0, 0x10000000 | (RAM_SIZE - 1));
+ jit_rshi_u(JIT_R1, JIT_V1, 28);
+ jit_andi(JIT_R2, JIT_V0, BIOS_SIZE - 1);
+ jit_addi(JIT_R2, JIT_R2, RAM_SIZE);
+ jit_movnr(JIT_V1, JIT_R2, JIT_R1);
+
+ /* If possible, use the code LUT */
+ if (!lut_is_32bit(state))
+ jit_lshi(JIT_V1, JIT_V1, 1);
+ jit_add_state(JIT_V1, JIT_V1);
+
+ offset = offsetof(struct lightrec_state, code_lut);
+ if (lut_is_32bit(state))
+ jit_ldxi_ui(JIT_V1, JIT_V1, offset);
+ else
+ jit_ldxi(JIT_V1, JIT_V1, offset);
/* If we get non-NULL, loop */
- jit_patch_at(jit_bnei(JIT_R0, 0), loop);
-
- /* Slow path: call C function get_next_block_func() */
- jit_patch(to_c);
-
- if (ENABLE_FIRST_PASS) {
- /* We may call the interpreter - update state->current_cycle */
- jit_ldxi_i(JIT_R2, LIGHTREC_REG_STATE,
- offsetof(struct lightrec_state, target_cycle));
- jit_subr(JIT_R1, JIT_R2, LIGHTREC_REG_CYCLE);
- jit_stxi_i(offsetof(struct lightrec_state, current_cycle),
- LIGHTREC_REG_STATE, JIT_R1);
- }
+ jit_patch_at(jit_bnei(JIT_V1, 0), loop);
/* The code LUT will be set to this address when the block at the target
* PC has been preprocessed but not yet compiled by the threaded
* recompiler */
addr = jit_indirect();
- /* Get the next block */
+ /* Slow path: call C function get_next_block_func() */
+
+ if (ENABLE_FIRST_PASS || OPT_DETECT_IMPOSSIBLE_BRANCHES) {
+ /* We may call the interpreter - update state->current_cycle */
+ update_cycle_counter_before_c(_jit);
+ }
+
jit_prepare();
jit_pushargr(LIGHTREC_REG_STATE);
jit_pushargr(JIT_V0);
+
+ /* Save the cycles register if needed */
+ if (!(ENABLE_FIRST_PASS || OPT_DETECT_IMPOSSIBLE_BRANCHES))
+ jit_movr(JIT_V0, LIGHTREC_REG_CYCLE);
+
+ /* Get the next block */
jit_finishi(&get_next_block_func);
- jit_retval(JIT_R0);
+ jit_retval(JIT_V1);
- if (ENABLE_FIRST_PASS) {
+ if (ENABLE_FIRST_PASS || OPT_DETECT_IMPOSSIBLE_BRANCHES) {
/* The interpreter may have updated state->current_cycle and
* state->target_cycle - recalc the delta */
- jit_ldxi_i(JIT_R1, LIGHTREC_REG_STATE,
- offsetof(struct lightrec_state, current_cycle));
- jit_ldxi_i(JIT_R2, LIGHTREC_REG_STATE,
- offsetof(struct lightrec_state, target_cycle));
- jit_subr(LIGHTREC_REG_CYCLE, JIT_R2, JIT_R1);
+ update_cycle_counter_after_c(_jit);
+ } else {
+ jit_movr(LIGHTREC_REG_CYCLE, JIT_V0);
}
- /* If we get non-NULL, loop */
- jit_patch_at(jit_bnei(JIT_R0, 0), loop);
+ /* Reset JIT_V0 to the next PC */
+ jit_ldxi_ui(JIT_V0, LIGHTREC_REG_STATE,
+ offsetof(struct lightrec_state, curr_pc));
- to_end2 = jit_jmpi();
+ /* If we get non-NULL, loop */
+ jit_patch_at(jit_bnei(JIT_V1, 0), loop);
/* When exiting, the recompiled code will jump to that address */
jit_note(__FILE__, __LINE__);
jit_patch(to_end);
- /* Store back the next_pc to the lightrec_state structure */
- offset = offsetof(struct lightrec_state, next_pc);
- jit_stxi_i(offset, LIGHTREC_REG_STATE, JIT_V0);
-
- jit_patch(to_end2);
-
jit_retr(LIGHTREC_REG_CYCLE);
jit_epilog();
- block->state = state;
block->_jit = _jit;
- block->function = jit_emit();
block->opcode_list = NULL;
- block->flags = 0;
+ block->flags = BLOCK_NO_OPCODE_LIST;
block->nb_ops = 0;
- jit_get_code(&code_size);
- lightrec_register(MEM_FOR_CODE, code_size);
-
- block->code_size = code_size;
+ block->function = lightrec_emit_code(state, block, _jit,
+ &block->code_size);
+ if (!block->function)
+ goto err_free_block;
state->eob_wrapper_func = jit_address(addr2);
+ if (OPT_DETECT_IMPOSSIBLE_BRANCHES)
+ state->interpreter_func = jit_address(addr4);
+ if (OPT_HANDLE_LOAD_DELAYS)
+ state->ds_check_func = jit_address(addr5);
+ if (OPT_REPLACE_MEMSET)
+ state->memset_func = jit_address(addr3);
state->get_next_block = jit_address(addr);
if (ENABLE_DISASSEMBLER) {
union code lightrec_read_opcode(struct lightrec_state *state, u32 pc)
{
- u32 addr, kunseg_pc = kunseg(pc);
- const u32 *code;
- const struct lightrec_mem_map *map = lightrec_get_map(state, kunseg_pc);
+ void *host = NULL;
- addr = kunseg_pc - map->pc;
+ lightrec_get_map(state, &host, kunseg(pc));
- while (map->mirror_of)
- map = map->mirror_of;
+ const u32 *code = (u32 *)host;
+ return (union code) LE32TOH(*code);
+}
+
+unsigned int lightrec_cycles_of_opcode(const struct lightrec_state *state,
+ union code code)
+{
+ return state->cycles_per_op;
+}
+
+void lightrec_free_opcode_list(struct lightrec_state *state, struct opcode *ops)
+{
+ struct opcode_list *list = container_of(ops, struct opcode_list, ops);
+
+ lightrec_free(state, MEM_FOR_IR,
+ sizeof(*list) + list->nb_ops * sizeof(struct opcode),
+ list);
+}
+
+static unsigned int lightrec_get_mips_block_len(const u32 *src)
+{
+ unsigned int i;
+ union code c;
- code = map->address + addr;
+ for (i = 1; ; i++) {
+ c.opcode = LE32TOH(*src++);
- return (union code) *code;
+ if (is_syscall(c))
+ return i;
+
+ if (is_unconditional_jump(c))
+ return i + 1;
+ }
+}
+
+static struct opcode * lightrec_disassemble(struct lightrec_state *state,
+ const u32 *src, unsigned int *len)
+{
+ struct opcode_list *list;
+ unsigned int i, length;
+
+ length = lightrec_get_mips_block_len(src);
+
+ list = lightrec_malloc(state, MEM_FOR_IR,
+ sizeof(*list) + sizeof(struct opcode) * length);
+ if (!list) {
+ pr_err("Unable to allocate memory\n");
+ return NULL;
+ }
+
+ list->nb_ops = (u16) length;
+
+ for (i = 0; i < length; i++) {
+ list->ops[i].opcode = LE32TOH(src[i]);
+ list->ops[i].flags = 0;
+ }
+
+ *len = length * sizeof(u32);
+
+ return list->ops;
}
static struct block * lightrec_precompile_block(struct lightrec_state *state,
{
struct opcode *list;
struct block *block;
- const u32 *code;
- u32 addr, kunseg_pc = kunseg(pc);
- const struct lightrec_mem_map *map = lightrec_get_map(state, kunseg_pc);
+ void *host, *addr;
+ const struct lightrec_mem_map *map = lightrec_get_map(state, &host, kunseg(pc));
+ const u32 *code = (u32 *) host;
unsigned int length;
+ bool fully_tagged;
+ u8 block_flags = 0;
if (!map)
return NULL;
- addr = kunseg_pc - map->pc;
-
- while (map->mirror_of)
- map = map->mirror_of;
-
- code = map->address + addr;
-
block = lightrec_malloc(state, MEM_FOR_IR, sizeof(*block));
if (!block) {
pr_err("Unable to recompile block: Out of memory\n");
}
block->pc = pc;
- block->state = state;
block->_jit = NULL;
block->function = NULL;
block->opcode_list = list;
- block->map = map;
+ block->code = code;
block->next = NULL;
block->flags = 0;
block->code_size = 0;
-#if ENABLE_THREADED_COMPILER
- block->op_list_freed = (atomic_flag)ATOMIC_FLAG_INIT;
-#endif
+ block->precompile_date = state->current_cycle;
block->nb_ops = length / sizeof(u32);
- lightrec_optimize(block);
+ lightrec_optimize(state, block);
length = block->nb_ops * sizeof(u32);
lightrec_register(MEM_FOR_MIPS_CODE, length);
if (ENABLE_DISASSEMBLER) {
- pr_debug("Disassembled block at PC: 0x%x\n", block->pc);
- lightrec_print_disassembly(block, code, length);
+ pr_debug("Disassembled block at PC: 0x%08x\n", block->pc);
+ lightrec_print_disassembly(block, code);
}
- pr_debug("Block size: %lu opcodes\n", block->nb_ops);
+ pr_debug("Block size: %hu opcodes\n", block->nb_ops);
+
+ fully_tagged = lightrec_block_is_fully_tagged(block);
+ if (fully_tagged)
+ block_flags |= BLOCK_FULLY_TAGGED;
- /* If the first opcode is an 'impossible' branch, never compile the
- * block */
- if (list->flags & LIGHTREC_EMULATE_BRANCH)
- block->flags |= BLOCK_NEVER_COMPILE;
+ if (block_flags)
+ block_set_flags(block, block_flags);
block->hash = lightrec_calculate_block_hash(block);
- pr_debug("Recompile count: %u\n", state->nb_precompile++);
+ if (OPT_REPLACE_MEMSET && block_has_flag(block, BLOCK_IS_MEMSET))
+ addr = state->memset_func;
+ else
+ addr = state->get_next_block;
+ lut_write(state, lut_offset(pc), addr);
+
+ pr_debug("Blocks created: %u\n", ++state->nb_precompile);
return block;
}
-static bool lightrec_block_is_fully_tagged(struct block *block)
+static bool lightrec_block_is_fully_tagged(const struct block *block)
{
- struct opcode *op;
+ const struct opcode *op;
+ unsigned int i;
- for (op = block->opcode_list; op; op = op->next) {
- /* Verify that all load/stores of the opcode list
- * Check all loads/stores of the opcode list and mark the
+ for (i = 0; i < block->nb_ops; i++) {
+ op = &block->opcode_list[i];
+
+ /* If we have one branch that must be emulated, we cannot trash
+ * the opcode list. */
+ if (should_emulate(op))
+ return false;
+
+ /* Check all loads/stores of the opcode list and mark the
* block as fully compiled if they all have been tagged. */
switch (op->c.i.op) {
case OP_LB:
case OP_SWR:
case OP_LWC2:
case OP_SWC2:
- if (!(op->flags & (LIGHTREC_DIRECT_IO |
- LIGHTREC_HW_IO)))
+ if (!LIGHTREC_FLAGS_GET_IO_MODE(op->flags))
return false;
- default: /* fall-through */
+ fallthrough;
+ default:
continue;
}
}
return true;
}
-static void lightrec_reap_block(void *data)
+static void lightrec_reap_block(struct lightrec_state *state, void *data)
{
struct block *block = data;
pr_debug("Reap dead block at PC 0x%08x\n", block->pc);
- lightrec_free_block(block);
+ lightrec_unregister_block(state->block_cache, block);
+ lightrec_free_block(state, block);
}
-static void lightrec_reap_jit(void *data)
+static void lightrec_reap_jit(struct lightrec_state *state, void *data)
{
_jit_destroy_state(data);
}
-int lightrec_compile_block(struct block *block)
+static void lightrec_free_function(struct lightrec_state *state, void *fn)
+{
+ if (ENABLE_CODE_BUFFER && state->tlsf) {
+ pr_debug("Freeing code block at 0x%" PRIxPTR "\n", (uintptr_t) fn);
+ lightrec_free_code(state, fn);
+ }
+}
+
+static void lightrec_reap_function(struct lightrec_state *state, void *data)
+{
+ lightrec_free_function(state, data);
+}
+
+static void lightrec_reap_opcode_list(struct lightrec_state *state, void *data)
+{
+ lightrec_free_opcode_list(state, data);
+}
+
+int lightrec_compile_block(struct lightrec_cstate *cstate,
+ struct block *block)
{
- struct lightrec_state *state = block->state;
+ struct lightrec_state *state = cstate->state;
struct lightrec_branch_target *target;
- bool op_list_freed = false, fully_tagged = false;
+ bool fully_tagged = false;
struct block *block2;
struct opcode *elm;
jit_state_t *_jit, *oldjit;
jit_node_t *start_of_block;
bool skip_next = false;
- jit_word_t code_size;
+ void *old_fn, *new_fn;
+ size_t old_code_size;
unsigned int i, j;
- u32 next_pc, offset;
+ u8 old_flags;
+ u32 offset;
fully_tagged = lightrec_block_is_fully_tagged(block);
if (fully_tagged)
- block->flags |= BLOCK_FULLY_TAGGED;
+ block_set_flags(block, BLOCK_FULLY_TAGGED);
_jit = jit_new_state();
if (!_jit)
return -ENOMEM;
oldjit = block->_jit;
+ old_fn = block->function;
+ old_code_size = block->code_size;
block->_jit = _jit;
- lightrec_regcache_reset(state->reg_cache);
- state->cycles = 0;
- state->nb_branches = 0;
- state->nb_local_branches = 0;
- state->nb_targets = 0;
+ lightrec_regcache_reset(cstate->reg_cache);
+
+ if (OPT_PRELOAD_PC && (block->flags & BLOCK_PRELOAD_PC))
+ lightrec_preload_pc(cstate->reg_cache, _jit);
+
+ cstate->cycles = 0;
+ cstate->nb_local_branches = 0;
+ cstate->nb_targets = 0;
+ cstate->no_load_delay = false;
jit_prolog();
jit_tramp(256);
start_of_block = jit_label();
- for (elm = block->opcode_list; elm; elm = elm->next) {
- next_pc = block->pc + elm->offset * sizeof(u32);
+ for (i = 0; i < block->nb_ops; i++) {
+ elm = &block->opcode_list[i];
if (skip_next) {
skip_next = false;
continue;
}
- state->cycles += lightrec_cycles_of_opcode(elm->c);
-
- if (elm->flags & LIGHTREC_EMULATE_BRANCH) {
+ if (should_emulate(elm)) {
pr_debug("Branch at offset 0x%x will be emulated\n",
- elm->offset << 2);
- lightrec_emit_eob(block, elm, next_pc);
- skip_next = !(elm->flags & LIGHTREC_NO_DS);
- } else if (elm->opcode) {
- lightrec_rec_opcode(block, elm, next_pc);
- skip_next = has_delay_slot(elm->c) &&
- !(elm->flags & LIGHTREC_NO_DS);
+ i << 2);
+
+ lightrec_emit_jump_to_interpreter(cstate, block, i);
+ skip_next = !op_flag_no_ds(elm->flags);
+ } else {
+ lightrec_rec_opcode(cstate, block, i);
+ skip_next = !op_flag_no_ds(elm->flags) && has_delay_slot(elm->c);
#if _WIN32
/* FIXME: GNU Lightning on Windows seems to use our
* mapped registers as temporaries. Until the actual bug
* is found and fixed, unconditionally mark our
* registers as live here. */
- lightrec_regcache_mark_live(state->reg_cache, _jit);
+ lightrec_regcache_mark_live(cstate->reg_cache, _jit);
#endif
}
- }
- for (i = 0; i < state->nb_branches; i++)
- jit_patch(state->branches[i]);
+ cstate->cycles += lightrec_cycles_of_opcode(state, elm->c);
+ }
- for (i = 0; i < state->nb_local_branches; i++) {
- struct lightrec_branch *branch = &state->local_branches[i];
+ for (i = 0; i < cstate->nb_local_branches; i++) {
+ struct lightrec_branch *branch = &cstate->local_branches[i];
pr_debug("Patch local branch to offset 0x%x\n",
branch->target << 2);
continue;
}
- for (j = 0; j < state->nb_targets; j++) {
- if (state->targets[j].offset == branch->target) {
+ for (j = 0; j < cstate->nb_targets; j++) {
+ if (cstate->targets[j].offset == branch->target) {
jit_patch_at(branch->branch,
- state->targets[j].label);
+ cstate->targets[j].label);
break;
}
}
- if (j == state->nb_targets)
+ if (j == cstate->nb_targets)
pr_err("Unable to find branch target\n");
}
- jit_ldxi(JIT_R0, LIGHTREC_REG_STATE,
- offsetof(struct lightrec_state, eob_wrapper_func));
-
- jit_jmpr(JIT_R0);
-
jit_ret();
jit_epilog();
- block->function = jit_emit();
- block->flags &= ~BLOCK_SHOULD_RECOMPILE;
+ new_fn = lightrec_emit_code(state, block, _jit, &block->code_size);
+ if (!new_fn) {
+ if (!ENABLE_THREADED_COMPILER)
+ pr_err("Unable to compile block!\n");
+ block->_jit = oldjit;
+ jit_clear_state();
+ _jit_destroy_state(_jit);
+ return -ENOMEM;
+ }
- /* Add compiled function to the LUT */
- state->code_lut[lut_offset(block->pc)] = block->function;
+ /* Pause the reaper, because lightrec_reset_lut_offset() may try to set
+ * the old block->function pointer to the code LUT. */
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_reaper_pause(state->reaper);
- /* Fill code LUT with the block's entry points */
- for (i = 0; i < state->nb_targets; i++) {
- target = &state->targets[i];
+ block->function = new_fn;
+ block_clear_flags(block, BLOCK_SHOULD_RECOMPILE);
- if (target->offset) {
- offset = lut_offset(block->pc) + target->offset;
- state->code_lut[offset] = jit_address(target->label);
- }
- }
+ /* Add compiled function to the LUT */
+ lut_write(state, lut_offset(block->pc), block->function);
+
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_reaper_continue(state->reaper);
/* Detect old blocks that have been covered by the new one */
- for (i = 0; i < state->nb_targets; i++) {
- target = &state->targets[i];
+ for (i = 0; i < cstate->nb_targets; i++) {
+ target = &cstate->targets[i];
if (!target->offset)
continue;
offset = block->pc + target->offset * sizeof(u32);
+
+ /* Pause the reaper while we search for the block until we set
+ * the BLOCK_IS_DEAD flag, otherwise the block may be removed
+ * under our feet. */
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_reaper_pause(state->reaper);
+
block2 = lightrec_find_block(state->block_cache, offset);
if (block2) {
/* No need to check if block2 is compilable - it must
* be, otherwise block wouldn't be compilable either */
- block2->flags |= BLOCK_IS_DEAD;
+ /* Set the "block dead" flag to prevent the dynarec from
+ * recompiling this block */
+ old_flags = block_set_flags(block2, BLOCK_IS_DEAD);
+ }
+
+ if (ENABLE_THREADED_COMPILER) {
+ lightrec_reaper_continue(state->reaper);
+ /* If block2 was pending for compilation, cancel it.
+ * If it's being compiled right now, wait until it
+ * finishes. */
+ if (block2)
+ lightrec_recompiler_remove(state->rec, block2);
+ }
+
+ /* We know from now on that block2 (if present) isn't going to
+ * be compiled. We can override the LUT entry with our new
+ * block's entry point. */
+ offset = lut_offset(block->pc) + target->offset;
+ lut_write(state, offset, jit_address(target->label));
+
+ if (block2) {
pr_debug("Reap block 0x%08x as it's covered by block "
"0x%08x\n", block2->pc, block->pc);
- lightrec_unregister_block(state->block_cache, block2);
-
- if (ENABLE_THREADED_COMPILER) {
- lightrec_recompiler_remove(state->rec, block2);
+ /* Finally, reap the block. */
+ if (!ENABLE_THREADED_COMPILER) {
+ lightrec_unregister_block(state->block_cache, block2);
+ lightrec_free_block(state, block2);
+ } else if (!(old_flags & BLOCK_IS_DEAD)) {
lightrec_reaper_add(state->reaper,
lightrec_reap_block,
block2);
- } else {
- lightrec_free_block(block2);
}
}
}
- jit_get_code(&code_size);
- lightrec_register(MEM_FOR_CODE, code_size);
-
- block->code_size = code_size;
-
if (ENABLE_DISASSEMBLER) {
- pr_debug("Compiling block at PC: 0x%x\n", block->pc);
+ pr_debug("Compiling block at PC: 0x%08x\n", block->pc);
jit_disassemble();
}
jit_clear_state();
-#if ENABLE_THREADED_COMPILER
if (fully_tagged)
- op_list_freed = atomic_flag_test_and_set(&block->op_list_freed);
-#endif
- if (fully_tagged && !op_list_freed) {
+ old_flags = block_set_flags(block, BLOCK_NO_OPCODE_LIST);
+
+ if (fully_tagged && !(old_flags & BLOCK_NO_OPCODE_LIST)) {
pr_debug("Block PC 0x%08x is fully tagged"
" - free opcode list\n", block->pc);
- lightrec_free_opcode_list(state, block->opcode_list);
- block->opcode_list = NULL;
+
+ if (ENABLE_THREADED_COMPILER) {
+ lightrec_reaper_add(state->reaper,
+ lightrec_reap_opcode_list,
+ block->opcode_list);
+ } else {
+ lightrec_free_opcode_list(state, block->opcode_list);
+ }
}
if (oldjit) {
pr_debug("Block 0x%08x recompiled, reaping old jit context.\n",
block->pc);
- if (ENABLE_THREADED_COMPILER)
+ if (ENABLE_THREADED_COMPILER) {
lightrec_reaper_add(state->reaper,
lightrec_reap_jit, oldjit);
- else
+ lightrec_reaper_add(state->reaper,
+ lightrec_reap_function, old_fn);
+ } else {
_jit_destroy_state(oldjit);
+ lightrec_free_function(state, old_fn);
+ }
+
+ lightrec_unregister(MEM_FOR_CODE, old_code_size);
}
+ pr_debug("Blocks compiled: %u\n", ++state->nb_compile);
+
return 0;
}
+static void lightrec_print_info(struct lightrec_state *state)
+{
+ if ((state->current_cycle & ~0xfffffff) != state->old_cycle_counter) {
+ pr_info("Lightrec RAM usage: IR %u KiB, CODE %u KiB, "
+ "MIPS %u KiB, TOTAL %u KiB, avg. IPI %f\n",
+ lightrec_get_mem_usage(MEM_FOR_IR) / 1024,
+ lightrec_get_mem_usage(MEM_FOR_CODE) / 1024,
+ lightrec_get_mem_usage(MEM_FOR_MIPS_CODE) / 1024,
+ lightrec_get_total_mem_usage() / 1024,
+ lightrec_get_average_ipi());
+ state->old_cycle_counter = state->current_cycle & ~0xfffffff;
+ }
+}
+
u32 lightrec_execute(struct lightrec_state *state, u32 pc, u32 target_cycle)
{
- s32 (*func)(void *, s32) = (void *)state->dispatcher->function;
+ s32 (*func)(struct lightrec_state *, u32, void *, s32) = (void *)state->dispatcher->function;
void *block_trace;
s32 cycles_delta;
target_cycle = UINT_MAX;
state->target_cycle = target_cycle;
+ state->curr_pc = pc;
block_trace = get_next_block_func(state, pc);
if (block_trace) {
cycles_delta = state->target_cycle - state->current_cycle;
- cycles_delta = (*func)(block_trace, cycles_delta);
+ cycles_delta = (*func)(state, state->curr_pc,
+ block_trace, cycles_delta);
state->current_cycle = state->target_cycle - cycles_delta;
}
if (ENABLE_THREADED_COMPILER)
lightrec_reaper_reap(state->reaper);
- return state->next_pc;
-}
+ if (LOG_LEVEL >= INFO_L)
+ lightrec_print_info(state);
-u32 lightrec_execute_one(struct lightrec_state *state, u32 pc)
-{
- return lightrec_execute(state, pc, state->current_cycle);
+ return state->curr_pc;
}
-u32 lightrec_run_interpreter(struct lightrec_state *state, u32 pc)
+u32 lightrec_run_interpreter(struct lightrec_state *state, u32 pc,
+ u32 target_cycle)
{
- struct block *block = lightrec_get_block(state, pc);
- if (!block)
- return 0;
+ struct block *block;
state->exit_flags = LIGHTREC_EXIT_NORMAL;
+ state->target_cycle = target_cycle;
+
+ do {
+ block = lightrec_get_block(state, pc);
+ if (!block)
+ break;
+
+ pc = lightrec_emulate_block(state, block, pc);
+
+ if (ENABLE_THREADED_COMPILER)
+ lightrec_reaper_reap(state->reaper);
+ } while (state->current_cycle < state->target_cycle);
+
+ if (LOG_LEVEL >= INFO_L)
+ lightrec_print_info(state);
- return lightrec_emulate_block(block, pc);
+ return pc;
}
-void lightrec_free_block(struct block *block)
+void lightrec_free_block(struct lightrec_state *state, struct block *block)
{
+ u8 old_flags;
+
lightrec_unregister(MEM_FOR_MIPS_CODE, block->nb_ops * sizeof(u32));
- if (block->opcode_list)
- lightrec_free_opcode_list(block->state, block->opcode_list);
+ old_flags = block_set_flags(block, BLOCK_NO_OPCODE_LIST);
+
+ if (!(old_flags & BLOCK_NO_OPCODE_LIST))
+ lightrec_free_opcode_list(state, block->opcode_list);
if (block->_jit)
_jit_destroy_state(block->_jit);
- lightrec_unregister(MEM_FOR_CODE, block->code_size);
- lightrec_free(block->state, MEM_FOR_IR, sizeof(*block), block);
+ if (block->function) {
+ lightrec_free_function(state, block->function);
+ lightrec_unregister(MEM_FOR_CODE, block->code_size);
+ }
+ lightrec_free(state, MEM_FOR_IR, sizeof(*block), block);
+}
+
+struct lightrec_cstate * lightrec_create_cstate(struct lightrec_state *state)
+{
+ struct lightrec_cstate *cstate;
+
+ cstate = lightrec_malloc(state, MEM_FOR_LIGHTREC, sizeof(*cstate));
+ if (!cstate)
+ return NULL;
+
+ cstate->reg_cache = lightrec_regcache_init(state);
+ if (!cstate->reg_cache) {
+ lightrec_free(state, MEM_FOR_LIGHTREC, sizeof(*cstate), cstate);
+ return NULL;
+ }
+
+ cstate->state = state;
+
+ return cstate;
+}
+
+void lightrec_free_cstate(struct lightrec_cstate *cstate)
+{
+ lightrec_free_regcache(cstate->reg_cache);
+ lightrec_free(cstate->state, MEM_FOR_LIGHTREC, sizeof(*cstate), cstate);
}
struct lightrec_state * lightrec_init(char *argv0,
size_t nb,
const struct lightrec_ops *ops)
{
+ const struct lightrec_mem_map *codebuf_map = &map[PSX_MAP_CODE_BUFFER];
struct lightrec_state *state;
+ uintptr_t addr;
+ void *tlsf = NULL;
+ bool with_32bit_lut = false;
+ size_t lut_size;
/* Sanity-check ops */
- if (!ops ||
- !ops->cop0_ops.mfc || !ops->cop0_ops.cfc || !ops->cop0_ops.mtc ||
- !ops->cop0_ops.ctc || !ops->cop0_ops.op ||
- !ops->cop2_ops.mfc || !ops->cop2_ops.cfc || !ops->cop2_ops.mtc ||
- !ops->cop2_ops.ctc || !ops->cop2_ops.op) {
+ if (!ops || !ops->cop2_op || !ops->enable_ram) {
pr_err("Missing callbacks in lightrec_ops structure\n");
return NULL;
}
+ if (ops->cop2_notify)
+ pr_debug("Optional cop2_notify callback in lightrec_ops\n");
+ else
+ pr_debug("No optional cop2_notify callback in lightrec_ops\n");
+
+ if (ENABLE_CODE_BUFFER && nb > PSX_MAP_CODE_BUFFER
+ && codebuf_map->address) {
+ tlsf = tlsf_create_with_pool(codebuf_map->address,
+ codebuf_map->length);
+ if (!tlsf) {
+ pr_err("Unable to initialize code buffer\n");
+ return NULL;
+ }
+
+ if (__WORDSIZE == 64) {
+ addr = (uintptr_t) codebuf_map->address + codebuf_map->length - 1;
+ with_32bit_lut = addr == (u32) addr;
+ }
+ }
+
+ if (with_32bit_lut)
+ lut_size = CODE_LUT_SIZE * 4;
+ else
+ lut_size = CODE_LUT_SIZE * sizeof(void *);
+
init_jit(argv0);
- state = calloc(1, sizeof(*state) +
- sizeof(*state->code_lut) * CODE_LUT_SIZE);
+ state = calloc(1, sizeof(*state) + lut_size);
if (!state)
goto err_finish_jit;
- lightrec_register(MEM_FOR_LIGHTREC, sizeof(*state) +
- sizeof(*state->code_lut) * CODE_LUT_SIZE);
+ lightrec_register(MEM_FOR_LIGHTREC, sizeof(*state) + lut_size);
-#if ENABLE_TINYMM
- state->tinymm = tinymm_init(malloc, free, 4096);
- if (!state->tinymm)
- goto err_free_state;
-#endif
+ state->tlsf = tlsf;
+ state->with_32bit_lut = with_32bit_lut;
+ state->in_delay_slot_n = 0xff;
+ state->cycles_per_op = 2;
state->block_cache = lightrec_blockcache_init(state);
if (!state->block_cache)
- goto err_free_tinymm;
-
- state->reg_cache = lightrec_regcache_init(state);
- if (!state->reg_cache)
- goto err_free_block_cache;
+ goto err_free_state;
if (ENABLE_THREADED_COMPILER) {
state->rec = lightrec_recompiler_init(state);
if (!state->rec)
- goto err_free_reg_cache;
+ goto err_free_block_cache;
state->reaper = lightrec_reaper_init(state);
if (!state->reaper)
goto err_free_recompiler;
+ } else {
+ state->cstate = lightrec_create_cstate(state);
+ if (!state->cstate)
+ goto err_free_block_cache;
}
state->nb_maps = nb;
if (!state->dispatcher)
goto err_free_reaper;
- state->rw_generic_wrapper = generate_wrapper(state,
- lightrec_rw_generic_cb,
- true);
- if (!state->rw_generic_wrapper)
+ state->c_wrapper_block = generate_wrapper(state);
+ if (!state->c_wrapper_block)
goto err_free_dispatcher;
- state->rw_wrapper = generate_wrapper(state, lightrec_rw_cb, false);
- if (!state->rw_wrapper)
- goto err_free_generic_rw_wrapper;
-
- state->mfc_wrapper = generate_wrapper(state, lightrec_mfc_cb, false);
- if (!state->mfc_wrapper)
- goto err_free_rw_wrapper;
-
- state->mtc_wrapper = generate_wrapper(state, lightrec_mtc_cb, false);
- if (!state->mtc_wrapper)
- goto err_free_mfc_wrapper;
-
- state->rfe_wrapper = generate_wrapper(state, lightrec_rfe_cb, false);
- if (!state->rfe_wrapper)
- goto err_free_mtc_wrapper;
-
- state->cp_wrapper = generate_wrapper(state, lightrec_cp_cb, false);
- if (!state->cp_wrapper)
- goto err_free_rfe_wrapper;
-
- state->syscall_wrapper = generate_wrapper(state, lightrec_syscall_cb,
- false);
- if (!state->syscall_wrapper)
- goto err_free_cp_wrapper;
-
- state->break_wrapper = generate_wrapper(state, lightrec_break_cb,
- false);
- if (!state->break_wrapper)
- goto err_free_syscall_wrapper;
-
- state->rw_generic_func = state->rw_generic_wrapper->function;
- state->rw_func = state->rw_wrapper->function;
- state->mfc_func = state->mfc_wrapper->function;
- state->mtc_func = state->mtc_wrapper->function;
- state->rfe_func = state->rfe_wrapper->function;
- state->cp_func = state->cp_wrapper->function;
- state->syscall_func = state->syscall_wrapper->function;
- state->break_func = state->break_wrapper->function;
+ state->c_wrappers[C_WRAPPER_RW] = lightrec_rw_cb;
+ state->c_wrappers[C_WRAPPER_RW_GENERIC] = lightrec_rw_generic_cb;
+ state->c_wrappers[C_WRAPPER_MFC] = lightrec_mfc_cb;
+ state->c_wrappers[C_WRAPPER_MTC] = lightrec_mtc_cb;
+ state->c_wrappers[C_WRAPPER_CP] = lightrec_cp_cb;
map = &state->maps[PSX_MAP_BIOS];
state->offset_bios = (uintptr_t)map->address - map->pc;
map = &state->maps[PSX_MAP_SCRATCH_PAD];
state->offset_scratch = (uintptr_t)map->address - map->pc;
+ map = &state->maps[PSX_MAP_HW_REGISTERS];
+ state->offset_io = (uintptr_t)map->address - map->pc;
+
map = &state->maps[PSX_MAP_KERNEL_USER_RAM];
state->offset_ram = (uintptr_t)map->address - map->pc;
state->maps[PSX_MAP_MIRROR3].address == map->address + 0x600000)
state->mirrors_mapped = true;
+ if (state->offset_bios == 0 &&
+ state->offset_scratch == 0 &&
+ state->offset_ram == 0 &&
+ state->offset_io == 0 &&
+ state->mirrors_mapped) {
+ pr_info("Memory map is perfect. Emitted code will be best.\n");
+ } else {
+ pr_info("Memory map is sub-par. Emitted code will be slow.\n");
+ }
+
+ if (state->with_32bit_lut)
+ pr_info("Using 32-bit LUT\n");
+
return state;
-err_free_syscall_wrapper:
- lightrec_free_block(state->syscall_wrapper);
-err_free_cp_wrapper:
- lightrec_free_block(state->cp_wrapper);
-err_free_rfe_wrapper:
- lightrec_free_block(state->rfe_wrapper);
-err_free_mtc_wrapper:
- lightrec_free_block(state->mtc_wrapper);
-err_free_mfc_wrapper:
- lightrec_free_block(state->mfc_wrapper);
-err_free_rw_wrapper:
- lightrec_free_block(state->rw_wrapper);
-err_free_generic_rw_wrapper:
- lightrec_free_block(state->rw_generic_wrapper);
err_free_dispatcher:
- lightrec_free_block(state->dispatcher);
+ lightrec_free_block(state, state->dispatcher);
err_free_reaper:
if (ENABLE_THREADED_COMPILER)
lightrec_reaper_destroy(state->reaper);
err_free_recompiler:
if (ENABLE_THREADED_COMPILER)
lightrec_free_recompiler(state->rec);
-err_free_reg_cache:
- lightrec_free_regcache(state->reg_cache);
+ else
+ lightrec_free_cstate(state->cstate);
err_free_block_cache:
lightrec_free_block_cache(state->block_cache);
-err_free_tinymm:
-#if ENABLE_TINYMM
- tinymm_shutdown(state->tinymm);
err_free_state:
-#endif
lightrec_unregister(MEM_FOR_LIGHTREC, sizeof(*state) +
- sizeof(*state->code_lut) * CODE_LUT_SIZE);
+ lut_elm_size(state) * CODE_LUT_SIZE);
free(state);
err_finish_jit:
finish_jit();
+ if (ENABLE_CODE_BUFFER && tlsf)
+ tlsf_destroy(tlsf);
return NULL;
}
void lightrec_destroy(struct lightrec_state *state)
{
+ /* Force a print info on destroy*/
+ state->current_cycle = ~state->current_cycle;
+ lightrec_print_info(state);
+
+ lightrec_free_block_cache(state->block_cache);
+ lightrec_free_block(state, state->dispatcher);
+ lightrec_free_block(state, state->c_wrapper_block);
+
if (ENABLE_THREADED_COMPILER) {
lightrec_free_recompiler(state->rec);
lightrec_reaper_destroy(state->reaper);
+ } else {
+ lightrec_free_cstate(state->cstate);
}
- lightrec_free_regcache(state->reg_cache);
- lightrec_free_block_cache(state->block_cache);
- lightrec_free_block(state->dispatcher);
- lightrec_free_block(state->rw_generic_wrapper);
- lightrec_free_block(state->rw_wrapper);
- lightrec_free_block(state->mfc_wrapper);
- lightrec_free_block(state->mtc_wrapper);
- lightrec_free_block(state->rfe_wrapper);
- lightrec_free_block(state->cp_wrapper);
- lightrec_free_block(state->syscall_wrapper);
- lightrec_free_block(state->break_wrapper);
finish_jit();
+ if (ENABLE_CODE_BUFFER && state->tlsf)
+ tlsf_destroy(state->tlsf);
-#if ENABLE_TINYMM
- tinymm_shutdown(state->tinymm);
-#endif
lightrec_unregister(MEM_FOR_LIGHTREC, sizeof(*state) +
- sizeof(*state->code_lut) * CODE_LUT_SIZE);
+ lut_elm_size(state) * CODE_LUT_SIZE);
free(state);
}
void lightrec_invalidate(struct lightrec_state *state, u32 addr, u32 len)
{
u32 kaddr = kunseg(addr & ~0x3);
- const struct lightrec_mem_map *map = lightrec_get_map(state, kaddr);
-
- if (map) {
- while (map->mirror_of)
- map = map->mirror_of;
-
- if (map != &state->maps[PSX_MAP_KERNEL_USER_RAM])
- return;
+ enum psx_map idx = lightrec_get_map_idx(state, kaddr);
+ switch (idx) {
+ case PSX_MAP_MIRROR1:
+ case PSX_MAP_MIRROR2:
+ case PSX_MAP_MIRROR3:
/* Handle mirrors */
- kaddr &= (state->maps[PSX_MAP_KERNEL_USER_RAM].length - 1);
-
- for (; len > 4; len -= 4, kaddr += 4)
- lightrec_invalidate_map(state, map, kaddr);
-
- lightrec_invalidate_map(state, map, kaddr);
+ kaddr &= RAM_SIZE - 1;
+ fallthrough;
+ case PSX_MAP_KERNEL_USER_RAM:
+ break;
+ default:
+ return;
}
+
+ memset(lut_address(state, lut_offset(kaddr)), 0,
+ ((len + 3) / 4) * lut_elm_size(state));
}
void lightrec_invalidate_all(struct lightrec_state *state)
{
- memset(state->code_lut, 0, sizeof(*state->code_lut) * CODE_LUT_SIZE);
+ memset(state->code_lut, 0, lut_elm_size(state) * CODE_LUT_SIZE);
}
-void lightrec_set_invalidate_mode(struct lightrec_state *state, bool dma_only)
+void lightrec_set_unsafe_opt_flags(struct lightrec_state *state, u32 flags)
{
- if (state->invalidate_from_dma_only != dma_only)
+ if ((flags ^ state->opt_flags) & LIGHTREC_OPT_INV_DMA_ONLY)
lightrec_invalidate_all(state);
- state->invalidate_from_dma_only = dma_only;
+ state->opt_flags = flags;
}
void lightrec_set_exit_flags(struct lightrec_state *state, u32 flags)
return state->exit_flags;
}
-void lightrec_dump_registers(struct lightrec_state *state, u32 regs[34])
-{
- memcpy(regs, state->native_reg_cache, sizeof(state->native_reg_cache));
-}
-
-void lightrec_restore_registers(struct lightrec_state *state, u32 regs[34])
-{
- memcpy(state->native_reg_cache, regs, sizeof(state->native_reg_cache));
-}
-
u32 lightrec_current_cycle_count(const struct lightrec_state *state)
{
return state->current_cycle;
state->target_cycle = cycles;
}
}
+
+struct lightrec_registers * lightrec_get_registers(struct lightrec_state *state)
+{
+ return &state->regs;
+}
+
+void lightrec_set_cycles_per_opcode(struct lightrec_state *state, u32 cycles)
+{
+ state->cycles_per_op = cycles;
+}