#include "lightrec-config.h"
#include "disassembler.h"
#include "lightrec.h"
+#include "regcache.h"
#if ENABLE_THREADED_COMPILER
#include <stdatomic.h>
#endif
+#ifdef _MSC_BUILD
+#include <immintrin.h>
+#endif
+
+#include <inttypes.h>
+#include <stdint.h>
+
+#define X32_FMT "0x%08"PRIx32
+#define PC_FMT "PC "X32_FMT
+
#define ARRAY_SIZE(x) (sizeof(x) ? sizeof(x) / sizeof((x)[0]) : 0)
+#define GENMASK(h, l) \
+ (((uintptr_t)-1 << (l)) & ((uintptr_t)-1 >> (__WORDSIZE - 1 - (h))))
+
#ifdef __GNUC__
# define likely(x) __builtin_expect(!!(x),1)
# define unlikely(x) __builtin_expect(!!(x),0)
#define SET_DEFAULT_ELM(table, value) [0] = NULL
#endif
-#define fallthrough do {} while (0) /* fall-through */
+#if __has_attribute(__fallthrough__)
+# define fallthrough __attribute__((__fallthrough__))
+#else
+# define fallthrough do {} while (0) /* fallthrough */
+#endif
+
+#define container_of(ptr, type, member) \
+ ((type *)((void *)(ptr) - offsetof(type, member)))
+
+#ifdef _MSC_BUILD
+# define popcount32(x) __popcnt(x)
+# define clz32(x) _lzcnt_u32(x)
+# define ctz32(x) _tzcnt_u32(x)
+#else
+# define popcount32(x) __builtin_popcount(x)
+# define clz32(x) __builtin_clz(x)
+# define ctz32(x) __builtin_ctz(x)
+#endif
/* Flags for (struct block *)->flags */
#define BLOCK_NEVER_COMPILE BIT(0)
#define BLOCK_FULLY_TAGGED BIT(2)
#define BLOCK_IS_DEAD BIT(3)
#define BLOCK_IS_MEMSET BIT(4)
+#define BLOCK_NO_OPCODE_LIST BIT(5)
+#define BLOCK_PRELOAD_PC BIT(6)
#define RAM_SIZE 0x200000
#define BIOS_SIZE 0x80000
#define REG_LO 32
#define REG_HI 33
+#define REG_TEMP (offsetof(struct lightrec_state, temp_reg) / sizeof(u32))
/* Definition of jit_state_t (avoids inclusion of <lightning.h>) */
struct jit_node;
struct opcode;
struct reaper;
+struct u16x2 {
+#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
+ u16 h, l;
+#else
+ u16 l, h;
+#endif
+};
+
struct block {
jit_state_t *_jit;
struct opcode *opcode_list;
u32 precompile_date;
unsigned int code_size;
u16 nb_ops;
- u8 flags;
#if ENABLE_THREADED_COMPILER
- atomic_flag op_list_freed;
+ _Atomic u8 flags;
+#else
+ u8 flags;
#endif
};
enum c_wrappers {
C_WRAPPER_RW,
C_WRAPPER_RW_GENERIC,
+ C_WRAPPER_MFC,
C_WRAPPER_MTC,
C_WRAPPER_CP,
- C_WRAPPER_SYSCALL,
- C_WRAPPER_BREAK,
C_WRAPPERS_COUNT,
};
struct lightrec_cstate {
struct lightrec_state *state;
- struct jit_node *branches[512];
struct lightrec_branch local_branches[512];
struct lightrec_branch_target targets[512];
- unsigned int nb_branches;
+ u16 movi_temp[32];
unsigned int nb_local_branches;
unsigned int nb_targets;
unsigned int cycles;
struct regcache *reg_cache;
+
+ _Bool no_load_delay;
};
struct lightrec_state {
struct lightrec_registers regs;
+ u32 temp_reg;
+ u32 curr_pc;
u32 next_pc;
+ uintptr_t wrapper_regs[NUM_TEMPS];
+ u8 in_delay_slot_n;
u32 current_cycle;
u32 target_cycle;
u32 exit_flags;
u32 old_cycle_counter;
- u32 c_wrapper_arg;
+ u32 cycles_per_op;
+ void *c_wrapper;
struct block *dispatcher, *c_wrapper_block;
void *c_wrappers[C_WRAPPERS_COUNT];
- void *wrappers_eps[C_WRAPPERS_COUNT];
struct blockcache *block_cache;
struct recompiler *rec;
struct lightrec_cstate *cstate;
struct reaper *reaper;
void *tlsf;
void (*eob_wrapper_func)(void);
+ void (*interpreter_func)(void);
+ void (*ds_check_func)(void);
void (*memset_func)(void);
void (*get_next_block)(void);
struct lightrec_ops ops;
unsigned int nb_precompile;
+ unsigned int nb_compile;
unsigned int nb_maps;
const struct lightrec_mem_map *maps;
- uintptr_t offset_ram, offset_bios, offset_scratch;
+ uintptr_t offset_ram, offset_bios, offset_scratch, offset_io;
+ u32 opt_flags;
_Bool with_32bit_lut;
_Bool mirrors_mapped;
- _Bool invalidate_from_dma_only;
void *code_lut[];
};
-u32 lightrec_rw(struct lightrec_state *state, union code op,
- u32 addr, u32 data, u16 *flags,
- struct block *block);
+#define lightrec_offset(ptr) \
+ offsetof(struct lightrec_state, ptr)
+
+u32 lightrec_rw(struct lightrec_state *state, union code op, u32 addr,
+ u32 data, u32 *flags, struct block *block, u16 offset);
void lightrec_free_block(struct lightrec_state *state, struct block *block);
void remove_from_code_lut(struct blockcache *cache, struct block *block);
-enum psx_map
-lightrec_get_map_idx(struct lightrec_state *state, u32 kaddr);
-
const struct lightrec_mem_map *
lightrec_get_map(struct lightrec_state *state, void **host, u32 kaddr);
{
u16 flags = block->opcode_list[offset].flags;
- offset += !!(OPT_SWITCH_DELAY_SLOTS && (flags & LIGHTREC_NO_DS));
+ offset += op_flag_no_ds(flags);
- return block->pc + (offset + imm << 2);
+ return block->pc + ((offset + imm) << 2);
}
static inline u32 get_branch_pc(const struct block *block, u16 offset, s16 imm)
{
u16 flags = block->opcode_list[offset].flags;
- offset -= !!(OPT_SWITCH_DELAY_SLOTS && (flags & LIGHTREC_NO_DS));
+ offset -= op_flag_no_ds(flags);
- return block->pc + (offset + imm << 2);
+ return block->pc + ((offset + imm) << 2);
}
-void lightrec_mtc(struct lightrec_state *state, union code op, u32 data);
+void lightrec_mtc(struct lightrec_state *state, union code op, u8 reg, u32 data);
u32 lightrec_mfc(struct lightrec_state *state, union code op);
void lightrec_rfe(struct lightrec_state *state);
void lightrec_cp(struct lightrec_state *state, union code op);
union code lightrec_read_opcode(struct lightrec_state *state, u32 pc);
-struct block * lightrec_get_block(struct lightrec_state *state, u32 pc);
int lightrec_compile_block(struct lightrec_cstate *cstate, struct block *block);
-void lightrec_free_opcode_list(struct lightrec_state *state, struct block *block);
+void lightrec_free_opcode_list(struct lightrec_state *state,
+ struct opcode *list);
-unsigned int lightrec_cycles_of_opcode(union code code);
+unsigned int lightrec_cycles_of_opcode(const struct lightrec_state *state,
+ union code code);
static inline u8 get_mult_div_lo(union code c)
{
return (OPT_FLAG_MULT_DIV && c.r.imm) ? c.r.imm : REG_HI;
}
+static inline s16 s16_max(s16 a, s16 b)
+{
+ return a > b ? a : b;
+}
+
+static inline _Bool block_has_flag(struct block *block, u8 flag)
+{
+#if ENABLE_THREADED_COMPILER
+ return atomic_load_explicit(&block->flags, memory_order_relaxed) & flag;
+#else
+ return block->flags & flag;
+#endif
+}
+
+static inline u8 block_set_flags(struct block *block, u8 mask)
+{
+#if ENABLE_THREADED_COMPILER
+ return atomic_fetch_or_explicit(&block->flags, mask,
+ memory_order_relaxed);
+#else
+ u8 flags = block->flags;
+
+ block->flags |= mask;
+
+ return flags;
+#endif
+}
+
+static inline u8 block_clear_flags(struct block *block, u8 mask)
+{
+#if ENABLE_THREADED_COMPILER
+ return atomic_fetch_and_explicit(&block->flags, ~mask,
+ memory_order_relaxed);
+#else
+ u8 flags = block->flags;
+
+ block->flags &= ~mask;
+
+ return flags;
+#endif
+}
+
+static inline _Bool can_sign_extend(s32 value, u8 order)
+{
+ return ((u32)(value >> (order - 1)) + 1) < 2;
+}
+
+static inline _Bool can_zero_extend(u32 value, u8 order)
+{
+ return (value >> order) == 0;
+}
+
+static inline const struct opcode *
+get_delay_slot(const struct opcode *list, u16 i)
+{
+ return op_flag_no_ds(list[i].flags) ? &list[i - 1] : &list[i + 1];
+}
+
+static inline _Bool lightrec_store_next_pc(void)
+{
+ return NUM_REGS + NUM_TEMPS <= 4;
+}
+
#endif /* __LIGHTREC_PRIVATE_H__ */