2 * (C) GraÅžvydas "notaz" Ignotas, 2011
4 * This work is licensed under the terms of any of these licenses
6 * - GNU GPL, version 2 or later.
7 * - GNU LGPL, version 2.1 or later.
8 * See the COPYING file in the top-level directory.
15 #define ARRAY_SIZE(x) (sizeof(x) / sizeof((x)[0]))
17 #define min(a, b) ((a) < (b) ? (a) : (b))
20 #define max(a, b) ((a) > (b) ? (a) : (b))
23 extern const unsigned char cmd_lengths[256];
24 #define command_lengths cmd_lengths
26 static unsigned int *ex_regs;
27 static int initialized;
30 #define SET_Ex(r, v) \
33 static __attribute__((noinline)) void
34 sync_enhancement_buffers(int x, int y, int w, int h);
36 #include "../gpulib/gpu.h"
37 #include "psx_gpu/psx_gpu.c"
38 #include "psx_gpu/psx_gpu_parse.c"
40 static psx_gpu_struct egpu __attribute__((aligned(256)));
42 int do_cmd_list(uint32_t *list, int count,
43 int *cycles_sum, int *cycles_last, int *last_cmd)
47 #if defined(__arm__) && defined(NEON_BUILD) && !defined(SIMD_BUILD)
48 // the asm doesn't bother to save callee-save vector regs, so do it here
49 __asm__ __volatile__("":::"q4","q5","q6","q7");
52 if (gpu.state.enhancement_active)
53 ret = gpu_parse_enhanced(&egpu, list, count * 4,
54 cycles_sum, cycles_last, (u32 *)last_cmd);
56 ret = gpu_parse(&egpu, list, count * 4,
57 cycles_sum, cycles_last, (u32 *)last_cmd);
59 #if defined(__arm__) && defined(NEON_BUILD) && !defined(SIMD_BUILD)
60 __asm__ __volatile__("":::"q4","q5","q6","q7");
64 ex_regs[1] |= egpu.texture_settings & 0x1ff;
68 #define ENHANCEMENT_BUF_SIZE (1024 * 1024 * 2 * 4 + 4096 * 2)
70 static void *get_enhancement_bufer(int *x, int *y, int *w, int *h,
73 uint16_t *ret = select_enhancement_buf_ptr(&egpu, *x, *y);
85 static void map_enhancement_buffer(void)
87 // currently we use 4x 1024*1024 buffers instead of single 2048*1024
88 // to be able to reuse 1024-width code better (triangle setup,
89 // dithering phase, lines).
90 egpu.enhancement_buf_ptr = gpu.mmap(ENHANCEMENT_BUF_SIZE);
91 if (egpu.enhancement_buf_ptr == NULL) {
92 fprintf(stderr, "failed to map enhancement buffer\n");
93 gpu.get_enhancement_bufer = NULL;
96 egpu.enhancement_buf_ptr += 4096 / 2;
97 gpu.get_enhancement_bufer = get_enhancement_bufer;
101 int renderer_init(void)
103 if (gpu.vram != NULL) {
104 initialize_psx_gpu(&egpu, gpu.vram);
108 if (gpu.mmap != NULL && egpu.enhancement_buf_ptr == NULL)
109 map_enhancement_buffer();
111 ex_regs = gpu.ex_regs;
115 void renderer_finish(void)
117 if (egpu.enhancement_buf_ptr != NULL) {
118 egpu.enhancement_buf_ptr -= 4096 / 2;
119 gpu.munmap(egpu.enhancement_buf_ptr, ENHANCEMENT_BUF_SIZE);
121 egpu.enhancement_buf_ptr = NULL;
122 egpu.enhancement_current_buf_ptr = NULL;
126 static __attribute__((noinline)) void
127 sync_enhancement_buffers(int x, int y, int w, int h)
129 int i, right = x + w, bottom = y + h;
130 const u16 *src = gpu.vram;
131 // use these because the scanout struct may hold reduced w, h
132 // due to intersection stuff, see the update_enhancement_buf_scanouts() mess
133 int s_w = max(gpu.screen.hres, gpu.screen.w);
134 int s_h = gpu.screen.vres;
135 if (gpu.screen.y < 0)
138 for (i = 0; i < ARRAY_SIZE(egpu.enhancement_scanouts); i++) {
139 const struct psx_gpu_scanout *s = &egpu.enhancement_scanouts[i];
140 u16 *dst = select_enhancement_buf_by_index(&egpu, i);
142 if (s->w == 0) continue;
143 if (s->x >= right) continue;
144 if (s->x + s_w <= x) continue;
145 if (s->y >= bottom) continue;
146 if (s->y + s_h <= y) continue;
148 x2 = min(right, s->x + s_w);
150 y2 = min(bottom, s->y + s_h);
151 // 16-byte align for the asm version
154 scale2x_tiles8(dst + y1 * 1024*2 + x1 * 2,
155 src + y1 * 1024 + x1, (x2 - x1 + 7) / 8u, y2 - y1);
159 void renderer_sync_ecmds(uint32_t *ecmds)
163 gpu_parse(&egpu, ecmds + 1, 6 * 4, &dummy0, &dummy0, &dummy1);
166 void renderer_update_caches(int x, int y, int w, int h, int state_changed)
168 update_texture_cache_region(&egpu, x, y, x + w - 1, y + h - 1);
170 if (gpu.state.enhancement_active) {
172 int vres = gpu.screen.vres;
173 if (gpu.screen.y < 0)
174 vres -= gpu.screen.y;
175 memset(egpu.enhancement_scanouts, 0, sizeof(egpu.enhancement_scanouts));
176 egpu.enhancement_scanout_eselect = 0;
177 update_enhancement_buf_scanouts(&egpu,
178 gpu.screen.src_x, gpu.screen.src_y, gpu.screen.hres, vres);
181 sync_enhancement_buffers(x, y, w, h);
185 void renderer_flush_queues(void)
187 flush_render_block_buffer(&egpu);
190 void renderer_set_interlace(int enable, int is_odd)
192 egpu.render_mode &= ~(RENDER_INTERLACE_ENABLED|RENDER_INTERLACE_ODD);
194 egpu.render_mode |= RENDER_INTERLACE_ENABLED;
196 egpu.render_mode |= RENDER_INTERLACE_ODD;
199 void renderer_notify_res_change(void)
201 renderer_notify_scanout_change(gpu.screen.src_x, gpu.screen.src_y);
204 void renderer_notify_scanout_change(int x, int y)
206 int vres = gpu.screen.vres;
207 if (!gpu.state.enhancement_active || !egpu.enhancement_buf_ptr)
210 if (gpu.screen.y < 0)
211 vres -= gpu.screen.y;
212 update_enhancement_buf_scanouts(&egpu, x, y, gpu.screen.hres, vres);
215 #include "../../frontend/plugin_lib.h"
217 void renderer_set_config(const struct rearmed_cbs *cbs)
220 initialize_psx_gpu(&egpu, gpu.vram);
223 if (cbs->pl_set_gpu_caps)
224 cbs->pl_set_gpu_caps(GPU_CAP_SUPPORTS_2X);
226 egpu.hack_disable_main = cbs->gpu_neon.enhancement_no_main;
227 egpu.hack_texture_adj = cbs->gpu_neon.enhancement_tex_adj;
228 if (gpu.state.enhancement_enable) {
229 if (gpu.mmap != NULL && egpu.enhancement_buf_ptr == NULL)
230 map_enhancement_buffer();
234 // vim:ts=2:sw=2:expandtab