frontend: update libpicofe, fix missed callbacks
[pcsx_rearmed.git] / plugins / gpu_neon / psx_gpu / psx_gpu_simd.c
... / ...
CommitLineData
1/*
2 * Copyright (C) 2011 Gilead Kutnick "Exophase" <exophase@gmail.com>
3 * Copyright (C) 2022 GraÅžvydas Ignotas "notaz" <notasas@gmail.com>
4 *
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License as
7 * published by the Free Software Foundation; either version 2 of
8 * the License, or (at your option) any later version.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * General Public License for more details.
14 */
15
16#include <string.h>
17#include "psx_gpu.h"
18#include "psx_gpu_simd.h"
19//#define ASM_PROTOTYPES
20//#include "psx_gpu_simd.h"
21#ifdef __SSE2__
22#include <x86intrin.h>
23#endif
24#ifndef SIMD_BUILD
25#error "please define SIMD_BUILD if you want this gpu_neon C simd implementation"
26#endif
27
28typedef u8 gvu8 __attribute__((vector_size(16)));
29typedef u16 gvu16 __attribute__((vector_size(16)));
30typedef u32 gvu32 __attribute__((vector_size(16)));
31typedef u64 gvu64 __attribute__((vector_size(16)));
32typedef s8 gvs8 __attribute__((vector_size(16)));
33typedef s16 gvs16 __attribute__((vector_size(16)));
34typedef s32 gvs32 __attribute__((vector_size(16)));
35typedef s64 gvs64 __attribute__((vector_size(16)));
36
37typedef u8 gvhu8 __attribute__((vector_size(8)));
38typedef u16 gvhu16 __attribute__((vector_size(8)));
39typedef u32 gvhu32 __attribute__((vector_size(8)));
40typedef u64 gvhu64 __attribute__((vector_size(8)));
41typedef s8 gvhs8 __attribute__((vector_size(8)));
42typedef s16 gvhs16 __attribute__((vector_size(8)));
43typedef s32 gvhs32 __attribute__((vector_size(8)));
44typedef s64 gvhs64 __attribute__((vector_size(8)));
45
46typedef union
47{
48 gvu8 u8;
49 gvu16 u16;
50 gvu32 u32;
51 gvu64 u64;
52 gvs8 s8;
53 gvs16 s16;
54 gvs32 s32;
55 gvs64 s64;
56#ifdef __SSE2__
57 __m128i m;
58#endif
59 // this may be tempting, but it causes gcc to do lots of stack spills
60 //gvhreg h[2];
61} gvreg;
62
63typedef gvreg gvreg_ua __attribute__((aligned(1)));
64typedef uint64_t uint64_t_ua __attribute__((aligned(1)));
65typedef gvu8 gvu8_ua __attribute__((aligned(1)));
66typedef gvu16 gvu16_ua __attribute__((aligned(1)));
67
68#if defined(__ARM_NEON) || defined(__ARM_NEON__)
69#include <arm_neon.h>
70
71typedef union
72{
73 gvhu8 u8;
74 gvhu16 u16;
75 gvhu32 u32;
76 gvhu64 u64;
77 //u64 u64;
78 //uint64x1_t u64;
79 gvhs8 s8;
80 gvhs16 s16;
81 gvhs32 s32;
82 gvhs64 s64;
83 //s64 s64;
84 //int64x1_t s64;
85} gvhreg;
86
87#define gvaddhn_u32(d, a, b) d.u16 = vaddhn_u32(a.u32, b.u32)
88#define gvaddw_s32(d, a, b) d.s64 = vaddw_s32(a.s64, b.s32)
89#define gvabsq_s32(d, s) d.s32 = vabsq_s32(s.s32)
90#define gvbic_n_u16(d, n) d.u16 = vbic_u16(d.u16, vmov_n_u16(n))
91#define gvbifq(d, a, b) d.u8 = vbslq_u8(b.u8, d.u8, a.u8)
92#define gvbit(d, a, b) d.u8 = vbsl_u8(b.u8, a.u8, d.u8)
93#define gvceqq_u16(d, a, b) d.u16 = vceqq_u16(a.u16, b.u16)
94#define gvcgt_s16(d, a, b) d.u16 = vcgt_s16(a.s16, b.s16)
95#define gvclt_s16(d, a, b) d.u16 = vclt_s16(a.s16, b.s16)
96#define gvcreate_s32(d, a, b) d.s32 = vcreate_s32((u32)(a) | ((u64)(b) << 32))
97#define gvcreate_u32(d, a, b) d.u32 = vcreate_u32((u32)(a) | ((u64)(b) << 32))
98#define gvcreate_s64(d, s) d.s64 = (gvhs64)vcreate_s64(s)
99#define gvcreate_u64(d, s) d.u64 = (gvhu64)vcreate_u64(s)
100#define gvcombine_u16(d, l, h) d.u16 = vcombine_u16(l.u16, h.u16)
101#define gvcombine_u32(d, l, h) d.u32 = vcombine_u32(l.u32, h.u32)
102#define gvcombine_s64(d, l, h) d.s64 = vcombine_s64((int64x1_t)l.s64, (int64x1_t)h.s64)
103#define gvdup_l_u8(d, s, l) d.u8 = vdup_lane_u8(s.u8, l)
104#define gvdup_l_u16(d, s, l) d.u16 = vdup_lane_u16(s.u16, l)
105#define gvdup_l_u32(d, s, l) d.u32 = vdup_lane_u32(s.u32, l)
106#define gvdupq_l_s64(d, s, l) d.s64 = vdupq_lane_s64((int64x1_t)s.s64, l)
107#define gvdupq_l_u32(d, s, l) d.u32 = vdupq_lane_u32(s.u32, l)
108#define gvdup_n_s64(d, n) d.s64 = vdup_n_s64(n)
109#define gvdup_n_u8(d, n) d.u8 = vdup_n_u8(n)
110#define gvdup_n_u16(d, n) d.u16 = vdup_n_u16(n)
111#define gvdup_n_u32(d, n) d.u32 = vdup_n_u32(n)
112#define gvdupq_n_u16(d, n) d.u16 = vdupq_n_u16(n)
113#define gvdupq_n_u32(d, n) d.u32 = vdupq_n_u32(n)
114#define gvdupq_n_s64(d, n) d.s64 = vdupq_n_s64(n)
115#define gvhaddq_u16(d, a, b) d.u16 = vhaddq_u16(a.u16, b.u16)
116#define gvmax_s16(d, a, b) d.s16 = vmax_s16(a.s16, b.s16)
117#define gvmin_s16(d, a, b) d.s16 = vmin_s16(a.s16, b.s16)
118#define gvmin_u16(d, a, b) d.u16 = vmin_u16(a.u16, b.u16)
119#define gvminq_u8(d, a, b) d.u8 = vminq_u8(a.u8, b.u8)
120#define gvminq_u16(d, a, b) d.u16 = vminq_u16(a.u16, b.u16)
121#define gvmla_s32(d, a, b) d.s32 = vmla_s32(d.s32, a.s32, b.s32)
122#define gvmla_u32(d, a, b) d.u32 = vmla_u32(d.u32, a.u32, b.u32)
123#define gvmlaq_s32(d, a, b) d.s32 = vmlaq_s32(d.s32, a.s32, b.s32)
124#define gvmlaq_u32(d, a, b) d.u32 = vmlaq_u32(d.u32, a.u32, b.u32)
125#define gvmlal_s32(d, a, b) d.s64 = vmlal_s32(d.s64, a.s32, b.s32)
126#define gvmlal_u8(d, a, b) d.u16 = vmlal_u8(d.u16, a.u8, b.u8)
127#define gvmlsq_s32(d, a, b) d.s32 = vmlsq_s32(d.s32, a.s32, b.s32)
128#define gvmlsq_l_s32(d, a, b, l) d.s32 = vmlsq_lane_s32(d.s32, a.s32, b.s32, l)
129#define gvmov_l_s32(d, s, l) d.s32 = vset_lane_s32(s, d.s32, l)
130#define gvmov_l_u32(d, s, l) d.u32 = vset_lane_u32(s, d.u32, l)
131#define gvmovl_u8(d, s) d.u16 = vmovl_u8(s.u8)
132#define gvmovl_s32(d, s) d.s64 = vmovl_s32(s.s32)
133#define gvmovn_u16(d, s) d.u8 = vmovn_u16(s.u16)
134#define gvmovn_u32(d, s) d.u16 = vmovn_u32(s.u32)
135#define gvmovn_u64(d, s) d.u32 = vmovn_u64(s.u64)
136#define gvmul_s32(d, a, b) d.s32 = vmul_s32(a.s32, b.s32)
137#define gvmull_s16(d, a, b) d.s32 = vmull_s16(a.s16, b.s16)
138#define gvmull_s32(d, a, b) d.s64 = vmull_s32(a.s32, b.s32)
139#define gvmull_u8(d, a, b) d.u16 = vmull_u8(a.u8, b.u8)
140#define gvmull_l_u32(d, a, b, l) d.u64 = vmull_lane_u32(a.u32, b.u32, l)
141#define gvmlsl_s16(d, a, b) d.s32 = vmlsl_s16(d.s32, a.s16, b.s16)
142#define gvneg_s32(d, s) d.s32 = vneg_s32(s.s32)
143#define gvqadd_u8(d, a, b) d.u8 = vqadd_u8(a.u8, b.u8)
144#define gvqsub_u8(d, a, b) d.u8 = vqsub_u8(a.u8, b.u8)
145#define gvshl_u16(d, a, b) d.u16 = vshl_u16(a.u16, b.s16)
146#define gvshlq_u64(d, a, b) d.u64 = vshlq_u64(a.u64, b.s64)
147#define gvshrq_n_s16(d, s, n) d.s16 = vshrq_n_s16(s.s16, n)
148#define gvshrq_n_u16(d, s, n) d.u16 = vshrq_n_u16(s.u16, n)
149#define gvshl_n_u32(d, s, n) d.u32 = vshl_n_u32(s.u32, n)
150#define gvshlq_n_u16(d, s, n) d.u16 = vshlq_n_u16(s.u16, n)
151#define gvshlq_n_u32(d, s, n) d.u32 = vshlq_n_u32(s.u32, n)
152#define gvshll_n_s8(d, s, n) d.s16 = vshll_n_s8(s.s8, n)
153#define gvshll_n_u8(d, s, n) d.u16 = vshll_n_u8(s.u8, n)
154#define gvshll_n_u16(d, s, n) d.u32 = vshll_n_u16(s.u16, n)
155#define gvshr_n_u8(d, s, n) d.u8 = vshr_n_u8(s.u8, n)
156#define gvshr_n_u16(d, s, n) d.u16 = vshr_n_u16(s.u16, n)
157#define gvshr_n_u32(d, s, n) d.u32 = vshr_n_u32(s.u32, n)
158#define gvshr_n_u64(d, s, n) d.u64 = (gvhu64)vshr_n_u64((uint64x1_t)s.u64, n)
159#define gvshrn_n_u16(d, s, n) d.u8 = vshrn_n_u16(s.u16, n)
160#define gvshrn_n_u32(d, s, n) d.u16 = vshrn_n_u32(s.u32, n)
161#define gvsli_n_u8(d, s, n) d.u8 = vsli_n_u8(d.u8, s.u8, n)
162#define gvsri_n_u8(d, s, n) d.u8 = vsri_n_u8(d.u8, s.u8, n)
163#define gvtstq_u16(d, a, b) d.u16 = vtstq_u16(a.u16, b.u16)
164#define gvqshrun_n_s16(d, s, n) d.u8 = vqshrun_n_s16(s.s16, n)
165#define gvqsubq_u8(d, a, b) d.u8 = vqsubq_u8(a.u8, b.u8)
166#define gvqsubq_u16(d, a, b) d.u16 = vqsubq_u16(a.u16, b.u16)
167
168#define gvmovn_top_u64(d, s) d.u32 = vshrn_n_u64(s.u64, 32)
169
170#define gvget_lo(d, s) d.u16 = vget_low_u16(s.u16)
171#define gvget_hi(d, s) d.u16 = vget_high_u16(s.u16)
172#define gvlo(s) ({gvhreg t_; gvget_lo(t_, s); t_;})
173#define gvhi(s) ({gvhreg t_; gvget_hi(t_, s); t_;})
174
175#define gvset_lo(d, s) d.u16 = vcombine_u16(s.u16, gvhi(d).u16)
176#define gvset_hi(d, s) d.u16 = vcombine_u16(gvlo(d).u16, s.u16)
177
178#define gvtbl2_u8(d, a, b) { \
179 uint8x8x2_t v_; \
180 v_.val[0] = vget_low_u8(a.u8); v_.val[1] = vget_high_u8(a.u8); \
181 d.u8 = vtbl2_u8(v_, b.u8); \
182}
183
184#define gvzip_u8(d, a, b) { \
185 uint8x8x2_t v_ = vzip_u8(a.u8, b.u8); \
186 d.u8 = vcombine_u8(v_.val[0], v_.val[1]); \
187}
188#define gvzipq_u16(d0, d1, s0, s1) { \
189 uint16x8x2_t v_ = vzipq_u16(s0.u16, s1.u16); \
190 d0.u16 = v_.val[0]; d1.u16 = v_.val[1]; \
191}
192
193#define gvld1_u8(d, s) d.u8 = vld1_u8(s)
194#define gvld1_u32(d, s) d.u32 = vld1_u32((const u32 *)(s))
195#define gvld1_u64(d, s) d.u64 = vld1_u64((const u64 *)(s))
196#define gvld1q_u8(d, s) d.u8 = vld1q_u8(s)
197#define gvld1q_u16(d, s) d.u16 = vld1q_u16(s)
198#define gvld1q_u32(d, s) d.u32 = vld1q_u32((const u32 *)(s))
199#define gvld2_u8_dup(v0, v1, p) { \
200 uint8x8x2_t v_ = vld2_dup_u8(p); \
201 v0.u8 = v_.val[0]; v1.u8 = v_.val[1]; \
202}
203#define gvld2q_u8(v0, v1, p) { \
204 uint8x16x2_t v_ = vld2q_u8(p); \
205 v0.u8 = v_.val[0]; v1.u8 = v_.val[1]; \
206}
207
208#define gvst1_u8(v, p) \
209 vst1_u8(p, v.u8)
210#define gvst1_u64(v, p) \
211 vst1_u64((u64 *)(p), v.u64)
212#define gvst1q_u16(v, p) \
213 vst1q_u16(p, v.u16)
214#define gvst1q_inc_u32(v, p, i) { \
215 vst1q_u32((u32 *)(p), v.u32); \
216 p += (i) / sizeof(*p); \
217}
218#define gvst2_u8(v0, v1, p) { \
219 uint8x8x2_t v_; \
220 v_.val[0] = v0.u8; v_.val[1] = v1.u8; \
221 vst2_u8(p, v_); \
222}
223#define gvst2_u16(v0, v1, p) { \
224 uint16x4x2_t v_; \
225 v_.val[0] = v0.u16; v_.val[1] = v1.u16; \
226 vst2_u16(p, v_); \
227}
228#define gvst2q_u8(v0, v1, p) { \
229 uint8x16x2_t v_; \
230 v_.val[0] = v0.u8; v_.val[1] = v1.u8; \
231 vst2q_u8(p, v_); \
232}
233#define gvst4_4_inc_u32(v0, v1, v2, v3, p, i) { \
234 uint32x2x4_t v_; \
235 v_.val[0] = v0.u32; v_.val[1] = v1.u32; v_.val[2] = v2.u32; v_.val[3] = v3.u32; \
236 vst4_u32(p, v_); p += (i) / sizeof(*p); \
237}
238#define gvst4_pi_u16(v0, v1, v2, v3, p) { \
239 uint16x4x4_t v_; \
240 v_.val[0] = v0.u16; v_.val[1] = v1.u16; v_.val[2] = v2.u16; v_.val[3] = v3.u16; \
241 vst4_u16((u16 *)(p), v_); p += sizeof(v_) / sizeof(*p); \
242}
243#define gvst1q_pi_u32(v, p) \
244 gvst1q_inc_u32(v, p, sizeof(v))
245// could use vst1q_u32_x2 but that's not always available
246#define gvst1q_2_pi_u32(v0, v1, p) { \
247 gvst1q_inc_u32(v0, p, sizeof(v0)); \
248 gvst1q_inc_u32(v1, p, sizeof(v1)); \
249}
250
251/* notes:
252 - gcc > 9: (arm32) int64x1_t type produces ops on gp regs
253 (also u64 __attribute__((vector_size(8))) :( )
254 - gcc <11: (arm32) handles '<vec> == 0' poorly
255*/
256
257#elif defined(__SSE2__)
258
259// use a full reg and discard the upper half
260#define gvhreg gvreg
261
262#define gv0() _mm_setzero_si128()
263
264#ifdef __x86_64__
265#define gvcreate_s32(d, a, b) d.m = _mm_cvtsi64_si128((u32)(a) | ((u64)(b) << 32))
266#define gvcreate_s64(d, s) d.m = _mm_cvtsi64_si128(s)
267#else
268#define gvcreate_s32(d, a, b) d.m = _mm_set_epi32(0, 0, b, a)
269#define gvcreate_s64(d, s) d.m = _mm_loadu_si64(&(s))
270#endif
271
272#define gvbic_n_u16(d, n) d.m = _mm_andnot_si128(_mm_set1_epi16(n), d.m)
273#define gvceqq_u16(d, a, b) d.u16 = vceqq_u16(a.u16, b.u16)
274#define gvcgt_s16(d, a, b) d.m = _mm_cmpgt_epi16(a.m, b.m)
275#define gvclt_s16(d, a, b) d.m = _mm_cmpgt_epi16(b.m, a.m)
276#define gvcreate_u32 gvcreate_s32
277#define gvcreate_u64 gvcreate_s64
278#define gvcombine_u16(d, l, h) d.m = _mm_unpacklo_epi64(l.m, h.m)
279#define gvcombine_u32 gvcombine_u16
280#define gvcombine_s64 gvcombine_u16
281#define gvdup_l_u8(d, s, l) d.u8 = vdup_lane_u8(s.u8, l)
282#define gvdup_l_u16(d, s, l) d.m = _mm_shufflelo_epi16(s.m, (l)|((l)<<2)|((l)<<4)|((l)<<6))
283#define gvdup_l_u32(d, s, l) d.m = vdup_lane_u32(s.u32, l)
284#define gvdupq_l_s64(d, s, l) d.m = _mm_unpacklo_epi64(s.m, s.m)
285#define gvdupq_l_u32(d, s, l) d.m = _mm_shuffle_epi32(s.m, (l)|((l)<<2)|((l)<<4)|((l)<<6))
286#define gvdup_n_s64(d, n) d.m = _mm_set1_epi64x(n)
287#define gvdup_n_u8(d, n) d.m = _mm_set1_epi8(n)
288#define gvdup_n_u16(d, n) d.m = _mm_set1_epi16(n)
289#define gvdup_n_u32(d, n) d.m = _mm_set1_epi32(n)
290#define gvdupq_n_u16(d, n) d.m = _mm_set1_epi16(n)
291#define gvdupq_n_u32(d, n) d.m = _mm_set1_epi32(n)
292#define gvdupq_n_s64(d, n) d.m = _mm_set1_epi64x(n)
293#define gvmax_s16(d, a, b) d.m = _mm_max_epi16(a.m, b.m)
294#define gvmin_s16(d, a, b) d.m = _mm_min_epi16(a.m, b.m)
295#define gvminq_u8(d, a, b) d.m = _mm_min_epu8(a.m, b.m)
296#define gvmovn_u64(d, s) d.m = _mm_shuffle_epi32(s.m, 0 | (2 << 2))
297#define gvmovn_top_u64(d, s) d.m = _mm_shuffle_epi32(s.m, 1 | (3 << 2))
298#define gvmull_s16(d, a, b) { \
299 __m128i lo_ = _mm_mullo_epi16(a.m, b.m); \
300 __m128i hi_ = _mm_mulhi_epi16(a.m, b.m); \
301 d.m = _mm_unpacklo_epi16(lo_, hi_); \
302}
303#define gvmull_l_u32(d, a, b, l) { \
304 __m128i a_ = _mm_unpacklo_epi32(a.m, a.m); /* lanes 0,1 -> 0,2 */ \
305 __m128i b_ = _mm_shuffle_epi32(b.m, (l) | ((l) << 4)); \
306 d.m = _mm_mul_epu32(a_, b_); \
307}
308#define gvmlsl_s16(d, a, b) { \
309 gvreg tmp_; \
310 gvmull_s16(tmp_, a, b); \
311 d.m = _mm_sub_epi32(d.m, tmp_.m); \
312}
313#define gvqadd_u8(d, a, b) d.m = _mm_adds_epu8(a.m, b.m)
314#define gvqsub_u8(d, a, b) d.m = _mm_subs_epu8(a.m, b.m)
315#define gvshrq_n_s16(d, s, n) d.m = _mm_srai_epi16(s.m, n)
316#define gvshrq_n_u16(d, s, n) d.m = _mm_srli_epi16(s.m, n)
317#define gvshrq_n_u32(d, s, n) d.m = _mm_srli_epi32(s.m, n)
318#define gvshl_n_u32(d, s, n) d.m = _mm_slli_epi32(s.m, n)
319#define gvshlq_n_u16(d, s, n) d.m = _mm_slli_epi16(s.m, n)
320#define gvshlq_n_u32(d, s, n) d.m = _mm_slli_epi32(s.m, n)
321#define gvshll_n_u16(d, s, n) d.m = _mm_slli_epi32(_mm_unpacklo_epi16(s.m, gv0()), n)
322#define gvshr_n_u16(d, s, n) d.m = _mm_srli_epi16(s.m, n)
323#define gvshr_n_u32(d, s, n) d.m = _mm_srli_epi32(s.m, n)
324#define gvshr_n_u64(d, s, n) d.m = _mm_srli_epi64(s.m, n)
325#define gvshrn_n_s64(d, s, n) { \
326 gvreg tmp_; \
327 gvshrq_n_s64(tmp_, s, n); \
328 d.m = _mm_shuffle_epi32(tmp_.m, 0 | (2 << 2)); \
329}
330#define gvqshrun_n_s16(d, s, n) { \
331 __m128i t_ = _mm_srai_epi16(s.m, n); \
332 d.m = _mm_packus_epi16(t_, t_); \
333}
334#define gvqsubq_u8(d, a, b) d.m = _mm_subs_epu8(a.m, b.m)
335#define gvqsubq_u16(d, a, b) d.m = _mm_subs_epu16(a.m, b.m)
336
337#ifdef __SSSE3__
338#define gvabsq_s32(d, s) d.m = _mm_abs_epi32(s.m)
339#define gvtbl2_u8(d, a, b) d.m = _mm_shuffle_epi8(a.m, b.m)
340#else
341// must supply these here or else gcc will produce something terrible with __builtin_shuffle
342#define gvmovn_u16(d, s) { \
343 __m128i t2_ = _mm_and_si128(s.m, _mm_set1_epi16(0xff)); \
344 d.m = _mm_packus_epi16(t2_, t2_); \
345}
346#define gvmovn_u32(d, s) { \
347 __m128i t2_; \
348 t2_ = _mm_shufflelo_epi16(s.m, (0 << 0) | (2 << 2)); \
349 t2_ = _mm_shufflehi_epi16(t2_, (0 << 0) | (2 << 2)); \
350 d.m = _mm_shuffle_epi32(t2_, (0 << 0) | (2 << 2)); \
351}
352#define gvmovn_top_u32(d, s) { \
353 __m128i t2_; \
354 t2_ = _mm_shufflelo_epi16(s.m, (1 << 0) | (3 << 2)); \
355 t2_ = _mm_shufflehi_epi16(t2_, (1 << 0) | (3 << 2)); \
356 d.m = _mm_shuffle_epi32(t2_, (0 << 0) | (2 << 2)); \
357}
358#endif // !__SSSE3__
359#ifdef __SSE4_1__
360#define gvmin_u16(d, a, b) d.m = _mm_min_epu16(a.m, b.m)
361#define gvminq_u16 gvmin_u16
362#define gvmovl_u8(d, s) d.m = _mm_cvtepu8_epi16(s.m)
363#define gvmovl_s8(d, s) d.m = _mm_cvtepi8_epi16(s.m)
364#define gvmovl_s32(d, s) d.m = _mm_cvtepi32_epi64(s.m)
365#define gvmull_s32(d, a, b) { \
366 __m128i a_ = _mm_unpacklo_epi32(a.m, a.m); /* lanes 0,1 -> 0,2 */ \
367 __m128i b_ = _mm_unpacklo_epi32(b.m, b.m); \
368 d.m = _mm_mul_epi32(a_, b_); \
369}
370#else
371#define gvmovl_u8(d, s) d.m = _mm_unpacklo_epi8(s.m, gv0())
372#define gvmovl_s8(d, s) d.m = _mm_unpacklo_epi8(s.m, _mm_cmpgt_epi8(gv0(), s.m))
373#define gvmovl_s32(d, s) d.m = _mm_unpacklo_epi32(s.m, _mm_srai_epi32(s.m, 31))
374#endif // !__SSE4_1__
375#ifndef __AVX2__
376#define gvshlq_u64(d, a, b) { \
377 gvreg t1_, t2_; \
378 t1_.m = _mm_sll_epi64(a.m, b.m); \
379 t2_.m = _mm_sll_epi64(a.m, _mm_shuffle_epi32(b.m, (2 << 0) | (3 << 2))); \
380 d.u64 = (gvu64){ t1_.u64[0], t2_.u64[1] }; \
381}
382#endif // __AVX2__
383
384#define gvlo(s) s
385#define gvhi(s) ((gvreg)_mm_shuffle_epi32(s.m, (2 << 0) | (3 << 2)))
386#define gvget_lo(d, s) d = gvlo(s)
387#define gvget_hi(d, s) d = gvhi(s)
388
389#define gvset_lo(d, s) d.m = _mm_unpacklo_epi64(s.m, gvhi(d).m)
390#define gvset_hi(d, s) d.m = _mm_unpacklo_epi64(d.m, s.m)
391
392#define gvld1_u8(d, s) d.m = _mm_loadu_si64(s)
393#define gvld1_u32 gvld1_u8
394#define gvld1_u64 gvld1_u8
395#define gvld1q_u8(d, s) d.m = _mm_loadu_si128((__m128i *)(s))
396#define gvld1q_u16 gvld1q_u8
397#define gvld1q_u32 gvld1q_u8
398
399#define gvst1_u8(v, p) _mm_storeu_si64(p, v.m)
400#define gvst1_u64 gvst1_u8
401
402#define gvst4_4_inc_u32(v0, v1, v2, v3, p, i) { \
403 __m128i t0 = _mm_unpacklo_epi32(v0.m, v1.m); \
404 __m128i t1 = _mm_unpacklo_epi32(v2.m, v3.m); \
405 _mm_storeu_si128(((__m128i *)(p)) + 0, _mm_unpacklo_epi64(t0, t1)); \
406 _mm_storeu_si128(((__m128i *)(p)) + 1, _mm_unpackhi_epi64(t0, t1)); \
407 p += (i) / sizeof(*p); \
408}
409#define gvst4_pi_u16(v0, v1, v2, v3, p) { \
410 __m128i t0 = _mm_unpacklo_epi16(v0.m, v1.m); \
411 __m128i t1 = _mm_unpacklo_epi16(v2.m, v3.m); \
412 _mm_storeu_si128(((__m128i *)(p)) + 0, _mm_unpacklo_epi32(t0, t1)); \
413 _mm_storeu_si128(((__m128i *)(p)) + 1, _mm_unpackhi_epi32(t0, t1)); \
414 p += sizeof(t0) * 2 / sizeof(*p); \
415}
416
417#else
418#error "arch not supported or SIMD support was not enabled by your compiler"
419#endif
420
421// the below have intrinsics but they evaluate to basic operations on both gcc and clang
422#define gvadd_s64(d, a, b) d.s64 = a.s64 + b.s64
423#define gvadd_u8(d, a, b) d.u8 = a.u8 + b.u8
424#define gvadd_u16(d, a, b) d.u16 = a.u16 + b.u16
425#define gvadd_u32(d, a, b) d.u32 = a.u32 + b.u32
426#define gvaddq_s64 gvadd_s64
427#define gvaddq_u16 gvadd_u16
428#define gvaddq_u32 gvadd_u32
429#define gvand(d, a, b) d.u32 = a.u32 & b.u32
430#define gvand_n_u32(d, n) d.u32 &= n
431#define gvbic(d, a, b) d.u32 = a.u32 & ~b.u32
432#define gvbicq gvbic
433#define gveor(d, a, b) d.u32 = a.u32 ^ b.u32
434#define gveorq gveor
435#define gvceqz_u16(d, s) d.u16 = s.u16 == 0
436#define gvceqzq_u16 gvceqz_u16
437#define gvcltz_s16(d, s) d.s16 = s.s16 < 0
438#define gvcltzq_s16 gvcltz_s16
439#define gvsub_u16(d, a, b) d.u16 = a.u16 - b.u16
440#define gvsub_u32(d, a, b) d.u32 = a.u32 - b.u32
441#define gvsubq_u16 gvsub_u16
442#define gvsubq_u32 gvsub_u32
443#define gvorr(d, a, b) d.u32 = a.u32 | b.u32
444#define gvorrq gvorr
445#define gvorr_n_u16(d, n) d.u16 |= n
446
447// fallbacks
448#if 1
449
450#ifndef gvaddhn_u32
451#define gvaddhn_u32(d, a, b) { \
452 gvreg tmp1_ = { .u32 = a.u32 + b.u32 }; \
453 gvmovn_top_u32(d, tmp1_); \
454}
455#endif
456#ifndef gvabsq_s32
457#define gvabsq_s32(d, s) { \
458 gvreg tmp1_ = { .s32 = (gvs32){} - s.s32 }; \
459 gvreg mask_ = { .s32 = s.s32 >> 31 }; \
460 gvbslq_(d, mask_, tmp1_, s); \
461}
462#endif
463#ifndef gvbit
464#define gvbslq_(d, s, a, b) d.u32 = (a.u32 & s.u32) | (b.u32 & ~s.u32)
465#define gvbifq(d, a, b) gvbslq_(d, b, d, a)
466#define gvbit(d, a, b) gvbslq_(d, b, a, d)
467#endif
468#ifndef gvaddw_s32
469#define gvaddw_s32(d, a, b) {gvreg t_; gvmovl_s32(t_, b); d.s64 += t_.s64;}
470#endif
471#ifndef gvhaddq_u16
472// can do this because the caller needs the msb clear
473#define gvhaddq_u16(d, a, b) d.u16 = (a.u16 + b.u16) >> 1
474#endif
475#ifndef gvmin_u16
476#define gvmin_u16(d, a, b) { \
477 gvu16 t_ = a.u16 < b.u16; \
478 d.u16 = (a.u16 & t_) | (b.u16 & ~t_); \
479}
480#define gvminq_u16 gvmin_u16
481#endif
482#ifndef gvmlsq_s32
483#define gvmlsq_s32(d, a, b) d.s32 -= a.s32 * b.s32
484#endif
485#ifndef gvmlsq_l_s32
486#define gvmlsq_l_s32(d, a, b, l){gvreg t_; gvdupq_l_u32(t_, b, l); d.s32 -= a.s32 * t_.s32;}
487#endif
488#ifndef gvmla_s32
489#define gvmla_s32(d, a, b) d.s32 += a.s32 * b.s32
490#endif
491#ifndef gvmla_u32
492#define gvmla_u32 gvmla_s32
493#endif
494#ifndef gvmlaq_s32
495#define gvmlaq_s32(d, a, b) d.s32 += a.s32 * b.s32
496#endif
497#ifndef gvmlaq_u32
498#define gvmlaq_u32 gvmlaq_s32
499#endif
500#ifndef gvmlal_u8
501#define gvmlal_u8(d, a, b) {gvreg t_; gvmull_u8(t_, a, b); d.u16 += t_.u16;}
502#endif
503#ifndef gvmlal_s32
504#define gvmlal_s32(d, a, b) {gvreg t_; gvmull_s32(t_, a, b); d.s64 += t_.s64;}
505#endif
506#ifndef gvmov_l_s32
507#define gvmov_l_s32(d, s, l) d.s32[l] = s
508#endif
509#ifndef gvmov_l_u32
510#define gvmov_l_u32(d, s, l) d.u32[l] = s
511#endif
512#ifndef gvmul_s32
513#define gvmul_s32(d, a, b) d.s32 = a.s32 * b.s32
514#endif
515#ifndef gvmull_u8
516#define gvmull_u8(d, a, b) { \
517 gvreg t1_, t2_; \
518 gvmovl_u8(t1_, a); \
519 gvmovl_u8(t2_, b); \
520 d.u16 = t1_.u16 * t2_.u16; \
521}
522#endif
523#ifndef gvmull_s32
524// note: compilers tend to use int regs here
525#define gvmull_s32(d, a, b) { \
526 d.s64[0] = (s64)a.s32[0] * b.s32[0]; \
527 d.s64[1] = (s64)a.s32[1] * b.s32[1]; \
528}
529#endif
530#ifndef gvneg_s32
531#define gvneg_s32(d, s) d.s32 = -s.s32
532#endif
533// x86 note: needs _mm_sllv_epi16 (avx512), else this sucks terribly
534#ifndef gvshl_u16
535#define gvshl_u16(d, a, b) d.u16 = a.u16 << b.u16
536#endif
537// x86 note: needs _mm_sllv_* (avx2)
538#ifndef gvshlq_u64
539#define gvshlq_u64(d, a, b) d.u64 = a.u64 << b.u64
540#endif
541#ifndef gvshll_n_s8
542#define gvshll_n_s8(d, s, n) {gvreg t_; gvmovl_s8(t_, s); gvshlq_n_u16(d, t_, n);}
543#endif
544#ifndef gvshll_n_u8
545#define gvshll_n_u8(d, s, n) {gvreg t_; gvmovl_u8(t_, s); gvshlq_n_u16(d, t_, n);}
546#endif
547#ifndef gvshr_n_u8
548#define gvshr_n_u8(d, s, n) d.u8 = s.u8 >> (n)
549#endif
550#ifndef gvshrq_n_s64
551#define gvshrq_n_s64(d, s, n) d.s64 = s.s64 >> (n)
552#endif
553#ifndef gvshrn_n_u16
554#define gvshrn_n_u16(d, s, n) {gvreg t_; gvshrq_n_u16(t_, s, n); gvmovn_u16(d, t_);}
555#endif
556#ifndef gvshrn_n_u32
557#define gvshrn_n_u32(d, s, n) {gvreg t_; gvshrq_n_u32(t_, s, n); gvmovn_u32(d, t_);}
558#endif
559#ifndef gvsli_n_u8
560#define gvsli_n_u8(d, s, n) d.u8 = (s.u8 << (n)) | (d.u8 & ((1u << (n)) - 1u))
561#endif
562#ifndef gvsri_n_u8
563#define gvsri_n_u8(d, s, n) d.u8 = (s.u8 >> (n)) | (d.u8 & ((0xff00u >> (n)) & 0xffu))
564#endif
565#ifndef gvtstq_u16
566#define gvtstq_u16(d, a, b) d.u16 = (a.u16 & b.u16) != 0
567#endif
568
569#ifndef gvld2_u8_dup
570#define gvld2_u8_dup(v0, v1, p) { \
571 gvdup_n_u8(v0, ((const u8 *)(p))[0]); \
572 gvdup_n_u8(v1, ((const u8 *)(p))[1]); \
573}
574#endif
575#ifndef gvst1_u8
576#define gvst1_u8(v, p) *(uint64_t_ua *)(p) = v.u64[0]
577#endif
578#ifndef gvst1q_u16
579#define gvst1q_u16(v, p) *(gvreg_ua *)(p) = v
580#endif
581#ifndef gvst1q_inc_u32
582#define gvst1q_inc_u32(v, p, i) {*(gvreg_ua *)(p) = v; p += (i) / sizeof(*p);}
583#endif
584#ifndef gvst1q_pi_u32
585#define gvst1q_pi_u32(v, p) gvst1q_inc_u32(v, p, sizeof(v))
586#endif
587#ifndef gvst1q_2_pi_u32
588#define gvst1q_2_pi_u32(v0, v1, p) { \
589 gvst1q_inc_u32(v0, p, sizeof(v0)); \
590 gvst1q_inc_u32(v1, p, sizeof(v1)); \
591}
592#endif
593#ifndef gvst2_u8
594#define gvst2_u8(v0, v1, p) {gvreg t_; gvzip_u8(t_, v0, v1); *(gvu8_ua *)(p) = t_.u8;}
595#endif
596#ifndef gvst2_u16
597#define gvst2_u16(v0, v1, p) {gvreg t_; gvzip_u16(t_, v0, v1); *(gvu16_ua *)(p) = t_.u16;}
598#endif
599
600// note: these shuffles assume sizeof(gvhreg) == 16 && sizeof(gvreg) == 16
601#ifndef __has_builtin
602#define __has_builtin(x) 0
603#endif
604
605// prefer __builtin_shuffle on gcc as it handles -1 poorly
606#if __has_builtin(__builtin_shufflevector) && !__has_builtin(__builtin_shuffle)
607
608#ifndef gvld2q_u8
609#define gvld2q_u8(v0, v1, p) { \
610 gvu8 v0_ = ((gvu8_ua *)(p))[0]; \
611 gvu8 v1_ = ((gvu8_ua *)(p))[1]; \
612 v0.u8 = __builtin_shufflevector(v0_, v1_, 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30); \
613 v1.u8 = __builtin_shufflevector(v0_, v1_, 1,3,5,7,9,11,13,15,17,19,21,23,25,27,29,31); \
614}
615#endif
616#ifndef gvmovn_u16
617#define gvmovn_u16(d, s) \
618 d.u8 = __builtin_shufflevector(s.u8, s.u8, 0,2,4,6,8,10,12,14,-1,-1,-1,-1,-1,-1,-1,-1)
619#endif
620#ifndef gvmovn_u32
621#define gvmovn_u32(d, s) \
622 d.u16 = __builtin_shufflevector(s.u16, s.u16, 0,2,4,6,-1,-1,-1,-1)
623#endif
624#ifndef gvmovn_top_u32
625#define gvmovn_top_u32(d, s) \
626 d.u16 = __builtin_shufflevector(s.u16, s.u16, 1,3,5,7,-1,-1,-1,-1)
627#endif
628#ifndef gvzip_u8
629#define gvzip_u8(d, a, b) \
630 d.u8 = __builtin_shufflevector(a.u8, b.u8, 0,16,1,17,2,18,3,19,4,20,5,21,6,22,7,23)
631#endif
632#ifndef gvzip_u16
633#define gvzip_u16(d, a, b) \
634 d.u16 = __builtin_shufflevector(a.u16, b.u16, 0,8,1,9,2,10,3,11)
635#endif
636#ifndef gvzipq_u16
637#define gvzipq_u16(d0, d1, s0, s1) { \
638 gvu16 t_ = __builtin_shufflevector(s0.u16, s1.u16, 0, 8, 1, 9, 2, 10, 3, 11); \
639 d1.u16 = __builtin_shufflevector(s0.u16, s1.u16, 4,12, 5,13, 6, 14, 7, 15); \
640 d0.u16 = t_; \
641}
642#endif
643
644#else // !__has_builtin(__builtin_shufflevector)
645
646#ifndef gvld2q_u8
647#define gvld2q_u8(v0, v1, p) { \
648 gvu8 v0_ = ((gvu8_ua *)(p))[0]; \
649 gvu8 v1_ = ((gvu8_ua *)(p))[1]; \
650 v0.u8 = __builtin_shuffle(v0_, v1_, (gvu8){0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30}); \
651 v1.u8 = __builtin_shuffle(v0_, v1_, (gvu8){1,3,5,7,9,11,13,15,17,19,21,23,25,27,29,31}); \
652}
653#endif
654#ifndef gvmovn_u16
655#define gvmovn_u16(d, s) \
656 d.u8 = __builtin_shuffle(s.u8, (gvu8){0,2,4,6,8,10,12,14,0,2,4,6,8,10,12,14})
657#endif
658#ifndef gvmovn_u32
659#define gvmovn_u32(d, s) \
660 d.u16 = __builtin_shuffle(s.u16, (gvu16){0,2,4,6,0,2,4,6})
661#endif
662#ifndef gvmovn_top_u32
663#define gvmovn_top_u32(d, s) \
664 d.u16 = __builtin_shuffle(s.u16, (gvu16){1,3,5,7,1,3,5,7})
665#endif
666#ifndef gvtbl2_u8
667#define gvtbl2_u8(d, a, b) d.u8 = __builtin_shuffle(a.u8, b.u8)
668#endif
669#ifndef gvzip_u8
670#define gvzip_u8(d, a, b) \
671 d.u8 = __builtin_shuffle(a.u8, b.u8, (gvu8){0,16,1,17,2,18,3,19,4,20,5,21,6,22,7,23})
672#endif
673#ifndef gvzip_u16
674#define gvzip_u16(d, a, b) \
675 d.u16 = __builtin_shuffle(a.u16, b.u16, (gvu16){0,8,1,9,2,10,3,11})
676#endif
677#ifndef gvzipq_u16
678#define gvzipq_u16(d0, d1, s0, s1) { \
679 gvu16 t_ = __builtin_shuffle(s0.u16, s1.u16, (gvu16){0, 8, 1, 9, 2, 10, 3, 11}); \
680 d1.u16 = __builtin_shuffle(s0.u16, s1.u16, (gvu16){4,12, 5,13, 6, 14, 7, 15}); \
681 d0.u16 = t_; \
682}
683#endif
684
685#endif // __builtin_shufflevector || __builtin_shuffle
686
687#ifndef gvtbl2_u8
688#define gvtbl2_u8(d, a, b) { \
689 int i_; \
690 for (i_ = 0; i_ < 16; i_++) \
691 d.u8[i_] = a.u8[b.u8[i_]]; \
692}
693#endif
694
695#endif // fallbacks
696
697#if defined(__arm__)
698
699#define gssub16(d, a, b) asm("ssub16 %0,%1,%2" : "=r"(d) : "r"(a), "r"(b))
700#define gsmusdx(d, a, b) asm("smusdx %0,%1,%2" : "=r"(d) : "r"(a), "r"(b))
701
702#if 0
703// gcc/config/arm/arm.c
704#undef gvadd_s64
705#define gvadd_s64(d, a, b) asm("vadd.i64 %P0,%P1,%P2" : "=w"(d.s64) : "w"(a.s64), "w"(b.s64))
706#endif
707
708#else
709
710#define gssub16(d, a, b) d = (u16)((a) - (b)) | ((((a) >> 16) - ((b) >> 16)) << 16)
711#define gsmusdx(d, a, b) d = ((s32)(s16)(a) * ((s32)(b) >> 16)) \
712 - (((s32)(a) >> 16) * (s16)(b))
713
714#endif
715
716// for compatibility with the original psx_gpu.c code
717#define vec_2x64s gvreg
718#define vec_2x64u gvreg
719#define vec_4x32s gvreg
720#define vec_4x32u gvreg
721#define vec_8x16s gvreg
722#define vec_8x16u gvreg
723#define vec_16x8s gvreg
724#define vec_16x8u gvreg
725#define vec_1x64s gvhreg
726#define vec_1x64u gvhreg
727#define vec_2x32s gvhreg
728#define vec_2x32u gvhreg
729#define vec_4x16s gvhreg
730#define vec_4x16u gvhreg
731#define vec_8x8s gvhreg
732#define vec_8x8u gvhreg
733
734#if 0
735#include <stdio.h>
736#include <stdlib.h>
737#include <unistd.h>
738static int ccount, dump_enabled;
739void cmpp(const char *name, const void *a_, const void *b_, size_t len)
740{
741 const uint32_t *a = a_, *b = b_, masks[] = { 0, 0xff, 0xffff, 0xffffff };
742 size_t i, left;
743 uint32_t mask;
744 for (i = 0; i < (len + 3)/4; i++) {
745 left = len - i*4;
746 mask = left >= 4 ? ~0u : masks[left];
747 if ((a[i] ^ b[i]) & mask) {
748 printf("%s: %08x %08x [%03zx/%zu] #%d\n",
749 name, a[i] & mask, b[i] & mask, i*4, i, ccount);
750 exit(1);
751 }
752 }
753 ccount++;
754}
755#define ccmpf(n) cmpp(#n, &psx_gpu->n, &n##_c, sizeof(n##_c))
756#define ccmpa(n,c) cmpp(#n, &psx_gpu->n, &n##_c, sizeof(n##_c[0]) * c)
757
758void dump_r_(const char *name, void *dump, int is_q)
759{
760 unsigned long long *u = dump;
761 if (!dump_enabled) return;
762 //if (ccount > 1) return;
763 printf("%20s %016llx ", name, u[0]);
764 if (is_q)
765 printf("%016llx", u[1]);
766 puts("");
767}
768void __attribute__((noinline,noclone)) dump_r_d(const char *name, void *dump)
769{ dump_r_(name, dump, 0); }
770void __attribute__((noinline,noclone)) dump_r_q(const char *name, void *dump)
771{ dump_r_(name, dump, 1); }
772#define dumprd(n) { u8 dump_[8]; gvst1_u8(n, dump_); dump_r_d(#n, dump_); }
773#define dumprq(n) { u16 dump_[8]; gvst1q_u16(n, dump_); dump_r_q(#n, dump_); }
774#endif
775
776void compute_all_gradients(psx_gpu_struct * __restrict__ psx_gpu,
777 const vertex_struct * __restrict__ a, const vertex_struct * __restrict__ b,
778 const vertex_struct * __restrict__ c)
779{
780 union { double d; struct { u32 l; u32 h; } i; } divident, divider;
781 union { double d; gvhreg v; } d30;
782
783#if 0
784 compute_all_gradients_(psx_gpu, a, b, c);
785 return;
786#endif
787 // First compute the triangle area reciprocal and shift. The division will
788 // happen concurrently with much of the work which follows.
789
790 // load exponent of 62 into upper half of double
791 u32 shift = __builtin_clz(psx_gpu->triangle_area);
792 u32 triangle_area_normalized = psx_gpu->triangle_area << shift;
793
794 // load area normalized into lower half of double
795 divident.i.l = triangle_area_normalized >> 10;
796 divident.i.h = (62 + 1023) << 20;
797
798 divider.i.l = triangle_area_normalized << 20;
799 divider.i.h = ((1022 + 31) << 20) + (triangle_area_normalized >> 11);
800
801 d30.d = divident.d / divider.d; // d30 = ((1 << 62) + ta_n) / ta_n
802
803 // ((x1 - x0) * (y2 - y1)) - ((x2 - x1) * (y1 - y0)) =
804 // ( d0 * d1 ) - ( d2 * d3 ) =
805 // ( m0 ) - ( m1 ) = gradient
806
807 // This is split to do 12 elements at a time over three sets: a, b, and c.
808 // Technically we only need to do 10 elements (uvrgb_x and uvrgb_y), so
809 // two of the slots are unused.
810
811 // Inputs are all 16-bit signed. The m0/m1 results are 32-bit signed, as
812 // is g.
813
814 // First type is: uvrg bxxx xxxx
815 // Second type is: yyyy ybyy uvrg
816 // Since x_a and y_c are the same the same variable is used for both.
817
818 gvreg v0;
819 gvreg v1;
820 gvreg v2;
821 gvreg uvrg_xxxx0;
822 gvreg uvrg_xxxx1;
823 gvreg uvrg_xxxx2;
824
825 gvreg y0_ab;
826 gvreg y1_ab;
827 gvreg y2_ab;
828
829 gvreg d0_ab;
830 gvreg d1_ab;
831 gvreg d2_ab;
832 gvreg d3_ab;
833
834 gvreg ga_uvrg_x;
835 gvreg ga_uvrg_y;
836 gvreg gw_rg_x;
837 gvreg gw_rg_y;
838 gvreg w_mask;
839 gvreg r_shift;
840 gvreg uvrg_dx2, uvrg_dx3;
841 gvreg uvrgb_phase;
842 gvhreg zero, tmp_lo, tmp_hi;
843
844 gvld1q_u8(v0, (u8 *)a); // v0 = { uvrg0, b0, x0, y0 }
845 gvld1q_u8(v1, (u8 *)b); // v1 = { uvrg1, b1, x1, y1 }
846 gvld1q_u8(v2, (u8 *)c); // v2 = { uvrg2, b2, x2, y2 }
847
848 gvmovl_u8(uvrg_xxxx0, gvlo(v0)); // uvrg_xxxx0 = { uv0, rg0, b0-, -- }
849 gvmovl_u8(uvrg_xxxx1, gvlo(v1)); // uvrg_xxxx1 = { uv1, rg1, b1-, -- }
850 gvmovl_u8(uvrg_xxxx2, gvlo(v2)); // uvrg_xxxx2 = { uv2, rg2, b2-, -- }
851
852 gvdup_l_u16(tmp_lo, gvhi(v0), 1); // yyyy0 = { yy0, yy0 }
853 gvcombine_u16(y0_ab, tmp_lo, gvlo(uvrg_xxxx0));
854
855 gvdup_l_u16(tmp_lo, gvhi(v0), 0); // xxxx0 = { xx0, xx0 }
856 gvset_hi(uvrg_xxxx0, tmp_lo);
857
858 u32 x1_x2 = (u16)b->x | (c->x << 16); // x1_x2 = { x1, x2 }
859 u32 x0_x1 = (u16)a->x | (b->x << 16); // x0_x1 = { x0, x1 }
860
861 gvdup_l_u16(tmp_lo, gvhi(v1), 1); // yyyy1 = { yy1, yy1 }
862 gvcombine_u16(y1_ab, tmp_lo, gvlo(uvrg_xxxx1));
863
864 gvdup_l_u16(tmp_lo, gvhi(v1), 0); // xxxx1 = { xx1, xx1 }
865 gvset_hi(uvrg_xxxx1, tmp_lo);
866
867 gvdup_l_u16(tmp_lo, gvhi(v2), 1); // yyyy2 = { yy2, yy2 }
868 gvcombine_u16(y2_ab, tmp_lo, gvlo(uvrg_xxxx2));
869
870 gvdup_l_u16(tmp_lo, gvhi(v2), 0); // xxxx2 = { xx2, xx2 }
871 gvset_hi(uvrg_xxxx2, tmp_lo);
872
873 u32 y0_y1 = (u16)a->y | (b->y << 16); // y0_y1 = { y0, y1 }
874 u32 y1_y2 = (u16)b->y | (c->y << 16); // y1_y2 = { y1, y2 }
875
876 gvsubq_u16(d0_ab, uvrg_xxxx1, uvrg_xxxx0);
877
878 u32 b1_b2 = b->b | (c->b << 16); // b1_b2 = { b1, b2 }
879
880 gvsubq_u16(d2_ab, uvrg_xxxx2, uvrg_xxxx1);
881
882 gvsubq_u16(d1_ab, y2_ab, y1_ab);
883
884 u32 b0_b1 = a->b | (b->b << 16); // b0_b1 = { b0, b1 }
885
886 u32 dx, dy, db;
887 gssub16(dx, x1_x2, x0_x1); // dx = { x1 - x0, x2 - x1 }
888 gssub16(dy, y1_y2, y0_y1); // dy = { y1 - y0, y2 - y1 }
889 gssub16(db, b1_b2, b0_b1); // db = { b1 - b0, b2 - b1 }
890
891 u32 ga_by, ga_bx;
892 gvsubq_u16(d3_ab, y1_ab, y0_ab);
893 gsmusdx(ga_by, dx, db); // ga_by = ((x1 - x0) * (b2 - b1)) -
894 // ((x2 - X1) * (b1 - b0))
895 gvmull_s16(ga_uvrg_x, gvlo(d0_ab), gvlo(d1_ab));
896 gsmusdx(ga_bx, db, dy); // ga_bx = ((b1 - b0) * (y2 - y1)) -
897 // ((b2 - b1) * (y1 - y0))
898 gvmlsl_s16(ga_uvrg_x, gvlo(d2_ab), gvlo(d3_ab));
899 u32 gs_bx = (s32)ga_bx >> 31; // movs
900
901 gvmull_s16(ga_uvrg_y, gvhi(d0_ab), gvhi(d1_ab));
902 if ((s32)gs_bx < 0) ga_bx = -ga_bx; // rsbmi
903
904 gvmlsl_s16(ga_uvrg_y, gvhi(d2_ab), gvhi(d3_ab));
905 u32 gs_by = (s32)ga_by >> 31; // movs
906
907 gvhreg d0;
908 gvshr_n_u64(d0, d30.v, 22); // note: on "d30 >> 22" gcc generates junk code
909
910 gvdupq_n_u32(uvrgb_phase, psx_gpu->uvrgb_phase);
911 u32 b_base = psx_gpu->uvrgb_phase + (a->b << 16);
912
913 if ((s32)gs_by < 0) ga_by = -ga_by; // rsbmi
914 gvreg gs_uvrg_x, gs_uvrg_y;
915 gs_uvrg_x.s32 = ga_uvrg_x.s32 < 0; // gs_uvrg_x = ga_uvrg_x < 0
916 gs_uvrg_y.s32 = ga_uvrg_y.s32 < 0; // gs_uvrg_y = ga_uvrg_y < 0
917
918 gvdupq_n_u32(w_mask, -psx_gpu->triangle_winding); // w_mask = { -w, -w, -w, -w }
919 shift -= 62 - 12; // shift -= (62 - FIXED_BITS)
920
921 gvreg uvrg_base;
922 gvshll_n_u16(uvrg_base, gvlo(uvrg_xxxx0), 16); // uvrg_base = uvrg0 << 16
923
924 gvaddq_u32(uvrg_base, uvrg_base, uvrgb_phase);
925 gvabsq_s32(ga_uvrg_x, ga_uvrg_x); // ga_uvrg_x = abs(ga_uvrg_x)
926
927 u32 area_r_s = d0.u32[0]; // area_r_s = triangle_reciprocal
928 gvabsq_s32(ga_uvrg_y, ga_uvrg_y); // ga_uvrg_y = abs(ga_uvrg_y)
929
930 gvmull_l_u32(gw_rg_x, gvhi(ga_uvrg_x), d0, 0);
931 gvmull_l_u32(ga_uvrg_x, gvlo(ga_uvrg_x), d0, 0);
932 gvmull_l_u32(gw_rg_y, gvhi(ga_uvrg_y), d0, 0);
933 gvmull_l_u32(ga_uvrg_y, gvlo(ga_uvrg_y), d0, 0);
934
935#if defined(__ARM_NEON) || defined(__ARM_NEON__)
936 gvdupq_n_s64(r_shift, shift); // r_shift = { shift, shift }
937 gvshlq_u64(gw_rg_x, gw_rg_x, r_shift);
938 gvshlq_u64(ga_uvrg_x, ga_uvrg_x, r_shift);
939 gvshlq_u64(gw_rg_y, gw_rg_y, r_shift);
940 gvshlq_u64(ga_uvrg_y, ga_uvrg_y, r_shift);
941#elif defined(__SSE2__)
942 r_shift.m = _mm_cvtsi32_si128(-shift);
943 gw_rg_x.m = _mm_srl_epi64(gw_rg_x.m, r_shift.m);
944 ga_uvrg_x.m = _mm_srl_epi64(ga_uvrg_x.m, r_shift.m);
945 gw_rg_y.m = _mm_srl_epi64(gw_rg_y.m, r_shift.m);
946 ga_uvrg_y.m = _mm_srl_epi64(ga_uvrg_y.m, r_shift.m);
947#else
948 gvdupq_n_s64(r_shift, -shift); // r_shift = { shift, shift }
949 gvshrq_u64(gw_rg_x, gw_rg_x, r_shift);
950 gvshrq_u64(ga_uvrg_x, ga_uvrg_x, r_shift);
951 gvshrq_u64(gw_rg_y, gw_rg_y, r_shift);
952 gvshrq_u64(ga_uvrg_y, ga_uvrg_y, r_shift);
953#endif
954
955 gveorq(gs_uvrg_x, gs_uvrg_x, w_mask);
956 gvmovn_u64(tmp_lo, ga_uvrg_x);
957
958 gveorq(gs_uvrg_y, gs_uvrg_y, w_mask);
959 gvmovn_u64(tmp_hi, gw_rg_x);
960
961 gvcombine_u32(ga_uvrg_x, tmp_lo, tmp_hi);
962
963 gveorq(ga_uvrg_x, ga_uvrg_x, gs_uvrg_x);
964 gvmovn_u64(tmp_lo, ga_uvrg_y);
965
966 gvsubq_u32(ga_uvrg_x, ga_uvrg_x, gs_uvrg_x);
967 gvmovn_u64(tmp_hi, gw_rg_y);
968
969 gvcombine_u32(ga_uvrg_y, tmp_lo, tmp_hi);
970
971 gveorq(ga_uvrg_y, ga_uvrg_y, gs_uvrg_y);
972 ga_bx = ga_bx << 13;
973
974 gvsubq_u32(ga_uvrg_y, ga_uvrg_y, gs_uvrg_y);
975 ga_by = ga_by << 13;
976
977 u32 gw_bx_h, gw_by_h;
978 gw_bx_h = (u64)ga_bx * area_r_s >> 32;
979
980 gvshlq_n_u32(ga_uvrg_x, ga_uvrg_x, 4);
981 gvshlq_n_u32(ga_uvrg_y, ga_uvrg_y, 4);
982
983 gw_by_h = (u64)ga_by * area_r_s >> 32;
984 gvdup_n_u32(tmp_lo, a->x);
985 gvmlsq_l_s32(uvrg_base, ga_uvrg_x, tmp_lo, 0);
986
987 gs_bx = gs_bx ^ -psx_gpu->triangle_winding;
988 gvaddq_u32(uvrg_dx2, ga_uvrg_x, ga_uvrg_x);
989
990 gs_by = gs_by ^ -psx_gpu->triangle_winding;
991
992 u32 r11 = -shift; // r11 = negative shift for scalar lsr
993 u32 *store_a = psx_gpu->uvrg.e;
994 r11 = r11 - (32 - 13);
995 u32 *store_b = store_a + 16 / sizeof(u32);
996
997 gvaddq_u32(uvrg_dx3, uvrg_dx2, ga_uvrg_x);
998 gvst1q_inc_u32(uvrg_base, store_a, 32);
999
1000 gvst1q_inc_u32(ga_uvrg_x, store_b, 32);
1001 u32 g_bx = (u32)gw_bx_h >> r11;
1002
1003 gvst1q_inc_u32(ga_uvrg_y, store_a, 32);
1004 u32 g_by = (u32)gw_by_h >> r11;
1005
1006 gvdup_n_u32(zero, 0);
1007
1008 gvst4_4_inc_u32(zero, gvlo(ga_uvrg_x), gvlo(uvrg_dx2), gvlo(uvrg_dx3), store_b, 32);
1009 g_bx = g_bx ^ gs_bx;
1010
1011 gvst4_4_inc_u32(zero, gvhi(ga_uvrg_x), gvhi(uvrg_dx2), gvhi(uvrg_dx3), store_b, 32);
1012 g_bx = g_bx - gs_bx;
1013
1014 g_bx = g_bx << 4;
1015 g_by = g_by ^ gs_by;
1016
1017 b_base -= g_bx * a->x;
1018 g_by = g_by - gs_by;
1019
1020 g_by = g_by << 4;
1021
1022 u32 g_bx2 = g_bx + g_bx;
1023 u32 g_bx3 = g_bx + g_bx2;
1024
1025 // 112
1026 store_b[0] = 0;
1027 store_b[1] = g_bx;
1028 store_b[2] = g_bx2;
1029 store_b[3] = g_bx3;
1030 store_b[4] = b_base;
1031 store_b[5] = g_by; // 132
1032}
1033
1034#define setup_spans_debug_check(span_edge_data_element) \
1035
1036#define setup_spans_prologue_alternate_yes() \
1037 vec_2x64s alternate_x; \
1038 vec_2x64s alternate_dx_dy; \
1039 vec_4x32s alternate_x_32; \
1040 vec_4x16u alternate_x_16; \
1041 \
1042 vec_4x16u alternate_select; \
1043 vec_4x16s y_mid_point; \
1044 \
1045 s32 y_b = v_b->y; \
1046 s64 edge_alt; \
1047 s32 edge_dx_dy_alt; \
1048 u32 edge_shift_alt \
1049
1050#define setup_spans_prologue_alternate_no() \
1051
1052#define setup_spans_prologue(alternate_active) \
1053 edge_data_struct *span_edge_data; \
1054 vec_4x32u *span_uvrg_offset; \
1055 u32 *span_b_offset; \
1056 \
1057 s32 clip; \
1058 vec_4x32u v_clip; \
1059 \
1060 vec_2x64s edges_xy; \
1061 vec_2x32s edges_dx_dy; \
1062 vec_2x32u edge_shifts; \
1063 \
1064 vec_2x64s left_x, right_x; \
1065 vec_2x64s left_dx_dy, right_dx_dy; \
1066 vec_4x32s left_x_32, right_x_32; \
1067 vec_2x32s left_x_32_lo, right_x_32_lo; \
1068 vec_2x32s left_x_32_hi, right_x_32_hi; \
1069 vec_4x16s left_right_x_16_lo, left_right_x_16_hi; \
1070 vec_4x16s y_x4; \
1071 vec_8x16s left_edge; \
1072 vec_8x16s right_edge; \
1073 vec_4x16u span_shift; \
1074 \
1075 vec_2x32u c_0x01; \
1076 vec_4x16u c_0x04; \
1077 vec_4x16u c_0xFFFE; \
1078 vec_4x16u c_0x07; \
1079 \
1080 vec_2x32s x_starts; \
1081 vec_2x32s x_ends; \
1082 \
1083 s32 x_a = v_a->x; \
1084 s32 x_b = v_b->x; \
1085 s32 x_c = v_c->x; \
1086 s32 y_a = v_a->y; \
1087 s32 y_c = v_c->y; \
1088 \
1089 vec_4x32u uvrg; \
1090 vec_4x32u uvrg_dy; \
1091 u32 b = psx_gpu->b; \
1092 u32 b_dy = psx_gpu->b_dy; \
1093 const u32 *reciprocal_table = psx_gpu->reciprocal_table_ptr; \
1094 \
1095 gvld1q_u32(uvrg, psx_gpu->uvrg.e); \
1096 gvld1q_u32(uvrg_dy, psx_gpu->uvrg_dy.e); \
1097 gvdup_n_u32(c_0x01, 0x01); \
1098 setup_spans_prologue_alternate_##alternate_active() \
1099
1100#define setup_spans_prologue_b() \
1101 span_edge_data = psx_gpu->span_edge_data; \
1102 span_uvrg_offset = (vec_4x32u *)psx_gpu->span_uvrg_offset; \
1103 span_b_offset = psx_gpu->span_b_offset; \
1104 \
1105 vec_8x16u c_0x0001; \
1106 vec_4x16u c_max_blocks_per_row; \
1107 \
1108 gvdupq_n_u16(c_0x0001, 0x0001); \
1109 gvdupq_n_u16(left_edge, psx_gpu->viewport_start_x); \
1110 gvdupq_n_u16(right_edge, psx_gpu->viewport_end_x); \
1111 gvaddq_u16(right_edge, right_edge, c_0x0001); \
1112 gvdup_n_u16(c_0x04, 0x04); \
1113 gvdup_n_u16(c_0x07, 0x07); \
1114 gvdup_n_u16(c_0xFFFE, 0xFFFE); \
1115 gvdup_n_u16(c_max_blocks_per_row, MAX_BLOCKS_PER_ROW); \
1116
1117#if defined(__ARM_NEON) || defined(__ARM_NEON__)
1118// better encoding, remaining bits are unused anyway
1119#define mask_edge_shifts(edge_shifts) \
1120 gvbic_n_u16(edge_shifts, 0xE0)
1121#else
1122#define mask_edge_shifts(edge_shifts) \
1123 gvand_n_u32(edge_shifts, 0x1F)
1124#endif
1125
1126#define compute_edge_delta_x2() \
1127{ \
1128 vec_2x32s heights; \
1129 vec_2x32s height_reciprocals; \
1130 vec_2x32s heights_b; \
1131 vec_2x32u widths; \
1132 \
1133 u32 edge_shift = reciprocal_table[height]; \
1134 \
1135 gvdup_n_u32(heights, height); \
1136 gvsub_u32(widths, x_ends, x_starts); \
1137 \
1138 gvdup_n_u32(edge_shifts, edge_shift); \
1139 gvsub_u32(heights_b, heights, c_0x01); \
1140 gvshr_n_u32(height_reciprocals, edge_shifts, 10); \
1141 \
1142 gvmla_s32(heights_b, x_starts, heights); \
1143 mask_edge_shifts(edge_shifts); \
1144 gvmul_s32(edges_dx_dy, widths, height_reciprocals); \
1145 gvmull_s32(edges_xy, heights_b, height_reciprocals); \
1146} \
1147
1148#define compute_edge_delta_x3(start_c, height_a, height_b) \
1149{ \
1150 vec_2x32s heights; \
1151 vec_2x32s height_reciprocals; \
1152 vec_2x32s heights_b; \
1153 vec_2x32u widths; \
1154 \
1155 u32 width_alt; \
1156 s32 height_b_alt; \
1157 u32 height_reciprocal_alt; \
1158 \
1159 gvcreate_u32(heights, height_a, height_b); \
1160 gvcreate_u32(edge_shifts, reciprocal_table[height_a], reciprocal_table[height_b]); \
1161 \
1162 edge_shift_alt = reciprocal_table[height_minor_b]; \
1163 \
1164 gvsub_u32(widths, x_ends, x_starts); \
1165 width_alt = x_c - start_c; \
1166 \
1167 gvshr_n_u32(height_reciprocals, edge_shifts, 10); \
1168 height_reciprocal_alt = edge_shift_alt >> 10; \
1169 \
1170 mask_edge_shifts(edge_shifts); \
1171 edge_shift_alt &= 0x1F; \
1172 \
1173 gvsub_u32(heights_b, heights, c_0x01); \
1174 height_b_alt = height_minor_b - 1; \
1175 \
1176 gvmla_s32(heights_b, x_starts, heights); \
1177 height_b_alt += height_minor_b * start_c; \
1178 \
1179 gvmull_s32(edges_xy, heights_b, height_reciprocals); \
1180 edge_alt = (s64)height_b_alt * height_reciprocal_alt; \
1181 \
1182 gvmul_s32(edges_dx_dy, widths, height_reciprocals); \
1183 edge_dx_dy_alt = width_alt * height_reciprocal_alt; \
1184} \
1185
1186
1187#define setup_spans_adjust_y_up() \
1188 gvsub_u32(y_x4, y_x4, c_0x04) \
1189
1190#define setup_spans_adjust_y_down() \
1191 gvadd_u32(y_x4, y_x4, c_0x04) \
1192
1193#define setup_spans_adjust_interpolants_up() \
1194 gvsubq_u32(uvrg, uvrg, uvrg_dy); \
1195 b -= b_dy \
1196
1197#define setup_spans_adjust_interpolants_down() \
1198 gvaddq_u32(uvrg, uvrg, uvrg_dy); \
1199 b += b_dy \
1200
1201
1202#define setup_spans_clip_interpolants_increment() \
1203 gvmlaq_s32(uvrg, uvrg_dy, v_clip); \
1204 b += b_dy * clip \
1205
1206#define setup_spans_clip_interpolants_decrement() \
1207 gvmlsq_s32(uvrg, uvrg_dy, v_clip); \
1208 b -= b_dy * clip \
1209
1210#define setup_spans_clip_alternate_yes() \
1211 edge_alt += edge_dx_dy_alt * (s64)(clip) \
1212
1213#define setup_spans_clip_alternate_no() \
1214
1215#define setup_spans_clip(direction, alternate_active) \
1216{ \
1217 gvdupq_n_u32(v_clip, clip); \
1218 gvmlal_s32(edges_xy, edges_dx_dy, gvlo(v_clip)); \
1219 setup_spans_clip_alternate_##alternate_active(); \
1220 setup_spans_clip_interpolants_##direction(); \
1221} \
1222
1223
1224#define setup_spans_adjust_edges_alternate_no(left_half, right_half) \
1225{ \
1226 vec_2x64s edge_shifts_64; \
1227 vec_2x64s edges_dx_dy_64; \
1228 vec_1x64s left_x_hi, right_x_hi; \
1229 \
1230 gvmovl_s32(edge_shifts_64, edge_shifts); \
1231 gvshlq_u64(edges_xy, edges_xy, edge_shifts_64); \
1232 \
1233 gvmovl_s32(edges_dx_dy_64, edges_dx_dy); \
1234 gvshlq_u64(edges_dx_dy_64, edges_dx_dy_64, edge_shifts_64); \
1235 \
1236 gvdupq_l_s64(left_x, gv##left_half(edges_xy), 0); \
1237 gvdupq_l_s64(right_x, gv##right_half(edges_xy), 0); \
1238 \
1239 gvdupq_l_s64(left_dx_dy, gv##left_half(edges_dx_dy_64), 0); \
1240 gvdupq_l_s64(right_dx_dy, gv##right_half(edges_dx_dy_64), 0); \
1241 \
1242 gvadd_s64(left_x_hi, gvlo(left_x), gvlo(left_dx_dy)); \
1243 gvadd_s64(right_x_hi, gvlo(right_x), gvlo(right_dx_dy)); \
1244 \
1245 gvset_hi(left_x, left_x_hi); \
1246 gvset_hi(right_x, right_x_hi); \
1247 \
1248 gvaddq_s64(left_dx_dy, left_dx_dy, left_dx_dy); \
1249 gvaddq_s64(right_dx_dy, right_dx_dy, right_dx_dy); \
1250} \
1251
1252#define setup_spans_adjust_edges_alternate_yes(left_half, right_half) \
1253{ \
1254 setup_spans_adjust_edges_alternate_no(left_half, right_half); \
1255 s64 edge_dx_dy_alt_64; \
1256 vec_1x64s alternate_x_hi; \
1257 \
1258 gvdup_n_u16(y_mid_point, y_b); \
1259 \
1260 edge_alt <<= edge_shift_alt; \
1261 edge_dx_dy_alt_64 = (s64)edge_dx_dy_alt << edge_shift_alt; \
1262 \
1263 gvdupq_n_s64(alternate_x, edge_alt); \
1264 gvdupq_n_s64(alternate_dx_dy, edge_dx_dy_alt_64); \
1265 \
1266 gvadd_s64(alternate_x_hi, gvlo(alternate_x), gvlo(alternate_dx_dy)); \
1267 gvaddq_s64(alternate_dx_dy, alternate_dx_dy, alternate_dx_dy); \
1268 gvset_hi(alternate_x, alternate_x_hi); \
1269} \
1270
1271
1272#define setup_spans_y_select_up() \
1273 gvclt_s16(alternate_select, y_x4, y_mid_point) \
1274
1275#define setup_spans_y_select_down() \
1276 gvcgt_s16(alternate_select, y_x4, y_mid_point) \
1277
1278#define setup_spans_y_select_alternate_yes(direction) \
1279 setup_spans_y_select_##direction() \
1280
1281#define setup_spans_y_select_alternate_no(direction) \
1282
1283#define setup_spans_alternate_select_left() \
1284 gvbit(left_right_x_16_lo, alternate_x_16, alternate_select); \
1285
1286#define setup_spans_alternate_select_right() \
1287 gvbit(left_right_x_16_hi, alternate_x_16, alternate_select); \
1288
1289#define setup_spans_alternate_select_none() \
1290
1291#define setup_spans_increment_alternate_yes() \
1292{ \
1293 vec_2x32s alternate_x_32_lo, alternate_x_32_hi; \
1294 gvmovn_top_u64(alternate_x_32_lo, alternate_x); \
1295 gvaddq_s64(alternate_x, alternate_x, alternate_dx_dy); \
1296 gvmovn_top_u64(alternate_x_32_hi, alternate_x); \
1297 gvaddq_s64(alternate_x, alternate_x, alternate_dx_dy); \
1298 gvcombine_u32(alternate_x_32, alternate_x_32_lo, alternate_x_32_hi); \
1299 gvmovn_u32(alternate_x_16, alternate_x_32); \
1300} \
1301
1302#define setup_spans_increment_alternate_no() \
1303
1304#if defined(__SSE2__) && !(defined(__AVX512BW__) && defined(__AVX512VL__))
1305#define setup_spans_make_span_shift(span_shift) { \
1306 gvreg tab1_ = { .u8 = { 0xfe, 0xfc, 0xf8, 0xf0, 0xe0, 0xc0, 0x80, 0x00 } }; \
1307 gvtbl2_u8(span_shift, tab1_, span_shift); \
1308 gvorr_n_u16(span_shift, 0xff00); \
1309 (void)c_0xFFFE; \
1310}
1311#else
1312#define setup_spans_make_span_shift(span_shift) \
1313 gvshl_u16(span_shift, c_0xFFFE, span_shift)
1314#endif
1315
1316#define setup_spans_set_x4(alternate, direction, alternate_active) \
1317{ \
1318 gvst1q_pi_u32(uvrg, span_uvrg_offset); \
1319 *span_b_offset++ = b; \
1320 setup_spans_adjust_interpolants_##direction(); \
1321 \
1322 gvst1q_pi_u32(uvrg, span_uvrg_offset); \
1323 *span_b_offset++ = b; \
1324 setup_spans_adjust_interpolants_##direction(); \
1325 \
1326 gvst1q_pi_u32(uvrg, span_uvrg_offset); \
1327 *span_b_offset++ = b; \
1328 setup_spans_adjust_interpolants_##direction(); \
1329 \
1330 gvst1q_pi_u32(uvrg, span_uvrg_offset); \
1331 *span_b_offset++ = b; \
1332 setup_spans_adjust_interpolants_##direction(); \
1333 \
1334 gvmovn_top_u64(left_x_32_lo, left_x); \
1335 gvmovn_top_u64(right_x_32_lo, right_x); \
1336 \
1337 gvaddq_s64(left_x, left_x, left_dx_dy); \
1338 gvaddq_s64(right_x, right_x, right_dx_dy); \
1339 \
1340 gvmovn_top_u64(left_x_32_hi, left_x); \
1341 gvmovn_top_u64(right_x_32_hi, right_x); \
1342 \
1343 gvaddq_s64(left_x, left_x, left_dx_dy); \
1344 gvaddq_s64(right_x, right_x, right_dx_dy); \
1345 \
1346 gvcombine_s64(left_x_32, left_x_32_lo, left_x_32_hi); \
1347 gvcombine_s64(right_x_32, right_x_32_lo, right_x_32_hi); \
1348 \
1349 gvmovn_u32(left_right_x_16_lo, left_x_32); \
1350 gvmovn_u32(left_right_x_16_hi, right_x_32); \
1351 \
1352 setup_spans_increment_alternate_##alternate_active(); \
1353 setup_spans_y_select_alternate_##alternate_active(direction); \
1354 setup_spans_alternate_select_##alternate(); \
1355 \
1356 gvmax_s16(left_right_x_16_lo, left_right_x_16_lo, gvlo(left_edge)); \
1357 gvmax_s16(left_right_x_16_hi, left_right_x_16_hi, gvhi(left_edge)); \
1358 gvmin_s16(left_right_x_16_lo, left_right_x_16_lo, gvlo(right_edge)); \
1359 gvmin_s16(left_right_x_16_hi, left_right_x_16_hi, gvhi(right_edge)); \
1360 \
1361 gvsub_u16(left_right_x_16_hi, left_right_x_16_hi, left_right_x_16_lo); \
1362 gvadd_u16(left_right_x_16_hi, left_right_x_16_hi, c_0x07); \
1363 gvand(span_shift, left_right_x_16_hi, c_0x07); \
1364 setup_spans_make_span_shift(span_shift); \
1365 gvshr_n_u16(left_right_x_16_hi, left_right_x_16_hi, 3); \
1366 gvmin_u16(left_right_x_16_hi, left_right_x_16_hi, c_max_blocks_per_row); \
1367 \
1368 gvst4_pi_u16(left_right_x_16_lo, left_right_x_16_hi, span_shift, y_x4, \
1369 span_edge_data); \
1370 \
1371 setup_spans_adjust_y_##direction(); \
1372} \
1373
1374
1375#define setup_spans_alternate_adjust_yes() \
1376 edge_alt -= edge_dx_dy_alt * (s64)height_minor_a \
1377
1378#define setup_spans_alternate_adjust_no() \
1379
1380
1381#define setup_spans_down(left_half, right_half, alternate, alternate_active) \
1382 setup_spans_alternate_adjust_##alternate_active(); \
1383 if(y_c > psx_gpu->viewport_end_y) \
1384 height -= y_c - psx_gpu->viewport_end_y - 1; \
1385 \
1386 clip = psx_gpu->viewport_start_y - y_a; \
1387 if(clip > 0) \
1388 { \
1389 height -= clip; \
1390 y_a += clip; \
1391 setup_spans_clip(increment, alternate_active); \
1392 } \
1393 \
1394 setup_spans_prologue_b(); \
1395 \
1396 if (height > 512) \
1397 height = 512; \
1398 if (height > 0) \
1399 { \
1400 u64 y_x4_ = ((u64)(y_a + 3) << 48) | ((u64)(u16)(y_a + 2) << 32) \
1401 | (u32)((y_a + 1) << 16) | (u16)y_a; \
1402 gvcreate_u64(y_x4, y_x4_); \
1403 setup_spans_adjust_edges_alternate_##alternate_active(left_half, right_half); \
1404 \
1405 psx_gpu->num_spans = height; \
1406 do \
1407 { \
1408 setup_spans_set_x4(alternate, down, alternate_active); \
1409 height -= 4; \
1410 } while(height > 0); \
1411 if (psx_gpu->hacks_active & (AHACK_TEXTURE_ADJ_U | AHACK_TEXTURE_ADJ_V)) \
1412 { \
1413 vec_2x32u tmp; \
1414 gvld1_u64(tmp, &span_uvrg_offset[height - 2]); \
1415 gvst1_u64(tmp, &span_uvrg_offset[height - 1]); \
1416 } \
1417 } \
1418
1419
1420#define setup_spans_alternate_pre_increment_yes() \
1421 edge_alt += edge_dx_dy_alt \
1422
1423#define setup_spans_alternate_pre_increment_no() \
1424
1425#define setup_spans_up_decrement_height_yes() \
1426 height-- \
1427
1428#define setup_spans_up_decrement_height_no() \
1429 {} \
1430
1431#define setup_spans_up(left_half, right_half, alternate, alternate_active) \
1432 setup_spans_alternate_adjust_##alternate_active(); \
1433 y_a--; \
1434 \
1435 if(y_c < psx_gpu->viewport_start_y) \
1436 height -= psx_gpu->viewport_start_y - y_c; \
1437 else \
1438 setup_spans_up_decrement_height_##alternate_active(); \
1439 \
1440 clip = y_a - psx_gpu->viewport_end_y; \
1441 if(clip > 0) \
1442 { \
1443 height -= clip; \
1444 y_a -= clip; \
1445 setup_spans_clip(decrement, alternate_active); \
1446 } \
1447 \
1448 setup_spans_prologue_b(); \
1449 \
1450 if (height > 512) \
1451 height = 512; \
1452 if (height > 0) \
1453 { \
1454 u64 y_x4_ = ((u64)(y_a - 3) << 48) | ((u64)(u16)(y_a - 2) << 32) \
1455 | (u32)((y_a - 1) << 16) | (u16)y_a; \
1456 gvcreate_u64(y_x4, y_x4_); \
1457 gvaddw_s32(edges_xy, edges_xy, edges_dx_dy); \
1458 setup_spans_alternate_pre_increment_##alternate_active(); \
1459 setup_spans_adjust_edges_alternate_##alternate_active(left_half, right_half); \
1460 setup_spans_adjust_interpolants_up(); \
1461 \
1462 psx_gpu->num_spans = height; \
1463 while(height > 0) \
1464 { \
1465 setup_spans_set_x4(alternate, up, alternate_active); \
1466 height -= 4; \
1467 } \
1468 if (psx_gpu->hacks_active & AHACK_TEXTURE_ADJ_V) \
1469 { \
1470 vec_2x32u tmp; \
1471 gvld1_u64(tmp, &psx_gpu->span_uvrg_offset[1]); \
1472 gvst1_u64(tmp, &psx_gpu->span_uvrg_offset[0]); \
1473 } \
1474 } \
1475
1476#define half_left lo
1477#define half_right hi
1478
1479#define setup_spans_up_up(minor, major) \
1480 setup_spans_prologue(yes); \
1481 s32 height_minor_a = y_a - y_b; \
1482 s32 height_minor_b = y_b - y_c; \
1483 s32 height = y_a - y_c; \
1484 \
1485 gvdup_n_u32(x_starts, x_a); \
1486 gvcreate_u32(x_ends, x_c, x_b); \
1487 \
1488 compute_edge_delta_x3(x_b, height, height_minor_a); \
1489 setup_spans_up(half_##major, half_##minor, minor, yes) \
1490
1491void setup_spans_up_left(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1492 vertex_struct *v_b, vertex_struct *v_c)
1493{
1494#if 0
1495 setup_spans_up_left_(psx_gpu, v_a, v_b, v_c);
1496 return;
1497#endif
1498 setup_spans_up_up(left, right)
1499}
1500
1501void setup_spans_up_right(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1502 vertex_struct *v_b, vertex_struct *v_c)
1503{
1504#if 0
1505 setup_spans_up_right_(psx_gpu, v_a, v_b, v_c);
1506 return;
1507#endif
1508 setup_spans_up_up(right, left)
1509}
1510
1511#define setup_spans_down_down(minor, major) \
1512 setup_spans_prologue(yes); \
1513 s32 height_minor_a = y_b - y_a; \
1514 s32 height_minor_b = y_c - y_b; \
1515 s32 height = y_c - y_a; \
1516 \
1517 gvdup_n_u32(x_starts, x_a); \
1518 gvcreate_u32(x_ends, x_c, x_b); \
1519 \
1520 compute_edge_delta_x3(x_b, height, height_minor_a); \
1521 setup_spans_down(half_##major, half_##minor, minor, yes) \
1522
1523void setup_spans_down_left(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1524 vertex_struct *v_b, vertex_struct *v_c)
1525{
1526#if 0
1527 setup_spans_down_left_(psx_gpu, v_a, v_b, v_c);
1528 return;
1529#endif
1530 setup_spans_down_down(left, right)
1531}
1532
1533void setup_spans_down_right(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1534 vertex_struct *v_b, vertex_struct *v_c)
1535{
1536#if 0
1537 setup_spans_down_right_(psx_gpu, v_a, v_b, v_c);
1538 return;
1539#endif
1540 setup_spans_down_down(right, left)
1541}
1542
1543#define setup_spans_up_flat() \
1544 s32 height = y_a - y_c; \
1545 \
1546 compute_edge_delta_x2(); \
1547 setup_spans_up(half_left, half_right, none, no) \
1548
1549void setup_spans_up_a(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1550 vertex_struct *v_b, vertex_struct *v_c)
1551{
1552#if 0
1553 setup_spans_up_a_(psx_gpu, v_a, v_b, v_c);
1554 return;
1555#endif
1556 setup_spans_prologue(no);
1557
1558 gvcreate_u32(x_starts, x_a, x_b);
1559 gvdup_n_u32(x_ends, x_c);
1560
1561 setup_spans_up_flat()
1562}
1563
1564void setup_spans_up_b(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1565 vertex_struct *v_b, vertex_struct *v_c)
1566{
1567#if 0
1568 setup_spans_up_b_(psx_gpu, v_a, v_b, v_c);
1569 return;
1570#endif
1571 setup_spans_prologue(no);
1572
1573 gvdup_n_u32(x_starts, x_a);
1574 gvcreate_u32(x_ends, x_b, x_c);
1575
1576 setup_spans_up_flat()
1577}
1578
1579#define setup_spans_down_flat() \
1580 s32 height = y_c - y_a; \
1581 \
1582 compute_edge_delta_x2(); \
1583 setup_spans_down(half_left, half_right, none, no) \
1584
1585void setup_spans_down_a(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1586 vertex_struct *v_b, vertex_struct *v_c)
1587{
1588#if 0
1589 setup_spans_down_a_(psx_gpu, v_a, v_b, v_c);
1590 return;
1591#endif
1592 setup_spans_prologue(no);
1593
1594 gvcreate_u32(x_starts, x_a, x_b);
1595 gvdup_n_u32(x_ends, x_c);
1596
1597 setup_spans_down_flat()
1598}
1599
1600void setup_spans_down_b(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1601 vertex_struct *v_b, vertex_struct *v_c)
1602{
1603#if 0
1604 setup_spans_down_b_(psx_gpu, v_a, v_b, v_c);
1605 return;
1606#endif
1607 setup_spans_prologue(no)
1608
1609 gvdup_n_u32(x_starts, x_a);
1610 gvcreate_u32(x_ends, x_b, x_c);
1611
1612 setup_spans_down_flat()
1613}
1614
1615void setup_spans_up_down(psx_gpu_struct *psx_gpu, vertex_struct *v_a,
1616 vertex_struct *v_b, vertex_struct *v_c)
1617{
1618#if 0
1619 setup_spans_up_down_(psx_gpu, v_a, v_b, v_c);
1620 return;
1621#endif
1622 setup_spans_prologue(no);
1623
1624 s32 y_b = v_b->y;
1625 s64 edge_alt;
1626 s32 edge_dx_dy_alt;
1627 u32 edge_shift_alt;
1628
1629 s32 middle_y = y_a;
1630 s32 height_minor_a = y_a - y_b;
1631 s32 height_minor_b = y_c - y_a;
1632 s32 height_major = y_c - y_b;
1633
1634 vec_2x64s edges_xy_b;
1635 vec_1x64s edges_xy_b_left;
1636 vec_2x32s edges_dx_dy_b;
1637 vec_2x32u edge_shifts_b;
1638
1639 vec_2x32s height_increment;
1640
1641 gvcreate_u32(x_starts, x_a, x_c);
1642 gvdup_n_u32(x_ends, x_b);
1643
1644 compute_edge_delta_x3(x_a, height_minor_a, height_major);
1645
1646 gvcreate_s32(height_increment, 0, height_minor_b);
1647
1648 gvmlal_s32(edges_xy, edges_dx_dy, height_increment);
1649
1650 gvcreate_s64(edges_xy_b_left, edge_alt);
1651 gvcombine_s64(edges_xy_b, edges_xy_b_left, gvhi(edges_xy));
1652
1653 edge_shifts_b = edge_shifts;
1654 gvmov_l_u32(edge_shifts_b, edge_shift_alt, 0);
1655
1656 gvneg_s32(edges_dx_dy_b, edges_dx_dy);
1657 gvmov_l_s32(edges_dx_dy_b, edge_dx_dy_alt, 0);
1658
1659 y_a--;
1660
1661 if(y_b < psx_gpu->viewport_start_y)
1662 height_minor_a -= psx_gpu->viewport_start_y - y_b;
1663
1664 clip = y_a - psx_gpu->viewport_end_y;
1665 if(clip > 0)
1666 {
1667 height_minor_a -= clip;
1668 y_a -= clip;
1669 setup_spans_clip(decrement, no);
1670 }
1671
1672 setup_spans_prologue_b();
1673
1674 if (height_minor_a > 512)
1675 height_minor_a = 512;
1676 if (height_minor_a > 0)
1677 {
1678 u64 y_x4_ = ((u64)(y_a - 3) << 48) | ((u64)(u16)(y_a - 2) << 32)
1679 | (u32)((y_a - 1) << 16) | (u16)y_a;
1680 gvcreate_u64(y_x4, y_x4_);
1681 gvaddw_s32(edges_xy, edges_xy, edges_dx_dy);
1682 setup_spans_adjust_edges_alternate_no(lo, hi);
1683 setup_spans_adjust_interpolants_up();
1684
1685 psx_gpu->num_spans = height_minor_a;
1686 while(height_minor_a > 0)
1687 {
1688 setup_spans_set_x4(none, up, no);
1689 height_minor_a -= 4;
1690 }
1691
1692 span_edge_data += height_minor_a;
1693 span_uvrg_offset += height_minor_a;
1694 span_b_offset += height_minor_a;
1695 }
1696
1697 edges_xy = edges_xy_b;
1698 edges_dx_dy = edges_dx_dy_b;
1699 edge_shifts = edge_shifts_b;
1700
1701 gvld1q_u32(uvrg, psx_gpu->uvrg.e);
1702 b = psx_gpu->b;
1703
1704 y_a = middle_y;
1705
1706 if(y_c > psx_gpu->viewport_end_y)
1707 height_minor_b -= y_c - psx_gpu->viewport_end_y - 1;
1708
1709 clip = psx_gpu->viewport_start_y - y_a;
1710 if(clip > 0)
1711 {
1712 height_minor_b -= clip;
1713 y_a += clip;
1714 setup_spans_clip(increment, no);
1715 }
1716
1717 if (height_minor_b > 512)
1718 height_minor_b = 512;
1719 if (height_minor_b > 0)
1720 {
1721 u64 y_x4_ = ((u64)(y_a + 3) << 48) | ((u64)(u16)(y_a + 2) << 32)
1722 | (u32)((y_a + 1) << 16) | (u16)y_a;
1723 gvcreate_u64(y_x4, y_x4_);
1724 setup_spans_adjust_edges_alternate_no(lo, hi);
1725
1726 // FIXME: overflow corner case
1727 if(psx_gpu->num_spans + height_minor_b == MAX_SPANS)
1728 height_minor_b &= ~3;
1729
1730 psx_gpu->num_spans += height_minor_b;
1731 while(height_minor_b > 0)
1732 {
1733 setup_spans_set_x4(none, down, no);
1734 height_minor_b -= 4;
1735 }
1736 if (psx_gpu->hacks_active & (AHACK_TEXTURE_ADJ_U | AHACK_TEXTURE_ADJ_V))
1737 {
1738 vec_2x32u tmp;
1739 gvld1_u64(tmp, &span_uvrg_offset[height_minor_b - 2]);
1740 gvst1_u64(tmp, &span_uvrg_offset[height_minor_b - 1]);
1741 }
1742 }
1743}
1744
1745
1746#define dither_table_entry_normal(value) \
1747 (value) \
1748
1749#define setup_blocks_load_msb_mask_indirect() \
1750
1751#define setup_blocks_load_msb_mask_direct() \
1752 vec_8x16u msb_mask; \
1753 gvdupq_n_u16(msb_mask, psx_gpu->mask_msb); \
1754
1755#define setup_blocks_variables_shaded_textured(target) \
1756 vec_4x32u u_block; \
1757 vec_4x32u v_block; \
1758 vec_4x32u r_block; \
1759 vec_4x32u g_block; \
1760 vec_4x32u b_block; \
1761 vec_4x32u uvrg_dx; \
1762 vec_4x32u uvrg_dx4; \
1763 vec_4x32u uvrg_dx8; \
1764 vec_4x32u uvrg; \
1765 vec_16x8u texture_mask; \
1766 vec_8x8u texture_mask_lo, texture_mask_hi; \
1767 u32 b_dx = psx_gpu->b_block_span.e[1]; \
1768 u32 b_dx4 = b_dx << 2; \
1769 u32 b_dx8 = b_dx << 3; \
1770 u32 b; \
1771 \
1772 gvld1q_u32(uvrg_dx, psx_gpu->uvrg_dx.e); \
1773 gvshlq_n_u32(uvrg_dx4, uvrg_dx, 2); \
1774 gvshlq_n_u32(uvrg_dx8, uvrg_dx, 3); \
1775 gvld2_u8_dup(texture_mask_lo, texture_mask_hi, &psx_gpu->texture_mask_width); \
1776 gvcombine_u16(texture_mask, texture_mask_lo, texture_mask_hi) \
1777
1778#define setup_blocks_variables_shaded_untextured(target) \
1779 vec_4x32u r_block; \
1780 vec_4x32u g_block; \
1781 vec_4x32u b_block; \
1782 vec_4x32u rgb_dx; \
1783 vec_2x32u rgb_dx_lo, rgb_dx_hi; \
1784 vec_4x32u rgb_dx4; \
1785 vec_4x32u rgb_dx8; \
1786 vec_4x32u rgb; \
1787 vec_2x32u rgb_lo, rgb_hi; \
1788 \
1789 vec_8x8u d64_0x07; \
1790 vec_8x8u d64_1; \
1791 vec_8x8u d64_4; \
1792 vec_8x8u d64_128; \
1793 \
1794 gvdup_n_u8(d64_0x07, 0x07); \
1795 gvdup_n_u8(d64_1, 1); \
1796 gvdup_n_u8(d64_4, 4); \
1797 gvdup_n_u8(d64_128, 128u); \
1798 \
1799 gvld1_u32(rgb_dx_lo, &psx_gpu->uvrg_dx.e[2]); \
1800 gvcreate_u32(rgb_dx_hi, psx_gpu->b_block_span.e[1], 0); \
1801 gvcombine_u32(rgb_dx, rgb_dx_lo, rgb_dx_hi); \
1802 gvshlq_n_u32(rgb_dx4, rgb_dx, 2); \
1803 gvshlq_n_u32(rgb_dx8, rgb_dx, 3) \
1804
1805#define setup_blocks_variables_unshaded_textured(target) \
1806 vec_4x32u u_block; \
1807 vec_4x32u v_block; \
1808 vec_2x32u uv_dx; \
1809 vec_2x32u uv_dx4; \
1810 vec_2x32u uv_dx8; \
1811 vec_2x32u uv; \
1812 vec_16x8u texture_mask; \
1813 vec_8x8u texture_mask_lo, texture_mask_hi; \
1814 \
1815 gvld1_u32(uv_dx, psx_gpu->uvrg_dx.e); \
1816 gvld1_u32(uv, psx_gpu->uvrg.e); \
1817 gvshl_n_u32(uv_dx4, uv_dx, 2); \
1818 gvshl_n_u32(uv_dx8, uv_dx, 3); \
1819 gvld2_u8_dup(texture_mask_lo, texture_mask_hi, &psx_gpu->texture_mask_width); \
1820 gvcombine_u16(texture_mask, texture_mask_lo, texture_mask_hi) \
1821
1822#define setup_blocks_variables_unshaded_untextured_direct() \
1823 gvorrq(colors, colors, msb_mask) \
1824
1825#define setup_blocks_variables_unshaded_untextured_indirect() \
1826
1827#define setup_blocks_variables_unshaded_untextured(target) \
1828 u32 color = psx_gpu->triangle_color; \
1829 vec_8x16u colors; \
1830 \
1831 u32 color_r = color & 0xFF; \
1832 u32 color_g = (color >> 8) & 0xFF; \
1833 u32 color_b = (color >> 16) & 0xFF; \
1834 \
1835 color = (color_r >> 3) | ((color_g >> 3) << 5) | \
1836 ((color_b >> 3) << 10); \
1837 gvdupq_n_u16(colors, color); \
1838 setup_blocks_variables_unshaded_untextured_##target() \
1839
1840#define setup_blocks_span_initialize_dithered_textured() \
1841 vec_8x16u dither_offsets; \
1842 gvshll_n_s8(dither_offsets, dither_offsets_short, 4) \
1843
1844#define setup_blocks_span_initialize_dithered_untextured() \
1845 vec_8x8u dither_offsets; \
1846 gvadd_u8(dither_offsets, dither_offsets_short, d64_4) \
1847
1848#define setup_blocks_span_initialize_dithered(texturing) \
1849 u32 dither_row = psx_gpu->dither_table[y & 0x3]; \
1850 u32 dither_shift = (span_edge_data->left_x & 0x3) * 8; \
1851 vec_8x8s dither_offsets_short; \
1852 \
1853 dither_row = \
1854 (dither_row >> dither_shift) | ((u64)dither_row << (32 - dither_shift)); \
1855 gvdup_n_u32(dither_offsets_short, dither_row); \
1856 setup_blocks_span_initialize_dithered_##texturing() \
1857
1858#define setup_blocks_span_initialize_undithered(texturing) \
1859
1860#define setup_blocks_span_initialize_shaded_textured() \
1861{ \
1862 u32 left_x = span_edge_data->left_x; \
1863 vec_4x32u block_span; \
1864 vec_4x32u v_left_x; \
1865 \
1866 gvld1q_u32(uvrg, span_uvrg_offset); \
1867 gvdupq_n_u32(v_left_x, left_x); \
1868 gvmlaq_u32(uvrg, uvrg_dx, v_left_x); \
1869 b = *span_b_offset; \
1870 b += b_dx * left_x; \
1871 \
1872 gvdupq_l_u32(u_block, gvlo(uvrg), 0); \
1873 gvdupq_l_u32(v_block, gvlo(uvrg), 1); \
1874 gvdupq_l_u32(r_block, gvhi(uvrg), 0); \
1875 gvdupq_l_u32(g_block, gvhi(uvrg), 1); \
1876 gvdupq_n_u32(b_block, b); \
1877 \
1878 gvld1q_u32(block_span, psx_gpu->u_block_span.e); \
1879 gvaddq_u32(u_block, u_block, block_span); \
1880 gvld1q_u32(block_span, psx_gpu->v_block_span.e); \
1881 gvaddq_u32(v_block, v_block, block_span); \
1882 gvld1q_u32(block_span, psx_gpu->r_block_span.e); \
1883 gvaddq_u32(r_block, r_block, block_span); \
1884 gvld1q_u32(block_span, psx_gpu->g_block_span.e); \
1885 gvaddq_u32(g_block, g_block, block_span); \
1886 gvld1q_u32(block_span, psx_gpu->b_block_span.e); \
1887 gvaddq_u32(b_block, b_block, block_span); \
1888}
1889
1890#define setup_blocks_span_initialize_shaded_untextured() \
1891{ \
1892 u32 left_x = span_edge_data->left_x; \
1893 u32 *span_uvrg_offset_high = (u32 *)span_uvrg_offset + 2; \
1894 vec_4x32u block_span; \
1895 vec_4x32u v_left_x; \
1896 \
1897 gvld1_u32(rgb_lo, span_uvrg_offset_high); \
1898 gvcreate_u32(rgb_hi, *span_b_offset, 0); \
1899 gvcombine_u32(rgb, rgb_lo, rgb_hi); \
1900 gvdupq_n_u32(v_left_x, left_x); \
1901 gvmlaq_u32(rgb, rgb_dx, v_left_x); \
1902 \
1903 gvdupq_l_u32(r_block, gvlo(rgb), 0); \
1904 gvdupq_l_u32(g_block, gvlo(rgb), 1); \
1905 gvdupq_l_u32(b_block, gvhi(rgb), 0); \
1906 \
1907 gvld1q_u32(block_span, psx_gpu->r_block_span.e); \
1908 gvaddq_u32(r_block, r_block, block_span); \
1909 gvld1q_u32(block_span, psx_gpu->g_block_span.e); \
1910 gvaddq_u32(g_block, g_block, block_span); \
1911 gvld1q_u32(block_span, psx_gpu->b_block_span.e); \
1912 gvaddq_u32(b_block, b_block, block_span); \
1913} \
1914
1915#define setup_blocks_span_initialize_unshaded_textured() \
1916{ \
1917 u32 left_x = span_edge_data->left_x; \
1918 vec_4x32u block_span; \
1919 vec_2x32u v_left_x; \
1920 \
1921 gvld1_u32(uv, span_uvrg_offset); \
1922 gvdup_n_u32(v_left_x, left_x); \
1923 gvmla_u32(uv, uv_dx, v_left_x); \
1924 \
1925 gvdupq_l_u32(u_block, uv, 0); \
1926 gvdupq_l_u32(v_block, uv, 1); \
1927 \
1928 gvld1q_u32(block_span, psx_gpu->u_block_span.e); \
1929 gvaddq_u32(u_block, u_block, block_span); \
1930 gvld1q_u32(block_span, psx_gpu->v_block_span.e); \
1931 gvaddq_u32(v_block, v_block, block_span); \
1932 (void)(span_b_offset); \
1933} \
1934
1935#define setup_blocks_span_initialize_unshaded_untextured() \
1936 (void)(span_uvrg_offset); \
1937 (void)(span_b_offset) \
1938
1939#define setup_blocks_texture_swizzled() \
1940{ \
1941 vec_8x8u u_saved = u; \
1942 gvsli_n_u8(u, v, 4); \
1943 gvsri_n_u8(v, u_saved, 4); \
1944} \
1945
1946#define setup_blocks_texture_unswizzled() \
1947
1948#define setup_blocks_store_shaded_textured(swizzling, dithering, target, \
1949 edge_type) \
1950{ \
1951 vec_8x16u u_whole; \
1952 vec_8x16u v_whole; \
1953 vec_8x16u r_whole; \
1954 vec_8x16u g_whole; \
1955 vec_8x16u b_whole; \
1956 vec_4x16u u_whole_lo, u_whole_hi; \
1957 vec_4x16u v_whole_lo, v_whole_hi; \
1958 vec_4x16u r_whole_lo, r_whole_hi; \
1959 vec_4x16u g_whole_lo, g_whole_hi; \
1960 vec_4x16u b_whole_lo, b_whole_hi; \
1961 \
1962 vec_8x8u u; \
1963 vec_8x8u v; \
1964 vec_8x8u r; \
1965 vec_8x8u g; \
1966 vec_8x8u b; \
1967 \
1968 vec_4x32u dx4; \
1969 vec_4x32u dx8; \
1970 \
1971 gvshrn_n_u32(u_whole_lo, u_block, 16); \
1972 gvshrn_n_u32(v_whole_lo, v_block, 16); \
1973 gvshrn_n_u32(r_whole_lo, r_block, 16); \
1974 gvshrn_n_u32(g_whole_lo, g_block, 16); \
1975 gvshrn_n_u32(b_whole_lo, b_block, 16); \
1976 \
1977 gvdupq_l_u32(dx4, gvlo(uvrg_dx4), 0); \
1978 gvaddhn_u32(u_whole_hi, u_block, dx4); \
1979 gvdupq_l_u32(dx4, gvlo(uvrg_dx4), 1); \
1980 gvaddhn_u32(v_whole_hi, v_block, dx4); \
1981 gvdupq_l_u32(dx4, gvhi(uvrg_dx4), 0); \
1982 gvaddhn_u32(r_whole_hi, r_block, dx4); \
1983 gvdupq_l_u32(dx4, gvhi(uvrg_dx4), 1); \
1984 gvaddhn_u32(g_whole_hi, g_block, dx4); \
1985 gvdupq_n_u32(dx4, b_dx4); \
1986 gvaddhn_u32(b_whole_hi, b_block, dx4); \
1987 \
1988 gvcombine_u16(u_whole, u_whole_lo, u_whole_hi); \
1989 gvcombine_u16(v_whole, v_whole_lo, v_whole_hi); \
1990 gvcombine_u16(r_whole, r_whole_lo, r_whole_hi); \
1991 gvcombine_u16(g_whole, g_whole_lo, g_whole_hi); \
1992 gvcombine_u16(b_whole, b_whole_lo, b_whole_hi); \
1993 gvmovn_u16(u, u_whole); \
1994 gvmovn_u16(v, v_whole); \
1995 gvmovn_u16(r, r_whole); \
1996 gvmovn_u16(g, g_whole); \
1997 gvmovn_u16(b, b_whole); \
1998 \
1999 gvdupq_l_u32(dx8, gvlo(uvrg_dx8), 0); \
2000 gvaddq_u32(u_block, u_block, dx8); \
2001 gvdupq_l_u32(dx8, gvlo(uvrg_dx8), 1); \
2002 gvaddq_u32(v_block, v_block, dx8); \
2003 gvdupq_l_u32(dx8, gvhi(uvrg_dx8), 0); \
2004 gvaddq_u32(r_block, r_block, dx8); \
2005 gvdupq_l_u32(dx8, gvhi(uvrg_dx8), 1); \
2006 gvaddq_u32(g_block, g_block, dx8); \
2007 gvdupq_n_u32(dx8, b_dx8); \
2008 gvaddq_u32(b_block, b_block, dx8); \
2009 \
2010 gvand(u, u, gvlo(texture_mask)); \
2011 gvand(v, v, gvhi(texture_mask)); \
2012 setup_blocks_texture_##swizzling(); \
2013 \
2014 gvst2_u8(u, v, (u8 *)block->uv.e); \
2015 gvst1_u8(r, block->r.e); \
2016 gvst1_u8(g, block->g.e); \
2017 gvst1_u8(b, block->b.e); \
2018 gvst1q_u16(dither_offsets, (u16 *)block->dither_offsets.e); \
2019 block->fb_ptr = fb_ptr; \
2020} \
2021
2022#define setup_blocks_store_unshaded_textured(swizzling, dithering, target, \
2023 edge_type) \
2024{ \
2025 vec_8x16u u_whole; \
2026 vec_8x16u v_whole; \
2027 vec_4x16u u_whole_lo, u_whole_hi; \
2028 vec_4x16u v_whole_lo, v_whole_hi; \
2029 \
2030 vec_8x8u u; \
2031 vec_8x8u v; \
2032 \
2033 vec_4x32u dx4; \
2034 vec_4x32u dx8; \
2035 \
2036 gvshrn_n_u32(u_whole_lo, u_block, 16); \
2037 gvshrn_n_u32(v_whole_lo, v_block, 16); \
2038 \
2039 gvdupq_l_u32(dx4, uv_dx4, 0); \
2040 gvaddhn_u32(u_whole_hi, u_block, dx4); \
2041 gvdupq_l_u32(dx4, uv_dx4, 1); \
2042 gvaddhn_u32(v_whole_hi, v_block, dx4); \
2043 \
2044 gvcombine_u16(u_whole, u_whole_lo, u_whole_hi); \
2045 gvcombine_u16(v_whole, v_whole_lo, v_whole_hi); \
2046 gvmovn_u16(u, u_whole); \
2047 gvmovn_u16(v, v_whole); \
2048 \
2049 gvdupq_l_u32(dx8, uv_dx8, 0); \
2050 gvaddq_u32(u_block, u_block, dx8); \
2051 gvdupq_l_u32(dx8, uv_dx8, 1); \
2052 gvaddq_u32(v_block, v_block, dx8); \
2053 \
2054 gvand(u, u, gvlo(texture_mask)); \
2055 gvand(v, v, gvhi(texture_mask)); \
2056 setup_blocks_texture_##swizzling(); \
2057 \
2058 gvst2_u8(u, v, (u8 *)block->uv.e); \
2059 gvst1q_u16(dither_offsets, (u16 *)block->dither_offsets.e); \
2060 block->fb_ptr = fb_ptr; \
2061} \
2062
2063#define setup_blocks_store_shaded_untextured_dithered() \
2064 gvqadd_u8(r, r, dither_offsets); \
2065 gvqadd_u8(g, g, dither_offsets); \
2066 gvqadd_u8(b, b, dither_offsets); \
2067 \
2068 gvqsub_u8(r, r, d64_4); \
2069 gvqsub_u8(g, g, d64_4); \
2070 gvqsub_u8(b, b, d64_4) \
2071
2072#define setup_blocks_store_shaded_untextured_undithered() \
2073
2074#define setup_blocks_store_untextured_pixels_indirect_full(_pixels) \
2075 gvst1q_u16(_pixels, block->pixels.e); \
2076 block->fb_ptr = fb_ptr \
2077
2078#define setup_blocks_store_untextured_pixels_indirect_edge(_pixels) \
2079 gvst1q_u16(_pixels, block->pixels.e); \
2080 block->fb_ptr = fb_ptr \
2081
2082#define setup_blocks_store_shaded_untextured_seed_pixels_indirect() \
2083 gvmull_u8(pixels, r, d64_1) \
2084
2085#define setup_blocks_store_untextured_pixels_direct_full(_pixels) \
2086 gvst1q_u16(_pixels, fb_ptr) \
2087
2088#define setup_blocks_store_untextured_pixels_direct_edge(_pixels) \
2089{ \
2090 vec_8x16u fb_pixels; \
2091 vec_8x16u draw_mask; \
2092 vec_8x16u test_mask; \
2093 \
2094 gvld1q_u16(test_mask, psx_gpu->test_mask.e); \
2095 gvld1q_u16(fb_pixels, fb_ptr); \
2096 gvdupq_n_u16(draw_mask, span_edge_data->right_mask); \
2097 gvtstq_u16(draw_mask, draw_mask, test_mask); \
2098 gvbifq(fb_pixels, _pixels, draw_mask); \
2099 gvst1q_u16(fb_pixels, fb_ptr); \
2100} \
2101
2102#define setup_blocks_store_shaded_untextured_seed_pixels_direct() \
2103 pixels = msb_mask; \
2104 gvmlal_u8(pixels, r, d64_1) \
2105
2106#define setup_blocks_store_shaded_untextured(swizzling, dithering, target, \
2107 edge_type) \
2108{ \
2109 vec_8x16u r_whole; \
2110 vec_8x16u g_whole; \
2111 vec_8x16u b_whole; \
2112 vec_4x16u r_whole_lo, r_whole_hi; \
2113 vec_4x16u g_whole_lo, g_whole_hi; \
2114 vec_4x16u b_whole_lo, b_whole_hi; \
2115 \
2116 vec_8x8u r; \
2117 vec_8x8u g; \
2118 vec_8x8u b; \
2119 \
2120 vec_4x32u dx4; \
2121 vec_4x32u dx8; \
2122 \
2123 vec_8x16u pixels; \
2124 \
2125 gvshrn_n_u32(r_whole_lo, r_block, 16); \
2126 gvshrn_n_u32(g_whole_lo, g_block, 16); \
2127 gvshrn_n_u32(b_whole_lo, b_block, 16); \
2128 \
2129 gvdupq_l_u32(dx4, gvlo(rgb_dx4), 0); \
2130 gvaddhn_u32(r_whole_hi, r_block, dx4); \
2131 gvdupq_l_u32(dx4, gvlo(rgb_dx4), 1); \
2132 gvaddhn_u32(g_whole_hi, g_block, dx4); \
2133 gvdupq_l_u32(dx4, gvhi(rgb_dx4), 0); \
2134 gvaddhn_u32(b_whole_hi, b_block, dx4); \
2135 \
2136 gvcombine_u16(r_whole, r_whole_lo, r_whole_hi); \
2137 gvcombine_u16(g_whole, g_whole_lo, g_whole_hi); \
2138 gvcombine_u16(b_whole, b_whole_lo, b_whole_hi); \
2139 gvmovn_u16(r, r_whole); \
2140 gvmovn_u16(g, g_whole); \
2141 gvmovn_u16(b, b_whole); \
2142 \
2143 gvdupq_l_u32(dx8, gvlo(rgb_dx8), 0); \
2144 gvaddq_u32(r_block, r_block, dx8); \
2145 gvdupq_l_u32(dx8, gvlo(rgb_dx8), 1); \
2146 gvaddq_u32(g_block, g_block, dx8); \
2147 gvdupq_l_u32(dx8, gvhi(rgb_dx8), 0); \
2148 gvaddq_u32(b_block, b_block, dx8); \
2149 \
2150 setup_blocks_store_shaded_untextured_##dithering(); \
2151 \
2152 gvshr_n_u8(r, r, 3); \
2153 gvbic(g, g, d64_0x07); \
2154 gvbic(b, b, d64_0x07); \
2155 \
2156 setup_blocks_store_shaded_untextured_seed_pixels_##target(); \
2157 gvmlal_u8(pixels, g, d64_4); \
2158 gvmlal_u8(pixels, b, d64_128); \
2159 \
2160 setup_blocks_store_untextured_pixels_##target##_##edge_type(pixels); \
2161} \
2162
2163#define setup_blocks_store_unshaded_untextured(swizzling, dithering, target, \
2164 edge_type) \
2165 setup_blocks_store_untextured_pixels_##target##_##edge_type(colors) \
2166
2167#define setup_blocks_store_draw_mask_textured_indirect(_block, bits) \
2168 (_block)->draw_mask_bits = bits \
2169
2170#define setup_blocks_store_draw_mask_untextured_indirect(_block, bits) \
2171{ \
2172 vec_8x16u bits_mask; \
2173 vec_8x16u test_mask; \
2174 \
2175 gvld1q_u16(test_mask, psx_gpu->test_mask.e); \
2176 gvdupq_n_u16(bits_mask, bits); \
2177 gvtstq_u16(bits_mask, bits_mask, test_mask); \
2178 gvst1q_u16(bits_mask, (_block)->draw_mask.e); \
2179} \
2180
2181#define setup_blocks_store_draw_mask_untextured_direct(_block, bits) \
2182 (void)(_block) \
2183
2184#define setup_blocks_uv_adj_hack_untextured(_block, edge_data, uvrg_offset) \
2185
2186#define setup_blocks_uv_adj_hack_textured(_block, edge_data, uvrg_offset) \
2187{ \
2188 u32 m_ = AHACK_TEXTURE_ADJ_U | AHACK_TEXTURE_ADJ_V; \
2189 if (unlikely(psx_gpu->hacks_active & m_)) \
2190 setup_blocks_uv_adj_hack(psx_gpu, _block, edge_data, (void *)uvrg_offset); \
2191} \
2192
2193#define setup_blocks_add_blocks_indirect() \
2194 num_blocks += span_num_blocks; \
2195 \
2196 if(num_blocks > MAX_BLOCKS) \
2197 { \
2198 psx_gpu->num_blocks = num_blocks - span_num_blocks; \
2199 flush_render_block_buffer(psx_gpu); \
2200 num_blocks = span_num_blocks; \
2201 block = psx_gpu->blocks; \
2202 } \
2203
2204#define setup_blocks_add_blocks_direct() \
2205
2206#define setup_blocks_do(shading, texturing, dithering, sw, target) \
2207 setup_blocks_load_msb_mask_##target(); \
2208 setup_blocks_variables_##shading##_##texturing(target); \
2209 \
2210 edge_data_struct *span_edge_data = psx_gpu->span_edge_data; \
2211 vec_4x32u *span_uvrg_offset = (vec_4x32u *)psx_gpu->span_uvrg_offset; \
2212 u32 *span_b_offset = psx_gpu->span_b_offset; \
2213 \
2214 block_struct *block = psx_gpu->blocks + psx_gpu->num_blocks; \
2215 \
2216 u32 num_spans = psx_gpu->num_spans; \
2217 \
2218 u16 * __restrict__ fb_ptr; \
2219 u32 y; \
2220 \
2221 u32 num_blocks = psx_gpu->num_blocks; \
2222 u32 span_num_blocks; \
2223 \
2224 while(num_spans) \
2225 { \
2226 span_num_blocks = span_edge_data->num_blocks; \
2227 if(span_num_blocks) \
2228 { \
2229 y = span_edge_data->y; \
2230 fb_ptr = psx_gpu->vram_out_ptr + span_edge_data->left_x + (y * 1024); \
2231 \
2232 setup_blocks_span_initialize_##shading##_##texturing(); \
2233 setup_blocks_span_initialize_##dithering(texturing); \
2234 \
2235 setup_blocks_add_blocks_##target(); \
2236 \
2237 span_num_blocks--; \
2238 while(span_num_blocks) \
2239 { \
2240 setup_blocks_store_##shading##_##texturing(sw, dithering, target, \
2241 full); \
2242 setup_blocks_store_draw_mask_##texturing##_##target(block, 0x00); \
2243 \
2244 fb_ptr += 8; \
2245 block++; \
2246 span_num_blocks--; \
2247 } \
2248 \
2249 setup_blocks_store_##shading##_##texturing(sw, dithering, target, edge); \
2250 setup_blocks_store_draw_mask_##texturing##_##target(block, \
2251 span_edge_data->right_mask); \
2252 setup_blocks_uv_adj_hack_##texturing(block, span_edge_data, \
2253 span_uvrg_offset); \
2254 \
2255 block++; \
2256 } \
2257 \
2258 num_spans--; \
2259 span_edge_data++; \
2260 span_uvrg_offset++; \
2261 span_b_offset++; \
2262 } \
2263 \
2264 psx_gpu->num_blocks = num_blocks \
2265
2266void setup_blocks_shaded_textured_dithered_swizzled_indirect(psx_gpu_struct
2267 *psx_gpu)
2268{
2269#if 0
2270 setup_blocks_shaded_textured_dithered_swizzled_indirect_(psx_gpu);
2271 return;
2272#endif
2273 setup_blocks_do(shaded, textured, dithered, swizzled, indirect);
2274}
2275
2276void setup_blocks_shaded_textured_dithered_unswizzled_indirect(psx_gpu_struct
2277 *psx_gpu)
2278{
2279#if 0
2280 setup_blocks_shaded_textured_dithered_unswizzled_indirect_(psx_gpu);
2281 return;
2282#endif
2283 setup_blocks_do(shaded, textured, dithered, unswizzled, indirect);
2284}
2285
2286void setup_blocks_unshaded_textured_dithered_swizzled_indirect(psx_gpu_struct
2287 *psx_gpu)
2288{
2289#if 0
2290 setup_blocks_unshaded_textured_dithered_swizzled_indirect_(psx_gpu);
2291 return;
2292#endif
2293 setup_blocks_do(unshaded, textured, dithered, swizzled, indirect);
2294}
2295
2296void setup_blocks_unshaded_textured_dithered_unswizzled_indirect(psx_gpu_struct
2297 *psx_gpu)
2298{
2299#if 0
2300 setup_blocks_unshaded_textured_dithered_unswizzled_indirect_(psx_gpu);
2301 return;
2302#endif
2303 setup_blocks_do(unshaded, textured, dithered, unswizzled, indirect);
2304}
2305
2306void setup_blocks_unshaded_untextured_undithered_unswizzled_indirect(
2307 psx_gpu_struct *psx_gpu)
2308{
2309#if 0
2310 setup_blocks_unshaded_untextured_undithered_unswizzled_indirect_(psx_gpu);
2311 return;
2312#endif
2313 setup_blocks_do(unshaded, untextured, undithered, unswizzled, indirect);
2314}
2315
2316void setup_blocks_unshaded_untextured_undithered_unswizzled_direct(
2317 psx_gpu_struct *psx_gpu)
2318{
2319#if 0
2320 setup_blocks_unshaded_untextured_undithered_unswizzled_direct_(psx_gpu);
2321 return;
2322#endif
2323 setup_blocks_do(unshaded, untextured, undithered, unswizzled, direct);
2324}
2325
2326void setup_blocks_shaded_untextured_undithered_unswizzled_indirect(psx_gpu_struct
2327 *psx_gpu)
2328{
2329#if 0
2330 setup_blocks_shaded_untextured_undithered_unswizzled_indirect_(psx_gpu);
2331 return;
2332#endif
2333 setup_blocks_do(shaded, untextured, undithered, unswizzled, indirect);
2334}
2335
2336void setup_blocks_shaded_untextured_dithered_unswizzled_indirect(psx_gpu_struct
2337 *psx_gpu)
2338{
2339#if 0
2340 setup_blocks_shaded_untextured_dithered_unswizzled_indirect_(psx_gpu);
2341 return;
2342#endif
2343 setup_blocks_do(shaded, untextured, dithered, unswizzled, indirect);
2344}
2345
2346void setup_blocks_shaded_untextured_undithered_unswizzled_direct(
2347 psx_gpu_struct *psx_gpu)
2348{
2349#if 0
2350 setup_blocks_shaded_untextured_undithered_unswizzled_direct_(psx_gpu);
2351 return;
2352#endif
2353 setup_blocks_do(shaded, untextured, undithered, unswizzled, direct);
2354}
2355
2356void setup_blocks_shaded_untextured_dithered_unswizzled_direct(psx_gpu_struct
2357 *psx_gpu)
2358{
2359#if 0
2360 setup_blocks_shaded_untextured_dithered_unswizzled_direct_(psx_gpu);
2361 return;
2362#endif
2363 setup_blocks_do(shaded, untextured, dithered, unswizzled, direct);
2364}
2365
2366static void update_texture_4bpp_cache(psx_gpu_struct *psx_gpu)
2367{
2368 u32 current_texture_page = psx_gpu->current_texture_page;
2369 u8 *texture_page_ptr = psx_gpu->texture_page_base;
2370 const u16 *vram_ptr = psx_gpu->vram_ptr;
2371 u32 tile_x, tile_y;
2372 u32 sub_y;
2373 vec_8x16u c_0x00f0;
2374
2375 vram_ptr += (current_texture_page >> 4) * 256 * 1024;
2376 vram_ptr += (current_texture_page & 0xF) * 64;
2377
2378 gvdupq_n_u16(c_0x00f0, 0x00f0);
2379
2380 psx_gpu->dirty_textures_4bpp_mask &= ~(psx_gpu->current_texture_mask);
2381
2382 for (tile_y = 16; tile_y; tile_y--)
2383 {
2384 for (tile_x = 16; tile_x; tile_x--)
2385 {
2386 for (sub_y = 8; sub_y; sub_y--)
2387 {
2388 vec_8x8u texel_block_a, texel_block_b;
2389 vec_8x16u texel_block_expanded_a, texel_block_expanded_b;
2390 vec_8x16u texel_block_expanded_c, texel_block_expanded_d;
2391 vec_8x16u texel_block_expanded_ab, texel_block_expanded_cd;
2392
2393 gvld1_u8(texel_block_a, (u8 *)vram_ptr); vram_ptr += 1024;
2394 gvld1_u8(texel_block_b, (u8 *)vram_ptr); vram_ptr += 1024;
2395
2396 gvmovl_u8(texel_block_expanded_a, texel_block_a);
2397 gvshll_n_u8(texel_block_expanded_b, texel_block_a, 4);
2398 gvmovl_u8(texel_block_expanded_c, texel_block_b);
2399 gvshll_n_u8(texel_block_expanded_d, texel_block_b, 4);
2400
2401 gvbicq(texel_block_expanded_a, texel_block_expanded_a, c_0x00f0);
2402 gvbicq(texel_block_expanded_b, texel_block_expanded_b, c_0x00f0);
2403 gvbicq(texel_block_expanded_c, texel_block_expanded_c, c_0x00f0);
2404 gvbicq(texel_block_expanded_d, texel_block_expanded_d, c_0x00f0);
2405
2406 gvorrq(texel_block_expanded_ab, texel_block_expanded_a, texel_block_expanded_b);
2407 gvorrq(texel_block_expanded_cd, texel_block_expanded_c, texel_block_expanded_d);
2408
2409 gvst1q_2_pi_u32(texel_block_expanded_ab, texel_block_expanded_cd, texture_page_ptr);
2410 }
2411
2412 vram_ptr -= (1024 * 16) - 4;
2413 }
2414
2415 vram_ptr += (16 * 1024) - (4 * 16);
2416 }
2417}
2418
2419void update_texture_8bpp_cache_slice(psx_gpu_struct *psx_gpu,
2420 u32 texture_page)
2421{
2422#if 0
2423 update_texture_8bpp_cache_slice_(psx_gpu, texture_page);
2424 return;
2425#endif
2426 u16 *texture_page_ptr = psx_gpu->texture_page_base;
2427 u16 *vram_ptr = psx_gpu->vram_ptr;
2428
2429 u32 tile_x, tile_y;
2430 u32 sub_y;
2431
2432 vram_ptr += (texture_page >> 4) * 256 * 1024;
2433 vram_ptr += (texture_page & 0xF) * 64;
2434
2435 if((texture_page ^ psx_gpu->current_texture_page) & 0x1)
2436 texture_page_ptr += (8 * 16) * 8;
2437
2438 for (tile_y = 16; tile_y; tile_y--)
2439 {
2440 for (tile_x = 8; tile_x; tile_x--)
2441 {
2442 for (sub_y = 4; sub_y; sub_y--)
2443 {
2444 vec_4x32u texels_a, texels_b, texels_c, texels_d = {};
2445 gvld1q_u32(texels_a, vram_ptr); vram_ptr += 1024;
2446 gvld1q_u32(texels_b, vram_ptr); vram_ptr += 1024;
2447 gvld1q_u32(texels_c, vram_ptr); vram_ptr += 1024;
2448 gvld1q_u32(texels_d, vram_ptr); vram_ptr += 1024;
2449
2450 gvst1q_2_pi_u32(texels_a, texels_b, texture_page_ptr);
2451 gvst1q_2_pi_u32(texels_c, texels_d, texture_page_ptr);
2452 }
2453
2454 vram_ptr -= (1024 * 16) - 8;
2455 }
2456
2457 vram_ptr -= (8 * 8);
2458 vram_ptr += (16 * 1024);
2459
2460 texture_page_ptr += (8 * 16) * 8;
2461 }
2462}
2463
2464void texture_blocks_untextured(psx_gpu_struct *psx_gpu)
2465{
2466}
2467
2468void texture_blocks_4bpp(psx_gpu_struct *psx_gpu)
2469{
2470#if 0
2471 texture_blocks_4bpp_(psx_gpu);
2472 return;
2473#endif
2474 block_struct *block = psx_gpu->blocks;
2475 u32 num_blocks = psx_gpu->num_blocks;
2476
2477 vec_8x8u texels_low;
2478 vec_8x8u texels_high;
2479
2480 vec_16x8u clut_low;
2481 vec_16x8u clut_high;
2482
2483 const u8 *texture_ptr_8bpp = psx_gpu->texture_page_ptr;
2484
2485 gvld2q_u8(clut_low, clut_high, (u8 *)psx_gpu->clut_ptr);
2486
2487 if(psx_gpu->current_texture_mask & psx_gpu->dirty_textures_4bpp_mask)
2488 update_texture_4bpp_cache(psx_gpu);
2489
2490 while(num_blocks)
2491 {
2492 vec_8x8u texels =
2493 {
2494 .u8 =
2495 {
2496 texture_ptr_8bpp[block->uv.e[0]],
2497 texture_ptr_8bpp[block->uv.e[1]],
2498 texture_ptr_8bpp[block->uv.e[2]],
2499 texture_ptr_8bpp[block->uv.e[3]],
2500 texture_ptr_8bpp[block->uv.e[4]],
2501 texture_ptr_8bpp[block->uv.e[5]],
2502 texture_ptr_8bpp[block->uv.e[6]],
2503 texture_ptr_8bpp[block->uv.e[7]]
2504 }
2505 };
2506
2507 gvtbl2_u8(texels_low, clut_low, texels);
2508 gvtbl2_u8(texels_high, clut_high, texels);
2509
2510 gvst2_u8(texels_low, texels_high, (u8 *)block->texels.e);
2511
2512 num_blocks--;
2513 block++;
2514 }
2515}
2516
2517void texture_blocks_8bpp(psx_gpu_struct *psx_gpu)
2518{
2519#if 0
2520 texture_blocks_8bpp_(psx_gpu);
2521 return;
2522#endif
2523 u32 num_blocks = psx_gpu->num_blocks;
2524
2525 if(psx_gpu->current_texture_mask & psx_gpu->dirty_textures_8bpp_mask)
2526 update_texture_8bpp_cache(psx_gpu);
2527
2528 const u8 * __restrict__ texture_ptr_8bpp = psx_gpu->texture_page_ptr;
2529 const u16 * __restrict__ clut_ptr = psx_gpu->clut_ptr;
2530 block_struct * __restrict__ block = psx_gpu->blocks;
2531
2532 while(num_blocks)
2533 {
2534 u16 offset;
2535 #define load_one(i_) \
2536 offset = block->uv.e[i_]; u16 texel##i_ = texture_ptr_8bpp[offset]
2537 #define store_one(i_) \
2538 block->texels.e[i_] = clut_ptr[texel##i_]
2539 load_one(0); load_one(1); load_one(2); load_one(3);
2540 load_one(4); load_one(5); load_one(6); load_one(7);
2541 store_one(0); store_one(1); store_one(2); store_one(3);
2542 store_one(4); store_one(5); store_one(6); store_one(7);
2543 #undef load_one
2544 #undef store_one
2545
2546 num_blocks--;
2547 block++;
2548 }
2549}
2550
2551void texture_blocks_16bpp(psx_gpu_struct *psx_gpu)
2552{
2553#if 0
2554 texture_blocks_16bpp_(psx_gpu);
2555 return;
2556#endif
2557 u32 num_blocks = psx_gpu->num_blocks;
2558 const u16 * __restrict__ texture_ptr_16bpp = psx_gpu->texture_page_ptr;
2559 block_struct * __restrict__ block = psx_gpu->blocks;
2560
2561 while(num_blocks)
2562 {
2563 u32 offset;
2564 #define load_one(i_) \
2565 offset = block->uv.e[i_]; \
2566 offset += ((offset & 0xFF00) * 3); \
2567 u16 texel##i_ = texture_ptr_16bpp[offset]
2568 #define store_one(i_) \
2569 block->texels.e[i_] = texel##i_
2570 load_one(0); load_one(1); load_one(2); load_one(3);
2571 load_one(4); load_one(5); load_one(6); load_one(7);
2572 store_one(0); store_one(1); store_one(2); store_one(3);
2573 store_one(4); store_one(5); store_one(6); store_one(7);
2574 #undef load_one
2575 #undef store_one
2576
2577 num_blocks--;
2578 block++;
2579 }
2580}
2581
2582#define shade_blocks_load_msb_mask_indirect() \
2583
2584#define shade_blocks_load_msb_mask_direct() \
2585 vec_8x16u msb_mask; \
2586 gvdupq_n_u16(msb_mask, psx_gpu->mask_msb); \
2587
2588#define shade_blocks_store_indirect(_draw_mask, _pixels) \
2589 gvst1q_u16(_draw_mask, block->draw_mask.e); \
2590 gvst1q_u16(_pixels, block->pixels.e); \
2591
2592#define shade_blocks_store_direct(_draw_mask, _pixels) \
2593{ \
2594 u16 * __restrict__ fb_ptr = block->fb_ptr; \
2595 vec_8x16u fb_pixels; \
2596 gvld1q_u16(fb_pixels, fb_ptr); \
2597 gvorrq(_pixels, _pixels, msb_mask); \
2598 gvbifq(fb_pixels, _pixels, _draw_mask); \
2599 gvst1q_u16(fb_pixels, fb_ptr); \
2600} \
2601
2602#define shade_blocks_textured_false_modulated_check_dithered(target) \
2603
2604#define shade_blocks_textured_false_modulated_check_undithered(target) \
2605 if(psx_gpu->triangle_color == 0x808080) \
2606 { \
2607 shade_blocks_textured_unmodulated_##target(psx_gpu); \
2608 return; \
2609 } \
2610
2611#define shade_blocks_textured_modulated_shaded_primitive_load(dithering, \
2612 target) \
2613
2614#define shade_blocks_textured_modulated_unshaded_primitive_load(dithering, \
2615 target) \
2616{ \
2617 u32 color = psx_gpu->triangle_color; \
2618 gvdup_n_u8(colors_r, color); \
2619 gvdup_n_u8(colors_g, color >> 8); \
2620 gvdup_n_u8(colors_b, color >> 16); \
2621 shade_blocks_textured_false_modulated_check_##dithering(target); \
2622} \
2623
2624#define shade_blocks_textured_modulated_shaded_block_load() \
2625 gvld1_u8(colors_r, block->r.e); \
2626 gvld1_u8(colors_g, block->g.e); \
2627 gvld1_u8(colors_b, block->b.e) \
2628
2629#define shade_blocks_textured_modulated_unshaded_block_load() \
2630
2631#define shade_blocks_textured_modulate_dithered(component) \
2632 gvld1q_u16(pixels_##component, block->dither_offsets.e); \
2633 gvmlal_u8(pixels_##component, texels_##component, colors_##component) \
2634
2635#define shade_blocks_textured_modulate_undithered(component) \
2636 gvmull_u8(pixels_##component, texels_##component, colors_##component) \
2637
2638#define shade_blocks_textured_modulated_do(shading, dithering, target) \
2639 block_struct * __restrict__ block = psx_gpu->blocks; \
2640 u32 num_blocks = psx_gpu->num_blocks; \
2641 vec_8x16u texels; \
2642 \
2643 vec_8x8u texels_r; \
2644 vec_8x8u texels_g; \
2645 vec_8x8u texels_b; \
2646 \
2647 vec_8x8u colors_r; \
2648 vec_8x8u colors_g; \
2649 vec_8x8u colors_b; \
2650 \
2651 vec_8x8u pixels_r_low; \
2652 vec_8x8u pixels_g_low; \
2653 vec_8x8u pixels_b_low; \
2654 vec_8x16u pixels; \
2655 \
2656 vec_8x16u pixels_r; \
2657 vec_8x16u pixels_g; \
2658 vec_8x16u pixels_b; \
2659 \
2660 vec_8x16u draw_mask; \
2661 vec_8x16u zero_mask; \
2662 \
2663 vec_8x8u d64_0x07; \
2664 vec_8x8u d64_0x1F; \
2665 vec_8x8u d64_1; \
2666 vec_8x8u d64_4; \
2667 vec_8x8u d64_128; \
2668 \
2669 vec_8x16u d128_0x8000; \
2670 \
2671 vec_8x16u test_mask; \
2672 u32 draw_mask_bits; \
2673 \
2674 gvld1q_u16(test_mask, psx_gpu->test_mask.e); \
2675 shade_blocks_load_msb_mask_##target(); \
2676 \
2677 gvdup_n_u8(d64_0x07, 0x07); \
2678 gvdup_n_u8(d64_0x1F, 0x1F); \
2679 gvdup_n_u8(d64_1, 1); \
2680 gvdup_n_u8(d64_4, 4); \
2681 gvdup_n_u8(d64_128, 128u); \
2682 \
2683 gvdupq_n_u16(d128_0x8000, 0x8000); \
2684 \
2685 shade_blocks_textured_modulated_##shading##_primitive_load(dithering, \
2686 target); \
2687 \
2688 while(num_blocks) \
2689 { \
2690 draw_mask_bits = block->draw_mask_bits; \
2691 gvdupq_n_u16(draw_mask, draw_mask_bits); \
2692 gvtstq_u16(draw_mask, draw_mask, test_mask); \
2693 \
2694 shade_blocks_textured_modulated_##shading##_block_load(); \
2695 \
2696 gvld1q_u16(texels, block->texels.e); \
2697 \
2698 gvmovn_u16(texels_r, texels); \
2699 gvshrn_n_u16(texels_g, texels, 5); \
2700 gvshrn_n_u16(texels_b, texels, 7); \
2701 \
2702 gvand(texels_r, texels_r, d64_0x1F); \
2703 gvand(texels_g, texels_g, d64_0x1F); \
2704 gvshr_n_u8(texels_b, texels_b, 3); \
2705 \
2706 shade_blocks_textured_modulate_##dithering(r); \
2707 shade_blocks_textured_modulate_##dithering(g); \
2708 shade_blocks_textured_modulate_##dithering(b); \
2709 \
2710 gvceqzq_u16(zero_mask, texels); \
2711 gvand(pixels, texels, d128_0x8000); \
2712 \
2713 gvqshrun_n_s16(pixels_r_low, pixels_r, 4); \
2714 gvqshrun_n_s16(pixels_g_low, pixels_g, 4); \
2715 gvqshrun_n_s16(pixels_b_low, pixels_b, 4); \
2716 \
2717 gvorrq(zero_mask, draw_mask, zero_mask); \
2718 \
2719 gvshr_n_u8(pixels_r_low, pixels_r_low, 3); \
2720 gvbic(pixels_g_low, pixels_g_low, d64_0x07); \
2721 gvbic(pixels_b_low, pixels_b_low, d64_0x07); \
2722 \
2723 gvmlal_u8(pixels, pixels_r_low, d64_1); \
2724 gvmlal_u8(pixels, pixels_g_low, d64_4); \
2725 gvmlal_u8(pixels, pixels_b_low, d64_128); \
2726 \
2727 shade_blocks_store_##target(zero_mask, pixels); \
2728 \
2729 num_blocks--; \
2730 block++; \
2731 } \
2732
2733void shade_blocks_shaded_textured_modulated_dithered_direct(psx_gpu_struct
2734 *psx_gpu)
2735{
2736#if 0
2737 shade_blocks_shaded_textured_modulated_dithered_direct_(psx_gpu);
2738 return;
2739#endif
2740 shade_blocks_textured_modulated_do(shaded, dithered, direct);
2741}
2742
2743void shade_blocks_shaded_textured_modulated_undithered_direct(psx_gpu_struct
2744 *psx_gpu)
2745{
2746#if 0
2747 shade_blocks_shaded_textured_modulated_undithered_direct_(psx_gpu);
2748 return;
2749#endif
2750 shade_blocks_textured_modulated_do(shaded, undithered, direct);
2751}
2752
2753void shade_blocks_unshaded_textured_modulated_dithered_direct(psx_gpu_struct
2754 *psx_gpu)
2755{
2756#if 0
2757 shade_blocks_unshaded_textured_modulated_dithered_direct_(psx_gpu);
2758 return;
2759#endif
2760 shade_blocks_textured_modulated_do(unshaded, dithered, direct);
2761}
2762
2763void shade_blocks_unshaded_textured_modulated_undithered_direct(psx_gpu_struct
2764 *psx_gpu)
2765{
2766#if 0
2767 shade_blocks_unshaded_textured_modulated_undithered_direct_(psx_gpu);
2768 return;
2769#endif
2770 shade_blocks_textured_modulated_do(unshaded, undithered, direct);
2771}
2772
2773void shade_blocks_shaded_textured_modulated_dithered_indirect(psx_gpu_struct
2774 *psx_gpu)
2775{
2776#if 0
2777 shade_blocks_shaded_textured_modulated_dithered_indirect_(psx_gpu);
2778 return;
2779#endif
2780 shade_blocks_textured_modulated_do(shaded, dithered, indirect);
2781}
2782
2783void shade_blocks_shaded_textured_modulated_undithered_indirect(psx_gpu_struct
2784 *psx_gpu)
2785{
2786#if 0
2787 shade_blocks_shaded_textured_modulated_undithered_indirect_(psx_gpu);
2788 return;
2789#endif
2790 shade_blocks_textured_modulated_do(shaded, undithered, indirect);
2791}
2792
2793void shade_blocks_unshaded_textured_modulated_dithered_indirect(psx_gpu_struct
2794 *psx_gpu)
2795{
2796#if 0
2797 shade_blocks_unshaded_textured_modulated_dithered_indirect_(psx_gpu);
2798 return;
2799#endif
2800 shade_blocks_textured_modulated_do(unshaded, dithered, indirect);
2801}
2802
2803void shade_blocks_unshaded_textured_modulated_undithered_indirect(psx_gpu_struct
2804 *psx_gpu)
2805{
2806#if 0
2807 shade_blocks_unshaded_textured_modulated_undithered_indirect_(psx_gpu);
2808 return;
2809#endif
2810 shade_blocks_textured_modulated_do(unshaded, undithered, indirect);
2811}
2812
2813#define shade_blocks_textured_unmodulated_do(target) \
2814 block_struct *block = psx_gpu->blocks; \
2815 u32 num_blocks = psx_gpu->num_blocks; \
2816 vec_8x16u draw_mask; \
2817 vec_8x16u test_mask; \
2818 u32 draw_mask_bits; \
2819 \
2820 vec_8x16u pixels; \
2821 \
2822 gvld1q_u16(test_mask, psx_gpu->test_mask.e); \
2823 shade_blocks_load_msb_mask_##target(); \
2824 \
2825 while(num_blocks) \
2826 { \
2827 vec_8x16u zero_mask; \
2828 \
2829 draw_mask_bits = block->draw_mask_bits; \
2830 gvdupq_n_u16(draw_mask, draw_mask_bits); \
2831 gvtstq_u16(draw_mask, draw_mask, test_mask); \
2832 \
2833 gvld1q_u16(pixels, block->texels.e); \
2834 \
2835 gvceqzq_u16(zero_mask, pixels); \
2836 gvorrq(zero_mask, draw_mask, zero_mask); \
2837 \
2838 shade_blocks_store_##target(zero_mask, pixels); \
2839 \
2840 num_blocks--; \
2841 block++; \
2842 } \
2843
2844void shade_blocks_textured_unmodulated_indirect(psx_gpu_struct *psx_gpu)
2845{
2846#if 0
2847 shade_blocks_textured_unmodulated_indirect_(psx_gpu);
2848 return;
2849#endif
2850 shade_blocks_textured_unmodulated_do(indirect)
2851}
2852
2853void shade_blocks_textured_unmodulated_direct(psx_gpu_struct *psx_gpu)
2854{
2855#if 0
2856 shade_blocks_textured_unmodulated_direct_(psx_gpu);
2857 return;
2858#endif
2859 shade_blocks_textured_unmodulated_do(direct)
2860}
2861
2862void shade_blocks_unshaded_untextured_indirect(psx_gpu_struct *psx_gpu)
2863{
2864}
2865
2866void shade_blocks_unshaded_untextured_direct(psx_gpu_struct *psx_gpu)
2867{
2868#if 0
2869 shade_blocks_unshaded_untextured_direct_(psx_gpu);
2870 return;
2871#endif
2872 block_struct *block = psx_gpu->blocks;
2873 u32 num_blocks = psx_gpu->num_blocks;
2874
2875 vec_8x16u pixels;
2876 gvld1q_u16(pixels, block->pixels.e);
2877 shade_blocks_load_msb_mask_direct();
2878
2879 while(num_blocks)
2880 {
2881 vec_8x16u draw_mask;
2882 gvld1q_u16(draw_mask, block->draw_mask.e);
2883 shade_blocks_store_direct(draw_mask, pixels);
2884
2885 num_blocks--;
2886 block++;
2887 }
2888}
2889
2890#define blend_blocks_mask_evaluate_on() \
2891 vec_8x16u mask_pixels; \
2892 gvcltzq_s16(mask_pixels, framebuffer_pixels); \
2893 gvorrq(draw_mask, draw_mask, mask_pixels) \
2894
2895#define blend_blocks_mask_evaluate_off() \
2896
2897#define blend_blocks_average() \
2898{ \
2899 vec_8x16u pixels_no_msb; \
2900 vec_8x16u fb_pixels_no_msb; \
2901 \
2902 vec_8x16u d128_0x0421; \
2903 \
2904 gvdupq_n_u16(d128_0x0421, 0x0421); \
2905 \
2906 gveorq(blend_pixels, pixels, framebuffer_pixels); \
2907 gvbicq(pixels_no_msb, pixels, d128_0x8000); \
2908 gvand(blend_pixels, blend_pixels, d128_0x0421); \
2909 gvsubq_u16(blend_pixels, pixels_no_msb, blend_pixels); \
2910 gvbicq(fb_pixels_no_msb, framebuffer_pixels, d128_0x8000); \
2911 gvhaddq_u16(blend_pixels, fb_pixels_no_msb, blend_pixels); \
2912} \
2913
2914#define blend_blocks_add() \
2915{ \
2916 vec_8x16u pixels_rb, pixels_g; \
2917 vec_8x16u fb_rb, fb_g; \
2918 \
2919 vec_8x16u d128_0x7C1F; \
2920 vec_8x16u d128_0x03E0; \
2921 \
2922 gvdupq_n_u16(d128_0x7C1F, 0x7C1F); \
2923 gvdupq_n_u16(d128_0x03E0, 0x03E0); \
2924 \
2925 gvand(pixels_rb, pixels, d128_0x7C1F); \
2926 gvand(pixels_g, pixels, d128_0x03E0); \
2927 \
2928 gvand(fb_rb, framebuffer_pixels, d128_0x7C1F); \
2929 gvand(fb_g, framebuffer_pixels, d128_0x03E0); \
2930 \
2931 gvaddq_u16(fb_rb, fb_rb, pixels_rb); \
2932 gvaddq_u16(fb_g, fb_g, pixels_g); \
2933 \
2934 gvminq_u8(fb_rb, fb_rb, d128_0x7C1F); \
2935 gvminq_u16(fb_g, fb_g, d128_0x03E0); \
2936 \
2937 gvorrq(blend_pixels, fb_rb, fb_g); \
2938} \
2939
2940#define blend_blocks_subtract() \
2941{ \
2942 vec_8x16u pixels_rb, pixels_g; \
2943 vec_8x16u fb_rb, fb_g; \
2944 \
2945 vec_8x16u d128_0x7C1F; \
2946 vec_8x16u d128_0x03E0; \
2947 \
2948 gvdupq_n_u16(d128_0x7C1F, 0x7C1F); \
2949 gvdupq_n_u16(d128_0x03E0, 0x03E0); \
2950 \
2951 gvand(pixels_rb, pixels, d128_0x7C1F); \
2952 gvand(pixels_g, pixels, d128_0x03E0); \
2953 \
2954 gvand(fb_rb, framebuffer_pixels, d128_0x7C1F); \
2955 gvand(fb_g, framebuffer_pixels, d128_0x03E0); \
2956 \
2957 gvqsubq_u8(fb_rb, fb_rb, pixels_rb); \
2958 gvqsubq_u16(fb_g, fb_g, pixels_g); \
2959 \
2960 gvorrq(blend_pixels, fb_rb, fb_g); \
2961} \
2962
2963#define blend_blocks_add_fourth() \
2964{ \
2965 vec_8x16u pixels_rb, pixels_g; \
2966 vec_8x16u pixels_fourth; \
2967 vec_8x16u fb_rb, fb_g; \
2968 \
2969 vec_8x16u d128_0x7C1F; \
2970 vec_8x16u d128_0x1C07; \
2971 vec_8x16u d128_0x03E0; \
2972 vec_8x16u d128_0x00E0; \
2973 \
2974 gvdupq_n_u16(d128_0x7C1F, 0x7C1F); \
2975 gvdupq_n_u16(d128_0x1C07, 0x1C07); \
2976 gvdupq_n_u16(d128_0x03E0, 0x03E0); \
2977 gvdupq_n_u16(d128_0x00E0, 0x00E0); \
2978 \
2979 gvshrq_n_u16(pixels_fourth, pixels, 2); \
2980 \
2981 gvand(fb_rb, framebuffer_pixels, d128_0x7C1F); \
2982 gvand(fb_g, framebuffer_pixels, d128_0x03E0); \
2983 \
2984 gvand(pixels_rb, pixels_fourth, d128_0x1C07); \
2985 gvand(pixels_g, pixels_fourth, d128_0x00E0); \
2986 \
2987 gvaddq_u16(fb_rb, fb_rb, pixels_rb); \
2988 gvaddq_u16(fb_g, fb_g, pixels_g); \
2989 \
2990 gvminq_u8(fb_rb, fb_rb, d128_0x7C1F); \
2991 gvminq_u16(fb_g, fb_g, d128_0x03E0); \
2992 \
2993 gvorrq(blend_pixels, fb_rb, fb_g); \
2994} \
2995
2996#define blend_blocks_blended_combine_textured() \
2997{ \
2998 vec_8x16u blend_mask; \
2999 gvcltzq_s16(blend_mask, pixels); \
3000 \
3001 gvorrq(blend_pixels, blend_pixels, d128_0x8000); \
3002 gvbifq(blend_pixels, pixels, blend_mask); \
3003} \
3004
3005#define blend_blocks_blended_combine_untextured() \
3006
3007#define blend_blocks_body_blend(blend_mode, texturing) \
3008{ \
3009 blend_blocks_##blend_mode(); \
3010 blend_blocks_blended_combine_##texturing(); \
3011} \
3012
3013#define blend_blocks_body_average(texturing) \
3014 blend_blocks_body_blend(average, texturing) \
3015
3016#define blend_blocks_body_add(texturing) \
3017 blend_blocks_body_blend(add, texturing) \
3018
3019#define blend_blocks_body_subtract(texturing) \
3020 blend_blocks_body_blend(subtract, texturing) \
3021
3022#define blend_blocks_body_add_fourth(texturing) \
3023 blend_blocks_body_blend(add_fourth, texturing) \
3024
3025#define blend_blocks_body_unblended(texturing) \
3026 blend_pixels = pixels \
3027
3028#define blend_blocks_do(texturing, blend_mode, mask_evaluate) \
3029 block_struct *block = psx_gpu->blocks; \
3030 u32 num_blocks = psx_gpu->num_blocks; \
3031 vec_8x16u draw_mask; \
3032 vec_8x16u pixels; \
3033 vec_8x16u blend_pixels; \
3034 vec_8x16u framebuffer_pixels; \
3035 vec_8x16u msb_mask; \
3036 vec_8x16u d128_0x8000; \
3037 \
3038 u16 *fb_ptr; \
3039 \
3040 gvdupq_n_u16(d128_0x8000, 0x8000); \
3041 gvdupq_n_u16(msb_mask, psx_gpu->mask_msb); \
3042 (void)d128_0x8000; /* sometimes unused */ \
3043 \
3044 while(num_blocks) \
3045 { \
3046 gvld1q_u16(pixels, block->pixels.e); \
3047 gvld1q_u16(draw_mask, block->draw_mask.e); \
3048 fb_ptr = block->fb_ptr; \
3049 \
3050 gvld1q_u16(framebuffer_pixels, fb_ptr); \
3051 \
3052 blend_blocks_mask_evaluate_##mask_evaluate(); \
3053 blend_blocks_body_##blend_mode(texturing); \
3054 \
3055 gvorrq(blend_pixels, blend_pixels, msb_mask); \
3056 gvbifq(framebuffer_pixels, blend_pixels, draw_mask); \
3057 gvst1q_u16(framebuffer_pixels, fb_ptr); \
3058 \
3059 num_blocks--; \
3060 block++; \
3061 } \
3062
3063
3064void blend_blocks_textured_average_off(psx_gpu_struct *psx_gpu)
3065{
3066#if 0
3067 blend_blocks_textured_average_off_(psx_gpu);
3068 return;
3069#endif
3070 blend_blocks_do(textured, average, off);
3071}
3072
3073void blend_blocks_untextured_average_off(psx_gpu_struct *psx_gpu)
3074{
3075#if 0
3076 blend_blocks_untextured_average_off_(psx_gpu);
3077 return;
3078#endif
3079 blend_blocks_do(untextured, average, off);
3080}
3081
3082void blend_blocks_textured_average_on(psx_gpu_struct *psx_gpu)
3083{
3084#if 0
3085 blend_blocks_textured_average_on_(psx_gpu);
3086 return;
3087#endif
3088 blend_blocks_do(textured, average, on);
3089}
3090
3091void blend_blocks_untextured_average_on(psx_gpu_struct *psx_gpu)
3092{
3093#if 0
3094 blend_blocks_untextured_average_on_(psx_gpu);
3095 return;
3096#endif
3097 blend_blocks_do(untextured, average, on);
3098}
3099
3100void blend_blocks_textured_add_off(psx_gpu_struct *psx_gpu)
3101{
3102#if 0
3103 blend_blocks_textured_add_off_(psx_gpu);
3104 return;
3105#endif
3106 blend_blocks_do(textured, add, off);
3107}
3108
3109void blend_blocks_textured_add_on(psx_gpu_struct *psx_gpu)
3110{
3111#if 0
3112 blend_blocks_textured_add_on_(psx_gpu);
3113 return;
3114#endif
3115 blend_blocks_do(textured, add, on);
3116}
3117
3118void blend_blocks_untextured_add_off(psx_gpu_struct *psx_gpu)
3119{
3120#if 0
3121 blend_blocks_untextured_add_off_(psx_gpu);
3122 return;
3123#endif
3124 blend_blocks_do(untextured, add, off);
3125}
3126
3127void blend_blocks_untextured_add_on(psx_gpu_struct *psx_gpu)
3128{
3129#if 0
3130 blend_blocks_untextured_add_on_(psx_gpu);
3131 return;
3132#endif
3133 blend_blocks_do(untextured, add, on);
3134}
3135
3136void blend_blocks_textured_subtract_off(psx_gpu_struct *psx_gpu)
3137{
3138#if 0
3139 blend_blocks_textured_subtract_off_(psx_gpu);
3140 return;
3141#endif
3142 blend_blocks_do(textured, subtract, off);
3143}
3144
3145void blend_blocks_textured_subtract_on(psx_gpu_struct *psx_gpu)
3146{
3147#if 0
3148 blend_blocks_textured_subtract_on_(psx_gpu);
3149 return;
3150#endif
3151 blend_blocks_do(textured, subtract, on);
3152}
3153
3154void blend_blocks_untextured_subtract_off(psx_gpu_struct *psx_gpu)
3155{
3156#if 0
3157 blend_blocks_untextured_subtract_off_(psx_gpu);
3158 return;
3159#endif
3160 blend_blocks_do(untextured, subtract, off);
3161}
3162
3163void blend_blocks_untextured_subtract_on(psx_gpu_struct *psx_gpu)
3164{
3165#if 0
3166 blend_blocks_untextured_subtract_on_(psx_gpu);
3167 return;
3168#endif
3169 blend_blocks_do(untextured, subtract, on);
3170}
3171
3172void blend_blocks_textured_add_fourth_off(psx_gpu_struct *psx_gpu)
3173{
3174#if 0
3175 blend_blocks_textured_add_fourth_off_(psx_gpu);
3176 return;
3177#endif
3178 blend_blocks_do(textured, add_fourth, off);
3179}
3180
3181void blend_blocks_textured_add_fourth_on(psx_gpu_struct *psx_gpu)
3182{
3183#if 0
3184 blend_blocks_textured_add_fourth_on_(psx_gpu);
3185 return;
3186#endif
3187 blend_blocks_do(textured, add_fourth, on);
3188}
3189
3190void blend_blocks_untextured_add_fourth_off(psx_gpu_struct *psx_gpu)
3191{
3192#if 0
3193 blend_blocks_untextured_add_fourth_off_(psx_gpu);
3194 return;
3195#endif
3196 blend_blocks_do(untextured, add_fourth, off);
3197}
3198
3199void blend_blocks_untextured_add_fourth_on(psx_gpu_struct *psx_gpu)
3200{
3201#if 0
3202 blend_blocks_untextured_add_fourth_on_(psx_gpu);
3203 return;
3204#endif
3205 blend_blocks_do(untextured, add_fourth, on);
3206}
3207
3208void blend_blocks_textured_unblended_on(psx_gpu_struct *psx_gpu)
3209{
3210#if 0
3211 blend_blocks_textured_unblended_on_(psx_gpu);
3212 return;
3213#endif
3214 blend_blocks_do(textured, unblended, on);
3215}
3216
3217void blend_blocks_textured_unblended_off(psx_gpu_struct *psx_gpu)
3218{
3219}
3220
3221void setup_sprite_untextured_512(psx_gpu_struct *psx_gpu, s32 x, s32 y, s32 u,
3222 s32 v, s32 width, s32 height, u32 color)
3223{
3224#if 0
3225 setup_sprite_untextured_512_(psx_gpu, x, y, u, v, width, height, color);
3226 return;
3227#endif
3228 u32 right_width = ((width - 1) & 0x7) + 1;
3229 u32 right_mask_bits = (0xFF << right_width);
3230 u16 *fb_ptr = psx_gpu->vram_out_ptr + (y * 1024) + x;
3231 u32 block_width = (width + 7) / 8;
3232 u32 fb_ptr_pitch = 1024 - ((block_width - 1) * 8);
3233 u32 blocks_remaining;
3234 u32 num_blocks = psx_gpu->num_blocks;
3235 block_struct *block = psx_gpu->blocks + num_blocks;
3236
3237 u32 color_r = color & 0xFF;
3238 u32 color_g = (color >> 8) & 0xFF;
3239 u32 color_b = (color >> 16) & 0xFF;
3240 vec_8x16u colors;
3241 vec_8x16u right_mask;
3242 vec_8x16u test_mask;
3243 vec_8x16u zero_mask;
3244
3245 gvld1q_u16(test_mask, psx_gpu->test_mask.e);
3246 color = (color_r >> 3) | ((color_g >> 3) << 5) | ((color_b >> 3) << 10);
3247
3248 gvdupq_n_u16(colors, color);
3249 gvdupq_n_u16(zero_mask, 0x00);
3250 gvdupq_n_u16(right_mask, right_mask_bits);
3251 gvtstq_u16(right_mask, right_mask, test_mask);
3252
3253 while(height)
3254 {
3255 blocks_remaining = block_width - 1;
3256 num_blocks += block_width;
3257
3258 if(num_blocks > MAX_BLOCKS)
3259 {
3260 flush_render_block_buffer(psx_gpu);
3261 num_blocks = block_width;
3262 block = psx_gpu->blocks;
3263 }
3264
3265 while(blocks_remaining)
3266 {
3267 gvst1q_u16(colors, block->pixels.e);
3268 gvst1q_u16(zero_mask, block->draw_mask.e);
3269 block->fb_ptr = fb_ptr;
3270
3271 fb_ptr += 8;
3272 block++;
3273 blocks_remaining--;
3274 }
3275
3276 gvst1q_u16(colors, block->pixels.e);
3277 gvst1q_u16(right_mask, block->draw_mask.e);
3278 block->fb_ptr = fb_ptr;
3279
3280 block++;
3281 fb_ptr += fb_ptr_pitch;
3282
3283 height--;
3284 psx_gpu->num_blocks = num_blocks;
3285 }
3286}
3287
3288#define setup_sprite_tiled_initialize_4bpp_clut() \
3289 vec_16x8u clut_low, clut_high; \
3290 \
3291 gvld2q_u8(clut_low, clut_high, (u8 *)psx_gpu->clut_ptr) \
3292
3293#define setup_sprite_tiled_initialize_4bpp() \
3294 setup_sprite_tiled_initialize_4bpp_clut(); \
3295 \
3296 if(psx_gpu->current_texture_mask & psx_gpu->dirty_textures_4bpp_mask) \
3297 update_texture_4bpp_cache(psx_gpu) \
3298
3299#define setup_sprite_tiled_initialize_8bpp() \
3300 if(psx_gpu->current_texture_mask & psx_gpu->dirty_textures_8bpp_mask) \
3301 update_texture_8bpp_cache(psx_gpu) \
3302
3303#define setup_sprite_tile_fetch_texel_block_8bpp(offset) \
3304 texture_block_ptr = psx_gpu->texture_page_ptr + \
3305 ((texture_offset + offset) & texture_mask); \
3306 \
3307 gvld1_u8(texels, (u8 *)texture_block_ptr) \
3308
3309#define setup_sprite_tile_add_blocks(tile_num_blocks) \
3310 num_blocks += tile_num_blocks; \
3311 \
3312 if(num_blocks > MAX_BLOCKS) \
3313 { \
3314 flush_render_block_buffer(psx_gpu); \
3315 num_blocks = tile_num_blocks; \
3316 block = psx_gpu->blocks; \
3317 } \
3318
3319#define setup_sprite_tile_full_4bpp(edge) \
3320{ \
3321 vec_8x8u texels_low, texels_high; \
3322 setup_sprite_tile_add_blocks(sub_tile_height * 2); \
3323 \
3324 while(sub_tile_height) \
3325 { \
3326 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3327 gvtbl2_u8(texels_low, clut_low, texels); \
3328 gvtbl2_u8(texels_high, clut_high, texels); \
3329 \
3330 gvst2_u8(texels_low, texels_high, (u8 *)block->texels.e); \
3331 block->draw_mask_bits = left_mask_bits; \
3332 block->fb_ptr = fb_ptr; \
3333 block++; \
3334 \
3335 setup_sprite_tile_fetch_texel_block_8bpp(8); \
3336 gvtbl2_u8(texels_low, clut_low, texels); \
3337 gvtbl2_u8(texels_high, clut_high, texels); \
3338 \
3339 gvst2_u8(texels_low, texels_high, (u8 *)block->texels.e); \
3340 block->draw_mask_bits = right_mask_bits; \
3341 block->fb_ptr = fb_ptr + 8; \
3342 block++; \
3343 \
3344 fb_ptr += 1024; \
3345 texture_offset += 0x10; \
3346 sub_tile_height--; \
3347 } \
3348 texture_offset += 0xF00; \
3349 psx_gpu->num_blocks = num_blocks; \
3350} \
3351
3352#define setup_sprite_tile_half_4bpp(edge) \
3353{ \
3354 vec_8x8u texels_low, texels_high; \
3355 setup_sprite_tile_add_blocks(sub_tile_height); \
3356 \
3357 while(sub_tile_height) \
3358 { \
3359 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3360 gvtbl2_u8(texels_low, clut_low, texels); \
3361 gvtbl2_u8(texels_high, clut_high, texels); \
3362 \
3363 gvst2_u8(texels_low, texels_high, (u8 *)block->texels.e); \
3364 block->draw_mask_bits = edge##_mask_bits; \
3365 block->fb_ptr = fb_ptr; \
3366 block++; \
3367 \
3368 fb_ptr += 1024; \
3369 texture_offset += 0x10; \
3370 sub_tile_height--; \
3371 } \
3372 texture_offset += 0xF00; \
3373 psx_gpu->num_blocks = num_blocks; \
3374} \
3375
3376#define setup_sprite_tile_full_8bpp(edge) \
3377{ \
3378 setup_sprite_tile_add_blocks(sub_tile_height * 2); \
3379 \
3380 while(sub_tile_height) \
3381 { \
3382 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3383 gvst1_u8(texels, block->r.e); \
3384 block->draw_mask_bits = left_mask_bits; \
3385 block->fb_ptr = fb_ptr; \
3386 block++; \
3387 \
3388 setup_sprite_tile_fetch_texel_block_8bpp(8); \
3389 gvst1_u8(texels, block->r.e); \
3390 block->draw_mask_bits = right_mask_bits; \
3391 block->fb_ptr = fb_ptr + 8; \
3392 block++; \
3393 \
3394 fb_ptr += 1024; \
3395 texture_offset += 0x10; \
3396 sub_tile_height--; \
3397 } \
3398 texture_offset += 0xF00; \
3399 psx_gpu->num_blocks = num_blocks; \
3400} \
3401
3402#define setup_sprite_tile_half_8bpp(edge) \
3403{ \
3404 setup_sprite_tile_add_blocks(sub_tile_height); \
3405 \
3406 while(sub_tile_height) \
3407 { \
3408 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3409 gvst1_u8(texels, block->r.e); \
3410 block->draw_mask_bits = edge##_mask_bits; \
3411 block->fb_ptr = fb_ptr; \
3412 block++; \
3413 \
3414 fb_ptr += 1024; \
3415 texture_offset += 0x10; \
3416 sub_tile_height--; \
3417 } \
3418 texture_offset += 0xF00; \
3419 psx_gpu->num_blocks = num_blocks; \
3420} \
3421
3422#define setup_sprite_tile_column_edge_pre_adjust_half_right() \
3423 texture_offset = texture_offset_base + 8; \
3424 fb_ptr += 8 \
3425
3426#define setup_sprite_tile_column_edge_pre_adjust_half_left() \
3427 texture_offset = texture_offset_base \
3428
3429#define setup_sprite_tile_column_edge_pre_adjust_half(edge) \
3430 setup_sprite_tile_column_edge_pre_adjust_half_##edge() \
3431
3432#define setup_sprite_tile_column_edge_pre_adjust_full(edge) \
3433 texture_offset = texture_offset_base \
3434
3435#define setup_sprite_tile_column_edge_post_adjust_half_right() \
3436 fb_ptr -= 8 \
3437
3438#define setup_sprite_tile_column_edge_post_adjust_half_left() \
3439
3440#define setup_sprite_tile_column_edge_post_adjust_half(edge) \
3441 setup_sprite_tile_column_edge_post_adjust_half_##edge() \
3442
3443#define setup_sprite_tile_column_edge_post_adjust_full(edge) \
3444
3445
3446#define setup_sprite_tile_column_height_single(edge_mode, edge, texture_mode, \
3447 x4mode) \
3448do \
3449{ \
3450 sub_tile_height = column_data; \
3451 setup_sprite_tile_column_edge_pre_adjust_##edge_mode##x4mode(edge); \
3452 setup_sprite_tile_##edge_mode##_##texture_mode##x4mode(edge); \
3453 setup_sprite_tile_column_edge_post_adjust_##edge_mode##x4mode(edge); \
3454} while(0) \
3455
3456#define setup_sprite_tile_column_height_multi(edge_mode, edge, texture_mode, \
3457 x4mode) \
3458do \
3459{ \
3460 u32 tiles_remaining = column_data >> 16; \
3461 sub_tile_height = column_data & 0xFF; \
3462 setup_sprite_tile_column_edge_pre_adjust_##edge_mode##x4mode(edge); \
3463 setup_sprite_tile_##edge_mode##_##texture_mode##x4mode(edge); \
3464 tiles_remaining -= 1; \
3465 \
3466 while(tiles_remaining) \
3467 { \
3468 sub_tile_height = 16; \
3469 setup_sprite_tile_##edge_mode##_##texture_mode##x4mode(edge); \
3470 tiles_remaining--; \
3471 } \
3472 \
3473 sub_tile_height = (column_data >> 8) & 0xFF; \
3474 setup_sprite_tile_##edge_mode##_##texture_mode##x4mode(edge); \
3475 setup_sprite_tile_column_edge_post_adjust_##edge_mode##x4mode(edge); \
3476} while(0) \
3477
3478
3479#define setup_sprite_column_data_single() \
3480 column_data = height \
3481
3482#define setup_sprite_column_data_multi() \
3483 column_data = 16 - offset_v; \
3484 column_data |= ((height_rounded & 0xF) + 1) << 8; \
3485 column_data |= (tile_height - 1) << 16 \
3486
3487#define RIGHT_MASK_BIT_SHIFT 8
3488#define RIGHT_MASK_BIT_SHIFT_4x 16
3489
3490#define setup_sprite_tile_column_width_single(texture_mode, multi_height, \
3491 edge_mode, edge, x4mode) \
3492{ \
3493 setup_sprite_column_data_##multi_height(); \
3494 left_mask_bits = left_block_mask | right_block_mask; \
3495 right_mask_bits = left_mask_bits >> RIGHT_MASK_BIT_SHIFT##x4mode; \
3496 \
3497 setup_sprite_tile_column_height_##multi_height(edge_mode, edge, \
3498 texture_mode, x4mode); \
3499} \
3500
3501#define setup_sprite_tiled_advance_column() \
3502 texture_offset_base += 0x100; \
3503 if((texture_offset_base & 0xF00) == 0) \
3504 texture_offset_base -= (0x100 + 0xF00) \
3505
3506#define FB_PTR_MULTIPLIER 1
3507#define FB_PTR_MULTIPLIER_4x 2
3508
3509#define setup_sprite_tile_column_width_multi(texture_mode, multi_height, \
3510 left_mode, right_mode, x4mode) \
3511{ \
3512 setup_sprite_column_data_##multi_height(); \
3513 s32 fb_ptr_advance_column = (16 - (1024 * height)) \
3514 * FB_PTR_MULTIPLIER##x4mode; \
3515 \
3516 tile_width -= 2; \
3517 left_mask_bits = left_block_mask; \
3518 right_mask_bits = left_mask_bits >> RIGHT_MASK_BIT_SHIFT##x4mode; \
3519 \
3520 setup_sprite_tile_column_height_##multi_height(left_mode, right, \
3521 texture_mode, x4mode); \
3522 fb_ptr += fb_ptr_advance_column; \
3523 \
3524 left_mask_bits = 0x00; \
3525 right_mask_bits = 0x00; \
3526 \
3527 while(tile_width) \
3528 { \
3529 setup_sprite_tiled_advance_column(); \
3530 setup_sprite_tile_column_height_##multi_height(full, none, \
3531 texture_mode, x4mode); \
3532 fb_ptr += fb_ptr_advance_column; \
3533 tile_width--; \
3534 } \
3535 \
3536 left_mask_bits = right_block_mask; \
3537 right_mask_bits = left_mask_bits >> RIGHT_MASK_BIT_SHIFT##x4mode; \
3538 \
3539 setup_sprite_tiled_advance_column(); \
3540 setup_sprite_tile_column_height_##multi_height(right_mode, left, \
3541 texture_mode, x4mode); \
3542} \
3543
3544
3545/* 4x stuff */
3546#define setup_sprite_tiled_initialize_4bpp_4x() \
3547 setup_sprite_tiled_initialize_4bpp_clut() \
3548
3549#define setup_sprite_tiled_initialize_8bpp_4x() \
3550
3551#define setup_sprite_tile_full_4bpp_4x(edge) \
3552{ \
3553 vec_8x8u texels_low, texels_high; \
3554 vec_8x16u pixels; \
3555 vec_4x16u pixels_half; \
3556 setup_sprite_tile_add_blocks(sub_tile_height * 2 * 4); \
3557 u32 left_mask_bits_a = left_mask_bits & 0xFF; \
3558 u32 left_mask_bits_b = left_mask_bits >> 8; \
3559 u32 right_mask_bits_a = right_mask_bits & 0xFF; \
3560 u32 right_mask_bits_b = right_mask_bits >> 8; \
3561 \
3562 while(sub_tile_height) \
3563 { \
3564 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3565 gvtbl2_u8(texels_low, clut_low, texels); \
3566 gvtbl2_u8(texels_high, clut_high, texels); \
3567 gvzip_u8(pixels, texels_low, texels_high); \
3568 \
3569 gvget_lo(pixels_half, pixels); \
3570 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3571 block->draw_mask_bits = left_mask_bits_a; \
3572 block->fb_ptr = fb_ptr; \
3573 block++; \
3574 \
3575 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3576 block->draw_mask_bits = left_mask_bits_a; \
3577 block->fb_ptr = fb_ptr + 1024; \
3578 block++; \
3579 \
3580 gvget_hi(pixels_half, pixels); \
3581 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3582 block->draw_mask_bits = left_mask_bits_b; \
3583 block->fb_ptr = fb_ptr + 8; \
3584 block++; \
3585 \
3586 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3587 block->draw_mask_bits = left_mask_bits_b; \
3588 block->fb_ptr = fb_ptr + 1024 + 8; \
3589 block++; \
3590 \
3591 setup_sprite_tile_fetch_texel_block_8bpp(8); \
3592 gvtbl2_u8(texels_low, clut_low, texels); \
3593 gvtbl2_u8(texels_high, clut_high, texels); \
3594 gvzip_u8(pixels, texels_low, texels_high); \
3595 \
3596 gvget_lo(pixels_half, pixels); \
3597 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3598 block->draw_mask_bits = right_mask_bits_a; \
3599 block->fb_ptr = fb_ptr + 16; \
3600 block++; \
3601 \
3602 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3603 block->draw_mask_bits = right_mask_bits_a; \
3604 block->fb_ptr = fb_ptr + 1024 + 16; \
3605 block++; \
3606 \
3607 gvget_hi(pixels_half, pixels); \
3608 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3609 block->draw_mask_bits = right_mask_bits_b; \
3610 block->fb_ptr = fb_ptr + 24; \
3611 block++; \
3612 \
3613 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3614 block->draw_mask_bits = right_mask_bits_b; \
3615 block->fb_ptr = fb_ptr + 1024 + 24; \
3616 block++; \
3617 \
3618 fb_ptr += 2048; \
3619 texture_offset += 0x10; \
3620 sub_tile_height--; \
3621 } \
3622 texture_offset += 0xF00; \
3623 psx_gpu->num_blocks = num_blocks; \
3624} \
3625
3626#define setup_sprite_tile_half_4bpp_4x(edge) \
3627{ \
3628 vec_8x8u texels_low, texels_high; \
3629 vec_8x16u pixels; \
3630 vec_4x16u pixels_half; \
3631 setup_sprite_tile_add_blocks(sub_tile_height * 4); \
3632 u32 edge##_mask_bits_a = edge##_mask_bits & 0xFF; \
3633 u32 edge##_mask_bits_b = edge##_mask_bits >> 8; \
3634 \
3635 while(sub_tile_height) \
3636 { \
3637 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3638 gvtbl2_u8(texels_low, clut_low, texels); \
3639 gvtbl2_u8(texels_high, clut_high, texels); \
3640 gvzip_u8(pixels, texels_low, texels_high); \
3641 \
3642 gvget_lo(pixels_half, pixels); \
3643 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3644 block->draw_mask_bits = edge##_mask_bits_a; \
3645 block->fb_ptr = fb_ptr; \
3646 block++; \
3647 \
3648 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3649 block->draw_mask_bits = edge##_mask_bits_a; \
3650 block->fb_ptr = fb_ptr + 1024; \
3651 block++; \
3652 \
3653 gvget_hi(pixels_half, pixels); \
3654 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3655 block->draw_mask_bits = edge##_mask_bits_b; \
3656 block->fb_ptr = fb_ptr + 8; \
3657 block++; \
3658 \
3659 gvst2_u16(pixels_half, pixels_half, block->texels.e); \
3660 block->draw_mask_bits = edge##_mask_bits_b; \
3661 block->fb_ptr = fb_ptr + 1024 + 8; \
3662 block++; \
3663 \
3664 fb_ptr += 2048; \
3665 texture_offset += 0x10; \
3666 sub_tile_height--; \
3667 } \
3668 texture_offset += 0xF00; \
3669 psx_gpu->num_blocks = num_blocks; \
3670} \
3671
3672#define setup_sprite_tile_full_8bpp_4x(edge) \
3673{ \
3674 setup_sprite_tile_add_blocks(sub_tile_height * 2 * 4); \
3675 vec_8x16u texels_wide; \
3676 vec_4x16u texels_half; \
3677 u32 left_mask_bits_a = left_mask_bits & 0xFF; \
3678 u32 left_mask_bits_b = left_mask_bits >> 8; \
3679 u32 right_mask_bits_a = right_mask_bits & 0xFF; \
3680 u32 right_mask_bits_b = right_mask_bits >> 8; \
3681 \
3682 while(sub_tile_height) \
3683 { \
3684 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3685 gvzip_u8(texels_wide, texels, texels); \
3686 gvget_lo(texels_half, texels_wide); \
3687 gvst1_u8(texels_half, block->r.e); \
3688 block->draw_mask_bits = left_mask_bits_a; \
3689 block->fb_ptr = fb_ptr; \
3690 block++; \
3691 \
3692 gvst1_u8(texels_half, block->r.e); \
3693 block->draw_mask_bits = left_mask_bits_a; \
3694 block->fb_ptr = fb_ptr + 1024; \
3695 block++; \
3696 \
3697 gvget_hi(texels_half, texels_wide); \
3698 gvst1_u8(texels_half, block->r.e); \
3699 block->draw_mask_bits = left_mask_bits_b; \
3700 block->fb_ptr = fb_ptr + 8; \
3701 block++; \
3702 \
3703 gvst1_u8(texels_half, block->r.e); \
3704 block->draw_mask_bits = left_mask_bits_b; \
3705 block->fb_ptr = fb_ptr + 1024 + 8; \
3706 block++; \
3707 \
3708 setup_sprite_tile_fetch_texel_block_8bpp(8); \
3709 gvzip_u8(texels_wide, texels, texels); \
3710 gvget_lo(texels_half, texels_wide); \
3711 gvst1_u8(texels_half, block->r.e); \
3712 block->draw_mask_bits = right_mask_bits_a; \
3713 block->fb_ptr = fb_ptr + 16; \
3714 block++; \
3715 \
3716 gvst1_u8(texels_half, block->r.e); \
3717 block->draw_mask_bits = right_mask_bits_a; \
3718 block->fb_ptr = fb_ptr + 1024 + 16; \
3719 block++; \
3720 \
3721 gvget_hi(texels_half, texels_wide); \
3722 gvst1_u8(texels_half, block->r.e); \
3723 block->draw_mask_bits = right_mask_bits_b; \
3724 block->fb_ptr = fb_ptr + 24; \
3725 block++; \
3726 \
3727 gvst1_u8(texels_half, block->r.e); \
3728 block->draw_mask_bits = right_mask_bits_b; \
3729 block->fb_ptr = fb_ptr + 24 + 1024; \
3730 block++; \
3731 \
3732 fb_ptr += 2048; \
3733 texture_offset += 0x10; \
3734 sub_tile_height--; \
3735 } \
3736 texture_offset += 0xF00; \
3737 psx_gpu->num_blocks = num_blocks; \
3738} \
3739
3740#define setup_sprite_tile_half_8bpp_4x(edge) \
3741{ \
3742 setup_sprite_tile_add_blocks(sub_tile_height * 4); \
3743 vec_8x16u texels_wide; \
3744 vec_4x16u texels_half; \
3745 u32 edge##_mask_bits_a = edge##_mask_bits & 0xFF; \
3746 u32 edge##_mask_bits_b = edge##_mask_bits >> 8; \
3747 \
3748 while(sub_tile_height) \
3749 { \
3750 setup_sprite_tile_fetch_texel_block_8bpp(0); \
3751 gvzip_u8(texels_wide, texels, texels); \
3752 gvget_lo(texels_half, texels_wide); \
3753 gvst1_u8(texels_half, block->r.e); \
3754 block->draw_mask_bits = edge##_mask_bits_a; \
3755 block->fb_ptr = fb_ptr; \
3756 block++; \
3757 \
3758 gvst1_u8(texels_half, block->r.e); \
3759 block->draw_mask_bits = edge##_mask_bits_a; \
3760 block->fb_ptr = fb_ptr + 1024; \
3761 block++; \
3762 \
3763 gvget_hi(texels_half, texels_wide); \
3764 gvst1_u8(texels_half, block->r.e); \
3765 block->draw_mask_bits = edge##_mask_bits_b; \
3766 block->fb_ptr = fb_ptr + 8; \
3767 block++; \
3768 \
3769 gvst1_u8(texels_half, block->r.e); \
3770 block->draw_mask_bits = edge##_mask_bits_b; \
3771 block->fb_ptr = fb_ptr + 8 + 1024; \
3772 block++; \
3773 \
3774 fb_ptr += 2048; \
3775 texture_offset += 0x10; \
3776 sub_tile_height--; \
3777 } \
3778 texture_offset += 0xF00; \
3779 psx_gpu->num_blocks = num_blocks; \
3780} \
3781
3782#define setup_sprite_tile_column_edge_pre_adjust_half_right_4x() \
3783 texture_offset = texture_offset_base + 8; \
3784 fb_ptr += 16 \
3785
3786#define setup_sprite_tile_column_edge_pre_adjust_half_left_4x() \
3787 texture_offset = texture_offset_base \
3788
3789#define setup_sprite_tile_column_edge_pre_adjust_half_4x(edge) \
3790 setup_sprite_tile_column_edge_pre_adjust_half_##edge##_4x() \
3791
3792#define setup_sprite_tile_column_edge_pre_adjust_full_4x(edge) \
3793 texture_offset = texture_offset_base \
3794
3795#define setup_sprite_tile_column_edge_post_adjust_half_right_4x() \
3796 fb_ptr -= 16 \
3797
3798#define setup_sprite_tile_column_edge_post_adjust_half_left_4x() \
3799
3800#define setup_sprite_tile_column_edge_post_adjust_half_4x(edge) \
3801 setup_sprite_tile_column_edge_post_adjust_half_##edge##_4x() \
3802
3803#define setup_sprite_tile_column_edge_post_adjust_full_4x(edge) \
3804
3805#define setup_sprite_offset_u_adjust() \
3806
3807#define setup_sprite_comapre_left_block_mask() \
3808 ((left_block_mask & 0xFF) == 0xFF) \
3809
3810#define setup_sprite_comapre_right_block_mask() \
3811 (((right_block_mask >> 8) & 0xFF) == 0xFF) \
3812
3813#define setup_sprite_offset_u_adjust_4x() \
3814 offset_u *= 2; \
3815 offset_u_right = offset_u_right * 2 + 1 \
3816
3817#define setup_sprite_comapre_left_block_mask_4x() \
3818 ((left_block_mask & 0xFFFF) == 0xFFFF) \
3819
3820#define setup_sprite_comapre_right_block_mask_4x() \
3821 (((right_block_mask >> 16) & 0xFFFF) == 0xFFFF) \
3822
3823
3824#define setup_sprite_tiled_do(texture_mode, x4mode) \
3825 s32 offset_u = u & 0xF; \
3826 s32 offset_v = v & 0xF; \
3827 \
3828 s32 width_rounded = offset_u + width + 15; \
3829 s32 height_rounded = offset_v + height + 15; \
3830 s32 tile_height = height_rounded / 16; \
3831 s32 tile_width = width_rounded / 16; \
3832 u32 offset_u_right = width_rounded & 0xF; \
3833 \
3834 setup_sprite_offset_u_adjust##x4mode(); \
3835 \
3836 u32 left_block_mask = ~(0xFFFFFFFF << offset_u); \
3837 u32 right_block_mask = 0xFFFFFFFE << offset_u_right; \
3838 \
3839 u32 left_mask_bits; \
3840 u32 right_mask_bits; \
3841 \
3842 u32 sub_tile_height; \
3843 u32 column_data; \
3844 \
3845 u32 texture_mask = (psx_gpu->texture_mask_width & 0xF) | \
3846 ((psx_gpu->texture_mask_height & 0xF) << 4) | \
3847 ((psx_gpu->texture_mask_width >> 4) << 8) | \
3848 ((psx_gpu->texture_mask_height >> 4) << 12); \
3849 u32 texture_offset = ((v & 0xF) << 4) | ((u & 0xF0) << 4) | \
3850 ((v & 0xF0) << 8); \
3851 u32 texture_offset_base = texture_offset; \
3852 u32 control_mask; \
3853 \
3854 u16 *fb_ptr = psx_gpu->vram_out_ptr + (y * 1024) + (x - offset_u); \
3855 u32 num_blocks = psx_gpu->num_blocks; \
3856 block_struct *block = psx_gpu->blocks + num_blocks; \
3857 \
3858 u16 *texture_block_ptr; \
3859 vec_8x8u texels; \
3860 \
3861 setup_sprite_tiled_initialize_##texture_mode##x4mode(); \
3862 \
3863 control_mask = tile_width == 1; \
3864 control_mask |= (tile_height == 1) << 1; \
3865 control_mask |= setup_sprite_comapre_left_block_mask##x4mode() << 2; \
3866 control_mask |= setup_sprite_comapre_right_block_mask##x4mode() << 3; \
3867 \
3868 switch(control_mask) \
3869 { \
3870 default: \
3871 case 0x0: \
3872 setup_sprite_tile_column_width_multi(texture_mode, multi, full, full, \
3873 x4mode); \
3874 break; \
3875 \
3876 case 0x1: \
3877 setup_sprite_tile_column_width_single(texture_mode, multi, full, none, \
3878 x4mode); \
3879 break; \
3880 \
3881 case 0x2: \
3882 setup_sprite_tile_column_width_multi(texture_mode, single, full, full, \
3883 x4mode); \
3884 break; \
3885 \
3886 case 0x3: \
3887 setup_sprite_tile_column_width_single(texture_mode, single, full, none, \
3888 x4mode); \
3889 break; \
3890 \
3891 case 0x4: \
3892 setup_sprite_tile_column_width_multi(texture_mode, multi, half, full, \
3893 x4mode); \
3894 break; \
3895 \
3896 case 0x5: \
3897 setup_sprite_tile_column_width_single(texture_mode, multi, half, right, \
3898 x4mode); \
3899 break; \
3900 \
3901 case 0x6: \
3902 setup_sprite_tile_column_width_multi(texture_mode, single, half, full, \
3903 x4mode); \
3904 break; \
3905 \
3906 case 0x7: \
3907 setup_sprite_tile_column_width_single(texture_mode, single, half, right, \
3908 x4mode); \
3909 break; \
3910 \
3911 case 0x8: \
3912 setup_sprite_tile_column_width_multi(texture_mode, multi, full, half, \
3913 x4mode); \
3914 break; \
3915 \
3916 case 0x9: \
3917 setup_sprite_tile_column_width_single(texture_mode, multi, half, left, \
3918 x4mode); \
3919 break; \
3920 \
3921 case 0xA: \
3922 setup_sprite_tile_column_width_multi(texture_mode, single, full, half, \
3923 x4mode); \
3924 break; \
3925 \
3926 case 0xB: \
3927 setup_sprite_tile_column_width_single(texture_mode, single, half, left, \
3928 x4mode); \
3929 break; \
3930 \
3931 case 0xC: \
3932 setup_sprite_tile_column_width_multi(texture_mode, multi, half, half, \
3933 x4mode); \
3934 break; \
3935 \
3936 case 0xE: \
3937 setup_sprite_tile_column_width_multi(texture_mode, single, half, half, \
3938 x4mode); \
3939 break; \
3940 } \
3941
3942void setup_sprite_4bpp(psx_gpu_struct *psx_gpu, s32 x, s32 y, s32 u, s32 v,
3943 s32 width, s32 height, u32 color)
3944{
3945#if 0
3946 setup_sprite_4bpp_(psx_gpu, x, y, u, v, width, height, color);
3947 return;
3948#endif
3949 setup_sprite_tiled_do(4bpp,)
3950}
3951
3952void setup_sprite_8bpp(psx_gpu_struct *psx_gpu, s32 x, s32 y, s32 u, s32 v,
3953 s32 width, s32 height, u32 color)
3954{
3955#if 0
3956 setup_sprite_8bpp_(psx_gpu, x, y, u, v, width, height, color);
3957 return;
3958#endif
3959 setup_sprite_tiled_do(8bpp,)
3960}
3961
3962#undef draw_mask_fb_ptr_left
3963#undef draw_mask_fb_ptr_right
3964
3965void setup_sprite_4bpp_4x(psx_gpu_struct *psx_gpu, s32 x, s32 y, s32 u, s32 v,
3966 s32 width, s32 height, u32 color)
3967{
3968#if 0
3969 setup_sprite_4bpp_4x_(psx_gpu, x, y, u, v, width, height, color);
3970 return;
3971#endif
3972 setup_sprite_tiled_do(4bpp, _4x)
3973}
3974
3975void setup_sprite_8bpp_4x(psx_gpu_struct *psx_gpu, s32 x, s32 y, s32 u, s32 v,
3976 s32 width, s32 height, u32 color)
3977{
3978#if 0
3979 setup_sprite_8bpp_4x_(psx_gpu, x, y, u, v, width, height, color);
3980 return;
3981#endif
3982 setup_sprite_tiled_do(8bpp, _4x)
3983}
3984
3985
3986void scale2x_tiles8(void * __restrict__ dst_, const void * __restrict__ src_, int w8, int h)
3987{
3988#if 0
3989 scale2x_tiles8_(dst_, src_, w8, h);
3990 return;
3991#endif
3992 const u16 * __restrict__ src = src_;
3993 const u16 * __restrict__ src1;
3994 u16 * __restrict__ dst = dst_;
3995 u16 * __restrict__ dst1;
3996 gvreg a, b;
3997 int w;
3998 for (; h > 0; h--, src += 1024, dst += 1024*2)
3999 {
4000 src1 = src;
4001 dst1 = dst;
4002 for (w = w8; w > 0; w--, src1 += 8, dst1 += 8*2)
4003 {
4004 gvld1q_u16(a, src1);
4005 gvzipq_u16(a, b, a, a);
4006 gvst1q_u16(a, dst1);
4007 gvst1q_u16(b, dst1 + 8);
4008 gvst1q_u16(a, dst1 + 1024);
4009 gvst1q_u16(b, dst1 + 1024 + 8);
4010 }
4011 }
4012}
4013
4014// vim:ts=2:sw=2:expandtab