\r
.arm\r
\r
-#include "neon_eagle2x.Sinc"\r
#include "neon_normalxx.Sinc"\r
+#include "neon_eagle2x.Sinc"\r
\r
.global neon_eagle2x_8_8\r
.global neon_eagle2x_16_16\r
add r6, r1, ip @ r6 = dst + dststride\r
sub sp, sp, #64 @ sp -= 64\r
sub r3, r3, r2 @ r3 = srcstride - width\r
- vst1.64 {d8-d11}, [sp:256] @ save q4,q5\r
+ vst1.64 {d8-d11}, [sp,:256] @ save q4,q5\r
add r7, sp, #32 @ r7 = sp + 32\r
sub ip, ip, r2 @ ip = dststride - width\r
- vst1.64 {d12-d15}, [r7:256] @ save q6,q7\r
+ vst1.64 {d12-d15}, [r7,:256] @ save q6,q7\r
lsl ip, #1 @ ip = 2 * dststride - 2 * width\r
mov r7, r2 @ r7 = width\r
sub r9, r9, #2 @ r9 = height - 2\r
neon_eagle2x_8_8_line last, r4, r0, r5, r7, r1, r6, r8, 0, 0\r
\r
add ip, sp, #32 @ ip = sp + 32\r
- vld1.64 {d8-d11}, [sp:256] @ restore q4,q5\r
+ vld1.64 {d8-d11}, [sp,:256] @ restore q4,q5\r
mov sp, r10 @ sp = oldsp\r
- vld1.64 {d12-d15}, [ip:256] @ restore q6,q7\r
+ vld1.64 {d12-d15}, [ip,:256] @ restore q6,q7\r
pop {r4-r10}\r
bx lr\r
\r
add r6, r1, ip @ r6 = dst + dststride\r
sub sp, sp, #64 @ sp -= 64\r
sub r3, r3, r2, lsl #1 @ r3 = srcstride - 2 * width\r
- vst1.64 {d8-d11}, [sp:256] @ save q4,q5\r
+ vst1.64 {d8-d11}, [sp,:256] @ save q4,q5\r
add r7, sp, #32 @ r7 = sp + 32\r
sub ip, ip, r2, lsl #1 @ ip = dststride - 2 * width\r
- vst1.64 {d12-d15}, [r7:256] @ save q6,q7\r
+ vst1.64 {d12-d15}, [r7,:256] @ save q6,q7\r
lsl ip, #1 @ ip = 2 * dststride - 4 * width\r
mov r7, r2 @ r7 = width\r
sub r9, r9, #2 @ r9 = height - 2\r
neon_eagle2x_16_16_line last, r4, r0, r5, r7, r1, r6, r8, 0, 0\r
\r
add ip, sp, #32 @ ip = sp + 32\r
- vld1.64 {d8-d11}, [sp:256] @ restore q4,q5\r
+ vld1.64 {d8-d11}, [sp,:256] @ restore q4,q5\r
mov sp, r10 @ sp = oldsp\r
- vld1.64 {d12-d15}, [ip:256] @ restore q6,q7\r
+ vld1.64 {d12-d15}, [ip,:256] @ restore q6,q7\r
pop {r4-r10}\r
bx lr\r
\r
mov lr, sp @ tmpline1 = sp\r
bic sp, sp, #31 @ align sp to 32 bytes\r
sub r8, sp, #64 @ r8 = sp - 64\r
- vst1.64 {d8-d11}, [r8:256] @ save q4,q5\r
+ vst1.64 {d8-d11}, [r8,:256] @ save q4,q5\r
sub r9, sp, #32 @ r9 = sp - 32\r
- vst1.64 {d12-d15}, [r9:256] @ save q6,q7\r
+ vst1.64 {d12-d15}, [r9,:256]@ save q6,q7\r
sub sp, sp, #(36 + 64) @ sp -= (36 + 64)\r
str r6, [sp] @ oldsp = r6\r
str r5, [sp, #4] @ height = r5\r
\r
add r6, sp, #36 @ r6 = sp + 36\r
ldr sp, [sp] @ sp = oldsp\r
- vld1.64 {d8-d11}, [r6:256] @ restore q4,q5\r
+ vld1.64 {d8-d11}, [r6,:256] @ restore q4,q5\r
add ip, r6, #32 @ ip = r6 + 32\r
- vld1.64 {d12-d15}, [ip:256] @ restore q6,q7\r
+ vld1.64 {d12-d15}, [ip,:256]@ restore q6,q7\r
pop {r4-r11,lr}\r
bx lr\r
\r
.else\r
\r
.ifeq \dstalign32\r
- _neon_eagle2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1:128, \src2:128, \src3:128, \dst1, \dst2\r
+ _neon_eagle2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1 A128, \src2 A128, \src3 A128, \dst1, \dst2\r
.else\r
- _neon_eagle2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1:128, \src2:128, \src3:128, \dst1:256, \dst2:256\r
+ _neon_eagle2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1 A128, \src2 A128, \src3 A128, \dst1 A256, \dst2 A256\r
.endif\r
\r
.endif\r
.else\r
\r
.ifeq \dstalign32\r
- _neon_scale2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1:128, \src2:128, \src3:128, \dst1, \dst2\r
+ _neon_scale2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1 A128, \src2 A128, \src3 A128, \dst1, \dst2\r
.else\r
- _neon_scale2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1:128, \src2:128, \src3:128, \dst1:256, \dst2:256\r
+ _neon_scale2x_16_16_line_\part \src1, \src2, \src3, \counter, \dst1, \dst2, \reg1, \src1 A128, \src2 A128, \src3 A128, \dst1 A256, \dst2 A256\r
.endif\r
\r
.endif\r