|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
%include "libavutil/x86/x86util.asm" |
|
|
|
SECTION_RODATA |
|
|
|
tm_shuf: times 8 db 0x03, 0x80 |
|
pw_ff00: times 8 dw 0xff00 |
|
plane_shuf: db -8, -7, -6, -5, -4, -3, -2, -1 |
|
db 1, 2, 3, 4, 5, 6, 7, 8 |
|
plane8_shuf: db -4, -3, -2, -1, 0, 0, 0, 0 |
|
db 1, 2, 3, 4, 0, 0, 0, 0 |
|
pw_0to7: dw 0, 1, 2, 3, 4, 5, 6, 7 |
|
pw_1to8: dw 1, 2, 3, 4, 5, 6, 7, 8 |
|
pw_m8tom1: dw -8, -7, -6, -5, -4, -3, -2, -1 |
|
pw_m4to4: dw -4, -3, -2, -1, 1, 2, 3, 4 |
|
|
|
SECTION .text |
|
|
|
cextern pb_1 |
|
cextern pb_3 |
|
cextern pw_4 |
|
cextern pw_8 |
|
|
|
|
|
|
|
|
|
|
|
INIT_XMM sse |
|
cglobal pred16x16_vertical_8, 2,3 |
|
sub r0, r1 |
|
mov r2, 4 |
|
movaps xmm0, [r0] |
|
.loop: |
|
movaps [r0+r1*1], xmm0 |
|
movaps [r0+r1*2], xmm0 |
|
lea r0, [r0+r1*2] |
|
movaps [r0+r1*1], xmm0 |
|
movaps [r0+r1*2], xmm0 |
|
lea r0, [r0+r1*2] |
|
dec r2 |
|
jg .loop |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
%macro PRED16x16_H 0 |
|
cglobal pred16x16_horizontal_8, 2,3 |
|
mov r2, 8 |
|
%if cpuflag(ssse3) |
|
mova m2, [pb_3] |
|
%endif |
|
.loop: |
|
movd m0, [r0+r1*0-4] |
|
movd m1, [r0+r1*1-4] |
|
|
|
%if cpuflag(ssse3) |
|
pshufb m0, m2 |
|
pshufb m1, m2 |
|
%else |
|
punpcklbw m0, m0 |
|
punpcklbw m1, m1 |
|
SPLATW m0, m0, 3 |
|
SPLATW m1, m1, 3 |
|
mova [r0+r1*0+8], m0 |
|
mova [r0+r1*1+8], m1 |
|
%endif |
|
|
|
mova [r0+r1*0], m0 |
|
mova [r0+r1*1], m1 |
|
lea r0, [r0+r1*2] |
|
dec r2 |
|
jg .loop |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX mmxext |
|
PRED16x16_H |
|
INIT_XMM ssse3 |
|
PRED16x16_H |
|
|
|
|
|
|
|
|
|
|
|
%macro PRED16x16_DC 0 |
|
cglobal pred16x16_dc_8, 2,7 |
|
mov r4, r0 |
|
sub r0, r1 |
|
pxor mm0, mm0 |
|
pxor mm1, mm1 |
|
psadbw mm0, [r0+0] |
|
psadbw mm1, [r0+8] |
|
dec r0 |
|
movzx r5d, byte [r0+r1*1] |
|
paddw mm0, mm1 |
|
movd r6d, mm0 |
|
lea r0, [r0+r1*2] |
|
%rep 7 |
|
movzx r2d, byte [r0+r1*0] |
|
movzx r3d, byte [r0+r1*1] |
|
add r5d, r2d |
|
add r6d, r3d |
|
lea r0, [r0+r1*2] |
|
%endrep |
|
movzx r2d, byte [r0+r1*0] |
|
add r5d, r6d |
|
lea r2d, [r2+r5+16] |
|
shr r2d, 5 |
|
%if cpuflag(ssse3) |
|
pxor m1, m1 |
|
%endif |
|
SPLATB_REG m0, r2, m1 |
|
|
|
mov r3d, 4 |
|
.loop: |
|
mova [r4+r1*0], m0 |
|
mova [r4+r1*1], m0 |
|
lea r4, [r4+r1*2] |
|
mova [r4+r1*0], m0 |
|
mova [r4+r1*1], m0 |
|
lea r4, [r4+r1*2] |
|
dec r3d |
|
jg .loop |
|
RET |
|
%endmacro |
|
|
|
INIT_XMM sse2 |
|
PRED16x16_DC |
|
INIT_XMM ssse3 |
|
PRED16x16_DC |
|
|
|
|
|
|
|
|
|
|
|
INIT_XMM sse2 |
|
cglobal pred16x16_tm_vp8_8, 2,6,6 |
|
sub r0, r1 |
|
pxor xmm2, xmm2 |
|
movdqa xmm0, [r0] |
|
movdqa xmm1, xmm0 |
|
punpcklbw xmm0, xmm2 |
|
punpckhbw xmm1, xmm2 |
|
movzx r4d, byte [r0-1] |
|
mov r5d, 8 |
|
.loop: |
|
movzx r2d, byte [r0+r1*1-1] |
|
movzx r3d, byte [r0+r1*2-1] |
|
sub r2d, r4d |
|
sub r3d, r4d |
|
movd xmm2, r2d |
|
movd xmm4, r3d |
|
pshuflw xmm2, xmm2, 0 |
|
pshuflw xmm4, xmm4, 0 |
|
punpcklqdq xmm2, xmm2 |
|
punpcklqdq xmm4, xmm4 |
|
movdqa xmm3, xmm2 |
|
movdqa xmm5, xmm4 |
|
paddw xmm2, xmm0 |
|
paddw xmm3, xmm1 |
|
paddw xmm4, xmm0 |
|
paddw xmm5, xmm1 |
|
packuswb xmm2, xmm3 |
|
packuswb xmm4, xmm5 |
|
movdqa [r0+r1*1], xmm2 |
|
movdqa [r0+r1*2], xmm4 |
|
lea r0, [r0+r1*2] |
|
dec r5d |
|
jg .loop |
|
RET |
|
|
|
%if HAVE_AVX2_EXTERNAL |
|
INIT_YMM avx2 |
|
cglobal pred16x16_tm_vp8_8, 2, 4, 5, dst, stride, stride3, iteration |
|
sub dstq, strideq |
|
pmovzxbw m0, [dstq] |
|
vpbroadcastb xm1, [r0-1] |
|
pmovzxbw m1, xm1 |
|
psubw m0, m1 |
|
mov iterationd, 4 |
|
lea stride3q, [strideq*3] |
|
.loop: |
|
vpbroadcastb xm1, [dstq+strideq*1-1] |
|
vpbroadcastb xm2, [dstq+strideq*2-1] |
|
vpbroadcastb xm3, [dstq+stride3q-1] |
|
vpbroadcastb xm4, [dstq+strideq*4-1] |
|
pmovzxbw m1, xm1 |
|
pmovzxbw m2, xm2 |
|
pmovzxbw m3, xm3 |
|
pmovzxbw m4, xm4 |
|
paddw m1, m0 |
|
paddw m2, m0 |
|
paddw m3, m0 |
|
paddw m4, m0 |
|
vpackuswb m1, m1, m2 |
|
vpackuswb m3, m3, m4 |
|
vpermq m1, m1, q3120 |
|
vpermq m3, m3, q3120 |
|
movdqa [dstq+strideq*1], xm1 |
|
vextracti128 [dstq+strideq*2], m1, 1 |
|
movdqa [dstq+stride3q*1], xm3 |
|
vextracti128 [dstq+strideq*4], m3, 1 |
|
lea dstq, [dstq+strideq*4] |
|
dec iterationd |
|
jg .loop |
|
RET |
|
%endif |
|
|
|
|
|
|
|
|
|
|
|
%macro H264_PRED16x16_PLANE 1 |
|
cglobal pred16x16_plane_%1_8, 2,9,7 |
|
mov r2, r1 |
|
neg r1 |
|
|
|
movh m0, [r0+r1 -1] |
|
%if cpuflag(ssse3) |
|
movhps m0, [r0+r1 +8] |
|
pmaddubsw m0, [plane_shuf] |
|
%else |
|
pxor m2, m2 |
|
movh m1, [r0+r1 +8] |
|
punpcklbw m0, m2 |
|
punpcklbw m1, m2 |
|
pmullw m0, [pw_m8tom1] |
|
pmullw m1, [pw_1to8] |
|
paddw m0, m1 |
|
%endif |
|
movhlps m1, m0 |
|
paddw m0, m1 |
|
PSHUFLW m1, m0, 0xE |
|
paddw m0, m1 |
|
PSHUFLW m1, m0, 0x1 |
|
paddw m0, m1 |
|
|
|
lea r4, [r0+r2*8-1] |
|
lea r3, [r0+r2*4-1] |
|
add r4, r2 |
|
|
|
%if ARCH_X86_64 |
|
%define e_reg r8 |
|
%else |
|
%define e_reg r0 |
|
%endif |
|
|
|
movzx e_reg, byte [r3+r2*2 ] |
|
movzx r5, byte [r4+r1 ] |
|
sub r5, e_reg |
|
|
|
movzx e_reg, byte [r3+r2 ] |
|
movzx r6, byte [r4 ] |
|
sub r6, e_reg |
|
lea r5, [r5+r6*2] |
|
|
|
movzx e_reg, byte [r3+r1 ] |
|
movzx r6, byte [r4+r2*2 ] |
|
sub r6, e_reg |
|
lea r5, [r5+r6*4] |
|
|
|
movzx e_reg, byte [r3 ] |
|
%if ARCH_X86_64 |
|
movzx r7, byte [r4+r2 ] |
|
sub r7, e_reg |
|
%else |
|
movzx r6, byte [r4+r2 ] |
|
sub r6, e_reg |
|
lea r5, [r5+r6*4] |
|
sub r5, r6 |
|
%endif |
|
|
|
lea e_reg, [r3+r1*4] |
|
lea r3, [r4+r2*4] |
|
|
|
movzx r4, byte [e_reg+r2 ] |
|
movzx r6, byte [r3 ] |
|
sub r6, r4 |
|
%if ARCH_X86_64 |
|
lea r6, [r7+r6*2] |
|
lea r5, [r5+r6*2] |
|
add r5, r6 |
|
%else |
|
lea r5, [r5+r6*4] |
|
lea r5, [r5+r6*2] |
|
%endif |
|
|
|
movzx r4, byte [e_reg ] |
|
%if ARCH_X86_64 |
|
movzx r7, byte [r3 +r2 ] |
|
sub r7, r4 |
|
sub r5, r7 |
|
%else |
|
movzx r6, byte [r3 +r2 ] |
|
sub r6, r4 |
|
lea r5, [r5+r6*8] |
|
sub r5, r6 |
|
%endif |
|
|
|
movzx r4, byte [e_reg+r1 ] |
|
movzx r6, byte [r3 +r2*2] |
|
sub r6, r4 |
|
%if ARCH_X86_64 |
|
add r6, r7 |
|
%endif |
|
lea r5, [r5+r6*8] |
|
|
|
movzx r4, byte [e_reg+r2*2] |
|
movzx r6, byte [r3 +r1 ] |
|
sub r6, r4 |
|
lea r5, [r5+r6*4] |
|
add r5, r6 |
|
|
|
%if ARCH_X86_64 == 0 |
|
mov r0, r0m |
|
%endif |
|
|
|
%ifidn %1, h264 |
|
lea r5, [r5*5+32] |
|
sar r5, 6 |
|
%elifidn %1, rv40 |
|
lea r5, [r5*5] |
|
sar r5, 6 |
|
%elifidn %1, svq3 |
|
test r5, r5 |
|
lea r6, [r5+3] |
|
cmovs r5, r6 |
|
sar r5, 2 |
|
lea r5, [r5*5] |
|
test r5, r5 |
|
lea r6, [r5+15] |
|
cmovs r5, r6 |
|
sar r5, 4 |
|
%endif |
|
|
|
movzx r4, byte [r0+r1 +15] |
|
movzx r3, byte [r3+r2*2 ] |
|
lea r3, [r3+r4+1] |
|
shl r3, 4 |
|
|
|
movd r1d, m0 |
|
movsx r1d, r1w |
|
%ifnidn %1, svq3 |
|
%ifidn %1, h264 |
|
lea r1d, [r1d*5+32] |
|
%else |
|
lea r1d, [r1d*5] |
|
%endif |
|
sar r1d, 6 |
|
%else |
|
test r1d, r1d |
|
lea r4d, [r1d+3] |
|
cmovs r1d, r4d |
|
sar r1d, 2 |
|
lea r1d, [r1d*5] |
|
test r1d, r1d |
|
lea r4d, [r1d+15] |
|
cmovs r1d, r4d |
|
sar r1d, 4 |
|
%endif |
|
movd m0, r1d |
|
|
|
add r1d, r5d |
|
add r3d, r1d |
|
shl r1d, 3 |
|
sub r3d, r1d |
|
|
|
movd m1, r5d |
|
movd m3, r3d |
|
SPLATW m0, m0, 0 |
|
SPLATW m1, m1, 0 |
|
SPLATW m3, m3, 0 |
|
%ifidn %1, svq3 |
|
SWAP 0, 1 |
|
%endif |
|
mova m2, m0 |
|
pmullw m0, [pw_0to7] |
|
psllw m2, 3 |
|
paddw m0, m3 |
|
paddw m2, m0 |
|
|
|
mov r4, 8 |
|
.loop: |
|
mova m3, m0 |
|
mova m4, m2 |
|
psraw m3, 5 |
|
psraw m4, 5 |
|
packuswb m3, m4 |
|
mova [r0], m3 |
|
paddw m0, m1 |
|
paddw m2, m1 |
|
|
|
mova m3, m0 |
|
mova m4, m2 |
|
psraw m3, 5 |
|
psraw m4, 5 |
|
packuswb m3, m4 |
|
mova [r0+r2], m3 |
|
paddw m0, m1 |
|
paddw m2, m1 |
|
|
|
lea r0, [r0+r2*2] |
|
dec r4 |
|
jg .loop |
|
RET |
|
%endmacro |
|
|
|
INIT_XMM sse2 |
|
H264_PRED16x16_PLANE h264 |
|
H264_PRED16x16_PLANE rv40 |
|
H264_PRED16x16_PLANE svq3 |
|
INIT_XMM ssse3 |
|
H264_PRED16x16_PLANE h264 |
|
H264_PRED16x16_PLANE rv40 |
|
H264_PRED16x16_PLANE svq3 |
|
|
|
|
|
|
|
|
|
|
|
%macro H264_PRED8x8_PLANE 0 |
|
cglobal pred8x8_plane_8, 2,9,7 |
|
mov r2, r1 |
|
neg r1 |
|
|
|
movd m0, [r0+r1 -1] |
|
%if cpuflag(ssse3) |
|
movhps m0, [r0+r1 +4] |
|
pmaddubsw m0, [plane8_shuf] |
|
%else |
|
pxor m2, m2 |
|
movd m1, [r0+r1 +4] |
|
punpckldq m0, m1 |
|
punpcklbw m0, m2 |
|
pmullw m0, [pw_m4to4] |
|
%endif |
|
movhlps m1, m0 |
|
paddw m0, m1 |
|
|
|
%if notcpuflag(ssse3) |
|
PSHUFLW m1, m0, 0xE |
|
paddw m0, m1 |
|
%endif |
|
|
|
PSHUFLW m1, m0, 0x1 |
|
paddw m0, m1 |
|
|
|
lea r4, [r0+r2*4-1] |
|
lea r3, [r0 -1] |
|
add r4, r2 |
|
|
|
%if ARCH_X86_64 |
|
%define e_reg r8 |
|
%else |
|
%define e_reg r0 |
|
%endif |
|
|
|
movzx e_reg, byte [r3+r2*2 ] |
|
movzx r5, byte [r4+r1 ] |
|
sub r5, e_reg |
|
|
|
movzx e_reg, byte [r3 ] |
|
%if ARCH_X86_64 |
|
movzx r7, byte [r4+r2 ] |
|
sub r7, e_reg |
|
sub r5, r7 |
|
%else |
|
movzx r6, byte [r4+r2 ] |
|
sub r6, e_reg |
|
lea r5, [r5+r6*4] |
|
sub r5, r6 |
|
%endif |
|
|
|
movzx e_reg, byte [r3+r1 ] |
|
movzx r6, byte [r4+r2*2 ] |
|
sub r6, e_reg |
|
%if ARCH_X86_64 |
|
add r6, r7 |
|
%endif |
|
lea r5, [r5+r6*4] |
|
|
|
movzx e_reg, byte [r3+r2 ] |
|
movzx r6, byte [r4 ] |
|
sub r6, e_reg |
|
lea r6, [r5+r6*2] |
|
|
|
lea r5, [r6*9+16] |
|
lea r5, [r5+r6*8] |
|
sar r5, 5 |
|
|
|
%if ARCH_X86_64 == 0 |
|
mov r0, r0m |
|
%endif |
|
|
|
movzx r3, byte [r4+r2*2 ] |
|
movzx r4, byte [r0+r1 +7] |
|
lea r3, [r3+r4+1] |
|
shl r3, 4 |
|
movd r1d, m0 |
|
movsx r1d, r1w |
|
imul r1d, 17 |
|
add r1d, 16 |
|
sar r1d, 5 |
|
movd m0, r1d |
|
add r1d, r5d |
|
sub r3d, r1d |
|
add r1d, r1d |
|
sub r3d, r1d |
|
|
|
movd m1, r5d |
|
movd m3, r3d |
|
SPLATW m0, m0, 0 |
|
SPLATW m1, m1, 0 |
|
SPLATW m3, m3, 0 |
|
pmullw m0, [pw_0to7] |
|
paddw m0, m3 |
|
|
|
mov r4, 4 |
|
ALIGN 16 |
|
.loop: |
|
mova m3, m0 |
|
paddw m0, m1 |
|
psraw m3, 5 |
|
mova m4, m0 |
|
paddw m0, m1 |
|
psraw m4, 5 |
|
packuswb m3, m4 |
|
movh [r0], m3 |
|
movhps [r0+r2], m3 |
|
|
|
lea r0, [r0+r2*2] |
|
dec r4 |
|
jg .loop |
|
RET |
|
%endmacro |
|
|
|
INIT_XMM sse2 |
|
H264_PRED8x8_PLANE |
|
INIT_XMM ssse3 |
|
H264_PRED8x8_PLANE |
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmx |
|
cglobal pred8x8_vertical_8, 2,2 |
|
sub r0, r1 |
|
movq mm0, [r0] |
|
%rep 3 |
|
movq [r0+r1*1], mm0 |
|
movq [r0+r1*2], mm0 |
|
lea r0, [r0+r1*2] |
|
%endrep |
|
movq [r0+r1*1], mm0 |
|
movq [r0+r1*2], mm0 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8_H 0 |
|
cglobal pred8x8_horizontal_8, 2,3 |
|
mov r2, 4 |
|
%if cpuflag(ssse3) |
|
mova m2, [pb_3] |
|
%endif |
|
.loop: |
|
SPLATB_LOAD m0, r0+r1*0-1, m2 |
|
SPLATB_LOAD m1, r0+r1*1-1, m2 |
|
mova [r0+r1*0], m0 |
|
mova [r0+r1*1], m1 |
|
lea r0, [r0+r1*2] |
|
dec r2 |
|
jg .loop |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX mmxext |
|
PRED8x8_H |
|
INIT_MMX ssse3 |
|
PRED8x8_H |
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred8x8_top_dc_8, 2,5 |
|
sub r0, r1 |
|
movq mm0, [r0] |
|
pxor mm1, mm1 |
|
pxor mm2, mm2 |
|
lea r2, [r0+r1*2] |
|
punpckhbw mm1, mm0 |
|
punpcklbw mm0, mm2 |
|
psadbw mm1, mm2 |
|
lea r3, [r2+r1*2] |
|
psadbw mm0, mm2 |
|
psrlw mm1, 1 |
|
psrlw mm0, 1 |
|
pavgw mm1, mm2 |
|
lea r4, [r3+r1*2] |
|
pavgw mm0, mm2 |
|
pshufw mm1, mm1, 0 |
|
pshufw mm0, mm0, 0 |
|
packuswb mm0, mm1 |
|
movq [r0+r1*1], mm0 |
|
movq [r0+r1*2], mm0 |
|
lea r0, [r3+r1*2] |
|
movq [r2+r1*1], mm0 |
|
movq [r2+r1*2], mm0 |
|
movq [r3+r1*1], mm0 |
|
movq [r3+r1*2], mm0 |
|
movq [r0+r1*1], mm0 |
|
movq [r0+r1*2], mm0 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred8x8_dc_8, 2,5 |
|
sub r0, r1 |
|
pxor m7, m7 |
|
movd m0, [r0+0] |
|
movd m1, [r0+4] |
|
psadbw m0, m7 |
|
mov r4, r0 |
|
psadbw m1, m7 |
|
|
|
movzx r2d, byte [r0+r1*1-1] |
|
movzx r3d, byte [r0+r1*2-1] |
|
lea r0, [r0+r1*2] |
|
add r2d, r3d |
|
movzx r3d, byte [r0+r1*1-1] |
|
add r2d, r3d |
|
movzx r3d, byte [r0+r1*2-1] |
|
add r2d, r3d |
|
lea r0, [r0+r1*2] |
|
movd m2, r2d |
|
movzx r2d, byte [r0+r1*1-1] |
|
movzx r3d, byte [r0+r1*2-1] |
|
lea r0, [r0+r1*2] |
|
add r2d, r3d |
|
movzx r3d, byte [r0+r1*1-1] |
|
add r2d, r3d |
|
movzx r3d, byte [r0+r1*2-1] |
|
add r2d, r3d |
|
movd m3, r2d |
|
|
|
punpcklwd m0, m1 |
|
mov r0, r4 |
|
punpcklwd m2, m3 |
|
punpckldq m0, m2 |
|
pshufw m3, m0, 11110110b |
|
lea r2, [r0+r1*2] |
|
pshufw m0, m0, 01110100b |
|
paddw m0, m3 |
|
lea r3, [r2+r1*2] |
|
psrlw m0, 2 |
|
pavgw m0, m7 |
|
lea r4, [r3+r1*2] |
|
packuswb m0, m0 |
|
punpcklbw m0, m0 |
|
movq m1, m0 |
|
punpcklbw m0, m0 |
|
punpckhbw m1, m1 |
|
movq [r0+r1*1], m0 |
|
movq [r0+r1*2], m0 |
|
movq [r2+r1*1], m0 |
|
movq [r2+r1*2], m0 |
|
movq [r3+r1*1], m1 |
|
movq [r3+r1*2], m1 |
|
movq [r4+r1*1], m1 |
|
movq [r4+r1*2], m1 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred8x8_dc_rv40_8, 2,7 |
|
mov r4, r0 |
|
sub r0, r1 |
|
pxor mm0, mm0 |
|
psadbw mm0, [r0] |
|
dec r0 |
|
movzx r5d, byte [r0+r1*1] |
|
movd r6d, mm0 |
|
lea r0, [r0+r1*2] |
|
%rep 3 |
|
movzx r2d, byte [r0+r1*0] |
|
movzx r3d, byte [r0+r1*1] |
|
add r5d, r2d |
|
add r6d, r3d |
|
lea r0, [r0+r1*2] |
|
%endrep |
|
movzx r2d, byte [r0+r1*0] |
|
add r5d, r6d |
|
lea r2d, [r2+r5+8] |
|
shr r2d, 4 |
|
movd mm0, r2d |
|
punpcklbw mm0, mm0 |
|
pshufw mm0, mm0, 0 |
|
mov r3d, 4 |
|
.loop: |
|
movq [r4+r1*0], mm0 |
|
movq [r4+r1*1], mm0 |
|
lea r4, [r4+r1*2] |
|
dec r3d |
|
jg .loop |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
INIT_XMM sse2 |
|
cglobal pred8x8_tm_vp8_8, 2,6,4 |
|
sub r0, r1 |
|
pxor xmm1, xmm1 |
|
movq xmm0, [r0] |
|
punpcklbw xmm0, xmm1 |
|
movzx r4d, byte [r0-1] |
|
mov r5d, 4 |
|
.loop: |
|
movzx r2d, byte [r0+r1*1-1] |
|
movzx r3d, byte [r0+r1*2-1] |
|
sub r2d, r4d |
|
sub r3d, r4d |
|
movd xmm2, r2d |
|
movd xmm3, r3d |
|
pshuflw xmm2, xmm2, 0 |
|
pshuflw xmm3, xmm3, 0 |
|
punpcklqdq xmm2, xmm2 |
|
punpcklqdq xmm3, xmm3 |
|
paddw xmm2, xmm0 |
|
paddw xmm3, xmm0 |
|
packuswb xmm2, xmm3 |
|
movq [r0+r1*1], xmm2 |
|
movhps [r0+r1*2], xmm2 |
|
lea r0, [r0+r1*2] |
|
dec r5d |
|
jg .loop |
|
RET |
|
|
|
INIT_XMM ssse3 |
|
cglobal pred8x8_tm_vp8_8, 2,3,6 |
|
sub r0, r1 |
|
movdqa xmm4, [tm_shuf] |
|
pxor xmm1, xmm1 |
|
movq xmm0, [r0] |
|
punpcklbw xmm0, xmm1 |
|
movd xmm5, [r0-4] |
|
pshufb xmm5, xmm4 |
|
mov r2d, 4 |
|
.loop: |
|
movd xmm2, [r0+r1*1-4] |
|
movd xmm3, [r0+r1*2-4] |
|
pshufb xmm2, xmm4 |
|
pshufb xmm3, xmm4 |
|
psubw xmm2, xmm5 |
|
psubw xmm3, xmm5 |
|
paddw xmm2, xmm0 |
|
paddw xmm3, xmm0 |
|
packuswb xmm2, xmm3 |
|
movq [r0+r1*1], xmm2 |
|
movhps [r0+r1*2], xmm2 |
|
lea r0, [r0+r1*2] |
|
dec r2d |
|
jg .loop |
|
RET |
|
|
|
|
|
|
|
%macro PRED4x4_LOWPASS 5 |
|
mova %5, %2 |
|
pavgb %2, %3 |
|
pxor %3, %5 |
|
mova %1, %4 |
|
pand %3, [pb_1] |
|
psubusb %2, %3 |
|
pavgb %1, %2 |
|
%endmacro |
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_TOP_DC 0 |
|
cglobal pred8x8l_top_dc_8, 4,4 |
|
sub r0, r3 |
|
pxor mm7, mm7 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
jmp .body |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .body |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
.body: |
|
PRED4x4_LOWPASS mm0, mm2, mm1, mm3, mm5 |
|
psadbw mm7, mm0 |
|
paddw mm7, [pw_4] |
|
psrlw mm7, 3 |
|
pshufw mm7, mm7, 0 |
|
packuswb mm7, mm7 |
|
%rep 3 |
|
movq [r0+r3*1], mm7 |
|
movq [r0+r3*2], mm7 |
|
lea r0, [r0+r3*2] |
|
%endrep |
|
movq [r0+r3*1], mm7 |
|
movq [r0+r3*2], mm7 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX mmxext |
|
PRED8x8L_TOP_DC |
|
INIT_MMX ssse3 |
|
PRED8x8L_TOP_DC |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_DC 0 |
|
cglobal pred8x8l_dc_8, 4,5 |
|
sub r0, r3 |
|
lea r4, [r0+r3*2] |
|
movq mm0, [r0+r3*1-8] |
|
punpckhbw mm0, [r0+r3*0-8] |
|
movq mm1, [r4+r3*1-8] |
|
punpckhbw mm1, [r0+r3*2-8] |
|
mov r4, r0 |
|
punpckhwd mm1, mm0 |
|
lea r0, [r0+r3*4] |
|
movq mm2, [r0+r3*1-8] |
|
punpckhbw mm2, [r0+r3*0-8] |
|
lea r0, [r0+r3*2] |
|
movq mm3, [r0+r3*1-8] |
|
punpckhbw mm3, [r0+r3*0-8] |
|
punpckhwd mm3, mm2 |
|
punpckhdq mm3, mm1 |
|
lea r0, [r0+r3*2] |
|
movq mm0, [r0+r3*0-8] |
|
movq mm1, [r4] |
|
mov r0, r4 |
|
movq mm4, mm3 |
|
movq mm2, mm3 |
|
PALIGNR mm4, mm0, 7, mm0 |
|
PALIGNR mm1, mm2, 1, mm2 |
|
test r1d, r1d |
|
jnz .do_left |
|
.fix_lt_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm4 |
|
psrlq mm5, 56 |
|
psllq mm5, 48 |
|
pxor mm1, mm5 |
|
jmp .do_left |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .body |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
jmp .body |
|
.do_left: |
|
movq mm0, mm4 |
|
PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 |
|
movq mm4, mm0 |
|
movq mm7, mm2 |
|
PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 |
|
psllq mm1, 56 |
|
PALIGNR mm7, mm1, 7, mm3 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
.body: |
|
lea r1, [r0+r3*2] |
|
PRED4x4_LOWPASS mm6, mm2, mm1, mm3, mm5 |
|
pxor mm0, mm0 |
|
pxor mm1, mm1 |
|
lea r2, [r1+r3*2] |
|
psadbw mm0, mm7 |
|
psadbw mm1, mm6 |
|
paddw mm0, [pw_8] |
|
paddw mm0, mm1 |
|
lea r4, [r2+r3*2] |
|
psrlw mm0, 4 |
|
pshufw mm0, mm0, 0 |
|
packuswb mm0, mm0 |
|
movq [r0+r3*1], mm0 |
|
movq [r0+r3*2], mm0 |
|
movq [r1+r3*1], mm0 |
|
movq [r1+r3*2], mm0 |
|
movq [r2+r3*1], mm0 |
|
movq [r2+r3*2], mm0 |
|
movq [r4+r3*1], mm0 |
|
movq [r4+r3*2], mm0 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX mmxext |
|
PRED8x8L_DC |
|
INIT_MMX ssse3 |
|
PRED8x8L_DC |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_HORIZONTAL 0 |
|
cglobal pred8x8l_horizontal_8, 4,4 |
|
sub r0, r3 |
|
lea r2, [r0+r3*2] |
|
movq mm0, [r0+r3*1-8] |
|
test r1d, r1d |
|
lea r1, [r0+r3] |
|
cmovnz r1, r0 |
|
punpckhbw mm0, [r1+r3*0-8] |
|
movq mm1, [r2+r3*1-8] |
|
punpckhbw mm1, [r0+r3*2-8] |
|
mov r2, r0 |
|
punpckhwd mm1, mm0 |
|
lea r0, [r0+r3*4] |
|
movq mm2, [r0+r3*1-8] |
|
punpckhbw mm2, [r0+r3*0-8] |
|
lea r0, [r0+r3*2] |
|
movq mm3, [r0+r3*1-8] |
|
punpckhbw mm3, [r0+r3*0-8] |
|
punpckhwd mm3, mm2 |
|
punpckhdq mm3, mm1 |
|
lea r0, [r0+r3*2] |
|
movq mm0, [r0+r3*0-8] |
|
movq mm1, [r1+r3*0-8] |
|
mov r0, r2 |
|
movq mm4, mm3 |
|
movq mm2, mm3 |
|
PALIGNR mm4, mm0, 7, mm0 |
|
PALIGNR mm1, mm2, 1, mm2 |
|
movq mm0, mm4 |
|
PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 |
|
movq mm4, mm0 |
|
movq mm7, mm2 |
|
PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 |
|
psllq mm1, 56 |
|
PALIGNR mm7, mm1, 7, mm3 |
|
movq mm3, mm7 |
|
lea r1, [r0+r3*2] |
|
movq mm7, mm3 |
|
punpckhbw mm3, mm3 |
|
punpcklbw mm7, mm7 |
|
pshufw mm0, mm3, 0xff |
|
pshufw mm1, mm3, 0xaa |
|
lea r2, [r1+r3*2] |
|
pshufw mm2, mm3, 0x55 |
|
pshufw mm3, mm3, 0x00 |
|
pshufw mm4, mm7, 0xff |
|
pshufw mm5, mm7, 0xaa |
|
pshufw mm6, mm7, 0x55 |
|
pshufw mm7, mm7, 0x00 |
|
movq [r0+r3*1], mm0 |
|
movq [r0+r3*2], mm1 |
|
movq [r1+r3*1], mm2 |
|
movq [r1+r3*2], mm3 |
|
movq [r2+r3*1], mm4 |
|
movq [r2+r3*2], mm5 |
|
lea r0, [r2+r3*2] |
|
movq [r0+r3*1], mm6 |
|
movq [r0+r3*2], mm7 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX mmxext |
|
PRED8x8L_HORIZONTAL |
|
INIT_MMX ssse3 |
|
PRED8x8L_HORIZONTAL |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_VERTICAL 0 |
|
cglobal pred8x8l_vertical_8, 4,4 |
|
sub r0, r3 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
jmp .body |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .body |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
.body: |
|
PRED4x4_LOWPASS mm0, mm2, mm1, mm3, mm5 |
|
%rep 3 |
|
movq [r0+r3*1], mm0 |
|
movq [r0+r3*2], mm0 |
|
lea r0, [r0+r3*2] |
|
%endrep |
|
movq [r0+r3*1], mm0 |
|
movq [r0+r3*2], mm0 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX mmxext |
|
PRED8x8L_VERTICAL |
|
INIT_MMX ssse3 |
|
PRED8x8L_VERTICAL |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_DOWN_LEFT 0 |
|
cglobal pred8x8l_down_left_8, 4,4 |
|
sub r0, r3 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
jmp .do_top |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .do_top |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
jmp .do_top |
|
.fix_tr_2: |
|
punpckhbw mm3, mm3 |
|
pshufw mm1, mm3, 0xFF |
|
jmp .do_topright |
|
.do_top: |
|
PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 |
|
movq2dq xmm3, mm4 |
|
test r2d, r2d |
|
jz .fix_tr_2 |
|
movq mm0, [r0+8] |
|
movq mm5, mm0 |
|
movq mm2, mm0 |
|
movq mm4, mm0 |
|
psrlq mm5, 56 |
|
PALIGNR mm2, mm3, 7, mm3 |
|
PALIGNR mm5, mm4, 1, mm4 |
|
PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4 |
|
.do_topright: |
|
movq2dq xmm4, mm1 |
|
psrlq mm1, 56 |
|
movq2dq xmm5, mm1 |
|
lea r1, [r0+r3*2] |
|
pslldq xmm4, 8 |
|
por xmm3, xmm4 |
|
movdqa xmm2, xmm3 |
|
psrldq xmm2, 1 |
|
pslldq xmm5, 15 |
|
por xmm2, xmm5 |
|
lea r2, [r1+r3*2] |
|
movdqa xmm1, xmm3 |
|
pslldq xmm1, 1 |
|
INIT_XMM cpuname |
|
PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm4 |
|
psrldq xmm0, 1 |
|
movq [r0+r3*1], xmm0 |
|
psrldq xmm0, 1 |
|
movq [r0+r3*2], xmm0 |
|
psrldq xmm0, 1 |
|
lea r0, [r2+r3*2] |
|
movq [r1+r3*1], xmm0 |
|
psrldq xmm0, 1 |
|
movq [r1+r3*2], xmm0 |
|
psrldq xmm0, 1 |
|
movq [r2+r3*1], xmm0 |
|
psrldq xmm0, 1 |
|
movq [r2+r3*2], xmm0 |
|
psrldq xmm0, 1 |
|
movq [r0+r3*1], xmm0 |
|
psrldq xmm0, 1 |
|
movq [r0+r3*2], xmm0 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX sse2 |
|
PRED8x8L_DOWN_LEFT |
|
INIT_MMX ssse3 |
|
PRED8x8L_DOWN_LEFT |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_DOWN_RIGHT 0 |
|
cglobal pred8x8l_down_right_8, 4,5 |
|
sub r0, r3 |
|
lea r4, [r0+r3*2] |
|
movq mm0, [r0+r3*1-8] |
|
punpckhbw mm0, [r0+r3*0-8] |
|
movq mm1, [r4+r3*1-8] |
|
punpckhbw mm1, [r0+r3*2-8] |
|
mov r4, r0 |
|
punpckhwd mm1, mm0 |
|
lea r0, [r0+r3*4] |
|
movq mm2, [r0+r3*1-8] |
|
punpckhbw mm2, [r0+r3*0-8] |
|
lea r0, [r0+r3*2] |
|
movq mm3, [r0+r3*1-8] |
|
punpckhbw mm3, [r0+r3*0-8] |
|
punpckhwd mm3, mm2 |
|
punpckhdq mm3, mm1 |
|
lea r0, [r0+r3*2] |
|
movq mm0, [r0+r3*0-8] |
|
movq mm1, [r4] |
|
mov r0, r4 |
|
movq mm4, mm3 |
|
movq mm2, mm3 |
|
PALIGNR mm4, mm0, 7, mm0 |
|
PALIGNR mm1, mm2, 1, mm2 |
|
test r1d, r1d |
|
jz .fix_lt_1 |
|
jmp .do_left |
|
.fix_lt_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm4 |
|
psrlq mm5, 56 |
|
psllq mm5, 48 |
|
pxor mm1, mm5 |
|
jmp .do_left |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .do_top |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
jmp .do_top |
|
.do_left: |
|
movq mm0, mm4 |
|
PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 |
|
movq mm4, mm0 |
|
movq mm7, mm2 |
|
movq2dq xmm3, mm2 |
|
PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 |
|
psllq mm1, 56 |
|
PALIGNR mm7, mm1, 7, mm3 |
|
movq2dq xmm1, mm7 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
.do_top: |
|
PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 |
|
movq2dq xmm4, mm4 |
|
lea r1, [r0+r3*2] |
|
movdqa xmm0, xmm3 |
|
pslldq xmm4, 8 |
|
por xmm3, xmm4 |
|
lea r2, [r1+r3*2] |
|
pslldq xmm4, 1 |
|
por xmm1, xmm4 |
|
psrldq xmm0, 7 |
|
pslldq xmm0, 15 |
|
psrldq xmm0, 7 |
|
por xmm1, xmm0 |
|
lea r0, [r2+r3*2] |
|
movdqa xmm2, xmm3 |
|
psrldq xmm2, 1 |
|
INIT_XMM cpuname |
|
PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm4 |
|
movdqa xmm1, xmm0 |
|
psrldq xmm1, 1 |
|
movq [r0+r3*2], xmm0 |
|
movq [r0+r3*1], xmm1 |
|
psrldq xmm0, 2 |
|
psrldq xmm1, 2 |
|
movq [r2+r3*2], xmm0 |
|
movq [r2+r3*1], xmm1 |
|
psrldq xmm0, 2 |
|
psrldq xmm1, 2 |
|
movq [r1+r3*2], xmm0 |
|
movq [r1+r3*1], xmm1 |
|
psrldq xmm0, 2 |
|
psrldq xmm1, 2 |
|
movq [r4+r3*2], xmm0 |
|
movq [r4+r3*1], xmm1 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX sse2 |
|
PRED8x8L_DOWN_RIGHT |
|
INIT_MMX ssse3 |
|
PRED8x8L_DOWN_RIGHT |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_VERTICAL_RIGHT 0 |
|
cglobal pred8x8l_vertical_right_8, 4,5,7 |
|
|
|
|
|
WIN64_SPILL_XMM 7 |
|
sub r0, r3 |
|
lea r4, [r0+r3*2] |
|
movq mm0, [r0+r3*1-8] |
|
punpckhbw mm0, [r0+r3*0-8] |
|
movq mm1, [r4+r3*1-8] |
|
punpckhbw mm1, [r0+r3*2-8] |
|
mov r4, r0 |
|
punpckhwd mm1, mm0 |
|
lea r0, [r0+r3*4] |
|
movq mm2, [r0+r3*1-8] |
|
punpckhbw mm2, [r0+r3*0-8] |
|
lea r0, [r0+r3*2] |
|
movq mm3, [r0+r3*1-8] |
|
punpckhbw mm3, [r0+r3*0-8] |
|
punpckhwd mm3, mm2 |
|
punpckhdq mm3, mm1 |
|
lea r0, [r0+r3*2] |
|
movq mm0, [r0+r3*0-8] |
|
movq mm1, [r4] |
|
mov r0, r4 |
|
movq mm4, mm3 |
|
movq mm2, mm3 |
|
PALIGNR mm4, mm0, 7, mm0 |
|
PALIGNR mm1, mm2, 1, mm2 |
|
test r1d, r1d |
|
jnz .do_left |
|
.fix_lt_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm4 |
|
psrlq mm5, 56 |
|
psllq mm5, 48 |
|
pxor mm1, mm5 |
|
jmp .do_left |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .do_top |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
jmp .do_top |
|
.do_left: |
|
movq mm0, mm4 |
|
PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 |
|
movq2dq xmm0, mm2 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
.do_top: |
|
PRED4x4_LOWPASS mm6, mm2, mm1, mm3, mm5 |
|
lea r1, [r0+r3*2] |
|
movq2dq xmm4, mm6 |
|
pslldq xmm4, 8 |
|
por xmm0, xmm4 |
|
movdqa xmm6, [pw_ff00] |
|
movdqa xmm1, xmm0 |
|
lea r2, [r1+r3*2] |
|
movdqa xmm2, xmm0 |
|
movdqa xmm3, xmm0 |
|
pslldq xmm0, 1 |
|
pslldq xmm1, 2 |
|
pavgb xmm2, xmm0 |
|
INIT_XMM cpuname |
|
PRED4x4_LOWPASS xmm4, xmm3, xmm1, xmm0, xmm5 |
|
pandn xmm6, xmm4 |
|
movdqa xmm5, xmm4 |
|
psrlw xmm4, 8 |
|
packuswb xmm6, xmm4 |
|
movhlps xmm4, xmm6 |
|
movhps [r0+r3*2], xmm5 |
|
movhps [r0+r3*1], xmm2 |
|
psrldq xmm5, 4 |
|
movss xmm5, xmm6 |
|
psrldq xmm2, 4 |
|
movss xmm2, xmm4 |
|
lea r0, [r2+r3*2] |
|
psrldq xmm5, 1 |
|
psrldq xmm2, 1 |
|
movq [r0+r3*2], xmm5 |
|
movq [r0+r3*1], xmm2 |
|
psrldq xmm5, 1 |
|
psrldq xmm2, 1 |
|
movq [r2+r3*2], xmm5 |
|
movq [r2+r3*1], xmm2 |
|
psrldq xmm5, 1 |
|
psrldq xmm2, 1 |
|
movq [r1+r3*2], xmm5 |
|
movq [r1+r3*1], xmm2 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX sse2 |
|
PRED8x8L_VERTICAL_RIGHT |
|
INIT_MMX ssse3 |
|
PRED8x8L_VERTICAL_RIGHT |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_VERTICAL_LEFT 0 |
|
cglobal pred8x8l_vertical_left_8, 4,4 |
|
sub r0, r3 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
jmp .do_top |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .do_top |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
jmp .do_top |
|
.fix_tr_2: |
|
punpckhbw mm3, mm3 |
|
pshufw mm1, mm3, 0xFF |
|
jmp .do_topright |
|
.do_top: |
|
PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 |
|
movq2dq xmm4, mm4 |
|
test r2d, r2d |
|
jz .fix_tr_2 |
|
movq mm0, [r0+8] |
|
movq mm5, mm0 |
|
movq mm2, mm0 |
|
movq mm4, mm0 |
|
psrlq mm5, 56 |
|
PALIGNR mm2, mm3, 7, mm3 |
|
PALIGNR mm5, mm4, 1, mm4 |
|
PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4 |
|
.do_topright: |
|
movq2dq xmm3, mm1 |
|
lea r1, [r0+r3*2] |
|
pslldq xmm3, 8 |
|
por xmm4, xmm3 |
|
movdqa xmm2, xmm4 |
|
movdqa xmm1, xmm4 |
|
movdqa xmm3, xmm4 |
|
psrldq xmm2, 1 |
|
pslldq xmm1, 1 |
|
pavgb xmm3, xmm2 |
|
lea r2, [r1+r3*2] |
|
INIT_XMM cpuname |
|
PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm4, xmm5 |
|
psrldq xmm0, 1 |
|
movq [r0+r3*1], xmm3 |
|
movq [r0+r3*2], xmm0 |
|
lea r0, [r2+r3*2] |
|
psrldq xmm3, 1 |
|
psrldq xmm0, 1 |
|
movq [r1+r3*1], xmm3 |
|
movq [r1+r3*2], xmm0 |
|
psrldq xmm3, 1 |
|
psrldq xmm0, 1 |
|
movq [r2+r3*1], xmm3 |
|
movq [r2+r3*2], xmm0 |
|
psrldq xmm3, 1 |
|
psrldq xmm0, 1 |
|
movq [r0+r3*1], xmm3 |
|
movq [r0+r3*2], xmm0 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX sse2 |
|
PRED8x8L_VERTICAL_LEFT |
|
INIT_MMX ssse3 |
|
PRED8x8L_VERTICAL_LEFT |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_HORIZONTAL_UP 0 |
|
cglobal pred8x8l_horizontal_up_8, 4,4 |
|
sub r0, r3 |
|
lea r2, [r0+r3*2] |
|
movq mm0, [r0+r3*1-8] |
|
test r1d, r1d |
|
lea r1, [r0+r3] |
|
cmovnz r1, r0 |
|
punpckhbw mm0, [r1+r3*0-8] |
|
movq mm1, [r2+r3*1-8] |
|
punpckhbw mm1, [r0+r3*2-8] |
|
mov r2, r0 |
|
punpckhwd mm1, mm0 |
|
lea r0, [r0+r3*4] |
|
movq mm2, [r0+r3*1-8] |
|
punpckhbw mm2, [r0+r3*0-8] |
|
lea r0, [r0+r3*2] |
|
movq mm3, [r0+r3*1-8] |
|
punpckhbw mm3, [r0+r3*0-8] |
|
punpckhwd mm3, mm2 |
|
punpckhdq mm3, mm1 |
|
lea r0, [r0+r3*2] |
|
movq mm0, [r0+r3*0-8] |
|
movq mm1, [r1+r3*0-8] |
|
mov r0, r2 |
|
movq mm4, mm3 |
|
movq mm2, mm3 |
|
PALIGNR mm4, mm0, 7, mm0 |
|
PALIGNR mm1, mm2, 1, mm2 |
|
movq mm0, mm4 |
|
PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 |
|
movq mm4, mm0 |
|
movq mm7, mm2 |
|
PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 |
|
psllq mm1, 56 |
|
PALIGNR mm7, mm1, 7, mm3 |
|
lea r1, [r0+r3*2] |
|
pshufw mm0, mm7, 00011011b |
|
psllq mm7, 56 |
|
movq mm2, mm0 |
|
psllw mm0, 8 |
|
psrlw mm2, 8 |
|
por mm2, mm0 |
|
movq mm3, mm2 |
|
movq mm4, mm2 |
|
movq mm5, mm2 |
|
psrlq mm2, 8 |
|
psrlq mm3, 16 |
|
lea r2, [r1+r3*2] |
|
por mm2, mm7 |
|
punpckhbw mm7, mm7 |
|
por mm3, mm7 |
|
pavgb mm4, mm2 |
|
PRED4x4_LOWPASS mm1, mm3, mm5, mm2, mm6 |
|
movq mm5, mm4 |
|
punpcklbw mm4, mm1 |
|
punpckhbw mm5, mm1 |
|
movq mm6, mm5 |
|
movq mm7, mm5 |
|
movq mm0, mm5 |
|
PALIGNR mm5, mm4, 2, mm1 |
|
pshufw mm1, mm6, 11111001b |
|
PALIGNR mm6, mm4, 4, mm2 |
|
pshufw mm2, mm7, 11111110b |
|
PALIGNR mm7, mm4, 6, mm3 |
|
pshufw mm3, mm0, 11111111b |
|
movq [r0+r3*1], mm4 |
|
movq [r0+r3*2], mm5 |
|
lea r0, [r2+r3*2] |
|
movq [r1+r3*1], mm6 |
|
movq [r1+r3*2], mm7 |
|
movq [r2+r3*1], mm0 |
|
movq [r2+r3*2], mm1 |
|
movq [r0+r3*1], mm2 |
|
movq [r0+r3*2], mm3 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX mmxext |
|
PRED8x8L_HORIZONTAL_UP |
|
INIT_MMX ssse3 |
|
PRED8x8L_HORIZONTAL_UP |
|
|
|
|
|
|
|
|
|
|
|
|
|
%macro PRED8x8L_HORIZONTAL_DOWN 0 |
|
cglobal pred8x8l_horizontal_down_8, 4,5 |
|
sub r0, r3 |
|
lea r4, [r0+r3*2] |
|
movq mm0, [r0+r3*1-8] |
|
punpckhbw mm0, [r0+r3*0-8] |
|
movq mm1, [r4+r3*1-8] |
|
punpckhbw mm1, [r0+r3*2-8] |
|
mov r4, r0 |
|
punpckhwd mm1, mm0 |
|
lea r0, [r0+r3*4] |
|
movq mm2, [r0+r3*1-8] |
|
punpckhbw mm2, [r0+r3*0-8] |
|
lea r0, [r0+r3*2] |
|
movq mm3, [r0+r3*1-8] |
|
punpckhbw mm3, [r0+r3*0-8] |
|
punpckhwd mm3, mm2 |
|
punpckhdq mm3, mm1 |
|
lea r0, [r0+r3*2] |
|
movq mm0, [r0+r3*0-8] |
|
movq mm1, [r4] |
|
mov r0, r4 |
|
movq mm4, mm3 |
|
movq mm2, mm3 |
|
PALIGNR mm4, mm0, 7, mm0 |
|
PALIGNR mm1, mm2, 1, mm2 |
|
test r1d, r1d |
|
jnz .do_left |
|
.fix_lt_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm4 |
|
psrlq mm5, 56 |
|
psllq mm5, 48 |
|
pxor mm1, mm5 |
|
jmp .do_left |
|
.fix_lt_2: |
|
movq mm5, mm3 |
|
pxor mm5, mm2 |
|
psllq mm5, 56 |
|
psrlq mm5, 56 |
|
pxor mm2, mm5 |
|
test r2d, r2d |
|
jnz .do_top |
|
.fix_tr_1: |
|
movq mm5, mm3 |
|
pxor mm5, mm1 |
|
psrlq mm5, 56 |
|
psllq mm5, 56 |
|
pxor mm1, mm5 |
|
jmp .do_top |
|
.fix_tr_2: |
|
punpckhbw mm3, mm3 |
|
pshufw mm1, mm3, 0xFF |
|
jmp .do_topright |
|
.do_left: |
|
movq mm0, mm4 |
|
PRED4x4_LOWPASS mm2, mm1, mm4, mm3, mm5 |
|
movq2dq xmm0, mm2 |
|
pslldq xmm0, 8 |
|
movq mm4, mm0 |
|
PRED4x4_LOWPASS mm1, mm3, mm0, mm4, mm5 |
|
movq2dq xmm2, mm1 |
|
pslldq xmm2, 15 |
|
psrldq xmm2, 8 |
|
por xmm0, xmm2 |
|
movq mm0, [r0-8] |
|
movq mm3, [r0] |
|
movq mm1, [r0+8] |
|
movq mm2, mm3 |
|
movq mm4, mm3 |
|
PALIGNR mm2, mm0, 7, mm0 |
|
PALIGNR mm1, mm4, 1, mm4 |
|
test r1d, r1d |
|
jz .fix_lt_2 |
|
test r2d, r2d |
|
jz .fix_tr_1 |
|
.do_top: |
|
PRED4x4_LOWPASS mm4, mm2, mm1, mm3, mm5 |
|
movq2dq xmm1, mm4 |
|
test r2d, r2d |
|
jz .fix_tr_2 |
|
movq mm0, [r0+8] |
|
movq mm5, mm0 |
|
movq mm2, mm0 |
|
movq mm4, mm0 |
|
psrlq mm5, 56 |
|
PALIGNR mm2, mm3, 7, mm3 |
|
PALIGNR mm5, mm4, 1, mm4 |
|
PRED4x4_LOWPASS mm1, mm2, mm5, mm0, mm4 |
|
.do_topright: |
|
movq2dq xmm5, mm1 |
|
pslldq xmm5, 8 |
|
por xmm1, xmm5 |
|
INIT_XMM cpuname |
|
lea r2, [r4+r3*2] |
|
movdqa xmm2, xmm1 |
|
movdqa xmm3, xmm1 |
|
PALIGNR xmm1, xmm0, 7, xmm4 |
|
PALIGNR xmm2, xmm0, 9, xmm5 |
|
lea r1, [r2+r3*2] |
|
PALIGNR xmm3, xmm0, 8, xmm0 |
|
movdqa xmm4, xmm1 |
|
pavgb xmm4, xmm3 |
|
lea r0, [r1+r3*2] |
|
PRED4x4_LOWPASS xmm0, xmm1, xmm2, xmm3, xmm5 |
|
punpcklbw xmm4, xmm0 |
|
movhlps xmm0, xmm4 |
|
movq [r0+r3*2], xmm4 |
|
movq [r2+r3*2], xmm0 |
|
psrldq xmm4, 2 |
|
psrldq xmm0, 2 |
|
movq [r0+r3*1], xmm4 |
|
movq [r2+r3*1], xmm0 |
|
psrldq xmm4, 2 |
|
psrldq xmm0, 2 |
|
movq [r1+r3*2], xmm4 |
|
movq [r4+r3*2], xmm0 |
|
psrldq xmm4, 2 |
|
psrldq xmm0, 2 |
|
movq [r1+r3*1], xmm4 |
|
movq [r4+r3*1], xmm0 |
|
RET |
|
%endmacro |
|
|
|
INIT_MMX sse2 |
|
PRED8x8L_HORIZONTAL_DOWN |
|
INIT_MMX ssse3 |
|
PRED8x8L_HORIZONTAL_DOWN |
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_dc_8, 3,5 |
|
pxor mm7, mm7 |
|
mov r4, r0 |
|
sub r0, r2 |
|
movd mm0, [r0] |
|
psadbw mm0, mm7 |
|
movzx r1d, byte [r0+r2*1-1] |
|
movd r3d, mm0 |
|
add r3d, r1d |
|
movzx r1d, byte [r0+r2*2-1] |
|
lea r0, [r0+r2*2] |
|
add r3d, r1d |
|
movzx r1d, byte [r0+r2*1-1] |
|
add r3d, r1d |
|
movzx r1d, byte [r0+r2*2-1] |
|
add r3d, r1d |
|
add r3d, 4 |
|
shr r3d, 3 |
|
imul r3d, 0x01010101 |
|
mov [r4+r2*0], r3d |
|
mov [r0+r2*0], r3d |
|
mov [r0+r2*1], r3d |
|
mov [r0+r2*2], r3d |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_tm_vp8_8, 3,6 |
|
sub r0, r2 |
|
pxor mm7, mm7 |
|
movd mm0, [r0] |
|
punpcklbw mm0, mm7 |
|
movzx r4d, byte [r0-1] |
|
mov r5d, 2 |
|
.loop: |
|
movzx r1d, byte [r0+r2*1-1] |
|
movzx r3d, byte [r0+r2*2-1] |
|
sub r1d, r4d |
|
sub r3d, r4d |
|
movd mm2, r1d |
|
movd mm4, r3d |
|
pshufw mm2, mm2, 0 |
|
pshufw mm4, mm4, 0 |
|
paddw mm2, mm0 |
|
paddw mm4, mm0 |
|
packuswb mm2, mm2 |
|
packuswb mm4, mm4 |
|
movd [r0+r2*1], mm2 |
|
movd [r0+r2*2], mm4 |
|
lea r0, [r0+r2*2] |
|
dec r5d |
|
jg .loop |
|
RET |
|
|
|
INIT_XMM ssse3 |
|
cglobal pred4x4_tm_vp8_8, 3,3 |
|
sub r0, r2 |
|
movq mm6, [tm_shuf] |
|
pxor mm1, mm1 |
|
movd mm0, [r0] |
|
punpcklbw mm0, mm1 |
|
movd mm7, [r0-4] |
|
pshufb mm7, mm6 |
|
lea r1, [r0+r2*2] |
|
movd mm2, [r0+r2*1-4] |
|
movd mm3, [r0+r2*2-4] |
|
movd mm4, [r1+r2*1-4] |
|
movd mm5, [r1+r2*2-4] |
|
pshufb mm2, mm6 |
|
pshufb mm3, mm6 |
|
pshufb mm4, mm6 |
|
pshufb mm5, mm6 |
|
psubw mm0, mm7 |
|
paddw mm2, mm0 |
|
paddw mm3, mm0 |
|
paddw mm4, mm0 |
|
paddw mm5, mm0 |
|
packuswb mm2, mm2 |
|
packuswb mm3, mm3 |
|
packuswb mm4, mm4 |
|
packuswb mm5, mm5 |
|
movd [r0+r2*1], mm2 |
|
movd [r0+r2*2], mm3 |
|
movd [r1+r2*1], mm4 |
|
movd [r1+r2*2], mm5 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_vertical_vp8_8, 3,3 |
|
sub r0, r2 |
|
movd m1, [r0-1] |
|
movd m0, [r0] |
|
mova m2, m0 |
|
punpckldq m0, [r1] |
|
lea r1, [r0+r2*2] |
|
psrlq m0, 8 |
|
PRED4x4_LOWPASS m3, m1, m0, m2, m4 |
|
movd [r0+r2*1], m3 |
|
movd [r0+r2*2], m3 |
|
movd [r1+r2*1], m3 |
|
movd [r1+r2*2], m3 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_down_left_8, 3,3 |
|
sub r0, r2 |
|
movq m1, [r0] |
|
punpckldq m1, [r1] |
|
movq m2, m1 |
|
movq m3, m1 |
|
psllq m1, 8 |
|
pxor m2, m1 |
|
psrlq m2, 8 |
|
pxor m2, m3 |
|
PRED4x4_LOWPASS m0, m1, m2, m3, m4 |
|
lea r1, [r0+r2*2] |
|
psrlq m0, 8 |
|
movd [r0+r2*1], m0 |
|
psrlq m0, 8 |
|
movd [r0+r2*2], m0 |
|
psrlq m0, 8 |
|
movd [r1+r2*1], m0 |
|
psrlq m0, 8 |
|
movd [r1+r2*2], m0 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_vertical_left_8, 3,3 |
|
sub r0, r2 |
|
movq m1, [r0] |
|
punpckldq m1, [r1] |
|
movq m3, m1 |
|
movq m2, m1 |
|
psrlq m3, 8 |
|
psrlq m2, 16 |
|
movq m4, m3 |
|
pavgb m4, m1 |
|
PRED4x4_LOWPASS m0, m1, m2, m3, m5 |
|
lea r1, [r0+r2*2] |
|
movh [r0+r2*1], m4 |
|
movh [r0+r2*2], m0 |
|
psrlq m4, 8 |
|
psrlq m0, 8 |
|
movh [r1+r2*1], m4 |
|
movh [r1+r2*2], m0 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_horizontal_up_8, 3,3 |
|
sub r0, r2 |
|
lea r1, [r0+r2*2] |
|
movd m0, [r0+r2*1-4] |
|
punpcklbw m0, [r0+r2*2-4] |
|
movd m1, [r1+r2*1-4] |
|
punpcklbw m1, [r1+r2*2-4] |
|
punpckhwd m0, m1 |
|
movq m1, m0 |
|
punpckhbw m1, m1 |
|
pshufw m1, m1, 0xFF |
|
punpckhdq m0, m1 |
|
movq m2, m0 |
|
movq m3, m0 |
|
movq m7, m0 |
|
psrlq m2, 16 |
|
psrlq m3, 8 |
|
pavgb m7, m3 |
|
PRED4x4_LOWPASS m4, m0, m2, m3, m5 |
|
punpcklbw m7, m4 |
|
movd [r0+r2*1], m7 |
|
psrlq m7, 16 |
|
movd [r0+r2*2], m7 |
|
psrlq m7, 16 |
|
movd [r1+r2*1], m7 |
|
movd [r1+r2*2], m1 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_horizontal_down_8, 3,3 |
|
sub r0, r2 |
|
lea r1, [r0+r2*2] |
|
movh m0, [r0-4] |
|
punpckldq m0, [r0] |
|
psllq m0, 8 |
|
movd m1, [r1+r2*2-4] |
|
punpcklbw m1, [r1+r2*1-4] |
|
movd m2, [r0+r2*2-4] |
|
punpcklbw m2, [r0+r2*1-4] |
|
punpckhwd m1, m2 |
|
punpckhdq m1, m0 |
|
movq m0, m1 |
|
movq m2, m1 |
|
movq m5, m1 |
|
psrlq m0, 16 |
|
psrlq m2, 8 |
|
pavgb m5, m2 |
|
PRED4x4_LOWPASS m3, m1, m0, m2, m4 |
|
punpcklbw m5, m3 |
|
psrlq m3, 32 |
|
PALIGNR m3, m5, 6, m4 |
|
movh [r1+r2*2], m5 |
|
psrlq m5, 16 |
|
movh [r1+r2*1], m5 |
|
psrlq m5, 16 |
|
movh [r0+r2*2], m5 |
|
movh [r0+r2*1], m3 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_vertical_right_8, 3,3 |
|
sub r0, r2 |
|
lea r1, [r0+r2*2] |
|
movh m0, [r0] |
|
movq m5, m0 |
|
PALIGNR m0, [r0-8], 7, m1 |
|
pavgb m5, m0 |
|
PALIGNR m0, [r0+r2*1-8], 7, m1 |
|
movq m1, m0 |
|
PALIGNR m0, [r0+r2*2-8], 7, m2 |
|
movq m2, m0 |
|
PALIGNR m0, [r1+r2*1-8], 7, m3 |
|
PRED4x4_LOWPASS m3, m1, m0, m2, m4 |
|
movq m1, m3 |
|
psrlq m3, 16 |
|
psllq m1, 48 |
|
movh [r0+r2*1], m5 |
|
movh [r0+r2*2], m3 |
|
PALIGNR m5, m1, 7, m2 |
|
psllq m1, 8 |
|
movh [r1+r2*1], m5 |
|
PALIGNR m3, m1, 7, m1 |
|
movh [r1+r2*2], m3 |
|
RET |
|
|
|
|
|
|
|
|
|
|
|
|
|
INIT_MMX mmxext |
|
cglobal pred4x4_down_right_8, 3,3 |
|
sub r0, r2 |
|
lea r1, [r0+r2*2] |
|
movq m1, [r1-8] |
|
movq m2, [r0+r2*1-8] |
|
punpckhbw m2, [r0-8] |
|
movh m3, [r0] |
|
punpckhwd m1, m2 |
|
PALIGNR m3, m1, 5, m1 |
|
movq m1, m3 |
|
PALIGNR m3, [r1+r2*1-8], 7, m4 |
|
movq m2, m3 |
|
PALIGNR m3, [r1+r2*2-8], 7, m4 |
|
PRED4x4_LOWPASS m0, m3, m1, m2, m4 |
|
movh [r1+r2*2], m0 |
|
psrlq m0, 8 |
|
movh [r1+r2*1], m0 |
|
psrlq m0, 8 |
|
movh [r0+r2*2], m0 |
|
psrlq m0, 8 |
|
movh [r0+r2*1], m0 |
|
RET |
|
|