|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.macro do_chroma_mc type unroll |
|
csrw vxrm, zero |
|
slli t2, a5, 3 |
|
mul t1, a5, a4 |
|
sh3add a5, a4, t2 |
|
slli a4, a4, 3 |
|
sub a5, t1, a5 |
|
sub a7, a4, t1 |
|
addi a6, a5, 64 |
|
sub t0, t2, t1 |
|
vsetvli t3, t6, e8, m1, ta, mu |
|
beqz t1, 2f |
|
blez a3, 8f |
|
li t4, 0 |
|
li t2, 0 |
|
li t5, 1 |
|
addi a5, t3, 1 |
|
slli t3, a2, (1 + \unroll) |
|
1: |
|
add a4, a1, t4 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
.ifc \unroll,1 |
|
addi t2, t2, 4 |
|
.else |
|
addi t2, t2, 2 |
|
.endif |
|
vle8.v v10, (a4) |
|
add a4, a4, a2 |
|
vslide1down.vx v11, v10, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v8, v10, a6 |
|
vwmaccu.vx v8, a7, v11 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vle8.v v12, (a4) |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
add a4, a4, a2 |
|
vwmaccu.vx v8, t0, v12 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vslide1down.vx v13, v12, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v10, v12, a6 |
|
vwmaccu.vx v8, t1, v13 |
|
vwmaccu.vx v10, a7, v13 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vle8.v v14, (a4) |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
add a4, a4, a2 |
|
vwmaccu.vx v10, t0, v14 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vslide1down.vx v15, v14, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v12, v14, a6 |
|
vwmaccu.vx v10, t1, v15 |
|
vwmaccu.vx v12, a7, v15 |
|
vnclipu.wi v15, v8, 6 |
|
.ifc \type,avg |
|
vle8.v v9, (a0) |
|
vaaddu.vv v15, v15, v9 |
|
.endif |
|
vse8.v v15, (a0) |
|
add a0, a0, a2 |
|
vnclipu.wi v8, v10, 6 |
|
.ifc \type,avg |
|
vle8.v v9, (a0) |
|
vaaddu.vv v8, v8, v9 |
|
.endif |
|
add t4, t4, t3 |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
.ifc \unroll,1 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vle8.v v14, (a4) |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
add a4, a4, a2 |
|
vwmaccu.vx v12, t0, v14 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vslide1down.vx v15, v14, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v16, v14, a6 |
|
vwmaccu.vx v12, t1, v15 |
|
vwmaccu.vx v16, a7, v15 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vle8.v v14, (a4) |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmaccu.vx v16, t0, v14 |
|
vsetvli zero, a5, e8, m1, ta, ma |
|
vslide1down.vx v14, v14, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmaccu.vx v16, t1, v14 |
|
vnclipu.wi v8, v12, 6 |
|
.ifc \type,avg |
|
vle8.v v9, (a0) |
|
vaaddu.vv v8, v8, v9 |
|
.endif |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
vnclipu.wi v8, v16, 6 |
|
.ifc \type,avg |
|
vle8.v v9, (a0) |
|
vaaddu.vv v8, v8, v9 |
|
.endif |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
.endif |
|
blt t2, a3, 1b |
|
j 8f |
|
2: |
|
bnez a4, 4f |
|
beqz t2, 4f |
|
blez a3, 8f |
|
li a4, 0 |
|
li t1, 0 |
|
slli a7, a2, (1 + \unroll) |
|
3: |
|
add a5, a1, a4 |
|
vsetvli zero, zero, e8, m1, ta, ma |
|
.ifc \unroll,1 |
|
addi t1, t1, 4 |
|
.else |
|
addi t1, t1, 2 |
|
.endif |
|
vle8.v v8, (a5) |
|
add a5, a5, a2 |
|
add t2, a5, a2 |
|
vwmulu.vx v10, v8, a6 |
|
vle8.v v8, (a5) |
|
vwmulu.vx v12, v8, a6 |
|
vle8.v v9, (t2) |
|
add t2, t2, a2 |
|
add a5, t2, a2 |
|
vwmaccu.vx v10, t0, v8 |
|
add a4, a4, a7 |
|
vwmaccu.vx v12, t0, v9 |
|
vnclipu.wi v15, v10, 6 |
|
vwmulu.vx v10, v9, a6 |
|
vnclipu.wi v9, v12, 6 |
|
.ifc \type,avg |
|
vle8.v v16, (a0) |
|
vaaddu.vv v15, v15, v16 |
|
.endif |
|
vse8.v v15, (a0) |
|
add a0, a0, a2 |
|
.ifc \type,avg |
|
vle8.v v16, (a0) |
|
vaaddu.vv v9, v9, v16 |
|
.endif |
|
vse8.v v9, (a0) |
|
add a0, a0, a2 |
|
.ifc \unroll,1 |
|
vle8.v v8, (t2) |
|
vle8.v v14, (a5) |
|
vwmaccu.vx v10, t0, v8 |
|
vwmulu.vx v12, v8, a6 |
|
vnclipu.wi v8, v10, 6 |
|
vwmaccu.vx v12, t0, v14 |
|
.ifc \type,avg |
|
vle8.v v16, (a0) |
|
vaaddu.vv v8, v8, v16 |
|
.endif |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
vnclipu.wi v8, v12, 6 |
|
.ifc \type,avg |
|
vle8.v v16, (a0) |
|
vaaddu.vv v8, v8, v16 |
|
.endif |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
.endif |
|
blt t1, a3, 3b |
|
j 8f |
|
4: |
|
beqz a4, 6f |
|
bnez t2, 6f |
|
blez a3, 8f |
|
li a4, 0 |
|
li t2, 0 |
|
addi t0, t3, 1 |
|
slli t1, a2, (1 + \unroll) |
|
5: |
|
add a5, a1, a4 |
|
vsetvli zero, t0, e8, m1, ta, ma |
|
.ifc \unroll,1 |
|
addi t2, t2, 4 |
|
.else |
|
addi t2, t2, 2 |
|
.endif |
|
vle8.v v8, (a5) |
|
add a5, a5, a2 |
|
vslide1down.vx v9, v8, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v10, v8, a6 |
|
vwmaccu.vx v10, a7, v9 |
|
vsetvli zero, t0, e8, m1, ta, ma |
|
vle8.v v8, (a5) |
|
add a5, a5, a2 |
|
vslide1down.vx v9, v8, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v12, v8, a6 |
|
vwmaccu.vx v12, a7, v9 |
|
vnclipu.wi v16, v10, 6 |
|
.ifc \type,avg |
|
vle8.v v18, (a0) |
|
vaaddu.vv v16, v16, v18 |
|
.endif |
|
vse8.v v16, (a0) |
|
add a0, a0, a2 |
|
vnclipu.wi v10, v12, 6 |
|
.ifc \type,avg |
|
vle8.v v18, (a0) |
|
vaaddu.vv v10, v10, v18 |
|
.endif |
|
add a4, a4, t1 |
|
vse8.v v10, (a0) |
|
add a0, a0, a2 |
|
.ifc \unroll,1 |
|
vsetvli zero, t0, e8, m1, ta, ma |
|
vle8.v v8, (a5) |
|
add a5, a5, a2 |
|
vslide1down.vx v9, v8, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v14, v8, a6 |
|
vwmaccu.vx v14, a7, v9 |
|
vsetvli zero, t0, e8, m1, ta, ma |
|
vle8.v v8, (a5) |
|
vslide1down.vx v9, v8, t5 |
|
vsetvli zero, t6, e8, m1, ta, ma |
|
vwmulu.vx v12, v8, a6 |
|
vnclipu.wi v8, v14, 6 |
|
vwmaccu.vx v12, a7, v9 |
|
.ifc \type,avg |
|
vle8.v v18, (a0) |
|
vaaddu.vv v8, v8, v18 |
|
.endif |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
vnclipu.wi v8, v12, 6 |
|
.ifc \type,avg |
|
vle8.v v18, (a0) |
|
vaaddu.vv v8, v8, v18 |
|
.endif |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
.endif |
|
blt t2, a3, 5b |
|
j 8f |
|
6: |
|
blez a3, 8f |
|
li a4, 0 |
|
li t2, 0 |
|
slli a7, a2, (1 + \unroll) |
|
7: |
|
add t0, a1, a4 |
|
vsetvli zero, zero, e8, m1, ta, ma |
|
add a5, a0, a4 |
|
add a4, a4, a7 |
|
.ifc \unroll,1 |
|
addi t2, t2, 4 |
|
.else |
|
addi t2, t2, 2 |
|
.endif |
|
vle8.v v8, (t0) |
|
add t0, t0, a2 |
|
add t1, t0, a2 |
|
vwmulu.vx v10, v8, a6 |
|
vle8.v v8, (t0) |
|
add t0, t1, a2 |
|
vnclipu.wi v13, v10, 6 |
|
vwmulu.vx v10, v8, a6 |
|
.ifc \type,avg |
|
vle8.v v18, (a5) |
|
vaaddu.vv v13, v13, v18 |
|
.endif |
|
vse8.v v13, (a5) |
|
add a5, a5, a2 |
|
vnclipu.wi v8, v10, 6 |
|
.ifc \type,avg |
|
vle8.v v18, (a5) |
|
vaaddu.vv v8, v8, v18 |
|
.endif |
|
vse8.v v8, (a5) |
|
add a5, a5, a2 |
|
.ifc \unroll,1 |
|
vle8.v v9, (t1) |
|
vle8.v v12, (t0) |
|
vwmulu.vx v10, v9, a6 |
|
vnclipu.wi v8, v10, 6 |
|
vwmulu.vx v10, v12, a6 |
|
.ifc \type,avg |
|
vle8.v v18, (a5) |
|
vaaddu.vv v8, v8, v18 |
|
.endif |
|
vse8.v v8, (a5) |
|
add a5, a5, a2 |
|
vnclipu.wi v8, v10, 6 |
|
.ifc \type,avg |
|
vle8.v v18, (a5) |
|
vaaddu.vv v8, v8, v18 |
|
.endif |
|
vse8.v v8, (a5) |
|
.endif |
|
blt t2, a3, 7b |
|
8: |
|
ret |
|
.endm |
|
|
|
func h264_put_chroma_mc_rvv, zve32x |
|
11: |
|
li a7, 3 |
|
blt a3, a7, 12f |
|
do_chroma_mc put 1 |
|
12: |
|
do_chroma_mc put 0 |
|
endfunc |
|
|
|
func h264_avg_chroma_mc_rvv, zve32x |
|
21: |
|
li a7, 3 |
|
blt a3, a7, 22f |
|
do_chroma_mc avg 1 |
|
22: |
|
do_chroma_mc avg 0 |
|
endfunc |
|
|
|
func h264_put_chroma_mc8_rvv, zve32x |
|
li t6, 8 |
|
j 11b |
|
endfunc |
|
|
|
func h264_put_chroma_mc4_rvv, zve32x |
|
li t6, 4 |
|
j 11b |
|
endfunc |
|
|
|
func h264_put_chroma_mc2_rvv, zve32x |
|
li t6, 2 |
|
j 11b |
|
endfunc |
|
|
|
func h264_avg_chroma_mc8_rvv, zve32x |
|
li t6, 8 |
|
j 21b |
|
endfunc |
|
|
|
func h264_avg_chroma_mc4_rvv, zve32x |
|
li t6, 4 |
|
j 21b |
|
endfunc |
|
|
|
func h264_avg_chroma_mc2_rvv, zve32x |
|
li t6, 2 |
|
j 21b |
|
endfunc |
|
|