[FFmpeg-cvslog] lavc/me_cmp: R-V V nsse
sunyuechi
git at videolan.org
Tue Feb 27 20:39:46 EET 2024
ffmpeg | branch: master | sunyuechi <sunyuechi at iscas.ac.cn> | Tue Feb 6 22:51:47 2024 +0800| [a7ad76fbbfae29eaf8c9372b34051edf4f835fd9] | committer: Rémi Denis-Courmont
lavc/me_cmp: R-V V nsse
C908:
nsse_0_c: 1990.0
nsse_0_rvv_i32: 572.0
nsse_1_c: 910.0
nsse_1_rvv_i32: 456.0
Signed-off-by: Rémi Denis-Courmont <remi at remlab.net>
> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=a7ad76fbbfae29eaf8c9372b34051edf4f835fd9
---
libavcodec/riscv/me_cmp_init.c | 25 +++++++++
libavcodec/riscv/me_cmp_rvv.S | 118 +++++++++++++++++++++++++++++++++++++++++
2 files changed, 143 insertions(+)
diff --git a/libavcodec/riscv/me_cmp_init.c b/libavcodec/riscv/me_cmp_init.c
index 24e78e3eeb..858e2ccdb8 100644
--- a/libavcodec/riscv/me_cmp_init.c
+++ b/libavcodec/riscv/me_cmp_init.c
@@ -54,6 +54,28 @@ int ff_vsad16_rvv(MpegEncContext *c, const uint8_t *s1, const uint8_t *s2, ptrdi
int ff_vsad8_rvv(MpegEncContext *c, const uint8_t *s1, const uint8_t *s2, ptrdiff_t stride, int h);
int ff_vsad_intra16_rvv(MpegEncContext *c, const uint8_t *s, const uint8_t *dummy, ptrdiff_t stride, int h);
int ff_vsad_intra8_rvv(MpegEncContext *c, const uint8_t *s, const uint8_t *dummy, ptrdiff_t stride, int h);
+int ff_nsse16_rvv(int multiplier, const uint8_t *s1, const uint8_t *s2,
+ ptrdiff_t stride, int h);
+int ff_nsse8_rvv(int multiplier, const uint8_t *s1, const uint8_t *s2,
+ ptrdiff_t stride, int h);
+
+static int nsse16_rvv_wrapper(MpegEncContext *c, const uint8_t *s1, const uint8_t *s2,
+ ptrdiff_t stride, int h)
+{
+ if (c)
+ return ff_nsse16_rvv(c->avctx->nsse_weight, s1, s2, stride, h);
+ else
+ return ff_nsse16_rvv(8, s1, s2, stride, h);
+}
+
+static int nsse8_rvv_wrapper(MpegEncContext *c, const uint8_t *s1, const uint8_t *s2,
+ ptrdiff_t stride, int h)
+{
+ if (c)
+ return ff_nsse8_rvv(c->avctx->nsse_weight, s1, s2, stride, h);
+ else
+ return ff_nsse8_rvv(8, s1, s2, stride, h);
+}
av_cold void ff_me_cmp_init_riscv(MECmpContext *c, AVCodecContext *avctx)
{
@@ -82,6 +104,9 @@ av_cold void ff_me_cmp_init_riscv(MECmpContext *c, AVCodecContext *avctx)
c->vsad[1] = ff_vsad8_rvv;
c->vsad[4] = ff_vsad_intra16_rvv;
c->vsad[5] = ff_vsad_intra8_rvv;
+
+ c->nsse[0] = nsse16_rvv_wrapper;
+ c->nsse[1] = nsse8_rvv_wrapper;
}
#endif
}
diff --git a/libavcodec/riscv/me_cmp_rvv.S b/libavcodec/riscv/me_cmp_rvv.S
index f32ae6b259..c9ae5bb6fc 100644
--- a/libavcodec/riscv/me_cmp_rvv.S
+++ b/libavcodec/riscv/me_cmp_rvv.S
@@ -407,3 +407,121 @@ endfunc
func ff_vsad_intra8_rvv, zve32x
vsad_vsse_intra8 abs
endfunc
+
+func ff_nsse16_rvv, zve32x
+ .macro squarediff16
+ vsetivli zero, 16, e8, m1, tu, ma
+ vle8.v v4, (a1)
+ vle8.v v12, (a2)
+ vwsubu.vv v16, v4, v12
+ vsetvli zero, zero, e16, m2, tu, ma
+ vwmacc.vv v24, v16, v16
+ .endm
+
+ .macro gradiff16 srcx srcv
+ vsetivli zero, 16, e8, m1, tu, ma
+ vle8.v v8, (\srcx)
+ vslide1down.vx v0, \srcv, t5
+ vslide1down.vx v16, v8, t5
+ vwsubu.vv v20, \srcv, v0
+ vwsubu.wv v0, v20, v8
+ vwaddu.wv v20, v0, v16
+ vsetivli zero, 15, e16, m2, tu, ma
+ vneg.v v0, v20
+ vmax.vv v0, v20, v0
+ .endm
+
+ csrwi vxrm, 0
+ vsetivli t0, 16, e32, m4, ta, ma
+ addi a4, a4, -1
+ li t5, 1
+ vmv.v.x v24, zero
+ vmv.v.x v28, zero
+1:
+ add t1, a1, a3
+ add t2, a2, a3
+ addi a4, a4, -1
+ squarediff16
+ gradiff16 t1, v4
+ vwaddu.wv v28, v28, v0
+ gradiff16 t2, v12
+ vwsubu.wv v28, v28, v0
+ add a1, a1, a3
+ add a2, a2, a3
+ bnez a4, 1b
+
+ squarediff16
+ vsetivli zero, 16, e32, m4, tu, ma
+ vmv.s.x v0, zero
+ vmv.s.x v4, zero
+ vredsum.vs v0, v24, v0
+ vredsum.vs v4, v28, v4
+ vmv.x.s t1, v0
+ vmv.x.s t2, v4
+ srai t3, t2, 31
+ xor t2, t3, t2
+ sub t2, t2, t3
+ mul t2, t2, a0
+ add a0, t2, t1
+
+ ret
+endfunc
+
+func ff_nsse8_rvv, zve32x
+ .macro squarediff8
+ vsetivli zero, 8, e8, mf2, tu, ma
+ vle8.v v4, (a1)
+ vle8.v v12, (a2)
+ vwsubu.vv v16, v4, v12
+ vsetvli zero, zero, e16, m1, tu, ma
+ vwmacc.vv v24, v16, v16
+ .endm
+
+ .macro gradiff8 srcx srcv
+ vsetivli zero, 8, e8, mf2, tu, ma
+ vle8.v v8, (\srcx)
+ vslide1down.vx v0, \srcv, t5
+ vslide1down.vx v16, v8, t5
+ vwsubu.vv v20, \srcv, v0
+ vwsubu.wv v0, v20, v8
+ vwaddu.wv v20, v0, v16
+ vsetivli zero, 7, e16, m1, tu, ma
+ vneg.v v0, v20
+ vmax.vv v0, v20, v0
+ .endm
+
+ csrwi vxrm, 0
+ vsetivli t0, 8, e32, m2, ta, ma
+ addi a4, a4, -1
+ li t5, 1
+ vmv.v.x v24, zero
+ vmv.v.x v28, zero
+1:
+ add t1, a1, a3
+ add t2, a2, a3
+ addi a4, a4, -1
+ squarediff8
+ gradiff8 t1, v4
+ vwaddu.wv v28, v28, v0
+ gradiff8 t2, v12
+ vwsubu.wv v28, v28, v0
+ add a1, a1, a3
+ add a2, a2, a3
+ bnez a4, 1b
+
+ squarediff8
+ vsetivli zero, 8, e32, m2, tu, ma
+ vmv.s.x v0, zero
+ vmv.s.x v4, zero
+ vredsum.vs v0, v24, v0
+ vredsum.vs v4, v28, v4
+ vmv.x.s t1, v0
+ vmv.x.s t2, v4
+ srai t3, t2, 31
+ xor t2, t3, t2
+ sub t2, t2, t3
+ mul t2, t2, a0
+ add a0, t2, t1
+
+ ret
+endfunc
More information about the ffmpeg-cvslog
mailing list