[FFmpeg-devel] [PATCH 3/4] x86/vvc_alf: avoid overwriting for non-16 aligned widths
Nuo Mi
nuomi2021 at gmail.com
Sat Jun 22 07:21:13 EEST 2024
Previously, the code allowed overwriting on 16-aligned blocks, which was suitable when there were
no picture's virtual boundaries because both CTU sizes and strides were 16-aligned. However, with
picture's virtual boundaries, each CTU is divided into four ALF blocks, leading to potential issues
with overwriting later CTUs.
In cases involving picture virtual boundaries, each ALF block is 8-pixel aligned.
For luma, we consistently ensure an 8-aligned width. For chroma in 4:2:0 format,
we need to account for a 4-aligned width.
---
libavcodec/x86/vvc/vvc_alf.asm | 85 ++++++++++++++++++++++++++++++----
1 file changed, 75 insertions(+), 10 deletions(-)
diff --git a/libavcodec/x86/vvc/vvc_alf.asm b/libavcodec/x86/vvc/vvc_alf.asm
index b35dd9b0e9..f69a69f05f 100644
--- a/libavcodec/x86/vvc/vvc_alf.asm
+++ b/libavcodec/x86/vvc/vvc_alf.asm
@@ -324,18 +324,69 @@ SECTION .text
%endif
%endmacro
-; STORE_PIXELS(dst, src)
-%macro STORE_PIXELS 2
+; STORE_PIXELS_W16(dst, src)
+%macro STORE_PIXELS_W16 2
%if ps == 2
- movu %1, m%2
+ movu [%1], m%2
%else
+ movu [%1], xm%2
+ %endif
+%endmacro
+
+%macro STORE_PIXELS_W8 2
+ %if ps == 2
+ movu [%1], xm%2
+ %else
+ movq [%1], xm%2
+ %endif
+%endmacro
+
+; STORE_PIXELS_W4(dst, src, offset)
+%macro STORE_PIXELS_W4 3
+ %if ps == 2
+ movq [%1 + %3 * ps], xm%2
+ %else
+ movd [%1 + %3], xm%2
+ %endif
+%endmacro
+
+%macro STORE_PIXELS_W8LE 3
+ cmp %3, 8
+ jl .w4
+ STORE_PIXELS_W8 %1, %2
+ cmp %3, 12
+ %if ps == 2
+ vpermq m%2, m%2, q0302
+ %else
+ vpermq m%2, m%2, q0101
+ %endif
+ jl .end
+ STORE_PIXELS_W4 %1, %2, 8
+ jmp .end
+.w4:
+ STORE_PIXELS_W4 %1, %2, 0
+.end:
+%endmacro
+
+; STORE_PIXELS(dst, src, width)
+%macro STORE_PIXELS 3
+ %if ps == 1
packuswb m%2, m%2
vpermq m%2, m%2, 0x8
- movu %1, xm%2
+ %endif
+
+ %ifidn %3, 16
+ STORE_PIXELS_W16 %1, %2
+ %else
+ %if LUMA
+ STORE_PIXELS_W8 %1, %2
+ %else
+ STORE_PIXELS_W8LE %1, %2, %3
+ %endif
%endif
%endmacro
-%macro FILTER_16x4 0
+%macro FILTER_16x4 1
%if LUMA
push clipq
push strideq
@@ -362,7 +413,7 @@ SECTION .text
; clip to pixel
CLIPW m0, m14, m15
- STORE_PIXELS [dstq], 0
+ STORE_PIXELS dstq, 0, %1
lea srcq, [srcq + src_strideq]
lea dstq, [dstq + dst_strideq]
@@ -399,7 +450,7 @@ SECTION .text
; const uint8_t *src, ptrdiff_t src_stride, const ptrdiff_t width, cosnt ptr_diff_t height,
; const int16_t *filter, const int16_t *clip, ptrdiff_t stride, ptrdiff_t vb_pos, ptrdiff_t pixel_max);
; ******************************
-cglobal vvc_alf_filter_%2_%1bpc, 11, 15, 16, 0-0x28, dst, dst_stride, src, src_stride, width, height, filter, clip, stride, vb_pos, pixel_max, \
+cglobal vvc_alf_filter_%2_%1bpc, 11, 15, 16, 0-0x30, dst, dst_stride, src, src_stride, width, height, filter, clip, stride, vb_pos, pixel_max, \
offset, x, s5, s6
%define ps (%1 / 8) ; pixel size
movd xm15, pixel_maxd
@@ -409,18 +460,32 @@ cglobal vvc_alf_filter_%2_%1bpc, 11, 15, 16, 0-0x28, dst, dst_stride, src, src_s
.loop:
push srcq
push dstq
+ push widthq
xor xq, xq
.loop_w:
+ cmp widthq, 16
+ jl .loop_w_end
+
LOAD_PARAMS
- FILTER_16x4
+ FILTER_16x4 16
add srcq, 16 * ps
add dstq, 16 * ps
add xq, 16
- cmp xq, widthq
- jl .loop_w
+ sub widthq, 16
+ jmp .loop_w
+
+.loop_w_end:
+ cmp widthq, 0
+ je .w_end
+
+ LOAD_PARAMS
+ FILTER_16x4 widthq
+
+.w_end:
+ pop widthq
pop dstq
pop srcq
lea srcq, [srcq + 4 * src_strideq]
--
2.34.1
More information about the ffmpeg-devel
mailing list