[FFmpeg-devel] [PATCH 2/2] swscale/aarch64: add vscale specializations
Martin Storsjö
martin at martin.st
Tue Jun 21 23:23:46 EEST 2022
On Mon, 13 Jun 2022, Swinney, Jonathan wrote:
> This commit adds new code paths for vscale when filterSize is 2, 4, or
> 8. By using specialized code with unrolling to match the filterSize we
> can improve performance. This patch also corrects the behavor for
> filterSize 1 which was previously failing the checkasm test.
>
> On AWS c7g (Graviton 3, Neoverse V1) instances:
> before after
> yuv2yuvX_2_0_512_neon: 355.1 270.4
> yuv2yuvX_4_0_512_neon: 621.4 436.6
> yuv2yuvX_8_0_512_neon: 1134.4 817.6
> yuv2yuvX_16_0_512_neon: 2071.9 1845.1
>
> Signed-off-by: Jonathan Swinney <jswinney at amazon.com>
> ---
> libswscale/aarch64/output.S | 188 +++++++++++++++++++++++++++++++++--
> libswscale/aarch64/swscale.c | 12 +++
> 2 files changed, 194 insertions(+), 6 deletions(-)
>
> diff --git a/libswscale/aarch64/output.S b/libswscale/aarch64/output.S
> index af71de6050..a18e2c7541 100644
> --- a/libswscale/aarch64/output.S
> +++ b/libswscale/aarch64/output.S
> @@ -21,29 +21,48 @@
> #include "libavutil/aarch64/asm.S"
>
> function ff_yuv2planeX_8_neon, export=1
> +// x0 - const int16_t *filter,
> +// x1 - int filterSize,
> +// x2 - const int16_t **src,
> +// x3 - uint8_t *dest,
> +// x4 - int dstW,
> +// x5 - const uint8_t *dither,
> +// x6 - int offset
> +
> ld1 {v0.8B}, [x5] // load 8x8-bit dither
> + and w6, w6, #7
> cbz w6, 1f // check if offsetting present
> ext v0.8B, v0.8B, v0.8B, #3 // honor offsetting which can be 0 or 3 only
> 1: uxtl v0.8H, v0.8B // extend dither to 16-bit
> ushll v1.4S, v0.4H, #12 // extend dither to 32-bit with left shift by 12 (part 1)
> ushll2 v2.4S, v0.8H, #12 // extend dither to 32-bit with left shift by 12 (part 2)
> + cmp w1, #8 // if filterSize == 8, branch to specialized version
> + b.eq 6f
> + cmp w1, #4 // if filterSize == 4, branch to specialized version
> + b.eq 8f
> + cmp w1, #2 // if filterSize == 2, branch to specialized version
> + b.eq 10f
> +
> +// The filter size does not match of the of specialized implementations. It is either even or odd. If it is even
> +// then use the first section below.
> mov x7, #0 // i = 0
> + tbnz w1, #0, 4f // if filterSize % 2 != 0 branch to specialized version
> +// fs % 2 == 0
> 2: mov v3.16B, v1.16B // initialize accumulator part 1 with dithering value
> mov v4.16B, v2.16B // initialize accumulator part 2 with dithering value
> mov w8, w1 // tmpfilterSize = filterSize
> mov x9, x2 // srcp = src
> mov x10, x0 // filterp = filter
> 3: ldp x11, x12, [x9], #16 // get 2 pointers: src[j] and src[j+1]
> + ldr s7, [x10], #4 // read 2x16-bit coeff X and Y at filter[j] and filter[j+1]
> add x11, x11, x7, lsl #1 // &src[j ][i]
> add x12, x12, x7, lsl #1 // &src[j+1][i]
> ld1 {v5.8H}, [x11] // read 8x16-bit @ src[j ][i + {0..7}]: A,B,C,D,E,F,G,H
> ld1 {v6.8H}, [x12] // read 8x16-bit @ src[j+1][i + {0..7}]: I,J,K,L,M,N,O,P
> - ld1r {v7.8H}, [x10], #2 // read 1x16-bit coeff X at filter[j ] and duplicate across lanes
> - ld1r {v16.8H}, [x10], #2 // read 1x16-bit coeff Y at filter[j+1] and duplicate across lanes
> - smlal v3.4S, v5.4H, v7.4H // val0 += {A,B,C,D} * X
> - smlal2 v4.4S, v5.8H, v7.8H // val1 += {E,F,G,H} * X
> - smlal v3.4S, v6.4H, v16.4H // val0 += {I,J,K,L} * Y
> - smlal2 v4.4S, v6.8H, v16.8H // val1 += {M,N,O,P} * Y
> + smlal v3.4S, v5.4H, v7.H[0] // val0 += {A,B,C,D} * X
> + smlal2 v4.4S, v5.8H, v7.H[0] // val1 += {E,F,G,H} * X
> + smlal v3.4S, v6.4H, v7.H[1] // val0 += {I,J,K,L} * Y
> + smlal2 v4.4S, v6.8H, v7.H[1] // val1 += {M,N,O,P} * Y
> subs w8, w8, #2 // tmpfilterSize -= 2
> b.gt 3b // loop until filterSize consumed
>
> @@ -55,4 +74,161 @@ function ff_yuv2planeX_8_neon, export=1
> add x7, x7, #8 // i += 8
> b.gt 2b // loop until width consumed
> ret
> +
> +// If filter size is odd (most likely == 1), then use this section.
> +// fs % 2 != 0
> +4: mov v3.16B, v1.16B // initialize accumulator part 1 with dithering value
> + mov v4.16B, v2.16B // initialize accumulator part 2 with dithering value
> + mov w8, w1 // tmpfilterSize = filterSize
> + mov x9, x2 // srcp = src
> + mov x10, x0 // filterp = filter
> +5: ldr x11, [x9], #8 // get 1 pointer: src[j]
> + ldr h6, [x10], #2 // read 1 16 bit coeff X at filter[j]
> + add x11, x11, x7, lsl #1 // &src[j ][i]
> + ld1 {v5.8H}, [x11] // read 8x16-bit @ src[j ][i + {0..7}]: A,B,C,D,E,F,G,H
> + smlal v3.4S, v5.4H, v6.H[0] // val0 += {A,B,C,D} * X
> + smlal2 v4.4S, v5.8H, v6.H[0] // val1 += {E,F,G,H} * X
> + subs w8, w8, #1 // tmpfilterSize -= 2
> + b.gt 5b // loop until filterSize consumed
Please move the subs to after the ld1 in this case.
> +
> + sqshrun v3.4h, v3.4s, #16 // clip16(val0>>16)
> + sqshrun2 v3.8h, v4.4s, #16 // clip16(val1>>16)
> + uqshrn v3.8b, v3.8h, #3 // clip8(val>>19)
> + st1 {v3.8b}, [x3], #8 // write to destination
> + subs w4, w4, #8 // dstW -= 8
> + add x7, x7, #8 // i += 8
> + b.gt 4b // loop until width consumed
Here, the subs is already further away from the b.gt, that's good, but
between uqshrn and st1 probably is even better
> + ret
> +
> +6: // fs=8
> + ldp x5, x6, [x2] // load 2 pointers: src[j ] and src[j+1]
> + ldp x7, x9, [x2, #16] // load 2 pointers: src[j+2] and src[j+3]
> + ldp x10, x11, [x2, #32] // load 2 pointers: src[j+4] and src[j+5]
> + ldp x12, x13, [x2, #48] // load 2 pointers: src[j+6] and src[j+7]
> +
> + // load 8x16-bit values for filter[j], where j=0..7
> + ld1 {v6.8H}, [x0]
> +7:
> + mov v3.16B, v1.16B // initialize accumulator part 1 with dithering value
> + mov v4.16B, v2.16B // initialize accumulator part 2 with dithering value
> +
> + ld1 {v24.8H}, [x5], #16 // load 8x16-bit values for src[j + 0][i + {0..7}]
> + ld1 {v25.8H}, [x6], #16 // load 8x16-bit values for src[j + 1][i + {0..7}]
> + ld1 {v26.8H}, [x7], #16 // load 8x16-bit values for src[j + 2][i + {0..7}]
> + ld1 {v27.8H}, [x9], #16 // load 8x16-bit values for src[j + 3][i + {0..7}]
> + ld1 {v28.8H}, [x10], #16 // load 8x16-bit values for src[j + 4][i + {0..7}]
> + ld1 {v29.8H}, [x11], #16 // load 8x16-bit values for src[j + 5][i + {0..7}]
> + ld1 {v30.8H}, [x12], #16 // load 8x16-bit values for src[j + 6][i + {0..7}]
> + ld1 {v31.8H}, [x13], #16 // load 8x16-bit values for src[j + 7][i + {0..7}]
> +
> + smlal v3.4S, v24.4H, v6.H[0] // val0 += src[0][i + {0..3}] * filter[0]
> + smlal2 v4.4S, v24.8H, v6.H[0] // val1 += src[0][i + {4..7}] * filter[0]
> + smlal v3.4S, v25.4H, v6.H[1] // val0 += src[1][i + {0..3}] * filter[1]
> + smlal2 v4.4S, v25.8H, v6.H[1] // val1 += src[1][i + {4..7}] * filter[1]
> + smlal v3.4S, v26.4H, v6.H[2] // val0 += src[2][i + {0..3}] * filter[2]
> + smlal2 v4.4S, v26.8H, v6.H[2] // val1 += src[2][i + {4..7}] * filter[2]
> + smlal v3.4S, v27.4H, v6.H[3] // val0 += src[3][i + {0..3}] * filter[3]
> + smlal2 v4.4S, v27.8H, v6.H[3] // val1 += src[3][i + {4..7}] * filter[3]
> + smlal v3.4S, v28.4H, v6.H[4] // val0 += src[4][i + {0..3}] * filter[4]
> + smlal2 v4.4S, v28.8H, v6.H[4] // val1 += src[4][i + {4..7}] * filter[4]
> + smlal v3.4S, v29.4H, v6.H[5] // val0 += src[5][i + {0..3}] * filter[5]
> + smlal2 v4.4S, v29.8H, v6.H[5] // val1 += src[5][i + {4..7}] * filter[5]
> + smlal v3.4S, v30.4H, v6.H[6] // val0 += src[6][i + {0..3}] * filter[6]
> + smlal2 v4.4S, v30.8H, v6.H[6] // val1 += src[6][i + {4..7}] * filter[6]
> + smlal v3.4S, v31.4H, v6.H[7] // val0 += src[7][i + {0..3}] * filter[7]
> + smlal2 v4.4S, v31.8H, v6.H[7] // val1 += src[7][i + {4..7}] * filter[7]
> +
> + sqshrun v3.4h, v3.4s, #16 // clip16(val0>>16)
> + sqshrun2 v3.8h, v4.4s, #16 // clip16(val1>>16)
> + uqshrn v3.8b, v3.8h, #3 // clip8(val>>19)
> + subs w4, w4, #8 // dstW -= 8
> + st1 {v3.8b}, [x3], #8 // write to destination
> + b.gt 7b // loop until width consumed
> + ret
> +
> +8: // fs=4
> + ldp x5, x6, [x2] // load 2 pointers: src[j ] and src[j+1]
> + ldp x7, x9, [x2, #16] // load 2 pointers: src[j+2] and src[j+3]
> +
> + // load 4x16-bit values for filter[j], where j=0..3 and replicated across lanes
> + ld1 {v6.4H}, [x0]
> +9:
> + mov v3.16B, v1.16B // initialize accumulator part 1 with dithering value
> + mov v4.16B, v2.16B // initialize accumulator part 2 with dithering value
> +
> + ld1 {v24.8H}, [x5], #16 // load 8x16-bit values for src[j + 0][i + {0..7}]
> + ld1 {v25.8H}, [x6], #16 // load 8x16-bit values for src[j + 1][i + {0..7}]
> + ld1 {v26.8H}, [x7], #16 // load 8x16-bit values for src[j + 2][i + {0..7}]
> + ld1 {v27.8H}, [x9], #16 // load 8x16-bit values for src[j + 3][i + {0..7}]
> +
> + smlal v3.4S, v24.4H, v6.H[0] // val0 += src[0][i + {0..3}] * filter[0]
> + smlal2 v4.4S, v24.8H, v6.H[0] // val1 += src[0][i + {4..7}] * filter[0]
> + smlal v3.4S, v25.4H, v6.H[1] // val0 += src[1][i + {0..3}] * filter[1]
> + smlal2 v4.4S, v25.8H, v6.H[1] // val1 += src[1][i + {4..7}] * filter[1]
> + smlal v3.4S, v26.4H, v6.H[2] // val0 += src[2][i + {0..3}] * filter[2]
> + smlal2 v4.4S, v26.8H, v6.H[2] // val1 += src[2][i + {4..7}] * filter[2]
> + smlal v3.4S, v27.4H, v6.H[3] // val0 += src[3][i + {0..3}] * filter[3]
> + smlal2 v4.4S, v27.8H, v6.H[3] // val1 += src[3][i + {4..7}] * filter[3]
> +
> + sqshrun v3.4h, v3.4s, #16 // clip16(val0>>16)
> + sqshrun2 v3.8h, v4.4s, #16 // clip16(val1>>16)
> + uqshrn v3.8b, v3.8h, #3 // clip8(val>>19)
> + st1 {v3.8b}, [x3], #8 // write to destination
> + subs w4, w4, #8 // dstW -= 8
> + b.gt 9b // loop until width consumed
Move subs up to above st1
Other than that, I think this seems fine.
// Martin
More information about the ffmpeg-devel
mailing list