31 #define PIXEL_STRIDE 16
33 #define randomize_buffers(src, dst, stride, coef) \
36 for (y = 0; y < 4; y++) { \
37 AV_WN32A((src) + y * (stride), rnd()); \
38 AV_WN32A((dst) + y * (stride), rnd()); \
39 for (x = 0; x < 4; x++) \
40 (coef)[y * 4 + x] = (src)[y * (stride) + x] - \
41 (dst)[y * (stride) + x]; \
48 for (i = 0; i < 4; i++) {
49 const int a1 = (coef[i*4 + 0] + coef[i*4 + 3]) * 8;
50 const int b1 = (coef[i*4 + 1] + coef[i*4 + 2]) * 8;
51 const int c1 = (coef[i*4 + 1] - coef[i*4 + 2]) * 8;
52 const int d1 = (coef[i*4 + 0] - coef[i*4 + 3]) * 8;
53 coef[i*4 + 0] = a1 + b1;
54 coef[i*4 + 1] = (c1 * 2217 + d1 * 5352 + 14500) >> 12;
55 coef[i*4 + 2] = a1 - b1;
56 coef[i*4 + 3] = (d1 * 2217 - c1 * 5352 + 7500) >> 12;
58 for (i = 0; i < 4; i++) {
59 const int a1 = coef[i + 0*4] + coef[i + 3*4];
60 const int b1 = coef[i + 1*4] + coef[i + 2*4];
61 const int c1 = coef[i + 1*4] - coef[i + 2*4];
62 const int d1 = coef[i + 0*4] - coef[i + 3*4];
63 coef[i + 0*4] = (a1 + b1 + 7) >> 4;
64 coef[i + 1*4] = ((c1 * 2217 + d1 * 5352 + 12000) >> 16) + !!d1;
65 coef[i + 2*4] = (a1 - b1 + 7) >> 4;
66 coef[i + 3*4] = (d1 * 2217 - c1 * 5352 + 51000) >> 16;
73 for (i = 0; i < 4; i++) {
74 int a1 = coef[0 * 4 + i];
75 int b1 = coef[1 * 4 + i];
76 int c1 = coef[2 * 4 + i];
77 int d1 = coef[3 * 4 + i];
91 for (i = 0; i < 4; i++) {
92 int a1 = coef[i * 4 + 0];
93 int b1 = coef[i * 4 + 1];
94 int c1 = coef[i * 4 + 2];
95 int d1 = coef[i * 4 + 3];
104 coef[i * 4 + 0] = a1 * 2;
105 coef[i * 4 + 1] = c1 * 2;
106 coef[i * 4 + 2] = d1 * 2;
107 coef[i * 4 + 3] = b1 * 2;
129 for (dc = 0; dc <= 1; dc++) {
134 memset(subcoef0, 0, 4 * 4 *
sizeof(int16_t));
135 subcoef0[0] = coef[0];
137 memcpy(subcoef0, coef, 4 * 4 *
sizeof(int16_t));
139 memcpy(dst0, dst, 4 * 4);
140 memcpy(dst1, dst, 4 * 4);
141 memcpy(subcoef1, subcoef0, 4 * 4 *
sizeof(int16_t));
147 if (memcmp(dst0, dst1, 4 * 4) ||
148 memcmp(subcoef0, subcoef1, 4 * 4 *
sizeof(int16_t)))
171 for (chroma = 0; chroma <= 1; chroma++) {
173 if (
check_func(idct4dc,
"vp8_idct_dc_add4%s", chroma ?
"uv" :
"y")) {
174 ptrdiff_t stride = chroma ? 8 : 16;
175 int w = chroma ? 2 : 4;
176 for (i = 0; i < 4; i++) {
177 int blockx = 4 * (i % w);
178 int blocky = 4 * (i / w);
179 randomize_buffers(
src + stride * blocky + blockx, dst + stride * blocky + blockx, stride, coef[i]);
181 memset(&coef[i][1], 0, 15 *
sizeof(int16_t));
184 memcpy(dst0, dst, 4 * 4 * 4);
185 memcpy(dst1, dst, 4 * 4 * 4);
186 memcpy(subcoef0, coef, 4 * 4 * 4 *
sizeof(int16_t));
187 memcpy(subcoef1, coef, 4 * 4 * 4 *
sizeof(int16_t));
190 if (memcmp(dst0, dst1, 4 * 4 * 4) ||
191 memcmp(subcoef0, subcoef1, 4 * 4 * 4 *
sizeof(int16_t)))
204 int16_t
block[4][4][16];
214 for (blocky = 0; blocky < 4; blocky++) {
215 for (blockx = 0; blockx < 4; blockx++) {
219 dct4x4(block[blocky][blockx]);
220 dc[blocky * 4 + blockx] = block[blocky][blockx][0];
221 block[blocky][blockx][0] =
rnd();
226 for (dc_only = 0; dc_only <= 1; dc_only++) {
231 memset(dc0, 0, 16 *
sizeof(int16_t));
234 memcpy(dc0, dc, 16 *
sizeof(int16_t));
236 memcpy(dc1, dc0, 16 *
sizeof(int16_t));
237 memcpy(block0, block, 4 * 4 * 16 *
sizeof(int16_t));
238 memcpy(
block1, block, 4 * 4 * 16 *
sizeof(int16_t));
241 if (memcmp(block0,
block1, 4 * 4 * 16 *
sizeof(int16_t)) ||
242 memcmp(dc0, dc1, 16 *
sizeof(int16_t)))
249 #define SRC_BUF_STRIDE 32
250 #define SRC_BUF_SIZE (((size << (size < 16)) + 5) * SRC_BUF_STRIDE)
254 #define src (buf + 2 * SRC_BUF_STRIDE + 2 + 1)
256 #undef randomize_buffers
257 #define randomize_buffers() \
260 for (k = 0; k < SRC_BUF_SIZE; k += 4) { \
261 AV_WN32A(buf + k, rnd()); \
276 for (type = 0; type < 2; type++) {
278 for (k = 1; k < 8; k++) {
280 int size = 16 >> hsize;
282 for (dy = 0; dy < 3; dy++) {
283 for (dx = 0; dx < 3; dx++) {
287 static const char *dx_names[] = {
"",
"h4",
"h6" };
288 static const char *dy_names[] = {
"",
"v4",
"v6" };
289 snprintf(str,
sizeof(str),
"epel%d_%s%s", size, dx_names[dx], dy_names[dy]);
291 snprintf(str,
sizeof(str),
"bilin%d_%s%s", size, dx ?
"h" :
"", dy ?
"v" :
"");
294 snprintf(str,
sizeof(str),
"pixels%d", size);
300 mx = dx == 2 ? 2 + 2 * (
rnd() % 3) : dx == 1 ? 1 + 2 * (
rnd() % 4) : 0;
301 my = dy == 2 ? 2 + 2 * (
rnd() % 3) : dy == 1 ? 1 + 2 * (
rnd() % 4) : 0;
303 mx = dx ? 1 + (
rnd() % 7) : 0;
304 my = dy ? 1 + (
rnd() % 7) : 0;
307 for (i = -2; i <= 3; i++) {
308 int val = (i == -1 || i == 2) ? 0 : 0xff;
316 if (memcmp(dst0, dst1, size * height))
326 #undef randomize_buffers
328 #define setpx(a, b, c) buf[(a) + (b) * jstride] = av_clip_uint8(c)
330 #define setdx(a, b, c, d) setpx(a, b, c - (d) + (rnd() % ((d) * 2 + 1)))
332 #define setdx2(a, b, o, c, d, e) setpx(a, b, o = c + ((d) + (rnd() % (e))) * (c >= 128 ? -1 : 1))
335 int dir,
int flim_E,
int flim_I,
339 uint32_t
mask = 0xff;
340 int off = dir ? lineoff : lineoff * str;
341 int istride = dir ? 1 : str;
342 int jstride = dir ? str : 1;
344 for (i = 0; i < 8; i += 2) {
349 int idx = off + i * istride, p2, p1, p0,
q0,
q1, q2;
351 if (i == 0 && force_hev >= 0 || force_hev > 0)
352 setdx2(idx, 1, q1, q0, hev_thresh + 1, flim_I - hev_thresh - 1);
354 setdx(idx, 1, q1 = q0, hev_thresh);
355 setdx(idx, 2, q2 = q1, flim_I);
356 setdx(idx, 3, q2, flim_I);
357 setdx(idx, -1, p0 = q0, flim_E >> 2);
358 if (i == 2 && force_hev >= 0 || force_hev > 0)
359 setdx2(idx, -2, p1, p0, hev_thresh + 1, flim_I - hev_thresh - 1);
361 setdx(idx, -2, p1 = p0, hev_thresh);
362 setdx(idx, -3, p2 = p1, flim_I);
363 setdx(idx, -4, p2, flim_I);
371 for (y = 0; y <
h; y++)
372 for (x = 0; x < w; x++)
373 buf[y * stride + x] =
rnd() & 0xff;
376 #define randomize_buffers(buf, lineoff, str, force_hev) \
377 randomize_loopfilter_buffers(lineoff, str, dir, flim_E, flim_I, hev_thresh, buf, force_hev)
384 int dir, edge, force_hev;
385 int flim_E = 20, flim_I = 10, hev_thresh = 7;
390 for (dir = 0; dir < 2; dir++) {
391 int midoff = dir ? 4 * 16 : 4;
392 int midoff_aligned = dir ? 4 * 16 : 16;
393 uint8_t *buf0 = base0 + midoff_aligned;
394 uint8_t *buf1 = base1 + midoff_aligned;
395 for (edge = 0; edge < 2; edge++) {
397 switch (dir << 1 | edge) {
403 if (
check_func(
func,
"vp8_loop_filter16y%s_%s", edge ?
"_inner" :
"", dir ?
"v" :
"h")) {
404 for (force_hev = -1; force_hev <= 1; force_hev++) {
408 memcpy(buf1 - midoff, buf0 - midoff, 16 * 16);
409 call_ref(buf0, 16, flim_E, flim_I, hev_thresh);
410 call_new(buf1, 16, flim_E, flim_I, hev_thresh);
411 if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16))
417 bench_new(buf0, 16, flim_E, flim_I, hev_thresh);
430 int dir, edge, force_hev;
431 int flim_E = 20, flim_I = 10, hev_thresh = 7;
436 for (dir = 0; dir < 2; dir++) {
437 int midoff = dir ? 4 * 16 : 4;
438 int midoff_aligned = dir ? 4 * 16 : 16;
439 uint8_t *buf0u = base0u + midoff_aligned;
440 uint8_t *buf0v = base0v + midoff_aligned;
441 uint8_t *buf1u = base1u + midoff_aligned;
442 uint8_t *buf1v = base1v + midoff_aligned;
443 for (edge = 0; edge < 2; edge++) {
445 switch (dir << 1 | edge) {
451 if (
check_func(
func,
"vp8_loop_filter8uv%s_%s", edge ?
"_inner" :
"", dir ?
"v" :
"h")) {
452 for (force_hev = -1; force_hev <= 1; force_hev++) {
457 memcpy(buf1u - midoff, buf0u - midoff, 16 * 16);
458 memcpy(buf1v - midoff, buf0v - midoff, 16 * 16);
460 call_ref(buf0u, buf0v, 16, flim_E, flim_I, hev_thresh);
461 call_new(buf1u, buf1v, 16, flim_E, flim_I, hev_thresh);
462 if (memcmp(buf0u - midoff, buf1u - midoff, 16 * 16) ||
463 memcmp(buf0v - midoff, buf1v - midoff, 16 * 16))
470 bench_new(buf0u, buf0v, 16, flim_E, flim_I, hev_thresh);
482 int flim_E = 20, flim_I = 30, hev_thresh = 0;
487 for (dir = 0; dir < 2; dir++) {
488 int midoff = dir ? 4 * 16 : 4;
489 int midoff_aligned = dir ? 4 * 16 : 16;
490 uint8_t *buf0 = base0 + midoff_aligned;
491 uint8_t *buf1 = base1 + midoff_aligned;
497 memcpy(buf1 - midoff, buf0 - midoff, 16 * 16);
500 if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16))
const char const char void * val
static void check_mc(void)
static void wht4x4(int16_t *coef)
static void idct(int16_t block[64])
static void check_loopfilter_8uv(void)
void(* vp8_v_loop_filter16y)(uint8_t *dst, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
static void check_idct(void)
vp8_mc_func put_vp8_bilinear_pixels_tab[3][3][3]
av_cold void ff_vp78dsp_init(VP8DSPContext *dsp)
static const uint8_t q1[256]
vp8_mc_func put_vp8_epel_pixels_tab[3][3][3]
first dimension: 4-log2(width) second dimension: 0 if no vertical interpolation is needed; 1 4-tap ve...
VP8 compatible video decoder.
void(* vp8_h_loop_filter8uv_inner)(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
static void check_luma_dc_wht(void)
static void check_idct_dc4(void)
#define randomize_buffers(src, dst, stride, coef)
static void check_loopfilter_simple(void)
void(* vp8_luma_dc_wht_dc)(int16_t block[4][4][16], int16_t dc[16])
static const uint16_t mask[17]
void(* vp8_h_loop_filter8uv)(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
void(* vp8_v_loop_filter8uv)(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
static const uint8_t q0[256]
static void check_loopfilter_16y(void)
typedef void(APIENTRY *FF_PFNGLACTIVETEXTUREPROC)(GLenum texture)
void(* vp8_idct_dc_add4y)(uint8_t *dst, int16_t block[4][16], ptrdiff_t stride)
void(* vp8_mc_func)(uint8_t *dst, ptrdiff_t dstStride, uint8_t *src, ptrdiff_t srcStride, int h, int x, int y)
void checkasm_check_vp8dsp(void)
#define declare_func_emms(cpu_flags, ret,...)
void(* vp8_v_loop_filter8uv_inner)(uint8_t *dstU, uint8_t *dstV, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
void(* vp8_h_loop_filter_simple)(uint8_t *dst, ptrdiff_t stride, int flim)
static void dct4x4(int16_t *coef)
void(* vp8_idct_add)(uint8_t *dst, int16_t block[16], ptrdiff_t stride)
Libavcodec external API header.
static void randomize_loopfilter_buffers(int lineoff, int str, int dir, int flim_E, int flim_I, int hev_thresh, uint8_t *buf, int force_hev)
#define AV_CPU_FLAG_MMX
standard MMX
#define check_func(func,...)
void(* vp8_idct_dc_add4uv)(uint8_t *dst, int16_t block[4][16], ptrdiff_t stride)
int(* func)(AVBPrint *dst, const char *in, const char *arg)
void(* vp8_h_loop_filter16y_inner)(uint8_t *dst, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
void(* vp8_luma_dc_wht)(int16_t block[4][4][16], int16_t dc[16])
GLint GLenum GLboolean GLsizei stride
void(* vp8_v_loop_filter_simple)(uint8_t *dst, ptrdiff_t stride, int flim)
common internal and external API header
#define setdx(a, b, c, d)
#define setdx2(a, b, o, c, d, e)
void(* vp8_h_loop_filter16y)(uint8_t *dst, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
static void fill_loopfilter_buffers(uint8_t *buf, ptrdiff_t stride, int w, int h)
void(* vp8_idct_dc_add)(uint8_t *dst, int16_t block[16], ptrdiff_t stride)
static const struct twinvq_data tab
static int16_t block1[64]
void(* vp8_v_loop_filter16y_inner)(uint8_t *dst, ptrdiff_t stride, int flim_E, int flim_I, int hev_thresh)
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(constuint8_t *) pi-0x80)*(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(constuint8_t *) pi-0x80)*(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(constint16_t *) pi >>8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t,*(constint16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t,*(constint16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(constint32_t *) pi >>24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t,*(constint32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t,*(constint32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(constfloat *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(constfloat *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(constfloat *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(constdouble *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(constdouble *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(constdouble *) pi *(1U<< 31))))#defineSET_CONV_FUNC_GROUP(ofmt, ifmt) staticvoidset_generic_function(AudioConvert *ac){}voidff_audio_convert_free(AudioConvert **ac){if(!*ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);}AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enumAVSampleFormatout_fmt, enumAVSampleFormatin_fmt, intchannels, intsample_rate, intapply_map){AudioConvert *ac;intin_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) returnNULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method!=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt)>2){ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc){av_free(ac);returnNULL;}returnac;}in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar){ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar?ac->channels:1;}elseif(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;elseac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);returnac;}intff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in){intuse_generic=1;intlen=in->nb_samples;intp;if(ac->dc){av_log(ac->avr, AV_LOG_TRACE,"%dsamples-audio_convert:%sto%s(dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));returnff_convert_dither(ac-> dc
#define LOCAL_ALIGNED_16(t, v,...)
void ff_vp8dsp_init(VP8DSPContext *c)