27 #ifndef AVCODEC_H264DSP_H 28 #define AVCODEC_H264DSP_H 49 int alpha,
int beta, int8_t *tc0);
51 int alpha,
int beta, int8_t *tc0);
53 int alpha,
int beta, int8_t *tc0);
62 int alpha,
int beta, int8_t *tc0);
64 int alpha,
int beta, int8_t *tc0);
76 int8_t
ref[2][40], int16_t
mv[2][40][2],
77 int bidir,
int edges,
int step,
78 int mask_mv0,
int mask_mv1,
int field);
84 int16_t *block ,
int stride);
86 int16_t *block ,
int stride);
88 int16_t *block ,
int stride);
91 int16_t *block ,
int stride,
94 int16_t *block ,
int stride,
97 int16_t *block ,
int stride,
103 int16_t *
input ,
int qmul);
121 const int chroma_format_idc);
123 const int chroma_format_idc);
125 const int chroma_format_idc);
127 const int chroma_format_idc);
129 const int chroma_format_idc);
131 const int chroma_format_idc);
void(* h264_biweight_func)(uint8_t *dst, uint8_t *src, ptrdiff_t stride, int height, int log2_denom, int weightd, int weights, int offset)
void(* h264_chroma_dc_dequant_idct)(int16_t *block, int qmul)
void ff_h264dsp_init(H264DSPContext *c, const int bit_depth, const int chroma_format_idc)
void(* h264_h_loop_filter_luma_intra)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta)
The exact code depends on how similar the blocks are and how related they are to the block
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
Undefined Behavior In the C some operations are like signed integer dereferencing freed accessing outside allocated Undefined Behavior must not occur in a C it is not safe even if the output of undefined operations is unused The unsafety may seem nit picking but Optimizing compilers have in fact optimized code on the assumption that no undefined Behavior occurs Optimizing code based on wrong assumptions can and has in some cases lead to effects beyond the output of computations The signed integer overflow problem in speed critical code Code which is highly optimized and works with signed integers sometimes has the problem that often the output of the computation does not c
void ff_h264dsp_init_arm(H264DSPContext *c, const int bit_depth, const int chroma_format_idc)
filter_frame For filters that do not use the this method is called when a frame is pushed to the filter s input It can be called at any time except in a reentrant way If the input frame is enough to produce output
void ff_h264dsp_init_mips(H264DSPContext *c, const int bit_depth, const int chroma_format_idc)
h264_weight_func weight_h264_pixels_tab[4]
void(* h264_idct_add16intra)(uint8_t *dst, const int *blockoffset, int16_t *block, int stride, const uint8_t nnzc[15 *8])
void(* h264_h_loop_filter_chroma_mbaff)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta, int8_t *tc0)
void(* h264_idct_add)(uint8_t *dst, int16_t *block, int stride)
void(* h264_h_loop_filter_chroma_intra)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta)
int(* startcode_find_candidate)(const uint8_t *buf, int size)
Search buf from the start for up to size bytes.
void(* h264_v_loop_filter_chroma_intra)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta)
void(* h264_idct8_dc_add)(uint8_t *dst, int16_t *block, int stride)
void(* h264_h_loop_filter_luma)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta, int8_t *tc0)
void ff_h264dsp_init_aarch64(H264DSPContext *c, const int bit_depth, const int chroma_format_idc)
static const int weights[]
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this field
void(* h264_v_loop_filter_luma_intra)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta)
void(* h264_h_loop_filter_chroma)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta, int8_t *tc0)
void(* h264_add_pixels8_clear)(uint8_t *dst, int16_t *block, int stride)
void(* h264_idct_add16)(uint8_t *dst, const int *blockoffset, int16_t *block, int stride, const uint8_t nnzc[15 *8])
h264_biweight_func biweight_h264_pixels_tab[4]
Context for storing H.264 DSP functions.
typedef void(APIENTRY *FF_PFNGLACTIVETEXTUREPROC)(GLenum texture)
static void bit_depth(AudioStatsContext *s, uint64_t mask, uint64_t imask, AVRational *depth)
void(* h264_h_loop_filter_luma_mbaff_intra)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta)
void ff_h264dsp_init_ppc(H264DSPContext *c, const int bit_depth, const int chroma_format_idc)
void(* h264_v_loop_filter_luma)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta, int8_t *tc0)
void(* h264_idct8_add)(uint8_t *dst, int16_t *block, int stride)
static const int8_t mv[256][2]
static const int16_t alpha[]
void(* h264_loop_filter_strength)(int16_t bS[2][4][4], uint8_t nnz[40], int8_t ref[2][40], int16_t mv[2][40][2], int bidir, int edges, int step, int mask_mv0, int mask_mv1, int field)
void(* h264_idct8_add4)(uint8_t *dst, const int *blockoffset, int16_t *block, int stride, const uint8_t nnzc[15 *8])
void(* h264_weight_func)(uint8_t *block, ptrdiff_t stride, int height, int log2_denom, int weight, int offset)
void(* h264_luma_dc_dequant_idct)(int16_t *output, int16_t *input, int qmul)
void(* h264_add_pixels4_clear)(uint8_t *dst, int16_t *block, int stride)
and forward the test the status of outputs and forward it to the corresponding return FFERROR_NOT_READY If the filters stores internally one or a few frame for some input
static int weight(int i, int blen, int offset)
static int ref[MAX_W *MAX_W]
void ff_h264dsp_init_x86(H264DSPContext *c, const int bit_depth, const int chroma_format_idc)
void(* h264_h_loop_filter_chroma_mbaff_intra)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta)
void(* h264_h_loop_filter_luma_mbaff)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta, int8_t *tc0)
void(* h264_idct_dc_add)(uint8_t *dst, int16_t *block, int stride)
void(* h264_idct_add8)(uint8_t **dst, const int *blockoffset, int16_t *block, int stride, const uint8_t nnzc[15 *8])
void(* h264_v_loop_filter_chroma)(uint8_t *pix, ptrdiff_t stride, int alpha, int beta, int8_t *tc0)
trying all byte sequences megabyte in length and selecting the best looking sequence will yield cases to try But a word about which is also called distortion Distortion can be quantified by almost any quality measurement one chooses the sum of squared differences is used but more complex methods that consider psychovisual effects can be used as well It makes no difference in this discussion First step