Go to the documentation of this file.
76 for (n = 0; n < 6; n++) {
95 ptrdiff_t linesize =
frame->linesize[0];
96 uint8_t *dest_y =
frame->data[0] + t->
mb_y * 16 * linesize + t->
mb_x * 16;
113 const int64_t qscale = (215 - 2*
quant)*5;
124 const uint8_t *buf = avpkt->
data;
125 int buf_size = avpkt->
size;
126 const uint8_t *buf_end = buf+buf_size;
152 (buf_end - buf) / 4);
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
void(* clear_blocks)(int16_t *blocks)
static int tqi_decode_mb(TqiContext *t, int16_t(*block)[64])
static av_cold int tqi_decode_end(AVCodecContext *avctx)
This structure describes decoded (raw) audio or video data.
int ff_set_dimensions(AVCodecContext *s, int width, int height)
Check that the provided frame dimensions are valid and set them on the codec context.
static int init_get_bits(GetBitContext *s, const uint8_t *buffer, int bit_size)
Initialize GetBitContext.
AVCodec p
The public AVCodec.
static int tqi_decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt)
int flags
AV_CODEC_FLAG_*.
void ff_ea_idct_put_c(uint8_t *dest, ptrdiff_t linesize, int16_t *block)
static int quant(float coef, const float Q, const float rounding)
Quantize one coefficient.
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
unsigned int bitstream_buf_size
#define FF_CODEC_DECODE_CB(func)
av_cold void ff_blockdsp_init(BlockDSPContext *c)
static void tqi_calculate_qtable(TqiContext *t, int quant)
void(* bswap_buf)(uint32_t *dst, const uint32_t *src, int w)
int(* init)(AVBSFContext *ctx)
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_RL16
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
#define CODEC_LONG_NAME(str)
av_cold void ff_bswapdsp_init(BswapDSPContext *c)
Rational number (pair of numerator and denominator).
static av_cold int tqi_decode_init(AVCodecContext *avctx)
int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Get a buffer for a frame.
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
#define AV_CODEC_FLAG_GRAY
Only decode/encode grayscale.
#define DECLARE_ALIGNED(n, t, v)
const uint16_t ff_mpeg1_default_intra_matrix[256]
const FFCodec ff_eatqi_decoder
int ff_mpeg1_decode_block_intra(GetBitContext *gb, const uint16_t *quant_matrix, const uint8_t *scantable, int last_dc[3], int16_t *block, int index, int qscale)
#define i(width, name, range_min, range_max)
av_cold void ff_mpeg12_init_vlcs(void)
void av_fast_padded_malloc(void *ptr, unsigned int *size, size_t min_size)
Same behaviour av_fast_malloc but the buffer has additional AV_INPUT_BUFFER_PADDING_SIZE at the end w...
const char * name
Name of the codec implementation.
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
const uint8_t ff_zigzag_direct[64]
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
const uint16_t ff_inv_aanscales[64]
main external API structure.
This structure stores compressed data.
uint16_t intra_matrix[64]
The exact code depends on how similar the blocks are and how related they are to the block
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
static void tqi_idct_put(AVCodecContext *avctx, AVFrame *frame, int16_t(*block)[64])