FFmpeg
fraps.c
Go to the documentation of this file.
1 /*
2  * Fraps FPS1 decoder
3  * Copyright (c) 2005 Roine Gustafsson
4  * Copyright (c) 2006 Konstantin Shishkov
5  *
6  * This file is part of FFmpeg.
7  *
8  * FFmpeg is free software; you can redistribute it and/or
9  * modify it under the terms of the GNU Lesser General Public
10  * License as published by the Free Software Foundation; either
11  * version 2.1 of the License, or (at your option) any later version.
12  *
13  * FFmpeg is distributed in the hope that it will be useful,
14  * but WITHOUT ANY WARRANTY; without even the implied warranty of
15  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16  * Lesser General Public License for more details.
17  *
18  * You should have received a copy of the GNU Lesser General Public
19  * License along with FFmpeg; if not, write to the Free Software
20  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
21  */
22 
23 /**
24  * @file
25  * Lossless Fraps 'FPS1' decoder
26  * @author Roine Gustafsson (roine at users sf net)
27  * @author Konstantin Shishkov
28  *
29  * Codec algorithm for version 0 is taken from Transcode <www.transcoding.org>
30  *
31  * Version 2 files support by Konstantin Shishkov
32  */
33 
34 #include "avcodec.h"
35 #include "get_bits.h"
36 #include "huffman.h"
37 #include "bytestream.h"
38 #include "bswapdsp.h"
39 #include "internal.h"
40 #include "thread.h"
41 
42 #define FPS_TAG MKTAG('F', 'P', 'S', 'x')
43 #define VLC_BITS 11
44 
45 /**
46  * local variable storage
47  */
48 typedef struct FrapsContext {
53 } FrapsContext;
54 
55 
56 /**
57  * initializes decoder
58  * @param avctx codec context
59  * @return 0 on success or negative if fails
60  */
62 {
63  FrapsContext * const s = avctx->priv_data;
64 
65  s->avctx = avctx;
66  s->tmpbuf = NULL;
67 
69 
70  return 0;
71 }
72 
73 /**
74  * Comparator - our nodes should ascend by count
75  * but with preserved symbol order
76  */
77 static int huff_cmp(const void *va, const void *vb)
78 {
79  const Node *a = va, *b = vb;
80  return (a->count - b->count)*256 + a->sym - b->sym;
81 }
82 
83 /**
84  * decode Fraps v2 packed plane
85  */
86 static int fraps2_decode_plane(FrapsContext *s, uint8_t *dst, int stride, int w,
87  int h, const uint8_t *src, int size, int Uoff,
88  const int step)
89 {
90  int i, j, ret;
91  GetBitContext gb;
92  VLC vlc;
93  Node nodes[512];
94 
95  for (i = 0; i < 256; i++)
96  nodes[i].count = bytestream_get_le32(&src);
97  size -= 1024;
98  if ((ret = ff_huff_build_tree(s->avctx, &vlc, 256, VLC_BITS,
99  nodes, huff_cmp,
101  return ret;
102  /* we have built Huffman table and are ready to decode plane */
103 
104  /* convert bits so they may be used by standard bitreader */
105  s->bdsp.bswap_buf((uint32_t *) s->tmpbuf,
106  (const uint32_t *) src, size >> 2);
107 
108  if ((ret = init_get_bits8(&gb, s->tmpbuf, size)) < 0)
109  return ret;
110 
111  for (j = 0; j < h; j++) {
112  for (i = 0; i < w*step; i += step) {
113  dst[i] = get_vlc2(&gb, vlc.table, VLC_BITS, 3);
114  /* lines are stored as deltas between previous lines
115  * and we need to add 0x80 to the first lines of chroma planes
116  */
117  if (j)
118  dst[i] += dst[i - stride];
119  else if (Uoff)
120  dst[i] += 0x80;
121  if (get_bits_left(&gb) < 0) {
122  ff_free_vlc(&vlc);
123  return AVERROR_INVALIDDATA;
124  }
125  }
126  dst += stride;
127  }
128  ff_free_vlc(&vlc);
129  return 0;
130 }
131 
133  void *data, int *got_frame,
134  AVPacket *avpkt)
135 {
136  FrapsContext * const s = avctx->priv_data;
137  const uint8_t *buf = avpkt->data;
138  int buf_size = avpkt->size;
139  ThreadFrame frame = { .f = data };
140  AVFrame * const f = data;
141  uint32_t header;
142  unsigned int version,header_size;
143  unsigned int x, y;
144  const uint32_t *buf32;
145  uint32_t *luma1,*luma2,*cb,*cr;
146  uint32_t offs[4];
147  int i, j, ret, is_chroma;
148  const int planes = 3;
149  int is_pal;
150  uint8_t *out;
151 
152  if (buf_size < 4) {
153  av_log(avctx, AV_LOG_ERROR, "Packet is too short\n");
154  return AVERROR_INVALIDDATA;
155  }
156 
157  header = AV_RL32(buf);
158  version = header & 0xff;
159  is_pal = buf[1] == 2 && version == 1;
160  header_size = (header & (1<<30))? 8 : 4; /* bit 30 means pad to 8 bytes */
161 
162  if (version > 5) {
163  avpriv_report_missing_feature(avctx, "Fraps version %u", version);
164  return AVERROR_PATCHWELCOME;
165  }
166 
167  buf += header_size;
168 
169  if (is_pal) {
170  unsigned needed_size = avctx->width * avctx->height + 1024;
171  needed_size += header_size;
172  if (buf_size != needed_size) {
173  av_log(avctx, AV_LOG_ERROR,
174  "Invalid frame length %d (should be %d)\n",
175  buf_size, needed_size);
176  return AVERROR_INVALIDDATA;
177  }
178  } else if (version < 2) {
179  unsigned needed_size = avctx->width * avctx->height * 3;
180  if (version == 0) needed_size /= 2;
181  needed_size += header_size;
182  /* bit 31 means same as previous pic */
183  if (header & (1U<<31)) {
184  *got_frame = 0;
185  return buf_size;
186  }
187  if (buf_size != needed_size) {
188  av_log(avctx, AV_LOG_ERROR,
189  "Invalid frame length %d (should be %d)\n",
190  buf_size, needed_size);
191  return AVERROR_INVALIDDATA;
192  }
193  } else {
194  /* skip frame */
195  if (buf_size == 8) {
196  *got_frame = 0;
197  return buf_size;
198  }
199  if (AV_RL32(buf) != FPS_TAG || buf_size < planes*1024 + 24) {
200  av_log(avctx, AV_LOG_ERROR, "error in data stream\n");
201  return AVERROR_INVALIDDATA;
202  }
203  for (i = 0; i < planes; i++) {
204  offs[i] = AV_RL32(buf + 4 + i * 4);
205  if (offs[i] >= buf_size - header_size || (i && offs[i] <= offs[i - 1] + 1024)) {
206  av_log(avctx, AV_LOG_ERROR, "plane %i offset is out of bounds\n", i);
207  return AVERROR_INVALIDDATA;
208  }
209  }
210  offs[planes] = buf_size - header_size;
211  for (i = 0; i < planes; i++) {
212  av_fast_padded_malloc(&s->tmpbuf, &s->tmpbuf_size, offs[i + 1] - offs[i] - 1024);
213  if (!s->tmpbuf)
214  return AVERROR(ENOMEM);
215  }
216  }
217 
219  f->key_frame = 1;
220 
221  avctx->pix_fmt = version & 1 ? is_pal ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_BGR24 : AV_PIX_FMT_YUVJ420P;
222  avctx->color_range = version & 1 ? AVCOL_RANGE_UNSPECIFIED
224  avctx->colorspace = version & 1 ? AVCOL_SPC_UNSPECIFIED : AVCOL_SPC_BT709;
225 
226  if ((ret = ff_thread_get_buffer(avctx, &frame, 0)) < 0)
227  return ret;
228 
229  switch (version) {
230  case 0:
231  default:
232  /* Fraps v0 is a reordered YUV420 */
233  if (((avctx->width % 8) != 0) || ((avctx->height % 2) != 0)) {
234  av_log(avctx, AV_LOG_ERROR, "Invalid frame size %dx%d\n",
235  avctx->width, avctx->height);
236  return AVERROR_INVALIDDATA;
237  }
238 
239  buf32 = (const uint32_t*)buf;
240  for (y = 0; y < avctx->height / 2; y++) {
241  luma1 = (uint32_t*)&f->data[0][ y * 2 * f->linesize[0] ];
242  luma2 = (uint32_t*)&f->data[0][ (y * 2 + 1) * f->linesize[0] ];
243  cr = (uint32_t*)&f->data[1][ y * f->linesize[1] ];
244  cb = (uint32_t*)&f->data[2][ y * f->linesize[2] ];
245  for (x = 0; x < avctx->width; x += 8) {
246  *luma1++ = *buf32++;
247  *luma1++ = *buf32++;
248  *luma2++ = *buf32++;
249  *luma2++ = *buf32++;
250  *cr++ = *buf32++;
251  *cb++ = *buf32++;
252  }
253  }
254  break;
255 
256  case 1:
257  if (is_pal) {
258  uint32_t *pal = (uint32_t *)f->data[1];
259 
260  for (y = 0; y < 256; y++) {
261  pal[y] = AV_RL32(buf) | 0xFF000000;
262  buf += 4;
263  }
264 
265  for (y = 0; y <avctx->height; y++)
266  memcpy(&f->data[0][y * f->linesize[0]],
267  &buf[y * avctx->width],
268  avctx->width);
269  } else {
270  /* Fraps v1 is an upside-down BGR24 */
271  for (y = 0; y<avctx->height; y++)
272  memcpy(&f->data[0][(avctx->height - y - 1) * f->linesize[0]],
273  &buf[y * avctx->width * 3],
274  3 * avctx->width);
275  }
276  break;
277 
278  case 2:
279  case 4:
280  /**
281  * Fraps v2 is Huffman-coded YUV420 planes
282  * Fraps v4 is virtually the same
283  */
284  for (i = 0; i < planes; i++) {
285  is_chroma = !!i;
286  if ((ret = fraps2_decode_plane(s, f->data[i], f->linesize[i],
287  avctx->width >> is_chroma,
288  avctx->height >> is_chroma,
289  buf + offs[i], offs[i + 1] - offs[i],
290  is_chroma, 1)) < 0) {
291  av_log(avctx, AV_LOG_ERROR, "Error decoding plane %i\n", i);
292  return ret;
293  }
294  }
295  break;
296  case 3:
297  case 5:
298  /* Virtually the same as version 4, but is for RGB24 */
299  for (i = 0; i < planes; i++) {
300  if ((ret = fraps2_decode_plane(s, f->data[0] + i + (f->linesize[0] * (avctx->height - 1)),
301  -f->linesize[0], avctx->width, avctx->height,
302  buf + offs[i], offs[i + 1] - offs[i], 0, 3)) < 0) {
303  av_log(avctx, AV_LOG_ERROR, "Error decoding plane %i\n", i);
304  return ret;
305  }
306  }
307  out = f->data[0];
308  // convert pseudo-YUV into real RGB
309  for (j = 0; j < avctx->height; j++) {
310  uint8_t *line_end = out + 3*avctx->width;
311  while (out < line_end) {
312  out[0] += out[1];
313  out[2] += out[1];
314  out += 3;
315  }
316  out += f->linesize[0] - 3*avctx->width;
317  }
318  break;
319  }
320 
321  *got_frame = 1;
322 
323  return buf_size;
324 }
325 
326 
327 /**
328  * closes decoder
329  * @param avctx codec context
330  * @return 0 on success or negative if fails
331  */
333 {
334  FrapsContext *s = (FrapsContext*)avctx->priv_data;
335 
336  av_freep(&s->tmpbuf);
337  return 0;
338 }
339 
340 
342  .name = "fraps",
343  .long_name = NULL_IF_CONFIG_SMALL("Fraps"),
344  .type = AVMEDIA_TYPE_VIDEO,
345  .id = AV_CODEC_ID_FRAPS,
346  .priv_data_size = sizeof(FrapsContext),
347  .init = decode_init,
348  .close = decode_end,
349  .decode = decode_frame,
351  .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE,
352 };
also ITU-R BT1361 / IEC 61966-2-4 xvYCC709 / SMPTE RP177 Annex B
Definition: pixfmt.h:499
#define NULL
Definition: coverity.c:32
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
Definition: error.h:59
BswapDSPContext bdsp
Definition: fraps.c:50
This structure describes decoded (raw) audio or video data.
Definition: frame.h:295
ptrdiff_t const GLvoid * data
Definition: opengl_enc.c:100
int ff_huff_build_tree(AVCodecContext *avctx, VLC *vlc, int nb_codes, int nb_bits, Node *nodes, HuffCmp cmp, int flags)
nodes size must be 2*nb_codes first nb_codes nodes.count must be set
Definition: huffman.c:157
static av_cold int init(AVCodecContext *avctx)
Definition: avrndec.c:35
enum AVColorRange color_range
MPEG vs JPEG YUV range.
Definition: avcodec.h:2203
int size
Definition: avcodec.h:1481
The reader does not expect b to be semantically here and if the code is changed by maybe adding a a division or other the signedness will almost certainly be mistaken To avoid this confusion a new type was SUINT is the C unsigned type but it holds a signed int to use the same example SUINT a
Definition: undefined.txt:36
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
Definition: avcodec.h:1778
void av_fast_padded_malloc(void *ptr, unsigned int *size, size_t min_size)
Same behaviour av_fast_malloc but the buffer has additional AV_INPUT_BUFFER_PADDING_SIZE at the end w...
Definition: utils.c:70
int version
Definition: avisynth_c.h:858
#define src
Definition: vp8dsp.c:254
AVCodec.
Definition: avcodec.h:3492
static void decode(AVCodecContext *dec_ctx, AVPacket *pkt, AVFrame *frame, FILE *outfile)
Definition: decode_audio.c:71
Definition: agm.c:910
static double cb(void *priv, double x, double y)
Definition: vf_geq.c:139
#define FF_CODEC_CAP_INIT_THREADSAFE
The codec does not modify any global variables in the init function, allowing to call the init functi...
Definition: internal.h:40
uint8_t
#define av_cold
Definition: attributes.h:82
8 bits with AV_PIX_FMT_RGB32 palette
Definition: pixfmt.h:77
void(* bswap_buf)(uint32_t *dst, const uint32_t *src, int w)
Definition: bswapdsp.h:25
#define f(width, name)
Definition: cbs_vp9.c:255
Multithreading support functions.
uint8_t * data
Definition: avcodec.h:1480
bitstream reader API header.
ptrdiff_t size
Definition: opengl_enc.c:100
static const uint8_t header[24]
Definition: sdr2.c:67
static int fraps2_decode_plane(FrapsContext *s, uint8_t *dst, int stride, int w, int h, const uint8_t *src, int size, int Uoff, const int step)
decode Fraps v2 packed plane
Definition: fraps.c:86
#define av_log(a,...)
#define U(x)
Definition: vp56_arith.h:37
static int get_bits_left(GetBitContext *gb)
Definition: get_bits.h:849
#define i(width, name, range_min, range_max)
Definition: cbs_h2645.c:259
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
Definition: log.h:176
#define FF_HUFFMAN_FLAG_ZERO_COUNT
Definition: huffman.h:39
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
Definition: internal.h:186
const char * name
Name of the codec implementation.
Definition: avcodec.h:3499
GLsizei count
Definition: opengl_enc.c:108
#define AV_CODEC_CAP_FRAME_THREADS
Codec supports frame-level multithreading.
Definition: avcodec.h:1040
Definition: vlc.h:26
static const struct @321 planes[]
static int huff_cmp(const void *va, const void *vb)
Comparator - our nodes should ascend by count but with preserved symbol order.
Definition: fraps.c:77
#define b
Definition: input.c:41
enum AVPictureType pict_type
Picture type of the frame.
Definition: frame.h:378
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
Definition: pixfmt.h:78
int width
picture width / height.
Definition: avcodec.h:1741
uint8_t w
Definition: llviddspenc.c:38
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
#define s(width, name)
Definition: cbs_vp9.c:257
static av_always_inline int get_vlc2(GetBitContext *s, VLC_TYPE(*table)[2], int bits, int max_depth)
Parse a vlc code.
Definition: get_bits.h:797
packed RGB 8:8:8, 24bpp, BGRBGR...
Definition: pixfmt.h:69
#define VLC_BITS
Definition: fraps.c:43
the normal 2^n-1 "JPEG" YUV ranges
Definition: pixfmt.h:523
#define AVERROR_PATCHWELCOME
Not yet implemented in FFmpeg, patches welcome.
Definition: error.h:62
static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame, AVPacket *avpkt)
Definition: fraps.c:132
AVCodecContext * avctx
Definition: fraps.c:49
Libavcodec external API header.
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:326
static int init_get_bits8(GetBitContext *s, const uint8_t *buffer, int byte_size)
Initialize GetBitContext.
Definition: get_bits.h:677
static av_cold int decode_end(AVCodecContext *avctx)
closes decoder
Definition: fraps.c:332
int ff_thread_get_buffer(AVCodecContext *avctx, ThreadFrame *f, int flags)
Wrapper around get_buffer() for frame-multithreaded codecs.
main external API structure.
Definition: avcodec.h:1568
static av_cold int decode_init(AVCodecContext *avctx)
initializes decoder
Definition: fraps.c:61
void * buf
Definition: avisynth_c.h:766
enum AVColorSpace colorspace
YUV colorspace type.
Definition: avcodec.h:2196
huffman tree builder and VLC generator
int16_t sym
Definition: huffman.h:33
void avpriv_report_missing_feature(void *avc, const char *msg,...) av_printf_format(2
Log a generic warning message about a missing feature.
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:309
GLint GLenum GLboolean GLsizei stride
Definition: opengl_enc.c:104
common internal api header.
AVCodec ff_fraps_decoder
Definition: fraps.c:341
uint8_t * tmpbuf
Definition: fraps.c:51
av_cold void ff_bswapdsp_init(BswapDSPContext *c)
Definition: bswapdsp.c:49
void * priv_data
Definition: avcodec.h:1595
VLC_TYPE(* table)[2]
code, bits
Definition: vlc.h:28
int key_frame
1 -> keyframe, 0-> not
Definition: frame.h:373
local variable storage
Definition: fraps.c:48
uint32_t count
Definition: huffman.h:35
FILE * out
Definition: movenc.c:54
#define av_freep(p)
#define stride
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later.That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another.Frame references ownership and permissions
static double cr(void *priv, double x, double y)
Definition: vf_geq.c:140
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_RL32
Definition: bytestream.h:87
This structure stores compressed data.
Definition: avcodec.h:1457
int tmpbuf_size
Definition: fraps.c:52
void ff_free_vlc(VLC *vlc)
Definition: bitstream.c:359
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() for allocating buffers and supports custom allocators.
Definition: avcodec.h:984
for(j=16;j >0;--j)
#define FPS_TAG
Definition: fraps.c:42
trying all byte sequences megabyte in length and selecting the best looking sequence will yield cases to try But a word about which is also called distortion Distortion can be quantified by almost any quality measurement one chooses the sum of squared differences is used but more complex methods that consider psychovisual effects can be used as well It makes no difference in this discussion First step