FFmpeg
ulti.c
Go to the documentation of this file.
1 /*
2  * IBM Ultimotion Video Decoder
3  * Copyright (C) 2004 Konstantin Shishkov
4  *
5  * This file is part of FFmpeg.
6  *
7  * FFmpeg is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2.1 of the License, or (at your option) any later version.
11  *
12  * FFmpeg is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with FFmpeg; if not, write to the Free Software
19  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20  */
21 
22 /**
23  * @file
24  * IBM Ultimotion Video Decoder.
25  */
26 
27 #include <stdio.h>
28 #include <stdlib.h>
29 #include <string.h>
30 
31 #include "avcodec.h"
32 #include "bytestream.h"
33 #include "internal.h"
34 
35 #include "ulti_cb.h"
36 
37 typedef struct UltimotionDecodeContext {
44 
46 {
48 
49  s->avctx = avctx;
50  s->width = avctx->width;
51  s->height = avctx->height;
52  s->blocks = (s->width / 8) * (s->height / 8);
53  if (s->blocks == 0)
54  return AVERROR_INVALIDDATA;
55  avctx->pix_fmt = AV_PIX_FMT_YUV410P;
57 
58  s->frame = av_frame_alloc();
59  if (!s->frame)
60  return AVERROR(ENOMEM);
61 
62  return 0;
63 }
64 
66 {
68 
69  av_frame_free(&s->frame);
70 
71  return 0;
72 }
73 
74 static const int block_coords[8] = // 4x4 block coords in 8x8 superblock
75  { 0, 0, 0, 4, 4, 4, 4, 0};
76 
77 static const int angle_by_index[4] = { 0, 2, 6, 12};
78 
79 /* Lookup tables for luma and chroma - used by ulti_convert_yuv() */
80 static const uint8_t ulti_lumas[64] =
81  { 0x10, 0x13, 0x17, 0x1A, 0x1E, 0x21, 0x25, 0x28,
82  0x2C, 0x2F, 0x33, 0x36, 0x3A, 0x3D, 0x41, 0x44,
83  0x48, 0x4B, 0x4F, 0x52, 0x56, 0x59, 0x5C, 0x60,
84  0x63, 0x67, 0x6A, 0x6E, 0x71, 0x75, 0x78, 0x7C,
85  0x7F, 0x83, 0x86, 0x8A, 0x8D, 0x91, 0x94, 0x98,
86  0x9B, 0x9F, 0xA2, 0xA5, 0xA9, 0xAC, 0xB0, 0xB3,
87  0xB7, 0xBA, 0xBE, 0xC1, 0xC5, 0xC8, 0xCC, 0xCF,
88  0xD3, 0xD6, 0xDA, 0xDD, 0xE1, 0xE4, 0xE8, 0xEB};
89 
90 static const uint8_t ulti_chromas[16] =
91  { 0x60, 0x67, 0x6D, 0x73, 0x7A, 0x80, 0x86, 0x8D,
92  0x93, 0x99, 0xA0, 0xA6, 0xAC, 0xB3, 0xB9, 0xC0};
93 
94 /* convert Ultimotion YUV block (sixteen 6-bit Y samples and
95  two 4-bit chroma samples) into standard YUV and put it into frame */
96 static void ulti_convert_yuv(AVFrame *frame, int x, int y,
97  uint8_t *luma,int chroma)
98 {
99  uint8_t *y_plane, *cr_plane, *cb_plane;
100  int i;
101 
102  y_plane = frame->data[0] + x + y * frame->linesize[0];
103  cr_plane = frame->data[1] + (x / 4) + (y / 4) * frame->linesize[1];
104  cb_plane = frame->data[2] + (x / 4) + (y / 4) * frame->linesize[2];
105 
106  cr_plane[0] = ulti_chromas[chroma >> 4];
107 
108  cb_plane[0] = ulti_chromas[chroma & 0xF];
109 
110 
111  for(i = 0; i < 16; i++){
112  y_plane[i & 3] = ulti_lumas[luma[i]];
113  if((i & 3) == 3) { //next row
114  y_plane += frame->linesize[0];
115  }
116  }
117 }
118 
119 /* generate block like in MS Video1 */
120 static void ulti_pattern(AVFrame *frame, int x, int y,
121  int f0, int f1, int Y0, int Y1, int chroma)
122 {
123  uint8_t Luma[16];
124  int mask, i;
125  for(mask = 0x80, i = 0; mask; mask >>= 1, i++) {
126  if(f0 & mask)
127  Luma[i] = Y1;
128  else
129  Luma[i] = Y0;
130  }
131 
132  for(mask = 0x80, i = 8; mask; mask >>= 1, i++) {
133  if(f1 & mask)
134  Luma[i] = Y1;
135  else
136  Luma[i] = Y0;
137  }
138 
139  ulti_convert_yuv(frame, x, y, Luma, chroma);
140 }
141 
142 /* fill block with some gradient */
143 static void ulti_grad(AVFrame *frame, int x, int y, uint8_t *Y, int chroma, int angle)
144 {
145  uint8_t Luma[16];
146  if(angle & 8) { //reverse order
147  int t;
148  angle &= 0x7;
149  t = Y[0];
150  Y[0] = Y[3];
151  Y[3] = t;
152  t = Y[1];
153  Y[1] = Y[2];
154  Y[2] = t;
155  }
156  switch(angle){
157  case 0:
158  Luma[0] = Y[0]; Luma[1] = Y[1]; Luma[2] = Y[2]; Luma[3] = Y[3];
159  Luma[4] = Y[0]; Luma[5] = Y[1]; Luma[6] = Y[2]; Luma[7] = Y[3];
160  Luma[8] = Y[0]; Luma[9] = Y[1]; Luma[10] = Y[2]; Luma[11] = Y[3];
161  Luma[12] = Y[0]; Luma[13] = Y[1]; Luma[14] = Y[2]; Luma[15] = Y[3];
162  break;
163  case 1:
164  Luma[0] = Y[1]; Luma[1] = Y[2]; Luma[2] = Y[3]; Luma[3] = Y[3];
165  Luma[4] = Y[0]; Luma[5] = Y[1]; Luma[6] = Y[2]; Luma[7] = Y[3];
166  Luma[8] = Y[0]; Luma[9] = Y[1]; Luma[10] = Y[2]; Luma[11] = Y[3];
167  Luma[12] = Y[0]; Luma[13] = Y[0]; Luma[14] = Y[1]; Luma[15] = Y[2];
168  break;
169  case 2:
170  Luma[0] = Y[1]; Luma[1] = Y[2]; Luma[2] = Y[3]; Luma[3] = Y[3];
171  Luma[4] = Y[1]; Luma[5] = Y[2]; Luma[6] = Y[2]; Luma[7] = Y[3];
172  Luma[8] = Y[0]; Luma[9] = Y[1]; Luma[10] = Y[1]; Luma[11] = Y[2];
173  Luma[12] = Y[0]; Luma[13] = Y[0]; Luma[14] = Y[1]; Luma[15] = Y[2];
174  break;
175  case 3:
176  Luma[0] = Y[2]; Luma[1] = Y[3]; Luma[2] = Y[3]; Luma[3] = Y[3];
177  Luma[4] = Y[1]; Luma[5] = Y[2]; Luma[6] = Y[2]; Luma[7] = Y[3];
178  Luma[8] = Y[0]; Luma[9] = Y[1]; Luma[10] = Y[1]; Luma[11] = Y[2];
179  Luma[12] = Y[0]; Luma[13] = Y[0]; Luma[14] = Y[0]; Luma[15] = Y[1];
180  break;
181  case 4:
182  Luma[0] = Y[3]; Luma[1] = Y[3]; Luma[2] = Y[3]; Luma[3] = Y[3];
183  Luma[4] = Y[2]; Luma[5] = Y[2]; Luma[6] = Y[2]; Luma[7] = Y[2];
184  Luma[8] = Y[1]; Luma[9] = Y[1]; Luma[10] = Y[1]; Luma[11] = Y[1];
185  Luma[12] = Y[0]; Luma[13] = Y[0]; Luma[14] = Y[0]; Luma[15] = Y[0];
186  break;
187  case 5:
188  Luma[0] = Y[3]; Luma[1] = Y[3]; Luma[2] = Y[3]; Luma[3] = Y[2];
189  Luma[4] = Y[3]; Luma[5] = Y[2]; Luma[6] = Y[2]; Luma[7] = Y[1];
190  Luma[8] = Y[2]; Luma[9] = Y[1]; Luma[10] = Y[1]; Luma[11] = Y[0];
191  Luma[12] = Y[1]; Luma[13] = Y[0]; Luma[14] = Y[0]; Luma[15] = Y[0];
192  break;
193  case 6:
194  Luma[0] = Y[3]; Luma[1] = Y[3]; Luma[2] = Y[2]; Luma[3] = Y[2];
195  Luma[4] = Y[3]; Luma[5] = Y[2]; Luma[6] = Y[1]; Luma[7] = Y[1];
196  Luma[8] = Y[2]; Luma[9] = Y[2]; Luma[10] = Y[1]; Luma[11] = Y[0];
197  Luma[12] = Y[1]; Luma[13] = Y[1]; Luma[14] = Y[0]; Luma[15] = Y[0];
198  break;
199  case 7:
200  Luma[0] = Y[3]; Luma[1] = Y[3]; Luma[2] = Y[2]; Luma[3] = Y[1];
201  Luma[4] = Y[3]; Luma[5] = Y[2]; Luma[6] = Y[1]; Luma[7] = Y[0];
202  Luma[8] = Y[3]; Luma[9] = Y[2]; Luma[10] = Y[1]; Luma[11] = Y[0];
203  Luma[12] = Y[2]; Luma[13] = Y[1]; Luma[14] = Y[0]; Luma[15] = Y[0];
204  break;
205  default:
206  Luma[0] = Y[0]; Luma[1] = Y[0]; Luma[2] = Y[1]; Luma[3] = Y[1];
207  Luma[4] = Y[0]; Luma[5] = Y[0]; Luma[6] = Y[1]; Luma[7] = Y[1];
208  Luma[8] = Y[2]; Luma[9] = Y[2]; Luma[10] = Y[3]; Luma[11] = Y[3];
209  Luma[12] = Y[2]; Luma[13] = Y[2]; Luma[14] = Y[3]; Luma[15] = Y[3];
210  break;
211  }
212 
213  ulti_convert_yuv(frame, x, y, Luma, chroma);
214 }
215 
217  void *data, int *got_frame,
218  AVPacket *avpkt)
219 {
220  const uint8_t *buf = avpkt->data;
221  int buf_size = avpkt->size;
223  int modifier = 0;
224  int uniq = 0;
225  int mode = 0;
226  int blocks = 0;
227  int done = 0;
228  int x = 0, y = 0;
229  int i, ret;
230  int skip;
231  int tmp;
232 
233  if ((ret = ff_reget_buffer(avctx, s->frame, 0)) < 0)
234  return ret;
235 
236  bytestream2_init(&s->gb, buf, buf_size);
237 
238  while(!done) {
239  int idx;
240  if(blocks >= s->blocks || y >= s->height)
241  break;//all blocks decoded
242 
243  if (bytestream2_get_bytes_left(&s->gb) < 1)
244  goto err;
245  idx = bytestream2_get_byteu(&s->gb);
246  if((idx & 0xF8) == 0x70) {
247  switch(idx) {
248  case 0x70: //change modifier
249  modifier = bytestream2_get_byte(&s->gb);
250  if(modifier>1)
251  av_log(avctx, AV_LOG_INFO, "warning: modifier must be 0 or 1, got %i\n", modifier);
252  break;
253  case 0x71: // set uniq flag
254  uniq = 1;
255  break;
256  case 0x72: //toggle mode
257  mode = !mode;
258  break;
259  case 0x73: //end-of-frame
260  done = 1;
261  break;
262  case 0x74: //skip some blocks
263  skip = bytestream2_get_byte(&s->gb);
264  if ((blocks + skip) >= s->blocks)
265  break;
266  blocks += skip;
267  x += skip * 8;
268  while(x >= s->width) {
269  x -= s->width;
270  y += 8;
271  }
272  break;
273  default:
274  av_log(avctx, AV_LOG_INFO, "warning: unknown escape 0x%02X\n", idx);
275  }
276  } else { //handle one block
277  int code;
278  int cf;
279  int angle = 0;
280  uint8_t Y[4]; // luma samples of block
281  int tx = 0, ty = 0; //coords of subblock
282  int chroma = 0;
283  if (mode || uniq) {
284  uniq = 0;
285  cf = 1;
286  chroma = 0;
287  } else {
288  cf = 0;
289  if (idx) {
290  chroma = bytestream2_get_byte(&s->gb);
291  }
292  }
293  for (i = 0; i < 4; i++) { // for every subblock
294  code = (idx >> (6 - i*2)) & 3; //extract 2 bits
295  if(!code) //skip subblock
296  continue;
297  if(cf) {
298  chroma = bytestream2_get_byte(&s->gb);
299  }
300  tx = x + block_coords[i * 2];
301  ty = y + block_coords[(i * 2) + 1];
302  switch(code) {
303  case 1:
304  tmp = bytestream2_get_byte(&s->gb);
305 
306  angle = angle_by_index[(tmp >> 6) & 0x3];
307 
308  Y[0] = tmp & 0x3F;
309  Y[1] = Y[0];
310 
311  if (angle) {
312  Y[2] = Y[0]+1;
313  if (Y[2] > 0x3F)
314  Y[2] = 0x3F;
315  Y[3] = Y[2];
316  } else {
317  Y[2] = Y[0];
318  Y[3] = Y[0];
319  }
320  break;
321 
322  case 2:
323  if (modifier) { // unpack four luma samples
324  tmp = bytestream2_get_be24(&s->gb);
325 
326  Y[0] = (tmp >> 18) & 0x3F;
327  Y[1] = (tmp >> 12) & 0x3F;
328  Y[2] = (tmp >> 6) & 0x3F;
329  Y[3] = tmp & 0x3F;
330  angle = 16;
331  } else { // retrieve luma samples from codebook
332  tmp = bytestream2_get_be16(&s->gb);
333 
334  angle = (tmp >> 12) & 0xF;
335  tmp &= 0xFFF;
336  tmp <<= 2;
337  Y[0] = s->ulti_codebook[tmp];
338  Y[1] = s->ulti_codebook[tmp + 1];
339  Y[2] = s->ulti_codebook[tmp + 2];
340  Y[3] = s->ulti_codebook[tmp + 3];
341  }
342  break;
343 
344  case 3:
345  if (modifier) { // all 16 luma samples
346  uint8_t Luma[16];
347 
348  if (bytestream2_get_bytes_left(&s->gb) < 12)
349  goto err;
350  tmp = bytestream2_get_be24u(&s->gb);
351  Luma[0] = (tmp >> 18) & 0x3F;
352  Luma[1] = (tmp >> 12) & 0x3F;
353  Luma[2] = (tmp >> 6) & 0x3F;
354  Luma[3] = tmp & 0x3F;
355 
356  tmp = bytestream2_get_be24u(&s->gb);
357  Luma[4] = (tmp >> 18) & 0x3F;
358  Luma[5] = (tmp >> 12) & 0x3F;
359  Luma[6] = (tmp >> 6) & 0x3F;
360  Luma[7] = tmp & 0x3F;
361 
362  tmp = bytestream2_get_be24u(&s->gb);
363  Luma[8] = (tmp >> 18) & 0x3F;
364  Luma[9] = (tmp >> 12) & 0x3F;
365  Luma[10] = (tmp >> 6) & 0x3F;
366  Luma[11] = tmp & 0x3F;
367 
368  tmp = bytestream2_get_be24u(&s->gb);
369  Luma[12] = (tmp >> 18) & 0x3F;
370  Luma[13] = (tmp >> 12) & 0x3F;
371  Luma[14] = (tmp >> 6) & 0x3F;
372  Luma[15] = tmp & 0x3F;
373 
374  ulti_convert_yuv(s->frame, tx, ty, Luma, chroma);
375  } else {
376  if (bytestream2_get_bytes_left(&s->gb) < 4)
377  goto err;
378  tmp = bytestream2_get_byteu(&s->gb);
379  if(tmp & 0x80) {
380  angle = (tmp >> 4) & 0x7;
381  tmp = (tmp << 8) + bytestream2_get_byteu(&s->gb);
382  Y[0] = (tmp >> 6) & 0x3F;
383  Y[1] = tmp & 0x3F;
384  Y[2] = bytestream2_get_byteu(&s->gb) & 0x3F;
385  Y[3] = bytestream2_get_byteu(&s->gb) & 0x3F;
386  ulti_grad(s->frame, tx, ty, Y, chroma, angle); //draw block
387  } else { // some patterns
388  int f0 = tmp;
389  int f1 = bytestream2_get_byteu(&s->gb);
390  Y[0] = bytestream2_get_byteu(&s->gb) & 0x3F;
391  Y[1] = bytestream2_get_byteu(&s->gb) & 0x3F;
392  ulti_pattern(s->frame, tx, ty, f0, f1, Y[0], Y[1], chroma);
393  }
394  }
395  break;
396  }
397  if(code != 3)
398  ulti_grad(s->frame, tx, ty, Y, chroma, angle); // draw block
399  }
400  blocks++;
401  x += 8;
402  if(x >= s->width) {
403  x = 0;
404  y += 8;
405  }
406  }
407  }
408 
409  *got_frame = 1;
410  if ((ret = av_frame_ref(data, s->frame)) < 0)
411  return ret;
412 
413  return buf_size;
414 
415 err:
416  av_log(avctx, AV_LOG_ERROR,
417  "Insufficient data\n");
418  return AVERROR_INVALIDDATA;
419 }
420 
422  .name = "ultimotion",
423  .long_name = NULL_IF_CONFIG_SMALL("IBM UltiMotion"),
424  .type = AVMEDIA_TYPE_VIDEO,
425  .id = AV_CODEC_ID_ULTI,
426  .priv_data_size = sizeof(UltimotionDecodeContext),
428  .close = ulti_decode_end,
430  .capabilities = AV_CODEC_CAP_DR1,
431 };
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
Definition: error.h:59
This structure describes decoded (raw) audio or video data.
Definition: frame.h:295
static av_cold int ulti_decode_end(AVCodecContext *avctx)
Definition: ulti.c:65
ptrdiff_t const GLvoid * data
Definition: opengl_enc.c:100
static av_cold int init(AVCodecContext *avctx)
Definition: avrndec.c:35
int size
Definition: avcodec.h:1481
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
Definition: avcodec.h:1778
static av_always_inline void bytestream2_init(GetByteContext *g, const uint8_t *buf, int buf_size)
Definition: bytestream.h:133
AVCodec.
Definition: avcodec.h:3492
static void decode(AVCodecContext *dec_ctx, AVPacket *pkt, AVFrame *frame, FILE *outfile)
Definition: decode_audio.c:71
AVCodec ff_ulti_decoder
Definition: ulti.c:421
uint8_t
#define av_cold
Definition: attributes.h:82
AVFrame * av_frame_alloc(void)
Allocate an AVFrame and set its fields to default values.
Definition: frame.c:189
AVFrame * frame
Definition: ulti.c:40
int av_frame_ref(AVFrame *dst, const AVFrame *src)
Set up a new reference to the data described by the source frame.
Definition: frame.c:443
GetByteContext gb
Definition: ulti.c:42
uint8_t * data
Definition: avcodec.h:1480
#define av_log(a,...)
#define i(width, name, range_min, range_max)
Definition: cbs_h2645.c:259
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
Definition: log.h:176
static const uint16_t mask[17]
Definition: lzw.c:38
static av_always_inline void chroma(WaveformContext *s, AVFrame *in, AVFrame *out, int component, int intensity, int offset_y, int offset_x, int column, int mirror, int jobnr, int nb_jobs)
Definition: vf_waveform.c:1542
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:202
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
Definition: internal.h:186
int ff_reget_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Identical in function to ff_get_buffer(), except it reuses the existing buffer if available...
Definition: decode.c:2015
static av_always_inline unsigned int bytestream2_get_bytes_left(GetByteContext *g)
Definition: bytestream.h:154
const char * name
Name of the codec implementation.
Definition: avcodec.h:3499
AVCodecContext * avctx
Definition: ulti.c:38
static const int angle_by_index[4]
Definition: ulti.c:77
#define Y
Definition: boxblur.h:38
int width
picture width / height.
Definition: avcodec.h:1741
#define s(width, name)
Definition: cbs_vp9.c:257
#define AV_LOG_INFO
Standard information.
Definition: log.h:187
static int ulti_decode_frame(AVCodecContext *avctx, void *data, int *got_frame, AVPacket *avpkt)
Definition: ulti.c:216
Libavcodec external API header.
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:326
main external API structure.
Definition: avcodec.h:1568
void * buf
Definition: avisynth_c.h:766
static av_cold int ulti_decode_init(AVCodecContext *avctx)
Definition: ulti.c:45
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:72
const uint8_t * ulti_codebook
Definition: ulti.c:41
static void ulti_convert_yuv(AVFrame *frame, int x, int y, uint8_t *luma, int chroma)
Definition: ulti.c:96
and forward the test the status of outputs and forward it to the corresponding return FFERROR_NOT_READY If the filters stores internally one or a few frame for some it can consider them to be part of the FIFO and delay acknowledging a status change accordingly Example code
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:309
common internal api header.
void * priv_data
Definition: avcodec.h:1595
static void ulti_grad(AVFrame *frame, int x, int y, uint8_t *Y, int chroma, int angle)
Definition: ulti.c:143
static void ulti_pattern(AVFrame *frame, int x, int y, int f0, int f1, int Y0, int Y1, int chroma)
Definition: ulti.c:120
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later.That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another.Frame references ownership and permissions
static const uint8_t ulti_chromas[16]
Definition: ulti.c:90
static const int block_coords[8]
Definition: ulti.c:74
This structure stores compressed data.
Definition: avcodec.h:1457
mode
Use these values in ebur128_init (or&#39;ed).
Definition: ebur128.h:83
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() for allocating buffers and supports custom allocators.
Definition: avcodec.h:984
static const uint8_t ulti_lumas[64]
Definition: ulti.c:80
static uint8_t tmp[11]
Definition: aes_ctr.c:26