FFmpeg
sga.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2021 Paul B Mahol
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20 
21 #include "libavutil/common.h"
22 #include "libavutil/mem.h"
23 #include "avcodec.h"
24 #include "get_bits.h"
25 #include "bytestream.h"
26 #include "codec_internal.h"
27 #include "decode.h"
28 
29 #define PALDATA_FOLLOWS_TILEDATA 4
30 #define HAVE_COMPRESSED_TILEMAP 32
31 #define HAVE_TILEMAP 128
32 
33 typedef struct SGAVideoContext {
35 
45 
46  int flags;
47  int nb_pal;
48  int nb_tiles;
50  int shift;
51  int plus;
52  int swap;
53 
54  uint32_t pal[256];
55  uint8_t *tileindex_data;
56  unsigned tileindex_size;
57  uint8_t *palmapindex_data;
58  unsigned palmapindex_size;
59  uint8_t uncompressed[65536];
61 
63 {
64  avctx->pix_fmt = AV_PIX_FMT_PAL8;
65  return 0;
66 }
67 
68 static int decode_palette(GetByteContext *gb, uint32_t *pal)
69 {
70  GetBitContext gbit;
71 
72  if (bytestream2_get_bytes_left(gb) < 18)
73  return AVERROR_INVALIDDATA;
74 
75  memset(pal, 0, 16 * sizeof(*pal));
76  (void)init_get_bits8(&gbit, gb->buffer, 18);
77 
78  for (int RGBIndex = 0; RGBIndex < 3; RGBIndex++) {
79  for (int index = 0; index < 16; index++) {
80  unsigned color = get_bits1(&gbit) << RGBIndex;
81  pal[15 - index] |= color << (5 + 16);
82  }
83  }
84 
85  for (int RGBIndex = 0; RGBIndex < 3; RGBIndex++) {
86  for (int index = 0; index < 16; index++) {
87  unsigned color = get_bits1(&gbit) << RGBIndex;
88  pal[15 - index] |= color << (5 + 8);
89  }
90  }
91 
92  for (int RGBIndex = 0; RGBIndex < 3; RGBIndex++) {
93  for (int index = 0; index < 16; index++) {
94  unsigned color = get_bits1(&gbit) << RGBIndex;
95  pal[15 - index] |= color << (5 + 0);
96  }
97  }
98 
99  for (int index = 0; index < 16; index++)
100  pal[index] = (0xFFU << 24) | pal[index] | (pal[index] >> 3);
101 
102  bytestream2_skip(gb, 18);
103 
104  return 0;
105 }
106 
108 {
109  const uint8_t *tt = s->tileindex_data;
110 
111  for (int y = 0; y < s->tiles_h; y++) {
112  for (int x = 0; x < s->tiles_w; x++) {
113  int pal_idx = s->palmapindex_data[y * s->tiles_w + x] * 16;
114  uint8_t *dst = frame->data[0] + y * 8 * frame->linesize[0] + x * 8;
115 
116  for (int yy = 0; yy < 8; yy++) {
117  for (int xx = 0; xx < 8; xx++)
118  dst[xx] = pal_idx + tt[xx];
119  tt += 8;
120 
121  dst += frame->linesize[0];
122  }
123  }
124  }
125 
126  return 0;
127 }
128 
130 {
131  GetByteContext *gb = &s->gb, gb2;
132 
133  bytestream2_seek(gb, s->tilemapdata_offset, SEEK_SET);
134  if (bytestream2_get_bytes_left(gb) < s->tilemapdata_size)
135  return AVERROR_INVALIDDATA;
136 
137  gb2 = *gb;
138 
139  for (int y = 0; y < s->tiles_h; y++) {
140  for (int x = 0; x < s->tiles_w; x++) {
141  uint8_t tile[64];
142  int tilemap = bytestream2_get_be16u(&gb2);
143  int flip_x = (tilemap >> 11) & 1;
144  int flip_y = (tilemap >> 12) & 1;
145  int tindex = av_clip((tilemap & 511) - 1, 0, s->nb_tiles - 1);
146  const uint8_t *tt = s->tileindex_data + tindex * 64;
147  int pal_idx = ((tilemap >> 13) & 3) * 16;
148  uint8_t *dst = frame->data[0] + y * 8 * frame->linesize[0] + x * 8;
149 
150  if (!flip_x && !flip_y) {
151  memcpy(tile, tt, 64);
152  } else if (flip_x && flip_y) {
153  for (int i = 0; i < 8; i++) {
154  for (int j = 0; j < 8; j++)
155  tile[i * 8 + j] = tt[(7 - i) * 8 + 7 - j];
156  }
157  } else if (flip_x) {
158  for (int i = 0; i < 8; i++) {
159  for (int j = 0; j < 8; j++)
160  tile[i * 8 + j] = tt[i * 8 + 7 - j];
161  }
162  } else {
163  for (int i = 0; i < 8; i++) {
164  for (int j = 0; j < 8; j++)
165  tile[i * 8 + j] = tt[(7 - i) * 8 + j];
166  }
167  }
168 
169  for (int yy = 0; yy < 8; yy++) {
170  for (int xx = 0; xx < 8; xx++)
171  dst[xx] = pal_idx + tile[xx + yy * 8];
172 
173  dst += frame->linesize[0];
174  }
175  }
176  }
177 
178  return 0;
179 }
180 
182 {
183  const uint8_t *src = s->tileindex_data;
184  uint8_t *dst = frame->data[0];
185 
186  for (int y = 0; y < frame->height; y += 8) {
187  for (int x = 0; x < frame->width; x += 8) {
188  for (int yy = 0; yy < 8; yy++) {
189  for (int xx = 0; xx < 8; xx++)
190  dst[x + xx + yy * frame->linesize[0]] = src[xx];
191  src += 8;
192  }
193  }
194 
195  dst += 8 * frame->linesize[0];
196  }
197 
198  return 0;
199 }
200 
201 static int lzss_decompress(AVCodecContext *avctx,
202  GetByteContext *gb, uint8_t *dst,
203  int dst_size, int shift, int plus)
204 {
205  int oi = 0;
206 
207  while (bytestream2_get_bytes_left(gb) > 0 && oi < dst_size) {
208  uint16_t displace, header = bytestream2_get_be16(gb);
209  int count, offset;
210 
211  for (int i = 0; i < 16; i++) {
212  switch (header >> 15) {
213  case 0:
214  if (oi + 2 < dst_size) {
215  dst[oi++] = bytestream2_get_byte(gb);
216  dst[oi++] = bytestream2_get_byte(gb);
217  }
218  break;
219  case 1:
220  displace = bytestream2_get_be16(gb);
221  count = displace >> shift;
222  offset = displace & ((1 << shift) - 1);
223 
224  if (displace == 0) {
225  while (bytestream2_get_bytes_left(gb) > 0 &&
226  oi < dst_size)
227  dst[oi++] = bytestream2_get_byte(gb);
228  return oi;
229  }
230 
231  count += plus;
232 
233  if (offset <= 0)
234  offset = 1;
235  if (oi < offset || oi + count * 2 > dst_size)
236  return AVERROR_INVALIDDATA;
237  for (int j = 0; j < count * 2; j++) {
238  dst[oi] = dst[oi - offset];
239  oi++;
240  }
241  break;
242  }
243 
244  header <<= 1;
245  }
246  }
247 
248  return AVERROR_INVALIDDATA;
249 }
250 
252 {
253  SGAVideoContext *s = avctx->priv_data;
254  const int bits = (s->nb_pal + 1) / 2;
255  GetByteContext *gb = &s->gb;
256  GetBitContext pm;
257  int ret;
258 
259  bytestream2_seek(gb, s->palmapdata_offset, SEEK_SET);
260  if (bytestream2_get_bytes_left(gb) < s->palmapdata_size)
261  return AVERROR_INVALIDDATA;
262  ret = init_get_bits8(&pm, gb->buffer, s->palmapdata_size);
263  av_assert1(ret >= 0);
264 
265  for (int y = 0; y < s->tiles_h; y++) {
266  uint8_t *dst = s->palmapindex_data + y * s->tiles_w;
267 
268  for (int x = 0; x < s->tiles_w; x++)
269  dst[x] = get_bits(&pm, bits);
270 
271  dst += s->tiles_w;
272  }
273 
274  return 0;
275 }
276 
277 static int decode_tiledata(AVCodecContext *avctx)
278 {
279  SGAVideoContext *s = avctx->priv_data;
280  GetByteContext *gb = &s->gb;
281  GetBitContext tm;
282  int ret;
283 
284  bytestream2_seek(gb, s->tiledata_offset, SEEK_SET);
285  if (bytestream2_get_bytes_left(gb) < s->tiledata_size)
286  return AVERROR_INVALIDDATA;
287  ret = init_get_bits8(&tm, gb->buffer, s->tiledata_size);
288  av_assert1(ret >= 0);
289 
290  for (int n = 0; n < s->nb_tiles; n++) {
291  uint8_t *dst = s->tileindex_data + n * 64;
292 
293  for (int yy = 0; yy < 8; yy++) {
294  for (int xx = 0; xx < 8; xx++)
295  dst[xx] = get_bits(&tm, 4);
296 
297  dst += 8;
298  }
299  }
300 
301  for (int i = 0; i < s->nb_tiles && s->swap; i++) {
302  uint8_t *dst = s->tileindex_data + i * 64;
303 
304  for (int j = 8; j < 64; j += 16) {
305  for (int k = 0; k < 8; k += 2)
306  FFSWAP(uint8_t, dst[j + k], dst[j+k+1]);
307  }
308  }
309 
310  return 0;
311 }
312 
314  int *got_frame, AVPacket *avpkt)
315 {
316  SGAVideoContext *s = avctx->priv_data;
317  GetByteContext *gb = &s->gb;
318  int ret, type;
319 
320  if (avpkt->size <= 14)
321  return AVERROR_INVALIDDATA;
322 
323  s->flags = avpkt->data[8];
324  s->nb_pal = avpkt->data[9];
325  s->tiles_w = avpkt->data[10];
326  s->tiles_h = avpkt->data[11];
327 
328  if (s->nb_pal > 4)
329  return AVERROR_INVALIDDATA;
330 
331  if ((ret = ff_set_dimensions(avctx,
332  s->tiles_w * 8,
333  s->tiles_h * 8)) < 0)
334  return ret;
335 
336  av_fast_padded_malloc(&s->tileindex_data, &s->tileindex_size,
337  avctx->width * avctx->height);
338  if (!s->tileindex_data)
339  return AVERROR(ENOMEM);
340 
341  av_fast_padded_malloc(&s->palmapindex_data, &s->palmapindex_size,
342  s->tiles_w * s->tiles_h);
343  if (!s->palmapindex_data)
344  return AVERROR(ENOMEM);
345 
346  if ((ret = ff_get_buffer(avctx, frame, 0)) < 0)
347  return ret;
348 
349  bytestream2_init(gb, avpkt->data, avpkt->size);
350 
351  type = bytestream2_get_byte(gb);
352  s->metadata_size = 12 + ((!!(s->flags & HAVE_TILEMAP)) * 2);
353  s->nb_tiles = s->flags & HAVE_TILEMAP ? AV_RB16(avpkt->data + 12) : s->tiles_w * s->tiles_h;
354  if (s->nb_tiles > s->tiles_w * s->tiles_h)
355  return AVERROR_INVALIDDATA;
356 
357  av_log(avctx, AV_LOG_DEBUG, "type: %X flags: %X nb_tiles: %d\n", type, s->flags, s->nb_tiles);
358 
359  switch (type) {
360  case 0xE7:
361  case 0xCB:
362  case 0xCD:
363  s->swap = 1;
364  s->shift = 12;
365  s->plus = 1;
366  break;
367  case 0xC9:
368  s->swap = 1;
369  s->shift = 13;
370  s->plus = 1;
371  break;
372  case 0xC8:
373  s->swap = 1;
374  s->shift = 13;
375  s->plus = 0;
376  break;
377  case 0xC7:
378  s->swap = 0;
379  s->shift = 13;
380  s->plus = 1;
381  break;
382  case 0xC6:
383  s->swap = 0;
384  s->shift = 13;
385  s->plus = 0;
386  break;
387  }
388 
389  if (type == 0xE7) {
390  int offset = s->metadata_size, left;
391  int sizes[3];
392 
393  bytestream2_seek(gb, s->metadata_size, SEEK_SET);
394 
395  for (int i = 0; i < 3; i++)
396  sizes[i] = bytestream2_get_be16(gb);
397 
398  for (int i = 0; i < 3; i++) {
399  int size = sizes[i];
400  int raw = size >> 15;
401 
402  size &= (1 << 15) - 1;
403 
404  if (raw) {
406  return AVERROR_INVALIDDATA;
407 
408  if (sizeof(s->uncompressed) - offset < size)
409  return AVERROR_INVALIDDATA;
410 
411  memcpy(s->uncompressed + offset, gb->buffer, size);
412  bytestream2_skip(gb, size);
413  } else {
414  GetByteContext gb2;
415 
417  return AVERROR_INVALIDDATA;
418 
419  bytestream2_init(&gb2, gb->buffer, size);
420  ret = lzss_decompress(avctx, &gb2, s->uncompressed + offset,
421  sizeof(s->uncompressed) - offset, s->shift, s->plus);
422  if (ret < 0)
423  return ret;
424  bytestream2_skip(gb, size);
425  size = ret;
426  }
427 
428  offset += size;
429  }
430 
432  if (sizeof(s->uncompressed) - offset < left)
433  return AVERROR_INVALIDDATA;
434 
435  bytestream2_get_buffer(gb, s->uncompressed + offset, left);
436 
437  offset += left;
438  bytestream2_init(gb, s->uncompressed, offset);
439  }
440 
441  switch (type) {
442  case 0xCD:
443  case 0xCB:
444  case 0xC9:
445  case 0xC8:
446  case 0xC7:
447  case 0xC6:
448  bytestream2_seek(gb, s->metadata_size, SEEK_SET);
449  ret = lzss_decompress(avctx, gb, s->uncompressed + s->metadata_size,
450  sizeof(s->uncompressed) - s->metadata_size, s->shift, s->plus);
451  if (ret < 0)
452  return ret;
453  bytestream2_init(gb, s->uncompressed, ret + s->metadata_size);
454  case 0xE7:
455  case 0xC1:
456  s->tiledata_size = s->nb_tiles * 32;
457  s->paldata_size = s->nb_pal * 18;
458  s->tiledata_offset = s->flags & PALDATA_FOLLOWS_TILEDATA ? s->metadata_size : s->metadata_size + s->paldata_size;
459  s->paldata_offset = s->flags & PALDATA_FOLLOWS_TILEDATA ? s->metadata_size + s->tiledata_size : s->metadata_size;
460  s->palmapdata_offset = (s->flags & HAVE_TILEMAP) ? -1 : s->paldata_offset + s->paldata_size;
461  s->palmapdata_size = (s->flags & HAVE_TILEMAP) || s->nb_pal < 2 ? 0 : (s->tiles_w * s->tiles_h * ((s->nb_pal + 1) / 2) + 7) / 8;
462  s->tilemapdata_size = (s->flags & HAVE_TILEMAP) ? s->tiles_w * s->tiles_h * 2 : 0;
463  s->tilemapdata_offset = (s->flags & HAVE_TILEMAP) ? s->paldata_offset + s->paldata_size: -1;
464 
465  bytestream2_seek(gb, s->paldata_offset, SEEK_SET);
466  for (int n = 0; n < s->nb_pal; n++) {
467  ret = decode_palette(gb, s->pal + 16 * n);
468  if (ret < 0)
469  return ret;
470  }
471 
472  if (s->tiledata_size > 0) {
473  ret = decode_tiledata(avctx);
474  if (ret < 0)
475  return ret;
476  }
477 
478  if (s->palmapdata_size > 0) {
479  ret = decode_palmapdata(avctx);
480  if (ret < 0)
481  return ret;
482  }
483 
484  if (s->palmapdata_size > 0 && s->tiledata_size > 0) {
486  if (ret < 0)
487  return ret;
488  } else if (s->tilemapdata_size > 0 && s->tiledata_size > 0) {
490  if (ret < 0)
491  return ret;
492  } else if (s->tiledata_size > 0) {
493  ret = decode_index(s, frame);
494  if (ret < 0)
495  return ret;
496  }
497  break;
498  default:
499  av_log(avctx, AV_LOG_ERROR, "Unknown type: %X\n", type);
500  return AVERROR_INVALIDDATA;
501  }
502 
503  memcpy(frame->data[1], s->pal, AVPALETTE_SIZE);
504 #if FF_API_PALETTE_HAS_CHANGED
506  frame->palette_has_changed = 1;
508 #endif
509  frame->pict_type = AV_PICTURE_TYPE_I;
510  frame->flags |= AV_FRAME_FLAG_KEY;
511 
512  *got_frame = 1;
513 
514  return avpkt->size;
515 }
516 
518 {
519  SGAVideoContext *s = avctx->priv_data;
520 
521  av_freep(&s->tileindex_data);
522  s->tileindex_size = 0;
523 
524  av_freep(&s->palmapindex_data);
525  s->palmapindex_size = 0;
526 
527  return 0;
528 }
529 
531  .p.name = "sga",
532  CODEC_LONG_NAME("Digital Pictures SGA Video"),
533  .p.type = AVMEDIA_TYPE_VIDEO,
534  .p.id = AV_CODEC_ID_SGA_VIDEO,
535  .priv_data_size = sizeof(SGAVideoContext),
538  .close = sga_decode_end,
539  .p.capabilities = AV_CODEC_CAP_DR1,
540 };
FF_ENABLE_DEPRECATION_WARNINGS
#define FF_ENABLE_DEPRECATION_WARNINGS
Definition: internal.h:73
SGAVideoContext::tileindex_data
uint8_t * tileindex_data
Definition: sga.c:55
av_clip
#define av_clip
Definition: common.h:100
AVERROR
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
SGAVideoContext::palmapindex_size
unsigned palmapindex_size
Definition: sga.c:58
color
Definition: vf_paletteuse.c:513
GetByteContext
Definition: bytestream.h:33
SGAVideoContext::pal
uint32_t pal[256]
Definition: sga.c:54
SGAVideoContext::nb_tiles
int nb_tiles
Definition: sga.c:48
bytestream2_seek
static av_always_inline int bytestream2_seek(GetByteContext *g, int offset, int whence)
Definition: bytestream.h:212
AVFrame
This structure describes decoded (raw) audio or video data.
Definition: frame.h:389
AVPacket::data
uint8_t * data
Definition: packet.h:539
SGAVideoContext::tiles_w
int tiles_w
Definition: sga.c:49
FFCodec
Definition: codec_internal.h:127
SGAVideoContext::plus
int plus
Definition: sga.c:51
ff_set_dimensions
int ff_set_dimensions(AVCodecContext *s, int width, int height)
Check that the provided frame dimensions are valid and set them on the codec context.
Definition: utils.c:94
SGAVideoContext::tiledata_offset
int tiledata_offset
Definition: sga.c:38
lzss_decompress
static int lzss_decompress(AVCodecContext *avctx, GetByteContext *gb, uint8_t *dst, int dst_size, int shift, int plus)
Definition: sga.c:201
SGAVideoContext::tileindex_size
unsigned tileindex_size
Definition: sga.c:56
bytestream2_skip
static av_always_inline void bytestream2_skip(GetByteContext *g, unsigned int size)
Definition: bytestream.h:168
get_bits
static unsigned int get_bits(GetBitContext *s, int n)
Read 1-25 bits.
Definition: get_bits.h:335
FFCodec::p
AVCodec p
The public AVCodec.
Definition: codec_internal.h:131
SGAVideoContext::metadata_size
int metadata_size
Definition: sga.c:36
GetBitContext
Definition: get_bits.h:108
type
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf type
Definition: writing_filters.txt:86
SGAVideoContext::tiledata_size
int tiledata_size
Definition: sga.c:37
sga_decode_end
static av_cold int sga_decode_end(AVCodecContext *avctx)
Definition: sga.c:517
HAVE_TILEMAP
#define HAVE_TILEMAP
Definition: sga.c:31
decode_palmapdata
static int decode_palmapdata(AVCodecContext *avctx)
Definition: sga.c:251
AV_LOG_ERROR
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
Definition: log.h:209
decode_tiledata
static int decode_tiledata(AVCodecContext *avctx)
Definition: sga.c:277
av_cold
#define av_cold
Definition: attributes.h:90
init_get_bits8
static int init_get_bits8(GetBitContext *s, const uint8_t *buffer, int byte_size)
Initialize GetBitContext.
Definition: get_bits.h:545
SGAVideoContext::gb
GetByteContext gb
Definition: sga.c:34
AV_FRAME_FLAG_KEY
#define AV_FRAME_FLAG_KEY
A flag to mark frames that are keyframes.
Definition: frame.h:640
FF_CODEC_DECODE_CB
#define FF_CODEC_DECODE_CB(func)
Definition: codec_internal.h:311
s
#define s(width, name)
Definition: cbs_vp9.c:198
AV_CODEC_ID_SGA_VIDEO
@ AV_CODEC_ID_SGA_VIDEO
Definition: codec_id.h:314
GetByteContext::buffer
const uint8_t * buffer
Definition: bytestream.h:34
bits
uint8_t bits
Definition: vp3data.h:128
decode_index
static int decode_index(SGAVideoContext *s, AVFrame *frame)
Definition: sga.c:181
AV_LOG_DEBUG
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
Definition: log.h:230
decode.h
get_bits.h
CODEC_LONG_NAME
#define CODEC_LONG_NAME(str)
Definition: codec_internal.h:296
sizes
static const int sizes[][2]
Definition: img2dec.c:60
AVPALETTE_SIZE
#define AVPALETTE_SIZE
Definition: pixfmt.h:32
AV_PICTURE_TYPE_I
@ AV_PICTURE_TYPE_I
Intra.
Definition: avutil.h:279
get_bits1
static unsigned int get_bits1(GetBitContext *s)
Definition: get_bits.h:388
ff_sga_decoder
const FFCodec ff_sga_decoder
Definition: sga.c:530
bytestream2_get_buffer
static av_always_inline unsigned int bytestream2_get_buffer(GetByteContext *g, uint8_t *dst, unsigned int size)
Definition: bytestream.h:267
SGAVideoContext::nb_pal
int nb_pal
Definition: sga.c:47
SGAVideoContext::palmapindex_data
uint8_t * palmapindex_data
Definition: sga.c:57
index
int index
Definition: gxfenc.c:90
decode_palette
static int decode_palette(GetByteContext *gb, uint32_t *pal)
Definition: sga.c:68
bytestream2_get_bytes_left
static av_always_inline int bytestream2_get_bytes_left(GetByteContext *g)
Definition: bytestream.h:158
ff_get_buffer
int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Get a buffer for a frame.
Definition: decode.c:1697
init
int(* init)(AVBSFContext *ctx)
Definition: dts2pts.c:368
AV_CODEC_CAP_DR1
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
Definition: codec.h:52
AVPacket::size
int size
Definition: packet.h:540
codec_internal.h
shift
static int shift(int a, int b)
Definition: bonk.c:261
dst
uint8_t ptrdiff_t const uint8_t ptrdiff_t int intptr_t intptr_t int int16_t * dst
Definition: dsp.h:83
size
int size
Definition: twinvq_data.h:10344
sga_decode_frame
static int sga_decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt)
Definition: sga.c:313
header
static const uint8_t header[24]
Definition: sdr2.c:68
SGAVideoContext::flags
int flags
Definition: sga.c:46
offset
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
Definition: writing_filters.txt:86
PALDATA_FOLLOWS_TILEDATA
#define PALDATA_FOLLOWS_TILEDATA
Definition: sga.c:29
SGAVideoContext
Definition: sga.c:33
decode_index_palmap
static int decode_index_palmap(SGAVideoContext *s, AVFrame *frame)
Definition: sga.c:107
SGAVideoContext::tilemapdata_offset
int tilemapdata_offset
Definition: sga.c:40
SGAVideoContext::tilemapdata_size
int tilemapdata_size
Definition: sga.c:39
SGAVideoContext::palmapdata_size
int palmapdata_size
Definition: sga.c:44
i
#define i(width, name, range_min, range_max)
Definition: cbs_h2645.c:256
SGAVideoContext::swap
int swap
Definition: sga.c:52
common.h
av_assert1
#define av_assert1(cond)
assert() equivalent, that does not lie in speed critical code.
Definition: avassert.h:56
decode_index_tilemap
static int decode_index_tilemap(SGAVideoContext *s, AVFrame *frame)
Definition: sga.c:129
av_fast_padded_malloc
void av_fast_padded_malloc(void *ptr, unsigned int *size, size_t min_size)
Same behaviour av_fast_malloc but the buffer has additional AV_INPUT_BUFFER_PADDING_SIZE at the end w...
Definition: utils.c:52
AVCodec::name
const char * name
Name of the codec implementation.
Definition: codec.h:194
AVCodecContext::height
int height
Definition: avcodec.h:624
AVCodecContext::pix_fmt
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
Definition: avcodec.h:663
avcodec.h
SGAVideoContext::shift
int shift
Definition: sga.c:50
AV_PIX_FMT_PAL8
@ AV_PIX_FMT_PAL8
8 bits with AV_PIX_FMT_RGB32 palette
Definition: pixfmt.h:84
ret
ret
Definition: filter_design.txt:187
FFSWAP
#define FFSWAP(type, a, b)
Definition: macros.h:52
frame
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
Definition: filter_design.txt:264
SGAVideoContext::palmapdata_offset
int palmapdata_offset
Definition: sga.c:43
left
Tag MUST be and< 10hcoeff half pel interpolation filter coefficients, hcoeff[0] are the 2 middle coefficients[1] are the next outer ones and so on, resulting in a filter like:...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2] ... the sign of the coefficients is not explicitly stored but alternates after each coeff and coeff[0] is positive, so ...,+,-,+,-,+,+,-,+,-,+,... hcoeff[0] is not explicitly stored but found by subtracting the sum of all stored coefficients with signs from 32 hcoeff[0]=32 - hcoeff[1] - hcoeff[2] - ... a good choice for hcoeff and htaps is htaps=6 hcoeff={40,-10, 2} an alternative which requires more computations at both encoder and decoder side and may or may not be better is htaps=8 hcoeff={42,-14, 6,-2}ref_frames minimum of the number of available reference frames and max_ref_frames for example the first frame after a key frame always has ref_frames=1spatial_decomposition_type wavelet type 0 is a 9/7 symmetric compact integer wavelet 1 is a 5/3 symmetric compact integer wavelet others are reserved stored as delta from last, last is reset to 0 if always_reset||keyframeqlog quality(logarithmic quantizer scale) stored as delta from last, last is reset to 0 if always_reset||keyframemv_scale stored as delta from last, last is reset to 0 if always_reset||keyframe FIXME check that everything works fine if this changes between framesqbias dequantization bias stored as delta from last, last is reset to 0 if always_reset||keyframeblock_max_depth maximum depth of the block tree stored as delta from last, last is reset to 0 if always_reset||keyframequant_table quantization tableHighlevel bitstream structure:==============================--------------------------------------------|Header|--------------------------------------------|------------------------------------|||Block0||||split?||||yes no||||......... intra?||||:Block01 :yes no||||:Block02 :....... ..........||||:Block03 ::y DC ::ref index:||||:Block04 ::cb DC ::motion x :||||......... :cr DC ::motion y :||||....... ..........|||------------------------------------||------------------------------------|||Block1|||...|--------------------------------------------|------------ ------------ ------------|||Y subbands||Cb subbands||Cr subbands||||--- ---||--- ---||--- ---|||||LL0||HL0||||LL0||HL0||||LL0||HL0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||LH0||HH0||||LH0||HH0||||LH0||HH0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HL1||LH1||||HL1||LH1||||HL1||LH1|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HH1||HL2||||HH1||HL2||||HH1||HL2|||||...||...||...|||------------ ------------ ------------|--------------------------------------------Decoding process:=================------------|||Subbands|------------||||------------|Intra DC||||LL0 subband prediction ------------|\ Dequantization ------------------- \||Reference frames|\ IDWT|------- -------|Motion \|||Frame 0||Frame 1||Compensation . OBMC v -------|------- -------|--------------. \------> Frame n output Frame Frame<----------------------------------/|...|------------------- Range Coder:============Binary Range Coder:------------------- The implemented range coder is an adapted version based upon "Range encoding: an algorithm for removing redundancy from a digitised message." by G. N. N. Martin. The symbols encoded by the Snow range coder are bits(0|1). The associated probabilities are not fix but change depending on the symbol mix seen so far. bit seen|new state ---------+----------------------------------------------- 0|256 - state_transition_table[256 - old_state];1|state_transition_table[old_state];state_transition_table={ 0, 0, 0, 0, 0, 0, 0, 0, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 190, 191, 192, 194, 194, 195, 196, 197, 198, 199, 200, 201, 202, 202, 204, 205, 206, 207, 208, 209, 209, 210, 211, 212, 213, 215, 215, 216, 217, 218, 219, 220, 220, 222, 223, 224, 225, 226, 227, 227, 229, 229, 230, 231, 232, 234, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 248, 0, 0, 0, 0, 0, 0, 0};FIXME Range Coding of integers:------------------------- FIXME Neighboring Blocks:===================left and top are set to the respective blocks unless they are outside of the image in which case they are set to the Null block top-left is set to the top left block unless it is outside of the image in which case it is set to the left block if this block has no larger parent block or it is at the left side of its parent block and the top right block is not outside of the image then the top right block is used for top-right else the top-left block is used Null block y, cb, cr are 128 level, ref, mx and my are 0 Motion Vector Prediction:=========================1. the motion vectors of all the neighboring blocks are scaled to compensate for the difference of reference frames scaled_mv=(mv *(256 *(current_reference+1)/(mv.reference+1))+128)> the median of the scaled left
Definition: snow.txt:386
AVCodecContext
main external API structure.
Definition: avcodec.h:451
SGAVideoContext::paldata_size
int paldata_size
Definition: sga.c:41
FF_DISABLE_DEPRECATION_WARNINGS
#define FF_DISABLE_DEPRECATION_WARNINGS
Definition: internal.h:72
AVMEDIA_TYPE_VIDEO
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:201
mem.h
AVPacket
This structure stores compressed data.
Definition: packet.h:516
AVCodecContext::priv_data
void * priv_data
Definition: avcodec.h:478
av_freep
#define av_freep(p)
Definition: tableprint_vlc.h:34
AVCodecContext::width
int width
picture width / height.
Definition: avcodec.h:624
bytestream.h
bytestream2_init
static av_always_inline void bytestream2_init(GetByteContext *g, const uint8_t *buf, int buf_size)
Definition: bytestream.h:137
av_log
#define av_log(a,...)
Definition: tableprint_vlc.h:27
AVERROR_INVALIDDATA
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
Definition: error.h:61
SGAVideoContext::tiles_h
int tiles_h
Definition: sga.c:49
sga_decode_init
static av_cold int sga_decode_init(AVCodecContext *avctx)
Definition: sga.c:62
SGAVideoContext::uncompressed
uint8_t uncompressed[65536]
Definition: sga.c:59
src
#define src
Definition: vp8dsp.c:248
AV_RB16
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_WL16 uint64_t_TMPL AV_WB64 unsigned int_TMPL AV_WB32 unsigned int_TMPL AV_WB24 unsigned int_TMPL AV_RB16
Definition: bytestream.h:98
SGAVideoContext::paldata_offset
int paldata_offset
Definition: sga.c:42