35 #define KMVC_KEYFRAME 0x80 36 #define KMVC_PALETTE 0x40 37 #define KMVC_METHOD 0x0F 38 #define MAX_PALSIZE 256 59 #define BLK(data, x, y) data[av_clip((x) + (y) * 320, 0, 320 * 200 -1)] 61 #define kmvc_init_getbits(bb, g) bb.bits = 7; bb.bitbuf = bytestream2_get_byte(g); 63 #define kmvc_getbit(bb, g, res) {\ 65 if (bb.bitbuf & (1 << bb.bits)) res = 1; \ 68 bb.bitbuf = bytestream2_get_byte(g); \ 79 int l0x, l1x, l0y, l1y;
84 for (by = 0; by <
h; by += 8)
85 for (bx = 0; bx <
w; bx += 8) {
92 val = bytestream2_get_byte(&ctx->
g);
93 for (i = 0; i < 64; i++)
94 BLK(ctx->
cur, bx + (i & 0x7), by + (i >> 3)) =
val;
96 for (i = 0; i < 4; i++) {
97 l0x = bx + (i & 1) * 4;
98 l0y = by + (i & 2) * 2;
103 val = bytestream2_get_byte(&ctx->
g);
104 for (j = 0; j < 16; j++)
105 BLK(ctx->
cur, l0x + (j & 3), l0y + (j >> 2)) =
val;
107 val = bytestream2_get_byte(&ctx->
g);
110 if ((l0x-mx) + 320*(l0y-my) < 0 || (l0x-mx) + 320*(l0y-my) > 320*197 - 4) {
114 for (j = 0; j < 16; j++)
115 BLK(ctx->
cur, l0x + (j & 3), l0y + (j >> 2)) =
116 BLK(ctx->
cur, l0x + (j & 3) - mx, l0y + (j >> 2) - my);
119 for (j = 0; j < 4; j++) {
120 l1x = l0x + (j & 1) * 2;
126 val = bytestream2_get_byte(&ctx->
g);
132 val = bytestream2_get_byte(&ctx->
g);
135 if ((l1x-mx) + 320*(l1y-my) < 0 || (l1x-mx) + 320*(l1y-my) > 320*199 - 2) {
139 BLK(ctx->
cur, l1x, l1y) =
BLK(ctx->
cur, l1x - mx, l1y - my);
140 BLK(ctx->
cur, l1x + 1, l1y) =
141 BLK(ctx->
cur, l1x + 1 - mx, l1y - my);
142 BLK(ctx->
cur, l1x, l1y + 1) =
143 BLK(ctx->
cur, l1x - mx, l1y + 1 - my);
144 BLK(ctx->
cur, l1x + 1, l1y + 1) =
145 BLK(ctx->
cur, l1x + 1 - mx, l1y + 1 - my);
148 BLK(ctx->
cur, l1x, l1y) = bytestream2_get_byte(&ctx->
g);
149 BLK(ctx->
cur, l1x + 1, l1y) = bytestream2_get_byte(&ctx->
g);
150 BLK(ctx->
cur, l1x, l1y + 1) = bytestream2_get_byte(&ctx->
g);
151 BLK(ctx->
cur, l1x + 1, l1y + 1) = bytestream2_get_byte(&ctx->
g);
168 int l0x, l1x, l0y, l1y;
173 for (by = 0; by <
h; by += 8)
174 for (bx = 0; bx <
w; bx += 8) {
183 val = bytestream2_get_byte(&ctx->
g);
184 for (i = 0; i < 64; i++)
185 BLK(ctx->
cur, bx + (i & 0x7), by + (i >> 3)) =
val;
187 for (i = 0; i < 64; i++)
188 BLK(ctx->
cur, bx + (i & 0x7), by + (i >> 3)) =
189 BLK(ctx->
prev, bx + (i & 0x7), by + (i >> 3));
196 for (i = 0; i < 4; i++) {
197 l0x = bx + (i & 1) * 4;
198 l0y = by + (i & 2) * 2;
203 val = bytestream2_get_byte(&ctx->
g);
204 for (j = 0; j < 16; j++)
205 BLK(ctx->
cur, l0x + (j & 3), l0y + (j >> 2)) =
val;
207 val = bytestream2_get_byte(&ctx->
g);
208 mx = (val & 0xF) - 8;
210 if ((l0x+mx) + 320*(l0y+my) < 0 || (l0x+mx) + 320*(l0y+my) > 320*197 - 4) {
214 for (j = 0; j < 16; j++)
215 BLK(ctx->
cur, l0x + (j & 3), l0y + (j >> 2)) =
216 BLK(ctx->
prev, l0x + (j & 3) + mx, l0y + (j >> 2) + my);
219 for (j = 0; j < 4; j++) {
220 l1x = l0x + (j & 1) * 2;
226 val = bytestream2_get_byte(&ctx->
g);
232 val = bytestream2_get_byte(&ctx->
g);
233 mx = (val & 0xF) - 8;
235 if ((l1x+mx) + 320*(l1y+my) < 0 || (l1x+mx) + 320*(l1y+my) > 320*199 - 2) {
239 BLK(ctx->
cur, l1x, l1y) =
BLK(ctx->
prev, l1x + mx, l1y + my);
240 BLK(ctx->
cur, l1x + 1, l1y) =
241 BLK(ctx->
prev, l1x + 1 + mx, l1y + my);
242 BLK(ctx->
cur, l1x, l1y + 1) =
243 BLK(ctx->
prev, l1x + mx, l1y + 1 + my);
244 BLK(ctx->
cur, l1x + 1, l1y + 1) =
245 BLK(ctx->
prev, l1x + 1 + mx, l1y + 1 + my);
248 BLK(ctx->
cur, l1x, l1y) = bytestream2_get_byte(&ctx->
g);
249 BLK(ctx->
cur, l1x + 1, l1y) = bytestream2_get_byte(&ctx->
g);
250 BLK(ctx->
cur, l1x, l1y + 1) = bytestream2_get_byte(&ctx->
g);
251 BLK(ctx->
cur, l1x + 1, l1y + 1) = bytestream2_get_byte(&ctx->
g);
279 header = bytestream2_get_byte(&ctx->
g);
282 if (bytestream2_peek_byte(&ctx->
g) == 127) {
284 for (i = 0; i < 127; i++) {
285 ctx->
pal[i + (header & 0x81)] = 0xFFU << 24 | bytestream2_get_be24(&ctx->
g);
302 for (i = 1; i <= ctx->
palsize; i++) {
303 ctx->
pal[
i] = 0xFF
U << 24 | bytestream2_get_be24(&ctx->
g);
320 memcpy(frame->
data[1], ctx->
pal, 1024);
322 blocksize = bytestream2_get_byte(&ctx->
g);
324 if (blocksize != 8 && blocksize != 127) {
328 memset(ctx->
cur, 0, 320 * 200);
332 memcpy(ctx->
cur, ctx->
prev, 320 * 200);
345 out = frame->
data[0];
347 for (i = 0; i < avctx->
height; i++) {
348 memcpy(out, src, avctx->
width);
388 for (i = 0; i < 256; i++) {
389 c->
pal[
i] = 0xFF
U << 24 | i * 0x10101;
394 "Extradata missing, decoding may not work properly...\n");
407 for (i = 0; i < 256; i++) {
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
This structure describes decoded (raw) audio or video data.
ptrdiff_t const GLvoid * data
#define AV_LOG_WARNING
Something somehow does not look correct.
static av_cold int init(AVCodecContext *avctx)
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
static av_always_inline void bytestream2_init(GetByteContext *g, const uint8_t *buf, int buf_size)
static int kmvc_decode_inter_8x8(KmvcContext *ctx, int w, int h)
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_RL16
static void decode(AVCodecContext *dec_ctx, AVPacket *pkt, AVFrame *frame, FILE *outfile)
8 bits with AV_PIX_FMT_RGB32 palette
Undefined Behavior In the C some operations are like signed integer dereferencing freed accessing outside allocated Undefined Behavior must not occur in a C it is not safe even if the output of undefined operations is unused The unsafety may seem nit picking but Optimizing compilers have in fact optimized code on the assumption that no undefined Behavior occurs Optimizing code based on wrong assumptions can and has in some cases lead to effects beyond the output of computations The signed integer overflow problem in speed critical code Code which is highly optimized and works with signed integers sometimes has the problem that often the output of the computation does not c
uint8_t * extradata
some codecs need / can use extradata like Huffman tables.
static const uint8_t header[24]
static av_always_inline int bytestream2_get_bytes_left(GetByteContext *g)
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
An AV_PKT_DATA_PALETTE side data packet contains exactly AVPALETTE_SIZE bytes worth of palette...
static int kmvc_decode_intra_8x8(KmvcContext *ctx, int w, int h)
static av_always_inline void bytestream2_skip(GetByteContext *g, unsigned int size)
uint8_t * av_packet_get_side_data(const AVPacket *pkt, enum AVPacketSideDataType type, int *size)
Get side information from packet.
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
const char * name
Name of the codec implementation.
#define kmvc_init_getbits(bb, g)
enum AVPictureType pict_type
Picture type of the frame.
int width
picture width / height.
#define kmvc_getbit(bb, g, res)
static av_cold int decode_init(AVCodecContext *avctx)
static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame, AVPacket *avpkt)
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
Libavcodec external API header.
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
main external API structure.
int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Get a buffer for a frame.
int palette_has_changed
Tell user application that palette has changed from previous frame.
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
common internal api header.
common internal and external API header
int key_frame
1 -> keyframe, 0-> not
static av_always_inline int bytestream2_seek(GetByteContext *g, int offset, int whence)
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later.That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another.Frame references ownership and permissions
uint32_t pal[MAX_PALSIZE]
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_RL32
static double val(void *priv, double ch)
This structure stores compressed data.
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() for allocating buffers and supports custom allocators.