}
if (avcodec_check_dimensions(avctx, avctx->width, avctx->height) < 0)
return -1;
- avctx->pix_fmt = PIX_FMT_YUV422P16;
+ avctx->pix_fmt = PIX_FMT_YUV422P16;
avctx->bits_per_raw_sample = 10;
- avctx->coded_frame = avcodec_alloc_frame();
+ avctx->coded_frame = avcodec_alloc_frame();
return 0;
}
-static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPacket *avpkt)
+static int decode_frame(AVCodecContext *avctx, void *data, int *data_size,
+ AVPacket *avpkt)
{
int h, w;
AVFrame *pic = avctx->coded_frame;
pic->pict_type = FF_I_TYPE;
pic->key_frame = 1;
-#define READ_PIXELS(a, b, c) \
- do { \
- val = le2me_32(*src++); \
+#define READ_PIXELS(a, b, c) \
+ do { \
+ val = le2me_32(*src++); \
*a++ = val << 6; \
*b++ = (val >> 4) & 0xFFC0; \
*c++ = (val >> 14) & 0xFFC0; \
}
psrc += stride;
- y += pic->linesize[0]/2 - avctx->width;
- u += pic->linesize[1]/2 - avctx->width/2;
- v += pic->linesize[2]/2 - avctx->width/2;
+ y += pic->linesize[0] / 2 - avctx->width;
+ u += pic->linesize[1] / 2 - avctx->width / 2;
+ v += pic->linesize[2] / 2 - avctx->width / 2;
}
*data_size = sizeof(AVFrame);
return 0;
}
-static int encode_frame(AVCodecContext *avctx, unsigned char *buf, int buf_size, void *data)
+static int encode_frame(AVCodecContext *avctx, unsigned char *buf,
+ int buf_size, void *data)
{
const AVFrame *pic = data;
int aligned_width = ((avctx->width + 47) / 48) * 48;
pdst += stride;
memset(p, 0, pdst - p);
p = pdst;
- y += pic->linesize[0]/2 - avctx->width;
- u += pic->linesize[1]/2 - avctx->width/2;
- v += pic->linesize[2]/2 - avctx->width/2;
+ y += pic->linesize[0] / 2 - avctx->width;
+ u += pic->linesize[1] / 2 - avctx->width / 2;
+ v += pic->linesize[2] / 2 - avctx->width / 2;
}
return p - buf;