| Index: libquicktime-1.2.4/plugins/ffmpeg/video.c |
| =================================================================== |
| --- libquicktime-1.2.4.orig/plugins/ffmpeg/video.c |
| +++ libquicktime-1.2.4/plugins/ffmpeg/video.c |
| @@ -37,10 +37,10 @@ |
| #endif |
| |
| |
| -#ifdef PIX_FMT_YUV422P10 |
| -#define PIX_FMT_YUV422P10_OR_DUMMY PIX_FMT_YUV422P10 |
| +#ifdef AV_PIX_FMT_YUV422P10 |
| +#define AV_PIX_FMT_YUV422P10_OR_DUMMY AV_PIX_FMT_YUV422P10 |
| #else |
| -#define PIX_FMT_YUV422P10_OR_DUMMY -1234 |
| +#define AV_PIX_FMT_YUV422P10_OR_DUMMY -1234 |
| #endif |
| |
| #if LIBAVCODEC_VERSION_INT >= ((54<<16)|(1<<8)|0) |
| @@ -90,9 +90,9 @@ typedef struct |
| int imx_bitrate; |
| int imx_strip_vbi; |
| |
| - /* In some cases FFMpeg would report something like PIX_FMT_YUV422P, while |
| - we would like to treat it as PIX_FMT_YUVJ422P. It's only used for decoding */ |
| - enum PixelFormat reinterpret_pix_fmt; |
| + /* In some cases FFMpeg would report something like AV_PIX_FMT_YUV422P, while |
| + we would like to treat it as AV_PIX_FMT_YUVJ422P. It's only used for decoding */ |
| + enum AVPixelFormat reinterpret_pix_fmt; |
| |
| int is_imx; |
| int y_offset; |
| @@ -137,42 +137,42 @@ typedef struct |
| |
| static const struct |
| { |
| - enum PixelFormat ffmpeg_id; |
| + enum AVPixelFormat ffmpeg_id; |
| int lqt_id; |
| int exact; |
| } |
| colormodels[] = |
| { |
| - { PIX_FMT_YUV420P, BC_YUV420P, 1 }, ///< Planar YUV 4:2:0 (1 Cr & Cb sample per 2x2 Y samples) |
| + { AV_PIX_FMT_YUV420P, BC_YUV420P, 1 }, ///< Planar YUV 4:2:0 (1 Cr & Cb sample per 2x2 Y samples) |
| #if LIBAVUTIL_VERSION_INT < (50<<16) |
| - { PIX_FMT_YUV422, BC_YUV422, 1 }, |
| + { AV_PIX_FMT_YUV422, BC_YUV422, 1 }, |
| #else |
| - { PIX_FMT_YUYV422, BC_YUV422, 1 }, |
| + { AV_PIX_FMT_YUYV422, BC_YUV422, 1 }, |
| #endif |
| - { PIX_FMT_RGB24, BC_RGB888, 1 }, ///< Packed pixel, 3 bytes per pixel, RGBRGB... |
| - { PIX_FMT_BGR24, BC_BGR888, 1 }, ///< Packed pixel, 3 bytes per pixel, BGRBGR... |
| - { PIX_FMT_YUV422P, BC_YUV422P, 1 }, ///< Planar YUV 4:2:2 (1 Cr & Cb sample per 2x1 Y samples) |
| - { PIX_FMT_YUV444P, BC_YUV444P, 1 }, ///< Planar YUV 4:4:4 (1 Cr & Cb sample per 1x1 Y samples) |
| - { PIX_FMT_YUV411P, BC_YUV411P, 1 }, ///< Planar YUV 4:1:1 (1 Cr & Cb sample per 4x1 Y samples) |
| - { PIX_FMT_YUV422P16, BC_YUV422P16, 1 }, ///< Planar 16 bit YUV 4:2:2 (1 Cr & Cb sample per 2x1 Y samples) |
| -#ifdef PIX_FMT_YUV422P10 |
| - { PIX_FMT_YUV422P10, BC_YUV422P10, 1 }, ///< 10 bit samples in uint16_t containers, planar 4:2:2 |
| -#endif |
| - { PIX_FMT_RGB565, BC_RGB565, 1 }, ///< always stored in cpu endianness |
| - { PIX_FMT_YUVJ420P, BC_YUVJ420P, 1 }, ///< Planar YUV 4:2:0 full scale (jpeg) |
| - { PIX_FMT_YUVJ422P, BC_YUVJ422P, 1 }, ///< Planar YUV 4:2:2 full scale (jpeg) |
| - { PIX_FMT_YUVJ444P, BC_YUVJ444P, 1 }, ///< Planar YUV 4:4:4 full scale (jpeg) |
| + { AV_PIX_FMT_RGB24, BC_RGB888, 1 }, ///< Packed pixel, 3 bytes per pixel, RGBRGB... |
| + { AV_PIX_FMT_BGR24, BC_BGR888, 1 }, ///< Packed pixel, 3 bytes per pixel, BGRBGR... |
| + { AV_PIX_FMT_YUV422P, BC_YUV422P, 1 }, ///< Planar YUV 4:2:2 (1 Cr & Cb sample per 2x1 Y samples) |
| + { AV_PIX_FMT_YUV444P, BC_YUV444P, 1 }, ///< Planar YUV 4:4:4 (1 Cr & Cb sample per 1x1 Y samples) |
| + { AV_PIX_FMT_YUV411P, BC_YUV411P, 1 }, ///< Planar YUV 4:1:1 (1 Cr & Cb sample per 4x1 Y samples) |
| + { AV_PIX_FMT_YUV422P16, BC_YUV422P16, 1 }, ///< Planar 16 bit YUV 4:2:2 (1 Cr & Cb sample per 2x1 Y samples) |
| +#ifdef AV_PIX_FMT_YUV422P10 |
| + { AV_PIX_FMT_YUV422P10, BC_YUV422P10, 1 }, ///< 10 bit samples in uint16_t containers, planar 4:2:2 |
| +#endif |
| + { AV_PIX_FMT_RGB565, BC_RGB565, 1 }, ///< always stored in cpu endianness |
| + { AV_PIX_FMT_YUVJ420P, BC_YUVJ420P, 1 }, ///< Planar YUV 4:2:0 full scale (jpeg) |
| + { AV_PIX_FMT_YUVJ422P, BC_YUVJ422P, 1 }, ///< Planar YUV 4:2:2 full scale (jpeg) |
| + { AV_PIX_FMT_YUVJ444P, BC_YUVJ444P, 1 }, ///< Planar YUV 4:4:4 full scale (jpeg) |
| #if LIBAVUTIL_VERSION_INT < (50<<16) |
| - { PIX_FMT_RGBA32, BC_RGBA8888, 0 }, ///< Packed pixel, 4 bytes per pixel, BGRABGRA... |
| + { AV_PIX_FMT_RGBA32, BC_RGBA8888, 0 }, ///< Packed pixel, 4 bytes per pixel, BGRABGRA... |
| #else |
| - { PIX_FMT_RGB32, BC_RGBA8888, 0 }, ///< Packed pixel, 4 bytes per pixel, BGRABGRA... |
| + { AV_PIX_FMT_RGB32, BC_RGBA8888, 0 }, ///< Packed pixel, 4 bytes per pixel, BGRABGRA... |
| #endif |
| - { PIX_FMT_RGB555, BC_RGB888, 0 }, ///< always stored in cpu endianness, most significant bit to 1 |
| - { PIX_FMT_GRAY8, BC_RGB888, 0 }, |
| - { PIX_FMT_MONOWHITE, BC_RGB888, 0 }, ///< 0 is white |
| - { PIX_FMT_MONOBLACK, BC_RGB888, 0 }, ///< 0 is black |
| - { PIX_FMT_PAL8, BC_RGB888, 0 }, ///< 8 bit with RGBA palette |
| - { PIX_FMT_YUV410P, BC_YUV420P, 0 }, ///< Planar YUV 4:1:0 (1 Cr & Cb sample per 4x4 Y samples) |
| + { AV_PIX_FMT_RGB555, BC_RGB888, 0 }, ///< always stored in cpu endianness, most significant bit to 1 |
| + { AV_PIX_FMT_GRAY8, BC_RGB888, 0 }, |
| + { AV_PIX_FMT_MONOWHITE, BC_RGB888, 0 }, ///< 0 is white |
| + { AV_PIX_FMT_MONOBLACK, BC_RGB888, 0 }, ///< 0 is black |
| + { AV_PIX_FMT_PAL8, BC_RGB888, 0 }, ///< 8 bit with RGBA palette |
| + { AV_PIX_FMT_YUV410P, BC_YUV420P, 0 }, ///< Planar YUV 4:1:0 (1 Cr & Cb sample per 4x4 Y samples) |
| }; |
| |
| static const struct |
| @@ -343,16 +343,16 @@ static int lqt_tenbit_dnxhd_supported(AV |
| if (!codec->pix_fmts) |
| return 0; |
| |
| - for (i = 0; codec->pix_fmts[i] != PIX_FMT_NONE; ++i) |
| + for (i = 0; codec->pix_fmts[i] != AV_PIX_FMT_NONE; ++i) |
| { |
| - if (codec->pix_fmts[i] == PIX_FMT_YUV422P10_OR_DUMMY) |
| + if (codec->pix_fmts[i] == AV_PIX_FMT_YUV422P10_OR_DUMMY) |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| -static enum PixelFormat lqt_ffmpeg_get_ffmpeg_colormodel(int id) |
| +static enum AVPixelFormat lqt_ffmpeg_get_ffmpeg_colormodel(int id) |
| { |
| int i; |
| |
| @@ -361,10 +361,10 @@ static enum PixelFormat lqt_ffmpeg_get_f |
| if(colormodels[i].lqt_id == id) |
| return colormodels[i].ffmpeg_id; |
| } |
| - return PIX_FMT_NB; |
| + return AV_PIX_FMT_NB; |
| } |
| |
| -static int lqt_ffmpeg_get_lqt_colormodel(enum PixelFormat id, int * exact) |
| +static int lqt_ffmpeg_get_lqt_colormodel(enum AVPixelFormat id, int * exact) |
| { |
| int i; |
| |
| @@ -402,24 +402,24 @@ static void lqt_ffmpeg_setup_decoding_co |
| /* First we try codec-specific colormodel matching. */ |
| if(codec->decoder->id == AV_CODEC_ID_DNXHD) |
| { |
| - /* FFMpeg supports PIX_FMT_YUV422P and PIX_FMT_YUV422P10 for DNxHD, which |
| - we sometimes interpret as PIX_FMT_YUVJ422P and PIX_FMT_YUVJ422P10. */ |
| - if (codec->avctx->pix_fmt == PIX_FMT_YUV422P || codec->avctx->pix_fmt == PIX_FMT_YUV422P10_OR_DUMMY) |
| + /* FFMpeg supports AV_PIX_FMT_YUV422P and AV_PIX_FMT_YUV422P10 for DNxHD, which |
| + we sometimes interpret as AV_PIX_FMT_YUVJ422P and AV_PIX_FMT_YUVJ422P10. */ |
| + if (codec->avctx->pix_fmt == AV_PIX_FMT_YUV422P || codec->avctx->pix_fmt == AV_PIX_FMT_YUV422P10_OR_DUMMY) |
| { |
| - int p10 = (codec->avctx->pix_fmt == PIX_FMT_YUV422P10_OR_DUMMY); |
| + int p10 = (codec->avctx->pix_fmt == AV_PIX_FMT_YUV422P10_OR_DUMMY); |
| *exact = 1; |
| if (lqt_ffmpeg_get_avid_yuv_range(vtrack->track) == AVID_FULL_YUV_RANGE) |
| { |
| vtrack->stream_cmodel = p10 ? BC_YUVJ422P10 : BC_YUVJ422P; |
| - codec->reinterpret_pix_fmt = p10 ? PIX_FMT_YUV422P10_OR_DUMMY : PIX_FMT_YUVJ422P; |
| - // Note: reinterpret_pix_fmt should really be PIX_FMT_YUVJ422P10, except |
| + codec->reinterpret_pix_fmt = p10 ? AV_PIX_FMT_YUV422P10_OR_DUMMY : AV_PIX_FMT_YUVJ422P; |
| + // Note: reinterpret_pix_fmt should really be AV_PIX_FMT_YUVJ422P10, except |
| // there is no such colormodel in FFMpeg. Fortunately, it's not a problem |
| // in this case, as reinterpret_pix_fmt is only used when *exact == 0. |
| } |
| else |
| { |
| vtrack->stream_cmodel = p10 ? BC_YUV422P10 : BC_YUV422P; |
| - codec->reinterpret_pix_fmt = p10 ? PIX_FMT_YUV422P10_OR_DUMMY : PIX_FMT_YUV422P; |
| + codec->reinterpret_pix_fmt = p10 ? AV_PIX_FMT_YUV422P10_OR_DUMMY : AV_PIX_FMT_YUV422P; |
| } |
| return; |
| } |
| @@ -440,14 +440,14 @@ static void lqt_ffmpeg_setup_encoding_co |
| |
| if (codec->encoder->id == AV_CODEC_ID_DNXHD) |
| { |
| - /* FFMpeg's DNxHD encoder only supports PIX_FMT_YUV422P and PIX_FMT_YUV422P10 |
| - and doesn't know anything about PIX_FMT_YUVJ422P and PIX_FMT_YUVJ422P10 |
| + /* FFMpeg's DNxHD encoder only supports AV_PIX_FMT_YUV422P and AV_PIX_FMT_YUV422P10 |
| + and doesn't know anything about AV_PIX_FMT_YUVJ422P and AV_PIX_FMT_YUVJ422P10 |
| (in fact, the latter doesn't even exist) */ |
| - codec->avctx->pix_fmt = PIX_FMT_YUV422P; |
| + codec->avctx->pix_fmt = AV_PIX_FMT_YUV422P; |
| if (vtrack->stream_cmodel == BC_YUV422P10 || vtrack->stream_cmodel == BC_YUVJ422P10) |
| { |
| if (lqt_tenbit_dnxhd_supported(codec->encoder)) |
| - codec->avctx->pix_fmt = PIX_FMT_YUV422P10_OR_DUMMY; |
| + codec->avctx->pix_fmt = AV_PIX_FMT_YUV422P10_OR_DUMMY; |
| } |
| } |
| } |
| @@ -458,7 +458,7 @@ static void lqt_ffmpeg_setup_encoding_co |
| /* From avcodec.h: */ |
| |
| /* |
| - * PIX_FMT_RGBA32 is handled in an endian-specific manner. A RGBA |
| + * AV_PIX_FMT_RGBA32 is handled in an endian-specific manner. A RGBA |
| * color is put together as: |
| * (A << 24) | (R << 16) | (G << 8) | B |
| * This is stored as BGRA on little endian CPU architectures and ARGB on |
| @@ -530,7 +530,7 @@ static void convert_rgba_to_argb(uint8_t |
| */ |
| |
| static void convert_image_decode(quicktime_ffmpeg_video_codec_t *codec, |
| - AVFrame * in_frame, enum PixelFormat in_format, |
| + AVFrame * in_frame, enum AVPixelFormat in_format, |
| unsigned char ** out_frame, int out_format, |
| int width, int height, int row_span, int row_span_uv) |
| { |
| @@ -547,9 +547,9 @@ static void convert_image_decode(quickti |
| * RGBA format like in ffmpeg?? |
| */ |
| #if LIBAVUTIL_VERSION_INT < (50<<16) |
| - if((in_format == PIX_FMT_RGBA32) && (out_format == BC_RGBA8888)) |
| + if((in_format == AV_PIX_FMT_RGBA32) && (out_format == BC_RGBA8888)) |
| #else |
| - if((in_format == PIX_FMT_RGB32) && (out_format == BC_RGBA8888)) |
| + if((in_format == AV_PIX_FMT_RGB32) && (out_format == BC_RGBA8888)) |
| #endif |
| { |
| convert_image_decode_rgba(in_frame, out_frame, width, height, codec->y_offset); |
| @@ -829,7 +829,7 @@ static int lqt_ffmpeg_decode_video(quick |
| if(avcodec_open2(codec->avctx, codec->decoder, NULL) != 0) |
| return -1; |
| #endif |
| - codec->frame = avcodec_alloc_frame(); |
| + codec->frame = av_frame_alloc(); |
| vtrack->stream_cmodel = LQT_COLORMODEL_NONE; |
| codec->initialized = 1; |
| } |
| @@ -929,10 +929,10 @@ static int lqt_ffmpeg_decode_video(quick |
| #ifdef HAVE_LIBSWSCALE |
| |
| #if LIBAVUTIL_VERSION_INT < (50<<16) |
| - if(!((codec->avctx->pix_fmt == PIX_FMT_RGBA32) && |
| + if(!((codec->avctx->pix_fmt == AV_PIX_FMT_RGBA32) && |
| (vtrack->stream_cmodel == BC_RGBA8888))) |
| #else |
| - if(!((codec->avctx->pix_fmt == PIX_FMT_RGB32) && |
| + if(!((codec->avctx->pix_fmt == AV_PIX_FMT_RGB32) && |
| (vtrack->stream_cmodel == BC_RGBA8888))) |
| #endif |
| { |
| @@ -1318,7 +1318,7 @@ static int lqt_ffmpeg_encode_video(quick |
| |
| if(!codec->initialized) |
| { |
| - codec->frame = avcodec_alloc_frame(); |
| + codec->frame = av_frame_alloc(); |
| |
| /* time_base is 1/framerate for constant framerate */ |
| |
| @@ -1396,9 +1396,9 @@ static int lqt_ffmpeg_encode_video(quick |
| if(vtrack->stream_cmodel == BC_RGBA8888) |
| { |
| /* Libquicktime doesn't natively support a color model equivalent |
| - to PIX_FMT_ARGB, which is required for QTRLE with alpha channel. |
| + to AV_PIX_FMT_ARGB, which is required for QTRLE with alpha channel. |
| So, we use BC_RGBA8888 and do ad hoc conversion below. */ |
| - codec->avctx->pix_fmt = PIX_FMT_ARGB; |
| + codec->avctx->pix_fmt = AV_PIX_FMT_ARGB; |
| vtrack->track->mdia.minf.stbl.stsd.table[0].depth = 32; |
| } |
| } |
| @@ -1467,7 +1467,7 @@ static int lqt_ffmpeg_encode_video(quick |
| } |
| // codec->lqt_colormodel = ffmepg_2_lqt(codec->com.ffcodec_enc); |
| |
| - if(codec->y_offset != 0 || codec->avctx->pix_fmt == PIX_FMT_ARGB) |
| + if(codec->y_offset != 0 || codec->avctx->pix_fmt == AV_PIX_FMT_ARGB) |
| { |
| if(!codec->tmp_rows) |
| { |
| @@ -1492,7 +1492,7 @@ static int lqt_ffmpeg_encode_video(quick |
| vtrack->stream_cmodel, |
| 0, 0, 0, codec->y_offset); |
| } |
| - else if(codec->avctx->pix_fmt == PIX_FMT_ARGB) |
| + else if(codec->avctx->pix_fmt == AV_PIX_FMT_ARGB) |
| { |
| convert_rgba_to_argb(row_pointers[0], vtrack->stream_row_span, |
| codec->tmp_rows[0], codec->tmp_row_span, |
| Index: libquicktime-1.2.4/plugins/ffmpeg/audio.c |
| =================================================================== |
| --- libquicktime-1.2.4.orig/plugins/ffmpeg/audio.c |
| +++ libquicktime-1.2.4/plugins/ffmpeg/audio.c |
| @@ -1266,7 +1266,7 @@ static int lqt_ffmpeg_encode_audio(quick |
| pkt.data = codec->chunk_buffer; |
| pkt.size = codec->chunk_buffer_alloc; |
| |
| - avcodec_get_frame_defaults(&f); |
| + av_frame_unref(&f); |
| f.nb_samples = codec->avctx->frame_size; |
| |
| avcodec_fill_audio_frame(&f, channels, codec->avctx->sample_fmt, |