--- OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegDecoderAudio.cpp~ 2015-06-01 15:40:20.000000000 +0200 +++ OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegDecoderAudio.cpp 2016-03-21 11:37:09.380644409 +0100 @@ -226,9 +226,12 @@ // Open codec if (avcodec_open2(m_context, p_codec, NULL) < 0) throw std::runtime_error("avcodec_open() failed"); - +#if LIBAVFORMAT_VERSION_INT < AV_VERSION_INT(55, 0, 0) m_context->get_buffer = avcodec_default_get_buffer; m_context->release_buffer = avcodec_default_release_buffer; +#else + m_context->get_buffer2 = avcodec_default_get_buffer2; +#endif } --- OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegDecoderVideo.cpp.orig 2016-03-21 11:37:53.888177241 +0100 +++ OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegDecoderVideo.cpp 2016-03-21 11:54:04.889892930 +0100 @@ -71,7 +71,7 @@ findAspectRatio(); // Find out whether we support Alpha channel - m_alpha_channel = (m_context->pix_fmt == PIX_FMT_YUVA420P); + m_alpha_channel = (m_context->pix_fmt == AV_PIX_FMT_YUVA420P); // Find out the framerate #if LIBAVCODEC_VERSION_MAJOR >= 56 @@ -95,20 +95,24 @@ throw std::runtime_error("avcodec_open() failed"); // Allocate video frame - m_frame.reset(avcodec_alloc_frame()); + m_frame.reset(av_frame_alloc()); // Allocate converted RGB frame - m_frame_rgba.reset(avcodec_alloc_frame()); - m_buffer_rgba[0].resize(avpicture_get_size(PIX_FMT_RGB24, width(), height())); + m_frame_rgba.reset(av_frame_alloc()); + m_buffer_rgba[0].resize(avpicture_get_size(AV_PIX_FMT_RGB24, width(), height())); m_buffer_rgba[1].resize(m_buffer_rgba[0].size()); // Assign appropriate parts of the buffer to image planes in m_frame_rgba - avpicture_fill((AVPicture *) (m_frame_rgba).get(), &(m_buffer_rgba[0])[0], PIX_FMT_RGB24, width(), height()); + avpicture_fill((AVPicture *) (m_frame_rgba).get(), &(m_buffer_rgba[0])[0], AV_PIX_FMT_RGB24, width(), height()); // Override get_buffer()/release_buffer() from codec context in order to retrieve the PTS of each frame. m_context->opaque = this; +#if LIBAVFORMAT_VERSION_INT < AV_VERSION_INT(55, 0, 0) m_context->get_buffer = getBuffer; m_context->release_buffer = releaseBuffer; +#else + m_context->get_buffer2 = getBuffer; +#endif } @@ -267,8 +271,8 @@ #ifdef USE_SWSCALE if (m_swscale_ctx==0) { - m_swscale_ctx = sws_getContext(src_width, src_height, (PixelFormat) src_pix_fmt, - src_width, src_height, (PixelFormat) dst_pix_fmt, + m_swscale_ctx = sws_getContext(src_width, src_height, (AVPixelFormat) src_pix_fmt, + src_width, src_height, (AVPixelFormat) dst_pix_fmt, /*SWS_BILINEAR*/ SWS_BICUBIC, NULL, NULL, NULL); } @@ -315,14 +319,14 @@ AVPicture * const dst = (AVPicture *) m_frame_rgba.get(); // Assign appropriate parts of the buffer to image planes in m_frame_rgba - avpicture_fill((AVPicture *) (m_frame_rgba).get(), &(m_buffer_rgba[m_writeBuffer])[0], PIX_FMT_RGB24, width(), height()); + avpicture_fill((AVPicture *) (m_frame_rgba).get(), &(m_buffer_rgba[m_writeBuffer])[0], AV_PIX_FMT_RGB24, width(), height()); // Convert YUVA420p (i.e. YUV420p plus alpha channel) using our own routine - if (m_context->pix_fmt == PIX_FMT_YUVA420P) + if (m_context->pix_fmt == AV_PIX_FMT_YUVA420P) yuva420pToRgba(dst, src, width(), height()); else - convert(dst, PIX_FMT_RGB24, src, m_context->pix_fmt, width(), height()); + convert(dst, AV_PIX_FMT_RGB24, src, m_context->pix_fmt, width(), height()); // Wait 'delay' seconds before publishing the picture. int i_delay = static_cast(delay * 1000000 + 0.5); @@ -349,7 +353,7 @@ void FFmpegDecoderVideo::yuva420pToRgba(AVPicture * const dst, AVPicture * const src, int width, int height) { - convert(dst, PIX_FMT_RGB24, src, m_context->pix_fmt, width, height); + convert(dst, AV_PIX_FMT_RGB24, src, m_context->pix_fmt, width, height); const size_t bpp = 4; @@ -369,6 +373,7 @@ +#if LIBAVFORMAT_VERSION_INT < AV_VERSION_INT(55, 0, 0) int FFmpegDecoderVideo::getBuffer(AVCodecContext * const context, AVFrame * const picture) { const FFmpegDecoderVideo * const this_ = reinterpret_cast(context->opaque); @@ -381,15 +386,31 @@ return result; } +#else +int FFmpegDecoderVideo::getBuffer(AVCodecContext * const context, AVFrame * const picture, int flags) +{ + const FFmpegDecoderVideo * const this_ = reinterpret_cast(context->opaque); + + const int result = avcodec_default_get_buffer2(context, picture, flags); + int64_t * p_pts = reinterpret_cast( av_malloc(sizeof(int64_t)) ); + + *p_pts = this_->m_packet_pts; + picture->opaque = p_pts; + + return result; +} +#endif void FFmpegDecoderVideo::releaseBuffer(AVCodecContext * const context, AVFrame * const picture) { +#if LIBAVFORMAT_VERSION_INT < AV_VERSION_INT(55, 0, 0) if (picture != 0) av_freep(&picture->opaque); avcodec_default_release_buffer(context, picture); +#endif } --- OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegDecoderVideo.hpp.orig 2016-03-21 11:55:42.261200729 +0100 +++ OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegDecoderVideo.hpp 2016-03-21 11:55:58.214746802 +0100 @@ -94,7 +94,11 @@ int src_pix_fmt, int src_width, int src_height); +#if LIBAVFORMAT_VERSION_INT < AV_VERSION_INT(55, 0, 0) static int getBuffer(AVCodecContext * context, AVFrame * picture); +#else + static int getBuffer(AVCodecContext * context, AVFrame * picture, int flags); +#endif static void releaseBuffer(AVCodecContext * context, AVFrame * picture); PacketQueue & m_packets; --- OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegParameters.cpp~ 2013-09-05 12:04:56.000000000 +0200 +++ OpenSceneGraph-3.5.1/src/osgPlugins/ffmpeg/FFmpegParameters.cpp 2016-03-21 11:56:17.585004570 +0100 @@ -19,7 +19,7 @@ #include } -inline PixelFormat osg_av_get_pix_fmt(const char *name) { return av_get_pix_fmt(name); } +inline AVPixelFormat osg_av_get_pix_fmt(const char *name) { return av_get_pix_fmt(name); } namespace osgFFmpeg {