/*
 * producer_avformat.c -- avformat producer
 * Copyright (C) 2003-2024 Meltytech, LLC
 *
 * This library is free software; you can redistribute it and/or
 * modify it under the terms of the GNU Lesser General Public
 * License as published by the Free Software Foundation; either
 * version 2.1 of the License, or (at your option) any later version.
 *
 * This library is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * Lesser General Public License for more details.
 *
 * You should have received a copy of the GNU Lesser General Public
 * License along with this library; if not, write to the Free Software
 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301  USA
 */

#if !defined(_POSIX_C_SOURCE) || _POSIX_C_SOURCE < 200809L
#undef _POSIX_C_SOURCE
#define _POSIX_C_SOURCE 200809L
#endif
#if !defined(_XOPEN_SOURCE) || _XOPEN_SOURCE < 500
#undef _XOPEN_SOURCE
#define _XOPEN_SOURCE 500
#endif

#include "common.h"

// MLT Header files
#include <framework/mlt_cache.h>
#include <framework/mlt_deque.h>
#include <framework/mlt_factory.h>
#include <framework/mlt_frame.h>
#include <framework/mlt_log.h>
#include <framework/mlt_producer.h>
#include <framework/mlt_profile.h>
#include <framework/mlt_slices.h>

// ffmpeg Header files
#include <libavcodec/avcodec.h>
#include <libavformat/avformat.h>
#include <libavutil/channel_layout.h>
#include <libavutil/dict.h>
#include <libavutil/imgutils.h>
#include <libavutil/opt.h>
#include <libavutil/pixdesc.h>
#include <libavutil/samplefmt.h>
#include <libavutil/version.h>
#include <libswscale/swscale.h>

#define USE_HWACCEL 1
#if USE_HWACCEL
#include <libavutil/hwcontext.h>
#endif

#ifdef AVFILTER
#include <libavfilter/avfilter.h>
#include <libavfilter/buffersink.h>
#include <libavfilter/buffersrc.h>
#endif

// System header files
#include <limits.h>
#include <math.h>
#include <pthread.h>
#include <stdatomic.h>
#include <stdlib.h>
#include <string.h>
#include <wchar.h>

#define POSITION_INITIAL (-2)
#define POSITION_INVALID (-1)

#define MAX_AUDIO_STREAMS (32)
#define MAX_AUDIO_FRAME_SIZE (192000) // 1 second of 48khz 32bit audio
#define IMAGE_ALIGN (1)
#define VFR_THRESHOLD \
    (3) // The minimum number of video frames with differing durations to be considered VFR.

struct producer_avformat_s
{
    mlt_producer parent;
    AVFormatContext *dummy_context;
    AVFormatContext *audio_format;
    AVFormatContext *video_format;
    AVCodecContext *audio_codec[MAX_AUDIO_STREAMS];
    AVCodecContext *video_codec;
    AVFrame *video_frame;
    AVFrame *audio_frame;
    AVPacket pkt;
    mlt_position audio_expected;
    mlt_position video_expected;
    int audio_index;
    int video_index;
    int64_t first_pts;
    atomic_int_fast64_t last_position;
    int video_seekable;
    int seekable; /// This one is used for both audio and file level seekability.
    atomic_int_fast64_t current_position;
    mlt_position nonseek_position;
    atomic_int top_field_first;
    int progressive;
    uint8_t *audio_buffer[MAX_AUDIO_STREAMS];
    int audio_buffer_size[MAX_AUDIO_STREAMS];
    uint8_t *decode_buffer[MAX_AUDIO_STREAMS];
    int audio_used[MAX_AUDIO_STREAMS];
    int audio_streams;
    int audio_max_stream;
    int total_channels;
    int max_channel;
    int max_frequency;
    unsigned int invalid_pts_counter;
    unsigned int invalid_dts_counter;
    mlt_cache image_cache;
    mlt_cache audio_cache;
    int yuv_colorspace, color_primaries, color_trc;
    int full_range;
    pthread_mutex_t video_mutex;
    pthread_mutex_t audio_mutex;
    mlt_deque apackets;
    mlt_deque vpackets;
    pthread_mutex_t packets_mutex;
    pthread_mutex_t open_mutex;
    pthread_mutex_t close_mutex;
    int is_mutex_init;
    pthread_t packets_thread;
    pthread_cond_t packets_cond;
    int packets_thread_ret; // latest non-zero non-EGAIN return on av_read_frame() in packets_thread
    int packets_thread_stop; // non-zero when packets_thread is to stop
    int is_thread_init;
    AVRational video_time_base;
    mlt_frame last_good_frame; // for video error concealment
    int last_good_position;    // for video error concealment
#ifdef AVFILTER
    AVFilterGraph *vfilter_graph;
    AVFilterContext *vfilter_in;
    AVFilterContext *vfilter_out;
#endif
    int autorotate;
    double rotation;
    int is_audio_synchronizing;
    int video_send_result;
    int reset_image_cache;
#if USE_HWACCEL
    struct
    {
        int pix_fmt;
        int device_type;
        char device[128];
        AVBufferRef *device_ctx;
    } hwaccel;
#endif
};
typedef struct producer_avformat_s *producer_avformat;

// Forward references.
static int list_components(char *file);
static int producer_open(
    producer_avformat self, mlt_profile profile, const char *URL, int take_lock, int test_open);
static int producer_get_frame(mlt_producer producer, mlt_frame_ptr frame, int index);
static int producer_probe(mlt_producer producer);
static void producer_avformat_close(producer_avformat);
static void producer_close(mlt_producer parent);
static void producer_set_up_video(producer_avformat self, mlt_frame frame);
static void producer_set_up_audio(producer_avformat self, mlt_frame frame);
static void apply_properties(void *obj, mlt_properties properties, int flags);
static int video_codec_init(producer_avformat self, int index, mlt_properties properties);
static void get_audio_streams_info(producer_avformat self);
static mlt_audio_format pick_audio_format(int sample_fmt);
static int pick_av_pixel_format(int *pix_fmt, int full_range);
static void property_changed(mlt_service owner, producer_avformat self, char *name);

static int absolute_stream_index(AVFormatContext *context, enum AVMediaType media_type, int relative)
{
    if (context) {
        int n = -1;
        for (int i = 0; i < context->nb_streams; i++) {
            AVCodecParameters *codec_params = context->streams[i]->codecpar;
            if (codec_params->codec_type == media_type && ++n == relative) {
                return i;
            }
        }
    }
    return -1;
}

static int relative_stream_index(AVFormatContext *context, enum AVMediaType media_type, int absolute)
{
    if (context) {
        int n = -1;
        for (int i = 0; i < context->nb_streams; i++) {
            AVCodecParameters *codec_params = context->streams[i]->codecpar;
            if (codec_params->codec_type == media_type) {
                ++n;
                if (absolute == i)
                    return n;
            }
        }
    }
    return -1;
}

/** Constructor for libavformat.
*/

mlt_producer producer_avformat_init(mlt_profile profile, const char *service, char *file)
{
    if (list_components(file))
        return NULL;

    mlt_producer producer = NULL;

    // Check that we have a non-NULL argument
    if (file) {
        // Construct the producer
        producer_avformat self = calloc(1, sizeof(struct producer_avformat_s));
        producer = calloc(1, sizeof(struct mlt_producer_s));

        // Initialise it
        if (mlt_producer_init(producer, self) == 0) {
            self->parent = producer;

            // Get the properties
            mlt_properties properties = MLT_PRODUCER_PROPERTIES(producer);

            // Set the resource property (required for all producers)
            mlt_properties_set(properties, "resource", file);

            // Register transport implementation with the producer
            producer->close = (mlt_destructor) producer_close;

            // Register our get_frame implementation
            producer->get_frame = producer_get_frame;

            // Register our probe implementation
            mlt_properties_set_data(properties, "mlt_producer_probe", producer_probe, 0, NULL, NULL);

            // Force the duration to be computed unless explicitly provided.
            mlt_properties_set_position(properties, "length", 0);
            mlt_properties_set_position(properties, "out", 0);

            if (strcmp(service, "avformat-novalidate")) {
                // Open the file
                if (producer_open(self, profile, mlt_properties_get(properties, "resource"), 1, 1)
                    != 0) {
                    // Clean up
                    producer_avformat_close(self);
                    mlt_producer_close(producer);
                    producer = NULL;
                } else if (self->seekable) {
                    // Close the file to release resources for large playlists - reopen later as needed
                    if (self->audio_format)
                        avformat_close_input(&self->audio_format);
                    if (self->video_format)
                        avformat_close_input(&self->video_format);
                }
            }
            if (producer) {
                // Default the user-selectable indices from the auto-detected indices
                mlt_properties_set_int(properties, "audio_index", self->audio_index);
                mlt_properties_set_int(properties, "video_index", self->video_index);
                mlt_service_cache_put(MLT_PRODUCER_SERVICE(producer),
                                      "producer_avformat",
                                      self,
                                      0,
                                      (mlt_destructor) producer_avformat_close);
                mlt_properties_set_int(properties, "mute_on_pause", 0);
                mlt_events_listen(properties,
                                  self,
                                  "property-changed",
                                  (mlt_listener) property_changed);
            }
        }
    }
    return producer;
}

int list_components(char *file)
{
    int skip = 0;

    // Report information about available demuxers and codecs as YAML Tiny
    if (file && strstr(file, "f-list")) {
        fprintf(stderr, "---\nformats:\n");
        void *state = NULL;
        const AVInputFormat *format = NULL;
        while ((format = av_demuxer_iterate(&state))) {
            fprintf(stderr, "  - %s\n", format->name);
        }
        fprintf(stderr, "...\n");
        skip = 1;
    }
    if (file && strstr(file, "acodec-list")) {
        fprintf(stderr, "---\naudio_codecs:\n");
        void *state = NULL;
        const AVCodec *codec = NULL;
        while ((codec = av_codec_iterate(&state))) {
            if (av_codec_is_decoder(codec) && codec->type == AVMEDIA_TYPE_AUDIO)
                fprintf(stderr, "  - %s\n", codec->name);
        }
        fprintf(stderr, "...\n");
        skip = 1;
    }
    if (file && strstr(file, "vcodec-list")) {
        fprintf(stderr, "---\nvideo_codecs:\n");
        void *state = NULL;
        const AVCodec *codec = NULL;
        while ((codec = av_codec_iterate(&state))) {
            if (av_codec_is_decoder(codec) && codec->type == AVMEDIA_TYPE_VIDEO)
                fprintf(stderr, "  - %s\n", codec->name);
        }
        fprintf(stderr, "...\n");
        skip = 1;
    }

    return skip;
}

static int first_video_index(producer_avformat self)
{
    AVFormatContext *context = self->video_format ? self->video_format : self->audio_format;
    int result = -1; // not found

    if (context) {
        unsigned int i;
        for (i = 0; i < context->nb_streams; i++) {
            if (context->streams[i]->codecpar
                && context->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_VIDEO)
                break;
        }
        if (i < context->nb_streams) {
            result = i;
        }
    }
    return result;
}

#include <libavutil/spherical.h>

static const char *get_projection(AVStream *st)
{
    const AVSphericalMapping *spherical
        = (const AVSphericalMapping *) av_stream_get_side_data(st, AV_PKT_DATA_SPHERICAL, NULL);

    if (spherical)
        return av_spherical_projection_name(spherical->projection);
    return NULL;
}

#include <libavutil/display.h>

static double get_rotation(mlt_properties properties, AVStream *st)
{
    AVDictionaryEntry *rotate_tag = av_dict_get(st->metadata, "rotate", NULL, 0);
    int has_rotate_metadata = rotate_tag && *rotate_tag->value && strcmp(rotate_tag->value, "0");
    uint8_t *displaymatrix = av_stream_get_side_data(st, AV_PKT_DATA_DISPLAYMATRIX, NULL);
    double theta = mlt_properties_get_double(properties, "rotate");
    int has_mlt_rotate = !!mlt_properties_get(properties, "rotate");

    if (has_rotate_metadata && !has_mlt_rotate) {
        char *tail;
        theta = strtod(rotate_tag->value, &tail);
        if (*tail) { // invalid
            theta = 0;
            has_rotate_metadata = 0;
        }
    }
    if (displaymatrix && !has_rotate_metadata && !has_mlt_rotate) {
        theta = -av_display_rotation_get((int32_t *) displaymatrix);
    }

    theta -= 360 * floor(theta / 360 + 0.9 / 360);

    return theta;
}

static char *filter_restricted(const char *in)
{
    if (!in)
        return NULL;
    size_t n = strlen(in);
    // https://github.com/bminor/glibc/commit/9bcd12d223a8990254b65e2dada54faa5d2742f3
    char *out = calloc(n + MB_CUR_MAX, 1);
    char *p = out;
    mbstate_t mbs;
    memset(&mbs, 0, sizeof(mbs));
    while (*in) {
        wchar_t w;
        size_t c = mbrtowc(&w, in, n, &mbs);
        if (c <= 0 || c > n)
            break;
        n -= c;
        in += c;
        if (w == 0x9 || w == 0xA || w == 0xD || (w >= 0x20 && w <= 0xD7FF)
            || (w >= 0xE000 && w <= 0xFFFD) || (w >= 0x10000 && w <= 0x10FFFF)) {
            mbstate_t ps;
            memset(&ps, 0, sizeof(ps));
            c = wcrtomb(p, w, &ps);
            if (c > 0)
                p += c;
        }
    }
    return out;
}

static AVRational guess_frame_rate(producer_avformat self, AVStream *stream)
{
    AVRational frame_rate = av_guess_frame_rate(self->video_format, stream, NULL);
    double fps = av_q2d(frame_rate);

    // Verify and sanitize the frame rate.
    if (isnan(fps) || isinf(fps) || fps == 0) {
        frame_rate = stream->avg_frame_rate;
        fps = av_q2d(frame_rate);
    }
    // With my samples when r_frame_rate != 1000 but avg_frame_rate is valid,
    // avg_frame_rate gives some approximate value that does not well match the media.
    // Also, on my sample where r_frame_rate = 1000, using avg_frame_rate directly
    // results in some very choppy output, but some value slightly different works
    // great.
    if (av_q2d(stream->r_frame_rate) >= 1000 && av_q2d(stream->avg_frame_rate) > 0) {
        frame_rate = av_d2q(av_q2d(stream->avg_frame_rate), 1024);
        fps = av_q2d(frame_rate);
    }
    // XXX frame rates less than 1 fps are not considered sane
    if (isnan(fps) || isinf(fps) || fps < 1.0) {
        // Get the frame rate from the codec.
        frame_rate.num = self->video_codec->time_base.den;
        frame_rate.den = self->video_codec->time_base.num * self->video_codec->ticks_per_frame;
        fps = av_q2d(frame_rate);
    }
    if (isnan(fps) || isinf(fps) || fps < 1.0) {
        // Use the profile frame rate if all else fails.
        mlt_profile profile = mlt_service_profile(MLT_PRODUCER_SERVICE(self->parent));
        frame_rate.num = profile->frame_rate_num;
        frame_rate.den = profile->frame_rate_den;
    }

    // Normalize broadcast frame rates for Matroska
    if (self->video_format->iformat->name && strstr(self->video_format->iformat->name, "matroska")) {
        switch (lrint(100000.0 * frame_rate.num / frame_rate.den)) {
        case 2997003:
            frame_rate.num = 30000;
            frame_rate.den = 1001;
            break;
        case 5994006:
            frame_rate.num = 60000;
            frame_rate.den = 1001;
            break;
        case 2397602:
            frame_rate.num = 24000;
            frame_rate.den = 1001;
            break;
        case 4795204:
            frame_rate.num = 48000;
            frame_rate.den = 1001;
            break;
        default:
            break;
        }
    }

    fps = mlt_properties_get_double(MLT_PRODUCER_PROPERTIES(self->parent),
                                    "meta.attr.com.android.capture.fps.markup");
    if (fps > 0.0 && isfinite(fps))
        frame_rate = av_d2q(fps, 1024);

    return frame_rate;
}

/** Find the default streams.
*/

static mlt_properties find_default_streams(producer_avformat self)
{
    unsigned int i;
    char key[200];
    AVDictionaryEntry *tag = NULL;
    AVFormatContext *context = self->video_format;
    mlt_properties meta_media = MLT_PRODUCER_PROPERTIES(self->parent);

    // Default to the first audio and video streams found
    self->audio_index = -1;
    int first_video_index = self->video_index = -1;

    mlt_properties_set_int(meta_media, "meta.media.nb_streams", context->nb_streams);

    // Allow for multiple audio and video streams in the file and select first of each (if available)
    for (i = 0; i < context->nb_streams; i++) {
        // Get the codec context
        AVFormatContext *vpx_context = NULL;
        AVStream *stream = context->streams[i];
        if (!stream)
            continue;
        AVCodecParameters *codec_params = stream->codecpar;
        const AVCodec *codec = avcodec_find_decoder(codec_params->codec_id);
        if (!codec)
            continue;
        int switch_to_vpx = 0;
        if (codec_params->codec_id == AV_CODEC_ID_VP9) {
            if (!(codec = avcodec_find_decoder_by_name("libvpx-vp9"))) {
                codec = avcodec_find_decoder(codec_params->codec_id);
            } else {
                switch_to_vpx = 1;
            }
        } else if (codec_params->codec_id == AV_CODEC_ID_VP8) {
            if (!(codec = avcodec_find_decoder_by_name("libvpx"))) {
                codec = avcodec_find_decoder(codec_params->codec_id);
            } else {
                switch_to_vpx = 1;
            }
        }
        if (switch_to_vpx) {
            // Use a temporary format context to get the real pixel format with the libvpx decoder,
            // since the native decoder incorreclty detects yuva420p as yuv420p
            int error = avformat_open_input(&vpx_context,
                                            mlt_properties_get(meta_media, "resource"),
                                            NULL,
                                            NULL);
            if (!error) {
                vpx_context->video_codec = codec;
                avformat_find_stream_info(vpx_context, NULL);
                AVStream *vpx_stream = vpx_context->streams[i];
                if (vpx_stream) {
                    codec_params = vpx_stream->codecpar;
                }
            }
        }

        snprintf(key, sizeof(key), "meta.media.%u.stream.type", i);

        // Determine the type and obtain the first index of each type
        switch (codec_params->codec_type) {
        case AVMEDIA_TYPE_VIDEO:
            // Save the first video stream
            if (first_video_index < 0)
                first_video_index = i;
            // Only set the video stream if not album art
            if (self->video_index < 0
                && !(context->streams[i]->disposition & AV_DISPOSITION_ATTACHED_PIC)) {
                self->video_index = i;
            }
            mlt_properties_set(meta_media, key, "video");
            snprintf(key, sizeof(key), "meta.media.%u.stream.frame_rate", i);
            AVRational frame_rate = guess_frame_rate(self, context->streams[i]);
            mlt_properties_set_double(meta_media, key, av_q2d(frame_rate));

            const char *projection = get_projection(context->streams[i]);
            if (projection) {
                snprintf(key, sizeof(key), "meta.media.%u.stream.projection", i);
                mlt_properties_set_string(meta_media, key, projection);
            }
            snprintf(key, sizeof(key), "meta.media.%u.stream.sample_aspect_ratio", i);
            mlt_properties_set_double(meta_media,
                                      key,
                                      av_q2d(context->streams[i]->sample_aspect_ratio));
            snprintf(key, sizeof(key), "meta.media.%u.codec.width", i);
            mlt_properties_set_int(meta_media, key, codec_params->width);
            snprintf(key, sizeof(key), "meta.media.%u.codec.height", i);
            mlt_properties_set_int(meta_media, key, codec_params->height);
            snprintf(key, sizeof(key), "meta.media.%u.codec.rotate", i);
            mlt_properties_set_int(meta_media, key, get_rotation(NULL, context->streams[i]));
            //				snprintf( key, sizeof(key), "meta.media.%u.codec.frame_rate", i );
            //				AVRational frame_rate = { codec_context->time_base.den, codec_context->time_base.num * codec_context->ticks_per_frame };
            //				mlt_properties_set_double( meta_media, key, av_q2d( frame_rate ) );
            snprintf(key, sizeof(key), "meta.media.%u.codec.pix_fmt", i);
            mlt_properties_set(meta_media, key, av_get_pix_fmt_name(codec_params->format));
            snprintf(key, sizeof(key), "meta.media.%u.codec.sample_aspect_ratio", i);
            mlt_properties_set_double(meta_media, key, av_q2d(codec_params->sample_aspect_ratio));
            snprintf(key, sizeof(key), "meta.media.%u.codec.colorspace", i);
            switch (codec_params->color_space) {
            case AVCOL_SPC_SMPTE240M:
                mlt_properties_set_int(meta_media, key, 240);
                break;
            case AVCOL_SPC_BT470BG:
            case AVCOL_SPC_SMPTE170M:
                mlt_properties_set_int(meta_media, key, 601);
                break;
            case AVCOL_SPC_BT709:
                mlt_properties_set_int(meta_media, key, 709);
                break;
            case AVCOL_SPC_UNSPECIFIED:
            case AVCOL_SPC_RESERVED:
                // This is a heuristic Charles Poynton suggests in "Digital Video and HDTV"
                mlt_properties_set_int(meta_media,
                                       key,
                                       codec_params->width * codec_params->height > 750000 ? 709
                                                                                           : 601);
                break;
            default:
                mlt_properties_set_int(meta_media, key, codec_params->color_space);
                break;
            }
            if (codec_params->color_trc && codec_params->color_trc != AVCOL_TRC_UNSPECIFIED) {
                snprintf(key, sizeof(key), "meta.media.%u.codec.color_trc", i);
                mlt_properties_set_double(meta_media, key, codec_params->color_trc);
            }
            break;
        case AVMEDIA_TYPE_AUDIO:
#if HAVE_FFMPEG_CH_LAYOUT
            if (!codec_params->ch_layout.nb_channels)
#else
            if (!codec_params->channels)
#endif
                break;
            // Use first audio stream
            if (self->audio_index < 0 && pick_audio_format(codec_params->format) != mlt_audio_none)
                self->audio_index = i;

            mlt_properties_set(meta_media, key, "audio");
            snprintf(key, sizeof(key), "meta.media.%u.codec.sample_fmt", i);
            mlt_properties_set(meta_media, key, av_get_sample_fmt_name(codec_params->format));
            snprintf(key, sizeof(key), "meta.media.%u.codec.sample_rate", i);
            mlt_properties_set_int(meta_media, key, codec_params->sample_rate);
            snprintf(key, sizeof(key), "meta.media.%u.codec.channels", i);
            mlt_channel_layout mlt_layout = mlt_channel_independent;
#if HAVE_FFMPEG_CH_LAYOUT
            mlt_properties_set_int(meta_media, key, codec_params->ch_layout.nb_channels);
            if (av_channel_layout_check(&codec_params->ch_layout)) {
                mlt_layout = av_channel_layout_to_mlt(&codec_params->ch_layout);
            } else {
                AVChannelLayout ch_layout;
                av_channel_layout_default(&ch_layout, codec_params->ch_layout.nb_channels);
                mlt_layout = av_channel_layout_to_mlt(&ch_layout);
                av_channel_layout_uninit(&ch_layout);
            }
#else
            mlt_properties_set_int(meta_media, key, codec_params->channels);
            if (codec_params->channel_layout == 0)
                mlt_layout = av_channel_layout_to_mlt(
                    av_get_default_channel_layout(codec_params->channels));
            else
                mlt_layout = av_channel_layout_to_mlt(codec_params->channel_layout);
#endif
            const char *layout = mlt_audio_channel_layout_name(mlt_layout);
            snprintf(key, sizeof(key), "meta.media.%u.codec.layout", i);
            mlt_properties_set(meta_media, key, layout);
            break;
        default:
            break;
        }
        //		snprintf( key, sizeof(key), "meta.media.%u.stream.time_base", i );
        //		mlt_properties_set_double( meta_media, key, av_q2d( context->streams[ i ]->time_base ) );
        snprintf(key, sizeof(key), "meta.media.%u.codec.name", i);
        mlt_properties_set(meta_media, key, codec->name);
        snprintf(key, sizeof(key), "meta.media.%u.codec.long_name", i);
        mlt_properties_set(meta_media, key, codec->long_name);
        snprintf(key, sizeof(key), "meta.media.%u.codec.bit_rate", i);
        mlt_properties_set_int64(meta_media, key, codec_params->bit_rate);
        // 		snprintf( key, sizeof(key), "meta.media.%u.codec.time_base", i );
        // 		mlt_properties_set_double( meta_media, key, av_q2d( codec_context->time_base ) );
        //		snprintf( key, sizeof(key), "meta.media.%u.codec.profile", i );
        //		mlt_properties_set_int( meta_media, key, codec_context->profile );
        //		snprintf( key, sizeof(key), "meta.media.%u.codec.level", i );
        //		mlt_properties_set_int( meta_media, key, codec_context->level );

        // Read Metadata
        while ((tag = av_dict_get(stream->metadata, "", tag, AV_DICT_IGNORE_SUFFIX))) {
            if (tag->value && strcmp(tag->value, "") && strcmp(tag->value, "und")) {
                snprintf(key, sizeof(key), "meta.attr.%u.stream.%s.markup", i, tag->key);
                char *value = filter_restricted(tag->value);
                mlt_properties_set(meta_media, key, value);
                free(value);
            }
        }
        if (vpx_context) {
            avformat_close_input(&vpx_context);
            vpx_context = NULL;
        }
    }

    // Use the album art if that is all we have
    if (self->video_index < 0 && first_video_index >= 0)
        self->video_index = first_video_index;

    while ((tag = av_dict_get(context->metadata, "", tag, AV_DICT_IGNORE_SUFFIX))) {
        if (tag->value && strcmp(tag->value, "") && strcmp(tag->value, "und")) {
            snprintf(key, sizeof(key), "meta.attr.%s.markup", tag->key);
            char *value = filter_restricted(tag->value);
            mlt_properties_set(meta_media, key, value);
            free(value);
        }
    }

    return meta_media;
}

static void get_aspect_ratio(mlt_properties properties,
                             AVStream *stream,
                             AVCodecParameters *codec_params)
{
    AVRational sar = stream->sample_aspect_ratio;
    if (sar.num <= 0 || sar.den <= 0)
        sar = codec_params->sample_aspect_ratio;
    if (sar.num <= 0 || sar.den <= 0)
        sar.num = sar.den = 1;
    mlt_properties_set_int(properties, "meta.media.sample_aspect_num", sar.num);
    mlt_properties_set_int(properties, "meta.media.sample_aspect_den", sar.den);
    mlt_properties_set_double(properties, "aspect_ratio", av_q2d(sar));
}

static char *parse_url(mlt_profile profile,
                       const char *URL,
                       const AVInputFormat **format,
                       AVDictionary **params)
{
    (void) profile; // unused
    if (!URL)
        return NULL;

    char *protocol = strdup(URL);
    char *url = strchr(protocol, ':');

    // Truncate protocol string
    if (url && (url - protocol) > 1 && avio_check(URL, 0) < 0) { // if defined and not a drive letter
        url[0] = '\0';
        ++url;
        mlt_log_debug(NULL, "%s: protocol=%s resource=%s\n", __FUNCTION__, protocol, url);

        // Lookup the format
        *format = av_find_input_format(protocol);
    } else {
        url = protocol;
    }

    // Eat the format designator
    char *result = url;

    // support for legacy width and height parameters
    char *width = NULL;
    char *height = NULL;

    // Parse out params
    char *query = strchr(url, '?');
    if (*format) {
        // Query string delimiter is '?'
        url = (query && query > url && query[-1] != '\\') ? query : NULL;
    } else {
        // Ignore unescaped question marks
        while (query && query > url && query[-1] != '\\') {
            query = strchr(query + 1, '?');
        }
        // Query string delimiter is '\?'
        url = (query && query > url && query[-1] == '\\') ? query : NULL;
        if (url)
            url[-1] = '\0'; // null the backslash
    }
    while (url) {
        url[0] = '\0';
        char *name = strdup(++url);
        char *value = strchr(name, '=');
        if (!value)
            // Also accept : as delimiter for backwards compatibility.
            value = strchr(name, ':');
        if (value) {
            value[0] = '\0';
            value++;
            char *t = strchr(value, '&');
            if (t)
                t[0] = 0;
            // translate old parameters to new av_dict names
            if (!strcmp(name, "frame_rate"))
                av_dict_set(params, "framerate", value, 0);
            else if (!strcmp(name, "pix_fmt"))
                av_dict_set(params, "pixel_format", value, 0);
            else if (!strcmp(name, "width"))
                width = strdup(value);
            else if (!strcmp(name, "height"))
                height = strdup(value);
            else
                // generic demux/device option support
                av_dict_set(params, name, value, 0);
        }
        free(name);
        url = strchr(url, '&');
    }
    // continued support for legacy width and height parameters
    if (width && height) {
        char *s = malloc(strlen(width) + strlen(height) + 2);
        strcpy(s, width);
        strcat(s, "x");
        strcat(s, height);
        av_dict_set(params, "video_size", s, 0);
        free(s);
    }
    free(width);
    free(height);

    result = strdup(result);
    free(protocol);
    mlt_log_debug(NULL, "[producer avformat] %s filename = %s\n", __FUNCTION__, result);
    return result;
}

static enum AVPixelFormat pick_pix_fmt(enum AVPixelFormat pix_fmt)
{
    switch (pix_fmt) {
    case AV_PIX_FMT_ARGB:
    case AV_PIX_FMT_RGBA:
    case AV_PIX_FMT_ABGR:
    case AV_PIX_FMT_BGRA:
        return AV_PIX_FMT_RGBA;
    case AV_PIX_FMT_BAYER_RGGB16LE:
        return AV_PIX_FMT_RGB24;
#if USE_HWACCEL
    case AV_PIX_FMT_VAAPI:
    case AV_PIX_FMT_CUDA:
    case AV_PIX_FMT_VIDEOTOOLBOX:
    case AV_PIX_FMT_DXVA2_VLD:
    case AV_PIX_FMT_D3D11:
        return AV_PIX_FMT_YUV420P;
#endif
    default:
        return AV_PIX_FMT_YUV422P;
    }
}

static mlt_image_format pick_image_format(enum AVPixelFormat pix_fmt,
                                          mlt_image_format current_format)
{
    if (current_format == mlt_image_none || current_format == mlt_image_movit
        || pix_fmt == AV_PIX_FMT_ARGB || pix_fmt == AV_PIX_FMT_RGBA || pix_fmt == AV_PIX_FMT_ABGR
        || pix_fmt == AV_PIX_FMT_BGRA) {
        switch (pix_fmt) {
        case AV_PIX_FMT_ARGB:
        case AV_PIX_FMT_RGBA:
        case AV_PIX_FMT_ABGR:
        case AV_PIX_FMT_BGRA:
            return mlt_image_rgba;
        case AV_PIX_FMT_YUV420P:
        case AV_PIX_FMT_YUVJ420P:
        case AV_PIX_FMT_YUVA420P:
            return mlt_image_yuv420p;
        case AV_PIX_FMT_RGB24:
        case AV_PIX_FMT_BGR24:
        case AV_PIX_FMT_GRAY8:
        case AV_PIX_FMT_MONOWHITE:
        case AV_PIX_FMT_MONOBLACK:
        case AV_PIX_FMT_RGB8:
        case AV_PIX_FMT_BGR8:
        case AV_PIX_FMT_BAYER_RGGB16LE:
            return mlt_image_rgb;
        case AV_PIX_FMT_YUV420P10LE:
            return mlt_image_yuv420p10;
        case AV_PIX_FMT_YUV422P10LE:
        case AV_PIX_FMT_YUV444P10LE:
            return mlt_image_yuv444p10;
        case AV_PIX_FMT_YUV422P16LE:
            return mlt_image_yuv422p16;
        default:
            current_format = mlt_image_yuv422;
        }
    }
    if (pix_fmt == AV_PIX_FMT_BAYER_RGGB16LE) {
        return mlt_image_rgb;
    } else if (pix_fmt == AV_PIX_FMT_YUVA444P10LE || pix_fmt == AV_PIX_FMT_GBRAP10LE
               || pix_fmt == AV_PIX_FMT_GBRAP12LE) {
        return mlt_image_rgba;
    }
    return current_format;
}

static int get_basic_info(producer_avformat self, mlt_profile profile, const char *filename)
{
    int error = 0;

    // Get the properties
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(self->parent);

    AVFormatContext *format = self->video_format;

    // Get the duration
    if (mlt_properties_get_position(properties, "length") <= 0
        || mlt_properties_get_position(properties, "out") <= 0) {
        if (format->duration != AV_NOPTS_VALUE) {
            // This isn't going to be accurate for all formats
            // We will treat everything with the producer fps.
            mlt_position frames = (mlt_position) lrint(format->duration * mlt_profile_fps(profile)
                                                       / AV_TIME_BASE);
            if (mlt_properties_get_position(properties, "out") <= 0)
                mlt_properties_set_position(properties, "out", frames - 1);
            if (mlt_properties_get_position(properties, "length") <= 0)
                mlt_properties_set_position(properties, "length", frames);
        } else if (format->nb_streams > 0 && format->streams[0]->codecpar
                   && format->streams[0]->codecpar->codec_id == AV_CODEC_ID_WEBP) {
            char *e = getenv("MLT_DEFAULT_PRODUCER_LENGTH");
            int p = e ? atoi(e) : 15000;
            mlt_properties_set_int(properties, "out", MAX(0, p - 1));
            mlt_properties_set_int(properties, "length", p);
        } else {
            // Set live sources to run forever
            if (mlt_properties_get_position(properties, "length") <= 0)
                mlt_properties_set_position(properties, "length", INT_MAX);
            if (mlt_properties_get_position(properties, "out") <= 0)
                mlt_properties_set_position(properties, "out", INT_MAX - 1);
            mlt_properties_set(properties, "eof", "loop");
        }
    }

    // Check if we're seekable
    // avdevices are typically AVFMT_NOFILE and not seekable
    self->seekable = !format->iformat || !(format->iformat->flags & AVFMT_NOFILE);
    if (format->pb) {
        // protocols can indicate if they support seeking
        self->seekable = format->pb->seekable;
    }
    if (self->seekable) {
        // Do a more rigorous test of seekable on a disposable context
        if (format->nb_streams > 0 && format->streams[0]->codecpar
            && format->streams[0]->codecpar->codec_id != AV_CODEC_ID_WEBP)
            self->seekable = av_seek_frame(format, -1, format->start_time, AVSEEK_FLAG_BACKWARD)
                             >= 0;
        mlt_properties_set_int(properties, "seekable", self->seekable);
        self->dummy_context = format;
        self->video_format = NULL;
        avformat_open_input(&self->video_format, filename, NULL, NULL);
        avformat_find_stream_info(self->video_format, NULL);
        format = self->video_format;
    }
    self->video_seekable = self->seekable;

    // Fetch the width, height and aspect ratio
    if (self->video_index != -1) {
        AVCodecParameters *codec_params = format->streams[self->video_index]->codecpar;
        mlt_properties_set_int(properties, "width", codec_params->width);
        mlt_properties_set_int(properties, "height", codec_params->height);
        get_aspect_ratio(properties, format->streams[self->video_index], codec_params);

#ifdef AVFILTER
        int pix_fmt = self->vfilter_out ? av_buffersink_get_format(self->vfilter_out)
                                        : codec_params->format;
#else
        int pix_fmt = codec_params->format;
#endif
        pick_av_pixel_format(&pix_fmt, self->full_range);
        if (pix_fmt != AV_PIX_FMT_NONE) {
            // Verify that we can convert this to one of our image formats.
            struct SwsContext *context = sws_getContext(codec_params->width,
                                                        codec_params->height,
                                                        pix_fmt,
                                                        codec_params->width,
                                                        codec_params->height,
                                                        pick_pix_fmt(pix_fmt),
                                                        SWS_BILINEAR,
                                                        NULL,
                                                        NULL,
                                                        NULL);
            if (context) {
                sws_freeContext(context);
                mlt_image_format format = pick_image_format(pix_fmt, mlt_image_yuv422);
                mlt_properties_set_int(properties, "format", format);
            } else
                error = 1;
        } else {
            self->video_index = -1;
        }
    }
    return error;
}

#ifdef AVFILTER
static int setup_video_filters(producer_avformat self)
{
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(self->parent);
    AVFormatContext *format = self->video_format;
    AVStream *stream = format->streams[self->video_index];
    AVCodecParameters *codec_params = stream->codecpar;

    self->vfilter_graph = avfilter_graph_alloc();

    // From ffplay.c:configure_video_filters().
    char buffersrc_args[256];
    AVRational frame_rate = guess_frame_rate(self, stream);
    snprintf(buffersrc_args,
             sizeof(buffersrc_args),
             "video_size=%dx%d:pix_fmt=%d:time_base=%d/%d:pixel_aspect=%d/%d:frame_rate=%d/%d",
             codec_params->width,
             codec_params->height,
             codec_params->format,
             stream->time_base.num,
             stream->time_base.den,
             mlt_properties_get_int(properties, "meta.media.sample_aspect_num"),
             FFMAX(mlt_properties_get_int(properties, "meta.media.sample_aspect_den"), 1),
             frame_rate.num,
             FFMAX(frame_rate.den, 1));

    int result = avfilter_graph_create_filter(&self->vfilter_in,
                                              avfilter_get_by_name("buffer"),
                                              "mlt_buffer",
                                              buffersrc_args,
                                              NULL,
                                              self->vfilter_graph);

    if (result >= 0) {
        result = avfilter_graph_create_filter(&self->vfilter_out,
                                              avfilter_get_by_name("buffersink"),
                                              "mlt_buffersink",
                                              NULL,
                                              NULL,
                                              self->vfilter_graph);
    }

    return result;
}

static int insert_filter(AVFilterGraph *graph,
                         AVFilterContext **last_filter,
                         const char *name,
                         const char *args)
{
    AVFilterContext *filt_ctx;
    int result = avfilter_graph_create_filter(&filt_ctx,
                                              avfilter_get_by_name(name),
                                              name,
                                              args,
                                              NULL,
                                              graph);
    if (result >= 0) {
        result = avfilter_link(filt_ctx, 0, *last_filter, 0);
        if (result >= 0)
            *last_filter = filt_ctx;
    }
    return result;
}

static int setup_filters(producer_avformat self)
{
    int error = 0;
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(self->parent);
    const char *filtergraph = mlt_properties_get(properties, "filtergraph");
    double theta = 0.0;

    if (self->video_index != -1 && self->autorotate) {
        theta = get_rotation(properties, self->video_format->streams[self->video_index]);
        if (self->vfilter_graph && theta != self->rotation) {
            // The rotation has changed. Force the filter graph to be rebuilt
            avfilter_graph_free(&self->vfilter_graph);
            self->vfilter_out = NULL;
            self->rotation = theta;
        }
    }

    if (!self->vfilter_graph && (self->autorotate || filtergraph) && self->video_index != -1) {
        AVFilterContext *last_filter = NULL;
        if (self->autorotate) {
            if (fabs(theta - 90) < 1.0) {
                error = (setup_video_filters(self) < 0);
                last_filter = self->vfilter_out;
                if (!error)
                    error = (insert_filter(self->vfilter_graph, &last_filter, "transpose", "clock")
                             < 0);
            } else if (fabs(theta - 180) < 1.0) {
                error = (setup_video_filters(self) < 0);
                last_filter = self->vfilter_out;
                if (!error)
                    error = (insert_filter(self->vfilter_graph, &last_filter, "hflip", NULL) < 0);
                if (!error)
                    error = (insert_filter(self->vfilter_graph, &last_filter, "vflip", NULL) < 0);
            } else if (fabs(theta - 270) < 1.0) {
                error = (setup_video_filters(self) < 0);
                last_filter = self->vfilter_out;
                if (!error)
                    error = (insert_filter(self->vfilter_graph, &last_filter, "transpose", "cclock")
                             < 0);
            }
        }
        if (filtergraph && !error) {
            if (!self->vfilter_graph) {
                error = (setup_video_filters(self) < 0);
                last_filter = self->vfilter_out;
            }
            AVFilterInOut *outputs = avfilter_inout_alloc();
            AVFilterInOut *inputs = avfilter_inout_alloc();

            outputs->name = av_strdup("in");
            outputs->filter_ctx = self->vfilter_in;
            outputs->pad_idx = 0;
            outputs->next = NULL;

            inputs->name = av_strdup("out");
            inputs->filter_ctx = last_filter;
            inputs->pad_idx = 0;
            inputs->next = NULL;

            if (!error)
                error
                    = (avfilter_graph_parse(self->vfilter_graph, filtergraph, inputs, outputs, NULL)
                       < 0);
        }
        if (self->vfilter_graph) {
            if (!error && !filtergraph)
                error = (avfilter_link(self->vfilter_in, 0, last_filter, 0) < 0);
            if (!error)
                error = (avfilter_graph_config(self->vfilter_graph, NULL) < 0);
        }
    }
    if (error && self->vfilter_graph) {
        avfilter_graph_free(&self->vfilter_graph);
    }
    return error;
}
#endif

static void set_up_discard(producer_avformat self, int audio_index, int video_index)
{
    // The open_mutex must be locked when this function is called
    if (self->audio_format) {
        for (int x = 0; x < self->audio_format->nb_streams; x++) {
            if (audio_index == INT_MAX || x == audio_index
                || (self->audio_format == self->video_format && x == video_index))
                self->audio_format->streams[x]->discard = AVDISCARD_DEFAULT;
            else
                self->audio_format->streams[x]->discard = AVDISCARD_ALL;
        }
    }

    if (self->video_format && self->video_format != self->audio_format) {
        for (int x = 0; x < self->video_format->nb_streams; x++) {
            if (x == video_index)
                self->video_format->streams[x]->discard = AVDISCARD_DEFAULT;
            else
                self->video_format->streams[x]->discard = AVDISCARD_ALL;
        }
    }
}

/** Open the file.
*/

static int producer_open(
    producer_avformat self, mlt_profile profile, const char *URL, int take_lock, int test_open)
{
    // Return an error code (0 == no error)
    int error = 0;
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(self->parent);

    if (!self->is_mutex_init) {
        pthread_mutexattr_t attr;
        pthread_mutexattr_init(&attr);
        pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_RECURSIVE);
        pthread_mutex_init(&self->audio_mutex, &attr);
        pthread_mutex_init(&self->video_mutex, &attr);
        pthread_mutex_init(&self->packets_mutex, &attr);
        pthread_mutex_init(&self->open_mutex, &attr);
        pthread_mutex_init(&self->close_mutex, &attr);
        self->is_mutex_init = 1;
    }

    // Lock the service
    if (take_lock) {
        pthread_mutex_lock(&self->audio_mutex);
        pthread_mutex_lock(&self->video_mutex);
    }
    mlt_events_block(properties, self->parent);

    // Parse URL
    const AVInputFormat *format = NULL;
    AVDictionary *params = NULL;
    char *filename = parse_url(profile, URL, &format, &params);

    // Now attempt to open the file or device with filename
    error = avformat_open_input(&self->video_format, filename, format, &params) < 0;
    if (error)
        // If the URL is a network stream URL, then we probably need to open with full URL
        error = avformat_open_input(&self->video_format, URL, format, &params) < 0;

    // Set MLT properties onto video AVFormatContext
    if (!error && self->video_format) {
        apply_properties(self->video_format, properties, AV_OPT_FLAG_DECODING_PARAM);
        if (self->video_format->iformat && self->video_format->iformat->priv_class
            && self->video_format->priv_data)
            apply_properties(self->video_format->priv_data, properties, AV_OPT_FLAG_DECODING_PARAM);
    }

    // If successful, then try to get additional info
    if (!error && self->video_format) {
        // Get the stream info
        error = avformat_find_stream_info(self->video_format, NULL) < 0;

        // Continue if no error
        if (!error && self->video_format) {
            // Find default audio and video streams
            find_default_streams(self);
            error = get_basic_info(self, profile, filename);

            // Initialize position info
            self->first_pts = AV_NOPTS_VALUE;
            self->last_position = POSITION_INITIAL;

#if USE_HWACCEL
            AVDictionaryEntry *hwaccel = av_dict_get(params, "hwaccel", NULL, 0);
            AVDictionaryEntry *hwaccel_device = av_dict_get(params, "hwaccel_device", NULL, 0);

            if (hwaccel && hwaccel->value) {
                // Leaving `device=NULL` will cause query string parameter `hwaccel_device` to be ignored
                char *device = NULL;
                if (!strcmp(hwaccel->value, "vaapi")) {
                    self->hwaccel.pix_fmt = AV_PIX_FMT_VAAPI;
                    self->hwaccel.device_type = AV_HWDEVICE_TYPE_VAAPI;
                    device = "/dev/dri/renderD128";
                } else if (!strcmp(hwaccel->value, "cuda") || !strcmp(hwaccel->value, "nvdec")) {
                    self->hwaccel.pix_fmt = AV_PIX_FMT_CUDA;
                    self->hwaccel.device_type = AV_HWDEVICE_TYPE_CUDA;
                    device = "0";
                } else if (!strcmp(hwaccel->value, "videotoolbox")) {
                    self->hwaccel.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX;
                    self->hwaccel.device_type = AV_HWDEVICE_TYPE_VIDEOTOOLBOX;
                } else if (!strcmp(hwaccel->value, "d3d11va")) {
                    self->hwaccel.pix_fmt = AV_PIX_FMT_D3D11;
                    self->hwaccel.device_type = AV_HWDEVICE_TYPE_D3D11VA;
                    device = "0";
                } else if (!strcmp(hwaccel->value, "dxva2")) {
                    self->hwaccel.pix_fmt = AV_PIX_FMT_DXVA2_VLD;
                    self->hwaccel.device_type = AV_HWDEVICE_TYPE_DXVA2;
                    device = "0";
                } else {
                    // TODO: init other hardware types
                }

                if (device) {
                    if (hwaccel_device && hwaccel_device->value)
                        device = hwaccel_device->value;
                    memcpy(self->hwaccel.device, device, strlen(device));
                }
            }
#endif

            if (!self->audio_format) {
                // We're going to cheat here - for seekable A/V files, we will have separate contexts
                // to support independent seeking of audio from video.
                // TODO: Is this really necessary?
                if (self->audio_index != -1 && self->video_index != -1) {
                    if (self->seekable) {
                        // And open again for our audio context
                        avformat_open_input(&self->audio_format, filename, NULL, NULL);
                        apply_properties(self->audio_format, properties, AV_OPT_FLAG_DECODING_PARAM);
                        if (self->audio_format->iformat && self->audio_format->iformat->priv_class
                            && self->audio_format->priv_data)
                            apply_properties(self->audio_format->priv_data,
                                             properties,
                                             AV_OPT_FLAG_DECODING_PARAM);
                        avformat_find_stream_info(self->audio_format, NULL);
                    } else {
                        self->audio_format = self->video_format;
                    }
                } else if (self->audio_index != -1) {
                    // We only have an audio context
                    self->audio_format = self->video_format;
                    self->video_format = NULL;
                } else if (self->video_index == -1) {
                    // Something has gone wrong
                    error = -1;
                }
                if (self->audio_format && !self->audio_streams)
                    get_audio_streams_info(self);

#ifdef AVFILTER
                if (!test_open) {
                    self->autorotate = !mlt_properties_get(properties, "autorotate")
                                       || mlt_properties_get_int(properties, "autorotate");
                    error = setup_filters(self);
                }
#endif
            }
        }
    }
    av_dict_free(&params);
    free(filename);
    if (!error) {
        self->apackets = mlt_deque_init();
        self->vpackets = mlt_deque_init();
    }

    if (self->dummy_context) {
        pthread_mutex_lock(&self->open_mutex);
        avformat_close_input(&self->dummy_context);
        self->dummy_context = NULL;
        pthread_mutex_unlock(&self->open_mutex);
    }

    // Unlock the service
    if (take_lock) {
        pthread_mutex_unlock(&self->audio_mutex);
        pthread_mutex_unlock(&self->video_mutex);
    }
    mlt_events_unblock(properties, self->parent);

    return error;
}

static void prepare_reopen(producer_avformat self)
{
    mlt_service_lock(MLT_PRODUCER_SERVICE(self->parent));
    pthread_mutex_lock(&self->audio_mutex);
    pthread_mutex_lock(&self->open_mutex);

    int i;
    for (i = 0; i < MAX_AUDIO_STREAMS; i++) {
        mlt_pool_release(self->audio_buffer[i]);
        self->audio_buffer[i] = NULL;
        av_free(self->decode_buffer[i]);
        self->decode_buffer[i] = NULL;
        avcodec_free_context(&self->audio_codec[i]);
    }
    avcodec_free_context(&self->video_codec);
    av_frame_unref(self->video_frame);
#if USE_HWACCEL
    av_buffer_unref(&self->hwaccel.device_ctx);
    self->hwaccel.device_ctx = NULL;
#endif
    if (self->seekable && self->audio_format)
        avformat_close_input(&self->audio_format);
    if (self->video_format)
        avformat_close_input(&self->video_format);
    self->audio_format = NULL;
    self->video_format = NULL;
#ifdef AVFILTER
    avfilter_graph_free(&self->vfilter_graph);
#endif
    pthread_mutex_unlock(&self->open_mutex);

    // Cleanup the packet queues
    AVPacket *pkt;
    if (self->apackets) {
        while ((pkt = mlt_deque_pop_back(self->apackets))) {
            av_packet_unref(pkt);
            free(pkt);
        }
        mlt_deque_close(self->apackets);
        self->apackets = NULL;
    }
    if (self->vpackets) {
        while ((pkt = mlt_deque_pop_back(self->vpackets))) {
            av_packet_unref(pkt);
            free(pkt);
        }
        mlt_deque_close(self->vpackets);
        self->vpackets = NULL;
    }
    pthread_mutex_unlock(&self->audio_mutex);
    mlt_service_unlock(MLT_PRODUCER_SERVICE(self->parent));
}

static int64_t best_pts(producer_avformat self, int64_t pts, int64_t dts)
{
    self->invalid_pts_counter += pts == AV_NOPTS_VALUE;
    self->invalid_dts_counter += dts == AV_NOPTS_VALUE;
    if ((self->invalid_pts_counter <= self->invalid_dts_counter || dts == AV_NOPTS_VALUE)
        && pts != AV_NOPTS_VALUE)
        return pts;
    else
        return dts;
}

static void find_first_pts(producer_avformat self, int video_index)
{
    // find initial PTS
    AVFormatContext *context = self->video_format ? self->video_format : self->audio_format;
    int ret = 0;
    int pkt_countdown = 500; // check max 500 packets for first video keyframe PTS
    int vfr_countdown = 20;  // check max 20 video frames for VFR
    int vfr_counter = 0;     // counts the number of frame duration changes
    AVPacket pkt;
    int64_t prev_pkt_duration = AV_NOPTS_VALUE;

    av_init_packet(&pkt);

    if (video_index == -1 && self->audio_index >= 0) {
        // Special case for audio only files.
        while (ret >= 0 && pkt_countdown-- > 0 && (self->first_pts == AV_NOPTS_VALUE)) {
            ret = av_read_frame(context, &pkt);
            if (ret >= 0 && pkt.stream_index == self->audio_index) {
                mlt_log_debug(MLT_PRODUCER_SERVICE(self->parent),
                              "first_pts %" PRId64 " dts %" PRId64 " pts_dts_delta %d\n",
                              pkt.pts,
                              pkt.dts,
                              (int) (pkt.pts - pkt.dts));
                self->first_pts = best_pts(self, pkt.pts, pkt.dts);
            }
            av_packet_unref(&pkt);
        }
        av_seek_frame(context, -1, 0, AVSEEK_FLAG_BACKWARD);
        return;
    }

    while (ret >= 0 && pkt_countdown-- > 0
           && (self->first_pts == AV_NOPTS_VALUE
               || (vfr_counter < VFR_THRESHOLD && vfr_countdown > 0))) {
        ret = av_read_frame(context, &pkt);
        if (ret >= 0 && pkt.stream_index == video_index) {
            // Variable frame rate check
            if (pkt.duration != AV_NOPTS_VALUE && pkt.duration != prev_pkt_duration) {
                mlt_log_verbose(MLT_PRODUCER_SERVICE(self->parent),
                                "checking VFR: pkt.duration %" PRId64 "\n",
                                pkt.duration);
                if (prev_pkt_duration != AV_NOPTS_VALUE)
                    ++vfr_counter;
            }
            prev_pkt_duration = pkt.duration;
            vfr_countdown--;

            // Finding PTS of first video key frame
            if ((pkt.flags & AV_PKT_FLAG_KEY) && self->first_pts == AV_NOPTS_VALUE) {
                mlt_log_debug(MLT_PRODUCER_SERVICE(self->parent),
                              "first_pts %" PRId64 " dts %" PRId64 " pts_dts_delta %d\n",
                              pkt.pts,
                              pkt.dts,
                              (int) (pkt.pts - pkt.dts));
                if (pkt.dts != AV_NOPTS_VALUE && pkt.dts < 0)
                    // Decoding Time Stamps with negative values are reported by ffmpeg code for
                    // (at least) MP4 files containing h.264 video using b-frames.
                    // For reasons not understood yet, the first PTS computed then is that of the
                    // third frame, causing MLT to display the third frame as if it was the first.
                    // This if-clause is meant to catch and work around this issue - if there is
                    // a valid, but negative DTS value, we just guess that the first valid
                    // Presentation Time Stamp is == 0.
                    self->first_pts = 0;
                else
                    self->first_pts = best_pts(self, pkt.pts, pkt.dts);
            }
        }
        av_packet_unref(&pkt);
    }

    // Determine if this is variable frame rate video
    int vfr = vfr_counter >= VFR_THRESHOLD;
    if (!vfr) {
        AVStream *stream = self->video_format ? self->video_format->streams[video_index] : NULL;
        if (stream) {
            int d = stream->avg_frame_rate.den;
            vfr = d != 0 && d != 1 && d != 2 && d != 125 && d != 1001
                  && av_cmp_q(stream->avg_frame_rate, stream->r_frame_rate);
        }
    }
    if (vfr) {
        mlt_properties_set_int(MLT_PRODUCER_PROPERTIES(self->parent),
                               "meta.media.variable_frame_rate",
                               1);
    }

    // Reset context position
    av_seek_frame(context, -1, 0, AVSEEK_FLAG_BACKWARD);
}

static int seek_video(producer_avformat self,
                      mlt_position position,
                      int64_t req_position,
                      int preseek)
{
    mlt_producer producer = self->parent;
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(producer);
    int paused = 0;
    int seek_threshold = mlt_properties_get_int(properties, "seek_threshold");
    if (seek_threshold <= 0)
        seek_threshold = 64;

    pthread_mutex_lock(&self->packets_mutex);

    if (self->video_seekable && (position != self->video_expected || self->last_position < 0)) {
        // Fetch the video format context
        AVFormatContext *context = self->video_format;

        // We may want to use the source fps if available
        double source_fps = mlt_properties_get_double(properties, "meta.media.frame_rate_num")
                            / mlt_properties_get_double(properties, "meta.media.frame_rate_den");

        if (self->first_pts == AV_NOPTS_VALUE && self->last_position == POSITION_INITIAL)
            find_first_pts(self, self->video_index);

        if (self->video_frame && position + 1 == self->video_expected) {
            // We're paused - use last image
            paused = 1;
        } else if (position < self->video_expected
                   || position - self->video_expected >= seek_threshold
                   || self->last_position < 0) {
            // Calculate the timestamp for the requested frame
            int64_t timestamp = req_position / (av_q2d(self->video_time_base) * source_fps);
            if (req_position <= 0)
                timestamp = 0;
            else if (self->first_pts != AV_NOPTS_VALUE)
                timestamp += self->first_pts;
            else if (context->start_time != AV_NOPTS_VALUE)
                timestamp += context->start_time;
            if (preseek && av_q2d(self->video_time_base) != 0)
                timestamp -= 2 / av_q2d(self->video_time_base);
            if (timestamp < 0)
                timestamp = 0;
            mlt_log_debug(MLT_PRODUCER_SERVICE(producer),
                          "seeking timestamp %" PRId64 " position " MLT_POSITION_FMT
                          " expected " MLT_POSITION_FMT " last_pos %" PRId64 "\n",
                          timestamp,
                          position,
                          self->video_expected,
                          self->last_position);

            // Seek to the timestamp
            self->video_codec->skip_loop_filter = AVDISCARD_NONREF;
            av_seek_frame(context, self->video_index, timestamp, AVSEEK_FLAG_BACKWARD);

            // flush any pictures still in decode buffer
            avcodec_flush_buffers(self->video_codec);
            self->video_send_result = 0;

            // let packets_worker know we handled EOF
            if (self->packets_thread_ret == AVERROR_EOF) {
                self->packets_thread_ret = 0;
            }

            // empty vpackets
            while (mlt_deque_count(self->vpackets) > 0) {
                AVPacket *tmp = (AVPacket *) mlt_deque_pop_front(self->vpackets);
                av_packet_free(&tmp);
            }

            pthread_cond_signal(&self->packets_cond);

            // Remove the cached info relating to the previous position
            self->current_position = POSITION_INVALID;
            self->last_position = POSITION_INVALID;
            av_frame_unref(self->video_frame);
        }
    }
    pthread_mutex_unlock(&self->packets_mutex);
    return paused;
}

/** Convert a frame position to a time code.
*/

static double producer_time_of_frame(mlt_producer producer, mlt_position position)
{
    return (double) position / mlt_producer_get_fps(producer);
}

// Collect information about all audio streams

static void get_audio_streams_info(producer_avformat self)
{
    // Fetch the audio format context
    AVFormatContext *context = self->audio_format;
    unsigned int i;

    for (i = 0; i < context->nb_streams; i++) {
        if (context->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) {
            AVCodecParameters *codec_params = context->streams[i]->codecpar;
            const AVCodec *codec = avcodec_find_decoder(codec_params->codec_id);

            // Setup the codec context
            AVCodecContext *codec_context = avcodec_alloc_context3(codec);
            if (!codec_context) {
                mlt_log_info(MLT_PRODUCER_SERVICE(self->parent),
                             "Failed to allocate the decoder context for stream #%d\n",
                             i);
                continue;
            }
            int ret = avcodec_parameters_to_context(codec_context, codec_params);
            if (ret < 0) {
                mlt_log_info(
                    MLT_PRODUCER_SERVICE(self->parent),
                    "Failed to copy decoder parameters to input decoder context for stream #%d\n",
                    i);
                continue;
            }

            // If we don't have a codec and we can't initialise it, we can't do much more...
            pthread_mutex_lock(&self->open_mutex);
            if (codec && avcodec_open2(codec_context, codec, NULL) >= 0) {
                self->audio_streams++;
                self->audio_max_stream = i;
#if HAVE_FFMPEG_CH_LAYOUT
                self->total_channels += codec_params->ch_layout.nb_channels;
                if (codec_params->ch_layout.nb_channels > self->max_channel)
                    self->max_channel = codec_params->ch_layout.nb_channels;
#else
                self->total_channels += codec_params->channels;
                if (codec_params->channels > self->max_channel)
                    self->max_channel = codec_params->channels;
#endif
                if (codec_params->sample_rate > self->max_frequency)
                    self->max_frequency = codec_params->sample_rate;
                avcodec_close(codec_context);
            }
            pthread_mutex_unlock(&self->open_mutex);
        }
    }
    mlt_log_verbose(NULL,
                    "[producer avformat] audio: total_streams %d max_stream %d total_channels %d "
                    "max_channels %d\n",
                    self->audio_streams,
                    self->audio_max_stream,
                    self->total_channels,
                    self->max_channel);
}

static mlt_audio_format pick_audio_format(int sample_fmt)
{
    switch (sample_fmt) {
    // interleaved
    case AV_SAMPLE_FMT_U8:
        return mlt_audio_u8;
    case AV_SAMPLE_FMT_S16:
        return mlt_audio_s16;
    case AV_SAMPLE_FMT_S32:
        return mlt_audio_s32le;
    case AV_SAMPLE_FMT_FLT:
        return mlt_audio_f32le;
    // planar - this producer converts planar to interleaved
    case AV_SAMPLE_FMT_U8P:
        return mlt_audio_u8;
    case AV_SAMPLE_FMT_S16P:
        return mlt_audio_s16;
    case AV_SAMPLE_FMT_S32P:
        return mlt_audio_s32le;
    case AV_SAMPLE_FMT_FLTP:
        return mlt_audio_f32le;
    default:
        return mlt_audio_none;
    }
}

/**
 * Handle deprecated pixel format (JPEG range in YUV420P for example).
 *
 * Replace pix_fmt with the official pixel format to use.
 * @return 0 if no pix_fmt replacement, 1 otherwise
 */
static int pick_av_pixel_format(int *pix_fmt, int full_range)
{
    switch (*pix_fmt) {
    case AV_PIX_FMT_YUV420P:
    case AV_PIX_FMT_YUVJ420P:
        *pix_fmt = full_range ? AV_PIX_FMT_YUVJ420P : AV_PIX_FMT_YUV420P;
        return 1;
    case AV_PIX_FMT_YUV411P:
    case AV_PIX_FMT_YUVJ411P:
        *pix_fmt = full_range ? AV_PIX_FMT_YUVJ411P : AV_PIX_FMT_YUV411P;
        return 1;
    case AV_PIX_FMT_YUV422P:
    case AV_PIX_FMT_YUVJ422P:
        *pix_fmt = full_range ? AV_PIX_FMT_YUVJ422P : AV_PIX_FMT_YUV422P;
        return 1;
    case AV_PIX_FMT_YUV444P:
    case AV_PIX_FMT_YUVJ444P:
        *pix_fmt = full_range ? AV_PIX_FMT_YUVJ444P : AV_PIX_FMT_YUV444P;
        return 1;
    case AV_PIX_FMT_YUV440P:
    case AV_PIX_FMT_YUVJ440P:
        *pix_fmt = full_range ? AV_PIX_FMT_YUVJ440P : AV_PIX_FMT_YUV440P;
        return 1;
    }
    return 0;
}

static void property_changed(mlt_service owner, producer_avformat self, char *name)
{
    if (self && name && self->parent) {
        mlt_properties properties = MLT_PRODUCER_PROPERTIES(self->parent);
        if (!strcmp("color_range", name)) {
            if (self->video_codec
                && !av_opt_set(self->video_codec,
                               name,
                               mlt_properties_get(properties, name),
                               AV_OPT_SEARCH_CHILDREN)) {
                if (self->full_range != (self->video_codec->color_range == AVCOL_RANGE_JPEG)) {
                    self->full_range = self->video_codec->color_range == AVCOL_RANGE_JPEG;
                    self->reset_image_cache = 1;
                }
            }
        } else if (!strcmp("force_full_range", name) || !strcmp("set.force_full_luma", name)) {
            if (self->full_range != mlt_properties_get_int(properties, name)) {
                self->full_range = mlt_properties_get_int(properties, name);
                self->reset_image_cache = 1;
            }
        } else if (!strcmp("force_progressive", name) || !strcmp("force_tff", name)) {
            self->reset_image_cache = 1;
        } else if (!strcmp("autorotate", name)) {
            self->autorotate = mlt_properties_get_int(properties, name);
            if (self->video_index != -1) {
                mlt_service_lock(MLT_PRODUCER_SERVICE(self->parent));
                avfilter_graph_free(&self->vfilter_graph);
                self->vfilter_out = NULL;
                self->rotation = 0.0;
                setup_filters(self);
                self->reset_image_cache = 1;
                mlt_service_unlock(MLT_PRODUCER_SERVICE(self->parent));
            }
        } else if (!strcmp("video_index", name) || !strcmp("vstream", name)) {
            if (mlt_properties_get_int(properties, "_probe_complete")) {
                mlt_properties_set_int(properties, "_probe_complete", 0);
            }
        }
    }
}

struct sliced_pix_fmt_conv_t
{
    int width, height, slice_w;
    AVFrame *frame;
    uint8_t *out_data[4];
    int out_stride[4];
    enum AVPixelFormat src_format, dst_format;
    const AVPixFmtDescriptor *src_desc, *dst_desc;
    int flags, src_colorspace, dst_colorspace, src_full_range, dst_full_range;
};

static int sliced_h_pix_fmt_conv_proc(int id, int idx, int jobs, void *cookie)
{
    uint8_t *out[4];
    const uint8_t *in[4];
    int in_stride[4], out_stride[4];
    int src_v_chr_pos = -513, dst_v_chr_pos = -513, ret, i, slice_x, slice_w, h, mul, field, slices,
        interlaced = 0;

    struct SwsContext *sws;
    struct sliced_pix_fmt_conv_t *ctx = (struct sliced_pix_fmt_conv_t *) cookie;

    interlaced = ctx->frame->interlaced_frame;
    field = (interlaced) ? (idx & 1) : 0;
    idx = (interlaced) ? (idx / 2) : idx;
    slices = (interlaced) ? (jobs / 2) : jobs;
    mul = (interlaced) ? 2 : 1;
    h = ctx->height >> !!interlaced;
    slice_w = ctx->slice_w;
    slice_x = slice_w * idx;
    slice_w = FFMIN(slice_w, ctx->width - slice_x);

    if (AV_PIX_FMT_YUV420P == ctx->src_format)
        src_v_chr_pos = (!interlaced) ? 128 : (!field) ? 64 : 192;

    if (AV_PIX_FMT_YUV420P == ctx->dst_format)
        dst_v_chr_pos = (!interlaced) ? 128 : (!field) ? 64 : 192;

    mlt_log_debug(NULL,
                  "%s:%d: [id=%d, idx=%d, jobs=%d], interlaced=%d, field=%d, slices=%d, mul=%d, "
                  "h=%d, slice_w=%d, slice_x=%d ctx->src_desc=[log2_chroma_h=%d, "
                  "log2_chroma_w=%d], src_v_chr_pos=%d, dst_v_chr_pos=%d\n",
                  __FUNCTION__,
                  __LINE__,
                  id,
                  idx,
                  jobs,
                  interlaced,
                  field,
                  slices,
                  mul,
                  h,
                  slice_w,
                  slice_x,
                  ctx->src_desc->log2_chroma_h,
                  ctx->src_desc->log2_chroma_w,
                  src_v_chr_pos,
                  dst_v_chr_pos);

    if (slice_w <= 0)
        return 0;

    sws = sws_alloc_context();

    av_opt_set_int(sws, "srcw", slice_w, 0);
    av_opt_set_int(sws, "srch", h, 0);
    av_opt_set_int(sws, "src_format", ctx->src_format, 0);
    av_opt_set_int(sws, "dstw", slice_w, 0);
    av_opt_set_int(sws, "dsth", h, 0);
    av_opt_set_int(sws, "dst_format", ctx->dst_format, 0);
    av_opt_set_int(sws, "sws_flags", ctx->flags, 0);

    av_opt_set_int(sws, "src_h_chr_pos", -513, 0);
    av_opt_set_int(sws, "src_v_chr_pos", src_v_chr_pos, 0);
    av_opt_set_int(sws, "dst_h_chr_pos", -513, 0);
    av_opt_set_int(sws, "dst_v_chr_pos", dst_v_chr_pos, 0);

    if ((ret = sws_init_context(sws, NULL, NULL)) < 0) {
        mlt_log_error(NULL, "%s:%d: sws_init_context failed, ret=%d\n", __FUNCTION__, __LINE__, ret);
        sws_freeContext(sws);
        return 0;
    }

    mlt_set_luma_transfer(sws,
                          ctx->src_colorspace,
                          ctx->dst_colorspace,
                          ctx->src_full_range,
                          ctx->dst_full_range);

#define PIX_DESC_BPP(DESC) (DESC.step)

    for (i = 0; i < 4; i++) {
        int in_offset = (AV_PIX_FMT_FLAG_PLANAR & ctx->src_desc->flags)
                            ? ((1 == i || 2 == i) ? (slice_x >> ctx->src_desc->log2_chroma_w)
                                                  : slice_x)
                            : ((0 == i) ? slice_x : 0);

        int out_offset = (AV_PIX_FMT_FLAG_PLANAR & ctx->dst_desc->flags)
                             ? ((1 == i || 2 == i) ? (slice_x >> ctx->dst_desc->log2_chroma_w)
                                                   : slice_x)
                             : ((0 == i) ? slice_x : 0);

        in_offset *= PIX_DESC_BPP(ctx->src_desc->comp[i]);
        out_offset *= PIX_DESC_BPP(ctx->dst_desc->comp[i]);

        in_stride[i] = ctx->frame->linesize[i] * mul;
        out_stride[i] = ctx->out_stride[i] * mul;

        in[i] = ctx->frame->data[i] + ctx->frame->linesize[i] * field + in_offset;
        out[i] = ctx->out_data[i] + ctx->out_stride[i] * field + out_offset;
    }

    sws_scale(sws, in, in_stride, 0, h, out, out_stride);

    sws_freeContext(sws);

    return 0;
}

static int convert_image_yuvp(producer_avformat self,
                              mlt_profile profile,
                              AVFrame *frame,
                              uint8_t *buffer,
                              mlt_image_format format,
                              int width,
                              int height,
                              int src_pix_fmt,
                              int dst_pix_fmt,
                              int dst_full_range)
{
    int result = self->yuv_colorspace;
    int flags = mlt_get_sws_flags(width, height, src_pix_fmt, width, height, dst_pix_fmt);
    struct SwsContext *context = sws_getContext(
        width, height, src_pix_fmt, width, height, dst_pix_fmt, flags, NULL, NULL, NULL);
    uint8_t *out_data[4];
    int out_stride[4];

    mlt_image_format_planes(format, width, height, buffer, out_data, out_stride);
    if (!mlt_set_luma_transfer(context,
                               self->yuv_colorspace,
                               profile->colorspace,
                               self->full_range,
                               dst_full_range))
        result = profile->colorspace;
    sws_scale(context,
              (const uint8_t *const *) frame->data,
              frame->linesize,
              0,
              height,
              out_data,
              out_stride);
    sws_freeContext(context);

    return result;
}

static void convert_image_rgb(producer_avformat self,
                              mlt_profile profile,
                              AVFrame *frame,
                              uint8_t *buffer,
                              mlt_image_format format,
                              int width,
                              int height,
                              int src_pix_fmt,
                              int dst_pix_fmt,
                              int dst_full_range)
{
    int flags = mlt_get_sws_flags(width, height, src_pix_fmt, width, height, dst_pix_fmt);
    uint8_t *out_data[4];
    int out_stride[4];

    if (src_pix_fmt == AV_PIX_FMT_YUV420P && frame->interlaced_frame) {
        // Perform field-aware conversion for 4:2:0
        int field_height = height / 2;
        const uint8_t *in_data[4];
        int in_stride[4];
        struct SwsContext *context = sws_getContext(width,
                                                    field_height,
                                                    src_pix_fmt,
                                                    width,
                                                    field_height,
                                                    dst_pix_fmt,
                                                    flags,
                                                    NULL,
                                                    NULL,
                                                    NULL);
        // libswscale wants the RGB colorspace to be SWS_CS_DEFAULT, which is = SWS_CS_ITU601.
        mlt_set_luma_transfer(context, self->yuv_colorspace, 601, self->full_range, 1);
        av_image_fill_arrays(out_data, out_stride, buffer, dst_pix_fmt, width, height, IMAGE_ALIGN);
        // Copy the input frame arrays
        for (int i = 0; i < 4; i++) {
            in_data[i] = frame->data[i];
            in_stride[i] = frame->linesize[i];
        }
        // Modify the strides to skip every other line
        for (int i = 0; i < 4; i++) {
            in_stride[i] *= 2;
            out_stride[i] *= 2;
        }
        // Convert the first field
        sws_scale(context, in_data, in_stride, 0, field_height, out_data, out_stride);
        // Offset the data to point at the second field
        for (int i = 0; i < 4; i++) {
            in_data[i] += in_stride[i] / 2;
            out_data[i] += out_stride[i] / 2;
        }
        // Convert the second field
        sws_scale(context, in_data, in_stride, 0, field_height, out_data, out_stride);
        sws_freeContext(context);
    } else {
        struct SwsContext *context = sws_getContext(
            width, height, src_pix_fmt, width, height, dst_pix_fmt, flags, NULL, NULL, NULL);
        av_image_fill_arrays(out_data, out_stride, buffer, dst_pix_fmt, width, height, IMAGE_ALIGN);
        // libswscale wants the RGB colorspace to be SWS_CS_DEFAULT, which is = SWS_CS_ITU601.
        mlt_set_luma_transfer(context, self->yuv_colorspace, 601, self->full_range, 1);
        sws_scale(context,
                  (const uint8_t *const *) frame->data,
                  frame->linesize,
                  0,
                  height,
                  out_data,
                  out_stride);
        sws_freeContext(context);
    }
}

// returns resulting YUV colorspace
static int convert_image(producer_avformat self,
                         AVFrame *frame,
                         uint8_t *buffer,
                         int pix_fmt,
                         mlt_image_format *format,
                         int width,
                         int height,
                         uint8_t **alpha,
                         int dst_full_range)
{
    mlt_profile profile = mlt_service_profile(MLT_PRODUCER_SERVICE(self->parent));
    int result = self->yuv_colorspace;

    mlt_log_timings_begin();

    mlt_log_debug(MLT_PRODUCER_SERVICE(self->parent),
                  "%s @ %dx%d space %d->%d\n",
                  mlt_image_format_name(*format),
                  width,
                  height,
                  self->yuv_colorspace,
                  profile->colorspace);

    // extract alpha from planar formats
    if ((pix_fmt == AV_PIX_FMT_YUVA420P || pix_fmt == AV_PIX_FMT_YUVA444P)
        && *format != mlt_image_rgba && frame->data[3] && frame->linesize[3]) {
        int i;
        uint8_t *src, *dst;

        dst = *alpha = mlt_pool_alloc(width * height);
        src = frame->data[3];

        for (i = 0; i < height; dst += width, src += frame->linesize[3], i++)
            memcpy(dst, src, FFMIN(width, frame->linesize[3]));
    }

    // Figure out source and destination pixel format
    int src_pix_fmt = pix_fmt;
    pick_av_pixel_format(&src_pix_fmt, self->full_range);
    int dst_pix_fmt = AV_PIX_FMT_NONE;
    switch (*format) {
    case mlt_image_yuv420p:
        dst_pix_fmt = AV_PIX_FMT_YUV420P;
        break;
    case mlt_image_yuv420p10:
        dst_pix_fmt = AV_PIX_FMT_YUV420P10LE;
        break;
    case mlt_image_yuv444p10:
        dst_pix_fmt = AV_PIX_FMT_YUV444P10LE;
        break;
    case mlt_image_yuv422p16:
        dst_pix_fmt = AV_PIX_FMT_YUV422P16LE;
        break;
    case mlt_image_rgb:
        dst_pix_fmt = AV_PIX_FMT_RGB24;
        break;
    case mlt_image_rgba:
        dst_pix_fmt = AV_PIX_FMT_RGBA;
        break;
    case mlt_image_none:
    case mlt_image_yuv422:
    case mlt_image_movit:
    case mlt_image_opengl_texture:
    case mlt_image_invalid:
        break;
    }

    // Convert
    if (mlt_image_rgb == *format || mlt_image_rgba == *format) {
        convert_image_rgb(self,
                          profile,
                          frame,
                          buffer,
                          *format,
                          width,
                          height,
                          src_pix_fmt,
                          dst_pix_fmt,
                          dst_full_range);
    } else if (dst_pix_fmt != AV_PIX_FMT_NONE) {
        result = convert_image_yuvp(self,
                                    profile,
                                    frame,
                                    buffer,
                                    *format,
                                    width,
                                    height,
                                    src_pix_fmt,
                                    dst_pix_fmt,
                                    dst_full_range);
    } else {
        int i, c;
        struct sliced_pix_fmt_conv_t ctx = {
            .width = width,
            .height = height,
            .frame = frame,
            .dst_format = AV_PIX_FMT_YUYV422,
            .src_colorspace = self->yuv_colorspace,
            .dst_colorspace = profile->colorspace,
            .src_full_range = self->full_range,
            .dst_full_range = dst_full_range,
        };
        ctx.src_format = (self->full_range && src_pix_fmt == AV_PIX_FMT_YUV422P)
                             ? AV_PIX_FMT_YUVJ422P
                             : src_pix_fmt;
        ctx.src_desc = av_pix_fmt_desc_get(ctx.src_format);
        ctx.dst_desc = av_pix_fmt_desc_get(ctx.dst_format);
        ctx.flags = mlt_get_sws_flags(width, height, ctx.src_format, width, height, ctx.dst_format);

        av_image_fill_arrays(ctx.out_data,
                             ctx.out_stride,
                             buffer,
                             ctx.dst_format,
                             width,
                             height,
                             IMAGE_ALIGN);

        int sliced = !getenv("MLT_AVFORMAT_SLICED_PIXFMT_DISABLE") && src_pix_fmt != ctx.dst_format;
        if (sliced) {
            ctx.slice_w = (width < 1000) ? (256 >> frame->interlaced_frame)
                                         : (512 >> frame->interlaced_frame);
        } else {
            ctx.slice_w = width;
        }

        c = (width + ctx.slice_w - 1) / ctx.slice_w;
        int last_slice_w = width - ctx.slice_w * (c - 1);

        if (sliced && (last_slice_w % 8) == 0
            && !(ctx.src_format == AV_PIX_FMT_YUV422P && last_slice_w % 16)) {
            c *= frame->interlaced_frame ? 2 : 1;
            mlt_slices_run_normal(c, sliced_h_pix_fmt_conv_proc, &ctx);
        } else {
            c = frame->interlaced_frame ? 2 : 1;
            ctx.slice_w = width;
            for (i = 0; i < c; i++)
                sliced_h_pix_fmt_conv_proc(i, i, c, &ctx);
        }

        result = profile->colorspace;
    }
    mlt_log_timings_end(NULL, __FUNCTION__);

    return result;
}

static void set_image_size(producer_avformat self, int *width, int *height)
{
#ifdef AVFILTER
    if (self->vfilter_out) {
        *width = av_buffersink_get_w(self->vfilter_out);
        *height = av_buffersink_get_h(self->vfilter_out);
    } else {
#endif
        double dar = mlt_profile_dar(mlt_service_profile(MLT_PRODUCER_SERVICE(self->parent)));
        *width = self->video_codec->width;
        // Workaround 1088 encodings missing cropping info.
        if (self->video_codec->height == 1088 && dar == 16.0 / 9.0)
            *height = 1080;
        else
            *height = self->video_codec->height;
#ifdef AVFILTER
    }
#endif
}

/** Allocate the image buffer and set it on the frame.
*/

static int allocate_buffer(mlt_frame frame,
                           AVCodecParameters *codec_params,
                           uint8_t **buffer,
                           mlt_image_format format,
                           int width,
                           int height)
{
    int size = 0;

    if (codec_params->width == 0 || codec_params->height == 0)
        return size;

    size = mlt_image_format_size(format, width, height, NULL);
    *buffer = mlt_pool_alloc(size);
    if (*buffer)
        mlt_frame_set_image(frame, *buffer, size, mlt_pool_release);
    else
        size = 0;

    return size;
}

static int ignore_send_packet_result(int result)
{
    return result >= 0 || result == AVERROR(EAGAIN) || result == AVERROR_EOF
           || result == AVERROR_INVALIDDATA || result == AVERROR(EINVAL);
}

static int is_album_art(producer_avformat self)
{
    return self->video_index >= 0
           && (self->video_format->streams[self->video_index]->disposition
               & AV_DISPOSITION_ATTACHED_PIC);
}

static void *packets_worker(void *param)
{
    producer_avformat self = param;
    AVPacket *pkt = av_packet_alloc();
    if (!pkt) {
        mlt_log_fatal(MLT_PRODUCER_SERVICE(self->parent), "av_packet_alloc failed\n");
        exit(EXIT_FAILURE);
    }

    pthread_mutex_lock(&self->packets_mutex);
    for (;;) {
    check_stop:
        if (self->packets_thread_stop) {
            av_packet_free(&pkt);
            pthread_mutex_unlock(&self->packets_mutex);
            return NULL;
        }

        if (mlt_deque_count(self->vpackets) >= 1 || self->packets_thread_ret < 0) {
            pthread_cond_wait(&self->packets_cond, &self->packets_mutex);
            goto check_stop;
        }

        int ret = av_read_frame(self->video_format, pkt);

        // don't bother signalling on EAGAIN
        if (ret != AVERROR(EAGAIN)) {
            self->packets_thread_ret = ret;

            if (ret == 0) {
                if (pkt->stream_index == self->video_index) {
                    mlt_deque_push_back(self->vpackets, av_packet_clone(pkt));
                } else if (!self->video_seekable && pkt->stream_index == self->audio_index
                           && !is_album_art(self)) {
                    mlt_deque_push_back(self->apackets, av_packet_clone(pkt));
                }
                av_packet_unref(pkt);
            } else if (ret != AVERROR_EOF) {
                mlt_log_verbose(MLT_PRODUCER_SERVICE(self->parent),
                                "av_read_frame returned error %d inside packets_worker\n",
                                ret);
            }

            pthread_cond_signal(&self->packets_cond);
        }
    }
}

static void init_cache(mlt_properties properties, mlt_cache *cache)
{
    // if cache size supplied by environment variable
    int cache_supplied = getenv("MLT_AVFORMAT_CACHE") != NULL;
    int cache_size = cache_supplied ? atoi(getenv("MLT_AVFORMAT_CACHE")) : 0;

    // cache size supplied via property
    if (mlt_properties_get(properties, "cache")) {
        cache_supplied = 1;
        cache_size = mlt_properties_get_int(properties, "cache");
    }
    if (mlt_properties_get_int(properties, "noimagecache")) {
        cache_supplied = 1;
        cache_size = 0;
    }
    // create cache if not disabled
    if (!cache_supplied || cache_size > 0)
        *cache = mlt_cache_init();
    // set cache size if supplied
    if (*cache && cache_supplied)
        mlt_cache_set_size(*cache, cache_size);
}

/** Get an image from a frame.
*/

static int producer_get_image(mlt_frame frame,
                              uint8_t **buffer,
                              mlt_image_format *format,
                              int *width,
                              int *height,
                              int writable)
{
    // Get the producer
    (void) writable; // unused
    producer_avformat self = mlt_frame_pop_service(frame);
    mlt_producer producer = self->parent;

    // Get the properties from the frame
    mlt_properties frame_properties = MLT_FRAME_PROPERTIES(frame);

    // Obtain the frame number of this frame
    mlt_position position = mlt_frame_original_position(frame);

    // Get the producer properties
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(producer);

    uint8_t *alpha = NULL;
    int got_picture = 0;
    int image_size = 0;
    const char *dst_color_range = mlt_properties_get(frame_properties, "consumer.color_range");
    int dst_full_range = mlt_image_full_range(dst_color_range);

    // if depth > 8 libswscale only changes range when scaling, not simple pix_fmt conversion
    const struct AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(self->video_codec->pix_fmt);
    if (pix_desc && pix_desc->nb_components > 0 && pix_desc->comp[0].depth > 8)
        dst_full_range = self->full_range;

    mlt_service_lock(MLT_PRODUCER_SERVICE(producer));
    pthread_mutex_lock(&self->video_mutex);
    mlt_log_timings_begin();

#ifdef AVFILTER
    if (self->autorotate && self->video_index != -1
        && get_rotation(properties, self->video_format->streams[self->video_index])
               != self->rotation) {
        // Rotation has changed. Clear any cached frames.
        self->reset_image_cache = 1;
    }
#endif

    if (self->reset_image_cache) {
        self->reset_image_cache = 0;
        mlt_cache_close(self->image_cache);
        self->image_cache = NULL;
        av_frame_free(&self->video_frame);
    }

    // Fetch the video format context
    AVFormatContext *context = self->video_format;
    AVCodecParameters *codec_params = NULL;
    if (!context)
        goto exit_get_image;

    // Get the video stream
    AVStream *stream = context->streams[self->video_index];
    codec_params = stream->codecpar;

    // Always use the image cache for album art.
    if (is_album_art(self))
        position = 0;

    // Get the image cache
    if (!self->image_cache) {
        init_cache(properties, &self->image_cache);
    }
    if (self->image_cache) {
        mlt_frame original = mlt_cache_get_frame(self->image_cache, position);
        if (original
            && (*format == mlt_image_none
                || *format == mlt_properties_get_int(MLT_FRAME_PROPERTIES(original), "format"))) {
            mlt_properties orig_props = MLT_FRAME_PROPERTIES(original);
            int size = 0;

            *buffer = mlt_frame_get_alpha_size(original, &size);
            if (*buffer)
                mlt_frame_set_alpha(frame, *buffer, size, NULL);
            *buffer = mlt_properties_get_data(orig_props, "image", &size);
            mlt_frame_set_image(frame, *buffer, size, NULL);
            mlt_properties_set_data(frame_properties,
                                    "avformat.image_cache",
                                    original,
                                    0,
                                    (mlt_destructor) mlt_frame_close,
                                    NULL);
            *format = mlt_properties_get_int(orig_props, "format");
            set_image_size(self, width, height);
            mlt_properties_pass_property(frame_properties, orig_props, "colorspace");
            mlt_properties_set_int(frame_properties, "full_range", dst_full_range);
            got_picture = 1;
            goto exit_get_image;
        } else {
            mlt_frame_close(original);
        }
    }
    // Cache miss

    // We may want to use the source fps if available
    double source_fps = mlt_properties_get_double(properties, "meta.media.frame_rate_num")
                        / mlt_properties_get_double(properties, "meta.media.frame_rate_den");

    // This is the physical frame position in the source
    int64_t req_position = (int64_t) (position / mlt_producer_get_fps(producer) * source_fps + 0.5);

    // Determines if we have to decode all frames in a sequence - when there temporal compression is used.
    const AVCodecDescriptor *descriptor = avcodec_descriptor_get(codec_params->codec_id);
    int must_decode = descriptor && !(descriptor->props & AV_CODEC_PROP_INTRA_ONLY);

    double delay = mlt_properties_get_double(properties, "video_delay");

    // Seek if necessary
    double speed = mlt_producer_get_speed(producer);
    int preseek = must_decode && self->video_codec->has_b_frames && speed >= 0.0 && speed <= 1.0;
    int paused = seek_video(self, position, req_position, preseek);

    // Seek might have reopened the file
    context = self->video_format;
    stream = context->streams[self->video_index];
    codec_params = stream->codecpar;

    // Only change the requested image format for special cases
#ifdef AVFILTER
    *format = pick_image_format(self->vfilter_out ? av_buffersink_get_format(self->vfilter_out)
                                                  : codec_params->format,
                                *format);
#else
    *format = pick_image_format(codec_params->format, *format);
#endif

    // Duplicate the last image if necessary
    if (self->video_frame && self->video_frame->linesize[0]
        && (self->pkt.stream_index == self->video_index)
        && (paused || self->current_position >= req_position)) {
        // Duplicate it
        set_image_size(self, width, height);
        if ((image_size = allocate_buffer(frame, codec_params, buffer, *format, *width, *height))) {
            int yuv_colorspace;
#if USE_HWACCEL
            yuv_colorspace = convert_image(self,
                                           self->video_frame,
                                           *buffer,
                                           self->video_frame->format,
                                           format,
                                           *width,
                                           *height,
                                           &alpha,
                                           dst_full_range);
#else
            yuv_colorspace = convert_image(self,
                                           self->video_frame,
                                           *buffer,
                                           codec_params->format,
                                           format,
                                           *width,
                                           *height,
                                           &alpha,
                                           dst_full_range);
#endif
            mlt_properties_set_int(frame_properties, "colorspace", yuv_colorspace);
            mlt_properties_set_int(frame_properties, "full_range", dst_full_range);
            got_picture = 1;
        }
    } else {
        int64_t int_position = 0;
        int decode_errors = 0;

        // Construct an AVFrame for YUV422 conversion
        if (!self->video_frame)
            self->video_frame = av_frame_alloc();
        else
            av_frame_unref(self->video_frame);

        if (!self->is_thread_init) {
            pthread_cond_init(&self->packets_cond, NULL);
            pthread_create(&self->packets_thread, NULL, packets_worker, self);
            self->is_thread_init = 1;
        }

        while (!got_picture && ignore_send_packet_result(self->video_send_result)) {
            if (self->video_send_result != AVERROR(EAGAIN)) {
                // Read a packet
                if (self->pkt.stream_index == self->video_index)
                    av_packet_unref(&self->pkt);
                av_init_packet(&self->pkt);
                pthread_mutex_lock(&self->packets_mutex);
                while (mlt_deque_count(self->vpackets) == 0 && self->packets_thread_ret == 0) {
                    pthread_cond_wait(&self->packets_cond, &self->packets_mutex);
                }

                if (self->packets_thread_ret == 0) {
                    AVPacket *tmp = (AVPacket *) mlt_deque_pop_front(self->vpackets);
                    av_packet_ref(&self->pkt, tmp);
                    av_packet_free(&tmp);
                    pthread_cond_signal(&self->packets_cond);
                } else {
                    if (self->packets_thread_ret == AVERROR_EOF) {
                        self->pkt.stream_index = self->video_index;
                    }

                    // notify packets_worker that we've seen the error
                    self->packets_thread_ret = 0;
                    pthread_cond_signal(&self->packets_cond);

                    if (!self->video_seekable && mlt_properties_get_int(properties, "reconnect")) {
                        // Try to reconnect to live sources by closing context and codecs,
                        // and letting next call to get_frame() reopen.
                        mlt_service_unlock(MLT_PRODUCER_SERVICE(producer));
                        prepare_reopen(self);
                        mlt_service_lock(MLT_PRODUCER_SERVICE(producer));
                        pthread_mutex_unlock(&self->packets_mutex);
                        goto exit_get_image;
                    }
                    if (!self->video_seekable
                        && mlt_properties_get_int(properties, "exit_on_disconnect")) {
                        mlt_log_fatal(MLT_PRODUCER_SERVICE(producer),
                                      "Exiting with error due to disconnected source.\n");
                        exit(EXIT_FAILURE);
                    }
                    // Send null packets to drain decoder.
                    self->pkt.size = 0;
                    self->pkt.data = NULL;
                }
                pthread_mutex_unlock(&self->packets_mutex);
            }

            // We only deal with video from the selected video_index
            if (self->pkt.stream_index == self->video_index) {
                int64_t pts = best_pts(self, self->pkt.pts, self->pkt.dts);
                if (pts != AV_NOPTS_VALUE) {
                    if (!self->video_seekable && self->first_pts == AV_NOPTS_VALUE)
                        self->first_pts = pts;
                    if (self->first_pts != AV_NOPTS_VALUE)
                        pts -= self->first_pts;
                    else if (context->start_time != AV_NOPTS_VALUE)
                        pts -= context->start_time;
                    int_position = (int64_t) ((av_q2d(self->video_time_base) * pts + delay)
                                                  * source_fps
                                              + 0.5);
                    if (int_position == self->last_position)
                        int_position = self->last_position + 1;
                }
                mlt_log_debug(MLT_PRODUCER_SERVICE(producer),
                              "V pkt.pts %" PRId64 " pkt.dts %" PRId64 " req_pos %" PRId64
                              " cur_pos %" PRId64 " pkt_pos %" PRId64 "\n",
                              self->pkt.pts,
                              self->pkt.dts,
                              req_position,
                              self->current_position,
                              int_position);

                // Make a dumb assumption on streams that contain wild timestamps
                if (llabs(req_position - int_position) > 999) {
                    mlt_log_verbose(MLT_PRODUCER_SERVICE(producer),
                                    " WILD TIMESTAMP: "
                                    "pkt.pts=[%" PRId64 "], pkt.dts=[%" PRId64
                                    "], req_position=[%" PRId64 "], "
                                    "current_position=[%" PRId64 "], int_position=[%" PRId64
                                    "], pts=[%" PRId64 "] \n",
                                    self->pkt.pts,
                                    self->pkt.dts,
                                    req_position,
                                    self->current_position,
                                    int_position,
                                    pts);
                    int_position = req_position;
                }
                self->last_position = int_position;

                // Decode the image
                if (must_decode || int_position >= req_position || !self->pkt.data) {
#if LIBAVCODEC_VERSION_MAJOR < 61
                    self->video_codec->reordered_opaque = int_position;
#endif
                    if (int_position >= req_position)
                        self->video_codec->skip_loop_filter = AVDISCARD_NONE;
                    self->video_send_result = avcodec_send_packet(self->video_codec, &self->pkt);
                    mlt_log_debug(MLT_PRODUCER_SERVICE(producer),
                                  "decoded video packet with size %d => %d\n",
                                  self->pkt.size,
                                  self->video_send_result);
                    // Note: decode may fail at the beginning of MPEGfile (B-frames referencing before first I-frame), so allow a few errors.
                    if (!ignore_send_packet_result(self->video_send_result)) {
                        mlt_log_warning(MLT_PRODUCER_SERVICE(producer),
                                        "video avcodec_send_packet failed with %d\n",
                                        self->video_send_result);
                    } else {
                        int error = avcodec_receive_frame(self->video_codec, self->video_frame);
                        if (error < 0) {
                            if (error != AVERROR(EAGAIN) && ++decode_errors > 10) {
                                mlt_log_warning(MLT_PRODUCER_SERVICE(producer),
                                                "video decoding error %d\n",
                                                error);
                                self->last_good_position = POSITION_INVALID;
                            }
                        } else {
#if USE_HWACCEL
                            if (self->hwaccel.device_ctx
                                && self->video_frame->format == self->hwaccel.pix_fmt) {
                                AVFrame *sw_video_frame = av_frame_alloc();
                                int transfer_data_result
                                    = av_hwframe_transfer_data(sw_video_frame, self->video_frame, 0);
                                if (transfer_data_result < 0) {
                                    mlt_log_error(MLT_PRODUCER_SERVICE(producer),
                                                  "av_hwframe_transfer_data() failed %d\n",
                                                  transfer_data_result);
                                    av_frame_free(&sw_video_frame);
                                    goto exit_get_image;
                                }
                                av_frame_copy_props(sw_video_frame, self->video_frame);
                                sw_video_frame->width = self->video_frame->width;
                                sw_video_frame->height = self->video_frame->height;

                                av_frame_unref(self->video_frame);
                                av_frame_move_ref(self->video_frame, sw_video_frame);
                                av_frame_free(&sw_video_frame);
                            }
#endif
                            got_picture = 1;
                            decode_errors = 0;
                        }
                    }
                }

                if (got_picture) {
                    // Get position of reordered frame
#if LIBAVCODEC_VERSION_MAJOR < 61
                    int_position = self->video_frame->reordered_opaque;
#endif
                    pts = best_pts(self, self->video_frame->pts, self->video_frame->pkt_dts);
                    if (pts != AV_NOPTS_VALUE) {
                        // Some streams are not marking their key frames even though
                        // there are I frames, and find_first_pts() fails as a result.
                        // Try to set first_pts here after getting pict_type.
                        if (self->first_pts == AV_NOPTS_VALUE
                            && (self->video_frame->key_frame
                                || self->video_frame->pict_type == AV_PICTURE_TYPE_I))
                            self->first_pts = pts;
                        if (self->first_pts != AV_NOPTS_VALUE)
                            pts -= self->first_pts;
                        else if (context->start_time != AV_NOPTS_VALUE)
                            pts -= context->start_time;
                        int_position = (int64_t) ((av_q2d(self->video_time_base) * pts + delay)
                                                      * source_fps
                                                  + 0.5);
                    }

                    if (int_position < req_position)
                        got_picture = 0;
                    else if (int_position >= req_position)
                        self->video_codec->skip_loop_filter = AVDISCARD_NONE;
                } else if (!self->pkt.data) // draining decoder with null packets
                {
                    self->video_send_result = -1;
                }
                mlt_log_debug(MLT_PRODUCER_SERVICE(producer),
                              " got_pic %d key %d send_result %d pkt_pos %" PRId64 "\n",
                              got_picture,
                              self->pkt.flags & AV_PKT_FLAG_KEY,
                              self->video_send_result,
                              int_position);
            }

            // Now handle the picture if we have one
            if (got_picture) {
                // Detect and correct scan type
                if (mlt_properties_get(properties, "force_progressive")) {
                    self->progressive = !!mlt_properties_get_int(properties, "force_progressive");
                } else if (self->video_frame && codec_params) {
                    self->progressive = !self->video_frame->interlaced_frame
                                        && (codec_params->field_order == AV_FIELD_PROGRESSIVE
                                            || codec_params->field_order == AV_FIELD_UNKNOWN);
                } else {
                    self->progressive = 0;
                }
                self->video_frame->interlaced_frame = !self->progressive;
                // Detect and correct field order
                if (mlt_properties_get(properties, "force_tff")) {
                    self->top_field_first = !!mlt_properties_get_int(properties, "force_tff");
                } else {
                    self->top_field_first = self->video_frame->top_field_first
                                            || codec_params->field_order == AV_FIELD_TT
                                            || codec_params->field_order == AV_FIELD_TB;
                }
                self->video_frame->top_field_first = self->top_field_first;
#ifdef AVFILTER
                if ((self->autorotate || mlt_properties_get(properties, "filtergraph"))
                    && !setup_filters(self) && self->vfilter_graph) {
                    int ret = av_buffersrc_add_frame(self->vfilter_in, self->video_frame);
                    if (ret < 0) {
                        got_picture = 0;
                        break;
                    }
                    while (ret >= 0) {
                        ret = av_buffersink_get_frame_flags(self->vfilter_out, self->video_frame, 0);
                        if (ret < 0) {
                            ret = 0;
                            break;
                        }
                    }
                }
#endif
                set_image_size(self, width, height);
                if ((image_size
                     = allocate_buffer(frame, codec_params, buffer, *format, *width, *height))) {
                    int yuv_colorspace;
#if USE_HWACCEL
                    // not sure why this is really needed, but doesn't seem to work otherwise
                    yuv_colorspace = convert_image(self,
                                                   self->video_frame,
                                                   *buffer,
                                                   self->video_frame->format,
                                                   format,
                                                   *width,
                                                   *height,
                                                   &alpha,
                                                   dst_full_range);
#else
                    yuv_colorspace = convert_image(self,
                                                   self->video_frame,
                                                   *buffer,
                                                   codec_params->format,
                                                   format,
                                                   *width,
                                                   *height,
                                                   &alpha,
                                                   dst_full_range);
#endif
                    mlt_properties_set_int(frame_properties, "colorspace", yuv_colorspace);
                    mlt_properties_set_int(frame_properties, "full_range", dst_full_range);
                    self->current_position = int_position;
                } else {
                    got_picture = 0;
                }
            }

            // Free packet data if not video and not live audio packet
            if (self->pkt.stream_index != self->video_index
                && !(!self->video_seekable && self->pkt.stream_index == self->audio_index))
                av_packet_unref(&self->pkt);
        }
    }

    // set alpha
    if (alpha)
        mlt_frame_set_alpha(frame, alpha, (*width) * (*height), mlt_pool_release);

    if (image_size > 0) {
        mlt_properties_set_int(frame_properties, "format", *format);
        // Cache the image for rapid repeated access.
        if (self->image_cache) {
            if (is_album_art(self)) {
                mlt_position original_pos = mlt_frame_original_position(frame);
                mlt_properties_set_position(frame_properties, "original_position", 0);
                mlt_cache_put_frame_image(self->image_cache, frame);
                mlt_properties_set_position(frame_properties, "original_position", original_pos);
            } else {
                mlt_cache_put_frame_image(self->image_cache, frame);
            }
        }
        // Clone frame for error concealment.
        if (self->current_position >= self->last_good_position) {
            self->last_good_position = self->current_position;
            if (self->last_good_frame)
                mlt_frame_close(self->last_good_frame);
            self->last_good_frame = mlt_frame_clone(frame, 1);
        }
    } else if (self->last_good_frame) {
        // Use last known good frame if there was a decoding failure.
        mlt_frame original = mlt_frame_clone(self->last_good_frame, 1);
        mlt_properties orig_props = MLT_FRAME_PROPERTIES(original);
        int size = 0;

        *buffer = mlt_frame_get_alpha_size(original, &size);
        if (*buffer)
            mlt_frame_set_alpha(frame, *buffer, size, NULL);
        *buffer = mlt_properties_get_data(orig_props, "image", &size);
        mlt_frame_set_image(frame, *buffer, size, NULL);
        mlt_properties_set_data(frame_properties,
                                "avformat.conceal_error",
                                original,
                                0,
                                (mlt_destructor) mlt_frame_close,
                                NULL);
        *format = mlt_properties_get_int(orig_props, "format");
        set_image_size(self, width, height);
        got_picture = 1;
    }

    // Regardless of speed, we expect to get the next frame (cos we ain't too bright)
    self->video_expected = position + 1;

exit_get_image:
    pthread_mutex_unlock(&self->video_mutex);

    mlt_properties_set_int(frame_properties, "progressive", self->progressive);
    mlt_properties_set_int(frame_properties, "top_field_first", self->top_field_first);

    // Set immutable properties of the selected track's (or overridden) source attributes.
    mlt_properties_set_int(properties, "meta.media.top_field_first", self->top_field_first);
    mlt_properties_set_int(properties, "meta.media.progressive", self->progressive);
    mlt_properties_set_int(properties, "_probe_complete", 1);
    mlt_service_unlock(MLT_PRODUCER_SERVICE(producer));

    mlt_log_timings_end(NULL, __FUNCTION__);

    return !got_picture;
}

/** Process properties as AVOptions and apply to AV context obj
*/

static void apply_properties(void *obj, mlt_properties properties, int flags)
{
    int i;
    int count = mlt_properties_count(properties);
    for (i = 0; i < count; i++) {
        const char *opt_name = mlt_properties_get_name(properties, i);
        int search_flags = AV_OPT_SEARCH_CHILDREN;
        const AVOption *opt = av_opt_find(obj, opt_name, NULL, flags, search_flags);
        if (opt_name && mlt_properties_get(properties, opt_name) && strcmp(opt_name, "seekable")) {
            if (opt)
                av_opt_set(obj, opt_name, mlt_properties_get(properties, opt_name), search_flags);
        }
    }
}

/** Initialize the video codec context.
 */

static int video_codec_init(producer_avformat self, int index, mlt_properties properties)
{
    // Initialise the codec if necessary
    if (!self->video_codec) {
        // Get the video stream
        AVStream *stream = self->video_format->streams[index];

        // Get codec context
        AVCodecParameters *codec_params = stream->codecpar;

        // Find the codec
        const AVCodec *codec = avcodec_find_decoder(codec_params->codec_id);
        if (mlt_properties_get(properties, "vcodec")) {
            if (!(codec = avcodec_find_decoder_by_name(mlt_properties_get(properties, "vcodec"))))
                codec = avcodec_find_decoder(codec_params->codec_id);
        } else if (codec_params->codec_id == AV_CODEC_ID_VP9) {
            if (!(codec = avcodec_find_decoder_by_name("libvpx-vp9")))
                codec = avcodec_find_decoder(codec_params->codec_id);
        } else if (codec_params->codec_id == AV_CODEC_ID_VP8) {
            if (!(codec = avcodec_find_decoder_by_name("libvpx")))
                codec = avcodec_find_decoder(codec_params->codec_id);
        }

        // Setup the codec context
        AVCodecContext *codec_context = avcodec_alloc_context3(codec);
        if (!codec_context) {
            mlt_log_error(MLT_PRODUCER_SERVICE(self->parent),
                          "Failed to allocate the decoder context for video stream #%d\n",
                          index);
            self->video_index = -1;
            return 0;
        }
        int ret = avcodec_parameters_to_context(codec_context, codec_params);
        if (ret < 0) {
            mlt_log_error(
                MLT_PRODUCER_SERVICE(self->parent),
                "Failed to copy decoder parameters to input decoder context for video stream #%d\n",
                index);
            self->video_index = -1;
            return 0;
        }

        // Initialise multi-threading
        int thread_count = mlt_properties_get_int(properties, "threads");
        if (thread_count == 0 && getenv("MLT_AVFORMAT_THREADS"))
            thread_count = atoi(getenv("MLT_AVFORMAT_THREADS"));
        if (thread_count >= 0)
            codec_context->thread_count = thread_count;

        // fix lowres if set too high
        int lowres = mlt_properties_get_int(properties, "lowres");
        if (lowres > codec_context->codec->max_lowres) {
            mlt_log_debug(MLT_PRODUCER_SERVICE(self->parent),
                          "clamping lowres=%i to max_lowres=%i\n",
                          lowres,
                          codec_context->codec->max_lowres);
            mlt_properties_set_int(properties, "lowres", codec_context->codec->max_lowres);
        }

#if USE_HWACCEL
        if (self->hwaccel.device_type == AV_HWDEVICE_TYPE_NONE
            || self->hwaccel.pix_fmt == AV_PIX_FMT_NONE) {
            mlt_log_debug(MLT_PRODUCER_SERVICE(self->parent),
                          "missing hwaccel parameters. skipping hardware initialization\n");
            goto skip_hwaccel;
        }

        int found_hw_pix_fmt = 0, i;
        for (i = 0;; i++) {
            const AVCodecHWConfig *config = avcodec_get_hw_config(codec, i);
            if (!config)
                break;

            if (config->methods & AV_CODEC_HW_CONFIG_METHOD_HW_DEVICE_CTX
                && config->device_type == self->hwaccel.device_type
                && config->pix_fmt == self->hwaccel.pix_fmt) {
                found_hw_pix_fmt = 1;
                break;
            }
        }

        if (found_hw_pix_fmt) {
            av_buffer_unref(&self->hwaccel.device_ctx);
            int ret = av_hwdevice_ctx_create(&self->hwaccel.device_ctx,
                                             self->hwaccel.device_type,
                                             self->hwaccel.device,
                                             NULL,
                                             0);
            if (ret >= 0) {
                codec_context->hw_device_ctx = av_buffer_ref(self->hwaccel.device_ctx);
                mlt_log_info(MLT_PRODUCER_SERVICE(self->parent),
                             "av_hwdevice_ctx_create() success %d\n",
                             codec_context->pix_fmt);
            } else {
                mlt_log_warning(MLT_PRODUCER_SERVICE(self->parent),
                                "av_hwdevice_ctx_create() failed %d\n",
                                ret);
            }
        } else {
            mlt_log_warning(MLT_PRODUCER_SERVICE(self->parent), "failed to find hw_pix_fmt\n");
        }

    skip_hwaccel:
#endif
        // If we don't have a codec and we can't initialise it, we can't do much more...
        pthread_mutex_lock(&self->open_mutex);
        if (codec && avcodec_open2(codec_context, codec, NULL) >= 0) {
            // Switch to the native vp8/vp9 decoder if not yuva420p
            if (codec_params->format != AV_PIX_FMT_YUVA420P
                && !mlt_properties_get(properties, "vcodec")
                && (!strcmp(codec->name, "libvpx") || !strcmp(codec->name, "libvpx-vp9"))) {
                codec = avcodec_find_decoder(codec_params->codec_id);
                if (codec && avcodec_open2(codec_context, codec, NULL) < 0) {
                    self->video_index = -1;
                    pthread_mutex_unlock(&self->open_mutex);
                    return 0;
                }
            }
            // Now store the codec with its destructor
            self->video_codec = codec_context;
        } else {
            // Remember that we can't use this later
            self->video_index = -1;
            pthread_mutex_unlock(&self->open_mutex);
            return 0;
        }
        pthread_mutex_unlock(&self->open_mutex);

        // Process properties as AVOptions
        apply_properties(codec_context,
                         properties,
                         AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM);
        if (codec && codec->priv_class && codec_context->priv_data)
            apply_properties(codec_context->priv_data,
                             properties,
                             AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM);

        // Reset some image properties
        mlt_properties_set_int(properties, "width", codec_params->width);
        mlt_properties_set_int(properties, "height", codec_params->height);
        get_aspect_ratio(properties, stream, codec_params);

        AVRational frame_rate = guess_frame_rate(self, stream);
        self->video_time_base = stream->time_base;
        if (mlt_properties_get(properties, "force_fps")) {
            AVRational force_fps = av_d2q(mlt_properties_get_double(properties, "force_fps"), 1024);
            self->video_time_base = av_mul_q(stream->time_base, av_div_q(frame_rate, force_fps));
            frame_rate = force_fps;
        }
        mlt_properties_set_int(properties, "meta.media.frame_rate_num", frame_rate.num);
        mlt_properties_set_int(properties, "meta.media.frame_rate_den", frame_rate.den);

        // Cover art is a single image at 90000 fps, which is not seekable.
        if (stream->disposition & AV_DISPOSITION_ATTACHED_PIC)
            self->video_seekable = 0;

        // Set the YUV colorspace from override or detect
        self->yuv_colorspace = mlt_properties_get_int(properties, "force_colorspace");
        if (!self->yuv_colorspace) {
            switch (self->video_codec->colorspace) {
            case AVCOL_SPC_SMPTE240M:
                self->yuv_colorspace = 240;
                break;
            case AVCOL_SPC_BT470BG:
            case AVCOL_SPC_SMPTE170M:
                self->yuv_colorspace = 601;
                break;
            case AVCOL_SPC_BT709:
                self->yuv_colorspace = 709;
                break;
            case AVCOL_SPC_BT2020_NCL:
                self->yuv_colorspace = 2020;
                break;
            case AVCOL_SPC_BT2020_CL:
                self->yuv_colorspace = 2021;
                break;
            default:
                // This is a heuristic Charles Poynton suggests in "Digital Video and HDTV"
                self->yuv_colorspace = self->video_codec->width * self->video_codec->height > 750000
                                           ? 709
                                           : 601;
                break;
            }
        }
        // Let apps get chosen colorspace
        mlt_properties_set_int(properties, "meta.media.colorspace", self->yuv_colorspace);

        // Get the color transfer characteristic (gamma).
        self->color_trc = mlt_properties_get_int(properties, "force_color_trc");
        if (!self->color_trc)
            self->color_trc = self->video_codec->color_trc;
        mlt_properties_set_int(properties, "meta.media.color_trc", self->color_trc);

        // Get the RGB color primaries.
        switch (self->video_codec->color_primaries) {
        case AVCOL_PRI_BT470BG:
            self->color_primaries = 601625;
            break;
        case AVCOL_PRI_SMPTE170M:
        case AVCOL_PRI_SMPTE240M:
            self->color_primaries = 601525;
            break;
        case AVCOL_PRI_BT2020:
            self->color_primaries = 2020;
            break;
        case AVCOL_PRI_BT709:
        case AVCOL_PRI_UNSPECIFIED:
        default:
            self->color_primaries = 709;
            break;
        }

        mlt_properties_set_int(properties,
                               "meta.media.has_b_frames",
                               self->video_codec->has_b_frames);

        self->full_range = codec_context->color_range == AVCOL_RANGE_JPEG;
        if (mlt_properties_get(properties, "force_full_range")) {
            self->full_range = mlt_properties_get_int(properties, "force_full_range");
        } else if (mlt_properties_get(properties, "set.force_full_luma")) { // deprecated
            self->full_range = mlt_properties_get_int(properties, "set.force_full_luma");
        }
    }
    return self->video_index > -1;
}

static int pick_video_stream(producer_avformat self)
{
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(self->parent);
    int absolute_index;

    if (self->video_format && mlt_properties_get(properties, "vstream")) {
        // Get the relative stream index
        absolute_index = absolute_stream_index(self->video_format,
                                               AVMEDIA_TYPE_VIDEO,
                                               mlt_properties_get_int(properties, "vstream"));
    } else {
        // Failover to the absolute index
        absolute_index = mlt_properties_get_int(properties, "video_index");
        if (self->video_format) {
            // Compute the relative stream index
            mlt_properties_set_int(properties,
                                   "vstream",
                                   relative_stream_index(self->video_format,
                                                         AVMEDIA_TYPE_VIDEO,
                                                         absolute_index));
        }
    }
    if (mlt_properties_get_int(properties, "video_index") != absolute_index) {
        // Update the absolute index
        mlt_properties_set_int(properties, "video_index", absolute_index);
        self->video_index = absolute_index;
    }
    return absolute_index;
}

/** Set up video handling.
*/

static void producer_set_up_video(producer_avformat self, mlt_frame frame)
{
    // Get the producer
    mlt_producer producer = self->parent;

    // Get the properties
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(producer);

    // Fetch the video format context
    AVFormatContext *context = self->video_format;

    // Get the video stream index
    int index = mlt_properties_get(properties, "vstream")
                    ? mlt_properties_get_int(properties, "vstream")
                    : mlt_properties_get_int(properties, "video_index");

    int unlock_needed = 0;

    // Reopen the file if necessary
    if (!context && index > -1) {
        unlock_needed = 1;
        pthread_mutex_lock(&self->video_mutex);
        producer_open(self,
                      mlt_service_profile(MLT_PRODUCER_SERVICE(producer)),
                      mlt_properties_get(properties, "resource"),
                      0,
                      0);
        context = self->video_format;
    }
    index = pick_video_stream(self);

    // Exception handling for video_index
    if (context && index >= (int) context->nb_streams) {
        // Get the last video stream
        for (index = context->nb_streams - 1;
             index >= 0 && context->streams[index]->codecpar->codec_type != AVMEDIA_TYPE_VIDEO;
             index--)
            ;
        mlt_properties_set_int(properties, "video_index", index);
        mlt_properties_set_int(properties,
                               "vstream",
                               relative_stream_index(context, AVMEDIA_TYPE_VIDEO, index));
    }
    if (context && index > -1
        && context->streams[index]->codecpar->codec_type != AVMEDIA_TYPE_VIDEO) {
        // Invalidate the video stream
        index = -1;
        mlt_properties_set_int(properties, "video_index", index);
        mlt_properties_set_int(properties,
                               "vstream",
                               relative_stream_index(context, AVMEDIA_TYPE_VIDEO, index));
    }

    // Update the video properties if the index changed
    if (context && index > -1 && index != self->video_index) {
        // Reset the video properties if the index changed
        self->video_index = index;
        mlt_properties_set_int(properties, "_probe_complete", 0);
        pthread_mutex_lock(&self->open_mutex);
        avcodec_free_context(&self->video_codec);
        set_up_discard(self, self->audio_index, index);
        pthread_mutex_unlock(&self->open_mutex);
    }

    // Get the frame properties
    mlt_properties frame_properties = MLT_FRAME_PROPERTIES(frame);

    // Get the codec
    if (context && index > -1 && video_codec_init(self, index, properties)) {
        // Set the frame properties
        double force_aspect_ratio = mlt_properties_get_double(properties, "force_aspect_ratio");
        double aspect_ratio = (force_aspect_ratio > 0.0)
                                  ? force_aspect_ratio
                                  : mlt_properties_get_double(properties, "aspect_ratio");

        // Set the width and height
        double dar = mlt_profile_dar(mlt_service_profile(MLT_PRODUCER_SERVICE(producer)));
        double theta = self->autorotate
                           ? get_rotation(properties, self->video_format->streams[index])
                           : 0.0;
        if (fabs(theta - 90.0) < 1.0 || fabs(theta - 270.0) < 1.0) {
            // Workaround 1088 encodings missing cropping info.
            if (self->video_codec->height == 1088 && dar == 16.0 / 9.0) {
                mlt_properties_set_int(frame_properties, "width", 1080);
                mlt_properties_set_int(properties, "meta.media.width", 1080);
                mlt_properties_set_int(properties, "width", 1080);
            } else {
                mlt_properties_set_int(frame_properties, "width", self->video_codec->height);
                mlt_properties_set_int(properties, "meta.media.width", self->video_codec->height);
                mlt_properties_set_int(properties, "width", self->video_codec->height);
            }
            mlt_properties_set_int(frame_properties, "height", self->video_codec->width);
            mlt_properties_set_int(properties, "meta.media.height", self->video_codec->width);
            mlt_properties_set_int(properties, "height", self->video_codec->width);
            aspect_ratio = (force_aspect_ratio > 0.0) ? force_aspect_ratio : 1.0 / aspect_ratio;
            mlt_properties_set_double(frame_properties, "aspect_ratio", 1.0 / aspect_ratio);
        } else {
            mlt_properties_set_int(frame_properties, "width", self->video_codec->width);
            mlt_properties_set_int(properties, "meta.media.width", self->video_codec->width);
            mlt_properties_set_int(properties, "width", self->video_codec->width);
            // Workaround 1088 encodings missing cropping info.
            if (self->video_codec->height == 1088 && dar == 16.0 / 9.0) {
                mlt_properties_set_int(frame_properties, "height", 1080);
                mlt_properties_set_int(properties, "meta.media.height", 1080);
                mlt_properties_set_int(properties, "height", 1080);
            } else {
                mlt_properties_set_int(frame_properties, "height", self->video_codec->height);
                mlt_properties_set_int(properties, "meta.media.height", self->video_codec->height);
                mlt_properties_set_int(properties, "height", self->video_codec->height);
            }
            mlt_properties_set_double(frame_properties, "aspect_ratio", aspect_ratio);
            mlt_properties_set_double(properties, "meta.media.aspect_ratio", aspect_ratio);
        }
        mlt_properties_set_int(frame_properties, "colorspace", self->yuv_colorspace);
        mlt_properties_set_int(frame_properties, "color_trc", self->color_trc);
        mlt_properties_set_int(frame_properties, "color_primaries", self->color_primaries);
        // full_range is the current state of frame
        mlt_properties_set_int(frame_properties, "full_range", self->full_range);
        // "full_luma" is deprecated but keep this for backwards compatibility for kdenlive
        mlt_properties_set_int(frame_properties, "full_luma", self->full_range);
        // meta.media.color_range is the range of this producer per video_index regardless of override
        mlt_properties_set(properties, "meta.media.color_range", self->full_range ? "full" : "mpeg");

        // Add our image operation
        mlt_frame_push_service(frame, self);
        mlt_frame_push_get_image(frame, producer_get_image);
    } else {
        // If something failed, use test card image
        mlt_properties_set_int(frame_properties, "test_image", 1);
    }
    if (unlock_needed)
        pthread_mutex_unlock(&self->video_mutex);
}

static int seek_audio(producer_avformat self, mlt_position position, double timecode)
{
    int paused = 0;

    pthread_mutex_lock(&self->packets_mutex);

    // Seek if necessary
    if (self->seekable && (position != self->audio_expected || self->last_position < 0)) {
        if (self->last_position == POSITION_INITIAL) {
            int video_index = self->video_index;
            if (video_index == -1)
                video_index = first_video_index(self);
            if (self->first_pts == AV_NOPTS_VALUE)
                find_first_pts(self, video_index);
        }

        if (position + 1 == self->audio_expected
            && mlt_properties_get_int(MLT_PRODUCER_PROPERTIES(self->parent), "mute_on_pause")) {
            // We're paused - silence required
            paused = 1;
        } else if (position < self->audio_expected || position - self->audio_expected >= 12) {
            AVFormatContext *context = self->audio_format;
            int64_t timestamp = llrint(timecode * AV_TIME_BASE);
            if (context->start_time != AV_NOPTS_VALUE)
                timestamp += context->start_time;
            if (timestamp < 0)
                timestamp = 0;

            // Set to the real timecode
            if (av_seek_frame(context, -1, timestamp, AVSEEK_FLAG_BACKWARD) != 0)
                paused = 1;

            // Clear the usage in the audio buffer
            int i = MAX_AUDIO_STREAMS + 1;
            while (--i)
                self->audio_used[i - 1] = 0;
        }
    }
    pthread_mutex_unlock(&self->packets_mutex);
    return paused;
}

static int sample_bytes(AVCodecContext *context)
{
    return av_get_bytes_per_sample(context->sample_fmt);
}

static void planar_to_interleaved(
    uint8_t *dest, AVFrame *src, int samples, int channels, int bytes_per_sample)
{
    int s, c;
    for (s = 0; s < samples; s++) {
        for (c = 0; c < channels; c++) {
            if (c < AV_NUM_DATA_POINTERS)
                memcpy(dest, &src->data[c][s * bytes_per_sample], bytes_per_sample);
            dest += bytes_per_sample;
        }
    }
}

static int decode_audio(producer_avformat self,
                        int *ignore,
                        const AVPacket *pkt,
                        int samples,
                        double timecode,
                        double fps)
{
    // Fetch the audio_format
    AVFormatContext *context = self->audio_format;

    // Get the current stream index
    int index = pkt->stream_index;

    // Get codec context
    AVCodecContext *codec_context = self->audio_codec[index];

    // Obtain the audio buffers
    uint8_t *audio_buffer = self->audio_buffer[index];
#if HAVE_FFMPEG_CH_LAYOUT
    int channels = codec_context->ch_layout.nb_channels;
#else
    int channels = codec_context->channels;
#endif
    int audio_used = self->audio_used[index];
    int audio_used_at_start = audio_used;
    int ret = 0;
    int discarded = 1;
    int sizeof_sample = sample_bytes(codec_context);

    // Decode the audio
    if (!self->audio_frame)
        self->audio_frame = av_frame_alloc();
    else
        av_frame_unref(self->audio_frame);
    int error = avcodec_send_packet(codec_context, pkt);
    mlt_log_debug(MLT_PRODUCER_SERVICE(self->parent),
                  "decoded audio packet with size %d => %d\n",
                  pkt->size,
                  error);
    if (!ignore_send_packet_result(error)) {
        mlt_log_warning(MLT_PRODUCER_SERVICE(self->parent),
                        "audio avcodec_send_packet failed with %d\n",
                        error);
    } else
        while (!error) {
            error = avcodec_receive_frame(codec_context, self->audio_frame);
            if (error) {
                if (error != AVERROR(EAGAIN)) {
                    mlt_log_warning(MLT_PRODUCER_SERVICE(self->parent),
                                    "audio decoding error %d\n",
                                    error);
                }
            } else {
                // Figure out how many samples will be needed after resampling
                int convert_samples = self->audio_frame->nb_samples;
#if HAVE_FFMPEG_CH_LAYOUT
                channels = codec_context->ch_layout.nb_channels;
#else
                channels = codec_context->channels;
#endif
                ret += convert_samples * channels * sizeof_sample;

                // Resize audio buffer to prevent overflow
                if ((audio_used + convert_samples) * channels * sizeof_sample
                    > self->audio_buffer_size[index]) {
                    self->audio_buffer_size[index] = (audio_used + convert_samples * 2) * channels
                                                     * sizeof_sample;
                    audio_buffer = self->audio_buffer[index]
                        = mlt_pool_realloc(audio_buffer, self->audio_buffer_size[index]);
                }
                uint8_t *dest = &audio_buffer[audio_used * channels * sizeof_sample];
                switch (codec_context->sample_fmt) {
                case AV_SAMPLE_FMT_U8P:
                case AV_SAMPLE_FMT_S16P:
                case AV_SAMPLE_FMT_S32P:
                case AV_SAMPLE_FMT_FLTP:
                    planar_to_interleaved(dest,
                                          self->audio_frame,
                                          convert_samples,
                                          channels,
                                          sizeof_sample);
                    break;
                default: {
                    int data_size = av_samples_get_buffer_size(NULL,
                                                               channels,
                                                               self->audio_frame->nb_samples,
                                                               codec_context->sample_fmt,
                                                               1);
                    // Straight copy to audio buffer
                    memcpy(dest, self->audio_frame->data[0], data_size);
                }
                }
                audio_used += convert_samples;
                discarded = 0;
            }
        }

    // Handle ignore
    if (*ignore > 0 && audio_used) {
        int n = FFMIN(audio_used, *ignore);
        *ignore -= n;
        audio_used -= n;
        audio_used_at_start -= n;
        memmove(audio_buffer,
                &audio_buffer[n * channels * sizeof_sample],
                audio_used * channels * sizeof_sample);
    }

    // If we're behind, ignore this packet
    // Skip this on non-seekable, audio-only inputs.
    if (!discarded && pkt->pts >= 0 && (self->seekable || self->video_format) && *ignore == 0
        && audio_used > samples / 2) {
        double timebase = av_q2d(context->streams[index]->time_base);
        int64_t pts_offset = lrint((double) audio_used_at_start / timebase
                                   / (double) codec_context->sample_rate);
        int64_t pts = pkt->pts - pts_offset;
        if (self->first_pts != AV_NOPTS_VALUE && self->video_index != -1)
            pts -= av_rescale_q(self->first_pts,
                                context->streams[self->video_index]->time_base,
                                context->streams[index]->time_base);
        else if (self->first_pts != AV_NOPTS_VALUE)
            pts -= self->first_pts;
        else if (context->start_time != AV_NOPTS_VALUE && self->video_index != -1)
            pts -= av_rescale_q(context->start_time,
                                AV_TIME_BASE_Q,
                                context->streams[index]->time_base);
        int64_t int_position = llrint(timebase * pts * fps);
        int64_t req_position = llrint(timecode * fps);
        int64_t req_pts = floor(timecode / timebase);

        mlt_log_debug(MLT_PRODUCER_SERVICE(self->parent),
                      "A pkt.pts %" PRId64 " pkt->dts %" PRId64 " req_pos %" PRId64
                      " cur_pos %" PRId64 " pkt_pos %" PRId64 "\n",
                      pkt->pts,
                      pkt->dts,
                      req_position,
                      self->current_position,
                      int_position);

        if (self->seekable || int_position > 0) {
            int64_t ahead_threshold = 2;
            if (codec_context->codec_id == AV_CODEC_ID_WMAPRO) {
                // WMAPro needs more tolerance for sync detection
                ahead_threshold = 4;
            }

            if (req_pts > pts) {
                // We are behind, so skip some
                *ignore = lrint(timebase * (req_pts - pts) * codec_context->sample_rate);
            } else if (self->audio_index != INT_MAX && int_position > req_position + ahead_threshold
                       && !self->is_audio_synchronizing) {
                // We are ahead, so seek backwards some more.
                // Supply -1 as the position to defeat the checks needed by for the other
                // call to seek_audio() at the beginning of producer_get_audio(). Otherwise,
                // more often than not, req_position will equal audio_expected.
                seek_audio(self, -1, timecode - 1.0);
                self->is_audio_synchronizing = 1;
            }
        }

        // Cancel the find_first_pts() in seek_audio()
        if (self->video_index == -1 && self->last_position == POSITION_INITIAL)
            self->last_position = int_position;
    }

    self->audio_used[index] = audio_used;

    return ret;
}

/** Get the audio from a frame.
*/
static int producer_get_audio(mlt_frame frame,
                              void **buffer,
                              mlt_audio_format *format,
                              int *frequency,
                              int *channels,
                              int *samples)
{
    // Get the producer
    producer_avformat self = mlt_frame_pop_audio(frame);

    pthread_mutex_lock(&self->audio_mutex);

    // Obtain the frame number of this frame
    mlt_position position = mlt_frame_original_position(frame);

    int paused = position + 1 == self->audio_expected
                 && mlt_properties_get_int(MLT_PRODUCER_PROPERTIES(self->parent), "mute_on_pause");

    if (!paused) {
        // Check the audio cache if not paused
        if (!self->audio_cache) {
            init_cache(MLT_PRODUCER_PROPERTIES(self->parent), &self->audio_cache);
        } else {
            mlt_frame original = mlt_cache_get_frame(self->audio_cache, position);
            if (original) {
                mlt_frame_get_audio(original, buffer, format, frequency, channels, samples);
                mlt_properties_set_data(MLT_FRAME_PROPERTIES(frame),
                                        "avformat.audio_cache",
                                        original,
                                        0,
                                        (mlt_destructor) mlt_frame_close,
                                        NULL);
                mlt_properties_pass_property(MLT_FRAME_PROPERTIES(frame),
                                             MLT_FRAME_PROPERTIES(original),
                                             "channel_layout");
                goto done_get_audio;
            }
        }
    }

    // Calculate the real time code
    double real_timecode = producer_time_of_frame(self->parent, position);

    // Get the producer fps
    double fps = mlt_producer_get_fps(self->parent);
    if (mlt_properties_get(MLT_FRAME_PROPERTIES(frame), "producer_consumer_fps"))
        fps = mlt_properties_get_double(MLT_FRAME_PROPERTIES(frame), "producer_consumer_fps");

    // Number of frames to ignore (for ffwd)
    int ignore[MAX_AUDIO_STREAMS] = {0};

    // Flag for paused (silence)
    double timecode = self->audio_expected > 0 ? real_timecode : FFMAX(real_timecode - 0.25, 0.0);
    paused = seek_audio(self, position, timecode);

    // Initialize ignore for all streams from the seek return value
    int i = MAX_AUDIO_STREAMS;
    while (i--)
        ignore[i] = ignore[0];

    // Fetch the audio_format
    AVFormatContext *context = self->audio_format;
    if (!context)
        goto exit_get_audio;

    int sizeof_sample = sizeof(int16_t);

    // Determine the tracks to use
    int index = self->audio_index;
    int index_max = self->audio_index + 1;
    if (self->audio_index == INT_MAX) {
        index = 0;
        index_max = FFMIN(MAX_AUDIO_STREAMS, context->nb_streams);
        *channels = self->total_channels;
        *samples = mlt_audio_calculate_frame_samples(fps, self->max_frequency, position);
        *frequency = self->max_frequency;
    }

    // Initialize the buffers
    for (; index < index_max && index < MAX_AUDIO_STREAMS; index++) {
        // Get codec context
        AVCodecContext *codec_context = self->audio_codec[index];

        if (codec_context && !self->audio_buffer[index]) {
#if !HAVE_FFMPEG_CH_LAYOUT
            if (self->audio_index != INT_MAX
                && !mlt_properties_get(MLT_PRODUCER_PROPERTIES(self->parent),
                                       "request_channel_layout"))
                codec_context->request_channel_layout = av_get_default_channel_layout(*channels);
#endif
            sizeof_sample = sample_bytes(codec_context);

            // Check for audio buffer and create if necessary
            self->audio_buffer_size[index] = MAX_AUDIO_FRAME_SIZE * sizeof_sample;
            self->audio_buffer[index] = mlt_pool_alloc(self->audio_buffer_size[index]);

            // Check for decoder buffer and create if necessary
            self->decode_buffer[index] = av_malloc(self->audio_buffer_size[index]);
        }
    }

    // Get the audio if required
    if (!paused && *frequency > 0) {
        int ret = 0;
        int got_audio = 0;
        AVPacket pkt;
        mlt_channel_layout mlt_layout = mlt_channel_auto;

        av_init_packet(&pkt);

        // Caller requested number samples based on requested sample rate.
        if (self->audio_index != INT_MAX)
            *samples = mlt_audio_calculate_frame_samples(fps,
                                                         self->audio_codec[self->audio_index]
                                                             ->sample_rate,
                                                         position);

        while (ret >= 0 && !got_audio) {
            // Check if the buffer already contains the samples required
            if (self->audio_index != INT_MAX && self->audio_used[self->audio_index] >= *samples
                && ignore[self->audio_index] == 0) {
                got_audio = 1;
                break;
            } else if (self->audio_index == INT_MAX) {
                // Check if there is enough audio for all streams
                got_audio = 1;
                for (index = 0; got_audio && index < index_max; index++)
                    if ((self->audio_codec[index] && self->audio_used[index] < *samples)
                        || ignore[index])
                        got_audio = 0;
                if (got_audio)
                    break;
            }

            // Read a packet
            pthread_mutex_lock(&self->packets_mutex);
            if (mlt_deque_count(self->apackets)) {
                AVPacket *tmp = (AVPacket *) mlt_deque_pop_front(self->apackets);
                av_packet_ref(&pkt, tmp);
                av_packet_free(&tmp);
            } else {
                ret = av_read_frame(context, &pkt);
                if (ret >= 0 && !self->seekable && pkt.stream_index == self->video_index) {
                    mlt_deque_push_back(self->vpackets, av_packet_clone(&pkt));
                } else if (ret == AVERROR(EAGAIN)) {
                    ret = 0;
                    pthread_mutex_unlock(&self->packets_mutex);
                    continue;
                } else if (ret < 0) {
                    mlt_producer producer = self->parent;
                    mlt_properties properties = MLT_PRODUCER_PROPERTIES(producer);
                    if (ret != AVERROR_EOF)
                        mlt_log_verbose(MLT_PRODUCER_SERVICE(producer),
                                        "av_read_frame returned error %d inside get_audio\n",
                                        ret);
                    if (!self->seekable && mlt_properties_get_int(properties, "reconnect")) {
                        // Try to reconnect to live sources by closing context and codecs,
                        // and letting next call to get_frame() reopen.
                        prepare_reopen(self);
                        pthread_mutex_unlock(&self->packets_mutex);
                        goto exit_get_audio;
                    }
                    if (!self->seekable
                        && mlt_properties_get_int(properties, "exit_on_disconnect")) {
                        mlt_log_fatal(MLT_PRODUCER_SERVICE(producer),
                                      "Exiting with error due to disconnected source.\n");
                        exit(EXIT_FAILURE);
                    }
                }
            }
            pthread_mutex_unlock(&self->packets_mutex);

            // We only deal with audio from the selected audio index
            index = pkt.stream_index;
            if (index < MAX_AUDIO_STREAMS && ret >= 0 && pkt.data && pkt.size > 0
                && (index == self->audio_index
                    || (self->audio_index == INT_MAX
                        && context->streams[index]->codecpar->codec_type == AVMEDIA_TYPE_AUDIO))) {
                ret = decode_audio(self, &ignore[index], &pkt, *samples, real_timecode, fps);
            }

            if (self->seekable || index != self->video_index)
                av_packet_unref(&pkt);
        }
        self->is_audio_synchronizing = 0;

        // Set some additional return values
        *format = mlt_audio_s16;
        if (self->audio_index != INT_MAX) {
            AVCodecContext *codec_ctx = self->audio_codec[self->audio_index];
            index = self->audio_index;
            *frequency = codec_ctx->sample_rate;
            *format = pick_audio_format(codec_ctx->sample_fmt);
            sizeof_sample = sample_bytes(codec_ctx);
#if HAVE_FFMPEG_CH_LAYOUT
            *channels = codec_ctx->ch_layout.nb_channels;
            if (av_channel_layout_check(&codec_ctx->ch_layout)) {
                mlt_layout = av_channel_layout_to_mlt(&codec_ctx->ch_layout);
            } else {
                AVChannelLayout ch_layout;
                av_channel_layout_default(&ch_layout, codec_ctx->ch_layout.nb_channels);
                mlt_layout = av_channel_layout_to_mlt(&ch_layout);
                av_channel_layout_uninit(&ch_layout);
            }
#else
            *channels = codec_ctx->channels;
            if (codec_ctx->channel_layout == 0)
                mlt_layout = av_channel_layout_to_mlt(
                    av_get_default_channel_layout(codec_ctx->channels));
            else
                mlt_layout = av_channel_layout_to_mlt(codec_ctx->channel_layout);
#endif
        } else if (self->audio_index == INT_MAX) {
            mlt_layout = mlt_channel_independent;
            for (index = 0; index < index_max; index++)
                if (self->audio_codec[index]) {
                    // XXX: This only works if all audio tracks have the same sample format.
                    *format = pick_audio_format(self->audio_codec[index]->sample_fmt);
                    sizeof_sample = sample_bytes(self->audio_codec[index]);
                    break;
                }
        }
        mlt_properties_set(MLT_FRAME_PROPERTIES(frame),
                           "channel_layout",
                           mlt_audio_channel_layout_name(mlt_layout));

        // Allocate and set the frame's audio buffer
        int size = mlt_audio_format_size(*format, *samples, *channels);
        *buffer = mlt_pool_alloc(size);
        mlt_frame_set_audio(frame, *buffer, *format, size, mlt_pool_release);

        // Interleave tracks if audio_index=all
        if (self->audio_index == INT_MAX) {
            uint8_t *dest = *buffer;
            int i;
            for (i = 0; i < *samples; i++) {
                for (index = 0; index < index_max; index++)
                    if (self->audio_codec[index]) {
#if HAVE_FFMPEG_CH_LAYOUT
                        int current_channels = self->audio_codec[index]->ch_layout.nb_channels;
#else
                        int current_channels = self->audio_codec[index]->channels;
#endif
                        uint8_t *src = self->audio_buffer[index]
                                       + i * current_channels * sizeof_sample;
                        memcpy(dest, src, current_channels * sizeof_sample);
                        dest += current_channels * sizeof_sample;
                    }
            }
            for (index = 0; index < index_max; index++)
                if (self->audio_codec[index] && self->audio_used[index] >= *samples) {
#if HAVE_FFMPEG_CH_LAYOUT
                    int current_channels = self->audio_codec[index]->ch_layout.nb_channels;
#else
                    int current_channels = self->audio_codec[index]->channels;
#endif
                    uint8_t *src = self->audio_buffer[index]
                                   + *samples * current_channels * sizeof_sample;
                    self->audio_used[index] -= *samples;
                    memmove(self->audio_buffer[index],
                            src,
                            self->audio_used[index] * current_channels * sizeof_sample);
                }
        }
        // Copy a single track to the output buffer
        else {
            index = self->audio_index;
            uint8_t silence = *format == mlt_audio_u8 ? 0x80 : 0;

            // Now handle the audio if we have enough
            if (self->audio_used[index] > 0) {
                uint8_t *src = self->audio_buffer[index];
                // copy samples from audio_buffer
                size = self->audio_used[index] < *samples ? self->audio_used[index] : *samples;
                memcpy(*buffer, src, size * *channels * sizeof_sample);
                // supply the remaining requested samples as silence
                if (*samples > self->audio_used[index])
                    memset(*buffer + size * *channels * sizeof_sample,
                           silence,
                           (*samples - self->audio_used[index]) * *channels * sizeof_sample);
                // reposition the samples within audio_buffer
                self->audio_used[index] -= size;
                memmove(src,
                        src + size * *channels * sizeof_sample,
                        self->audio_used[index] * *channels * sizeof_sample);
            } else {
                // Otherwise fill with silence
                memset(*buffer, silence, *samples * *channels * sizeof_sample);
            }
        }

        mlt_properties_set_int(MLT_FRAME_PROPERTIES(frame), "audio_samples", *samples);

        if (self->audio_cache) {
            mlt_cache_put_frame_audio(self->audio_cache, frame);
        }

    } else {
    exit_get_audio:
        if (*format == mlt_audio_none) {
            *format = mlt_audio_s16;
        }
        // Get silence and don't touch the context
        mlt_frame_get_audio(frame, buffer, format, frequency, channels, samples);
    }

done_get_audio:

    // Regardless of speed (other than paused), we expect to get the next frame
    if (!paused)
        self->audio_expected = position + 1;

    pthread_mutex_unlock(&self->audio_mutex);

    return 0;
}

/** Initialize the audio codec context.
*/

static int audio_codec_init(producer_avformat self, int index, mlt_properties properties)
{
    // Initialise the codec if necessary
    if (!self->audio_codec[index]) {
        // Get the video stream
        AVStream *stream = self->audio_format->streams[index];

        // Get codec context
        AVCodecParameters *codec_params = stream->codecpar;

        // Find the codec
        const AVCodec *codec = avcodec_find_decoder(codec_params->codec_id);
        if (mlt_properties_get(properties, "acodec")) {
            if (!(codec = avcodec_find_decoder_by_name(mlt_properties_get(properties, "acodec"))))
                codec = avcodec_find_decoder(codec_params->codec_id);
        }

        // Setup the codec context
        AVCodecContext *codec_context = avcodec_alloc_context3(codec);
        if (!codec_context) {
            mlt_log_error(MLT_PRODUCER_SERVICE(self->parent),
                          "Failed to allocate the decoder context for audio stream #%d\n",
                          index);
            self->audio_index = -1;
            return 0;
        }
        int ret = avcodec_parameters_to_context(codec_context, codec_params);
        if (ret < 0) {
            mlt_log_error(
                MLT_PRODUCER_SERVICE(self->parent),
                "Failed to copy decoder parameters to input decoder context for audio stream #%d\n",
                index);
            self->audio_index = -1;
            return 0;
        }

        // If we don't have a codec and we can't initialise it, we can't do much more...
        pthread_mutex_lock(&self->open_mutex);
        if (codec && avcodec_open2(codec_context, codec, NULL) >= 0) {
            // Now store the codec with its destructor
            if (self->audio_codec[index])
                avcodec_close(self->audio_codec[index]);
            self->audio_codec[index] = codec_context;
            self->audio_index = index;
        } else {
            // Remember that we can't use self later
            self->audio_index = -1;
        }
        pthread_mutex_unlock(&self->open_mutex);

        // Process properties as AVOptions
        apply_properties(codec_context,
                         properties,
                         AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_DECODING_PARAM);
        if (codec && codec->priv_class && codec_context->priv_data)
            apply_properties(codec_context->priv_data,
                             properties,
                             AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_DECODING_PARAM);
    }
    return self->audio_codec[index] && self->audio_index > -1;
}

static int pick_audio_stream(producer_avformat self)
{
    AVFormatContext *context = self->audio_format;
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(self->parent);
    int absolute_index = -1;

    // Handle all audio tracks
    if (self->audio_index > -1) {
        if (mlt_properties_get(properties, "audio_index")
            && !strcmp(mlt_properties_get(properties, "audio_index"), "all")) {
            absolute_index = INT_MAX;
            mlt_properties_set(properties, "astream", "all");
        }
        if (mlt_properties_get(properties, "astream")
            && !strcmp(mlt_properties_get(properties, "astream"), "all")) {
            absolute_index = INT_MAX;
            mlt_properties_set(properties, "audio_index", "all");
        }
    }

    // Are both audio_index != all and astream != all?
    if (absolute_index == -1) {
        if (context && mlt_properties_get(properties, "astream")) {
            // Get the relative stream index
            absolute_index = absolute_stream_index(context,
                                                   AVMEDIA_TYPE_AUDIO,
                                                   mlt_properties_get_int(properties, "astream"));
        } else {
            // Failover to the absolute index
            absolute_index = mlt_properties_get_int(properties, "audio_index");
            if (context) {
                // Compute the relative stream index
                mlt_properties_set_int(properties,
                                       "astream",
                                       relative_stream_index(context,
                                                             AVMEDIA_TYPE_AUDIO,
                                                             absolute_index));
            }
        }
        if (mlt_properties_get_int(properties, "audio_index") != absolute_index) {
            // Update the absolute index
            mlt_properties_set_int(properties, "audio_index", absolute_index);
            self->audio_index = absolute_index;
        }
    }

    // Exception handling for audio_index
    if (context && absolute_index >= (int) context->nb_streams && absolute_index < INT_MAX) {
        for (absolute_index = context->nb_streams - 1;
             absolute_index >= 0
             && context->streams[absolute_index]->codecpar->codec_type != AVMEDIA_TYPE_AUDIO;
             absolute_index--)
            ;
        mlt_properties_set_int(properties, "audio_index", absolute_index);
        mlt_properties_set_int(properties,
                               "astream",
                               relative_stream_index(context, AVMEDIA_TYPE_AUDIO, absolute_index));
    }
    if (context && absolute_index > -1 && absolute_index < INT_MAX
        && context->streams[absolute_index]->codecpar->codec_type != AVMEDIA_TYPE_AUDIO) {
        absolute_index = self->audio_index;
        mlt_properties_set_int(properties, "audio_index", absolute_index);
        mlt_properties_set_int(properties,
                               "astream",
                               relative_stream_index(context, AVMEDIA_TYPE_AUDIO, absolute_index));
    }
    if (context && absolute_index > -1 && absolute_index < INT_MAX
        && pick_audio_format(context->streams[absolute_index]->codecpar->format) == mlt_audio_none) {
        absolute_index = -1;
    }

    return absolute_index;
}

/** Set up audio handling.
*/

static void producer_set_up_audio(producer_avformat self, mlt_frame frame)
{
    // Get the producer
    mlt_producer producer = self->parent;

    // Get the properties
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(producer);

    // Fetch the audio format context
    AVFormatContext *context = self->audio_format;

    mlt_properties frame_properties = MLT_FRAME_PROPERTIES(frame);

    // Get the audio index
    int index = mlt_properties_get(properties, "astream")
                    ? mlt_properties_get_int(properties, "astream")
                    : mlt_properties_get_int(properties, "audio_index");

    // Reopen the file if necessary
    if (!context && self->audio_index > -1 && index > -1) {
        producer_open(self,
                      mlt_service_profile(MLT_PRODUCER_SERVICE(producer)),
                      mlt_properties_get(properties, "resource"),
                      1,
                      0);
        context = self->audio_format;
        mlt_properties_set_int(properties, "_probe_complete", 0);
    }
    index = pick_audio_stream(self);

    // Update the audio properties if the index changed
    if (context && self->audio_index > -1 && index != self->audio_index) {
        self->audio_index = index;
        pthread_mutex_lock(&self->open_mutex);
        unsigned i = 0;
        int index_max = FFMIN(MAX_AUDIO_STREAMS, context->nb_streams);
        for (i = 0; i < index_max; i++) {
            avcodec_free_context(&self->audio_codec[i]);
        }
        set_up_discard(self, index, self->video_index);
        mlt_cache_close(self->audio_cache);
        self->audio_cache = NULL;
        pthread_mutex_unlock(&self->open_mutex);
    }

    // Get the codec(s)
    if (context && index == INT_MAX) {
        unsigned int index;
        mlt_properties_set_int(frame_properties, "audio_frequency", self->max_frequency);
        mlt_properties_set_int(frame_properties, "audio_channels", self->total_channels);
        for (index = 0; index < context->nb_streams && index < MAX_AUDIO_STREAMS; index++) {
            if (context->streams[index]->codecpar->codec_type == AVMEDIA_TYPE_AUDIO)
                audio_codec_init(self, index, properties);
        }
        self->audio_index = INT_MAX;
    } else if (context && index > -1 && index < MAX_AUDIO_STREAMS
               && audio_codec_init(self, index, properties)) {
        mlt_properties_set_int(frame_properties,
                               "audio_frequency",
                               self->audio_codec[index]->sample_rate);
        mlt_properties_set_int(frame_properties,
                               "audio_channels",
#if HAVE_FFMPEG_CH_LAYOUT
                               self->audio_codec[index]->ch_layout.nb_channels);
#else
                               self->audio_codec[index]->channels);
#endif
    }
    if (context && index > -1) {
        // Add our audio operation
        mlt_frame_push_audio(frame, self);
        mlt_frame_push_audio(frame, producer_get_audio);
    }
}

/** Our get frame implementation.
*/

static int producer_get_frame(mlt_producer producer, mlt_frame_ptr frame, int index)
{
    // Access the private data
    (void) index; // unused
    mlt_service service = MLT_PRODUCER_SERVICE(producer);
    mlt_cache_item cache_item = mlt_service_cache_get(service, "producer_avformat");
    producer_avformat self = mlt_cache_item_data(cache_item, NULL);

    // If cache miss
    if (!self) {
        self = calloc(1, sizeof(struct producer_avformat_s));
        self->parent = producer;
        mlt_service_cache_put(service,
                              "producer_avformat",
                              self,
                              0,
                              (mlt_destructor) producer_avformat_close);
        cache_item = mlt_service_cache_get(service, "producer_avformat");
    }

    // Create an empty frame
    *frame = mlt_frame_init(service);

    if (*frame == NULL) {
        mlt_cache_item_close(cache_item);
        return 1;
    }

    mlt_properties frame_properties = MLT_FRAME_PROPERTIES(*frame);
    mlt_properties_set_data(frame_properties,
                            "avformat_cache",
                            cache_item,
                            0,
                            (mlt_destructor) mlt_cache_item_close,
                            NULL);

    // Update timecode on the frame we're creating
    mlt_frame_set_position(*frame, mlt_producer_position(producer));

    // Set up the video
    producer_set_up_video(self, *frame);

    // Set up the audio
    producer_set_up_audio(self, *frame);

    mlt_properties_set_int(frame_properties,
                           "format",
                           mlt_properties_get_int(MLT_PRODUCER_PROPERTIES(producer), "format"));

    // Set the position of this producer
    mlt_position position = mlt_producer_frame(producer);
    mlt_properties_set_position(frame_properties, "original_position", position);

    if (!mlt_properties_get_int(MLT_PRODUCER_PROPERTIES(producer), "_probe_complete")
        && self->video_index < 0) {
        // If video index is valid, get_image() must be called before the probe is complete
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.width");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.height");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.aspect_ratio");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.color_range");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.aspect_ratio");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.progressive");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.sample_aspect_num");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.sample_aspect_den");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.top_field_first");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.variable_frame_rate");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.colorspace");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.color_trc");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "meta.media.has_b_frames");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "aspect_ratio");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "width");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "height");
        mlt_properties_clear(MLT_PRODUCER_PROPERTIES(producer), "format");
        mlt_properties_set_int(MLT_PRODUCER_PROPERTIES(producer), "_probe_complete", 1);
    }

    // Calculate the next timecode
    mlt_producer_prepare_next(producer);

    return 0;
}

static int producer_probe(mlt_producer producer)
{
    int error = 0;
    mlt_properties properties = MLT_PRODUCER_PROPERTIES(producer);

    if (mlt_properties_get_int(properties, "_probe_complete")) {
        return error;
    }

    if (!mlt_properties_exists(properties, "_probe_complete")) {
        // If metadata was loaded from XML, no need to probe
        int video_in_use = mlt_properties_get_int(properties, "vstream") > -1;
        if (video_in_use && mlt_properties_exists(properties, "meta.media.progressive")) {
            return error;
        } else if (!video_in_use && mlt_properties_exists(properties, "meta.media.nb_streams")) {
            return error;
        }
    }

    mlt_frame fr = NULL;
    mlt_position save_position = mlt_producer_position(producer);

    // Call producer_get_frame() directly so that the underlying service will not attach any normalizers
    mlt_service_lock(MLT_PRODUCER_SERVICE(producer));
    error = producer_get_frame(producer, &fr, 0);
    mlt_service_unlock(MLT_PRODUCER_SERVICE(producer));
    if (!error && fr && mlt_properties_get_int(properties, "vstream") > -1) {
        // Some video metadata is not exposed until after the first get_image call.
        uint8_t *buffer = NULL;
        mlt_image_format fmt = mlt_image_none;
        int w = 0;
        int h = 0;
        error = mlt_frame_get_image(fr, &buffer, &fmt, &w, &h, 0);
    }
    mlt_frame_close(fr);
    mlt_producer_seek(producer, save_position);

    return error;
}

static void producer_avformat_close(producer_avformat self)
{
    mlt_log_debug(NULL, "producer_avformat_close\n");

    pthread_mutex_lock(&self->close_mutex);
    if (self->parent && self->parent->close)
        mlt_events_disconnect(MLT_PRODUCER_PROPERTIES(self->parent), self);
    pthread_mutex_unlock(&self->close_mutex);

    // Cleanup av contexts
    av_packet_unref(&self->pkt);
    av_frame_free(&self->video_frame);
    av_frame_free(&self->audio_frame);

#if USE_HWACCEL
    av_buffer_unref(&self->hwaccel.device_ctx);
#endif

    if (self->is_mutex_init)
        pthread_mutex_lock(&self->open_mutex);
    int i;
    for (i = 0; i < MAX_AUDIO_STREAMS; i++) {
        mlt_pool_release(self->audio_buffer[i]);
        av_free(self->decode_buffer[i]);
        avcodec_free_context(&self->audio_codec[i]);
    }
    avcodec_free_context(&self->video_codec);
    // Close the file
    if (self->is_thread_init) {
        pthread_mutex_lock(&self->packets_mutex);
        self->packets_thread_stop = 1;
        pthread_cond_signal(&self->packets_cond);
        pthread_mutex_unlock(&self->packets_mutex);
        pthread_join(self->packets_thread, NULL);
        pthread_cond_destroy(&self->packets_cond);
    }
    if (self->dummy_context)
        avformat_close_input(&self->dummy_context);
    if (self->seekable && self->audio_format)
        avformat_close_input(&self->audio_format);
    if (self->video_format)
        avformat_close_input(&self->video_format);
    if (self->is_mutex_init)
        pthread_mutex_unlock(&self->open_mutex);
#ifdef AVFILTER
    avfilter_graph_free(&self->vfilter_graph);
#endif

    // Cleanup caches.
    mlt_cache_close(self->image_cache);
    mlt_cache_close(self->audio_cache);
    if (self->last_good_frame)
        mlt_frame_close(self->last_good_frame);

    // Cleanup the mutexes
    if (self->is_mutex_init) {
        pthread_mutex_destroy(&self->audio_mutex);
        pthread_mutex_destroy(&self->video_mutex);
        pthread_mutex_destroy(&self->packets_mutex);
        pthread_mutex_destroy(&self->open_mutex);
        pthread_mutex_destroy(&self->close_mutex);
    }

    // Cleanup the packet queues
    AVPacket *pkt;
    if (self->apackets) {
        while ((pkt = mlt_deque_pop_back(self->apackets))) {
            av_packet_free(&pkt);
        }
        mlt_deque_close(self->apackets);
        self->apackets = NULL;
    }
    if (self->vpackets) {
        while ((pkt = mlt_deque_pop_back(self->vpackets))) {
            av_packet_free(&pkt);
        }
        mlt_deque_close(self->vpackets);
        self->vpackets = NULL;
    }

    free(self);
}

static void producer_close(mlt_producer parent)
{
    // Remove this instance from the cache
    mlt_service_cache_purge(MLT_PRODUCER_SERVICE(parent));

    // Detach the producer_avformat struct
    mlt_cache_item cache_item = mlt_service_cache_get(MLT_PRODUCER_SERVICE(parent),
                                                      "producer_avformat");
    producer_avformat self = mlt_cache_item_data(cache_item, NULL);
    if (self) {
        pthread_mutex_lock(&self->close_mutex);
        self->parent = NULL;
        parent->close = NULL;
        pthread_mutex_unlock(&self->close_mutex);
    } else {
        parent->close = NULL;
    }
    mlt_cache_item_close(cache_item);

    // Close the parent
    mlt_producer_close(parent);

    // Free the memory
    free(parent);
}
