ffmpeg-hls實現分析


[時間:2018-05] [狀態:Open]
[關鍵詞:流媒體,stream,HLS, ffmpeg, demuxer,avformat]

0 引言

本文將關注於FFmpeg中的HLS相關實現,相關代碼在libavformat/hls.c中(我所使用的Ffmpeg版本是4.0的),分析hls_demuxer的主要實現邏輯。

本文作為我之前的HLS綜述的后續文章,也是ffmpeg框架分析的后續文章。前者介紹了HLS協議相關的理論部分,后者介紹了FFmpeg主要框架分析(本文主要關注demuxer);要是你對此感興趣建議了解下。

1 ff_hls_demuxer的主要對外接口

#define OFFSET(x) offsetof(HLSContext, x)
static const AVOption hls_options[] = {
    {"live_start_index", "segment index to start live streams at (negative values are from the end)",
        OFFSET(live_start_index), AV_OPT_TYPE_INT, {.i64 = -3}, INT_MIN, INT_MAX, FLAGS},
	// ... HLS demuxer支持的參數,有刪減
    {NULL}
};

static const AVClass hls_class = {
    .class_name = "hls,applehttp",
    .item_name  = av_default_item_name,
    .option     = hls_options,
    .version    = LIBAVUTIL_VERSION_INT,
};

AVInputFormat ff_hls_demuxer = {
    .name           = "hls,applehttp",
    .long_name      = NULL_IF_CONFIG_SMALL("Apple HTTP Live Streaming"),
    .priv_class     = &hls_class,
    .priv_data_size = sizeof(HLSContext),
    .flags          = AVFMT_NOGENSEARCH,
    .read_probe     = hls_probe, // 媒體格式探測
    .read_header    = hls_read_header, // 讀取協議頭並獲取節目信息
    .read_packet    = hls_read_packet, // 讀取音視頻包
    .read_close     = hls_close, // 關閉HLS通信
    .read_seek      = hls_read_seek, // 實現HLS的seek操作
};

通常分析FFmpeg中的demuxer,我們主要關注其read_probe、read_header、read_packet、read_close、read_seek這五個函數指針所對應的實現代碼。FFmpeg在實現demuxer時將其協議解析的部分完整封裝了demuxer中。
hls_demuxer中主要的結構是HLSContext,定義如下:

typedef struct HLSContext {
    AVClass *class;
    AVFormatContext *ctx;
    int n_variants;
    struct variant **variants;/* master playlist中有多個variants */
    int n_playlists;
    struct playlist **playlists;/* playlists中包含一個segment的列表 */
    int n_renditions;
    struct rendition **renditions;

    int cur_seq_no;
    int live_start_index;
    int first_packet;
    int64_t first_timestamp;
    int64_t cur_timestamp;
    AVIOInterruptCB *interrupt_callback;
    AVDictionary *avio_opts;
    int strict_std_compliance;
    char *allowed_extensions;
    int max_reload;
    int http_persistent;
    int http_multiple;
    AVIOContext *playlist_pb;
} HLSContext;

接下來我們將逐個函數查看。

2 hls_demuxer核心接口實現

2.1 read_probe -- hls_probe

這個函數主要是HLS格式探測,其實現代碼如下:

static int hls_probe(AVProbeData *p)
{
    /* HLS協議要求必須以#EXTM3U打頭,並至少有下面三個字段之一存在 */
    if (strncmp(p->buf, "#EXTM3U", 7))
        return 0;

    if (strstr(p->buf, "#EXT-X-STREAM-INF:")     ||
        strstr(p->buf, "#EXT-X-TARGETDURATION:") ||
        strstr(p->buf, "#EXT-X-MEDIA-SEQUENCE:"))
        return AVPROBE_SCORE_MAX;
    return 0;
}

這段代碼邏輯比較簡單,都是關於字符串匹配的處理,也是按照HLS協議規定的要求進行媒體格式探測的。

2.2 read_header -- hls_read_header

hls_read_header將是比較復雜的處理邏輯,因為這里涉及到master playlist解析、節目信息獲取,同時初始化hls解析相關的結構。這個函數實現有300行左右。我們將分為三部分:playlist解析、hls相關初始化、提取節目信息。

2.2.1 playlist解析

hls_read_header第一部分代碼如下:

static int hls_read_header(AVFormatContext *s)
{
    HLSContext *c = s->priv_data;
    int ret = 0, i;

    c->ctx                = s;
    c->interrupt_callback = &s->interrupt_callback;
    c->strict_std_compliance = s->strict_std_compliance;
    c->first_packet = 1;
    c->first_timestamp = AV_NOPTS_VALUE;
    c->cur_timestamp = AV_NOPTS_VALUE;
	
	/* 解析m3u8 */
    if ((ret = parse_playlist(c, s->url, NULL, s->pb)) < 0)
        goto fail;

    if (c->n_variants == 0) {
        av_log(NULL, AV_LOG_WARNING, "Empty playlist\n");
        ret = AVERROR_EOF;
        goto fail;
    }
    /* 對於master playlist,逐個解析其中的playlist */
    if (c->n_playlists > 1 || c->playlists[0]->n_segments == 0) {
        for (i = 0; i < c->n_playlists; i++) {
            struct playlist *pls = c->playlists[i];
            if ((ret = parse_playlist(c, pls->url, pls, NULL)) < 0)
                goto fail;
        }
    }
	/* 必須有至少一個variant */
    if (c->variants[0]->playlists[0]->n_segments == 0) {
        av_log(NULL, AV_LOG_WARNING, "Empty playlist\n");
        ret = AVERROR_EOF;
        goto fail;
    }
	// ... 部分代碼,有刪減

從上面代碼來看最主要的邏輯就是調用parse_playlist函數來解析m3u8文件。解析部分主要參考HLS協議就可以。其代碼如下:

static int parse_playlist(HLSContext *c, const char *url,
                          struct playlist *pls, AVIOContext *in)
{
    int ret = 0, is_segment = 0, is_variant = 0;
    int64_t duration = 0;
    char line[MAX_URL_SIZE];
    const char *ptr;
    int64_t seg_offset = 0;
    int64_t seg_size = -1;
    uint8_t *new_url = NULL;
    struct variant_info variant_info;
    char tmp_str[MAX_URL_SIZE];
    struct segment *cur_init_section = NULL;

    if (!in) { /* 創建用於HTTP請求的AVIO */
        AVDictionary *opts = NULL;
        av_dict_copy(&opts, c->avio_opts, 0);

        ret = c->ctx->io_open(c->ctx, &in, url, AVIO_FLAG_READ, &opts);
        av_dict_free(&opts);
        if (ret < 0)
            return ret;
    }
	/* HTTP-URL重定向 */
    if (av_opt_get(in, "location", AV_OPT_SEARCH_CHILDREN, &new_url) >= 0)
        url = new_url;

    ff_get_chomp_line(in, line, sizeof(line));
    if (strcmp(line, "#EXTM3U")) {/* HLS協議標志起始頭 */
        ret = AVERROR_INVALIDDATA;
        goto fail;
    }
	/* 釋放已經存在的pls及segment */
    if (pls) {
        free_segment_list(pls);
        pls->finished = 0;
        pls->type = PLS_TYPE_UNSPECIFIED;
    }/* 以下是具體協議的解析 */
    while (!avio_feof(in)) {
        ff_get_chomp_line(in, line, sizeof(line));
        if (av_strstart(line, "#EXT-X-STREAM-INF:", &ptr)) {
            is_variant = 1;
            memset(&variant_info, 0, sizeof(variant_info));
            ff_parse_key_value(ptr, (ff_parse_key_val_cb) handle_variant_args,
                               &variant_info);
        } else if (av_strstart(line, "#EXT-X-MEDIA:", &ptr)) {
            struct rendition_info info = {{0}};
            ff_parse_key_value(ptr, (ff_parse_key_val_cb) handle_rendition_args,
                               &info);
            new_rendition(c, &info, url);
        } else if (av_strstart(line, "#EXT-X-TARGETDURATION:", &ptr)) {
            ret = ensure_playlist(c, &pls, url);
            if (ret < 0)
                goto fail; /* 最大分片時長 */
            pls->target_duration = strtoll(ptr, NULL, 10) * AV_TIME_BASE;
        } else if (av_strstart(line, "#EXT-X-MEDIA-SEQUENCE:", &ptr)) {
            ret = ensure_playlist(c, &pls, url);
            if (ret < 0)
                goto fail;
            pls->start_seq_no = atoi(ptr); /* 起始segment number */
        } else if (av_strstart(line, "#EXT-X-PLAYLIST-TYPE:", &ptr)) {
            ret = ensure_playlist(c, &pls, url);
            if (ret < 0)
                goto fail;
            if (!strcmp(ptr, "EVENT"))/* HLS類型:VOD/EVENT */
                pls->type = PLS_TYPE_EVENT;
            else if (!strcmp(ptr, "VOD"))
                pls->type = PLS_TYPE_VOD;
        } else if (av_strstart(line, "#EXT-X-ENDLIST", &ptr)) {
            if (pls) /* playlist結束標志,表示VOD */
                pls->finished = 1;
        } else if (av_strstart(line, "#EXTINF:", &ptr)) {
            is_segment = 1;
            duration   = atof(ptr) * AV_TIME_BASE;
        } else if (av_strstart(line, "#EXT-X-BYTERANGE:", &ptr)) {
            seg_size = strtoll(ptr, NULL, 10);
            ptr = strchr(ptr, '@');
            if (ptr)/* 使用字節划分的m3u8 */
                seg_offset = strtoll(ptr+1, NULL, 10);
        } else if (av_strstart(line, "#", NULL)) {
            continue; /* 忽略無法識別的字段 */
        } else if (line[0]) {
            if (is_variant) { /* 針對variant的處理,下一行一般是URL */
                if (!new_variant(c, &variant_info, line, url)) {
                    ret = AVERROR(ENOMEM);
                    goto fail;
                }
                is_variant = 0;
            }
            if (is_segment) { /* 針對segment的處理,需要拼接segment的URL */
                struct segment *seg;
                if (!pls) {
                    if (!new_variant(c, 0, url, NULL)) {
                        ret = AVERROR(ENOMEM);
                        goto fail;
                    }
                    pls = c->playlists[c->n_playlists - 1];
                }
                seg = av_malloc(sizeof(struct segment));
                if (!seg) {
                    ret = AVERROR(ENOMEM);
                    goto fail;
                }
                seg->duration = duration;
                seg->key_type = KEY_NONE;
                seg->key = NULL;

                ff_make_absolute_url(tmp_str, sizeof(tmp_str), url, line);
                seg->url = av_strdup(tmp_str);
                if (!seg->url) {
                    av_free(seg->key);
                    av_free(seg);
                    ret = AVERROR(ENOMEM);
                    goto fail;
                }

                dynarray_add(&pls->segments, &pls->n_segments, seg);
                is_segment = 0;

                seg->size = seg_size;
                if (seg_size >= 0) {
                    seg->url_offset = seg_offset;
                    seg_offset += seg_size;
                    seg_size = -1;
                } else {
                    seg->url_offset = 0;
                    seg_offset = 0;
                }

                seg->init_section = cur_init_section;
            }
        }
    }
    if (pls)
        pls->last_load_time = av_gettime_relative();

fail:
    av_free(new_url);
    ff_format_io_close(c->ctx, &in);
    c->ctx->ctx_flags = c->ctx->ctx_flags & ~(unsigned)AVFMTCTX_UNSEEKABLE;
    if (!c->n_variants || !c->variants[0]->n_playlists ||
        !(c->variants[0]->playlists[0]->finished ||
          c->variants[0]->playlists[0]->type == PLS_TYPE_EVENT))
        c->ctx->ctx_flags |= AVFMTCTX_UNSEEKABLE;
    return ret;
}

這樣通過對m3u8的parse之后,我們已經創建了好了playlist、segment、variant以及rendition。接下來是第二部分。

2.2.2 hls相關初始化

主要相關代碼如下:

	/* 對於非直播流,根據segment的時長計算總的節目時長 */
    if (c->variants[0]->playlists[0]->finished) {
        int64_t duration = 0;
        for (i = 0; i < c->variants[0]->playlists[0]->n_segments; i++)
            duration += c->variants[0]->playlists[0]->segments[i]->duration;
        s->duration = duration;
    }

    /* 將renditions與variants關聯起來 */
    for (i = 0; i < c->n_variants; i++) {
        struct variant *var = c->variants[i];

        if (var->audio_group[0])
            add_renditions_to_variant(c, var, AVMEDIA_TYPE_AUDIO, var->audio_group);
        if (var->video_group[0])
            add_renditions_to_variant(c, var, AVMEDIA_TYPE_VIDEO, var->video_group);
        if (var->subtitles_group[0])
            add_renditions_to_variant(c, var, AVMEDIA_TYPE_SUBTITLE, var->subtitles_group);
    }

    /* 為每個variant創建一個節目 */
    for (i = 0; i < c->n_variants; i++) {
        struct variant *v = c->variants[i];
        AVProgram *program = av_new_program(s, i);
        if (!program)
            goto fail;
        av_dict_set_int(&program->metadata, "variant_bitrate", v->bandwidth, 0);
    }

    /* 選擇起始segment的索引 */
    for (i = 0; i < c->n_playlists; i++) {
        struct playlist *pls = c->playlists[i];

        if (pls->n_segments == 0)
            continue;

        pls->cur_seq_no = select_cur_seq_no(c, pls);
        highest_cur_seq_no = FFMAX(highest_cur_seq_no, pls->cur_seq_no);
    }

我們看一下如何選擇起始segment的索引,代碼如下:

static int select_cur_seq_no(HLSContext *c, struct playlist *pls)
{
    int seq_no;
	/* 直播情況下,定期更新m3u8 */
    if (!pls->finished && !c->first_packet &&
        av_gettime_relative() - pls->last_load_time >= default_reload_interval(pls))
        parse_playlist(c, pls->url, pls, NULL);

    /* 對於非直播的情況,直接通過時長查找對應的segment索引號(seek時比較常用的邏輯) */
    if (pls->finished && c->cur_timestamp != AV_NOPTS_VALUE) {
        find_timestamp_in_playlist(c, pls, c->cur_timestamp, &seq_no);
        return seq_no;
    }

    if (!pls->finished) {
        if (!c->first_packet && /* 是在播放中選擇segment */
            c->cur_seq_no >= pls->start_seq_no &&
            c->cur_seq_no < pls->start_seq_no + pls->n_segments)
            return c->cur_seq_no;

        /* 直播情況下,需要參考live_start_index調整下 */
        if (c->live_start_index < 0)
            return pls->start_seq_no + FFMAX(pls->n_segments + c->live_start_index, 0);
        else
            return pls->start_seq_no + FFMIN(c->live_start_index, pls->n_segments - 1);
    }

    /* 其他情況直接返回起始segment索引號*/
    return pls->start_seq_no;
}

經過上面的處理,我們已經創建好FFmpeg框架上需要的AVProgram等信息了,接下來就需要填充AVProgram,並獲得所有的AVStream信息。

2.2.3 提取節目信息

在開始介紹前,我們先看一下playlist結構體的定義:

/* 每個playlist都有自己的demuxer,如果該playlist是在用的,它還會有AVIOContext和AVPacket */
struct playlist {
    char url[MAX_URL_SIZE];
    AVIOContext pb;
    uint8_t* read_buffer;
    AVIOContext *input;
    int input_read_done;
    AVIOContext *input_next;
    int input_next_requested;
    AVFormatContext *parent;// 這個將指向公用的AVFormatContext
    int index;
    AVFormatContext *ctx; // 這個將用於解析當前playlist的所有segment
    AVPacket pkt;
    int has_noheader_flag;

    /* 當前playlist中包含的AVStream信息 */
    AVStream **main_streams;
    int n_main_streams;

    int finished; /* segment讀取狀態的相關參數 */
    enum PlaylistType type;
    int64_t target_duration;
    int start_seq_no;
    int n_segments; /* 當前playlist中的所有segment數組 */
    struct segment **segments;
    int needed;
    int cur_seq_no;
    int64_t cur_seg_offset;
    int64_t last_load_time;

    /* Media Initialization Section */
    struct segment *cur_init_section;
    uint8_t *init_sec_buf;
    unsigned int init_sec_buf_size;
    unsigned int init_sec_data_len;
    unsigned int init_sec_buf_read_offset;

    char key_url[MAX_URL_SIZE]; /* HLS解密密鑰對應的URL */
    uint8_t key[16];

    int64_t seek_timestamp; /* seek相關的參數 */
    int seek_flags;
    int seek_stream_index; /* into subdemuxer stream array */

    /* 和當前playlist相關的Renditions(可選) */
    int n_renditions;
    struct rendition **renditions;

    /* Media Initialization Sections (EXT-X-MAP)(可選) */
    int n_init_sections;
    struct segment **init_sections;
};

hls_read_header后續邏輯與playlist結構體相關比較大,節目信息提取的代碼如下:

    /* 對每個playlist打開其demuxer */
    for (i = 0; i < c->n_playlists; i++) {
        struct playlist *pls = c->playlists[i];
        AVInputFormat *in_fmt = NULL;

        if (!(pls->ctx = avformat_alloc_context())) {
            ret = AVERROR(ENOMEM);
            goto fail;
        }

        if (pls->n_segments == 0)
            continue;

        pls->index  = i;
        pls->needed = 1;
        pls->parent = s;

        /* 調整直播流的起播索引號,以保證所有playlist是同步的 */
        if (!pls->finished && pls->cur_seq_no == highest_cur_seq_no - 1 &&
            highest_cur_seq_no < pls->start_seq_no + pls->n_segments) {
            pls->cur_seq_no = highest_cur_seq_no;
        }

        pls->read_buffer = av_malloc(INITIAL_BUFFER_SIZE);
        if (!pls->read_buffer){
            ret = AVERROR(ENOMEM);
            avformat_free_context(pls->ctx);
            pls->ctx = NULL;
            goto fail;
        }/* 這里初始化了AVIOContext,留意read_data函數,這將是后續讀包的核心 */
        ffio_init_context(&pls->pb, pls->read_buffer, INITIAL_BUFFER_SIZE, 0, pls,
                          read_data, NULL, NULL);
        pls->pb.seekable = 0;
        ret = av_probe_input_buffer(&pls->pb, &in_fmt, pls->segments[0]->url,
                                    NULL, 0, 0);
        if (ret < 0) {
            av_log(s, AV_LOG_ERROR, "Error when loading first segment '%s'\n", pls->segments[0]->url);
            avformat_free_context(pls->ctx);
            pls->ctx = NULL;
            goto fail;
        }
        pls->ctx->pb       = &pls->pb;
        pls->ctx->io_open  = nested_io_open;
        pls->ctx->flags   |= s->flags & ~AVFMT_FLAG_CUSTOM_IO;

		/* 接下來就是打開流,獲取其中AVStream信息了 */
        ret = avformat_open_input(&pls->ctx, pls->segments[0]->url, in_fmt, NULL);
        if (ret < 0)
            goto fail;

        ret = avformat_find_stream_info(pls->ctx, NULL);
        if (ret < 0)
            goto fail;

        pls->has_noheader_flag = !!(pls->ctx->ctx_flags & AVFMTCTX_NOHEADER);

        /* Create new AVStreams for each stream in this playlist */
        ret = update_streams_from_subdemuxer(s, pls);
        if (ret < 0)
            goto fail;

        if (pls->n_main_streams)
            av_dict_copy(&pls->main_streams[0]->metadata, pls->ctx->metadata, 0);

        add_metadata_from_renditions(s, pls, AVMEDIA_TYPE_AUDIO);
        add_metadata_from_renditions(s, pls, AVMEDIA_TYPE_VIDEO);
        add_metadata_from_renditions(s, pls, AVMEDIA_TYPE_SUBTITLE);
    }

    update_noheader_flag(s);

    return 0;
fail:
    hls_close(s);
    return ret;
}

2.3 read_close -- hls_close

鑒於read_close實現相對簡單,我們先了解下該函數的實現代碼:

static void free_playlist_list(HLSContext *c)
{
    int i; /* 針對每個playlist釋放動態分配的內存 */
    for (i = 0; i < c->n_playlists; i++) {
        struct playlist *pls = c->playlists[i];
        free_segment_list(pls);
        free_init_section_list(pls);
        av_freep(&pls->main_streams);
        av_freep(&pls->renditions);
        av_freep(&pls->init_sec_buf);
        av_packet_unref(&pls->pkt);
        av_freep(&pls->pb.buffer);
        if (pls->input)
            ff_format_io_close(c->ctx, &pls->input);
        pls->input_read_done = 0;
        if (pls->input_next)
            ff_format_io_close(c->ctx, &pls->input_next);
        pls->input_next_requested = 0;
        if (pls->ctx) {
            pls->ctx->pb = NULL;
            avformat_close_input(&pls->ctx);
        }
        av_free(pls);
    }
    av_freep(&c->playlists);
    c->n_playlists = 0;
}

static void free_variant_list(HLSContext *c)
{
    int i;
    for (i = 0; i < c->n_variants; i++) {
        struct variant *var = c->variants[i];
        av_freep(&var->playlists);
        av_free(var);
    }
    av_freep(&c->variants);
    c->n_variants = 0;
}

static void free_rendition_list(HLSContext *c)
{
    int i;
    for (i = 0; i < c->n_renditions; i++)
        av_freep(&c->renditions[i]);
    av_freep(&c->renditions);
    c->n_renditions = 0;
}

static int hls_close(AVFormatContext *s)
{
    HLSContext *c = s->priv_data;
	/* 基本邏輯就是依次釋放:playlist, variant, rendition */
    free_playlist_list(c); /* 最復雜的釋放 */
    free_variant_list(c);
    free_rendition_list(c);

    av_dict_free(&c->avio_opts);
    ff_format_io_close(c->ctx, &c->playlist_pb);

    return 0;
}

2.4 read_packet -- hls_read_packet

該函數主要功能是讀取原始的數據,並解析為AVPacket。代碼大約有150行,主要邏輯如下:

static int hls_read_packet(AVFormatContext *s, AVPacket *pkt)
{
    HLSContext *c = s->priv_data;
    int ret, i, minplaylist = -1;
	/* 這個函數會根據AVStream.discard標志來判斷關閉不需要的HTTP下載 */
    recheck_discard_flags(s, c->first_packet);
    c->first_packet = 0;

    for (i = 0; i < c->n_playlists; i++) {
        struct playlist *pls = c->playlists[i];
        /* 從每一個已打開的playlist上讀取一個AVPacket */
        if (pls->needed && !pls->pkt.data) {
            while (1) {
                int64_t ts_diff;
                AVRational tb;
                ret = av_read_frame(pls->ctx, &pls->pkt);
                if (ret < 0) {
                    if (!avio_feof(&pls->pb) && ret != AVERROR_EOF)
                        return ret;
                    reset_packet(&pls->pkt);
                    break;
                } else {
                    if (c->first_timestamp == AV_NOPTS_VALUE &&
                        pls->pkt.dts       != AV_NOPTS_VALUE)
                        c->first_timestamp = av_rescale_q(pls->pkt.dts,
                            get_timebase(pls), AV_TIME_BASE_Q);
                }
				/* 以下部分跟seek有關,無seek操作無需后續判斷 */
                if (pls->seek_timestamp == AV_NOPTS_VALUE)
                    break;

                if (pls->seek_stream_index < 0 ||
                    pls->seek_stream_index == pls->pkt.stream_index) {

                    if (pls->pkt.dts == AV_NOPTS_VALUE) {
                        pls->seek_timestamp = AV_NOPTS_VALUE;
                        break;
                    }

                    tb = get_timebase(pls);
                    ts_diff = av_rescale_rnd(pls->pkt.dts, AV_TIME_BASE,
                                            tb.den, AV_ROUND_DOWN) -
                            pls->seek_timestamp;
                    if (ts_diff >= 0 && (pls->seek_flags  & AVSEEK_FLAG_ANY ||
                                        pls->pkt.flags & AV_PKT_FLAG_KEY)) {
                        pls->seek_timestamp = AV_NOPTS_VALUE;
                        break;
                    }
                }/* 到此,說面當前AVPacket是需要丟棄的,重新讀取 */
                av_packet_unref(&pls->pkt);
                reset_packet(&pls->pkt);
            }
        }
        /* 從所有AVStream中找到dts最小的一個,記錄該索引值 */
        if (pls->pkt.data) {
            struct playlist *minpls = minplaylist < 0 ?
                                     NULL : c->playlists[minplaylist];
            if (minplaylist < 0) {
                minplaylist = i;
            } else {
                int64_t dts     =    pls->pkt.dts;
                int64_t mindts  = minpls->pkt.dts;

                if (dts == AV_NOPTS_VALUE ||
                    (mindts != AV_NOPTS_VALUE && compare_ts_with_wrapdetect(dts, pls, mindts, minpls) < 0))
                    minplaylist = i;
            }
        }
    }

    /* 成功讀取AVPacket,需要返回給上層調用者 */
    if (minplaylist >= 0) {
        struct playlist *pls = c->playlists[minplaylist];
        AVStream *ist;
        AVStream *st;

		/* 判斷stream_index有效性 */
        if (pls->pkt.stream_index >= pls->n_main_streams) {
            av_packet_unref(&pls->pkt);
            reset_packet(&pls->pkt);
            return AVERROR_BUG;
        }

        ist = pls->ctx->streams[pls->pkt.stream_index];
        st = pls->main_streams[pls->pkt.stream_index];

        *pkt = pls->pkt;
        pkt->stream_index = st->index;
        reset_packet(&c->playlists[minplaylist]->pkt);
		/* 更新當前HLS的讀取位置 */
        if (pkt->dts != AV_NOPTS_VALUE)
            c->cur_timestamp = av_rescale_q(pkt->dts,
                                            ist->time_base,
                                            AV_TIME_BASE_Q);

        return 0;
    }
    return AVERROR_EOF;
}

從這個代碼上來看很奇怪,這里沒有包含任何http請求、playlist更新、segment的處理邏輯,那么真正的邏輯隱藏在哪里呢?不知道大家還記得2.2節中hls_read_header函數中對AVIOContext的初始化邏輯,其實最終的解析主要部分就在其中一個不起眼的函數中——read_data。上面的av_read_frame通過FFmpeg提供的一系列機制,最終調用read_data(你要是對此感興趣,建議研究下雷神的自定義AVIO的示例)。這個函數長度在150行左右,但其內部是一個帶有多個標簽的循環邏輯(實際上這部分代碼是最復雜的,使用了c語言中不推薦使用的goto語句),詳細代碼如下:

static int read_data(void *opaque, uint8_t *buf, int buf_size)
{
    struct playlist *v = opaque;
    HLSContext *c = v->parent->priv_data;
    int ret;
    int just_opened = 0;
    int reload_count = 0;
    struct segment *seg;

restart: /* 標簽:重啟新的下載 */
    if (!v->needed)
        return AVERROR_EOF;

    if (!v->input || (c->http_persistent && v->input_read_done)) {
        int64_t reload_interval;

        /* 檢查playlist是否需要下載,如果不需要了,直接返回EOF */
        v->needed = playlist_needed(v);
        if (!v->needed) {
            return AVERROR_EOF;
        }

        /* 對於直播流,定期更新playlist */
        reload_interval = default_reload_interval(v);

reload: /* 標簽:重新下載playlist */
        reload_count++;
        if (reload_count > c->max_reload)
            return AVERROR_EOF;
        if (!v->finished &&
            av_gettime_relative() - v->last_load_time >= reload_interval) {
            if ((ret = parse_playlist(c, v->url, v, NULL)) < 0)
                return ret;

            /* 按照HLS協議規定,如果還需要請求playlist,請求間隔可以設置為segment時長的一半 */
            reload_interval = v->target_duration / 2;
        }
        if (v->cur_seq_no < v->start_seq_no) {
            /* 這種情況下客戶端因為某些原因下載過慢,直接調整cur_seq_no,跳過一些segment */
            v->cur_seq_no = v->start_seq_no;
        }
        if (v->cur_seq_no >= v->start_seq_no + v->n_segments) {
            if (v->finished) /* 點播的情況下,表示讀取完成 */
                return AVERROR_EOF;
            while (av_gettime_relative() - v->last_load_time < reload_interval) {
                if (ff_check_interrupt(c->interrupt_callback))
                    return AVERROR_EXIT;
                av_usleep(100*1000);
            }
            /* 上面的循環用於等待給定的時間,然后重新加載playlist */
            goto reload;
        }
		/* 所有的前提都滿足,可以開始讀segment了 */
        v->input_read_done = 0;
        seg = current_segment(v);

        /* 如果存在,加載並更新Media Initialization Section */
        ret = update_init_section(v, seg);
        if (ret)
            return ret;
		
        if (c->http_multiple == 1 && v->input_next_requested) {
            FFSWAP(AVIOContext *, v->input, v->input_next);
            v->input_next_requested = 0;
            ret = 0;
        } else { /* 這里發起了新的HTTP請求 */
            ret = open_input(c, v, seg, &v->input);
        }
        if (ret < 0) { /* 請求失敗的情況下的處理,默認直接跳過當前segment */
            if (ff_check_interrupt(c->interrupt_callback))
                return AVERROR_EXIT;
            v->cur_seq_no += 1;
            goto reload;
        }
        just_opened = 1;
    }

    if (c->http_multiple == -1) {
        uint8_t *http_version_opt = NULL;
        int r = av_opt_get(v->input, "http_version", AV_OPT_SEARCH_CHILDREN, &http_version_opt);
        if (r >= 0) {
            c->http_multiple = strncmp((const char *)http_version_opt, "1.1", 3) == 0;
            av_freep(&http_version_opt);
        }
    }

    seg = next_segment(v); /* 這是使用單HTTP發起請求的處理 */
    if (c->http_multiple == 1 && !v->input_next_requested &&
        seg && seg->key_type == KEY_NONE && av_strstart(seg->url, "http", NULL)) {
        ret = open_input(c, v, seg, &v->input_next);
        if (ret < 0) {
            if (ff_check_interrupt(c->interrupt_callback))
                return AVERROR_EXIT;
            av_log(v->parent, AV_LOG_WARNING, "Failed to open segment %d of playlist %d\n",
                   v->cur_seq_no + 1,
                   v->index);
        } else {
            v->input_next_requested = 1;
        }
    }

    if (v->init_sec_buf_read_offset < v->init_sec_data_len) {
        /* 在任何實際數據返回之前,首先返回init section(解碼器可能會依賴這些數據做初始化) */
        int copy_size = FFMIN(v->init_sec_data_len - v->init_sec_buf_read_offset, buf_size);
        memcpy(buf, v->init_sec_buf, copy_size);
        v->init_sec_buf_read_offset += copy_size;
        return copy_size;
    }

    seg = current_segment(v);
	/* 這是實際通過HTTP讀取buf_size長度的數據,這個數據是交織后的數據,從server上獲取的 */
    ret = read_from_url(v, seg, buf, buf_size);
    if (ret > 0)
        return ret;
	
	/* 讀取完成了,可以關閉segment所使用的HTTP資源了 */
    if (c->http_persistent &&
        seg->key_type == KEY_NONE && av_strstart(seg->url, "http", NULL)) {
        v->input_read_done = 1;
    } else {
        ff_format_io_close(v->parent, &v->input);
    }

	/* 更新當前讀取的seg_no,同時更新全局的,保證流切換之后可以同步 */
    v->cur_seq_no++;

    c->cur_seq_no = v->cur_seq_no;

    goto restart; /* 回到上面繼續循環了 */
}

這是需要強調下,read_data就是完成一個I/O該做的事情,從server上去數據,放到緩沖區中,真正的數據解析是由FFmpeg的通用框架實現的。換句話來說,HLS demuxer僅僅做了HLS協議相關的解析,關於mpeg-ts/mp4的解析實際上有它們各自的demuxer完成的。

2.5 read_seek -- hls_read_seek

seek的基本思路就是按照給定的時間點(timestamp)找到對應的流的讀取位置,然后繼續讀取數據。所以在執行seek之前,需要清理下之前緩存的數據。實際代碼如下:

static int hls_read_seek(AVFormatContext *s, int stream_index,
                               int64_t timestamp, int flags)
{
    HLSContext *c = s->priv_data;
    struct playlist *seek_pls = NULL;
    int i, seq_no;
    int j;
    int stream_subdemuxer_index;
    int64_t first_timestamp, seek_timestamp, duration;

    if ((flags & AVSEEK_FLAG_BYTE) || (c->ctx->ctx_flags & AVFMTCTX_UNSEEKABLE))
        return AVERROR(ENOSYS);

    first_timestamp = c->first_timestamp == AV_NOPTS_VALUE ?
                      0 : c->first_timestamp;

    seek_timestamp = av_rescale_rnd(timestamp, AV_TIME_BASE,
                                    s->streams[stream_index]->time_base.den,
                                    flags & AVSEEK_FLAG_BACKWARD ?
                                    AV_ROUND_DOWN : AV_ROUND_UP);

    duration = s->duration == AV_NOPTS_VALUE ?
               0 : s->duration;
	/* 檢查seek位置的有效性 */
    if (0 < duration && duration < seek_timestamp - first_timestamp)
        return AVERROR(EIO);

    /* 找到stream_index對應的playlist */
    for (i = 0; i < c->n_playlists; i++) {
        struct playlist *pls = c->playlists[i];
        for (j = 0; j < pls->n_main_streams; j++) {
            if (pls->main_streams[j] == s->streams[stream_index]) {
                seek_pls = pls;
                stream_subdemuxer_index = j;
                break;
            }
        }
    }
    /* 檢查給定的seek timestamp對指定的流是否有效 */
    if (!seek_pls || !find_timestamp_in_playlist(c, seek_pls, seek_timestamp, &seq_no))
        return AVERROR(EIO);

    /* 所有參數都有效了,這就可以設置目標位置了 */
    seek_pls->cur_seq_no = seq_no;
    seek_pls->seek_stream_index = stream_subdemuxer_index;
	/* 下面是對正在讀取的流的處理 */
    for (i = 0; i < c->n_playlists; i++) {
        struct playlist *pls = c->playlists[i];
        if (pls->input)
            ff_format_io_close(pls->parent, &pls->input);
        pls->input_read_done = 0;
        if (pls->input_next)
            ff_format_io_close(pls->parent, &pls->input_next);
        pls->input_next_requested = 0;
        av_packet_unref(&pls->pkt);
        reset_packet(&pls->pkt);
        pls->pb.eof_reached = 0;
        /* 清空所有緩存的數據 */
        pls->pb.buf_end = pls->pb.buf_ptr = pls->pb.buffer;
        /* 重置讀取位置,以確保demuxer知道需要seek */
        pls->pb.pos = 0;
        /* 清空subdemuxer緩存的所有AVPacket隊列 */
        ff_read_frame_flush(pls->ctx);

        pls->seek_timestamp = seek_timestamp;
        pls->seek_flags = flags;

        if (pls != seek_pls) {
            /* 對於不是seek的playlist,將其讀取位置設置到seek目標點的位置 */
            find_timestamp_in_playlist(c, pls, seek_timestamp, &pls->cur_seq_no);
            pls->seek_stream_index = -1; /* 這兩個標志將在read_packet中使用 */
            pls->seek_flags |= AVSEEK_FLAG_ANY;
        }
    }
	/* 最后記錄seek目標位置 */
    c->cur_timestamp = seek_timestamp;

    return 0;
}

讀完seek的代碼,基本上是處理了下正在讀取的流,並沒有處理新收到的數據。對新收到數據的過濾處理明顯是在hls_read_packet中完成(通過seek_stream_index、seek_flags、seek_timestamp等參數)。需要說明的是seek操作和read_packet必須位於同一個線程,否則從上面的實現來看,明顯存在多線程邏輯上的問題。

3 小結

本文主要參考FFmpeg/libavformat/hls.c,對其代碼邏輯做了簡單收集及整理。整體來說,本文總結了ffmpeg中hls_demxuer的實現邏輯,希望對讀者有所幫助。

HLS協議中還涉及一些比較細節的部分,比如subtitle、rendetion、group、init_section、fragment mp4,對這些感興趣的建議參考HLS官方標准。


免責聲明!

本站轉載的文章為個人學習借鑒使用,本站對版權不負任何法律責任。如果侵犯了您的隱私權益,請聯系本站郵箱yoyou2525@163.com刪除。



 
粵ICP備18138465號   © 2018-2025 CODEPRJ.COM