|
@@ -209,6 +209,11 @@ int ff_qsv_enc_init(AVCodecContext *avctx, QSVEncContext *q)
|
|
|
q->param.IOPattern = MFX_IOPATTERN_IN_SYSTEM_MEMORY;
|
|
|
q->param.AsyncDepth = q->async_depth;
|
|
|
|
|
|
+ q->async_fifo = av_fifo_alloc((1 + q->async_depth) *
|
|
|
+ (sizeof(AVPacket) + sizeof(mfxSyncPoint) + sizeof(mfxBitstream*)));
|
|
|
+ if (!q->async_fifo)
|
|
|
+ return AVERROR(ENOMEM);
|
|
|
+
|
|
|
if (avctx->hwaccel_context) {
|
|
|
AVQSVContext *qsv = avctx->hwaccel_context;
|
|
|
|
|
@@ -386,7 +391,8 @@ static void print_interlace_msg(AVCodecContext *avctx, QSVEncContext *q)
|
|
|
int ff_qsv_encode(AVCodecContext *avctx, QSVEncContext *q,
|
|
|
AVPacket *pkt, const AVFrame *frame, int *got_packet)
|
|
|
{
|
|
|
- mfxBitstream bs = { { { 0 } } };
|
|
|
+ AVPacket new_pkt = { 0 };
|
|
|
+ mfxBitstream *bs;
|
|
|
|
|
|
mfxFrameSurface1 *surf = NULL;
|
|
|
mfxSyncPoint sync = NULL;
|
|
@@ -400,16 +406,22 @@ int ff_qsv_encode(AVCodecContext *avctx, QSVEncContext *q,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- ret = ff_alloc_packet(pkt, q->packet_size);
|
|
|
+ ret = av_new_packet(&new_pkt, q->packet_size);
|
|
|
if (ret < 0) {
|
|
|
av_log(avctx, AV_LOG_ERROR, "Error allocating the output packet\n");
|
|
|
return ret;
|
|
|
}
|
|
|
- bs.Data = pkt->data;
|
|
|
- bs.MaxLength = pkt->size;
|
|
|
+
|
|
|
+ bs = av_mallocz(sizeof(*bs));
|
|
|
+ if (!bs) {
|
|
|
+ av_packet_unref(&new_pkt);
|
|
|
+ return AVERROR(ENOMEM);
|
|
|
+ }
|
|
|
+ bs->Data = new_pkt.data;
|
|
|
+ bs->MaxLength = new_pkt.size;
|
|
|
|
|
|
do {
|
|
|
- ret = MFXVideoENCODE_EncodeFrameAsync(q->session, NULL, surf, &bs, &sync);
|
|
|
+ ret = MFXVideoENCODE_EncodeFrameAsync(q->session, NULL, surf, bs, &sync);
|
|
|
if (ret == MFX_WRN_DEVICE_BUSY) {
|
|
|
av_usleep(1);
|
|
|
continue;
|
|
@@ -418,6 +430,8 @@ int ff_qsv_encode(AVCodecContext *avctx, QSVEncContext *q,
|
|
|
} while ( 1 );
|
|
|
|
|
|
if (ret < 0) {
|
|
|
+ av_packet_unref(&new_pkt);
|
|
|
+ av_freep(&bs);
|
|
|
if (ret == MFX_ERR_MORE_DATA)
|
|
|
return 0;
|
|
|
av_log(avctx, AV_LOG_ERROR, "EncodeFrameAsync returned %d\n", ret);
|
|
@@ -432,22 +446,56 @@ int ff_qsv_encode(AVCodecContext *avctx, QSVEncContext *q,
|
|
|
"EncodeFrameAsync returned 'incompatible param' code\n");
|
|
|
}
|
|
|
if (sync) {
|
|
|
+ av_fifo_generic_write(q->async_fifo, &new_pkt, sizeof(new_pkt), NULL);
|
|
|
+ av_fifo_generic_write(q->async_fifo, &sync, sizeof(sync), NULL);
|
|
|
+ av_fifo_generic_write(q->async_fifo, &bs, sizeof(bs), NULL);
|
|
|
+ } else {
|
|
|
+ av_packet_unref(&new_pkt);
|
|
|
+ av_freep(&bs);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (!av_fifo_space(q->async_fifo) ||
|
|
|
+ (!frame && av_fifo_size(q->async_fifo))) {
|
|
|
+ av_fifo_generic_read(q->async_fifo, &new_pkt, sizeof(new_pkt), NULL);
|
|
|
+ av_fifo_generic_read(q->async_fifo, &sync, sizeof(sync), NULL);
|
|
|
+ av_fifo_generic_read(q->async_fifo, &bs, sizeof(bs), NULL);
|
|
|
+
|
|
|
MFXVideoCORE_SyncOperation(q->session, sync, 60000);
|
|
|
|
|
|
- if (bs.FrameType & MFX_FRAMETYPE_I || bs.FrameType & MFX_FRAMETYPE_xI)
|
|
|
+ new_pkt.dts = av_rescale_q(bs->DecodeTimeStamp, (AVRational){1, 90000}, avctx->time_base);
|
|
|
+ new_pkt.pts = av_rescale_q(bs->TimeStamp, (AVRational){1, 90000}, avctx->time_base);
|
|
|
+ new_pkt.size = bs->DataLength;
|
|
|
+
|
|
|
+ if (bs->FrameType & MFX_FRAMETYPE_IDR ||
|
|
|
+ bs->FrameType & MFX_FRAMETYPE_xIDR)
|
|
|
+ new_pkt.flags |= AV_PKT_FLAG_KEY;
|
|
|
+
|
|
|
+ if (bs->FrameType & MFX_FRAMETYPE_I || bs->FrameType & MFX_FRAMETYPE_xI)
|
|
|
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_I;
|
|
|
- else if (bs.FrameType & MFX_FRAMETYPE_P || bs.FrameType & MFX_FRAMETYPE_xP)
|
|
|
+ else if (bs->FrameType & MFX_FRAMETYPE_P || bs->FrameType & MFX_FRAMETYPE_xP)
|
|
|
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_P;
|
|
|
- else if (bs.FrameType & MFX_FRAMETYPE_B || bs.FrameType & MFX_FRAMETYPE_xB)
|
|
|
+ else if (bs->FrameType & MFX_FRAMETYPE_B || bs->FrameType & MFX_FRAMETYPE_xB)
|
|
|
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_B;
|
|
|
|
|
|
- pkt->dts = av_rescale_q(bs.DecodeTimeStamp, (AVRational){1, 90000}, avctx->time_base);
|
|
|
- pkt->pts = av_rescale_q(bs.TimeStamp, (AVRational){1, 90000}, avctx->time_base);
|
|
|
- pkt->size = bs.DataLength;
|
|
|
+ av_freep(&bs);
|
|
|
+
|
|
|
+ if (pkt->data) {
|
|
|
+ if (pkt->size < new_pkt.size) {
|
|
|
+ av_log(avctx, AV_LOG_ERROR, "Submitted buffer not large enough: %d < %d\n",
|
|
|
+ pkt->size, new_pkt.size);
|
|
|
+ av_packet_unref(&new_pkt);
|
|
|
+ return AVERROR(EINVAL);
|
|
|
+ }
|
|
|
+
|
|
|
+ memcpy(pkt->data, new_pkt.data, new_pkt.size);
|
|
|
+ pkt->size = new_pkt.size;
|
|
|
|
|
|
- if (bs.FrameType & MFX_FRAMETYPE_IDR ||
|
|
|
- bs.FrameType & MFX_FRAMETYPE_xIDR)
|
|
|
- pkt->flags |= AV_PKT_FLAG_KEY;
|
|
|
+ ret = av_packet_copy_props(pkt, &new_pkt);
|
|
|
+ av_packet_unref(&new_pkt);
|
|
|
+ if (ret < 0)
|
|
|
+ return ret;
|
|
|
+ } else
|
|
|
+ *pkt = new_pkt;
|
|
|
|
|
|
*got_packet = 1;
|
|
|
}
|
|
@@ -472,6 +520,21 @@ int ff_qsv_enc_close(AVCodecContext *avctx, QSVEncContext *q)
|
|
|
cur = q->work_frames;
|
|
|
}
|
|
|
|
|
|
+ while (q->async_fifo && av_fifo_size(q->async_fifo)) {
|
|
|
+ AVPacket pkt;
|
|
|
+ mfxSyncPoint sync;
|
|
|
+ mfxBitstream *bs;
|
|
|
+
|
|
|
+ av_fifo_generic_read(q->async_fifo, &pkt, sizeof(pkt), NULL);
|
|
|
+ av_fifo_generic_read(q->async_fifo, &sync, sizeof(sync), NULL);
|
|
|
+ av_fifo_generic_read(q->async_fifo, &bs, sizeof(bs), NULL);
|
|
|
+
|
|
|
+ av_freep(&bs);
|
|
|
+ av_packet_unref(&pkt);
|
|
|
+ }
|
|
|
+ av_fifo_free(q->async_fifo);
|
|
|
+ q->async_fifo = NULL;
|
|
|
+
|
|
|
av_frame_free(&avctx->coded_frame);
|
|
|
|
|
|
return 0;
|