X-Git-Url: https://git.ao2.it/libam7xxx.git/blobdiff_plain/a648584dd366450bc62a820541338fa40e93069e..refs/heads/master:/examples/am7xxx-play.c?ds=inline diff --git a/examples/am7xxx-play.c b/examples/am7xxx-play.c index 88d398a..81aff84 100644 --- a/examples/am7xxx-play.c +++ b/examples/am7xxx-play.c @@ -55,7 +55,6 @@ static int video_input_init(struct video_input_ctx *input_ctx, AVCodecContext *input_codec_ctx; AVCodec *input_codec; int video_index; - unsigned int i; int ret; avdevice_register_all(); @@ -100,29 +99,13 @@ static int video_input_init(struct video_input_ctx *input_ctx, av_dump_format(input_format_ctx, 0, input_path, 0); /* look for the first video_stream */ - video_index = -1; - for (i = 0; i < input_format_ctx->nb_streams; i++) - if (input_format_ctx->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) { - video_index = i; - break; - } - if (video_index == -1) { + video_index = av_find_best_stream(input_format_ctx, AVMEDIA_TYPE_VIDEO, -1, -1, &input_codec, 0); + if (video_index < 0) { fprintf(stderr, "cannot find any video streams\n"); ret = -EINVAL; goto cleanup; } - /* get a pointer to the codec parameters for the video stream */ - input_codec_params = input_format_ctx->streams[video_index]->codecpar; - - /* find the decoder for the video stream */ - input_codec = avcodec_find_decoder(input_codec_params->codec_id); - if (input_codec == NULL) { - fprintf(stderr, "input_codec is NULL!\n"); - ret = -EINVAL; - goto cleanup; - } - input_codec_ctx = avcodec_alloc_context3(input_codec); if (input_codec_ctx == NULL) { fprintf(stderr, "failed to allocate the input codec context\n"); @@ -130,6 +113,7 @@ static int video_input_init(struct video_input_ctx *input_ctx, goto cleanup; } + input_codec_params = input_format_ctx->streams[video_index]->codecpar; ret = avcodec_parameters_to_context(input_codec_ctx, input_codec_params); if (ret < 0) { fprintf(stderr, "cannot copy parameters to input codec context\n"); @@ -193,7 +177,7 @@ static int video_output_init(struct video_output_ctx *output_ctx, goto out; } - /* Calculate the new output dimension so the original picture is shown + /* Calculate the new output dimension so the original frame is shown * in its entirety */ ret = am7xxx_calc_scaled_image_dimensions(dev, upscale, @@ -227,6 +211,7 @@ static int video_output_init(struct video_output_ctx *output_ctx, goto out; } + /* YUVJ420P is deprecated in swscaler, but mjpeg still relies on it. */ output_codec_ctx->pix_fmt = AV_PIX_FMT_YUVJ420P; output_codec_ctx->codec_id = AV_CODEC_ID_MJPEG; output_codec_ctx->codec_type = AVMEDIA_TYPE_VIDEO; @@ -272,6 +257,67 @@ out: } +/* + * Wrap the new avcodec API from FFMpeg 3.1 to minimize the changes in the + * user code. + * + * If the use of the wrappers were to be made conditional, a check like the + * following could be used: + * + * #if (LIBAVCODEC_VERSION_INT >= AV_VERSION_INT(57, 48, 101)) + * + * As derived from the APIchanges document: + * https://github.com/FFmpeg/FFmpeg/blob/master/doc/APIchanges + * + * The wrapper implementation has been taken from: + * https://blogs.gentoo.org/lu_zero/2016/03/29/new-avcodec-api/ + */ +static int decode(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *pkt) +{ + int ret; + + *got_frame = 0; + + if (pkt) { + ret = avcodec_send_packet(avctx, pkt); + /* + * In particular, we don't expect AVERROR(EAGAIN), because we + * read all decoded frames with avcodec_receive_frame() until + * done. + */ + if (ret < 0) + return ret == AVERROR_EOF ? 0 : ret; + } + + ret = avcodec_receive_frame(avctx, frame); + if (ret < 0 && ret != AVERROR(EAGAIN) && ret != AVERROR_EOF) + return ret; + if (ret >= 0) + *got_frame = 1; + + return 0; +} + +static int encode(AVCodecContext *avctx, AVPacket *pkt, int *got_packet, AVFrame *frame) +{ + int ret; + + *got_packet = 0; + + ret = avcodec_send_frame(avctx, frame); + if (ret < 0) + return ret; + + ret = avcodec_receive_packet(avctx, pkt); + if (!ret) + *got_packet = 1; + if (ret == AVERROR(EAGAIN)) + return 0; + + return ret; +} + + static int am7xxx_play(const char *input_format_string, AVDictionary **input_options, const char *input_path, @@ -284,16 +330,16 @@ static int am7xxx_play(const char *input_format_string, { struct video_input_ctx input_ctx; struct video_output_ctx output_ctx; - AVFrame *picture_raw; - AVFrame *picture_scaled; + AVFrame *frame_raw; + AVFrame *frame_scaled; int out_buf_size; uint8_t *out_buf; - int out_picture_size; - uint8_t *out_picture; + int out_frame_size; + uint8_t *out_frame; struct SwsContext *sw_scale_ctx; AVPacket in_packet; AVPacket out_packet; - int got_picture; + int got_frame; int got_packet; int ret; @@ -310,23 +356,23 @@ static int am7xxx_play(const char *input_format_string, } /* allocate an input frame */ - picture_raw = av_frame_alloc(); - if (picture_raw == NULL) { - fprintf(stderr, "cannot allocate the raw picture frame!\n"); + frame_raw = av_frame_alloc(); + if (frame_raw == NULL) { + fprintf(stderr, "cannot allocate the raw frame!\n"); ret = -ENOMEM; goto cleanup_output; } /* allocate output frame */ - picture_scaled = av_frame_alloc(); - if (picture_scaled == NULL) { - fprintf(stderr, "cannot allocate the scaled picture!\n"); + frame_scaled = av_frame_alloc(); + if (frame_scaled == NULL) { + fprintf(stderr, "cannot allocate the scaled frame!\n"); ret = -ENOMEM; - goto cleanup_picture_raw; + goto cleanup_frame_raw; } - picture_scaled->format = (output_ctx.codec_ctx)->pix_fmt; - picture_scaled->width = (output_ctx.codec_ctx)->width; - picture_scaled->height = (output_ctx.codec_ctx)->height; + frame_scaled->format = (output_ctx.codec_ctx)->pix_fmt; + frame_scaled->width = (output_ctx.codec_ctx)->width; + frame_scaled->height = (output_ctx.codec_ctx)->height; /* calculate the bytes needed for the output image and create buffer for the output image */ out_buf_size = av_image_get_buffer_size((output_ctx.codec_ctx)->pix_fmt, @@ -337,12 +383,12 @@ static int am7xxx_play(const char *input_format_string, if (out_buf == NULL) { fprintf(stderr, "cannot allocate output data buffer!\n"); ret = -ENOMEM; - goto cleanup_picture_scaled; + goto cleanup_frame_scaled; } - /* assign appropriate parts of buffer to image planes in picture_scaled */ - av_image_fill_arrays(picture_scaled->data, - picture_scaled->linesize, + /* assign appropriate parts of buffer to image planes in frame_scaled */ + av_image_fill_arrays(frame_scaled->data, + frame_scaled->linesize, out_buf, (output_ctx.codec_ctx)->pix_fmt, (output_ctx.codec_ctx)->width, @@ -384,8 +430,8 @@ static int am7xxx_play(const char *input_format_string, } /* decode */ - got_picture = 0; - ret = avcodec_decode_video2(input_ctx.codec_ctx, picture_raw, &got_picture, &in_packet); + got_frame = 0; + ret = decode(input_ctx.codec_ctx, frame_raw, &got_frame, &in_packet); if (ret < 0) { fprintf(stderr, "cannot decode video\n"); run = 0; @@ -393,41 +439,41 @@ static int am7xxx_play(const char *input_format_string, } /* if we got the complete frame */ - if (got_picture) { - /* - * Rescaling the picture also changes its pixel format + if (got_frame) { + /* + * Rescaling the frame also changes its pixel format * to the raw format supported by the projector if * this was set in video_output_init() */ sws_scale(sw_scale_ctx, - (const uint8_t * const *)picture_raw->data, - picture_raw->linesize, + (const uint8_t * const *)frame_raw->data, + frame_raw->linesize, 0, (input_ctx.codec_ctx)->height, - picture_scaled->data, - picture_scaled->linesize); + frame_scaled->data, + frame_scaled->linesize); if (output_ctx.raw_output) { - out_picture = out_buf; - out_picture_size = out_buf_size; + out_frame = out_buf; + out_frame_size = out_buf_size; } else { - picture_scaled->quality = (output_ctx.codec_ctx)->global_quality; + frame_scaled->quality = (output_ctx.codec_ctx)->global_quality; av_init_packet(&out_packet); out_packet.data = NULL; out_packet.size = 0; got_packet = 0; - ret = avcodec_encode_video2(output_ctx.codec_ctx, - &out_packet, - picture_scaled, - &got_packet); + ret = encode(output_ctx.codec_ctx, + &out_packet, + &got_packet, + frame_scaled); if (ret < 0 || !got_packet) { fprintf(stderr, "cannot encode video\n"); run = 0; goto end_while; } - out_picture = out_packet.data; - out_picture_size = out_packet.size; + out_frame = out_packet.data; + out_frame_size = out_packet.size; } #ifdef DEBUG @@ -439,7 +485,7 @@ static int am7xxx_play(const char *input_format_string, else snprintf(filename, NAME_MAX, "out.raw"); file = fopen(filename, "wb"); - fwrite(out_picture, 1, out_picture_size, file); + fwrite(out_frame, 1, out_frame_size, file); fclose(file); } #else @@ -450,8 +496,8 @@ static int am7xxx_play(const char *input_format_string, image_format, (output_ctx.codec_ctx)->width, (output_ctx.codec_ctx)->height, - out_picture, - out_picture_size); + out_frame, + out_frame_size); if (ret < 0) { perror("am7xxx_send_image_async"); run = 0; @@ -467,10 +513,10 @@ end_while: sws_freeContext(sw_scale_ctx); cleanup_out_buf: av_free(out_buf); -cleanup_picture_scaled: - av_frame_free(&picture_scaled); -cleanup_picture_raw: - av_frame_free(&picture_raw); +cleanup_frame_scaled: + av_frame_free(&frame_scaled); +cleanup_frame_raw: + av_frame_free(&frame_raw); cleanup_output: /* Freeing the codec context is needed as well,