ffmpeg 内存转码
来源:互联网 发布:单簧管和长笛 知乎 编辑:程序博客网 时间:2024/05/17 04:22
文章的末尾有原文的地址, 这里只是记录源码方便自己时不时的看看代码
FFmpeg读写内存的关键点有2个:
1. 初始化自定义的AVIOContext,指定自定义的回调函数。
2. 自己写回调函数。注意函数的参数和返回值(尤其是返回值)。
流程
程序的流程图如下图所示。从图中可以看出,首先分别初始化了输入和输出的AVFormatContext。然后首先解码输入的AVPacket,得到存储像素数据(YUV420P格式)的AVFrame,然后编码AVFrame为H.264的AVPacket,最后将编码后的AVPacket输出。#include <stdio.h>#include <libavcodec/avcodec.h>#include <libavformat/avformat.h>#include <libavutil/avutil.h>#include <libavutil/opt.h>#include <libavutil/pixdesc.h>FILE *fp_open;FILE *fp_write;//Read Fileint read_buffer(void *opaque, uint8_t *buf, int buf_size){ if(!feof(fp_open)) { int true_size=fread(buf,1,buf_size,fp_open); return true_size; } else { return -1; }}//Write Fileint write_buffer(void *opaque, uint8_t *buf, int buf_size){ if(!feof(fp_write)) { int true_size=fwrite(buf,1,buf_size,fp_write); return true_size; } else { return -1; }}int flush_encoder(AVFormatContext *fmt_ctx,unsigned int stream_index){ int ret; int got_frame; AVPacket enc_pkt; if (!(fmt_ctx->streams[stream_index]->codec->codec->capabilities & CODEC_CAP_DELAY)) return 0; while (1) { av_log(NULL, AV_LOG_INFO, "Flushing stream #%u encoder\n", stream_index); //ret = encode_write_frame(NULL, stream_index, &got_frame); enc_pkt.data = NULL; enc_pkt.size = 0; av_init_packet(&enc_pkt); ret = avcodec_encode_video2 (fmt_ctx->streams[stream_index]->codec, &enc_pkt,NULL, &got_frame); av_frame_free(NULL); if (ret < 0) break; if (!got_frame) {ret=0;break;} /* prepare packet for muxing */ enc_pkt.stream_index = stream_index; enc_pkt.dts = av_rescale_q_rnd(enc_pkt.dts, fmt_ctx->streams[stream_index]->codec->time_base, fmt_ctx->streams[stream_index]->time_base, (AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); enc_pkt.pts = av_rescale_q_rnd(enc_pkt.pts, fmt_ctx->streams[stream_index]->codec->time_base, fmt_ctx->streams[stream_index]->time_base, (AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); enc_pkt.duration = av_rescale_q(enc_pkt.duration, fmt_ctx->streams[stream_index]->codec->time_base, fmt_ctx->streams[stream_index]->time_base); av_log(NULL, AV_LOG_DEBUG, "Muxing frame\n"); /* mux encoded frame */ ret = av_write_frame(fmt_ctx, &enc_pkt); if (ret < 0) break; } return ret;}int main(int argc, char* argv[]){ int ret; AVFormatContext* ifmt_ctx=NULL; AVFormatContext* ofmt_ctx=NULL; AVPacket packet,enc_pkt; AVFrame *frame = NULL; enum AVMediaType type; unsigned int stream_index; unsigned int i=0; int got_frame,enc_got_frame; AVStream *out_stream; AVStream *in_stream; AVCodecContext *dec_ctx, *enc_ctx; AVCodec *encoder; fp_open = fopen("/home/gh/video/segment-001.ts", "rb");//视频源文件 fp_write=fopen("./cuc60anniversary_start.h264","wb+"); //输出文件 av_register_all(); ifmt_ctx=avformat_alloc_context(); avformat_alloc_output_context2(&ofmt_ctx, NULL, "h264", NULL); unsigned char* inbuffer=NULL; unsigned char* outbuffer=NULL; inbuffer=(unsigned char*)av_malloc(32768); outbuffer=(unsigned char*)av_malloc(32768); /*open input file*/ AVIOContext *avio_in =avio_alloc_context(inbuffer, 32768,0,NULL,read_buffer,NULL,NULL); if(avio_in==NULL) goto end; ifmt_ctx->pb=avio_in; ifmt_ctx->flags=AVFMT_FLAG_CUSTOM_IO; if ((ret = avformat_open_input(&ifmt_ctx, "whatever", NULL, NULL)) < 0) { av_log(NULL, AV_LOG_ERROR, "Cannot open input file\n"); return ret; } if ((ret = avformat_find_stream_info(ifmt_ctx, NULL)) < 0) { av_log(NULL, AV_LOG_ERROR, "Cannot find stream information\n"); return ret; } for (i = 0; i < ifmt_ctx->nb_streams; i++) { AVStream *stream; AVCodecContext *codec_ctx; stream = ifmt_ctx->streams[i]; codec_ctx = stream->codec; /* Reencode video & audio and remux subtitles etc. */ if (codec_ctx->codec_type == AVMEDIA_TYPE_VIDEO) { /* Open decoder */ ret = avcodec_open2(codec_ctx, avcodec_find_decoder(codec_ctx->codec_id), NULL); if (ret < 0) { av_log(NULL, AV_LOG_ERROR, "Failed to open decoder for stream #%u\n", i); return ret; } } } //av_dump_format(ifmt_ctx, 0, "whatever", 0); /*open output file*/ AVIOContext *avio_out =avio_alloc_context(outbuffer, 32768,0,NULL,NULL,write_buffer,NULL); if(avio_out==NULL) goto end; //avio_out->write_packet=write_packet; ofmt_ctx->pb=avio_out; ofmt_ctx->flags=AVFMT_FLAG_CUSTOM_IO; for (i = 0; i < 1; i++) { out_stream = avformat_new_stream(ofmt_ctx, NULL); if (!out_stream) { av_log(NULL, AV_LOG_ERROR, "Failed allocating output stream\n"); return AVERROR_UNKNOWN; } in_stream = ifmt_ctx->streams[i]; dec_ctx = in_stream->codec; enc_ctx = out_stream->codec; if (dec_ctx->codec_type == AVMEDIA_TYPE_VIDEO) { encoder = avcodec_find_encoder(AV_CODEC_ID_H264); enc_ctx->height = dec_ctx->height; enc_ctx->width = dec_ctx->width; enc_ctx->sample_aspect_ratio = dec_ctx->sample_aspect_ratio; enc_ctx->pix_fmt = encoder->pix_fmts[0]; enc_ctx->time_base = dec_ctx->time_base; //enc_ctx->time_base.num = 1; //enc_ctx->time_base.den = 25; //H264的必备选项,没有就会错 enc_ctx->me_range=16; enc_ctx->max_qdiff = 4; enc_ctx->qmin = 10; enc_ctx->qmax = 51; enc_ctx->qcompress = 0.6; enc_ctx->refs=3; enc_ctx->bit_rate = 500000; ret = avcodec_open2(enc_ctx, encoder, NULL); if (ret < 0) { av_log(NULL, AV_LOG_ERROR, "Cannot open video encoder for stream #%u\n", i); return ret; } } else if (dec_ctx->codec_type == AVMEDIA_TYPE_UNKNOWN) { av_log(NULL, AV_LOG_FATAL, "Elementary stream #%d is of unknown type, cannot proceed\n", i); return AVERROR_INVALIDDATA; } else { /* if this stream must be remuxed */ ret = avcodec_copy_context(ofmt_ctx->streams[i]->codec, ifmt_ctx->streams[i]->codec); if (ret < 0) { av_log(NULL, AV_LOG_ERROR, "Copying stream context failed\n"); return ret; } } if (ofmt_ctx->oformat->flags & AVFMT_GLOBALHEADER) enc_ctx->flags |= CODEC_FLAG_GLOBAL_HEADER; } //av_dump_format(ofmt_ctx, 0, "whatever", 1); /* init muxer, write output file header */ ret = avformat_write_header(ofmt_ctx, NULL); if (ret < 0) { av_log(NULL, AV_LOG_ERROR, "Error occurred when opening output file\n"); return ret; } i=0; /* read all packets */ while (1) { i++; if ((ret = av_read_frame(ifmt_ctx, &packet)) < 0) break; stream_index = packet.stream_index; if(stream_index!=0) continue; type = ifmt_ctx->streams[packet.stream_index]->codec->codec_type; av_log(NULL, AV_LOG_DEBUG, "Demuxer gave frame of stream_index %u\n", stream_index); av_log(NULL, AV_LOG_DEBUG, "Going to reencode the frame\n"); frame = av_frame_alloc(); if (!frame) { ret = AVERROR(ENOMEM); break; } packet.dts = av_rescale_q_rnd(packet.dts, ifmt_ctx->streams[stream_index]->time_base, ifmt_ctx->streams[stream_index]->codec->time_base, (AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); packet.pts = av_rescale_q_rnd(packet.pts, ifmt_ctx->streams[stream_index]->time_base, ifmt_ctx->streams[stream_index]->codec->time_base, (AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); ret = avcodec_decode_video2(ifmt_ctx->streams[stream_index]->codec, frame, &got_frame, &packet); printf("Decode 1 Packet\tsize:%d\tpts:%d\n",packet.size,packet.pts); if (ret < 0) { av_frame_free(&frame); av_log(NULL, AV_LOG_ERROR, "Decoding failed\n"); break; } if (got_frame) { frame->pts = av_frame_get_best_effort_timestamp(frame); frame->pict_type=AV_PICTURE_TYPE_NONE; enc_pkt.data = NULL; enc_pkt.size = 0; av_init_packet(&enc_pkt); ret = avcodec_encode_video2 (ofmt_ctx->streams[stream_index]->codec, &enc_pkt, frame, &enc_got_frame); printf("Encode 1 Packet\tsize:%d\tpts:%d\n",enc_pkt.size,enc_pkt.pts); av_frame_free(&frame); if (ret < 0) goto end; if (!enc_got_frame) continue; /* prepare packet for muxing */ enc_pkt.stream_index = stream_index; enc_pkt.dts = av_rescale_q_rnd(enc_pkt.dts, ofmt_ctx->streams[stream_index]->codec->time_base, ofmt_ctx->streams[stream_index]->time_base, (AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); enc_pkt.pts = av_rescale_q_rnd(enc_pkt.pts, ofmt_ctx->streams[stream_index]->codec->time_base, ofmt_ctx->streams[stream_index]->time_base, (AV_ROUND_NEAR_INF|AV_ROUND_PASS_MINMAX)); enc_pkt.duration = av_rescale_q(enc_pkt.duration, ofmt_ctx->streams[stream_index]->codec->time_base, ofmt_ctx->streams[stream_index]->time_base); av_log(NULL, AV_LOG_INFO, "Muxing frame %d\n",i); /* mux encoded frame */ av_write_frame(ofmt_ctx,&enc_pkt); if (ret < 0) goto end; } else { av_frame_free(&frame); } av_free_packet(&packet); } /* flush encoders */ for (i = 0; i < 1; i++) { /* flush encoder */ ret = flush_encoder(ofmt_ctx,i); if (ret < 0) { av_log(NULL, AV_LOG_ERROR, "Flushing encoder failed\n"); goto end; } } av_write_trailer(ofmt_ctx);end: av_freep(avio_in); av_freep(avio_out); av_free(inbuffer); av_free(outbuffer); av_free_packet(&packet); av_frame_free(&frame); avformat_close_input(&ifmt_ctx); avformat_free_context(ofmt_ctx); fcloseall(); if (ret < 0) av_log(NULL, AV_LOG_ERROR, "Error occurred\n"); return (ret? 1:0);}
原文地址:http://blog.csdn.net/leixiaohua1020/article/details/39759623
0 0
- ffmpeg 内存转码
- FFMPEG基于内存的转码实例
- ffmpeg 库yuv420转jpeg(内存)
- FFMPEG基于内存的转码实例——输入输出视频均在内存
- ffmpeg 内存 demo
- ffmpeg内存释放问题
- ffmpeg解码内存缓冲区
- ffmpeg 内存读写相关
- FFMPEG中的内存读取
- ffmpeg 转码笔记
- ffmpeg视频转码
- ffmpeg转码参数
- FFmpeg 本地化转码
- ffmpeg多线程转码
- ffmpeg转码参数
- ffmpeg转码封装
- ffmpeg 转码命令
- ffmpeg()音频转码
- Spring框架4的改进
- Unity3D与Android相互传递消息
- OutputStream转InputStream的两种方案
- java枚举类型详解
- SSH三大框架的作用。
- ffmpeg 内存转码
- 知识共享平台开发-BUG[2014-11-27]
- 上班无聊,可以自己学习下别的知识或看看自己以前写的代码有没有好的改进的地方
- 真机调试报错gen_entitlements.py: No such file or directory
- Java---快排
- windows 下 搭建redis数据库 和 php安装redis 扩展
- The project was not built since its build path is incomplete. Can not find
- Jquery EasyUI 按钮上文字改为中文
- oracle里BLOB字段的HIBERNATE,spring操作访问