ffmpeg + ios 移植ffmpeg 到ios 平台上
来源:互联网 发布:外汇储备最新数据 编辑:程序博客网 时间:2024/06/05 03:29
声明:本文章原创,禁止转载。分享出来,主要是在csdn上收获了这么多,该贡献点了
说实话移植ffmpeg 到mac ox 和 模拟器 上基本没啥难度,基本就是config && make && make install,也没有遇到啥难度,在模拟器上 运行没啥问题,问题的重点在移植到ios 平台上。
以前也知道交叉编译是怎么一回事,但是没有实战过,ide 用多了,对gcc 那一套选项根本就动不了,对ffmpeg 的 编译选项也是动不了,怎么办呢,慢慢啃吧,这是咋程序员的命。
那么就要对这些知识温故一下才有法了,要不然 编译的时候copy人家得脚本,然后一出问题就找不到 南北了。精简版 支持 mp4 + m3u8 的编译脚本如下,我的imac 不是i386处理器,是x86_64, 这个也是个坑,网上都说模拟器是 i386 处理器,尼玛我的就怎么不一样,编译脚本如下,如果你懂原理 基本没啥难度。
#!/bin/sh# FFMpeg,SDK版本号VERSION="2.1"SDKVERSION="8.2"#最低支持的SDK版本号MINSDKVERSION="7.0"# 源文件路径SRCDIR="$(pwd)"BUILDDIR="${SRCDIR}/build"mkdir -p $BUILDDIR# 获取xcode开发环境安装路径DEVELOPER=`xcode-select -print-path`# 要编译的架构列表ARCHS="x86_64,armv7,arm64"for ARCH in ${ARCHS}doif [ "${ARCH}" == "x86_64" ];thenPLATFORM="iPhoneSimulator"EXTRA_CFLAGS="-arch x86_64"EXTRA_LDFLAGS="-arch x86_64 -mfpu=neon"EXTRA_CONFIG="--arch=x86_64 --cpu=x86_64"elsePLATFORM="iPhoneOS"EXTRA_CFLAGS="-arch ${ARCH} -mfloat-abi=softfp"EXTRA_LDFLAGS="-arch ${ARCH} -mfpu=neon -mfloat-abi=softfp"EXTRA_CONFIG="--arch=${ARCH} --disable-armv5te"fimake clean# 根据需要裁减./configure --prefix="${BUILDDIR}/${ARCH}" \--disable-doc \--disable-ffprobe \--disable-ffmpeg \--disable-ffplay \--disable-ffserver \--disable-debug \--disable-zlib \--disable-encoders \--disable-muxers \--disable-devices \--disable-filters \--disable-bsfs \--disable-demuxers \--enable-demuxer=hls \--enable-demuxer=mpegts \--enable-demuxer=mpegtsraw \--enable-demuxer=h264 \--enable-demuxer=m4v \--enable-demuxer=mov \--enable-demuxer=mpegvideo \--enable-demuxer=yuv4mpegpipe \--disable-protocols \--enable-protocol=file \--enable-protocol=http \--enable-protocol=hls \--disable-decoders \--enable-decoder=h264 \--enable-decoder=h264_crystalhd \--enable-decoder=h264_vda \--enable-decoder=h264_vdpau \--enable-decoder=aac \--enable-decoder=aac_latm \--enable-decoder=tscc \--enable-decoder=tscc2 \--enable-decoder=mts2 \--disable-parsers \--enable-parser=aac \--enable-parser=aac_latm \--enable-parser=h264 \--enable-parser=mpeg4video \--enable-cross-compile \--enable-pic \--disable-asm \--target-os=darwin \--enable-small \${EXTRA_CONFIG} \--cc="${DEVELOPER}/Platforms/${PLATFORM}.platform/Developer/usr/bin/gcc" \--as="/usr/bin/gas-preprocessor.pl" \--sysroot="${DEVELOPER}/Platforms/${PLATFORM}.platform/Developer/SDKs/${PLATFORM}${SDKVERSION}.sdk" \--extra-cflags="-miphoneos-version-min=${MINSDKVERSION} ${EXTRA_CFLAGS}" \--extra-ldflags="-miphoneos-version-min=${MINSDKVERSION} ${EXTRA_LDFLAGS} -isysroot ${DEVELOPER}/Platforms/${PLATFORM}.platform/Developer/SDKs/${PLATFORM}${SDKVERSION}.sdk"make && make install && make cleandone是否开启硬件加速,这个要根据处理器来,在 cflags 和 ldflags 中,指定选项。 需要合并*.a 库的 lipo 工具可以帮到你。
好吧,好人做到底,贴上测试 播放器代码 opengl + openal。也可以用苹果的CoreGraphics + audioqueuf,这个你随意。
//// CorePlayer.c// player//// Created by KuaiyuIOS on 15/4/16.// Copyright (c) 2015年 KuaiyuD. All rights reserved.//#include "CorePlayer.h"#include <stdlib.h>#include <string.h>#include <pthread/pthread.h>#include <OpenAL/al.h>#include <OpenAL/alc.h>#include <unistd.h>#include "libavcodec/avcodec.h"#include "libavformat/avformat.h"#include "libswresample/swresample.h"#include "libswscale/swscale.h"#include "libavutil/time.h"#define kTwoConverImgThread 0#define QueueSize 26#define AudioBufferMaxSize 26#define SYN_PTS 500#define VideoType PIX_FMT_RGBA#define AudioType AV_SAMPLE_FMT_S16#define MAX_AUDIO_FRAME_SIZE 192000 // 1 second of 48khz 32bit audio//typedef struct videoItem{// uint8_t *rgbData;// int width;// int height;// int length;// int64_t pts;//} VideoItem;typedef struct rgbVideoItem{ uint8_t *rgbData; int widht; int height; int length; int64_t pts;}RGBVideoItem;typedef struct packetquque { int size; int lenght; void **items; //指向具体的item void (*free_cb)(void *); double pts; pthread_mutex_t mutex; pthread_cond_t cond;}PacketQueue;//openal 无需持有 一旦有的data 马上让 opanal 持有typedef struct audioQueue{ int inUse[AudioBufferMaxSize]; ALuint alBuffers[AudioBufferMaxSize]; int64_t pts[AudioBufferMaxSize]; int itemIndex; pthread_mutex_t mutex; pthread_cond_t cond;} AudioQueue;struct CPContext { //base char url[256]; double duratioin; double progress; double volume; void *userinfo; void (*dis_callback)(uint8_t *data,int lenght,int width,int height,void *info); void (*dis_audio_callback)(uint8_t *audioDta,int lenght,void *info); void (*init_callback)(int width,int height,void *info); void (*finish_callback)(int flag,void *info); CPContex **p_ref; int eof; int seek; int seekPos; PacketQueue *video_q; PacketQueue *picture_q; AudioQueue *audio_q; pthread_t read_t; pthread_t decode_audio_t; pthread_t decode_video_t; pthread_t decode_video_other_t; pthread_t display_picture_t; pthread_mutex_t lock_muxt; pthread_cond_t lock_cond; int play; int last_play; int quit; //时钟 int64_t audio_pts_curr; //openal ALCcontext *al_ctx; ALCdevice *al_device; ALuint al_source; int videoLength; int audioLength; struct SwsContext *img_convert_ctx; struct SwrContext *aud_convert_ctx; AVCodecContext *decode_video_ctx; AVCodecContext *decode_audio_ctx; };void seconds_sleep(unsigned seconds){ struct timeval tv = {seconds,0}; int err; do{ err = select(0, NULL, NULL, NULL, &tv); }while (err < 0 && errno == EINTR);}void suseconds_sleep(double suseconds){ struct timeval tv = {0,(int)(suseconds * 1000)}; int err; do{ err = select(0, NULL, NULL, NULL, &tv); }while (err < 0 && errno == EINTR);}#pragma mark ==============================#pragma mark ==[quque item free callback]==#pragma mark ==============================void packet_quque_free_callback(void *item){ AVPacket *packet = (AVPacket *)item; av_free_packet(packet);}void frame_queue_free_callback(void *item){ AVFrame *frame = (AVFrame *)item; av_frame_free(&frame);}void rgb_video_quque_free_callback(void *item){ RGBVideoItem *videoitem = (RGBVideoItem *)item; av_free(videoitem->rgbData); av_free(item);}#pragma mark =========================#pragma mark ==[quque operation]==#pragma mark =========================int queue_init(PacketQueue **q,void (*free_cb)(void *)){ PacketQueue *tempq = av_mallocz(sizeof(PacketQueue)); tempq->size = QueueSize; tempq->items = (void **)av_mallocz(sizeof(void *) * QueueSize); tempq->free_cb = free_cb; pthread_mutex_init(&tempq->mutex, NULL); pthread_cond_init(&tempq->cond, NULL); *q = tempq; return 0;}int queue_flush(PacketQueue *q){ pthread_mutex_lock(&q->mutex); for(int i = 0;i < q->lenght;i++) { q->free_cb(q->items[i]); } q->lenght = 0; pthread_mutex_unlock(&q->mutex); return 0;}int queue_uninit(PacketQueue **q){ PacketQueue *tempq = *q; pthread_cond_broadcast(&tempq->cond); queue_flush(tempq); av_free((void *)tempq->items); pthread_mutex_destroy(&tempq->mutex); pthread_cond_destroy(&tempq->cond); av_free(tempq); *q = NULL; return 0;}int queue_push(PacketQueue *q,void *item){ pthread_mutex_lock(&q->mutex); while (q->lenght + 1 >= q->size) { pthread_cond_wait(&q->cond, &q->mutex); } q->items[q->lenght] = item; q->lenght ++; pthread_mutex_unlock(&q->mutex); pthread_cond_signal(&q->cond); return 0;}int queue_get(PacketQueue *q,void **item ,void *info){ CPContex *ctx = (CPContex *)info; pthread_mutex_lock(&q->mutex); while (q->lenght <= 0) { pthread_cond_wait(&q->cond, &q->mutex); if(ctx->quit || ctx->eof) { *item = NULL; pthread_mutex_unlock(&q->mutex); return -1; } } q->lenght --; *item = q->items[0]; if(q->lenght > 0) //前移动 { memcpy(&q->items[0], &q->items[1], sizeof(void *) * q->lenght); } q->items[q->lenght] = NULL; pthread_mutex_unlock(&q->mutex); pthread_cond_signal(&q->cond); return 0;}int audio_queue_init(AudioQueue **q){ AudioQueue *tempq = av_mallocz(sizeof(AudioQueue)); pthread_mutex_init(&tempq->mutex, NULL); pthread_cond_init(&tempq->cond, NULL); *q = tempq; return 0;}int audio_queue_uninit(AudioQueue **q){ AudioQueue *tempq = *q; pthread_cond_broadcast(&tempq->cond); pthread_mutex_lock(&tempq->mutex); for(int i = 0;i < AudioBufferMaxSize;i++){ if(tempq->alBuffers[i] != 0) { alDeleteBuffers(1, &tempq->alBuffers[i]); } } pthread_mutex_unlock(&tempq->mutex); pthread_mutex_destroy(&tempq->mutex); pthread_cond_destroy(&tempq->cond); *q = NULL; return 0;}int audio_queue_put(AudioQueue *q,uint8_t *audioData,int length,int64_t pts,void *info){ CPContex *ctx = (CPContex *)info; pthread_mutex_lock(&q->mutex); while (q->inUse[q->itemIndex]) { pthread_cond_wait(&q->cond, &q->mutex); if(ctx->quit || ctx->eof) { pthread_mutex_unlock(&q->mutex); return -1; } } q->inUse[q->itemIndex] = 1; pthread_mutex_unlock(&q->mutex); ALuint buffid = q->alBuffers[q->itemIndex]; if(buffid <= 0){ alGenBuffers(1, &buffid); q->alBuffers[q->itemIndex] = buffid; } alBufferData(buffid, AL_FORMAT_STEREO16, audioData, length, ctx ->decode_audio_ctx->sample_rate); alSourceQueueBuffers(ctx->al_source, 1, &buffid); q->pts[q->itemIndex] = pts;// ctx->audio_pts_curr = pts; if(ctx->dis_audio_callback) { ctx->dis_audio_callback(audioData,length,ctx->userinfo); } q->itemIndex ++; if(q->itemIndex >= AudioBufferMaxSize) q->itemIndex = 0; return 0;}int audio_queue_flush(AudioQueue *q){ return 0;}#pragma mark =========================#pragma mark ==[three thread]==#pragma mark =========================static const char * read_thread_name = "com.CorePlayer.read";static const char * decode_audio_thread_name = "com.CorePlayer.decodeAudio";static const char * decode_video_thread_name = "com.CorePlayer.decodeVideo";static const char * display_picture_thread_name = "com.CorePlayer.displayPicture";void * decode_audio_thread(void *info){ pthread_setname_np(decode_audio_thread_name); CPContex *ctx = (CPContex *)info; while (ctx->quit == 0 && ctx->eof == 0) { pthread_mutex_lock(&ctx->lock_muxt); while (!ctx->play) { pthread_cond_wait(&ctx->lock_cond, &ctx->lock_muxt); } pthread_mutex_unlock(&ctx->lock_muxt); if(ctx->quit || ctx->eof) break; //暂停 suseconds_sleep(0.75); if(ctx->quit || ctx->eof) break; if(!ctx->play) continue; //清空 ALint stateValue; int processed,queued; ALfloat sec_off; alGetSourcei(ctx->al_source, AL_BUFFERS_PROCESSED, &processed); alGetSourcei(ctx->al_source, AL_BUFFERS_QUEUED, &queued); alGetSourcei(ctx->al_source, AL_SOURCE_STATE, &stateValue); alGetSourcef(ctx->al_source, AL_SEC_OFFSET, &sec_off); ctx->progress += sec_off; while (processed) { ALuint buffId; alSourceUnqueueBuffers(ctx->al_source, 1, &buffId); processed --; int buffIndex = -1; for (int i = 0; i < AudioBufferMaxSize;i ++) { if(buffId == ctx->audio_q->alBuffers[i]) { buffIndex = i; break; } } if(buffIndex != -1) { ctx->audio_pts_curr = ctx->audio_q->pts[buffIndex]; pthread_mutex_lock(&ctx->audio_q->mutex); ctx->audio_q->inUse[buffIndex] = 0; pthread_cond_signal(&ctx->audio_q->cond); pthread_mutex_unlock(&ctx->audio_q->mutex); } } if(stateValue != AL_PLAYING && queued >= 5){ alSourcePlay(ctx->al_source); } else if(stateValue == AL_PLAYING && queued < 5){ alSourcePause(ctx->al_source); } } return NULL;}void * display_picture_thread(void *info){ pthread_setname_np(display_picture_thread_name); CPContex *ctx = (CPContex *)info; while (ctx->quit == 0 && ctx->eof == 0) { pthread_mutex_lock(&ctx->lock_muxt); while (!ctx->play) { pthread_cond_wait(&ctx->lock_cond, &ctx->lock_muxt); } pthread_mutex_unlock(&ctx->lock_muxt); if(ctx->quit || ctx->eof) break; void *item = NULL; queue_get(ctx->picture_q, (void **)&item,info); if(item != NULL) { RGBVideoItem *newItem = (RGBVideoItem *)item; if(ctx->dis_callback /*&& newItem->pts != AV_NOPTS_VALUE*/ ) ctx->dis_callback(newItem->rgbData ,newItem->length,newItem->widht,newItem->height,ctx->userinfo); ctx->picture_q->free_cb(item); } } return NULL;}void decode_video_inline(void *info){ CPContex *ctx = (CPContex *)info; AVFrame *pframeRGB = av_frame_alloc(); while (ctx->quit == 0 && ctx->eof == 0) { pthread_mutex_lock(&ctx->lock_muxt); while (!ctx->play) { pthread_cond_wait(&ctx->lock_cond, &ctx->lock_muxt); } pthread_mutex_unlock(&ctx->lock_muxt); if(ctx->quit || ctx->eof) break; while (ctx->seek){ //等待清空完毕 suseconds_sleep(0.2); } if(ctx->quit || ctx->eof) break; AVFrame *pframe = NULL; queue_get(ctx->video_q, (void **)&pframe,info); if(pframe != NULL){ uint8_t *buff = (uint8_t *)av_mallocz(ctx->videoLength); avpicture_fill((AVPicture *)pframeRGB, buff, VideoType, ctx->decode_video_ctx->width, ctx->decode_video_ctx->height); sws_scale(ctx->img_convert_ctx, (const uint8_t * const *)pframe->data, pframe->linesize, 0, ctx->decode_video_ctx->height, pframeRGB->data, pframeRGB->linesize); RGBVideoItem *item = av_mallocz(sizeof(RGBVideoItem)); item->rgbData = buff; item->widht = ctx->decode_video_ctx->width; item->height = ctx->decode_video_ctx->height; item->length = ctx->videoLength; item->pts = pframe->pts == AV_NOPTS_VALUE ? (pframe->pkt_pts == AV_NOPTS_VALUE ? pframe->pkt_dts : pframe->pkt_pts) : pframe->pts; queue_push(ctx->picture_q, item); ctx->video_q->free_cb(pframe); } } av_frame_free(&pframeRGB);}#if kTwoConverImgThreadstatic const char * decode_video_thread_other = "com.CorePlayer.other.decodeVideo";void *decode_video_thread_two(void *info){ pthread_setname_np(decode_video_thread_other); decode_video_inline(info); return NULL;}#endifvoid *decode_video_thread(void *info){ pthread_setname_np(decode_video_thread_name); CPContex *ctx = (CPContex *)info;#if kTwoConverImgThread pthread_create(&ctx->decode_video_other_t, NULL, decode_video_thread_two , info);#endif queue_init(&ctx->picture_q, rgb_video_quque_free_callback); pthread_create(&ctx->display_picture_t, NULL, display_picture_thread, info); decode_video_inline(info); #if kTwoConverImgThread pthread_join(ctx->decode_video_other_t, NULL);#endif queue_uninit(&ctx->picture_q); pthread_join(ctx->display_picture_t, NULL); return NULL;}void * read_thread(void *info){ pthread_setname_np(read_thread_name); pthread_detach(pthread_self()); CPContex *ctx = (CPContex *)info; //完成必要的初始化 ctx->al_device = alcOpenDevice(NULL); if(ctx->al_device != 0){ ctx->al_ctx = alcCreateContext(ctx->al_device, NULL); alcMakeContextCurrent(ctx->al_ctx); } alGenSources(1, &ctx->al_source); alSourcei(ctx->al_source, AL_LOOPING, AL_FALSE); alSourcei(ctx->al_source, AL_GAIN, 1.0f); //声音 alSpeedOfSound(1.0f); queue_init(&ctx->video_q, frame_queue_free_callback); audio_queue_init(&ctx->audio_q); pthread_mutex_init(&ctx->lock_muxt, NULL); pthread_cond_init(&ctx->lock_cond, NULL); av_register_all(); avformat_network_init(); int ret = -1,stream_index_a = -1,stream_index_v = -1; AVFormatContext *fmt_ctx = NULL; AVCodec *dec_a = NULL,*dec_v = NULL; AVCodecContext *dec_ctx_a = NULL,*dec_ctx_v = NULL; AVPacket pkt,*pktptr = &pkt; AVFrame *pframe = av_frame_alloc(),*pframeRGB = av_frame_alloc(); int frameFinished = 0,audioLength = 0,videoLength = 0; struct SwsContext *img_convert_ctx = NULL; struct SwrContext *aud_convert_ctx = NULL; pthread_mutex_t wait_muxt; pthread_cond_t wait_cond; pthread_mutex_init(&wait_muxt, NULL); pthread_cond_init(&wait_cond, NULL); if((ret = avformat_open_input(&fmt_ctx, ctx->url, NULL, NULL)) < 0){ printf("open file fail\n"); goto lable; } if((ret = avformat_find_stream_info(fmt_ctx, NULL)) < 0){ printf("not more stream info \n"); goto lable; } ctx->duratioin = (double)(fmt_ctx->duration / AV_TIME_BASE); ret = av_find_best_stream(fmt_ctx, AVMEDIA_TYPE_VIDEO, -1, -1, &dec_v, 0); if(ret != AVERROR_STREAM_NOT_FOUND){ stream_index_v = ret; dec_ctx_v = fmt_ctx->streams[ret]->codec; ret = avcodec_open2(dec_ctx_v, dec_v, NULL); if(ret < 0) { printf("open video decode fail\n"); } } ret = av_find_best_stream(fmt_ctx, AVMEDIA_TYPE_AUDIO, -1, -1, &dec_a, 0); if(ret != AVERROR_STREAM_NOT_FOUND){ stream_index_a = ret; dec_ctx_a = fmt_ctx->streams[ret]->codec; ret = avcodec_open2(dec_ctx_a, dec_a, NULL); if(ret < 0){ printf("open audio decode fail\n"); } } if(stream_index_v == -1 && stream_index_a == -1) { printf("no effect stream \n"); goto lable; } if(stream_index_v >= 0){ videoLength = avpicture_get_size(VideoType, dec_ctx_v->width, dec_ctx_v->height); img_convert_ctx = sws_getContext(dec_ctx_v->width, dec_ctx_v->height, dec_ctx_v->pix_fmt, dec_ctx_v->width, dec_ctx_v->height, VideoType, SWS_FAST_BILINEAR, NULL, NULL, NULL); ctx->decode_video_ctx = dec_ctx_v; ctx->videoLength = videoLength; ctx->img_convert_ctx = img_convert_ctx; pthread_create(&ctx->decode_video_t, NULL, decode_video_thread, info); } if(stream_index_a >= 0){ audioLength = av_samples_get_buffer_size(NULL, av_get_channel_layout_nb_channels(AV_CH_LAYOUT_STEREO), dec_ctx_a->frame_size, AudioType, 1); aud_convert_ctx = swr_alloc_set_opts(NULL, AV_CH_LAYOUT_STEREO, AudioType, dec_ctx_a->sample_rate, dec_ctx_a->channel_layout, dec_ctx_a->sample_fmt, dec_ctx_a->sample_rate, 0, NULL); swr_init(aud_convert_ctx); ctx->aud_convert_ctx = aud_convert_ctx; ctx->decode_audio_ctx = dec_ctx_a; ctx->audioLength = audioLength; pthread_create(&ctx->decode_audio_t, NULL, decode_audio_thread, info); } if(/* DISABLES CODE */ (0)){ av_dump_format(fmt_ctx, 0, ctx->url, 0); } if(ctx->init_callback){ ctx->init_callback(dec_ctx_v->width,dec_ctx_v->height,ctx->userinfo); } pthread_mutex_lock(&ctx->lock_muxt); while (!ctx->play) { pthread_cond_wait(&ctx->lock_cond, &ctx->lock_muxt); } pthread_mutex_unlock(&ctx->lock_muxt); while (ctx->quit == 0 && ctx->eof == 0) { //是否需要暂停网络 if(ctx->play != ctx->last_play) { ctx->last_play = ctx->play; if(ctx->play) av_read_pause(fmt_ctx); else av_read_play(fmt_ctx); } //是否暂停 pthread_mutex_lock(&ctx->lock_muxt); while (!ctx->play) { pthread_cond_wait(&ctx->lock_cond, &ctx->lock_muxt); } pthread_mutex_unlock(&ctx->lock_muxt); if(ctx->quit && ctx->eof) break; //是否快进 if(ctx->seek) { //new api for seek int seek_ret = avformat_seek_file(fmt_ctx, -1, (ctx->seekPos - 10) * AV_TIME_BASE, (ctx->seekPos) * AV_TIME_BASE, (ctx->seekPos + 10) *AV_TIME_BASE, AVSEEK_FLAG_ANY); if(seek_ret < 0){ printf("error while seeking\n"); }else{ if(stream_index_v >= 0) { queue_flush(ctx->video_q); } if(stream_index_a >= 0) { audio_queue_flush(ctx->audio_q); ctx->progress = ctx->seekPos; } //设置 exl clocl } ctx->seek = 0; ctx->eof = 0; } frameFinished = 0; ret = av_read_frame(fmt_ctx, pktptr); if(ret < 0){ if((ret == AVERROR_EOF || avio_feof(fmt_ctx->pb)) && !ctx->eof){ ctx->eof = 1; } if(fmt_ctx->pb && fmt_ctx->pb->error) break; /* sleep */ struct timespec tm = {10,0}; pthread_mutex_lock(&wait_muxt); pthread_cond_timedwait(&wait_cond, &wait_muxt, &tm); pthread_mutex_unlock(&wait_muxt); continue; } else{ ctx->eof = 0; } if(pktptr->stream_index == stream_index_v) { ret = avcodec_decode_video2(ctx->decode_video_ctx,pframe, &frameFinished, pktptr); if(ret < 0) break; if(frameFinished) { AVFrame *newFrame = av_frame_alloc(); av_frame_copy_props(newFrame, pframe); newFrame->format = pframe->format; newFrame->width = dec_ctx_v->width; newFrame->height = dec_ctx_v->height; ret = av_frame_get_buffer(newFrame, 1); av_frame_copy(newFrame, pframe); queue_push(ctx->video_q, newFrame); } } else if(pktptr->stream_index == stream_index_a) { ret = avcodec_decode_audio4(dec_ctx_a, pframe, &frameFinished, pktptr); if(ret < 0) break; if(frameFinished){ uint8_t *buff = (uint8_t *)av_mallocz(MAX_AUDIO_FRAME_SIZE * 2); swr_convert(aud_convert_ctx, &buff, MAX_AUDIO_FRAME_SIZE, (const uint8_t **)pframe->data, pframe->nb_samples); int64_t pts = (pframe->pts == AV_NOPTS_VALUE ? pframe->pkt_pts : pframe->pts); audio_queue_put(ctx->audio_q, buff, audioLength, pts,info); av_free(buff); } } av_free_packet(pktptr); } lable: avformat_network_deinit(); pthread_mutex_destroy(&wait_muxt); pthread_cond_destroy(&wait_cond); av_frame_free(&pframe); av_frame_free(&pframeRGB); pthread_mutex_destroy(&ctx->lock_muxt); pthread_cond_destroy(&ctx->lock_cond); ctx->eof = 1; audio_queue_uninit(&ctx->audio_q); queue_uninit(&ctx->video_q); if(ctx->decode_video_t != 0) pthread_join(ctx->decode_video_t, NULL); if(ctx->decode_audio_t != 0) pthread_join(ctx->decode_audio_t, NULL); if(aud_convert_ctx != NULL) swr_free(&aud_convert_ctx); if(img_convert_ctx != NULL) sws_freeContext(img_convert_ctx); if(dec_ctx_a) avcodec_close(dec_ctx_a); if(dec_ctx_v) avcodec_close(dec_ctx_v); if(fmt_ctx) avformat_close_input(&fmt_ctx); alSourceStop(ctx->al_source); alDeleteSources(1, &ctx->al_source); if(ctx->al_device){ alcCloseDevice(ctx->al_device); } if(ctx->finish_callback){ int quit_flag = (ctx->quit == 1) ? 1 : 0; ctx->finish_callback(quit_flag,ctx->userinfo); } *(ctx->p_ref) = NULL; av_free(ctx); return NULL;}#pragma mark =========================#pragma mark ==[out controll method]==#pragma mark =========================int CP_start(CPContex *cp_ctx){ if(! cp_ctx ) return -1; pthread_mutex_lock(&cp_ctx->lock_muxt); cp_ctx->play = 1; pthread_cond_broadcast(&cp_ctx->lock_cond); pthread_mutex_unlock(&cp_ctx->lock_muxt); alSourcePlay(cp_ctx->al_source); return 0;}int CP_pasue(CPContex *cp_ctx){ if(! cp_ctx ) return -1; pthread_mutex_lock(&cp_ctx->lock_muxt); cp_ctx->play = 0; pthread_mutex_unlock(&cp_ctx->lock_muxt); alSourcePause(cp_ctx->al_source); return 0;}int CP_stop(CPContex *cp_ctx){ if(! cp_ctx ) return -1; cp_ctx->quit = 1; if(cp_ctx->play == 0){ pthread_mutex_lock(&cp_ctx->lock_muxt); cp_ctx->play = 1; pthread_cond_broadcast(&cp_ctx->lock_cond); pthread_mutex_unlock(&cp_ctx->lock_muxt); } return 0;}int CP_init(CPContex **cp_ctx, const char *url, void (*dis_callback)(uint8_t *data,int lenght,int width,int height,void *info), void (*dis_audio_callback)(uint8_t *audioData,int length,void *info), void (*init_callback)(int width,int height,void *info), void *userinfo){ // checkurl CPContex *temp = av_mallocz(sizeof(CPContex)); strcpy(temp->url, url); temp->userinfo = userinfo; temp->dis_callback = dis_callback; temp->dis_audio_callback = dis_audio_callback; temp->init_callback = init_callback; temp->p_ref = cp_ctx; // pthread_create(&temp->read_t, NULL, read_thread, temp); *cp_ctx = temp; return 0;}int CP_set_finish_callback(CPContex *cp_ctx,void (*finish_callback)(int flag,void*info)){ if(! cp_ctx ) return -1; cp_ctx->finish_callback = finish_callback; return 0;}int CP_seek(CPContex *cp_ctx,double pos,int flag){ if(! cp_ctx ) return -1; if(cp_ctx->seek == 1) return -2; if(flag == 0){ cp_ctx->seek = 1; cp_ctx->seekPos = pos; } return 0;}int CP_get_progress(CPContex *cp_ctx,double *progress){ if(! cp_ctx ) return -1; *progress = cp_ctx->progress; return 0;}int CP_get_duration(CPContex *cp_ctx,double *duration){ if(! cp_ctx ) return -1; *duration = cp_ctx->duratioin; return 0;}
以下为真机和模拟器测试图片
模拟器 iphone5
0 0
- ffmpeg + ios 移植ffmpeg 到ios 平台上
- IOS移植ffmpeg
- FFmpeg 在iOS平台上的配置
- FFmpeg编译ios平台
- 移植 ffmpeg 到 ARM 平台
- ffmpeg移植到android平台
- FFmpeg 移植到 Android 平台
- FFmpeg 移植到 Android 平台
- ffmpeg移植到arm上
- iOS —— FFmpeg移植
- ffmpeg---移植到MAC、IOS系统的步骤总结
- iOS平台编译FFmpeg最新版
- FFmpeg框架在iOS平台上的编译和使用
- FFmpeg框架在iOS平台上的编译和使用
- ffmpeg在android平台上的移植
- OpenCV ffmpeg移植到ARM平台
- 如何将ffmpeg移植到Android平台
- FFmpeg 移植到 Android 平台(2)
- Libnids中tcp流处理"tcp.c"
- math.random()
- xcode更改SVN地址
- HttpClient 教程 (二)
- jtable——数据列的type
- ffmpeg + ios 移植ffmpeg 到ios 平台上
- Java之synchronized加鎖範圍
- HttpClient 教程 (三)
- Java 线程问题
- 理解事务——原子性、一致性、隔离性和持久性
- libnids中主要数据结构和函数“nids.h”
- Oracle查看一个表的生成时间
- QTableWidget 获取item为0/选中错位问题
- contentprovider数据共享机制