读取笔记本的摄像头的原始yuv数据,通过libav(ffmpeg编码)
来源:互联网 发布:淘宝客推广有什么要求 编辑:程序博客网 时间:2024/06/05 02:50
一、程序的逻辑主要分两部分:
1、通过video4linux2读取摄像头的V4L2_PIX_FMT_YUYV格式的原始数据
2、把V4L2_PIX_FMT_YUYV格式的数据转换成AV_PIX_FMT_YUV422P格式的yuv数据,并存放在AVFrame结构中; 把AVFrame结构送到编码器;
收取编码后的h264数据流,并存到文件中
二、代码中主要用到的库:
1、通过一个叫uvccapture-0.5库(我直接改了里面的代码),获取摄像头的V4L2_PIX_FMT_YUYV格式数据。
2、通过libav编码
三、主要代码:
/* * main.cpp * * Created on: 2014-3-2 * Author: xy */#include <unistd.h>#include <stdio.h>#include <stdlib.h>#include <time.h>#include <linux/videodev2.h>//ffplay -f video4linux2 -framerate 30 -video_size hd720 /dev/video0#ifdef __cplusplusextern "C" {#endif#define __STDC_CONSTANT_MACROS#ifdef _STDINT_H#undef _STDINT_H#endif# include <stdint.h>#include "libavformat/avformat.h"#include "libavdevice/avdevice.h"#include <libavutil/imgutils.h>#include <libavutil/opt.h>#ifdef __cplusplus}#endif//输入设备#include "v4l2uvc.h"int main() {//v4l2 valchar *videodevice = "/dev/video0";int width = 640; //320;int height = 480; //240;int brightness = 0, contrast = 0, saturation = 0, gain = 0;int quality = 95;int format = V4L2_PIX_FMT_YUYV;struct vdIn *videoIn;int grabmethod = 1;//video encoder initavcodec_register_all();AVCodec *codec;AVCodecContext *c = NULL;int i, ret, x, y, got_output;FILE *f;AVFrame *frame;AVPacket pkt;uint8_t endcode[] = { 0, 0, 1, 0xb7 };char filename[] = "test.264";printf("Encode video file %s\n", filename);/* find the mpeg1 video encoder */codec = avcodec_find_encoder(AV_CODEC_ID_H264);if (!codec) {fprintf(stderr, "Codec not found\n");exit(1);}c = avcodec_alloc_context3(codec);if (!c) {fprintf(stderr, "Could not allocate video codec context\n");exit(1);}/* put sample parameters */c->bit_rate = 400000;/* resolution must be a multiple of two */c->width = width; // 352;c->height = height; // 288;/* frames per second */c->time_base = (AVRational) {1,10/*25*/};c->gop_size = 10; /* emit one intra frame every ten frames */c->max_b_frames = 0; //1c->pix_fmt = AV_PIX_FMT_YUV422P; //v4l2是这个格式 AV_PIX_FMT_YUV420P;//av_opt_set(c->priv_data, "preset", "slow", 0);av_opt_set(c->priv_data, "tune", "zerolatency", 0); //这个可以让libav不缓存视频帧/******************************************************************************** * 有两个地方影响libav是不是缓存编码后的视频帧,也就是影响实时性: * 1、av_opt_set(c->priv_data, "tune", "zerolatency", 0);这个比较主要。 * 2、参数中有c->max_b_frames = 1;如果这个帧设为0,就没有B帧了,编码会很快的。 ********************************************************************************//* open it */if (avcodec_open2(c, codec, NULL) < 0) {fprintf(stderr, "Could not open codec\n");exit(1);}f = fopen(filename, "wb");if (!f) {fprintf(stderr, "Could not open %s\n", filename);exit(1);}frame = av_frame_alloc();if (!frame) {fprintf(stderr, "Could not allocate video frame\n");exit(1);}frame->format = c->pix_fmt;frame->width = c->width;frame->height = c->height;/* the image can be allocated by any means and av_image_alloc() is * just the most convenient way if av_malloc() is to be used */ret = av_image_alloc(frame->data, frame->linesize, c->width, c->height,c->pix_fmt, 32);if (ret < 0) {fprintf(stderr, "Could not allocate raw picture buffer\n");exit(1);}//v4l2 initvideoIn = (struct vdIn *) calloc(1, sizeof(struct vdIn));if (init_videoIn(videoIn, (char *) videodevice, width, height, format,grabmethod) < 0)exit(1);printf("w:%d,h:%d\n", c->width, c->height);time_t timep;timep = time(NULL);printf("%s\n", asctime(gmtime(&timep)));for (i = 0; i < 100; i++) {//usleep(200000);//从v4l2中获取数据格式为AV_PIX_FMT_YUV422Pif (uvcGrab(videoIn) < 0) {fprintf(stderr, "Error grabbing\n");close_v4l2(videoIn);free(videoIn);exit(1);}unsigned char *yuyv = videoIn->framebuffer;//把数据复制到libav想要的结构中av_init_packet(&pkt);pkt.data = NULL; // packet data will be allocated by the encoderpkt.size = 0;printf("debug!!!");fflush(stdout);#if 0/* prepare a dummy image *//* Y */for (y = 0; y < c->height; y++) {for (x = 0; x < c->width; x++) {frame->data[0][y * frame->linesize[0] + x] = x + y + i * 3;}}/* Cb and Cr */for (y = 0; y < c->height/2; y++) {for (x = 0; x < c->width/2; x++) {frame->data[1][y * frame->linesize[1] + x] = 128 + y + i * 2;frame->data[2][y * frame->linesize[2] + x] = 64 + x + i * 5;}}#else/* prepare image *//* Y Cb Rb */for (y = 0; y < c->height; y++) {for (x = 0; x < c->width; x++) {frame->data[0][y * frame->linesize[0] + x] = yuyv[y* frame->linesize[0] * 2 + 2 * x];}}/* Cb and Cr */for (y = 0; y < c->height; y++) {for (x = 0; x < c->width / 2; x++) {//frame->data[0][y * frame->linesize[0] + 2*x ] = yuyv[y*frame->linesize[0]*4+4*x];//frame->data[0][y * frame->linesize[0] + 2*x+1] = yuyv[y*frame->linesize[0]*4+4*x+2];frame->data[1][y * frame->linesize[1] + x] = yuyv[y* frame->linesize[1] * 4 + 4 * x + 1];frame->data[2][y * frame->linesize[2] + x] = yuyv[y* frame->linesize[2] * 4 + 4 * x + 3];}}#endifframe->pts = i;/* encode the image */ret = avcodec_encode_video2(c, &pkt, frame, &got_output);if (ret < 0) {fprintf(stderr, "Error encoding frame\n");exit(1);}if (got_output) {printf("Write frame %3d (size=%5d)\n", i, pkt.size);fwrite(pkt.data, 1, pkt.size, f);av_free_packet(&pkt);}//编码}/* get the delayed frames */for (got_output = 1; got_output; i++) {fflush(stdout);ret = avcodec_encode_video2(c, &pkt, NULL, &got_output);if (ret < 0) {fprintf(stderr, "Error encoding frame\n");exit(1);}if (got_output) {printf("Write frame %3d (size=%5d)\n", i, pkt.size);fwrite(pkt.data, 1, pkt.size, f);av_free_packet(&pkt);}}timep = time(NULL);printf("%s\n", asctime(gmtime(&timep)));/* add sequence end code to have a real mpeg file */fwrite(endcode, 1, sizeof(endcode), f);fclose(f);avcodec_close(c);av_free(c);av_freep(&frame->data[0]);av_frame_free(&frame);printf("111\n");free(videoIn);}
想运行代码的朋友可以通过这个文章:
http://blog.csdn.net/xyyangkun/article/details/20456725
编译用到的libav库
在通过改工程中的makefile:
LIBAVDIR=/home/xy/mywork/av/libav-2014-03-02
指定libav库的位置就可以了。
使用的读摄像头数据的代码出处:
http://staticwave.ca/source/uvccapture/uvccapture-0.5.tar.bz2
库的工程在代码在:
http://download.csdn.net/detail/xyyangkun/6990791
所在工程已上传到github:
https://github.com/xyyangkun/read_encoder_sender
提交版本:2e60986a438e731aa53ca7d54bc492c521e7b5bc
0 0
- 读取笔记本的摄像头的原始yuv数据,通过libav(ffmpeg编码)
- 读取笔记本的摄像头的原始yuv数据,通过libav(ffmpeg编码)
- 在Android上使用FFmpeg将摄像头采集的YUV裸流编码为h264。
- 利用ffmpeg进行摄像头提取视频编码为h264通过RTP发送数据到指定的rtp地址
- 利用ffmpeg进行摄像头提取视频编码为h264通过RTP发送数据到指定的rtp地址
- 【libav】libav的常用方法
- FFmpeg读取Mac摄像头的问题
- 摄像头的YUV
- 摄像头的YUV
- ffmpeg转换和显示yuv,rgb等原始图像数据
- 解决ffmpeg/libav 编译后ffplay没有生成的问题
- 04 ffmpeg摄像头数据H264编码
- ffmpeg最简单的解码保存YUV数据
- ffmpeg最简单的解码保存YUV数据
- ffmpeg最简单的解码保存YUV数据
- yuv文件的读取
- FFmpeg / libav GPU-accelerated video processing 编码/解码/转码
- FFMpeg读取摄像头数据流
- 从子界面获取数据返回给主界面
- 老字号,没那么容易倒——wifi共享精灵
- U盘文件被损坏如何才能恢复
- 用mysql数据库写的分词算法代码
- bool 和BOOL的区别
- 读取笔记本的摄像头的原始yuv数据,通过libav(ffmpeg编码)
- Git 过滤文件,控制上传
- 不常用到的sql语句记录
- yii上传图片、yii上传文件、yii控件activeFileField使用
- vi
- 对象左列表增加非对象关联表的字段,添加两个以上字段
- plsql developer中 保存/删除 登录的用户名和密码
- 不定期搬运二次元福利
- hibernate查询