用jni实现基于opengl的yuv格式的视频渲染

来源:互联网 发布:绘制结构图软件 编辑:程序博客网 时间:2024/05/07 14:01

由于项目需要,需要在android上面实现视频流的解码显示,综合考虑决定使用ffmpeg解码,opengl渲染视频。

技术选型确定以后,开始写demo,不做不知道,一做才发现网上的东西太不靠谱了,基于jni实现的opengl不是直接渲染yuv格式的数据,都是yuv转rgb以后在显示的,有实现的资料都是在java层做的,我不是java出生,所以对那个不感冒,综合考虑之后决定自己通过jni来实现,由于以前基于webrtc开发了一款产品,用的是webrtc的c++接口开发(现在的webrtc都基于浏览器开发了,更加成熟了,接口也更加简单,^_^我觉得还是挖c++代码出来自己实现接口层有意思,我那个项目就是这样搞的),废话不多说,开始讲述实现步骤。


注意:android2.3.3版本才开始支持opengl。

写jni的时候需要在Android.mk里面加上opengl的库连接,这里我发一个我的Android.mk出来供大家参考一下:

LOCAL_PATH := $(call my-dir)MY_LIBS_PATH := /Users/chenjianjun/Documents/work/ffmpeg-android/build/libMY_INCLUDE_PATH := /Users/chenjianjun/Documents/work/ffmpeg-android/build/includeinclude $(CLEAR_VARS)LOCAL_MODULE := libavcodecLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libavcodec.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE := libavfilterLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libavfilter.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE := libavformatLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libavformat.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE := libavresampleLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libavresample.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE := libavutilLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libavutil.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE := libpostprocLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libpostproc.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE := libswresampleLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libswresample.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE := libswscaleLOCAL_SRC_FILES :=  $(MY_LIBS_PATH)/libswscale.ainclude $(PREBUILT_STATIC_LIBRARY)include $(CLEAR_VARS)LOCAL_MODULE_TAGS := MICloudPubLOCAL_MODULE := libMICloudPubLOCAL_SRC_FILES := H264Decoder.cpp \              #我的H264基于ffmpeg的解码接口代码                   render_opengles20.cpp \        #opengl的渲染代码                   test.cpp                       #测试接口代码LOCAL_CFLAGS :=LOCAL_C_INCLUDES := $(MY_INCLUDE_PATH)LOCAL_CPP_INCLUDES := $(MY_INCLUDE_PATH)LOCAL_LDLIBS := \    -llog \    -lgcc \    <span style="font-size:32px;color:#FF0000;">-lGLESv2 \</span>    -lz    LOCAL_WHOLE_STATIC_LIBRARIES := \    libavcodec \    libavfilter \    libavformat \    libavresample \    libavutil \    libpostproc \    libswresample \    libswscaleinclude $(BUILD_SHARED_LIBRARY) 
上面红色的是opengl的库,我是mac电脑上面编译的,其他系统的不知道是不是叫这个名字哈(不过这么弱智的应该不会变哈).


第一步:

写java代码(主要是为了jni里面的代码回调java的代码实现,其中的妙用大家后面便知)

我把webrtc里面的代码拿出来改动了一下,就没自己去写了(不用重复造轮子)

ViEAndroidGLES20.java

package hzcw.opengl;import java.util.concurrent.locks.ReentrantLock;import javax.microedition.khronos.egl.EGL10;import javax.microedition.khronos.egl.EGLConfig;import javax.microedition.khronos.egl.EGLContext;import javax.microedition.khronos.egl.EGLDisplay;import javax.microedition.khronos.opengles.GL10;import android.app.ActivityManager;import android.content.Context;import android.content.pm.ConfigurationInfo;import android.graphics.PixelFormat;import android.opengl.GLSurfaceView;import android.util.Log;public class ViEAndroidGLES20 extends GLSurfaceView implements GLSurfaceView.Renderer{    private static String TAG = "MICloudPub";    private static final boolean DEBUG = false;    // True if onSurfaceCreated has been called.    private boolean surfaceCreated = false;    private boolean openGLCreated = false;    // True if NativeFunctionsRegistered has been called.    private boolean nativeFunctionsRegisted = false;    private ReentrantLock nativeFunctionLock = new ReentrantLock();    // Address of Native object that will do the drawing.    private long nativeObject = 0;    private int viewWidth = 0;    private int viewHeight = 0;    public static boolean UseOpenGL2(Object renderWindow) {        return ViEAndroidGLES20.class.isInstance(renderWindow);    }    public ViEAndroidGLES20(Context context) {        super(context);        init(false, 0, 0);    }    public ViEAndroidGLES20(Context context, boolean translucent,            int depth, int stencil) {        super(context);        init(translucent, depth, stencil);    }    private void init(boolean translucent, int depth, int stencil) {        // By default, GLSurfaceView() creates a RGB_565 opaque surface.        // If we want a translucent one, we should change the surface's        // format here, using PixelFormat.TRANSLUCENT for GL Surfaces        // is interpreted as any 32-bit surface with alpha by SurfaceFlinger.        if (translucent) {            this.getHolder().setFormat(PixelFormat.TRANSLUCENT);        }        // Setup the context factory for 2.0 rendering.        // See ContextFactory class definition below        setEGLContextFactory(new ContextFactory());        // We need to choose an EGLConfig that matches the format of        // our surface exactly. This is going to be done in our        // custom config chooser. See ConfigChooser class definition        // below.        setEGLConfigChooser( translucent ?                             new ConfigChooser(8, 8, 8, 8, depth, stencil) :                             new ConfigChooser(5, 6, 5, 0, depth, stencil) );        // Set the renderer responsible for frame rendering        this.setRenderer(this);        this.setRenderMode(GLSurfaceView.RENDERMODE_WHEN_DIRTY);    }    private static class ContextFactory implements GLSurfaceView.EGLContextFactory {        private static int EGL_CONTEXT_CLIENT_VERSION = 0x3098;        public EGLContext createContext(EGL10 egl, EGLDisplay display, EGLConfig eglConfig) {            Log.w(TAG, "creating OpenGL ES 2.0 context");            checkEglError("Before eglCreateContext", egl);            int[] attrib_list = {EGL_CONTEXT_CLIENT_VERSION, 2, EGL10.EGL_NONE };            EGLContext context = egl.eglCreateContext(display, eglConfig,                    EGL10.EGL_NO_CONTEXT, attrib_list);            checkEglError("After eglCreateContext", egl);            return context;        }        public void destroyContext(EGL10 egl, EGLDisplay display, EGLContext context) {            egl.eglDestroyContext(display, context);        }    }    private static void checkEglError(String prompt, EGL10 egl) {        int error;        while ((error = egl.eglGetError()) != EGL10.EGL_SUCCESS) {            Log.e(TAG, String.format("%s: EGL error: 0x%x", prompt, error));        }    }    private static class ConfigChooser implements GLSurfaceView.EGLConfigChooser {        public ConfigChooser(int r, int g, int b, int a, int depth, int stencil) {            mRedSize = r;            mGreenSize = g;            mBlueSize = b;            mAlphaSize = a;            mDepthSize = depth;            mStencilSize = stencil;        }        // This EGL config specification is used to specify 2.0 rendering.        // We use a minimum size of 4 bits for red/green/blue, but will        // perform actual matching in chooseConfig() below.        private static int EGL_OPENGL_ES2_BIT = 4;        private static int[] s_configAttribs2 =        {            EGL10.EGL_RED_SIZE, 4,            EGL10.EGL_GREEN_SIZE, 4,            EGL10.EGL_BLUE_SIZE, 4,            EGL10.EGL_RENDERABLE_TYPE, EGL_OPENGL_ES2_BIT,            EGL10.EGL_NONE        };        public EGLConfig chooseConfig(EGL10 egl, EGLDisplay display) {            // Get the number of minimally matching EGL configurations            int[] num_config = new int[1];            egl.eglChooseConfig(display, s_configAttribs2, null, 0, num_config);            int numConfigs = num_config[0];            if (numConfigs <= 0) {                throw new IllegalArgumentException("No configs match configSpec");            }            // Allocate then read the array of minimally matching EGL configs            EGLConfig[] configs = new EGLConfig[numConfigs];            egl.eglChooseConfig(display, s_configAttribs2, configs, numConfigs, num_config);            if (DEBUG) {                printConfigs(egl, display, configs);            }            // Now return the "best" one            return chooseConfig(egl, display, configs);        }        public EGLConfig chooseConfig(EGL10 egl, EGLDisplay display,                EGLConfig[] configs) {            for(EGLConfig config : configs) {                int d = findConfigAttrib(egl, display, config,                        EGL10.EGL_DEPTH_SIZE, 0);                int s = findConfigAttrib(egl, display, config,                        EGL10.EGL_STENCIL_SIZE, 0);                // We need at least mDepthSize and mStencilSize bits                if (d < mDepthSize || s < mStencilSize)                    continue;                // We want an *exact* match for red/green/blue/alpha                int r = findConfigAttrib(egl, display, config,                        EGL10.EGL_RED_SIZE, 0);                int g = findConfigAttrib(egl, display, config,                            EGL10.EGL_GREEN_SIZE, 0);                int b = findConfigAttrib(egl, display, config,                            EGL10.EGL_BLUE_SIZE, 0);                int a = findConfigAttrib(egl, display, config,                        EGL10.EGL_ALPHA_SIZE, 0);                if (r == mRedSize && g == mGreenSize && b == mBlueSize && a == mAlphaSize)                    return config;            }            return null;        }        private int findConfigAttrib(EGL10 egl, EGLDisplay display,                EGLConfig config, int attribute, int defaultValue) {            if (egl.eglGetConfigAttrib(display, config, attribute, mValue)) {                return mValue[0];            }            return defaultValue;        }        private void printConfigs(EGL10 egl, EGLDisplay display,            EGLConfig[] configs) {            int numConfigs = configs.length;            Log.w(TAG, String.format("%d configurations", numConfigs));            for (int i = 0; i < numConfigs; i++) {                Log.w(TAG, String.format("Configuration %d:\n", i));                printConfig(egl, display, configs[i]);            }        }        private void printConfig(EGL10 egl, EGLDisplay display,                EGLConfig config) {            int[] attributes = {                    EGL10.EGL_BUFFER_SIZE,                    EGL10.EGL_ALPHA_SIZE,                    EGL10.EGL_BLUE_SIZE,                    EGL10.EGL_GREEN_SIZE,                    EGL10.EGL_RED_SIZE,                    EGL10.EGL_DEPTH_SIZE,                    EGL10.EGL_STENCIL_SIZE,                    EGL10.EGL_CONFIG_CAVEAT,                    EGL10.EGL_CONFIG_ID,                    EGL10.EGL_LEVEL,                    EGL10.EGL_MAX_PBUFFER_HEIGHT,                    EGL10.EGL_MAX_PBUFFER_PIXELS,                    EGL10.EGL_MAX_PBUFFER_WIDTH,                    EGL10.EGL_NATIVE_RENDERABLE,                    EGL10.EGL_NATIVE_VISUAL_ID,                    EGL10.EGL_NATIVE_VISUAL_TYPE,                    0x3030, // EGL10.EGL_PRESERVED_RESOURCES,                    EGL10.EGL_SAMPLES,                    EGL10.EGL_SAMPLE_BUFFERS,                    EGL10.EGL_SURFACE_TYPE,                    EGL10.EGL_TRANSPARENT_TYPE,                    EGL10.EGL_TRANSPARENT_RED_VALUE,                    EGL10.EGL_TRANSPARENT_GREEN_VALUE,                    EGL10.EGL_TRANSPARENT_BLUE_VALUE,                    0x3039, // EGL10.EGL_BIND_TO_TEXTURE_RGB,                    0x303A, // EGL10.EGL_BIND_TO_TEXTURE_RGBA,                    0x303B, // EGL10.EGL_MIN_SWAP_INTERVAL,                    0x303C, // EGL10.EGL_MAX_SWAP_INTERVAL,                    EGL10.EGL_LUMINANCE_SIZE,                    EGL10.EGL_ALPHA_MASK_SIZE,                    EGL10.EGL_COLOR_BUFFER_TYPE,                    EGL10.EGL_RENDERABLE_TYPE,                    0x3042 // EGL10.EGL_CONFORMANT            };            String[] names = {                    "EGL_BUFFER_SIZE",                    "EGL_ALPHA_SIZE",                    "EGL_BLUE_SIZE",                    "EGL_GREEN_SIZE",                    "EGL_RED_SIZE",                    "EGL_DEPTH_SIZE",                    "EGL_STENCIL_SIZE",                    "EGL_CONFIG_CAVEAT",                    "EGL_CONFIG_ID",                    "EGL_LEVEL",                    "EGL_MAX_PBUFFER_HEIGHT",                    "EGL_MAX_PBUFFER_PIXELS",                    "EGL_MAX_PBUFFER_WIDTH",                    "EGL_NATIVE_RENDERABLE",                    "EGL_NATIVE_VISUAL_ID",                    "EGL_NATIVE_VISUAL_TYPE",                    "EGL_PRESERVED_RESOURCES",                    "EGL_SAMPLES",                    "EGL_SAMPLE_BUFFERS",                    "EGL_SURFACE_TYPE",                    "EGL_TRANSPARENT_TYPE",                    "EGL_TRANSPARENT_RED_VALUE",                    "EGL_TRANSPARENT_GREEN_VALUE",                    "EGL_TRANSPARENT_BLUE_VALUE",                    "EGL_BIND_TO_TEXTURE_RGB",                    "EGL_BIND_TO_TEXTURE_RGBA",                    "EGL_MIN_SWAP_INTERVAL",                    "EGL_MAX_SWAP_INTERVAL",                    "EGL_LUMINANCE_SIZE",                    "EGL_ALPHA_MASK_SIZE",                    "EGL_COLOR_BUFFER_TYPE",                    "EGL_RENDERABLE_TYPE",                    "EGL_CONFORMANT"            };            int[] value = new int[1];            for (int i = 0; i < attributes.length; i++) {                int attribute = attributes[i];                String name = names[i];                if (egl.eglGetConfigAttrib(display, config, attribute, value)) {                    Log.w(TAG, String.format("  %s: %d\n", name, value[0]));                } else {                    // Log.w(TAG, String.format("  %s: failed\n", name));                    while (egl.eglGetError() != EGL10.EGL_SUCCESS);                }            }        }        // Subclasses can adjust these values:        protected int mRedSize;        protected int mGreenSize;        protected int mBlueSize;        protected int mAlphaSize;        protected int mDepthSize;        protected int mStencilSize;        private int[] mValue = new int[1];    }    // IsSupported    // Return true if this device support Open GL ES 2.0 rendering.    public static boolean IsSupported(Context context) {        ActivityManager am =                (ActivityManager) context.getSystemService(Context.ACTIVITY_SERVICE);        ConfigurationInfo info = am.getDeviceConfigurationInfo();        if(info.reqGlEsVersion >= 0x20000) {            // Open GL ES 2.0 is supported.            return true;        }        return false;    }    public void onDrawFrame(GL10 gl) {        nativeFunctionLock.lock();        if(!nativeFunctionsRegisted || !surfaceCreated) {            nativeFunctionLock.unlock();            return;        }        if(!openGLCreated) {            if(0 != CreateOpenGLNative(nativeObject, viewWidth, viewHeight)) {                return; // Failed to create OpenGL            }            openGLCreated = true; // Created OpenGL successfully        }        DrawNative(nativeObject); // Draw the new frame        nativeFunctionLock.unlock();    }    public void onSurfaceChanged(GL10 gl, int width, int height) {        surfaceCreated = true;        viewWidth = width;        viewHeight = height;        nativeFunctionLock.lock();        if(nativeFunctionsRegisted) {            if(CreateOpenGLNative(nativeObject,width,height) == 0)                openGLCreated = true;        }        nativeFunctionLock.unlock();    }    public void onSurfaceCreated(GL10 gl, EGLConfig config) {    }    public void RegisterNativeObject(long nativeObject) {        nativeFunctionLock.lock();        this.nativeObject = nativeObject;        nativeFunctionsRegisted = true;        nativeFunctionLock.unlock();    }    public void DeRegisterNativeObject() {        nativeFunctionLock.lock();        nativeFunctionsRegisted = false;        openGLCreated = false;        this.nativeObject = 0;        nativeFunctionLock.unlock();    }    public void ReDraw() {// jni层解码以后的数据回调,然后由系统调用onDrawFrame显示        if(surfaceCreated) {            // Request the renderer to redraw using the render thread context.            this.requestRender();        }    }    private native int CreateOpenGLNative(long nativeObject, int width, int height);    private native void DrawNative(long nativeObject);}
ViERenderer.java
package hzcw.opengl;import android.content.Context;import android.view.SurfaceView;public class ViERenderer{    public static SurfaceView CreateRenderer(Context context) {        return CreateRenderer(context, false);    }    public static SurfaceView CreateRenderer(Context context,            boolean useOpenGLES2) {        if(useOpenGLES2 == true && ViEAndroidGLES20.IsSupported(context))            return new ViEAndroidGLES20(context);        else            return null;    }}

GL2JNILib.java (native接口代码)

package com.example.filltriangle;public class GL2JNILib {      static {        System.loadLibrary("MICloudPub");  }     /**   */    public static native void init(Object glSurface);  public static native void step(String filepath);}

第二步:写jni代码

com_example_filltriangle_GL2JNILib.h (javah自动生成的)

/* DO NOT EDIT THIS FILE - it is machine generated */#include <jni.h>/* Header for class com_example_filltriangle_GL2JNILib */#ifndef _Included_com_example_filltriangle_GL2JNILib#define _Included_com_example_filltriangle_GL2JNILib#ifdef __cplusplusextern "C" {#endif/* * Class:     com_example_filltriangle_GL2JNILib * Method:    init * Signature: (II)V */JNIEXPORT void JNICALL Java_com_example_filltriangle_GL2JNILib_init  (JNIEnv *, jclass, jobject);/* * Class:     com_example_filltriangle_GL2JNILib * Method:    step * Signature: ()V */JNIEXPORT void JNICALL Java_com_example_filltriangle_GL2JNILib_step    (JNIEnv *, jclass, jstring);#ifdef __cplusplus}#endif#endif

test.cpp

#include <jni.h>#include <stdlib.h>#include <stdio.h>#include "render_opengles20.h"#include "com_example_filltriangle_GL2JNILib.h"#include "H264Decoder.h"class AndroidNativeOpenGl2Channel{public:    AndroidNativeOpenGl2Channel(JavaVM* jvm,                                void* window)    {        _jvm = jvm;        _ptrWindow = window;        _buffer = (uint8_t*)malloc(1024000);    }        ~AndroidNativeOpenGl2Channel()    {        if (_jvm)        {            bool isAttached = false;            JNIEnv* env = NULL;            if (_jvm->GetEnv((void**) &env, JNI_VERSION_1_4) != JNI_OK) {                // try to attach the thread and get the env                // Attach this thread to JVM                jint res = _jvm->AttachCurrentThread(&env, NULL);                                // Get the JNI env for this thread                if ((res < 0) || !env) {                    WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                                 "%s: Could not attach thread to JVM (%d, %p)",                                 __FUNCTION__, res, env);                    env = NULL;                } else {                    isAttached = true;                }            }                        if (env && _deRegisterNativeCID) {                env->CallVoidMethod(_javaRenderObj, _deRegisterNativeCID);            }                        env->DeleteGlobalRef(_javaRenderObj);            env->DeleteGlobalRef(_javaRenderClass);                        if (isAttached) {                if (_jvm->DetachCurrentThread() < 0) {                    WEBRTC_TRACE(kTraceWarning, kTraceVideoRenderer, _id,                                 "%s: Could not detach thread from JVM",                                 __FUNCTION__);                }            }        }                free(_buffer);    }        int32_t Init()    {        if (!_ptrWindow)        {            WEBRTC_TRACE(kTraceWarning, kTraceVideoRenderer, _id,                         "(%s): No window have been provided.", __FUNCTION__);            return -1;        }                if (!_jvm)        {            WEBRTC_TRACE(kTraceWarning, kTraceVideoRenderer, _id,                         "(%s): No JavaVM have been provided.", __FUNCTION__);            return -1;        }                // get the JNI env for this thread        bool isAttached = false;        JNIEnv* env = NULL;        if (_jvm->GetEnv((void**) &env, JNI_VERSION_1_4) != JNI_OK) {            // try to attach the thread and get the env            // Attach this thread to JVM            jint res = _jvm->AttachCurrentThread(&env, NULL);                        // Get the JNI env for this thread            if ((res < 0) || !env) {                WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                             "%s: Could not attach thread to JVM (%d, %p)",                             __FUNCTION__, res, env);                return -1;            }            isAttached = true;        }                // get the ViEAndroidGLES20 class        jclass javaRenderClassLocal = reinterpret_cast<jclass> (env->FindClass("hzcw/opengl/ViEAndroidGLES20"));        if (!javaRenderClassLocal) {            WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                         "%s: could not find ViEAndroidGLES20", __FUNCTION__);            return -1;        }                _javaRenderClass = reinterpret_cast<jclass> (env->NewGlobalRef(javaRenderClassLocal));        if (!_javaRenderClass) {            WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                         "%s: could not create Java SurfaceHolder class reference",                         __FUNCTION__);            return -1;        }        // Delete local class ref, we only use the global ref        env->DeleteLocalRef(javaRenderClassLocal);        jmethodID cidUseOpenGL = env->GetStaticMethodID(_javaRenderClass,                                                        "UseOpenGL2",                                                        "(Ljava/lang/Object;)Z");        if (cidUseOpenGL == NULL) {            WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, -1,                         "%s: could not get UseOpenGL ID", __FUNCTION__);            return false;        }        jboolean res = env->CallStaticBooleanMethod(_javaRenderClass,                                                    cidUseOpenGL, (jobject) _ptrWindow);                // create a reference to the object (to tell JNI that we are referencing it        // after this function has returned)        _javaRenderObj = reinterpret_cast<jobject> (env->NewGlobalRef((jobject)_ptrWindow));        if (!_javaRenderObj)        {            WEBRTC_TRACE(                         kTraceError,                         kTraceVideoRenderer,                         _id,                         "%s: could not create Java SurfaceRender object reference",                         __FUNCTION__);            return -1;        }                // get the method ID for the ReDraw function        _redrawCid = env->GetMethodID(_javaRenderClass, "ReDraw", "()V");        if (_redrawCid == NULL) {            WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                         "%s: could not get ReDraw ID", __FUNCTION__);            return -1;        }                _registerNativeCID = env->GetMethodID(_javaRenderClass,                                              "RegisterNativeObject", "(J)V");        if (_registerNativeCID == NULL) {            WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                         "%s: could not get RegisterNativeObject ID", __FUNCTION__);            return -1;        }                _deRegisterNativeCID = env->GetMethodID(_javaRenderClass,                                                "DeRegisterNativeObject", "()V");        if (_deRegisterNativeCID == NULL) {            WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                         "%s: could not get DeRegisterNativeObject ID",                         __FUNCTION__);            return -1;        }                JNINativeMethod nativeFunctions[2] = {            { "DrawNative",                "(J)V",                (void*) &AndroidNativeOpenGl2Channel::DrawNativeStatic, },            { "CreateOpenGLNative",                "(JII)I",                (void*) &AndroidNativeOpenGl2Channel::CreateOpenGLNativeStatic },        };        if (env->RegisterNatives(_javaRenderClass, nativeFunctions, 2) == 0) {            WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, -1,                         "%s: Registered native functions", __FUNCTION__);        }        else {            WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, -1,                         "%s: Failed to register native functions", __FUNCTION__);            return -1;        }                env->CallVoidMethod(_javaRenderObj, _registerNativeCID, (jlong) this);                if (isAttached) {            if (_jvm->DetachCurrentThread() < 0) {                WEBRTC_TRACE(kTraceWarning, kTraceVideoRenderer, _id,                             "%s: Could not detach thread from JVM", __FUNCTION__);            }        }                WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, "%s done",                     __FUNCTION__);        //        if (_openGLRenderer.SetCoordinates(zOrder, left, top, right, bottom) != 0) {//            return -1;//        }                return 0;    }        void DeliverFrame(int32_t widht, int32_t height)    {        if (_jvm)        {            bool isAttached = false;            JNIEnv* env = NULL;            if (_jvm->GetEnv((void**) &env, JNI_VERSION_1_4) != JNI_OK) {                // try to attach the thread and get the env                // Attach this thread to JVM                jint res = _jvm->AttachCurrentThread(&env, NULL);                                // Get the JNI env for this thread                if ((res < 0) || !env) {                    WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                                 "%s: Could not attach thread to JVM (%d, %p)",                                 __FUNCTION__, res, env);                    env = NULL;                } else {                    isAttached = true;                }            }                        if (env && _redrawCid)            {                _widht = widht;                _height = height;                                env->CallVoidMethod(_javaRenderObj, _redrawCid);            }                        if (isAttached) {                if (_jvm->DetachCurrentThread() < 0) {                    WEBRTC_TRACE(kTraceWarning, kTraceVideoRenderer, _id,                                 "%s: Could not detach thread from JVM",                                 __FUNCTION__);                }            }        }    }        void GetDataBuf(uint8_t*& pbuf, int32_t& isize)    {        pbuf = _buffer;        isize = 1024000;    }        static jint CreateOpenGLNativeStatic(JNIEnv * env,                                        jobject,                                        jlong context,                                        jint width,                                        jint height)    {        AndroidNativeOpenGl2Channel* renderChannel =        reinterpret_cast<AndroidNativeOpenGl2Channel*> (context);                WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "%s:", __FUNCTION__);                return renderChannel->CreateOpenGLNative(width, height);    }        static void DrawNativeStatic(JNIEnv * env,jobject, jlong context)    {        AndroidNativeOpenGl2Channel* renderChannel =        reinterpret_cast<AndroidNativeOpenGl2Channel*>(context);        renderChannel->DrawNative();    }        jint CreateOpenGLNative(int width, int height)    {        return _openGLRenderer.Setup(width, height);    }        void DrawNative()    {        _openGLRenderer.Render(_buffer, _widht, _height);    }    private:    JavaVM*     _jvm;    void* _ptrWindow;        jobject _javaRenderObj;    jclass _javaRenderClass;    JNIEnv* _javaRenderJniEnv;        jmethodID      _redrawCid;    jmethodID      _registerNativeCID;    jmethodID      _deRegisterNativeCID;        RenderOpenGles20 _openGLRenderer;        uint8_t* _buffer;    int32_t _widht;    int32_t _height;};static JavaVM* g_jvm = NULL;static AndroidNativeOpenGl2Channel* p_opengl_channel = NULL;extern "C"{    JNIEXPORT jint JNI_OnLoad(JavaVM* vm, void *reserved)    {        JNIEnv* env = NULL;        jint result = -1;                if (vm->GetEnv((void**) &env, JNI_VERSION_1_4) != JNI_OK)            return -1;                g_jvm = vm;                return JNI_VERSION_1_4;    }}extern "C"{    int mTrans = 0x0F0F0F0F;    int MergeBuffer(uint8_t *NalBuf, int NalBufUsed, uint8_t *SockBuf, int SockBufUsed, int SockRemain)    {        //把读取的数剧分割成NAL块        int i = 0;        char Temp;                for (i = 0; i < SockRemain; i++) {            Temp = SockBuf[i + SockBufUsed];            NalBuf[i + NalBufUsed] = Temp;                        mTrans <<= 8;            mTrans |= Temp;                        if (mTrans == 1) // 找到一个开始字            {                i++;                break;            }        }                return i;    }        JNIEXPORT void JNICALL Java_com_example_filltriangle_GL2JNILib_init    (JNIEnv *env, jclass oclass, jobject glSurface)    {        if (p_opengl_channel)        {            WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "初期化失败[%d].", __LINE__);            return;        }                p_opengl_channel = new AndroidNativeOpenGl2Channel(g_jvm, glSurface);        if (p_opengl_channel->Init() != 0)        {            WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "初期化失败[%d].", __LINE__);            return;        }    }        JNIEXPORT void JNICALL Java_com_example_filltriangle_GL2JNILib_step(JNIEnv* env, jclass tis, jstring filepath)    {        const char *filename = env->GetStringUTFChars(filepath, NULL);                WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "step[%d].", __LINE__);                FILE *_imgFileHandle =  fopen(filename, "rb");        if (_imgFileHandle == NULL)        {            WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "File No Exist[%s][%d].", filename, __LINE__);            return;        }                H264Decoder* pMyH264 = new H264Decoder();        X264_DECODER_H handle = pMyH264->X264Decoder_Init();        if (handle <= 0)        {            WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "X264Decoder_Init Error[%d].", __LINE__);            return;        }                int iTemp = 0;        int nalLen;        int bytesRead = 0;        int NalBufUsed = 0;        int SockBufUsed = 0;                bool bFirst = true;        bool bFindPPS = true;                uint8_t *SockBuf = (uint8_t *)malloc(204800);        uint8_t *NalBuf = (uint8_t *)malloc(4098000);                        int nWidth, nHeight;        memset(SockBuf, 0, 204800);                uint8_t *buffOut = NULL;        int outSize = 0;        p_opengl_channel->GetDataBuf(buffOut, outSize);                        uint8_t *IIBuf = (uint8_t *)malloc(204800);        int IILen = 0;                        do {            bytesRead = fread(SockBuf, 1, 204800, _imgFileHandle);            WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "bytesRead  = %d", bytesRead);            if (bytesRead <= 0) {                break;            }                        SockBufUsed = 0;            while (bytesRead - SockBufUsed > 0) {                nalLen = MergeBuffer(NalBuf, NalBufUsed, SockBuf, SockBufUsed,                                     bytesRead - SockBufUsed);                NalBufUsed += nalLen;                SockBufUsed += nalLen;                                while (mTrans == 1) {                    mTrans = 0xFFFFFFFF;                    if (bFirst == true) // the first start flag                    {                        bFirst = false;                    }                    else // a complete NAL data, include 0x00000001 trail.                    {                        if (bFindPPS == true) // true                        {                            if ((NalBuf[4] & 0x1F) == 7 || (NalBuf[4] & 0x1F) == 8)                            {                                bFindPPS = false;                            }                            else                            {                                NalBuf[0] = 0;                                NalBuf[1] = 0;                                NalBuf[2] = 0;                                NalBuf[3] = 1;                                                                NalBufUsed = 4;                                                                break;                            }                        }                        if (NalBufUsed == 16 || NalBufUsed == 10 || NalBufUsed == 54 || NalBufUsed == 12 || NalBufUsed == 20) {                            memcpy(IIBuf + IILen, NalBuf, NalBufUsed);                            IILen += NalBufUsed;                        }                        else                        {                            memcpy(IIBuf + IILen, NalBuf, NalBufUsed);                            IILen += NalBufUsed;                                                        //decode nal                            iTemp = pMyH264->X264Decoder_Decode(handle, (uint8_t *)IIBuf,                                                                IILen, (uint8_t *)buffOut,                                                                outSize, &nWidth, &nHeight);                            if (iTemp == 0) {                                WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "解码成功,宽度:%d高度:%d,解码数据长度:%d.", nWidth, nHeight, iTemp);//                                [self.glView setVideoSize:nWidth height:nHeight];//                                [self.glView displayYUV420pData:buffOut//                                                          width:nWidth//                                                         height:nHeight];                                p_opengl_channel->DeliverFrame(nWidth, nHeight);                            }                            else                            {                                WEBRTC_TRACE(kTraceInfo, kTraceVideoRenderer, -1, "解码失败.");                            }                                                        IILen = 0;                        }                    }                                        NalBuf[0]=0;                    NalBuf[1]=0;                    NalBuf[2]=0;                    NalBuf[3]=1;                                        NalBufUsed=4;                }            }        }while (bytesRead>0);                        fclose(_imgFileHandle);        pMyH264->X264Decoder_UnInit(handle);                free(SockBuf);        free(NalBuf);        delete pMyH264;                env->ReleaseStringUTFChars(filepath, filename);    }}

render_opengles20.cpp

#include <GLES2/gl2.h>#include <GLES2/gl2ext.h>#include <stdio.h>#include <stdlib.h>#include <stdio.h>#include "render_opengles20.h"const char RenderOpenGles20::g_indices[] = { 0, 3, 2, 0, 2, 1 };const char RenderOpenGles20::g_vertextShader[] = {    "attribute vec4 aPosition;\n"    "attribute vec2 aTextureCoord;\n"    "varying vec2 vTextureCoord;\n"    "void main() {\n"    "  gl_Position = aPosition;\n"    "  vTextureCoord = aTextureCoord;\n"    "}\n" };// The fragment shader.// Do YUV to RGB565 conversion.const char RenderOpenGles20::g_fragmentShader[] = {    "precision mediump float;\n"    "uniform sampler2D Ytex;\n"    "uniform sampler2D Utex,Vtex;\n"    "varying vec2 vTextureCoord;\n"    "void main(void) {\n"    "  float nx,ny,r,g,b,y,u,v;\n"    "  mediump vec4 txl,ux,vx;"    "  nx=vTextureCoord[0];\n"    "  ny=vTextureCoord[1];\n"    "  y=texture2D(Ytex,vec2(nx,ny)).r;\n"    "  u=texture2D(Utex,vec2(nx,ny)).r;\n"    "  v=texture2D(Vtex,vec2(nx,ny)).r;\n"        //"  y = v;\n"+    "  y=1.1643*(y-0.0625);\n"    "  u=u-0.5;\n"    "  v=v-0.5;\n"        "  r=y+1.5958*v;\n"    "  g=y-0.39173*u-0.81290*v;\n"    "  b=y+2.017*u;\n"    "  gl_FragColor=vec4(r,g,b,1.0);\n"    "}\n" };RenderOpenGles20::RenderOpenGles20() :_id(0),_textureWidth(-1),_textureHeight(-1){    WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, "%s: id %d",                 __FUNCTION__, (int) _id);        const GLfloat vertices[20] = {        // X, Y, Z, U, V        -1, -1, 0, 1, 0, // Bottom Left        1, -1, 0, 0, 0, //Bottom Right        1, 1, 0, 0, 1, //Top Right        -1, 1, 0, 1, 1 }; //Top Left        memcpy(_vertices, vertices, sizeof(_vertices));}RenderOpenGles20::~RenderOpenGles20() {    glDeleteTextures(3, _textureIds);}int32_t RenderOpenGles20::Setup(int32_t width, int32_t height) {    WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id,                 "%s: width %d, height %d", __FUNCTION__, (int) width,                 (int) height);        printGLString("Version", GL_VERSION);    printGLString("Vendor", GL_VENDOR);    printGLString("Renderer", GL_RENDERER);    printGLString("Extensions", GL_EXTENSIONS);        int maxTextureImageUnits[2];    int maxTextureSize[2];    glGetIntegerv(GL_MAX_TEXTURE_IMAGE_UNITS, maxTextureImageUnits);    glGetIntegerv(GL_MAX_TEXTURE_SIZE, maxTextureSize);        WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id,                 "%s: number of textures %d, size %d", __FUNCTION__,                 (int) maxTextureImageUnits[0], (int) maxTextureSize[0]);        _program = createProgram(g_vertextShader, g_fragmentShader);    if (!_program) {        WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                     "%s: Could not create program", __FUNCTION__);        return -1;    }        int positionHandle = glGetAttribLocation(_program, "aPosition");    checkGlError("glGetAttribLocation aPosition");    if (positionHandle == -1) {        WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                     "%s: Could not get aPosition handle", __FUNCTION__);        return -1;    }        int textureHandle = glGetAttribLocation(_program, "aTextureCoord");    checkGlError("glGetAttribLocation aTextureCoord");    if (textureHandle == -1) {        WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                     "%s: Could not get aTextureCoord handle", __FUNCTION__);        return -1;    }        // set the vertices array in the shader    // _vertices contains 4 vertices with 5 coordinates.    // 3 for (xyz) for the vertices and 2 for the texture    glVertexAttribPointer(positionHandle, 3, GL_FLOAT, false,                          5 * sizeof(GLfloat), _vertices);    checkGlError("glVertexAttribPointer aPosition");        glEnableVertexAttribArray(positionHandle);    checkGlError("glEnableVertexAttribArray positionHandle");        // set the texture coordinate array in the shader    // _vertices contains 4 vertices with 5 coordinates.    // 3 for (xyz) for the vertices and 2 for the texture    glVertexAttribPointer(textureHandle, 2, GL_FLOAT, false, 5                          * sizeof(GLfloat), &_vertices[3]);    checkGlError("glVertexAttribPointer maTextureHandle");    glEnableVertexAttribArray(textureHandle);    checkGlError("glEnableVertexAttribArray textureHandle");        glUseProgram(_program);    int i = glGetUniformLocation(_program, "Ytex");    checkGlError("glGetUniformLocation");    glUniform1i(i, 0); /* Bind Ytex to texture unit 0 */    checkGlError("glUniform1i Ytex");        i = glGetUniformLocation(_program, "Utex");    checkGlError("glGetUniformLocation Utex");    glUniform1i(i, 1); /* Bind Utex to texture unit 1 */    checkGlError("glUniform1i Utex");        i = glGetUniformLocation(_program, "Vtex");    checkGlError("glGetUniformLocation");    glUniform1i(i, 2); /* Bind Vtex to texture unit 2 */    checkGlError("glUniform1i");        glViewport(0, 0, width, height);    checkGlError("glViewport");    return 0;}// SetCoordinates// Sets the coordinates where the stream shall be rendered.// Values must be between 0 and 1.int32_t RenderOpenGles20::SetCoordinates(int32_t zOrder,                                         const float left,                                         const float top,                                         const float right,                                         const float bottom) {    if ((top > 1 || top < 0) || (right > 1 || right < 0) ||        (bottom > 1 || bottom < 0) || (left > 1 || left < 0)) {        WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                     "%s: Wrong coordinates", __FUNCTION__);        return -1;    }        //  X, Y, Z, U, V    // -1, -1, 0, 0, 1, // Bottom Left    //  1, -1, 0, 1, 1, //Bottom Right    //  1,  1, 0, 1, 0, //Top Right    // -1,  1, 0, 0, 0  //Top Left        // Bottom Left    _vertices[0] = (left * 2) - 1;    _vertices[1] = -1 * (2 * bottom) + 1;    _vertices[2] = zOrder;        //Bottom Right    _vertices[5] = (right * 2) - 1;    _vertices[6] = -1 * (2 * bottom) + 1;    _vertices[7] = zOrder;        //Top Right    _vertices[10] = (right * 2) - 1;    _vertices[11] = -1 * (2 * top) + 1;    _vertices[12] = zOrder;        //Top Left    _vertices[15] = (left * 2) - 1;    _vertices[16] = -1 * (2 * top) + 1;    _vertices[17] = zOrder;        return 0;}GLuint RenderOpenGles20::loadShader(GLenum shaderType, const char* pSource){    GLuint shader = glCreateShader(shaderType);    if (shader) {        glShaderSource(shader, 1, &pSource, NULL);        glCompileShader(shader);        GLint compiled = 0;        glGetShaderiv(shader, GL_COMPILE_STATUS, &compiled);        if (!compiled) {            GLint infoLen = 0;            glGetShaderiv(shader, GL_INFO_LOG_LENGTH, &infoLen);            if (infoLen) {                char* buf = (char*) malloc(infoLen);                if (buf) {                    glGetShaderInfoLog(shader, infoLen, NULL, buf);                    WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                                 "%s: Could not compile shader %d: %s",                                 __FUNCTION__, shaderType, buf);                    free(buf);                }                glDeleteShader(shader);                shader = 0;            }        }    }    return shader;}GLuint RenderOpenGles20::createProgram(const char* pVertexSource,                                       const char* pFragmentSource) {    GLuint vertexShader = loadShader(GL_VERTEX_SHADER, pVertexSource);    if (!vertexShader) {        return 0;    }        GLuint pixelShader = loadShader(GL_FRAGMENT_SHADER, pFragmentSource);    if (!pixelShader) {        return 0;    }        GLuint program = glCreateProgram();    if (program) {        glAttachShader(program, vertexShader);        checkGlError("glAttachShader");        glAttachShader(program, pixelShader);        checkGlError("glAttachShader");        glLinkProgram(program);        GLint linkStatus = GL_FALSE;        glGetProgramiv(program, GL_LINK_STATUS, &linkStatus);        if (linkStatus != GL_TRUE) {            GLint bufLength = 0;            glGetProgramiv(program, GL_INFO_LOG_LENGTH, &bufLength);            if (bufLength) {                char* buf = (char*) malloc(bufLength);                if (buf) {                    glGetProgramInfoLog(program, bufLength, NULL, buf);                    WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                                 "%s: Could not link program: %s",                                 __FUNCTION__, buf);                    free(buf);                }            }            glDeleteProgram(program);            program = 0;        }    }    return program;}void RenderOpenGles20::printGLString(const char *name, GLenum s) {    const char *v = (const char *) glGetString(s);    WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, "GL %s = %s\n",                 name, v);}void RenderOpenGles20::checkGlError(const char* op) {#ifdef ANDROID_LOG    for (GLint error = glGetError(); error; error = glGetError()) {        WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id,                     "after %s() glError (0x%x)\n", op, error);    }#else    return;#endif}static void InitializeTexture(int name, int id, int width, int height) {    glActiveTexture(name);    glBindTexture(GL_TEXTURE_2D, id);    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST);    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR);    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);    glTexImage2D(GL_TEXTURE_2D, 0, GL_LUMINANCE, width, height, 0,                 GL_LUMINANCE, GL_UNSIGNED_BYTE, NULL);}// Uploads a plane of pixel data, accounting for stride != width*bpp.static void GlTexSubImage2D(GLsizei width, GLsizei height, int stride,                            const uint8_t* plane) {    if (stride == width) {        // Yay!  We can upload the entire plane in a single GL call.        glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, width, height, GL_LUMINANCE,                        GL_UNSIGNED_BYTE,                        static_cast<const GLvoid*>(plane));    } else {        // Boo!  Since GLES2 doesn't have GL_UNPACK_ROW_LENGTH and Android doesn't        // have GL_EXT_unpack_subimage we have to upload a row at a time.  Ick.        for (int row = 0; row < height; ++row) {            glTexSubImage2D(GL_TEXTURE_2D, 0, 0, row, width, 1, GL_LUMINANCE,                            GL_UNSIGNED_BYTE,                            static_cast<const GLvoid*>(plane + (row * stride)));        }    }}int32_t RenderOpenGles20::Render(void * data, int32_t widht, int32_t height){    WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, "%s: id %d",                 __FUNCTION__, (int) _id);        glUseProgram(_program);    checkGlError("glUseProgram");        if (_textureWidth != (GLsizei) widht || _textureHeight != (GLsizei) height) {        SetupTextures(widht, height);    }    UpdateTextures(data, widht, height);        glDrawElements(GL_TRIANGLES, 6, GL_UNSIGNED_BYTE, g_indices);    checkGlError("glDrawArrays");        return 0;}void RenderOpenGles20::SetupTextures(int32_t width, int32_t height){    glDeleteTextures(3, _textureIds);    glGenTextures(3, _textureIds); //Generate  the Y, U and V texture    InitializeTexture(GL_TEXTURE0, _textureIds[0], width, height);    InitializeTexture(GL_TEXTURE1, _textureIds[1], width / 2, height / 2);    InitializeTexture(GL_TEXTURE2, _textureIds[2], width / 2, height / 2);        checkGlError("SetupTextures");        _textureWidth = width;    _textureHeight = height;}void RenderOpenGles20::UpdateTextures(void* data, int32_t widht, int32_t height){    glActiveTexture(GL_TEXTURE0);    glBindTexture(GL_TEXTURE_2D, _textureIds[0]);    glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, widht, height, GL_LUMINANCE, GL_UNSIGNED_BYTE,                    data);        glActiveTexture(GL_TEXTURE1);    glBindTexture(GL_TEXTURE_2D, _textureIds[1]);    glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, widht / 2, height / 2, GL_LUMINANCE,                    GL_UNSIGNED_BYTE, (char *)data + widht * height);        glActiveTexture(GL_TEXTURE2);    glBindTexture(GL_TEXTURE_2D, _textureIds[2]);    glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, widht / 2, height / 2, GL_LUMINANCE,                    GL_UNSIGNED_BYTE, (char *)data + widht * height * 5 / 4);        checkGlError("UpdateTextures");}

H264Decoder.cpp (解码代码,前面的博客贴过代码,这里就不贴了)


第三步:编译jni,生成so文件

第四步:把生成的so文件拷贝到android工程里面去,这里贴一下我的Activity代码,如下:

package com.example.filltriangle;import java.io.IOException;import java.io.InputStream;import hzcw.opengl.ViERenderer;import android.app.Activity;  import android.os.Bundle;   import android.os.Environment;import android.util.Log;import android.view.SurfaceView;  public class FillTriangle extends Activity {    private SurfaceView mView = null;static {        System.loadLibrary("MICloudPub");}      @Override protected void onCreate(Bundle icicle) {          super.onCreate(icicle);          mView = ViERenderer.CreateRenderer(this, true);if (mView == null) {Log.i("test", "mView is null");}        setContentView(mView);                GL2JNILib.init(mView);                new MyThread().start();    }                public class MyThread extends Thread {          public void run() {        GL2JNILib.step("/sdcard/test.264");        }      }}

这个demo就是读一个视频文件,解码以后在界面显示出来。便于运行,最后上效果图哈,免得有人怀疑项目真实性。


0 0