public final class

VideoDecoderGLSurfaceView

extends GLSurfaceView

implements VideoDecoderOutputBufferRenderer

 java.lang.Object

↳GLSurfaceView

↳androidx.media3.exoplayer.video.VideoDecoderGLSurfaceView

Gradle dependencies

compile group: 'androidx.media3', name: 'media3-exoplayer', version: '1.5.0-alpha01'

  • groupId: androidx.media3
  • artifactId: media3-exoplayer
  • version: 1.5.0-alpha01

Artifact androidx.media3:media3-exoplayer:1.5.0-alpha01 it located at Google repository (https://maven.google.com/)

Overview

GLSurfaceView implementing VideoDecoderOutputBufferRenderer for rendering VideoDecoderOutputBuffers.

This view is intended for use only with decoders that produce VideoDecoderOutputBuffers. For other use cases a or should be used instead.

Summary

Constructors
publicVideoDecoderGLSurfaceView(Context context)

publicVideoDecoderGLSurfaceView(Context context, AttributeSet attrs)

Methods
public VideoDecoderOutputBufferRenderergetVideoDecoderOutputBufferRenderer()

public voidsetOutputBuffer(VideoDecoderOutputBuffer outputBuffer)

from java.lang.Objectclone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait

Constructors

public VideoDecoderGLSurfaceView(Context context)

Parameters:

context: A .

public VideoDecoderGLSurfaceView(Context context, AttributeSet attrs)

Parameters:

context: A .
attrs: Custom attributes.

Methods

public void setOutputBuffer(VideoDecoderOutputBuffer outputBuffer)

public VideoDecoderOutputBufferRenderer getVideoDecoderOutputBufferRenderer()

Deprecated: This class implements VideoDecoderOutputBufferRenderer directly.

Source

/*
 * Copyright (C) 2016 The Android Open Source Project
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *      http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package androidx.media3.exoplayer.video;

import static androidx.media3.common.util.Assertions.checkNotNull;

import android.content.Context;
import android.opengl.GLES20;
import android.opengl.GLSurfaceView;
import android.util.AttributeSet;
import android.util.Log;
import androidx.annotation.Nullable;
import androidx.media3.common.util.Assertions;
import androidx.media3.common.util.GlProgram;
import androidx.media3.common.util.GlUtil;
import androidx.media3.common.util.NullableType;
import androidx.media3.common.util.UnstableApi;
import androidx.media3.decoder.VideoDecoderOutputBuffer;
import java.nio.ByteBuffer;
import java.nio.FloatBuffer;
import java.util.concurrent.atomic.AtomicReference;
import javax.microedition.khronos.egl.EGLConfig;
import javax.microedition.khronos.opengles.GL10;
import org.checkerframework.checker.nullness.qual.MonotonicNonNull;
import org.checkerframework.checker.nullness.qual.RequiresNonNull;

/**
 * GLSurfaceView implementing {@link VideoDecoderOutputBufferRenderer} for rendering {@link
 * VideoDecoderOutputBuffer VideoDecoderOutputBuffers}.
 *
 * <p>This view is intended for use only with decoders that produce {@link VideoDecoderOutputBuffer
 * VideoDecoderOutputBuffers}. For other use cases a {@link android.view.SurfaceView} or {@link
 * android.view.TextureView} should be used instead.
 */
@UnstableApi
public final class VideoDecoderGLSurfaceView extends GLSurfaceView
    implements VideoDecoderOutputBufferRenderer {

  private static final String TAG = "VideoDecoderGLSV";

  private final Renderer renderer;

  /**
   * @param context A {@link Context}.
   */
  public VideoDecoderGLSurfaceView(Context context) {
    this(context, /* attrs= */ null);
  }

  /**
   * @param context A {@link Context}.
   * @param attrs Custom attributes.
   */
  @SuppressWarnings({"nullness:assignment", "nullness:argument", "nullness:method.invocation"})
  public VideoDecoderGLSurfaceView(Context context, @Nullable AttributeSet attrs) {
    super(context, attrs);
    renderer = new Renderer(/* surfaceView= */ this);
    setPreserveEGLContextOnPause(true);
    setEGLContextClientVersion(2);
    setRenderer(renderer);
    setRenderMode(GLSurfaceView.RENDERMODE_WHEN_DIRTY);
  }

  @Override
  public void setOutputBuffer(VideoDecoderOutputBuffer outputBuffer) {
    renderer.setOutputBuffer(outputBuffer);
  }

  /**
   * @deprecated This class implements {@link VideoDecoderOutputBufferRenderer} directly.
   */
  @Deprecated
  public VideoDecoderOutputBufferRenderer getVideoDecoderOutputBufferRenderer() {
    return this;
  }

  private static final class Renderer implements GLSurfaceView.Renderer {

    private static final float[] kColorConversion601 = {
      1.164f, 1.164f, 1.164f,
      0.0f, -0.392f, 2.017f,
      1.596f, -0.813f, 0.0f,
    };

    private static final float[] kColorConversion709 = {
      1.164f, 1.164f, 1.164f,
      0.0f, -0.213f, 2.112f,
      1.793f, -0.533f, 0.0f,
    };

    private static final float[] kColorConversion2020 = {
      1.168f, 1.168f, 1.168f,
      0.0f, -0.188f, 2.148f,
      1.683f, -0.652f, 0.0f,
    };

    private static final String VERTEX_SHADER =
        "varying vec2 interp_tc_y;\n"
            + "varying vec2 interp_tc_u;\n"
            + "varying vec2 interp_tc_v;\n"
            + "attribute vec4 in_pos;\n"
            + "attribute vec2 in_tc_y;\n"
            + "attribute vec2 in_tc_u;\n"
            + "attribute vec2 in_tc_v;\n"
            + "void main() {\n"
            + "  gl_Position = in_pos;\n"
            + "  interp_tc_y = in_tc_y;\n"
            + "  interp_tc_u = in_tc_u;\n"
            + "  interp_tc_v = in_tc_v;\n"
            + "}\n";
    private static final String[] TEXTURE_UNIFORMS = {"y_tex", "u_tex", "v_tex"};
    private static final String FRAGMENT_SHADER =
        "precision mediump float;\n"
            + "varying vec2 interp_tc_y;\n"
            + "varying vec2 interp_tc_u;\n"
            + "varying vec2 interp_tc_v;\n"
            + "uniform sampler2D y_tex;\n"
            + "uniform sampler2D u_tex;\n"
            + "uniform sampler2D v_tex;\n"
            + "uniform mat3 mColorConversion;\n"
            + "void main() {\n"
            + "  vec3 yuv;\n"
            + "  yuv.x = texture2D(y_tex, interp_tc_y).r - 0.0625;\n"
            + "  yuv.y = texture2D(u_tex, interp_tc_u).r - 0.5;\n"
            + "  yuv.z = texture2D(v_tex, interp_tc_v).r - 0.5;\n"
            + "  gl_FragColor = vec4(mColorConversion * yuv, 1.0);\n"
            + "}\n";

    private static final FloatBuffer TEXTURE_VERTICES =
        GlUtil.createBuffer(new float[] {-1.0f, 1.0f, -1.0f, -1.0f, 1.0f, 1.0f, 1.0f, -1.0f});

    private final GLSurfaceView surfaceView;
    private final int[] yuvTextures;
    private final int[] texLocations;
    private final int[] previousWidths;
    private final int[] previousStrides;
    private final AtomicReference<@NullableType VideoDecoderOutputBuffer>
        pendingOutputBufferReference;

    // Kept in field rather than a local variable in order not to get garbage collected before
    // glDrawArrays uses it.
    private final FloatBuffer[] textureCoords;

    private @MonotonicNonNull GlProgram program;
    private int colorMatrixLocation;

    // Accessed only from the GL thread.
    private @MonotonicNonNull VideoDecoderOutputBuffer renderedOutputBuffer;

    public Renderer(GLSurfaceView surfaceView) {
      this.surfaceView = surfaceView;
      yuvTextures = new int[3];
      texLocations = new int[3];
      previousWidths = new int[3];
      previousStrides = new int[3];
      pendingOutputBufferReference = new AtomicReference<>();
      textureCoords = new FloatBuffer[3];
      for (int i = 0; i < 3; i++) {
        previousWidths[i] = previousStrides[i] = -1;
      }
    }

    @Override
    public void onSurfaceCreated(GL10 unused, EGLConfig config) {
      try {
        program = new GlProgram(VERTEX_SHADER, FRAGMENT_SHADER);
        int posLocation = program.getAttributeArrayLocationAndEnable("in_pos");
        GLES20.glVertexAttribPointer(
            posLocation,
            2,
            GLES20.GL_FLOAT,
            /* normalized= */ false,
            /* stride= */ 0,
            TEXTURE_VERTICES);
        texLocations[0] = program.getAttributeArrayLocationAndEnable("in_tc_y");
        texLocations[1] = program.getAttributeArrayLocationAndEnable("in_tc_u");
        texLocations[2] = program.getAttributeArrayLocationAndEnable("in_tc_v");
        colorMatrixLocation = program.getUniformLocation("mColorConversion");
        GlUtil.checkGlError();
        setupTextures();
        GlUtil.checkGlError();
      } catch (GlUtil.GlException e) {
        Log.e(TAG, "Failed to set up the textures and program", e);
      }
    }

    @Override
    public void onSurfaceChanged(GL10 unused, int width, int height) {
      GLES20.glViewport(0, 0, width, height);
    }

    @Override
    public void onDrawFrame(GL10 unused) {
      @Nullable
      VideoDecoderOutputBuffer pendingOutputBuffer =
          pendingOutputBufferReference.getAndSet(/* newValue= */ null);
      if (pendingOutputBuffer == null && renderedOutputBuffer == null) {
        // There is no output buffer to render at the moment.
        return;
      }
      if (pendingOutputBuffer != null) {
        if (renderedOutputBuffer != null) {
          renderedOutputBuffer.release();
        }
        renderedOutputBuffer = pendingOutputBuffer;
      }

      VideoDecoderOutputBuffer outputBuffer = checkNotNull(renderedOutputBuffer);

      // Set color matrix. Assume BT709 if the color space is unknown.
      float[] colorConversion = kColorConversion709;
      switch (outputBuffer.colorspace) {
        case VideoDecoderOutputBuffer.COLORSPACE_BT601:
          colorConversion = kColorConversion601;
          break;
        case VideoDecoderOutputBuffer.COLORSPACE_BT2020:
          colorConversion = kColorConversion2020;
          break;
        case VideoDecoderOutputBuffer.COLORSPACE_BT709:
        default:
          // Do nothing.
          break;
      }
      GLES20.glUniformMatrix3fv(
          colorMatrixLocation,
          /* color= */ 1,
          /* transpose= */ false,
          colorConversion,
          /* offset= */ 0);

      int[] yuvStrides = checkNotNull(outputBuffer.yuvStrides);
      ByteBuffer[] yuvPlanes = checkNotNull(outputBuffer.yuvPlanes);

      for (int i = 0; i < 3; i++) {
        int h = (i == 0) ? outputBuffer.height : (outputBuffer.height + 1) / 2;
        GLES20.glActiveTexture(GLES20.GL_TEXTURE0 + i);
        GLES20.glBindTexture(GLES20.GL_TEXTURE_2D, yuvTextures[i]);
        GLES20.glPixelStorei(GLES20.GL_UNPACK_ALIGNMENT, 1);
        GLES20.glTexImage2D(
            GLES20.GL_TEXTURE_2D,
            /* level= */ 0,
            GLES20.GL_LUMINANCE,
            yuvStrides[i],
            h,
            /* border= */ 0,
            GLES20.GL_LUMINANCE,
            GLES20.GL_UNSIGNED_BYTE,
            yuvPlanes[i]);
      }

      int[] widths = new int[3];
      widths[0] = outputBuffer.width;
      // TODO(b/142097774): Handle streams where chroma channels are not stored at half width and
      // height compared to the luma channel. U and V planes are being stored at half width compared
      // to Y.
      widths[1] = widths[2] = (widths[0] + 1) / 2;
      for (int i = 0; i < 3; i++) {
        // Set cropping of stride if either width or stride has changed.
        if (previousWidths[i] != widths[i] || previousStrides[i] != yuvStrides[i]) {
          Assertions.checkState(yuvStrides[i] != 0);
          float widthRatio = (float) widths[i] / yuvStrides[i];
          // These buffers are consumed during each call to glDrawArrays. They need to be member
          // variables rather than local variables in order not to get garbage collected.
          textureCoords[i] =
              GlUtil.createBuffer(
                  new float[] {0.0f, 0.0f, 0.0f, 1.0f, widthRatio, 0.0f, widthRatio, 1.0f});
          GLES20.glVertexAttribPointer(
              texLocations[i],
              /* size= */ 2,
              GLES20.GL_FLOAT,
              /* normalized= */ false,
              /* stride= */ 0,
              textureCoords[i]);
          previousWidths[i] = widths[i];
          previousStrides[i] = yuvStrides[i];
        }
      }

      GLES20.glClear(GLES20.GL_COLOR_BUFFER_BIT);
      GLES20.glDrawArrays(GLES20.GL_TRIANGLE_STRIP, /* first= */ 0, /* count= */ 4);
      try {
        GlUtil.checkGlError();
      } catch (GlUtil.GlException e) {
        Log.e(TAG, "Failed to draw a frame", e);
      }
    }

    public void setOutputBuffer(VideoDecoderOutputBuffer outputBuffer) {
      @Nullable
      VideoDecoderOutputBuffer oldPendingOutputBuffer =
          pendingOutputBufferReference.getAndSet(outputBuffer);
      if (oldPendingOutputBuffer != null) {
        // The old pending output buffer will never be used for rendering, so release it now.
        oldPendingOutputBuffer.release();
      }
      surfaceView.requestRender();
    }

    @RequiresNonNull("program")
    private void setupTextures() {
      try {
        GLES20.glGenTextures(/* n= */ 3, yuvTextures, /* offset= */ 0);
        for (int i = 0; i < 3; i++) {
          GLES20.glUniform1i(program.getUniformLocation(TEXTURE_UNIFORMS[i]), i);
          GLES20.glActiveTexture(GLES20.GL_TEXTURE0 + i);
          GlUtil.bindTexture(GLES20.GL_TEXTURE_2D, yuvTextures[i], GLES20.GL_LINEAR);
        }
        GlUtil.checkGlError();
      } catch (GlUtil.GlException e) {
        Log.e(TAG, "Failed to set up the textures", e);
      }
    }
  }
}