/*
 *  Copyright 2013 The WebRTC project authors. All Rights Reserved.
 *
 *  Use of this source code is governed by a BSD-style license
 *  that can be found in the LICENSE file in the root of the source
 *  tree. An additional intellectual property rights grant can be found
 *  in the file PATENTS.  All contributing project authors may
 *  be found in the AUTHORS file in the root of the source tree.
 */

package org.webrtc;

import android.annotation.TargetApi;
import android.graphics.Matrix;
import android.media.MediaCodec;
import android.media.MediaCodecInfo;
import android.media.MediaCodecInfo.CodecCapabilities;
import android.media.MediaCodecList;
import android.media.MediaFormat;
import android.opengl.GLES20;
import android.os.Build;
import android.os.Bundle;
import android.view.Surface;
import java.nio.ByteBuffer;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.HashSet;
import java.util.List;
import java.util.Set;
import java.util.concurrent.CountDownLatch;
import java.util.concurrent.TimeUnit;
import org.webrtc.EglBase14;
import org.webrtc.VideoFrame;

// Java-side of peerconnection.cc:MediaCodecVideoEncoder.
// This class is an implementation detail of the Java PeerConnection API.
@TargetApi(19)
@SuppressWarnings("deprecation")
@JNINamespace("webrtc::jni")
public class MediaCodecVideoEncoder {
	// This class is constructed, operated, and destroyed by its C++ incarnation,
	// so the class and its methods have non-public visibility.  The API this
	// class exposes aims to mimic the webrtc::VideoEncoder API as closely as
	// possibly to minimize the amount of translation work necessary.

	private static final String TAG = "MediaCodecVideoEncoder";

	// Tracks webrtc::VideoCodecType.
	public enum VideoCodecType {
		VIDEO_CODEC_VP8,
		VIDEO_CODEC_VP9,
		VIDEO_CODEC_H264;

		@CalledByNative("VideoCodecType")
		static VideoCodecType fromNativeIndex(int nativeIndex) {
			return values()[nativeIndex];
		}
	}

	private static final int MEDIA_CODEC_RELEASE_TIMEOUT_MS = 5000; // Timeout for codec releasing.
	private static final int DEQUEUE_TIMEOUT = 0; // Non-blocking, no wait.
	private static final int BITRATE_ADJUSTMENT_FPS = 30;
	private static final int MAXIMUM_INITIAL_FPS = 30;
	private static final double BITRATE_CORRECTION_SEC = 3.0;
	// Maximum bitrate correction scale - no more than 4 times.
	private static final double BITRATE_CORRECTION_MAX_SCALE = 4;
	// Amount of correction steps to reach correction maximum scale.
	private static final int BITRATE_CORRECTION_STEPS = 20;
	// Forced key frame interval - used to reduce color distortions on Qualcomm platform.
	private static final long QCOM_VP8_KEY_FRAME_INTERVAL_ANDROID_L_MS = 15000;
	private static final long QCOM_VP8_KEY_FRAME_INTERVAL_ANDROID_M_MS = 20000;
	private static final long QCOM_VP8_KEY_FRAME_INTERVAL_ANDROID_N_MS = 15000;

	// Active running encoder instance. Set in initEncode() (called from native code)
	// and reset to null in release() call.
	private static MediaCodecVideoEncoder runningInstance = null;
	private static MediaCodecVideoEncoderErrorCallback errorCallback = null;
	private static int codecErrors = 0;
	// List of disabled codec types - can be set from application.
	private static Set<String> hwEncoderDisabledTypes = new HashSet<String>();

	private Thread mediaCodecThread;
	private MediaCodec mediaCodec;
	private ByteBuffer[] outputBuffers;
	private EglBase14 eglBase;
	private int profile;
	private int width;
	private int height;
	private Surface inputSurface;
	private GlRectDrawer drawer;

	private static final String VP8_MIME_TYPE = "video/x-vnd.on2.vp8";
	private static final String VP9_MIME_TYPE = "video/x-vnd.on2.vp9";
	private static final String H264_MIME_TYPE = "video/avc";

	private static final int VIDEO_AVCProfileHigh = 8;
	private static final int VIDEO_AVCLevel3 = 0x100;

	// Type of bitrate adjustment for video encoder.
	public enum BitrateAdjustmentType {
		// No adjustment - video encoder has no known bitrate problem.
		NO_ADJUSTMENT,
		// Framerate based bitrate adjustment is required - HW encoder does not use frame
		// timestamps to calculate frame bitrate budget and instead is relying on initial
		// fps configuration assuming that all frames are coming at fixed initial frame rate.
		FRAMERATE_ADJUSTMENT,
		// Dynamic bitrate adjustment is required - HW encoder used frame timestamps, but actual
		// bitrate deviates too much from the target value.
		DYNAMIC_ADJUSTMENT
	}

	// Should be in sync with webrtc::H264::Profile.
	public static enum H264Profile {
		CONSTRAINED_BASELINE(0),
		BASELINE(1),
		MAIN(2),
		CONSTRAINED_HIGH(3),
		HIGH(4);

		private final int value;

		H264Profile(int value) {
			this.value = value;
		}

		public int getValue() {
			return value;
		}
	}

	// Class describing supported media codec properties.
	private static class MediaCodecProperties {
		public final String codecPrefix;
		// Minimum Android SDK required for this codec to be used.
		public final int minSdk;
		// Flag if encoder implementation does not use frame timestamps to calculate frame bitrate
		// budget and instead is relying on initial fps configuration assuming that all frames are
		// coming at fixed initial frame rate. Bitrate adjustment is required for this case.
		public final BitrateAdjustmentType bitrateAdjustmentType;

		MediaCodecProperties(
				String codecPrefix, int minSdk, BitrateAdjustmentType bitrateAdjustmentType) {
			this.codecPrefix = codecPrefix;
			this.minSdk = minSdk;
			this.bitrateAdjustmentType = bitrateAdjustmentType;
		}
	}

	// List of supported HW VP8 encoders.
	private static final MediaCodecProperties qcomVp8HwProperties = new MediaCodecProperties(
			"OMX.qcom.", Build.VERSION_CODES.KITKAT, BitrateAdjustmentType.NO_ADJUSTMENT);
	private static final MediaCodecProperties exynosVp8HwProperties = new MediaCodecProperties(
			"OMX.Exynos.", Build.VERSION_CODES.M, BitrateAdjustmentType.DYNAMIC_ADJUSTMENT);
	private static final MediaCodecProperties intelVp8HwProperties = new MediaCodecProperties(
			"OMX.Intel.", Build.VERSION_CODES.LOLLIPOP, BitrateAdjustmentType.NO_ADJUSTMENT);

	private static MediaCodecProperties[] vp8HwList() {
		final ArrayList<MediaCodecProperties> supported_codecs = new ArrayList<MediaCodecProperties>();
		supported_codecs.add(qcomVp8HwProperties);
		supported_codecs.add(exynosVp8HwProperties);
		if (PeerConnectionFactory.fieldTrialsFindFullName("WebRTC-IntelVP8").equals("Enabled")) {
			supported_codecs.add(intelVp8HwProperties);
		}
		return supported_codecs.toArray(new MediaCodecProperties[supported_codecs.size()]);
	}

	// List of supported HW VP9 encoders.
	private static final MediaCodecProperties qcomVp9HwProperties = new MediaCodecProperties(
			"OMX.qcom.", Build.VERSION_CODES.N, BitrateAdjustmentType.NO_ADJUSTMENT);
	private static final MediaCodecProperties exynosVp9HwProperties = new MediaCodecProperties(
			"OMX.Exynos.", Build.VERSION_CODES.N, BitrateAdjustmentType.FRAMERATE_ADJUSTMENT);
	private static final MediaCodecProperties[] vp9HwList =
			new MediaCodecProperties[] {qcomVp9HwProperties, exynosVp9HwProperties};

	// List of supported HW H.264 encoders.
	private static final MediaCodecProperties qcomH264HwProperties = new MediaCodecProperties(
			"OMX.qcom.", Build.VERSION_CODES.KITKAT, BitrateAdjustmentType.NO_ADJUSTMENT);
	private static final MediaCodecProperties exynosH264HwProperties = new MediaCodecProperties(
			"OMX.Exynos.", Build.VERSION_CODES.LOLLIPOP, BitrateAdjustmentType.FRAMERATE_ADJUSTMENT);


	private static final MediaCodecProperties googleH264HwProperties = new MediaCodecProperties(
			"OMX.google.", 21, MediaCodecVideoEncoder.BitrateAdjustmentType.FRAMERATE_ADJUSTMENT);
	private static final MediaCodecProperties hisiH264HwProperties = new MediaCodecProperties(
			"OMX.hisi.", 21, MediaCodecVideoEncoder.BitrateAdjustmentType.FRAMERATE_ADJUSTMENT);

	private static final MediaCodecProperties[] h264HwList =
			new MediaCodecProperties[] {qcomH264HwProperties, exynosH264HwProperties,googleH264HwProperties,hisiH264HwProperties};

	// List of supported HW H.264 high profile encoders.
	private static final MediaCodecProperties exynosH264HighProfileHwProperties =
			new MediaCodecProperties(
					"OMX.Exynos.", Build.VERSION_CODES.M, BitrateAdjustmentType.FRAMERATE_ADJUSTMENT);
	private static final MediaCodecProperties[] h264HighProfileHwList =
			new MediaCodecProperties[] {exynosH264HighProfileHwProperties};

	// List of devices with poor H.264 encoder quality.
	// HW H.264 encoder on below devices has poor bitrate control - actual
	// bitrates deviates a lot from the target value.
	private static final String[] H264_HW_EXCEPTION_MODELS =
			new String[] {"SAMSUNG-SGH-I337", "Nexus 7", "Nexus 4"};

	// Bitrate modes - should be in sync with OMX_VIDEO_CONTROLRATETYPE defined
	// in OMX_Video.h
	private static final int VIDEO_ControlRateConstant = 2;
	// NV12 color format supported by QCOM codec, but not declared in MediaCodec -
	// see /hardware/qcom/media/mm-core/inc/OMX_QCOMExtns.h
	private static final int COLOR_QCOM_FORMATYUV420PackedSemiPlanar32m = 0x7FA30C04;
	// Allowable color formats supported by codec - in order of preference.
	private static final int[] supportedColorList = {CodecCapabilities.COLOR_FormatYUV420Planar,
			CodecCapabilities.COLOR_FormatYUV420SemiPlanar,
			CodecCapabilities.COLOR_QCOM_FormatYUV420SemiPlanar,
			COLOR_QCOM_FORMATYUV420PackedSemiPlanar32m};
	private static final int[] supportedSurfaceColorList = {CodecCapabilities.COLOR_FormatSurface};
	private VideoCodecType type;
	private int colorFormat;

	// Variables used for dynamic bitrate adjustment.
	private BitrateAdjustmentType bitrateAdjustmentType = BitrateAdjustmentType.NO_ADJUSTMENT;
	private double bitrateAccumulator;
	private double bitrateAccumulatorMax;
	private double bitrateObservationTimeMs;
	private int bitrateAdjustmentScaleExp;
	private int targetBitrateBps;
	private int targetFps;

	// Interval in ms to force key frame generation. Used to reduce the time of color distortions
	// happened sometime when using Qualcomm video encoder.
	private long forcedKeyFrameMs;
	private long lastKeyFrameMs;

	// SPS and PPS NALs (Config frame) for H.264.
	private ByteBuffer configData = null;

	// MediaCodec error handler - invoked when critical error happens which may prevent
	// further use of media codec API. Now it means that one of media codec instances
	// is hanging and can no longer be used in the next call.
	public static interface MediaCodecVideoEncoderErrorCallback {
		void onMediaCodecVideoEncoderCriticalError(int codecErrors);
	}

	public static void setErrorCallback(MediaCodecVideoEncoderErrorCallback errorCallback) {
		Logging.d(TAG, "Set error callback");
		MediaCodecVideoEncoder.errorCallback = errorCallback;
	}

	// Functions to disable HW encoding - can be called from applications for platforms
	// which have known HW decoding problems.
	public static void disableVp8HwCodec() {
		Logging.w(TAG, "VP8 encoding is disabled by application.");
		hwEncoderDisabledTypes.add(VP8_MIME_TYPE);
	}

	public static void disableVp9HwCodec() {
		Logging.w(TAG, "VP9 encoding is disabled by application.");
		hwEncoderDisabledTypes.add(VP9_MIME_TYPE);
	}

	public static void disableH264HwCodec() {
		Logging.w(TAG, "H.264 encoding is disabled by application.");
		hwEncoderDisabledTypes.add(H264_MIME_TYPE);
	}

	// Functions to query if HW encoding is supported.
	@CalledByNative
	public static boolean isVp8HwSupported() {
		return !hwEncoderDisabledTypes.contains(VP8_MIME_TYPE)
				&& (findHwEncoder(VP8_MIME_TYPE, vp8HwList(), supportedColorList) != null);
	}

	public static EncoderProperties vp8HwEncoderProperties() {
		if (hwEncoderDisabledTypes.contains(VP8_MIME_TYPE)) {
			return null;
		} else {
			return findHwEncoder(VP8_MIME_TYPE, vp8HwList(), supportedColorList);
		}
	}

	@CalledByNative
	public static boolean isVp9HwSupported() {
		return !hwEncoderDisabledTypes.contains(VP9_MIME_TYPE)
				&& (findHwEncoder(VP9_MIME_TYPE, vp9HwList, supportedColorList) != null);
	}

	@CalledByNative
	public static boolean isH264HwSupported() {
		return !hwEncoderDisabledTypes.contains(H264_MIME_TYPE)
				&& (findHwEncoder(H264_MIME_TYPE, h264HwList, supportedColorList) != null);
	}

	public static boolean isH264HighProfileHwSupported() {
		return !hwEncoderDisabledTypes.contains(H264_MIME_TYPE)
				&& (findHwEncoder(H264_MIME_TYPE, h264HighProfileHwList, supportedColorList) != null);
	}

	public static boolean isVp8HwSupportedUsingTextures() {
		return !hwEncoderDisabledTypes.contains(VP8_MIME_TYPE)
				&& (findHwEncoder(VP8_MIME_TYPE, vp8HwList(), supportedSurfaceColorList) != null);
	}

	public static boolean isVp9HwSupportedUsingTextures() {
		return !hwEncoderDisabledTypes.contains(VP9_MIME_TYPE)
				&& (findHwEncoder(VP9_MIME_TYPE, vp9HwList, supportedSurfaceColorList) != null);
	}

	public static boolean isH264HwSupportedUsingTextures() {
		return !hwEncoderDisabledTypes.contains(H264_MIME_TYPE)
				&& (findHwEncoder(H264_MIME_TYPE, h264HwList, supportedSurfaceColorList) != null);
	}

	// Helper struct for findHwEncoder() below.
	public static class EncoderProperties {
		public EncoderProperties(
				String codecName, int colorFormat, BitrateAdjustmentType bitrateAdjustmentType) {
			this.codecName = codecName;
			this.colorFormat = colorFormat;
			this.bitrateAdjustmentType = bitrateAdjustmentType;
		}
		public final String codecName; // OpenMax component name for HW codec.
		public final int colorFormat; // Color format supported by codec.
		public final BitrateAdjustmentType bitrateAdjustmentType; // Bitrate adjustment type
	}

	private static EncoderProperties findHwEncoder(
			String mime, MediaCodecProperties[] supportedHwCodecProperties, int[] colorList) {
		// MediaCodec.setParameters is missing for JB and below, so bitrate
		// can not be adjusted dynamically.
		if (Build.VERSION.SDK_INT < Build.VERSION_CODES.KITKAT) {
			return null;
		}

		// Check if device is in H.264 exception list.
		if (mime.equals(H264_MIME_TYPE)) {
			List<String> exceptionModels = Arrays.asList(H264_HW_EXCEPTION_MODELS);
			if (exceptionModels.contains(Build.MODEL)) {
				Logging.w(TAG, "Model: " + Build.MODEL + " has black listed H.264 encoder.");
				return null;
			}
		}

		for (int i = 0; i < MediaCodecList.getCodecCount(); ++i) {
			MediaCodecInfo info = null;
			try {
				info = MediaCodecList.getCodecInfoAt(i);
			} catch (IllegalArgumentException e) {
				Logging.e(TAG, "Cannot retrieve encoder codec info", e);
			}
			if (info == null || !info.isEncoder()) {
				continue;
			}
			String name = null;
			for (String mimeType : info.getSupportedTypes()) {
				if (mimeType.equals(mime)) {
					name = info.getName();
					break;
				}
			}
			if (name == null) {
				continue; // No HW support in this codec; try the next one.
			}
			Logging.v(TAG, "Found candidate encoder " + name);

			// Check if this is supported HW encoder.
			boolean supportedCodec = false;
			BitrateAdjustmentType bitrateAdjustmentType = BitrateAdjustmentType.NO_ADJUSTMENT;
			for (MediaCodecProperties codecProperties : supportedHwCodecProperties) {
				if (name.startsWith(codecProperties.codecPrefix)) {
					if (Build.VERSION.SDK_INT < codecProperties.minSdk) {
						Logging.w(
								TAG, "Codec " + name + " is disabled due to SDK version " + Build.VERSION.SDK_INT);
						continue;
					}
					if (codecProperties.bitrateAdjustmentType != BitrateAdjustmentType.NO_ADJUSTMENT) {
						bitrateAdjustmentType = codecProperties.bitrateAdjustmentType;
						Logging.w(
								TAG, "Codec " + name + " requires bitrate adjustment: " + bitrateAdjustmentType);
					}
					supportedCodec = true;
					break;
				}
			}
			if (!supportedCodec) {
				continue;
			}

			// Check if HW codec supports known color format.
			CodecCapabilities capabilities;
			try {
				capabilities = info.getCapabilitiesForType(mime);
			} catch (IllegalArgumentException e) {
				Logging.e(TAG, "Cannot retrieve encoder capabilities", e);
				continue;
			}
			for (int colorFormat : capabilities.colorFormats) {
				Logging.v(TAG, "   Color: 0x" + Integer.toHexString(colorFormat));
			}

			for (int supportedColorFormat : colorList) {
				for (int codecColorFormat : capabilities.colorFormats) {
					if (codecColorFormat == supportedColorFormat) {
						// Found supported HW encoder.
						Logging.d(TAG, "Found target encoder for mime " + mime + " : " + name + ". Color: 0x"
								+ Integer.toHexString(codecColorFormat) + ". Bitrate adjustment: "
								+ bitrateAdjustmentType);
						return new EncoderProperties(name, codecColorFormat, bitrateAdjustmentType);
					}
				}
			}
		}
		return null; // No HW encoder.
	}

	@CalledByNative
	MediaCodecVideoEncoder() {}

	private void checkOnMediaCodecThread() {
		if (mediaCodecThread.getId() != Thread.currentThread().getId()) {
			throw new RuntimeException("MediaCodecVideoEncoder previously operated on " + mediaCodecThread
					+ " but is now called on " + Thread.currentThread());
		}
	}

	public static void printStackTrace() {
		if (runningInstance != null && runningInstance.mediaCodecThread != null) {
			StackTraceElement[] mediaCodecStackTraces = runningInstance.mediaCodecThread.getStackTrace();
			if (mediaCodecStackTraces.length > 0) {
				Logging.d(TAG, "MediaCodecVideoEncoder stacks trace:");
				for (StackTraceElement stackTrace : mediaCodecStackTraces) {
					Logging.d(TAG, stackTrace.toString());
				}
			}
		}
	}

	static MediaCodec createByCodecName(String codecName) {
		try {
			// In the L-SDK this call can throw IOException so in order to work in
			// both cases catch an exception.
			return MediaCodec.createByCodecName(codecName);
		} catch (Exception e) {
			return null;
		}
	}

	@CalledByNativeUnchecked
	boolean initEncode(VideoCodecType type, int profile, int width, int height, int kbps, int fps,
	                   EglBase14.Context sharedContext) {
		final boolean useSurface = sharedContext != null;
		Logging.d(TAG,
				"Java initEncode: " + type + ". Profile: " + profile + " : " + width + " x " + height
						+ ". @ " + kbps + " kbps. Fps: " + fps + ". Encode from texture : " + useSurface);

		this.profile = profile;
		this.width = width;
		this.height = height;
		if (mediaCodecThread != null) {
			throw new RuntimeException("Forgot to release()?");
		}
		EncoderProperties properties = null;
		String mime = null;
		int keyFrameIntervalSec = 0;
		boolean configureH264HighProfile = false;
		if (type == VideoCodecType.VIDEO_CODEC_VP8) {
			mime = VP8_MIME_TYPE;
			properties = findHwEncoder(
					VP8_MIME_TYPE, vp8HwList(), useSurface ? supportedSurfaceColorList : supportedColorList);
			keyFrameIntervalSec = 100;
		} else if (type == VideoCodecType.VIDEO_CODEC_VP9) {
			mime = VP9_MIME_TYPE;
			properties = findHwEncoder(
					VP9_MIME_TYPE, vp9HwList, useSurface ? supportedSurfaceColorList : supportedColorList);
			keyFrameIntervalSec = 100;
		} else if (type == VideoCodecType.VIDEO_CODEC_H264) {
			mime = H264_MIME_TYPE;
			properties = findHwEncoder(
					H264_MIME_TYPE, h264HwList, useSurface ? supportedSurfaceColorList : supportedColorList);
			if (profile == H264Profile.CONSTRAINED_HIGH.getValue()) {
				EncoderProperties h264HighProfileProperties = findHwEncoder(H264_MIME_TYPE,
						h264HighProfileHwList, useSurface ? supportedSurfaceColorList : supportedColorList);
				if (h264HighProfileProperties != null) {
					Logging.d(TAG, "High profile H.264 encoder supported.");
					configureH264HighProfile = true;
				} else {
					Logging.d(TAG, "High profile H.264 encoder requested, but not supported. Use baseline.");
				}
			}
			keyFrameIntervalSec = 20;
		}
		if (properties == null) {
			throw new RuntimeException("Can not find HW encoder for " + type);
		}
		runningInstance = this; // Encoder is now running and can be queried for stack traces.
		colorFormat = properties.colorFormat;
		bitrateAdjustmentType = properties.bitrateAdjustmentType;
		if (bitrateAdjustmentType == BitrateAdjustmentType.FRAMERATE_ADJUSTMENT) {
			fps = BITRATE_ADJUSTMENT_FPS;
		} else {
			fps = Math.min(fps, MAXIMUM_INITIAL_FPS);
		}

		forcedKeyFrameMs = 0;
		lastKeyFrameMs = -1;
		if (type == VideoCodecType.VIDEO_CODEC_VP8
				&& properties.codecName.startsWith(qcomVp8HwProperties.codecPrefix)) {
			if (Build.VERSION.SDK_INT == Build.VERSION_CODES.LOLLIPOP
					|| Build.VERSION.SDK_INT == Build.VERSION_CODES.LOLLIPOP_MR1) {
				forcedKeyFrameMs = QCOM_VP8_KEY_FRAME_INTERVAL_ANDROID_L_MS;
			} else if (Build.VERSION.SDK_INT == Build.VERSION_CODES.M) {
				forcedKeyFrameMs = QCOM_VP8_KEY_FRAME_INTERVAL_ANDROID_M_MS;
			} else if (Build.VERSION.SDK_INT > Build.VERSION_CODES.M) {
				forcedKeyFrameMs = QCOM_VP8_KEY_FRAME_INTERVAL_ANDROID_N_MS;
			}
		}

		Logging.d(TAG, "Color format: " + colorFormat + ". Bitrate adjustment: " + bitrateAdjustmentType
				+ ". Key frame interval: " + forcedKeyFrameMs + " . Initial fps: " + fps);
		targetBitrateBps = 1000 * kbps;
		targetFps = fps;
		bitrateAccumulatorMax = targetBitrateBps / 8.0;
		bitrateAccumulator = 0;
		bitrateObservationTimeMs = 0;
		bitrateAdjustmentScaleExp = 0;

		mediaCodecThread = Thread.currentThread();
		try {
			MediaFormat format = MediaFormat.createVideoFormat(mime, width, height);
			format.setInteger(MediaFormat.KEY_BIT_RATE, targetBitrateBps);
			format.setInteger("bitrate-mode", VIDEO_ControlRateConstant);
			format.setInteger(MediaFormat.KEY_COLOR_FORMAT, properties.colorFormat);
			format.setInteger(MediaFormat.KEY_FRAME_RATE, targetFps);
			format.setInteger(MediaFormat.KEY_I_FRAME_INTERVAL, keyFrameIntervalSec);
			if (configureH264HighProfile) {
				format.setInteger("profile", VIDEO_AVCProfileHigh);
				format.setInteger("level", VIDEO_AVCLevel3);
			}
			Logging.d(TAG, "  Format: " + format);
			mediaCodec = createByCodecName(properties.codecName);
			this.type = type;
			if (mediaCodec == null) {
				Logging.e(TAG, "Can not create media encoder");
				release();
				return false;
			}
			mediaCodec.configure(format, null, null, MediaCodec.CONFIGURE_FLAG_ENCODE);

			if (useSurface) {
				eglBase = new EglBase14(sharedContext, EglBase.CONFIG_RECORDABLE);
				// Create an input surface and keep a reference since we must release the surface when done.
				inputSurface = mediaCodec.createInputSurface();
				eglBase.createSurface(inputSurface);
				drawer = new GlRectDrawer();
			}
			mediaCodec.start();
			outputBuffers = mediaCodec.getOutputBuffers();
			Logging.d(TAG, "Output buffers: " + outputBuffers.length);

		} catch (IllegalStateException e) {
			Logging.e(TAG, "initEncode failed", e);
			release();
			return false;
		}
		return true;
	}

	@CalledByNativeUnchecked
	ByteBuffer[] getInputBuffers() {
		ByteBuffer[] inputBuffers = mediaCodec.getInputBuffers();
		Logging.d(TAG, "Input buffers: " + inputBuffers.length);
		return inputBuffers;
	}

	void checkKeyFrameRequired(boolean requestedKeyFrame, long presentationTimestampUs) {
		long presentationTimestampMs = (presentationTimestampUs + 500) / 1000;
		if (lastKeyFrameMs < 0) {
			lastKeyFrameMs = presentationTimestampMs;
		}
		boolean forcedKeyFrame = false;
		if (!requestedKeyFrame && forcedKeyFrameMs > 0
				&& presentationTimestampMs > lastKeyFrameMs + forcedKeyFrameMs) {
			forcedKeyFrame = true;
		}
		if (requestedKeyFrame || forcedKeyFrame) {
			// Ideally MediaCodec would honor BUFFER_FLAG_SYNC_FRAME so we could
			// indicate this in queueInputBuffer() below and guarantee _this_ frame
			// be encoded as a key frame, but sadly that flag is ignored.  Instead,
			// we request a key frame "soon".
			if (requestedKeyFrame) {
				Logging.d(TAG, "Sync frame request");
			} else {
				Logging.d(TAG, "Sync frame forced");
			}
			Bundle b = new Bundle();
			b.putInt(MediaCodec.PARAMETER_KEY_REQUEST_SYNC_FRAME, 0);
			mediaCodec.setParameters(b);
			lastKeyFrameMs = presentationTimestampMs;
		}
	}

	@CalledByNativeUnchecked
	boolean encodeBuffer(
			boolean isKeyframe, int inputBuffer, int size, long presentationTimestampUs) {
		checkOnMediaCodecThread();
		try {
			checkKeyFrameRequired(isKeyframe, presentationTimestampUs);
			mediaCodec.queueInputBuffer(inputBuffer, 0, size, presentationTimestampUs, 0);
			return true;
		} catch (IllegalStateException e) {
			Logging.e(TAG, "encodeBuffer failed", e);
			return false;
		}
	}

	@CalledByNativeUnchecked
	boolean encodeTexture(boolean isKeyframe, int oesTextureId, float[] transformationMatrix,
	                      long presentationTimestampUs) {
		checkOnMediaCodecThread();
		try {
			checkKeyFrameRequired(isKeyframe, presentationTimestampUs);
			eglBase.makeCurrent();
			// TODO(perkj): glClear() shouldn't be necessary since every pixel is covered anyway,
			// but it's a workaround for bug webrtc:5147.
			GLES20.glClear(GLES20.GL_COLOR_BUFFER_BIT);
			drawer.drawOes(oesTextureId, transformationMatrix, width, height, 0, 0, width, height);
			eglBase.swapBuffers(TimeUnit.MICROSECONDS.toNanos(presentationTimestampUs));
			return true;
		} catch (RuntimeException e) {
			Logging.e(TAG, "encodeTexture failed", e);
			return false;
		}
	}

	/**
	 * Encodes a new style VideoFrame. |bufferIndex| is -1 if we are not encoding in surface mode.
	 */
	@CalledByNativeUnchecked
	boolean encodeFrame(long nativeEncoder, boolean isKeyframe, VideoFrame frame, int bufferIndex) {
		checkOnMediaCodecThread();
		try {
			long presentationTimestampUs = TimeUnit.NANOSECONDS.toMicros(frame.getTimestampNs());
			checkKeyFrameRequired(isKeyframe, presentationTimestampUs);

			VideoFrame.Buffer buffer = frame.getBuffer();
			if (buffer instanceof VideoFrame.TextureBuffer) {
				VideoFrame.TextureBuffer textureBuffer = (VideoFrame.TextureBuffer) buffer;
				eglBase.makeCurrent();
				// TODO(perkj): glClear() shouldn't be necessary since every pixel is covered anyway,
				// but it's a workaround for bug webrtc:5147.
				GLES20.glClear(GLES20.GL_COLOR_BUFFER_BIT);
				VideoFrameDrawer.drawTexture(drawer, textureBuffer, new Matrix() /* renderMatrix */, width,
						height, 0 /* viewportX */, 0 /* viewportY */, width, height);
				eglBase.swapBuffers(frame.getTimestampNs());
			} else {
				VideoFrame.I420Buffer i420Buffer = buffer.toI420();
				final int chromaHeight = (height + 1) / 2;
				final ByteBuffer dataY = i420Buffer.getDataY();
				final ByteBuffer dataU = i420Buffer.getDataU();
				final ByteBuffer dataV = i420Buffer.getDataV();
				final int strideY = i420Buffer.getStrideY();
				final int strideU = i420Buffer.getStrideU();
				final int strideV = i420Buffer.getStrideV();
				if (dataY.capacity() < strideY * height) {
					throw new RuntimeException("Y-plane buffer size too small.");
				}
				if (dataU.capacity() < strideU * chromaHeight) {
					throw new RuntimeException("U-plane buffer size too small.");
				}
				if (dataV.capacity() < strideV * chromaHeight) {
					throw new RuntimeException("V-plane buffer size too small.");
				}
				nativeFillInputBuffer(
						nativeEncoder, bufferIndex, dataY, strideY, dataU, strideU, dataV, strideV);
				i420Buffer.release();
				// I420 consists of one full-resolution and two half-resolution planes.
				// 1 + 1 / 4 + 1 / 4 = 3 / 2
				int yuvSize = width * height * 3 / 2;
				mediaCodec.queueInputBuffer(bufferIndex, 0, yuvSize, presentationTimestampUs, 0);
			}
			return true;
		} catch (RuntimeException e) {
			Logging.e(TAG, "encodeFrame failed", e);
			return false;
		}
	}

	@CalledByNativeUnchecked
	void release() {
		Logging.d(TAG, "Java releaseEncoder");
		checkOnMediaCodecThread();

		class CaughtException {
			Exception e;
		}
		final CaughtException caughtException = new CaughtException();
		boolean stopHung = false;

		if (mediaCodec != null) {
			// Run Mediacodec stop() and release() on separate thread since sometime
			// Mediacodec.stop() may hang.
			final CountDownLatch releaseDone = new CountDownLatch(1);

			Runnable runMediaCodecRelease = new Runnable() {
				@Override
				public void run() {
					Logging.d(TAG, "Java releaseEncoder on release thread");
					try {
						mediaCodec.stop();
					} catch (Exception e) {
						Logging.e(TAG, "Media encoder stop failed", e);
					}
					try {
						mediaCodec.release();
					} catch (Exception e) {
						Logging.e(TAG, "Media encoder release failed", e);
						caughtException.e = e;
					}
					Logging.d(TAG, "Java releaseEncoder on release thread done");

					releaseDone.countDown();
				}
			};
			new Thread(runMediaCodecRelease).start();

			if (!ThreadUtils.awaitUninterruptibly(releaseDone, MEDIA_CODEC_RELEASE_TIMEOUT_MS)) {
				Logging.e(TAG, "Media encoder release timeout");
				stopHung = true;
			}

			mediaCodec = null;
		}

		mediaCodecThread = null;
		if (drawer != null) {
			drawer.release();
			drawer = null;
		}
		if (eglBase != null) {
			eglBase.release();
			eglBase = null;
		}
		if (inputSurface != null) {
			inputSurface.release();
			inputSurface = null;
		}
		runningInstance = null;

		if (stopHung) {
			codecErrors++;
			if (errorCallback != null) {
				Logging.e(TAG, "Invoke codec error callback. Errors: " + codecErrors);
				errorCallback.onMediaCodecVideoEncoderCriticalError(codecErrors);
			}
			throw new RuntimeException("Media encoder release timeout.");
		}

		// Re-throw any runtime exception caught inside the other thread. Since this is an invoke, add
		// stack trace for the waiting thread as well.
		if (caughtException.e != null) {
			final RuntimeException runtimeException = new RuntimeException(caughtException.e);
			runtimeException.setStackTrace(ThreadUtils.concatStackTraces(
					caughtException.e.getStackTrace(), runtimeException.getStackTrace()));
			throw runtimeException;
		}

		Logging.d(TAG, "Java releaseEncoder done");
	}

	@CalledByNativeUnchecked
	private boolean setRates(int kbps, int frameRate) {
		checkOnMediaCodecThread();

		int codecBitrateBps = 1000 * kbps;
		if (bitrateAdjustmentType == BitrateAdjustmentType.DYNAMIC_ADJUSTMENT) {
			bitrateAccumulatorMax = codecBitrateBps / 8.0;
			if (targetBitrateBps > 0 && codecBitrateBps < targetBitrateBps) {
				// Rescale the accumulator level if the accumulator max decreases
				bitrateAccumulator = bitrateAccumulator * codecBitrateBps / targetBitrateBps;
			}
		}
		targetBitrateBps = codecBitrateBps;
		targetFps = frameRate;

		// Adjust actual encoder bitrate based on bitrate adjustment type.
		if (bitrateAdjustmentType == BitrateAdjustmentType.FRAMERATE_ADJUSTMENT && targetFps > 0) {
			codecBitrateBps = BITRATE_ADJUSTMENT_FPS * targetBitrateBps / targetFps;
			Logging.v(TAG,
					"setRates: " + kbps + " -> " + (codecBitrateBps / 1000) + " kbps. Fps: " + targetFps);
		} else if (bitrateAdjustmentType == BitrateAdjustmentType.DYNAMIC_ADJUSTMENT) {
			Logging.v(TAG, "setRates: " + kbps + " kbps. Fps: " + targetFps + ". ExpScale: "
					+ bitrateAdjustmentScaleExp);
			if (bitrateAdjustmentScaleExp != 0) {
				codecBitrateBps = (int) (codecBitrateBps * getBitrateScale(bitrateAdjustmentScaleExp));
			}
		} else {
			Logging.v(TAG, "setRates: " + kbps + " kbps. Fps: " + targetFps);
		}

		try {
			Bundle params = new Bundle();
			params.putInt(MediaCodec.PARAMETER_KEY_VIDEO_BITRATE, codecBitrateBps);
			mediaCodec.setParameters(params);
			return true;
		} catch (IllegalStateException e) {
			Logging.e(TAG, "setRates failed", e);
			return false;
		}
	}

	// Dequeue an input buffer and return its index, -1 if no input buffer is
	// available, or -2 if the codec is no longer operative.
	@CalledByNativeUnchecked
	int dequeueInputBuffer() {
		checkOnMediaCodecThread();
		try {
			return mediaCodec.dequeueInputBuffer(DEQUEUE_TIMEOUT);
		} catch (IllegalStateException e) {
			Logging.e(TAG, "dequeueIntputBuffer failed", e);
			return -2;
		}
	}

	// Helper struct for dequeueOutputBuffer() below.
	static class OutputBufferInfo {
		public OutputBufferInfo(
				int index, ByteBuffer buffer, boolean isKeyFrame, long presentationTimestampUs) {
			this.index = index;
			this.buffer = buffer;
			this.isKeyFrame = isKeyFrame;
			this.presentationTimestampUs = presentationTimestampUs;
		}

		public final int index;
		public final ByteBuffer buffer;
		public final boolean isKeyFrame;
		public final long presentationTimestampUs;

		@CalledByNative("OutputBufferInfo")
		int getIndex() {
			return index;
		}

		@CalledByNative("OutputBufferInfo")
		ByteBuffer getBuffer() {
			return buffer;
		}

		@CalledByNative("OutputBufferInfo")
		boolean isKeyFrame() {
			return isKeyFrame;
		}

		@CalledByNative("OutputBufferInfo")
		long getPresentationTimestampUs() {
			return presentationTimestampUs;
		}
	}

	// Dequeue and return an output buffer, or null if no output is ready.  Return
	// a fake OutputBufferInfo with index -1 if the codec is no longer operable.
	@CalledByNativeUnchecked
	OutputBufferInfo dequeueOutputBuffer() {
		checkOnMediaCodecThread();
		try {
			MediaCodec.BufferInfo info = new MediaCodec.BufferInfo();
			int result = mediaCodec.dequeueOutputBuffer(info, DEQUEUE_TIMEOUT);
			// Check if this is config frame and save configuration data.
			if (result >= 0) {
				boolean isConfigFrame = (info.flags & MediaCodec.BUFFER_FLAG_CODEC_CONFIG) != 0;
				if (isConfigFrame) {
					Logging.d(TAG, "Config frame generated. Offset: " + info.offset + ". Size: " + info.size);
					configData = ByteBuffer.allocateDirect(info.size);
					outputBuffers[result].position(info.offset);
					outputBuffers[result].limit(info.offset + info.size);
					configData.put(outputBuffers[result]);
					// Log few SPS header bytes to check profile and level.
					String spsData = "";
					for (int i = 0; i < (info.size < 8 ? info.size : 8); i++) {
						spsData += Integer.toHexString(configData.get(i) & 0xff) + " ";
					}
					Logging.d(TAG, spsData);
					// Release buffer back.
					mediaCodec.releaseOutputBuffer(result, false);
					// Query next output.
					result = mediaCodec.dequeueOutputBuffer(info, DEQUEUE_TIMEOUT);
				}
			}
			if (result >= 0) {
				// MediaCodec doesn't care about Buffer position/remaining/etc so we can
				// mess with them to get a slice and avoid having to pass extra
				// (BufferInfo-related) parameters back to C++.
				ByteBuffer outputBuffer = outputBuffers[result].duplicate();
				outputBuffer.position(info.offset);
				outputBuffer.limit(info.offset + info.size);
				reportEncodedFrame(info.size);

				// Check key frame flag.
				boolean isKeyFrame = (info.flags & MediaCodec.BUFFER_FLAG_SYNC_FRAME) != 0;
				if (isKeyFrame) {
					Logging.d(TAG, "Sync frame generated");
				}
				if (isKeyFrame && type == VideoCodecType.VIDEO_CODEC_H264) {
					Logging.d(TAG, "Appending config frame of size " + configData.capacity()
							+ " to output buffer with offset " + info.offset + ", size " + info.size);
					// For H.264 key frame append SPS and PPS NALs at the start
					ByteBuffer keyFrameBuffer = ByteBuffer.allocateDirect(configData.capacity() + info.size);
					configData.rewind();
					keyFrameBuffer.put(configData);
					keyFrameBuffer.put(outputBuffer);
					keyFrameBuffer.position(0);
					return new OutputBufferInfo(result, keyFrameBuffer, isKeyFrame, info.presentationTimeUs);
				} else {
					return new OutputBufferInfo(
							result, outputBuffer.slice(), isKeyFrame, info.presentationTimeUs);
				}
			} else if (result == MediaCodec.INFO_OUTPUT_BUFFERS_CHANGED) {
				outputBuffers = mediaCodec.getOutputBuffers();
				return dequeueOutputBuffer();
			} else if (result == MediaCodec.INFO_OUTPUT_FORMAT_CHANGED) {
				return dequeueOutputBuffer();
			} else if (result == MediaCodec.INFO_TRY_AGAIN_LATER) {
				return null;
			}
			throw new RuntimeException("dequeueOutputBuffer: " + result);
		} catch (IllegalStateException e) {
			Logging.e(TAG, "dequeueOutputBuffer failed", e);
			return new OutputBufferInfo(-1, null, false, -1);
		}
	}

	private double getBitrateScale(int bitrateAdjustmentScaleExp) {
		return Math.pow(BITRATE_CORRECTION_MAX_SCALE,
				(double) bitrateAdjustmentScaleExp / BITRATE_CORRECTION_STEPS);
	}

	private void reportEncodedFrame(int size) {
		if (targetFps == 0 || bitrateAdjustmentType != BitrateAdjustmentType.DYNAMIC_ADJUSTMENT) {
			return;
		}

		// Accumulate the difference between actial and expected frame sizes.
		double expectedBytesPerFrame = targetBitrateBps / (8.0 * targetFps);
		bitrateAccumulator += (size - expectedBytesPerFrame);
		bitrateObservationTimeMs += 1000.0 / targetFps;

		// Put a cap on the accumulator, i.e., don't let it grow beyond some level to avoid
		// using too old data for bitrate adjustment.
		double bitrateAccumulatorCap = BITRATE_CORRECTION_SEC * bitrateAccumulatorMax;
		bitrateAccumulator = Math.min(bitrateAccumulator, bitrateAccumulatorCap);
		bitrateAccumulator = Math.max(bitrateAccumulator, -bitrateAccumulatorCap);

		// Do bitrate adjustment every 3 seconds if actual encoder bitrate deviates too much
		// form the target value.
		if (bitrateObservationTimeMs > 1000 * BITRATE_CORRECTION_SEC) {
			Logging.d(TAG, "Acc: " + (int) bitrateAccumulator + ". Max: " + (int) bitrateAccumulatorMax
					+ ". ExpScale: " + bitrateAdjustmentScaleExp);
			boolean bitrateAdjustmentScaleChanged = false;
			if (bitrateAccumulator > bitrateAccumulatorMax) {
				// Encoder generates too high bitrate - need to reduce the scale.
				int bitrateAdjustmentInc = (int) (bitrateAccumulator / bitrateAccumulatorMax + 0.5);
				bitrateAdjustmentScaleExp -= bitrateAdjustmentInc;
				bitrateAccumulator = bitrateAccumulatorMax;
				bitrateAdjustmentScaleChanged = true;
			} else if (bitrateAccumulator < -bitrateAccumulatorMax) {
				// Encoder generates too low bitrate - need to increase the scale.
				int bitrateAdjustmentInc = (int) (-bitrateAccumulator / bitrateAccumulatorMax + 0.5);
				bitrateAdjustmentScaleExp += bitrateAdjustmentInc;
				bitrateAccumulator = -bitrateAccumulatorMax;
				bitrateAdjustmentScaleChanged = true;
			}
			if (bitrateAdjustmentScaleChanged) {
				bitrateAdjustmentScaleExp = Math.min(bitrateAdjustmentScaleExp, BITRATE_CORRECTION_STEPS);
				bitrateAdjustmentScaleExp = Math.max(bitrateAdjustmentScaleExp, -BITRATE_CORRECTION_STEPS);
				Logging.d(TAG, "Adjusting bitrate scale to " + bitrateAdjustmentScaleExp + ". Value: "
						+ getBitrateScale(bitrateAdjustmentScaleExp));
				setRates(targetBitrateBps / 1000, targetFps);
			}
			bitrateObservationTimeMs = 0;
		}
	}

	// Release a dequeued output buffer back to the codec for re-use.  Return
	// false if the codec is no longer operable.
	@CalledByNativeUnchecked
	boolean releaseOutputBuffer(int index) {
		checkOnMediaCodecThread();
		try {
			mediaCodec.releaseOutputBuffer(index, false);
			return true;
		} catch (IllegalStateException e) {
			Logging.e(TAG, "releaseOutputBuffer failed", e);
			return false;
		}
	}

	@CalledByNative
	int getColorFormat() {
		return colorFormat;
	}

	@CalledByNative
	static boolean isTextureBuffer(VideoFrame.Buffer buffer) {
		return buffer instanceof VideoFrame.TextureBuffer;
	}

	/** Fills an inputBuffer with the given index with data from the byte buffers. */
	private static native void nativeFillInputBuffer(long encoder, int inputBuffer, ByteBuffer dataY,
	                                                 int strideY, ByteBuffer dataU, int strideU, ByteBuffer dataV, int strideV);
}
