import android.content.Context;
import android.graphics.ImageFormat;
import android.graphics.SurfaceTexture;
-import android.graphics.SurfaceTexture.OnFrameAvailableListener;
import android.hardware.Camera;
import android.hardware.Camera.PreviewCallback;
+import android.hardware.Camera.Size;
import android.opengl.GLES20;
import android.util.Log;
import android.view.Surface;
import org.chromium.base.JNINamespace;
import java.io.IOException;
-import java.util.Iterator;
+import java.util.ArrayList;
import java.util.List;
import java.util.concurrent.locks.ReentrantLock;
* are invoked by this owner, including the callback OnPreviewFrame().
**/
@JNINamespace("media")
-public class VideoCapture implements PreviewCallback, OnFrameAvailableListener {
- static class CaptureCapability {
+public class VideoCapture implements PreviewCallback {
+ static class CaptureFormat {
+ public CaptureFormat(
+ int width, int height, int framerate, int pixelformat) {
+ mWidth = width;
+ mHeight = height;
+ mFramerate = framerate;
+ mPixelFormat = pixelformat;
+ }
public int mWidth;
public int mHeight;
- public int mDesiredFps;
+ public final int mFramerate;
+ public final int mPixelFormat;
+ @CalledByNative("CaptureFormat")
+ public int getWidth() {
+ return mWidth;
+ }
+ @CalledByNative("CaptureFormat")
+ public int getHeight() {
+ return mHeight;
+ }
+ @CalledByNative("CaptureFormat")
+ public int getFramerate() {
+ return mFramerate;
+ }
+ @CalledByNative("CaptureFormat")
+ public int getPixelFormat() {
+ return mPixelFormat;
+ }
}
// Some devices don't support YV12 format correctly, even with JELLY_BEAN or
"ODROID-U2",
};
- static void applyMinDimensions(CaptureCapability capability) {
+ static void applyMinDimensions(CaptureFormat format) {
// NOTE: this can discard requested aspect ratio considerations.
for (IdAndSizes buggyDevice : s_CAPTURESIZE_BUGGY_DEVICE_LIST) {
if (buggyDevice.mModel.contentEquals(android.os.Build.MODEL) &&
buggyDevice.mDevice.contentEquals(android.os.Build.DEVICE)) {
- capability.mWidth = (buggyDevice.mMinWidth > capability.mWidth)
+ format.mWidth = (buggyDevice.mMinWidth > format.mWidth)
? buggyDevice.mMinWidth
- : capability.mWidth;
- capability.mHeight = (buggyDevice.mMinHeight > capability.mHeight)
+ : format.mWidth;
+ format.mHeight = (buggyDevice.mMinHeight > format.mHeight)
? buggyDevice.mMinHeight
- : capability.mHeight;
+ : format.mHeight;
}
}
}
private Camera mCamera;
public ReentrantLock mPreviewBufferLock = new ReentrantLock();
- private int mImageFormat = ImageFormat.YV12;
- private byte[] mColorPlane = null;
private Context mContext = null;
// True when native code has started capture.
private boolean mIsRunning = false;
private int mCameraFacing = 0;
private int mDeviceOrientation = 0;
- CaptureCapability mCurrentCapability = null;
+ CaptureFormat mCaptureFormat = null;
private static final String TAG = "VideoCapture";
@CalledByNative
return new VideoCapture(context, id, nativeVideoCaptureDeviceAndroid);
}
+ @CalledByNative
+ public static CaptureFormat[] getDeviceSupportedFormats(int id) {
+ Camera camera;
+ try {
+ camera = Camera.open(id);
+ } catch (RuntimeException ex) {
+ Log.e(TAG, "Camera.open: " + ex);
+ return null;
+ }
+ Camera.Parameters parameters = camera.getParameters();
+
+ ArrayList<CaptureFormat> formatList = new ArrayList<CaptureFormat>();
+ // getSupportedPreview{Formats,FpsRange,PreviewSizes}() returns Lists
+ // with at least one element, but when the camera is in bad state, they
+ // can return null pointers; in that case we use a 0 entry, so we can
+ // retrieve as much information as possible.
+ List<Integer> pixelFormats = parameters.getSupportedPreviewFormats();
+ if (pixelFormats == null) {
+ pixelFormats = new ArrayList<Integer>();
+ }
+ if (pixelFormats.size() == 0) {
+ pixelFormats.add(ImageFormat.UNKNOWN);
+ }
+ for (Integer previewFormat : pixelFormats) {
+ int pixelFormat =
+ AndroidImageFormatList.ANDROID_IMAGEFORMAT_UNKNOWN;
+ if (previewFormat == ImageFormat.YV12) {
+ pixelFormat = AndroidImageFormatList.ANDROID_IMAGEFORMAT_YV12;
+ } else if (previewFormat == ImageFormat.NV21) {
+ continue;
+ }
+
+ List<int[]> listFpsRange = parameters.getSupportedPreviewFpsRange();
+ if (listFpsRange == null) {
+ listFpsRange = new ArrayList<int[]>();
+ }
+ if (listFpsRange.size() == 0) {
+ listFpsRange.add(new int[] {0, 0});
+ }
+ for (int[] fpsRange : listFpsRange) {
+ List<Camera.Size> supportedSizes =
+ parameters.getSupportedPreviewSizes();
+ if (supportedSizes == null) {
+ supportedSizes = new ArrayList<Camera.Size>();
+ }
+ if (supportedSizes.size() == 0) {
+ supportedSizes.add(camera.new Size(0, 0));
+ }
+ for (Camera.Size size : supportedSizes) {
+ formatList.add(new CaptureFormat(size.width, size.height,
+ (fpsRange[0] + 999 ) / 1000, pixelFormat));
+ }
+ }
+ }
+ camera.release();
+ return formatList.toArray(new CaptureFormat[formatList.size()]);
+ }
+
public VideoCapture(
Context context, int id, long nativeVideoCaptureDeviceAndroid) {
mContext = context;
// Returns true on success, false otherwise.
@CalledByNative
public boolean allocate(int width, int height, int frameRate) {
- Log.d(TAG, "allocate: requested width=" + width +
- ", height=" + height + ", frameRate=" + frameRate);
+ Log.d(TAG, "allocate: requested (" + width + "x" + height + ")@" +
+ frameRate + "fps");
try {
mCamera = Camera.open(mId);
} catch (RuntimeException ex) {
- Log.e(TAG, "allocate:Camera.open: " + ex);
+ Log.e(TAG, "allocate: Camera.open: " + ex);
return false;
}
- try {
- Camera.CameraInfo cameraInfo = new Camera.CameraInfo();
- Camera.getCameraInfo(mId, cameraInfo);
- mCameraOrientation = cameraInfo.orientation;
- mCameraFacing = cameraInfo.facing;
- mDeviceOrientation = getDeviceOrientation();
- Log.d(TAG, "allocate: device orientation=" + mDeviceOrientation +
- ", camera orientation=" + mCameraOrientation +
- ", facing=" + mCameraFacing);
-
- Camera.Parameters parameters = mCamera.getParameters();
-
- // Calculate fps.
- List<int[]> listFpsRange = parameters.getSupportedPreviewFpsRange();
- if (listFpsRange == null || listFpsRange.size() == 0) {
- Log.e(TAG, "allocate: no fps range found");
- return false;
- }
- int frameRateInMs = frameRate * 1000;
- Iterator itFpsRange = listFpsRange.iterator();
- int[] fpsRange = (int[]) itFpsRange.next();
- // Use the first range as default.
- int fpsMin = fpsRange[0];
- int fpsMax = fpsRange[1];
- int newFrameRate = (fpsMin + 999) / 1000;
- while (itFpsRange.hasNext()) {
- fpsRange = (int[]) itFpsRange.next();
- if (fpsRange[0] <= frameRateInMs &&
- frameRateInMs <= fpsRange[1]) {
- fpsMin = fpsRange[0];
- fpsMax = fpsRange[1];
- newFrameRate = frameRate;
- break;
- }
- }
- frameRate = newFrameRate;
- Log.d(TAG, "allocate: fps set to " + frameRate);
-
- mCurrentCapability = new CaptureCapability();
- mCurrentCapability.mDesiredFps = frameRate;
-
- // Calculate size.
- List<Camera.Size> listCameraSize =
- parameters.getSupportedPreviewSizes();
- int minDiff = Integer.MAX_VALUE;
- int matchedWidth = width;
- int matchedHeight = height;
- Iterator itCameraSize = listCameraSize.iterator();
- while (itCameraSize.hasNext()) {
- Camera.Size size = (Camera.Size) itCameraSize.next();
- int diff = Math.abs(size.width - width) +
- Math.abs(size.height - height);
- Log.d(TAG, "allocate: support resolution (" +
- size.width + ", " + size.height + "), diff=" + diff);
- // TODO(wjia): Remove this hack (forcing width to be multiple
- // of 32) by supporting stride in video frame buffer.
- // Right now, VideoCaptureController requires compact YV12
- // (i.e., with no padding).
- if (diff < minDiff && (size.width % 32 == 0)) {
- minDiff = diff;
- matchedWidth = size.width;
- matchedHeight = size.height;
- }
- }
- if (minDiff == Integer.MAX_VALUE) {
- Log.e(TAG, "allocate: can not find a resolution whose width " +
- "is multiple of 32");
- return false;
+ Camera.CameraInfo cameraInfo = new Camera.CameraInfo();
+ Camera.getCameraInfo(mId, cameraInfo);
+ mCameraOrientation = cameraInfo.orientation;
+ mCameraFacing = cameraInfo.facing;
+ mDeviceOrientation = getDeviceOrientation();
+ Log.d(TAG, "allocate: orientation dev=" + mDeviceOrientation +
+ ", cam=" + mCameraOrientation + ", facing=" + mCameraFacing);
+
+ Camera.Parameters parameters = mCamera.getParameters();
+
+ // getSupportedPreviewFpsRange() returns a List with at least one
+ // element, but when camera is in bad state, it can return null pointer.
+ List<int[]> listFpsRange = parameters.getSupportedPreviewFpsRange();
+ if (listFpsRange == null || listFpsRange.size() == 0) {
+ Log.e(TAG, "allocate: no fps range found");
+ return false;
+ }
+ int frameRateInMs = frameRate * 1000;
+ // Use the first range as default.
+ int[] fpsMinMax = listFpsRange.get(0);
+ int newFrameRate = (fpsMinMax[0] + 999) / 1000;
+ for (int[] fpsRange : listFpsRange) {
+ if (fpsRange[0] <= frameRateInMs && frameRateInMs <= fpsRange[1]) {
+ fpsMinMax = fpsRange;
+ newFrameRate = frameRate;
+ break;
}
- mCurrentCapability.mWidth = matchedWidth;
- mCurrentCapability.mHeight = matchedHeight;
- // Hack to avoid certain capture resolutions under a minimum one,
- // see http://crbug.com/305294
- BuggyDeviceHack.applyMinDimensions(mCurrentCapability);
-
- Log.d(TAG, "allocate: matched width=" + mCurrentCapability.mWidth +
- ", height=" + mCurrentCapability.mHeight);
-
- mImageFormat = BuggyDeviceHack.getImageFormat();
-
- if (parameters.isVideoStabilizationSupported()) {
- Log.d(TAG, "Image stabilization supported, currently: "
- + parameters.getVideoStabilization() + ", setting it.");
- parameters.setVideoStabilization(true);
- } else {
- Log.d(TAG, "Image stabilization not supported.");
+ }
+ frameRate = newFrameRate;
+ Log.d(TAG, "allocate: fps set to " + frameRate);
+
+ // Calculate size.
+ List<Camera.Size> listCameraSize =
+ parameters.getSupportedPreviewSizes();
+ int minDiff = Integer.MAX_VALUE;
+ int matchedWidth = width;
+ int matchedHeight = height;
+ for (Camera.Size size : listCameraSize) {
+ int diff = Math.abs(size.width - width) +
+ Math.abs(size.height - height);
+ Log.d(TAG, "allocate: supported (" +
+ size.width + ", " + size.height + "), diff=" + diff);
+ // TODO(wjia): Remove this hack (forcing width to be multiple
+ // of 32) by supporting stride in video frame buffer.
+ // Right now, VideoCaptureController requires compact YV12
+ // (i.e., with no padding).
+ if (diff < minDiff && (size.width % 32 == 0)) {
+ minDiff = diff;
+ matchedWidth = size.width;
+ matchedHeight = size.height;
}
+ }
+ if (minDiff == Integer.MAX_VALUE) {
+ Log.e(TAG, "allocate: can not find a multiple-of-32 resolution");
+ return false;
+ }
- parameters.setPreviewSize(mCurrentCapability.mWidth,
- mCurrentCapability.mHeight);
- parameters.setPreviewFormat(mImageFormat);
- parameters.setPreviewFpsRange(fpsMin, fpsMax);
- mCamera.setParameters(parameters);
-
- // Set SurfaceTexture.
- mGlTextures = new int[1];
- // Generate one texture pointer and bind it as an external texture.
- GLES20.glGenTextures(1, mGlTextures, 0);
- GLES20.glBindTexture(GL_TEXTURE_EXTERNAL_OES, mGlTextures[0]);
- // No mip-mapping with camera source.
- GLES20.glTexParameterf(GL_TEXTURE_EXTERNAL_OES,
- GLES20.GL_TEXTURE_MIN_FILTER, GLES20.GL_LINEAR);
- GLES20.glTexParameterf(GL_TEXTURE_EXTERNAL_OES,
- GLES20.GL_TEXTURE_MAG_FILTER, GLES20.GL_LINEAR);
- // Clamp to edge is only option.
- GLES20.glTexParameteri(GL_TEXTURE_EXTERNAL_OES,
- GLES20.GL_TEXTURE_WRAP_S, GLES20.GL_CLAMP_TO_EDGE);
- GLES20.glTexParameteri(GL_TEXTURE_EXTERNAL_OES,
- GLES20.GL_TEXTURE_WRAP_T, GLES20.GL_CLAMP_TO_EDGE);
-
- mSurfaceTexture = new SurfaceTexture(mGlTextures[0]);
- mSurfaceTexture.setOnFrameAvailableListener(null);
+ mCaptureFormat = new CaptureFormat(
+ matchedWidth, matchedHeight, frameRate,
+ BuggyDeviceHack.getImageFormat());
+ // Hack to avoid certain capture resolutions under a minimum one,
+ // see http://crbug.com/305294
+ BuggyDeviceHack.applyMinDimensions(mCaptureFormat);
+ Log.d(TAG, "allocate: matched (" + mCaptureFormat.mWidth + "x" +
+ mCaptureFormat.mHeight + ")");
+
+ if (parameters.isVideoStabilizationSupported()) {
+ Log.d(TAG, "Image stabilization supported, currently: "
+ + parameters.getVideoStabilization() + ", setting it.");
+ parameters.setVideoStabilization(true);
+ } else {
+ Log.d(TAG, "Image stabilization not supported.");
+ }
+ parameters.setPreviewSize(mCaptureFormat.mWidth,
+ mCaptureFormat.mHeight);
+ parameters.setPreviewFormat(mCaptureFormat.mPixelFormat);
+ parameters.setPreviewFpsRange(fpsMinMax[0], fpsMinMax[1]);
+ mCamera.setParameters(parameters);
+
+ // Set SurfaceTexture. Android Capture needs a SurfaceTexture even if
+ // it is not going to be used.
+ mGlTextures = new int[1];
+ // Generate one texture pointer and bind it as an external texture.
+ GLES20.glGenTextures(1, mGlTextures, 0);
+ GLES20.glBindTexture(GL_TEXTURE_EXTERNAL_OES, mGlTextures[0]);
+ // No mip-mapping with camera source.
+ GLES20.glTexParameterf(GL_TEXTURE_EXTERNAL_OES,
+ GLES20.GL_TEXTURE_MIN_FILTER, GLES20.GL_LINEAR);
+ GLES20.glTexParameterf(GL_TEXTURE_EXTERNAL_OES,
+ GLES20.GL_TEXTURE_MAG_FILTER, GLES20.GL_LINEAR);
+ // Clamp to edge is only option.
+ GLES20.glTexParameteri(GL_TEXTURE_EXTERNAL_OES,
+ GLES20.GL_TEXTURE_WRAP_S, GLES20.GL_CLAMP_TO_EDGE);
+ GLES20.glTexParameteri(GL_TEXTURE_EXTERNAL_OES,
+ GLES20.GL_TEXTURE_WRAP_T, GLES20.GL_CLAMP_TO_EDGE);
+
+ mSurfaceTexture = new SurfaceTexture(mGlTextures[0]);
+ mSurfaceTexture.setOnFrameAvailableListener(null);
+ try {
mCamera.setPreviewTexture(mSurfaceTexture);
-
- int bufSize = mCurrentCapability.mWidth *
- mCurrentCapability.mHeight *
- ImageFormat.getBitsPerPixel(mImageFormat) / 8;
- for (int i = 0; i < NUM_CAPTURE_BUFFERS; i++) {
- byte[] buffer = new byte[bufSize];
- mCamera.addCallbackBuffer(buffer);
- }
- mExpectedFrameSize = bufSize;
} catch (IOException ex) {
Log.e(TAG, "allocate: " + ex);
return false;
}
+ int bufSize = mCaptureFormat.mWidth *
+ mCaptureFormat.mHeight *
+ ImageFormat.getBitsPerPixel(
+ mCaptureFormat.mPixelFormat) / 8;
+ for (int i = 0; i < NUM_CAPTURE_BUFFERS; i++) {
+ byte[] buffer = new byte[bufSize];
+ mCamera.addCallbackBuffer(buffer);
+ }
+ mExpectedFrameSize = bufSize;
+
return true;
}
@CalledByNative
public int queryWidth() {
- return mCurrentCapability.mWidth;
+ return mCaptureFormat.mWidth;
}
@CalledByNative
public int queryHeight() {
- return mCurrentCapability.mHeight;
+ return mCaptureFormat.mHeight;
}
@CalledByNative
public int queryFrameRate() {
- return mCurrentCapability.mDesiredFps;
+ return mCaptureFormat.mFramerate;
}
@CalledByNative
public int getColorspace() {
- switch (mImageFormat) {
+ switch (mCaptureFormat.mPixelFormat) {
case ImageFormat.YV12:
return AndroidImageFormatList.ANDROID_IMAGEFORMAT_YV12;
case ImageFormat.NV21:
return AndroidImageFormatList.ANDROID_IMAGEFORMAT_NV21;
- case ImageFormat.YUY2:
- return AndroidImageFormatList.ANDROID_IMAGEFORMAT_YUY2;
- case ImageFormat.NV16:
- return AndroidImageFormatList.ANDROID_IMAGEFORMAT_NV16;
- case ImageFormat.JPEG:
- return AndroidImageFormatList.ANDROID_IMAGEFORMAT_JPEG;
- case ImageFormat.RGB_565:
- return AndroidImageFormatList.ANDROID_IMAGEFORMAT_RGB_565;
case ImageFormat.UNKNOWN:
default:
return AndroidImageFormatList.ANDROID_IMAGEFORMAT_UNKNOWN;
mCamera.setPreviewTexture(null);
if (mGlTextures != null)
GLES20.glDeleteTextures(1, mGlTextures, 0);
- mCurrentCapability = null;
+ mCaptureFormat = null;
mCamera.release();
mCamera = null;
} catch (IOException ex) {
}
// TODO(wjia): investigate whether reading from texture could give better
- // performance and frame rate.
- @Override
- public void onFrameAvailable(SurfaceTexture surfaceTexture) { }
+ // performance and frame rate, using onFrameAvailable().
private static class ChromiumCameraInfo {
private final int mId;