Loading core/java/android/hardware/camera2/CameraCharacteristics.java +113 −226 Original line number Diff line number Diff line Loading @@ -127,74 +127,58 @@ public final class CameraCharacteristics extends CameraMetadata { *~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~*/ /** * <p> * Which set of antibanding modes are * supported * </p> * <p>Which set of antibanding modes are * supported</p> */ public static final Key<byte[]> CONTROL_AE_AVAILABLE_ANTIBANDING_MODES = new Key<byte[]>("android.control.aeAvailableAntibandingModes", byte[].class); /** * <p> * List of frame rate ranges supported by the * AE algorithm/hardware * </p> * <p>List of frame rate ranges supported by the * AE algorithm/hardware</p> */ public static final Key<int[]> CONTROL_AE_AVAILABLE_TARGET_FPS_RANGES = new Key<int[]>("android.control.aeAvailableTargetFpsRanges", int[].class); /** * <p> * Maximum and minimum exposure compensation * <p>Maximum and minimum exposure compensation * setting, in counts of * android.control.aeCompensationStepSize * </p> * android.control.aeCompensationStepSize</p> */ public static final Key<int[]> CONTROL_AE_COMPENSATION_RANGE = new Key<int[]>("android.control.aeCompensationRange", int[].class); /** * <p> * Smallest step by which exposure compensation * can be changed * </p> * <p>Smallest step by which exposure compensation * can be changed</p> */ public static final Key<Rational> CONTROL_AE_COMPENSATION_STEP = new Key<Rational>("android.control.aeCompensationStep", Rational.class); /** * <p> * List of AF modes that can be * selected * </p> * <p>List of AF modes that can be * selected</p> */ public static final Key<byte[]> CONTROL_AF_AVAILABLE_MODES = new Key<byte[]>("android.control.afAvailableModes", byte[].class); /** * <p> * what subset of the full color effect enum * list is supported * </p> * <p>what subset of the full color effect enum * list is supported</p> */ public static final Key<byte[]> CONTROL_AVAILABLE_EFFECTS = new Key<byte[]>("android.control.availableEffects", byte[].class); /** * <p> * what subset of the scene mode enum list is * supported. * </p> * <p>what subset of the scene mode enum list is * supported.</p> */ public static final Key<byte[]> CONTROL_AVAILABLE_SCENE_MODES = new Key<byte[]>("android.control.availableSceneModes", byte[].class); /** * <p> * List of video stabilization modes that can * be supported * </p> * <p>List of video stabilization modes that can * be supported</p> */ public static final Key<byte[]> CONTROL_AVAILABLE_VIDEO_STABILIZATION_MODES = new Key<byte[]>("android.control.availableVideoStabilizationModes", byte[].class); Loading @@ -205,127 +189,93 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<byte[]>("android.control.awbAvailableModes", byte[].class); /** * <p> * For AE, AWB, and AF, how many individual * regions can be listed for metering? * </p> * <p>For AE, AWB, and AF, how many individual * regions can be listed for metering?</p> */ public static final Key<Integer> CONTROL_MAX_REGIONS = new Key<Integer>("android.control.maxRegions", int.class); /** * <p> * Whether this camera has a * flash * </p> * <p> * If no flash, none of the flash controls do * anything. All other metadata should return 0 * </p> * <p>Whether this camera has a * flash</p> * <p>If no flash, none of the flash controls do * anything. All other metadata should return 0</p> */ public static final Key<Byte> FLASH_INFO_AVAILABLE = new Key<Byte>("android.flash.info.available", byte.class); /** * <p> * Supported resolutions for the JPEG * thumbnail * </p> * <p>Supported resolutions for the JPEG * thumbnail</p> */ public static final Key<android.hardware.camera2.Size[]> JPEG_AVAILABLE_THUMBNAIL_SIZES = new Key<android.hardware.camera2.Size[]>("android.jpeg.availableThumbnailSizes", android.hardware.camera2.Size[].class); /** * <p> * List of supported aperture * values * </p> * <p> * If variable aperture not available, only setting * should be for the fixed aperture * </p> * <p>List of supported aperture * values</p> * <p>If variable aperture not available, only setting * should be for the fixed aperture</p> */ public static final Key<float[]> LENS_INFO_AVAILABLE_APERTURES = new Key<float[]>("android.lens.info.availableApertures", float[].class); /** * <p> * List of supported ND filter * values * </p> * <p> * If not available, only setting is 0. Otherwise, * <p>List of supported ND filter * values</p> * <p>If not available, only setting is 0. Otherwise, * lists the available exposure index values for dimming * (2 would mean the filter is set to reduce incoming * light by two stops) * </p> * light by two stops)</p> */ public static final Key<float[]> LENS_INFO_AVAILABLE_FILTER_DENSITIES = new Key<float[]>("android.lens.info.availableFilterDensities", float[].class); /** * <p> * If fitted with optical zoom, what focal * <p>If fitted with optical zoom, what focal * lengths are available. If not, the static focal * length * </p> * <p> * If optical zoom not supported, only one value * should be reported * </p> * length</p> * <p>If optical zoom not supported, only one value * should be reported</p> */ public static final Key<float[]> LENS_INFO_AVAILABLE_FOCAL_LENGTHS = new Key<float[]>("android.lens.info.availableFocalLengths", float[].class); /** * <p> * List of supported optical image * stabilization modes * </p> * <p>List of supported optical image * stabilization modes</p> */ public static final Key<byte[]> LENS_INFO_AVAILABLE_OPTICAL_STABILIZATION = new Key<byte[]>("android.lens.info.availableOpticalStabilization", byte[].class); /** * <p> * Hyperfocal distance for this lens; set to * 0 if fixed focus * </p> * <p> * The hyperfocal distance is used for the old * API's 'fixed' setting * </p> * <p>Hyperfocal distance for this lens; set to * 0 if fixed focus</p> * <p>The hyperfocal distance is used for the old * API's 'fixed' setting</p> */ public static final Key<Float> LENS_INFO_HYPERFOCAL_DISTANCE = new Key<Float>("android.lens.info.hyperfocalDistance", float.class); /** * <p> * Shortest distance from frontmost surface * of the lens that can be focused correctly * </p> * <p> * If the lens is fixed-focus, this should be * 0 * </p> * <p>Shortest distance from frontmost surface * of the lens that can be focused correctly</p> * <p>If the lens is fixed-focus, this should be * 0</p> */ public static final Key<Float> LENS_INFO_MINIMUM_FOCUS_DISTANCE = new Key<Float>("android.lens.info.minimumFocusDistance", float.class); /** * <p> * Dimensions of lens shading * map * </p> * <p>Dimensions of lens shading * map</p> */ public static final Key<android.hardware.camera2.Size> LENS_INFO_SHADING_MAP_SIZE = new Key<android.hardware.camera2.Size>("android.lens.info.shadingMapSize", android.hardware.camera2.Size.class); /** * <p> * Direction the camera faces relative to * device screen * </p> * <p>Direction the camera faces relative to * device screen</p> * @see #LENS_FACING_FRONT * @see #LENS_FACING_BACK */ Loading @@ -333,18 +283,14 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Integer>("android.lens.facing", int.class); /** * <p> * If set to 1, the HAL will always split result * <p>If set to 1, the HAL will always split result * metadata for a single capture into multiple buffers, * returned using multiple process_capture_result calls. * </p> * <p> * Does not need to be listed in static * returned using multiple process_capture_result calls.</p> * <p>Does not need to be listed in static * metadata. Support for partial results will be reworked in * future versions of camera service. This quirk will stop * working at that point; DO NOT USE without careful * consideration of future support. * </p> * consideration of future support.</p> * * <b>Optional</b> - This value may be null on some devices. * Loading @@ -354,160 +300,122 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Byte>("android.quirks.usePartialResult", byte.class); /** * <p> * How many output streams can be allocated at * the same time for each type of stream * </p> * <p> * Video snapshot with preview callbacks requires 3 * <p>How many output streams can be allocated at * the same time for each type of stream</p> * <p>Video snapshot with preview callbacks requires 3 * processed streams (preview, record, app callbacks) and * one JPEG stream (snapshot) * </p> * one JPEG stream (snapshot)</p> */ public static final Key<int[]> REQUEST_MAX_NUM_OUTPUT_STREAMS = new Key<int[]>("android.request.maxNumOutputStreams", int[].class); /** * <p> * List of app-visible formats * </p> * <p>List of app-visible formats</p> */ public static final Key<int[]> SCALER_AVAILABLE_FORMATS = new Key<int[]>("android.scaler.availableFormats", int[].class); /** * <p> * The minimum frame duration that is supported * <p>The minimum frame duration that is supported * for each resolution in availableJpegSizes. Should * correspond to the frame duration when only that JPEG * stream is active and captured in a burst, with all * processing set to FAST * </p> * <p> * When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations) * </p> * processing set to FAST</p> * <p>When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations)</p> */ public static final Key<long[]> SCALER_AVAILABLE_JPEG_MIN_DURATIONS = new Key<long[]>("android.scaler.availableJpegMinDurations", long[].class); /** * <p> * The resolutions available for output from * the JPEG block. Listed as width x height * </p> * <p>The resolutions available for output from * the JPEG block. Listed as width x height</p> */ public static final Key<android.hardware.camera2.Size[]> SCALER_AVAILABLE_JPEG_SIZES = new Key<android.hardware.camera2.Size[]>("android.scaler.availableJpegSizes", android.hardware.camera2.Size[].class); /** * <p> * The maximum ratio between active area width * <p>The maximum ratio between active area width * and crop region width, or between active area height and * crop region height, if the crop region height is larger * than width * </p> * than width</p> */ public static final Key<Float> SCALER_AVAILABLE_MAX_DIGITAL_ZOOM = new Key<Float>("android.scaler.availableMaxDigitalZoom", float.class); /** * <p> * The minimum frame duration that is supported * <p>The minimum frame duration that is supported * for each resolution in availableProcessedSizes. Should * correspond to the frame duration when only that processed * stream is active, with all processing set to * FAST * </p> * <p> * When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations) * </p> * FAST</p> * <p>When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations)</p> */ public static final Key<long[]> SCALER_AVAILABLE_PROCESSED_MIN_DURATIONS = new Key<long[]>("android.scaler.availableProcessedMinDurations", long[].class); /** * <p> * The resolutions available for use with * <p>The resolutions available for use with * processed output streams, such as YV12, NV12, and * platform opaque YUV/RGB streams to the GPU or video * encoders. Listed as width, height * </p> * <p> * The actual supported resolution list may be limited by * encoders. Listed as width, height</p> * <p>The actual supported resolution list may be limited by * consumer end points for different use cases. For example, for * recording use case, the largest supported resolution may be * limited by max supported size from encoder, for preview use * case, the largest supported resolution may be limited by max * resolution SurfaceTexture/SurfaceView can support. * </p> * resolution SurfaceTexture/SurfaceView can support.</p> */ public static final Key<android.hardware.camera2.Size[]> SCALER_AVAILABLE_PROCESSED_SIZES = new Key<android.hardware.camera2.Size[]>("android.scaler.availableProcessedSizes", android.hardware.camera2.Size[].class); /** * <p> * Area of raw data which corresponds to only * <p>Area of raw data which corresponds to only * active pixels; smaller or equal to * pixelArraySize. * </p> * pixelArraySize.</p> */ public static final Key<android.graphics.Rect> SENSOR_INFO_ACTIVE_ARRAY_SIZE = new Key<android.graphics.Rect>("android.sensor.info.activeArraySize", android.graphics.Rect.class); /** * <p> * Range of valid sensitivities * </p> * <p>Range of valid sensitivities</p> */ public static final Key<int[]> SENSOR_INFO_SENSITIVITY_RANGE = new Key<int[]>("android.sensor.info.sensitivityRange", int[].class); /** * <p> * Range of valid exposure * times * </p> * <p>Range of valid exposure * times</p> */ public static final Key<long[]> SENSOR_INFO_EXPOSURE_TIME_RANGE = new Key<long[]>("android.sensor.info.exposureTimeRange", long[].class); /** * <p> * Maximum possible frame duration (minimum frame * rate) * </p> * <p> * Minimum duration is a function of resolution, * <p>Maximum possible frame duration (minimum frame * rate)</p> * <p>Minimum duration is a function of resolution, * processing settings. See * android.scaler.availableProcessedMinDurations * android.scaler.availableJpegMinDurations * android.scaler.availableRawMinDurations * </p> * android.scaler.availableRawMinDurations</p> */ public static final Key<Long> SENSOR_INFO_MAX_FRAME_DURATION = new Key<Long>("android.sensor.info.maxFrameDuration", long.class); /** * <p> * The physical dimensions of the full pixel * array * </p> * <p> * Needed for FOV calculation for old API * </p> * <p>The physical dimensions of the full pixel * array</p> * <p>Needed for FOV calculation for old API</p> */ public static final Key<float[]> SENSOR_INFO_PHYSICAL_SIZE = new Key<float[]>("android.sensor.info.physicalSize", float[].class); /** * <p> * Gain factor from electrons to raw units when * ISO=100 * </p> * <p>Gain factor from electrons to raw units when * ISO=100</p> * * <b>Optional</b> - This value may be null on some devices. * Loading @@ -519,16 +427,12 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Rational>("android.sensor.baseGainFactor", Rational.class); /** * <p> * Maximum sensitivity that is implemented * purely through analog gain * </p> * <p> * For android.sensor.sensitivity values less than or * <p>Maximum sensitivity that is implemented * purely through analog gain</p> * <p>For android.sensor.sensitivity values less than or * equal to this, all applied gain must be analog. For * values above this, it can be a mix of analog and * digital * </p> * digital</p> * * <b>Optional</b> - This value may be null on some devices. * Loading @@ -540,60 +444,46 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Integer>("android.sensor.maxAnalogSensitivity", int.class); /** * <p> * Clockwise angle through which the output * <p>Clockwise angle through which the output * image needs to be rotated to be upright on the device * screen in its native orientation. Also defines the * direction of rolling shutter readout, which is from top * to bottom in the sensor's coordinate system * </p> * to bottom in the sensor's coordinate system</p> */ public static final Key<Integer> SENSOR_ORIENTATION = new Key<Integer>("android.sensor.orientation", int.class); /** * <p> * Which face detection modes are available, * if any * </p> * <p> * OFF means face detection is disabled, it must * be included in the list. * </p><p> * SIMPLE means the device supports the * <p>Which face detection modes are available, * if any</p> * <p>OFF means face detection is disabled, it must * be included in the list.</p> * <p>SIMPLE means the device supports the * android.statistics.faceRectangles and * android.statistics.faceScores outputs. * </p><p> * FULL means the device additionally supports the * android.statistics.faceScores outputs.</p> * <p>FULL means the device additionally supports the * android.statistics.faceIds and * android.statistics.faceLandmarks outputs. * </p> * android.statistics.faceLandmarks outputs.</p> */ public static final Key<byte[]> STATISTICS_INFO_AVAILABLE_FACE_DETECT_MODES = new Key<byte[]>("android.statistics.info.availableFaceDetectModes", byte[].class); /** * <p> * Maximum number of simultaneously detectable * faces * </p> * <p>Maximum number of simultaneously detectable * faces</p> */ public static final Key<Integer> STATISTICS_INFO_MAX_FACE_COUNT = new Key<Integer>("android.statistics.info.maxFaceCount", int.class); /** * <p> * Maximum number of supported points in the * tonemap curve * </p> * <p>Maximum number of supported points in the * tonemap curve</p> */ public static final Key<Integer> TONEMAP_MAX_CURVE_POINTS = new Key<Integer>("android.tonemap.maxCurvePoints", int.class); /** * <p> * A list of camera LEDs that are available on this system. * </p> * <p>A list of camera LEDs that are available on this system.</p> * @see #LED_AVAILABLE_LEDS_TRANSMIT * * @hide Loading @@ -602,17 +492,14 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<int[]>("android.led.availableLeds", int[].class); /** * <p> * The camera 3 HAL device can implement one of two possible * <p>The camera 3 HAL device can implement one of two possible * operational modes; limited and full. Full support is * expected from new higher-end devices. Limited mode has * hardware requirements roughly in line with those for a * camera HAL device v1 implementation, and is expected from * older or inexpensive devices. Full is a strict superset of * limited, and they share the same essential operational flow. * </p><p> * For full details refer to "S3. Operational Modes" in camera3.h * </p> * limited, and they share the same essential operational flow.</p> * <p>For full details refer to "S3. Operational Modes" in camera3.h</p> * @see #INFO_SUPPORTED_HARDWARE_LEVEL_LIMITED * @see #INFO_SUPPORTED_HARDWARE_LEVEL_FULL */ Loading Loading
core/java/android/hardware/camera2/CameraCharacteristics.java +113 −226 Original line number Diff line number Diff line Loading @@ -127,74 +127,58 @@ public final class CameraCharacteristics extends CameraMetadata { *~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~*/ /** * <p> * Which set of antibanding modes are * supported * </p> * <p>Which set of antibanding modes are * supported</p> */ public static final Key<byte[]> CONTROL_AE_AVAILABLE_ANTIBANDING_MODES = new Key<byte[]>("android.control.aeAvailableAntibandingModes", byte[].class); /** * <p> * List of frame rate ranges supported by the * AE algorithm/hardware * </p> * <p>List of frame rate ranges supported by the * AE algorithm/hardware</p> */ public static final Key<int[]> CONTROL_AE_AVAILABLE_TARGET_FPS_RANGES = new Key<int[]>("android.control.aeAvailableTargetFpsRanges", int[].class); /** * <p> * Maximum and minimum exposure compensation * <p>Maximum and minimum exposure compensation * setting, in counts of * android.control.aeCompensationStepSize * </p> * android.control.aeCompensationStepSize</p> */ public static final Key<int[]> CONTROL_AE_COMPENSATION_RANGE = new Key<int[]>("android.control.aeCompensationRange", int[].class); /** * <p> * Smallest step by which exposure compensation * can be changed * </p> * <p>Smallest step by which exposure compensation * can be changed</p> */ public static final Key<Rational> CONTROL_AE_COMPENSATION_STEP = new Key<Rational>("android.control.aeCompensationStep", Rational.class); /** * <p> * List of AF modes that can be * selected * </p> * <p>List of AF modes that can be * selected</p> */ public static final Key<byte[]> CONTROL_AF_AVAILABLE_MODES = new Key<byte[]>("android.control.afAvailableModes", byte[].class); /** * <p> * what subset of the full color effect enum * list is supported * </p> * <p>what subset of the full color effect enum * list is supported</p> */ public static final Key<byte[]> CONTROL_AVAILABLE_EFFECTS = new Key<byte[]>("android.control.availableEffects", byte[].class); /** * <p> * what subset of the scene mode enum list is * supported. * </p> * <p>what subset of the scene mode enum list is * supported.</p> */ public static final Key<byte[]> CONTROL_AVAILABLE_SCENE_MODES = new Key<byte[]>("android.control.availableSceneModes", byte[].class); /** * <p> * List of video stabilization modes that can * be supported * </p> * <p>List of video stabilization modes that can * be supported</p> */ public static final Key<byte[]> CONTROL_AVAILABLE_VIDEO_STABILIZATION_MODES = new Key<byte[]>("android.control.availableVideoStabilizationModes", byte[].class); Loading @@ -205,127 +189,93 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<byte[]>("android.control.awbAvailableModes", byte[].class); /** * <p> * For AE, AWB, and AF, how many individual * regions can be listed for metering? * </p> * <p>For AE, AWB, and AF, how many individual * regions can be listed for metering?</p> */ public static final Key<Integer> CONTROL_MAX_REGIONS = new Key<Integer>("android.control.maxRegions", int.class); /** * <p> * Whether this camera has a * flash * </p> * <p> * If no flash, none of the flash controls do * anything. All other metadata should return 0 * </p> * <p>Whether this camera has a * flash</p> * <p>If no flash, none of the flash controls do * anything. All other metadata should return 0</p> */ public static final Key<Byte> FLASH_INFO_AVAILABLE = new Key<Byte>("android.flash.info.available", byte.class); /** * <p> * Supported resolutions for the JPEG * thumbnail * </p> * <p>Supported resolutions for the JPEG * thumbnail</p> */ public static final Key<android.hardware.camera2.Size[]> JPEG_AVAILABLE_THUMBNAIL_SIZES = new Key<android.hardware.camera2.Size[]>("android.jpeg.availableThumbnailSizes", android.hardware.camera2.Size[].class); /** * <p> * List of supported aperture * values * </p> * <p> * If variable aperture not available, only setting * should be for the fixed aperture * </p> * <p>List of supported aperture * values</p> * <p>If variable aperture not available, only setting * should be for the fixed aperture</p> */ public static final Key<float[]> LENS_INFO_AVAILABLE_APERTURES = new Key<float[]>("android.lens.info.availableApertures", float[].class); /** * <p> * List of supported ND filter * values * </p> * <p> * If not available, only setting is 0. Otherwise, * <p>List of supported ND filter * values</p> * <p>If not available, only setting is 0. Otherwise, * lists the available exposure index values for dimming * (2 would mean the filter is set to reduce incoming * light by two stops) * </p> * light by two stops)</p> */ public static final Key<float[]> LENS_INFO_AVAILABLE_FILTER_DENSITIES = new Key<float[]>("android.lens.info.availableFilterDensities", float[].class); /** * <p> * If fitted with optical zoom, what focal * <p>If fitted with optical zoom, what focal * lengths are available. If not, the static focal * length * </p> * <p> * If optical zoom not supported, only one value * should be reported * </p> * length</p> * <p>If optical zoom not supported, only one value * should be reported</p> */ public static final Key<float[]> LENS_INFO_AVAILABLE_FOCAL_LENGTHS = new Key<float[]>("android.lens.info.availableFocalLengths", float[].class); /** * <p> * List of supported optical image * stabilization modes * </p> * <p>List of supported optical image * stabilization modes</p> */ public static final Key<byte[]> LENS_INFO_AVAILABLE_OPTICAL_STABILIZATION = new Key<byte[]>("android.lens.info.availableOpticalStabilization", byte[].class); /** * <p> * Hyperfocal distance for this lens; set to * 0 if fixed focus * </p> * <p> * The hyperfocal distance is used for the old * API's 'fixed' setting * </p> * <p>Hyperfocal distance for this lens; set to * 0 if fixed focus</p> * <p>The hyperfocal distance is used for the old * API's 'fixed' setting</p> */ public static final Key<Float> LENS_INFO_HYPERFOCAL_DISTANCE = new Key<Float>("android.lens.info.hyperfocalDistance", float.class); /** * <p> * Shortest distance from frontmost surface * of the lens that can be focused correctly * </p> * <p> * If the lens is fixed-focus, this should be * 0 * </p> * <p>Shortest distance from frontmost surface * of the lens that can be focused correctly</p> * <p>If the lens is fixed-focus, this should be * 0</p> */ public static final Key<Float> LENS_INFO_MINIMUM_FOCUS_DISTANCE = new Key<Float>("android.lens.info.minimumFocusDistance", float.class); /** * <p> * Dimensions of lens shading * map * </p> * <p>Dimensions of lens shading * map</p> */ public static final Key<android.hardware.camera2.Size> LENS_INFO_SHADING_MAP_SIZE = new Key<android.hardware.camera2.Size>("android.lens.info.shadingMapSize", android.hardware.camera2.Size.class); /** * <p> * Direction the camera faces relative to * device screen * </p> * <p>Direction the camera faces relative to * device screen</p> * @see #LENS_FACING_FRONT * @see #LENS_FACING_BACK */ Loading @@ -333,18 +283,14 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Integer>("android.lens.facing", int.class); /** * <p> * If set to 1, the HAL will always split result * <p>If set to 1, the HAL will always split result * metadata for a single capture into multiple buffers, * returned using multiple process_capture_result calls. * </p> * <p> * Does not need to be listed in static * returned using multiple process_capture_result calls.</p> * <p>Does not need to be listed in static * metadata. Support for partial results will be reworked in * future versions of camera service. This quirk will stop * working at that point; DO NOT USE without careful * consideration of future support. * </p> * consideration of future support.</p> * * <b>Optional</b> - This value may be null on some devices. * Loading @@ -354,160 +300,122 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Byte>("android.quirks.usePartialResult", byte.class); /** * <p> * How many output streams can be allocated at * the same time for each type of stream * </p> * <p> * Video snapshot with preview callbacks requires 3 * <p>How many output streams can be allocated at * the same time for each type of stream</p> * <p>Video snapshot with preview callbacks requires 3 * processed streams (preview, record, app callbacks) and * one JPEG stream (snapshot) * </p> * one JPEG stream (snapshot)</p> */ public static final Key<int[]> REQUEST_MAX_NUM_OUTPUT_STREAMS = new Key<int[]>("android.request.maxNumOutputStreams", int[].class); /** * <p> * List of app-visible formats * </p> * <p>List of app-visible formats</p> */ public static final Key<int[]> SCALER_AVAILABLE_FORMATS = new Key<int[]>("android.scaler.availableFormats", int[].class); /** * <p> * The minimum frame duration that is supported * <p>The minimum frame duration that is supported * for each resolution in availableJpegSizes. Should * correspond to the frame duration when only that JPEG * stream is active and captured in a burst, with all * processing set to FAST * </p> * <p> * When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations) * </p> * processing set to FAST</p> * <p>When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations)</p> */ public static final Key<long[]> SCALER_AVAILABLE_JPEG_MIN_DURATIONS = new Key<long[]>("android.scaler.availableJpegMinDurations", long[].class); /** * <p> * The resolutions available for output from * the JPEG block. Listed as width x height * </p> * <p>The resolutions available for output from * the JPEG block. Listed as width x height</p> */ public static final Key<android.hardware.camera2.Size[]> SCALER_AVAILABLE_JPEG_SIZES = new Key<android.hardware.camera2.Size[]>("android.scaler.availableJpegSizes", android.hardware.camera2.Size[].class); /** * <p> * The maximum ratio between active area width * <p>The maximum ratio between active area width * and crop region width, or between active area height and * crop region height, if the crop region height is larger * than width * </p> * than width</p> */ public static final Key<Float> SCALER_AVAILABLE_MAX_DIGITAL_ZOOM = new Key<Float>("android.scaler.availableMaxDigitalZoom", float.class); /** * <p> * The minimum frame duration that is supported * <p>The minimum frame duration that is supported * for each resolution in availableProcessedSizes. Should * correspond to the frame duration when only that processed * stream is active, with all processing set to * FAST * </p> * <p> * When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations) * </p> * FAST</p> * <p>When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations)</p> */ public static final Key<long[]> SCALER_AVAILABLE_PROCESSED_MIN_DURATIONS = new Key<long[]>("android.scaler.availableProcessedMinDurations", long[].class); /** * <p> * The resolutions available for use with * <p>The resolutions available for use with * processed output streams, such as YV12, NV12, and * platform opaque YUV/RGB streams to the GPU or video * encoders. Listed as width, height * </p> * <p> * The actual supported resolution list may be limited by * encoders. Listed as width, height</p> * <p>The actual supported resolution list may be limited by * consumer end points for different use cases. For example, for * recording use case, the largest supported resolution may be * limited by max supported size from encoder, for preview use * case, the largest supported resolution may be limited by max * resolution SurfaceTexture/SurfaceView can support. * </p> * resolution SurfaceTexture/SurfaceView can support.</p> */ public static final Key<android.hardware.camera2.Size[]> SCALER_AVAILABLE_PROCESSED_SIZES = new Key<android.hardware.camera2.Size[]>("android.scaler.availableProcessedSizes", android.hardware.camera2.Size[].class); /** * <p> * Area of raw data which corresponds to only * <p>Area of raw data which corresponds to only * active pixels; smaller or equal to * pixelArraySize. * </p> * pixelArraySize.</p> */ public static final Key<android.graphics.Rect> SENSOR_INFO_ACTIVE_ARRAY_SIZE = new Key<android.graphics.Rect>("android.sensor.info.activeArraySize", android.graphics.Rect.class); /** * <p> * Range of valid sensitivities * </p> * <p>Range of valid sensitivities</p> */ public static final Key<int[]> SENSOR_INFO_SENSITIVITY_RANGE = new Key<int[]>("android.sensor.info.sensitivityRange", int[].class); /** * <p> * Range of valid exposure * times * </p> * <p>Range of valid exposure * times</p> */ public static final Key<long[]> SENSOR_INFO_EXPOSURE_TIME_RANGE = new Key<long[]>("android.sensor.info.exposureTimeRange", long[].class); /** * <p> * Maximum possible frame duration (minimum frame * rate) * </p> * <p> * Minimum duration is a function of resolution, * <p>Maximum possible frame duration (minimum frame * rate)</p> * <p>Minimum duration is a function of resolution, * processing settings. See * android.scaler.availableProcessedMinDurations * android.scaler.availableJpegMinDurations * android.scaler.availableRawMinDurations * </p> * android.scaler.availableRawMinDurations</p> */ public static final Key<Long> SENSOR_INFO_MAX_FRAME_DURATION = new Key<Long>("android.sensor.info.maxFrameDuration", long.class); /** * <p> * The physical dimensions of the full pixel * array * </p> * <p> * Needed for FOV calculation for old API * </p> * <p>The physical dimensions of the full pixel * array</p> * <p>Needed for FOV calculation for old API</p> */ public static final Key<float[]> SENSOR_INFO_PHYSICAL_SIZE = new Key<float[]>("android.sensor.info.physicalSize", float[].class); /** * <p> * Gain factor from electrons to raw units when * ISO=100 * </p> * <p>Gain factor from electrons to raw units when * ISO=100</p> * * <b>Optional</b> - This value may be null on some devices. * Loading @@ -519,16 +427,12 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Rational>("android.sensor.baseGainFactor", Rational.class); /** * <p> * Maximum sensitivity that is implemented * purely through analog gain * </p> * <p> * For android.sensor.sensitivity values less than or * <p>Maximum sensitivity that is implemented * purely through analog gain</p> * <p>For android.sensor.sensitivity values less than or * equal to this, all applied gain must be analog. For * values above this, it can be a mix of analog and * digital * </p> * digital</p> * * <b>Optional</b> - This value may be null on some devices. * Loading @@ -540,60 +444,46 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<Integer>("android.sensor.maxAnalogSensitivity", int.class); /** * <p> * Clockwise angle through which the output * <p>Clockwise angle through which the output * image needs to be rotated to be upright on the device * screen in its native orientation. Also defines the * direction of rolling shutter readout, which is from top * to bottom in the sensor's coordinate system * </p> * to bottom in the sensor's coordinate system</p> */ public static final Key<Integer> SENSOR_ORIENTATION = new Key<Integer>("android.sensor.orientation", int.class); /** * <p> * Which face detection modes are available, * if any * </p> * <p> * OFF means face detection is disabled, it must * be included in the list. * </p><p> * SIMPLE means the device supports the * <p>Which face detection modes are available, * if any</p> * <p>OFF means face detection is disabled, it must * be included in the list.</p> * <p>SIMPLE means the device supports the * android.statistics.faceRectangles and * android.statistics.faceScores outputs. * </p><p> * FULL means the device additionally supports the * android.statistics.faceScores outputs.</p> * <p>FULL means the device additionally supports the * android.statistics.faceIds and * android.statistics.faceLandmarks outputs. * </p> * android.statistics.faceLandmarks outputs.</p> */ public static final Key<byte[]> STATISTICS_INFO_AVAILABLE_FACE_DETECT_MODES = new Key<byte[]>("android.statistics.info.availableFaceDetectModes", byte[].class); /** * <p> * Maximum number of simultaneously detectable * faces * </p> * <p>Maximum number of simultaneously detectable * faces</p> */ public static final Key<Integer> STATISTICS_INFO_MAX_FACE_COUNT = new Key<Integer>("android.statistics.info.maxFaceCount", int.class); /** * <p> * Maximum number of supported points in the * tonemap curve * </p> * <p>Maximum number of supported points in the * tonemap curve</p> */ public static final Key<Integer> TONEMAP_MAX_CURVE_POINTS = new Key<Integer>("android.tonemap.maxCurvePoints", int.class); /** * <p> * A list of camera LEDs that are available on this system. * </p> * <p>A list of camera LEDs that are available on this system.</p> * @see #LED_AVAILABLE_LEDS_TRANSMIT * * @hide Loading @@ -602,17 +492,14 @@ public final class CameraCharacteristics extends CameraMetadata { new Key<int[]>("android.led.availableLeds", int[].class); /** * <p> * The camera 3 HAL device can implement one of two possible * <p>The camera 3 HAL device can implement one of two possible * operational modes; limited and full. Full support is * expected from new higher-end devices. Limited mode has * hardware requirements roughly in line with those for a * camera HAL device v1 implementation, and is expected from * older or inexpensive devices. Full is a strict superset of * limited, and they share the same essential operational flow. * </p><p> * For full details refer to "S3. Operational Modes" in camera3.h * </p> * limited, and they share the same essential operational flow.</p> * <p>For full details refer to "S3. Operational Modes" in camera3.h</p> * @see #INFO_SUPPORTED_HARDWARE_LEVEL_LIMITED * @see #INFO_SUPPORTED_HARDWARE_LEVEL_FULL */ Loading