Loading apex/TEST_MAPPING +25 −0 Original line number Diff line number Diff line Loading @@ -3,5 +3,30 @@ { "path": "system/apex/tests" } ], "presubmit": [ // The following tests validate codec and drm path. { "name": "GtsMediaTestCases", "options" : [ { "include-annotation": "android.platform.test.annotations.Presubmit" }, { "include-filter": "com.google.android.media.gts.WidevineGenericOpsTests" } ] }, { "name": "GtsExoPlayerTestCases", "options" : [ { "include-annotation": "android.platform.test.annotations.SocPresubmit" }, { "include-filter": "com.google.android.exoplayer.gts.DashTest#testWidevine23FpsH264Fixed" } ] } ] } camera/ndk/include/camera/NdkCameraMetadataTags.h +17 −7 Original line number Diff line number Diff line Loading @@ -4383,8 +4383,8 @@ typedef enum acamera_metadata_tag { ACAMERA_SENSOR_AVAILABLE_TEST_PATTERN_MODES = // int32[n] ACAMERA_SENSOR_START + 25, /** * <p>Duration between the start of first row exposure * and the start of last row exposure.</p> * <p>Duration between the start of exposure for the first row of the image sensor, * and the start of exposure for one past the last row of the image sensor.</p> * * <p>Type: int64</p> * Loading @@ -4393,12 +4393,22 @@ typedef enum acamera_metadata_tag { * <li>ACameraMetadata from ACameraCaptureSession_captureCallback_result callbacks</li> * </ul></p> * * <p>This is the exposure time skew between the first and last * row exposure start times. The first row and the last row are * the first and last rows inside of the * <p>This is the exposure time skew between the first and <code>(last+1)</code> row exposure start times. The * first row and the last row are the first and last rows inside of the * ACAMERA_SENSOR_INFO_ACTIVE_ARRAY_SIZE.</p> * <p>For typical camera sensors that use rolling shutters, this is also equivalent * to the frame readout time.</p> * <p>For typical camera sensors that use rolling shutters, this is also equivalent to the frame * readout time.</p> * <p>If the image sensor is operating in a binned or cropped mode due to the current output * target resolutions, it's possible this skew is reported to be larger than the exposure * time, for example, since it is based on the full array even if a partial array is read * out. Be sure to scale the number to cover the section of the sensor actually being used * for the outputs you care about. So if your output covers N rows of the active array of * height H, scale this value by N/H to get the total skew for that viewport.</p> * <p><em>Note:</em> Prior to Android 11, this field was described as measuring duration from * first to last row of the image sensor, which is not equal to the frame readout time for a * rolling shutter sensor. Implementations generally reported the latter value, so to resolve * the inconsistency, the description has been updated to range from (first, last+1) row * exposure start, instead.</p> * * @see ACAMERA_SENSOR_INFO_ACTIVE_ARRAY_SIZE */ Loading media/codec2/sfplugin/CCodecBufferChannel.cpp +23 −20 Original line number Diff line number Diff line Loading @@ -851,6 +851,10 @@ status_t CCodecBufferChannel::renderOutputBuffer( if (hdrStaticInfo || hdr10PlusInfo) { HdrMetadata hdr; if (hdrStaticInfo) { // If mastering max and min luminance fields are 0, do not use them. // It indicates the value may not be present in the stream. if (hdrStaticInfo->mastering.maxLuminance > 0.0f && hdrStaticInfo->mastering.minLuminance > 0.0f) { struct android_smpte2086_metadata smpte2086_meta = { .displayPrimaryRed = { hdrStaticInfo->mastering.red.x, hdrStaticInfo->mastering.red.y Loading @@ -867,10 +871,9 @@ status_t CCodecBufferChannel::renderOutputBuffer( .maxLuminance = hdrStaticInfo->mastering.maxLuminance, .minLuminance = hdrStaticInfo->mastering.minLuminance, }; hdr.validTypes = HdrMetadata::SMPTE2086; hdr.smpte2086 = smpte2086_meta; } // If the content light level fields are 0, do not use them, it // indicates the value may not be present in the stream. if (hdrStaticInfo->maxCll > 0.0f && hdrStaticInfo->maxFall > 0.0f) { Loading Loading
apex/TEST_MAPPING +25 −0 Original line number Diff line number Diff line Loading @@ -3,5 +3,30 @@ { "path": "system/apex/tests" } ], "presubmit": [ // The following tests validate codec and drm path. { "name": "GtsMediaTestCases", "options" : [ { "include-annotation": "android.platform.test.annotations.Presubmit" }, { "include-filter": "com.google.android.media.gts.WidevineGenericOpsTests" } ] }, { "name": "GtsExoPlayerTestCases", "options" : [ { "include-annotation": "android.platform.test.annotations.SocPresubmit" }, { "include-filter": "com.google.android.exoplayer.gts.DashTest#testWidevine23FpsH264Fixed" } ] } ] }
camera/ndk/include/camera/NdkCameraMetadataTags.h +17 −7 Original line number Diff line number Diff line Loading @@ -4383,8 +4383,8 @@ typedef enum acamera_metadata_tag { ACAMERA_SENSOR_AVAILABLE_TEST_PATTERN_MODES = // int32[n] ACAMERA_SENSOR_START + 25, /** * <p>Duration between the start of first row exposure * and the start of last row exposure.</p> * <p>Duration between the start of exposure for the first row of the image sensor, * and the start of exposure for one past the last row of the image sensor.</p> * * <p>Type: int64</p> * Loading @@ -4393,12 +4393,22 @@ typedef enum acamera_metadata_tag { * <li>ACameraMetadata from ACameraCaptureSession_captureCallback_result callbacks</li> * </ul></p> * * <p>This is the exposure time skew between the first and last * row exposure start times. The first row and the last row are * the first and last rows inside of the * <p>This is the exposure time skew between the first and <code>(last+1)</code> row exposure start times. The * first row and the last row are the first and last rows inside of the * ACAMERA_SENSOR_INFO_ACTIVE_ARRAY_SIZE.</p> * <p>For typical camera sensors that use rolling shutters, this is also equivalent * to the frame readout time.</p> * <p>For typical camera sensors that use rolling shutters, this is also equivalent to the frame * readout time.</p> * <p>If the image sensor is operating in a binned or cropped mode due to the current output * target resolutions, it's possible this skew is reported to be larger than the exposure * time, for example, since it is based on the full array even if a partial array is read * out. Be sure to scale the number to cover the section of the sensor actually being used * for the outputs you care about. So if your output covers N rows of the active array of * height H, scale this value by N/H to get the total skew for that viewport.</p> * <p><em>Note:</em> Prior to Android 11, this field was described as measuring duration from * first to last row of the image sensor, which is not equal to the frame readout time for a * rolling shutter sensor. Implementations generally reported the latter value, so to resolve * the inconsistency, the description has been updated to range from (first, last+1) row * exposure start, instead.</p> * * @see ACAMERA_SENSOR_INFO_ACTIVE_ARRAY_SIZE */ Loading
media/codec2/sfplugin/CCodecBufferChannel.cpp +23 −20 Original line number Diff line number Diff line Loading @@ -851,6 +851,10 @@ status_t CCodecBufferChannel::renderOutputBuffer( if (hdrStaticInfo || hdr10PlusInfo) { HdrMetadata hdr; if (hdrStaticInfo) { // If mastering max and min luminance fields are 0, do not use them. // It indicates the value may not be present in the stream. if (hdrStaticInfo->mastering.maxLuminance > 0.0f && hdrStaticInfo->mastering.minLuminance > 0.0f) { struct android_smpte2086_metadata smpte2086_meta = { .displayPrimaryRed = { hdrStaticInfo->mastering.red.x, hdrStaticInfo->mastering.red.y Loading @@ -867,10 +871,9 @@ status_t CCodecBufferChannel::renderOutputBuffer( .maxLuminance = hdrStaticInfo->mastering.maxLuminance, .minLuminance = hdrStaticInfo->mastering.minLuminance, }; hdr.validTypes = HdrMetadata::SMPTE2086; hdr.smpte2086 = smpte2086_meta; } // If the content light level fields are 0, do not use them, it // indicates the value may not be present in the stream. if (hdrStaticInfo->maxCll > 0.0f && hdrStaticInfo->maxFall > 0.0f) { Loading