1121 lines
44 KiB
C++
1121 lines
44 KiB
C++
|
|
/*
|
||
|
|
* Copyright (C) 2019 The Android Open Source Project
|
||
|
|
*
|
||
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||
|
|
* you may not use this file except in compliance with the License.
|
||
|
|
* You may obtain a copy of the License at
|
||
|
|
*
|
||
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||
|
|
*
|
||
|
|
* Unless required by applicable law or agreed to in writing, software
|
||
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
|
|
* See the License for the specific language governing permissions and
|
||
|
|
* limitations under the License.
|
||
|
|
*/
|
||
|
|
|
||
|
|
//#define LOG_NDEBUG 0
|
||
|
|
#define LOG_TAG "C2SoftGav1Dec"
|
||
|
|
#include "C2SoftGav1Dec.h"
|
||
|
|
|
||
|
|
#include <C2Debug.h>
|
||
|
|
#include <C2PlatformSupport.h>
|
||
|
|
#include <Codec2BufferUtils.h>
|
||
|
|
#include <Codec2CommonUtils.h>
|
||
|
|
#include <Codec2Mapper.h>
|
||
|
|
#include <SimpleC2Interface.h>
|
||
|
|
#include <libyuv.h>
|
||
|
|
#include <log/log.h>
|
||
|
|
#include <media/stagefright/foundation/AUtils.h>
|
||
|
|
#include <media/stagefright/foundation/MediaDefs.h>
|
||
|
|
|
||
|
|
// libyuv version required for I410ToAB30Matrix and I210ToAB30Matrix.
|
||
|
|
#if LIBYUV_VERSION >= 1780
|
||
|
|
#include <algorithm>
|
||
|
|
#define HAVE_LIBYUV_I410_I210_TO_AB30 1
|
||
|
|
#else
|
||
|
|
#define HAVE_LIBYUV_I410_I210_TO_AB30 0
|
||
|
|
#endif
|
||
|
|
|
||
|
|
namespace android {
|
||
|
|
|
||
|
|
// codecname set and passed in as a compile flag from Android.bp
|
||
|
|
constexpr char COMPONENT_NAME[] = CODECNAME;
|
||
|
|
|
||
|
|
constexpr size_t kMinInputBufferSize = 2 * 1024 * 1024;
|
||
|
|
|
||
|
|
class C2SoftGav1Dec::IntfImpl : public SimpleInterface<void>::BaseParams {
|
||
|
|
public:
|
||
|
|
explicit IntfImpl(const std::shared_ptr<C2ReflectorHelper> &helper)
|
||
|
|
: SimpleInterface<void>::BaseParams(
|
||
|
|
helper, COMPONENT_NAME, C2Component::KIND_DECODER,
|
||
|
|
C2Component::DOMAIN_VIDEO, MEDIA_MIMETYPE_VIDEO_AV1) {
|
||
|
|
noPrivateBuffers(); // TODO: account for our buffers here.
|
||
|
|
noInputReferences();
|
||
|
|
noOutputReferences();
|
||
|
|
noInputLatency();
|
||
|
|
noTimeStretch();
|
||
|
|
|
||
|
|
addParameter(DefineParam(mAttrib, C2_PARAMKEY_COMPONENT_ATTRIBUTES)
|
||
|
|
.withConstValue(new C2ComponentAttributesSetting(
|
||
|
|
C2Component::ATTRIB_IS_TEMPORAL))
|
||
|
|
.build());
|
||
|
|
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mSize, C2_PARAMKEY_PICTURE_SIZE)
|
||
|
|
.withDefault(new C2StreamPictureSizeInfo::output(0u, 320, 240))
|
||
|
|
.withFields({
|
||
|
|
C2F(mSize, width).inRange(2, 4096),
|
||
|
|
C2F(mSize, height).inRange(2, 4096),
|
||
|
|
})
|
||
|
|
.withSetter(SizeSetter)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
addParameter(DefineParam(mProfileLevel, C2_PARAMKEY_PROFILE_LEVEL)
|
||
|
|
.withDefault(new C2StreamProfileLevelInfo::input(
|
||
|
|
0u, C2Config::PROFILE_AV1_0, C2Config::LEVEL_AV1_2_1))
|
||
|
|
.withFields({C2F(mProfileLevel, profile)
|
||
|
|
.oneOf({C2Config::PROFILE_AV1_0,
|
||
|
|
C2Config::PROFILE_AV1_1}),
|
||
|
|
C2F(mProfileLevel, level)
|
||
|
|
.oneOf({
|
||
|
|
C2Config::LEVEL_AV1_2, C2Config::LEVEL_AV1_2_1,
|
||
|
|
C2Config::LEVEL_AV1_2_2, C2Config::LEVEL_AV1_2_3,
|
||
|
|
C2Config::LEVEL_AV1_3, C2Config::LEVEL_AV1_3_1,
|
||
|
|
C2Config::LEVEL_AV1_3_2, C2Config::LEVEL_AV1_3_3,
|
||
|
|
C2Config::LEVEL_AV1_4, C2Config::LEVEL_AV1_4_1,
|
||
|
|
C2Config::LEVEL_AV1_4_2, C2Config::LEVEL_AV1_4_3,
|
||
|
|
C2Config::LEVEL_AV1_5, C2Config::LEVEL_AV1_5_1,
|
||
|
|
C2Config::LEVEL_AV1_5_2, C2Config::LEVEL_AV1_5_3,
|
||
|
|
})})
|
||
|
|
.withSetter(ProfileLevelSetter, mSize)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
mHdr10PlusInfoInput = C2StreamHdr10PlusInfo::input::AllocShared(0);
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mHdr10PlusInfoInput, C2_PARAMKEY_INPUT_HDR10_PLUS_INFO)
|
||
|
|
.withDefault(mHdr10PlusInfoInput)
|
||
|
|
.withFields({
|
||
|
|
C2F(mHdr10PlusInfoInput, m.value).any(),
|
||
|
|
})
|
||
|
|
.withSetter(Hdr10PlusInfoInputSetter)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
mHdr10PlusInfoOutput = C2StreamHdr10PlusInfo::output::AllocShared(0);
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mHdr10PlusInfoOutput, C2_PARAMKEY_OUTPUT_HDR10_PLUS_INFO)
|
||
|
|
.withDefault(mHdr10PlusInfoOutput)
|
||
|
|
.withFields({
|
||
|
|
C2F(mHdr10PlusInfoOutput, m.value).any(),
|
||
|
|
})
|
||
|
|
.withSetter(Hdr10PlusInfoOutputSetter)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
// default static info
|
||
|
|
C2HdrStaticMetadataStruct defaultStaticInfo{};
|
||
|
|
helper->addStructDescriptors<C2MasteringDisplayColorVolumeStruct, C2ColorXyStruct>();
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mHdrStaticInfo, C2_PARAMKEY_HDR_STATIC_INFO)
|
||
|
|
.withDefault(new C2StreamHdrStaticInfo::output(0u, defaultStaticInfo))
|
||
|
|
.withFields({
|
||
|
|
C2F(mHdrStaticInfo, mastering.red.x).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.red.y).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.green.x).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.green.y).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.blue.x).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.blue.y).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.white.x).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.white.x).inRange(0, 1),
|
||
|
|
C2F(mHdrStaticInfo, mastering.maxLuminance).inRange(0, 65535),
|
||
|
|
C2F(mHdrStaticInfo, mastering.minLuminance).inRange(0, 6.5535),
|
||
|
|
C2F(mHdrStaticInfo, maxCll).inRange(0, 0XFFFF),
|
||
|
|
C2F(mHdrStaticInfo, maxFall).inRange(0, 0XFFFF)
|
||
|
|
})
|
||
|
|
.withSetter(HdrStaticInfoSetter)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mMaxSize, C2_PARAMKEY_MAX_PICTURE_SIZE)
|
||
|
|
.withDefault(new C2StreamMaxPictureSizeTuning::output(0u, 320, 240))
|
||
|
|
.withFields({
|
||
|
|
C2F(mSize, width).inRange(2, 2048, 2),
|
||
|
|
C2F(mSize, height).inRange(2, 2048, 2),
|
||
|
|
})
|
||
|
|
.withSetter(MaxPictureSizeSetter, mSize)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
addParameter(DefineParam(mMaxInputSize, C2_PARAMKEY_INPUT_MAX_BUFFER_SIZE)
|
||
|
|
.withDefault(new C2StreamMaxBufferSizeInfo::input(0u, kMinInputBufferSize))
|
||
|
|
.withFields({
|
||
|
|
C2F(mMaxInputSize, value).any(),
|
||
|
|
})
|
||
|
|
.calculatedAs(MaxInputSizeSetter, mMaxSize)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
C2ChromaOffsetStruct locations[1] = {C2ChromaOffsetStruct::ITU_YUV_420_0()};
|
||
|
|
std::shared_ptr<C2StreamColorInfo::output> defaultColorInfo =
|
||
|
|
C2StreamColorInfo::output::AllocShared(1u, 0u, 8u /* bitDepth */,
|
||
|
|
C2Color::YUV_420);
|
||
|
|
memcpy(defaultColorInfo->m.locations, locations, sizeof(locations));
|
||
|
|
|
||
|
|
defaultColorInfo = C2StreamColorInfo::output::AllocShared(
|
||
|
|
{C2ChromaOffsetStruct::ITU_YUV_420_0()}, 0u, 8u /* bitDepth */,
|
||
|
|
C2Color::YUV_420);
|
||
|
|
helper->addStructDescriptors<C2ChromaOffsetStruct>();
|
||
|
|
|
||
|
|
addParameter(DefineParam(mColorInfo, C2_PARAMKEY_CODED_COLOR_INFO)
|
||
|
|
.withConstValue(defaultColorInfo)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mDefaultColorAspects, C2_PARAMKEY_DEFAULT_COLOR_ASPECTS)
|
||
|
|
.withDefault(new C2StreamColorAspectsTuning::output(
|
||
|
|
0u, C2Color::RANGE_UNSPECIFIED, C2Color::PRIMARIES_UNSPECIFIED,
|
||
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED))
|
||
|
|
.withFields(
|
||
|
|
{C2F(mDefaultColorAspects, range)
|
||
|
|
.inRange(C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER),
|
||
|
|
C2F(mDefaultColorAspects, primaries)
|
||
|
|
.inRange(C2Color::PRIMARIES_UNSPECIFIED,
|
||
|
|
C2Color::PRIMARIES_OTHER),
|
||
|
|
C2F(mDefaultColorAspects, transfer)
|
||
|
|
.inRange(C2Color::TRANSFER_UNSPECIFIED,
|
||
|
|
C2Color::TRANSFER_OTHER),
|
||
|
|
C2F(mDefaultColorAspects, matrix)
|
||
|
|
.inRange(C2Color::MATRIX_UNSPECIFIED,
|
||
|
|
C2Color::MATRIX_OTHER)})
|
||
|
|
.withSetter(DefaultColorAspectsSetter)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mCodedColorAspects, C2_PARAMKEY_VUI_COLOR_ASPECTS)
|
||
|
|
.withDefault(new C2StreamColorAspectsInfo::input(
|
||
|
|
0u, C2Color::RANGE_LIMITED, C2Color::PRIMARIES_UNSPECIFIED,
|
||
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED))
|
||
|
|
.withFields({
|
||
|
|
C2F(mCodedColorAspects, range).inRange(
|
||
|
|
C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER),
|
||
|
|
C2F(mCodedColorAspects, primaries).inRange(
|
||
|
|
C2Color::PRIMARIES_UNSPECIFIED, C2Color::PRIMARIES_OTHER),
|
||
|
|
C2F(mCodedColorAspects, transfer).inRange(
|
||
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::TRANSFER_OTHER),
|
||
|
|
C2F(mCodedColorAspects, matrix).inRange(
|
||
|
|
C2Color::MATRIX_UNSPECIFIED, C2Color::MATRIX_OTHER)
|
||
|
|
})
|
||
|
|
.withSetter(CodedColorAspectsSetter)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mColorAspects, C2_PARAMKEY_COLOR_ASPECTS)
|
||
|
|
.withDefault(new C2StreamColorAspectsInfo::output(
|
||
|
|
0u, C2Color::RANGE_UNSPECIFIED, C2Color::PRIMARIES_UNSPECIFIED,
|
||
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED))
|
||
|
|
.withFields({
|
||
|
|
C2F(mColorAspects, range).inRange(
|
||
|
|
C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER),
|
||
|
|
C2F(mColorAspects, primaries).inRange(
|
||
|
|
C2Color::PRIMARIES_UNSPECIFIED, C2Color::PRIMARIES_OTHER),
|
||
|
|
C2F(mColorAspects, transfer).inRange(
|
||
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::TRANSFER_OTHER),
|
||
|
|
C2F(mColorAspects, matrix).inRange(
|
||
|
|
C2Color::MATRIX_UNSPECIFIED, C2Color::MATRIX_OTHER)
|
||
|
|
})
|
||
|
|
.withSetter(ColorAspectsSetter, mDefaultColorAspects, mCodedColorAspects)
|
||
|
|
.build());
|
||
|
|
|
||
|
|
std::vector<uint32_t> pixelFormats = {HAL_PIXEL_FORMAT_YCBCR_420_888};
|
||
|
|
if (isHalPixelFormatSupported((AHardwareBuffer_Format)HAL_PIXEL_FORMAT_YCBCR_P010)) {
|
||
|
|
pixelFormats.push_back(HAL_PIXEL_FORMAT_YCBCR_P010);
|
||
|
|
}
|
||
|
|
// If color format surface isn't added to supported formats, there is no way to know
|
||
|
|
// when the color-format is configured to surface. This is necessary to be able to
|
||
|
|
// choose 10-bit format while decoding 10-bit clips in surface mode.
|
||
|
|
pixelFormats.push_back(HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED);
|
||
|
|
|
||
|
|
// TODO: support more formats?
|
||
|
|
addParameter(
|
||
|
|
DefineParam(mPixelFormat, C2_PARAMKEY_PIXEL_FORMAT)
|
||
|
|
.withDefault(new C2StreamPixelFormatInfo::output(
|
||
|
|
0u, HAL_PIXEL_FORMAT_YCBCR_420_888))
|
||
|
|
.withFields({C2F(mPixelFormat, value).oneOf(pixelFormats)})
|
||
|
|
.withSetter((Setter<decltype(*mPixelFormat)>::StrictValueWithNoDeps))
|
||
|
|
.build());
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R SizeSetter(bool mayBlock,
|
||
|
|
const C2P<C2StreamPictureSizeInfo::output> &oldMe,
|
||
|
|
C2P<C2StreamPictureSizeInfo::output> &me) {
|
||
|
|
(void)mayBlock;
|
||
|
|
C2R res = C2R::Ok();
|
||
|
|
if (!me.F(me.v.width).supportsAtAll(me.v.width)) {
|
||
|
|
res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.width)));
|
||
|
|
me.set().width = oldMe.v.width;
|
||
|
|
}
|
||
|
|
if (!me.F(me.v.height).supportsAtAll(me.v.height)) {
|
||
|
|
res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.height)));
|
||
|
|
me.set().height = oldMe.v.height;
|
||
|
|
}
|
||
|
|
return res;
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R MaxPictureSizeSetter(
|
||
|
|
bool mayBlock, C2P<C2StreamMaxPictureSizeTuning::output> &me,
|
||
|
|
const C2P<C2StreamPictureSizeInfo::output> &size) {
|
||
|
|
(void)mayBlock;
|
||
|
|
// TODO: get max width/height from the size's field helpers vs.
|
||
|
|
// hardcoding
|
||
|
|
me.set().width = c2_min(c2_max(me.v.width, size.v.width), 4096u);
|
||
|
|
me.set().height = c2_min(c2_max(me.v.height, size.v.height), 4096u);
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R MaxInputSizeSetter(
|
||
|
|
bool mayBlock, C2P<C2StreamMaxBufferSizeInfo::input> &me,
|
||
|
|
const C2P<C2StreamMaxPictureSizeTuning::output> &maxSize) {
|
||
|
|
(void)mayBlock;
|
||
|
|
// assume compression ratio of 2, but enforce a floor
|
||
|
|
me.set().value = c2_max((((maxSize.v.width + 63) / 64)
|
||
|
|
* ((maxSize.v.height + 63) / 64) * 3072), kMinInputBufferSize);
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R DefaultColorAspectsSetter(
|
||
|
|
bool mayBlock, C2P<C2StreamColorAspectsTuning::output> &me) {
|
||
|
|
(void)mayBlock;
|
||
|
|
if (me.v.range > C2Color::RANGE_OTHER) {
|
||
|
|
me.set().range = C2Color::RANGE_OTHER;
|
||
|
|
}
|
||
|
|
if (me.v.primaries > C2Color::PRIMARIES_OTHER) {
|
||
|
|
me.set().primaries = C2Color::PRIMARIES_OTHER;
|
||
|
|
}
|
||
|
|
if (me.v.transfer > C2Color::TRANSFER_OTHER) {
|
||
|
|
me.set().transfer = C2Color::TRANSFER_OTHER;
|
||
|
|
}
|
||
|
|
if (me.v.matrix > C2Color::MATRIX_OTHER) {
|
||
|
|
me.set().matrix = C2Color::MATRIX_OTHER;
|
||
|
|
}
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R CodedColorAspectsSetter(bool mayBlock, C2P<C2StreamColorAspectsInfo::input> &me) {
|
||
|
|
(void)mayBlock;
|
||
|
|
if (me.v.range > C2Color::RANGE_OTHER) {
|
||
|
|
me.set().range = C2Color::RANGE_OTHER;
|
||
|
|
}
|
||
|
|
if (me.v.primaries > C2Color::PRIMARIES_OTHER) {
|
||
|
|
me.set().primaries = C2Color::PRIMARIES_OTHER;
|
||
|
|
}
|
||
|
|
if (me.v.transfer > C2Color::TRANSFER_OTHER) {
|
||
|
|
me.set().transfer = C2Color::TRANSFER_OTHER;
|
||
|
|
}
|
||
|
|
if (me.v.matrix > C2Color::MATRIX_OTHER) {
|
||
|
|
me.set().matrix = C2Color::MATRIX_OTHER;
|
||
|
|
}
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R ColorAspectsSetter(bool mayBlock, C2P<C2StreamColorAspectsInfo::output> &me,
|
||
|
|
const C2P<C2StreamColorAspectsTuning::output> &def,
|
||
|
|
const C2P<C2StreamColorAspectsInfo::input> &coded) {
|
||
|
|
(void)mayBlock;
|
||
|
|
// take default values for all unspecified fields, and coded values for specified ones
|
||
|
|
me.set().range = coded.v.range == RANGE_UNSPECIFIED ? def.v.range : coded.v.range;
|
||
|
|
me.set().primaries = coded.v.primaries == PRIMARIES_UNSPECIFIED
|
||
|
|
? def.v.primaries : coded.v.primaries;
|
||
|
|
me.set().transfer = coded.v.transfer == TRANSFER_UNSPECIFIED
|
||
|
|
? def.v.transfer : coded.v.transfer;
|
||
|
|
me.set().matrix = coded.v.matrix == MATRIX_UNSPECIFIED ? def.v.matrix : coded.v.matrix;
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R ProfileLevelSetter(
|
||
|
|
bool mayBlock, C2P<C2StreamProfileLevelInfo::input> &me,
|
||
|
|
const C2P<C2StreamPictureSizeInfo::output> &size) {
|
||
|
|
(void)mayBlock;
|
||
|
|
(void)size;
|
||
|
|
(void)me; // TODO: validate
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
std::shared_ptr<C2StreamColorAspectsTuning::output>
|
||
|
|
getDefaultColorAspects_l() {
|
||
|
|
return mDefaultColorAspects;
|
||
|
|
}
|
||
|
|
|
||
|
|
std::shared_ptr<C2StreamColorAspectsInfo::output> getColorAspects_l() {
|
||
|
|
return mColorAspects;
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R Hdr10PlusInfoInputSetter(bool mayBlock,
|
||
|
|
C2P<C2StreamHdr10PlusInfo::input> &me) {
|
||
|
|
(void)mayBlock;
|
||
|
|
(void)me; // TODO: validate
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
static C2R Hdr10PlusInfoOutputSetter(bool mayBlock,
|
||
|
|
C2P<C2StreamHdr10PlusInfo::output> &me) {
|
||
|
|
(void)mayBlock;
|
||
|
|
(void)me; // TODO: validate
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
// unsafe getters
|
||
|
|
std::shared_ptr<C2StreamPixelFormatInfo::output> getPixelFormat_l() const { return mPixelFormat; }
|
||
|
|
|
||
|
|
static C2R HdrStaticInfoSetter(bool mayBlock, C2P<C2StreamHdrStaticInfo::output> &me) {
|
||
|
|
(void)mayBlock;
|
||
|
|
if (me.v.mastering.red.x > 1) {
|
||
|
|
me.set().mastering.red.x = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.red.y > 1) {
|
||
|
|
me.set().mastering.red.y = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.green.x > 1) {
|
||
|
|
me.set().mastering.green.x = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.green.y > 1) {
|
||
|
|
me.set().mastering.green.y = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.blue.x > 1) {
|
||
|
|
me.set().mastering.blue.x = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.blue.y > 1) {
|
||
|
|
me.set().mastering.blue.y = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.white.x > 1) {
|
||
|
|
me.set().mastering.white.x = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.white.y > 1) {
|
||
|
|
me.set().mastering.white.y = 1;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.maxLuminance > 65535.0) {
|
||
|
|
me.set().mastering.maxLuminance = 65535.0;
|
||
|
|
}
|
||
|
|
if (me.v.mastering.minLuminance > 6.5535) {
|
||
|
|
me.set().mastering.minLuminance = 6.5535;
|
||
|
|
}
|
||
|
|
if (me.v.maxCll > 65535.0) {
|
||
|
|
me.set().maxCll = 65535.0;
|
||
|
|
}
|
||
|
|
if (me.v.maxFall > 65535.0) {
|
||
|
|
me.set().maxFall = 65535.0;
|
||
|
|
}
|
||
|
|
return C2R::Ok();
|
||
|
|
}
|
||
|
|
|
||
|
|
private:
|
||
|
|
std::shared_ptr<C2StreamProfileLevelInfo::input> mProfileLevel;
|
||
|
|
std::shared_ptr<C2StreamPictureSizeInfo::output> mSize;
|
||
|
|
std::shared_ptr<C2StreamMaxPictureSizeTuning::output> mMaxSize;
|
||
|
|
std::shared_ptr<C2StreamMaxBufferSizeInfo::input> mMaxInputSize;
|
||
|
|
std::shared_ptr<C2StreamColorInfo::output> mColorInfo;
|
||
|
|
std::shared_ptr<C2StreamPixelFormatInfo::output> mPixelFormat;
|
||
|
|
std::shared_ptr<C2StreamColorAspectsTuning::output> mDefaultColorAspects;
|
||
|
|
std::shared_ptr<C2StreamColorAspectsInfo::input> mCodedColorAspects;
|
||
|
|
std::shared_ptr<C2StreamColorAspectsInfo::output> mColorAspects;
|
||
|
|
std::shared_ptr<C2StreamHdr10PlusInfo::input> mHdr10PlusInfoInput;
|
||
|
|
std::shared_ptr<C2StreamHdr10PlusInfo::output> mHdr10PlusInfoOutput;
|
||
|
|
std::shared_ptr<C2StreamHdrStaticInfo::output> mHdrStaticInfo;
|
||
|
|
};
|
||
|
|
|
||
|
|
C2SoftGav1Dec::C2SoftGav1Dec(const char *name, c2_node_id_t id,
|
||
|
|
const std::shared_ptr<IntfImpl> &intfImpl)
|
||
|
|
: SimpleC2Component(
|
||
|
|
std::make_shared<SimpleInterface<IntfImpl>>(name, id, intfImpl)),
|
||
|
|
mIntf(intfImpl),
|
||
|
|
mCodecCtx(nullptr) {
|
||
|
|
mTimeStart = mTimeEnd = systemTime();
|
||
|
|
}
|
||
|
|
|
||
|
|
C2SoftGav1Dec::~C2SoftGav1Dec() { onRelease(); }
|
||
|
|
|
||
|
|
c2_status_t C2SoftGav1Dec::onInit() {
|
||
|
|
return initDecoder() ? C2_OK : C2_CORRUPTED;
|
||
|
|
}
|
||
|
|
|
||
|
|
c2_status_t C2SoftGav1Dec::onStop() {
|
||
|
|
mSignalledError = false;
|
||
|
|
mSignalledOutputEos = false;
|
||
|
|
return C2_OK;
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::onReset() {
|
||
|
|
(void)onStop();
|
||
|
|
c2_status_t err = onFlush_sm();
|
||
|
|
if (err != C2_OK) {
|
||
|
|
ALOGW("Failed to flush the av1 decoder. Trying to hard reset.");
|
||
|
|
destroyDecoder();
|
||
|
|
if (!initDecoder()) {
|
||
|
|
ALOGE("Hard reset failed.");
|
||
|
|
}
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::onRelease() { destroyDecoder(); }
|
||
|
|
|
||
|
|
c2_status_t C2SoftGav1Dec::onFlush_sm() {
|
||
|
|
Libgav1StatusCode status = mCodecCtx->SignalEOS();
|
||
|
|
if (status != kLibgav1StatusOk) {
|
||
|
|
ALOGE("Failed to flush av1 decoder. status: %d.", status);
|
||
|
|
return C2_CORRUPTED;
|
||
|
|
}
|
||
|
|
|
||
|
|
// Dequeue frame (if any) that was enqueued previously.
|
||
|
|
const libgav1::DecoderBuffer *buffer;
|
||
|
|
status = mCodecCtx->DequeueFrame(&buffer);
|
||
|
|
if (status != kLibgav1StatusOk && status != kLibgav1StatusNothingToDequeue) {
|
||
|
|
ALOGE("Failed to dequeue frame after flushing the av1 decoder. status: %d",
|
||
|
|
status);
|
||
|
|
return C2_CORRUPTED;
|
||
|
|
}
|
||
|
|
|
||
|
|
mSignalledError = false;
|
||
|
|
mSignalledOutputEos = false;
|
||
|
|
|
||
|
|
return C2_OK;
|
||
|
|
}
|
||
|
|
|
||
|
|
static int GetCPUCoreCount() {
|
||
|
|
int cpuCoreCount = 1;
|
||
|
|
#if defined(_SC_NPROCESSORS_ONLN)
|
||
|
|
cpuCoreCount = sysconf(_SC_NPROCESSORS_ONLN);
|
||
|
|
#else
|
||
|
|
// _SC_NPROC_ONLN must be defined...
|
||
|
|
cpuCoreCount = sysconf(_SC_NPROC_ONLN);
|
||
|
|
#endif
|
||
|
|
CHECK(cpuCoreCount >= 1);
|
||
|
|
ALOGV("Number of CPU cores: %d", cpuCoreCount);
|
||
|
|
return cpuCoreCount;
|
||
|
|
}
|
||
|
|
|
||
|
|
bool C2SoftGav1Dec::initDecoder() {
|
||
|
|
mSignalledError = false;
|
||
|
|
mSignalledOutputEos = false;
|
||
|
|
mHalPixelFormat = HAL_PIXEL_FORMAT_YV12;
|
||
|
|
{
|
||
|
|
IntfImpl::Lock lock = mIntf->lock();
|
||
|
|
mPixelFormatInfo = mIntf->getPixelFormat_l();
|
||
|
|
}
|
||
|
|
mCodecCtx.reset(new libgav1::Decoder());
|
||
|
|
|
||
|
|
if (mCodecCtx == nullptr) {
|
||
|
|
ALOGE("mCodecCtx is null");
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
|
||
|
|
libgav1::DecoderSettings settings = {};
|
||
|
|
settings.threads = GetCPUCoreCount();
|
||
|
|
|
||
|
|
ALOGV("Using libgav1 AV1 software decoder.");
|
||
|
|
Libgav1StatusCode status = mCodecCtx->Init(&settings);
|
||
|
|
if (status != kLibgav1StatusOk) {
|
||
|
|
ALOGE("av1 decoder failed to initialize. status: %d.", status);
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
|
||
|
|
return true;
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::destroyDecoder() { mCodecCtx = nullptr; }
|
||
|
|
|
||
|
|
void fillEmptyWork(const std::unique_ptr<C2Work> &work) {
|
||
|
|
uint32_t flags = 0;
|
||
|
|
if (work->input.flags & C2FrameData::FLAG_END_OF_STREAM) {
|
||
|
|
flags |= C2FrameData::FLAG_END_OF_STREAM;
|
||
|
|
ALOGV("signalling eos");
|
||
|
|
}
|
||
|
|
work->worklets.front()->output.flags = (C2FrameData::flags_t)flags;
|
||
|
|
work->worklets.front()->output.buffers.clear();
|
||
|
|
work->worklets.front()->output.ordinal = work->input.ordinal;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::finishWork(uint64_t index,
|
||
|
|
const std::unique_ptr<C2Work> &work,
|
||
|
|
const std::shared_ptr<C2GraphicBlock> &block) {
|
||
|
|
std::shared_ptr<C2Buffer> buffer =
|
||
|
|
createGraphicBuffer(block, C2Rect(mWidth, mHeight));
|
||
|
|
{
|
||
|
|
IntfImpl::Lock lock = mIntf->lock();
|
||
|
|
buffer->setInfo(mIntf->getColorAspects_l());
|
||
|
|
}
|
||
|
|
auto fillWork = [buffer, index](const std::unique_ptr<C2Work> &work) {
|
||
|
|
uint32_t flags = 0;
|
||
|
|
if ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) &&
|
||
|
|
(c2_cntr64_t(index) == work->input.ordinal.frameIndex)) {
|
||
|
|
flags |= C2FrameData::FLAG_END_OF_STREAM;
|
||
|
|
ALOGV("signalling eos");
|
||
|
|
}
|
||
|
|
work->worklets.front()->output.flags = (C2FrameData::flags_t)flags;
|
||
|
|
work->worklets.front()->output.buffers.clear();
|
||
|
|
work->worklets.front()->output.buffers.push_back(buffer);
|
||
|
|
work->worklets.front()->output.ordinal = work->input.ordinal;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
};
|
||
|
|
if (work && c2_cntr64_t(index) == work->input.ordinal.frameIndex) {
|
||
|
|
fillWork(work);
|
||
|
|
} else {
|
||
|
|
finish(index, fillWork);
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::process(const std::unique_ptr<C2Work> &work,
|
||
|
|
const std::shared_ptr<C2BlockPool> &pool) {
|
||
|
|
work->result = C2_OK;
|
||
|
|
work->workletsProcessed = 0u;
|
||
|
|
work->worklets.front()->output.configUpdate.clear();
|
||
|
|
work->worklets.front()->output.flags = work->input.flags;
|
||
|
|
if (mSignalledError || mSignalledOutputEos) {
|
||
|
|
work->result = C2_BAD_VALUE;
|
||
|
|
return;
|
||
|
|
}
|
||
|
|
|
||
|
|
size_t inOffset = 0u;
|
||
|
|
size_t inSize = 0u;
|
||
|
|
C2ReadView rView = mDummyReadView;
|
||
|
|
if (!work->input.buffers.empty()) {
|
||
|
|
rView = work->input.buffers[0]->data().linearBlocks().front().map().get();
|
||
|
|
inSize = rView.capacity();
|
||
|
|
if (inSize && rView.error()) {
|
||
|
|
ALOGE("read view map failed %d", rView.error());
|
||
|
|
work->result = C2_CORRUPTED;
|
||
|
|
return;
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
bool codecConfig =
|
||
|
|
((work->input.flags & C2FrameData::FLAG_CODEC_CONFIG) != 0);
|
||
|
|
bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0);
|
||
|
|
|
||
|
|
ALOGV("in buffer attr. size %zu timestamp %d frameindex %d, flags %x", inSize,
|
||
|
|
(int)work->input.ordinal.timestamp.peeku(),
|
||
|
|
(int)work->input.ordinal.frameIndex.peeku(), work->input.flags);
|
||
|
|
|
||
|
|
if (codecConfig) {
|
||
|
|
fillEmptyWork(work);
|
||
|
|
return;
|
||
|
|
}
|
||
|
|
|
||
|
|
int64_t frameIndex = work->input.ordinal.frameIndex.peekll();
|
||
|
|
if (inSize) {
|
||
|
|
uint8_t *bitstream = const_cast<uint8_t *>(rView.data() + inOffset);
|
||
|
|
|
||
|
|
mTimeStart = systemTime();
|
||
|
|
nsecs_t delay = mTimeStart - mTimeEnd;
|
||
|
|
|
||
|
|
const Libgav1StatusCode status =
|
||
|
|
mCodecCtx->EnqueueFrame(bitstream, inSize, frameIndex,
|
||
|
|
/*buffer_private_data=*/nullptr);
|
||
|
|
|
||
|
|
mTimeEnd = systemTime();
|
||
|
|
nsecs_t decodeTime = mTimeEnd - mTimeStart;
|
||
|
|
ALOGV("decodeTime=%4" PRId64 " delay=%4" PRId64 "\n", decodeTime, delay);
|
||
|
|
|
||
|
|
if (status != kLibgav1StatusOk) {
|
||
|
|
ALOGE("av1 decoder failed to decode frame. status: %d.", status);
|
||
|
|
work->result = C2_CORRUPTED;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
mSignalledError = true;
|
||
|
|
return;
|
||
|
|
}
|
||
|
|
|
||
|
|
}
|
||
|
|
|
||
|
|
(void)outputBuffer(pool, work);
|
||
|
|
|
||
|
|
if (eos) {
|
||
|
|
drainInternal(DRAIN_COMPONENT_WITH_EOS, pool, work);
|
||
|
|
mSignalledOutputEos = true;
|
||
|
|
} else if (!inSize) {
|
||
|
|
fillEmptyWork(work);
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::getHDRStaticParams(const libgav1::DecoderBuffer *buffer,
|
||
|
|
const std::unique_ptr<C2Work> &work) {
|
||
|
|
C2StreamHdrStaticMetadataInfo::output hdrStaticMetadataInfo{};
|
||
|
|
bool infoPresent = false;
|
||
|
|
if (buffer->has_hdr_mdcv) {
|
||
|
|
// hdr_mdcv.primary_chromaticity_* values are in 0.16 fixed-point format.
|
||
|
|
hdrStaticMetadataInfo.mastering.red.x = buffer->hdr_mdcv.primary_chromaticity_x[0] / 65536.0;
|
||
|
|
hdrStaticMetadataInfo.mastering.red.y = buffer->hdr_mdcv.primary_chromaticity_y[0] / 65536.0;
|
||
|
|
|
||
|
|
hdrStaticMetadataInfo.mastering.green.x = buffer->hdr_mdcv.primary_chromaticity_x[1] / 65536.0;
|
||
|
|
hdrStaticMetadataInfo.mastering.green.y = buffer->hdr_mdcv.primary_chromaticity_y[1] / 65536.0;
|
||
|
|
|
||
|
|
hdrStaticMetadataInfo.mastering.blue.x = buffer->hdr_mdcv.primary_chromaticity_x[2] / 65536.0;
|
||
|
|
hdrStaticMetadataInfo.mastering.blue.y = buffer->hdr_mdcv.primary_chromaticity_y[2] / 65536.0;
|
||
|
|
|
||
|
|
// hdr_mdcv.white_point_chromaticity_* values are in 0.16 fixed-point format.
|
||
|
|
hdrStaticMetadataInfo.mastering.white.x = buffer->hdr_mdcv.white_point_chromaticity_x / 65536.0;
|
||
|
|
hdrStaticMetadataInfo.mastering.white.y = buffer->hdr_mdcv.white_point_chromaticity_y / 65536.0;
|
||
|
|
|
||
|
|
// hdr_mdcv.luminance_max is in 24.8 fixed-point format.
|
||
|
|
hdrStaticMetadataInfo.mastering.maxLuminance = buffer->hdr_mdcv.luminance_max / 256.0;
|
||
|
|
// hdr_mdcv.luminance_min is in 18.14 format.
|
||
|
|
hdrStaticMetadataInfo.mastering.minLuminance = buffer->hdr_mdcv.luminance_min / 16384.0;
|
||
|
|
infoPresent = true;
|
||
|
|
}
|
||
|
|
|
||
|
|
if (buffer->has_hdr_cll) {
|
||
|
|
hdrStaticMetadataInfo.maxCll = buffer->hdr_cll.max_cll;
|
||
|
|
hdrStaticMetadataInfo.maxFall = buffer->hdr_cll.max_fall;
|
||
|
|
infoPresent = true;
|
||
|
|
}
|
||
|
|
// config if static info has changed
|
||
|
|
if (infoPresent && !(hdrStaticMetadataInfo == mHdrStaticMetadataInfo)) {
|
||
|
|
mHdrStaticMetadataInfo = hdrStaticMetadataInfo;
|
||
|
|
work->worklets.front()->output.configUpdate.push_back(C2Param::Copy(mHdrStaticMetadataInfo));
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::getHDR10PlusInfoData(const libgav1::DecoderBuffer *buffer,
|
||
|
|
const std::unique_ptr<C2Work> &work) {
|
||
|
|
if (buffer->has_itut_t35) {
|
||
|
|
std::vector<uint8_t> payload;
|
||
|
|
size_t payloadSize = buffer->itut_t35.payload_size;
|
||
|
|
if (payloadSize > 0) {
|
||
|
|
payload.push_back(buffer->itut_t35.country_code);
|
||
|
|
if (buffer->itut_t35.country_code == 0xFF) {
|
||
|
|
payload.push_back(buffer->itut_t35.country_code_extension_byte);
|
||
|
|
}
|
||
|
|
payload.insert(payload.end(), buffer->itut_t35.payload_bytes,
|
||
|
|
buffer->itut_t35.payload_bytes + buffer->itut_t35.payload_size);
|
||
|
|
}
|
||
|
|
|
||
|
|
std::unique_ptr<C2StreamHdr10PlusInfo::output> hdr10PlusInfo =
|
||
|
|
C2StreamHdr10PlusInfo::output::AllocUnique(payload.size());
|
||
|
|
if (!hdr10PlusInfo) {
|
||
|
|
ALOGE("Hdr10PlusInfo allocation failed");
|
||
|
|
mSignalledError = true;
|
||
|
|
work->result = C2_NO_MEMORY;
|
||
|
|
return;
|
||
|
|
}
|
||
|
|
memcpy(hdr10PlusInfo->m.value, payload.data(), payload.size());
|
||
|
|
|
||
|
|
// config if hdr10Plus info has changed
|
||
|
|
if (nullptr == mHdr10PlusInfo || !(*hdr10PlusInfo == *mHdr10PlusInfo)) {
|
||
|
|
mHdr10PlusInfo = std::move(hdr10PlusInfo);
|
||
|
|
work->worklets.front()->output.configUpdate.push_back(std::move(mHdr10PlusInfo));
|
||
|
|
}
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::getVuiParams(const libgav1::DecoderBuffer *buffer) {
|
||
|
|
VuiColorAspects vuiColorAspects;
|
||
|
|
vuiColorAspects.primaries = buffer->color_primary;
|
||
|
|
vuiColorAspects.transfer = buffer->transfer_characteristics;
|
||
|
|
vuiColorAspects.coeffs = buffer->matrix_coefficients;
|
||
|
|
vuiColorAspects.fullRange = buffer->color_range;
|
||
|
|
|
||
|
|
// convert vui aspects to C2 values if changed
|
||
|
|
if (!(vuiColorAspects == mBitstreamColorAspects)) {
|
||
|
|
mBitstreamColorAspects = vuiColorAspects;
|
||
|
|
ColorAspects sfAspects;
|
||
|
|
C2StreamColorAspectsInfo::input codedAspects = { 0u };
|
||
|
|
ColorUtils::convertIsoColorAspectsToCodecAspects(
|
||
|
|
vuiColorAspects.primaries, vuiColorAspects.transfer, vuiColorAspects.coeffs,
|
||
|
|
vuiColorAspects.fullRange, sfAspects);
|
||
|
|
if (!C2Mapper::map(sfAspects.mPrimaries, &codedAspects.primaries)) {
|
||
|
|
codedAspects.primaries = C2Color::PRIMARIES_UNSPECIFIED;
|
||
|
|
}
|
||
|
|
if (!C2Mapper::map(sfAspects.mRange, &codedAspects.range)) {
|
||
|
|
codedAspects.range = C2Color::RANGE_UNSPECIFIED;
|
||
|
|
}
|
||
|
|
if (!C2Mapper::map(sfAspects.mMatrixCoeffs, &codedAspects.matrix)) {
|
||
|
|
codedAspects.matrix = C2Color::MATRIX_UNSPECIFIED;
|
||
|
|
}
|
||
|
|
if (!C2Mapper::map(sfAspects.mTransfer, &codedAspects.transfer)) {
|
||
|
|
codedAspects.transfer = C2Color::TRANSFER_UNSPECIFIED;
|
||
|
|
}
|
||
|
|
std::vector<std::unique_ptr<C2SettingResult>> failures;
|
||
|
|
mIntf->config({&codedAspects}, C2_MAY_BLOCK, &failures);
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
void C2SoftGav1Dec::setError(const std::unique_ptr<C2Work> &work, c2_status_t error) {
|
||
|
|
mSignalledError = true;
|
||
|
|
work->result = error;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
}
|
||
|
|
|
||
|
|
bool C2SoftGav1Dec::allocTmpFrameBuffer(size_t size) {
|
||
|
|
if (size > mTmpFrameBufferSize) {
|
||
|
|
mTmpFrameBuffer = std::make_unique<uint16_t[]>(size);
|
||
|
|
if (mTmpFrameBuffer == nullptr) {
|
||
|
|
mTmpFrameBufferSize = 0;
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
mTmpFrameBufferSize = size;
|
||
|
|
}
|
||
|
|
return true;
|
||
|
|
}
|
||
|
|
|
||
|
|
bool C2SoftGav1Dec::outputBuffer(const std::shared_ptr<C2BlockPool> &pool,
|
||
|
|
const std::unique_ptr<C2Work> &work) {
|
||
|
|
if (!(work && pool)) return false;
|
||
|
|
|
||
|
|
const libgav1::DecoderBuffer *buffer;
|
||
|
|
const Libgav1StatusCode status = mCodecCtx->DequeueFrame(&buffer);
|
||
|
|
|
||
|
|
if (status != kLibgav1StatusOk && status != kLibgav1StatusNothingToDequeue) {
|
||
|
|
ALOGE("av1 decoder DequeueFrame failed. status: %d.", status);
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
|
||
|
|
// |buffer| can be NULL if status was equal to kLibgav1StatusOk or
|
||
|
|
// kLibgav1StatusNothingToDequeue. This is not an error. This could mean one
|
||
|
|
// of two things:
|
||
|
|
// - The EnqueueFrame() call was either a flush (called with nullptr).
|
||
|
|
// - The enqueued frame did not have any displayable frames.
|
||
|
|
if (!buffer) {
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
|
||
|
|
const int width = buffer->displayed_width[0];
|
||
|
|
const int height = buffer->displayed_height[0];
|
||
|
|
if (width != mWidth || height != mHeight) {
|
||
|
|
mWidth = width;
|
||
|
|
mHeight = height;
|
||
|
|
|
||
|
|
C2StreamPictureSizeInfo::output size(0u, mWidth, mHeight);
|
||
|
|
std::vector<std::unique_ptr<C2SettingResult>> failures;
|
||
|
|
c2_status_t err = mIntf->config({&size}, C2_MAY_BLOCK, &failures);
|
||
|
|
if (err == C2_OK) {
|
||
|
|
work->worklets.front()->output.configUpdate.push_back(
|
||
|
|
C2Param::Copy(size));
|
||
|
|
} else {
|
||
|
|
ALOGE("Config update size failed");
|
||
|
|
mSignalledError = true;
|
||
|
|
work->result = C2_CORRUPTED;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
}
|
||
|
|
|
||
|
|
getVuiParams(buffer);
|
||
|
|
getHDRStaticParams(buffer, work);
|
||
|
|
getHDR10PlusInfoData(buffer, work);
|
||
|
|
|
||
|
|
#if LIBYUV_VERSION < 1779
|
||
|
|
if (buffer->bitdepth == 10 &&
|
||
|
|
!(buffer->image_format == libgav1::kImageFormatYuv420 ||
|
||
|
|
buffer->image_format == libgav1::kImageFormatMonochrome400)) {
|
||
|
|
ALOGE("image_format %d not supported for 10bit", buffer->image_format);
|
||
|
|
mSignalledError = true;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
work->result = C2_CORRUPTED;
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
#endif
|
||
|
|
|
||
|
|
const bool isMonochrome =
|
||
|
|
buffer->image_format == libgav1::kImageFormatMonochrome400;
|
||
|
|
|
||
|
|
std::shared_ptr<C2GraphicBlock> block;
|
||
|
|
uint32_t format = HAL_PIXEL_FORMAT_YV12;
|
||
|
|
std::shared_ptr<C2StreamColorAspectsInfo::output> codedColorAspects;
|
||
|
|
if (buffer->bitdepth == 10 && mPixelFormatInfo->value != HAL_PIXEL_FORMAT_YCBCR_420_888) {
|
||
|
|
IntfImpl::Lock lock = mIntf->lock();
|
||
|
|
codedColorAspects = mIntf->getColorAspects_l();
|
||
|
|
bool allowRGBA1010102 = false;
|
||
|
|
if (codedColorAspects->primaries == C2Color::PRIMARIES_BT2020 &&
|
||
|
|
codedColorAspects->matrix == C2Color::MATRIX_BT2020 &&
|
||
|
|
codedColorAspects->transfer == C2Color::TRANSFER_ST2084) {
|
||
|
|
allowRGBA1010102 = true;
|
||
|
|
}
|
||
|
|
format = getHalPixelFormatForBitDepth10(allowRGBA1010102);
|
||
|
|
#if !HAVE_LIBYUV_I410_I210_TO_AB30
|
||
|
|
if ((format == HAL_PIXEL_FORMAT_RGBA_1010102) &&
|
||
|
|
(buffer->image_format != libgav1::kImageFormatYuv420)) {
|
||
|
|
ALOGE("Only YUV420 output is supported when targeting RGBA_1010102");
|
||
|
|
mSignalledError = true;
|
||
|
|
work->result = C2_OMITTED;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
#endif
|
||
|
|
}
|
||
|
|
|
||
|
|
if (mHalPixelFormat != format) {
|
||
|
|
C2StreamPixelFormatInfo::output pixelFormat(0u, format);
|
||
|
|
std::vector<std::unique_ptr<C2SettingResult>> failures;
|
||
|
|
c2_status_t err = mIntf->config({&pixelFormat }, C2_MAY_BLOCK, &failures);
|
||
|
|
if (err == C2_OK) {
|
||
|
|
work->worklets.front()->output.configUpdate.push_back(
|
||
|
|
C2Param::Copy(pixelFormat));
|
||
|
|
} else {
|
||
|
|
ALOGE("Config update pixelFormat failed");
|
||
|
|
mSignalledError = true;
|
||
|
|
work->workletsProcessed = 1u;
|
||
|
|
work->result = C2_CORRUPTED;
|
||
|
|
return UNKNOWN_ERROR;
|
||
|
|
}
|
||
|
|
mHalPixelFormat = format;
|
||
|
|
}
|
||
|
|
|
||
|
|
C2MemoryUsage usage = {C2MemoryUsage::CPU_READ, C2MemoryUsage::CPU_WRITE};
|
||
|
|
|
||
|
|
// We always create a graphic block that is width aligned to 16 and height
|
||
|
|
// aligned to 2. We set the correct "crop" value of the image in the call to
|
||
|
|
// createGraphicBuffer() by setting the correct image dimensions.
|
||
|
|
c2_status_t err = pool->fetchGraphicBlock(align(mWidth, 16),
|
||
|
|
align(mHeight, 2), format, usage,
|
||
|
|
&block);
|
||
|
|
|
||
|
|
if (err != C2_OK) {
|
||
|
|
ALOGE("fetchGraphicBlock for Output failed with status %d", err);
|
||
|
|
work->result = err;
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
|
||
|
|
C2GraphicView wView = block->map().get();
|
||
|
|
|
||
|
|
if (wView.error()) {
|
||
|
|
ALOGE("graphic view map failed %d", wView.error());
|
||
|
|
work->result = C2_CORRUPTED;
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
|
||
|
|
ALOGV("provided (%dx%d) required (%dx%d), out frameindex %d", block->width(),
|
||
|
|
block->height(), mWidth, mHeight, (int)buffer->user_private_data);
|
||
|
|
|
||
|
|
uint8_t *dstY = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_Y]);
|
||
|
|
uint8_t *dstU = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_U]);
|
||
|
|
uint8_t *dstV = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_V]);
|
||
|
|
|
||
|
|
C2PlanarLayout layout = wView.layout();
|
||
|
|
size_t dstYStride = layout.planes[C2PlanarLayout::PLANE_Y].rowInc;
|
||
|
|
size_t dstUStride = layout.planes[C2PlanarLayout::PLANE_U].rowInc;
|
||
|
|
size_t dstVStride = layout.planes[C2PlanarLayout::PLANE_V].rowInc;
|
||
|
|
|
||
|
|
if (buffer->bitdepth == 10) {
|
||
|
|
const uint16_t *srcY = (const uint16_t *)buffer->plane[0];
|
||
|
|
const uint16_t *srcU = (const uint16_t *)buffer->plane[1];
|
||
|
|
const uint16_t *srcV = (const uint16_t *)buffer->plane[2];
|
||
|
|
size_t srcYStride = buffer->stride[0] / 2;
|
||
|
|
size_t srcUStride = buffer->stride[1] / 2;
|
||
|
|
size_t srcVStride = buffer->stride[2] / 2;
|
||
|
|
|
||
|
|
if (format == HAL_PIXEL_FORMAT_RGBA_1010102) {
|
||
|
|
bool processed = false;
|
||
|
|
#if HAVE_LIBYUV_I410_I210_TO_AB30
|
||
|
|
if (buffer->image_format == libgav1::kImageFormatYuv444) {
|
||
|
|
libyuv::I410ToAB30Matrix(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
dstY, dstYStride, &libyuv::kYuvV2020Constants,
|
||
|
|
mWidth, mHeight);
|
||
|
|
processed = true;
|
||
|
|
} else if (buffer->image_format == libgav1::kImageFormatYuv422) {
|
||
|
|
libyuv::I210ToAB30Matrix(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
dstY, dstYStride, &libyuv::kYuvV2020Constants,
|
||
|
|
mWidth, mHeight);
|
||
|
|
processed = true;
|
||
|
|
}
|
||
|
|
#endif // HAVE_LIBYUV_I410_I210_TO_AB30
|
||
|
|
if (!processed) {
|
||
|
|
if (isMonochrome) {
|
||
|
|
const size_t tmpSize = mWidth;
|
||
|
|
const bool needFill = tmpSize > mTmpFrameBufferSize;
|
||
|
|
if (!allocTmpFrameBuffer(tmpSize)) {
|
||
|
|
ALOGE("Error allocating temp conversion buffer (%zu bytes)", tmpSize);
|
||
|
|
setError(work, C2_NO_MEMORY);
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
srcU = srcV = mTmpFrameBuffer.get();
|
||
|
|
srcUStride = srcVStride = 0;
|
||
|
|
if (needFill) {
|
||
|
|
std::fill_n(mTmpFrameBuffer.get(), tmpSize, 512);
|
||
|
|
}
|
||
|
|
}
|
||
|
|
convertYUV420Planar16ToY410OrRGBA1010102(
|
||
|
|
(uint32_t *)dstY, srcY, srcU, srcV, srcYStride,
|
||
|
|
srcUStride, srcVStride,
|
||
|
|
dstYStride / sizeof(uint32_t), mWidth, mHeight,
|
||
|
|
std::static_pointer_cast<const C2ColorAspectsStruct>(codedColorAspects));
|
||
|
|
}
|
||
|
|
} else if (format == HAL_PIXEL_FORMAT_YCBCR_P010) {
|
||
|
|
dstYStride /= 2;
|
||
|
|
dstUStride /= 2;
|
||
|
|
dstVStride /= 2;
|
||
|
|
#if LIBYUV_VERSION >= 1779
|
||
|
|
if (buffer->image_format == libgav1::kImageFormatYuv444 ||
|
||
|
|
buffer->image_format == libgav1::kImageFormatYuv422) {
|
||
|
|
// TODO(https://crbug.com/libyuv/952): replace this block with libyuv::I410ToP010 and
|
||
|
|
// libyuv::I210ToP010 when they are available.
|
||
|
|
// Note it may be safe to alias dstY in I010ToP010, but the libyuv API doesn't make any
|
||
|
|
// guarantees.
|
||
|
|
const size_t tmpSize = dstYStride * mHeight + dstUStride * align(mHeight, 2);
|
||
|
|
if (!allocTmpFrameBuffer(tmpSize)) {
|
||
|
|
ALOGE("Error allocating temp conversion buffer (%zu bytes)", tmpSize);
|
||
|
|
setError(work, C2_NO_MEMORY);
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
uint16_t *const tmpY = mTmpFrameBuffer.get();
|
||
|
|
uint16_t *const tmpU = tmpY + dstYStride * mHeight;
|
||
|
|
uint16_t *const tmpV = tmpU + dstUStride * align(mHeight, 2) / 2;
|
||
|
|
if (buffer->image_format == libgav1::kImageFormatYuv444) {
|
||
|
|
libyuv::I410ToI010(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
tmpY, dstYStride, tmpU, dstUStride, tmpV, dstUStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
} else {
|
||
|
|
libyuv::I210ToI010(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
tmpY, dstYStride, tmpU, dstUStride, tmpV, dstUStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
}
|
||
|
|
libyuv::I010ToP010(tmpY, dstYStride, tmpU, dstUStride, tmpV, dstVStride,
|
||
|
|
(uint16_t*)dstY, dstYStride, (uint16_t*)dstU, dstUStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
} else {
|
||
|
|
convertYUV420Planar16ToP010((uint16_t *)dstY, (uint16_t *)dstU, srcY, srcU, srcV,
|
||
|
|
srcYStride, srcUStride, srcVStride, dstYStride,
|
||
|
|
dstUStride, mWidth, mHeight, isMonochrome);
|
||
|
|
}
|
||
|
|
#else // LIBYUV_VERSION < 1779
|
||
|
|
convertYUV420Planar16ToP010((uint16_t *)dstY, (uint16_t *)dstU, srcY, srcU, srcV,
|
||
|
|
srcYStride, srcUStride, srcVStride, dstYStride,
|
||
|
|
dstUStride, mWidth, mHeight, isMonochrome);
|
||
|
|
#endif // LIBYUV_VERSION >= 1779
|
||
|
|
} else {
|
||
|
|
#if LIBYUV_VERSION >= 1779
|
||
|
|
if (buffer->image_format == libgav1::kImageFormatYuv444) {
|
||
|
|
// TODO(https://crbug.com/libyuv/950): replace this block with libyuv::I410ToI420 when
|
||
|
|
// it's available.
|
||
|
|
const size_t tmpSize = dstYStride * mHeight + dstUStride * align(mHeight, 2);
|
||
|
|
if (!allocTmpFrameBuffer(tmpSize)) {
|
||
|
|
ALOGE("Error allocating temp conversion buffer (%zu bytes)", tmpSize);
|
||
|
|
setError(work, C2_NO_MEMORY);
|
||
|
|
return false;
|
||
|
|
}
|
||
|
|
uint16_t *const tmpY = mTmpFrameBuffer.get();
|
||
|
|
uint16_t *const tmpU = tmpY + dstYStride * mHeight;
|
||
|
|
uint16_t *const tmpV = tmpU + dstUStride * align(mHeight, 2) / 2;
|
||
|
|
libyuv::I410ToI010(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
tmpY, dstYStride, tmpU, dstUStride, tmpV, dstVStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
libyuv::I010ToI420(tmpY, dstYStride, tmpU, dstUStride, tmpV, dstUStride,
|
||
|
|
dstY, dstYStride, dstU, dstUStride, dstV, dstVStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
} else if (buffer->image_format == libgav1::kImageFormatYuv422) {
|
||
|
|
libyuv::I210ToI420(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
dstY, dstYStride, dstU, dstUStride, dstV, dstVStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
} else {
|
||
|
|
convertYUV420Planar16ToYV12(dstY, dstU, dstV, srcY, srcU, srcV, srcYStride,
|
||
|
|
srcUStride, srcVStride, dstYStride, dstUStride,
|
||
|
|
mWidth, mHeight, isMonochrome);
|
||
|
|
}
|
||
|
|
#else // LIBYUV_VERSION < 1779
|
||
|
|
convertYUV420Planar16ToYV12(dstY, dstU, dstV, srcY, srcU, srcV, srcYStride,
|
||
|
|
srcUStride, srcVStride, dstYStride, dstUStride,
|
||
|
|
mWidth, mHeight, isMonochrome);
|
||
|
|
#endif // LIBYUV_VERSION >= 1779
|
||
|
|
}
|
||
|
|
} else {
|
||
|
|
const uint8_t *srcY = (const uint8_t *)buffer->plane[0];
|
||
|
|
const uint8_t *srcU = (const uint8_t *)buffer->plane[1];
|
||
|
|
const uint8_t *srcV = (const uint8_t *)buffer->plane[2];
|
||
|
|
size_t srcYStride = buffer->stride[0];
|
||
|
|
size_t srcUStride = buffer->stride[1];
|
||
|
|
size_t srcVStride = buffer->stride[2];
|
||
|
|
|
||
|
|
if (buffer->image_format == libgav1::kImageFormatYuv444) {
|
||
|
|
libyuv::I444ToI420(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
dstY, dstYStride, dstU, dstUStride, dstV, dstVStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
} else if (buffer->image_format == libgav1::kImageFormatYuv422) {
|
||
|
|
libyuv::I422ToI420(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride,
|
||
|
|
dstY, dstYStride, dstU, dstUStride, dstV, dstVStride,
|
||
|
|
mWidth, mHeight);
|
||
|
|
} else {
|
||
|
|
convertYUV420Planar8ToYV12(dstY, dstU, dstV, srcY, srcU, srcV, srcYStride, srcUStride,
|
||
|
|
srcVStride, dstYStride, dstUStride, dstVStride, mWidth, mHeight,
|
||
|
|
isMonochrome);
|
||
|
|
}
|
||
|
|
}
|
||
|
|
finishWork(buffer->user_private_data, work, std::move(block));
|
||
|
|
block = nullptr;
|
||
|
|
return true;
|
||
|
|
}
|
||
|
|
|
||
|
|
c2_status_t C2SoftGav1Dec::drainInternal(
|
||
|
|
uint32_t drainMode, const std::shared_ptr<C2BlockPool> &pool,
|
||
|
|
const std::unique_ptr<C2Work> &work) {
|
||
|
|
if (drainMode == NO_DRAIN) {
|
||
|
|
ALOGW("drain with NO_DRAIN: no-op");
|
||
|
|
return C2_OK;
|
||
|
|
}
|
||
|
|
if (drainMode == DRAIN_CHAIN) {
|
||
|
|
ALOGW("DRAIN_CHAIN not supported");
|
||
|
|
return C2_OMITTED;
|
||
|
|
}
|
||
|
|
|
||
|
|
const Libgav1StatusCode status = mCodecCtx->SignalEOS();
|
||
|
|
if (status != kLibgav1StatusOk) {
|
||
|
|
ALOGE("Failed to flush av1 decoder. status: %d.", status);
|
||
|
|
return C2_CORRUPTED;
|
||
|
|
}
|
||
|
|
|
||
|
|
while (outputBuffer(pool, work)) {
|
||
|
|
}
|
||
|
|
|
||
|
|
if (drainMode == DRAIN_COMPONENT_WITH_EOS && work &&
|
||
|
|
work->workletsProcessed == 0u) {
|
||
|
|
fillEmptyWork(work);
|
||
|
|
}
|
||
|
|
|
||
|
|
return C2_OK;
|
||
|
|
}
|
||
|
|
|
||
|
|
c2_status_t C2SoftGav1Dec::drain(uint32_t drainMode,
|
||
|
|
const std::shared_ptr<C2BlockPool> &pool) {
|
||
|
|
return drainInternal(drainMode, pool, nullptr);
|
||
|
|
}
|
||
|
|
|
||
|
|
class C2SoftGav1Factory : public C2ComponentFactory {
|
||
|
|
public:
|
||
|
|
C2SoftGav1Factory()
|
||
|
|
: mHelper(std::static_pointer_cast<C2ReflectorHelper>(
|
||
|
|
GetCodec2PlatformComponentStore()->getParamReflector())) {}
|
||
|
|
|
||
|
|
virtual c2_status_t createComponent(
|
||
|
|
c2_node_id_t id, std::shared_ptr<C2Component> *const component,
|
||
|
|
std::function<void(C2Component *)> deleter) override {
|
||
|
|
*component = std::shared_ptr<C2Component>(
|
||
|
|
new C2SoftGav1Dec(COMPONENT_NAME, id,
|
||
|
|
std::make_shared<C2SoftGav1Dec::IntfImpl>(mHelper)),
|
||
|
|
deleter);
|
||
|
|
return C2_OK;
|
||
|
|
}
|
||
|
|
|
||
|
|
virtual c2_status_t createInterface(
|
||
|
|
c2_node_id_t id, std::shared_ptr<C2ComponentInterface> *const interface,
|
||
|
|
std::function<void(C2ComponentInterface *)> deleter) override {
|
||
|
|
*interface = std::shared_ptr<C2ComponentInterface>(
|
||
|
|
new SimpleInterface<C2SoftGav1Dec::IntfImpl>(
|
||
|
|
COMPONENT_NAME, id,
|
||
|
|
std::make_shared<C2SoftGav1Dec::IntfImpl>(mHelper)),
|
||
|
|
deleter);
|
||
|
|
return C2_OK;
|
||
|
|
}
|
||
|
|
|
||
|
|
virtual ~C2SoftGav1Factory() override = default;
|
||
|
|
|
||
|
|
private:
|
||
|
|
std::shared_ptr<C2ReflectorHelper> mHelper;
|
||
|
|
};
|
||
|
|
|
||
|
|
} // namespace android
|
||
|
|
|
||
|
|
__attribute__((cfi_canonical_jump_table))
|
||
|
|
extern "C" ::C2ComponentFactory *CreateCodec2Factory() {
|
||
|
|
ALOGV("in %s", __func__);
|
||
|
|
return new ::android::C2SoftGav1Factory();
|
||
|
|
}
|
||
|
|
|
||
|
|
__attribute__((cfi_canonical_jump_table))
|
||
|
|
extern "C" void DestroyCodec2Factory(::C2ComponentFactory *factory) {
|
||
|
|
ALOGV("in %s", __func__);
|
||
|
|
delete factory;
|
||
|
|
}
|