blob: a1929e7d5e088dc2f03057f268c8012f8b787882 [file] [log] [blame]
Vignesh Venkatasubramanianb6d383d2019-06-10 15:11:58 -07001/*
2 * Copyright (C) 2019 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17//#define LOG_NDEBUG 0
18#define LOG_TAG "C2SoftGav1Dec"
19#include "C2SoftGav1Dec.h"
20
21#include <C2Debug.h>
22#include <C2PlatformSupport.h>
23#include <SimpleC2Interface.h>
24#include <log/log.h>
25#include <media/stagefright/foundation/AUtils.h>
26#include <media/stagefright/foundation/MediaDefs.h>
27
28namespace android {
29
Ray Essickc2cc4372019-08-21 14:02:28 -070030// codecname set and passed in as a compile flag from Android.bp
31constexpr char COMPONENT_NAME[] = CODECNAME;
Vignesh Venkatasubramanianb6d383d2019-06-10 15:11:58 -070032
33class C2SoftGav1Dec::IntfImpl : public SimpleInterface<void>::BaseParams {
34 public:
35 explicit IntfImpl(const std::shared_ptr<C2ReflectorHelper> &helper)
36 : SimpleInterface<void>::BaseParams(
37 helper, COMPONENT_NAME, C2Component::KIND_DECODER,
38 C2Component::DOMAIN_VIDEO, MEDIA_MIMETYPE_VIDEO_AV1) {
39 noPrivateBuffers(); // TODO: account for our buffers here.
40 noInputReferences();
41 noOutputReferences();
42 noInputLatency();
43 noTimeStretch();
44
45 addParameter(DefineParam(mAttrib, C2_PARAMKEY_COMPONENT_ATTRIBUTES)
46 .withConstValue(new C2ComponentAttributesSetting(
47 C2Component::ATTRIB_IS_TEMPORAL))
48 .build());
49
50 addParameter(
51 DefineParam(mSize, C2_PARAMKEY_PICTURE_SIZE)
52 .withDefault(new C2StreamPictureSizeInfo::output(0u, 320, 240))
53 .withFields({
54 C2F(mSize, width).inRange(2, 2048, 2),
55 C2F(mSize, height).inRange(2, 2048, 2),
56 })
57 .withSetter(SizeSetter)
58 .build());
59
60 addParameter(DefineParam(mProfileLevel, C2_PARAMKEY_PROFILE_LEVEL)
61 .withDefault(new C2StreamProfileLevelInfo::input(
62 0u, C2Config::PROFILE_AV1_0, C2Config::LEVEL_AV1_2_1))
63 .withFields({C2F(mProfileLevel, profile)
64 .oneOf({C2Config::PROFILE_AV1_0,
65 C2Config::PROFILE_AV1_1}),
66 C2F(mProfileLevel, level)
67 .oneOf({
68 C2Config::LEVEL_AV1_2,
69 C2Config::LEVEL_AV1_2_1,
70 C2Config::LEVEL_AV1_2_2,
71 C2Config::LEVEL_AV1_3,
72 C2Config::LEVEL_AV1_3_1,
73 C2Config::LEVEL_AV1_3_2,
74 })})
75 .withSetter(ProfileLevelSetter, mSize)
76 .build());
77
78 mHdr10PlusInfoInput = C2StreamHdr10PlusInfo::input::AllocShared(0);
79 addParameter(
80 DefineParam(mHdr10PlusInfoInput, C2_PARAMKEY_INPUT_HDR10_PLUS_INFO)
81 .withDefault(mHdr10PlusInfoInput)
82 .withFields({
83 C2F(mHdr10PlusInfoInput, m.value).any(),
84 })
85 .withSetter(Hdr10PlusInfoInputSetter)
86 .build());
87
88 mHdr10PlusInfoOutput = C2StreamHdr10PlusInfo::output::AllocShared(0);
89 addParameter(
90 DefineParam(mHdr10PlusInfoOutput, C2_PARAMKEY_OUTPUT_HDR10_PLUS_INFO)
91 .withDefault(mHdr10PlusInfoOutput)
92 .withFields({
93 C2F(mHdr10PlusInfoOutput, m.value).any(),
94 })
95 .withSetter(Hdr10PlusInfoOutputSetter)
96 .build());
97
98 addParameter(
99 DefineParam(mMaxSize, C2_PARAMKEY_MAX_PICTURE_SIZE)
100 .withDefault(new C2StreamMaxPictureSizeTuning::output(0u, 320, 240))
101 .withFields({
102 C2F(mSize, width).inRange(2, 2048, 2),
103 C2F(mSize, height).inRange(2, 2048, 2),
104 })
105 .withSetter(MaxPictureSizeSetter, mSize)
106 .build());
107
108 addParameter(DefineParam(mMaxInputSize, C2_PARAMKEY_INPUT_MAX_BUFFER_SIZE)
109 .withDefault(new C2StreamMaxBufferSizeInfo::input(
110 0u, 320 * 240 * 3 / 4))
111 .withFields({
112 C2F(mMaxInputSize, value).any(),
113 })
114 .calculatedAs(MaxInputSizeSetter, mMaxSize)
115 .build());
116
117 C2ChromaOffsetStruct locations[1] = {C2ChromaOffsetStruct::ITU_YUV_420_0()};
118 std::shared_ptr<C2StreamColorInfo::output> defaultColorInfo =
119 C2StreamColorInfo::output::AllocShared(1u, 0u, 8u /* bitDepth */,
120 C2Color::YUV_420);
121 memcpy(defaultColorInfo->m.locations, locations, sizeof(locations));
122
123 defaultColorInfo = C2StreamColorInfo::output::AllocShared(
124 {C2ChromaOffsetStruct::ITU_YUV_420_0()}, 0u, 8u /* bitDepth */,
125 C2Color::YUV_420);
126 helper->addStructDescriptors<C2ChromaOffsetStruct>();
127
128 addParameter(DefineParam(mColorInfo, C2_PARAMKEY_CODED_COLOR_INFO)
129 .withConstValue(defaultColorInfo)
130 .build());
131
132 addParameter(
133 DefineParam(mDefaultColorAspects, C2_PARAMKEY_DEFAULT_COLOR_ASPECTS)
134 .withDefault(new C2StreamColorAspectsTuning::output(
135 0u, C2Color::RANGE_UNSPECIFIED, C2Color::PRIMARIES_UNSPECIFIED,
136 C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED))
137 .withFields(
138 {C2F(mDefaultColorAspects, range)
139 .inRange(C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER),
140 C2F(mDefaultColorAspects, primaries)
141 .inRange(C2Color::PRIMARIES_UNSPECIFIED,
142 C2Color::PRIMARIES_OTHER),
143 C2F(mDefaultColorAspects, transfer)
144 .inRange(C2Color::TRANSFER_UNSPECIFIED,
145 C2Color::TRANSFER_OTHER),
146 C2F(mDefaultColorAspects, matrix)
147 .inRange(C2Color::MATRIX_UNSPECIFIED,
148 C2Color::MATRIX_OTHER)})
149 .withSetter(DefaultColorAspectsSetter)
150 .build());
151
152 // TODO: support more formats?
153 addParameter(DefineParam(mPixelFormat, C2_PARAMKEY_PIXEL_FORMAT)
154 .withConstValue(new C2StreamPixelFormatInfo::output(
155 0u, HAL_PIXEL_FORMAT_YCBCR_420_888))
156 .build());
157 }
158
159 static C2R SizeSetter(bool mayBlock,
160 const C2P<C2StreamPictureSizeInfo::output> &oldMe,
161 C2P<C2StreamPictureSizeInfo::output> &me) {
162 (void)mayBlock;
163 C2R res = C2R::Ok();
164 if (!me.F(me.v.width).supportsAtAll(me.v.width)) {
165 res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.width)));
166 me.set().width = oldMe.v.width;
167 }
168 if (!me.F(me.v.height).supportsAtAll(me.v.height)) {
169 res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.height)));
170 me.set().height = oldMe.v.height;
171 }
172 return res;
173 }
174
175 static C2R MaxPictureSizeSetter(
176 bool mayBlock, C2P<C2StreamMaxPictureSizeTuning::output> &me,
177 const C2P<C2StreamPictureSizeInfo::output> &size) {
178 (void)mayBlock;
179 // TODO: get max width/height from the size's field helpers vs.
180 // hardcoding
181 me.set().width = c2_min(c2_max(me.v.width, size.v.width), 4096u);
182 me.set().height = c2_min(c2_max(me.v.height, size.v.height), 4096u);
183 return C2R::Ok();
184 }
185
186 static C2R MaxInputSizeSetter(
187 bool mayBlock, C2P<C2StreamMaxBufferSizeInfo::input> &me,
188 const C2P<C2StreamMaxPictureSizeTuning::output> &maxSize) {
189 (void)mayBlock;
190 // assume compression ratio of 2
191 me.set().value =
192 (((maxSize.v.width + 63) / 64) * ((maxSize.v.height + 63) / 64) * 3072);
193 return C2R::Ok();
194 }
195
196 static C2R DefaultColorAspectsSetter(
197 bool mayBlock, C2P<C2StreamColorAspectsTuning::output> &me) {
198 (void)mayBlock;
199 if (me.v.range > C2Color::RANGE_OTHER) {
200 me.set().range = C2Color::RANGE_OTHER;
201 }
202 if (me.v.primaries > C2Color::PRIMARIES_OTHER) {
203 me.set().primaries = C2Color::PRIMARIES_OTHER;
204 }
205 if (me.v.transfer > C2Color::TRANSFER_OTHER) {
206 me.set().transfer = C2Color::TRANSFER_OTHER;
207 }
208 if (me.v.matrix > C2Color::MATRIX_OTHER) {
209 me.set().matrix = C2Color::MATRIX_OTHER;
210 }
211 return C2R::Ok();
212 }
213
214 static C2R ProfileLevelSetter(
215 bool mayBlock, C2P<C2StreamProfileLevelInfo::input> &me,
216 const C2P<C2StreamPictureSizeInfo::output> &size) {
217 (void)mayBlock;
218 (void)size;
219 (void)me; // TODO: validate
220 return C2R::Ok();
221 }
222
223 std::shared_ptr<C2StreamColorAspectsTuning::output>
224 getDefaultColorAspects_l() {
225 return mDefaultColorAspects;
226 }
227
228 static C2R Hdr10PlusInfoInputSetter(bool mayBlock,
229 C2P<C2StreamHdr10PlusInfo::input> &me) {
230 (void)mayBlock;
231 (void)me; // TODO: validate
232 return C2R::Ok();
233 }
234
235 static C2R Hdr10PlusInfoOutputSetter(bool mayBlock,
236 C2P<C2StreamHdr10PlusInfo::output> &me) {
237 (void)mayBlock;
238 (void)me; // TODO: validate
239 return C2R::Ok();
240 }
241
242 private:
243 std::shared_ptr<C2StreamProfileLevelInfo::input> mProfileLevel;
244 std::shared_ptr<C2StreamPictureSizeInfo::output> mSize;
245 std::shared_ptr<C2StreamMaxPictureSizeTuning::output> mMaxSize;
246 std::shared_ptr<C2StreamMaxBufferSizeInfo::input> mMaxInputSize;
247 std::shared_ptr<C2StreamColorInfo::output> mColorInfo;
248 std::shared_ptr<C2StreamPixelFormatInfo::output> mPixelFormat;
249 std::shared_ptr<C2StreamColorAspectsTuning::output> mDefaultColorAspects;
250 std::shared_ptr<C2StreamHdr10PlusInfo::input> mHdr10PlusInfoInput;
251 std::shared_ptr<C2StreamHdr10PlusInfo::output> mHdr10PlusInfoOutput;
252};
253
254C2SoftGav1Dec::C2SoftGav1Dec(const char *name, c2_node_id_t id,
255 const std::shared_ptr<IntfImpl> &intfImpl)
256 : SimpleC2Component(
257 std::make_shared<SimpleInterface<IntfImpl>>(name, id, intfImpl)),
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700258 mIntf(intfImpl),
259 mCodecCtx(nullptr) {
260 gettimeofday(&mTimeStart, nullptr);
261 gettimeofday(&mTimeEnd, nullptr);
262}
Vignesh Venkatasubramanianb6d383d2019-06-10 15:11:58 -0700263
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700264C2SoftGav1Dec::~C2SoftGav1Dec() { onRelease(); }
265
266c2_status_t C2SoftGav1Dec::onInit() {
267 return initDecoder() ? C2_OK : C2_CORRUPTED;
268}
269
270c2_status_t C2SoftGav1Dec::onStop() {
271 mSignalledError = false;
272 mSignalledOutputEos = false;
Vignesh Venkatasubramanianb6d383d2019-06-10 15:11:58 -0700273 return C2_OK;
274}
275
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700276void C2SoftGav1Dec::onReset() {
277 (void)onStop();
278 c2_status_t err = onFlush_sm();
279 if (err != C2_OK) {
280 ALOGW("Failed to flush the av1 decoder. Trying to hard reset.");
281 destroyDecoder();
282 if (!initDecoder()) {
283 ALOGE("Hard reset failed.");
284 }
285 }
286}
287
288void C2SoftGav1Dec::onRelease() { destroyDecoder(); }
289
290c2_status_t C2SoftGav1Dec::onFlush_sm() {
James Zernb7aee6e2020-06-26 13:49:53 -0700291 Libgav1StatusCode status = mCodecCtx->SignalEOS();
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700292 if (status != kLibgav1StatusOk) {
293 ALOGE("Failed to flush av1 decoder. status: %d.", status);
294 return C2_CORRUPTED;
295 }
296
297 // Dequeue frame (if any) that was enqueued previously.
298 const libgav1::DecoderBuffer *buffer;
299 status = mCodecCtx->DequeueFrame(&buffer);
James Zernb7aee6e2020-06-26 13:49:53 -0700300 if (status != kLibgav1StatusOk && status != kLibgav1StatusNothingToDequeue) {
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700301 ALOGE("Failed to dequeue frame after flushing the av1 decoder. status: %d",
302 status);
303 return C2_CORRUPTED;
304 }
305
306 mSignalledError = false;
307 mSignalledOutputEos = false;
308
309 return C2_OK;
310}
311
312static int GetCPUCoreCount() {
313 int cpuCoreCount = 1;
314#if defined(_SC_NPROCESSORS_ONLN)
315 cpuCoreCount = sysconf(_SC_NPROCESSORS_ONLN);
316#else
317 // _SC_NPROC_ONLN must be defined...
318 cpuCoreCount = sysconf(_SC_NPROC_ONLN);
319#endif
320 CHECK(cpuCoreCount >= 1);
321 ALOGV("Number of CPU cores: %d", cpuCoreCount);
322 return cpuCoreCount;
323}
324
325bool C2SoftGav1Dec::initDecoder() {
326 mSignalledError = false;
327 mSignalledOutputEos = false;
328 mCodecCtx.reset(new libgav1::Decoder());
329
330 if (mCodecCtx == nullptr) {
331 ALOGE("mCodecCtx is null");
332 return false;
333 }
334
335 libgav1::DecoderSettings settings = {};
336 settings.threads = GetCPUCoreCount();
337
Vignesh Venkatasubramanian61ba2cf2019-06-24 10:04:00 -0700338 ALOGV("Using libgav1 AV1 software decoder.");
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700339 Libgav1StatusCode status = mCodecCtx->Init(&settings);
340 if (status != kLibgav1StatusOk) {
341 ALOGE("av1 decoder failed to initialize. status: %d.", status);
342 return false;
343 }
344
345 return true;
346}
347
348void C2SoftGav1Dec::destroyDecoder() { mCodecCtx = nullptr; }
349
350void fillEmptyWork(const std::unique_ptr<C2Work> &work) {
351 uint32_t flags = 0;
352 if (work->input.flags & C2FrameData::FLAG_END_OF_STREAM) {
353 flags |= C2FrameData::FLAG_END_OF_STREAM;
354 ALOGV("signalling eos");
355 }
356 work->worklets.front()->output.flags = (C2FrameData::flags_t)flags;
357 work->worklets.front()->output.buffers.clear();
358 work->worklets.front()->output.ordinal = work->input.ordinal;
359 work->workletsProcessed = 1u;
360}
361
362void C2SoftGav1Dec::finishWork(uint64_t index,
363 const std::unique_ptr<C2Work> &work,
364 const std::shared_ptr<C2GraphicBlock> &block) {
365 std::shared_ptr<C2Buffer> buffer =
366 createGraphicBuffer(block, C2Rect(mWidth, mHeight));
367 auto fillWork = [buffer, index](const std::unique_ptr<C2Work> &work) {
368 uint32_t flags = 0;
369 if ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) &&
370 (c2_cntr64_t(index) == work->input.ordinal.frameIndex)) {
371 flags |= C2FrameData::FLAG_END_OF_STREAM;
372 ALOGV("signalling eos");
373 }
374 work->worklets.front()->output.flags = (C2FrameData::flags_t)flags;
375 work->worklets.front()->output.buffers.clear();
376 work->worklets.front()->output.buffers.push_back(buffer);
377 work->worklets.front()->output.ordinal = work->input.ordinal;
378 work->workletsProcessed = 1u;
379 };
380 if (work && c2_cntr64_t(index) == work->input.ordinal.frameIndex) {
381 fillWork(work);
382 } else {
383 finish(index, fillWork);
384 }
385}
386
387void C2SoftGav1Dec::process(const std::unique_ptr<C2Work> &work,
388 const std::shared_ptr<C2BlockPool> &pool) {
389 work->result = C2_OK;
390 work->workletsProcessed = 0u;
391 work->worklets.front()->output.configUpdate.clear();
392 work->worklets.front()->output.flags = work->input.flags;
393 if (mSignalledError || mSignalledOutputEos) {
394 work->result = C2_BAD_VALUE;
395 return;
396 }
397
398 size_t inOffset = 0u;
399 size_t inSize = 0u;
400 C2ReadView rView = mDummyReadView;
401 if (!work->input.buffers.empty()) {
402 rView = work->input.buffers[0]->data().linearBlocks().front().map().get();
403 inSize = rView.capacity();
404 if (inSize && rView.error()) {
405 ALOGE("read view map failed %d", rView.error());
406 work->result = C2_CORRUPTED;
407 return;
408 }
409 }
410
411 bool codecConfig =
412 ((work->input.flags & C2FrameData::FLAG_CODEC_CONFIG) != 0);
413 bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0);
414
415 ALOGV("in buffer attr. size %zu timestamp %d frameindex %d, flags %x", inSize,
416 (int)work->input.ordinal.timestamp.peeku(),
417 (int)work->input.ordinal.frameIndex.peeku(), work->input.flags);
418
419 if (codecConfig) {
420 fillEmptyWork(work);
421 return;
422 }
423
424 int64_t frameIndex = work->input.ordinal.frameIndex.peekll();
425 if (inSize) {
426 uint8_t *bitstream = const_cast<uint8_t *>(rView.data() + inOffset);
427 int32_t decodeTime = 0;
428 int32_t delay = 0;
429
430 GETTIME(&mTimeStart, nullptr);
431 TIME_DIFF(mTimeEnd, mTimeStart, delay);
432
433 const Libgav1StatusCode status =
James Zernb7aee6e2020-06-26 13:49:53 -0700434 mCodecCtx->EnqueueFrame(bitstream, inSize, frameIndex,
435 /*buffer_private_data=*/nullptr);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700436
437 GETTIME(&mTimeEnd, nullptr);
438 TIME_DIFF(mTimeStart, mTimeEnd, decodeTime);
439 ALOGV("decodeTime=%4d delay=%4d\n", decodeTime, delay);
440
441 if (status != kLibgav1StatusOk) {
442 ALOGE("av1 decoder failed to decode frame. status: %d.", status);
443 work->result = C2_CORRUPTED;
444 work->workletsProcessed = 1u;
445 mSignalledError = true;
446 return;
447 }
448
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700449 }
450
451 (void)outputBuffer(pool, work);
452
453 if (eos) {
454 drainInternal(DRAIN_COMPONENT_WITH_EOS, pool, work);
455 mSignalledOutputEos = true;
456 } else if (!inSize) {
457 fillEmptyWork(work);
458 }
459}
460
ming.zhouac19c3d2019-10-11 11:14:07 +0800461static void copyOutputBufferToYV12Frame(uint8_t *dstY, uint8_t *dstU, uint8_t *dstV,
462 const uint8_t *srcY, const uint8_t *srcU, const uint8_t *srcV,
463 size_t srcYStride, size_t srcUStride, size_t srcVStride,
464 size_t dstYStride, size_t dstUVStride,
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700465 uint32_t width, uint32_t height) {
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700466
467 for (size_t i = 0; i < height; ++i) {
ming.zhouac19c3d2019-10-11 11:14:07 +0800468 memcpy(dstY, srcY, width);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700469 srcY += srcYStride;
ming.zhouac19c3d2019-10-11 11:14:07 +0800470 dstY += dstYStride;
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700471 }
472
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700473 for (size_t i = 0; i < height / 2; ++i) {
ming.zhouac19c3d2019-10-11 11:14:07 +0800474 memcpy(dstV, srcV, width / 2);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700475 srcV += srcVStride;
ming.zhouac19c3d2019-10-11 11:14:07 +0800476 dstV += dstUVStride;
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700477 }
478
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700479 for (size_t i = 0; i < height / 2; ++i) {
ming.zhouac19c3d2019-10-11 11:14:07 +0800480 memcpy(dstU, srcU, width / 2);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700481 srcU += srcUStride;
ming.zhouac19c3d2019-10-11 11:14:07 +0800482 dstU += dstUVStride;
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700483 }
484}
485
486static void convertYUV420Planar16ToY410(uint32_t *dst, const uint16_t *srcY,
487 const uint16_t *srcU,
488 const uint16_t *srcV, size_t srcYStride,
489 size_t srcUStride, size_t srcVStride,
490 size_t dstStride, size_t width,
491 size_t height) {
492 // Converting two lines at a time, slightly faster
493 for (size_t y = 0; y < height; y += 2) {
494 uint32_t *dstTop = (uint32_t *)dst;
495 uint32_t *dstBot = (uint32_t *)(dst + dstStride);
496 uint16_t *ySrcTop = (uint16_t *)srcY;
497 uint16_t *ySrcBot = (uint16_t *)(srcY + srcYStride);
498 uint16_t *uSrc = (uint16_t *)srcU;
499 uint16_t *vSrc = (uint16_t *)srcV;
500
501 uint32_t u01, v01, y01, y23, y45, y67, uv0, uv1;
502 size_t x = 0;
503 for (; x < width - 3; x += 4) {
504 u01 = *((uint32_t *)uSrc);
505 uSrc += 2;
506 v01 = *((uint32_t *)vSrc);
507 vSrc += 2;
508
509 y01 = *((uint32_t *)ySrcTop);
510 ySrcTop += 2;
511 y23 = *((uint32_t *)ySrcTop);
512 ySrcTop += 2;
513 y45 = *((uint32_t *)ySrcBot);
514 ySrcBot += 2;
515 y67 = *((uint32_t *)ySrcBot);
516 ySrcBot += 2;
517
518 uv0 = (u01 & 0x3FF) | ((v01 & 0x3FF) << 20);
519 uv1 = (u01 >> 16) | ((v01 >> 16) << 20);
520
521 *dstTop++ = 3 << 30 | ((y01 & 0x3FF) << 10) | uv0;
522 *dstTop++ = 3 << 30 | ((y01 >> 16) << 10) | uv0;
523 *dstTop++ = 3 << 30 | ((y23 & 0x3FF) << 10) | uv1;
524 *dstTop++ = 3 << 30 | ((y23 >> 16) << 10) | uv1;
525
526 *dstBot++ = 3 << 30 | ((y45 & 0x3FF) << 10) | uv0;
527 *dstBot++ = 3 << 30 | ((y45 >> 16) << 10) | uv0;
528 *dstBot++ = 3 << 30 | ((y67 & 0x3FF) << 10) | uv1;
529 *dstBot++ = 3 << 30 | ((y67 >> 16) << 10) | uv1;
530 }
531
532 // There should be at most 2 more pixels to process. Note that we don't
533 // need to consider odd case as the buffer is always aligned to even.
534 if (x < width) {
535 u01 = *uSrc;
536 v01 = *vSrc;
537 y01 = *((uint32_t *)ySrcTop);
538 y45 = *((uint32_t *)ySrcBot);
539 uv0 = (u01 & 0x3FF) | ((v01 & 0x3FF) << 20);
540 *dstTop++ = ((y01 & 0x3FF) << 10) | uv0;
541 *dstTop++ = ((y01 >> 16) << 10) | uv0;
542 *dstBot++ = ((y45 & 0x3FF) << 10) | uv0;
543 *dstBot++ = ((y45 >> 16) << 10) | uv0;
544 }
545
546 srcY += srcYStride * 2;
547 srcU += srcUStride;
548 srcV += srcVStride;
549 dst += dstStride * 2;
550 }
551}
552
553static void convertYUV420Planar16ToYUV420Planar(
ming.zhouac19c3d2019-10-11 11:14:07 +0800554 uint8_t *dstY, uint8_t *dstU, uint8_t *dstV,
555 const uint16_t *srcY, const uint16_t *srcU, const uint16_t *srcV,
556 size_t srcYStride, size_t srcUStride, size_t srcVStride,
557 size_t dstYStride, size_t dstUVStride,
558 size_t width, size_t height) {
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700559
560 for (size_t y = 0; y < height; ++y) {
561 for (size_t x = 0; x < width; ++x) {
562 dstY[x] = (uint8_t)(srcY[x] >> 2);
563 }
564
565 srcY += srcYStride;
ming.zhouac19c3d2019-10-11 11:14:07 +0800566 dstY += dstYStride;
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700567 }
568
569 for (size_t y = 0; y < (height + 1) / 2; ++y) {
570 for (size_t x = 0; x < (width + 1) / 2; ++x) {
571 dstU[x] = (uint8_t)(srcU[x] >> 2);
572 dstV[x] = (uint8_t)(srcV[x] >> 2);
573 }
574
575 srcU += srcUStride;
576 srcV += srcVStride;
577 dstU += dstUVStride;
578 dstV += dstUVStride;
579 }
580}
581
582bool C2SoftGav1Dec::outputBuffer(const std::shared_ptr<C2BlockPool> &pool,
583 const std::unique_ptr<C2Work> &work) {
584 if (!(work && pool)) return false;
585
586 const libgav1::DecoderBuffer *buffer;
587 const Libgav1StatusCode status = mCodecCtx->DequeueFrame(&buffer);
588
James Zernb7aee6e2020-06-26 13:49:53 -0700589 if (status != kLibgav1StatusOk && status != kLibgav1StatusNothingToDequeue) {
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700590 ALOGE("av1 decoder DequeueFrame failed. status: %d.", status);
591 return false;
592 }
593
James Zernb7aee6e2020-06-26 13:49:53 -0700594 // |buffer| can be NULL if status was equal to kLibgav1StatusOk or
595 // kLibgav1StatusNothingToDequeue. This is not an error. This could mean one
596 // of two things:
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700597 // - The EnqueueFrame() call was either a flush (called with nullptr).
598 // - The enqueued frame did not have any displayable frames.
599 if (!buffer) {
600 return false;
601 }
602
603 const int width = buffer->displayed_width[0];
604 const int height = buffer->displayed_height[0];
605 if (width != mWidth || height != mHeight) {
606 mWidth = width;
607 mHeight = height;
608
609 C2StreamPictureSizeInfo::output size(0u, mWidth, mHeight);
610 std::vector<std::unique_ptr<C2SettingResult>> failures;
611 c2_status_t err = mIntf->config({&size}, C2_MAY_BLOCK, &failures);
612 if (err == C2_OK) {
613 work->worklets.front()->output.configUpdate.push_back(
614 C2Param::Copy(size));
615 } else {
616 ALOGE("Config update size failed");
617 mSignalledError = true;
618 work->result = C2_CORRUPTED;
619 work->workletsProcessed = 1u;
620 return false;
621 }
622 }
623
624 // TODO(vigneshv): Add support for monochrome videos since AV1 supports it.
625 CHECK(buffer->image_format == libgav1::kImageFormatYuv420);
626
627 std::shared_ptr<C2GraphicBlock> block;
628 uint32_t format = HAL_PIXEL_FORMAT_YV12;
629 if (buffer->bitdepth == 10) {
630 IntfImpl::Lock lock = mIntf->lock();
631 std::shared_ptr<C2StreamColorAspectsTuning::output> defaultColorAspects =
632 mIntf->getDefaultColorAspects_l();
633
634 if (defaultColorAspects->primaries == C2Color::PRIMARIES_BT2020 &&
635 defaultColorAspects->matrix == C2Color::MATRIX_BT2020 &&
636 defaultColorAspects->transfer == C2Color::TRANSFER_ST2084) {
637 format = HAL_PIXEL_FORMAT_RGBA_1010102;
638 }
639 }
640 C2MemoryUsage usage = {C2MemoryUsage::CPU_READ, C2MemoryUsage::CPU_WRITE};
641
642 c2_status_t err = pool->fetchGraphicBlock(align(mWidth, 16), mHeight, format,
643 usage, &block);
644
645 if (err != C2_OK) {
646 ALOGE("fetchGraphicBlock for Output failed with status %d", err);
647 work->result = err;
648 return false;
649 }
650
651 C2GraphicView wView = block->map().get();
652
653 if (wView.error()) {
654 ALOGE("graphic view map failed %d", wView.error());
655 work->result = C2_CORRUPTED;
656 return false;
657 }
658
659 ALOGV("provided (%dx%d) required (%dx%d), out frameindex %d", block->width(),
660 block->height(), mWidth, mHeight, (int)buffer->user_private_data);
661
ming.zhouac19c3d2019-10-11 11:14:07 +0800662 uint8_t *dstY = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_Y]);
663 uint8_t *dstU = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_U]);
664 uint8_t *dstV = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_V]);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700665 size_t srcYStride = buffer->stride[0];
666 size_t srcUStride = buffer->stride[1];
667 size_t srcVStride = buffer->stride[2];
668
ming.zhouac19c3d2019-10-11 11:14:07 +0800669 C2PlanarLayout layout = wView.layout();
670 size_t dstYStride = layout.planes[C2PlanarLayout::PLANE_Y].rowInc;
671 size_t dstUVStride = layout.planes[C2PlanarLayout::PLANE_U].rowInc;
672
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700673 if (buffer->bitdepth == 10) {
674 const uint16_t *srcY = (const uint16_t *)buffer->plane[0];
675 const uint16_t *srcU = (const uint16_t *)buffer->plane[1];
676 const uint16_t *srcV = (const uint16_t *)buffer->plane[2];
677
678 if (format == HAL_PIXEL_FORMAT_RGBA_1010102) {
679 convertYUV420Planar16ToY410(
ming.zhouac19c3d2019-10-11 11:14:07 +0800680 (uint32_t *)dstY, srcY, srcU, srcV, srcYStride / 2, srcUStride / 2,
James Zern8dfb3f22020-08-07 18:49:51 -0700681 srcVStride / 2, dstYStride / sizeof(uint32_t), mWidth, mHeight);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700682 } else {
ming.zhouac19c3d2019-10-11 11:14:07 +0800683 convertYUV420Planar16ToYUV420Planar(dstY, dstU, dstV,
684 srcY, srcU, srcV,
685 srcYStride / 2, srcUStride / 2, srcVStride / 2,
686 dstYStride, dstUVStride,
687 mWidth, mHeight);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700688 }
689 } else {
690 const uint8_t *srcY = (const uint8_t *)buffer->plane[0];
691 const uint8_t *srcU = (const uint8_t *)buffer->plane[1];
692 const uint8_t *srcV = (const uint8_t *)buffer->plane[2];
ming.zhouac19c3d2019-10-11 11:14:07 +0800693 copyOutputBufferToYV12Frame(dstY, dstU, dstV,
694 srcY, srcU, srcV,
695 srcYStride, srcUStride, srcVStride,
696 dstYStride, dstUVStride,
697 mWidth, mHeight);
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700698 }
699 finishWork(buffer->user_private_data, work, std::move(block));
700 block = nullptr;
701 return true;
702}
703
704c2_status_t C2SoftGav1Dec::drainInternal(
705 uint32_t drainMode, const std::shared_ptr<C2BlockPool> &pool,
706 const std::unique_ptr<C2Work> &work) {
707 if (drainMode == NO_DRAIN) {
708 ALOGW("drain with NO_DRAIN: no-op");
709 return C2_OK;
710 }
711 if (drainMode == DRAIN_CHAIN) {
712 ALOGW("DRAIN_CHAIN not supported");
713 return C2_OMITTED;
714 }
715
James Zernb7aee6e2020-06-26 13:49:53 -0700716 const Libgav1StatusCode status = mCodecCtx->SignalEOS();
Vignesh Venkatasubramanian0f3e7422019-06-17 16:21:36 -0700717 if (status != kLibgav1StatusOk) {
718 ALOGE("Failed to flush av1 decoder. status: %d.", status);
719 return C2_CORRUPTED;
720 }
721
722 while (outputBuffer(pool, work)) {
723 }
724
725 if (drainMode == DRAIN_COMPONENT_WITH_EOS && work &&
726 work->workletsProcessed == 0u) {
727 fillEmptyWork(work);
728 }
729
730 return C2_OK;
731}
732
733c2_status_t C2SoftGav1Dec::drain(uint32_t drainMode,
734 const std::shared_ptr<C2BlockPool> &pool) {
735 return drainInternal(drainMode, pool, nullptr);
736}
737
Vignesh Venkatasubramanianb6d383d2019-06-10 15:11:58 -0700738class C2SoftGav1Factory : public C2ComponentFactory {
739 public:
740 C2SoftGav1Factory()
741 : mHelper(std::static_pointer_cast<C2ReflectorHelper>(
742 GetCodec2PlatformComponentStore()->getParamReflector())) {}
743
744 virtual c2_status_t createComponent(
745 c2_node_id_t id, std::shared_ptr<C2Component> *const component,
746 std::function<void(C2Component *)> deleter) override {
747 *component = std::shared_ptr<C2Component>(
748 new C2SoftGav1Dec(COMPONENT_NAME, id,
749 std::make_shared<C2SoftGav1Dec::IntfImpl>(mHelper)),
750 deleter);
751 return C2_OK;
752 }
753
754 virtual c2_status_t createInterface(
755 c2_node_id_t id, std::shared_ptr<C2ComponentInterface> *const interface,
756 std::function<void(C2ComponentInterface *)> deleter) override {
757 *interface = std::shared_ptr<C2ComponentInterface>(
758 new SimpleInterface<C2SoftGav1Dec::IntfImpl>(
759 COMPONENT_NAME, id,
760 std::make_shared<C2SoftGav1Dec::IntfImpl>(mHelper)),
761 deleter);
762 return C2_OK;
763 }
764
765 virtual ~C2SoftGav1Factory() override = default;
766
767 private:
768 std::shared_ptr<C2ReflectorHelper> mHelper;
769};
770
771} // namespace android
772
773extern "C" ::C2ComponentFactory *CreateCodec2Factory() {
774 ALOGV("in %s", __func__);
775 return new ::android::C2SoftGav1Factory();
776}
777
778extern "C" void DestroyCodec2Factory(::C2ComponentFactory *factory) {
779 ALOGV("in %s", __func__);
780 delete factory;
781}