/* * Copyright 2018 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ //#define LOG_NDEBUG 0 #ifdef MPEG4 #define LOG_TAG "C2SoftMpeg4Enc" #else #define LOG_TAG "C2SoftH263Enc" #endif #include <log/log.h> #include <inttypes.h> #include <media/hardware/VideoAPI.h> #include <media/stagefright/foundation/AUtils.h> #include <media/stagefright/MediaDefs.h> #include <utils/misc.h> #include <C2Debug.h> #include <C2PlatformSupport.h> #include <SimpleC2Interface.h> #include <util/C2InterfaceHelper.h> #include "C2SoftMpeg4Enc.h" #include "mp4enc_api.h" namespace android { #ifdef MPEG4 constexpr char COMPONENT_NAME[] = "c2.android.mpeg4.encoder"; #else constexpr char COMPONENT_NAME[] = "c2.android.h263.encoder"; #endif class C2SoftMpeg4Enc::IntfImpl : public C2InterfaceHelper { public: explicit IntfImpl(const std::shared_ptr<C2ReflectorHelper>& helper) : C2InterfaceHelper(helper) { setDerivedInstance(this); addParameter( DefineParam(mInputFormat, C2_NAME_INPUT_STREAM_FORMAT_SETTING) .withConstValue( new C2StreamFormatConfig::input(0u, C2FormatVideo)) .build()); addParameter( DefineParam(mOutputFormat, C2_NAME_OUTPUT_STREAM_FORMAT_SETTING) .withConstValue( new C2StreamFormatConfig::output(0u, C2FormatCompressed)) .build()); addParameter( DefineParam(mInputMediaType, C2_NAME_INPUT_PORT_MIME_SETTING) .withConstValue(AllocSharedString<C2PortMimeConfig::input>( MEDIA_MIMETYPE_VIDEO_RAW)) .build()); addParameter( DefineParam(mOutputMediaType, C2_NAME_OUTPUT_PORT_MIME_SETTING) .withConstValue(AllocSharedString<C2PortMimeConfig::output>( #ifdef MPEG4 MEDIA_MIMETYPE_VIDEO_MPEG4 #else MEDIA_MIMETYPE_VIDEO_H263 #endif )) .build()); addParameter(DefineParam(mUsage, C2_NAME_INPUT_STREAM_USAGE_SETTING) .withConstValue(new C2StreamUsageTuning::input( 0u, (uint64_t)C2MemoryUsage::CPU_READ)) .build()); addParameter( DefineParam(mSize, C2_NAME_STREAM_VIDEO_SIZE_SETTING) .withDefault(new C2VideoSizeStreamTuning::input(0u, 176, 144)) .withFields({ C2F(mSize, width).inRange(16, 176, 16), C2F(mSize, height).inRange(16, 144, 16), }) .withSetter(SizeSetter) .build()); addParameter( DefineParam(mFrameRate, C2_NAME_STREAM_FRAME_RATE_SETTING) .withDefault(new C2StreamFrameRateInfo::output(0u, 17.)) // TODO: More restriction? .withFields({C2F(mFrameRate, value).greaterThan(0.)}) .withSetter( Setter<decltype(*mFrameRate)>::StrictValueWithNoDeps) .build()); addParameter( DefineParam(mBitrate, C2_NAME_STREAM_BITRATE_SETTING) .withDefault(new C2BitrateTuning::output(0u, 64000)) .withFields({C2F(mBitrate, value).inRange(1, 12000000)}) .withSetter( Setter<decltype(*mBitrate)>::NonStrictValueWithNoDeps) .build()); } static C2R SizeSetter(bool mayBlock, C2P<C2VideoSizeStreamTuning::input>& me) { (void)mayBlock; // TODO: maybe apply block limit? C2R res = C2R::Ok(); if (!me.F(me.v.width).supportsAtAll(me.v.width)) { res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.width))); } if (!me.F(me.v.height).supportsAtAll(me.v.height)) { res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.height))); } return res; } uint32_t getWidth() const { return mSize->width; } uint32_t getHeight() const { return mSize->height; } float getFrameRate() const { return mFrameRate->value; } uint32_t getBitrate() const { return mBitrate->value; } private: std::shared_ptr<C2StreamFormatConfig::input> mInputFormat; std::shared_ptr<C2StreamFormatConfig::output> mOutputFormat; std::shared_ptr<C2PortMimeConfig::input> mInputMediaType; std::shared_ptr<C2PortMimeConfig::output> mOutputMediaType; std::shared_ptr<C2StreamUsageTuning::input> mUsage; std::shared_ptr<C2VideoSizeStreamTuning::input> mSize; std::shared_ptr<C2StreamFrameRateInfo::output> mFrameRate; std::shared_ptr<C2BitrateTuning::output> mBitrate; }; C2SoftMpeg4Enc::C2SoftMpeg4Enc(const char* name, c2_node_id_t id, const std::shared_ptr<IntfImpl>& intfImpl) : SimpleC2Component( std::make_shared<SimpleInterface<IntfImpl>>(name, id, intfImpl)), mIntf(intfImpl), mHandle(nullptr), mEncParams(nullptr), mStarted(false), mOutBufferSize(524288), mKeyFrameInterval(10) { } C2SoftMpeg4Enc::~C2SoftMpeg4Enc() { onRelease(); } c2_status_t C2SoftMpeg4Enc::onInit() { #ifdef MPEG4 mEncodeMode = COMBINE_MODE_WITH_ERR_RES; #else mEncodeMode = H263_MODE; #endif if (!mHandle) { mHandle = new tagvideoEncControls; } if (!mEncParams) { mEncParams = new tagvideoEncOptions; } mSignalledOutputEos = false; mSignalledError = false; return initEncoder(); } c2_status_t C2SoftMpeg4Enc::onStop() { if (!mStarted) { return C2_OK; } if (mHandle) { (void)PVCleanUpVideoEncoder(mHandle); } mStarted = false; mSignalledOutputEos = false; mSignalledError = false; return C2_OK; } void C2SoftMpeg4Enc::onReset() { onStop(); initEncoder(); } void C2SoftMpeg4Enc::onRelease() { onStop(); if (mEncParams) { delete mEncParams; mEncParams = nullptr; } if (mHandle) { delete mHandle; mHandle = nullptr; } } c2_status_t C2SoftMpeg4Enc::onFlush_sm() { return C2_OK; } static void fillEmptyWork(const std::unique_ptr<C2Work> &work) { uint32_t flags = 0; if (work->input.flags & C2FrameData::FLAG_END_OF_STREAM) { flags |= C2FrameData::FLAG_END_OF_STREAM; ALOGV("signalling eos"); } work->worklets.front()->output.flags = (C2FrameData::flags_t)flags; work->worklets.front()->output.buffers.clear(); work->worklets.front()->output.ordinal = work->input.ordinal; work->workletsProcessed = 1u; } c2_status_t C2SoftMpeg4Enc::initEncParams() { if (mHandle) { memset(mHandle, 0, sizeof(tagvideoEncControls)); } else return C2_CORRUPTED; if (mEncParams) { memset(mEncParams, 0, sizeof(tagvideoEncOptions)); } else return C2_CORRUPTED; if (!PVGetDefaultEncOption(mEncParams, 0)) { ALOGE("Failed to get default encoding parameters"); return C2_CORRUPTED; } if (mIntf->getFrameRate() == 0) { ALOGE("Framerate should not be 0"); return C2_BAD_VALUE; } mEncParams->encMode = mEncodeMode; mEncParams->encWidth[0] = mIntf->getWidth(); mEncParams->encHeight[0] = mIntf->getHeight(); mEncParams->encFrameRate[0] = mIntf->getFrameRate(); mEncParams->rcType = VBR_1; mEncParams->vbvDelay = 5.0f; mEncParams->profile_level = CORE_PROFILE_LEVEL2; mEncParams->packetSize = 32; mEncParams->rvlcEnable = PV_OFF; mEncParams->numLayers = 1; mEncParams->timeIncRes = 1000; mEncParams->tickPerSrc = mEncParams->timeIncRes / mIntf->getFrameRate(); mEncParams->bitRate[0] = mIntf->getBitrate(); mEncParams->iQuant[0] = 15; mEncParams->pQuant[0] = 12; mEncParams->quantType[0] = 0; mEncParams->noFrameSkipped = PV_OFF; // PV's MPEG4 encoder requires the video dimension of multiple if (mIntf->getWidth() % 16 != 0 || mIntf->getHeight() % 16 != 0) { ALOGE("Video frame size %dx%d must be a multiple of 16", mIntf->getWidth(), mIntf->getHeight()); return C2_BAD_VALUE; } // Set IDR frame refresh interval mEncParams->intraPeriod = mKeyFrameInterval; mEncParams->numIntraMB = 0; mEncParams->sceneDetect = PV_ON; mEncParams->searchRange = 16; mEncParams->mv8x8Enable = PV_OFF; mEncParams->gobHeaderInterval = 0; mEncParams->useACPred = PV_ON; mEncParams->intraDCVlcTh = 0; return C2_OK; } c2_status_t C2SoftMpeg4Enc::initEncoder() { if (mStarted) { return C2_OK; } c2_status_t err = initEncParams(); if (C2_OK != err) { ALOGE("Failed to initialized encoder params"); mSignalledError = true; return err; } if (!PVInitVideoEncoder(mHandle, mEncParams)) { ALOGE("Failed to initialize the encoder"); mSignalledError = true; return C2_CORRUPTED; } // 1st buffer for codec specific data mNumInputFrames = -1; mStarted = true; return C2_OK; } void C2SoftMpeg4Enc::process( const std::unique_ptr<C2Work> &work, const std::shared_ptr<C2BlockPool> &pool) { work->result = C2_OK; work->workletsProcessed = 0u; if (mSignalledError || mSignalledOutputEos) { work->result = C2_BAD_VALUE; return; } // Initialize encoder if not already initialized if (!mStarted && C2_OK != initEncoder()) { ALOGE("Failed to initialize encoder"); mSignalledError = true; work->result = C2_CORRUPTED; return; } std::shared_ptr<C2LinearBlock> block; C2MemoryUsage usage = { C2MemoryUsage::CPU_READ, C2MemoryUsage::CPU_WRITE }; c2_status_t err = pool->fetchLinearBlock(mOutBufferSize, usage, &block); if (err != C2_OK) { ALOGE("fetchLinearBlock for Output failed with status %d", err); work->result = C2_NO_MEMORY; return; } C2WriteView wView = block->map().get(); if (wView.error()) { ALOGE("write view map failed %d", wView.error()); work->result = wView.error(); return; } uint8_t *outPtr = (uint8_t *)wView.data(); if (mNumInputFrames < 0) { // The very first thing we want to output is the codec specific data. int32_t outputSize = mOutBufferSize; if (!PVGetVolHeader(mHandle, outPtr, &outputSize, 0)) { ALOGE("Failed to get VOL header"); work->result = C2_CORRUPTED; mSignalledError = true; return; } else { ALOGV("Bytes Generated in header %d\n", outputSize); } ++mNumInputFrames; std::unique_ptr<C2StreamCsdInfo::output> csd = C2StreamCsdInfo::output::AllocUnique(outputSize, 0u); memcpy(csd->m.value, outPtr, outputSize); work->worklets.front()->output.configUpdate.push_back(std::move(csd)); } std::shared_ptr<const C2GraphicView> rView; std::shared_ptr<C2Buffer> inputBuffer; if (!work->input.buffers.empty()) { inputBuffer = work->input.buffers[0]; rView = std::make_shared<const C2GraphicView>( inputBuffer->data().graphicBlocks().front().map().get()); if (rView->error() != C2_OK) { ALOGE("graphic view map err = %d", rView->error()); work->result = rView->error(); return; } } else { ALOGE("Empty input Buffer"); work->result = C2_BAD_VALUE; return; } uint64_t inputTimeStamp = work->input.ordinal.timestamp.peekull(); bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0); const C2ConstGraphicBlock inBuffer = inputBuffer->data().graphicBlocks().front(); if (inBuffer.width() == 0 || inBuffer.height() == 0) { fillEmptyWork(work); if (eos) { mSignalledOutputEos = true; ALOGV("signalled EOS"); } return; } if (inBuffer.width() < mIntf->getWidth() || inBuffer.height() < mIntf->getHeight()) { /* Expect width height to be configured */ ALOGW("unexpected Capacity Aspect %d(%d) x %d(%d)", inBuffer.width(), mIntf->getWidth(), inBuffer.height(), mIntf->getHeight()); work->result = C2_BAD_VALUE; return; } const C2PlanarLayout &layout = rView->layout(); uint8_t *yPlane = const_cast<uint8_t *>(rView->data()[C2PlanarLayout::PLANE_Y]); uint8_t *uPlane = const_cast<uint8_t *>(rView->data()[C2PlanarLayout::PLANE_U]); uint8_t *vPlane = const_cast<uint8_t *>(rView->data()[C2PlanarLayout::PLANE_V]); int32_t yStride = layout.planes[C2PlanarLayout::PLANE_Y].rowInc; int32_t uStride = layout.planes[C2PlanarLayout::PLANE_U].rowInc; int32_t vStride = layout.planes[C2PlanarLayout::PLANE_V].rowInc; uint32_t width = mIntf->getWidth(); uint32_t height = mIntf->getHeight(); // width and height are always even (as block size is 16x16) CHECK_EQ((width & 1u), 0u); CHECK_EQ((height & 1u), 0u); size_t yPlaneSize = width * height; switch (layout.type) { case C2PlanarLayout::TYPE_RGB: // fall-through case C2PlanarLayout::TYPE_RGBA: { MemoryBlock conversionBuffer = mConversionBuffers.fetch(yPlaneSize * 3 / 2); mConversionBuffersInUse.emplace(conversionBuffer.data(), conversionBuffer); yPlane = conversionBuffer.data(); uPlane = yPlane + yPlaneSize; vPlane = uPlane + yPlaneSize / 4; yStride = width; uStride = vStride = width / 2; ConvertRGBToPlanarYUV(yPlane, yStride, height, conversionBuffer.size(), *rView.get()); break; } case C2PlanarLayout::TYPE_YUV: { if (!IsYUV420(*rView)) { ALOGE("input is not YUV420"); work->result = C2_BAD_VALUE; break; } if (layout.planes[layout.PLANE_Y].colInc == 1 && layout.planes[layout.PLANE_U].colInc == 1 && layout.planes[layout.PLANE_V].colInc == 1) { // I420 compatible - planes are already set up above break; } // copy to I420 MemoryBlock conversionBuffer = mConversionBuffers.fetch(yPlaneSize * 3 / 2); mConversionBuffersInUse.emplace(conversionBuffer.data(), conversionBuffer); MediaImage2 img = CreateYUV420PlanarMediaImage2(width, height, width, height); status_t err = ImageCopy(conversionBuffer.data(), &img, *rView); if (err != OK) { ALOGE("Buffer conversion failed: %d", err); work->result = C2_BAD_VALUE; return; } yPlane = conversionBuffer.data(); uPlane = yPlane + yPlaneSize; vPlane = uPlane + yPlaneSize / 4; yStride = width; uStride = vStride = width / 2; break; } case C2PlanarLayout::TYPE_YUVA: ALOGE("YUVA plane type is not supported"); work->result = C2_BAD_VALUE; return; default: ALOGE("Unrecognized plane type: %d", layout.type); work->result = C2_BAD_VALUE; return; } CHECK(NULL != yPlane); /* Encode frames */ VideoEncFrameIO vin, vout; memset(&vin, 0, sizeof(vin)); memset(&vout, 0, sizeof(vout)); vin.yChan = yPlane; vin.uChan = uPlane; vin.vChan = vPlane; vin.timestamp = (inputTimeStamp + 500) / 1000; // in ms vin.height = align(height, 16); vin.pitch = align(width, 16); uint32_t modTimeMs = 0; int32_t nLayer = 0; MP4HintTrack hintTrack; int32_t outputSize = mOutBufferSize; if (!PVEncodeVideoFrame(mHandle, &vin, &vout, &modTimeMs, outPtr, &outputSize, &nLayer) || !PVGetHintTrack(mHandle, &hintTrack)) { ALOGE("Failed to encode frame or get hint track at frame %" PRId64, mNumInputFrames); mSignalledError = true; work->result = C2_CORRUPTED; return; } ALOGV("outputSize filled : %d", outputSize); ++mNumInputFrames; CHECK(NULL == PVGetOverrunBuffer(mHandle)); fillEmptyWork(work); if (outputSize) { std::shared_ptr<C2Buffer> buffer = createLinearBuffer(block, 0, outputSize); work->worklets.front()->output.buffers.push_back(buffer); work->worklets.front()->output.ordinal.timestamp = inputTimeStamp; if (hintTrack.CodeType == 0) { buffer->setInfo(std::make_shared<C2StreamPictureTypeMaskInfo::output>( 0u /* stream id */, C2PictureTypeKeyFrame)); } } if (eos) { mSignalledOutputEos = true; } mConversionBuffersInUse.erase(yPlane); } c2_status_t C2SoftMpeg4Enc::drain( uint32_t drainMode, const std::shared_ptr<C2BlockPool> &pool) { (void)pool; if (drainMode == NO_DRAIN) { ALOGW("drain with NO_DRAIN: no-op"); return C2_OK; } if (drainMode == DRAIN_CHAIN) { ALOGW("DRAIN_CHAIN not supported"); return C2_OMITTED; } return C2_OK; } class C2SoftMpeg4EncFactory : public C2ComponentFactory { public: C2SoftMpeg4EncFactory() : mHelper(std::static_pointer_cast<C2ReflectorHelper>( GetCodec2PlatformComponentStore()->getParamReflector())) {} virtual c2_status_t createComponent( c2_node_id_t id, std::shared_ptr<C2Component>* const component, std::function<void(C2Component*)> deleter) override { *component = std::shared_ptr<C2Component>( new C2SoftMpeg4Enc( COMPONENT_NAME, id, std::make_shared<C2SoftMpeg4Enc::IntfImpl>(mHelper)), deleter); return C2_OK; } virtual c2_status_t createInterface( c2_node_id_t id, std::shared_ptr<C2ComponentInterface>* const interface, std::function<void(C2ComponentInterface*)> deleter) override { *interface = std::shared_ptr<C2ComponentInterface>( new SimpleInterface<C2SoftMpeg4Enc::IntfImpl>( COMPONENT_NAME, id, std::make_shared<C2SoftMpeg4Enc::IntfImpl>(mHelper)), deleter); return C2_OK; } virtual ~C2SoftMpeg4EncFactory() override = default; private: std::shared_ptr<C2ReflectorHelper> mHelper; }; } // namespace android extern "C" ::C2ComponentFactory* CreateCodec2Factory() { ALOGV("in %s", __func__); return new ::android::C2SoftMpeg4EncFactory(); } extern "C" void DestroyCodec2Factory(::C2ComponentFactory* factory) { ALOGV("in %s", __func__); delete factory; }