/* * Copyright 2018 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ //#define LOG_NDEBUG 0 #define LOG_TAG "C2SoftVpxEnc" #include <log/log.h> #include <utils/misc.h> #include <media/hardware/VideoAPI.h> #include <Codec2BufferUtils.h> #include <C2Debug.h> #include "C2SoftVpxEnc.h" #ifndef INT32_MAX #define INT32_MAX 2147483647 #endif namespace android { static size_t getCpuCoreCount() { long cpuCoreCount = 1; #if defined(_SC_NPROCESSORS_ONLN) cpuCoreCount = sysconf(_SC_NPROCESSORS_ONLN); #else // _SC_NPROC_ONLN must be defined... cpuCoreCount = sysconf(_SC_NPROC_ONLN); #endif CHECK(cpuCoreCount >= 1); ALOGV("Number of CPU cores: %ld", cpuCoreCount); return (size_t)cpuCoreCount; } C2SoftVpxEnc::C2SoftVpxEnc(const char* name, c2_node_id_t id, const std::shared_ptr<IntfImpl>& intfImpl) : SimpleC2Component( std::make_shared<SimpleInterface<IntfImpl>>(name, id, intfImpl)), mIntf(intfImpl), mCodecContext(nullptr), mCodecConfiguration(nullptr), mCodecInterface(nullptr), mStrideAlign(2), mColorFormat(VPX_IMG_FMT_I420), mBitrateUpdated(false), mBitrateControlMode(VPX_VBR), mErrorResilience(false), mKeyFrameInterval(0), mMinQuantizer(0), mMaxQuantizer(0), mTemporalLayers(0), mTemporalPatternType(VPXTemporalLayerPatternNone), mTemporalPatternLength(0), mTemporalPatternIdx(0), mLastTimestamp(0x7FFFFFFFFFFFFFFFull), mKeyFrameRequested(false), mSignalledOutputEos(false), mSignalledError(false) { memset(mTemporalLayerBitrateRatio, 0, sizeof(mTemporalLayerBitrateRatio)); mTemporalLayerBitrateRatio[0] = 100; } C2SoftVpxEnc::~C2SoftVpxEnc() { onRelease(); } c2_status_t C2SoftVpxEnc::onInit() { status_t err = initEncoder(); return err == OK ? C2_OK : C2_CORRUPTED; } void C2SoftVpxEnc::onRelease() { if (mCodecContext) { vpx_codec_destroy(mCodecContext); delete mCodecContext; mCodecContext = nullptr; } if (mCodecConfiguration) { delete mCodecConfiguration; mCodecConfiguration = nullptr; } // this one is not allocated by us mCodecInterface = nullptr; } c2_status_t C2SoftVpxEnc::onStop() { onRelease(); mLastTimestamp = 0x7FFFFFFFFFFFFFFFLL; mSignalledOutputEos = false; mSignalledError = false; return C2_OK; } void C2SoftVpxEnc::onReset() { (void)onStop(); } c2_status_t C2SoftVpxEnc::onFlush_sm() { return onStop(); } status_t C2SoftVpxEnc::initEncoder() { vpx_codec_err_t codec_return; status_t result = UNKNOWN_ERROR; setCodecSpecificInterface(); if (!mCodecInterface) goto CleanUp; ALOGD("VPx: initEncoder. BRMode: %u. TSLayers: %zu. KF: %u. QP: %u - %u", (uint32_t)mBitrateControlMode, mTemporalLayers, mKeyFrameInterval, mMinQuantizer, mMaxQuantizer); mCodecConfiguration = new vpx_codec_enc_cfg_t; if (!mCodecConfiguration) goto CleanUp; codec_return = vpx_codec_enc_config_default(mCodecInterface, mCodecConfiguration, 0); if (codec_return != VPX_CODEC_OK) { ALOGE("Error populating default configuration for vpx encoder."); goto CleanUp; } mCodecConfiguration->g_w = mIntf->getWidth(); mCodecConfiguration->g_h = mIntf->getHeight(); mCodecConfiguration->g_threads = getCpuCoreCount(); mCodecConfiguration->g_error_resilient = mErrorResilience; // timebase unit is microsecond // g_timebase is in seconds (i.e. 1/1000000 seconds) mCodecConfiguration->g_timebase.num = 1; mCodecConfiguration->g_timebase.den = 1000000; // rc_target_bitrate is in kbps, mBitrate in bps mCodecConfiguration->rc_target_bitrate = (mIntf->getBitrate() + 500) / 1000; mCodecConfiguration->rc_end_usage = mBitrateControlMode; // Disable frame drop - not allowed in MediaCodec now. mCodecConfiguration->rc_dropframe_thresh = 0; // Disable lagged encoding. mCodecConfiguration->g_lag_in_frames = 0; if (mBitrateControlMode == VPX_CBR) { // Disable spatial resizing. mCodecConfiguration->rc_resize_allowed = 0; // Single-pass mode. mCodecConfiguration->g_pass = VPX_RC_ONE_PASS; // Maximum amount of bits that can be subtracted from the target // bitrate - expressed as percentage of the target bitrate. mCodecConfiguration->rc_undershoot_pct = 100; // Maximum amount of bits that can be added to the target // bitrate - expressed as percentage of the target bitrate. mCodecConfiguration->rc_overshoot_pct = 15; // Initial value of the buffer level in ms. mCodecConfiguration->rc_buf_initial_sz = 500; // Amount of data that the encoder should try to maintain in ms. mCodecConfiguration->rc_buf_optimal_sz = 600; // The amount of data that may be buffered by the decoding // application in ms. mCodecConfiguration->rc_buf_sz = 1000; // Enable error resilience - needed for packet loss. mCodecConfiguration->g_error_resilient = 1; // Maximum key frame interval - for CBR boost to 3000 mCodecConfiguration->kf_max_dist = 3000; // Encoder determines optimal key frame placement automatically. mCodecConfiguration->kf_mode = VPX_KF_AUTO; } // Frames temporal pattern - for now WebRTC like pattern is only supported. switch (mTemporalLayers) { case 0: mTemporalPatternLength = 0; break; case 1: mCodecConfiguration->ts_number_layers = 1; mCodecConfiguration->ts_rate_decimator[0] = 1; mCodecConfiguration->ts_periodicity = 1; mCodecConfiguration->ts_layer_id[0] = 0; mTemporalPattern[0] = kTemporalUpdateLastRefAll; mTemporalPatternLength = 1; break; case 2: mCodecConfiguration->ts_number_layers = 2; mCodecConfiguration->ts_rate_decimator[0] = 2; mCodecConfiguration->ts_rate_decimator[1] = 1; mCodecConfiguration->ts_periodicity = 2; mCodecConfiguration->ts_layer_id[0] = 0; mCodecConfiguration->ts_layer_id[1] = 1; mTemporalPattern[0] = kTemporalUpdateLastAndGoldenRefAltRef; mTemporalPattern[1] = kTemporalUpdateGoldenWithoutDependencyRefAltRef; mTemporalPattern[2] = kTemporalUpdateLastRefAltRef; mTemporalPattern[3] = kTemporalUpdateGoldenRefAltRef; mTemporalPattern[4] = kTemporalUpdateLastRefAltRef; mTemporalPattern[5] = kTemporalUpdateGoldenRefAltRef; mTemporalPattern[6] = kTemporalUpdateLastRefAltRef; mTemporalPattern[7] = kTemporalUpdateNone; mTemporalPatternLength = 8; break; case 3: mCodecConfiguration->ts_number_layers = 3; mCodecConfiguration->ts_rate_decimator[0] = 4; mCodecConfiguration->ts_rate_decimator[1] = 2; mCodecConfiguration->ts_rate_decimator[2] = 1; mCodecConfiguration->ts_periodicity = 4; mCodecConfiguration->ts_layer_id[0] = 0; mCodecConfiguration->ts_layer_id[1] = 2; mCodecConfiguration->ts_layer_id[2] = 1; mCodecConfiguration->ts_layer_id[3] = 2; mTemporalPattern[0] = kTemporalUpdateLastAndGoldenRefAltRef; mTemporalPattern[1] = kTemporalUpdateNoneNoRefGoldenRefAltRef; mTemporalPattern[2] = kTemporalUpdateGoldenWithoutDependencyRefAltRef; mTemporalPattern[3] = kTemporalUpdateNone; mTemporalPattern[4] = kTemporalUpdateLastRefAltRef; mTemporalPattern[5] = kTemporalUpdateNone; mTemporalPattern[6] = kTemporalUpdateGoldenRefAltRef; mTemporalPattern[7] = kTemporalUpdateNone; mTemporalPatternLength = 8; break; default: ALOGE("Wrong number of temporal layers %zu", mTemporalLayers); goto CleanUp; } // Set bitrate values for each layer for (size_t i = 0; i < mCodecConfiguration->ts_number_layers; i++) { mCodecConfiguration->ts_target_bitrate[i] = mCodecConfiguration->rc_target_bitrate * mTemporalLayerBitrateRatio[i] / 100; } if (mKeyFrameInterval > 0) { mCodecConfiguration->kf_max_dist = mKeyFrameInterval; mCodecConfiguration->kf_min_dist = mKeyFrameInterval; mCodecConfiguration->kf_mode = VPX_KF_AUTO; } if (mMinQuantizer > 0) { mCodecConfiguration->rc_min_quantizer = mMinQuantizer; } if (mMaxQuantizer > 0) { mCodecConfiguration->rc_max_quantizer = mMaxQuantizer; } setCodecSpecificConfiguration(); mCodecContext = new vpx_codec_ctx_t; if (!mCodecContext) goto CleanUp; codec_return = vpx_codec_enc_init(mCodecContext, mCodecInterface, mCodecConfiguration, 0); // flags if (codec_return != VPX_CODEC_OK) { ALOGE("Error initializing vpx encoder"); goto CleanUp; } // Extra CBR settings if (mBitrateControlMode == VPX_CBR) { codec_return = vpx_codec_control(mCodecContext, VP8E_SET_STATIC_THRESHOLD, 1); if (codec_return == VPX_CODEC_OK) { uint32_t rc_max_intra_target = mCodecConfiguration->rc_buf_optimal_sz * ((uint32_t)mIntf->getFrameRate() >> 1) / 10; // Don't go below 3 times per frame bandwidth. if (rc_max_intra_target < 300) { rc_max_intra_target = 300; } codec_return = vpx_codec_control(mCodecContext, VP8E_SET_MAX_INTRA_BITRATE_PCT, rc_max_intra_target); } if (codec_return == VPX_CODEC_OK) { codec_return = vpx_codec_control(mCodecContext, VP8E_SET_CPUUSED, -8); } if (codec_return != VPX_CODEC_OK) { ALOGE("Error setting cbr parameters for vpx encoder."); goto CleanUp; } } codec_return = setCodecSpecificControls(); if (codec_return != VPX_CODEC_OK) goto CleanUp; { uint32_t width = mIntf->getWidth(); uint32_t height = mIntf->getHeight(); if (((uint64_t)width * height) > ((uint64_t)INT32_MAX / 3)) { ALOGE("b/25812794, Buffer size is too big, width=%u, height=%u.", width, height); } else { uint32_t stride = (width + mStrideAlign - 1) & ~(mStrideAlign - 1); uint32_t vstride = (height + mStrideAlign - 1) & ~(mStrideAlign - 1); mConversionBuffer = MemoryBlock::Allocate(stride * vstride * 3 / 2); if (!mConversionBuffer.size()) { ALOGE("Allocating conversion buffer failed."); } else { mNumInputFrames = -1; return OK; } } } CleanUp: onRelease(); return result; } vpx_enc_frame_flags_t C2SoftVpxEnc::getEncodeFlags() { vpx_enc_frame_flags_t flags = 0; if (mTemporalPatternLength > 0) { int patternIdx = mTemporalPatternIdx % mTemporalPatternLength; mTemporalPatternIdx++; switch (mTemporalPattern[patternIdx]) { case kTemporalUpdateLast: flags |= VP8_EFLAG_NO_UPD_GF; flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_REF_GF; flags |= VP8_EFLAG_NO_REF_ARF; break; case kTemporalUpdateGoldenWithoutDependency: flags |= VP8_EFLAG_NO_REF_GF; // Deliberately no break here. case kTemporalUpdateGolden: flags |= VP8_EFLAG_NO_REF_ARF; flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_UPD_LAST; break; case kTemporalUpdateAltrefWithoutDependency: flags |= VP8_EFLAG_NO_REF_ARF; flags |= VP8_EFLAG_NO_REF_GF; // Deliberately no break here. case kTemporalUpdateAltref: flags |= VP8_EFLAG_NO_UPD_GF; flags |= VP8_EFLAG_NO_UPD_LAST; break; case kTemporalUpdateNoneNoRefAltref: flags |= VP8_EFLAG_NO_REF_ARF; // Deliberately no break here. case kTemporalUpdateNone: flags |= VP8_EFLAG_NO_UPD_GF; flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_UPD_LAST; flags |= VP8_EFLAG_NO_UPD_ENTROPY; break; case kTemporalUpdateNoneNoRefGoldenRefAltRef: flags |= VP8_EFLAG_NO_REF_GF; flags |= VP8_EFLAG_NO_UPD_GF; flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_UPD_LAST; flags |= VP8_EFLAG_NO_UPD_ENTROPY; break; case kTemporalUpdateGoldenWithoutDependencyRefAltRef: flags |= VP8_EFLAG_NO_REF_GF; flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_UPD_LAST; break; case kTemporalUpdateLastRefAltRef: flags |= VP8_EFLAG_NO_UPD_GF; flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_REF_GF; break; case kTemporalUpdateGoldenRefAltRef: flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_UPD_LAST; break; case kTemporalUpdateLastAndGoldenRefAltRef: flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_REF_GF; break; case kTemporalUpdateLastRefAll: flags |= VP8_EFLAG_NO_UPD_ARF; flags |= VP8_EFLAG_NO_UPD_GF; break; } } return flags; } // TODO: add support for YUV input color formats // TODO: add support for SVC, ARF. SVC and ARF returns multiple frames // (hierarchical / noshow) in one call. These frames should be combined in to // a single buffer and sent back to the client void C2SoftVpxEnc::process( const std::unique_ptr<C2Work> &work, const std::shared_ptr<C2BlockPool> &pool) { work->result = C2_OK; work->workletsProcessed = 1u; if (mSignalledError || mSignalledOutputEos) { work->result = C2_BAD_VALUE; return; } // Initialize encoder if not already if (!mCodecContext && OK != initEncoder()) { ALOGE("Failed to initialize encoder"); mSignalledError = true; work->result = C2_CORRUPTED; return; } if (mNumInputFrames < 0) { ++mNumInputFrames; std::unique_ptr<C2StreamCsdInfo::output> csd = C2StreamCsdInfo::output::AllocUnique(0, 0u); work->worklets.front()->output.configUpdate.push_back(std::move(csd)); } std::shared_ptr<const C2GraphicView> rView; std::shared_ptr<C2Buffer> inputBuffer; if (!work->input.buffers.empty()) { inputBuffer = work->input.buffers[0]; rView = std::make_shared<const C2GraphicView>( inputBuffer->data().graphicBlocks().front().map().get()); if (rView->error() != C2_OK) { ALOGE("graphic view map err = %d", rView->error()); work->result = C2_CORRUPTED; return; } } else { ALOGE("Empty input Buffer"); work->result = C2_BAD_VALUE; return; } const C2ConstGraphicBlock inBuffer = inputBuffer->data().graphicBlocks().front(); if (inBuffer.width() != mIntf->getWidth() || inBuffer.height() != mIntf->getHeight()) { ALOGE("unexpected Input buffer attributes %d(%d) x %d(%d)", inBuffer.width(), mIntf->getWidth(), inBuffer.height(), mIntf->getHeight()); work->result = C2_BAD_VALUE; return; } bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0); vpx_image_t raw_frame; const C2PlanarLayout &layout = rView->layout(); uint32_t width = rView->width(); uint32_t height = rView->height(); if (width > 0x8000 || height > 0x8000) { ALOGE("Image too big: %u x %u", width, height); work->result = C2_BAD_VALUE; return; } uint32_t stride = (width + mStrideAlign - 1) & ~(mStrideAlign - 1); uint32_t vstride = (height + mStrideAlign - 1) & ~(mStrideAlign - 1); switch (layout.type) { case C2PlanarLayout::TYPE_RGB: case C2PlanarLayout::TYPE_RGBA: { ConvertRGBToPlanarYUV(mConversionBuffer.data(), stride, vstride, mConversionBuffer.size(), *rView.get()); vpx_img_wrap(&raw_frame, VPX_IMG_FMT_I420, width, height, mStrideAlign, mConversionBuffer.data()); break; } case C2PlanarLayout::TYPE_YUV: { if (!IsYUV420(*rView)) { ALOGE("input is not YUV420"); work->result = C2_BAD_VALUE; return; } if (layout.planes[layout.PLANE_Y].colInc == 1 && layout.planes[layout.PLANE_U].colInc == 1 && layout.planes[layout.PLANE_V].colInc == 1) { // I420 compatible - though with custom offset and stride vpx_img_wrap(&raw_frame, VPX_IMG_FMT_I420, width, height, mStrideAlign, (uint8_t*)rView->data()[0]); raw_frame.planes[1] = (uint8_t*)rView->data()[1]; raw_frame.planes[2] = (uint8_t*)rView->data()[2]; raw_frame.stride[0] = layout.planes[layout.PLANE_Y].rowInc; raw_frame.stride[1] = layout.planes[layout.PLANE_U].rowInc; raw_frame.stride[2] = layout.planes[layout.PLANE_V].rowInc; } else { // copy to I420 MediaImage2 img = CreateYUV420PlanarMediaImage2(width, height, stride, vstride); if (mConversionBuffer.size() >= stride * vstride * 3 / 2) { status_t err = ImageCopy(mConversionBuffer.data(), &img, *rView); if (err != OK) { ALOGE("Buffer conversion failed: %d", err); work->result = C2_BAD_VALUE; return; } vpx_img_wrap(&raw_frame, VPX_IMG_FMT_I420, stride, vstride, mStrideAlign, (uint8_t*)rView->data()[0]); vpx_img_set_rect(&raw_frame, 0, 0, width, height); } else { ALOGE("Conversion buffer is too small: %u x %u for %zu", stride, vstride, mConversionBuffer.size()); work->result = C2_BAD_VALUE; return; } } break; } default: ALOGE("Unrecognized plane type: %d", layout.type); work->result = C2_BAD_VALUE; return; } vpx_enc_frame_flags_t flags = getEncodeFlags(); if (mKeyFrameRequested) { flags |= VPX_EFLAG_FORCE_KF; mKeyFrameRequested = false; } uint64_t inputTimeStamp = work->input.ordinal.timestamp.peekull(); uint32_t frameDuration; if (inputTimeStamp > mLastTimestamp) { frameDuration = (uint32_t)(inputTimeStamp - mLastTimestamp); } else { // Use default of 30 fps in case of 0 frame rate. uint32_t framerate = mIntf->getFrameRate() ?: 30; frameDuration = (uint32_t)((uint64_t)1000000 / framerate); } mLastTimestamp = inputTimeStamp; if (mBitrateUpdated) { mCodecConfiguration->rc_target_bitrate = (mIntf->getBitrate() + 500) / 1000; vpx_codec_err_t res = vpx_codec_enc_config_set(mCodecContext, mCodecConfiguration); if (res != VPX_CODEC_OK) { ALOGE("vpx encoder failed to update bitrate: %s", vpx_codec_err_to_string(res)); work->result = C2_CORRUPTED; return; } mBitrateUpdated = false; } vpx_codec_err_t codec_return = vpx_codec_encode(mCodecContext, &raw_frame, inputTimeStamp, frameDuration, flags, VPX_DL_REALTIME); if (codec_return != VPX_CODEC_OK) { ALOGE("vpx encoder failed to encode frame"); work->result = C2_CORRUPTED; return; } bool populated = false; vpx_codec_iter_t encoded_packet_iterator = nullptr; const vpx_codec_cx_pkt_t* encoded_packet; while ((encoded_packet = vpx_codec_get_cx_data( mCodecContext, &encoded_packet_iterator))) { if (encoded_packet->kind == VPX_CODEC_CX_FRAME_PKT) { std::shared_ptr<C2LinearBlock> block; C2MemoryUsage usage = { C2MemoryUsage::CPU_READ, C2MemoryUsage::CPU_WRITE }; c2_status_t err = pool->fetchLinearBlock(encoded_packet->data.frame.sz, usage, &block); if (err != C2_OK) { ALOGE("fetchLinearBlock for Output failed with status %d", err); work->result = C2_NO_MEMORY; return; } C2WriteView wView = block->map().get(); if (wView.error()) { ALOGE("write view map failed %d", wView.error()); work->result = C2_CORRUPTED; return; } memcpy(wView.data(), encoded_packet->data.frame.buf, encoded_packet->data.frame.sz); ++mNumInputFrames; ALOGD("bytes generated %zu", encoded_packet->data.frame.sz); uint32_t flags = 0; if (eos) { flags |= C2FrameData::FLAG_END_OF_STREAM; } work->worklets.front()->output.flags = (C2FrameData::flags_t)flags; work->worklets.front()->output.buffers.clear(); work->worklets.front()->output.buffers.push_back(createLinearBuffer(block)); work->worklets.front()->output.ordinal = work->input.ordinal; work->worklets.front()->output.ordinal.timestamp = encoded_packet->data.frame.pts; work->workletsProcessed = 1u; populated = true; if (eos) { mSignalledOutputEos = true; ALOGV("signalled EOS"); } } } if (!populated) { work->workletsProcessed = 0u; } } c2_status_t C2SoftVpxEnc::drain( uint32_t drainMode, const std::shared_ptr<C2BlockPool> &pool) { (void)pool; if (drainMode == NO_DRAIN) { ALOGW("drain with NO_DRAIN: no-op"); return C2_OK; } if (drainMode == DRAIN_CHAIN) { ALOGW("DRAIN_CHAIN not supported"); return C2_OMITTED; } return C2_OK; } } // namespace android