/* * Copyright (c) 2009-2011 Intel Corporation. All rights reserved. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ //#define LOG_NDEBUG 0 #define LOG_TAG "OMXVideoDecoder" #include <wrs_omxil_core/log.h> #include <nativebase/nativebase.h> #include <hardware/gralloc.h> #include <va/va_android.h> #include "OMXVideoDecoderBase.h" #include "ProtectedDataBuffer.h" static const char* VA_RAW_MIME_TYPE = "video/x-raw-va"; static const uint32_t VA_COLOR_FORMAT = 0x7FA00E00; OMXVideoDecoderBase::OMXVideoDecoderBase() : mRotationDegrees(0), #ifdef TARGET_HAS_ISV mVppBufferNum(0), #endif mCodecPriority(1), mOperatingRate(0), mVideoDecoder(NULL), mNativeBufferCount(OUTPORT_NATIVE_BUFFER_COUNT), mWorkingMode(RAWDATA_MODE), mErrorReportEnabled (false), mAPMode(LEGACY_MODE), mFlushMode(false), mFormatChanged(false) { mOMXBufferHeaderTypePtrNum = 0; mMetaDataBuffersNum = 0; memset(&mGraphicBufferParam, 0, sizeof(mGraphicBufferParam)); } OMXVideoDecoderBase::~OMXVideoDecoderBase() { releaseVideoDecoder(mVideoDecoder); if (this->ports) { if (this->ports[INPORT_INDEX]) { delete this->ports[INPORT_INDEX]; this->ports[INPORT_INDEX] = NULL; } if (this->ports[OUTPORT_INDEX]) { delete this->ports[OUTPORT_INDEX]; this->ports[OUTPORT_INDEX] = NULL; } } } OMX_ERRORTYPE OMXVideoDecoderBase::InitInputPort(void) { this->ports[INPORT_INDEX] = new PortVideo; if (this->ports[INPORT_INDEX] == NULL) { return OMX_ErrorInsufficientResources; } PortVideo *port = static_cast<PortVideo *>(this->ports[INPORT_INDEX]); // OMX_PARAM_PORTDEFINITIONTYPE OMX_PARAM_PORTDEFINITIONTYPE paramPortDefinitionInput; memset(¶mPortDefinitionInput, 0, sizeof(paramPortDefinitionInput)); SetTypeHeader(¶mPortDefinitionInput, sizeof(paramPortDefinitionInput)); paramPortDefinitionInput.nPortIndex = INPORT_INDEX; paramPortDefinitionInput.eDir = OMX_DirInput; paramPortDefinitionInput.nBufferCountActual = INPORT_ACTUAL_BUFFER_COUNT; paramPortDefinitionInput.nBufferCountMin = INPORT_MIN_BUFFER_COUNT; paramPortDefinitionInput.nBufferSize = INPORT_BUFFER_SIZE; paramPortDefinitionInput.bEnabled = OMX_TRUE; paramPortDefinitionInput.bPopulated = OMX_FALSE; paramPortDefinitionInput.eDomain = OMX_PortDomainVideo; paramPortDefinitionInput.format.video.cMIMEType = NULL; // to be overridden paramPortDefinitionInput.format.video.pNativeRender = NULL; paramPortDefinitionInput.format.video.nFrameWidth = 176; paramPortDefinitionInput.format.video.nFrameHeight = 144; paramPortDefinitionInput.format.video.nStride = 0; paramPortDefinitionInput.format.video.nSliceHeight = 0; paramPortDefinitionInput.format.video.nBitrate = 64000; paramPortDefinitionInput.format.video.xFramerate = 15 << 16; // TODO: check if we need to set bFlagErrorConcealment to OMX_TRUE paramPortDefinitionInput.format.video.bFlagErrorConcealment = OMX_FALSE; paramPortDefinitionInput.format.video.eCompressionFormat = OMX_VIDEO_CodingUnused; // to be overridden paramPortDefinitionInput.format.video.eColorFormat = OMX_COLOR_FormatUnused; paramPortDefinitionInput.format.video.pNativeWindow = NULL; paramPortDefinitionInput.bBuffersContiguous = OMX_FALSE; paramPortDefinitionInput.nBufferAlignment = 0; // Derived class must implement this interface and override any field if needed. // eCompressionFormat and and cMIMEType must be overridden InitInputPortFormatSpecific(¶mPortDefinitionInput); port->SetPortDefinition(¶mPortDefinitionInput, true); // OMX_VIDEO_PARAM_PORTFORMATTYPE OMX_VIDEO_PARAM_PORTFORMATTYPE paramPortFormat; memset(¶mPortFormat, 0, sizeof(paramPortFormat)); SetTypeHeader(¶mPortFormat, sizeof(paramPortFormat)); paramPortFormat.nPortIndex = INPORT_INDEX; paramPortFormat.nIndex = 0; paramPortFormat.eCompressionFormat = paramPortDefinitionInput.format.video.eCompressionFormat; paramPortFormat.eColorFormat = paramPortDefinitionInput.format.video.eColorFormat; paramPortFormat.xFramerate = paramPortDefinitionInput.format.video.xFramerate; port->SetPortVideoParam(¶mPortFormat, true); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::InitOutputPort(void) { this->ports[OUTPORT_INDEX] = new PortVideo; if (this->ports[OUTPORT_INDEX] == NULL) { return OMX_ErrorInsufficientResources; } PortVideo *port = static_cast<PortVideo *>(this->ports[OUTPORT_INDEX]); // OMX_PARAM_PORTDEFINITIONTYPE OMX_PARAM_PORTDEFINITIONTYPE paramPortDefinitionOutput; memset(¶mPortDefinitionOutput, 0, sizeof(paramPortDefinitionOutput)); SetTypeHeader(¶mPortDefinitionOutput, sizeof(paramPortDefinitionOutput)); paramPortDefinitionOutput.nPortIndex = OUTPORT_INDEX; paramPortDefinitionOutput.eDir = OMX_DirOutput; paramPortDefinitionOutput.nBufferCountActual = OUTPORT_ACTUAL_BUFFER_COUNT; paramPortDefinitionOutput.nBufferCountMin = OUTPORT_MIN_BUFFER_COUNT; paramPortDefinitionOutput.nBufferSize = sizeof(VideoRenderBuffer); paramPortDefinitionOutput.bEnabled = OMX_TRUE; paramPortDefinitionOutput.bPopulated = OMX_FALSE; paramPortDefinitionOutput.eDomain = OMX_PortDomainVideo; paramPortDefinitionOutput.format.video.cMIMEType = (OMX_STRING)VA_RAW_MIME_TYPE; paramPortDefinitionOutput.format.video.pNativeRender = NULL; paramPortDefinitionOutput.format.video.nFrameWidth = 176; paramPortDefinitionOutput.format.video.nFrameHeight = 144; paramPortDefinitionOutput.format.video.nStride = 176; paramPortDefinitionOutput.format.video.nSliceHeight = 144; paramPortDefinitionOutput.format.video.nBitrate = 64000; paramPortDefinitionOutput.format.video.xFramerate = 15 << 16; paramPortDefinitionOutput.format.video.bFlagErrorConcealment = OMX_FALSE; paramPortDefinitionOutput.format.video.eCompressionFormat = OMX_VIDEO_CodingUnused; paramPortDefinitionOutput.format.video.eColorFormat = OMX_COLOR_FormatYUV420SemiPlanar; paramPortDefinitionOutput.format.video.pNativeWindow = NULL; paramPortDefinitionOutput.bBuffersContiguous = OMX_FALSE; paramPortDefinitionOutput.nBufferAlignment = 0; // no format specific to initialize output port InitOutputPortFormatSpecific(¶mPortDefinitionOutput); port->SetPortDefinition(¶mPortDefinitionOutput, true); // OMX_VIDEO_PARAM_PORTFORMATTYPE OMX_VIDEO_PARAM_PORTFORMATTYPE paramPortFormat; SetTypeHeader(¶mPortFormat, sizeof(paramPortFormat)); paramPortFormat.nPortIndex = OUTPORT_INDEX; paramPortFormat.nIndex = 0; paramPortFormat.eCompressionFormat = paramPortDefinitionOutput.format.video.eCompressionFormat; paramPortFormat.eColorFormat = paramPortDefinitionOutput.format.video.eColorFormat; paramPortFormat.xFramerate = paramPortDefinitionOutput.format.video.xFramerate; port->SetPortVideoParam(¶mPortFormat, true); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::InitOutputPortFormatSpecific(OMX_PARAM_PORTDEFINITIONTYPE *) { // no format specific to initialize output port return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorInit(void) { OMX_ERRORTYPE ret; ret = OMXComponentCodecBase::ProcessorInit(); CHECK_RETURN_VALUE("OMXComponentCodecBase::ProcessorInit"); if (mVideoDecoder == NULL) { LOGE("ProcessorInit: Video decoder is not created."); return OMX_ErrorDynamicResourcesUnavailable; } VideoConfigBuffer configBuffer; ret = PrepareConfigBuffer(&configBuffer); CHECK_RETURN_VALUE("PrepareConfigBuffer"); //pthread_mutex_lock(&mSerializationLock); Decode_Status status = mVideoDecoder->start(&configBuffer); //pthread_mutex_unlock(&mSerializationLock); if (status != DECODE_SUCCESS) { return TranslateDecodeStatus(status); } return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorReset(void) { OMX_ERRORTYPE ret; VideoConfigBuffer configBuffer; // reset the configbuffer and set it to mix ret = PrepareConfigBuffer(&configBuffer); CHECK_RETURN_VALUE("PrepareConfigBuffer"); mVideoDecoder->reset(&configBuffer); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorDeinit(void) { if (mWorkingMode != GRAPHICBUFFER_MODE) { if (mVideoDecoder == NULL) { LOGE("ProcessorDeinit: Video decoder is not created."); return OMX_ErrorDynamicResourcesUnavailable; } mVideoDecoder->stop(); } mOMXBufferHeaderTypePtrNum = 0; memset(&mGraphicBufferParam, 0, sizeof(mGraphicBufferParam)); mRotationDegrees = 0; #ifdef TARGET_HAS_ISV mVppBufferNum = 0; #endif return OMXComponentCodecBase::ProcessorDeinit(); } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorStart(void) { return OMXComponentCodecBase::ProcessorStart(); } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorStop(void) { // There is no need to return all retained buffers as we don't accumulate buffer //this->ports[INPORT_INDEX]->ReturnAllRetainedBuffers(); // TODO: this is new code ProcessorFlush(OMX_ALL); if (mWorkingMode == GRAPHICBUFFER_MODE) { // for GRAPHICBUFFER_MODE mode, va_destroySurface need to lock the graphicbuffer, // Make sure va_destroySurface is called(ExecutingToIdle) before graphicbuffer is freed(IdleToLoaded). if (mVideoDecoder == NULL) { LOGE("ProcessorStop: Video decoder is not created."); return OMX_ErrorDynamicResourcesUnavailable; } mVideoDecoder->stop(); } return OMXComponentCodecBase::ProcessorStop(); } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorPause(void) { return OMXComponentCodecBase::ProcessorPause(); } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorResume(void) { return OMXComponentCodecBase::ProcessorResume(); } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorFlush(OMX_U32 portIndex) { LOGI("Flushing port# %u.", portIndex); if (mVideoDecoder == NULL) { LOGE("ProcessorFlush: Video decoder is not created."); return OMX_ErrorDynamicResourcesUnavailable; } // Portbase has returned all retained buffers. if (portIndex == INPORT_INDEX || portIndex == OMX_ALL) { //pthread_mutex_lock(&mSerializationLock); LOGW("Flushing video pipeline."); mVideoDecoder->flush(); //pthread_mutex_unlock(&mSerializationLock); } // TODO: do we need to flush output port? return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorPreFreeBuffer(OMX_U32 nPortIndex, OMX_BUFFERHEADERTYPE * pBuffer) { if (mWorkingMode == GRAPHICBUFFER_MODE) return OMX_ErrorNone; if (nPortIndex == OUTPORT_INDEX && pBuffer->pPlatformPrivate) { VideoRenderBuffer *p = (VideoRenderBuffer *)pBuffer->pPlatformPrivate; p->renderDone = true; pBuffer->pPlatformPrivate = NULL; } return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorPreFillBuffer(OMX_BUFFERHEADERTYPE* buffer) { if (mWorkingMode == GRAPHICBUFFER_MODE && buffer->nOutputPortIndex == OUTPORT_INDEX){ Decode_Status status; if(mVideoDecoder == NULL){ LOGW("ProcessorPreFillBuffer: Video decoder is not created"); return OMX_ErrorDynamicResourcesUnavailable; } if (mAPMode == METADATA_MODE) { bool found = false; if (mOMXBufferHeaderTypePtrNum < mMetaDataBuffersNum) { for (uint32_t i = 0; i < mOMXBufferHeaderTypePtrNum; i++) { if (mOMXBufferHeaderTypePtrArray[i] == buffer) { found = true; break; } } if (!found) { mOMXBufferHeaderTypePtrArray[mOMXBufferHeaderTypePtrNum] = buffer; mOMXBufferHeaderTypePtrNum++; } } VideoDecoderOutputMetaData *metadata = (VideoDecoderOutputMetaData *)(buffer->pBuffer); status = mVideoDecoder->signalRenderDone((void *)(metadata->pHandle), !found); } else { status = mVideoDecoder->signalRenderDone(buffer->pBuffer); } if (status != DECODE_SUCCESS) { LOGW("ProcessorPreFillBuffer:: signalRenderDone return error"); return TranslateDecodeStatus(status); } } else if (buffer->pPlatformPrivate && buffer->nOutputPortIndex == OUTPORT_INDEX){ VideoRenderBuffer *p = (VideoRenderBuffer *)buffer->pPlatformPrivate; p->renderDone = true; buffer->pPlatformPrivate = NULL; } return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::ProcessorProcess( OMX_BUFFERHEADERTYPE ***pBuffers, buffer_retain_t *retains, OMX_U32) { OMX_ERRORTYPE ret; Decode_Status status; OMX_BOOL isResolutionChange = OMX_FALSE; // fill render buffer without draining decoder output queue ret = FillRenderBuffer(pBuffers[OUTPORT_INDEX], &retains[OUTPORT_INDEX], 0, &isResolutionChange); if (ret == OMX_ErrorNone) { retains[INPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; if (isResolutionChange) { HandleFormatChange(); } if (mFlushMode) { LOGI("in mFlushMode, do HandleFormatChange."); HandleFormatChange(); } else { // Actually, if mAPMode is set, mWorkingMode should be GRAPHICBUFFER_MODE. if (((mAPMode == METADATA_MODE) && (mWorkingMode == GRAPHICBUFFER_MODE)) && mFormatChanged) { if (((*pBuffers[OUTPORT_INDEX])->nFlags & OMX_BUFFERFLAG_EOS) || (mVideoDecoder->getOutputQueueLength() == 0)) { // Format changed, set mFlushMode, clear eos mFlushMode = true; mFormatChanged = false; (*pBuffers[OUTPORT_INDEX])->nFlags &= ~OMX_BUFFERFLAG_EOS; } } } // TODO: continue decoding return ret; } else if (ret != OMX_ErrorNotReady) { return ret; } VideoDecodeBuffer decodeBuffer; // PrepareDecodeBuffer will set retain to either BUFFER_RETAIN_GETAGAIN or BUFFER_RETAIN_NOT_RETAIN ret = PrepareDecodeBuffer(*pBuffers[INPORT_INDEX], &retains[INPORT_INDEX], &decodeBuffer); if (ret == OMX_ErrorNotReady) { retains[OUTPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; return OMX_ErrorNone; } else if (ret != OMX_ErrorNone) { return ret; } if (decodeBuffer.size != 0) { //pthread_mutex_lock(&mSerializationLock); status = mVideoDecoder->decode(&decodeBuffer); //pthread_mutex_unlock(&mSerializationLock); if (status == DECODE_FORMAT_CHANGE) { if ((mAPMode == METADATA_MODE) && (mWorkingMode == GRAPHICBUFFER_MODE)) { mFormatChanged = true; retains[INPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; } else { ret = HandleFormatChange(); CHECK_RETURN_VALUE("HandleFormatChange"); ((*pBuffers[OUTPORT_INDEX]))->nFilledLen = 0; retains[OUTPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; retains[INPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; // real dynamic resolution change will be handled later // Here is just a temporary workaround // don't use the output buffer if format is changed. return OMX_ErrorNone; } } else if (status == DECODE_NO_CONFIG) { LOGW("Decoder returns DECODE_NO_CONFIG."); retains[OUTPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; return OMX_ErrorNone; } else if (status == DECODE_NO_REFERENCE) { LOGW("Decoder returns DECODE_NO_REFERENCE."); //retains[OUTPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; //return OMX_ErrorNone; } else if (status == DECODE_MULTIPLE_FRAME){ if (decodeBuffer.ext != NULL && decodeBuffer.ext->extType == PACKED_FRAME_TYPE && decodeBuffer.ext->extData != NULL) { PackedFrameData* nextFrame = (PackedFrameData*)decodeBuffer.ext->extData; (*pBuffers[INPORT_INDEX])->nOffset += nextFrame->offSet; (*pBuffers[INPORT_INDEX])->nTimeStamp = nextFrame->timestamp; (*pBuffers[INPORT_INDEX])->nFilledLen -= nextFrame->offSet; retains[INPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; LOGW("Find multiple frames in a buffer, next frame offset = %d, timestamp = %lld", (*pBuffers[INPORT_INDEX])->nOffset, (*pBuffers[INPORT_INDEX])->nTimeStamp); } } else if (status != DECODE_SUCCESS && status != DECODE_FRAME_DROPPED) { if (checkFatalDecoderError(status)) { return TranslateDecodeStatus(status); } else { // For decoder errors that could be omitted, not throw error and continue to decode. TranslateDecodeStatus(status); ((*pBuffers[OUTPORT_INDEX]))->nFilledLen = 0; // Do not return, and try to drain the output queue // retains[OUTPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; // return OMX_ErrorNone; } } } // drain the decoder output queue when in EOS state and fill the render buffer ret = FillRenderBuffer(pBuffers[OUTPORT_INDEX], &retains[OUTPORT_INDEX], ((*pBuffers[INPORT_INDEX]))->nFlags,&isResolutionChange); if (isResolutionChange) { HandleFormatChange(); } if (mFlushMode) { LOGI("in mFlushMode, do HandleFormatChange."); HandleFormatChange(); } else { if (((mAPMode == METADATA_MODE) && (mWorkingMode == GRAPHICBUFFER_MODE)) && mFormatChanged) { if (((*pBuffers[OUTPORT_INDEX])->nFlags & OMX_BUFFERFLAG_EOS) || (mVideoDecoder->getOutputQueueLength() == 0)) { // Format changed, set mFlushMode, clear eos. mFlushMode = true; mFormatChanged = false; (*pBuffers[OUTPORT_INDEX])->nFlags &= ~OMX_BUFFERFLAG_EOS; } } } bool inputEoS = ((*pBuffers[INPORT_INDEX])->nFlags & OMX_BUFFERFLAG_EOS); bool outputEoS = ((*pBuffers[OUTPORT_INDEX])->nFlags & OMX_BUFFERFLAG_EOS); // if output port is not eos, retain the input buffer until all the output buffers are drained. if (inputEoS && !outputEoS) { retains[INPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; // the input buffer is retained for draining purpose. Set nFilledLen to 0 so buffer will not be decoded again. (*pBuffers[INPORT_INDEX])->nFilledLen = 0; } if (ret == OMX_ErrorNotReady) { retains[OUTPORT_INDEX] = BUFFER_RETAIN_GETAGAIN; ret = OMX_ErrorNone; } return ret; } bool OMXVideoDecoderBase::IsAllBufferAvailable(void) { bool b = ComponentBase::IsAllBufferAvailable(); if (b == false) { return false; } PortVideo *port = NULL; port = static_cast<PortVideo *>(this->ports[OUTPORT_INDEX]); const OMX_PARAM_PORTDEFINITIONTYPE* port_def = port->GetPortDefinition(); // if output port is disabled, retain the input buffer if (!port_def->bEnabled) { return false; } if (mVideoDecoder) { return mVideoDecoder->checkBufferAvail(); } return false; } OMX_ERRORTYPE OMXVideoDecoderBase::PrepareConfigBuffer(VideoConfigBuffer *p) { // default config buffer preparation memset(p, 0, sizeof(VideoConfigBuffer)); const OMX_PARAM_PORTDEFINITIONTYPE *paramPortDefinitionInput = this->ports[INPORT_INDEX]->GetPortDefinition(); if (paramPortDefinitionInput == NULL) { return OMX_ErrorBadParameter; } if (mWorkingMode == GRAPHICBUFFER_MODE) { if (mAPMode == METADATA_MODE) { const OMX_PARAM_PORTDEFINITIONTYPE *def_output = this->ports[OUTPORT_INDEX]->GetPortDefinition(); if (def_output == NULL) { return OMX_ErrorBadParameter; } mMetaDataBuffersNum = def_output->nBufferCountActual; mOMXBufferHeaderTypePtrNum = 0; mGraphicBufferParam.graphicBufferColorFormat = def_output->format.video.eColorFormat; mGraphicBufferParam.graphicBufferHStride = getStride(def_output->format.video.nFrameWidth); mGraphicBufferParam.graphicBufferVStride = (def_output->format.video.nFrameHeight + 0x1f) & ~0x1f; mGraphicBufferParam.graphicBufferWidth = def_output->format.video.nFrameWidth; mGraphicBufferParam.graphicBufferHeight = def_output->format.video.nFrameHeight; p->surfaceNumber = mMetaDataBuffersNum; for (int i = 0; i < MAX_GRAPHIC_BUFFER_NUM; i++) { p->graphicBufferHandler[i] = NULL; } p->flag |= WANT_STORE_META_DATA; } else { p->surfaceNumber = mOMXBufferHeaderTypePtrNum; for (uint32_t i = 0; i < mOMXBufferHeaderTypePtrNum; i++){ OMX_BUFFERHEADERTYPE *buffer_hdr = mOMXBufferHeaderTypePtrArray[i]; p->graphicBufferHandler[i] = buffer_hdr->pBuffer; LOGV("PrepareConfigBuffer bufferid = %p, handle = %p", buffer_hdr, buffer_hdr->pBuffer); } } p->flag |= USE_NATIVE_GRAPHIC_BUFFER; p->graphicBufferHStride = mGraphicBufferParam.graphicBufferHStride; p->graphicBufferVStride = mGraphicBufferParam.graphicBufferVStride; p->graphicBufferWidth = mGraphicBufferParam.graphicBufferWidth; p->graphicBufferHeight = mGraphicBufferParam.graphicBufferHeight; p->graphicBufferColorFormat = mGraphicBufferParam.graphicBufferColorFormat; if (p->graphicBufferColorFormat == OMX_INTEL_COLOR_FormatYUV420PackedSemiPlanar_Tiled #ifdef USE_GEN_HW || p->graphicBufferColorFormat == HAL_PIXEL_FORMAT_NV12_X_TILED_INTEL #endif ) p->flag |= USE_TILING_MEMORY; if (mEnableAdaptivePlayback) p->flag |= WANT_ADAPTIVE_PLAYBACK; PortVideo *port = NULL; port = static_cast<PortVideo *>(this->ports[INPORT_INDEX]); OMX_PARAM_PORTDEFINITIONTYPE port_def; memcpy(&port_def, port->GetPortDefinition(), sizeof(port_def)); if (port_def.format.video.pNativeWindow != NULL) { p->nativeWindow = port_def.format.video.pNativeWindow; LOGD("NativeWindow = %p", p->nativeWindow); } } p->rotationDegrees = mRotationDegrees; #ifdef TARGET_HAS_ISV p->vppBufferNum = mVppBufferNum; #endif p->width = paramPortDefinitionInput->format.video.nFrameWidth; p->height = paramPortDefinitionInput->format.video.nFrameHeight; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::PrepareDecodeBuffer(OMX_BUFFERHEADERTYPE *buffer, buffer_retain_t *retain, VideoDecodeBuffer *p) { // default decode buffer preparation memset(p, 0, sizeof(VideoDecodeBuffer)); if (buffer->nFilledLen == 0) { LOGW("Len of filled data to decode is 0."); return OMX_ErrorNone; //OMX_ErrorBadParameter; } if (buffer->pBuffer == NULL) { LOGE("Buffer to decode is empty."); return OMX_ErrorBadParameter; } if (buffer->nFlags & OMX_BUFFERFLAG_CODECCONFIG) { LOGI("Buffer has OMX_BUFFERFLAG_CODECCONFIG flag."); } if (buffer->nFlags & OMX_BUFFERFLAG_DECODEONLY) { // TODO: Handle OMX_BUFFERFLAG_DECODEONLY : drop the decoded frame without rendering it. LOGW("Buffer has OMX_BUFFERFLAG_DECODEONLY flag."); } p->data = buffer->pBuffer + buffer->nOffset; p->size = buffer->nFilledLen; p->timeStamp = buffer->nTimeStamp; if (buffer->nFlags & (OMX_BUFFERFLAG_ENDOFFRAME | OMX_BUFFERFLAG_EOS)) { // TODO: OMX_BUFFERFLAG_ENDOFFRAME can be used to indicate end of a NAL unit. // setting this flag may cause corruption if buffer does not contain end-of-frame data. p->flag = HAS_COMPLETE_FRAME; } if (buffer->nFlags & OMX_BUFFERFLAG_SYNCFRAME) { p->flag |= IS_SYNC_FRAME; } if (buffer->pInputPortPrivate) { uint32_t degree = 0; memcpy ((void *) °ree, buffer->pInputPortPrivate, sizeof(uint32_t)); p->rotationDegrees = degree; LOGV("rotationDegrees = %d", p->rotationDegrees); } else { p->rotationDegrees = mRotationDegrees; } *retain= BUFFER_RETAIN_NOT_RETAIN; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::PrepareDecodeNativeHandleBuffer(OMX_BUFFERHEADERTYPE *buffer, buffer_retain_t *retain, VideoDecodeBuffer *p) { // default decode buffer preparation memset(p, 0, sizeof(VideoDecodeBuffer)); if (buffer->nFilledLen == 0) { LOGW("Len of filled data to decode is 0."); return OMX_ErrorNone; //OMX_ErrorBadParameter; } if (buffer->pBuffer == NULL) { LOGE("Buffer to decode is empty."); return OMX_ErrorBadParameter; } if (buffer->nFlags & OMX_BUFFERFLAG_CODECCONFIG) { LOGI("Buffer has OMX_BUFFERFLAG_CODECCONFIG flag."); } if (buffer->nFlags & OMX_BUFFERFLAG_DECODEONLY) { // TODO: Handle OMX_BUFFERFLAG_DECODEONLY : drop the decoded frame without rendering it. LOGW("Buffer has OMX_BUFFERFLAG_DECODEONLY flag."); } //Get data pointer from native_handle native_handle_t *native_handle = (native_handle_t *)buffer->pBuffer; ProtectedDataBuffer *dataBuffer = (ProtectedDataBuffer *) native_handle->data[1]; p->data = dataBuffer->data + buffer->nOffset; p->size = buffer->nFilledLen; p->timeStamp = buffer->nTimeStamp; if (buffer->nFlags & (OMX_BUFFERFLAG_ENDOFFRAME | OMX_BUFFERFLAG_EOS)) { // TODO: OMX_BUFFERFLAG_ENDOFFRAME can be used to indicate end of a NAL unit. // setting this flag may cause corruption if buffer does not contain end-of-frame data. p->flag = HAS_COMPLETE_FRAME; } if (buffer->nFlags & OMX_BUFFERFLAG_SYNCFRAME) { p->flag |= IS_SYNC_FRAME; } if (buffer->pInputPortPrivate) { uint32_t degree = 0; memcpy ((void *) °ree, buffer->pInputPortPrivate, sizeof(uint32_t)); p->rotationDegrees = degree; LOGV("rotationDegrees = %d", p->rotationDegrees); } else { p->rotationDegrees = mRotationDegrees; } *retain= BUFFER_RETAIN_NOT_RETAIN; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::FillRenderBuffer(OMX_BUFFERHEADERTYPE **pBuffer, buffer_retain_t *retain, OMX_U32 inportBufferFlags, OMX_BOOL *isResolutionChange) { OMX_BUFFERHEADERTYPE *buffer = *pBuffer; OMX_BUFFERHEADERTYPE *buffer_orign = buffer; VideoErrorBuffer *ErrBufPtr = NULL; if (mWorkingMode != GRAPHICBUFFER_MODE && buffer->pPlatformPrivate) { VideoRenderBuffer *p = (VideoRenderBuffer *)buffer->pPlatformPrivate; p->renderDone = true; buffer->pPlatformPrivate = NULL; } if (mWorkingMode == GRAPHICBUFFER_MODE && mErrorReportEnabled) { if (buffer->pOutputPortPrivate == NULL) LOGE("The App doesn't provide the output buffer for error reporting"); else ErrBufPtr = (VideoErrorBuffer *)buffer->pOutputPortPrivate; } bool draining = (inportBufferFlags & OMX_BUFFERFLAG_EOS); //pthread_mutex_lock(&mSerializationLock); const VideoRenderBuffer *renderBuffer = NULL; //pthread_mutex_unlock(&mSerializationLock); // in mFlushMode, provide empty buffer. if (mFlushMode) { buffer->nFilledLen = 0; return OMX_ErrorNone; } if (((mAPMode == METADATA_MODE) && (mWorkingMode == GRAPHICBUFFER_MODE)) && mFormatChanged) { renderBuffer = mVideoDecoder->getOutput(true, ErrBufPtr); } else { renderBuffer = mVideoDecoder->getOutput(draining, ErrBufPtr); } if (renderBuffer == NULL) { buffer->nFilledLen = 0; if (draining) { LOGI("output EOS received"); buffer->nFlags = OMX_BUFFERFLAG_EOS; return OMX_ErrorNone; } return OMX_ErrorNotReady; } if (mWorkingMode == GRAPHICBUFFER_MODE) { buffer = *pBuffer = mOMXBufferHeaderTypePtrArray[renderBuffer->graphicBufferIndex]; } buffer->nFlags = OMX_BUFFERFLAG_ENDOFFRAME; #ifdef DEINTERLACE_EXT if (renderBuffer->scanFormat & (VA_TOP_FIELD | VA_BOTTOM_FIELD)) buffer->nFlags |= OMX_BUFFERFLAG_TFF; #endif buffer->nTimeStamp = renderBuffer->timeStamp; if (renderBuffer->flag & IS_EOS) { buffer->nFlags |= OMX_BUFFERFLAG_EOS; } *isResolutionChange = (renderBuffer->flag & IS_RESOLUTION_CHANGE)? OMX_TRUE: OMX_FALSE; if (mWorkingMode == GRAPHICBUFFER_MODE) { if (buffer_orign != buffer) { VideoErrorBuffer *ErrBufOutPtr = NULL; ErrBufOutPtr = (VideoErrorBuffer *)buffer->pOutputPortPrivate; if (ErrBufPtr && ErrBufOutPtr) { memcpy(ErrBufOutPtr, ErrBufPtr, sizeof(VideoErrorBuffer)); memset(ErrBufPtr, 0, sizeof(VideoErrorBuffer)); } *retain = BUFFER_RETAIN_OVERRIDDEN; } buffer->nFilledLen = sizeof(OMX_U8*); } else { uint32_t size = 0; Decode_Status status = mVideoDecoder->getRawDataFromSurface(const_cast<VideoRenderBuffer *>(renderBuffer), buffer->pBuffer + buffer->nOffset, &size, false); if (status != DECODE_SUCCESS) { return TranslateDecodeStatus(status); } buffer->nFilledLen = size; buffer->pPlatformPrivate = (void *)renderBuffer; } return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::HandleFormatChange(void) { LOGW("Video format is changed."); //pthread_mutex_lock(&mSerializationLock); const VideoFormatInfo *formatInfo = mVideoDecoder->getFormatInfo(); //pthread_mutex_unlock(&mSerializationLock); // Sync port definition as it may change. OMX_PARAM_PORTDEFINITIONTYPE paramPortDefinitionInput, paramPortDefinitionOutput; memcpy(¶mPortDefinitionInput, this->ports[INPORT_INDEX]->GetPortDefinition(), sizeof(paramPortDefinitionInput)); memcpy(¶mPortDefinitionOutput, this->ports[OUTPORT_INDEX]->GetPortDefinition(), sizeof(paramPortDefinitionOutput)); uint32_t width = formatInfo->width; uint32_t height = formatInfo->height; uint32_t stride = formatInfo->width; uint32_t sliceHeight = formatInfo->height; uint32_t widthCropped = formatInfo->width - formatInfo->cropLeft - formatInfo->cropRight; uint32_t heightCropped = formatInfo->height - formatInfo->cropTop - formatInfo->cropBottom; if (strcasecmp(formatInfo->mimeType,"video/avc") == 0 || strcasecmp(formatInfo->mimeType,"video/avc-secure") == 0 || strcasecmp(formatInfo->mimeType,"video/h264") == 0) { heightCropped = formatInfo->height; widthCropped = formatInfo->width; } uint32_t strideCropped = widthCropped; uint32_t sliceHeightCropped = heightCropped; int force_realloc = 0; bool isVP8 = false; if (mAPMode == METADATA_MODE && mWorkingMode == GRAPHICBUFFER_MODE) { #ifdef TARGET_HAS_ISV if (paramPortDefinitionOutput.nBufferCountActual - mVppBufferNum < formatInfo->actualBufferNeeded) { #else if (paramPortDefinitionOutput.nBufferCountActual < formatInfo->actualBufferNeeded) { #endif paramPortDefinitionOutput.nBufferCountActual = mNativeBufferCount = formatInfo->actualBufferNeeded; paramPortDefinitionOutput.nBufferCountMin = formatInfo->actualBufferNeeded - 4; } // input port paramPortDefinitionInput.format.video.nFrameWidth = width; paramPortDefinitionInput.format.video.nFrameHeight = height; paramPortDefinitionInput.format.video.nStride = stride; paramPortDefinitionInput.format.video.nSliceHeight = sliceHeight; // output port paramPortDefinitionOutput.format.video.nFrameWidth = width; paramPortDefinitionOutput.format.video.nFrameHeight = height; paramPortDefinitionOutput.format.video.eColorFormat = GetOutputColorFormat(paramPortDefinitionOutput.format.video.nFrameWidth); paramPortDefinitionOutput.format.video.nStride = stride; paramPortDefinitionOutput.format.video.nSliceHeight = sliceHeight; paramPortDefinitionOutput.bEnabled = (OMX_BOOL)false; mOMXBufferHeaderTypePtrNum = 0; memset(&mGraphicBufferParam, 0, sizeof(mGraphicBufferParam)); mMetaDataBuffersNum = 0; this->ports[INPORT_INDEX]->SetPortDefinition(¶mPortDefinitionInput, true); this->ports[OUTPORT_INDEX]->SetPortDefinition(¶mPortDefinitionOutput, true); ProcessorFlush(INPORT_INDEX); mVideoDecoder->freeSurfaceBuffers(); this->ports[OUTPORT_INDEX]->ReportPortSettingsChanged(); mFlushMode = false; return OMX_ErrorNone; } #ifdef TARGET_HAS_ISV LOGI("============== mVppBufferNum = %d\n", mVppBufferNum); if (paramPortDefinitionOutput.nBufferCountActual - mVppBufferNum < formatInfo->actualBufferNeeded) { #else if (paramPortDefinitionOutput.nBufferCountActual < formatInfo->actualBufferNeeded) { #endif if (mWorkingMode == GRAPHICBUFFER_MODE) { LOGV("output port buffer number is not enough: %d to %d", paramPortDefinitionOutput.nBufferCountActual, formatInfo->actualBufferNeeded); paramPortDefinitionOutput.nBufferCountActual = mNativeBufferCount = formatInfo->actualBufferNeeded; paramPortDefinitionOutput.nBufferCountMin = mNativeBufferCount; force_realloc = 1; } } LOGV("Original size = %u x %u, new size = %d x %d, cropped size = %d x %d", paramPortDefinitionInput.format.video.nFrameWidth, paramPortDefinitionInput.format.video.nFrameHeight, width, height, widthCropped, heightCropped); if (paramPortDefinitionInput.format.video.eCompressionFormat == OMX_VIDEO_CodingVP8) { isVP8 = true; } if (!force_realloc && widthCropped == paramPortDefinitionOutput.format.video.nFrameWidth && heightCropped == paramPortDefinitionOutput.format.video.nFrameHeight) { if (mWorkingMode == RAWDATA_MODE) { LOGW("Change of portsetting is not reported as size is not changed."); return OMX_ErrorNone; } } paramPortDefinitionInput.format.video.nFrameWidth = width; paramPortDefinitionInput.format.video.nFrameHeight = height; paramPortDefinitionInput.format.video.nStride = stride; paramPortDefinitionInput.format.video.nSliceHeight = sliceHeight; if (mWorkingMode == RAWDATA_MODE) { paramPortDefinitionOutput.format.video.nFrameWidth = widthCropped; paramPortDefinitionOutput.format.video.nFrameHeight = heightCropped; paramPortDefinitionOutput.format.video.nStride = strideCropped; paramPortDefinitionOutput.format.video.nSliceHeight = sliceHeightCropped; } else if (mWorkingMode == GRAPHICBUFFER_MODE) { // when the width and height ES parse are not larger than allocated graphic buffer in outport, // there is no need to reallocate graphic buffer,just report the crop info to omx client if (!force_realloc && width <= formatInfo->surfaceWidth && height <= formatInfo->surfaceHeight) { this->ports[INPORT_INDEX]->SetPortDefinition(¶mPortDefinitionInput, true); this->ports[OUTPORT_INDEX]->ReportOutputCrop(); return OMX_ErrorNone; } if (isVP8 || width > formatInfo->surfaceWidth || height > formatInfo->surfaceHeight) { // update the real decoded resolution to outport instead of display resolution for graphic buffer reallocation // when the width and height parsed from ES are larger than allocated graphic buffer in outport, paramPortDefinitionOutput.format.video.nFrameWidth = width; paramPortDefinitionOutput.format.video.nFrameHeight = height; paramPortDefinitionOutput.format.video.eColorFormat = GetOutputColorFormat( paramPortDefinitionOutput.format.video.nFrameWidth); paramPortDefinitionOutput.format.video.nStride = stride; paramPortDefinitionOutput.format.video.nSliceHeight = sliceHeight; } } paramPortDefinitionOutput.bEnabled = (OMX_BOOL)false; mOMXBufferHeaderTypePtrNum = 0; memset(&mGraphicBufferParam, 0, sizeof(mGraphicBufferParam)); this->ports[INPORT_INDEX]->SetPortDefinition(¶mPortDefinitionInput, true); this->ports[OUTPORT_INDEX]->SetPortDefinition(¶mPortDefinitionOutput, true); if (mWorkingMode == GRAPHICBUFFER_MODE) { // Make sure va_destroySurface is called before graphicbuffer is freed in case of port setting changed mVideoDecoder->freeSurfaceBuffers(); // Also make sure all the reference frames are flushed ProcessorFlush(INPORT_INDEX); } this->ports[OUTPORT_INDEX]->ReportPortSettingsChanged(); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::TranslateDecodeStatus(Decode_Status status) { switch (status) { case DECODE_NEED_RESTART: LOGE("Decoder returned DECODE_NEED_RESTART"); return (OMX_ERRORTYPE)OMX_ErrorIntelVideoNotPermitted; case DECODE_NO_CONFIG: LOGE("Decoder returned DECODE_NO_CONFIG"); return (OMX_ERRORTYPE)OMX_ErrorIntelMissingConfig; case DECODE_NO_SURFACE: LOGE("Decoder returned DECODE_NO_SURFACE"); return OMX_ErrorDynamicResourcesUnavailable; case DECODE_NO_REFERENCE: LOGE("Decoder returned DECODE_NO_REFERENCE"); return OMX_ErrorDynamicResourcesUnavailable; // TO DO case DECODE_NO_PARSER: LOGE("Decoder returned DECODE_NO_PARSER"); return OMX_ErrorDynamicResourcesUnavailable; case DECODE_INVALID_DATA: LOGE("Decoder returned DECODE_INVALID_DATA"); return OMX_ErrorBadParameter; case DECODE_DRIVER_FAIL: LOGE("Decoder returned DECODE_DRIVER_FAIL"); return OMX_ErrorHardware; case DECODE_PARSER_FAIL: LOGE("Decoder returned DECODE_PARSER_FAIL"); return (OMX_ERRORTYPE)OMX_ErrorIntelProcessStream; // OMX_ErrorStreamCorrupt case DECODE_MEMORY_FAIL: LOGE("Decoder returned DECODE_MEMORY_FAIL"); return OMX_ErrorInsufficientResources; case DECODE_FAIL: LOGE("Decoder returned DECODE_FAIL"); return OMX_ErrorUndefined; case DECODE_SUCCESS: return OMX_ErrorNone; case DECODE_FORMAT_CHANGE: LOGW("Decoder returned DECODE_FORMAT_CHANGE"); return OMX_ErrorNone; case DECODE_FRAME_DROPPED: LOGI("Decoder returned DECODE_FRAME_DROPPED"); return OMX_ErrorNone; default: LOGW("Decoder returned unknown error"); return OMX_ErrorUndefined; } } OMX_ERRORTYPE OMXVideoDecoderBase::BuildHandlerList(void) { OMXComponentCodecBase::BuildHandlerList(); AddHandler(OMX_IndexParamVideoPortFormat, GetParamVideoPortFormat, SetParamVideoPortFormat); AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexExtGetNativeBufferUsage), GetNativeBufferUsage, SetNativeBufferUsage); AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexExtUseNativeBuffer), GetNativeBuffer, SetNativeBuffer); AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexExtEnableNativeBuffer), GetNativeBufferMode, SetNativeBufferMode); AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexExtRotationDegrees), GetDecoderRotation, SetDecoderRotation); #ifdef TARGET_HAS_ISV AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexExtVppBufferNum), GetDecoderVppBufferNum, SetDecoderVppBufferNum); #endif AddHandler(OMX_IndexConfigCommonOutputCrop, GetDecoderOutputCrop, SetDecoderOutputCrop); #ifdef USE_META_DATA AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexStoreMetaDataInBuffers), GetStoreMetaDataMode, SetStoreMetaDataMode); #endif AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexExtEnableErrorReport), GetErrorReportMode, SetErrorReportMode); AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexConfigPriority), GetCodecPriority, SetCodecPriority); AddHandler(static_cast<OMX_INDEXTYPE>(OMX_IndexConfigOperatingRate), GetDecoderOperatingRate, SetDecoderOperatingRate); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetParamVideoPortFormat(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_VIDEO_PARAM_PORTFORMATTYPE *p = (OMX_VIDEO_PARAM_PORTFORMATTYPE *)pStructure; CHECK_TYPE_HEADER(p); CHECK_PORT_INDEX_RANGE(p); CHECK_ENUMERATION_RANGE(p->nIndex, 1); PortVideo *port = NULL; port = static_cast<PortVideo *>(this->ports[p->nPortIndex]); memcpy(p, port->GetPortVideoParam(), sizeof(*p)); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::SetParamVideoPortFormat(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_VIDEO_PARAM_PORTFORMATTYPE *p = (OMX_VIDEO_PARAM_PORTFORMATTYPE *)pStructure; CHECK_TYPE_HEADER(p); CHECK_PORT_INDEX_RANGE(p); CHECK_SET_PARAM_STATE(); // TODO: do we need to check if port is enabled? PortVideo *port = NULL; port = static_cast<PortVideo *>(this->ports[p->nPortIndex]); port->SetPortVideoParam(p, false); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetNativeBufferUsageSpecific(OMX_PTR pStructure) { OMX_ERRORTYPE ret; GetAndroidNativeBufferUsageParams *param = (GetAndroidNativeBufferUsageParams*)pStructure; CHECK_TYPE_HEADER(param); // hardware usage: consumed by GLES and HWC param->nUsage |= GRALLOC_USAGE_HW_TEXTURE | GRALLOC_USAGE_HW_COMPOSER; // software usage: can be read/written by apps param->nUsage |= GRALLOC_USAGE_SW_READ_RARELY | GRALLOC_USAGE_SW_WRITE_RARELY; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::SetNativeBufferUsageSpecific(OMX_PTR) { CHECK_SET_PARAM_STATE(); return OMX_ErrorBadParameter; } OMX_ERRORTYPE OMXVideoDecoderBase::GetNativeBufferUsage(OMX_PTR pStructure) { return this->GetNativeBufferUsageSpecific(pStructure); } OMX_ERRORTYPE OMXVideoDecoderBase::SetNativeBufferUsage(OMX_PTR pStructure) { return this->SetNativeBufferUsageSpecific(pStructure); } OMX_ERRORTYPE OMXVideoDecoderBase::GetNativeBuffer(OMX_PTR) { return OMX_ErrorBadParameter; } OMX_ERRORTYPE OMXVideoDecoderBase::SetNativeBuffer(OMX_PTR pStructure) { OMX_ERRORTYPE ret; UseAndroidNativeBufferParams *param = (UseAndroidNativeBufferParams*)pStructure; CHECK_TYPE_HEADER(param); if (param->nPortIndex != OUTPORT_INDEX) return OMX_ErrorBadParameter; OMX_BUFFERHEADERTYPE **buf_hdr = NULL; mOMXBufferHeaderTypePtrNum++; if (mOMXBufferHeaderTypePtrNum > MAX_GRAPHIC_BUFFER_NUM) return OMX_ErrorOverflow; buf_hdr = &mOMXBufferHeaderTypePtrArray[mOMXBufferHeaderTypePtrNum-1]; ret = this->ports[OUTPORT_INDEX]->UseBuffer(buf_hdr, OUTPORT_INDEX, param->pAppPrivate, sizeof(OMX_U8*), const_cast<OMX_U8*>(reinterpret_cast<const OMX_U8*>(param->nativeBuffer->handle))); if (ret != OMX_ErrorNone) return ret; if (mOMXBufferHeaderTypePtrNum == 1) { mGraphicBufferParam.graphicBufferColorFormat = param->nativeBuffer->format; mGraphicBufferParam.graphicBufferHStride = param->nativeBuffer->stride; // FIXME: use IMG_native_handle_t->aiVStride[0] instead.. mGraphicBufferParam.graphicBufferVStride = param->nativeBuffer->height; mGraphicBufferParam.graphicBufferWidth = param->nativeBuffer->width; mGraphicBufferParam.graphicBufferHeight = param->nativeBuffer->height; } *(param->bufferHeader) = *buf_hdr; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetNativeBufferMode(OMX_PTR pStructure) { return this->GetNativeBufferModeSpecific(pStructure); } OMX_ERRORTYPE OMXVideoDecoderBase::SetNativeBufferMode(OMX_PTR pStructure) { return this->SetNativeBufferModeSpecific(pStructure); } OMX_ERRORTYPE OMXVideoDecoderBase::GetNativeBufferModeSpecific(OMX_PTR) { LOGE("GetNativeBufferMode is not implemented"); return OMX_ErrorNotImplemented; } OMX_ERRORTYPE OMXVideoDecoderBase::SetNativeBufferModeSpecific(OMX_PTR pStructure) { OMX_ERRORTYPE ret; EnableAndroidNativeBuffersParams *param = (EnableAndroidNativeBuffersParams*)pStructure; CHECK_TYPE_HEADER(param); CHECK_PORT_INDEX_RANGE(param); CHECK_SET_PARAM_STATE(); PortVideo *port = NULL; port = static_cast<PortVideo *>(this->ports[OUTPORT_INDEX]); OMX_PARAM_PORTDEFINITIONTYPE port_def; memcpy(&port_def,port->GetPortDefinition(),sizeof(port_def)); if (!param->enable) { mWorkingMode = RAWDATA_MODE; // If it is fallback from native mode the color format has been // already set to INTEL format. // We need to set back the default color format and Native stuff. port_def.format.video.eColorFormat = OMX_COLOR_FormatYUV420SemiPlanar; port_def.format.video.pNativeRender = NULL; port_def.format.video.pNativeWindow = NULL; port->SetPortDefinition(&port_def,true); return OMX_ErrorNone; } mWorkingMode = GRAPHICBUFFER_MODE; port_def.nBufferCountMin = mNativeBufferCount; if (mEnableAdaptivePlayback) { SetMaxOutputBufferCount(&port_def); } else { port_def.nBufferCountActual = mNativeBufferCount; } port_def.format.video.cMIMEType = (OMX_STRING)VA_VED_RAW_MIME_TYPE; port_def.format.video.eColorFormat = OMX_INTEL_COLOR_FormatYUV420PackedSemiPlanar; port_def.format.video.nFrameHeight = port_def.format.video.nFrameHeight; port_def.format.video.eColorFormat = GetOutputColorFormat( port_def.format.video.nFrameWidth); port->SetPortDefinition(&port_def,true); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetStoreMetaDataMode(OMX_PTR) { ALOGE("GetMetaDataMode is not implemented"); return OMX_ErrorNotImplemented; } OMX_ERRORTYPE OMXVideoDecoderBase::SetStoreMetaDataMode(OMX_PTR pStructure) { #ifndef USE_META_DATA OMX_PARAM_PORTDEFINITIONTYPE defInput; memcpy(&defInput, this->ports[INPORT_INDEX]->GetPortDefinition(), sizeof(defInput)); if (defInput.format.video.eCompressionFormat == OMX_VIDEO_CodingVP9) { ALOGE("SetMetaDataMode for VP9 is not implemented"); return OMX_ErrorNotImplemented; } #endif OMX_ERRORTYPE ret; StoreMetaDataInBuffersParams *param = (StoreMetaDataInBuffersParams*)pStructure; CHECK_TYPE_HEADER(param); CHECK_PORT_INDEX(param, OUTPORT_INDEX); CHECK_SET_PARAM_STATE(); if (!param->bStoreMetaData) { mAPMode = LEGACY_MODE; // Don't return error which otherwise may cause framework crash return OMX_ErrorNone; } mAPMode = METADATA_MODE; ALOGI("We are in meta data mode!!!"); return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetDecoderRotation(OMX_PTR) { return OMX_ErrorBadParameter; } OMX_ERRORTYPE OMXVideoDecoderBase::SetDecoderRotation(OMX_PTR pStructure) { CHECK_SET_PARAM_STATE(); int32_t rotationDegrees = 0; if (pStructure) { rotationDegrees = *(static_cast<int32_t*>(pStructure)); mRotationDegrees = rotationDegrees; LOGI("Rotation Degree = %d", rotationDegrees); return OMX_ErrorNone; } else { return OMX_ErrorBadParameter; } } #ifdef TARGET_HAS_ISV OMX_ERRORTYPE OMXVideoDecoderBase::GetDecoderVppBufferNum(OMX_PTR) { return OMX_ErrorBadParameter; } OMX_ERRORTYPE OMXVideoDecoderBase::SetDecoderVppBufferNum(OMX_PTR pStructure) { CHECK_SET_PARAM_STATE(); int32_t num = 0; num = *(static_cast<int32_t*>(pStructure)); mVppBufferNum = num; return OMX_ErrorNone; } #endif OMX_ERRORTYPE OMXVideoDecoderBase::GetDecoderOutputCropSpecific(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_CONFIG_RECTTYPE *rectParams = (OMX_CONFIG_RECTTYPE *)pStructure; CHECK_TYPE_HEADER(rectParams); if (rectParams->nPortIndex != OUTPORT_INDEX) { return OMX_ErrorUndefined; } PortVideo *port = NULL; port = static_cast<PortVideo *>(this->ports[OUTPORT_INDEX]); OMX_PARAM_PORTDEFINITIONTYPE port_def; memcpy(&port_def,port->GetPortDefinition(),sizeof(port_def)); const VideoFormatInfo *formatInfo = mVideoDecoder->getFormatInfo(); if (formatInfo->valid == true) { rectParams->nLeft = formatInfo->cropLeft; rectParams->nTop = formatInfo->cropTop; rectParams->nWidth = formatInfo->width - formatInfo->cropLeft - formatInfo->cropRight; rectParams->nHeight = formatInfo->height - formatInfo->cropTop - formatInfo->cropBottom; if (strcasecmp(formatInfo->mimeType,"video/avc") == 0 || strcasecmp(formatInfo->mimeType,"video/avc-secure") == 0 || strcasecmp(formatInfo->mimeType,"video/h264") == 0) { rectParams->nHeight = formatInfo->height; rectParams->nWidth = formatInfo->width; } // if port width parsed from extractor is not as same as from SPS/PPS nalu header, // align it. if (port_def.format.video.nFrameWidth != rectParams->nWidth) { port_def.format.video.nFrameWidth = rectParams->nWidth; } port->SetPortDefinition(&port_def,true); return OMX_ErrorNone; } else { return OMX_ErrorFormatNotDetected; } } OMX_ERRORTYPE OMXVideoDecoderBase::SetDecoderOutputCropSpecific(OMX_PTR) { return OMX_ErrorUnsupportedSetting; } OMX_ERRORTYPE OMXVideoDecoderBase::SetDecoderOutputCrop(OMX_PTR pStructure) { return this->SetDecoderOutputCropSpecific(pStructure); } OMX_ERRORTYPE OMXVideoDecoderBase::GetDecoderOutputCrop(OMX_PTR pStructure) { return this->GetDecoderOutputCropSpecific(pStructure); } OMX_ERRORTYPE OMXVideoDecoderBase::SetCodecPriority(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_PARAM_U32TYPE *priorityParam = (OMX_PARAM_U32TYPE *)pStructure; mCodecPriority = priorityParam->nU32; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetCodecPriority(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_PARAM_U32TYPE *priorityParam = (OMX_PARAM_U32TYPE *)pStructure; CHECK_TYPE_HEADER(priorityParam); priorityParam->nU32 = mCodecPriority; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::SetDecoderOperatingRate(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_PARAM_U32TYPE *operatingRateParam = (OMX_PARAM_U32TYPE *)pStructure; CHECK_TYPE_HEADER(operatingRateParam); mOperatingRate = operatingRateParam->nU32; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetDecoderOperatingRate(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_PARAM_U32TYPE *operatingRateParam = (OMX_PARAM_U32TYPE *)pStructure; CHECK_TYPE_HEADER(operatingRateParam); operatingRateParam->nU32 = mOperatingRate; return OMX_ErrorNone; } OMX_ERRORTYPE OMXVideoDecoderBase::GetErrorReportMode(OMX_PTR) { LOGE("GetErrorReportMode is not implemented"); return OMX_ErrorNotImplemented; } OMX_ERRORTYPE OMXVideoDecoderBase::SetErrorReportMode(OMX_PTR pStructure) { OMX_ERRORTYPE ret; OMX_VIDEO_CONFIG_INTEL_ERROR_REPORT *p = (OMX_VIDEO_CONFIG_INTEL_ERROR_REPORT *)pStructure; CHECK_TYPE_HEADER(p); CHECK_PORT_INDEX(p, OUTPORT_INDEX); mErrorReportEnabled = p->bEnable; LOGD("Error reporting is %s", mErrorReportEnabled ? "enabled" : "disabled"); mVideoDecoder->enableErrorReport(mErrorReportEnabled); return OMX_ErrorNone; } OMX_COLOR_FORMATTYPE OMXVideoDecoderBase::GetOutputColorFormat(int width) { #ifndef VED_TILING return OMX_INTEL_COLOR_FormatYUV420PackedSemiPlanar; #else if (width > 1280 && width <= 2048) { LOGI("HD Video and use tiled format"); return OMX_INTEL_COLOR_FormatYUV420PackedSemiPlanar_Tiled; } else { return OMX_INTEL_COLOR_FormatYUV420PackedSemiPlanar; } #endif } OMX_ERRORTYPE OMXVideoDecoderBase::SetMaxOutputBufferCount(OMX_PARAM_PORTDEFINITIONTYPE *) { return OMX_ErrorNone; } uint32_t OMXVideoDecoderBase::getStride(uint32_t width) { uint32_t stride = 0; if (width <= 512) stride = 512; else if (width <= 1024) stride = 1024; else if (width <= 1280) { stride = 1280; } else if (width <= 2048) stride = 2048; else if (width <= 4096) stride = 4096; else stride = (width + 0x3f) & ~0x3f; return stride; }