/*
**
** Copyright 2008, The Android Open Source Project
** Copyright 2010, Samsung Electronics Co. LTD
**
** Licensed under the Apache License, Version 2.0 (the "License");
** you may not use this file except in compliance with the License.
** You may obtain a copy of the License at
**
** http://www.apache.org/licenses/LICENSE-2.0
**
** Unless required by applicable law or agreed to in writing, software
** distributed under the License is distributed on an "AS IS" BASIS,
** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
** See the License for the specific language governing permissions and
** limitations under the License.
*/
//#define LOG_NDEBUG 0
#define LOG_TAG "CameraHardwareSec"
#include <utils/Log.h>
#include "SecCameraHWInterface.h"
#include <utils/threads.h>
#include <fcntl.h>
#include <sys/mman.h>
#include <camera/Camera.h>
#include <MetadataBufferType.h>
#define VIDEO_COMMENT_MARKER_H 0xFFBE
#define VIDEO_COMMENT_MARKER_L 0xFFBF
#define VIDEO_COMMENT_MARKER_LENGTH 4
#define JPEG_EOI_MARKER 0xFFD9
#define HIBYTE(x) (((x) >> 8) & 0xFF)
#define LOBYTE(x) ((x) & 0xFF)
#define BACK_CAMERA_AUTO_FOCUS_DISTANCES_STR "0.10,1.20,Infinity"
#define BACK_CAMERA_MACRO_FOCUS_DISTANCES_STR "0.10,0.20,Infinity"
#define BACK_CAMERA_INFINITY_FOCUS_DISTANCES_STR "0.10,1.20,Infinity"
#define FRONT_CAMERA_FOCUS_DISTANCES_STR "0.20,0.25,Infinity"
// FIXME:
// -- The actual preview color is set to YV12. The preview frames
// returned via preview callback must be generated by color
// conversion if the requested preview color format for the
// preview frames is _not_ YV12. The reason that YV12 is used
// for actual preview is because that is the only color format
// supported by gralloc. Matching the preview cor format with
// gralloc color format improves performance since no color
// conversion is needed for preview.
//
// -- we only support two preview color formats that client
// applications can set: NV21 and YUV420/YV12.
namespace android {
struct addrs {
uint32_t type; // make sure that this is 4 byte.
unsigned int addr_y;
unsigned int addr_cbcr;
unsigned int buf_index;
unsigned int reserved;
};
struct addrs_cap {
unsigned int addr_y;
unsigned int width;
unsigned int height;
};
static const int INITIAL_SKIP_FRAME = 3;
static const int EFFECT_SKIP_FRAME = 1;
gralloc_module_t const* CameraHardwareSec::mGrallocHal;
CameraHardwareSec::CameraHardwareSec(int cameraId, camera_device_t *dev)
:
mCaptureInProgress(false),
mParameters(),
mCameraSensorName(NULL),
mSkipFrame(0),
mNotifyCb(0),
mDataCb(0),
mDataCbTimestamp(0),
mCallbackCookie(0),
mMsgEnabled(0),
mRecordRunning(false),
mPostViewWidth(0),
mPostViewHeight(0),
mPostViewSize(0),
mHalDevice(dev)
{
ALOGV("%s :", __func__);
int ret = 0;
mPreviewWindow = NULL;
mSecCamera = SecCamera::createInstance();
mRawHeap = NULL;
mPreviewHeap = NULL;
mRecordHeap = NULL;
if (!mGrallocHal) {
ret = hw_get_module(GRALLOC_HARDWARE_MODULE_ID, (const hw_module_t **)&mGrallocHal);
if (ret)
ALOGE("ERR(%s):Fail on loading gralloc HAL", __func__);
}
ret = mSecCamera->initCamera(cameraId);
if (ret < 0) {
ALOGE("ERR(%s):Fail on mSecCamera init", __func__);
}
mSecCamera->getPostViewConfig(&mPostViewWidth, &mPostViewHeight, &mPostViewSize);
ALOGV("mPostViewWidth = %d mPostViewHeight = %d mPostViewSize = %d",
mPostViewWidth,mPostViewHeight,mPostViewSize);
initDefaultParameters(cameraId);
mExitAutoFocusThread = false;
mExitPreviewThread = false;
/* whether the PreviewThread is active in preview or stopped. we
* create the thread but it is initially in stopped state.
*/
mPreviewRunning = false;
mPreviewStartDeferred = false;
mPreviewThread = new PreviewThread(this);
mAutoFocusThread = new AutoFocusThread(this);
mPictureThread = new PictureThread(this);
}
int CameraHardwareSec::getCameraId() const
{
return mSecCamera->getCameraId();
}
void CameraHardwareSec::initDefaultParameters(int cameraId)
{
if (mSecCamera == NULL) {
ALOGE("ERR(%s):mSecCamera object is NULL", __func__);
return;
}
CameraParameters p;
CameraParameters ip;
mCameraSensorName = mSecCamera->getCameraSensorName();
ALOGV("CameraSensorName: %s", mCameraSensorName);
int preview_max_width = 0;
int preview_max_height = 0;
int snapshot_max_width = 0;
int snapshot_max_height = 0;
if (cameraId == SecCamera::CAMERA_ID_BACK) {
p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_SIZES,
"720x480,640x480,352x288,176x144");
p.set(CameraParameters::KEY_SUPPORTED_PICTURE_SIZES,
"2560x1920,2048x1536,1600x1200,1280x960,640x480");
} else {
p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_SIZES,
"640x480,320x240,176x144");
p.set(CameraParameters::KEY_SUPPORTED_PICTURE_SIZES,
"640x480");
}
p.getSupportedPreviewSizes(mSupportedPreviewSizes);
// If these fail, then we are using an invalid cameraId and we'll leave the
// sizes at zero to catch the error.
if (mSecCamera->getPreviewMaxSize(&preview_max_width,
&preview_max_height) < 0)
ALOGE("getPreviewMaxSize fail (%d / %d) \n",
preview_max_width, preview_max_height);
if (mSecCamera->getSnapshotMaxSize(&snapshot_max_width,
&snapshot_max_height) < 0)
ALOGE("getSnapshotMaxSize fail (%d / %d) \n",
snapshot_max_width, snapshot_max_height);
String8 previewColorString;
previewColorString = CameraParameters::PIXEL_FORMAT_YUV420SP;
previewColorString.append(",");
previewColorString.append(CameraParameters::PIXEL_FORMAT_YUV420P);
p.setPreviewFormat(CameraParameters::PIXEL_FORMAT_YUV420SP);
p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FORMATS, previewColorString.string());
p.set(CameraParameters::KEY_VIDEO_FRAME_FORMAT, CameraParameters::PIXEL_FORMAT_YUV420P);
p.setPreviewSize(preview_max_width, preview_max_height);
p.setPictureFormat(CameraParameters::PIXEL_FORMAT_JPEG);
p.setPictureSize(snapshot_max_width, snapshot_max_height);
p.set(CameraParameters::KEY_JPEG_QUALITY, "100"); // maximum quality
p.set(CameraParameters::KEY_SUPPORTED_PICTURE_FORMATS,
CameraParameters::PIXEL_FORMAT_JPEG);
String8 parameterString;
if (cameraId == SecCamera::CAMERA_ID_BACK) {
parameterString = CameraParameters::FOCUS_MODE_AUTO;
parameterString.append(",");
parameterString.append(CameraParameters::FOCUS_MODE_INFINITY);
parameterString.append(",");
parameterString.append(CameraParameters::FOCUS_MODE_MACRO);
p.set(CameraParameters::KEY_SUPPORTED_FOCUS_MODES,
parameterString.string());
p.set(CameraParameters::KEY_FOCUS_MODE,
CameraParameters::FOCUS_MODE_AUTO);
p.set(CameraParameters::KEY_FOCUS_DISTANCES,
BACK_CAMERA_AUTO_FOCUS_DISTANCES_STR);
p.set(CameraParameters::KEY_SUPPORTED_JPEG_THUMBNAIL_SIZES,
"320x240,0x0");
p.set(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH, "320");
p.set(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT, "240");
p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FRAME_RATES, "30");
p.setPreviewFrameRate(30);
} else {
parameterString = CameraParameters::FOCUS_MODE_FIXED;
p.set(CameraParameters::KEY_SUPPORTED_FOCUS_MODES,
parameterString.string());
p.set(CameraParameters::KEY_FOCUS_MODE,
CameraParameters::FOCUS_MODE_FIXED);
p.set(CameraParameters::KEY_FOCUS_DISTANCES,
FRONT_CAMERA_FOCUS_DISTANCES_STR);
p.set(CameraParameters::KEY_SUPPORTED_JPEG_THUMBNAIL_SIZES,
"160x120,0x0");
p.set(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH, "160");
p.set(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT, "120");
p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FRAME_RATES, "15");
p.setPreviewFrameRate(15);
}
parameterString = CameraParameters::EFFECT_NONE;
parameterString.append(",");
parameterString.append(CameraParameters::EFFECT_MONO);
parameterString.append(",");
parameterString.append(CameraParameters::EFFECT_NEGATIVE);
parameterString.append(",");
parameterString.append(CameraParameters::EFFECT_SEPIA);
p.set(CameraParameters::KEY_SUPPORTED_EFFECTS, parameterString.string());
if (cameraId == SecCamera::CAMERA_ID_BACK) {
parameterString = CameraParameters::FLASH_MODE_ON;
parameterString.append(",");
parameterString.append(CameraParameters::FLASH_MODE_OFF);
parameterString.append(",");
parameterString.append(CameraParameters::FLASH_MODE_AUTO);
parameterString.append(",");
parameterString.append(CameraParameters::FLASH_MODE_TORCH);
p.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES,
parameterString.string());
p.set(CameraParameters::KEY_FLASH_MODE,
CameraParameters::FLASH_MODE_OFF);
parameterString = CameraParameters::SCENE_MODE_AUTO;
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_PORTRAIT);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_LANDSCAPE);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_NIGHT);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_BEACH);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_SNOW);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_SUNSET);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_FIREWORKS);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_SPORTS);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_PARTY);
parameterString.append(",");
parameterString.append(CameraParameters::SCENE_MODE_CANDLELIGHT);
p.set(CameraParameters::KEY_SUPPORTED_SCENE_MODES,
parameterString.string());
p.set(CameraParameters::KEY_SCENE_MODE,
CameraParameters::SCENE_MODE_AUTO);
/* we have two ranges, 4-30fps for night mode and
* 15-30fps for all others
*/
p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(15000,30000)");
p.set(CameraParameters::KEY_PREVIEW_FPS_RANGE, "15000,30000");
p.set(CameraParameters::KEY_FOCAL_LENGTH, "3.43");
} else {
p.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(7500,30000)");
p.set(CameraParameters::KEY_PREVIEW_FPS_RANGE, "7500,30000");
p.set(CameraParameters::KEY_FOCAL_LENGTH, "0.9");
}
parameterString = CameraParameters::WHITE_BALANCE_AUTO;
parameterString.append(",");
parameterString.append(CameraParameters::WHITE_BALANCE_INCANDESCENT);
parameterString.append(",");
parameterString.append(CameraParameters::WHITE_BALANCE_FLUORESCENT);
parameterString.append(",");
parameterString.append(CameraParameters::WHITE_BALANCE_DAYLIGHT);
parameterString.append(",");
parameterString.append(CameraParameters::WHITE_BALANCE_CLOUDY_DAYLIGHT);
p.set(CameraParameters::KEY_SUPPORTED_WHITE_BALANCE,
parameterString.string());
ip.set("sharpness-min", 0);
ip.set("sharpness-max", 4);
ip.set("saturation-min", 0);
ip.set("saturation-max", 4);
ip.set("contrast-min", 0);
ip.set("contrast-max", 4);
p.set(CameraParameters::KEY_JPEG_THUMBNAIL_QUALITY, "100");
p.set(CameraParameters::KEY_ROTATION, 0);
p.set(CameraParameters::KEY_WHITE_BALANCE, CameraParameters::WHITE_BALANCE_AUTO);
p.set(CameraParameters::KEY_EFFECT, CameraParameters::EFFECT_NONE);
ip.set("sharpness", SHARPNESS_DEFAULT);
ip.set("contrast", CONTRAST_DEFAULT);
ip.set("saturation", SATURATION_DEFAULT);
ip.set("iso", "auto");
ip.set("metering", "center");
ip.set("wdr", 0);
ip.set("chk_dataline", 0);
if (cameraId == SecCamera::CAMERA_ID_FRONT) {
ip.set("vtmode", 0);
ip.set("blur", 0);
}
p.set(CameraParameters::KEY_HORIZONTAL_VIEW_ANGLE, "51.2");
p.set(CameraParameters::KEY_VERTICAL_VIEW_ANGLE, "39.4");
p.set(CameraParameters::KEY_EXPOSURE_COMPENSATION, "0");
p.set(CameraParameters::KEY_MAX_EXPOSURE_COMPENSATION, "4");
p.set(CameraParameters::KEY_MIN_EXPOSURE_COMPENSATION, "-4");
p.set(CameraParameters::KEY_EXPOSURE_COMPENSATION_STEP, "0.5");
mParameters = p;
mInternalParameters = ip;
/* make sure mSecCamera has all the settings we do. applications
* aren't required to call setParameters themselves (only if they
* want to change something.
*/
setParameters(p);
mSecCamera->setISO(ISO_AUTO);
mSecCamera->setMetering(METERING_CENTER);
mSecCamera->setContrast(CONTRAST_DEFAULT);
mSecCamera->setSharpness(SHARPNESS_DEFAULT);
mSecCamera->setSaturation(SATURATION_DEFAULT);
if (cameraId == SecCamera::CAMERA_ID_BACK)
mSecCamera->setFrameRate(30);
else
mSecCamera->setFrameRate(15);
}
CameraHardwareSec::~CameraHardwareSec()
{
ALOGV("%s", __func__);
mSecCamera->DeinitCamera();
}
status_t CameraHardwareSec::setPreviewWindow(preview_stream_ops *w)
{
int min_bufs;
mPreviewWindow = w;
ALOGV("%s: mPreviewWindow %p", __func__, mPreviewWindow);
if (!w) {
ALOGE("preview window is NULL!");
return OK;
}
mPreviewLock.lock();
if (mPreviewRunning && !mPreviewStartDeferred) {
ALOGI("stop preview (window change)");
stopPreviewInternal();
}
if (w->get_min_undequeued_buffer_count(w, &min_bufs)) {
ALOGE("%s: could not retrieve min undequeued buffer count", __func__);
return INVALID_OPERATION;
}
if (min_bufs >= kBufferCount) {
ALOGE("%s: min undequeued buffer count %d is too high (expecting at most %d)", __func__,
min_bufs, kBufferCount - 1);
}
ALOGV("%s: setting buffer count to %d", __func__, kBufferCount);
if (w->set_buffer_count(w, kBufferCount)) {
ALOGE("%s: could not set buffer count", __func__);
return INVALID_OPERATION;
}
int preview_width;
int preview_height;
mParameters.getPreviewSize(&preview_width, &preview_height);
int hal_pixel_format = HAL_PIXEL_FORMAT_YV12;
const char *str_preview_format = mParameters.getPreviewFormat();
ALOGV("%s: preview format %s", __func__, str_preview_format);
if (w->set_usage(w, GRALLOC_USAGE_SW_WRITE_OFTEN)) {
ALOGE("%s: could not set usage on gralloc buffer", __func__);
return INVALID_OPERATION;
}
if (w->set_buffers_geometry(w,
preview_width, preview_height,
hal_pixel_format)) {
ALOGE("%s: could not set buffers geometry to %s",
__func__, str_preview_format);
return INVALID_OPERATION;
}
if (mPreviewRunning && mPreviewStartDeferred) {
ALOGV("start/resume preview");
status_t ret = startPreviewInternal();
if (ret == OK) {
mPreviewStartDeferred = false;
mPreviewCondition.signal();
}
}
mPreviewLock.unlock();
return OK;
}
void CameraHardwareSec::setCallbacks(camera_notify_callback notify_cb,
camera_data_callback data_cb,
camera_data_timestamp_callback data_cb_timestamp,
camera_request_memory get_memory,
void *user)
{
mNotifyCb = notify_cb;
mDataCb = data_cb;
mDataCbTimestamp = data_cb_timestamp;
mGetMemoryCb = get_memory;
mCallbackCookie = user;
}
void CameraHardwareSec::enableMsgType(int32_t msgType)
{
ALOGV("%s : msgType = 0x%x, mMsgEnabled before = 0x%x",
__func__, msgType, mMsgEnabled);
mMsgEnabled |= msgType;
ALOGV("%s : mMsgEnabled = 0x%x", __func__, mMsgEnabled);
}
void CameraHardwareSec::disableMsgType(int32_t msgType)
{
ALOGV("%s : msgType = 0x%x, mMsgEnabled before = 0x%x",
__func__, msgType, mMsgEnabled);
mMsgEnabled &= ~msgType;
ALOGV("%s : mMsgEnabled = 0x%x", __func__, mMsgEnabled);
}
bool CameraHardwareSec::msgTypeEnabled(int32_t msgType)
{
return (mMsgEnabled & msgType);
}
// ---------------------------------------------------------------------------
void CameraHardwareSec::setSkipFrame(int frame)
{
Mutex::Autolock lock(mSkipFrameLock);
if (frame < mSkipFrame)
return;
mSkipFrame = frame;
}
int CameraHardwareSec::previewThreadWrapper()
{
ALOGI("%s: starting", __func__);
while (1) {
mPreviewLock.lock();
while (!mPreviewRunning) {
ALOGI("%s: calling mSecCamera->stopPreview() and waiting", __func__);
mSecCamera->stopPreview();
/* signal that we're stopping */
mPreviewStoppedCondition.signal();
mPreviewCondition.wait(mPreviewLock);
ALOGI("%s: return from wait", __func__);
}
mPreviewLock.unlock();
if (mExitPreviewThread) {
ALOGI("%s: exiting", __func__);
mSecCamera->stopPreview();
return 0;
}
previewThread();
}
}
int CameraHardwareSec::previewThread()
{
int index;
nsecs_t timestamp;
unsigned int phyYAddr;
unsigned int phyCAddr;
struct addrs *addrs;
index = mSecCamera->getPreview();
if (index < 0) {
ALOGE("ERR(%s):Fail on SecCamera->getPreview()", __func__);
return UNKNOWN_ERROR;
}
// ALOGV("%s: index %d", __func__, index);
mSkipFrameLock.lock();
if (mSkipFrame > 0) {
mSkipFrame--;
mSkipFrameLock.unlock();
ALOGV("%s: index %d skipping frame", __func__, index);
return NO_ERROR;
}
mSkipFrameLock.unlock();
timestamp = systemTime(SYSTEM_TIME_MONOTONIC);
phyYAddr = mSecCamera->getPhyAddrY(index);
phyCAddr = mSecCamera->getPhyAddrC(index);
if (phyYAddr == 0xffffffff || phyCAddr == 0xffffffff) {
ALOGE("ERR(%s):Fail on SecCamera getPhyAddr Y addr = %0x C addr = %0x",
__func__, phyYAddr, phyCAddr);
return UNKNOWN_ERROR;
}
int width, height, frame_size, offset;
mSecCamera->getPreviewSize(&width, &height, &frame_size);
offset = frame_size * index;
if (mPreviewWindow && mGrallocHal) {
buffer_handle_t *buf_handle;
int stride;
if (0 != mPreviewWindow->dequeue_buffer(mPreviewWindow, &buf_handle, &stride)) {
ALOGE("Could not dequeue gralloc buffer!\n");
goto callbacks;
}
void *vaddr;
if (!mGrallocHal->lock(mGrallocHal,
*buf_handle,
GRALLOC_USAGE_SW_WRITE_OFTEN,
0, 0, width, height, &vaddr)) {
char *frame = ((char *)mPreviewHeap->data) + offset;
// the code below assumes YUV, not RGB
{
int h;
char *src = frame;
char *ptr = (char *)vaddr;
// Copy the Y plane, while observing the stride
for (h = 0; h < height; h++) {
memcpy(ptr, src, width);
ptr += stride;
src += width;
}
{
// U
char *v = ptr;
ptr += stride * height / 4;
for (h = 0; h < height / 2; h++) {
memcpy(ptr, src, width / 2);
ptr += stride / 2;
src += width / 2;
}
// V
ptr = v;
for (h = 0; h < height / 2; h++) {
memcpy(ptr, src, width / 2);
ptr += stride / 2;
src += width / 2;
}
}
}
mGrallocHal->unlock(mGrallocHal, *buf_handle);
}
else
ALOGE("%s: could not obtain gralloc buffer", __func__);
if (0 != mPreviewWindow->enqueue_buffer(mPreviewWindow, buf_handle)) {
ALOGE("Could not enqueue gralloc buffer!\n");
goto callbacks;
}
}
callbacks:
// Notify the client of a new frame.
if (mMsgEnabled & CAMERA_MSG_PREVIEW_FRAME) {
const char * preview_format = mParameters.getPreviewFormat();
if (!strcmp(preview_format, CameraParameters::PIXEL_FORMAT_YUV420SP)) {
// Color conversion from YUV420 to NV21
char *vu = ((char *)mPreviewHeap->data) + offset + width * height;
const int uv_size = (width * height) >> 1;
char saved_uv[uv_size];
memcpy(saved_uv, vu, uv_size);
char *u = saved_uv;
char *v = u + (uv_size >> 1);
int h = 0;
while (h < width * height / 4) {
*vu++ = *v++;
*vu++ = *u++;
++h;
}
}
mDataCb(CAMERA_MSG_PREVIEW_FRAME, mPreviewHeap, index, NULL, mCallbackCookie);
}
Mutex::Autolock lock(mRecordLock);
if (mRecordRunning == true) {
index = mSecCamera->getRecordFrame();
if (index < 0) {
ALOGE("ERR(%s):Fail on SecCamera->getRecord()", __func__);
return UNKNOWN_ERROR;
}
phyYAddr = mSecCamera->getRecPhyAddrY(index);
phyCAddr = mSecCamera->getRecPhyAddrC(index);
if (phyYAddr == 0xffffffff || phyCAddr == 0xffffffff) {
ALOGE("ERR(%s):Fail on SecCamera getRectPhyAddr Y addr = %0x C addr = %0x", __func__,
phyYAddr, phyCAddr);
return UNKNOWN_ERROR;
}
addrs = (struct addrs *)mRecordHeap->data;
addrs[index].type = kMetadataBufferTypeCameraSource;
addrs[index].addr_y = phyYAddr;
addrs[index].addr_cbcr = phyCAddr;
addrs[index].buf_index = index;
// Notify the client of a new frame.
if (mMsgEnabled & CAMERA_MSG_VIDEO_FRAME) {
mDataCbTimestamp(timestamp, CAMERA_MSG_VIDEO_FRAME,
mRecordHeap, index, mCallbackCookie);
} else {
mSecCamera->releaseRecordFrame(index);
}
}
return NO_ERROR;
}
status_t CameraHardwareSec::startPreview()
{
int ret = 0; //s1 [Apply factory standard]
ALOGV("%s :", __func__);
if (waitCaptureCompletion() != NO_ERROR) {
return TIMED_OUT;
}
mPreviewLock.lock();
if (mPreviewRunning) {
// already running
ALOGE("%s : preview thread already running", __func__);
mPreviewLock.unlock();
return INVALID_OPERATION;
}
mPreviewRunning = true;
mPreviewStartDeferred = false;
if (!mPreviewWindow) {
ALOGI("%s : deferring", __func__);
mPreviewStartDeferred = true;
mPreviewLock.unlock();
return NO_ERROR;
}
ret = startPreviewInternal();
if (ret == OK)
mPreviewCondition.signal();
mPreviewLock.unlock();
return ret;
}
status_t CameraHardwareSec::startPreviewInternal()
{
ALOGV("%s", __func__);
int ret = mSecCamera->startPreview();
ALOGV("%s : mSecCamera->startPreview() returned %d", __func__, ret);
if (ret < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->startPreview()", __func__);
return UNKNOWN_ERROR;
}
setSkipFrame(INITIAL_SKIP_FRAME);
int width, height, frame_size;
mSecCamera->getPreviewSize(&width, &height, &frame_size);
ALOGD("mPreviewHeap(fd(%d), size(%d), width(%d), height(%d))",
mSecCamera->getCameraFd(), frame_size, width, height);
if (mPreviewHeap) {
mPreviewHeap->release(mPreviewHeap);
mPreviewHeap = 0;
}
mPreviewHeap = mGetMemoryCb((int)mSecCamera->getCameraFd(),
frame_size,
kBufferCount,
0); // no cookie
mSecCamera->getPostViewConfig(&mPostViewWidth, &mPostViewHeight, &mPostViewSize);
ALOGV("CameraHardwareSec: mPostViewWidth = %d mPostViewHeight = %d mPostViewSize = %d",
mPostViewWidth,mPostViewHeight,mPostViewSize);
return NO_ERROR;
}
void CameraHardwareSec::stopPreviewInternal()
{
ALOGV("%s :", __func__);
/* request that the preview thread stop. */
if (mPreviewRunning) {
mPreviewRunning = false;
if (!mPreviewStartDeferred) {
mPreviewCondition.signal();
/* wait until preview thread is stopped */
mPreviewStoppedCondition.wait(mPreviewLock);
}
else
ALOGV("%s : preview running but deferred, doing nothing", __func__);
} else
ALOGI("%s : preview not running, doing nothing", __func__);
}
void CameraHardwareSec::stopPreview()
{
ALOGV("%s :", __func__);
/* request that the preview thread stop. */
mPreviewLock.lock();
stopPreviewInternal();
mPreviewLock.unlock();
}
bool CameraHardwareSec::previewEnabled()
{
Mutex::Autolock lock(mPreviewLock);
ALOGV("%s : %d", __func__, mPreviewRunning);
return mPreviewRunning;
}
// ---------------------------------------------------------------------------
status_t CameraHardwareSec::startRecording()
{
ALOGV("%s :", __func__);
Mutex::Autolock lock(mRecordLock);
if (mRecordHeap) {
mRecordHeap->release(mRecordHeap);
mRecordHeap = 0;
}
mRecordHeap = mGetMemoryCb(-1, sizeof(struct addrs), kBufferCount, NULL);
if (!mRecordHeap) {
ALOGE("ERR(%s): Record heap creation fail", __func__);
return UNKNOWN_ERROR;
}
if (mRecordRunning == false) {
if (mSecCamera->startRecord() < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->startRecord()", __func__);
return UNKNOWN_ERROR;
}
mRecordRunning = true;
}
return NO_ERROR;
}
void CameraHardwareSec::stopRecording()
{
ALOGV("%s :", __func__);
Mutex::Autolock lock(mRecordLock);
if (mRecordRunning == true) {
if (mSecCamera->stopRecord() < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->stopRecord()", __func__);
return;
}
mRecordRunning = false;
}
}
bool CameraHardwareSec::recordingEnabled()
{
ALOGV("%s :", __func__);
return mRecordRunning;
}
void CameraHardwareSec::releaseRecordingFrame(const void *opaque)
{
struct addrs *addrs = (struct addrs *)opaque;
mSecCamera->releaseRecordFrame(addrs->buf_index);
}
// ---------------------------------------------------------------------------
int CameraHardwareSec::autoFocusThread()
{
int count =0;
int af_status =0 ;
ALOGV("%s : starting", __func__);
/* block until we're told to start. we don't want to use
* a restartable thread and requestExitAndWait() in cancelAutoFocus()
* because it would cause deadlock between our callbacks and the
* caller of cancelAutoFocus() which both want to grab the same lock
* in CameraServices layer.
*/
mFocusLock.lock();
/* check early exit request */
if (mExitAutoFocusThread) {
mFocusLock.unlock();
ALOGV("%s : exiting on request0", __func__);
return NO_ERROR;
}
mFocusCondition.wait(mFocusLock);
/* check early exit request */
if (mExitAutoFocusThread) {
mFocusLock.unlock();
ALOGV("%s : exiting on request1", __func__);
return NO_ERROR;
}
mFocusLock.unlock();
ALOGV("%s : calling setAutoFocus", __func__);
if (mSecCamera->setAutofocus() < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setAutofocus()", __func__);
return UNKNOWN_ERROR;
}
af_status = mSecCamera->getAutoFocusResult();
if (af_status == 0x01) {
ALOGV("%s : AF Success!!", __func__);
if (mMsgEnabled & CAMERA_MSG_FOCUS)
mNotifyCb(CAMERA_MSG_FOCUS, true, 0, mCallbackCookie);
} else if (af_status == 0x02) {
ALOGV("%s : AF Cancelled !!", __func__);
if (mMsgEnabled & CAMERA_MSG_FOCUS) {
/* CAMERA_MSG_FOCUS only takes a bool. true for
* finished and false for failure. cancel is still
* considered a true result.
*/
mNotifyCb(CAMERA_MSG_FOCUS, true, 0, mCallbackCookie);
}
} else {
ALOGV("%s : AF Fail !!", __func__);
ALOGV("%s : mMsgEnabled = 0x%x", __func__, mMsgEnabled);
if (mMsgEnabled & CAMERA_MSG_FOCUS)
mNotifyCb(CAMERA_MSG_FOCUS, false, 0, mCallbackCookie);
}
ALOGV("%s : exiting with no error", __func__);
return NO_ERROR;
}
status_t CameraHardwareSec::autoFocus()
{
ALOGV("%s :", __func__);
/* signal autoFocusThread to run once */
mFocusCondition.signal();
return NO_ERROR;
}
/* 2009.10.14 by icarus for added interface */
status_t CameraHardwareSec::cancelAutoFocus()
{
ALOGV("%s :", __func__);
// If preview is not running, cancel autofocus can still be called.
// Since the camera subsystem is completely reset on preview start,
// cancel AF is a no-op.
if (!mPreviewRunning) return NO_ERROR;
// cancelAutoFocus should be allowed after preview is started. But if
// the preview is deferred, cancelAutoFocus will fail. Ignore it if that is
// the case.
if (mPreviewRunning && mPreviewStartDeferred) return NO_ERROR;
if (mSecCamera->cancelAutofocus() < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->cancelAutofocus()", __func__);
return UNKNOWN_ERROR;
}
return NO_ERROR;
}
int CameraHardwareSec::save_jpeg( unsigned char *real_jpeg, int jpeg_size)
{
FILE *yuv_fp = NULL;
char filename[100], *buffer = NULL;
/* file create/open, note to "wb" */
yuv_fp = fopen("/data/camera_dump.jpeg", "wb");
if (yuv_fp == NULL) {
ALOGE("Save jpeg file open error");
return -1;
}
ALOGV("[BestIQ] real_jpeg size ========> %d\n", jpeg_size);
buffer = (char *) malloc(jpeg_size);
if (buffer == NULL) {
ALOGE("Save YUV] buffer alloc failed");
if (yuv_fp)
fclose(yuv_fp);
return -1;
}
memcpy(buffer, real_jpeg, jpeg_size);
fflush(stdout);
fwrite(buffer, 1, jpeg_size, yuv_fp);
fflush(yuv_fp);
if (yuv_fp)
fclose(yuv_fp);
if (buffer)
free(buffer);
return 0;
}
void CameraHardwareSec::save_postview(const char *fname, uint8_t *buf, uint32_t size)
{
int nw;
int cnt = 0;
uint32_t written = 0;
ALOGD("opening file [%s]\n", fname);
int fd = open(fname, O_RDWR | O_CREAT);
if (fd < 0) {
ALOGE("failed to create file [%s]: %s", fname, strerror(errno));
return;
}
ALOGD("writing %d bytes to file [%s]\n", size, fname);
while (written < size) {
nw = ::write(fd, buf + written, size - written);
if (nw < 0) {
ALOGE("failed to write to file %d [%s]: %s",written,fname, strerror(errno));
break;
}
written += nw;
cnt++;
}
ALOGD("done writing %d bytes to file [%s] in %d passes\n",size, fname, cnt);
::close(fd);
}
bool CameraHardwareSec::scaleDownYuv422(char *srcBuf, uint32_t srcWidth, uint32_t srcHeight,
char *dstBuf, uint32_t dstWidth, uint32_t dstHeight)
{
int32_t step_x, step_y;
int32_t iXsrc, iXdst;
int32_t x, y, src_y_start_pos, dst_pos, src_pos;
if (dstWidth % 2 != 0 || dstHeight % 2 != 0){
ALOGE("scale_down_yuv422: invalid width, height for scaling");
return false;
}
step_x = srcWidth / dstWidth;
step_y = srcHeight / dstHeight;
dst_pos = 0;
for (uint32_t y = 0; y < dstHeight; y++) {
src_y_start_pos = (y * step_y * (srcWidth * 2));
for (uint32_t x = 0; x < dstWidth; x += 2) {
src_pos = src_y_start_pos + (x * (step_x * 2));
dstBuf[dst_pos++] = srcBuf[src_pos ];
dstBuf[dst_pos++] = srcBuf[src_pos + 1];
dstBuf[dst_pos++] = srcBuf[src_pos + 2];
dstBuf[dst_pos++] = srcBuf[src_pos + 3];
}
}
return true;
}
bool CameraHardwareSec::YUY2toNV21(void *srcBuf, void *dstBuf, uint32_t srcWidth, uint32_t srcHeight)
{
int32_t x, y, src_y_start_pos, dst_cbcr_pos, dst_pos, src_pos;
unsigned char *srcBufPointer = (unsigned char *)srcBuf;
unsigned char *dstBufPointer = (unsigned char *)dstBuf;
dst_pos = 0;
dst_cbcr_pos = srcWidth*srcHeight;
for (uint32_t y = 0; y < srcHeight; y++) {
src_y_start_pos = (y * (srcWidth * 2));
for (uint32_t x = 0; x < (srcWidth * 2); x += 2) {
src_pos = src_y_start_pos + x;
dstBufPointer[dst_pos++] = srcBufPointer[src_pos];
}
}
for (uint32_t y = 0; y < srcHeight; y += 2) {
src_y_start_pos = (y * (srcWidth * 2));
for (uint32_t x = 0; x < (srcWidth * 2); x += 4) {
src_pos = src_y_start_pos + x;
dstBufPointer[dst_cbcr_pos++] = srcBufPointer[src_pos + 3];
dstBufPointer[dst_cbcr_pos++] = srcBufPointer[src_pos + 1];
}
}
return true;
}
int CameraHardwareSec::pictureThread()
{
ALOGV("%s :", __func__);
int jpeg_size = 0;
int ret = NO_ERROR;
unsigned char *jpeg_data = NULL;
int postview_offset = 0;
unsigned char *postview_data = NULL;
unsigned char *addr = NULL;
int mPostViewWidth, mPostViewHeight, mPostViewSize;
int mThumbWidth, mThumbHeight, mThumbSize;
int cap_width, cap_height, cap_frame_size;
int JpegImageSize, JpegExifSize;
bool isLSISensor = false;
unsigned int output_size = 0;
mSecCamera->getPostViewConfig(&mPostViewWidth, &mPostViewHeight, &mPostViewSize);
mSecCamera->getThumbnailConfig(&mThumbWidth, &mThumbHeight, &mThumbSize);
int postviewHeapSize = mPostViewSize;
mSecCamera->getSnapshotSize(&cap_width, &cap_height, &cap_frame_size);
int mJpegHeapSize;
if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK)
mJpegHeapSize = cap_frame_size * SecCamera::getJpegRatio();
else
mJpegHeapSize = cap_frame_size;
LOG_TIME_DEFINE(0)
LOG_TIME_START(0)
// sp<MemoryBase> buffer = new MemoryBase(mRawHeap, 0, mPostViewSize + 8);
struct addrs_cap *addrs = (struct addrs_cap *)mRawHeap->data;
addrs[0].width = mPostViewWidth;
addrs[0].height = mPostViewHeight;
ALOGV("[5B] mPostViewWidth = %d mPostViewHeight = %d\n",mPostViewWidth,mPostViewHeight);
camera_memory_t *JpegHeap = mGetMemoryCb(-1, mJpegHeapSize, 1, 0);
sp<MemoryHeapBase> PostviewHeap = new MemoryHeapBase(mPostViewSize);
sp<MemoryHeapBase> ThumbnailHeap = new MemoryHeapBase(mThumbSize);
LOG_TIME_DEFINE(1)
LOG_TIME_START(1)
int picture_size, picture_width, picture_height;
mSecCamera->getSnapshotSize(&picture_width, &picture_height, &picture_size);
int picture_format = mSecCamera->getSnapshotPixelFormat();
unsigned int phyAddr;
// Modified the shutter sound timing for Jpeg capture
if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK)
mSecCamera->setSnapshotCmd();
if (mMsgEnabled & CAMERA_MSG_SHUTTER) {
mNotifyCb(CAMERA_MSG_SHUTTER, 0, 0, mCallbackCookie);
}
if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK){
jpeg_data = mSecCamera->getJpeg(&jpeg_size, &phyAddr);
if (jpeg_data == NULL) {
ALOGE("ERR(%s):Fail on SecCamera->getSnapshot()", __func__);
ret = UNKNOWN_ERROR;
goto out;
}
} else {
if (mSecCamera->getSnapshotAndJpeg((unsigned char*)PostviewHeap->base(),
(unsigned char*)JpegHeap->data, &output_size) < 0) {
ret = UNKNOWN_ERROR;
goto out;
}
ALOGI("snapshotandjpeg done\n");
}
LOG_TIME_END(1)
LOG_CAMERA("getSnapshotAndJpeg interval: %lu us", LOG_TIME(1));
if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) {
isLSISensor = !strncmp((const char*)mCameraSensorName, "S5K4ECGX", 8);
if(isLSISensor) {
ALOGI("== Camera Sensor Detect %s - Samsung LSI SOC 5M ==\n", mCameraSensorName);
// LSI 5M SOC
if (!SplitFrame(jpeg_data, SecCamera::getInterleaveDataSize(),
SecCamera::getJpegLineLength(),
mPostViewWidth * 2, mPostViewWidth,
JpegHeap->data, &JpegImageSize,
PostviewHeap->base(), &mPostViewSize)) {
ret = UNKNOWN_ERROR;
goto out;
}
} else {
ALOGI("== Camera Sensor Detect %s Sony SOC 5M ==\n", mCameraSensorName);
decodeInterleaveData(jpeg_data,
SecCamera::getInterleaveDataSize(),
mPostViewWidth, mPostViewHeight,
&JpegImageSize, JpegHeap->data, PostviewHeap->base());
}
} else {
JpegImageSize = static_cast<int>(output_size);
}
scaleDownYuv422((char *)PostviewHeap->base(), mPostViewWidth, mPostViewHeight,
(char *)ThumbnailHeap->base(), mThumbWidth, mThumbHeight);
memcpy(mRawHeap->data, PostviewHeap->base(), postviewHeapSize);
if (mMsgEnabled & CAMERA_MSG_RAW_IMAGE) {
mDataCb(CAMERA_MSG_RAW_IMAGE, mRawHeap, 0, NULL, mCallbackCookie);
} else if (mMsgEnabled & CAMERA_MSG_RAW_IMAGE_NOTIFY) {
mNotifyCb(CAMERA_MSG_RAW_IMAGE_NOTIFY, 0, 0, mCallbackCookie);
}
if (mMsgEnabled & CAMERA_MSG_COMPRESSED_IMAGE) {
camera_memory_t *ExifHeap =
mGetMemoryCb(-1, EXIF_FILE_SIZE + JPG_STREAM_BUF_SIZE, 1, 0);
JpegExifSize = mSecCamera->getExif((unsigned char *)ExifHeap->data,
(unsigned char *)ThumbnailHeap->base());
ALOGV("JpegExifSize=%d", JpegExifSize);
if (JpegExifSize < 0) {
ret = UNKNOWN_ERROR;
ExifHeap->release(ExifHeap);
goto out;
}
camera_memory_t *mem = mGetMemoryCb(-1, JpegImageSize + JpegExifSize, 1, 0);
uint8_t *ptr = (uint8_t *) mem->data;
memcpy(ptr, JpegHeap->data, 2); ptr += 2;
memcpy(ptr, ExifHeap->data, JpegExifSize); ptr += JpegExifSize;
memcpy(ptr, (uint8_t *) JpegHeap->data + 2, JpegImageSize - 2);
mDataCb(CAMERA_MSG_COMPRESSED_IMAGE, mem, 0, NULL, mCallbackCookie);
mem->release(mem);
ExifHeap->release(ExifHeap);
}
LOG_TIME_END(0)
LOG_CAMERA("pictureThread interval: %lu us", LOG_TIME(0));
ALOGV("%s : pictureThread end", __func__);
out:
JpegHeap->release(JpegHeap);
mSecCamera->endSnapshot();
mCaptureLock.lock();
mCaptureInProgress = false;
mCaptureCondition.broadcast();
mCaptureLock.unlock();
return ret;
}
status_t CameraHardwareSec::waitCaptureCompletion() {
// 5 seconds timeout
nsecs_t endTime = 5000000000LL + systemTime(SYSTEM_TIME_MONOTONIC);
Mutex::Autolock lock(mCaptureLock);
while (mCaptureInProgress) {
nsecs_t remainingTime = endTime - systemTime(SYSTEM_TIME_MONOTONIC);
if (remainingTime <= 0) {
ALOGE("Timed out waiting picture thread.");
return TIMED_OUT;
}
ALOGD("Waiting for picture thread to complete.");
mCaptureCondition.waitRelative(mCaptureLock, remainingTime);
}
return NO_ERROR;
}
status_t CameraHardwareSec::takePicture()
{
ALOGV("%s :", __func__);
stopPreview();
if (!mRawHeap) {
int rawHeapSize = mPostViewSize;
ALOGV("mRawHeap : MemoryHeapBase(previewHeapSize(%d))", rawHeapSize);
mRawHeap = mGetMemoryCb(-1, rawHeapSize, 1, 0);
if (!mRawHeap) {
ALOGE("ERR(%s): Raw heap creation fail", __func__);
}
}
if (waitCaptureCompletion() != NO_ERROR) {
return TIMED_OUT;
}
if (mPictureThread->run("CameraPictureThread", PRIORITY_DEFAULT) != NO_ERROR) {
ALOGE("%s : couldn't run picture thread", __func__);
return INVALID_OPERATION;
}
mCaptureLock.lock();
mCaptureInProgress = true;
mCaptureLock.unlock();
return NO_ERROR;
}
status_t CameraHardwareSec::cancelPicture()
{
ALOGV("%s", __func__);
if (mPictureThread.get()) {
ALOGV("%s: waiting for picture thread to exit", __func__);
mPictureThread->requestExitAndWait();
ALOGV("%s: picture thread has exited", __func__);
}
return NO_ERROR;
}
bool CameraHardwareSec::CheckVideoStartMarker(unsigned char *pBuf)
{
if (!pBuf) {
ALOGE("CheckVideoStartMarker() => pBuf is NULL\n");
return false;
}
if (HIBYTE(VIDEO_COMMENT_MARKER_H) == * pBuf && LOBYTE(VIDEO_COMMENT_MARKER_H) == *(pBuf + 1) &&
HIBYTE(VIDEO_COMMENT_MARKER_L) == *(pBuf + 2) && LOBYTE(VIDEO_COMMENT_MARKER_L) == *(pBuf + 3))
return true;
return false;
}
bool CameraHardwareSec::CheckEOIMarker(unsigned char *pBuf)
{
if (!pBuf) {
ALOGE("CheckEOIMarker() => pBuf is NULL\n");
return false;
}
// EOI marker [FF D9]
if (HIBYTE(JPEG_EOI_MARKER) == *pBuf && LOBYTE(JPEG_EOI_MARKER) == *(pBuf + 1))
return true;
return false;
}
bool CameraHardwareSec::FindEOIMarkerInJPEG(unsigned char *pBuf, int dwBufSize, int *pnJPEGsize)
{
if (NULL == pBuf || 0 >= dwBufSize) {
ALOGE("FindEOIMarkerInJPEG() => There is no contents.");
return false;
}
unsigned char *pBufEnd = pBuf + dwBufSize;
while (pBuf < pBufEnd) {
if (CheckEOIMarker(pBuf++))
return true;
(*pnJPEGsize)++;
}
return false;
}
bool CameraHardwareSec::SplitFrame(unsigned char *pFrame, int dwSize,
int dwJPEGLineLength, int dwVideoLineLength, int dwVideoHeight,
void *pJPEG, int *pdwJPEGSize,
void *pVideo, int *pdwVideoSize)
{
ALOGV("===========SplitFrame Start==============");
if (NULL == pFrame || 0 >= dwSize) {
ALOGE("There is no contents (pFrame=%p, dwSize=%d", pFrame, dwSize);
return false;
}
if (0 == dwJPEGLineLength || 0 == dwVideoLineLength) {
ALOGE("There in no input information for decoding interleaved jpeg");
return false;
}
unsigned char *pSrc = pFrame;
unsigned char *pSrcEnd = pFrame + dwSize;
unsigned char *pJ = (unsigned char *)pJPEG;
int dwJSize = 0;
unsigned char *pV = (unsigned char *)pVideo;
int dwVSize = 0;
bool bRet = false;
bool isFinishJpeg = false;
while (pSrc < pSrcEnd) {
// Check video start marker
if (CheckVideoStartMarker(pSrc)) {
int copyLength;
if (pSrc + dwVideoLineLength <= pSrcEnd)
copyLength = dwVideoLineLength;
else
copyLength = pSrcEnd - pSrc - VIDEO_COMMENT_MARKER_LENGTH;
// Copy video data
if (pV) {
memcpy(pV, pSrc + VIDEO_COMMENT_MARKER_LENGTH, copyLength);
pV += copyLength;
dwVSize += copyLength;
}
pSrc += copyLength + VIDEO_COMMENT_MARKER_LENGTH;
} else {
// Copy pure JPEG data
int size = 0;
int dwCopyBufLen = dwJPEGLineLength <= pSrcEnd-pSrc ? dwJPEGLineLength : pSrcEnd - pSrc;
if (FindEOIMarkerInJPEG((unsigned char *)pSrc, dwCopyBufLen, &size)) {
isFinishJpeg = true;
size += 2; // to count EOF marker size
} else {
if ((dwCopyBufLen == 1) && (pJPEG < pJ)) {
unsigned char checkBuf[2] = { *(pJ - 1), *pSrc };
if (CheckEOIMarker(checkBuf))
isFinishJpeg = true;
}
size = dwCopyBufLen;
}
memcpy(pJ, pSrc, size);
dwJSize += size;
pJ += dwCopyBufLen;
pSrc += dwCopyBufLen;
}
if (isFinishJpeg)
break;
}
if (isFinishJpeg) {
bRet = true;
if(pdwJPEGSize)
*pdwJPEGSize = dwJSize;
if(pdwVideoSize)
*pdwVideoSize = dwVSize;
} else {
ALOGE("DecodeInterleaveJPEG_WithOutDT() => Can not find EOI");
bRet = false;
if(pdwJPEGSize)
*pdwJPEGSize = 0;
if(pdwVideoSize)
*pdwVideoSize = 0;
}
ALOGV("===========SplitFrame end==============");
return bRet;
}
int CameraHardwareSec::decodeInterleaveData(unsigned char *pInterleaveData,
int interleaveDataSize,
int yuvWidth,
int yuvHeight,
int *pJpegSize,
void *pJpegData,
void *pYuvData)
{
if (pInterleaveData == NULL)
return false;
bool ret = true;
unsigned int *interleave_ptr = (unsigned int *)pInterleaveData;
unsigned char *jpeg_ptr = (unsigned char *)pJpegData;
unsigned char *yuv_ptr = (unsigned char *)pYuvData;
unsigned char *p;
int jpeg_size = 0;
int yuv_size = 0;
int i = 0;
ALOGV("decodeInterleaveData Start~~~");
while (i < interleaveDataSize) {
if ((*interleave_ptr == 0xFFFFFFFF) || (*interleave_ptr == 0x02FFFFFF) ||
(*interleave_ptr == 0xFF02FFFF)) {
// Padding Data
// ALOGE("%d(%x) padding data\n", i, *interleave_ptr);
interleave_ptr++;
i += 4;
}
else if ((*interleave_ptr & 0xFFFF) == 0x05FF) {
// Start-code of YUV Data
// ALOGE("%d(%x) yuv data\n", i, *interleave_ptr);
p = (unsigned char *)interleave_ptr;
p += 2;
i += 2;
// Extract YUV Data
if (pYuvData != NULL) {
memcpy(yuv_ptr, p, yuvWidth * 2);
yuv_ptr += yuvWidth * 2;
yuv_size += yuvWidth * 2;
}
p += yuvWidth * 2;
i += yuvWidth * 2;
// Check End-code of YUV Data
if ((*p == 0xFF) && (*(p + 1) == 0x06)) {
interleave_ptr = (unsigned int *)(p + 2);
i += 2;
} else {
ret = false;
break;
}
} else {
// Extract JPEG Data
// ALOGE("%d(%x) jpg data, jpeg_size = %d bytes\n", i, *interleave_ptr, jpeg_size);
if (pJpegData != NULL) {
memcpy(jpeg_ptr, interleave_ptr, 4);
jpeg_ptr += 4;
jpeg_size += 4;
}
interleave_ptr++;
i += 4;
}
}
if (ret) {
if (pJpegData != NULL) {
// Remove Padding after EOI
for (i = 0; i < 3; i++) {
if (*(--jpeg_ptr) != 0xFF) {
break;
}
jpeg_size--;
}
*pJpegSize = jpeg_size;
}
// Check YUV Data Size
if (pYuvData != NULL) {
if (yuv_size != (yuvWidth * yuvHeight * 2)) {
ret = false;
}
}
}
ALOGV("decodeInterleaveData End~~~");
return ret;
}
status_t CameraHardwareSec::dump(int fd) const
{
const size_t SIZE = 256;
char buffer[SIZE];
String8 result;
const Vector<String16> args;
if (mSecCamera != 0) {
mSecCamera->dump(fd);
mParameters.dump(fd, args);
mInternalParameters.dump(fd, args);
snprintf(buffer, 255, " preview running(%s)\n", mPreviewRunning?"true": "false");
result.append(buffer);
} else {
result.append("No camera client yet.\n");
}
write(fd, result.string(), result.size());
return NO_ERROR;
}
bool CameraHardwareSec::isSupportedPreviewSize(const int width,
const int height) const
{
unsigned int i;
for (i = 0; i < mSupportedPreviewSizes.size(); i++) {
if (mSupportedPreviewSizes[i].width == width &&
mSupportedPreviewSizes[i].height == height)
return true;
}
return false;
}
bool CameraHardwareSec::isSupportedParameter(const char * const parm,
const char * const supported_parm) const
{
const char *pStart;
const char *pEnd;
if (!parm || !supported_parm)
return false;
pStart = supported_parm;
while (true) {
pEnd = strchr(pStart, ',');
if (!pEnd) {
if (!strcmp(parm, pStart))
return true;
else
return false;
}
if (!strncmp(parm, pStart, pEnd - pStart)) {
return true;
}
pStart = pEnd + 1;
}
/* NOTREACHED */
}
status_t CameraHardwareSec::setParameters(const CameraParameters& params)
{
ALOGV("%s :", __func__);
status_t ret = NO_ERROR;
/* if someone calls us while picture thread is running, it could screw
* up the sensor quite a bit so return error.
*/
if (waitCaptureCompletion() != NO_ERROR) {
return TIMED_OUT;
}
// preview size
int new_preview_width = 0;
int new_preview_height = 0;
params.getPreviewSize(&new_preview_width, &new_preview_height);
const char *new_str_preview_format = params.getPreviewFormat();
ALOGV("%s : new_preview_width x new_preview_height = %dx%d, format = %s",
__func__, new_preview_width, new_preview_height, new_str_preview_format);
if (strcmp(new_str_preview_format, CameraParameters::PIXEL_FORMAT_YUV420SP) &&
strcmp(new_str_preview_format, CameraParameters::PIXEL_FORMAT_YUV420P)) {
ALOGE("Unsupported preview color format: %s", new_str_preview_format);
return BAD_VALUE;
}
if (0 < new_preview_width && 0 < new_preview_height &&
new_str_preview_format != NULL &&
isSupportedPreviewSize(new_preview_width, new_preview_height)) {
int new_preview_format = V4L2_PIX_FMT_YUV420;
int current_preview_width, current_preview_height, current_frame_size;
mSecCamera->getPreviewSize(¤t_preview_width,
¤t_preview_height,
¤t_frame_size);
int current_pixel_format = mSecCamera->getPreviewPixelFormat();
if (current_preview_width != new_preview_width ||
current_preview_height != new_preview_height ||
current_pixel_format != new_preview_format) {
if (mSecCamera->setPreviewSize(new_preview_width, new_preview_height,
new_preview_format) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setPreviewSize(width(%d), height(%d), format(%d))",
__func__, new_preview_width, new_preview_height, new_preview_format);
ret = UNKNOWN_ERROR;
} else {
if (mPreviewWindow) {
if (mPreviewRunning && !mPreviewStartDeferred) {
ALOGE("ERR(%s): preview is running, cannot change size and format!",
__func__);
ret = INVALID_OPERATION;
}
ALOGV("%s: mPreviewWindow (%p) set_buffers_geometry", __func__, mPreviewWindow);
ALOGV("%s: mPreviewWindow->set_buffers_geometry (%p)", __func__,
mPreviewWindow->set_buffers_geometry);
mPreviewWindow->set_buffers_geometry(mPreviewWindow,
new_preview_width, new_preview_height,
new_preview_format);
ALOGV("%s: DONE mPreviewWindow (%p) set_buffers_geometry", __func__, mPreviewWindow);
}
mParameters.setPreviewSize(new_preview_width, new_preview_height);
mParameters.setPreviewFormat(new_str_preview_format);
}
}
else ALOGV("%s: preview size and format has not changed", __func__);
} else {
ALOGE("%s: Invalid preview size(%dx%d)",
__func__, new_preview_width, new_preview_height);
ret = INVALID_OPERATION;
}
int new_picture_width = 0;
int new_picture_height = 0;
params.getPictureSize(&new_picture_width, &new_picture_height);
ALOGV("%s : new_picture_width x new_picture_height = %dx%d", __func__, new_picture_width, new_picture_height);
if (0 < new_picture_width && 0 < new_picture_height) {
ALOGV("%s: setSnapshotSize", __func__);
if (mSecCamera->setSnapshotSize(new_picture_width, new_picture_height) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setSnapshotSize(width(%d), height(%d))",
__func__, new_picture_width, new_picture_height);
ret = UNKNOWN_ERROR;
} else {
mParameters.setPictureSize(new_picture_width, new_picture_height);
}
}
// picture format
const char *new_str_picture_format = params.getPictureFormat();
ALOGV("%s : new_str_picture_format %s", __func__, new_str_picture_format);
if (new_str_picture_format != NULL) {
int new_picture_format = 0;
if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_RGB565))
new_picture_format = V4L2_PIX_FMT_RGB565;
else if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_RGBA8888))
new_picture_format = V4L2_PIX_FMT_RGB32;
else if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_YUV420SP))
new_picture_format = V4L2_PIX_FMT_NV21;
else if (!strcmp(new_str_picture_format, "yuv420sp_custom"))
new_picture_format = V4L2_PIX_FMT_NV12T;
else if (!strcmp(new_str_picture_format, "yuv420p"))
new_picture_format = V4L2_PIX_FMT_YUV420;
else if (!strcmp(new_str_picture_format, "yuv422i"))
new_picture_format = V4L2_PIX_FMT_YUYV;
else if (!strcmp(new_str_picture_format, "uyv422i_custom")) //Zero copy UYVY format
new_picture_format = V4L2_PIX_FMT_UYVY;
else if (!strcmp(new_str_picture_format, "uyv422i")) //Non-zero copy UYVY format
new_picture_format = V4L2_PIX_FMT_UYVY;
else if (!strcmp(new_str_picture_format, CameraParameters::PIXEL_FORMAT_JPEG))
new_picture_format = V4L2_PIX_FMT_YUYV;
else if (!strcmp(new_str_picture_format, "yuv422p"))
new_picture_format = V4L2_PIX_FMT_YUV422P;
else
new_picture_format = V4L2_PIX_FMT_NV21; //for 3rd party
if (mSecCamera->setSnapshotPixelFormat(new_picture_format) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setSnapshotPixelFormat(format(%d))", __func__, new_picture_format);
ret = UNKNOWN_ERROR;
} else {
mParameters.setPictureFormat(new_str_picture_format);
}
}
//JPEG image quality
int new_jpeg_quality = params.getInt(CameraParameters::KEY_JPEG_QUALITY);
ALOGV("%s : new_jpeg_quality %d", __func__, new_jpeg_quality);
/* we ignore bad values */
if (new_jpeg_quality >=1 && new_jpeg_quality <= 100) {
if (mSecCamera->setJpegQuality(new_jpeg_quality) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setJpegQuality(quality(%d))", __func__, new_jpeg_quality);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_JPEG_QUALITY, new_jpeg_quality);
}
}
// JPEG thumbnail size
int new_jpeg_thumbnail_width = params.getInt(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH);
int new_jpeg_thumbnail_height= params.getInt(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT);
if (0 <= new_jpeg_thumbnail_width && 0 <= new_jpeg_thumbnail_height) {
if (mSecCamera->setJpegThumbnailSize(new_jpeg_thumbnail_width, new_jpeg_thumbnail_height) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setJpegThumbnailSize(width(%d), height(%d))", __func__, new_jpeg_thumbnail_width, new_jpeg_thumbnail_height);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_JPEG_THUMBNAIL_WIDTH, new_jpeg_thumbnail_width);
mParameters.set(CameraParameters::KEY_JPEG_THUMBNAIL_HEIGHT, new_jpeg_thumbnail_height);
}
}
// frame rate
int new_frame_rate = params.getPreviewFrameRate();
/* ignore any fps request, we're determine fps automatically based
* on scene mode. don't return an error because it causes CTS failure.
*/
if (new_frame_rate != mParameters.getPreviewFrameRate()) {
ALOGW("WARN(%s): request for preview frame %d not allowed, != %d\n",
__func__, new_frame_rate, mParameters.getPreviewFrameRate());
}
// rotation
int new_rotation = params.getInt(CameraParameters::KEY_ROTATION);
ALOGV("%s : new_rotation %d", __func__, new_rotation);
if (0 <= new_rotation) {
ALOGV("%s : set orientation:%d\n", __func__, new_rotation);
if (mSecCamera->setExifOrientationInfo(new_rotation) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setExifOrientationInfo(%d)", __func__, new_rotation);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_ROTATION, new_rotation);
}
}
// brightness
int new_exposure_compensation = params.getInt(CameraParameters::KEY_EXPOSURE_COMPENSATION);
int max_exposure_compensation = params.getInt(CameraParameters::KEY_MAX_EXPOSURE_COMPENSATION);
int min_exposure_compensation = params.getInt(CameraParameters::KEY_MIN_EXPOSURE_COMPENSATION);
ALOGV("%s : new_exposure_compensation %d", __func__, new_exposure_compensation);
if ((min_exposure_compensation <= new_exposure_compensation) &&
(max_exposure_compensation >= new_exposure_compensation)) {
if (mSecCamera->setBrightness(new_exposure_compensation) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setBrightness(brightness(%d))", __func__, new_exposure_compensation);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_EXPOSURE_COMPENSATION, new_exposure_compensation);
}
}
// whitebalance
const char *new_white_str = params.get(CameraParameters::KEY_WHITE_BALANCE);
ALOGV("%s : new_white_str %s", __func__, new_white_str);
if (new_white_str != NULL) {
int new_white = -1;
if (!strcmp(new_white_str, CameraParameters::WHITE_BALANCE_AUTO))
new_white = WHITE_BALANCE_AUTO;
else if (!strcmp(new_white_str,
CameraParameters::WHITE_BALANCE_DAYLIGHT))
new_white = WHITE_BALANCE_SUNNY;
else if (!strcmp(new_white_str,
CameraParameters::WHITE_BALANCE_CLOUDY_DAYLIGHT))
new_white = WHITE_BALANCE_CLOUDY;
else if (!strcmp(new_white_str,
CameraParameters::WHITE_BALANCE_FLUORESCENT))
new_white = WHITE_BALANCE_FLUORESCENT;
else if (!strcmp(new_white_str,
CameraParameters::WHITE_BALANCE_INCANDESCENT))
new_white = WHITE_BALANCE_TUNGSTEN;
else {
ALOGE("ERR(%s):Invalid white balance(%s)", __func__, new_white_str); //twilight, shade, warm_flourescent
ret = UNKNOWN_ERROR;
}
if (0 <= new_white) {
if (mSecCamera->setWhiteBalance(new_white) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setWhiteBalance(white(%d))", __func__, new_white);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_WHITE_BALANCE, new_white_str);
}
}
}
// scene mode
const char *new_scene_mode_str = params.get(CameraParameters::KEY_SCENE_MODE);
const char *current_scene_mode_str = mParameters.get(CameraParameters::KEY_SCENE_MODE);
// fps range
int new_min_fps = 0;
int new_max_fps = 0;
int current_min_fps, current_max_fps;
params.getPreviewFpsRange(&new_min_fps, &new_max_fps);
mParameters.getPreviewFpsRange(¤t_min_fps, ¤t_max_fps);
/* our fps range is determined by the sensor, reject any request
* that isn't exactly what we're already at.
* but the check is performed when requesting only changing fps range
*/
if (new_scene_mode_str && current_scene_mode_str) {
if (!strcmp(new_scene_mode_str, current_scene_mode_str)) {
if ((new_min_fps != current_min_fps) || (new_max_fps != current_max_fps)) {
ALOGW("%s : requested new_min_fps = %d, new_max_fps = %d not allowed",
__func__, new_min_fps, new_max_fps);
ALOGE("%s : current_min_fps = %d, current_max_fps = %d",
__func__, current_min_fps, current_max_fps);
ret = UNKNOWN_ERROR;
}
}
} else {
/* Check basic validation if scene mode is different */
if ((new_min_fps > new_max_fps) ||
(new_min_fps < 0) || (new_max_fps < 0))
ret = UNKNOWN_ERROR;
}
const char *new_focus_mode_str = params.get(CameraParameters::KEY_FOCUS_MODE);
if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) {
int new_scene_mode = -1;
const char *new_flash_mode_str = params.get(CameraParameters::KEY_FLASH_MODE);
// fps range is (15000,30000) by default.
mParameters.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(15000,30000)");
mParameters.set(CameraParameters::KEY_PREVIEW_FPS_RANGE,
"15000,30000");
if (!strcmp(new_scene_mode_str, CameraParameters::SCENE_MODE_AUTO)) {
new_scene_mode = SCENE_MODE_NONE;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "on,off,auto,torch");
} else {
// defaults for non-auto scene modes
if (mSecCamera->getCameraId() == SecCamera::CAMERA_ID_BACK) {
new_focus_mode_str = CameraParameters::FOCUS_MODE_AUTO;
}
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_PORTRAIT)) {
new_scene_mode = SCENE_MODE_PORTRAIT;
new_flash_mode_str = CameraParameters::FLASH_MODE_AUTO;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "auto");
} else if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_LANDSCAPE)) {
new_scene_mode = SCENE_MODE_LANDSCAPE;
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off");
} else if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_SPORTS)) {
new_scene_mode = SCENE_MODE_SPORTS;
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off");
} else if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_PARTY)) {
new_scene_mode = SCENE_MODE_PARTY_INDOOR;
new_flash_mode_str = CameraParameters::FLASH_MODE_AUTO;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "auto");
} else if ((!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_BEACH)) ||
(!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_SNOW))) {
new_scene_mode = SCENE_MODE_BEACH_SNOW;
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off");
} else if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_SUNSET)) {
new_scene_mode = SCENE_MODE_SUNSET;
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off");
} else if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_NIGHT)) {
new_scene_mode = SCENE_MODE_NIGHTSHOT;
mParameters.set(CameraParameters::KEY_SUPPORTED_PREVIEW_FPS_RANGE, "(4000,30000)");
mParameters.set(CameraParameters::KEY_PREVIEW_FPS_RANGE,
"4000,30000");
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off");
} else if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_FIREWORKS)) {
new_scene_mode = SCENE_MODE_FIREWORKS;
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off");
} else if (!strcmp(new_scene_mode_str,
CameraParameters::SCENE_MODE_CANDLELIGHT)) {
new_scene_mode = SCENE_MODE_CANDLE_LIGHT;
new_flash_mode_str = CameraParameters::FLASH_MODE_OFF;
mParameters.set(CameraParameters::KEY_SUPPORTED_FLASH_MODES, "off");
} else {
ALOGE("%s::unmatched scene_mode(%s)",
__func__, new_scene_mode_str); //action, night-portrait, theatre, steadyphoto
ret = UNKNOWN_ERROR;
}
}
// focus mode
if (new_focus_mode_str != NULL) {
int new_focus_mode = -1;
if (!strcmp(new_focus_mode_str,
CameraParameters::FOCUS_MODE_AUTO)) {
new_focus_mode = FOCUS_MODE_AUTO;
mParameters.set(CameraParameters::KEY_FOCUS_DISTANCES,
BACK_CAMERA_AUTO_FOCUS_DISTANCES_STR);
}
else if (!strcmp(new_focus_mode_str,
CameraParameters::FOCUS_MODE_MACRO)) {
new_focus_mode = FOCUS_MODE_MACRO;
mParameters.set(CameraParameters::KEY_FOCUS_DISTANCES,
BACK_CAMERA_MACRO_FOCUS_DISTANCES_STR);
}
else if (!strcmp(new_focus_mode_str,
CameraParameters::FOCUS_MODE_INFINITY)) {
new_focus_mode = FOCUS_MODE_INFINITY;
mParameters.set(CameraParameters::KEY_FOCUS_DISTANCES,
BACK_CAMERA_INFINITY_FOCUS_DISTANCES_STR);
}
else {
ALOGE("%s::unmatched focus_mode(%s)", __func__, new_focus_mode_str);
ret = UNKNOWN_ERROR;
}
if (0 <= new_focus_mode) {
if (mSecCamera->setFocusMode(new_focus_mode) < 0) {
ALOGE("%s::mSecCamera->setFocusMode(%d) fail", __func__, new_focus_mode);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_FOCUS_MODE, new_focus_mode_str);
}
}
}
// flash..
if (new_flash_mode_str != NULL) {
int new_flash_mode = -1;
if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_OFF))
new_flash_mode = FLASH_MODE_OFF;
else if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_AUTO))
new_flash_mode = FLASH_MODE_AUTO;
else if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_ON))
new_flash_mode = FLASH_MODE_ON;
else if (!strcmp(new_flash_mode_str, CameraParameters::FLASH_MODE_TORCH))
new_flash_mode = FLASH_MODE_TORCH;
else {
ALOGE("%s::unmatched flash_mode(%s)", __func__, new_flash_mode_str); //red-eye
ret = UNKNOWN_ERROR;
}
if (0 <= new_flash_mode) {
if (mSecCamera->setFlashMode(new_flash_mode) < 0) {
ALOGE("%s::mSecCamera->setFlashMode(%d) fail", __func__, new_flash_mode);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_FLASH_MODE, new_flash_mode_str);
}
}
}
// scene..
if (0 <= new_scene_mode) {
if (mSecCamera->setSceneMode(new_scene_mode) < 0) {
ALOGE("%s::mSecCamera->setSceneMode(%d) fail", __func__, new_scene_mode);
ret = UNKNOWN_ERROR;
} else {
mParameters.set(CameraParameters::KEY_SCENE_MODE, new_scene_mode_str);
}
}
} else {
if (!isSupportedParameter(new_focus_mode_str,
mParameters.get(CameraParameters::KEY_SUPPORTED_FOCUS_MODES))) {
ALOGE("%s: Unsupported focus mode: %s", __func__, new_focus_mode_str);
ret = UNKNOWN_ERROR;
}
}
// ---------------------------------------------------------------------------
// image effect
const char *new_image_effect_str = params.get(CameraParameters::KEY_EFFECT);
if (new_image_effect_str != NULL) {
int new_image_effect = -1;
if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_NONE))
new_image_effect = IMAGE_EFFECT_NONE;
else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_MONO))
new_image_effect = IMAGE_EFFECT_BNW;
else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_SEPIA))
new_image_effect = IMAGE_EFFECT_SEPIA;
else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_AQUA))
new_image_effect = IMAGE_EFFECT_AQUA;
else if (!strcmp(new_image_effect_str, CameraParameters::EFFECT_NEGATIVE))
new_image_effect = IMAGE_EFFECT_NEGATIVE;
else {
//posterize, whiteboard, blackboard, solarize
ALOGE("ERR(%s):Invalid effect(%s)", __func__, new_image_effect_str);
ret = UNKNOWN_ERROR;
}
if (new_image_effect >= 0) {
if (mSecCamera->setImageEffect(new_image_effect) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setImageEffect(effect(%d))", __func__, new_image_effect);
ret = UNKNOWN_ERROR;
} else {
const char *old_image_effect_str = mParameters.get(CameraParameters::KEY_EFFECT);
if (old_image_effect_str) {
if (strcmp(old_image_effect_str, new_image_effect_str)) {
setSkipFrame(EFFECT_SKIP_FRAME);
}
}
mParameters.set(CameraParameters::KEY_EFFECT, new_image_effect_str);
}
}
}
//vt mode
int new_vtmode = mInternalParameters.getInt("vtmode");
if (0 <= new_vtmode) {
if (mSecCamera->setVTmode(new_vtmode) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setVTMode(%d)", __func__, new_vtmode);
ret = UNKNOWN_ERROR;
}
}
//contrast
int new_contrast = mInternalParameters.getInt("contrast");
if (0 <= new_contrast) {
if (mSecCamera->setContrast(new_contrast) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setContrast(%d)", __func__, new_contrast);
ret = UNKNOWN_ERROR;
}
}
//WDR
int new_wdr = mInternalParameters.getInt("wdr");
if (0 <= new_wdr) {
if (mSecCamera->setWDR(new_wdr) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setWDR(%d)", __func__, new_wdr);
ret = UNKNOWN_ERROR;
}
}
//anti shake
int new_anti_shake = mInternalParameters.getInt("anti-shake");
if (0 <= new_anti_shake) {
if (mSecCamera->setAntiShake(new_anti_shake) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setWDR(%d)", __func__, new_anti_shake);
ret = UNKNOWN_ERROR;
}
}
// gps latitude
const char *new_gps_latitude_str = params.get(CameraParameters::KEY_GPS_LATITUDE);
if (mSecCamera->setGPSLatitude(new_gps_latitude_str) < 0) {
ALOGE("%s::mSecCamera->setGPSLatitude(%s) fail", __func__, new_gps_latitude_str);
ret = UNKNOWN_ERROR;
} else {
if (new_gps_latitude_str) {
mParameters.set(CameraParameters::KEY_GPS_LATITUDE, new_gps_latitude_str);
} else {
mParameters.remove(CameraParameters::KEY_GPS_LATITUDE);
}
}
// gps longitude
const char *new_gps_longitude_str = params.get(CameraParameters::KEY_GPS_LONGITUDE);
if (mSecCamera->setGPSLongitude(new_gps_longitude_str) < 0) {
ALOGE("%s::mSecCamera->setGPSLongitude(%s) fail", __func__, new_gps_longitude_str);
ret = UNKNOWN_ERROR;
} else {
if (new_gps_longitude_str) {
mParameters.set(CameraParameters::KEY_GPS_LONGITUDE, new_gps_longitude_str);
} else {
mParameters.remove(CameraParameters::KEY_GPS_LONGITUDE);
}
}
// gps altitude
const char *new_gps_altitude_str = params.get(CameraParameters::KEY_GPS_ALTITUDE);
if (mSecCamera->setGPSAltitude(new_gps_altitude_str) < 0) {
ALOGE("%s::mSecCamera->setGPSAltitude(%s) fail", __func__, new_gps_altitude_str);
ret = UNKNOWN_ERROR;
} else {
if (new_gps_altitude_str) {
mParameters.set(CameraParameters::KEY_GPS_ALTITUDE, new_gps_altitude_str);
} else {
mParameters.remove(CameraParameters::KEY_GPS_ALTITUDE);
}
}
// gps timestamp
const char *new_gps_timestamp_str = params.get(CameraParameters::KEY_GPS_TIMESTAMP);
if (mSecCamera->setGPSTimeStamp(new_gps_timestamp_str) < 0) {
ALOGE("%s::mSecCamera->setGPSTimeStamp(%s) fail", __func__, new_gps_timestamp_str);
ret = UNKNOWN_ERROR;
} else {
if (new_gps_timestamp_str) {
mParameters.set(CameraParameters::KEY_GPS_TIMESTAMP, new_gps_timestamp_str);
} else {
mParameters.remove(CameraParameters::KEY_GPS_TIMESTAMP);
}
}
// gps processing method
const char *new_gps_processing_method_str = params.get(CameraParameters::KEY_GPS_PROCESSING_METHOD);
if (mSecCamera->setGPSProcessingMethod(new_gps_processing_method_str) < 0) {
ALOGE("%s::mSecCamera->setGPSProcessingMethod(%s) fail", __func__, new_gps_processing_method_str);
ret = UNKNOWN_ERROR;
} else {
if (new_gps_processing_method_str) {
mParameters.set(CameraParameters::KEY_GPS_PROCESSING_METHOD, new_gps_processing_method_str);
} else {
mParameters.remove(CameraParameters::KEY_GPS_PROCESSING_METHOD);
}
}
// Recording size
int new_recording_width = mInternalParameters.getInt("recording-size-width");
int new_recording_height= mInternalParameters.getInt("recording-size-height");
if (0 < new_recording_width && 0 < new_recording_height) {
if (mSecCamera->setRecordingSize(new_recording_width, new_recording_height) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setRecordingSize(width(%d), height(%d))", __func__, new_recording_width, new_recording_height);
ret = UNKNOWN_ERROR;
}
} else {
if (mSecCamera->setRecordingSize(new_preview_width, new_preview_height) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setRecordingSize(width(%d), height(%d))", __func__, new_preview_width, new_preview_height);
ret = UNKNOWN_ERROR;
}
}
//gamma
const char *new_gamma_str = mInternalParameters.get("video_recording_gamma");
if (new_gamma_str != NULL) {
int new_gamma = -1;
if (!strcmp(new_gamma_str, "off"))
new_gamma = GAMMA_OFF;
else if (!strcmp(new_gamma_str, "on"))
new_gamma = GAMMA_ON;
else {
ALOGE("%s::unmatched gamma(%s)", __func__, new_gamma_str);
ret = UNKNOWN_ERROR;
}
if (0 <= new_gamma) {
if (mSecCamera->setGamma(new_gamma) < 0) {
ALOGE("%s::mSecCamera->setGamma(%d) fail", __func__, new_gamma);
ret = UNKNOWN_ERROR;
}
}
}
//slow ae
const char *new_slow_ae_str = mInternalParameters.get("slow_ae");
if (new_slow_ae_str != NULL) {
int new_slow_ae = -1;
if (!strcmp(new_slow_ae_str, "off"))
new_slow_ae = SLOW_AE_OFF;
else if (!strcmp(new_slow_ae_str, "on"))
new_slow_ae = SLOW_AE_ON;
else {
ALOGE("%s::unmatched slow_ae(%s)", __func__, new_slow_ae_str);
ret = UNKNOWN_ERROR;
}
if (0 <= new_slow_ae) {
if (mSecCamera->setSlowAE(new_slow_ae) < 0) {
ALOGE("%s::mSecCamera->setSlowAE(%d) fail", __func__, new_slow_ae);
ret = UNKNOWN_ERROR;
}
}
}
/*Camcorder fix fps*/
int new_sensor_mode = mInternalParameters.getInt("cam_mode");
if (0 <= new_sensor_mode) {
if (mSecCamera->setSensorMode(new_sensor_mode) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setSensorMode(%d)", __func__, new_sensor_mode);
ret = UNKNOWN_ERROR;
}
} else {
new_sensor_mode=0;
}
/*Shot mode*/
int new_shot_mode = mInternalParameters.getInt("shot_mode");
if (0 <= new_shot_mode) {
if (mSecCamera->setShotMode(new_shot_mode) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setShotMode(%d)", __func__, new_shot_mode);
ret = UNKNOWN_ERROR;
}
} else {
new_shot_mode=0;
}
//blur for Video call
int new_blur_level = mInternalParameters.getInt("blur");
if (0 <= new_blur_level) {
if (mSecCamera->setBlur(new_blur_level) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setBlur(%d)", __func__, new_blur_level);
ret = UNKNOWN_ERROR;
}
}
// chk_dataline
int new_dataline = mInternalParameters.getInt("chk_dataline");
if (0 <= new_dataline) {
if (mSecCamera->setDataLineCheck(new_dataline) < 0) {
ALOGE("ERR(%s):Fail on mSecCamera->setDataLineCheck(%d)", __func__, new_dataline);
ret = UNKNOWN_ERROR;
}
}
ALOGV("%s return ret = %d", __func__, ret);
return ret;
}
CameraParameters CameraHardwareSec::getParameters() const
{
ALOGV("%s :", __func__);
return mParameters;
}
status_t CameraHardwareSec::sendCommand(int32_t command, int32_t arg1, int32_t arg2)
{
return BAD_VALUE;
}
void CameraHardwareSec::release()
{
ALOGV("%s", __func__);
/* shut down any threads we have that might be running. do it here
* instead of the destructor. we're guaranteed to be on another thread
* than the ones below. if we used the destructor, since the threads
* have a reference to this object, we could wind up trying to wait
* for ourself to exit, which is a deadlock.
*/
if (mPreviewThread != NULL) {
/* this thread is normally already in it's threadLoop but blocked
* on the condition variable or running. signal it so it wakes
* up and can exit.
*/
mPreviewThread->requestExit();
mExitPreviewThread = true;
mPreviewRunning = true; /* let it run so it can exit */
mPreviewCondition.signal();
mPreviewThread->requestExitAndWait();
mPreviewThread.clear();
}
if (mAutoFocusThread != NULL) {
/* this thread is normally already in it's threadLoop but blocked
* on the condition variable. signal it so it wakes up and can exit.
*/
mFocusLock.lock();
mAutoFocusThread->requestExit();
mExitAutoFocusThread = true;
mFocusCondition.signal();
mFocusLock.unlock();
mAutoFocusThread->requestExitAndWait();
mAutoFocusThread.clear();
}
if (mPictureThread != NULL) {
mPictureThread->requestExitAndWait();
mPictureThread.clear();
}
if (mRawHeap) {
mRawHeap->release(mRawHeap);
mRawHeap = 0;
}
if (mPreviewHeap) {
mPreviewHeap->release(mPreviewHeap);
mPreviewHeap = 0;
}
if (mRecordHeap) {
mRecordHeap->release(mRecordHeap);
mRecordHeap = 0;
}
/* close after all the heaps are cleared since those
* could have dup'd our file descriptor.
*/
mSecCamera->DeinitCamera();
}
status_t CameraHardwareSec::storeMetaDataInBuffers(bool enable)
{
// FIXME:
// metadata buffer mode can be turned on or off.
// Samsung needs to fix this.
if (!enable) {
ALOGE("Non-metadata buffer mode is not supported!");
return INVALID_OPERATION;
}
return OK;
}
static CameraInfo sCameraInfo[] = {
{
CAMERA_FACING_BACK,
90, /* orientation */
},
{
CAMERA_FACING_FRONT,
270, /* orientation */
}
};
/** Close this device */
static camera_device_t *g_cam_device;
static int HAL_camera_device_close(struct hw_device_t* device)
{
ALOGI("%s", __func__);
if (device) {
camera_device_t *cam_device = (camera_device_t *)device;
delete static_cast<CameraHardwareSec *>(cam_device->priv);
free(cam_device);
g_cam_device = 0;
}
return 0;
}
static inline CameraHardwareSec *obj(struct camera_device *dev)
{
return reinterpret_cast<CameraHardwareSec *>(dev->priv);
}
/** Set the preview_stream_ops to which preview frames are sent */
static int HAL_camera_device_set_preview_window(struct camera_device *dev,
struct preview_stream_ops *buf)
{
ALOGV("%s", __func__);
return obj(dev)->setPreviewWindow(buf);
}
/** Set the notification and data callbacks */
static void HAL_camera_device_set_callbacks(struct camera_device *dev,
camera_notify_callback notify_cb,
camera_data_callback data_cb,
camera_data_timestamp_callback data_cb_timestamp,
camera_request_memory get_memory,
void* user)
{
ALOGV("%s", __func__);
obj(dev)->setCallbacks(notify_cb, data_cb, data_cb_timestamp,
get_memory,
user);
}
/**
* The following three functions all take a msg_type, which is a bitmask of
* the messages defined in include/ui/Camera.h
*/
/**
* Enable a message, or set of messages.
*/
static void HAL_camera_device_enable_msg_type(struct camera_device *dev, int32_t msg_type)
{
ALOGV("%s", __func__);
obj(dev)->enableMsgType(msg_type);
}
/**
* Disable a message, or a set of messages.
*
* Once received a call to disableMsgType(CAMERA_MSG_VIDEO_FRAME), camera
* HAL should not rely on its client to call releaseRecordingFrame() to
* release video recording frames sent out by the cameral HAL before and
* after the disableMsgType(CAMERA_MSG_VIDEO_FRAME) call. Camera HAL
* clients must not modify/access any video recording frame after calling
* disableMsgType(CAMERA_MSG_VIDEO_FRAME).
*/
static void HAL_camera_device_disable_msg_type(struct camera_device *dev, int32_t msg_type)
{
ALOGV("%s", __func__);
obj(dev)->disableMsgType(msg_type);
}
/**
* Query whether a message, or a set of messages, is enabled. Note that
* this is operates as an AND, if any of the messages queried are off, this
* will return false.
*/
static int HAL_camera_device_msg_type_enabled(struct camera_device *dev, int32_t msg_type)
{
ALOGV("%s", __func__);
return obj(dev)->msgTypeEnabled(msg_type);
}
/**
* Start preview mode.
*/
static int HAL_camera_device_start_preview(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->startPreview();
}
/**
* Stop a previously started preview.
*/
static void HAL_camera_device_stop_preview(struct camera_device *dev)
{
ALOGV("%s", __func__);
obj(dev)->stopPreview();
}
/**
* Returns true if preview is enabled.
*/
static int HAL_camera_device_preview_enabled(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->previewEnabled();
}
/**
* Request the camera HAL to store meta data or real YUV data in the video
* buffers sent out via CAMERA_MSG_VIDEO_FRAME for a recording session. If
* it is not called, the default camera HAL behavior is to store real YUV
* data in the video buffers.
*
* This method should be called before startRecording() in order to be
* effective.
*
* If meta data is stored in the video buffers, it is up to the receiver of
* the video buffers to interpret the contents and to find the actual frame
* data with the help of the meta data in the buffer. How this is done is
* outside of the scope of this method.
*
* Some camera HALs may not support storing meta data in the video buffers,
* but all camera HALs should support storing real YUV data in the video
* buffers. If the camera HAL does not support storing the meta data in the
* video buffers when it is requested to do do, INVALID_OPERATION must be
* returned. It is very useful for the camera HAL to pass meta data rather
* than the actual frame data directly to the video encoder, since the
* amount of the uncompressed frame data can be very large if video size is
* large.
*
* @param enable if true to instruct the camera HAL to store
* meta data in the video buffers; false to instruct
* the camera HAL to store real YUV data in the video
* buffers.
*
* @return OK on success.
*/
static int HAL_camera_device_store_meta_data_in_buffers(struct camera_device *dev, int enable)
{
ALOGV("%s", __func__);
return obj(dev)->storeMetaDataInBuffers(enable);
}
/**
* Start record mode. When a record image is available, a
* CAMERA_MSG_VIDEO_FRAME message is sent with the corresponding
* frame. Every record frame must be released by a camera HAL client via
* releaseRecordingFrame() before the client calls
* disableMsgType(CAMERA_MSG_VIDEO_FRAME). After the client calls
* disableMsgType(CAMERA_MSG_VIDEO_FRAME), it is the camera HAL's
* responsibility to manage the life-cycle of the video recording frames,
* and the client must not modify/access any video recording frames.
*/
static int HAL_camera_device_start_recording(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->startRecording();
}
/**
* Stop a previously started recording.
*/
static void HAL_camera_device_stop_recording(struct camera_device *dev)
{
ALOGV("%s", __func__);
obj(dev)->stopRecording();
}
/**
* Returns true if recording is enabled.
*/
static int HAL_camera_device_recording_enabled(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->recordingEnabled();
}
/**
* Release a record frame previously returned by CAMERA_MSG_VIDEO_FRAME.
*
* It is camera HAL client's responsibility to release video recording
* frames sent out by the camera HAL before the camera HAL receives a call
* to disableMsgType(CAMERA_MSG_VIDEO_FRAME). After it receives the call to
* disableMsgType(CAMERA_MSG_VIDEO_FRAME), it is the camera HAL's
* responsibility to manage the life-cycle of the video recording frames.
*/
static void HAL_camera_device_release_recording_frame(struct camera_device *dev,
const void *opaque)
{
ALOGV("%s", __func__);
obj(dev)->releaseRecordingFrame(opaque);
}
/**
* Start auto focus, the notification callback routine is called with
* CAMERA_MSG_FOCUS once when focusing is complete. autoFocus() will be
* called again if another auto focus is needed.
*/
static int HAL_camera_device_auto_focus(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->autoFocus();
}
/**
* Cancels auto-focus function. If the auto-focus is still in progress,
* this function will cancel it. Whether the auto-focus is in progress or
* not, this function will return the focus position to the default. If
* the camera does not support auto-focus, this is a no-op.
*/
static int HAL_camera_device_cancel_auto_focus(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->cancelAutoFocus();
}
/**
* Take a picture.
*/
static int HAL_camera_device_take_picture(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->takePicture();
}
/**
* Cancel a picture that was started with takePicture. Calling this method
* when no picture is being taken is a no-op.
*/
static int HAL_camera_device_cancel_picture(struct camera_device *dev)
{
ALOGV("%s", __func__);
return obj(dev)->cancelPicture();
}
/**
* Set the camera parameters. This returns BAD_VALUE if any parameter is
* invalid or not supported.
*/
static int HAL_camera_device_set_parameters(struct camera_device *dev,
const char *parms)
{
ALOGV("%s", __func__);
String8 str(parms);
CameraParameters p(str);
return obj(dev)->setParameters(p);
}
/** Return the camera parameters. */
char *HAL_camera_device_get_parameters(struct camera_device *dev)
{
ALOGV("%s", __func__);
String8 str;
CameraParameters parms = obj(dev)->getParameters();
str = parms.flatten();
return strdup(str.string());
}
void HAL_camera_device_put_parameters(struct camera_device *dev, char *parms)
{
ALOGV("%s", __func__);
free(parms);
}
/**
* Send command to camera driver.
*/
static int HAL_camera_device_send_command(struct camera_device *dev,
int32_t cmd, int32_t arg1, int32_t arg2)
{
ALOGV("%s", __func__);
return obj(dev)->sendCommand(cmd, arg1, arg2);
}
/**
* Release the hardware resources owned by this object. Note that this is
* *not* done in the destructor.
*/
static void HAL_camera_device_release(struct camera_device *dev)
{
ALOGV("%s", __func__);
obj(dev)->release();
}
/**
* Dump state of the camera hardware
*/
static int HAL_camera_device_dump(struct camera_device *dev, int fd)
{
ALOGV("%s", __func__);
return obj(dev)->dump(fd);
}
static int HAL_getNumberOfCameras()
{
ALOGV("%s", __func__);
return sizeof(sCameraInfo) / sizeof(sCameraInfo[0]);
}
static int HAL_getCameraInfo(int cameraId, struct camera_info *cameraInfo)
{
ALOGV("%s", __func__);
memcpy(cameraInfo, &sCameraInfo[cameraId], sizeof(CameraInfo));
return 0;
}
#define SET_METHOD(m) m : HAL_camera_device_##m
static camera_device_ops_t camera_device_ops = {
SET_METHOD(set_preview_window),
SET_METHOD(set_callbacks),
SET_METHOD(enable_msg_type),
SET_METHOD(disable_msg_type),
SET_METHOD(msg_type_enabled),
SET_METHOD(start_preview),
SET_METHOD(stop_preview),
SET_METHOD(preview_enabled),
SET_METHOD(store_meta_data_in_buffers),
SET_METHOD(start_recording),
SET_METHOD(stop_recording),
SET_METHOD(recording_enabled),
SET_METHOD(release_recording_frame),
SET_METHOD(auto_focus),
SET_METHOD(cancel_auto_focus),
SET_METHOD(take_picture),
SET_METHOD(cancel_picture),
SET_METHOD(set_parameters),
SET_METHOD(get_parameters),
SET_METHOD(put_parameters),
SET_METHOD(send_command),
SET_METHOD(release),
SET_METHOD(dump),
};
#undef SET_METHOD
static int HAL_camera_device_open(const struct hw_module_t* module,
const char *id,
struct hw_device_t** device)
{
ALOGV("%s", __func__);
int cameraId = atoi(id);
if (cameraId < 0 || cameraId >= HAL_getNumberOfCameras()) {
ALOGE("Invalid camera ID %s", id);
return -EINVAL;
}
if (g_cam_device) {
if (obj(g_cam_device)->getCameraId() == cameraId) {
ALOGV("returning existing camera ID %s", id);
goto done;
} else {
ALOGE("Cannot open camera %d. camera %d is already running!",
cameraId, obj(g_cam_device)->getCameraId());
return -ENOSYS;
}
}
g_cam_device = (camera_device_t *)malloc(sizeof(camera_device_t));
if (!g_cam_device)
return -ENOMEM;
g_cam_device->common.tag = HARDWARE_DEVICE_TAG;
g_cam_device->common.version = 1;
g_cam_device->common.module = const_cast<hw_module_t *>(module);
g_cam_device->common.close = HAL_camera_device_close;
g_cam_device->ops = &camera_device_ops;
ALOGI("%s: open camera %s", __func__, id);
g_cam_device->priv = new CameraHardwareSec(cameraId, g_cam_device);
done:
*device = (hw_device_t *)g_cam_device;
ALOGI("%s: opened camera %s (%p)", __func__, id, *device);
return 0;
}
static hw_module_methods_t camera_module_methods = {
open : HAL_camera_device_open
};
extern "C" {
struct camera_module HAL_MODULE_INFO_SYM = {
common : {
tag : HARDWARE_MODULE_TAG,
version_major : 1,
version_minor : 0,
id : CAMERA_HARDWARE_MODULE_ID,
name : "Crespo camera HAL",
author : "Samsung Corporation",
methods : &camera_module_methods,
},
get_number_of_cameras : HAL_getNumberOfCameras,
get_camera_info : HAL_getCameraInfo
};
}
}; // namespace android