| /* |
| * Copyright (C) 2012 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define LOG_TAG "Camera2-CaptureSequencer" |
| #define ATRACE_TAG ATRACE_TAG_CAMERA |
| //#define LOG_NDEBUG 0 |
| |
| #include <inttypes.h> |
| |
| #include <utils/Log.h> |
| #include <utils/Trace.h> |
| #include <utils/Vector.h> |
| |
| #include "api1/Camera2Client.h" |
| #include "api1/client2/CaptureSequencer.h" |
| #include "api1/client2/BurstCapture.h" |
| #include "api1/client2/Parameters.h" |
| #include "api1/client2/ZslProcessorInterface.h" |
| |
| namespace android { |
| namespace camera2 { |
| |
| /** Public members */ |
| |
| CaptureSequencer::CaptureSequencer(wp<Camera2Client> client): |
| Thread(false), |
| mStartCapture(false), |
| mBusy(false), |
| mNewAEState(false), |
| mNewFrameReceived(false), |
| mNewCaptureReceived(false), |
| mShutterNotified(false), |
| mClient(client), |
| mCaptureState(IDLE), |
| mStateTransitionCount(0), |
| mTriggerId(0), |
| mTimeoutCount(0), |
| mCaptureId(Camera2Client::kCaptureRequestIdStart), |
| mMsgType(0) { |
| ALOGV("%s", __FUNCTION__); |
| } |
| |
| CaptureSequencer::~CaptureSequencer() { |
| ALOGV("%s: Exit", __FUNCTION__); |
| } |
| |
| void CaptureSequencer::setZslProcessor(wp<ZslProcessorInterface> processor) { |
| Mutex::Autolock l(mInputMutex); |
| mZslProcessor = processor; |
| } |
| |
| status_t CaptureSequencer::startCapture(int msgType) { |
| ALOGV("%s", __FUNCTION__); |
| ATRACE_CALL(); |
| Mutex::Autolock l(mInputMutex); |
| if (mBusy) { |
| ALOGE("%s: Already busy capturing!", __FUNCTION__); |
| return INVALID_OPERATION; |
| } |
| if (!mStartCapture) { |
| mMsgType = msgType; |
| mStartCapture = true; |
| mStartCaptureSignal.signal(); |
| } |
| return OK; |
| } |
| |
| status_t CaptureSequencer::waitUntilIdle(nsecs_t timeout) { |
| ATRACE_CALL(); |
| ALOGV("%s: Waiting for idle", __FUNCTION__); |
| Mutex::Autolock l(mStateMutex); |
| status_t res = -1; |
| while (mCaptureState != IDLE) { |
| nsecs_t startTime = systemTime(); |
| |
| res = mStateChanged.waitRelative(mStateMutex, timeout); |
| if (res != OK) return res; |
| |
| timeout -= (systemTime() - startTime); |
| } |
| ALOGV("%s: Now idle", __FUNCTION__); |
| return OK; |
| } |
| |
| void CaptureSequencer::notifyAutoExposure(uint8_t newState, int triggerId) { |
| ATRACE_CALL(); |
| Mutex::Autolock l(mInputMutex); |
| mAEState = newState; |
| mAETriggerId = triggerId; |
| if (!mNewAEState) { |
| mNewAEState = true; |
| mNewNotifySignal.signal(); |
| } |
| } |
| |
| void CaptureSequencer::onResultAvailable(const CaptureResult &result) { |
| ATRACE_CALL(); |
| ALOGV("%s: New result available.", __FUNCTION__); |
| Mutex::Autolock l(mInputMutex); |
| mNewFrameId = result.mResultExtras.requestId; |
| mNewFrame = result.mMetadata; |
| if (!mNewFrameReceived) { |
| mNewFrameReceived = true; |
| mNewFrameSignal.signal(); |
| } |
| } |
| |
| void CaptureSequencer::onCaptureAvailable(nsecs_t timestamp, |
| sp<MemoryBase> captureBuffer) { |
| ATRACE_CALL(); |
| ALOGV("%s", __FUNCTION__); |
| Mutex::Autolock l(mInputMutex); |
| mCaptureTimestamp = timestamp; |
| mCaptureBuffer = captureBuffer; |
| if (!mNewCaptureReceived) { |
| mNewCaptureReceived = true; |
| mNewCaptureSignal.signal(); |
| } |
| } |
| |
| |
| void CaptureSequencer::dump(int fd, const Vector<String16>& /*args*/) { |
| String8 result; |
| if (mCaptureRequest.entryCount() != 0) { |
| result = " Capture request:\n"; |
| write(fd, result.string(), result.size()); |
| mCaptureRequest.dump(fd, 2, 6); |
| } else { |
| result = " Capture request: undefined\n"; |
| write(fd, result.string(), result.size()); |
| } |
| result = String8::format(" Current capture state: %s\n", |
| kStateNames[mCaptureState]); |
| result.append(" Latest captured frame:\n"); |
| write(fd, result.string(), result.size()); |
| mNewFrame.dump(fd, 2, 6); |
| } |
| |
| /** Private members */ |
| |
| const char* CaptureSequencer::kStateNames[CaptureSequencer::NUM_CAPTURE_STATES+1] = |
| { |
| "IDLE", |
| "START", |
| "ZSL_START", |
| "ZSL_WAITING", |
| "ZSL_REPROCESSING", |
| "STANDARD_START", |
| "STANDARD_PRECAPTURE_WAIT", |
| "STANDARD_CAPTURE", |
| "STANDARD_CAPTURE_WAIT", |
| "BURST_CAPTURE_START", |
| "BURST_CAPTURE_WAIT", |
| "DONE", |
| "ERROR", |
| "UNKNOWN" |
| }; |
| |
| const CaptureSequencer::StateManager |
| CaptureSequencer::kStateManagers[CaptureSequencer::NUM_CAPTURE_STATES-1] = { |
| &CaptureSequencer::manageIdle, |
| &CaptureSequencer::manageStart, |
| &CaptureSequencer::manageZslStart, |
| &CaptureSequencer::manageZslWaiting, |
| &CaptureSequencer::manageZslReprocessing, |
| &CaptureSequencer::manageStandardStart, |
| &CaptureSequencer::manageStandardPrecaptureWait, |
| &CaptureSequencer::manageStandardCapture, |
| &CaptureSequencer::manageStandardCaptureWait, |
| &CaptureSequencer::manageBurstCaptureStart, |
| &CaptureSequencer::manageBurstCaptureWait, |
| &CaptureSequencer::manageDone, |
| }; |
| |
| bool CaptureSequencer::threadLoop() { |
| |
| sp<Camera2Client> client = mClient.promote(); |
| if (client == 0) return false; |
| |
| CaptureState currentState; |
| { |
| Mutex::Autolock l(mStateMutex); |
| currentState = mCaptureState; |
| } |
| |
| currentState = (this->*kStateManagers[currentState])(client); |
| |
| Mutex::Autolock l(mStateMutex); |
| if (currentState != mCaptureState) { |
| if (mCaptureState != IDLE) { |
| ATRACE_ASYNC_END(kStateNames[mCaptureState], mStateTransitionCount); |
| } |
| mCaptureState = currentState; |
| mStateTransitionCount++; |
| if (mCaptureState != IDLE) { |
| ATRACE_ASYNC_BEGIN(kStateNames[mCaptureState], mStateTransitionCount); |
| } |
| ALOGV("Camera %d: New capture state %s", |
| client->getCameraId(), kStateNames[mCaptureState]); |
| mStateChanged.signal(); |
| } |
| |
| if (mCaptureState == ERROR) { |
| ALOGE("Camera %d: Stopping capture sequencer due to error", |
| client->getCameraId()); |
| return false; |
| } |
| |
| return true; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageIdle( |
| sp<Camera2Client> &/*client*/) { |
| status_t res; |
| Mutex::Autolock l(mInputMutex); |
| while (!mStartCapture) { |
| res = mStartCaptureSignal.waitRelative(mInputMutex, |
| kWaitDuration); |
| if (res == TIMED_OUT) break; |
| } |
| if (mStartCapture) { |
| mStartCapture = false; |
| mBusy = true; |
| return START; |
| } |
| return IDLE; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageDone(sp<Camera2Client> &client) { |
| status_t res = OK; |
| ATRACE_CALL(); |
| mCaptureId++; |
| if (mCaptureId >= Camera2Client::kCaptureRequestIdEnd) { |
| mCaptureId = Camera2Client::kCaptureRequestIdStart; |
| } |
| { |
| Mutex::Autolock l(mInputMutex); |
| mBusy = false; |
| } |
| |
| int takePictureCounter = 0; |
| { |
| SharedParameters::Lock l(client->getParameters()); |
| switch (l.mParameters.state) { |
| case Parameters::DISCONNECTED: |
| ALOGW("%s: Camera %d: Discarding image data during shutdown ", |
| __FUNCTION__, client->getCameraId()); |
| res = INVALID_OPERATION; |
| break; |
| case Parameters::STILL_CAPTURE: |
| res = client->getCameraDevice()->waitUntilDrained(); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Can't idle after still capture: " |
| "%s (%d)", __FUNCTION__, client->getCameraId(), |
| strerror(-res), res); |
| } |
| l.mParameters.state = Parameters::STOPPED; |
| break; |
| case Parameters::VIDEO_SNAPSHOT: |
| l.mParameters.state = Parameters::RECORD; |
| break; |
| default: |
| ALOGE("%s: Camera %d: Still image produced unexpectedly " |
| "in state %s!", |
| __FUNCTION__, client->getCameraId(), |
| Parameters::getStateName(l.mParameters.state)); |
| res = INVALID_OPERATION; |
| } |
| takePictureCounter = l.mParameters.takePictureCounter; |
| } |
| sp<ZslProcessorInterface> processor = mZslProcessor.promote(); |
| if (processor != 0) { |
| ALOGV("%s: Memory optimization, clearing ZSL queue", |
| __FUNCTION__); |
| processor->clearZslQueue(); |
| } |
| |
| /** |
| * Fire the jpegCallback in Camera#takePicture(..., jpegCallback) |
| */ |
| if (mCaptureBuffer != 0 && res == OK) { |
| ATRACE_ASYNC_END(Camera2Client::kTakepictureLabel, takePictureCounter); |
| |
| Camera2Client::SharedCameraCallbacks::Lock |
| l(client->mSharedCameraCallbacks); |
| ALOGV("%s: Sending still image to client", __FUNCTION__); |
| if (l.mRemoteCallback != 0) { |
| l.mRemoteCallback->dataCallback(CAMERA_MSG_COMPRESSED_IMAGE, |
| mCaptureBuffer, NULL); |
| } else { |
| ALOGV("%s: No client!", __FUNCTION__); |
| } |
| } |
| mCaptureBuffer.clear(); |
| |
| return IDLE; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageStart( |
| sp<Camera2Client> &client) { |
| ALOGV("%s", __FUNCTION__); |
| status_t res; |
| ATRACE_CALL(); |
| SharedParameters::Lock l(client->getParameters()); |
| CaptureState nextState = DONE; |
| |
| res = updateCaptureRequest(l.mParameters, client); |
| if (res != OK ) { |
| ALOGE("%s: Camera %d: Can't update still image capture request: %s (%d)", |
| __FUNCTION__, client->getCameraId(), strerror(-res), res); |
| return DONE; |
| } |
| |
| if(l.mParameters.lightFx != Parameters::LIGHTFX_NONE && |
| l.mParameters.state == Parameters::STILL_CAPTURE) { |
| nextState = BURST_CAPTURE_START; |
| } |
| else if (l.mParameters.zslMode && |
| l.mParameters.state == Parameters::STILL_CAPTURE && |
| l.mParameters.flashMode != Parameters::FLASH_MODE_ON) { |
| nextState = ZSL_START; |
| } else { |
| nextState = STANDARD_START; |
| } |
| mShutterNotified = false; |
| |
| return nextState; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageZslStart( |
| sp<Camera2Client> &client) { |
| ALOGV("%s", __FUNCTION__); |
| status_t res; |
| sp<ZslProcessorInterface> processor = mZslProcessor.promote(); |
| if (processor == 0) { |
| ALOGE("%s: No ZSL queue to use!", __FUNCTION__); |
| return DONE; |
| } |
| |
| // We don't want to get partial results for ZSL capture. |
| client->registerFrameListener(mCaptureId, mCaptureId + 1, |
| this, |
| /*sendPartials*/false); |
| |
| // TODO: Actually select the right thing here. |
| res = processor->pushToReprocess(mCaptureId); |
| if (res != OK) { |
| if (res == NOT_ENOUGH_DATA) { |
| ALOGV("%s: Camera %d: ZSL queue doesn't have good frame, " |
| "falling back to normal capture", __FUNCTION__, |
| client->getCameraId()); |
| } else { |
| ALOGE("%s: Camera %d: Error in ZSL queue: %s (%d)", |
| __FUNCTION__, client->getCameraId(), strerror(-res), res); |
| } |
| return STANDARD_START; |
| } |
| |
| SharedParameters::Lock l(client->getParameters()); |
| /* warning: this also locks a SharedCameraCallbacks */ |
| shutterNotifyLocked(l.mParameters, client, mMsgType); |
| mShutterNotified = true; |
| mTimeoutCount = kMaxTimeoutsForCaptureEnd; |
| return STANDARD_CAPTURE_WAIT; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageZslWaiting( |
| sp<Camera2Client> &/*client*/) { |
| ALOGV("%s", __FUNCTION__); |
| return DONE; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageZslReprocessing( |
| sp<Camera2Client> &/*client*/) { |
| ALOGV("%s", __FUNCTION__); |
| return START; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageStandardStart( |
| sp<Camera2Client> &client) { |
| ATRACE_CALL(); |
| |
| bool isAeConverged = false; |
| // Get the onFrameAvailable callback when the requestID == mCaptureId |
| // We don't want to get partial results for normal capture, as we need |
| // Get ANDROID_SENSOR_TIMESTAMP from the capture result, but partial |
| // result doesn't have to have this metadata available. |
| // TODO: Update to use the HALv3 shutter notification for remove the |
| // need for this listener and make it faster. see bug 12530628. |
| client->registerFrameListener(mCaptureId, mCaptureId + 1, |
| this, |
| /*sendPartials*/false); |
| |
| { |
| Mutex::Autolock l(mInputMutex); |
| isAeConverged = (mAEState == ANDROID_CONTROL_AE_STATE_CONVERGED); |
| } |
| |
| { |
| SharedParameters::Lock l(client->getParameters()); |
| // Skip AE precapture when it is already converged and not in force flash mode. |
| if (l.mParameters.flashMode != Parameters::FLASH_MODE_ON && isAeConverged) { |
| return STANDARD_CAPTURE; |
| } |
| |
| mTriggerId = l.mParameters.precaptureTriggerCounter++; |
| } |
| client->getCameraDevice()->triggerPrecaptureMetering(mTriggerId); |
| |
| mAeInPrecapture = false; |
| mTimeoutCount = kMaxTimeoutsForPrecaptureStart; |
| return STANDARD_PRECAPTURE_WAIT; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageStandardPrecaptureWait( |
| sp<Camera2Client> &/*client*/) { |
| status_t res; |
| ATRACE_CALL(); |
| Mutex::Autolock l(mInputMutex); |
| while (!mNewAEState) { |
| res = mNewNotifySignal.waitRelative(mInputMutex, kWaitDuration); |
| if (res == TIMED_OUT) { |
| mTimeoutCount--; |
| break; |
| } |
| } |
| if (mTimeoutCount <= 0) { |
| ALOGW("Timed out waiting for precapture %s", |
| mAeInPrecapture ? "end" : "start"); |
| return STANDARD_CAPTURE; |
| } |
| if (mNewAEState) { |
| if (!mAeInPrecapture) { |
| // Waiting to see PRECAPTURE state |
| if (mAETriggerId == mTriggerId) { |
| if (mAEState == ANDROID_CONTROL_AE_STATE_PRECAPTURE) { |
| ALOGV("%s: Got precapture start", __FUNCTION__); |
| mAeInPrecapture = true; |
| mTimeoutCount = kMaxTimeoutsForPrecaptureEnd; |
| } else if (mAEState == ANDROID_CONTROL_AE_STATE_CONVERGED || |
| mAEState == ANDROID_CONTROL_AE_STATE_FLASH_REQUIRED) { |
| // It is legal to transit to CONVERGED or FLASH_REQUIRED |
| // directly after a trigger. |
| ALOGV("%s: AE is already in good state, start capture", __FUNCTION__); |
| return STANDARD_CAPTURE; |
| } |
| } |
| } else { |
| // Waiting to see PRECAPTURE state end |
| if (mAETriggerId == mTriggerId && |
| mAEState != ANDROID_CONTROL_AE_STATE_PRECAPTURE) { |
| ALOGV("%s: Got precapture end", __FUNCTION__); |
| return STANDARD_CAPTURE; |
| } |
| } |
| mNewAEState = false; |
| } |
| return STANDARD_PRECAPTURE_WAIT; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageStandardCapture( |
| sp<Camera2Client> &client) { |
| status_t res; |
| ATRACE_CALL(); |
| SharedParameters::Lock l(client->getParameters()); |
| Vector<int32_t> outputStreams; |
| uint8_t captureIntent = static_cast<uint8_t>(ANDROID_CONTROL_CAPTURE_INTENT_STILL_CAPTURE); |
| |
| /** |
| * Set up output streams in the request |
| * - preview |
| * - capture/jpeg |
| * - callback (if preview callbacks enabled) |
| * - recording (if recording enabled) |
| */ |
| outputStreams.push(client->getPreviewStreamId()); |
| outputStreams.push(client->getCaptureStreamId()); |
| |
| if (l.mParameters.previewCallbackFlags & |
| CAMERA_FRAME_CALLBACK_FLAG_ENABLE_MASK) { |
| outputStreams.push(client->getCallbackStreamId()); |
| } |
| |
| if (l.mParameters.state == Parameters::VIDEO_SNAPSHOT) { |
| outputStreams.push(client->getRecordingStreamId()); |
| captureIntent = static_cast<uint8_t>(ANDROID_CONTROL_CAPTURE_INTENT_VIDEO_SNAPSHOT); |
| } |
| |
| res = mCaptureRequest.update(ANDROID_REQUEST_OUTPUT_STREAMS, |
| outputStreams); |
| if (res == OK) { |
| res = mCaptureRequest.update(ANDROID_REQUEST_ID, |
| &mCaptureId, 1); |
| } |
| if (res == OK) { |
| res = mCaptureRequest.update(ANDROID_CONTROL_CAPTURE_INTENT, |
| &captureIntent, 1); |
| } |
| if (res == OK) { |
| res = mCaptureRequest.sort(); |
| } |
| |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to set up still capture request: %s (%d)", |
| __FUNCTION__, client->getCameraId(), strerror(-res), res); |
| return DONE; |
| } |
| |
| // Create a capture copy since CameraDeviceBase#capture takes ownership |
| CameraMetadata captureCopy = mCaptureRequest; |
| if (captureCopy.entryCount() == 0) { |
| ALOGE("%s: Camera %d: Unable to copy capture request for HAL device", |
| __FUNCTION__, client->getCameraId()); |
| return DONE; |
| } |
| |
| /** |
| * Clear the streaming request for still-capture pictures |
| * (as opposed to i.e. video snapshots) |
| */ |
| if (l.mParameters.state == Parameters::STILL_CAPTURE) { |
| // API definition of takePicture() - stop preview before taking pic |
| res = client->stopStream(); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to stop preview for still capture: " |
| "%s (%d)", |
| __FUNCTION__, client->getCameraId(), strerror(-res), res); |
| return DONE; |
| } |
| } |
| // TODO: Capture should be atomic with setStreamingRequest here |
| res = client->getCameraDevice()->capture(captureCopy); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to submit still image capture request: " |
| "%s (%d)", |
| __FUNCTION__, client->getCameraId(), strerror(-res), res); |
| return DONE; |
| } |
| |
| mTimeoutCount = kMaxTimeoutsForCaptureEnd; |
| return STANDARD_CAPTURE_WAIT; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageStandardCaptureWait( |
| sp<Camera2Client> &client) { |
| status_t res; |
| ATRACE_CALL(); |
| Mutex::Autolock l(mInputMutex); |
| |
| // Wait for new metadata result (mNewFrame) |
| while (!mNewFrameReceived) { |
| res = mNewFrameSignal.waitRelative(mInputMutex, kWaitDuration); |
| if (res == TIMED_OUT) { |
| mTimeoutCount--; |
| break; |
| } |
| } |
| |
| // Approximation of the shutter being closed |
| // - TODO: use the hal3 exposure callback in Camera3Device instead |
| if (mNewFrameReceived && !mShutterNotified) { |
| SharedParameters::Lock l(client->getParameters()); |
| /* warning: this also locks a SharedCameraCallbacks */ |
| shutterNotifyLocked(l.mParameters, client, mMsgType); |
| mShutterNotified = true; |
| } |
| |
| // Wait until jpeg was captured by JpegProcessor |
| while (mNewFrameReceived && !mNewCaptureReceived) { |
| res = mNewCaptureSignal.waitRelative(mInputMutex, kWaitDuration); |
| if (res == TIMED_OUT) { |
| mTimeoutCount--; |
| break; |
| } |
| } |
| if (mTimeoutCount <= 0) { |
| ALOGW("Timed out waiting for capture to complete"); |
| return DONE; |
| } |
| if (mNewFrameReceived && mNewCaptureReceived) { |
| if (mNewFrameId != mCaptureId) { |
| ALOGW("Mismatched capture frame IDs: Expected %d, got %d", |
| mCaptureId, mNewFrameId); |
| } |
| camera_metadata_entry_t entry; |
| entry = mNewFrame.find(ANDROID_SENSOR_TIMESTAMP); |
| if (entry.count == 0) { |
| ALOGE("No timestamp field in capture frame!"); |
| } else if (entry.count == 1) { |
| if (entry.data.i64[0] != mCaptureTimestamp) { |
| ALOGW("Mismatched capture timestamps: Metadata frame %" PRId64 "," |
| " captured buffer %" PRId64, |
| entry.data.i64[0], |
| mCaptureTimestamp); |
| } |
| } else { |
| ALOGE("Timestamp metadata is malformed!"); |
| } |
| client->removeFrameListener(mCaptureId, mCaptureId + 1, this); |
| |
| mNewFrameReceived = false; |
| mNewCaptureReceived = false; |
| return DONE; |
| } |
| return STANDARD_CAPTURE_WAIT; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageBurstCaptureStart( |
| sp<Camera2Client> &client) { |
| ALOGV("%s", __FUNCTION__); |
| status_t res; |
| ATRACE_CALL(); |
| |
| // check which burst mode is set, create respective burst object |
| { |
| SharedParameters::Lock l(client->getParameters()); |
| |
| res = updateCaptureRequest(l.mParameters, client); |
| if(res != OK) { |
| return DONE; |
| } |
| |
| // |
| // check for burst mode type in mParameters here |
| // |
| mBurstCapture = new BurstCapture(client, this); |
| } |
| |
| res = mCaptureRequest.update(ANDROID_REQUEST_ID, &mCaptureId, 1); |
| if (res == OK) { |
| res = mCaptureRequest.sort(); |
| } |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to set up still capture request: %s (%d)", |
| __FUNCTION__, client->getCameraId(), strerror(-res), res); |
| return DONE; |
| } |
| |
| CameraMetadata captureCopy = mCaptureRequest; |
| if (captureCopy.entryCount() == 0) { |
| ALOGE("%s: Camera %d: Unable to copy capture request for HAL device", |
| __FUNCTION__, client->getCameraId()); |
| return DONE; |
| } |
| |
| Vector<CameraMetadata> requests; |
| requests.push(mCaptureRequest); |
| res = mBurstCapture->start(requests, mCaptureId); |
| mTimeoutCount = kMaxTimeoutsForCaptureEnd * 10; |
| return BURST_CAPTURE_WAIT; |
| } |
| |
| CaptureSequencer::CaptureState CaptureSequencer::manageBurstCaptureWait( |
| sp<Camera2Client> &/*client*/) { |
| status_t res; |
| ATRACE_CALL(); |
| |
| while (!mNewCaptureReceived) { |
| res = mNewCaptureSignal.waitRelative(mInputMutex, kWaitDuration); |
| if (res == TIMED_OUT) { |
| mTimeoutCount--; |
| break; |
| } |
| } |
| |
| if (mTimeoutCount <= 0) { |
| ALOGW("Timed out waiting for burst capture to complete"); |
| return DONE; |
| } |
| if (mNewCaptureReceived) { |
| mNewCaptureReceived = false; |
| // TODO: update mCaptureId to last burst's capture ID + 1? |
| return DONE; |
| } |
| |
| return BURST_CAPTURE_WAIT; |
| } |
| |
| status_t CaptureSequencer::updateCaptureRequest(const Parameters ¶ms, |
| sp<Camera2Client> &client) { |
| ATRACE_CALL(); |
| status_t res; |
| if (mCaptureRequest.entryCount() == 0) { |
| res = client->getCameraDevice()->createDefaultRequest( |
| CAMERA2_TEMPLATE_STILL_CAPTURE, |
| &mCaptureRequest); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to create default still image request:" |
| " %s (%d)", __FUNCTION__, client->getCameraId(), |
| strerror(-res), res); |
| return res; |
| } |
| } |
| |
| res = params.updateRequest(&mCaptureRequest); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update common entries of capture " |
| "request: %s (%d)", __FUNCTION__, client->getCameraId(), |
| strerror(-res), res); |
| return res; |
| } |
| |
| res = params.updateRequestJpeg(&mCaptureRequest); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update JPEG entries of capture " |
| "request: %s (%d)", __FUNCTION__, client->getCameraId(), |
| strerror(-res), res); |
| return res; |
| } |
| |
| return OK; |
| } |
| |
| /*static*/ void CaptureSequencer::shutterNotifyLocked(const Parameters ¶ms, |
| sp<Camera2Client> client, int msgType) { |
| ATRACE_CALL(); |
| |
| if (params.state == Parameters::STILL_CAPTURE |
| && params.playShutterSound |
| && (msgType & CAMERA_MSG_SHUTTER)) { |
| client->getCameraService()->playSound(CameraService::SOUND_SHUTTER); |
| } |
| |
| { |
| Camera2Client::SharedCameraCallbacks::Lock |
| l(client->mSharedCameraCallbacks); |
| |
| ALOGV("%s: Notifying of shutter close to client", __FUNCTION__); |
| if (l.mRemoteCallback != 0) { |
| // ShutterCallback |
| l.mRemoteCallback->notifyCallback(CAMERA_MSG_SHUTTER, |
| /*ext1*/0, /*ext2*/0); |
| |
| // RawCallback with null buffer |
| l.mRemoteCallback->notifyCallback(CAMERA_MSG_RAW_IMAGE_NOTIFY, |
| /*ext1*/0, /*ext2*/0); |
| } else { |
| ALOGV("%s: No client!", __FUNCTION__); |
| } |
| } |
| } |
| |
| |
| }; // namespace camera2 |
| }; // namespace android |