/* * Copyright (C) 2019 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "Enumerator.h" #include "HalCamera.h" #include "VirtualCamera.h" #include #include #include namespace android { namespace automotive { namespace evs { namespace V1_1 { namespace implementation { // TODO(changyeon): // We need to hook up death monitoring to detect stream death so we can attempt a reconnect using ::android::base::StringAppendF; using ::android::base::WriteStringToFd; HalCamera::~HalCamera() { // Reports the usage statistics before the destruction // EvsUsageStatsReported atom is defined in // frameworks/proto_logging/stats/atoms.proto mUsageStats->writeStats(); } sp HalCamera::makeVirtualCamera() { // Create the client camera interface object std::vector> sourceCameras; sourceCameras.reserve(1); sourceCameras.emplace_back(this); sp client = new VirtualCamera(sourceCameras); if (client == nullptr) { LOG(ERROR) << "Failed to create client camera object"; return nullptr; } if (!ownVirtualCamera(client)) { LOG(ERROR) << "Failed to own a client camera object"; client = nullptr; } return client; } bool HalCamera::ownVirtualCamera(sp virtualCamera) { if (virtualCamera == nullptr) { LOG(ERROR) << "Failed to create virtualCamera camera object"; return false; } // Make sure we have enough buffers available for all our clients if (!changeFramesInFlight(virtualCamera->getAllowedBuffers())) { // Gah! We couldn't get enough buffers, so we can't support this virtualCamera // Null the pointer, dropping our reference, thus destroying the virtualCamera object return false; } // Add this virtualCamera to our ownership list via weak pointer mClients.emplace_back(virtualCamera); // Update statistics mUsageStats->updateNumClients(mClients.size()); return true; } void HalCamera::disownVirtualCamera(sp virtualCamera) { // Ignore calls with null pointers if (virtualCamera.get() == nullptr) { LOG(WARNING) << "Ignoring disownVirtualCamera call with null pointer"; return; } // Remove the virtual camera from our client list unsigned clientCount = mClients.size(); mClients.remove(virtualCamera); if (clientCount != mClients.size() + 1) { LOG(ERROR) << "Couldn't find camera in our client list to remove it; " << "this client may be removed already."; } // Recompute the number of buffers required with the target camera removed from the list if (!changeFramesInFlight(0)) { LOG(ERROR) << "Error when trying to reduce the in flight buffer count"; } // Update statistics mUsageStats->updateNumClients(mClients.size()); } bool HalCamera::changeFramesInFlight(int delta) { // Walk all our clients and count their currently required frames unsigned bufferCount = 0; for (auto&& client : mClients) { sp virtCam = client.promote(); if (virtCam != nullptr) { bufferCount += virtCam->getAllowedBuffers(); } } // Add the requested delta bufferCount += delta; // Never drop below 1 buffer -- even if all client cameras get closed if (bufferCount < 1) { bufferCount = 1; } // Ask the hardware for the resulting buffer count Return result = mHwCamera->setMaxFramesInFlight(bufferCount); bool success = (result.isOk() && result == EvsResult::OK); // Update the size of our array of outstanding frame records if (success) { std::vector newRecords; newRecords.reserve(bufferCount); // Copy and compact the old records that are still active for (const auto& rec : mFrames) { if (rec.refCount > 0) { newRecords.emplace_back(rec); } } if (newRecords.size() > (unsigned)bufferCount) { LOG(WARNING) << "We found more frames in use than requested."; } mFrames.swap(newRecords); } return success; } bool HalCamera::changeFramesInFlight(const hidl_vec& buffers, int* delta) { // Return immediately if a list is empty. if (buffers.size() < 1) { LOG(DEBUG) << "No external buffers to add."; return true; } // Walk all our clients and count their currently required frames auto bufferCount = 0; for (auto&& client : mClients) { sp virtCam = client.promote(); if (virtCam != nullptr) { bufferCount += virtCam->getAllowedBuffers(); } } EvsResult status = EvsResult::OK; // Ask the hardware for the resulting buffer count mHwCamera->importExternalBuffers(buffers, [&](auto result, auto added) { status = result; *delta = added; }); if (status != EvsResult::OK) { LOG(ERROR) << "Failed to add external capture buffers."; return false; } bufferCount += *delta; // Update the size of our array of outstanding frame records std::vector newRecords; newRecords.reserve(bufferCount); // Copy and compact the old records that are still active for (const auto& rec : mFrames) { if (rec.refCount > 0) { newRecords.emplace_back(rec); } } if (newRecords.size() > (unsigned)bufferCount) { LOG(WARNING) << "We found more frames in use than requested."; } mFrames.swap(newRecords); return true; } void HalCamera::requestNewFrame(sp client, const int64_t lastTimestamp) { FrameRequest req; req.client = client; req.timestamp = lastTimestamp; std::lock_guard lock(mFrameMutex); mNextRequests->push_back(req); } Return HalCamera::clientStreamStarting() { Return result = EvsResult::OK; if (mStreamState == STOPPED) { mStreamState = RUNNING; result = mHwCamera->startVideoStream(this); } return result; } void HalCamera::cancelCaptureRequestFromClientLocked(std::deque* requests, const VirtualCamera* client) { auto it = requests->begin(); while (it != requests->end()) { if (it->client == client) { requests->erase(it); return; } ++it; } } void HalCamera::clientStreamEnding(const VirtualCamera* client) { { std::lock_guard lock(mFrameMutex); cancelCaptureRequestFromClientLocked(mNextRequests, client); cancelCaptureRequestFromClientLocked(mCurrentRequests, client); } // Do we still have a running client? bool stillRunning = false; for (auto&& client : mClients) { sp virtCam = client.promote(); if (virtCam != nullptr) { stillRunning |= virtCam->isStreaming(); } } // If not, then stop the hardware stream if (!stillRunning) { mStreamState = STOPPING; mHwCamera->stopVideoStream(); } } Return HalCamera::doneWithFrame(const BufferDesc_1_0& buffer) { // Find this frame in our list of outstanding frames unsigned i; for (i = 0; i < mFrames.size(); i++) { if (mFrames[i].frameId == buffer.bufferId) { break; } } if (i == mFrames.size()) { LOG(ERROR) << "We got a frame back with an ID we don't recognize!"; } else { // Are there still clients using this buffer? mFrames[i].refCount--; if (mFrames[i].refCount <= 0) { // Since all our clients are done with this buffer, return it to the device layer mHwCamera->doneWithFrame(buffer); // Counts a returned buffer mUsageStats->framesReturned(); } } return Void(); } Return HalCamera::doneWithFrame(const BufferDesc_1_1& buffer) { // Find this frame in our list of outstanding frames unsigned i; for (i = 0; i < mFrames.size(); i++) { if (mFrames[i].frameId == buffer.bufferId) { break; } } if (i == mFrames.size()) { LOG(ERROR) << "We got a frame back with an ID we don't recognize!"; } else { // Are there still clients using this buffer? mFrames[i].refCount--; if (mFrames[i].refCount <= 0) { // Since all our clients are done with this buffer, return it to the device layer hardware::hidl_vec returnedBuffers; returnedBuffers.resize(1); returnedBuffers[0] = buffer; mHwCamera->doneWithFrame_1_1(returnedBuffers); // Counts a returned buffer mUsageStats->framesReturned(returnedBuffers); } } return Void(); } // Methods from ::android::hardware::automotive::evs::V1_0::IEvsCameraStream follow. Return HalCamera::deliverFrame(const BufferDesc_1_0& buffer) { /* Frames are delivered via deliverFrame_1_1 callback for clients that implement * IEvsCameraStream v1.1 interfaces and therefore this method must not be * used. */ LOG(INFO) << "A delivered frame from EVS v1.0 HW module is rejected."; mHwCamera->doneWithFrame(buffer); // Reports a received and returned buffer mUsageStats->framesReceived(); mUsageStats->framesReturned(); return Void(); } // Methods from ::android::hardware::automotive::evs::V1_1::IEvsCameraStream follow. Return HalCamera::deliverFrame_1_1(const hardware::hidl_vec& buffer) { LOG(VERBOSE) << "Received a frame"; // Frames are being forwarded to v1.1 clients only who requested new frame. const auto timestamp = buffer[0].timestamp; // TODO(b/145750636): For now, we are using a approximately half of 1 seconds / 30 frames = 33ms // but this must be derived from current framerate. constexpr int64_t kThreshold = 16 * 1e+3; // ms unsigned frameDeliveriesV1 = 0; { // Handle frame requests from v1.1 clients std::lock_guard lock(mFrameMutex); std::swap(mCurrentRequests, mNextRequests); while (!mCurrentRequests->empty()) { auto req = mCurrentRequests->front(); mCurrentRequests->pop_front(); sp vCam = req.client.promote(); if (vCam == nullptr) { // Ignore a client already dead. continue; } else if (timestamp - req.timestamp < kThreshold) { // Skip current frame because it arrives too soon. LOG(DEBUG) << "Skips a frame from " << getId(); mNextRequests->push_back(req); // Reports a skipped frame mUsageStats->framesSkippedToSync(); } else if (vCam != nullptr) { if (!vCam->deliverFrame(buffer[0])) { LOG(WARNING) << getId() << " failed to forward the buffer to " << vCam.get(); } else { LOG(ERROR) << getId() << " forwarded the buffer #" << buffer[0].bufferId << " to " << vCam.get() << " from " << this; ++frameDeliveriesV1; } } } } // Reports the number of received buffers mUsageStats->framesReceived(buffer); // Frames are being forwarded to active v1.0 clients and v1.1 clients if we // failed to create a timeline. unsigned frameDeliveries = 0; for (auto&& client : mClients) { sp vCam = client.promote(); if (vCam == nullptr || vCam->getVersion() > 0) { continue; } if (vCam->deliverFrame(buffer[0])) { ++frameDeliveries; } } frameDeliveries += frameDeliveriesV1; if (frameDeliveries < 1) { // If none of our clients could accept the frame, then return it // right away. LOG(INFO) << "Trivially rejecting frame (" << buffer[0].bufferId << ") from " << getId() << " with no acceptance"; mHwCamera->doneWithFrame_1_1(buffer); // Reports a returned buffer mUsageStats->framesReturned(buffer); } else { // Add an entry for this frame in our tracking list. unsigned i; for (i = 0; i < mFrames.size(); ++i) { if (mFrames[i].refCount == 0) { break; } } if (i == mFrames.size()) { mFrames.emplace_back(buffer[0].bufferId); } else { mFrames[i].frameId = buffer[0].bufferId; } mFrames[i].refCount = frameDeliveries; } return Void(); } Return HalCamera::notify(const EvsEventDesc& event) { LOG(DEBUG) << "Received an event id: " << static_cast(event.aType); if(event.aType == EvsEventType::STREAM_STOPPED) { // This event happens only when there is no more active client. if (mStreamState != STOPPING) { LOG(WARNING) << "Stream stopped unexpectedly"; } mStreamState = STOPPED; } // Forward all other events to the clients for (auto&& client : mClients) { sp vCam = client.promote(); if (vCam != nullptr) { if (!vCam->notify(event)) { LOG(INFO) << "Failed to forward an event"; } } } return Void(); } Return HalCamera::setMaster(sp virtualCamera) { if (mPrimaryClient == nullptr) { LOG(DEBUG) << __FUNCTION__ << ": " << virtualCamera.get() << " becomes a primary client."; mPrimaryClient = virtualCamera; return EvsResult::OK; } else { LOG(INFO) << "This camera already has a primary client."; return EvsResult::OWNERSHIP_LOST; } } Return HalCamera::forceMaster(sp virtualCamera) { sp prevPrimary = mPrimaryClient.promote(); if (prevPrimary == virtualCamera) { LOG(DEBUG) << "Client " << virtualCamera.get() << " is already a primary client"; } else { mPrimaryClient = virtualCamera; if (prevPrimary != nullptr) { LOG(INFO) << "High priority client " << virtualCamera.get() << " steals a primary role from " << prevPrimary.get(); /* Notify a previous primary client the loss of a primary role */ EvsEventDesc event; event.aType = EvsEventType::MASTER_RELEASED; if (!prevPrimary->notify(event)) { LOG(ERROR) << "Fail to deliver a primary role lost notification"; } } } return EvsResult::OK; } Return HalCamera::unsetMaster(const VirtualCamera* virtualCamera) { if (mPrimaryClient.promote() != virtualCamera) { return EvsResult::INVALID_ARG; } else { LOG(INFO) << "Unset a primary camera client"; mPrimaryClient = nullptr; /* Notify other clients that a primary role becomes available. */ EvsEventDesc event; event.aType = EvsEventType::MASTER_RELEASED; auto cbResult = this->notify(event); if (!cbResult.isOk()) { LOG(ERROR) << "Fail to deliver a parameter change notification"; } return EvsResult::OK; } } Return HalCamera::setParameter(sp virtualCamera, CameraParam id, int32_t& value) { EvsResult result = EvsResult::INVALID_ARG; if (virtualCamera == mPrimaryClient.promote()) { mHwCamera->setIntParameter(id, value, [&result, &value](auto status, auto readValue) { result = status; value = readValue[0]; }); if (result == EvsResult::OK) { /* Notify a parameter change */ EvsEventDesc event; event.aType = EvsEventType::PARAMETER_CHANGED; event.payload[0] = static_cast(id); event.payload[1] = static_cast(value); auto cbResult = this->notify(event); if (!cbResult.isOk()) { LOG(ERROR) << "Fail to deliver a parameter change notification"; } } } else { LOG(WARNING) << "A parameter change request from the non-primary client is declined."; /* Read a current value of a requested camera parameter */ getParameter(id, value); } return result; } Return HalCamera::getParameter(CameraParam id, int32_t& value) { EvsResult result = EvsResult::OK; mHwCamera->getIntParameter(id, [&result, &value](auto status, auto readValue) { result = status; if (result == EvsResult::OK) { value = readValue[0]; } }); return result; } CameraUsageStatsRecord HalCamera::getStats() const { return mUsageStats->snapshot(); } Stream HalCamera::getStreamConfiguration() const { return mStreamConfig; } std::string HalCamera::toString(const char* indent) const { std::string buffer; const auto timeElapsedMs = android::uptimeMillis() - mTimeCreatedMs; StringAppendF(&buffer, "%sCreated: @%" PRId64 " (elapsed %" PRId64 " ms)\n", indent, mTimeCreatedMs, timeElapsedMs); std::string double_indent(indent); double_indent += indent; buffer += CameraUsageStats::toString(getStats(), double_indent.c_str()); for (auto&& client : mClients) { auto handle = client.promote(); if (!handle) { continue; } StringAppendF(&buffer, "%sClient %p\n", indent, handle.get()); buffer += handle->toString(double_indent.c_str()); } StringAppendF(&buffer, "%sPrimary client: %p\n", indent, mPrimaryClient.promote().get()); buffer += HalCamera::toString(mStreamConfig, indent); return buffer; } std::string HalCamera::toString(Stream configuration, const char* indent) { std::string streamInfo; std::string double_indent(indent); double_indent += indent; StringAppendF(&streamInfo, "%sActive Stream Configuration\n" "%sid: %d\n" "%swidth: %d\n" "%sheight: %d\n" "%sformat: 0x%X\n" "%susage: 0x%" PRIx64 "\n" "%srotation: 0x%X\n\n", indent, double_indent.c_str(), configuration.id, double_indent.c_str(), configuration.width, double_indent.c_str(), configuration.height, double_indent.c_str(), configuration.format, double_indent.c_str(), configuration.usage, double_indent.c_str(), configuration.rotation); return streamInfo; } } // namespace implementation } // namespace V1_1 } // namespace evs } // namespace automotive } // namespace android