blob: 20d91f02e155e41c348fc02703af517eefe8376f [file] [log] [blame]
// Copyright 2014 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include <fcntl.h>
#include <linux/videodev2.h>
#include <poll.h>
#include <sys/eventfd.h>
#include <sys/ioctl.h>
#include <sys/mman.h>
#include "base/bind.h"
#include "base/bind_helpers.h"
#include "base/callback.h"
#include "base/message_loop/message_loop_proxy.h"
#include "base/numerics/safe_conversions.h"
#include "content/common/gpu/media/v4l2_image_processor.h"
#include "media/base/bind_to_current_loop.h"
#define NOTIFY_ERROR() \
do { \
LOG(ERROR) << "calling NotifyError()"; \
NotifyError(); \
} while (0)
#define IOCTL_OR_ERROR_RETURN_VALUE(type, arg, value) \
do { \
if (device_->Ioctl(type, arg) != 0) { \
PLOG(ERROR) << __func__ << "(): ioctl() failed: " << #type; \
return value; \
} \
} while (0)
#define IOCTL_OR_ERROR_RETURN(type, arg) \
IOCTL_OR_ERROR_RETURN_VALUE(type, arg, ((void)0))
#define IOCTL_OR_ERROR_RETURN_FALSE(type, arg) \
IOCTL_OR_ERROR_RETURN_VALUE(type, arg, false)
#define IOCTL_OR_LOG_ERROR(type, arg) \
do { \
if (device_->Ioctl(type, arg) != 0) \
PLOG(ERROR) << __func__ << "(): ioctl() failed: " << #type; \
} while (0)
namespace content {
V4L2ImageProcessor::InputRecord::InputRecord() : at_device(false) {
}
V4L2ImageProcessor::OutputRecord::OutputRecord()
: at_device(false), at_client(false) {
}
V4L2ImageProcessor::JobRecord::JobRecord() {
}
V4L2ImageProcessor::V4L2ImageProcessor(scoped_ptr<V4L2Device> device)
: input_format_(media::VideoFrame::UNKNOWN),
output_format_(media::VideoFrame::UNKNOWN),
input_format_fourcc_(0),
output_format_fourcc_(0),
input_planes_count_(0),
output_planes_count_(0),
child_message_loop_proxy_(base::MessageLoopProxy::current()),
device_(device.Pass()),
device_thread_("V4L2ImageProcessorThread"),
device_poll_thread_("V4L2ImageProcessorDevicePollThread"),
input_streamon_(false),
input_buffer_queued_count_(0),
output_streamon_(false),
output_buffer_queued_count_(0),
device_weak_factory_(this) {
}
V4L2ImageProcessor::~V4L2ImageProcessor() {
DCHECK(child_message_loop_proxy_->BelongsToCurrentThread());
DCHECK(!device_thread_.IsRunning());
DCHECK(!device_poll_thread_.IsRunning());
DestroyInputBuffers();
DestroyOutputBuffers();
}
void V4L2ImageProcessor::NotifyError() {
if (!child_message_loop_proxy_->BelongsToCurrentThread())
child_message_loop_proxy_->PostTask(FROM_HERE, error_cb_);
else
error_cb_.Run();
}
bool V4L2ImageProcessor::Initialize(media::VideoFrame::Format input_format,
media::VideoFrame::Format output_format,
gfx::Size input_visible_size,
gfx::Size output_visible_size,
gfx::Size output_allocated_size,
const base::Closure& error_cb) {
DCHECK(!error_cb.is_null());
error_cb_ = error_cb;
// TODO(posciak): Replace Exynos-specific format/parameter hardcoding in this
// class with proper capability enumeration.
DCHECK_EQ(input_format, media::VideoFrame::I420);
DCHECK_EQ(output_format, media::VideoFrame::NV12);
input_format_ = input_format;
output_format_ = output_format;
input_format_fourcc_ = V4L2Device::VideoFrameFormatToV4L2PixFmt(input_format);
output_format_fourcc_ =
V4L2Device::VideoFrameFormatToV4L2PixFmt(output_format);
if (!input_format_fourcc_ || !output_format_fourcc_) {
LOG(ERROR) << "Unrecognized format(s)";
return false;
}
input_visible_size_ = input_visible_size;
output_visible_size_ = output_visible_size;
output_allocated_size_ = output_allocated_size;
input_planes_count_ = media::VideoFrame::NumPlanes(input_format);
DCHECK_LE(input_planes_count_, static_cast<size_t>(VIDEO_MAX_PLANES));
output_planes_count_ = media::VideoFrame::NumPlanes(output_format);
DCHECK_LE(output_planes_count_, static_cast<size_t>(VIDEO_MAX_PLANES));
// Capabilities check.
struct v4l2_capability caps;
memset(&caps, 0, sizeof(caps));
const __u32 kCapsRequired = V4L2_CAP_VIDEO_CAPTURE_MPLANE |
V4L2_CAP_VIDEO_OUTPUT_MPLANE | V4L2_CAP_STREAMING;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QUERYCAP, &caps);
if ((caps.capabilities & kCapsRequired) != kCapsRequired) {
LOG(ERROR) << "Initialize(): ioctl() failed: VIDIOC_QUERYCAP: "
"caps check failed: 0x" << std::hex << caps.capabilities;
return false;
}
if (!CreateInputBuffers() || !CreateOutputBuffers())
return false;
if (!device_thread_.Start()) {
LOG(ERROR) << "Initialize(): encoder thread failed to start";
return false;
}
// StartDevicePoll will NOTIFY_ERROR on failure, so IgnoreResult is fine here.
device_thread_.message_loop()->PostTask(
FROM_HERE,
base::Bind(base::IgnoreResult(&V4L2ImageProcessor::StartDevicePoll),
base::Unretained(this)));
DVLOG(1) << "V4L2ImageProcessor initialized for "
<< " input_format:"
<< media::VideoFrame::FormatToString(input_format)
<< ", output_format:"
<< media::VideoFrame::FormatToString(output_format)
<< ", input_visible_size: " << input_visible_size.ToString()
<< ", input_allocated_size: " << input_allocated_size_.ToString()
<< ", output_visible_size: " << output_visible_size.ToString()
<< ", output_allocated_size: " << output_allocated_size.ToString();
return true;
}
void V4L2ImageProcessor::Process(const scoped_refptr<media::VideoFrame>& frame,
const FrameReadyCB& cb) {
DVLOG(3) << __func__ << ": ts=" << frame->timestamp().InMilliseconds();
scoped_ptr<JobRecord> job_record(new JobRecord());
job_record->frame = frame;
job_record->ready_cb = cb;
device_thread_.message_loop()->PostTask(
FROM_HERE,
base::Bind(&V4L2ImageProcessor::ProcessTask,
base::Unretained(this),
base::Passed(&job_record)));
}
void V4L2ImageProcessor::ProcessTask(scoped_ptr<JobRecord> job_record) {
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
input_queue_.push(make_linked_ptr(job_record.release()));
Enqueue();
}
void V4L2ImageProcessor::Destroy() {
DVLOG(3) << __func__;
DCHECK(child_message_loop_proxy_->BelongsToCurrentThread());
// If the device thread is running, destroy using posted task.
if (device_thread_.IsRunning()) {
device_thread_.message_loop()->PostTask(
FROM_HERE,
base::Bind(&V4L2ImageProcessor::DestroyTask, base::Unretained(this)));
// Wait for tasks to finish/early-exit.
device_thread_.Stop();
} else {
// Otherwise DestroyTask() is not needed.
DCHECK(!device_poll_thread_.IsRunning());
DCHECK(!device_weak_factory_.HasWeakPtrs());
}
delete this;
}
void V4L2ImageProcessor::DestroyTask() {
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
device_weak_factory_.InvalidateWeakPtrs();
// Stop streaming and the device_poll_thread_.
StopDevicePoll();
}
bool V4L2ImageProcessor::CreateInputBuffers() {
DVLOG(3) << __func__;
DCHECK(child_message_loop_proxy_->BelongsToCurrentThread());
DCHECK(!input_streamon_);
struct v4l2_control control;
memset(&control, 0, sizeof(control));
control.id = V4L2_CID_ROTATE;
control.value = 0;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CTRL, &control);
memset(&control, 0, sizeof(control));
control.id = V4L2_CID_HFLIP;
control.value = 0;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CTRL, &control);
memset(&control, 0, sizeof(control));
control.id = V4L2_CID_VFLIP;
control.value = 0;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CTRL, &control);
memset(&control, 0, sizeof(control));
control.id = V4L2_CID_ALPHA_COMPONENT;
control.value = 255;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CTRL, &control);
struct v4l2_format format;
memset(&format, 0, sizeof(format));
format.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
format.fmt.pix_mp.width = input_visible_size_.width();
format.fmt.pix_mp.height = input_visible_size_.height();
format.fmt.pix_mp.pixelformat = input_format_fourcc_;
format.fmt.pix_mp.num_planes = input_planes_count_;
for (size_t i = 0; i < input_planes_count_; ++i) {
format.fmt.pix_mp.plane_fmt[i].sizeimage =
media::VideoFrame::PlaneAllocationSize(
input_format_, i, input_allocated_size_);
format.fmt.pix_mp.plane_fmt[i].bytesperline =
base::checked_cast<__u32>(input_allocated_size_.width());
}
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_FMT, &format);
input_allocated_size_ = V4L2Device::CodedSizeFromV4L2Format(format);
DCHECK(gfx::Rect(input_allocated_size_).Contains(
gfx::Rect(input_visible_size_)));
struct v4l2_crop crop;
memset(&crop, 0, sizeof(crop));
crop.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
crop.c.left = 0;
crop.c.top = 0;
crop.c.width = base::checked_cast<__u32>(input_visible_size_.width());
crop.c.height = base::checked_cast<__u32>(input_visible_size_.height());
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CROP, &crop);
struct v4l2_requestbuffers reqbufs;
memset(&reqbufs, 0, sizeof(reqbufs));
reqbufs.count = kInputBufferCount;
reqbufs.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
reqbufs.memory = V4L2_MEMORY_USERPTR;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_REQBUFS, &reqbufs);
DCHECK(input_buffer_map_.empty());
input_buffer_map_.resize(reqbufs.count);
for (size_t i = 0; i < input_buffer_map_.size(); ++i)
free_input_buffers_.push_back(i);
return true;
}
bool V4L2ImageProcessor::CreateOutputBuffers() {
DVLOG(3) << __func__;
DCHECK(child_message_loop_proxy_->BelongsToCurrentThread());
DCHECK(!output_streamon_);
struct v4l2_format format;
memset(&format, 0, sizeof(format));
format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
format.fmt.pix_mp.width = output_allocated_size_.width();
format.fmt.pix_mp.height = output_allocated_size_.height();
format.fmt.pix_mp.pixelformat = output_format_fourcc_;
format.fmt.pix_mp.num_planes = output_planes_count_;
for (size_t i = 0; i < output_planes_count_; ++i) {
format.fmt.pix_mp.plane_fmt[i].sizeimage =
media::VideoFrame::PlaneAllocationSize(
output_format_, i, output_allocated_size_);
format.fmt.pix_mp.plane_fmt[i].bytesperline =
base::checked_cast<__u32>(output_allocated_size_.width());
}
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_FMT, &format);
gfx::Size adjusted_allocated_size =
V4L2Device::CodedSizeFromV4L2Format(format);
DCHECK(gfx::Rect(adjusted_allocated_size).Contains(
gfx::Rect(output_allocated_size_)));
output_allocated_size_ = adjusted_allocated_size;
struct v4l2_crop crop;
memset(&crop, 0, sizeof(crop));
crop.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
crop.c.left = 0;
crop.c.top = 0;
crop.c.width = base::checked_cast<__u32>(output_visible_size_.width());
crop.c.height = base::checked_cast<__u32>(output_visible_size_.height());
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CROP, &crop);
struct v4l2_requestbuffers reqbufs;
memset(&reqbufs, 0, sizeof(reqbufs));
reqbufs.count = kOutputBufferCount;
reqbufs.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
reqbufs.memory = V4L2_MEMORY_MMAP;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_REQBUFS, &reqbufs);
DCHECK(output_buffer_map_.empty());
output_buffer_map_.resize(reqbufs.count);
for (size_t i = 0; i < output_buffer_map_.size(); ++i) {
OutputRecord& output_record = output_buffer_map_[i];
output_record.fds.resize(output_planes_count_);
for (size_t j = 0; j < output_planes_count_; ++j) {
struct v4l2_exportbuffer expbuf;
memset(&expbuf, 0, sizeof(expbuf));
expbuf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
expbuf.index = i;
expbuf.plane = j;
expbuf.flags = O_CLOEXEC;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_EXPBUF, &expbuf);
output_record.fds[j] = expbuf.fd;
}
free_output_buffers_.push_back(i);
}
return true;
}
void V4L2ImageProcessor::DestroyInputBuffers() {
DCHECK(child_message_loop_proxy_->BelongsToCurrentThread());
DCHECK(!input_streamon_);
struct v4l2_requestbuffers reqbufs;
memset(&reqbufs, 0, sizeof(reqbufs));
reqbufs.count = 0;
reqbufs.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
reqbufs.memory = V4L2_MEMORY_USERPTR;
IOCTL_OR_LOG_ERROR(VIDIOC_REQBUFS, &reqbufs);
input_buffer_map_.clear();
free_input_buffers_.clear();
}
void V4L2ImageProcessor::DestroyOutputBuffers() {
DCHECK(child_message_loop_proxy_->BelongsToCurrentThread());
DCHECK(!output_streamon_);
for (size_t buf = 0; buf < output_buffer_map_.size(); ++buf) {
OutputRecord& output_record = output_buffer_map_[buf];
for (size_t plane = 0; plane < output_record.fds.size(); ++plane)
close(output_record.fds[plane]);
output_record.fds.clear();
}
struct v4l2_requestbuffers reqbufs;
memset(&reqbufs, 0, sizeof(reqbufs));
reqbufs.count = 0;
reqbufs.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
reqbufs.memory = V4L2_MEMORY_MMAP;
IOCTL_OR_LOG_ERROR(VIDIOC_REQBUFS, &reqbufs);
output_buffer_map_.clear();
free_output_buffers_.clear();
}
void V4L2ImageProcessor::DevicePollTask(bool poll_device) {
DCHECK_EQ(device_poll_thread_.message_loop(), base::MessageLoop::current());
bool event_pending;
if (!device_->Poll(poll_device, &event_pending)) {
NOTIFY_ERROR();
return;
}
// All processing should happen on ServiceDeviceTask(), since we shouldn't
// touch encoder state from this thread.
device_thread_.message_loop()->PostTask(
FROM_HERE,
base::Bind(&V4L2ImageProcessor::ServiceDeviceTask,
base::Unretained(this)));
}
void V4L2ImageProcessor::ServiceDeviceTask() {
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
// ServiceDeviceTask() should only ever be scheduled from DevicePollTask(),
// so either:
// * device_poll_thread_ is running normally
// * device_poll_thread_ scheduled us, but then a DestroyTask() shut it down,
// in which case we should early-out.
if (!device_poll_thread_.message_loop())
return;
Dequeue();
Enqueue();
if (!device_->ClearDevicePollInterrupt())
return;
bool poll_device =
(input_buffer_queued_count_ > 0 && output_buffer_queued_count_ > 0);
device_poll_thread_.message_loop()->PostTask(
FROM_HERE,
base::Bind(&V4L2ImageProcessor::DevicePollTask,
base::Unretained(this),
poll_device));
DVLOG(2) << __func__ << ": buffer counts: INPUT["
<< input_queue_.size() << "] => DEVICE["
<< free_input_buffers_.size() << "+"
<< input_buffer_queued_count_ << "/"
<< input_buffer_map_.size() << "->"
<< free_output_buffers_.size() << "+"
<< output_buffer_queued_count_ << "/"
<< output_buffer_map_.size() << "] => CLIENT["
<< output_buffer_map_.size() - output_buffer_queued_count_ -
free_output_buffers_.size() << "]";
}
void V4L2ImageProcessor::Enqueue() {
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
const int old_inputs_queued = input_buffer_queued_count_;
while (!input_queue_.empty() && !free_input_buffers_.empty()) {
if (!EnqueueInputRecord())
return;
}
if (old_inputs_queued == 0 && input_buffer_queued_count_ != 0) {
// We started up a previously empty queue.
// Queue state changed; signal interrupt.
if (!device_->SetDevicePollInterrupt())
return;
// VIDIOC_STREAMON if we haven't yet.
if (!input_streamon_) {
__u32 type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
IOCTL_OR_ERROR_RETURN(VIDIOC_STREAMON, &type);
input_streamon_ = true;
}
}
// TODO(posciak): Fix this to be non-Exynos specific.
// Exynos GSC is liable to race conditions if more than one output buffer is
// simultaneously enqueued, so enqueue just one.
if (output_buffer_queued_count_ == 0 && !free_output_buffers_.empty()) {
const int old_outputs_queued = output_buffer_queued_count_;
if (!EnqueueOutputRecord())
return;
if (old_outputs_queued == 0 && output_buffer_queued_count_ != 0) {
// We just started up a previously empty queue.
// Queue state changed; signal interrupt.
if (!device_->SetDevicePollInterrupt())
return;
// Start VIDIOC_STREAMON if we haven't yet.
if (!output_streamon_) {
__u32 type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
IOCTL_OR_ERROR_RETURN(VIDIOC_STREAMON, &type);
output_streamon_ = true;
}
}
}
DCHECK_LE(output_buffer_queued_count_, 1);
}
void V4L2ImageProcessor::Dequeue() {
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
// Dequeue completed input (VIDEO_OUTPUT) buffers,
// and recycle to the free list.
struct v4l2_buffer dqbuf;
struct v4l2_plane planes[VIDEO_MAX_PLANES];
while (input_buffer_queued_count_ > 0) {
DCHECK(input_streamon_);
memset(&dqbuf, 0, sizeof(dqbuf));
memset(&planes, 0, sizeof(planes));
dqbuf.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
dqbuf.memory = V4L2_MEMORY_USERPTR;
dqbuf.m.planes = planes;
dqbuf.length = input_planes_count_;
if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) {
if (errno == EAGAIN) {
// EAGAIN if we're just out of buffers to dequeue.
break;
}
PLOG(ERROR) << "ioctl() failed: VIDIOC_DQBUF";
NOTIFY_ERROR();
return;
}
InputRecord& input_record = input_buffer_map_[dqbuf.index];
DCHECK(input_record.at_device);
input_record.at_device = false;
input_record.frame = NULL;
free_input_buffers_.push_back(dqbuf.index);
input_buffer_queued_count_--;
}
// Dequeue completed output (VIDEO_CAPTURE) buffers, recycle to the free list.
// Return the finished buffer to the client via the job ready callback.
while (output_buffer_queued_count_ > 0) {
DCHECK(output_streamon_);
memset(&dqbuf, 0, sizeof(dqbuf));
memset(&planes, 0, sizeof(planes));
dqbuf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
dqbuf.memory = V4L2_MEMORY_DMABUF;
dqbuf.m.planes = planes;
dqbuf.length = output_planes_count_;
if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) {
if (errno == EAGAIN) {
// EAGAIN if we're just out of buffers to dequeue.
break;
}
PLOG(ERROR) << "ioctl() failed: VIDIOC_DQBUF";
NOTIFY_ERROR();
return;
}
OutputRecord& output_record = output_buffer_map_[dqbuf.index];
DCHECK(output_record.at_device);
output_record.at_device = false;
output_record.at_client = true;
output_buffer_queued_count_--;
// Jobs are always processed in FIFO order.
DCHECK(!running_jobs_.empty());
linked_ptr<JobRecord> job_record = running_jobs_.front();
running_jobs_.pop();
scoped_refptr<media::VideoFrame> output_frame =
media::VideoFrame::WrapExternalDmabufs(
output_format_,
output_allocated_size_,
gfx::Rect(output_visible_size_),
output_visible_size_,
output_record.fds,
job_record->frame->timestamp(),
media::BindToCurrentLoop(
base::Bind(&V4L2ImageProcessor::ReuseOutputBuffer,
device_weak_factory_.GetWeakPtr(),
dqbuf.index)));
DVLOG(3) << "Processing finished, returning frame, ts="
<< output_frame->timestamp().InMilliseconds();
child_message_loop_proxy_->PostTask(
FROM_HERE, base::Bind(job_record->ready_cb, output_frame));
}
}
void V4L2ImageProcessor::ReuseOutputBuffer(int index) {
DVLOG(3) << "Reusing output buffer, index=" << index;
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
OutputRecord& output_record = output_buffer_map_[index];
DCHECK(output_record.at_client);
DCHECK(!output_record.at_device);
output_record.at_client = false;
free_output_buffers_.push_back(index);
Enqueue();
}
bool V4L2ImageProcessor::EnqueueInputRecord() {
DCHECK(!input_queue_.empty());
DCHECK(!free_input_buffers_.empty());
// Enqueue an input (VIDEO_OUTPUT) buffer for an input video frame.
linked_ptr<JobRecord> job_record = input_queue_.front();
input_queue_.pop();
const int index = free_input_buffers_.back();
InputRecord& input_record = input_buffer_map_[index];
DCHECK(!input_record.at_device);
input_record.frame = job_record->frame;
struct v4l2_buffer qbuf;
struct v4l2_plane qbuf_planes[VIDEO_MAX_PLANES];
memset(&qbuf, 0, sizeof(qbuf));
memset(qbuf_planes, 0, sizeof(qbuf_planes));
qbuf.index = index;
qbuf.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
qbuf.memory = V4L2_MEMORY_USERPTR;
qbuf.m.planes = qbuf_planes;
qbuf.length = input_planes_count_;
for (size_t i = 0; i < input_planes_count_; ++i) {
qbuf.m.planes[i].bytesused = media::VideoFrame::PlaneAllocationSize(
input_record.frame->format(), i, input_allocated_size_);
qbuf.m.planes[i].length = qbuf.m.planes[i].bytesused;
qbuf.m.planes[i].m.userptr =
reinterpret_cast<unsigned long>(input_record.frame->data(i));
}
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QBUF, &qbuf);
input_record.at_device = true;
running_jobs_.push(job_record);
free_input_buffers_.pop_back();
input_buffer_queued_count_++;
DVLOG(3) << __func__ << ": enqueued frame ts="
<< job_record->frame->timestamp().InMilliseconds() << " to device.";
return true;
}
bool V4L2ImageProcessor::EnqueueOutputRecord() {
DCHECK(!free_output_buffers_.empty());
// Enqueue an output (VIDEO_CAPTURE) buffer.
const int index = free_output_buffers_.back();
OutputRecord& output_record = output_buffer_map_[index];
DCHECK(!output_record.at_device);
struct v4l2_buffer qbuf;
struct v4l2_plane qbuf_planes[VIDEO_MAX_PLANES];
memset(&qbuf, 0, sizeof(qbuf));
memset(qbuf_planes, 0, sizeof(qbuf_planes));
qbuf.index = index;
qbuf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
qbuf.memory = V4L2_MEMORY_MMAP;
qbuf.m.planes = qbuf_planes;
qbuf.length = output_planes_count_;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QBUF, &qbuf);
output_record.at_device = true;
free_output_buffers_.pop_back();
output_buffer_queued_count_++;
return true;
}
bool V4L2ImageProcessor::StartDevicePoll() {
DVLOG(3) << __func__ << ": starting device poll";
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
DCHECK(!device_poll_thread_.IsRunning());
// Start up the device poll thread and schedule its first DevicePollTask().
if (!device_poll_thread_.Start()) {
LOG(ERROR) << "StartDevicePoll(): Device thread failed to start";
NOTIFY_ERROR();
return false;
}
// Enqueue a poll task with no devices to poll on - will wait only for the
// poll interrupt
device_poll_thread_.message_loop()->PostTask(
FROM_HERE,
base::Bind(
&V4L2ImageProcessor::DevicePollTask, base::Unretained(this), false));
return true;
}
bool V4L2ImageProcessor::StopDevicePoll() {
DVLOG(3) << __func__ << ": stopping device poll";
if (device_thread_.IsRunning())
DCHECK_EQ(device_thread_.message_loop(), base::MessageLoop::current());
// Signal the DevicePollTask() to stop, and stop the device poll thread.
if (!device_->SetDevicePollInterrupt())
return false;
device_poll_thread_.Stop();
// Clear the interrupt now, to be sure.
if (!device_->ClearDevicePollInterrupt())
return false;
if (input_streamon_) {
__u32 type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_STREAMOFF, &type);
}
input_streamon_ = false;
if (output_streamon_) {
__u32 type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_STREAMOFF, &type);
}
output_streamon_ = false;
// Reset all our accounting info.
while (!input_queue_.empty())
input_queue_.pop();
while (!running_jobs_.empty())
running_jobs_.pop();
free_input_buffers_.clear();
for (size_t i = 0; i < input_buffer_map_.size(); ++i) {
InputRecord& input_record = input_buffer_map_[i];
input_record.at_device = false;
input_record.frame = NULL;
free_input_buffers_.push_back(i);
}
input_buffer_queued_count_ = 0;
free_output_buffers_.clear();
for (size_t i = 0; i < output_buffer_map_.size(); ++i) {
OutputRecord& output_record = output_buffer_map_[i];
output_record.at_device = false;
if (!output_record.at_client)
free_output_buffers_.push_back(i);
}
output_buffer_queued_count_ = 0;
return true;
}
} // namespace content