| // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "media/base/video_frame.h" |
| |
| #include <algorithm> |
| |
| #include "base/bind.h" |
| #include "base/callback_helpers.h" |
| #include "base/logging.h" |
| #include "base/memory/aligned_memory.h" |
| #include "base/strings/string_piece.h" |
| #include "gpu/command_buffer/common/mailbox_holder.h" |
| #include "media/base/limits.h" |
| #include "media/base/video_util.h" |
| #include "ui/gfx/point.h" |
| |
| #if !defined(MEDIA_FOR_CAST_IOS) |
| #include "third_party/skia/include/core/SkBitmap.h" |
| #endif |
| |
| namespace media { |
| |
| static bool IsPowerOfTwo(size_t x) { |
| return x != 0 && (x & (x - 1)) == 0; |
| } |
| |
| static inline size_t RoundUp(size_t value, size_t alignment) { |
| DCHECK(IsPowerOfTwo(alignment)); |
| return ((value + (alignment - 1)) & ~(alignment - 1)); |
| } |
| |
| static inline size_t RoundDown(size_t value, size_t alignment) { |
| DCHECK(IsPowerOfTwo(alignment)); |
| return value & ~(alignment - 1); |
| } |
| |
| // Returns the pixel size per element for given |plane| and |format|. E.g. 2x2 |
| // for the U-plane in I420. |
| static gfx::Size SampleSize(VideoFrame::Format format, size_t plane) { |
| DCHECK(VideoFrame::IsValidPlane(plane, format)); |
| |
| switch (plane) { |
| case VideoFrame::kYPlane: |
| case VideoFrame::kAPlane: |
| return gfx::Size(1, 1); |
| |
| case VideoFrame::kUPlane: // and VideoFrame::kUVPlane: |
| case VideoFrame::kVPlane: |
| switch (format) { |
| case VideoFrame::YV24: |
| return gfx::Size(1, 1); |
| |
| case VideoFrame::YV16: |
| return gfx::Size(2, 1); |
| |
| case VideoFrame::YV12: |
| case VideoFrame::YV12J: |
| case VideoFrame::I420: |
| case VideoFrame::YV12A: |
| case VideoFrame::NV12: |
| return gfx::Size(2, 2); |
| |
| case VideoFrame::UNKNOWN: |
| #if defined(VIDEO_HOLE) |
| case VideoFrame::HOLE: |
| #endif // defined(VIDEO_HOLE) |
| case VideoFrame::NATIVE_TEXTURE: |
| break; |
| } |
| } |
| NOTREACHED(); |
| return gfx::Size(); |
| } |
| |
| // Return the alignment for the whole frame, calculated as the max of the |
| // alignment for each individual plane. |
| static gfx::Size CommonAlignment(VideoFrame::Format format) { |
| int max_sample_width = 0; |
| int max_sample_height = 0; |
| for (size_t plane = 0; plane < VideoFrame::NumPlanes(format); ++plane) { |
| const gfx::Size sample_size = SampleSize(format, plane); |
| max_sample_width = std::max(max_sample_width, sample_size.width()); |
| max_sample_height = std::max(max_sample_height, sample_size.height()); |
| } |
| return gfx::Size(max_sample_width, max_sample_height); |
| } |
| |
| // Returns the number of bytes per element for given |plane| and |format|. E.g. |
| // 2 for the UV plane in NV12. |
| static int BytesPerElement(VideoFrame::Format format, size_t plane) { |
| DCHECK(VideoFrame::IsValidPlane(plane, format)); |
| return (format == VideoFrame::NV12 && plane == VideoFrame::kUVPlane) ? 2 : 1; |
| } |
| |
| // Rounds up |coded_size| if necessary for |format|. |
| static gfx::Size AdjustCodedSize(VideoFrame::Format format, |
| const gfx::Size& coded_size) { |
| const gfx::Size alignment = CommonAlignment(format); |
| return gfx::Size(RoundUp(coded_size.width(), alignment.width()), |
| RoundUp(coded_size.height(), alignment.height())); |
| } |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::CreateFrame( |
| VideoFrame::Format format, |
| const gfx::Size& coded_size, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size, |
| base::TimeDelta timestamp) { |
| DCHECK(format != VideoFrame::UNKNOWN && |
| format != VideoFrame::NV12 && |
| format != VideoFrame::NATIVE_TEXTURE); |
| #if defined(VIDEO_HOLE) |
| DCHECK(format != VideoFrame::HOLE); |
| #endif // defined(VIDEO_HOLE) |
| |
| // Since we're creating a new YUV frame (and allocating memory for it |
| // ourselves), we can pad the requested |coded_size| if necessary if the |
| // request does not line up on sample boundaries. |
| const gfx::Size new_coded_size = AdjustCodedSize(format, coded_size); |
| DCHECK(IsValidConfig(format, new_coded_size, visible_rect, natural_size)); |
| |
| scoped_refptr<VideoFrame> frame( |
| new VideoFrame(format, |
| new_coded_size, |
| visible_rect, |
| natural_size, |
| scoped_ptr<gpu::MailboxHolder>(), |
| timestamp, |
| false)); |
| frame->AllocateYUV(); |
| return frame; |
| } |
| |
| // static |
| std::string VideoFrame::FormatToString(VideoFrame::Format format) { |
| switch (format) { |
| case VideoFrame::UNKNOWN: |
| return "UNKNOWN"; |
| case VideoFrame::YV12: |
| return "YV12"; |
| case VideoFrame::YV16: |
| return "YV16"; |
| case VideoFrame::I420: |
| return "I420"; |
| case VideoFrame::NATIVE_TEXTURE: |
| return "NATIVE_TEXTURE"; |
| #if defined(VIDEO_HOLE) |
| case VideoFrame::HOLE: |
| return "HOLE"; |
| #endif // defined(VIDEO_HOLE) |
| case VideoFrame::YV12A: |
| return "YV12A"; |
| case VideoFrame::YV12J: |
| return "YV12J"; |
| case VideoFrame::NV12: |
| return "NV12"; |
| case VideoFrame::YV24: |
| return "YV24"; |
| } |
| NOTREACHED() << "Invalid videoframe format provided: " << format; |
| return ""; |
| } |
| |
| // static |
| bool VideoFrame::IsValidConfig(VideoFrame::Format format, |
| const gfx::Size& coded_size, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size) { |
| // Check maximum limits for all formats. |
| if (coded_size.GetArea() > limits::kMaxCanvas || |
| coded_size.width() > limits::kMaxDimension || |
| coded_size.height() > limits::kMaxDimension || |
| visible_rect.x() < 0 || visible_rect.y() < 0 || |
| visible_rect.right() > coded_size.width() || |
| visible_rect.bottom() > coded_size.height() || |
| natural_size.GetArea() > limits::kMaxCanvas || |
| natural_size.width() > limits::kMaxDimension || |
| natural_size.height() > limits::kMaxDimension) |
| return false; |
| |
| // Check format-specific width/height requirements. |
| switch (format) { |
| case VideoFrame::UNKNOWN: |
| return (coded_size.IsEmpty() && visible_rect.IsEmpty() && |
| natural_size.IsEmpty()); |
| |
| // NATIVE_TEXTURE and HOLE have no software-allocated buffers and are |
| // allowed to skip the below check. |
| case VideoFrame::NATIVE_TEXTURE: |
| #if defined(VIDEO_HOLE) |
| case VideoFrame::HOLE: |
| #endif // defined(VIDEO_HOLE) |
| return true; |
| |
| case VideoFrame::YV24: |
| case VideoFrame::YV12: |
| case VideoFrame::YV12J: |
| case VideoFrame::I420: |
| case VideoFrame::YV12A: |
| case VideoFrame::NV12: |
| case VideoFrame::YV16: |
| // Check that software-allocated buffer formats are aligned correctly and |
| // not empty. |
| const gfx::Size alignment = CommonAlignment(format); |
| return RoundUp(visible_rect.right(), alignment.width()) <= |
| static_cast<size_t>(coded_size.width()) && |
| RoundUp(visible_rect.bottom(), alignment.height()) <= |
| static_cast<size_t>(coded_size.height()) && |
| !coded_size.IsEmpty() && !visible_rect.IsEmpty() && |
| !natural_size.IsEmpty(); |
| } |
| |
| NOTREACHED(); |
| return false; |
| } |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::WrapNativeTexture( |
| scoped_ptr<gpu::MailboxHolder> mailbox_holder, |
| const ReleaseMailboxCB& mailbox_holder_release_cb, |
| const gfx::Size& coded_size, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size, |
| base::TimeDelta timestamp, |
| const ReadPixelsCB& read_pixels_cb) { |
| scoped_refptr<VideoFrame> frame(new VideoFrame(NATIVE_TEXTURE, |
| coded_size, |
| visible_rect, |
| natural_size, |
| mailbox_holder.Pass(), |
| timestamp, |
| false)); |
| frame->mailbox_holder_release_cb_ = mailbox_holder_release_cb; |
| frame->read_pixels_cb_ = read_pixels_cb; |
| |
| return frame; |
| } |
| |
| #if !defined(MEDIA_FOR_CAST_IOS) |
| void VideoFrame::ReadPixelsFromNativeTexture(const SkBitmap& pixels) { |
| DCHECK_EQ(format_, NATIVE_TEXTURE); |
| if (!read_pixels_cb_.is_null()) |
| read_pixels_cb_.Run(pixels); |
| } |
| #endif |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::WrapExternalPackedMemory( |
| Format format, |
| const gfx::Size& coded_size, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size, |
| uint8* data, |
| size_t data_size, |
| base::SharedMemoryHandle handle, |
| base::TimeDelta timestamp, |
| const base::Closure& no_longer_needed_cb) { |
| const gfx::Size new_coded_size = AdjustCodedSize(format, coded_size); |
| |
| if (!IsValidConfig(format, new_coded_size, visible_rect, natural_size)) |
| return NULL; |
| if (data_size < AllocationSize(format, new_coded_size)) |
| return NULL; |
| |
| switch (format) { |
| case VideoFrame::I420: { |
| scoped_refptr<VideoFrame> frame( |
| new VideoFrame(format, |
| new_coded_size, |
| visible_rect, |
| natural_size, |
| scoped_ptr<gpu::MailboxHolder>(), |
| timestamp, |
| false)); |
| frame->shared_memory_handle_ = handle; |
| frame->strides_[kYPlane] = new_coded_size.width(); |
| frame->strides_[kUPlane] = new_coded_size.width() / 2; |
| frame->strides_[kVPlane] = new_coded_size.width() / 2; |
| frame->data_[kYPlane] = data; |
| frame->data_[kUPlane] = data + new_coded_size.GetArea(); |
| frame->data_[kVPlane] = data + (new_coded_size.GetArea() * 5 / 4); |
| frame->no_longer_needed_cb_ = no_longer_needed_cb; |
| return frame; |
| } |
| default: |
| NOTIMPLEMENTED(); |
| return NULL; |
| } |
| } |
| |
| #if defined(OS_POSIX) |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::WrapExternalDmabufs( |
| Format format, |
| const gfx::Size& coded_size, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size, |
| const std::vector<int> dmabuf_fds, |
| base::TimeDelta timestamp, |
| const base::Closure& no_longer_needed_cb) { |
| if (!IsValidConfig(format, coded_size, visible_rect, natural_size)) |
| return NULL; |
| |
| if (dmabuf_fds.size() != NumPlanes(format)) { |
| LOG(FATAL) << "Not enough dmabuf fds provided!"; |
| return NULL; |
| } |
| |
| scoped_refptr<VideoFrame> frame( |
| new VideoFrame(format, |
| coded_size, |
| visible_rect, |
| natural_size, |
| scoped_ptr<gpu::MailboxHolder>(), |
| timestamp, |
| false)); |
| |
| for (size_t i = 0; i < dmabuf_fds.size(); ++i) { |
| int duped_fd = HANDLE_EINTR(dup(dmabuf_fds[i])); |
| if (duped_fd == -1) { |
| // The already-duped in previous iterations fds will be closed when |
| // the partially-created frame drops out of scope here. |
| DLOG(ERROR) << "Failed duplicating a dmabuf fd"; |
| return NULL; |
| } |
| |
| frame->dmabuf_fds_[i].reset(duped_fd); |
| // Data is accessible only via fds. |
| frame->data_[i] = NULL; |
| frame->strides_[i] = 0; |
| } |
| |
| frame->no_longer_needed_cb_ = no_longer_needed_cb; |
| return frame; |
| } |
| #endif |
| |
| #if defined(OS_MACOSX) |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::WrapCVPixelBuffer( |
| CVPixelBufferRef cv_pixel_buffer, |
| base::TimeDelta timestamp) { |
| DCHECK(cv_pixel_buffer); |
| DCHECK(CFGetTypeID(cv_pixel_buffer) == CVPixelBufferGetTypeID()); |
| |
| const OSType cv_format = CVPixelBufferGetPixelFormatType(cv_pixel_buffer); |
| Format format; |
| // There are very few compatible CV pixel formats, so just check each. |
| if (cv_format == kCVPixelFormatType_420YpCbCr8Planar) { |
| format = Format::I420; |
| } else if (cv_format == kCVPixelFormatType_444YpCbCr8) { |
| format = Format::YV24; |
| } else if (cv_format == '420v') { |
| // TODO(jfroy): Use kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange when the |
| // minimum OS X and iOS SDKs permits it. |
| format = Format::NV12; |
| } else { |
| DLOG(ERROR) << "CVPixelBuffer format not supported: " << cv_format; |
| return NULL; |
| } |
| |
| const gfx::Size coded_size(CVImageBufferGetEncodedSize(cv_pixel_buffer)); |
| const gfx::Rect visible_rect(CVImageBufferGetCleanRect(cv_pixel_buffer)); |
| const gfx::Size natural_size(CVImageBufferGetDisplaySize(cv_pixel_buffer)); |
| |
| if (!IsValidConfig(format, coded_size, visible_rect, natural_size)) |
| return NULL; |
| |
| scoped_refptr<VideoFrame> frame( |
| new VideoFrame(format, |
| coded_size, |
| visible_rect, |
| natural_size, |
| scoped_ptr<gpu::MailboxHolder>(), |
| timestamp, |
| false)); |
| |
| frame->cv_pixel_buffer_.reset(cv_pixel_buffer, base::scoped_policy::RETAIN); |
| return frame; |
| } |
| #endif |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::WrapExternalYuvData( |
| Format format, |
| const gfx::Size& coded_size, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size, |
| int32 y_stride, |
| int32 u_stride, |
| int32 v_stride, |
| uint8* y_data, |
| uint8* u_data, |
| uint8* v_data, |
| base::TimeDelta timestamp, |
| const base::Closure& no_longer_needed_cb) { |
| const gfx::Size new_coded_size = AdjustCodedSize(format, coded_size); |
| CHECK(IsValidConfig(format, new_coded_size, visible_rect, natural_size)); |
| |
| scoped_refptr<VideoFrame> frame( |
| new VideoFrame(format, |
| new_coded_size, |
| visible_rect, |
| natural_size, |
| scoped_ptr<gpu::MailboxHolder>(), |
| timestamp, |
| false)); |
| frame->strides_[kYPlane] = y_stride; |
| frame->strides_[kUPlane] = u_stride; |
| frame->strides_[kVPlane] = v_stride; |
| frame->data_[kYPlane] = y_data; |
| frame->data_[kUPlane] = u_data; |
| frame->data_[kVPlane] = v_data; |
| frame->no_longer_needed_cb_ = no_longer_needed_cb; |
| return frame; |
| } |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::WrapVideoFrame( |
| const scoped_refptr<VideoFrame>& frame, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size, |
| const base::Closure& no_longer_needed_cb) { |
| // NATIVE_TEXTURE frames need mailbox info propagated, and there's no support |
| // for that here yet, see http://crbug/362521. |
| CHECK_NE(frame->format(), NATIVE_TEXTURE); |
| |
| DCHECK(frame->visible_rect().Contains(visible_rect)); |
| scoped_refptr<VideoFrame> wrapped_frame( |
| new VideoFrame(frame->format(), |
| frame->coded_size(), |
| visible_rect, |
| natural_size, |
| scoped_ptr<gpu::MailboxHolder>(), |
| frame->timestamp(), |
| frame->end_of_stream())); |
| |
| for (size_t i = 0; i < NumPlanes(frame->format()); ++i) { |
| wrapped_frame->strides_[i] = frame->stride(i); |
| wrapped_frame->data_[i] = frame->data(i); |
| } |
| |
| wrapped_frame->no_longer_needed_cb_ = no_longer_needed_cb; |
| return wrapped_frame; |
| } |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::CreateEOSFrame() { |
| return new VideoFrame(VideoFrame::UNKNOWN, |
| gfx::Size(), |
| gfx::Rect(), |
| gfx::Size(), |
| scoped_ptr<gpu::MailboxHolder>(), |
| kNoTimestamp(), |
| true); |
| } |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::CreateColorFrame( |
| const gfx::Size& size, |
| uint8 y, uint8 u, uint8 v, |
| base::TimeDelta timestamp) { |
| scoped_refptr<VideoFrame> frame = VideoFrame::CreateFrame( |
| VideoFrame::YV12, size, gfx::Rect(size), size, timestamp); |
| FillYUV(frame.get(), y, u, v); |
| return frame; |
| } |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::CreateBlackFrame(const gfx::Size& size) { |
| const uint8 kBlackY = 0x00; |
| const uint8 kBlackUV = 0x80; |
| const base::TimeDelta kZero; |
| return CreateColorFrame(size, kBlackY, kBlackUV, kBlackUV, kZero); |
| } |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::CreateTransparentFrame( |
| const gfx::Size& size) { |
| const uint8 kBlackY = 0x00; |
| const uint8 kBlackUV = 0x00; |
| const uint8 kTransparentA = 0x00; |
| const base::TimeDelta kZero; |
| scoped_refptr<VideoFrame> frame = VideoFrame::CreateFrame( |
| VideoFrame::YV12A, size, gfx::Rect(size), size, kZero); |
| FillYUVA(frame.get(), kBlackY, kBlackUV, kBlackUV, kTransparentA); |
| return frame; |
| } |
| |
| #if defined(VIDEO_HOLE) |
| // This block and other blocks wrapped around #if defined(VIDEO_HOLE) is not |
| // maintained by the general compositor team. Please contact the following |
| // people instead: |
| // |
| // wonsik@chromium.org |
| // ycheo@chromium.org |
| |
| // static |
| scoped_refptr<VideoFrame> VideoFrame::CreateHoleFrame( |
| const gfx::Size& size) { |
| DCHECK(IsValidConfig(VideoFrame::HOLE, size, gfx::Rect(size), size)); |
| scoped_refptr<VideoFrame> frame( |
| new VideoFrame(VideoFrame::HOLE, |
| size, |
| gfx::Rect(size), |
| size, |
| scoped_ptr<gpu::MailboxHolder>(), |
| base::TimeDelta(), |
| false)); |
| return frame; |
| } |
| #endif // defined(VIDEO_HOLE) |
| |
| // static |
| size_t VideoFrame::NumPlanes(Format format) { |
| switch (format) { |
| case VideoFrame::NATIVE_TEXTURE: |
| #if defined(VIDEO_HOLE) |
| case VideoFrame::HOLE: |
| #endif // defined(VIDEO_HOLE) |
| return 0; |
| case VideoFrame::NV12: |
| return 2; |
| case VideoFrame::YV12: |
| case VideoFrame::YV16: |
| case VideoFrame::I420: |
| case VideoFrame::YV12J: |
| case VideoFrame::YV24: |
| return 3; |
| case VideoFrame::YV12A: |
| return 4; |
| case VideoFrame::UNKNOWN: |
| break; |
| } |
| NOTREACHED() << "Unsupported video frame format: " << format; |
| return 0; |
| } |
| |
| |
| // static |
| size_t VideoFrame::AllocationSize(Format format, const gfx::Size& coded_size) { |
| size_t total = 0; |
| for (size_t i = 0; i < NumPlanes(format); ++i) |
| total += PlaneAllocationSize(format, i, coded_size); |
| return total; |
| } |
| |
| // static |
| gfx::Size VideoFrame::PlaneSize(Format format, |
| size_t plane, |
| const gfx::Size& coded_size) { |
| DCHECK(IsValidPlane(plane, format)); |
| |
| // Align to multiple-of-two size overall. This ensures that non-subsampled |
| // planes can be addressed by pixel with the same scaling as the subsampled |
| // planes. |
| const int width = RoundUp(coded_size.width(), 2); |
| const int height = RoundUp(coded_size.height(), 2); |
| |
| const gfx::Size subsample = SampleSize(format, plane); |
| DCHECK(width % subsample.width() == 0); |
| DCHECK(height % subsample.height() == 0); |
| return gfx::Size(BytesPerElement(format, plane) * width / subsample.width(), |
| height / subsample.height()); |
| } |
| |
| size_t VideoFrame::PlaneAllocationSize(Format format, |
| size_t plane, |
| const gfx::Size& coded_size) { |
| // VideoFrame formats are (so far) all YUV and 1 byte per sample. |
| return PlaneSize(format, plane, coded_size).GetArea(); |
| } |
| |
| // static |
| int VideoFrame::PlaneHorizontalBitsPerPixel(Format format, size_t plane) { |
| DCHECK(IsValidPlane(plane, format)); |
| const int bits_per_element = 8 * BytesPerElement(format, plane); |
| const int pixels_per_element = SampleSize(format, plane).GetArea(); |
| DCHECK(bits_per_element % pixels_per_element == 0); |
| return bits_per_element / pixels_per_element; |
| } |
| |
| // Release data allocated by AllocateYUV(). |
| static void ReleaseData(uint8* data) { |
| DCHECK(data); |
| base::AlignedFree(data); |
| } |
| |
| void VideoFrame::AllocateYUV() { |
| DCHECK(format_ == YV12 || format_ == YV16 || format_ == YV12A || |
| format_ == I420 || format_ == YV12J || format_ == YV24); |
| COMPILE_ASSERT(0 == kYPlane, y_plane_data_must_be_index_0); |
| |
| size_t data_size = 0; |
| size_t offset[kMaxPlanes]; |
| for (size_t plane = 0; plane < VideoFrame::NumPlanes(format_); ++plane) { |
| // The *2 in alignment for height is because some formats (e.g. h264) allow |
| // interlaced coding, and then the size needs to be a multiple of two |
| // macroblocks (vertically). See |
| // libavcodec/utils.c:avcodec_align_dimensions2(). |
| const size_t height = RoundUp(rows(plane), kFrameSizeAlignment * 2); |
| strides_[plane] = RoundUp(row_bytes(plane), kFrameSizeAlignment); |
| offset[plane] = data_size; |
| data_size += height * strides_[plane]; |
| } |
| |
| // The extra line of UV being allocated is because h264 chroma MC |
| // overreads by one line in some cases, see libavcodec/utils.c: |
| // avcodec_align_dimensions2() and libavcodec/x86/h264_chromamc.asm: |
| // put_h264_chroma_mc4_ssse3(). |
| DCHECK(IsValidPlane(kUPlane, format_)); |
| data_size += strides_[kUPlane] + kFrameSizePadding; |
| |
| // FFmpeg expects the initialize allocation to be zero-initialized. Failure |
| // to do so can lead to unitialized value usage. See http://crbug.com/390941 |
| uint8* data = reinterpret_cast<uint8*>( |
| base::AlignedAlloc(data_size, kFrameAddressAlignment)); |
| memset(data, 0, data_size); |
| |
| for (size_t plane = 0; plane < VideoFrame::NumPlanes(format_); ++plane) |
| data_[plane] = data + offset[plane]; |
| |
| no_longer_needed_cb_ = base::Bind(&ReleaseData, data); |
| } |
| |
| VideoFrame::VideoFrame(VideoFrame::Format format, |
| const gfx::Size& coded_size, |
| const gfx::Rect& visible_rect, |
| const gfx::Size& natural_size, |
| scoped_ptr<gpu::MailboxHolder> mailbox_holder, |
| base::TimeDelta timestamp, |
| bool end_of_stream) |
| : format_(format), |
| coded_size_(coded_size), |
| visible_rect_(visible_rect), |
| natural_size_(natural_size), |
| mailbox_holder_(mailbox_holder.Pass()), |
| shared_memory_handle_(base::SharedMemory::NULLHandle()), |
| timestamp_(timestamp), |
| release_sync_point_(0), |
| end_of_stream_(end_of_stream) { |
| DCHECK(IsValidConfig(format_, coded_size_, visible_rect_, natural_size_)); |
| |
| memset(&strides_, 0, sizeof(strides_)); |
| memset(&data_, 0, sizeof(data_)); |
| } |
| |
| VideoFrame::~VideoFrame() { |
| if (!mailbox_holder_release_cb_.is_null()) { |
| uint32 release_sync_point; |
| { |
| // To ensure that changes to |release_sync_point_| are visible on this |
| // thread (imply a memory barrier). |
| base::AutoLock locker(release_sync_point_lock_); |
| release_sync_point = release_sync_point_; |
| } |
| base::ResetAndReturn(&mailbox_holder_release_cb_).Run(release_sync_point); |
| } |
| if (!no_longer_needed_cb_.is_null()) |
| base::ResetAndReturn(&no_longer_needed_cb_).Run(); |
| } |
| |
| // static |
| bool VideoFrame::IsValidPlane(size_t plane, VideoFrame::Format format) { |
| return (plane < NumPlanes(format)); |
| } |
| |
| int VideoFrame::stride(size_t plane) const { |
| DCHECK(IsValidPlane(plane, format_)); |
| return strides_[plane]; |
| } |
| |
| // static |
| size_t VideoFrame::RowBytes(size_t plane, |
| VideoFrame::Format format, |
| int width) { |
| DCHECK(IsValidPlane(plane, format)); |
| return BytesPerElement(format, plane) * Columns(plane, format, width); |
| } |
| |
| int VideoFrame::row_bytes(size_t plane) const { |
| return RowBytes(plane, format_, coded_size_.width()); |
| } |
| |
| // static |
| size_t VideoFrame::Rows(size_t plane, VideoFrame::Format format, int height) { |
| DCHECK(IsValidPlane(plane, format)); |
| const int sample_height = SampleSize(format, plane).height(); |
| return RoundUp(height, sample_height) / sample_height; |
| } |
| |
| // static |
| size_t VideoFrame::Columns(size_t plane, Format format, int width) { |
| DCHECK(IsValidPlane(plane, format)); |
| const int sample_width = SampleSize(format, plane).width(); |
| return RoundUp(width, sample_width) / sample_width; |
| } |
| |
| int VideoFrame::rows(size_t plane) const { |
| return Rows(plane, format_, coded_size_.height()); |
| } |
| |
| const uint8* VideoFrame::data(size_t plane) const { |
| DCHECK(IsValidPlane(plane, format_)); |
| return data_[plane]; |
| } |
| |
| uint8* VideoFrame::data(size_t plane) { |
| DCHECK(IsValidPlane(plane, format_)); |
| return data_[plane]; |
| } |
| |
| const uint8* VideoFrame::visible_data(size_t plane) const { |
| DCHECK(IsValidPlane(plane, format_)); |
| |
| // Calculate an offset that is properly aligned for all planes. |
| const gfx::Size alignment = CommonAlignment(format_); |
| const gfx::Point offset(RoundDown(visible_rect_.x(), alignment.width()), |
| RoundDown(visible_rect_.y(), alignment.height())); |
| |
| const gfx::Size subsample = SampleSize(format_, plane); |
| DCHECK(offset.x() % subsample.width() == 0); |
| DCHECK(offset.y() % subsample.height() == 0); |
| return data(plane) + |
| stride(plane) * (offset.y() / subsample.height()) + // Row offset. |
| BytesPerElement(format_, plane) * // Column offset. |
| (offset.x() / subsample.width()); |
| } |
| |
| uint8* VideoFrame::visible_data(size_t plane) { |
| return const_cast<uint8*>( |
| static_cast<const VideoFrame*>(this)->visible_data(plane)); |
| } |
| |
| const gpu::MailboxHolder* VideoFrame::mailbox_holder() const { |
| DCHECK_EQ(format_, NATIVE_TEXTURE); |
| return mailbox_holder_.get(); |
| } |
| |
| base::SharedMemoryHandle VideoFrame::shared_memory_handle() const { |
| return shared_memory_handle_; |
| } |
| |
| void VideoFrame::UpdateReleaseSyncPoint(SyncPointClient* client) { |
| DCHECK_EQ(format_, NATIVE_TEXTURE); |
| base::AutoLock locker(release_sync_point_lock_); |
| // Must wait on the previous sync point before inserting a new sync point so |
| // that |mailbox_holder_release_cb_| guarantees the previous sync point |
| // occurred when it waits on |release_sync_point_|. |
| if (release_sync_point_) |
| client->WaitSyncPoint(release_sync_point_); |
| release_sync_point_ = client->InsertSyncPoint(); |
| } |
| |
| #if defined(OS_POSIX) |
| int VideoFrame::dmabuf_fd(size_t plane) const { |
| return dmabuf_fds_[plane].get(); |
| } |
| #endif |
| |
| #if defined(OS_MACOSX) |
| CVPixelBufferRef VideoFrame::cv_pixel_buffer() const { |
| return cv_pixel_buffer_.get(); |
| } |
| #endif |
| |
| void VideoFrame::HashFrameForTesting(base::MD5Context* context) { |
| for (size_t plane = 0; plane < NumPlanes(format_); ++plane) { |
| for (int row = 0; row < rows(plane); ++row) { |
| base::MD5Update(context, base::StringPiece( |
| reinterpret_cast<char*>(data(plane) + stride(plane) * row), |
| row_bytes(plane))); |
| } |
| } |
| } |
| |
| } // namespace media |