blob: a19726cb26478c6e8578aacaa7b8fa554fa94170 [file] [log] [blame]
/*
* Copyright (C) 2012 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package android.media;
import android.annotation.NonNull;
import android.annotation.Nullable;
import android.util.Log;
import android.util.Pair;
import android.util.Range;
import android.util.Rational;
import android.util.Size;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.HashMap;
import java.util.Map;
import java.util.Set;
import static android.media.Utils.intersectSortedDistinctRanges;
import static android.media.Utils.sortDistinctRanges;
/**
* Provides information about a given media codec available on the device. You can
* iterate through all codecs available by querying {@link MediaCodecList}. For example,
* here's how to find an encoder that supports a given MIME type:
* <pre>
* private static MediaCodecInfo selectCodec(String mimeType) {
* int numCodecs = MediaCodecList.getCodecCount();
* for (int i = 0; i &lt; numCodecs; i++) {
* MediaCodecInfo codecInfo = MediaCodecList.getCodecInfoAt(i);
*
* if (!codecInfo.isEncoder()) {
* continue;
* }
*
* String[] types = codecInfo.getSupportedTypes();
* for (int j = 0; j &lt; types.length; j++) {
* if (types[j].equalsIgnoreCase(mimeType)) {
* return codecInfo;
* }
* }
* }
* return null;
* }</pre>
*
*/
public final class MediaCodecInfo {
private boolean mIsEncoder;
private String mName;
private Map<String, CodecCapabilities> mCaps;
/* package private */ MediaCodecInfo(
String name, boolean isEncoder, CodecCapabilities[] caps) {
mName = name;
mIsEncoder = isEncoder;
mCaps = new HashMap<String, CodecCapabilities>();
for (CodecCapabilities c: caps) {
mCaps.put(c.getMimeType(), c);
}
}
/**
* Retrieve the codec name.
*/
public final String getName() {
return mName;
}
/**
* Query if the codec is an encoder.
*/
public final boolean isEncoder() {
return mIsEncoder;
}
/**
* Query the media types supported by the codec.
*/
public final String[] getSupportedTypes() {
Set<String> typeSet = mCaps.keySet();
String[] types = typeSet.toArray(new String[typeSet.size()]);
Arrays.sort(types);
return types;
}
private static int checkPowerOfTwo(int value, String message) {
if ((value & (value - 1)) != 0) {
throw new IllegalArgumentException(message);
}
return value;
}
private static class Feature {
public String mName;
public int mValue;
public boolean mDefault;
public Feature(String name, int value, boolean def) {
mName = name;
mValue = value;
mDefault = def;
}
}
// COMMON CONSTANTS
private static final Range<Integer> POSITIVE_INTEGERS =
Range.create(1, Integer.MAX_VALUE);
private static final Range<Long> POSITIVE_LONGS =
Range.create(1l, Long.MAX_VALUE);
private static final Range<Rational> POSITIVE_RATIONALS =
Range.create(new Rational(1, Integer.MAX_VALUE),
new Rational(Integer.MAX_VALUE, 1));
private static final Range<Integer> SIZE_RANGE = Range.create(1, 32768);
private static final Range<Integer> FRAME_RATE_RANGE = Range.create(0, 960);
private static final Range<Integer> BITRATE_RANGE = Range.create(0, 500000000);
private static final int DEFAULT_MAX_SUPPORTED_INSTANCES = 32;
private static final int MAX_SUPPORTED_INSTANCES_LIMIT = 256;
// found stuff that is not supported by framework (=> this should not happen)
private static final int ERROR_UNRECOGNIZED = (1 << 0);
// found profile/level for which we don't have capability estimates
private static final int ERROR_UNSUPPORTED = (1 << 1);
// have not found any profile/level for which we don't have capability estimate
private static final int ERROR_NONE_SUPPORTED = (1 << 2);
/**
* Encapsulates the capabilities of a given codec component.
* For example, what profile/level combinations it supports and what colorspaces
* it is capable of providing the decoded data in, as well as some
* codec-type specific capability flags.
* <p>You can get an instance for a given {@link MediaCodecInfo} object with
* {@link MediaCodecInfo#getCapabilitiesForType getCapabilitiesForType()}, passing a MIME type.
*/
public static final class CodecCapabilities {
public CodecCapabilities() {
}
// CLASSIFICATION
private String mMime;
private int mMaxSupportedInstances;
// LEGACY FIELDS
// Enumerates supported profile/level combinations as defined
// by the type of encoded data. These combinations impose restrictions
// on video resolution, bitrate... and limit the available encoder tools
// such as B-frame support, arithmetic coding...
public CodecProfileLevel[] profileLevels; // NOTE this array is modifiable by user
// from OMX_COLOR_FORMATTYPE
/** @deprecated Use {@link #COLOR_Format24bitBGR888}. */
public static final int COLOR_FormatMonochrome = 1;
/** @deprecated Use {@link #COLOR_Format24bitBGR888}. */
public static final int COLOR_Format8bitRGB332 = 2;
/** @deprecated Use {@link #COLOR_Format24bitBGR888}. */
public static final int COLOR_Format12bitRGB444 = 3;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format16bitARGB4444 = 4;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format16bitARGB1555 = 5;
/**
* 16 bits per pixel RGB color format, with 5-bit red & blue and 6-bit green component.
* <p>
* Using 16-bit little-endian representation, colors stored as Red 15:11, Green 10:5, Blue 4:0.
* <pre>
* byte byte
* <--------- i --------> | <------ i + 1 ------>
* +--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+
* | BLUE | GREEN | RED |
* +--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+
* 0 4 5 7 0 2 3 7
* bit
* </pre>
*
* This format corresponds to {@link android.graphics.PixelFormat#RGB_565} and
* {@link android.graphics.ImageFormat#RGB_565}.
*/
public static final int COLOR_Format16bitRGB565 = 6;
/** @deprecated Use {@link #COLOR_Format16bitRGB565}. */
public static final int COLOR_Format16bitBGR565 = 7;
/** @deprecated Use {@link #COLOR_Format24bitBGR888}. */
public static final int COLOR_Format18bitRGB666 = 8;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format18bitARGB1665 = 9;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format19bitARGB1666 = 10;
/** @deprecated Use {@link #COLOR_Format24bitBGR888} or {@link #COLOR_FormatRGBFlexible}. */
public static final int COLOR_Format24bitRGB888 = 11;
/**
* 24 bits per pixel RGB color format, with 8-bit red, green & blue components.
* <p>
* Using 24-bit little-endian representation, colors stored as Red 7:0, Green 15:8, Blue 23:16.
* <pre>
* byte byte byte
* <------ i -----> | <---- i+1 ----> | <---- i+2 ----->
* +-----------------+-----------------+-----------------+
* | RED | GREEN | BLUE |
* +-----------------+-----------------+-----------------+
* </pre>
*
* This format corresponds to {@link android.graphics.PixelFormat#RGB_888}, and can also be
* represented as a flexible format by {@link #COLOR_FormatRGBFlexible}.
*/
public static final int COLOR_Format24bitBGR888 = 12;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format24bitARGB1887 = 13;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format25bitARGB1888 = 14;
/**
* @deprecated Use {@link #COLOR_Format32bitABGR8888} Or {@link #COLOR_FormatRGBAFlexible}.
*/
public static final int COLOR_Format32bitBGRA8888 = 15;
/**
* @deprecated Use {@link #COLOR_Format32bitABGR8888} Or {@link #COLOR_FormatRGBAFlexible}.
*/
public static final int COLOR_Format32bitARGB8888 = 16;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_FormatYUV411Planar = 17;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_FormatYUV411PackedPlanar = 18;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_FormatYUV420Planar = 19;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_FormatYUV420PackedPlanar = 20;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_FormatYUV420SemiPlanar = 21;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatYUV422Planar = 22;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatYUV422PackedPlanar = 23;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatYUV422SemiPlanar = 24;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatYCbYCr = 25;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatYCrYCb = 26;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatCbYCrY = 27;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatCrYCbY = 28;
/** @deprecated Use {@link #COLOR_FormatYUV444Flexible}. */
public static final int COLOR_FormatYUV444Interleaved = 29;
/**
* SMIA 8-bit Bayer format.
* Each byte represents the top 8-bits of a 10-bit signal.
*/
public static final int COLOR_FormatRawBayer8bit = 30;
/**
* SMIA 10-bit Bayer format.
*/
public static final int COLOR_FormatRawBayer10bit = 31;
/**
* SMIA 8-bit compressed Bayer format.
* Each byte represents a sample from the 10-bit signal that is compressed into 8-bits
* using DPCM/PCM compression, as defined by the SMIA Functional Specification.
*/
public static final int COLOR_FormatRawBayer8bitcompressed = 32;
/** @deprecated Use {@link #COLOR_FormatL8}. */
public static final int COLOR_FormatL2 = 33;
/** @deprecated Use {@link #COLOR_FormatL8}. */
public static final int COLOR_FormatL4 = 34;
/**
* 8 bits per pixel Y color format.
* <p>
* Each byte contains a single pixel.
* This format corresponds to {@link android.graphics.PixelFormat#L_8}.
*/
public static final int COLOR_FormatL8 = 35;
/**
* 16 bits per pixel, little-endian Y color format.
* <p>
* <pre>
* byte byte
* <--------- i --------> | <------ i + 1 ------>
* +--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+
* | Y |
* +--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+--+
* 0 7 0 7
* bit
* </pre>
*/
public static final int COLOR_FormatL16 = 36;
/** @deprecated Use {@link #COLOR_FormatL16}. */
public static final int COLOR_FormatL24 = 37;
/**
* 32 bits per pixel, little-endian Y color format.
* <p>
* <pre>
* byte byte byte byte
* <------ i -----> | <---- i+1 ----> | <---- i+2 ----> | <---- i+3 ----->
* +-----------------+-----------------+-----------------+-----------------+
* | Y |
* +-----------------+-----------------+-----------------+-----------------+
* 0 7 0 7 0 7 0 7
* bit
* </pre>
*
* @deprecated Use {@link #COLOR_FormatL16}.
*/
public static final int COLOR_FormatL32 = 38;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_FormatYUV420PackedSemiPlanar = 39;
/** @deprecated Use {@link #COLOR_FormatYUV422Flexible}. */
public static final int COLOR_FormatYUV422PackedSemiPlanar = 40;
/** @deprecated Use {@link #COLOR_Format24bitBGR888}. */
public static final int COLOR_Format18BitBGR666 = 41;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format24BitARGB6666 = 42;
/** @deprecated Use {@link #COLOR_Format32bitABGR8888}. */
public static final int COLOR_Format24BitABGR6666 = 43;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_TI_FormatYUV420PackedSemiPlanar = 0x7f000100;
// COLOR_FormatSurface indicates that the data will be a GraphicBuffer metadata reference.
// In OMX this is called OMX_COLOR_FormatAndroidOpaque.
public static final int COLOR_FormatSurface = 0x7F000789;
/**
* 32 bits per pixel RGBA color format, with 8-bit red, green, blue, and alpha components.
* <p>
* Using 32-bit little-endian representation, colors stored as Red 7:0, Green 15:8,
* Blue 23:16, and Alpha 31:24.
* <pre>
* byte byte byte byte
* <------ i -----> | <---- i+1 ----> | <---- i+2 ----> | <---- i+3 ----->
* +-----------------+-----------------+-----------------+-----------------+
* | RED | GREEN | BLUE | ALPHA |
* +-----------------+-----------------+-----------------+-----------------+
* </pre>
*
* This corresponds to {@link android.graphics.PixelFormat#RGBA_8888}.
*/
public static final int COLOR_Format32bitABGR8888 = 0x7F00A000;
/**
* Flexible 12 bits per pixel, subsampled YUV color format with 8-bit chroma and luma
* components.
* <p>
* Chroma planes are subsampled by 2 both horizontally and vertically.
* Use this format with {@link Image}.
* This format corresponds to {@link android.graphics.ImageFormat#YUV_420_888},
* and can represent the {@link #COLOR_FormatYUV411Planar},
* {@link #COLOR_FormatYUV411PackedPlanar}, {@link #COLOR_FormatYUV420Planar},
* {@link #COLOR_FormatYUV420PackedPlanar}, {@link #COLOR_FormatYUV420SemiPlanar}
* and {@link #COLOR_FormatYUV420PackedSemiPlanar} formats.
*
* @see Image#getFormat
*/
public static final int COLOR_FormatYUV420Flexible = 0x7F420888;
/**
* Flexible 16 bits per pixel, subsampled YUV color format with 8-bit chroma and luma
* components.
* <p>
* Chroma planes are horizontally subsampled by 2. Use this format with {@link Image}.
* This format corresponds to {@link android.graphics.ImageFormat#YUV_422_888},
* and can represent the {@link #COLOR_FormatYCbYCr}, {@link #COLOR_FormatYCrYCb},
* {@link #COLOR_FormatCbYCrY}, {@link #COLOR_FormatCrYCbY},
* {@link #COLOR_FormatYUV422Planar}, {@link #COLOR_FormatYUV422PackedPlanar},
* {@link #COLOR_FormatYUV422SemiPlanar} and {@link #COLOR_FormatYUV422PackedSemiPlanar}
* formats.
*
* @see Image#getFormat
*/
public static final int COLOR_FormatYUV422Flexible = 0x7F422888;
/**
* Flexible 24 bits per pixel YUV color format with 8-bit chroma and luma
* components.
* <p>
* Chroma planes are not subsampled. Use this format with {@link Image}.
* This format corresponds to {@link android.graphics.ImageFormat#YUV_444_888},
* and can represent the {@link #COLOR_FormatYUV444Interleaved} format.
* @see Image#getFormat
*/
public static final int COLOR_FormatYUV444Flexible = 0x7F444888;
/**
* Flexible 24 bits per pixel RGB color format with 8-bit red, green and blue
* components.
* <p>
* Use this format with {@link Image}. This format corresponds to
* {@link android.graphics.ImageFormat#FLEX_RGB_888}, and can represent
* {@link #COLOR_Format24bitBGR888} and {@link #COLOR_Format24bitRGB888} formats.
* @see Image#getFormat.
*/
public static final int COLOR_FormatRGBFlexible = 0x7F36B888;
/**
* Flexible 32 bits per pixel RGBA color format with 8-bit red, green, blue, and alpha
* components.
* <p>
* Use this format with {@link Image}. This format corresponds to
* {@link android.graphics.ImageFormat#FLEX_RGBA_8888}, and can represent
* {@link #COLOR_Format32bitBGRA8888}, {@link #COLOR_Format32bitABGR8888} and
* {@link #COLOR_Format32bitARGB8888} formats.
*
* @see Image#getFormat
*/
public static final int COLOR_FormatRGBAFlexible = 0x7F36A888;
/** @deprecated Use {@link #COLOR_FormatYUV420Flexible}. */
public static final int COLOR_QCOM_FormatYUV420SemiPlanar = 0x7fa30c00;
/**
* Defined in the OpenMAX IL specs, color format values are drawn from
* OMX_COLOR_FORMATTYPE.
*/
public int[] colorFormats; // NOTE this array is modifiable by user
// FEATURES
private int mFlagsSupported;
private int mFlagsRequired;
private int mFlagsVerified;
/**
* <b>video decoder only</b>: codec supports seamless resolution changes.
*/
public static final String FEATURE_AdaptivePlayback = "adaptive-playback";
/**
* <b>video decoder only</b>: codec supports secure decryption.
*/
public static final String FEATURE_SecurePlayback = "secure-playback";
/**
* <b>video or audio decoder only</b>: codec supports tunneled playback.
*/
public static final String FEATURE_TunneledPlayback = "tunneled-playback";
/**
* <b>video decoder only</b>: codec supports queuing partial frames.
*/
public static final String FEATURE_PartialFrame = "partial-frame";
/**
* <b>video encoder only</b>: codec supports intra refresh.
*/
public static final String FEATURE_IntraRefresh = "intra-refresh";
/**
* Query codec feature capabilities.
* <p>
* These features are supported to be used by the codec. These
* include optional features that can be turned on, as well as
* features that are always on.
*/
public final boolean isFeatureSupported(String name) {
return checkFeature(name, mFlagsSupported);
}
/**
* Query codec feature requirements.
* <p>
* These features are required to be used by the codec, and as such,
* they are always turned on.
*/
public final boolean isFeatureRequired(String name) {
return checkFeature(name, mFlagsRequired);
}
private static final Feature[] decoderFeatures = {
new Feature(FEATURE_AdaptivePlayback, (1 << 0), true),
new Feature(FEATURE_SecurePlayback, (1 << 1), false),
new Feature(FEATURE_TunneledPlayback, (1 << 2), false),
new Feature(FEATURE_PartialFrame, (1 << 3), false),
};
private static final Feature[] encoderFeatures = {
new Feature(FEATURE_IntraRefresh, (1 << 0), false),
};
/** @hide */
public String[] validFeatures() {
Feature[] features = getValidFeatures();
String[] res = new String[features.length];
for (int i = 0; i < res.length; i++) {
res[i] = features[i].mName;
}
return res;
}
private Feature[] getValidFeatures() {
if (!isEncoder()) {
return decoderFeatures;
}
return encoderFeatures;
}
private boolean checkFeature(String name, int flags) {
for (Feature feat: getValidFeatures()) {
if (feat.mName.equals(name)) {
return (flags & feat.mValue) != 0;
}
}
return false;
}
/** @hide */
public boolean isRegular() {
// regular codecs only require default features
for (Feature feat: getValidFeatures()) {
if (!feat.mDefault && isFeatureRequired(feat.mName)) {
return false;
}
}
return true;
}
/**
* Query whether codec supports a given {@link MediaFormat}.
*
* <p class=note>
* <strong>Note:</strong> On {@link android.os.Build.VERSION_CODES#LOLLIPOP},
* {@code format} must not contain a {@linkplain MediaFormat#KEY_FRAME_RATE
* frame rate}. Use
* <code class=prettyprint>format.setString(MediaFormat.KEY_FRAME_RATE, null)</code>
* to clear any existing frame rate setting in the format.
* <p>
*
* The following table summarizes the format keys considered by this method.
*
* <table style="width: 0%">
* <thead>
* <tr>
* <th rowspan=3>OS Version(s)</th>
* <td colspan=3>{@code MediaFormat} keys considered for</th>
* </tr><tr>
* <th>Audio Codecs</th>
* <th>Video Codecs</th>
* <th>Encoders</th>
* </tr>
* </thead>
* <tbody>
* <tr>
* <td>{@link android.os.Build.VERSION_CODES#LOLLIPOP}</th>
* <td rowspan=3>{@link MediaFormat#KEY_MIME}<sup>*</sup>,<br>
* {@link MediaFormat#KEY_SAMPLE_RATE},<br>
* {@link MediaFormat#KEY_CHANNEL_COUNT},</td>
* <td>{@link MediaFormat#KEY_MIME}<sup>*</sup>,<br>
* {@link CodecCapabilities#FEATURE_AdaptivePlayback}<sup>D</sup>,<br>
* {@link CodecCapabilities#FEATURE_SecurePlayback}<sup>D</sup>,<br>
* {@link CodecCapabilities#FEATURE_TunneledPlayback}<sup>D</sup>,<br>
* {@link MediaFormat#KEY_WIDTH},<br>
* {@link MediaFormat#KEY_HEIGHT},<br>
* <strong>no</strong> {@code KEY_FRAME_RATE}</td>
* <td rowspan=4>{@link MediaFormat#KEY_BITRATE_MODE},<br>
* {@link MediaFormat#KEY_PROFILE}
* (and/or {@link MediaFormat#KEY_AAC_PROFILE}<sup>~</sup>),<br>
* <!-- {link MediaFormat#KEY_QUALITY},<br> -->
* {@link MediaFormat#KEY_COMPLEXITY}
* (and/or {@link MediaFormat#KEY_FLAC_COMPRESSION_LEVEL}<sup>~</sup>)</td>
* </tr><tr>
* <td>{@link android.os.Build.VERSION_CODES#LOLLIPOP_MR1}</th>
* <td rowspan=2>as above, plus<br>
* {@link MediaFormat#KEY_FRAME_RATE}</td>
* </tr><tr>
* <td>{@link android.os.Build.VERSION_CODES#M}</th>
* </tr><tr>
* <td>{@link android.os.Build.VERSION_CODES#N}</th>
* <td>as above, plus<br>
* {@link MediaFormat#KEY_PROFILE},<br>
* <!-- {link MediaFormat#KEY_MAX_BIT_RATE},<br> -->
* {@link MediaFormat#KEY_BIT_RATE}</td>
* <td>as above, plus<br>
* {@link MediaFormat#KEY_PROFILE},<br>
* {@link MediaFormat#KEY_LEVEL}<sup>+</sup>,<br>
* <!-- {link MediaFormat#KEY_MAX_BIT_RATE},<br> -->
* {@link MediaFormat#KEY_BIT_RATE},<br>
* {@link CodecCapabilities#FEATURE_IntraRefresh}<sup>E</sup></td>
* </tr>
* <tr>
* <td colspan=4>
* <p class=note><strong>Notes:</strong><br>
* *: must be specified; otherwise, method returns {@code false}.<br>
* +: method does not verify that the format parameters are supported
* by the specified level.<br>
* D: decoders only<br>
* E: encoders only<br>
* ~: if both keys are provided values must match
* </td>
* </tr>
* </tbody>
* </table>
*
* @param format media format with optional feature directives.
* @throws IllegalArgumentException if format is not a valid media format.
* @return whether the codec capabilities support the given format
* and feature requests.
*/
public final boolean isFormatSupported(MediaFormat format) {
final Map<String, Object> map = format.getMap();
final String mime = (String)map.get(MediaFormat.KEY_MIME);
// mime must match if present
if (mime != null && !mMime.equalsIgnoreCase(mime)) {
return false;
}
// check feature support
for (Feature feat: getValidFeatures()) {
Integer yesNo = (Integer)map.get(MediaFormat.KEY_FEATURE_ + feat.mName);
if (yesNo == null) {
continue;
}
if ((yesNo == 1 && !isFeatureSupported(feat.mName)) ||
(yesNo == 0 && isFeatureRequired(feat.mName))) {
return false;
}
}
Integer profile = (Integer)map.get(MediaFormat.KEY_PROFILE);
Integer level = (Integer)map.get(MediaFormat.KEY_LEVEL);
if (profile != null) {
if (!supportsProfileLevel(profile, level)) {
return false;
}
// If we recognize this profile, check that this format is supported by the
// highest level supported by the codec for that profile. (Ignore specified
// level beyond the above profile/level check as level is only used as a
// guidance. E.g. AVC Level 1 CIF format is supported if codec supports level 1.1
// even though max size for Level 1 is QCIF. However, MPEG2 Simple Profile
// 1080p format is not supported even if codec supports Main Profile Level High,
// as Simple Profile does not support 1080p.
CodecCapabilities levelCaps = null;
int maxLevel = 0;
for (CodecProfileLevel pl : profileLevels) {
if (pl.profile == profile && pl.level > maxLevel) {
maxLevel = pl.level;
}
}
levelCaps = createFromProfileLevel(mMime, profile, maxLevel);
// remove profile from this format otherwise levelCaps.isFormatSupported will
// get into this same conditon and loop forever.
Map<String, Object> mapWithoutProfile = new HashMap<>(map);
mapWithoutProfile.remove(MediaFormat.KEY_PROFILE);
MediaFormat formatWithoutProfile = new MediaFormat(mapWithoutProfile);
if (levelCaps != null && !levelCaps.isFormatSupported(formatWithoutProfile)) {
return false;
}
}
if (mAudioCaps != null && !mAudioCaps.supportsFormat(format)) {
return false;
}
if (mVideoCaps != null && !mVideoCaps.supportsFormat(format)) {
return false;
}
if (mEncoderCaps != null && !mEncoderCaps.supportsFormat(format)) {
return false;
}
return true;
}
private static boolean supportsBitrate(
Range<Integer> bitrateRange, MediaFormat format) {
Map<String, Object> map = format.getMap();
// consider max bitrate over average bitrate for support
Integer maxBitrate = (Integer)map.get(MediaFormat.KEY_MAX_BIT_RATE);
Integer bitrate = (Integer)map.get(MediaFormat.KEY_BIT_RATE);
if (bitrate == null) {
bitrate = maxBitrate;
} else if (maxBitrate != null) {
bitrate = Math.max(bitrate, maxBitrate);
}
if (bitrate != null && bitrate > 0) {
return bitrateRange.contains(bitrate);
}
return true;
}
private boolean supportsProfileLevel(int profile, Integer level) {
for (CodecProfileLevel pl: profileLevels) {
if (pl.profile != profile) {
continue;
}
// AAC does not use levels
if (level == null || mMime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_AAC)) {
return true;
}
// H.263 levels are not completely ordered:
// Level45 support only implies Level10 support
if (mMime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_H263)) {
if (pl.level != level && pl.level == CodecProfileLevel.H263Level45
&& level > CodecProfileLevel.H263Level10) {
continue;
}
}
// MPEG4 levels are not completely ordered:
// Level1 support only implies Level0 (and not Level0b) support
if (mMime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_MPEG4)) {
if (pl.level != level && pl.level == CodecProfileLevel.MPEG4Level1
&& level > CodecProfileLevel.MPEG4Level0) {
continue;
}
}
// HEVC levels incorporate both tiers and levels. Verify tier support.
if (mMime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_HEVC)) {
boolean supportsHighTier =
(pl.level & CodecProfileLevel.HEVCHighTierLevels) != 0;
boolean checkingHighTier = (level & CodecProfileLevel.HEVCHighTierLevels) != 0;
// high tier levels are only supported by other high tier levels
if (checkingHighTier && !supportsHighTier) {
continue;
}
}
if (pl.level >= level) {
// if we recognize the listed profile/level, we must also recognize the
// profile/level arguments.
if (createFromProfileLevel(mMime, profile, pl.level) != null) {
return createFromProfileLevel(mMime, profile, level) != null;
}
return true;
}
}
return false;
}
// errors while reading profile levels - accessed from sister capabilities
int mError;
private static final String TAG = "CodecCapabilities";
// NEW-STYLE CAPABILITIES
private AudioCapabilities mAudioCaps;
private VideoCapabilities mVideoCaps;
private EncoderCapabilities mEncoderCaps;
private MediaFormat mDefaultFormat;
/**
* Returns a MediaFormat object with default values for configurations that have
* defaults.
*/
public MediaFormat getDefaultFormat() {
return mDefaultFormat;
}
/**
* Returns the mime type for which this codec-capability object was created.
*/
public String getMimeType() {
return mMime;
}
/**
* Returns the max number of the supported concurrent codec instances.
* <p>
* This is a hint for an upper bound. Applications should not expect to successfully
* operate more instances than the returned value, but the actual number of
* concurrently operable instances may be less as it depends on the available
* resources at time of use.
*/
public int getMaxSupportedInstances() {
return mMaxSupportedInstances;
}
private boolean isAudio() {
return mAudioCaps != null;
}
/**
* Returns the audio capabilities or {@code null} if this is not an audio codec.
*/
public AudioCapabilities getAudioCapabilities() {
return mAudioCaps;
}
private boolean isEncoder() {
return mEncoderCaps != null;
}
/**
* Returns the encoding capabilities or {@code null} if this is not an encoder.
*/
public EncoderCapabilities getEncoderCapabilities() {
return mEncoderCaps;
}
private boolean isVideo() {
return mVideoCaps != null;
}
/**
* Returns the video capabilities or {@code null} if this is not a video codec.
*/
public VideoCapabilities getVideoCapabilities() {
return mVideoCaps;
}
/** @hide */
public CodecCapabilities dup() {
return new CodecCapabilities(
// clone writable arrays
Arrays.copyOf(profileLevels, profileLevels.length),
Arrays.copyOf(colorFormats, colorFormats.length),
isEncoder(),
mFlagsVerified,
mDefaultFormat,
mCapabilitiesInfo);
}
/**
* Retrieve the codec capabilities for a certain {@code mime type}, {@code
* profile} and {@code level}. If the type, or profile-level combination
* is not understood by the framework, it returns null.
* <p class=note> In {@link android.os.Build.VERSION_CODES#M}, calling this
* method without calling any method of the {@link MediaCodecList} class beforehand
* results in a {@link NullPointerException}.</p>
*/
public static CodecCapabilities createFromProfileLevel(
String mime, int profile, int level) {
CodecProfileLevel pl = new CodecProfileLevel();
pl.profile = profile;
pl.level = level;
MediaFormat defaultFormat = new MediaFormat();
defaultFormat.setString(MediaFormat.KEY_MIME, mime);
CodecCapabilities ret = new CodecCapabilities(
new CodecProfileLevel[] { pl }, new int[0], true /* encoder */,
0 /* flags */, defaultFormat, new MediaFormat() /* info */);
if (ret.mError != 0) {
return null;
}
return ret;
}
/* package private */ CodecCapabilities(
CodecProfileLevel[] profLevs, int[] colFmts,
boolean encoder, int flags,
Map<String, Object>defaultFormatMap,
Map<String, Object>capabilitiesMap) {
this(profLevs, colFmts, encoder, flags,
new MediaFormat(defaultFormatMap),
new MediaFormat(capabilitiesMap));
}
private MediaFormat mCapabilitiesInfo;
/* package private */ CodecCapabilities(
CodecProfileLevel[] profLevs, int[] colFmts, boolean encoder, int flags,
MediaFormat defaultFormat, MediaFormat info) {
final Map<String, Object> map = info.getMap();
colorFormats = colFmts;
mFlagsVerified = flags;
mDefaultFormat = defaultFormat;
mCapabilitiesInfo = info;
mMime = mDefaultFormat.getString(MediaFormat.KEY_MIME);
/* VP9 introduced profiles around 2016, so some VP9 codecs may not advertise any
supported profiles. Determine the level for them using the info they provide. */
if (profLevs.length == 0 && mMime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_VP9)) {
CodecProfileLevel profLev = new CodecProfileLevel();
profLev.profile = CodecProfileLevel.VP9Profile0;
profLev.level = VideoCapabilities.equivalentVP9Level(info);
profLevs = new CodecProfileLevel[] { profLev };
}
profileLevels = profLevs;
if (mMime.toLowerCase().startsWith("audio/")) {
mAudioCaps = AudioCapabilities.create(info, this);
mAudioCaps.setDefaultFormat(mDefaultFormat);
} else if (mMime.toLowerCase().startsWith("video/")) {
mVideoCaps = VideoCapabilities.create(info, this);
}
if (encoder) {
mEncoderCaps = EncoderCapabilities.create(info, this);
mEncoderCaps.setDefaultFormat(mDefaultFormat);
}
final Map<String, Object> global = MediaCodecList.getGlobalSettings();
mMaxSupportedInstances = Utils.parseIntSafely(
global.get("max-concurrent-instances"), DEFAULT_MAX_SUPPORTED_INSTANCES);
int maxInstances = Utils.parseIntSafely(
map.get("max-concurrent-instances"), mMaxSupportedInstances);
mMaxSupportedInstances =
Range.create(1, MAX_SUPPORTED_INSTANCES_LIMIT).clamp(maxInstances);
for (Feature feat: getValidFeatures()) {
String key = MediaFormat.KEY_FEATURE_ + feat.mName;
Integer yesNo = (Integer)map.get(key);
if (yesNo == null) {
continue;
}
if (yesNo > 0) {
mFlagsRequired |= feat.mValue;
}
mFlagsSupported |= feat.mValue;
mDefaultFormat.setInteger(key, 1);
// TODO restrict features by mFlagsVerified once all codecs reliably verify them
}
}
}
/**
* A class that supports querying the audio capabilities of a codec.
*/
public static final class AudioCapabilities {
private static final String TAG = "AudioCapabilities";
private CodecCapabilities mParent;
private Range<Integer> mBitrateRange;
private int[] mSampleRates;
private Range<Integer>[] mSampleRateRanges;
private int mMaxInputChannelCount;
private static final int MAX_INPUT_CHANNEL_COUNT = 30;
/**
* Returns the range of supported bitrates in bits/second.
*/
public Range<Integer> getBitrateRange() {
return mBitrateRange;
}
/**
* Returns the array of supported sample rates if the codec
* supports only discrete values. Otherwise, it returns
* {@code null}. The array is sorted in ascending order.
*/
public int[] getSupportedSampleRates() {
return Arrays.copyOf(mSampleRates, mSampleRates.length);
}
/**
* Returns the array of supported sample rate ranges. The
* array is sorted in ascending order, and the ranges are
* distinct.
*/
public Range<Integer>[] getSupportedSampleRateRanges() {
return Arrays.copyOf(mSampleRateRanges, mSampleRateRanges.length);
}
/**
* Returns the maximum number of input channels supported. The codec
* supports any number of channels between 1 and this maximum value.
*/
public int getMaxInputChannelCount() {
return mMaxInputChannelCount;
}
/* no public constructor */
private AudioCapabilities() { }
/** @hide */
public static AudioCapabilities create(
MediaFormat info, CodecCapabilities parent) {
AudioCapabilities caps = new AudioCapabilities();
caps.init(info, parent);
return caps;
}
/** @hide */
public void init(MediaFormat info, CodecCapabilities parent) {
mParent = parent;
initWithPlatformLimits();
applyLevelLimits();
parseFromInfo(info);
}
private void initWithPlatformLimits() {
mBitrateRange = Range.create(0, Integer.MAX_VALUE);
mMaxInputChannelCount = MAX_INPUT_CHANNEL_COUNT;
// mBitrateRange = Range.create(1, 320000);
mSampleRateRanges = new Range[] { Range.create(8000, 96000) };
mSampleRates = null;
}
private boolean supports(Integer sampleRate, Integer inputChannels) {
// channels and sample rates are checked orthogonally
if (inputChannels != null &&
(inputChannels < 1 || inputChannels > mMaxInputChannelCount)) {
return false;
}
if (sampleRate != null) {
int ix = Utils.binarySearchDistinctRanges(
mSampleRateRanges, sampleRate);
if (ix < 0) {
return false;
}
}
return true;
}
/**
* Query whether the sample rate is supported by the codec.
*/
public boolean isSampleRateSupported(int sampleRate) {
return supports(sampleRate, null);
}
/** modifies rates */
private void limitSampleRates(int[] rates) {
Arrays.sort(rates);
ArrayList<Range<Integer>> ranges = new ArrayList<Range<Integer>>();
for (int rate: rates) {
if (supports(rate, null /* channels */)) {
ranges.add(Range.create(rate, rate));
}
}
mSampleRateRanges = ranges.toArray(new Range[ranges.size()]);
createDiscreteSampleRates();
}
private void createDiscreteSampleRates() {
mSampleRates = new int[mSampleRateRanges.length];
for (int i = 0; i < mSampleRateRanges.length; i++) {
mSampleRates[i] = mSampleRateRanges[i].getLower();
}
}
/** modifies rateRanges */
private void limitSampleRates(Range<Integer>[] rateRanges) {
sortDistinctRanges(rateRanges);
mSampleRateRanges = intersectSortedDistinctRanges(mSampleRateRanges, rateRanges);
// check if all values are discrete
for (Range<Integer> range: mSampleRateRanges) {
if (!range.getLower().equals(range.getUpper())) {
mSampleRates = null;
return;
}
}
createDiscreteSampleRates();
}
private void applyLevelLimits() {
int[] sampleRates = null;
Range<Integer> sampleRateRange = null, bitRates = null;
int maxChannels = MAX_INPUT_CHANNEL_COUNT;
String mime = mParent.getMimeType();
if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_MPEG)) {
sampleRates = new int[] {
8000, 11025, 12000,
16000, 22050, 24000,
32000, 44100, 48000 };
bitRates = Range.create(8000, 320000);
maxChannels = 2;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_AMR_NB)) {
sampleRates = new int[] { 8000 };
bitRates = Range.create(4750, 12200);
maxChannels = 1;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_AMR_WB)) {
sampleRates = new int[] { 16000 };
bitRates = Range.create(6600, 23850);
maxChannels = 1;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_AAC)) {
sampleRates = new int[] {
7350, 8000,
11025, 12000, 16000,
22050, 24000, 32000,
44100, 48000, 64000,
88200, 96000 };
bitRates = Range.create(8000, 510000);
maxChannels = 48;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_VORBIS)) {
bitRates = Range.create(32000, 500000);
sampleRateRange = Range.create(8000, 192000);
maxChannels = 255;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_OPUS)) {
bitRates = Range.create(6000, 510000);
sampleRates = new int[] { 8000, 12000, 16000, 24000, 48000 };
maxChannels = 255;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_RAW)) {
sampleRateRange = Range.create(1, 96000);
bitRates = Range.create(1, 10000000);
maxChannels = AudioTrack.CHANNEL_COUNT_MAX;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_FLAC)) {
sampleRateRange = Range.create(1, 655350);
// lossless codec, so bitrate is ignored
maxChannels = 255;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_G711_ALAW)
|| mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_G711_MLAW)) {
sampleRates = new int[] { 8000 };
bitRates = Range.create(64000, 64000);
// platform allows multiple channels for this format
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_MSGSM)) {
sampleRates = new int[] { 8000 };
bitRates = Range.create(13000, 13000);
maxChannels = 1;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_AC3)) {
maxChannels = 6;
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_EAC3)) {
maxChannels = 16;
} else {
Log.w(TAG, "Unsupported mime " + mime);
mParent.mError |= ERROR_UNSUPPORTED;
}
// restrict ranges
if (sampleRates != null) {
limitSampleRates(sampleRates);
} else if (sampleRateRange != null) {
limitSampleRates(new Range[] { sampleRateRange });
}
applyLimits(maxChannels, bitRates);
}
private void applyLimits(int maxInputChannels, Range<Integer> bitRates) {
mMaxInputChannelCount = Range.create(1, mMaxInputChannelCount)
.clamp(maxInputChannels);
if (bitRates != null) {
mBitrateRange = mBitrateRange.intersect(bitRates);
}
}
private void parseFromInfo(MediaFormat info) {
int maxInputChannels = MAX_INPUT_CHANNEL_COUNT;
Range<Integer> bitRates = POSITIVE_INTEGERS;
if (info.containsKey("sample-rate-ranges")) {
String[] rateStrings = info.getString("sample-rate-ranges").split(",");
Range<Integer>[] rateRanges = new Range[rateStrings.length];
for (int i = 0; i < rateStrings.length; i++) {
rateRanges[i] = Utils.parseIntRange(rateStrings[i], null);
}
limitSampleRates(rateRanges);
}
if (info.containsKey("max-channel-count")) {
maxInputChannels = Utils.parseIntSafely(
info.getString("max-channel-count"), maxInputChannels);
} else if ((mParent.mError & ERROR_UNSUPPORTED) != 0) {
maxInputChannels = 0;
}
if (info.containsKey("bitrate-range")) {
bitRates = bitRates.intersect(
Utils.parseIntRange(info.getString("bitrate-range"), bitRates));
}
applyLimits(maxInputChannels, bitRates);
}
/** @hide */
public void setDefaultFormat(MediaFormat format) {
// report settings that have only a single choice
if (mBitrateRange.getLower().equals(mBitrateRange.getUpper())) {
format.setInteger(MediaFormat.KEY_BIT_RATE, mBitrateRange.getLower());
}
if (mMaxInputChannelCount == 1) {
// mono-only format
format.setInteger(MediaFormat.KEY_CHANNEL_COUNT, 1);
}
if (mSampleRates != null && mSampleRates.length == 1) {
format.setInteger(MediaFormat.KEY_SAMPLE_RATE, mSampleRates[0]);
}
}
/** @hide */
public boolean supportsFormat(MediaFormat format) {
Map<String, Object> map = format.getMap();
Integer sampleRate = (Integer)map.get(MediaFormat.KEY_SAMPLE_RATE);
Integer channels = (Integer)map.get(MediaFormat.KEY_CHANNEL_COUNT);
if (!supports(sampleRate, channels)) {
return false;
}
if (!CodecCapabilities.supportsBitrate(mBitrateRange, format)) {
return false;
}
// nothing to do for:
// KEY_CHANNEL_MASK: codecs don't get this
// KEY_IS_ADTS: required feature for all AAC decoders
return true;
}
}
/**
* A class that supports querying the video capabilities of a codec.
*/
public static final class VideoCapabilities {
private static final String TAG = "VideoCapabilities";
private CodecCapabilities mParent;
private Range<Integer> mBitrateRange;
private Range<Integer> mHeightRange;
private Range<Integer> mWidthRange;
private Range<Integer> mBlockCountRange;
private Range<Integer> mHorizontalBlockRange;
private Range<Integer> mVerticalBlockRange;
private Range<Rational> mAspectRatioRange;
private Range<Rational> mBlockAspectRatioRange;
private Range<Long> mBlocksPerSecondRange;
private Map<Size, Range<Long>> mMeasuredFrameRates;
private Range<Integer> mFrameRateRange;
private int mBlockWidth;
private int mBlockHeight;
private int mWidthAlignment;
private int mHeightAlignment;
private int mSmallerDimensionUpperLimit;
private boolean mAllowMbOverride; // allow XML to override calculated limits
/**
* Returns the range of supported bitrates in bits/second.
*/
public Range<Integer> getBitrateRange() {
return mBitrateRange;
}
/**
* Returns the range of supported video widths.
*/
public Range<Integer> getSupportedWidths() {
return mWidthRange;
}
/**
* Returns the range of supported video heights.
*/
public Range<Integer> getSupportedHeights() {
return mHeightRange;
}
/**
* Returns the alignment requirement for video width (in pixels).
*
* This is a power-of-2 value that video width must be a
* multiple of.
*/
public int getWidthAlignment() {
return mWidthAlignment;
}
/**
* Returns the alignment requirement for video height (in pixels).
*
* This is a power-of-2 value that video height must be a
* multiple of.
*/
public int getHeightAlignment() {
return mHeightAlignment;
}
/**
* Return the upper limit on the smaller dimension of width or height.
* <p></p>
* Some codecs have a limit on the smaller dimension, whether it be
* the width or the height. E.g. a codec may only be able to handle
* up to 1920x1080 both in landscape and portrait mode (1080x1920).
* In this case the maximum width and height are both 1920, but the
* smaller dimension limit will be 1080. For other codecs, this is
* {@code Math.min(getSupportedWidths().getUpper(),
* getSupportedHeights().getUpper())}.
*
* @hide
*/
public int getSmallerDimensionUpperLimit() {
return mSmallerDimensionUpperLimit;
}
/**
* Returns the range of supported frame rates.
* <p>
* This is not a performance indicator. Rather, it expresses the
* limits specified in the coding standard, based on the complexities
* of encoding material for later playback at a certain frame rate,
* or the decoding of such material in non-realtime.
*/
public Range<Integer> getSupportedFrameRates() {
return mFrameRateRange;
}
/**
* Returns the range of supported video widths for a video height.
* @param height the height of the video
*/
public Range<Integer> getSupportedWidthsFor(int height) {
try {
Range<Integer> range = mWidthRange;
if (!mHeightRange.contains(height)
|| (height % mHeightAlignment) != 0) {
throw new IllegalArgumentException("unsupported height");
}
final int heightInBlocks = Utils.divUp(height, mBlockHeight);
// constrain by block count and by block aspect ratio
final int minWidthInBlocks = Math.max(
Utils.divUp(mBlockCountRange.getLower(), heightInBlocks),
(int)Math.ceil(mBlockAspectRatioRange.getLower().doubleValue()
* heightInBlocks));
final int maxWidthInBlocks = Math.min(
mBlockCountRange.getUpper() / heightInBlocks,
(int)(mBlockAspectRatioRange.getUpper().doubleValue()
* heightInBlocks));
range = range.intersect(
(minWidthInBlocks - 1) * mBlockWidth + mWidthAlignment,
maxWidthInBlocks * mBlockWidth);
// constrain by smaller dimension limit
if (height > mSmallerDimensionUpperLimit) {
range = range.intersect(1, mSmallerDimensionUpperLimit);
}
// constrain by aspect ratio
range = range.intersect(
(int)Math.ceil(mAspectRatioRange.getLower().doubleValue()
* height),
(int)(mAspectRatioRange.getUpper().doubleValue() * height));
return range;
} catch (IllegalArgumentException e) {
// height is not supported because there are no suitable widths
Log.v(TAG, "could not get supported widths for " + height);
throw new IllegalArgumentException("unsupported height");
}
}
/**
* Returns the range of supported video heights for a video width
* @param width the width of the video
*/
public Range<Integer> getSupportedHeightsFor(int width) {
try {
Range<Integer> range = mHeightRange;
if (!mWidthRange.contains(width)
|| (width % mWidthAlignment) != 0) {
throw new IllegalArgumentException("unsupported width");
}
final int widthInBlocks = Utils.divUp(width, mBlockWidth);
// constrain by block count and by block aspect ratio
final int minHeightInBlocks = Math.max(
Utils.divUp(mBlockCountRange.getLower(), widthInBlocks),
(int)Math.ceil(widthInBlocks /
mBlockAspectRatioRange.getUpper().doubleValue()));
final int maxHeightInBlocks = Math.min(
mBlockCountRange.getUpper() / widthInBlocks,
(int)(widthInBlocks /
mBlockAspectRatioRange.getLower().doubleValue()));
range = range.intersect(
(minHeightInBlocks - 1) * mBlockHeight + mHeightAlignment,
maxHeightInBlocks * mBlockHeight);
// constrain by smaller dimension limit
if (width > mSmallerDimensionUpperLimit) {
range = range.intersect(1, mSmallerDimensionUpperLimit);
}
// constrain by aspect ratio
range = range.intersect(
(int)Math.ceil(width /
mAspectRatioRange.getUpper().doubleValue()),
(int)(width / mAspectRatioRange.getLower().doubleValue()));
return range;
} catch (IllegalArgumentException e) {
// width is not supported because there are no suitable heights
Log.v(TAG, "could not get supported heights for " + width);
throw new IllegalArgumentException("unsupported width");
}
}
/**
* Returns the range of supported video frame rates for a video size.
* <p>
* This is not a performance indicator. Rather, it expresses the limits specified in
* the coding standard, based on the complexities of encoding material of a given
* size for later playback at a certain frame rate, or the decoding of such material
* in non-realtime.
* @param width the width of the video
* @param height the height of the video
*/
public Range<Double> getSupportedFrameRatesFor(int width, int height) {
Range<Integer> range = mHeightRange;
if (!supports(width, height, null)) {
throw new IllegalArgumentException("unsupported size");
}
final int blockCount =
Utils.divUp(width, mBlockWidth) * Utils.divUp(height, mBlockHeight);
return Range.create(
Math.max(mBlocksPerSecondRange.getLower() / (double) blockCount,
(double) mFrameRateRange.getLower()),
Math.min(mBlocksPerSecondRange.getUpper() / (double) blockCount,
(double) mFrameRateRange.getUpper()));
}
private int getBlockCount(int width, int height) {
return Utils.divUp(width, mBlockWidth) * Utils.divUp(height, mBlockHeight);
}
@NonNull
private Size findClosestSize(int width, int height) {
int targetBlockCount = getBlockCount(width, height);
Size closestSize = null;
int minDiff = Integer.MAX_VALUE;
for (Size size : mMeasuredFrameRates.keySet()) {
int diff = Math.abs(targetBlockCount -
getBlockCount(size.getWidth(), size.getHeight()));
if (diff < minDiff) {
minDiff = diff;
closestSize = size;
}
}
return closestSize;
}
private Range<Double> estimateFrameRatesFor(int width, int height) {
Size size = findClosestSize(width, height);
Range<Long> range = mMeasuredFrameRates.get(size);
Double ratio = getBlockCount(size.getWidth(), size.getHeight())
/ (double)Math.max(getBlockCount(width, height), 1);
return Range.create(range.getLower() * ratio, range.getUpper() * ratio);
}
/**
* Returns the range of achievable video frame rates for a video size.
* May return {@code null}, if the codec did not publish any measurement
* data.
* <p>
* This is a performance estimate provided by the device manufacturer based on statistical
* sampling of full-speed decoding and encoding measurements in various configurations
* of common video sizes supported by the codec. As such it should only be used to
* compare individual codecs on the device. The value is not suitable for comparing
* different devices or even different android releases for the same device.
* <p>
* <em>On {@link android.os.Build.VERSION_CODES#M} release</em> the returned range
* corresponds to the fastest frame rates achieved in the tested configurations. As
* such, it should not be used to gauge guaranteed or even average codec performance
* on the device.
* <p>
* <em>On {@link android.os.Build.VERSION_CODES#N} release</em> the returned range
* corresponds closer to sustained performance <em>in tested configurations</em>.
* One can expect to achieve sustained performance higher than the lower limit more than
* 50% of the time, and higher than half of the lower limit at least 90% of the time
* <em>in tested configurations</em>.
* Conversely, one can expect performance lower than twice the upper limit at least
* 90% of the time.
* <p class=note>
* Tested configurations use a single active codec. For use cases where multiple
* codecs are active, applications can expect lower and in most cases significantly lower
* performance.
* <p class=note>
* The returned range value is interpolated from the nearest frame size(s) tested.
* Codec performance is severely impacted by other activity on the device as well
* as environmental factors (such as battery level, temperature or power source), and can
* vary significantly even in a steady environment.
* <p class=note>
* Use this method in cases where only codec performance matters, e.g. to evaluate if
* a codec has any chance of meeting a performance target. Codecs are listed
* in {@link MediaCodecList} in the preferred order as defined by the device
* manufacturer. As such, applications should use the first suitable codec in the
* list to achieve the best balance between power use and performance.
*
* @param width the width of the video
* @param height the height of the video
*
* @throws IllegalArgumentException if the video size is not supported.
*/
@Nullable
public Range<Double> getAchievableFrameRatesFor(int width, int height) {
if (!supports(width, height, null)) {
throw new IllegalArgumentException("unsupported size");
}
if (mMeasuredFrameRates == null || mMeasuredFrameRates.size() <= 0) {
Log.w(TAG, "Codec did not publish any measurement data.");
return null;
}
return estimateFrameRatesFor(width, height);
}
/**
* Returns whether a given video size ({@code width} and
* {@code height}) and {@code frameRate} combination is supported.
*/
public boolean areSizeAndRateSupported(
int width, int height, double frameRate) {
return supports(width, height, frameRate);
}
/**
* Returns whether a given video size ({@code width} and
* {@code height}) is supported.
*/
public boolean isSizeSupported(int width, int height) {
return supports(width, height, null);
}
private boolean supports(Integer width, Integer height, Number rate) {
boolean ok = true;
if (ok && width != null) {
ok = mWidthRange.contains(width)
&& (width % mWidthAlignment == 0);
}
if (ok && height != null) {
ok = mHeightRange.contains(height)
&& (height % mHeightAlignment == 0);
}
if (ok && rate != null) {
ok = mFrameRateRange.contains(Utils.intRangeFor(rate.doubleValue()));
}
if (ok && height != null && width != null) {
ok = Math.min(height, width) <= mSmallerDimensionUpperLimit;
final int widthInBlocks = Utils.divUp(width, mBlockWidth);
final int heightInBlocks = Utils.divUp(height, mBlockHeight);
final int blockCount = widthInBlocks * heightInBlocks;
ok = ok && mBlockCountRange.contains(blockCount)
&& mBlockAspectRatioRange.contains(
new Rational(widthInBlocks, heightInBlocks))
&& mAspectRatioRange.contains(new Rational(width, height));
if (ok && rate != null) {
double blocksPerSec = blockCount * rate.doubleValue();
ok = mBlocksPerSecondRange.contains(
Utils.longRangeFor(blocksPerSec));
}
}
return ok;
}
/**
* @hide
* @throws java.lang.ClassCastException */
public boolean supportsFormat(MediaFormat format) {
final Map<String, Object> map = format.getMap();
Integer width = (Integer)map.get(MediaFormat.KEY_WIDTH);
Integer height = (Integer)map.get(MediaFormat.KEY_HEIGHT);
Number rate = (Number)map.get(MediaFormat.KEY_FRAME_RATE);
if (!supports(width, height, rate)) {
return false;
}
if (!CodecCapabilities.supportsBitrate(mBitrateRange, format)) {
return false;
}
// we ignore color-format for now as it is not reliably reported by codec
return true;
}
/* no public constructor */
private VideoCapabilities() { }
/** @hide */
public static VideoCapabilities create(
MediaFormat info, CodecCapabilities parent) {
VideoCapabilities caps = new VideoCapabilities();
caps.init(info, parent);
return caps;
}
/** @hide */
public void init(MediaFormat info, CodecCapabilities parent) {
mParent = parent;
initWithPlatformLimits();
applyLevelLimits();
parseFromInfo(info);
updateLimits();
}
/** @hide */
public Size getBlockSize() {
return new Size(mBlockWidth, mBlockHeight);
}
/** @hide */
public Range<Integer> getBlockCountRange() {
return mBlockCountRange;
}
/** @hide */
public Range<Long> getBlocksPerSecondRange() {
return mBlocksPerSecondRange;
}
/** @hide */
public Range<Rational> getAspectRatioRange(boolean blocks) {
return blocks ? mBlockAspectRatioRange : mAspectRatioRange;
}
private void initWithPlatformLimits() {
mBitrateRange = BITRATE_RANGE;
mWidthRange = SIZE_RANGE;
mHeightRange = SIZE_RANGE;
mFrameRateRange = FRAME_RATE_RANGE;
mHorizontalBlockRange = SIZE_RANGE;
mVerticalBlockRange = SIZE_RANGE;
// full positive ranges are supported as these get calculated
mBlockCountRange = POSITIVE_INTEGERS;
mBlocksPerSecondRange = POSITIVE_LONGS;
mBlockAspectRatioRange = POSITIVE_RATIONALS;
mAspectRatioRange = POSITIVE_RATIONALS;
// YUV 4:2:0 requires 2:2 alignment
mWidthAlignment = 2;
mHeightAlignment = 2;
mBlockWidth = 2;
mBlockHeight = 2;
mSmallerDimensionUpperLimit = SIZE_RANGE.getUpper();
}
private Map<Size, Range<Long>> getMeasuredFrameRates(Map<String, Object> map) {
Map<Size, Range<Long>> ret = new HashMap<Size, Range<Long>>();
final String prefix = "measured-frame-rate-";
Set<String> keys = map.keySet();
for (String key : keys) {
// looking for: measured-frame-rate-WIDTHxHEIGHT-range
if (!key.startsWith(prefix)) {
continue;
}
String subKey = key.substring(prefix.length());
String[] temp = key.split("-");
if (temp.length != 5) {
continue;
}
String sizeStr = temp[3];
Size size = Utils.parseSize(sizeStr, null);
if (size == null || size.getWidth() * size.getHeight() <= 0) {
continue;
}
Range<Long> range = Utils.parseLongRange(map.get(key), null);
if (range == null || range.getLower() < 0 || range.getUpper() < 0) {
continue;
}
ret.put(size, range);
}
return ret;
}
private static Pair<Range<Integer>, Range<Integer>> parseWidthHeightRanges(Object o) {
Pair<Size, Size> range = Utils.parseSizeRange(o);
if (range != null) {
try {
return Pair.create(
Range.create(range.first.getWidth(), range.second.getWidth()),
Range.create(range.first.getHeight(), range.second.getHeight()));
} catch (IllegalArgumentException e) {
Log.w(TAG, "could not parse size range '" + o + "'");
}
}
return null;
}
/** @hide */
public static int equivalentVP9Level(MediaFormat info) {
final Map<String, Object> map = info.getMap();
Size blockSize = Utils.parseSize(map.get("block-size"), new Size(8, 8));
int BS = blockSize.getWidth() * blockSize.getHeight();
Range<Integer> counts = Utils.parseIntRange(map.get("block-count-range"), null);
int FS = counts == null ? 0 : BS * counts.getUpper();
Range<Long> blockRates =
Utils.parseLongRange(map.get("blocks-per-second-range"), null);
long SR = blockRates == null ? 0 : BS * blockRates.getUpper();
Pair<Range<Integer>, Range<Integer>> dimensionRanges =
parseWidthHeightRanges(map.get("size-range"));
int D = dimensionRanges == null ? 0 : Math.max(
dimensionRanges.first.getUpper(), dimensionRanges.second.getUpper());
Range<Integer> bitRates = Utils.parseIntRange(map.get("bitrate-range"), null);
int BR = bitRates == null ? 0 : Utils.divUp(bitRates.getUpper(), 1000);
if (SR <= 829440 && FS <= 36864 && BR <= 200 && D <= 512)
return CodecProfileLevel.VP9Level1;
if (SR <= 2764800 && FS <= 73728 && BR <= 800 && D <= 768)
return CodecProfileLevel.VP9Level11;
if (SR <= 4608000 && FS <= 122880 && BR <= 1800 && D <= 960)
return CodecProfileLevel.VP9Level2;
if (SR <= 9216000 && FS <= 245760 && BR <= 3600 && D <= 1344)
return CodecProfileLevel.VP9Level21;
if (SR <= 20736000 && FS <= 552960 && BR <= 7200 && D <= 2048)
return CodecProfileLevel.VP9Level3;
if (SR <= 36864000 && FS <= 983040 && BR <= 12000 && D <= 2752)
return CodecProfileLevel.VP9Level31;
if (SR <= 83558400 && FS <= 2228224 && BR <= 18000 && D <= 4160)
return CodecProfileLevel.VP9Level4;
if (SR <= 160432128 && FS <= 2228224 && BR <= 30000 && D <= 4160)
return CodecProfileLevel.VP9Level41;
if (SR <= 311951360 && FS <= 8912896 && BR <= 60000 && D <= 8384)
return CodecProfileLevel.VP9Level5;
if (SR <= 588251136 && FS <= 8912896 && BR <= 120000 && D <= 8384)
return CodecProfileLevel.VP9Level51;
if (SR <= 1176502272 && FS <= 8912896 && BR <= 180000 && D <= 8384)
return CodecProfileLevel.VP9Level52;
if (SR <= 1176502272 && FS <= 35651584 && BR <= 180000 && D <= 16832)
return CodecProfileLevel.VP9Level6;
if (SR <= 2353004544L && FS <= 35651584 && BR <= 240000 && D <= 16832)
return CodecProfileLevel.VP9Level61;
if (SR <= 4706009088L && FS <= 35651584 && BR <= 480000 && D <= 16832)
return CodecProfileLevel.VP9Level62;
// returning largest level
return CodecProfileLevel.VP9Level62;
}
private void parseFromInfo(MediaFormat info) {
final Map<String, Object> map = info.getMap();
Size blockSize = new Size(mBlockWidth, mBlockHeight);
Size alignment = new Size(mWidthAlignment, mHeightAlignment);
Range<Integer> counts = null, widths = null, heights = null;
Range<Integer> frameRates = null, bitRates = null;
Range<Long> blockRates = null;
Range<Rational> ratios = null, blockRatios = null;
blockSize = Utils.parseSize(map.get("block-size"), blockSize);
alignment = Utils.parseSize(map.get("alignment"), alignment);
counts = Utils.parseIntRange(map.get("block-count-range"), null);
blockRates =
Utils.parseLongRange(map.get("blocks-per-second-range"), null);
mMeasuredFrameRates = getMeasuredFrameRates(map);
Pair<Range<Integer>, Range<Integer>> sizeRanges =
parseWidthHeightRanges(map.get("size-range"));
if (sizeRanges != null) {
widths = sizeRanges.first;
heights = sizeRanges.second;
}
// for now this just means using the smaller max size as 2nd
// upper limit.
// for now we are keeping the profile specific "width/height
// in macroblocks" limits.
if (map.containsKey("feature-can-swap-width-height")) {
if (widths != null) {
mSmallerDimensionUpperLimit =
Math.min(widths.getUpper(), heights.getUpper());
widths = heights = widths.extend(heights);
} else {
Log.w(TAG, "feature can-swap-width-height is best used with size-range");
mSmallerDimensionUpperLimit =
Math.min(mWidthRange.getUpper(), mHeightRange.getUpper());
mWidthRange = mHeightRange = mWidthRange.extend(mHeightRange);
}
}
ratios = Utils.parseRationalRange(
map.get("block-aspect-ratio-range"), null);
blockRatios = Utils.parseRationalRange(
map.get("pixel-aspect-ratio-range"), null);
frameRates = Utils.parseIntRange(map.get("frame-rate-range"), null);
if (frameRates != null) {
try {
frameRates = frameRates.intersect(FRAME_RATE_RANGE);
} catch (IllegalArgumentException e) {
Log.w(TAG, "frame rate range (" + frameRates
+ ") is out of limits: " + FRAME_RATE_RANGE);
frameRates = null;
}
}
bitRates = Utils.parseIntRange(map.get("bitrate-range"), null);
if (bitRates != null) {
try {
bitRates = bitRates.intersect(BITRATE_RANGE);
} catch (IllegalArgumentException e) {
Log.w(TAG, "bitrate range (" + bitRates
+ ") is out of limits: " + BITRATE_RANGE);
bitRates = null;
}
}
checkPowerOfTwo(
blockSize.getWidth(), "block-size width must be power of two");
checkPowerOfTwo(
blockSize.getHeight(), "block-size height must be power of two");
checkPowerOfTwo(
alignment.getWidth(), "alignment width must be power of two");
checkPowerOfTwo(
alignment.getHeight(), "alignment height must be power of two");
// update block-size and alignment
applyMacroBlockLimits(
Integer.MAX_VALUE, Integer.MAX_VALUE, Integer.MAX_VALUE,
Long.MAX_VALUE, blockSize.getWidth(), blockSize.getHeight(),
alignment.getWidth(), alignment.getHeight());
if ((mParent.mError & ERROR_UNSUPPORTED) != 0 || mAllowMbOverride) {
// codec supports profiles that we don't know.
// Use supplied values clipped to platform limits
if (widths != null) {
mWidthRange = SIZE_RANGE.intersect(widths);
}
if (heights != null) {
mHeightRange = SIZE_RANGE.intersect(heights);
}
if (counts != null) {
mBlockCountRange = POSITIVE_INTEGERS.intersect(
Utils.factorRange(counts, mBlockWidth * mBlockHeight
/ blockSize.getWidth() / blockSize.getHeight()));
}
if (blockRates != null) {
mBlocksPerSecondRange = POSITIVE_LONGS.intersect(
Utils.factorRange(blockRates, mBlockWidth * mBlockHeight
/ blockSize.getWidth() / blockSize.getHeight()));
}
if (blockRatios != null) {
mBlockAspectRatioRange = POSITIVE_RATIONALS.intersect(
Utils.scaleRange(blockRatios,
mBlockHeight / blockSize.getHeight(),
mBlockWidth / blockSize.getWidth()));
}
if (ratios != null) {
mAspectRatioRange = POSITIVE_RATIONALS.intersect(ratios);
}
if (frameRates != null) {
mFrameRateRange = FRAME_RATE_RANGE.intersect(frameRates);
}
if (bitRates != null) {
// only allow bitrate override if unsupported profiles were encountered
if ((mParent.mError & ERROR_UNSUPPORTED) != 0) {
mBitrateRange = BITRATE_RANGE.intersect(bitRates);
} else {
mBitrateRange = mBitrateRange.intersect(bitRates);
}
}
} else {
// no unsupported profile/levels, so restrict values to known limits
if (widths != null) {
mWidthRange = mWidthRange.intersect(widths);
}
if (heights != null) {
mHeightRange = mHeightRange.intersect(heights);
}
if (counts != null) {
mBlockCountRange = mBlockCountRange.intersect(
Utils.factorRange(counts, mBlockWidth * mBlockHeight
/ blockSize.getWidth() / blockSize.getHeight()));
}
if (blockRates != null) {
mBlocksPerSecondRange = mBlocksPerSecondRange.intersect(
Utils.factorRange(blockRates, mBlockWidth * mBlockHeight
/ blockSize.getWidth() / blockSize.getHeight()));
}
if (blockRatios != null) {
mBlockAspectRatioRange = mBlockAspectRatioRange.intersect(
Utils.scaleRange(blockRatios,
mBlockHeight / blockSize.getHeight(),
mBlockWidth / blockSize.getWidth()));
}
if (ratios != null) {
mAspectRatioRange = mAspectRatioRange.intersect(ratios);
}
if (frameRates != null) {
mFrameRateRange = mFrameRateRange.intersect(frameRates);
}
if (bitRates != null) {
mBitrateRange = mBitrateRange.intersect(bitRates);
}
}
updateLimits();
}
private void applyBlockLimits(
int blockWidth, int blockHeight,
Range<Integer> counts, Range<Long> rates, Range<Rational> ratios) {
checkPowerOfTwo(blockWidth, "blockWidth must be a power of two");
checkPowerOfTwo(blockHeight, "blockHeight must be a power of two");
final int newBlockWidth = Math.max(blockWidth, mBlockWidth);
final int newBlockHeight = Math.max(blockHeight, mBlockHeight);
// factor will always be a power-of-2
int factor =
newBlockWidth * newBlockHeight / mBlockWidth / mBlockHeight;
if (factor != 1) {
mBlockCountRange = Utils.factorRange(mBlockCountRange, factor);
mBlocksPerSecondRange = Utils.factorRange(
mBlocksPerSecondRange, factor);
mBlockAspectRatioRange = Utils.scaleRange(
mBlockAspectRatioRange,
newBlockHeight / mBlockHeight,
newBlockWidth / mBlockWidth);
mHorizontalBlockRange = Utils.factorRange(
mHorizontalBlockRange, newBlockWidth / mBlockWidth);
mVerticalBlockRange = Utils.factorRange(
mVerticalBlockRange, newBlockHeight / mBlockHeight);
}
factor = newBlockWidth * newBlockHeight / blockWidth / blockHeight;
if (factor != 1) {
counts = Utils.factorRange(counts, factor);
rates = Utils.factorRange(rates, factor);
ratios = Utils.scaleRange(
ratios, newBlockHeight / blockHeight,
newBlockWidth / blockWidth);
}
mBlockCountRange = mBlockCountRange.intersect(counts);
mBlocksPerSecondRange = mBlocksPerSecondRange.intersect(rates);
mBlockAspectRatioRange = mBlockAspectRatioRange.intersect(ratios);
mBlockWidth = newBlockWidth;
mBlockHeight = newBlockHeight;
}
private void applyAlignment(int widthAlignment, int heightAlignment) {
checkPowerOfTwo(widthAlignment, "widthAlignment must be a power of two");
checkPowerOfTwo(heightAlignment, "heightAlignment must be a power of two");
if (widthAlignment > mBlockWidth || heightAlignment > mBlockHeight) {
// maintain assumption that 0 < alignment <= block-size
applyBlockLimits(
Math.max(widthAlignment, mBlockWidth),
Math.max(heightAlignment, mBlockHeight),
POSITIVE_INTEGERS, POSITIVE_LONGS, POSITIVE_RATIONALS);
}
mWidthAlignment = Math.max(widthAlignment, mWidthAlignment);
mHeightAlignment = Math.max(heightAlignment, mHeightAlignment);
mWidthRange = Utils.alignRange(mWidthRange, mWidthAlignment);
mHeightRange = Utils.alignRange(mHeightRange, mHeightAlignment);
}
private void updateLimits() {
// pixels -> blocks <- counts
mHorizontalBlockRange = mHorizontalBlockRange.intersect(
Utils.factorRange(mWidthRange, mBlockWidth));
mHorizontalBlockRange = mHorizontalBlockRange.intersect(
Range.create(
mBlockCountRange.getLower() / mVerticalBlockRange.getUpper(),
mBlockCountRange.getUpper() / mVerticalBlockRange.getLower()));
mVerticalBlockRange = mVerticalBlockRange.intersect(
Utils.factorRange(mHeightRange, mBlockHeight));
mVerticalBlockRange = mVerticalBlockRange.intersect(
Range.create(
mBlockCountRange.getLower() / mHorizontalBlockRange.getUpper(),
mBlockCountRange.getUpper() / mHorizontalBlockRange.getLower()));
mBlockCountRange = mBlockCountRange.intersect(
Range.create(
mHorizontalBlockRange.getLower()
* mVerticalBlockRange.getLower(),
mHorizontalBlockRange.getUpper()
* mVerticalBlockRange.getUpper()));
mBlockAspectRatioRange = mBlockAspectRatioRange.intersect(
new Rational(
mHorizontalBlockRange.getLower(), mVerticalBlockRange.getUpper()),
new Rational(
mHorizontalBlockRange.getUpper(), mVerticalBlockRange.getLower()));
// blocks -> pixels
mWidthRange = mWidthRange.intersect(
(mHorizontalBlockRange.getLower() - 1) * mBlockWidth + mWidthAlignment,
mHorizontalBlockRange.getUpper() * mBlockWidth);
mHeightRange = mHeightRange.intersect(
(mVerticalBlockRange.getLower() - 1) * mBlockHeight + mHeightAlignment,
mVerticalBlockRange.getUpper() * mBlockHeight);
mAspectRatioRange = mAspectRatioRange.intersect(
new Rational(mWidthRange.getLower(), mHeightRange.getUpper()),
new Rational(mWidthRange.getUpper(), mHeightRange.getLower()));
mSmallerDimensionUpperLimit = Math.min(
mSmallerDimensionUpperLimit,
Math.min(mWidthRange.getUpper(), mHeightRange.getUpper()));
// blocks -> rate
mBlocksPerSecondRange = mBlocksPerSecondRange.intersect(
mBlockCountRange.getLower() * (long)mFrameRateRange.getLower(),
mBlockCountRange.getUpper() * (long)mFrameRateRange.getUpper());
mFrameRateRange = mFrameRateRange.intersect(
(int)(mBlocksPerSecondRange.getLower()
/ mBlockCountRange.getUpper()),
(int)(mBlocksPerSecondRange.getUpper()
/ (double)mBlockCountRange.getLower()));
}
private void applyMacroBlockLimits(
int maxHorizontalBlocks, int maxVerticalBlocks,
int maxBlocks, long maxBlocksPerSecond,
int blockWidth, int blockHeight,
int widthAlignment, int heightAlignment) {
applyMacroBlockLimits(
1 /* minHorizontalBlocks */, 1 /* minVerticalBlocks */,
maxHorizontalBlocks, maxVerticalBlocks,
maxBlocks, maxBlocksPerSecond,
blockWidth, blockHeight, widthAlignment, heightAlignment);
}
private void applyMacroBlockLimits(
int minHorizontalBlocks, int minVerticalBlocks,
int maxHorizontalBlocks, int maxVerticalBlocks,
int maxBlocks, long maxBlocksPerSecond,
int blockWidth, int blockHeight,
int widthAlignment, int heightAlignment) {
applyAlignment(widthAlignment, heightAlignment);
applyBlockLimits(
blockWidth, blockHeight, Range.create(1, maxBlocks),
Range.create(1L, maxBlocksPerSecond),
Range.create(
new Rational(1, maxVerticalBlocks),
new Rational(maxHorizontalBlocks, 1)));
mHorizontalBlockRange =
mHorizontalBlockRange.intersect(
Utils.divUp(minHorizontalBlocks, (mBlockWidth / blockWidth)),
maxHorizontalBlocks / (mBlockWidth / blockWidth));
mVerticalBlockRange =
mVerticalBlockRange.intersect(
Utils.divUp(minVerticalBlocks, (mBlockHeight / blockHeight)),
maxVerticalBlocks / (mBlockHeight / blockHeight));
}
private void applyLevelLimits() {
long maxBlocksPerSecond = 0;
int maxBlocks = 0;
int maxBps = 0;
int maxDPBBlocks = 0;
int errors = ERROR_NONE_SUPPORTED;
CodecProfileLevel[] profileLevels = mParent.profileLevels;
String mime = mParent.getMimeType();
if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_AVC)) {
maxBlocks = 99;
maxBlocksPerSecond = 1485;
maxBps = 64000;
maxDPBBlocks = 396;
for (CodecProfileLevel profileLevel: profileLevels) {
int MBPS = 0, FS = 0, BR = 0, DPB = 0;
boolean supported = true;
switch (profileLevel.level) {
case CodecProfileLevel.AVCLevel1:
MBPS = 1485; FS = 99; BR = 64; DPB = 396; break;
case CodecProfileLevel.AVCLevel1b:
MBPS = 1485; FS = 99; BR = 128; DPB = 396; break;
case CodecProfileLevel.AVCLevel11:
MBPS = 3000; FS = 396; BR = 192; DPB = 900; break;
case CodecProfileLevel.AVCLevel12:
MBPS = 6000; FS = 396; BR = 384; DPB = 2376; break;
case CodecProfileLevel.AVCLevel13:
MBPS = 11880; FS = 396; BR = 768; DPB = 2376; break;
case CodecProfileLevel.AVCLevel2:
MBPS = 11880; FS = 396; BR = 2000; DPB = 2376; break;
case CodecProfileLevel.AVCLevel21:
MBPS = 19800; FS = 792; BR = 4000; DPB = 4752; break;
case CodecProfileLevel.AVCLevel22:
MBPS = 20250; FS = 1620; BR = 4000; DPB = 8100; break;
case CodecProfileLevel.AVCLevel3:
MBPS = 40500; FS = 1620; BR = 10000; DPB = 8100; break;
case CodecProfileLevel.AVCLevel31:
MBPS = 108000; FS = 3600; BR = 14000; DPB = 18000; break;
case CodecProfileLevel.AVCLevel32:
MBPS = 216000; FS = 5120; BR = 20000; DPB = 20480; break;
case CodecProfileLevel.AVCLevel4:
MBPS = 245760; FS = 8192; BR = 20000; DPB = 32768; break;
case CodecProfileLevel.AVCLevel41:
MBPS = 245760; FS = 8192; BR = 50000; DPB = 32768; break;
case CodecProfileLevel.AVCLevel42:
MBPS = 522240; FS = 8704; BR = 50000; DPB = 34816; break;
case CodecProfileLevel.AVCLevel5:
MBPS = 589824; FS = 22080; BR = 135000; DPB = 110400; break;
case CodecProfileLevel.AVCLevel51:
MBPS = 983040; FS = 36864; BR = 240000; DPB = 184320; break;
case CodecProfileLevel.AVCLevel52:
MBPS = 2073600; FS = 36864; BR = 240000; DPB = 184320; break;
default:
Log.w(TAG, "Unrecognized level "
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
switch (profileLevel.profile) {
case CodecProfileLevel.AVCProfileHigh:
BR *= 1250; break;
case CodecProfileLevel.AVCProfileHigh10:
BR *= 3000; break;
case CodecProfileLevel.AVCProfileExtended:
case CodecProfileLevel.AVCProfileHigh422:
case CodecProfileLevel.AVCProfileHigh444:
Log.w(TAG, "Unsupported profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNSUPPORTED;
supported = false;
// fall through - treat as base profile
case CodecProfileLevel.AVCProfileBaseline:
case CodecProfileLevel.AVCProfileMain:
BR *= 1000; break;
default:
Log.w(TAG, "Unrecognized profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
BR *= 1000;
}
if (supported) {
errors &= ~ERROR_NONE_SUPPORTED;
}
maxBlocksPerSecond = Math.max(MBPS, maxBlocksPerSecond);
maxBlocks = Math.max(FS, maxBlocks);
maxBps = Math.max(BR, maxBps);
maxDPBBlocks = Math.max(maxDPBBlocks, DPB);
}
int maxLengthInBlocks = (int)(Math.sqrt(maxBlocks * 8));
applyMacroBlockLimits(
maxLengthInBlocks, maxLengthInBlocks,
maxBlocks, maxBlocksPerSecond,
16 /* blockWidth */, 16 /* blockHeight */,
1 /* widthAlignment */, 1 /* heightAlignment */);
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_MPEG2)) {
int maxWidth = 11, maxHeight = 9, maxRate = 15;
maxBlocks = 99;
maxBlocksPerSecond = 1485;
maxBps = 64000;
for (CodecProfileLevel profileLevel: profileLevels) {
int MBPS = 0, FS = 0, BR = 0, FR = 0, W = 0, H = 0;
boolean supported = true;
switch (profileLevel.profile) {
case CodecProfileLevel.MPEG2ProfileSimple:
switch (profileLevel.level) {
case CodecProfileLevel.MPEG2LevelML:
FR = 30; W = 45; H = 36; MBPS = 40500; FS = 1620; BR = 15000; break;
default:
Log.w(TAG, "Unrecognized profile/level "
+ profileLevel.profile + "/"
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
break;
case CodecProfileLevel.MPEG2ProfileMain:
switch (profileLevel.level) {
case CodecProfileLevel.MPEG2LevelLL:
FR = 30; W = 22; H = 18; MBPS = 11880; FS = 396; BR = 4000; break;
case CodecProfileLevel.MPEG2LevelML:
FR = 30; W = 45; H = 36; MBPS = 40500; FS = 1620; BR = 15000; break;
case CodecProfileLevel.MPEG2LevelH14:
FR = 60; W = 90; H = 68; MBPS = 183600; FS = 6120; BR = 60000; break;
case CodecProfileLevel.MPEG2LevelHL:
FR = 60; W = 120; H = 68; MBPS = 244800; FS = 8160; BR = 80000; break;
case CodecProfileLevel.MPEG2LevelHP:
FR = 60; W = 120; H = 68; MBPS = 489600; FS = 8160; BR = 80000; break;
default:
Log.w(TAG, "Unrecognized profile/level "
+ profileLevel.profile + "/"
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
break;
case CodecProfileLevel.MPEG2Profile422:
case CodecProfileLevel.MPEG2ProfileSNR:
case CodecProfileLevel.MPEG2ProfileSpatial:
case CodecProfileLevel.MPEG2ProfileHigh:
Log.i(TAG, "Unsupported profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNSUPPORTED;
supported = false;
break;
default:
Log.w(TAG, "Unrecognized profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
if (supported) {
errors &= ~ERROR_NONE_SUPPORTED;
}
maxBlocksPerSecond = Math.max(MBPS, maxBlocksPerSecond);
maxBlocks = Math.max(FS, maxBlocks);
maxBps = Math.max(BR * 1000, maxBps);
maxWidth = Math.max(W, maxWidth);
maxHeight = Math.max(H, maxHeight);
maxRate = Math.max(FR, maxRate);
}
applyMacroBlockLimits(maxWidth, maxHeight,
maxBlocks, maxBlocksPerSecond,
16 /* blockWidth */, 16 /* blockHeight */,
1 /* widthAlignment */, 1 /* heightAlignment */);
mFrameRateRange = mFrameRateRange.intersect(12, maxRate);
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_MPEG4)) {
int maxWidth = 11, maxHeight = 9, maxRate = 15;
maxBlocks = 99;
maxBlocksPerSecond = 1485;
maxBps = 64000;
for (CodecProfileLevel profileLevel: profileLevels) {
int MBPS = 0, FS = 0, BR = 0, FR = 0, W = 0, H = 0;
boolean strict = false; // true: W, H and FR are individual max limits
boolean supported = true;
switch (profileLevel.profile) {
case CodecProfileLevel.MPEG4ProfileSimple:
switch (profileLevel.level) {
case CodecProfileLevel.MPEG4Level0:
strict = true;
FR = 15; W = 11; H = 9; MBPS = 1485; FS = 99; BR = 64; break;
case CodecProfileLevel.MPEG4Level1:
FR = 30; W = 11; H = 9; MBPS = 1485; FS = 99; BR = 64; break;
case CodecProfileLevel.MPEG4Level0b:
strict = true;
FR = 15; W = 11; H = 9; MBPS = 1485; FS = 99; BR = 128; break;
case CodecProfileLevel.MPEG4Level2:
FR = 30; W = 22; H = 18; MBPS = 5940; FS = 396; BR = 128; break;
case CodecProfileLevel.MPEG4Level3:
FR = 30; W = 22; H = 18; MBPS = 11880; FS = 396; BR = 384; break;
case CodecProfileLevel.MPEG4Level4a:
FR = 30; W = 40; H = 30; MBPS = 36000; FS = 1200; BR = 4000; break;
case CodecProfileLevel.MPEG4Level5:
FR = 30; W = 45; H = 36; MBPS = 40500; FS = 1620; BR = 8000; break;
case CodecProfileLevel.MPEG4Level6:
FR = 30; W = 80; H = 45; MBPS = 108000; FS = 3600; BR = 12000; break;
default:
Log.w(TAG, "Unrecognized profile/level "
+ profileLevel.profile + "/"
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
break;
case CodecProfileLevel.MPEG4ProfileAdvancedSimple:
switch (profileLevel.level) {
case CodecProfileLevel.MPEG4Level0:
case CodecProfileLevel.MPEG4Level1:
FR = 30; W = 11; H = 9; MBPS = 2970; FS = 99; BR = 128; break;
case CodecProfileLevel.MPEG4Level2:
FR = 30; W = 22; H = 18; MBPS = 5940; FS = 396; BR = 384; break;
case CodecProfileLevel.MPEG4Level3:
FR = 30; W = 22; H = 18; MBPS = 11880; FS = 396; BR = 768; break;
case CodecProfileLevel.MPEG4Level3b:
FR = 30; W = 22; H = 18; MBPS = 11880; FS = 396; BR = 1500; break;
case CodecProfileLevel.MPEG4Level4:
FR = 30; W = 44; H = 36; MBPS = 23760; FS = 792; BR = 3000; break;
case CodecProfileLevel.MPEG4Level5:
FR = 30; W = 45; H = 36; MBPS = 48600; FS = 1620; BR = 8000; break;
default:
Log.w(TAG, "Unrecognized profile/level "
+ profileLevel.profile + "/"
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
break;
case CodecProfileLevel.MPEG4ProfileMain: // 2-4
case CodecProfileLevel.MPEG4ProfileNbit: // 2
case CodecProfileLevel.MPEG4ProfileAdvancedRealTime: // 1-4
case CodecProfileLevel.MPEG4ProfileCoreScalable: // 1-3
case CodecProfileLevel.MPEG4ProfileAdvancedCoding: // 1-4
case CodecProfileLevel.MPEG4ProfileCore: // 1-2
case CodecProfileLevel.MPEG4ProfileAdvancedCore: // 1-4
case CodecProfileLevel.MPEG4ProfileSimpleScalable: // 0-2
case CodecProfileLevel.MPEG4ProfileHybrid: // 1-2
// Studio profiles are not supported by our codecs.
// Only profiles that can decode simple object types are considered.
// The following profiles are not able to.
case CodecProfileLevel.MPEG4ProfileBasicAnimated: // 1-2
case CodecProfileLevel.MPEG4ProfileScalableTexture: // 1
case CodecProfileLevel.MPEG4ProfileSimpleFace: // 1-2
case CodecProfileLevel.MPEG4ProfileAdvancedScalable: // 1-3
case CodecProfileLevel.MPEG4ProfileSimpleFBA: // 1-2
Log.i(TAG, "Unsupported profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNSUPPORTED;
supported = false;
break;
default:
Log.w(TAG, "Unrecognized profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
if (supported) {
errors &= ~ERROR_NONE_SUPPORTED;
}
maxBlocksPerSecond = Math.max(MBPS, maxBlocksPerSecond);
maxBlocks = Math.max(FS, maxBlocks);
maxBps = Math.max(BR * 1000, maxBps);
if (strict) {
maxWidth = Math.max(W, maxWidth);
maxHeight = Math.max(H, maxHeight);
maxRate = Math.max(FR, maxRate);
} else {
// assuming max 60 fps frame rate and 1:2 aspect ratio
int maxDim = (int)Math.sqrt(FS * 2);
maxWidth = Math.max(maxDim, maxWidth);
maxHeight = Math.max(maxDim, maxHeight);
maxRate = Math.max(Math.max(FR, 60), maxRate);
}
}
applyMacroBlockLimits(maxWidth, maxHeight,
maxBlocks, maxBlocksPerSecond,
16 /* blockWidth */, 16 /* blockHeight */,
1 /* widthAlignment */, 1 /* heightAlignment */);
mFrameRateRange = mFrameRateRange.intersect(12, maxRate);
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_H263)) {
int maxWidth = 11, maxHeight = 9, maxRate = 15;
int minWidth = maxWidth, minHeight = maxHeight;
int minAlignment = 16;
maxBlocks = 99;
maxBlocksPerSecond = 1485;
maxBps = 64000;
for (CodecProfileLevel profileLevel: profileLevels) {
int MBPS = 0, BR = 0, FR = 0, W = 0, H = 0, minW = minWidth, minH = minHeight;
boolean strict = false; // true: support only sQCIF, QCIF (maybe CIF)
switch (profileLevel.level) {
case CodecProfileLevel.H263Level10:
strict = true; // only supports sQCIF & QCIF
FR = 15; W = 11; H = 9; BR = 1; MBPS = W * H * FR; break;
case CodecProfileLevel.H263Level20:
strict = true; // only supports sQCIF, QCIF & CIF
FR = 30; W = 22; H = 18; BR = 2; MBPS = W * H * 15; break;
case CodecProfileLevel.H263Level30:
strict = true; // only supports sQCIF, QCIF & CIF
FR = 30; W = 22; H = 18; BR = 6; MBPS = W * H * FR; break;
case CodecProfileLevel.H263Level40:
strict = true; // only supports sQCIF, QCIF & CIF
FR = 30; W = 22; H = 18; BR = 32; MBPS = W * H * FR; break;
case CodecProfileLevel.H263Level45:
// only implies level 10 support
strict = profileLevel.profile == CodecProfileLevel.H263ProfileBaseline
|| profileLevel.profile ==
CodecProfileLevel.H263ProfileBackwardCompatible;
if (!strict) {
minW = 1; minH = 1; minAlignment = 4;
}
FR = 15; W = 11; H = 9; BR = 2; MBPS = W * H * FR; break;
case CodecProfileLevel.H263Level50:
// only supports 50fps for H > 15
minW = 1; minH = 1; minAlignment = 4;
FR = 60; W = 22; H = 18; BR = 64; MBPS = W * H * 50; break;
case CodecProfileLevel.H263Level60:
// only supports 50fps for H > 15
minW = 1; minH = 1; minAlignment = 4;
FR = 60; W = 45; H = 18; BR = 128; MBPS = W * H * 50; break;
case CodecProfileLevel.H263Level70:
// only supports 50fps for H > 30
minW = 1; minH = 1; minAlignment = 4;
FR = 60; W = 45; H = 36; BR = 256; MBPS = W * H * 50; break;
default:
Log.w(TAG, "Unrecognized profile/level " + profileLevel.profile
+ "/" + profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
switch (profileLevel.profile) {
case CodecProfileLevel.H263ProfileBackwardCompatible:
case CodecProfileLevel.H263ProfileBaseline:
case CodecProfileLevel.H263ProfileH320Coding:
case CodecProfileLevel.H263ProfileHighCompression:
case CodecProfileLevel.H263ProfileHighLatency:
case CodecProfileLevel.H263ProfileInterlace:
case CodecProfileLevel.H263ProfileInternet:
case CodecProfileLevel.H263ProfileISWV2:
case CodecProfileLevel.H263ProfileISWV3:
break;
default:
Log.w(TAG, "Unrecognized profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
if (strict) {
// Strict levels define sub-QCIF min size and enumerated sizes. We cannot
// express support for "only sQCIF & QCIF (& CIF)" using VideoCapabilities
// but we can express "only QCIF (& CIF)", so set minimume size at QCIF.
// minW = 8; minH = 6;
minW = 11; minH = 9;
} else {
// any support for non-strict levels (including unrecognized profiles or
// levels) allow custom frame size support beyond supported limits
// (other than bitrate)
mAllowMbOverride = true;
}
errors &= ~ERROR_NONE_SUPPORTED;
maxBlocksPerSecond = Math.max(MBPS, maxBlocksPerSecond);
maxBlocks = Math.max(W * H, maxBlocks);
maxBps = Math.max(BR * 64000, maxBps);
maxWidth = Math.max(W, maxWidth);
maxHeight = Math.max(H, maxHeight);
maxRate = Math.max(FR, maxRate);
minWidth = Math.min(minW, minWidth);
minHeight = Math.min(minH, minHeight);
}
// unless we encountered custom frame size support, limit size to QCIF and CIF
// using aspect ratio.
if (!mAllowMbOverride) {
mBlockAspectRatioRange =
Range.create(new Rational(11, 9), new Rational(11, 9));
}
applyMacroBlockLimits(
minWidth, minHeight,
maxWidth, maxHeight,
maxBlocks, maxBlocksPerSecond,
16 /* blockWidth */, 16 /* blockHeight */,
minAlignment /* widthAlignment */, minAlignment /* heightAlignment */);
mFrameRateRange = Range.create(1, maxRate);
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_VP8)) {
maxBlocks = Integer.MAX_VALUE;
maxBlocksPerSecond = Integer.MAX_VALUE;
// TODO: set to 100Mbps for now, need a number for VP8
maxBps = 100000000;
// profile levels are not indicative for VPx, but verify
// them nonetheless
for (CodecProfileLevel profileLevel: profileLevels) {
switch (profileLevel.level) {
case CodecProfileLevel.VP8Level_Version0:
case CodecProfileLevel.VP8Level_Version1:
case CodecProfileLevel.VP8Level_Version2:
case CodecProfileLevel.VP8Level_Version3:
break;
default:
Log.w(TAG, "Unrecognized level "
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
switch (profileLevel.profile) {
case CodecProfileLevel.VP8ProfileMain:
break;
default:
Log.w(TAG, "Unrecognized profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
errors &= ~ERROR_NONE_SUPPORTED;
}
final int blockSize = 16;
applyMacroBlockLimits(Short.MAX_VALUE, Short.MAX_VALUE,
maxBlocks, maxBlocksPerSecond, blockSize, blockSize,
1 /* widthAlignment */, 1 /* heightAlignment */);
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_VP9)) {
maxBlocksPerSecond = 829440;
maxBlocks = 36864;
maxBps = 200000;
int maxDim = 512;
for (CodecProfileLevel profileLevel: profileLevels) {
long SR = 0; // luma sample rate
int FS = 0; // luma picture size
int BR = 0; // bit rate kbps
int D = 0; // luma dimension
switch (profileLevel.level) {
case CodecProfileLevel.VP9Level1:
SR = 829440; FS = 36864; BR = 200; D = 512; break;
case CodecProfileLevel.VP9Level11:
SR = 2764800; FS = 73728; BR = 800; D = 768; break;
case CodecProfileLevel.VP9Level2:
SR = 4608000; FS = 122880; BR = 1800; D = 960; break;
case CodecProfileLevel.VP9Level21:
SR = 9216000; FS = 245760; BR = 3600; D = 1344; break;
case CodecProfileLevel.VP9Level3:
SR = 20736000; FS = 552960; BR = 7200; D = 2048; break;
case CodecProfileLevel.VP9Level31:
SR = 36864000; FS = 983040; BR = 12000; D = 2752; break;
case CodecProfileLevel.VP9Level4:
SR = 83558400; FS = 2228224; BR = 18000; D = 4160; break;
case CodecProfileLevel.VP9Level41:
SR = 160432128; FS = 2228224; BR = 30000; D = 4160; break;
case CodecProfileLevel.VP9Level5:
SR = 311951360; FS = 8912896; BR = 60000; D = 8384; break;
case CodecProfileLevel.VP9Level51:
SR = 588251136; FS = 8912896; BR = 120000; D = 8384; break;
case CodecProfileLevel.VP9Level52:
SR = 1176502272; FS = 8912896; BR = 180000; D = 8384; break;
case CodecProfileLevel.VP9Level6:
SR = 1176502272; FS = 35651584; BR = 180000; D = 16832; break;
case CodecProfileLevel.VP9Level61:
SR = 2353004544L; FS = 35651584; BR = 240000; D = 16832; break;
case CodecProfileLevel.VP9Level62:
SR = 4706009088L; FS = 35651584; BR = 480000; D = 16832; break;
default:
Log.w(TAG, "Unrecognized level "
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
switch (profileLevel.profile) {
case CodecProfileLevel.VP9Profile0:
case CodecProfileLevel.VP9Profile1:
case CodecProfileLevel.VP9Profile2:
case CodecProfileLevel.VP9Profile3:
case CodecProfileLevel.VP9Profile2HDR:
case CodecProfileLevel.VP9Profile3HDR:
break;
default:
Log.w(TAG, "Unrecognized profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
errors &= ~ERROR_NONE_SUPPORTED;
maxBlocksPerSecond = Math.max(SR, maxBlocksPerSecond);
maxBlocks = Math.max(FS, maxBlocks);
maxBps = Math.max(BR * 1000, maxBps);
maxDim = Math.max(D, maxDim);
}
final int blockSize = 8;
int maxLengthInBlocks = Utils.divUp(maxDim, blockSize);
maxBlocks = Utils.divUp(maxBlocks, blockSize * blockSize);
maxBlocksPerSecond = Utils.divUp(maxBlocksPerSecond, blockSize * blockSize);
applyMacroBlockLimits(
maxLengthInBlocks, maxLengthInBlocks,
maxBlocks, maxBlocksPerSecond,
blockSize, blockSize,
1 /* widthAlignment */, 1 /* heightAlignment */);
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_VIDEO_HEVC)) {
// CTBs are at least 8x8 so use 8x8 block size
maxBlocks = 36864 >> 6; // 192x192 pixels == 576 8x8 blocks
maxBlocksPerSecond = maxBlocks * 15;
maxBps = 128000;
for (CodecProfileLevel profileLevel: profileLevels) {
double FR = 0;
int FS = 0;
int BR = 0;
switch (profileLevel.level) {
/* The HEVC spec talks only in a very convoluted manner about the
existence of levels 1-3.1 for High tier, which could also be
understood as 'decoders and encoders should treat these levels
as if they were Main tier', so we do that. */
case CodecProfileLevel.HEVCMainTierLevel1:
case CodecProfileLevel.HEVCHighTierLevel1:
FR = 15; FS = 36864; BR = 128; break;
case CodecProfileLevel.HEVCMainTierLevel2:
case CodecProfileLevel.HEVCHighTierLevel2:
FR = 30; FS = 122880; BR = 1500; break;
case CodecProfileLevel.HEVCMainTierLevel21:
case CodecProfileLevel.HEVCHighTierLevel21:
FR = 30; FS = 245760; BR = 3000; break;
case CodecProfileLevel.HEVCMainTierLevel3:
case CodecProfileLevel.HEVCHighTierLevel3:
FR = 30; FS = 552960; BR = 6000; break;
case CodecProfileLevel.HEVCMainTierLevel31:
case CodecProfileLevel.HEVCHighTierLevel31:
FR = 33.75; FS = 983040; BR = 10000; break;
case CodecProfileLevel.HEVCMainTierLevel4:
FR = 30; FS = 2228224; BR = 12000; break;
case CodecProfileLevel.HEVCHighTierLevel4:
FR = 30; FS = 2228224; BR = 30000; break;
case CodecProfileLevel.HEVCMainTierLevel41:
FR = 60; FS = 2228224; BR = 20000; break;
case CodecProfileLevel.HEVCHighTierLevel41:
FR = 60; FS = 2228224; BR = 50000; break;
case CodecProfileLevel.HEVCMainTierLevel5:
FR = 30; FS = 8912896; BR = 25000; break;
case CodecProfileLevel.HEVCHighTierLevel5:
FR = 30; FS = 8912896; BR = 100000; break;
case CodecProfileLevel.HEVCMainTierLevel51:
FR = 60; FS = 8912896; BR = 40000; break;
case CodecProfileLevel.HEVCHighTierLevel51:
FR = 60; FS = 8912896; BR = 160000; break;
case CodecProfileLevel.HEVCMainTierLevel52:
FR = 120; FS = 8912896; BR = 60000; break;
case CodecProfileLevel.HEVCHighTierLevel52:
FR = 120; FS = 8912896; BR = 240000; break;
case CodecProfileLevel.HEVCMainTierLevel6:
FR = 30; FS = 35651584; BR = 60000; break;
case CodecProfileLevel.HEVCHighTierLevel6:
FR = 30; FS = 35651584; BR = 240000; break;
case CodecProfileLevel.HEVCMainTierLevel61:
FR = 60; FS = 35651584; BR = 120000; break;
case CodecProfileLevel.HEVCHighTierLevel61:
FR = 60; FS = 35651584; BR = 480000; break;
case CodecProfileLevel.HEVCMainTierLevel62:
FR = 120; FS = 35651584; BR = 240000; break;
case CodecProfileLevel.HEVCHighTierLevel62:
FR = 120; FS = 35651584; BR = 800000; break;
default:
Log.w(TAG, "Unrecognized level "
+ profileLevel.level + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
switch (profileLevel.profile) {
case CodecProfileLevel.HEVCProfileMain:
case CodecProfileLevel.HEVCProfileMain10:
case CodecProfileLevel.HEVCProfileMain10HDR10:
break;
default:
Log.w(TAG, "Unrecognized profile "
+ profileLevel.profile + " for " + mime);
errors |= ERROR_UNRECOGNIZED;
}
/* DPB logic:
if (width * height <= FS / 4) DPB = 16;
else if (width * height <= FS / 2) DPB = 12;
else if (width * height <= FS * 0.75) DPB = 8;
else DPB = 6;
*/
FS >>= 6; // convert pixels to blocks
errors &= ~ERROR_NONE_SUPPORTED;
maxBlocksPerSecond = Math.max((int)(FR * FS), maxBlocksPerSecond);
maxBlocks = Math.max(FS, maxBlocks);
maxBps = Math.max(BR * 1000, maxBps);
}
int maxLengthInBlocks = (int)(Math.sqrt(maxBlocks * 8));
applyMacroBlockLimits(
maxLengthInBlocks, maxLengthInBlocks,
maxBlocks, maxBlocksPerSecond,
8 /* blockWidth */, 8 /* blockHeight */,
1 /* widthAlignment */, 1 /* heightAlignment */);
} else {
Log.w(TAG, "Unsupported mime " + mime);
// using minimal bitrate here. should be overriden by
// info from media_codecs.xml
maxBps = 64000;
errors |= ERROR_UNSUPPORTED;
}
mBitrateRange = Range.create(1, maxBps);
mParent.mError |= errors;
}
}
/**
* A class that supports querying the encoding capabilities of a codec.
*/
public static final class EncoderCapabilities {
/**
* Returns the supported range of quality values.
*
* @hide
*/
public Range<Integer> getQualityRange() {
return mQualityRange;
}
/**
* Returns the supported range of encoder complexity values.
* <p>
* Some codecs may support multiple complexity levels, where higher
* complexity values use more encoder tools (e.g. perform more
* intensive calculations) to improve the quality or the compression
* ratio. Use a lower value to save power and/or time.
*/
public Range<Integer> getComplexityRange() {
return mComplexityRange;
}
/** Constant quality mode */
public static final int BITRATE_MODE_CQ = 0;
/** Variable bitrate mode */
public static final int BITRATE_MODE_VBR = 1;
/** Constant bitrate mode */
public static final int BITRATE_MODE_CBR = 2;
private static final Feature[] bitrates = new Feature[] {
new Feature("VBR", BITRATE_MODE_VBR, true),
new Feature("CBR", BITRATE_MODE_CBR, false),
new Feature("CQ", BITRATE_MODE_CQ, false)
};
private static int parseBitrateMode(String mode) {
for (Feature feat: bitrates) {
if (feat.mName.equalsIgnoreCase(mode)) {
return feat.mValue;
}
}
return 0;
}
/**
* Query whether a bitrate mode is supported.
*/
public boolean isBitrateModeSupported(int mode) {
for (Feature feat: bitrates) {
if (mode == feat.mValue) {
return (mBitControl & (1 << mode)) != 0;
}
}
return false;
}
private Range<Integer> mQualityRange;
private Range<Integer> mComplexityRange;
private CodecCapabilities mParent;
/* no public constructor */
private EncoderCapabilities() { }
/** @hide */
public static EncoderCapabilities create(
MediaFormat info, CodecCapabilities parent) {
EncoderCapabilities caps = new EncoderCapabilities();
caps.init(info, parent);
return caps;
}
/** @hide */
public void init(MediaFormat info, CodecCapabilities parent) {
// no support for complexity or quality yet
mParent = parent;
mComplexityRange = Range.create(0, 0);
mQualityRange = Range.create(0, 0);
mBitControl = (1 << BITRATE_MODE_VBR);
applyLevelLimits();
parseFromInfo(info);
}
private void applyLevelLimits() {
String mime = mParent.getMimeType();
if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_FLAC)) {
mComplexityRange = Range.create(0, 8);
mBitControl = (1 << BITRATE_MODE_CQ);
} else if (mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_AMR_NB)
|| mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_AMR_WB)
|| mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_G711_ALAW)
|| mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_G711_MLAW)
|| mime.equalsIgnoreCase(MediaFormat.MIMETYPE_AUDIO_MSGSM)) {
mBitControl = (1 << BITRATE_MODE_CBR);
}
}
private int mBitControl;
private Integer mDefaultComplexity;
private Integer mDefaultQuality;
private String mQualityScale;
private void parseFromInfo(MediaFormat info) {
Map<String, Object> map = info.getMap();
if (info.containsKey("complexity-range")) {
mComplexityRange = Utils
.parseIntRange(info.getString("complexity-range"), mComplexityRange);
// TODO should we limit this to level limits?
}
if (info.containsKey("quality-range")) {
mQualityRange = Utils
.parseIntRange(info.getString("quality-range"), mQualityRange);
}
if (info.containsKey("feature-bitrate-control")) {
for (String mode: info.getString("feature-bitrate-control").split(",")) {
mBitControl |= parseBitrateMode(mode);
}
}
try {
mDefaultComplexity = Integer.parseInt((String)map.get("complexity-default"));
} catch (NumberFormatException e) { }
try {
mDefaultQuality = Integer.parseInt((String)map.get("quality-default"));
} catch (NumberFormatException e) { }
mQualityScale = (String)map.get("quality-scale");
}
private boolean supports(
Integer complexity, Integer quality, Integer profile) {
boolean ok = true;
if (ok && complexity != null) {
ok = mComplexityRange.contains(complexity);
}
if (ok && quality != null) {
ok = mQualityRange.contains(quality);
}
if (ok && profile != null) {
for (CodecProfileLevel pl: mParent.profileLevels) {
if (pl.profile == profile) {
profile = null;
break;
}
}
ok = profile == null;
}
return ok;
}
/** @hide */
public void setDefaultFormat(MediaFormat format) {
// don't list trivial quality/complexity as default for now
if (!mQualityRange.getUpper().equals(mQualityRange.getLower())
&& mDefaultQuality != null) {
format.setInteger(MediaFormat.KEY_QUALITY, mDefaultQuality);
}
if (!mComplexityRange.getUpper().equals(mComplexityRange.getLower())
&& mDefaultComplexity != null) {
format.setInteger(MediaFormat.KEY_COMPLEXITY, mDefaultComplexity);
}
// bitrates are listed in order of preference
for (Feature feat: bitrates) {
if ((mBitControl & (1 << feat.mValue)) != 0) {
format.setInteger(MediaFormat.KEY_BITRATE_MODE, feat.mValue);
break;
}
}
}
/** @hide */
public boolean supportsFormat(MediaFormat format) {
final Map<String, Object> map = format.getMap();
final String mime = mParent.getMimeType();
Integer mode = (Integer)map.get(MediaFormat.KEY_BITRATE_MODE);
if (mode != null && !isBitrateModeSupported(mode)) {
return false;
}
Integer complexity = (Integer)map.get(MediaFormat.KEY_COMPLEXITY);
if (MediaFormat.MIMETYPE_AUDIO_FLAC.equalsIgnoreCase(mime)) {
Integer flacComplexity =
(Integer)map.get(MediaFormat.KEY_FLAC_COMPRESSION_LEVEL);
if (complexity == null) {
complexity = flacComplexity;
} else if (flacComplexity != null && !complexity.equals(flacComplexity)) {
throw new IllegalArgumentException(
"conflicting values for complexity and " +
"flac-compression-level");
}
}
// other audio parameters
Integer profile = (Integer)map.get(MediaFormat.KEY_PROFILE);
if (MediaFormat.MIMETYPE_AUDIO_AAC.equalsIgnoreCase(mime)) {
Integer aacProfile = (Integer)map.get(MediaFormat.KEY_AAC_PROFILE);
if (profile == null) {
profile = aacProfile;
} else if (aacProfile != null && !aacProfile.equals(profile)) {
throw new IllegalArgumentException(
"conflicting values for profile and aac-profile");
}
}
Integer quality = (Integer)map.get(MediaFormat.KEY_QUALITY);
return supports(complexity, quality, profile);
}
};
/**
* Encapsulates the profiles available for a codec component.
* <p>You can get a set of {@link MediaCodecInfo.CodecProfileLevel} objects for a given
* {@link MediaCodecInfo} object from the
* {@link MediaCodecInfo.CodecCapabilities#profileLevels} field.
*/
public static final class CodecProfileLevel {
// from OMX_VIDEO_AVCPROFILETYPE
public static final int AVCProfileBaseline = 0x01;
public static final int AVCProfileMain = 0x02;
public static final int AVCProfileExtended = 0x04;
public static final int AVCProfileHigh = 0x08;
public static final int AVCProfileHigh10 = 0x10;
public static final int AVCProfileHigh422 = 0x20;
public static final int AVCProfileHigh444 = 0x40;
// from OMX_VIDEO_AVCLEVELTYPE
public static final int AVCLevel1 = 0x01;
public static final int AVCLevel1b = 0x02;
public static final int AVCLevel11 = 0x04;
public static final int AVCLevel12 = 0x08;
public static final int AVCLevel13 = 0x10;
public static final int AVCLevel2 = 0x20;
public static final int AVCLevel21 = 0x40;
public static final int AVCLevel22 = 0x80;
public static final int AVCLevel3 = 0x100;
public static final int AVCLevel31 = 0x200;
public static final int AVCLevel32 = 0x400;
public static final int AVCLevel4 = 0x800;
public static final int AVCLevel41 = 0x1000;
public static final int AVCLevel42 = 0x2000;
public static final int AVCLevel5 = 0x4000;
public static final int AVCLevel51 = 0x8000;
public static final int AVCLevel52 = 0x10000;
// from OMX_VIDEO_H263PROFILETYPE
public static final int H263ProfileBaseline = 0x01;
public static final int H263ProfileH320Coding = 0x02;
public static final int H263ProfileBackwardCompatible = 0x04;
public static final int H263ProfileISWV2 = 0x08;
public static final int H263ProfileISWV3 = 0x10;
public static final int H263ProfileHighCompression = 0x20;
public static final int H263ProfileInternet = 0x40;
public static final int H263ProfileInterlace = 0x80;
public static final int H263ProfileHighLatency = 0x100;
// from OMX_VIDEO_H263LEVELTYPE
public static final int H263Level10 = 0x01;
public static final int H263Level20 = 0x02;
public static final int H263Level30 = 0x04;
public static final int H263Level40 = 0x08;
public static final int H263Level45 = 0x10;
public static final int H263Level50 = 0x20;
public static final int H263Level60 = 0x40;
public static final int H263Level70 = 0x80;
// from OMX_VIDEO_MPEG4PROFILETYPE
public static final int MPEG4ProfileSimple = 0x01;
public static final int MPEG4ProfileSimpleScalable = 0x02;
public static final int MPEG4ProfileCore = 0x04;
public static final int MPEG4ProfileMain = 0x08;
public static final int MPEG4ProfileNbit = 0x10;
public static final int MPEG4ProfileScalableTexture = 0x20;
public static final int MPEG4ProfileSimpleFace = 0x40;
public static final int MPEG4ProfileSimpleFBA = 0x80;
public static final int MPEG4ProfileBasicAnimated = 0x100;
public static final int MPEG4ProfileHybrid = 0x200;
public static final int MPEG4ProfileAdvancedRealTime = 0x400;
public static final int MPEG4ProfileCoreScalable = 0x800;
public static final int MPEG4ProfileAdvancedCoding = 0x1000;
public static final int MPEG4ProfileAdvancedCore = 0x2000;
public static final int MPEG4ProfileAdvancedScalable = 0x4000;
public static final int MPEG4ProfileAdvancedSimple = 0x8000;
// from OMX_VIDEO_MPEG4LEVELTYPE
public static final int MPEG4Level0 = 0x01;
public static final int MPEG4Level0b = 0x02;
public static final int MPEG4Level1 = 0x04;
public static final int MPEG4Level2 = 0x08;
public static final int MPEG4Level3 = 0x10;
public static final int MPEG4Level3b = 0x18;
public static final int MPEG4Level4 = 0x20;
public static final int MPEG4Level4a = 0x40;
public static final int MPEG4Level5 = 0x80;
public static final int MPEG4Level6 = 0x100;
// from OMX_VIDEO_MPEG2PROFILETYPE
public static final int MPEG2ProfileSimple = 0x00;
public static final int MPEG2ProfileMain = 0x01;
public static final int MPEG2Profile422 = 0x02;
public static final int MPEG2ProfileSNR = 0x03;
public static final int MPEG2ProfileSpatial = 0x04;
public static final int MPEG2ProfileHigh = 0x05;
// from OMX_VIDEO_MPEG2LEVELTYPE
public static final int MPEG2LevelLL = 0x00;
public static final int MPEG2LevelML = 0x01;
public static final int MPEG2LevelH14 = 0x02;
public static final int MPEG2LevelHL = 0x03;
public static final int MPEG2LevelHP = 0x04;
// from OMX_AUDIO_AACPROFILETYPE
public static final int AACObjectMain = 1;
public static final int AACObjectLC = 2;
public static final int AACObjectSSR = 3;
public static final int AACObjectLTP = 4;
public static final int AACObjectHE = 5;
public static final int AACObjectScalable = 6;
public static final int AACObjectERLC = 17;
public static final int AACObjectERScalable = 20;
public static final int AACObjectLD = 23;
public static final int AACObjectHE_PS = 29;
public static final int AACObjectELD = 39;
// from OMX_VIDEO_VP8LEVELTYPE
public static final int VP8Level_Version0 = 0x01;
public static final int VP8Level_Version1 = 0x02;
public static final int VP8Level_Version2 = 0x04;
public static final int VP8Level_Version3 = 0x08;
// from OMX_VIDEO_VP8PROFILETYPE
public static final int VP8ProfileMain = 0x01;
// from OMX_VIDEO_VP9PROFILETYPE
public static final int VP9Profile0 = 0x01;
public static final int VP9Profile1 = 0x02;
public static final int VP9Profile2 = 0x04;
public static final int VP9Profile3 = 0x08;
// HDR profiles also support passing HDR metadata
public static final int VP9Profile2HDR = 0x1000;
public static final int VP9Profile3HDR = 0x2000;
// from OMX_VIDEO_VP9LEVELTYPE
public static final int VP9Level1 = 0x1;
public static final int VP9Level11 = 0x2;
public static final int VP9Level2 = 0x4;
public static final int VP9Level21 = 0x8;
public static final int VP9Level3 = 0x10;
public static final int VP9Level31 = 0x20;
public static final int VP9Level4 = 0x40;
public static final int VP9Level41 = 0x80;
public static final int VP9Level5 = 0x100;
public static final int VP9Level51 = 0x200;
public static final int VP9Level52 = 0x400;
public static final int VP9Level6 = 0x800;
public static final int VP9Level61 = 0x1000;
public static final int VP9Level62 = 0x2000;
// from OMX_VIDEO_HEVCPROFILETYPE
public static final int HEVCProfileMain = 0x01;
public static final int HEVCProfileMain10 = 0x02;
public static final int HEVCProfileMain10HDR10 = 0x1000;
// from OMX_VIDEO_HEVCLEVELTYPE
public static final int HEVCMainTierLevel1 = 0x1;
public static final int HEVCHighTierLevel1 = 0x2;
public static final int HEVCMainTierLevel2 = 0x4;
public static final int HEVCHighTierLevel2 = 0x8;
public static final int HEVCMainTierLevel21 = 0x10;
public static final int HEVCHighTierLevel21 = 0x20;
public static final int HEVCMainTierLevel3 = 0x40;
public static final int HEVCHighTierLevel3 = 0x80;
public static final int HEVCMainTierLevel31 = 0x100;
public static final int HEVCHighTierLevel31 = 0x200;
public static final int HEVCMainTierLevel4 = 0x400;
public static final int HEVCHighTierLevel4 = 0x800;
public static final int HEVCMainTierLevel41 = 0x1000;
public static final int HEVCHighTierLevel41 = 0x2000;
public static final int HEVCMainTierLevel5 = 0x4000;
public static final int HEVCHighTierLevel5 = 0x8000;
public static final int HEVCMainTierLevel51 = 0x10000;
public static final int HEVCHighTierLevel51 = 0x20000;
public static final int HEVCMainTierLevel52 = 0x40000;
public static final int HEVCHighTierLevel52 = 0x80000;
public static final int HEVCMainTierLevel6 = 0x100000;
public static final int HEVCHighTierLevel6 = 0x200000;
public static final int HEVCMainTierLevel61 = 0x400000;
public static final int HEVCHighTierLevel61 = 0x800000;
public static final int HEVCMainTierLevel62 = 0x1000000;
public static final int HEVCHighTierLevel62 = 0x2000000;
private static final int HEVCHighTierLevels =
HEVCHighTierLevel1 | HEVCHighTierLevel2 | HEVCHighTierLevel21 | HEVCHighTierLevel3 |
HEVCHighTierLevel31 | HEVCHighTierLevel4 | HEVCHighTierLevel41 | HEVCHighTierLevel5 |
HEVCHighTierLevel51 | HEVCHighTierLevel52 | HEVCHighTierLevel6 | HEVCHighTierLevel61 |
HEVCHighTierLevel62;
// from OMX_VIDEO_DOLBYVISIONPROFILETYPE
public static final int DolbyVisionProfileDvavPer = 0x1;
public static final int DolbyVisionProfileDvavPen = 0x2;
public static final int DolbyVisionProfileDvheDer = 0x4;
public static final int DolbyVisionProfileDvheDen = 0x8;
public static final int DolbyVisionProfileDvheDtr = 0x10;
public static final int DolbyVisionProfileDvheStn = 0x20;
public static final int DolbyVisionProfileDvheDth = 0x40;
public static final int DolbyVisionProfileDvheDtb = 0x80;
// from OMX_VIDEO_DOLBYVISIONLEVELTYPE
public static final int DolbyVisionLevelHd24 = 0x1;
public static final int DolbyVisionLevelHd30 = 0x2;
public static final int DolbyVisionLevelFhd24 = 0x4;
public static final int DolbyVisionLevelFhd30 = 0x8;
public static final int DolbyVisionLevelFhd60 = 0x10;
public static final int DolbyVisionLevelUhd24 = 0x20;
public static final int DolbyVisionLevelUhd30 = 0x40;
public static final int DolbyVisionLevelUhd48 = 0x80;
public static final int DolbyVisionLevelUhd60 = 0x100;
/**
* Defined in the OpenMAX IL specs, depending on the type of media
* this can be OMX_VIDEO_AVCPROFILETYPE, OMX_VIDEO_H263PROFILETYPE,
* OMX_VIDEO_MPEG4PROFILETYPE, OMX_VIDEO_VP8PROFILETYPE or OMX_VIDEO_VP9PROFILETYPE.
*/
public int profile;
/**
* Defined in the OpenMAX IL specs, depending on the type of media
* this can be OMX_VIDEO_AVCLEVELTYPE, OMX_VIDEO_H263LEVELTYPE
* OMX_VIDEO_MPEG4LEVELTYPE, OMX_VIDEO_VP8LEVELTYPE or OMX_VIDEO_VP9LEVELTYPE.
*
* Note that VP9 decoder on platforms before {@link android.os.Build.VERSION_CODES#N} may
* not advertise a profile level support. For those VP9 decoders, please use
* {@link VideoCapabilities} to determine the codec capabilities.
*/
public int level;
};
/**
* Enumerates the capabilities of the codec component. Since a single
* component can support data of a variety of types, the type has to be
* specified to yield a meaningful result.
* @param type The MIME type to query
*/
public final CodecCapabilities getCapabilitiesForType(
String type) {
CodecCapabilities caps = mCaps.get(type);
if (caps == null) {
throw new IllegalArgumentException("codec does not support type");
}
// clone writable object
return caps.dup();
}
/** @hide */
public MediaCodecInfo makeRegular() {
ArrayList<CodecCapabilities> caps = new ArrayList<CodecCapabilities>();
for (CodecCapabilities c: mCaps.values()) {
if (c.isRegular()) {
caps.add(c);
}
}
if (caps.size() == 0) {
return null;
} else if (caps.size() == mCaps.size()) {
return this;
}
return new MediaCodecInfo(
mName, mIsEncoder,
caps.toArray(new CodecCapabilities[caps.size()]));
}
}