blob: bd4ac3881e0487d6ff47ddb1369209a266595c19 [file] [log] [blame]
/*
* Copyright (C) 2019 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
//#define LOG_NDEBUG 0
#define LOG_TAG "AudioAnalytics"
#include <android-base/logging.h>
#include <utils/Log.h>
#include "AudioAnalytics.h"
#include <aaudio/AAudio.h> // error codes
#include <audio_utils/clock.h> // clock conversions
#include <cutils/properties.h>
#include <stats_media_metrics.h> // statsd
#include <system/audio.h>
#include "AudioTypes.h" // string to int conversions
#include "MediaMetricsService.h" // package info
#include "StringUtils.h"
#include "ValidateId.h"
#define PROP_AUDIO_ANALYTICS_CLOUD_ENABLED "persist.audio.analytics.cloud.enabled"
namespace android::mediametrics {
// Enable for testing of delivery to statsd. Caution if this is enabled, all protos MUST exist.
#define STATSD_ENABLE
#ifdef STATSD_ENABLE
#define CONDITION(INT_VALUE) (INT_VALUE) // allow value
#else
#define CONDITION(INT_VALUE) (int(0)) // mask value since the proto may not be defined yet.
#endif
// Maximum length of a device name.
// static constexpr size_t STATSD_DEVICE_NAME_MAX_LENGTH = 32; // unused since we suppress
// Transmit Enums to statsd in integer or strings (this must match the atoms.proto)
static constexpr bool STATSD_USE_INT_FOR_ENUM = false;
// derive types based on integer or strings.
using short_enum_type_t = std::conditional_t<STATSD_USE_INT_FOR_ENUM, int32_t, std::string>;
using long_enum_type_t = std::conditional_t<STATSD_USE_INT_FOR_ENUM, int64_t, std::string>;
// Convert std::string to char *
template <typename T>
auto ENUM_EXTRACT(const T& x) {
if constexpr (std::is_same_v<std::decay_t<T>, std::string>) {
return x.c_str();
} else {
return x;
}
}
// The status variable contains status_t codes which are used by
// the core audio framework. We also consider AAudio status codes.
//
// Compare with mediametrics::statusToStatusString
//
inline constexpr const char* extendedStatusToStatusString(status_t status) {
switch (status) {
case BAD_VALUE: // status_t
case AAUDIO_ERROR_ILLEGAL_ARGUMENT:
case AAUDIO_ERROR_INVALID_FORMAT:
case AAUDIO_ERROR_INVALID_RATE:
case AAUDIO_ERROR_NULL:
case AAUDIO_ERROR_OUT_OF_RANGE:
return AMEDIAMETRICS_PROP_STATUS_VALUE_ARGUMENT;
case DEAD_OBJECT: // status_t
case FAILED_TRANSACTION: // status_t
case AAUDIO_ERROR_DISCONNECTED:
case AAUDIO_ERROR_INVALID_HANDLE:
case AAUDIO_ERROR_NO_SERVICE:
return AMEDIAMETRICS_PROP_STATUS_VALUE_IO;
case NO_MEMORY: // status_t
case AAUDIO_ERROR_NO_FREE_HANDLES:
case AAUDIO_ERROR_NO_MEMORY:
return AMEDIAMETRICS_PROP_STATUS_VALUE_MEMORY;
case PERMISSION_DENIED: // status_t
return AMEDIAMETRICS_PROP_STATUS_VALUE_SECURITY;
case INVALID_OPERATION: // status_t
case NO_INIT: // status_t
case AAUDIO_ERROR_INVALID_STATE:
case AAUDIO_ERROR_UNAVAILABLE:
case AAUDIO_ERROR_UNIMPLEMENTED:
return AMEDIAMETRICS_PROP_STATUS_VALUE_STATE;
case WOULD_BLOCK: // status_t
case AAUDIO_ERROR_TIMEOUT:
case AAUDIO_ERROR_WOULD_BLOCK:
return AMEDIAMETRICS_PROP_STATUS_VALUE_TIMEOUT;
default:
if (status >= 0) return AMEDIAMETRICS_PROP_STATUS_VALUE_OK; // non-negative values "OK"
[[fallthrough]]; // negative values are error.
case UNKNOWN_ERROR: // status_t
return AMEDIAMETRICS_PROP_STATUS_VALUE_UNKNOWN;
}
}
static constexpr const auto LOG_LEVEL = android::base::VERBOSE;
static constexpr int PREVIOUS_STATE_EXPIRE_SEC = 60 * 60; // 1 hour.
static constexpr const char * SUPPRESSED = "SUPPRESSED";
/*
* For logging purposes, we list all of the MediaMetrics atom fields,
* which can then be associated with consecutive arguments to the statsd write.
*/
static constexpr const char * const AudioRecordDeviceUsageFields[] = {
"mediametrics_audiorecorddeviceusage_reported", // proto number
"devices",
"device_names",
"device_time_nanos",
"encoding",
"frame_count",
"interval_count",
"sample_rate",
"flags",
"package_name",
"selected_device_id",
"caller",
"source",
"log_session_id",
};
static constexpr const char * const AudioThreadDeviceUsageFields[] = {
"mediametrics_audiothreaddeviceusage_reported",
"devices",
"device_names",
"device_time_nanos",
"encoding",
"frame_count",
"interval_count",
"sample_rate",
"flags",
"xruns",
"type",
};
static constexpr const char * const AudioTrackDeviceUsageFields[] = {
"mediametrics_audiotrackdeviceusage_reported",
"devices",
"device_names",
"device_time_nanos",
"encoding",
"frame_count",
"interval_count",
"sample_rate",
"flags",
"xruns",
"package_name",
"device_latency_millis",
"device_startup_millis",
"device_volume",
"selected_device_id",
"stream_type",
"usage",
"content_type",
"caller",
"traits",
"log_session_id",
};
static constexpr const char * const AudioRecordStatusFields[] {
"mediametrics_audiorecordstatus_reported",
"status",
"debug_message",
"status_subcode",
"uid",
"event",
"input_flags",
"source",
"encoding",
"channel_mask",
"buffer_frame_count",
"sample_rate",
};
static constexpr const char * const AudioTrackStatusFields[] {
"mediametrics_audiotrackstatus_reported",
"status",
"debug_message",
"status_subcode",
"uid",
"event",
"output_flags",
"content_type",
"usage",
"encoding",
"channel_mask",
"buffer_frame_count",
"sample_rate",
"speed",
"pitch",
};
static constexpr const char * const AudioDeviceConnectionFields[] = {
"mediametrics_audiodeviceconnection_reported",
"input_devices",
"output_devices",
"device_names",
"result",
"time_to_connect_millis",
"connection_count",
};
static constexpr const char * const AAudioStreamFields[] {
"mediametrics_aaudiostream_reported",
"path",
"direction",
"frames_per_burst",
"buffer_size",
"buffer_capacity",
"channel_count",
"total_frames_transferred",
"perf_mode_requested",
"perf_mode_actual",
"sharing",
"xrun_count",
"device_type",
"format_app",
"format_device",
"log_session_id",
"sample_rate",
"content_type",
"sharing_requested",
"format_hardware",
"channel_count_hardware",
"sample_rate_hardware",
"uid",
"sample_rate_client",
};
static constexpr const char * HeadTrackerDeviceEnabledFields[] {
"mediametrics_headtrackerdeviceenabled_reported",
"type",
"event",
"enabled",
};
static constexpr const char * HeadTrackerDeviceSupportedFields[] {
"mediametrics_headtrackerdevicesupported_reported",
"type",
"event",
"supported",
};
static constexpr const char * SpatializerCapabilitiesFields[] {
"mediametrics_spatializer_reported",
"head_tracking_modes",
"spatializer_levels",
"spatializer_modes",
"channel_masks",
};
static constexpr const char * SpatializerDeviceEnabledFields[] {
"mediametrics_spatializerdeviceenabled_reported",
"type",
"event",
"enabled",
};
static constexpr const char * const MidiDeviceCloseFields[] {
"mediametrics_midi_device_close_reported",
"uid",
"midi_device_id",
"input_port_count",
"output_port_count",
"device_type",
"is_shared",
"supports_ump",
"using_alsa",
"duration_ns",
"opened_count",
"closed_count",
"device_disconnected",
"total_input_bytes",
"total_output_bytes",
};
/**
* printFields is a helper method that prints the fields and corresponding values
* in a human readable style.
*/
template <size_t N, typename ...Types>
std::string printFields(const char * const (& fields)[N], Types ... args)
{
std::stringstream ss;
ss << " { ";
stringutils::fieldPrint(ss, fields, args...);
ss << "}";
return ss.str();
}
/**
* sendToStatsd is a helper method that sends the arguments to statsd
*/
template <typename ...Types>
int sendToStatsd(Types ... args)
{
int result = 0;
#ifdef STATSD_ENABLE
result = stats::media_metrics::stats_write(args...);
#endif
return result;
}
/**
* sendToStatsd is a helper method that sends the arguments to statsd
* and returns a pair { result, summary_string }.
*/
template <size_t N, typename ...Types>
std::pair<int, std::string> sendToStatsd(const char * const (& fields)[N], Types ... args)
{
int result = 0;
std::stringstream ss;
#ifdef STATSD_ENABLE
result = stats::media_metrics::stats_write(args...);
ss << "result:" << result;
#endif
ss << " { ";
stringutils::fieldPrint(ss, fields, args...);
ss << "}";
return { result, ss.str() };
}
AudioAnalytics::AudioAnalytics(const std::shared_ptr<StatsdLog>& statsdLog)
: mDeliverStatistics(property_get_bool(PROP_AUDIO_ANALYTICS_CLOUD_ENABLED, true))
, mStatsdLog(statsdLog)
, mAudioPowerUsage(this, statsdLog)
{
SetMinimumLogSeverity(android::base::DEBUG); // for LOG().
ALOGD("%s", __func__);
// Add action to save AnalyticsState if audioserver is restarted.
// This triggers on AudioFlinger or AudioPolicy ctors and onFirstRef,
// as well as TimeCheck events.
mActions.addAction(
AMEDIAMETRICS_KEY_AUDIO_FLINGER "." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_CTOR),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mHealth.onAudioServerStart(Health::Module::AUDIOFLINGER, item);
}));
mActions.addAction(
AMEDIAMETRICS_KEY_AUDIO_POLICY "." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_CTOR),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mHealth.onAudioServerStart(Health::Module::AUDIOPOLICY, item);
}));
mActions.addAction(
AMEDIAMETRICS_KEY_AUDIO_FLINGER "." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_TIMEOUT),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mHealth.onAudioServerTimeout(Health::Module::AUDIOFLINGER, item);
}));
mActions.addAction(
AMEDIAMETRICS_KEY_AUDIO_POLICY "." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_TIMEOUT),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mHealth.onAudioServerTimeout(Health::Module::AUDIOPOLICY, item);
}));
// Handle legacy aaudio playback stream statistics
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAAUDIOSTREAM),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item) {
mAAudioStreamInfo.endAAudioStream(item, AAudioStreamInfo::CALLER_PATH_LEGACY);
}));
// Handle legacy aaudio capture stream statistics
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAAUDIOSTREAM),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item) {
mAAudioStreamInfo.endAAudioStream(item, AAudioStreamInfo::CALLER_PATH_LEGACY);
}));
// Handle mmap aaudio stream statistics
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_STREAM "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAAUDIOSTREAM),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item) {
mAAudioStreamInfo.endAAudioStream(item, AAudioStreamInfo::CALLER_PATH_MMAP);
}));
// Handle device use record statistics
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAUDIOINTERVALGROUP),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mDeviceUse.endAudioIntervalGroup(item, DeviceUse::RECORD);
}));
// Handle device use thread statistics
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_THREAD "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAUDIOINTERVALGROUP),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mDeviceUse.endAudioIntervalGroup(item, DeviceUse::THREAD);
}));
// Handle device use track statistics
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAUDIOINTERVALGROUP),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mDeviceUse.endAudioIntervalGroup(item, DeviceUse::TRACK);
}));
// Handle device connection statistics
// We track connections (not disconnections) for the time to connect.
// TODO: consider BT requests in their A2dp service
// AudioManager.setBluetoothA2dpDeviceConnectionStateSuppressNoisyIntent
// AudioDeviceBroker.postBluetoothA2dpDeviceConnectionStateSuppressNoisyIntent
// AudioDeviceBroker.postA2dpActiveDeviceChange
mActions.addAction(
"audio.device.a2dp.state",
"connected",
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mDeviceConnection.a2dpConnected(item);
}));
// If audio is active, we expect to see a createAudioPatch after the device is connected.
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_THREAD "*." AMEDIAMETRICS_PROP_EVENT,
std::string("createAudioPatch"),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mDeviceConnection.createPatch(item);
}));
// Called from BT service
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_DEVICE
"postBluetoothA2dpDeviceConnectionStateSuppressNoisyIntent"
"." AMEDIAMETRICS_PROP_STATE,
"connected",
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mDeviceConnection.postBluetoothA2dpDeviceConnectionStateSuppressNoisyIntent(item);
}));
// Handle power usage
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAUDIOINTERVALGROUP),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mAudioPowerUsage.checkTrackRecord(item, true /* isTrack */);
}));
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD "*." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_ENDAUDIOINTERVALGROUP),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mAudioPowerUsage.checkTrackRecord(item, false /* isTrack */);
}));
mActions.addAction(
AMEDIAMETRICS_KEY_AUDIO_FLINGER "." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_SETMODE),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
// ALOGD("(key=%s) Audioflinger setMode", item->getKey().c_str());
mAudioPowerUsage.checkMode(item);
}));
mActions.addAction(
AMEDIAMETRICS_KEY_AUDIO_FLINGER "." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_SETVOICEVOLUME),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
// ALOGD("(key=%s) Audioflinger setVoiceVolume", item->getKey().c_str());
mAudioPowerUsage.checkVoiceVolume(item);
}));
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_THREAD "*." AMEDIAMETRICS_PROP_EVENT,
std::string("createAudioPatch"),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mAudioPowerUsage.checkCreatePatch(item);
}));
// Handle Spatializer - these keys are prefixed by "audio.spatializer."
mActions.addAction(
AMEDIAMETRICS_KEY_PREFIX_AUDIO_SPATIALIZER "*." AMEDIAMETRICS_PROP_EVENT,
std::monostate{}, /* match any event */
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item){
mSpatializer.onEvent(item);
}));
// Handle MIDI
mActions.addAction(
AMEDIAMETRICS_KEY_AUDIO_MIDI "." AMEDIAMETRICS_PROP_EVENT,
std::string(AMEDIAMETRICS_PROP_EVENT_VALUE_DEVICECLOSED),
std::make_shared<AnalyticsActions::Function>(
[this](const std::shared_ptr<const android::mediametrics::Item> &item) {
mMidiLogging.onEvent(item);
}));
}
AudioAnalytics::~AudioAnalytics()
{
ALOGD("%s", __func__);
mTimedAction.quit(); // ensure no deferred access during destructor.
}
status_t AudioAnalytics::submit(
const std::shared_ptr<const mediametrics::Item>& item, bool isTrusted)
{
if (!startsWith(item->getKey(), AMEDIAMETRICS_KEY_PREFIX_AUDIO)) return BAD_VALUE;
status_t status = mAnalyticsState->submit(item, isTrusted);
// Status is selectively authenticated.
processStatus(item);
if (status != NO_ERROR) return status; // may not be permitted.
// Only if the item was successfully submitted (permission)
// do we check triggered actions.
processActions(item);
return NO_ERROR;
}
std::pair<std::string, int32_t> AudioAnalytics::dump(
int32_t lines, int64_t sinceNs, const char *prefix) const
{
std::stringstream ss;
int32_t ll = lines;
if (ll > 0) {
auto [s, l] = mAnalyticsState->dump(ll, sinceNs, prefix);
ss << s;
ll -= l;
}
if (ll > 0) {
ss << "Prior audioserver state:\n";
--ll;
}
if (ll > 0) {
auto [s, l] = mPreviousAnalyticsState->dump(ll, sinceNs, prefix);
ss << s;
ll -= l;
}
if (ll > 0 && prefix == nullptr) {
auto [s, l] = mAudioPowerUsage.dump(ll);
ss << s;
ll -= l;
}
return { ss.str(), lines - ll };
}
void AudioAnalytics::processActions(const std::shared_ptr<const mediametrics::Item>& item)
{
auto actions = mActions.getActionsForItem(item); // internally locked.
// Execute actions with no lock held.
for (const auto& action : actions) {
(*action)(item);
}
}
void AudioAnalytics::processStatus(const std::shared_ptr<const mediametrics::Item>& item)
{
int32_t status;
if (!item->get(AMEDIAMETRICS_PROP_STATUS, &status)) return;
// Any record with a status will automatically be added to a heat map.
// Standard information.
const auto key = item->getKey();
const auto uid = item->getUid();
// from audio.track.10 -> prefix = audio.track, suffix = 10
// from audio.track.error -> prefix = audio.track, suffix = error
const auto [prefixKey, suffixKey] = stringutils::splitPrefixKey(key);
std::string message;
item->get(AMEDIAMETRICS_PROP_STATUSMESSAGE, &message); // optional
int32_t subCode = 0; // not used
(void)item->get(AMEDIAMETRICS_PROP_STATUSSUBCODE, &subCode); // optional
std::string eventStr; // optional
item->get(AMEDIAMETRICS_PROP_EVENT, &eventStr);
const std::string statusString = extendedStatusToStatusString(status);
// Add to the heat map - we automatically track every item's status to see
// the types of errors and the frequency of errors.
mHeatMap.add(prefixKey, suffixKey, eventStr, statusString, uid, message, subCode);
// Certain keys/event pairs are sent to statsd. If we get a match (true) we return early.
if (reportAudioRecordStatus(item, key, eventStr, statusString, uid, message, subCode)) return;
if (reportAudioTrackStatus(item, key, eventStr, statusString, uid, message, subCode)) return;
}
bool AudioAnalytics::reportAudioRecordStatus(
const std::shared_ptr<const mediametrics::Item>& item,
const std::string& key, const std::string& eventStr,
const std::string& statusString, uid_t uid, const std::string& message,
int32_t subCode) const
{
// Note that the prefixes often end with a '.' so we use startsWith.
if (!startsWith(key, AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD)) return false;
if (eventStr == AMEDIAMETRICS_PROP_EVENT_VALUE_CREATE) {
const int atom_status = types::lookup<types::STATUS, int32_t>(statusString);
// currently we only send create status events.
const int32_t event = stats::media_metrics::
MEDIAMETRICS_AUDIO_RECORD_STATUS_REPORTED__EVENT__AUDIO_RECORD_EVENT_CREATE;
// The following fields should all be present in a create event.
std::string flagsStr;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_ORIGINALFLAGS, &flagsStr),
"%s: %s missing %s field", __func__,
AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD, AMEDIAMETRICS_PROP_ORIGINALFLAGS);
const auto flags = types::lookup<types::INPUT_FLAG, int32_t>(flagsStr);
// AMEDIAMETRICS_PROP_SESSIONID omitted from atom
std::string sourceStr;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_SOURCE, &sourceStr),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD, AMEDIAMETRICS_PROP_SOURCE);
const int32_t source = types::lookup<types::SOURCE_TYPE, int32_t>(sourceStr);
// AMEDIAMETRICS_PROP_SELECTEDDEVICEID omitted from atom
std::string encodingStr;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_ENCODING, &encodingStr),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD, AMEDIAMETRICS_PROP_ENCODING);
const auto encoding = types::lookup<types::ENCODING, int32_t>(encodingStr);
int32_t channelMask = 0;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_CHANNELMASK, &channelMask),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD, AMEDIAMETRICS_PROP_CHANNELMASK);
int32_t frameCount = 0;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_FRAMECOUNT, &frameCount),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD, AMEDIAMETRICS_PROP_FRAMECOUNT);
int32_t sampleRate = 0;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_SAMPLERATE, &sampleRate),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD, AMEDIAMETRICS_PROP_SAMPLERATE);
const auto [ result, str ] = sendToStatsd(AudioRecordStatusFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIORECORDSTATUS_REPORTED)
, atom_status
, message.c_str()
, subCode
, uid
, event
, flags
, source
, encoding
, (int64_t)channelMask
, frameCount
, sampleRate
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mStatsdLog->log(stats::media_metrics::MEDIAMETRICS_AUDIORECORDSTATUS_REPORTED, str);
return true;
}
return false;
}
bool AudioAnalytics::reportAudioTrackStatus(
const std::shared_ptr<const mediametrics::Item>& item,
const std::string& key, const std::string& eventStr,
const std::string& statusString, uid_t uid, const std::string& message,
int32_t subCode) const
{
// Note that the prefixes often end with a '.' so we use startsWith.
if (!startsWith(key, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK)) return false;
if (eventStr == AMEDIAMETRICS_PROP_EVENT_VALUE_CREATE) {
const int atom_status = types::lookup<types::STATUS, int32_t>(statusString);
// currently we only send create status events.
const int32_t event = stats::media_metrics::
MEDIAMETRICS_AUDIO_TRACK_STATUS_REPORTED__EVENT__AUDIO_TRACK_EVENT_CREATE;
// The following fields should all be present in a create event.
std::string flagsStr;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_ORIGINALFLAGS, &flagsStr),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_ORIGINALFLAGS);
const auto flags = types::lookup<types::OUTPUT_FLAG, int32_t>(flagsStr);
// AMEDIAMETRICS_PROP_SESSIONID omitted from atom
std::string contentTypeStr;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_CONTENTTYPE, &contentTypeStr),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_CONTENTTYPE);
const auto contentType = types::lookup<types::CONTENT_TYPE, int32_t>(contentTypeStr);
std::string usageStr;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_USAGE, &usageStr),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_USAGE);
const auto usage = types::lookup<types::USAGE, int32_t>(usageStr);
// AMEDIAMETRICS_PROP_SELECTEDDEVICEID omitted from atom
std::string encodingStr;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_ENCODING, &encodingStr),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_ENCODING);
const auto encoding = types::lookup<types::ENCODING, int32_t>(encodingStr);
int32_t channelMask = 0;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_CHANNELMASK, &channelMask),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_CHANNELMASK);
int32_t frameCount = 0;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_FRAMECOUNT, &frameCount),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_FRAMECOUNT);
int32_t sampleRate = 0;
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_SAMPLERATE, &sampleRate),
"%s: %s missing %s field",
__func__, AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_SAMPLERATE);
double speed = 0.f; // default is 1.f
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_PLAYBACK_SPEED, &speed),
"%s: %s missing %s field",
__func__,
AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_PLAYBACK_SPEED);
double pitch = 0.f; // default is 1.f
ALOGD_IF(!item->get(AMEDIAMETRICS_PROP_PLAYBACK_PITCH, &pitch),
"%s: %s missing %s field",
__func__,
AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK, AMEDIAMETRICS_PROP_PLAYBACK_PITCH);
const auto [ result, str ] = sendToStatsd(AudioTrackStatusFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIOTRACKSTATUS_REPORTED)
, atom_status
, message.c_str()
, subCode
, uid
, event
, flags
, contentType
, usage
, encoding
, (int64_t)channelMask
, frameCount
, sampleRate
, (float)speed
, (float)pitch
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mStatsdLog->log(stats::media_metrics::MEDIAMETRICS_AUDIOTRACKSTATUS_REPORTED, str);
return true;
}
return false;
}
// HELPER METHODS
std::string AudioAnalytics::getThreadFromTrack(const std::string& track) const
{
int32_t threadId_int32{};
if (mAnalyticsState->timeMachine().get(
track, AMEDIAMETRICS_PROP_THREADID, &threadId_int32) != NO_ERROR) {
return {};
}
return std::string(AMEDIAMETRICS_KEY_PREFIX_AUDIO_THREAD) + std::to_string(threadId_int32);
}
// DeviceUse helper class.
void AudioAnalytics::DeviceUse::endAudioIntervalGroup(
const std::shared_ptr<const android::mediametrics::Item> &item, ItemType itemType) const {
const std::string& key = item->getKey();
const std::string id = key.substr(
(itemType == THREAD ? sizeof(AMEDIAMETRICS_KEY_PREFIX_AUDIO_THREAD)
: itemType == TRACK ? sizeof(AMEDIAMETRICS_KEY_PREFIX_AUDIO_TRACK)
: sizeof(AMEDIAMETRICS_KEY_PREFIX_AUDIO_RECORD))
- 1);
// deliver statistics
int64_t deviceTimeNs = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DEVICETIMENS, &deviceTimeNs);
std::string encoding;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_ENCODING, &encoding);
int32_t frameCount = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_FRAMECOUNT, &frameCount);
int32_t intervalCount = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_INTERVALCOUNT, &intervalCount);
int32_t sampleRate = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SAMPLERATE, &sampleRate);
std::string flags;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_FLAGS, &flags);
switch (itemType) {
case RECORD: {
std::string inputDevicePairs;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_INPUTDEVICES, &inputDevicePairs);
const auto [ inputDeviceStatsd, inputDevices ] =
stringutils::parseInputDevicePairs(inputDevicePairs);
const std::string inputDeviceNames; // not filled currently.
std::string callerName;
const bool clientCalled = mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CALLERNAME, &callerName) == OK;
std::string packageName;
int64_t versionCode = 0;
int32_t uid = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_ALLOWUID, &uid);
if (uid != -1) {
std::tie(packageName, versionCode) =
MediaMetricsService::getSanitizedPackageNameAndVersionCode(uid);
}
int32_t selectedDeviceId = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SELECTEDDEVICEID, &selectedDeviceId);
std::string source;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SOURCE, &source);
// Android S
std::string logSessionId;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_LOGSESSIONID, &logSessionId);
const auto callerNameForStats =
types::lookup<types::CALLER_NAME, short_enum_type_t>(callerName);
const auto encodingForStats = types::lookup<types::ENCODING, short_enum_type_t>(encoding);
const auto flagsForStats = types::lookup<types::INPUT_FLAG, short_enum_type_t>(flags);
const auto sourceForStats = types::lookup<types::SOURCE_TYPE, short_enum_type_t>(source);
// Android S
const auto logSessionIdForStats = ValidateId::get()->validateId(logSessionId);
LOG(LOG_LEVEL) << "key:" << key
<< " id:" << id
<< " inputDevices:" << inputDevices << "(" << inputDeviceStatsd
<< ") inputDeviceNames:" << inputDeviceNames
<< " deviceTimeNs:" << deviceTimeNs
<< " encoding:" << encoding << "(" << encodingForStats
<< ") frameCount:" << frameCount
<< " intervalCount:" << intervalCount
<< " sampleRate:" << sampleRate
<< " flags:" << flags << "(" << flagsForStats
<< ") packageName:" << packageName
<< " selectedDeviceId:" << selectedDeviceId
<< " callerName:" << callerName << "(" << callerNameForStats
<< ") source:" << source << "(" << sourceForStats
<< ") logSessionId:" << logSessionId << "(" << logSessionIdForStats
<< ")";
if (clientCalled // only log if client app called AudioRecord.
&& mAudioAnalytics.mDeliverStatistics) {
const auto [ result, str ] = sendToStatsd(AudioRecordDeviceUsageFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIORECORDDEVICEUSAGE_REPORTED)
, ENUM_EXTRACT(inputDeviceStatsd)
, inputDeviceNames.c_str()
, deviceTimeNs
, ENUM_EXTRACT(encodingForStats)
, frameCount
, intervalCount
, sampleRate
, ENUM_EXTRACT(flagsForStats)
, packageName.c_str()
, selectedDeviceId
, ENUM_EXTRACT(callerNameForStats)
, ENUM_EXTRACT(sourceForStats)
, logSessionIdForStats.c_str()
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_AUDIORECORDDEVICEUSAGE_REPORTED, str);
}
} break;
case THREAD: {
std::string type;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_TYPE, &type);
int32_t underrun = 0; // zero for record types
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_UNDERRUN, &underrun);
const bool isInput = types::isInputThreadType(type);
// get device information
std::string devicePairs;
std::string deviceStatsd;
std::string devices;
std::string deviceNames;
if (isInput) {
// Note we get the "last" device which is the one associated with group.
item->get(AMEDIAMETRICS_PROP_PREFIX_LAST AMEDIAMETRICS_PROP_INPUTDEVICES,
&devicePairs);
std::tie(deviceStatsd, devices) = stringutils::parseInputDevicePairs(devicePairs);
} else {
// Note we get the "last" device which is the one associated with group.
item->get(AMEDIAMETRICS_PROP_PREFIX_LAST AMEDIAMETRICS_PROP_OUTPUTDEVICES,
&devicePairs);
std::tie(deviceStatsd, devices) = stringutils::parseOutputDevicePairs(devicePairs);
deviceNames = mAudioAnalytics.getDeviceNamesFromOutputDevices(devices);
}
const auto encodingForStats = types::lookup<types::ENCODING, short_enum_type_t>(encoding);
const auto flagsForStats =
(isInput ? types::lookup<types::INPUT_FLAG, short_enum_type_t>(flags)
: types::lookup<types::OUTPUT_FLAG, short_enum_type_t>(flags));
const auto typeForStats = types::lookup<types::THREAD_TYPE, short_enum_type_t>(type);
LOG(LOG_LEVEL) << "key:" << key
<< " id:" << id
<< " devices:" << devices << "(" << deviceStatsd
<< ") deviceNames:" << deviceNames
<< " deviceTimeNs:" << deviceTimeNs
<< " encoding:" << encoding << "(" << encodingForStats
<< ") frameCount:" << frameCount
<< " intervalCount:" << intervalCount
<< " sampleRate:" << sampleRate
<< " underrun:" << underrun
<< " flags:" << flags << "(" << flagsForStats
<< ") type:" << type << "(" << typeForStats
<< ")";
if (mAudioAnalytics.mDeliverStatistics) {
const auto [ result, str ] = sendToStatsd(AudioThreadDeviceUsageFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIOTHREADDEVICEUSAGE_REPORTED)
, ENUM_EXTRACT(deviceStatsd)
, deviceNames.c_str()
, deviceTimeNs
, ENUM_EXTRACT(encodingForStats)
, frameCount
, intervalCount
, sampleRate
, ENUM_EXTRACT(flagsForStats)
, underrun
, ENUM_EXTRACT(typeForStats)
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_AUDIOTHREADDEVICEUSAGE_REPORTED, str);
}
} break;
case TRACK: {
std::string outputDevicePairs;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_OUTPUTDEVICES, &outputDevicePairs);
const auto [ outputDeviceStatsd, outputDevices ] =
stringutils::parseOutputDevicePairs(outputDevicePairs);
const std::string outputDeviceNames =
mAudioAnalytics.getDeviceNamesFromOutputDevices(outputDevices);
std::string callerName;
const bool clientCalled = mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CALLERNAME, &callerName) == OK;
std::string contentType;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CONTENTTYPE, &contentType);
double deviceLatencyMs = 0.;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DEVICELATENCYMS, &deviceLatencyMs);
double deviceStartupMs = 0.;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DEVICESTARTUPMS, &deviceStartupMs);
double deviceVolume = 0.;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DEVICEVOLUME, &deviceVolume);
std::string packageName;
int64_t versionCode = 0;
int32_t uid = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_ALLOWUID, &uid);
if (uid != -1) {
std::tie(packageName, versionCode) =
MediaMetricsService::getSanitizedPackageNameAndVersionCode(uid);
}
double playbackPitch = 0.;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_PLAYBACK_PITCH, &playbackPitch);
double playbackSpeed = 0.;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_PLAYBACK_SPEED, &playbackSpeed);
int32_t selectedDeviceId = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SELECTEDDEVICEID, &selectedDeviceId);
std::string streamType;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_STREAMTYPE, &streamType);
std::string traits;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_TRAITS, &traits);
int32_t underrun = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_UNDERRUN, &underrun);
std::string usage;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_USAGE, &usage);
// Android S
std::string logSessionId;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_LOGSESSIONID, &logSessionId);
const auto callerNameForStats =
types::lookup<types::CALLER_NAME, short_enum_type_t>(callerName);
const auto contentTypeForStats =
types::lookup<types::CONTENT_TYPE, short_enum_type_t>(contentType);
const auto encodingForStats = types::lookup<types::ENCODING, short_enum_type_t>(encoding);
const auto flagsForStats = types::lookup<types::OUTPUT_FLAG, short_enum_type_t>(flags);
const auto streamTypeForStats =
types::lookup<types::STREAM_TYPE, short_enum_type_t>(streamType);
const auto traitsForStats =
types::lookup<types::TRACK_TRAITS, short_enum_type_t>(traits);
const auto usageForStats = types::lookup<types::USAGE, short_enum_type_t>(usage);
// Android S
const auto logSessionIdForStats = ValidateId::get()->validateId(logSessionId);
LOG(LOG_LEVEL) << "key:" << key
<< " id:" << id
<< " outputDevices:" << outputDevices << "(" << outputDeviceStatsd
<< ") outputDeviceNames:" << outputDeviceNames
<< " deviceTimeNs:" << deviceTimeNs
<< " encoding:" << encoding << "(" << encodingForStats
<< ") frameCount:" << frameCount
<< " intervalCount:" << intervalCount
<< " sampleRate:" << sampleRate
<< " underrun:" << underrun
<< " flags:" << flags << "(" << flagsForStats
<< ") callerName:" << callerName << "(" << callerNameForStats
<< ") contentType:" << contentType << "(" << contentTypeForStats
<< ") deviceLatencyMs:" << deviceLatencyMs
<< " deviceStartupMs:" << deviceStartupMs
<< " deviceVolume:" << deviceVolume
<< " packageName:" << packageName
<< " playbackPitch:" << playbackPitch
<< " playbackSpeed:" << playbackSpeed
<< " selectedDeviceId:" << selectedDeviceId
<< " streamType:" << streamType << "(" << streamTypeForStats
<< ") traits:" << traits << "(" << traitsForStats
<< ") usage:" << usage << "(" << usageForStats
<< ") logSessionId:" << logSessionId << "(" << logSessionIdForStats
<< ")";
if (clientCalled // only log if client app called AudioTracks
&& mAudioAnalytics.mDeliverStatistics) {
const auto [ result, str ] = sendToStatsd(AudioTrackDeviceUsageFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIOTRACKDEVICEUSAGE_REPORTED)
, ENUM_EXTRACT(outputDeviceStatsd)
, outputDeviceNames.c_str()
, deviceTimeNs
, ENUM_EXTRACT(encodingForStats)
, frameCount
, intervalCount
, sampleRate
, ENUM_EXTRACT(flagsForStats)
, underrun
, packageName.c_str()
, (float)deviceLatencyMs
, (float)deviceStartupMs
, (float)deviceVolume
, selectedDeviceId
, ENUM_EXTRACT(streamTypeForStats)
, ENUM_EXTRACT(usageForStats)
, ENUM_EXTRACT(contentTypeForStats)
, ENUM_EXTRACT(callerNameForStats)
, ENUM_EXTRACT(traitsForStats)
, logSessionIdForStats.c_str()
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_AUDIOTRACKDEVICEUSAGE_REPORTED, str);
}
} break;
}
}
// DeviceConnection helper class.
void AudioAnalytics::DeviceConnection::a2dpConnected(
const std::shared_ptr<const android::mediametrics::Item> &item) {
const std::string& key = item->getKey();
const int64_t atNs = item->getTimestamp();
{
std::lock_guard l(mLock);
mA2dpConnectionServiceNs = atNs;
++mA2dpConnectionServices;
if (mA2dpConnectionRequestNs == 0) {
mAudioAnalytics.mTimedAction.postIn(std::chrono::seconds(5), [this](){ expire(); });
}
// This sets the time we were connected. Now we look for the delta in the future.
}
std::string name;
item->get(AMEDIAMETRICS_PROP_NAME, &name);
ALOGD("(key=%s) a2dp connected device:%s atNs:%lld",
key.c_str(), name.c_str(), (long long)atNs);
}
void AudioAnalytics::DeviceConnection::createPatch(
const std::shared_ptr<const android::mediametrics::Item> &item) {
std::lock_guard l(mLock);
if (mA2dpConnectionServiceNs == 0) return; // patch unrelated to us.
const std::string& key = item->getKey();
std::string outputDevices;
item->get(AMEDIAMETRICS_PROP_OUTPUTDEVICES, &outputDevices);
if (outputDevices.find("AUDIO_DEVICE_OUT_BLUETOOTH_A2DP") != std::string::npos) {
// TODO compare address
int64_t timeDiffNs = item->getTimestamp();
if (mA2dpConnectionRequestNs == 0) {
ALOGD("%s: A2DP create patch didn't see a connection request", __func__);
timeDiffNs -= mA2dpConnectionServiceNs;
} else {
timeDiffNs -= mA2dpConnectionRequestNs;
}
mA2dpConnectionRequestNs = 0;
mA2dpConnectionServiceNs = 0;
++mA2dpConnectionSuccesses;
const auto connectionTimeMs = float((double)timeDiffNs * 1e-6);
const auto outputDeviceBits = types::lookup<types::OUTPUT_DEVICE, long_enum_type_t>(
"AUDIO_DEVICE_OUT_BLUETOOTH_A2DP");
LOG(LOG_LEVEL) << "key:" << key
<< " A2DP SUCCESS"
<< " outputDevices:" << outputDeviceBits
<< " deviceName:" << mA2dpDeviceName
<< " connectionTimeMs:" << connectionTimeMs;
if (mAudioAnalytics.mDeliverStatistics) {
const long_enum_type_t inputDeviceBits{};
const auto [ result, str ] = sendToStatsd(AudioDeviceConnectionFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIODEVICECONNECTION_REPORTED)
, ENUM_EXTRACT(inputDeviceBits)
, ENUM_EXTRACT(outputDeviceBits)
, mA2dpDeviceName.c_str()
, types::DEVICE_CONNECTION_RESULT_SUCCESS
, connectionTimeMs
, /* connection_count */ 1
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_AUDIODEVICECONNECTION_REPORTED, str);
}
}
}
// Called through AudioManager when the BT service wants to enable
void AudioAnalytics::DeviceConnection::postBluetoothA2dpDeviceConnectionStateSuppressNoisyIntent(
const std::shared_ptr<const android::mediametrics::Item> &item) {
const int64_t atNs = item->getTimestamp();
const std::string& key = item->getKey();
std::string state;
item->get(AMEDIAMETRICS_PROP_STATE, &state);
if (state != "connected") return;
std::string name;
item->get(AMEDIAMETRICS_PROP_NAME, &name);
{
std::lock_guard l(mLock);
mA2dpConnectionRequestNs = atNs;
++mA2dpConnectionRequests;
mA2dpDeviceName = SUPPRESSED; // TODO(b/161554630) sanitize name
}
ALOGD("(key=%s) a2dp connection name:%s request atNs:%lld",
key.c_str(), name.c_str(), (long long)atNs);
// TODO: attempt to cancel a timed event, rather than let it expire.
mAudioAnalytics.mTimedAction.postIn(std::chrono::seconds(5), [this](){ expire(); });
}
void AudioAnalytics::DeviceConnection::expire() {
std::lock_guard l(mLock);
if (mA2dpConnectionRequestNs == 0) return; // ignore (this was an internal connection).
const long_enum_type_t inputDeviceBits{};
const auto outputDeviceBits = types::lookup<types::OUTPUT_DEVICE, long_enum_type_t>(
"AUDIO_DEVICE_OUT_BLUETOOTH_A2DP");
if (mA2dpConnectionServiceNs == 0) {
++mA2dpConnectionJavaServiceCancels; // service did not connect to A2DP
LOG(LOG_LEVEL) << "A2DP CANCEL"
<< " outputDevices:" << outputDeviceBits
<< " deviceName:" << mA2dpDeviceName;
if (mAudioAnalytics.mDeliverStatistics) {
const auto [ result, str ] = sendToStatsd(AudioDeviceConnectionFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIODEVICECONNECTION_REPORTED)
, ENUM_EXTRACT(inputDeviceBits)
, ENUM_EXTRACT(outputDeviceBits)
, mA2dpDeviceName.c_str()
, types::DEVICE_CONNECTION_RESULT_JAVA_SERVICE_CANCEL
, /* connection_time_ms */ 0.f
, /* connection_count */ 1
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_AUDIODEVICECONNECTION_REPORTED, str);
}
return;
}
// AudioFlinger didn't play - an expiration may occur because there is no audio playing.
// Should we check elsewhere?
// TODO: disambiguate this case.
mA2dpConnectionRequestNs = 0;
mA2dpConnectionServiceNs = 0;
++mA2dpConnectionUnknowns; // connection result unknown
LOG(LOG_LEVEL) << "A2DP UNKNOWN"
<< " outputDevices:" << outputDeviceBits
<< " deviceName:" << mA2dpDeviceName;
if (mAudioAnalytics.mDeliverStatistics) {
const auto [ result, str ] = sendToStatsd(AudioDeviceConnectionFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AUDIODEVICECONNECTION_REPORTED)
, ENUM_EXTRACT(inputDeviceBits)
, ENUM_EXTRACT(outputDeviceBits)
, mA2dpDeviceName.c_str()
, types::DEVICE_CONNECTION_RESULT_UNKNOWN
, /* connection_time_ms */ 0.f
, /* connection_count */ 1
);
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_AUDIODEVICECONNECTION_REPORTED, str);
}
}
void AudioAnalytics::AAudioStreamInfo::endAAudioStream(
const std::shared_ptr<const android::mediametrics::Item> &item, CallerPath path) const {
const std::string& key = item->getKey();
std::string directionStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DIRECTION, &directionStr);
const auto direction = types::lookup<types::AAUDIO_DIRECTION, int32_t>(directionStr);
int32_t framesPerBurst = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_BURSTFRAMES, &framesPerBurst);
int32_t bufferSizeInFrames = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_BUFFERSIZEFRAMES, &bufferSizeInFrames);
int32_t bufferCapacityInFrames = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_BUFFERCAPACITYFRAMES, &bufferCapacityInFrames);
int32_t channelCount = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CHANNELCOUNT, &channelCount);
if (channelCount == -1) {
// Try to get channel count from channel mask. From the legacy path,
// only channel mask are logged.
int32_t channelMask = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CHANNELMASK, &channelMask);
if (channelMask != 0) {
switch (direction) {
case 1: // Output, keep sync with AudioTypes#getAAudioDirection()
channelCount = (int32_t)audio_channel_count_from_out_mask(channelMask);
break;
case 2: // Input, keep sync with AudioTypes#getAAudioDirection()
channelCount = (int32_t)audio_channel_count_from_in_mask(channelMask);
break;
default:
ALOGW("Invalid direction %d", direction);
}
}
}
int64_t totalFramesTransferred = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_FRAMESTRANSFERRED, &totalFramesTransferred);
std::string perfModeRequestedStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_PERFORMANCEMODE, &perfModeRequestedStr);
const auto perfModeRequested =
types::lookup<types::AAUDIO_PERFORMANCE_MODE, int32_t>(perfModeRequestedStr);
std::string perfModeActualStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_PERFORMANCEMODEACTUAL, &perfModeActualStr);
const auto perfModeActual =
types::lookup<types::AAUDIO_PERFORMANCE_MODE, int32_t>(perfModeActualStr);
std::string sharingModeActualStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SHARINGMODEACTUAL, &sharingModeActualStr);
const auto sharingModeActual =
types::lookup<types::AAUDIO_SHARING_MODE, int32_t>(sharingModeActualStr);
int32_t xrunCount = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_UNDERRUN, &xrunCount);
std::string serializedDeviceTypes;
// TODO: only routed device id is logged, but no device type
std::string formatAppStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_ENCODINGCLIENT, &formatAppStr);
const auto formatApp = types::lookup<types::ENCODING, int32_t>(formatAppStr);
std::string formatDeviceStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_ENCODING, &formatDeviceStr);
const auto formatDevice = types::lookup<types::ENCODING, int32_t>(formatDeviceStr);
std::string logSessionId;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_LOGSESSIONID, &logSessionId);
int32_t sampleRate = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SAMPLERATE, &sampleRate);
std::string contentTypeStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CONTENTTYPE, &contentTypeStr);
const auto contentType = types::lookup<types::CONTENT_TYPE, int32_t>(contentTypeStr);
std::string sharingModeRequestedStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SHARINGMODE, &sharingModeRequestedStr);
const auto sharingModeRequested =
types::lookup<types::AAUDIO_SHARING_MODE, int32_t>(sharingModeRequestedStr);
std::string formatHardwareStr;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_ENCODINGHARDWARE, &formatHardwareStr);
const auto formatHardware = types::lookup<types::ENCODING, int32_t>(formatHardwareStr);
int32_t channelCountHardware = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CHANNELCOUNTHARDWARE, &channelCountHardware);
int32_t sampleRateHardware = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SAMPLERATEHARDWARE, &sampleRateHardware);
const auto uid = item->getUid();
int32_t sampleRateClient = 0;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SAMPLERATECLIENT, &sampleRateClient);
LOG(LOG_LEVEL) << "key:" << key
<< " path:" << path
<< " direction:" << direction << "(" << directionStr << ")"
<< " frames_per_burst:" << framesPerBurst
<< " buffer_size:" << bufferSizeInFrames
<< " buffer_capacity:" << bufferCapacityInFrames
<< " channel_count:" << channelCount
<< " total_frames_transferred:" << totalFramesTransferred
<< " perf_mode_requested:" << perfModeRequested << "(" << perfModeRequestedStr << ")"
<< " perf_mode_actual:" << perfModeActual << "(" << perfModeActualStr << ")"
<< " sharing:" << sharingModeActual << "(" << sharingModeActualStr << ")"
<< " xrun_count:" << xrunCount
<< " device_type:" << serializedDeviceTypes
<< " format_app:" << formatApp << "(" << formatAppStr << ")"
<< " format_device: " << formatDevice << "(" << formatDeviceStr << ")"
<< " log_session_id: " << logSessionId
<< " sample_rate: " << sampleRate
<< " content_type: " << contentType << "(" << contentTypeStr << ")"
<< " sharing_requested:" << sharingModeRequested
<< "(" << sharingModeRequestedStr << ")"
<< " format_hardware:" << formatHardware << "(" << formatHardwareStr << ")"
<< " channel_count_hardware:" << channelCountHardware
<< " sample_rate_hardware: " << sampleRateHardware
<< " uid: " << uid
<< " sample_rate_client: " << sampleRateClient;
if (mAudioAnalytics.mDeliverStatistics) {
const stats::media_metrics::BytesField bf_serialized(
serializedDeviceTypes.c_str(), serializedDeviceTypes.size());
const auto result = sendToStatsd(
CONDITION(stats::media_metrics::MEDIAMETRICS_AAUDIOSTREAM_REPORTED)
, path
, direction
, framesPerBurst
, bufferSizeInFrames
, bufferCapacityInFrames
, channelCount
, totalFramesTransferred
, perfModeRequested
, perfModeActual
, sharingModeActual
, xrunCount
, bf_serialized
, formatApp
, formatDevice
, logSessionId.c_str()
, sampleRate
, contentType
, sharingModeRequested
, formatHardware
, channelCountHardware
, sampleRateHardware
, uid
, sampleRateClient
);
std::stringstream ss;
ss << "result:" << result;
const auto fieldsStr = printFields(AAudioStreamFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_AAUDIOSTREAM_REPORTED)
, path
, direction
, framesPerBurst
, bufferSizeInFrames
, bufferCapacityInFrames
, channelCount
, totalFramesTransferred
, perfModeRequested
, perfModeActual
, sharingModeActual
, xrunCount
, serializedDeviceTypes.c_str()
, formatApp
, formatDevice
, logSessionId.c_str()
, sampleRate
, contentType
, sharingModeRequested
, formatHardware
, channelCountHardware
, sampleRateHardware
, uid
, sampleRateClient
);
ss << " " << fieldsStr;
std::string str = ss.str();
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(stats::media_metrics::MEDIAMETRICS_AAUDIOSTREAM_REPORTED, str);
}
}
// Create new state, typically occurs after an AudioFlinger ctor event.
void AudioAnalytics::newState()
{
mPreviousAnalyticsState.set(std::make_shared<AnalyticsState>(
*mAnalyticsState.get()));
// Note: get returns shared_ptr temp, whose lifetime is extended
// to end of full expression.
mAnalyticsState->clear(); // TODO: filter the analytics state.
// Perhaps report this.
// Set up a timer to expire the previous audio state to save space.
// Use the transaction log size as a cookie to see if it is the
// same as before. A benign race is possible where a state is cleared early.
const size_t size = mPreviousAnalyticsState->transactionLog().size();
mTimedAction.postIn(
std::chrono::seconds(PREVIOUS_STATE_EXPIRE_SEC), [this, size](){
if (mPreviousAnalyticsState->transactionLog().size() == size) {
ALOGD("expiring previous audio state after %d seconds.",
PREVIOUS_STATE_EXPIRE_SEC);
mPreviousAnalyticsState->clear(); // removes data from the state.
}
});
}
void AudioAnalytics::Health::onAudioServerStart(Module module,
const std::shared_ptr<const android::mediametrics::Item> &item)
{
const auto nowTime = std::chrono::system_clock::now();
if (module == Module::AUDIOFLINGER) {
{
std::lock_guard lg(mLock);
// reset state on AudioFlinger construction.
// AudioPolicy is created after AudioFlinger.
mAudioFlingerCtorTime = nowTime;
mSimpleLog.log("AudioFlinger ctor");
}
mAudioAnalytics.newState();
return;
}
if (module == Module::AUDIOPOLICY) {
// A start event occurs when audioserver
//
// (1) Starts the first time
// (2) Restarts because of the TimeCheck watchdog
// (3) Restarts not because of the TimeCheck watchdog.
int64_t executionTimeNs = 0;
(void)item->get(AMEDIAMETRICS_PROP_EXECUTIONTIMENS, &executionTimeNs);
const float loadTimeMs = executionTimeNs * 1e-6f;
std::lock_guard lg(mLock);
const int64_t restarts = mStartCount;
if (mStopCount == mStartCount) {
mAudioPolicyCtorTime = nowTime;
++mStartCount;
if (mStopCount == 0) {
// (1) First time initialization.
ALOGW("%s: (key=%s) AudioPolicy ctor, loadTimeMs:%f",
__func__, item->getKey().c_str(), loadTimeMs);
mSimpleLog.log("AudioPolicy ctor, loadTimeMs:%f", loadTimeMs);
} else {
// (2) Previous failure caught due to TimeCheck. We know how long restart takes.
const float restartMs =
std::chrono::duration_cast<std::chrono::duration<float, std::milli>>(
mAudioFlingerCtorTime - mStopTime).count();
ALOGW("%s: (key=%s) AudioPolicy ctor, "
"restarts:%lld restartMs:%f loadTimeMs:%f",
__func__, item->getKey().c_str(),
(long long)restarts, restartMs, loadTimeMs);
mSimpleLog.log("AudioPolicy ctor restarts:%lld restartMs:%f loadTimeMs:%f",
(long long)restarts, restartMs, loadTimeMs);
}
} else {
// (3) Previous failure is NOT due to TimeCheck, so we don't know the restart time.
// However we can estimate the uptime from the delta time from previous ctor.
const float uptimeMs =
std::chrono::duration_cast<std::chrono::duration<float, std::milli>>(
nowTime - mAudioFlingerCtorTime).count();
mStopCount = mStartCount;
mAudioPolicyCtorTime = nowTime;
++mStartCount;
ALOGW("%s: (key=%s) AudioPolicy ctor after uncaught failure, "
"mStartCount:%lld mStopCount:%lld uptimeMs:%f loadTimeMs:%f",
__func__, item->getKey().c_str(),
(long long)mStartCount, (long long)mStopCount, uptimeMs, loadTimeMs);
mSimpleLog.log("AudioPolicy ctor after uncaught failure, "
"restarts:%lld uptimeMs:%f loadTimeMs:%f",
(long long)restarts, uptimeMs, loadTimeMs);
}
}
}
void AudioAnalytics::Health::onAudioServerTimeout(Module module,
const std::shared_ptr<const android::mediametrics::Item> &item)
{
std::string moduleName = getModuleName(module);
int64_t methodCode{};
std::string methodName;
(void)item->get(AMEDIAMETRICS_PROP_METHODCODE, &methodCode);
(void)item->get(AMEDIAMETRICS_PROP_METHODNAME, &methodName);
std::lock_guard lg(mLock);
if (mStopCount >= mStartCount) {
ALOGD("%s: (key=%s) %s timeout %s(%lld) "
"unmatched mStopCount(%lld) >= mStartCount(%lld), ignoring",
__func__, item->getKey().c_str(), moduleName.c_str(),
methodName.c_str(), (long long)methodCode,
(long long)mStopCount, (long long)mStartCount);
return;
}
const int64_t restarts = mStartCount - 1;
++mStopCount;
mStopTime = std::chrono::system_clock::now();
const float uptimeMs = std::chrono::duration_cast<std::chrono::duration<float, std::milli>>(
mStopTime - mAudioFlingerCtorTime).count();
ALOGW("%s: (key=%s) %s timeout %s(%lld) restarts:%lld uptimeMs:%f",
__func__, item->getKey().c_str(), moduleName.c_str(),
methodName.c_str(), (long long)methodCode,
(long long)restarts, uptimeMs);
mSimpleLog.log("%s timeout %s(%lld) restarts:%lld uptimeMs:%f",
moduleName.c_str(), methodName.c_str(), (long long)methodCode,
(long long)restarts, uptimeMs);
}
std::pair<std::string, int32_t> AudioAnalytics::Health::dump(
int32_t lines, const char *prefix) const
{
std::lock_guard lg(mLock);
std::string s = mSimpleLog.dumpToString(prefix == nullptr ? "" : prefix, lines);
size_t n = std::count(s.begin(), s.end(), '\n');
return { s, n };
}
// Classifies the setting event for statsd (use generated statsd enums.proto constants).
static int32_t classifySettingEvent(bool isSetAlready, bool withinBoot) {
if (isSetAlready) {
return stats::media_metrics::MEDIAMETRICS_SPATIALIZER_DEVICE_ENABLED_REPORTED__EVENT__SPATIALIZER_SETTING_EVENT_NORMAL;
}
if (withinBoot) {
return stats::media_metrics::MEDIAMETRICS_SPATIALIZER_DEVICE_ENABLED_REPORTED__EVENT__SPATIALIZER_SETTING_EVENT_BOOT;
}
return stats::media_metrics::MEDIAMETRICS_SPATIALIZER_DEVICE_ENABLED_REPORTED__EVENT__SPATIALIZER_SETTING_EVENT_FIRST;
}
void AudioAnalytics::Spatializer::onEvent(
const std::shared_ptr<const android::mediametrics::Item> &item)
{
const auto key = item->getKey();
if (!startsWith(key, AMEDIAMETRICS_KEY_PREFIX_AUDIO_SPATIALIZER)) return;
const std::string suffix =
key.substr(std::size(AMEDIAMETRICS_KEY_PREFIX_AUDIO_SPATIALIZER) - 1);
std::string eventStr; // optional - find the actual event string.
(void)item->get(AMEDIAMETRICS_PROP_EVENT, &eventStr);
const size_t delim = suffix.find('.'); // note could use split.
if (delim == suffix.npos) {
// on create with suffix == "0" for the first spatializer effect.
std::string headTrackingModes;
(void)item->get(AMEDIAMETRICS_PROP_HEADTRACKINGMODES, &headTrackingModes);
std::string levels;
(void)item->get(AMEDIAMETRICS_PROP_LEVELS, &levels);
std::string modes;
(void)item->get(AMEDIAMETRICS_PROP_MODES, &modes);
std::string channelMasks;
(void)item->get(AMEDIAMETRICS_PROP_CHANNELMASKS, &channelMasks);
LOG(LOG_LEVEL) << "key:" << key
<< " headTrackingModes:" << headTrackingModes
<< " levels:" << levels
<< " modes:" << modes
<< " channelMasks:" << channelMasks
;
const std::vector<int32_t> headTrackingModesVector =
types::vectorFromMap(headTrackingModes, types::getHeadTrackingModeMap());
const std::vector<int32_t> levelsVector =
types::vectorFromMap(levels, types::getSpatializerLevelMap());
const std::vector<int32_t> modesVector =
types::vectorFromMap(modes, types::getSpatializerModeMap());
const std::vector<int64_t> channelMasksVector =
types::channelMaskVectorFromString(channelMasks);
const auto [ result, str ] = sendToStatsd(SpatializerCapabilitiesFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_SPATIALIZERCAPABILITIES_REPORTED)
, headTrackingModesVector
, levelsVector
, modesVector
, channelMasksVector
);
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_SPATIALIZERCAPABILITIES_REPORTED, str);
std::lock_guard lg(mLock);
if (mFirstCreateTimeNs == 0) {
// Only update the create time once to prevent audioserver restart
// from looking like a boot.
mFirstCreateTimeNs = item->getTimestamp();
}
mSimpleLog.log("%s suffix: %s item: %s",
__func__, suffix.c_str(), item->toString().c_str());
} else {
std::string subtype = suffix.substr(0, delim);
if (subtype != "device") return; // not a device.
const std::string deviceType = suffix.substr(std::size("device.") - 1);
const int32_t deviceTypeStatsd =
types::lookup<types::AUDIO_DEVICE_INFO_TYPE, int32_t>(deviceType);
std::string address;
(void)item->get(AMEDIAMETRICS_PROP_ADDRESS, &address);
std::string enabled;
(void)item->get(AMEDIAMETRICS_PROP_ENABLED, &enabled);
std::string hasHeadTracker;
(void)item->get(AMEDIAMETRICS_PROP_HASHEADTRACKER, &hasHeadTracker);
std::string headTrackerEnabled;
(void)item->get(AMEDIAMETRICS_PROP_HEADTRACKERENABLED, &headTrackerEnabled);
std::lock_guard lg(mLock);
// Validate from our cached state
// Our deviceKey takes the device type and appends the address if any.
// This distinguishes different wireless devices for the purposes of tracking.
std::string deviceKey(deviceType);
deviceKey.append("_").append(address);
DeviceState& deviceState = mDeviceStateMap[deviceKey];
// check whether the settings event is within a certain time of spatializer creation.
const bool withinBoot =
item->getTimestamp() - mFirstCreateTimeNs < kBootDurationThreshold;
if (!enabled.empty()) {
if (enabled != deviceState.enabled) {
const int32_t settingEventStatsd =
classifySettingEvent(!deviceState.enabled.empty(), withinBoot);
deviceState.enabled = enabled;
const bool enabledStatsd = enabled == "true";
const auto [ result, str ] = sendToStatsd(SpatializerDeviceEnabledFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_SPATIALIZERDEVICEENABLED_REPORTED)
, deviceTypeStatsd
, settingEventStatsd
, enabledStatsd
);
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_SPATIALIZERDEVICEENABLED_REPORTED, str);
}
}
if (!hasHeadTracker.empty()) {
if (hasHeadTracker != deviceState.hasHeadTracker) {
const int32_t settingEventStatsd =
classifySettingEvent(!deviceState.hasHeadTracker.empty(), withinBoot);
deviceState.hasHeadTracker = hasHeadTracker;
const bool supportedStatsd = hasHeadTracker == "true";
const auto [ result, str ] = sendToStatsd(HeadTrackerDeviceSupportedFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_HEADTRACKERDEVICESUPPORTED_REPORTED)
, deviceTypeStatsd
, settingEventStatsd
, supportedStatsd
);
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_HEADTRACKERDEVICESUPPORTED_REPORTED, str);
}
}
if (!headTrackerEnabled.empty()) {
if (headTrackerEnabled != deviceState.headTrackerEnabled) {
const int32_t settingEventStatsd =
classifySettingEvent(!deviceState.headTrackerEnabled.empty(), withinBoot);
deviceState.headTrackerEnabled = headTrackerEnabled;
const bool enabledStatsd = headTrackerEnabled == "true";
const auto [ result, str ] = sendToStatsd(HeadTrackerDeviceEnabledFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_HEADTRACKERDEVICEENABLED_REPORTED)
, deviceTypeStatsd
, settingEventStatsd
, enabledStatsd
);
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_HEADTRACKERDEVICEENABLED_REPORTED, str);
}
}
mSimpleLog.log("%s deviceKey: %s item: %s",
__func__, deviceKey.c_str(), item->toString().c_str());
}
}
std::pair<std::string, int32_t> AudioAnalytics::Spatializer::dump(
int32_t lines, const char *prefix) const
{
std::lock_guard lg(mLock);
std::string s = mSimpleLog.dumpToString(prefix == nullptr ? "" : prefix, lines);
size_t n = std::count(s.begin(), s.end(), '\n');
return { s, n };
}
void AudioAnalytics::MidiLogging::onEvent(
const std::shared_ptr<const android::mediametrics::Item> &item) const {
const std::string& key = item->getKey();
const auto uid = item->getUid();
int32_t deviceId = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DEVICEID, &deviceId);
int32_t inputPortCount = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_INPUTPORTCOUNT, &inputPortCount);
int32_t outputPortCount = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_OUTPUTPORTCOUNT, &outputPortCount);
int32_t hardwareType = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_HARDWARETYPE, &hardwareType);
std::string isSharedString;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_ISSHARED, &isSharedString);
const bool isShared = (isSharedString == "true");
std::string supportsMidiUmpString;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_SUPPORTSMIDIUMP, &supportsMidiUmpString);
const bool supportsMidiUmp = (supportsMidiUmpString == "true");
std::string usingAlsaString;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_USINGALSA, &usingAlsaString);
const bool usingAlsa = (usingAlsaString == "true");
int64_t durationNs = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DURATIONNS, &durationNs);
int32_t openedCount = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_OPENEDCOUNT, &openedCount);
int32_t closedCount = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_CLOSEDCOUNT, &closedCount);
std::string deviceDisconnectedString;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_DEVICEDISCONNECTED, &deviceDisconnectedString);
const bool deviceDisconnected = (deviceDisconnectedString == "true");
int32_t totalInputBytes = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_TOTALINPUTBYTES, &totalInputBytes);
int32_t totalOutputBytes = -1;
mAudioAnalytics.mAnalyticsState->timeMachine().get(
key, AMEDIAMETRICS_PROP_TOTALOUTPUTBYTES, &totalOutputBytes);
LOG(LOG_LEVEL) << "key:" << key
<< " uid:" << uid
<< " id:" << deviceId
<< " input_port_count:" << inputPortCount
<< " output_port_count:" << outputPortCount
<< " device_type:" << hardwareType
<< " is_shared:" << isSharedString
<< " supports_ump:" << supportsMidiUmpString
<< " using_alsa:" << usingAlsaString
<< " duration_opened_ms:" << durationNs
<< " opened_count:" << openedCount
<< " closed_count:" << closedCount
<< " device_disconnected:" << deviceDisconnectedString
<< " total_input_bytes:" << totalInputBytes
<< " total_output_bytes:" << totalOutputBytes;
if (mAudioAnalytics.mDeliverStatistics) {
const auto result = sendToStatsd(
CONDITION(stats::media_metrics::MEDIAMETRICS_MIDI_DEVICE_CLOSE_REPORTED)
, uid
, deviceId
, inputPortCount
, outputPortCount
, hardwareType
, isShared
, supportsMidiUmp
, usingAlsa
, durationNs
, openedCount
, closedCount
, deviceDisconnected
, totalInputBytes
, totalOutputBytes);
std::stringstream ss;
ss << "result:" << result;
const auto fieldsStr = printFields(MidiDeviceCloseFields,
CONDITION(stats::media_metrics::MEDIAMETRICS_MIDI_DEVICE_CLOSE_REPORTED)
, uid
, deviceId
, inputPortCount
, outputPortCount
, hardwareType
, isShared
, supportsMidiUmp
, usingAlsa
, durationNs
, openedCount
, closedCount
, deviceDisconnected
, totalInputBytes
, totalOutputBytes);
ss << " " << fieldsStr;
std::string str = ss.str();
ALOGV("%s: statsd %s", __func__, str.c_str());
mAudioAnalytics.mStatsdLog->log(
stats::media_metrics::MEDIAMETRICS_MIDI_DEVICE_CLOSE_REPORTED, str);
}
}
// This method currently suppresses the name.
std::string AudioAnalytics::getDeviceNamesFromOutputDevices(std::string_view devices) const {
std::string deviceNames;
if (stringutils::hasBluetoothOutputDevice(devices)) {
deviceNames = SUPPRESSED;
#if 0 // TODO(b/161554630) sanitize name
mAudioAnalytics.mAnalyticsState->timeMachine().get(
"audio.device.bt_a2dp", AMEDIAMETRICS_PROP_NAME, &deviceNames);
// Remove | if present
stringutils::replace(deviceNames, "|", '?');
if (deviceNames.size() > STATSD_DEVICE_NAME_MAX_LENGTH) {
deviceNames.resize(STATSD_DEVICE_NAME_MAX_LENGTH); // truncate
}
#endif
}
return deviceNames;
}
} // namespace android::mediametrics