blob: c616685ba302de29c45f6533ce062a57f24abe81 [file] [log] [blame]
/**
* Copyright (C) 2014 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License"); you may not
* use this file except in compliance with the License. You may obtain a copy
* of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
* License for the specific language governing permissions and limitations
* under the License.
*/
package com.android.server.usage;
import android.app.usage.TimeSparseArray;
import android.app.usage.UsageEvents;
import android.app.usage.UsageStats;
import android.app.usage.UsageStatsManager;
import android.os.Build;
import android.os.SystemProperties;
import android.util.AtomicFile;
import android.util.Slog;
import android.util.TimeUtils;
import com.android.internal.annotations.VisibleForTesting;
import libcore.io.IoUtils;
import java.io.BufferedReader;
import java.io.BufferedWriter;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.DataInputStream;
import java.io.DataOutputStream;
import java.io.File;
import java.io.FileInputStream;
import java.io.FileNotFoundException;
import java.io.FileOutputStream;
import java.io.FileReader;
import java.io.FileWriter;
import java.io.FilenameFilter;
import java.io.InputStream;
import java.io.IOException;
import java.io.OutputStream;
import java.nio.file.Files;
import java.nio.file.StandardCopyOption;
import java.util.ArrayList;
import java.util.List;
/**
* Provides an interface to query for UsageStat data from a Protocol Buffer database.
*
* Prior to version 4, UsageStatsDatabase used XML to store Usage Stats data to disk.
* When the UsageStatsDatabase version is upgraded, the files on disk are migrated to the new
* version on init. The steps of migration are as follows:
* 1) Check if version upgrade breadcrumb exists on disk, if so skip to step 4.
* 2) Copy current files to versioned backup files.
* 3) Write a temporary breadcrumb file with some info about the backed up files.
* 4) Deserialize a versioned backup file using the info written to the breadcrumb for the
* correct deserialization methodology.
* 5) Reserialize the data read from the file with the new version format and replace the old files
* 6) Repeat Step 3 and 4 for each versioned backup file matching the breadcrumb file.
* 7) Update the version file with the new version and build fingerprint.
* 8) Delete the versioned backup files (unless flagged to be kept).
* 9) Delete the breadcrumb file.
*
* Performing the upgrade steps in this order, protects against unexpected shutdowns mid upgrade
*
* A versioned backup file is simply a copy of a Usage Stats file with some extra info embedded in
* the file name. The structure of the versioned backup filename is as followed:
* (original file name).(backup timestamp).(original file version).vak
*
* During the version upgrade process, the new upgraded file will have it's name set to the original
* file name. The backup timestamp helps distinguish between versioned backups if multiple upgrades
* and downgrades have taken place. The original file version denotes how to parse the file.
*/
public class UsageStatsDatabase {
private static final int DEFAULT_CURRENT_VERSION = 4;
// Current version of the backup schema
static final int BACKUP_VERSION = 1;
// Key under which the payload blob is stored
// same as UsageStatsBackupHelper.KEY_USAGE_STATS
static final String KEY_USAGE_STATS = "usage_stats";
// Persist versioned backup files.
// Should be false, except when testing new versions
// STOPSHIP: b/111422946 this should be false on launch
static final boolean KEEP_VAK_FILES = true;
private static final String TAG = "UsageStatsDatabase";
// STOPSHIP: b/111422946 this should be boolean DEBUG = UsageStatsService.DEBUG; on launch
private static final boolean DEBUG = true;
private static final String BAK_SUFFIX = ".bak";
private static final String VERSIONED_BAK_SUFFIX = ".vak";
private static final String CHECKED_IN_SUFFIX = UsageStatsXml.CHECKED_IN_SUFFIX;
private static final String RETENTION_LEN_KEY = "ro.usagestats.chooser.retention";
private static final int SELECTION_LOG_RETENTION_LEN =
SystemProperties.getInt(RETENTION_LEN_KEY, 14);
private final Object mLock = new Object();
private final File[] mIntervalDirs;
private final TimeSparseArray<AtomicFile>[] mSortedStatFiles;
private final UnixCalendar mCal;
private final File mVersionFile;
// If this file exists on disk, UsageStatsDatabase is in the middle of migrating files to a new
// version. If this file exists on boot, the upgrade was interrupted and needs to be picked up
// where it left off.
private final File mUpdateBreadcrumb;
// Current version of the database files schema
private final int mCurrentVersion;
private boolean mFirstUpdate;
private boolean mNewUpdate;
/**
* UsageStatsDatabase constructor that allows setting the version number.
* This should only be used for testing.
*
* @hide
*/
@VisibleForTesting
public UsageStatsDatabase(File dir, int version) {
mIntervalDirs = new File[]{
new File(dir, "daily"),
new File(dir, "weekly"),
new File(dir, "monthly"),
new File(dir, "yearly"),
};
mCurrentVersion = version;
mVersionFile = new File(dir, "version");
mUpdateBreadcrumb = new File(dir, "breadcrumb");
mSortedStatFiles = new TimeSparseArray[mIntervalDirs.length];
mCal = new UnixCalendar(0);
}
public UsageStatsDatabase(File dir) {
this(dir, DEFAULT_CURRENT_VERSION);
}
/**
* Initialize any directories required and index what stats are available.
*/
public void init(long currentTimeMillis) {
synchronized (mLock) {
for (File f : mIntervalDirs) {
f.mkdirs();
if (!f.exists()) {
throw new IllegalStateException("Failed to create directory "
+ f.getAbsolutePath());
}
}
checkVersionAndBuildLocked();
indexFilesLocked();
// Delete files that are in the future.
for (TimeSparseArray<AtomicFile> files : mSortedStatFiles) {
final int startIndex = files.closestIndexOnOrAfter(currentTimeMillis);
if (startIndex < 0) {
continue;
}
final int fileCount = files.size();
for (int i = startIndex; i < fileCount; i++) {
files.valueAt(i).delete();
}
// Remove in a separate loop because any accesses (valueAt)
// will cause a gc in the SparseArray and mess up the order.
for (int i = startIndex; i < fileCount; i++) {
files.removeAt(i);
}
}
}
}
public interface CheckinAction {
boolean checkin(IntervalStats stats);
}
/**
* Calls {@link CheckinAction#checkin(IntervalStats)} on the given {@link CheckinAction}
* for all {@link IntervalStats} that haven't been checked-in.
* If any of the calls to {@link CheckinAction#checkin(IntervalStats)} returns false or throws
* an exception, the check-in will be aborted.
*
* @param checkinAction The callback to run when checking-in {@link IntervalStats}.
* @return true if the check-in succeeded.
*/
public boolean checkinDailyFiles(CheckinAction checkinAction) {
synchronized (mLock) {
final TimeSparseArray<AtomicFile> files =
mSortedStatFiles[UsageStatsManager.INTERVAL_DAILY];
final int fileCount = files.size();
// We may have holes in the checkin (if there was an error)
// so find the last checked-in file and go from there.
int lastCheckin = -1;
for (int i = 0; i < fileCount - 1; i++) {
if (files.valueAt(i).getBaseFile().getPath().endsWith(CHECKED_IN_SUFFIX)) {
lastCheckin = i;
}
}
final int start = lastCheckin + 1;
if (start == fileCount - 1) {
return true;
}
try {
IntervalStats stats = new IntervalStats();
for (int i = start; i < fileCount - 1; i++) {
readLocked(files.valueAt(i), stats);
if (!checkinAction.checkin(stats)) {
return false;
}
}
} catch (IOException e) {
Slog.e(TAG, "Failed to check-in", e);
return false;
}
// We have successfully checked-in the stats, so rename the files so that they
// are marked as checked-in.
for (int i = start; i < fileCount - 1; i++) {
final AtomicFile file = files.valueAt(i);
final File checkedInFile = new File(
file.getBaseFile().getPath() + CHECKED_IN_SUFFIX);
if (!file.getBaseFile().renameTo(checkedInFile)) {
// We must return success, as we've already marked some files as checked-in.
// It's better to repeat ourselves than to lose data.
Slog.e(TAG, "Failed to mark file " + file.getBaseFile().getPath()
+ " as checked-in");
return true;
}
// AtomicFile needs to set a new backup path with the same -c extension, so
// we replace the old AtomicFile with the updated one.
files.setValueAt(i, new AtomicFile(checkedInFile));
}
}
return true;
}
private void indexFilesLocked() {
final FilenameFilter backupFileFilter = new FilenameFilter() {
@Override
public boolean accept(File dir, String name) {
return !name.endsWith(BAK_SUFFIX) && !name.endsWith(VERSIONED_BAK_SUFFIX);
}
};
// Index the available usage stat files on disk.
for (int i = 0; i < mSortedStatFiles.length; i++) {
if (mSortedStatFiles[i] == null) {
mSortedStatFiles[i] = new TimeSparseArray<>();
} else {
mSortedStatFiles[i].clear();
}
File[] files = mIntervalDirs[i].listFiles(backupFileFilter);
if (files != null) {
if (DEBUG) {
Slog.d(TAG, "Found " + files.length + " stat files for interval " + i);
}
for (File f : files) {
final AtomicFile af = new AtomicFile(f);
try {
mSortedStatFiles[i].put(parseBeginTime(af), af);
} catch (IOException e) {
Slog.e(TAG, "failed to index file: " + f, e);
}
}
}
}
}
/**
* Is this the first update to the system from L to M?
*/
boolean isFirstUpdate() {
return mFirstUpdate;
}
/**
* Is this a system update since we started tracking build fingerprint in the version file?
*/
boolean isNewUpdate() {
return mNewUpdate;
}
private void checkVersionAndBuildLocked() {
int version;
String buildFingerprint;
String currentFingerprint = getBuildFingerprint();
mFirstUpdate = true;
mNewUpdate = true;
try (BufferedReader reader = new BufferedReader(new FileReader(mVersionFile))) {
version = Integer.parseInt(reader.readLine());
buildFingerprint = reader.readLine();
if (buildFingerprint != null) {
mFirstUpdate = false;
}
if (currentFingerprint.equals(buildFingerprint)) {
mNewUpdate = false;
}
} catch (NumberFormatException | IOException e) {
version = 0;
}
if (version != mCurrentVersion) {
Slog.i(TAG, "Upgrading from version " + version + " to " + mCurrentVersion);
if (!mUpdateBreadcrumb.exists()) {
doUpgradeLocked(version);
} else {
Slog.i(TAG, "Version upgrade breadcrumb found on disk! Continuing version upgrade");
}
if (mUpdateBreadcrumb.exists()) {
int previousVersion;
long token;
try (BufferedReader reader = new BufferedReader(
new FileReader(mUpdateBreadcrumb))) {
token = Long.parseLong(reader.readLine());
previousVersion = Integer.parseInt(reader.readLine());
} catch (NumberFormatException | IOException e) {
Slog.e(TAG, "Failed read version upgrade breadcrumb");
throw new RuntimeException(e);
}
continueUpgradeLocked(previousVersion, token);
}
}
if (version != mCurrentVersion || mNewUpdate) {
try (BufferedWriter writer = new BufferedWriter(new FileWriter(mVersionFile))) {
writer.write(Integer.toString(mCurrentVersion));
writer.write("\n");
writer.write(currentFingerprint);
writer.write("\n");
writer.flush();
} catch (IOException e) {
Slog.e(TAG, "Failed to write new version");
throw new RuntimeException(e);
}
}
if (mUpdateBreadcrumb.exists()) {
// Files should be up to date with current version. Clear the version update breadcrumb
if (!KEEP_VAK_FILES) {
removeVersionedBackupFiles();
}
mUpdateBreadcrumb.delete();
}
}
private String getBuildFingerprint() {
return Build.VERSION.RELEASE + ";"
+ Build.VERSION.CODENAME + ";"
+ Build.VERSION.INCREMENTAL;
}
private void doUpgradeLocked(int thisVersion) {
if (thisVersion < 2) {
// Delete all files if we are version 0. This is a pre-release version,
// so this is fine.
Slog.i(TAG, "Deleting all usage stats files");
for (int i = 0; i < mIntervalDirs.length; i++) {
File[] files = mIntervalDirs[i].listFiles();
if (files != null) {
for (File f : files) {
f.delete();
}
}
}
} else {
// Turn all current usage stats files into versioned backup files
final long token = System.currentTimeMillis();
final FilenameFilter backupFileFilter = new FilenameFilter() {
@Override
public boolean accept(File dir, String name) {
return !name.endsWith(BAK_SUFFIX) && !name.endsWith(VERSIONED_BAK_SUFFIX);
}
};
for (int i = 0; i < mIntervalDirs.length; i++) {
File[] files = mIntervalDirs[i].listFiles(backupFileFilter);
if (files != null) {
for (int j = 0; j < files.length; j++) {
final File backupFile = new File(
files[j].toString() + "." + Long.toString(token) + "."
+ Integer.toString(thisVersion) + VERSIONED_BAK_SUFFIX);
if (DEBUG) {
Slog.d(TAG, "Creating versioned (" + Integer.toString(thisVersion)
+ ") backup of " + files[j].toString()
+ " stat files for interval "
+ i + " to " + backupFile.toString());
}
try {
// Backup file should not already exist, but make sure it doesn't
Files.deleteIfExists(backupFile.toPath());
Files.move(files[j].toPath(), backupFile.toPath(),
StandardCopyOption.ATOMIC_MOVE);
} catch (IOException e) {
Slog.e(TAG, "Failed to back up file : " + files[j].toString());
throw new RuntimeException(e);
}
}
}
}
// Leave a breadcrumb behind noting that all the usage stats have been copied to a
// versioned backup.
BufferedWriter writer = null;
try {
writer = new BufferedWriter(new FileWriter(mUpdateBreadcrumb));
writer.write(Long.toString(token));
writer.write("\n");
writer.write(Integer.toString(thisVersion));
writer.write("\n");
writer.flush();
} catch (IOException e) {
Slog.e(TAG, "Failed to write new version upgrade breadcrumb");
throw new RuntimeException(e);
} finally {
IoUtils.closeQuietly(writer);
}
}
}
private void continueUpgradeLocked(int version, long token) {
// Read all the backed ups for the specified version and rewrite them with the current
// version's file format.
final FilenameFilter versionedBackupFileFilter = new FilenameFilter() {
@Override
public boolean accept(File dir, String name) {
return name.endsWith("." + Long.toString(token) + "." + Integer.toString(version)
+ VERSIONED_BAK_SUFFIX);
}
};
for (int i = 0; i < mIntervalDirs.length; i++) {
File[] files = mIntervalDirs[i].listFiles(versionedBackupFileFilter);
if (files != null) {
for (int j = 0; j < files.length; j++) {
if (DEBUG) {
Slog.d(TAG,
"Upgrading " + files[j].toString() + " to version ("
+ Integer.toString(
mCurrentVersion) + ") for interval " + i);
}
try {
IntervalStats stats = new IntervalStats();
readLocked(new AtomicFile(files[j]), stats, version);
writeLocked(new AtomicFile(new File(mIntervalDirs[i],
Long.toString(stats.beginTime))), stats, mCurrentVersion);
} catch (IOException e) {
Slog.e(TAG,
"Failed to upgrade versioned backup file : " + files[j].toString());
throw new RuntimeException(e);
}
}
}
}
}
private void removeVersionedBackupFiles() {
final FilenameFilter versionedBackupFileFilter = new FilenameFilter() {
@Override
public boolean accept(File dir, String name) {
return name.endsWith(VERSIONED_BAK_SUFFIX);
}
};
for (int i = 0; i < mIntervalDirs.length; i++) {
File[] files = mIntervalDirs[i].listFiles(versionedBackupFileFilter);
if (files != null) {
for (int j = 0; j < files.length; j++) {
if (DEBUG) {
Slog.d(TAG,
"Removing " + files[j].toString() + " for interval " + i);
}
if (!files[j].delete()) {
Slog.e(TAG, "Failed to delete file : " + files[j].toString());
}
}
}
}
}
public void onTimeChanged(long timeDiffMillis) {
synchronized (mLock) {
StringBuilder logBuilder = new StringBuilder();
logBuilder.append("Time changed by ");
TimeUtils.formatDuration(timeDiffMillis, logBuilder);
logBuilder.append(".");
int filesDeleted = 0;
int filesMoved = 0;
for (TimeSparseArray<AtomicFile> files : mSortedStatFiles) {
final int fileCount = files.size();
for (int i = 0; i < fileCount; i++) {
final AtomicFile file = files.valueAt(i);
final long newTime = files.keyAt(i) + timeDiffMillis;
if (newTime < 0) {
filesDeleted++;
file.delete();
} else {
try {
file.openRead().close();
} catch (IOException e) {
// Ignore, this is just to make sure there are no backups.
}
String newName = Long.toString(newTime);
if (file.getBaseFile().getName().endsWith(CHECKED_IN_SUFFIX)) {
newName = newName + CHECKED_IN_SUFFIX;
}
final File newFile = new File(file.getBaseFile().getParentFile(), newName);
filesMoved++;
file.getBaseFile().renameTo(newFile);
}
}
files.clear();
}
logBuilder.append(" files deleted: ").append(filesDeleted);
logBuilder.append(" files moved: ").append(filesMoved);
Slog.i(TAG, logBuilder.toString());
// Now re-index the new files.
indexFilesLocked();
}
}
/**
* Get the latest stats that exist for this interval type.
*/
public IntervalStats getLatestUsageStats(int intervalType) {
synchronized (mLock) {
if (intervalType < 0 || intervalType >= mIntervalDirs.length) {
throw new IllegalArgumentException("Bad interval type " + intervalType);
}
final int fileCount = mSortedStatFiles[intervalType].size();
if (fileCount == 0) {
return null;
}
try {
final AtomicFile f = mSortedStatFiles[intervalType].valueAt(fileCount - 1);
IntervalStats stats = new IntervalStats();
readLocked(f, stats);
return stats;
} catch (IOException e) {
Slog.e(TAG, "Failed to read usage stats file", e);
}
}
return null;
}
/**
* Figures out what to extract from the given IntervalStats object.
*/
public interface StatCombiner<T> {
/**
* Implementations should extract interesting from <code>stats</code> and add it
* to the <code>accumulatedResult</code> list.
*
* If the <code>stats</code> object is mutable, <code>mutable</code> will be true,
* which means you should make a copy of the data before adding it to the
* <code>accumulatedResult</code> list.
*
* @param stats The {@link IntervalStats} object selected.
* @param mutable Whether or not the data inside the stats object is mutable.
* @param accumulatedResult The list to which to add extracted data.
*/
void combine(IntervalStats stats, boolean mutable, List<T> accumulatedResult);
}
/**
* Find all {@link IntervalStats} for the given range and interval type.
*/
public <T> List<T> queryUsageStats(int intervalType, long beginTime, long endTime,
StatCombiner<T> combiner) {
synchronized (mLock) {
if (intervalType < 0 || intervalType >= mIntervalDirs.length) {
throw new IllegalArgumentException("Bad interval type " + intervalType);
}
final TimeSparseArray<AtomicFile> intervalStats = mSortedStatFiles[intervalType];
if (endTime <= beginTime) {
if (DEBUG) {
Slog.d(TAG, "endTime(" + endTime + ") <= beginTime(" + beginTime + ")");
}
return null;
}
int startIndex = intervalStats.closestIndexOnOrBefore(beginTime);
if (startIndex < 0) {
// All the stats available have timestamps after beginTime, which means they all
// match.
startIndex = 0;
}
int endIndex = intervalStats.closestIndexOnOrBefore(endTime);
if (endIndex < 0) {
// All the stats start after this range ends, so nothing matches.
if (DEBUG) {
Slog.d(TAG, "No results for this range. All stats start after.");
}
return null;
}
if (intervalStats.keyAt(endIndex) == endTime) {
// The endTime is exclusive, so if we matched exactly take the one before.
endIndex--;
if (endIndex < 0) {
// All the stats start after this range ends, so nothing matches.
if (DEBUG) {
Slog.d(TAG, "No results for this range. All stats start after.");
}
return null;
}
}
final IntervalStats stats = new IntervalStats();
final ArrayList<T> results = new ArrayList<>();
for (int i = startIndex; i <= endIndex; i++) {
final AtomicFile f = intervalStats.valueAt(i);
if (DEBUG) {
Slog.d(TAG, "Reading stat file " + f.getBaseFile().getAbsolutePath());
}
try {
readLocked(f, stats);
if (beginTime < stats.endTime) {
combiner.combine(stats, false, results);
}
} catch (IOException e) {
Slog.e(TAG, "Failed to read usage stats file", e);
// We continue so that we return results that are not
// corrupt.
}
}
return results;
}
}
/**
* Find the interval that best matches this range.
*
* TODO(adamlesinski): Use endTimeStamp in best fit calculation.
*/
public int findBestFitBucket(long beginTimeStamp, long endTimeStamp) {
synchronized (mLock) {
int bestBucket = -1;
long smallestDiff = Long.MAX_VALUE;
for (int i = mSortedStatFiles.length - 1; i >= 0; i--) {
final int index = mSortedStatFiles[i].closestIndexOnOrBefore(beginTimeStamp);
int size = mSortedStatFiles[i].size();
if (index >= 0 && index < size) {
// We have some results here, check if they are better than our current match.
long diff = Math.abs(mSortedStatFiles[i].keyAt(index) - beginTimeStamp);
if (diff < smallestDiff) {
smallestDiff = diff;
bestBucket = i;
}
}
}
return bestBucket;
}
}
/**
* Remove any usage stat files that are too old.
*/
public void prune(final long currentTimeMillis) {
synchronized (mLock) {
mCal.setTimeInMillis(currentTimeMillis);
mCal.addYears(-3);
pruneFilesOlderThan(mIntervalDirs[UsageStatsManager.INTERVAL_YEARLY],
mCal.getTimeInMillis());
mCal.setTimeInMillis(currentTimeMillis);
mCal.addMonths(-6);
pruneFilesOlderThan(mIntervalDirs[UsageStatsManager.INTERVAL_MONTHLY],
mCal.getTimeInMillis());
mCal.setTimeInMillis(currentTimeMillis);
mCal.addWeeks(-4);
pruneFilesOlderThan(mIntervalDirs[UsageStatsManager.INTERVAL_WEEKLY],
mCal.getTimeInMillis());
mCal.setTimeInMillis(currentTimeMillis);
mCal.addDays(-10);
pruneFilesOlderThan(mIntervalDirs[UsageStatsManager.INTERVAL_DAILY],
mCal.getTimeInMillis());
mCal.setTimeInMillis(currentTimeMillis);
mCal.addDays(-SELECTION_LOG_RETENTION_LEN);
for (int i = 0; i < mIntervalDirs.length; ++i) {
pruneChooserCountsOlderThan(mIntervalDirs[i], mCal.getTimeInMillis());
}
// We must re-index our file list or we will be trying to read
// deleted files.
indexFilesLocked();
}
}
private static void pruneFilesOlderThan(File dir, long expiryTime) {
File[] files = dir.listFiles();
if (files != null) {
for (File f : files) {
long beginTime;
try {
beginTime = parseBeginTime(f);
} catch (IOException e) {
beginTime = 0;
}
if (beginTime < expiryTime) {
new AtomicFile(f).delete();
}
}
}
}
private void pruneChooserCountsOlderThan(File dir, long expiryTime) {
File[] files = dir.listFiles();
if (files != null) {
for (File f : files) {
long beginTime;
try {
beginTime = parseBeginTime(f);
} catch (IOException e) {
beginTime = 0;
}
if (beginTime < expiryTime) {
try {
final AtomicFile af = new AtomicFile(f);
final IntervalStats stats = new IntervalStats();
readLocked(af, stats);
final int pkgCount = stats.packageStats.size();
for (int i = 0; i < pkgCount; i++) {
UsageStats pkgStats = stats.packageStats.valueAt(i);
if (pkgStats.mChooserCounts != null) {
pkgStats.mChooserCounts.clear();
}
}
writeLocked(af, stats);
} catch (IOException e) {
Slog.e(TAG, "Failed to delete chooser counts from usage stats file", e);
}
}
}
}
}
private static long parseBeginTime(AtomicFile file) throws IOException {
return parseBeginTime(file.getBaseFile());
}
private static long parseBeginTime(File file) throws IOException {
String name = file.getName();
// Parse out the digits from the the front of the file name
for (int i = 0; i < name.length(); i++) {
final char c = name.charAt(i);
if (c < '0' || c > '9') {
// found first char that is not a digit.
name = name.substring(0, i);
break;
}
}
try {
return Long.parseLong(name);
} catch (NumberFormatException e) {
throw new IOException(e);
}
}
private void writeLocked(AtomicFile file, IntervalStats stats) throws IOException {
writeLocked(file, stats, mCurrentVersion);
}
private static void writeLocked(AtomicFile file, IntervalStats stats, int version)
throws IOException {
FileOutputStream fos = file.startWrite();
try {
writeLocked(fos, stats, version);
file.finishWrite(fos);
fos = null;
} finally {
// When fos is null (successful write), this will no-op
file.failWrite(fos);
}
}
private void writeLocked(OutputStream out, IntervalStats stats) throws IOException {
writeLocked(out, stats, mCurrentVersion);
}
private static void writeLocked(OutputStream out, IntervalStats stats, int version)
throws IOException {
switch (version) {
case 1:
case 2:
case 3:
UsageStatsXml.write(out, stats);
break;
case 4:
UsageStatsProto.write(out, stats);
break;
default:
throw new RuntimeException(
"Unhandled UsageStatsDatabase version: " + Integer.toString(version)
+ " on write.");
}
}
private void readLocked(AtomicFile file, IntervalStats statsOut) throws IOException {
readLocked(file, statsOut, mCurrentVersion);
}
private static void readLocked(AtomicFile file, IntervalStats statsOut, int version)
throws IOException {
try {
FileInputStream in = file.openRead();
try {
statsOut.beginTime = parseBeginTime(file);
readLocked(in, statsOut, version);
statsOut.lastTimeSaved = file.getLastModifiedTime();
} finally {
try {
in.close();
} catch (IOException e) {
// Empty
}
}
} catch (FileNotFoundException e) {
Slog.e(TAG, "UsageStatsDatabase", e);
throw e;
}
// STOPSHIP: b/111422946, b/115429334
// Everything below this comment is sanity check against the new database version.
// After the new version has soaked for some time the following should removed.
// The goal of this check is to make sure the the ProtoInputStream is properly reading from
// the UsageStats files.
final StringBuilder sb = new StringBuilder();
final int failureLogLimit = 10;
int failures = 0;
final int packagesSize = statsOut.packageStats.size();
for (int i = 0; i < packagesSize; i++) {
final UsageStats stat = statsOut.packageStats.valueAt(i);
if (stat == null) {
// ArrayMap may contain null values, skip them
continue;
}
if (stat.mPackageName.isEmpty()) {
if (failures++ < failureLogLimit) {
sb.append("\nUnexpected empty usage stats package name loaded");
}
}
if (stat.mBeginTimeStamp > statsOut.endTime) {
if (failures++ < failureLogLimit) {
sb.append("\nUnreasonable usage stats stat begin timestamp ");
sb.append(stat.mBeginTimeStamp);
sb.append(" loaded (beginTime : ");
sb.append(statsOut.beginTime);
sb.append(", endTime : ");
sb.append(statsOut.endTime);
sb.append(")");
}
}
if (stat.mEndTimeStamp > statsOut.endTime) {
if (failures++ < failureLogLimit) {
sb.append("\nUnreasonable usage stats stat end timestamp ");
sb.append(stat.mEndTimeStamp);
sb.append(" loaded (beginTime : ");
sb.append(statsOut.beginTime);
sb.append(", endTime : ");
sb.append(statsOut.endTime);
sb.append(")");
}
}
if (stat.mLastTimeUsed > statsOut.endTime) {
if (failures++ < failureLogLimit) {
sb.append("\nUnreasonable usage stats stat last used timestamp ");
sb.append(stat.mLastTimeUsed);
sb.append(" loaded (beginTime : ");
sb.append(statsOut.beginTime);
sb.append(", endTime : ");
sb.append(statsOut.endTime);
sb.append(")");
}
}
}
if (statsOut.events != null) {
final int eventSize = statsOut.events.size();
for (int i = 0; i < eventSize; i++) {
final UsageEvents.Event event = statsOut.events.get(i);
if (event.mPackage.isEmpty()) {
if (failures++ < failureLogLimit) {
sb.append("\nUnexpected empty empty package name loaded");
}
}
if (event.mTimeStamp < statsOut.beginTime || event.mTimeStamp > statsOut.endTime) {
if (failures++ < failureLogLimit) {
sb.append("\nUnexpected event timestamp ");
sb.append(event.mTimeStamp);
sb.append(" loaded (beginTime : ");
sb.append(statsOut.beginTime);
sb.append(", endTime : ");
sb.append(statsOut.endTime);
sb.append(")");
}
}
if (event.mEventType < 0 || event.mEventType > UsageEvents.Event.MAX_EVENT_TYPE) {
if (failures++ < failureLogLimit) {
sb.append("\nUnexpected event type ");
sb.append(event.mEventType);
sb.append(" loaded");
}
}
if ((event.mFlags & ~UsageEvents.Event.VALID_FLAG_BITS) != 0) {
if (failures++ < failureLogLimit) {
sb.append("\nUnexpected event flag bit 0b");
sb.append(Integer.toBinaryString(event.mFlags));
sb.append(" loaded");
}
}
}
}
if (failures != 0) {
if (failures > failureLogLimit) {
sb.append("\nFailure log limited (");
sb.append(failures);
sb.append(" total failures found!)");
}
sb.append("\nError found in:\n");
sb.append(file.getBaseFile().getAbsolutePath());
sb.append("\nPlease go to b/115429334 to help root cause this issue");
Slog.wtf(TAG,sb.toString());
}
}
private void readLocked(InputStream in, IntervalStats statsOut) throws IOException {
readLocked(in, statsOut, mCurrentVersion);
}
private static void readLocked(InputStream in, IntervalStats statsOut, int version)
throws IOException {
switch (version) {
case 1:
case 2:
case 3:
UsageStatsXml.read(in, statsOut);
break;
case 4:
UsageStatsProto.read(in, statsOut);
break;
default:
throw new RuntimeException(
"Unhandled UsageStatsDatabase version: " + Integer.toString(version)
+ " on read.");
}
}
/**
* Update the stats in the database. They may not be written to disk immediately.
*/
public void putUsageStats(int intervalType, IntervalStats stats) throws IOException {
if (stats == null) return;
synchronized (mLock) {
if (intervalType < 0 || intervalType >= mIntervalDirs.length) {
throw new IllegalArgumentException("Bad interval type " + intervalType);
}
AtomicFile f = mSortedStatFiles[intervalType].get(stats.beginTime);
if (f == null) {
f = new AtomicFile(new File(mIntervalDirs[intervalType],
Long.toString(stats.beginTime)));
mSortedStatFiles[intervalType].put(stats.beginTime, f);
}
writeLocked(f, stats);
stats.lastTimeSaved = f.getLastModifiedTime();
}
}
/* Backup/Restore Code */
byte[] getBackupPayload(String key) {
synchronized (mLock) {
ByteArrayOutputStream baos = new ByteArrayOutputStream();
if (KEY_USAGE_STATS.equals(key)) {
prune(System.currentTimeMillis());
DataOutputStream out = new DataOutputStream(baos);
try {
out.writeInt(BACKUP_VERSION);
out.writeInt(mSortedStatFiles[UsageStatsManager.INTERVAL_DAILY].size());
for (int i = 0; i < mSortedStatFiles[UsageStatsManager.INTERVAL_DAILY].size();
i++) {
writeIntervalStatsToStream(out,
mSortedStatFiles[UsageStatsManager.INTERVAL_DAILY].valueAt(i));
}
out.writeInt(mSortedStatFiles[UsageStatsManager.INTERVAL_WEEKLY].size());
for (int i = 0; i < mSortedStatFiles[UsageStatsManager.INTERVAL_WEEKLY].size();
i++) {
writeIntervalStatsToStream(out,
mSortedStatFiles[UsageStatsManager.INTERVAL_WEEKLY].valueAt(i));
}
out.writeInt(mSortedStatFiles[UsageStatsManager.INTERVAL_MONTHLY].size());
for (int i = 0; i < mSortedStatFiles[UsageStatsManager.INTERVAL_MONTHLY].size();
i++) {
writeIntervalStatsToStream(out,
mSortedStatFiles[UsageStatsManager.INTERVAL_MONTHLY].valueAt(i));
}
out.writeInt(mSortedStatFiles[UsageStatsManager.INTERVAL_YEARLY].size());
for (int i = 0; i < mSortedStatFiles[UsageStatsManager.INTERVAL_YEARLY].size();
i++) {
writeIntervalStatsToStream(out,
mSortedStatFiles[UsageStatsManager.INTERVAL_YEARLY].valueAt(i));
}
if (DEBUG) Slog.i(TAG, "Written " + baos.size() + " bytes of data");
} catch (IOException ioe) {
Slog.d(TAG, "Failed to write data to output stream", ioe);
baos.reset();
}
}
return baos.toByteArray();
}
}
void applyRestoredPayload(String key, byte[] payload) {
synchronized (mLock) {
if (KEY_USAGE_STATS.equals(key)) {
// Read stats files for the current device configs
IntervalStats dailyConfigSource =
getLatestUsageStats(UsageStatsManager.INTERVAL_DAILY);
IntervalStats weeklyConfigSource =
getLatestUsageStats(UsageStatsManager.INTERVAL_WEEKLY);
IntervalStats monthlyConfigSource =
getLatestUsageStats(UsageStatsManager.INTERVAL_MONTHLY);
IntervalStats yearlyConfigSource =
getLatestUsageStats(UsageStatsManager.INTERVAL_YEARLY);
try {
DataInputStream in = new DataInputStream(new ByteArrayInputStream(payload));
int backupDataVersion = in.readInt();
// Can't handle this backup set
if (backupDataVersion < 1 || backupDataVersion > BACKUP_VERSION) return;
// Delete all stats files
// Do this after reading version and before actually restoring
for (int i = 0; i < mIntervalDirs.length; i++) {
deleteDirectoryContents(mIntervalDirs[i]);
}
int fileCount = in.readInt();
for (int i = 0; i < fileCount; i++) {
IntervalStats stats = deserializeIntervalStats(getIntervalStatsBytes(in));
stats = mergeStats(stats, dailyConfigSource);
putUsageStats(UsageStatsManager.INTERVAL_DAILY, stats);
}
fileCount = in.readInt();
for (int i = 0; i < fileCount; i++) {
IntervalStats stats = deserializeIntervalStats(getIntervalStatsBytes(in));
stats = mergeStats(stats, weeklyConfigSource);
putUsageStats(UsageStatsManager.INTERVAL_WEEKLY, stats);
}
fileCount = in.readInt();
for (int i = 0; i < fileCount; i++) {
IntervalStats stats = deserializeIntervalStats(getIntervalStatsBytes(in));
stats = mergeStats(stats, monthlyConfigSource);
putUsageStats(UsageStatsManager.INTERVAL_MONTHLY, stats);
}
fileCount = in.readInt();
for (int i = 0; i < fileCount; i++) {
IntervalStats stats = deserializeIntervalStats(getIntervalStatsBytes(in));
stats = mergeStats(stats, yearlyConfigSource);
putUsageStats(UsageStatsManager.INTERVAL_YEARLY, stats);
}
if (DEBUG) Slog.i(TAG, "Completed Restoring UsageStats");
} catch (IOException ioe) {
Slog.d(TAG, "Failed to read data from input stream", ioe);
} finally {
indexFilesLocked();
}
}
}
}
/**
* Get the Configuration Statistics from the current device statistics and merge them
* with the backed up usage statistics.
*/
private IntervalStats mergeStats(IntervalStats beingRestored, IntervalStats onDevice) {
if (onDevice == null) return beingRestored;
if (beingRestored == null) return null;
beingRestored.activeConfiguration = onDevice.activeConfiguration;
beingRestored.configurations.putAll(onDevice.configurations);
beingRestored.events = onDevice.events;
return beingRestored;
}
private void writeIntervalStatsToStream(DataOutputStream out, AtomicFile statsFile)
throws IOException {
IntervalStats stats = new IntervalStats();
try {
readLocked(statsFile, stats);
} catch (IOException e) {
Slog.e(TAG, "Failed to read usage stats file", e);
out.writeInt(0);
return;
}
sanitizeIntervalStatsForBackup(stats);
byte[] data = serializeIntervalStats(stats);
out.writeInt(data.length);
out.write(data);
}
private static byte[] getIntervalStatsBytes(DataInputStream in) throws IOException {
int length = in.readInt();
byte[] buffer = new byte[length];
in.read(buffer, 0, length);
return buffer;
}
private static void sanitizeIntervalStatsForBackup(IntervalStats stats) {
if (stats == null) return;
stats.activeConfiguration = null;
stats.configurations.clear();
if (stats.events != null) stats.events.clear();
}
private byte[] serializeIntervalStats(IntervalStats stats) {
ByteArrayOutputStream baos = new ByteArrayOutputStream();
DataOutputStream out = new DataOutputStream(baos);
try {
out.writeLong(stats.beginTime);
writeLocked(out, stats);
} catch (IOException ioe) {
Slog.d(TAG, "Serializing IntervalStats Failed", ioe);
baos.reset();
}
return baos.toByteArray();
}
private IntervalStats deserializeIntervalStats(byte[] data) {
ByteArrayInputStream bais = new ByteArrayInputStream(data);
DataInputStream in = new DataInputStream(bais);
IntervalStats stats = new IntervalStats();
try {
stats.beginTime = in.readLong();
readLocked(in, stats);
} catch (IOException ioe) {
Slog.d(TAG, "DeSerializing IntervalStats Failed", ioe);
stats = null;
}
return stats;
}
private static void deleteDirectoryContents(File directory) {
File[] files = directory.listFiles();
for (File file : files) {
deleteDirectory(file);
}
}
private static void deleteDirectory(File directory) {
File[] files = directory.listFiles();
if (files != null) {
for (File file : files) {
if (!file.isDirectory()) {
file.delete();
} else {
deleteDirectory(file);
}
}
}
directory.delete();
}
}