| /* |
| * Copyright (C) 2005 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define LOG_TAG "hw-Parcel" |
| //#define LOG_NDEBUG 0 |
| |
| #include <errno.h> |
| #include <fcntl.h> |
| #include <inttypes.h> |
| #include <pthread.h> |
| #include <stdint.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <sys/mman.h> |
| #include <sys/stat.h> |
| #include <sys/types.h> |
| #include <sys/resource.h> |
| #include <unistd.h> |
| |
| #include <hwbinder/Binder.h> |
| #include <hwbinder/BpHwBinder.h> |
| #include <hwbinder/IPCThreadState.h> |
| #include <hwbinder/Parcel.h> |
| #include <hwbinder/ProcessState.h> |
| |
| #include <cutils/ashmem.h> |
| #include <utils/Log.h> |
| #include <utils/misc.h> |
| #include <utils/String8.h> |
| #include <utils/String16.h> |
| |
| #include "binder_kernel.h" |
| #include <hwbinder/Static.h> |
| #include "TextOutput.h" |
| #include "Utils.h" |
| |
| #include <atomic> |
| |
| #define LOG_REFS(...) |
| //#define LOG_REFS(...) ALOG(LOG_DEBUG, LOG_TAG, __VA_ARGS__) |
| #define LOG_ALLOC(...) |
| //#define LOG_ALLOC(...) ALOG(LOG_DEBUG, LOG_TAG, __VA_ARGS__) |
| #define LOG_BUFFER(...) |
| // #define LOG_BUFFER(...) ALOG(LOG_DEBUG, LOG_TAG, __VA_ARGS__) |
| |
| // --------------------------------------------------------------------------- |
| |
| // This macro should never be used at runtime, as a too large value |
| // of s could cause an integer overflow. Instead, you should always |
| // use the wrapper function pad_size() |
| #define PAD_SIZE_UNSAFE(s) (((s)+3)&~3) |
| |
| static size_t pad_size(size_t s) { |
| if (s > (std::numeric_limits<size_t>::max() - 3)) { |
| LOG_ALWAYS_FATAL("pad size too big %zu", s); |
| } |
| return PAD_SIZE_UNSAFE(s); |
| } |
| |
| // Note: must be kept in sync with android/os/StrictMode.java's PENALTY_GATHER |
| #define STRICT_MODE_PENALTY_GATHER (0x40 << 16) |
| |
| namespace android { |
| namespace hardware { |
| |
| static std::atomic<size_t> gParcelGlobalAllocCount; |
| static std::atomic<size_t> gParcelGlobalAllocSize; |
| |
| static size_t gMaxFds = 0; |
| |
| void acquire_binder_object(const sp<ProcessState>& proc, |
| const flat_binder_object& obj, const void* who) |
| { |
| switch (obj.hdr.type) { |
| case BINDER_TYPE_BINDER: |
| if (obj.binder) { |
| LOG_REFS("Parcel %p acquiring reference on local %llu", who, obj.cookie); |
| reinterpret_cast<IBinder*>(obj.cookie)->incStrong(who); |
| } |
| return; |
| case BINDER_TYPE_WEAK_BINDER: |
| if (obj.binder) |
| reinterpret_cast<RefBase::weakref_type*>(obj.binder)->incWeak(who); |
| return; |
| case BINDER_TYPE_HANDLE: { |
| const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle); |
| if (b != nullptr) { |
| LOG_REFS("Parcel %p acquiring reference on remote %p", who, b.get()); |
| b->incStrong(who); |
| } |
| return; |
| } |
| case BINDER_TYPE_WEAK_HANDLE: { |
| const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle); |
| if (b != nullptr) b.get_refs()->incWeak(who); |
| return; |
| } |
| } |
| |
| ALOGD("Invalid object type 0x%08x", obj.hdr.type); |
| } |
| |
| void acquire_object(const sp<ProcessState>& proc, const binder_object_header& obj, |
| const void *who) { |
| switch (obj.type) { |
| case BINDER_TYPE_BINDER: |
| case BINDER_TYPE_WEAK_BINDER: |
| case BINDER_TYPE_HANDLE: |
| case BINDER_TYPE_WEAK_HANDLE: { |
| const flat_binder_object& fbo = reinterpret_cast<const flat_binder_object&>(obj); |
| acquire_binder_object(proc, fbo, who); |
| break; |
| } |
| } |
| } |
| |
| void release_object(const sp<ProcessState>& proc, |
| const flat_binder_object& obj, const void* who) |
| { |
| switch (obj.hdr.type) { |
| case BINDER_TYPE_BINDER: |
| if (obj.binder) { |
| LOG_REFS("Parcel %p releasing reference on local %llu", who, obj.cookie); |
| reinterpret_cast<IBinder*>(obj.cookie)->decStrong(who); |
| } |
| return; |
| case BINDER_TYPE_WEAK_BINDER: |
| if (obj.binder) |
| reinterpret_cast<RefBase::weakref_type*>(obj.binder)->decWeak(who); |
| return; |
| case BINDER_TYPE_HANDLE: { |
| const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle); |
| if (b != nullptr) { |
| LOG_REFS("Parcel %p releasing reference on remote %p", who, b.get()); |
| b->decStrong(who); |
| } |
| return; |
| } |
| case BINDER_TYPE_WEAK_HANDLE: { |
| const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle); |
| if (b != nullptr) b.get_refs()->decWeak(who); |
| return; |
| } |
| case BINDER_TYPE_FD: { |
| if (obj.cookie != 0) { // owned |
| close(obj.handle); |
| } |
| return; |
| } |
| case BINDER_TYPE_PTR: { |
| // The relevant buffer is part of the transaction buffer and will be freed that way |
| return; |
| } |
| case BINDER_TYPE_FDA: { |
| // The enclosed file descriptors are closed in the kernel |
| return; |
| } |
| } |
| |
| ALOGE("Invalid object type 0x%08x", obj.hdr.type); |
| } |
| |
| inline static status_t finish_flatten_binder( |
| const sp<IBinder>& /*binder*/, const flat_binder_object& flat, Parcel* out) |
| { |
| return out->writeObject(flat); |
| } |
| |
| status_t flatten_binder(const sp<ProcessState>& /*proc*/, |
| const sp<IBinder>& binder, Parcel* out) |
| { |
| flat_binder_object obj = {}; |
| |
| if (binder != nullptr) { |
| BHwBinder *local = binder->localBinder(); |
| if (!local) { |
| BpHwBinder *proxy = binder->remoteBinder(); |
| if (proxy == nullptr) { |
| ALOGE("null proxy"); |
| } |
| const int32_t handle = proxy ? proxy->handle() : 0; |
| obj.hdr.type = BINDER_TYPE_HANDLE; |
| obj.flags = FLAT_BINDER_FLAG_ACCEPTS_FDS; |
| obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */ |
| obj.handle = handle; |
| obj.cookie = 0; |
| } else { |
| // Get policy and convert it |
| int policy = local->getMinSchedulingPolicy(); |
| int priority = local->getMinSchedulingPriority(); |
| |
| obj.flags = priority & FLAT_BINDER_FLAG_PRIORITY_MASK; |
| obj.flags |= FLAT_BINDER_FLAG_ACCEPTS_FDS | FLAT_BINDER_FLAG_INHERIT_RT; |
| obj.flags |= (policy & 3) << FLAT_BINDER_FLAG_SCHED_POLICY_SHIFT; |
| if (local->isRequestingSid()) { |
| obj.flags |= FLAT_BINDER_FLAG_TXN_SECURITY_CTX; |
| } |
| obj.hdr.type = BINDER_TYPE_BINDER; |
| obj.binder = reinterpret_cast<uintptr_t>(local->getWeakRefs()); |
| obj.cookie = reinterpret_cast<uintptr_t>(local); |
| } |
| } else { |
| obj.hdr.type = BINDER_TYPE_BINDER; |
| obj.binder = 0; |
| obj.cookie = 0; |
| } |
| |
| return finish_flatten_binder(binder, obj, out); |
| } |
| |
| inline static status_t finish_unflatten_binder( |
| BpHwBinder* /*proxy*/, const flat_binder_object& /*flat*/, |
| const Parcel& /*in*/) |
| { |
| return NO_ERROR; |
| } |
| |
| status_t unflatten_binder(const sp<ProcessState>& proc, |
| const Parcel& in, sp<IBinder>* out) |
| { |
| const flat_binder_object* flat = in.readObject<flat_binder_object>(); |
| |
| if (flat) { |
| switch (flat->hdr.type) { |
| case BINDER_TYPE_BINDER: |
| *out = reinterpret_cast<IBinder*>(flat->cookie); |
| return finish_unflatten_binder(nullptr, *flat, in); |
| case BINDER_TYPE_HANDLE: |
| *out = proc->getStrongProxyForHandle(flat->handle); |
| return finish_unflatten_binder( |
| static_cast<BpHwBinder*>(out->get()), *flat, in); |
| } |
| } |
| return BAD_TYPE; |
| } |
| |
| // --------------------------------------------------------------------------- |
| |
| Parcel::Parcel() |
| { |
| LOG_ALLOC("Parcel %p: constructing", this); |
| initState(); |
| } |
| |
| Parcel::~Parcel() |
| { |
| freeDataNoInit(); |
| LOG_ALLOC("Parcel %p: destroyed", this); |
| } |
| |
| size_t Parcel::getGlobalAllocSize() { |
| return gParcelGlobalAllocSize.load(); |
| } |
| |
| size_t Parcel::getGlobalAllocCount() { |
| return gParcelGlobalAllocCount.load(); |
| } |
| |
| const uint8_t* Parcel::data() const |
| { |
| return mData; |
| } |
| |
| size_t Parcel::dataSize() const |
| { |
| return (mDataSize > mDataPos ? mDataSize : mDataPos); |
| } |
| |
| size_t Parcel::dataAvail() const |
| { |
| size_t result = dataSize() - dataPosition(); |
| if (result > INT32_MAX) { |
| LOG_ALWAYS_FATAL("result too big: %zu", result); |
| } |
| return result; |
| } |
| |
| size_t Parcel::dataPosition() const |
| { |
| return mDataPos; |
| } |
| |
| size_t Parcel::dataCapacity() const |
| { |
| return mDataCapacity; |
| } |
| |
| status_t Parcel::setDataSize(size_t size) |
| { |
| if (size > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| status_t err; |
| err = continueWrite(size); |
| if (err == NO_ERROR) { |
| mDataSize = size; |
| ALOGV("setDataSize Setting data size of %p to %zu", this, mDataSize); |
| } |
| return err; |
| } |
| |
| void Parcel::setDataPosition(size_t pos) const |
| { |
| if (pos > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| LOG_ALWAYS_FATAL("pos too big: %zu", pos); |
| } |
| |
| mDataPos = pos; |
| mNextObjectHint = 0; |
| } |
| |
| status_t Parcel::setDataCapacity(size_t size) |
| { |
| if (size > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| if (size > mDataCapacity) return continueWrite(size); |
| return NO_ERROR; |
| } |
| |
| status_t Parcel::setData(const uint8_t* buffer, size_t len) |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| status_t err = restartWrite(len); |
| if (err == NO_ERROR) { |
| memcpy(const_cast<uint8_t*>(data()), buffer, len); |
| mDataSize = len; |
| mFdsKnown = false; |
| } |
| return err; |
| } |
| |
| void Parcel::markSensitive() const |
| { |
| mDeallocZero = true; |
| } |
| |
| // Write RPC headers. (previously just the interface token) |
| status_t Parcel::writeInterfaceToken(const char* interface) |
| { |
| // currently the interface identification token is just its name as a string |
| return writeCString(interface); |
| } |
| |
| bool Parcel::enforceInterface(const char* interface) const |
| { |
| const char* str = readCString(); |
| if (str != nullptr && strcmp(str, interface) == 0) { |
| return true; |
| } else { |
| ALOGW("**** enforceInterface() expected '%s' but read '%s'", |
| interface, (str ? str : "<empty string>")); |
| return false; |
| } |
| } |
| |
| const binder_size_t* Parcel::objects() const |
| { |
| return mObjects; |
| } |
| |
| size_t Parcel::objectsCount() const |
| { |
| return mObjectsSize; |
| } |
| |
| status_t Parcel::errorCheck() const |
| { |
| return mError; |
| } |
| |
| void Parcel::setError(status_t err) |
| { |
| mError = err; |
| } |
| |
| status_t Parcel::finishWrite(size_t len) |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| //printf("Finish write of %d\n", len); |
| mDataPos += len; |
| ALOGV("finishWrite Setting data pos of %p to %zu", this, mDataPos); |
| if (mDataPos > mDataSize) { |
| mDataSize = mDataPos; |
| ALOGV("finishWrite Setting data size of %p to %zu", this, mDataSize); |
| } |
| //printf("New pos=%d, size=%d\n", mDataPos, mDataSize); |
| return NO_ERROR; |
| } |
| |
| status_t Parcel::writeUnpadded(const void* data, size_t len) |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| size_t end = mDataPos + len; |
| if (end < mDataPos) { |
| // integer overflow |
| return BAD_VALUE; |
| } |
| |
| if (end <= mDataCapacity) { |
| restart_write: |
| memcpy(mData+mDataPos, data, len); |
| return finishWrite(len); |
| } |
| |
| status_t err = growData(len); |
| if (err == NO_ERROR) goto restart_write; |
| return err; |
| } |
| |
| status_t Parcel::write(const void* data, size_t len) |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| void* const d = writeInplace(len); |
| if (d) { |
| memcpy(d, data, len); |
| return NO_ERROR; |
| } |
| return mError; |
| } |
| |
| void* Parcel::writeInplace(size_t len) |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return nullptr; |
| } |
| |
| const size_t padded = pad_size(len); |
| |
| // validate for integer overflow |
| if (mDataPos+padded < mDataPos) { |
| return nullptr; |
| } |
| |
| if ((mDataPos+padded) <= mDataCapacity) { |
| restart_write: |
| //printf("Writing %ld bytes, padded to %ld\n", len, padded); |
| uint8_t* const data = mData+mDataPos; |
| |
| // Need to pad at end? |
| if (padded != len) { |
| #if BYTE_ORDER == BIG_ENDIAN |
| static const uint32_t mask[4] = { |
| 0x00000000, 0xffffff00, 0xffff0000, 0xff000000 |
| }; |
| #endif |
| #if BYTE_ORDER == LITTLE_ENDIAN |
| static const uint32_t mask[4] = { |
| 0x00000000, 0x00ffffff, 0x0000ffff, 0x000000ff |
| }; |
| #endif |
| //printf("Applying pad mask: %p to %p\n", (void*)mask[padded-len], |
| // *reinterpret_cast<void**>(data+padded-4)); |
| *reinterpret_cast<uint32_t*>(data+padded-4) &= mask[padded-len]; |
| } |
| |
| finishWrite(padded); |
| return data; |
| } |
| |
| status_t err = growData(padded); |
| if (err == NO_ERROR) goto restart_write; |
| return nullptr; |
| } |
| |
| status_t Parcel::writeInt8(int8_t val) |
| { |
| return write(&val, sizeof(val)); |
| } |
| |
| status_t Parcel::writeUint8(uint8_t val) |
| { |
| return write(&val, sizeof(val)); |
| } |
| |
| status_t Parcel::writeInt16(int16_t val) |
| { |
| return write(&val, sizeof(val)); |
| } |
| |
| status_t Parcel::writeUint16(uint16_t val) |
| { |
| return write(&val, sizeof(val)); |
| } |
| |
| status_t Parcel::writeInt32(int32_t val) |
| { |
| return writeAligned(val); |
| } |
| |
| status_t Parcel::writeUint32(uint32_t val) |
| { |
| return writeAligned(val); |
| } |
| |
| status_t Parcel::writeBool(bool val) |
| { |
| return writeInt8(int8_t(val)); |
| } |
| status_t Parcel::writeInt64(int64_t val) |
| { |
| return writeAligned(val); |
| } |
| |
| status_t Parcel::writeUint64(uint64_t val) |
| { |
| return writeAligned(val); |
| } |
| |
| status_t Parcel::writePointer(uintptr_t val) |
| { |
| return writeAligned<binder_uintptr_t>(val); |
| } |
| |
| status_t Parcel::writeFloat(float val) |
| { |
| return writeAligned(val); |
| } |
| |
| #if defined(__mips__) && defined(__mips_hard_float) |
| |
| status_t Parcel::writeDouble(double val) |
| { |
| union { |
| double d; |
| unsigned long long ll; |
| } u; |
| u.d = val; |
| return writeAligned(u.ll); |
| } |
| |
| #else |
| |
| status_t Parcel::writeDouble(double val) |
| { |
| return writeAligned(val); |
| } |
| |
| #endif |
| |
| status_t Parcel::writeCString(const char* str) |
| { |
| return write(str, strlen(str)+1); |
| } |
| status_t Parcel::writeString16(const std::unique_ptr<String16>& str) |
| { |
| if (!str) { |
| return writeInt32(-1); |
| } |
| |
| return writeString16(*str); |
| } |
| |
| status_t Parcel::writeString16(const String16& str) |
| { |
| return writeString16(str.c_str(), str.size()); |
| } |
| |
| status_t Parcel::writeString16(const char16_t* str, size_t len) |
| { |
| if (str == nullptr) return writeInt32(-1); |
| |
| status_t err = writeInt32(len); |
| if (err == NO_ERROR) { |
| len *= sizeof(char16_t); |
| uint8_t* data = (uint8_t*)writeInplace(len+sizeof(char16_t)); |
| if (data) { |
| memcpy(data, str, len); |
| *reinterpret_cast<char16_t*>(data+len) = 0; |
| return NO_ERROR; |
| } |
| err = mError; |
| } |
| return err; |
| } |
| status_t Parcel::writeStrongBinder(const sp<IBinder>& val) |
| { |
| return flatten_binder(ProcessState::self(), val, this); |
| } |
| |
| template <typename T> |
| status_t Parcel::writeObject(const T& val) |
| { |
| const bool enoughData = (mDataPos+sizeof(val)) <= mDataCapacity; |
| const bool enoughObjects = mObjectsSize < mObjectsCapacity; |
| if (enoughData && enoughObjects) { |
| restart_write: |
| *reinterpret_cast<T*>(mData+mDataPos) = val; |
| |
| const binder_object_header* hdr = reinterpret_cast<binder_object_header*>(mData+mDataPos); |
| switch (hdr->type) { |
| case BINDER_TYPE_BINDER: |
| case BINDER_TYPE_WEAK_BINDER: |
| case BINDER_TYPE_HANDLE: |
| case BINDER_TYPE_WEAK_HANDLE: { |
| const flat_binder_object *fbo = reinterpret_cast<const flat_binder_object*>(hdr); |
| if (fbo->binder != 0) { |
| mObjects[mObjectsSize++] = mDataPos; |
| acquire_binder_object(ProcessState::self(), *fbo, this); |
| } |
| break; |
| } |
| case BINDER_TYPE_FD: { |
| // remember if it's a file descriptor |
| if (!mAllowFds) { |
| // fail before modifying our object index |
| return FDS_NOT_ALLOWED; |
| } |
| mHasFds = mFdsKnown = true; |
| mObjects[mObjectsSize++] = mDataPos; |
| break; |
| } |
| case BINDER_TYPE_FDA: |
| mObjects[mObjectsSize++] = mDataPos; |
| break; |
| case BINDER_TYPE_PTR: { |
| const binder_buffer_object *buffer_obj = reinterpret_cast< |
| const binder_buffer_object*>(hdr); |
| if ((void *)buffer_obj->buffer != nullptr) { |
| mObjects[mObjectsSize++] = mDataPos; |
| } |
| break; |
| } |
| default: { |
| ALOGE("writeObject: unknown type %d", hdr->type); |
| break; |
| } |
| } |
| return finishWrite(sizeof(val)); |
| } |
| |
| if (!enoughData) { |
| const status_t err = growData(sizeof(val)); |
| if (err != NO_ERROR) return err; |
| } |
| if (!enoughObjects) { |
| if (mObjectsSize > SIZE_MAX - 2) return NO_MEMORY; // overflow |
| if (mObjectsSize + 2 > SIZE_MAX / 3) return NO_MEMORY; // overflow |
| size_t newSize = ((mObjectsSize+2)*3)/2; |
| if (newSize > SIZE_MAX / sizeof(binder_size_t)) return NO_MEMORY; // overflow |
| binder_size_t* objects = (binder_size_t*)realloc(mObjects, newSize*sizeof(binder_size_t)); |
| if (objects == nullptr) return NO_MEMORY; |
| mObjects = objects; |
| mObjectsCapacity = newSize; |
| } |
| |
| goto restart_write; |
| } |
| |
| template status_t Parcel::writeObject<flat_binder_object>(const flat_binder_object& val); |
| template status_t Parcel::writeObject<binder_fd_object>(const binder_fd_object& val); |
| template status_t Parcel::writeObject<binder_buffer_object>(const binder_buffer_object& val); |
| template status_t Parcel::writeObject<binder_fd_array_object>(const binder_fd_array_object& val); |
| |
| bool Parcel::validateBufferChild(size_t child_buffer_handle, |
| size_t child_offset) const { |
| if (child_buffer_handle >= mObjectsSize) |
| return false; |
| binder_buffer_object *child = reinterpret_cast<binder_buffer_object*> |
| (mData + mObjects[child_buffer_handle]); |
| if (child->hdr.type != BINDER_TYPE_PTR || child_offset > child->length) { |
| // Parent object not a buffer, or not large enough |
| LOG_BUFFER("writeEmbeddedReference found weird child. " |
| "child_offset = %zu, child->length = %zu", |
| child_offset, (size_t)child->length); |
| return false; |
| } |
| return true; |
| } |
| |
| bool Parcel::validateBufferParent(size_t parent_buffer_handle, |
| size_t parent_offset) const { |
| if (parent_buffer_handle >= mObjectsSize) |
| return false; |
| binder_buffer_object *parent = reinterpret_cast<binder_buffer_object*> |
| (mData + mObjects[parent_buffer_handle]); |
| if (parent->hdr.type != BINDER_TYPE_PTR || |
| sizeof(binder_uintptr_t) > parent->length || |
| parent_offset > parent->length - sizeof(binder_uintptr_t)) { |
| // Parent object not a buffer, or not large enough |
| return false; |
| } |
| return true; |
| } |
| status_t Parcel::writeEmbeddedBuffer( |
| const void *buffer, size_t length, size_t *handle, |
| size_t parent_buffer_handle, size_t parent_offset) { |
| LOG_BUFFER("writeEmbeddedBuffer(%p, %zu, parent = (%zu, %zu)) -> %zu", |
| buffer, length, parent_buffer_handle, |
| parent_offset, mObjectsSize); |
| if(!validateBufferParent(parent_buffer_handle, parent_offset)) |
| return BAD_VALUE; |
| binder_buffer_object obj = { |
| .hdr = { .type = BINDER_TYPE_PTR }, |
| .flags = BINDER_BUFFER_FLAG_HAS_PARENT, |
| .buffer = reinterpret_cast<binder_uintptr_t>(buffer), |
| .length = length, |
| .parent = parent_buffer_handle, |
| .parent_offset = parent_offset, |
| }; |
| if (handle != nullptr) { |
| // We use an index into mObjects as a handle |
| *handle = mObjectsSize; |
| } |
| return writeObject(obj); |
| } |
| |
| status_t Parcel::writeBuffer(const void *buffer, size_t length, size_t *handle) |
| { |
| LOG_BUFFER("writeBuffer(%p, %zu) -> %zu", |
| buffer, length, mObjectsSize); |
| binder_buffer_object obj { |
| .hdr = { .type = BINDER_TYPE_PTR }, |
| .flags = 0, |
| .buffer = reinterpret_cast<binder_uintptr_t>(buffer), |
| .length = length, |
| }; |
| if (handle != nullptr) { |
| // We use an index into mObjects as a handle |
| *handle = mObjectsSize; |
| } |
| return writeObject(obj); |
| } |
| |
| void Parcel::clearCache() const { |
| LOG_BUFFER("clearing cache."); |
| mBufCachePos = 0; |
| mBufCache.clear(); |
| } |
| |
| void Parcel::updateCache() const { |
| if(mBufCachePos == mObjectsSize) |
| return; |
| LOG_BUFFER("updating cache from %zu to %zu", mBufCachePos, mObjectsSize); |
| for(size_t i = mBufCachePos; i < mObjectsSize; i++) { |
| binder_size_t dataPos = mObjects[i]; |
| binder_buffer_object *obj = |
| reinterpret_cast<binder_buffer_object*>(mData+dataPos); |
| if(obj->hdr.type != BINDER_TYPE_PTR) |
| continue; |
| BufferInfo ifo; |
| ifo.index = i; |
| ifo.buffer = obj->buffer; |
| ifo.bufend = obj->buffer + obj->length; |
| mBufCache.push_back(ifo); |
| } |
| mBufCachePos = mObjectsSize; |
| } |
| |
| /* O(n) (n=#buffers) to find a buffer that contains the given addr */ |
| status_t Parcel::findBuffer(const void *ptr, size_t length, bool *found, |
| size_t *handle, size_t *offset) const { |
| if(found == nullptr) |
| return UNKNOWN_ERROR; |
| updateCache(); |
| binder_uintptr_t ptrVal = reinterpret_cast<binder_uintptr_t>(ptr); |
| // true if the pointer is in some buffer, but the length is too big |
| // so that ptr + length doesn't fit into the buffer. |
| bool suspectRejectBadPointer = false; |
| LOG_BUFFER("findBuffer examining %zu objects.", mObjectsSize); |
| for(auto entry = mBufCache.rbegin(); entry != mBufCache.rend(); ++entry ) { |
| if(entry->buffer <= ptrVal && ptrVal < entry->bufend) { |
| // might have found it. |
| if(ptrVal + length <= entry->bufend) { |
| *found = true; |
| if(handle != nullptr) *handle = entry->index; |
| if(offset != nullptr) *offset = ptrVal - entry->buffer; |
| LOG_BUFFER(" findBuffer has a match at %zu!", entry->index); |
| return OK; |
| } else { |
| suspectRejectBadPointer = true; |
| } |
| } |
| } |
| LOG_BUFFER("findBuffer did not find for ptr = %p.", ptr); |
| *found = false; |
| return suspectRejectBadPointer ? BAD_VALUE : OK; |
| } |
| |
| /* findBuffer with the assumption that ptr = .buffer (so it points to top |
| * of the buffer, aka offset 0). |
| * */ |
| status_t Parcel::quickFindBuffer(const void *ptr, size_t *handle) const { |
| updateCache(); |
| binder_uintptr_t ptrVal = reinterpret_cast<binder_uintptr_t>(ptr); |
| LOG_BUFFER("quickFindBuffer examining %zu objects.", mObjectsSize); |
| for(auto entry = mBufCache.rbegin(); entry != mBufCache.rend(); ++entry ) { |
| if(entry->buffer == ptrVal) { |
| if(handle != nullptr) *handle = entry->index; |
| return OK; |
| } |
| } |
| LOG_BUFFER("quickFindBuffer did not find for ptr = %p.", ptr); |
| return NO_INIT; |
| } |
| |
| status_t Parcel::writeNativeHandleNoDup(const native_handle_t *handle, |
| bool embedded, |
| size_t parent_buffer_handle, |
| size_t parent_offset) |
| { |
| size_t buffer_handle; |
| status_t status = OK; |
| |
| if (handle == nullptr) { |
| status = writeUint64(0); |
| return status; |
| } |
| |
| size_t native_handle_size = sizeof(native_handle_t) |
| + handle->numFds * sizeof(int) + handle->numInts * sizeof(int); |
| writeUint64(native_handle_size); |
| |
| if (embedded) { |
| status = writeEmbeddedBuffer((void*) handle, |
| native_handle_size, &buffer_handle, |
| parent_buffer_handle, parent_offset); |
| } else { |
| status = writeBuffer((void*) handle, native_handle_size, &buffer_handle); |
| } |
| |
| if (status != OK) { |
| return status; |
| } |
| |
| struct binder_fd_array_object fd_array { |
| .hdr = { .type = BINDER_TYPE_FDA }, |
| .num_fds = static_cast<binder_size_t>(handle->numFds), |
| .parent = buffer_handle, |
| .parent_offset = offsetof(native_handle_t, data), |
| }; |
| |
| return writeObject(fd_array); |
| } |
| |
| status_t Parcel::writeNativeHandleNoDup(const native_handle_t *handle) |
| { |
| return writeNativeHandleNoDup(handle, false /* embedded */); |
| } |
| |
| status_t Parcel::writeEmbeddedNativeHandle(const native_handle_t *handle, |
| size_t parent_buffer_handle, |
| size_t parent_offset) |
| { |
| return writeNativeHandleNoDup(handle, true /* embedded */, |
| parent_buffer_handle, parent_offset); |
| } |
| |
| status_t Parcel::read(void* outData, size_t len) const |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| if ((mDataPos+pad_size(len)) >= mDataPos && (mDataPos+pad_size(len)) <= mDataSize |
| && len <= pad_size(len)) { |
| memcpy(outData, mData+mDataPos, len); |
| mDataPos += pad_size(len); |
| ALOGV("read Setting data pos of %p to %zu", this, mDataPos); |
| return NO_ERROR; |
| } |
| return NOT_ENOUGH_DATA; |
| } |
| |
| const void* Parcel::readInplace(size_t len) const |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return nullptr; |
| } |
| |
| if ((mDataPos+pad_size(len)) >= mDataPos && (mDataPos+pad_size(len)) <= mDataSize |
| && len <= pad_size(len)) { |
| const void* data = mData+mDataPos; |
| mDataPos += pad_size(len); |
| ALOGV("readInplace Setting data pos of %p to %zu", this, mDataPos); |
| return data; |
| } |
| return nullptr; |
| } |
| |
| template<class T> |
| status_t Parcel::readAligned(T *pArg) const { |
| static_assert(PAD_SIZE_UNSAFE(sizeof(T)) == sizeof(T)); |
| |
| if ((mDataPos+sizeof(T)) <= mDataSize) { |
| const void* data = mData+mDataPos; |
| mDataPos += sizeof(T); |
| *pArg = *reinterpret_cast<const T*>(data); |
| return NO_ERROR; |
| } else { |
| return NOT_ENOUGH_DATA; |
| } |
| } |
| |
| template<class T> |
| T Parcel::readAligned() const { |
| T result; |
| if (readAligned(&result) != NO_ERROR) { |
| result = 0; |
| } |
| |
| return result; |
| } |
| |
| template<class T> |
| status_t Parcel::writeAligned(T val) { |
| static_assert(PAD_SIZE_UNSAFE(sizeof(T)) == sizeof(T)); |
| |
| if ((mDataPos+sizeof(val)) <= mDataCapacity) { |
| restart_write: |
| *reinterpret_cast<T*>(mData+mDataPos) = val; |
| return finishWrite(sizeof(val)); |
| } |
| |
| status_t err = growData(sizeof(val)); |
| if (err == NO_ERROR) goto restart_write; |
| return err; |
| } |
| |
| status_t Parcel::readInt8(int8_t *pArg) const |
| { |
| return read(pArg, sizeof(*pArg)); |
| } |
| |
| status_t Parcel::readUint8(uint8_t *pArg) const |
| { |
| return read(pArg, sizeof(*pArg)); |
| } |
| |
| status_t Parcel::readInt16(int16_t *pArg) const |
| { |
| return read(pArg, sizeof(*pArg)); |
| } |
| |
| status_t Parcel::readUint16(uint16_t *pArg) const |
| { |
| return read(pArg, sizeof(*pArg)); |
| } |
| |
| status_t Parcel::readInt32(int32_t *pArg) const |
| { |
| return readAligned(pArg); |
| } |
| |
| int32_t Parcel::readInt32() const |
| { |
| return readAligned<int32_t>(); |
| } |
| |
| status_t Parcel::readUint32(uint32_t *pArg) const |
| { |
| return readAligned(pArg); |
| } |
| |
| uint32_t Parcel::readUint32() const |
| { |
| return readAligned<uint32_t>(); |
| } |
| |
| status_t Parcel::readInt64(int64_t *pArg) const |
| { |
| return readAligned(pArg); |
| } |
| |
| int64_t Parcel::readInt64() const |
| { |
| return readAligned<int64_t>(); |
| } |
| |
| status_t Parcel::readUint64(uint64_t *pArg) const |
| { |
| return readAligned(pArg); |
| } |
| |
| uint64_t Parcel::readUint64() const |
| { |
| return readAligned<uint64_t>(); |
| } |
| |
| status_t Parcel::readPointer(uintptr_t *pArg) const |
| { |
| status_t ret; |
| binder_uintptr_t ptr; |
| ret = readAligned(&ptr); |
| if (!ret) |
| *pArg = ptr; |
| return ret; |
| } |
| |
| uintptr_t Parcel::readPointer() const |
| { |
| return readAligned<binder_uintptr_t>(); |
| } |
| |
| |
| status_t Parcel::readFloat(float *pArg) const |
| { |
| return readAligned(pArg); |
| } |
| |
| |
| float Parcel::readFloat() const |
| { |
| return readAligned<float>(); |
| } |
| |
| #if defined(__mips__) && defined(__mips_hard_float) |
| |
| status_t Parcel::readDouble(double *pArg) const |
| { |
| union { |
| double d; |
| unsigned long long ll; |
| } u; |
| u.d = 0; |
| status_t status; |
| status = readAligned(&u.ll); |
| *pArg = u.d; |
| return status; |
| } |
| |
| double Parcel::readDouble() const |
| { |
| union { |
| double d; |
| unsigned long long ll; |
| } u; |
| u.ll = readAligned<unsigned long long>(); |
| return u.d; |
| } |
| |
| #else |
| |
| status_t Parcel::readDouble(double *pArg) const |
| { |
| return readAligned(pArg); |
| } |
| |
| double Parcel::readDouble() const |
| { |
| return readAligned<double>(); |
| } |
| |
| #endif |
| |
| status_t Parcel::readBool(bool *pArg) const |
| { |
| int8_t tmp; |
| status_t ret = readInt8(&tmp); |
| *pArg = (tmp != 0); |
| return ret; |
| } |
| |
| bool Parcel::readBool() const |
| { |
| int8_t tmp; |
| status_t err = readInt8(&tmp); |
| |
| if (err != OK) { |
| return 0; |
| } |
| |
| return tmp != 0; |
| } |
| |
| const char* Parcel::readCString() const |
| { |
| if (mDataPos < mDataSize) { |
| const size_t avail = mDataSize-mDataPos; |
| const char* str = reinterpret_cast<const char*>(mData+mDataPos); |
| // is the string's trailing NUL within the parcel's valid bounds? |
| const char* eos = reinterpret_cast<const char*>(memchr(str, 0, avail)); |
| if (eos) { |
| const size_t len = eos - str; |
| mDataPos += pad_size(len+1); |
| ALOGV("readCString Setting data pos of %p to %zu", this, mDataPos); |
| return str; |
| } |
| } |
| return nullptr; |
| } |
| String16 Parcel::readString16() const |
| { |
| size_t len; |
| const char16_t* str = readString16Inplace(&len); |
| if (str) return String16(str, len); |
| ALOGE("Reading a NULL string not supported here."); |
| return String16(); |
| } |
| |
| status_t Parcel::readString16(std::unique_ptr<String16>* pArg) const |
| { |
| const int32_t start = dataPosition(); |
| int32_t size; |
| status_t status = readInt32(&size); |
| pArg->reset(); |
| |
| if (status != OK || size < 0) { |
| return status; |
| } |
| |
| setDataPosition(start); |
| pArg->reset(new (std::nothrow) String16()); |
| |
| status = readString16(pArg->get()); |
| |
| if (status != OK) { |
| pArg->reset(); |
| } |
| |
| return status; |
| } |
| |
| status_t Parcel::readString16(String16* pArg) const |
| { |
| size_t len; |
| const char16_t* str = readString16Inplace(&len); |
| if (str) { |
| pArg->setTo(str, len); |
| return 0; |
| } else { |
| *pArg = String16(); |
| return UNEXPECTED_NULL; |
| } |
| } |
| |
| const char16_t* Parcel::readString16Inplace(size_t* outLen) const |
| { |
| int32_t size = readInt32(); |
| // watch for potential int overflow from size+1 |
| if (size >= 0 && size < INT32_MAX) { |
| *outLen = size; |
| const char16_t* str = (const char16_t*)readInplace((size+1)*sizeof(char16_t)); |
| if (str != nullptr) { |
| return str; |
| } |
| } |
| *outLen = 0; |
| return nullptr; |
| } |
| status_t Parcel::readStrongBinder(sp<IBinder>* val) const |
| { |
| status_t status = readNullableStrongBinder(val); |
| if (status == OK && !val->get()) { |
| status = UNEXPECTED_NULL; |
| } |
| return status; |
| } |
| |
| status_t Parcel::readNullableStrongBinder(sp<IBinder>* val) const |
| { |
| return unflatten_binder(ProcessState::self(), *this, val); |
| } |
| |
| sp<IBinder> Parcel::readStrongBinder() const |
| { |
| sp<IBinder> val; |
| // Note that a lot of code in Android reads binders by hand with this |
| // method, and that code has historically been ok with getting nullptr |
| // back (while ignoring error codes). |
| readNullableStrongBinder(&val); |
| return val; |
| } |
| |
| template<typename T> |
| const T* Parcel::readObject(size_t *objects_offset) const |
| { |
| const size_t DPOS = mDataPos; |
| if (objects_offset != nullptr) { |
| *objects_offset = 0; |
| } |
| |
| if ((DPOS+sizeof(T)) <= mDataSize) { |
| const T* obj = reinterpret_cast<const T*>(mData+DPOS); |
| mDataPos = DPOS + sizeof(T); |
| const binder_object_header *hdr = reinterpret_cast<const binder_object_header*>(obj); |
| switch (hdr->type) { |
| case BINDER_TYPE_BINDER: |
| case BINDER_TYPE_WEAK_BINDER: |
| case BINDER_TYPE_HANDLE: |
| case BINDER_TYPE_WEAK_HANDLE: { |
| const flat_binder_object *flat_obj = |
| reinterpret_cast<const flat_binder_object*>(hdr); |
| if (flat_obj->cookie == 0 && flat_obj->binder == 0) { |
| // When transferring a NULL binder object, we don't write it into |
| // the object list, so we don't want to check for it when |
| // reading. |
| ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); |
| return obj; |
| } |
| break; |
| } |
| case BINDER_TYPE_FD: |
| case BINDER_TYPE_FDA: |
| // fd (-arrays) must always appear in the meta-data list (eg touched by the kernel) |
| break; |
| case BINDER_TYPE_PTR: { |
| const binder_buffer_object *buffer_obj = |
| reinterpret_cast<const binder_buffer_object*>(hdr); |
| if ((void *)buffer_obj->buffer == nullptr) { |
| // null pointers can be returned directly - they're not written in the |
| // object list. All non-null buffers must appear in the objects list. |
| return obj; |
| } |
| break; |
| } |
| } |
| // Ensure that this object is valid... |
| binder_size_t* const OBJS = mObjects; |
| const size_t N = mObjectsSize; |
| size_t opos = mNextObjectHint; |
| |
| if (N > 0) { |
| ALOGV("Parcel %p looking for obj at %zu, hint=%zu", |
| this, DPOS, opos); |
| |
| // Start at the current hint position, looking for an object at |
| // the current data position. |
| if (opos < N) { |
| while (opos < (N-1) && OBJS[opos] < DPOS) { |
| opos++; |
| } |
| } else { |
| opos = N-1; |
| } |
| if (OBJS[opos] == DPOS) { |
| // Found it! |
| ALOGV("Parcel %p found obj %zu at index %zu with forward search", |
| this, DPOS, opos); |
| mNextObjectHint = opos+1; |
| ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); |
| if (objects_offset != nullptr) { |
| *objects_offset = opos; |
| } |
| return obj; |
| } |
| |
| // Look backwards for it... |
| while (opos > 0 && OBJS[opos] > DPOS) { |
| opos--; |
| } |
| if (OBJS[opos] == DPOS) { |
| // Found it! |
| ALOGV("Parcel %p found obj %zu at index %zu with backward search", |
| this, DPOS, opos); |
| mNextObjectHint = opos+1; |
| ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); |
| if (objects_offset != nullptr) { |
| *objects_offset = opos; |
| } |
| return obj; |
| } |
| } |
| ALOGW("Attempt to read object from Parcel %p at offset %zu that is not in the object list", |
| this, DPOS); |
| } |
| return nullptr; |
| } |
| |
| template const flat_binder_object* Parcel::readObject<flat_binder_object>(size_t *objects_offset) const; |
| |
| template const binder_fd_object* Parcel::readObject<binder_fd_object>(size_t *objects_offset) const; |
| |
| template const binder_buffer_object* Parcel::readObject<binder_buffer_object>(size_t *objects_offset) const; |
| |
| template const binder_fd_array_object* Parcel::readObject<binder_fd_array_object>(size_t *objects_offset) const; |
| |
| bool Parcel::verifyBufferObject(const binder_buffer_object *buffer_obj, |
| size_t size, uint32_t flags, size_t parent, |
| size_t parentOffset) const { |
| if (buffer_obj->length != size) { |
| ALOGE("Buffer length %" PRIu64 " does not match expected size %zu.", |
| static_cast<uint64_t>(buffer_obj->length), size); |
| return false; |
| } |
| |
| if (buffer_obj->flags != flags) { |
| ALOGE("Buffer flags 0x%02X do not match expected flags 0x%02X.", buffer_obj->flags, flags); |
| return false; |
| } |
| |
| if (flags & BINDER_BUFFER_FLAG_HAS_PARENT) { |
| if (buffer_obj->parent != parent) { |
| ALOGE("Buffer parent %" PRIu64 " does not match expected parent %zu.", |
| static_cast<uint64_t>(buffer_obj->parent), parent); |
| return false; |
| } |
| if (buffer_obj->parent_offset != parentOffset) { |
| ALOGE("Buffer parent offset %" PRIu64 " does not match expected offset %zu.", |
| static_cast<uint64_t>(buffer_obj->parent_offset), parentOffset); |
| return false; |
| } |
| |
| // checked by kernel driver, but needed for fuzzer |
| if (parent >= mObjectsSize) { |
| ALOGE("Parent index %zu but only have %zu objects", parent, mObjectsSize); |
| return false; |
| } |
| |
| binder_buffer_object *parentBuffer = |
| reinterpret_cast<binder_buffer_object*>(mData + mObjects[parent]); |
| void* bufferInParent = *reinterpret_cast<void**>( |
| reinterpret_cast<uint8_t*>(parentBuffer->buffer) + parentOffset); |
| void* childBuffer = reinterpret_cast<void*>(buffer_obj->buffer); |
| |
| if (bufferInParent != childBuffer) { |
| ALOGE("Buffer in parent %p differs from embedded buffer %p", |
| bufferInParent, childBuffer); |
| android_errorWriteLog(0x534e4554, "179289794"); |
| return false; |
| } |
| } |
| |
| return true; |
| } |
| |
| status_t Parcel::readBuffer(size_t buffer_size, size_t *buffer_handle, |
| uint32_t flags, size_t parent, size_t parentOffset, |
| const void **buffer_out) const { |
| |
| const binder_buffer_object* buffer_obj = readObject<binder_buffer_object>(buffer_handle); |
| |
| if (buffer_obj == nullptr || buffer_obj->hdr.type != BINDER_TYPE_PTR) { |
| return BAD_VALUE; |
| } |
| |
| if (!verifyBufferObject(buffer_obj, buffer_size, flags, parent, parentOffset)) { |
| return BAD_VALUE; |
| } |
| |
| // in read side, always use .buffer and .length. |
| *buffer_out = reinterpret_cast<void*>(buffer_obj->buffer); |
| |
| return OK; |
| } |
| |
| status_t Parcel::readNullableBuffer(size_t buffer_size, size_t *buffer_handle, |
| const void **buffer_out) const |
| { |
| return readBuffer(buffer_size, buffer_handle, |
| 0 /* flags */, 0 /* parent */, 0 /* parentOffset */, |
| buffer_out); |
| } |
| |
| status_t Parcel::readBuffer(size_t buffer_size, size_t *buffer_handle, |
| const void **buffer_out) const |
| { |
| status_t status = readNullableBuffer(buffer_size, buffer_handle, buffer_out); |
| if (status == OK && *buffer_out == nullptr) { |
| return UNEXPECTED_NULL; |
| } |
| return status; |
| } |
| |
| |
| status_t Parcel::readEmbeddedBuffer(size_t buffer_size, |
| size_t *buffer_handle, |
| size_t parent_buffer_handle, |
| size_t parent_offset, |
| const void **buffer_out) const |
| { |
| status_t status = readNullableEmbeddedBuffer(buffer_size, buffer_handle, |
| parent_buffer_handle, |
| parent_offset, buffer_out); |
| if (status == OK && *buffer_out == nullptr) { |
| return UNEXPECTED_NULL; |
| } |
| return status; |
| } |
| |
| status_t Parcel::readNullableEmbeddedBuffer(size_t buffer_size, |
| size_t *buffer_handle, |
| size_t parent_buffer_handle, |
| size_t parent_offset, |
| const void **buffer_out) const |
| { |
| return readBuffer(buffer_size, buffer_handle, BINDER_BUFFER_FLAG_HAS_PARENT, |
| parent_buffer_handle, parent_offset, buffer_out); |
| } |
| |
| status_t Parcel::readEmbeddedNativeHandle(size_t parent_buffer_handle, |
| size_t parent_offset, |
| const native_handle_t **handle) const |
| { |
| status_t status = readNullableEmbeddedNativeHandle(parent_buffer_handle, parent_offset, handle); |
| if (status == OK && *handle == nullptr) { |
| return UNEXPECTED_NULL; |
| } |
| return status; |
| } |
| |
| status_t Parcel::readNullableNativeHandleNoDup(const native_handle_t **handle, |
| bool embedded, |
| size_t parent_buffer_handle, |
| size_t parent_offset) const |
| { |
| uint64_t nativeHandleSize; |
| status_t status = readUint64(&nativeHandleSize); |
| if (status != OK) { |
| return BAD_VALUE; |
| } |
| |
| if (nativeHandleSize == 0) { |
| // If !embedded, then parent_* vars are 0 and don't actually correspond |
| // to anything. In that case, we're actually reading this data into |
| // writable memory, and the handle returned from here will actually be |
| // used (rather than be ignored). |
| if (embedded) { |
| if(!validateBufferParent(parent_buffer_handle, parent_offset)) { |
| ALOGE("Buffer in parent %zu offset %zu invalid.", parent_buffer_handle, parent_offset); |
| return BAD_VALUE; |
| } |
| |
| binder_buffer_object *parentBuffer = |
| reinterpret_cast<binder_buffer_object*>(mData + mObjects[parent_buffer_handle]); |
| |
| void* bufferInParent = *reinterpret_cast<void**>( |
| reinterpret_cast<uint8_t*>(parentBuffer->buffer) + parent_offset); |
| |
| if (bufferInParent != nullptr) { |
| ALOGE("Buffer in (handle) parent %p is not nullptr.", bufferInParent); |
| android_errorWriteLog(0x534e4554, "179289794"); |
| return BAD_VALUE; |
| } |
| } |
| |
| *handle = nullptr; |
| return status; |
| } |
| |
| if (nativeHandleSize < sizeof(native_handle_t) || nativeHandleSize > std::numeric_limits<uint32_t>::max()) { |
| ALOGE("Invalid native_handle_t size: %" PRIu64, nativeHandleSize); |
| return BAD_VALUE; |
| } |
| |
| size_t fdaParent; |
| if (embedded) { |
| status = readNullableEmbeddedBuffer(nativeHandleSize, &fdaParent, |
| parent_buffer_handle, parent_offset, |
| reinterpret_cast<const void**>(handle)); |
| } else { |
| status = readNullableBuffer(nativeHandleSize, &fdaParent, |
| reinterpret_cast<const void**>(handle)); |
| } |
| |
| if (status != OK) { |
| return status; |
| } |
| |
| if (*handle == nullptr) { |
| // null handle already read above |
| ALOGE("Expecting non-null handle buffer"); |
| return BAD_VALUE; |
| } |
| |
| int numFds = (*handle)->numFds; |
| int numInts = (*handle)->numInts; |
| |
| if (numFds < 0 || numFds > NATIVE_HANDLE_MAX_FDS) { |
| ALOGE("Received native_handle with invalid number of fds."); |
| return BAD_VALUE; |
| } |
| |
| if (numInts < 0 || numInts > NATIVE_HANDLE_MAX_INTS) { |
| ALOGE("Received native_handle with invalid number of ints."); |
| return BAD_VALUE; |
| } |
| |
| if (nativeHandleSize != (sizeof(native_handle_t) + ((numFds + numInts) * sizeof(int)))) { |
| ALOGE("Size of native_handle doesn't match."); |
| return BAD_VALUE; |
| } |
| |
| const binder_fd_array_object* fd_array_obj = readObject<binder_fd_array_object>(); |
| |
| if (fd_array_obj == nullptr || fd_array_obj->hdr.type != BINDER_TYPE_FDA) { |
| ALOGE("Can't find file-descriptor array object."); |
| return BAD_VALUE; |
| } |
| |
| if (static_cast<int>(fd_array_obj->num_fds) != numFds) { |
| ALOGE("Number of native handles does not match."); |
| return BAD_VALUE; |
| } |
| |
| if (fd_array_obj->parent != fdaParent) { |
| ALOGE("Parent handle of file-descriptor array not correct."); |
| return BAD_VALUE; |
| } |
| |
| if (fd_array_obj->parent_offset != offsetof(native_handle_t, data)) { |
| ALOGE("FD array object not properly offset in parent."); |
| return BAD_VALUE; |
| } |
| |
| return OK; |
| } |
| |
| status_t Parcel::readNullableEmbeddedNativeHandle(size_t parent_buffer_handle, |
| size_t parent_offset, |
| const native_handle_t **handle) const |
| { |
| return readNullableNativeHandleNoDup(handle, true /* embedded */, parent_buffer_handle, |
| parent_offset); |
| } |
| |
| status_t Parcel::readNativeHandleNoDup(const native_handle_t **handle) const |
| { |
| status_t status = readNullableNativeHandleNoDup(handle); |
| if (status == OK && *handle == nullptr) { |
| return UNEXPECTED_NULL; |
| } |
| return status; |
| } |
| |
| status_t Parcel::readNullableNativeHandleNoDup(const native_handle_t **handle) const |
| { |
| return readNullableNativeHandleNoDup(handle, false /* embedded */); |
| } |
| |
| void Parcel::closeFileDescriptors() |
| { |
| size_t i = mObjectsSize; |
| if (i > 0) { |
| //ALOGI("Closing file descriptors for %zu objects...", i); |
| } |
| while (i > 0) { |
| i--; |
| const flat_binder_object* flat |
| = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]); |
| if (flat->hdr.type == BINDER_TYPE_FD) { |
| //ALOGI("Closing fd: %ld", flat->handle); |
| close(flat->handle); |
| } |
| } |
| } |
| |
| uintptr_t Parcel::ipcData() const |
| { |
| return reinterpret_cast<uintptr_t>(mData); |
| } |
| |
| size_t Parcel::ipcDataSize() const |
| { |
| return mDataSize > mDataPos ? mDataSize : mDataPos; |
| } |
| |
| uintptr_t Parcel::ipcObjects() const |
| { |
| return reinterpret_cast<uintptr_t>(mObjects); |
| } |
| |
| size_t Parcel::ipcObjectsCount() const |
| { |
| return mObjectsSize; |
| } |
| |
| #define BUFFER_ALIGNMENT_BYTES 8 |
| size_t Parcel::ipcBufferSize() const |
| { |
| size_t totalBuffersSize = 0; |
| // Add size for BINDER_TYPE_PTR |
| size_t i = mObjectsSize; |
| while (i > 0) { |
| i--; |
| const binder_buffer_object* buffer |
| = reinterpret_cast<binder_buffer_object*>(mData+mObjects[i]); |
| if (buffer->hdr.type == BINDER_TYPE_PTR) { |
| /* The binder kernel driver requires each buffer to be 8-byte |
| * aligned */ |
| size_t alignedSize = (buffer->length + (BUFFER_ALIGNMENT_BYTES - 1)) |
| & ~(BUFFER_ALIGNMENT_BYTES - 1); |
| if (alignedSize > SIZE_MAX - totalBuffersSize) { |
| ALOGE("ipcBuffersSize(): invalid buffer sizes."); |
| return 0; |
| } |
| totalBuffersSize += alignedSize; |
| } |
| } |
| return totalBuffersSize; |
| } |
| |
| void Parcel::ipcSetDataReference(const uint8_t* data, size_t dataSize, |
| const binder_size_t* objects, size_t objectsCount, release_func relFunc, void* relCookie) |
| { |
| binder_size_t minOffset = 0; |
| freeDataNoInit(); |
| mError = NO_ERROR; |
| mData = const_cast<uint8_t*>(data); |
| mDataSize = mDataCapacity = dataSize; |
| //ALOGI("setDataReference Setting data size of %p to %lu (pid=%d)", this, mDataSize, getpid()); |
| mDataPos = 0; |
| ALOGV("setDataReference Setting data pos of %p to %zu", this, mDataPos); |
| mObjects = const_cast<binder_size_t*>(objects); |
| mObjectsSize = mObjectsCapacity = objectsCount; |
| mNextObjectHint = 0; |
| clearCache(); |
| mOwner = relFunc; |
| mOwnerCookie = relCookie; |
| for (size_t i = 0; i < mObjectsSize; i++) { |
| binder_size_t offset = mObjects[i]; |
| if (offset < minOffset) { |
| ALOGE("%s: bad object offset %" PRIu64 " < %" PRIu64 "\n", |
| __func__, (uint64_t)offset, (uint64_t)minOffset); |
| mObjectsSize = 0; |
| break; |
| } |
| minOffset = offset + sizeof(flat_binder_object); |
| } |
| scanForFds(); |
| } |
| |
| void Parcel::print(TextOutput& to, uint32_t /*flags*/) const |
| { |
| to << "Parcel("; |
| |
| if (errorCheck() != NO_ERROR) { |
| const status_t err = errorCheck(); |
| to << "Error: " << (void*)(intptr_t)err << " \"" << strerror(-err) << "\""; |
| } else if (dataSize() > 0) { |
| const uint8_t* DATA = data(); |
| to << indent << HexDump(DATA, dataSize()) << dedent; |
| const binder_size_t* OBJS = objects(); |
| const size_t N = objectsCount(); |
| for (size_t i=0; i<N; i++) { |
| const flat_binder_object* flat |
| = reinterpret_cast<const flat_binder_object*>(DATA+OBJS[i]); |
| if (flat->hdr.type == BINDER_TYPE_PTR) { |
| const binder_buffer_object* buffer |
| = reinterpret_cast<const binder_buffer_object*>(DATA+OBJS[i]); |
| HexDump bufferDump((const uint8_t*)buffer->buffer, (size_t)buffer->length); |
| bufferDump.setSingleLineCutoff(0); |
| to << endl << "Object #" << i << " @ " << (void*)OBJS[i] << " (buffer size " << buffer->length << "):"; |
| to << indent << bufferDump << dedent; |
| } else { |
| to << endl << "Object #" << i << " @ " << (void*)OBJS[i] << ": " |
| << TypeCode(flat->hdr.type & 0x7f7f7f00) |
| << " = " << flat->binder; |
| } |
| } |
| } else { |
| to << "NULL"; |
| } |
| |
| to << ")"; |
| } |
| |
| void Parcel::releaseObjects() |
| { |
| const sp<ProcessState> proc(ProcessState::self()); |
| size_t i = mObjectsSize; |
| uint8_t* const data = mData; |
| binder_size_t* const objects = mObjects; |
| while (i > 0) { |
| i--; |
| const flat_binder_object* flat |
| = reinterpret_cast<flat_binder_object*>(data+objects[i]); |
| release_object(proc, *flat, this); |
| } |
| } |
| |
| void Parcel::acquireObjects() |
| { |
| const sp<ProcessState> proc(ProcessState::self()); |
| size_t i = mObjectsSize; |
| uint8_t* const data = mData; |
| binder_size_t* const objects = mObjects; |
| while (i > 0) { |
| i--; |
| const binder_object_header* flat |
| = reinterpret_cast<binder_object_header*>(data+objects[i]); |
| acquire_object(proc, *flat, this); |
| } |
| } |
| |
| void Parcel::freeData() |
| { |
| freeDataNoInit(); |
| initState(); |
| } |
| |
| void Parcel::freeDataNoInit() |
| { |
| if (mOwner) { |
| LOG_ALLOC("Parcel %p: freeing other owner data", this); |
| //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid()); |
| mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie); |
| } else { |
| LOG_ALLOC("Parcel %p: freeing allocated data", this); |
| releaseObjects(); |
| if (mData) { |
| LOG_ALLOC("Parcel %p: freeing with %zu capacity", this, mDataCapacity); |
| gParcelGlobalAllocSize -= mDataCapacity; |
| gParcelGlobalAllocCount--; |
| if (mDeallocZero) { |
| zeroMemory(mData, mDataSize); |
| } |
| free(mData); |
| } |
| if (mObjects) free(mObjects); |
| } |
| } |
| |
| status_t Parcel::growData(size_t len) |
| { |
| if (len > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| if (len > SIZE_MAX - mDataSize) return NO_MEMORY; // overflow |
| if (mDataSize + len > SIZE_MAX / 3) return NO_MEMORY; // overflow |
| size_t newSize = ((mDataSize+len)*3)/2; |
| return continueWrite(newSize); |
| } |
| |
| static uint8_t* reallocZeroFree(uint8_t* data, size_t oldCapacity, size_t newCapacity, bool zero) { |
| if (!zero) { |
| return (uint8_t*)realloc(data, newCapacity); |
| } |
| uint8_t* newData = (uint8_t*)malloc(newCapacity); |
| if (!newData) { |
| return nullptr; |
| } |
| |
| memcpy(newData, data, std::min(oldCapacity, newCapacity)); |
| zeroMemory(data, oldCapacity); |
| free(data); |
| return newData; |
| } |
| |
| status_t Parcel::restartWrite(size_t desired) |
| { |
| if (desired > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| if (mOwner) { |
| freeData(); |
| return continueWrite(desired); |
| } |
| |
| uint8_t* data = reallocZeroFree(mData, mDataCapacity, desired, mDeallocZero); |
| if (!data && desired > mDataCapacity) { |
| mError = NO_MEMORY; |
| return NO_MEMORY; |
| } |
| |
| releaseObjects(); |
| |
| if (data || desired == 0) { |
| LOG_ALLOC("Parcel %p: restart from %zu to %zu capacity", this, mDataCapacity, desired); |
| if (mDataCapacity > desired) { |
| gParcelGlobalAllocSize -= (mDataCapacity - desired); |
| } else { |
| gParcelGlobalAllocSize += (desired - mDataCapacity); |
| } |
| |
| if (!mData) { |
| gParcelGlobalAllocCount++; |
| } |
| mData = data; |
| mDataCapacity = desired; |
| } |
| |
| mDataSize = mDataPos = 0; |
| ALOGV("restartWrite Setting data size of %p to %zu", this, mDataSize); |
| ALOGV("restartWrite Setting data pos of %p to %zu", this, mDataPos); |
| |
| free(mObjects); |
| mObjects = nullptr; |
| mObjectsSize = mObjectsCapacity = 0; |
| mNextObjectHint = 0; |
| mHasFds = false; |
| clearCache(); |
| mFdsKnown = true; |
| mAllowFds = true; |
| |
| return NO_ERROR; |
| } |
| |
| status_t Parcel::continueWrite(size_t desired) |
| { |
| if (desired > INT32_MAX) { |
| // don't accept size_t values which may have come from an |
| // inadvertent conversion from a negative int. |
| return BAD_VALUE; |
| } |
| |
| // If shrinking, first adjust for any objects that appear |
| // after the new data size. |
| size_t objectsSize = mObjectsSize; |
| if (desired < mDataSize) { |
| if (desired == 0) { |
| objectsSize = 0; |
| } else { |
| while (objectsSize > 0) { |
| if (mObjects[objectsSize-1] < desired) |
| break; |
| objectsSize--; |
| } |
| } |
| } |
| |
| if (mOwner) { |
| // If the size is going to zero, just release the owner's data. |
| if (desired == 0) { |
| freeData(); |
| return NO_ERROR; |
| } |
| |
| // If there is a different owner, we need to take |
| // posession. |
| uint8_t* data = (uint8_t*)malloc(desired); |
| if (!data) { |
| mError = NO_MEMORY; |
| return NO_MEMORY; |
| } |
| binder_size_t* objects = nullptr; |
| |
| if (objectsSize) { |
| objects = (binder_size_t*)calloc(objectsSize, sizeof(binder_size_t)); |
| if (!objects) { |
| free(data); |
| |
| mError = NO_MEMORY; |
| return NO_MEMORY; |
| } |
| |
| // Little hack to only acquire references on objects |
| // we will be keeping. |
| size_t oldObjectsSize = mObjectsSize; |
| mObjectsSize = objectsSize; |
| acquireObjects(); |
| mObjectsSize = oldObjectsSize; |
| } |
| |
| if (mData) { |
| memcpy(data, mData, mDataSize < desired ? mDataSize : desired); |
| } |
| if (objects && mObjects) { |
| memcpy(objects, mObjects, objectsSize*sizeof(binder_size_t)); |
| } |
| //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid()); |
| mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie); |
| mOwner = nullptr; |
| |
| LOG_ALLOC("Parcel %p: taking ownership of %zu capacity", this, desired); |
| gParcelGlobalAllocSize += desired; |
| gParcelGlobalAllocCount++; |
| |
| mData = data; |
| mObjects = objects; |
| mDataSize = (mDataSize < desired) ? mDataSize : desired; |
| ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); |
| mDataCapacity = desired; |
| mObjectsSize = mObjectsCapacity = objectsSize; |
| mNextObjectHint = 0; |
| |
| clearCache(); |
| } else if (mData) { |
| if (objectsSize < mObjectsSize) { |
| // Need to release refs on any objects we are dropping. |
| const sp<ProcessState> proc(ProcessState::self()); |
| for (size_t i=objectsSize; i<mObjectsSize; i++) { |
| const flat_binder_object* flat |
| = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]); |
| if (flat->hdr.type == BINDER_TYPE_FD) { |
| // will need to rescan because we may have lopped off the only FDs |
| mFdsKnown = false; |
| } |
| release_object(proc, *flat, this); |
| } |
| |
| if (objectsSize == 0) { |
| free(mObjects); |
| mObjects = nullptr; |
| } else { |
| binder_size_t* objects = |
| (binder_size_t*)realloc(mObjects, objectsSize*sizeof(binder_size_t)); |
| if (objects) { |
| mObjects = objects; |
| } |
| } |
| mObjectsSize = objectsSize; |
| mNextObjectHint = 0; |
| |
| clearCache(); |
| } |
| |
| // We own the data, so we can just do a realloc(). |
| if (desired > mDataCapacity) { |
| uint8_t* data = reallocZeroFree(mData, mDataCapacity, desired, mDeallocZero); |
| if (data) { |
| LOG_ALLOC("Parcel %p: continue from %zu to %zu capacity", this, mDataCapacity, |
| desired); |
| gParcelGlobalAllocSize += desired; |
| gParcelGlobalAllocSize -= mDataCapacity; |
| mData = data; |
| mDataCapacity = desired; |
| } else { |
| mError = NO_MEMORY; |
| return NO_MEMORY; |
| } |
| } else { |
| if (mDataSize > desired) { |
| mDataSize = desired; |
| ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); |
| } |
| if (mDataPos > desired) { |
| mDataPos = desired; |
| ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos); |
| } |
| } |
| |
| } else { |
| // This is the first data. Easy! |
| uint8_t* data = (uint8_t*)malloc(desired); |
| if (!data) { |
| mError = NO_MEMORY; |
| return NO_MEMORY; |
| } |
| |
| if(!(mDataCapacity == 0 && mObjects == nullptr |
| && mObjectsCapacity == 0)) { |
| ALOGE("continueWrite: %zu/%p/%zu/%zu", mDataCapacity, mObjects, mObjectsCapacity, desired); |
| } |
| |
| LOG_ALLOC("Parcel %p: allocating with %zu capacity", this, desired); |
| gParcelGlobalAllocSize += desired; |
| gParcelGlobalAllocCount++; |
| |
| mData = data; |
| mDataSize = mDataPos = 0; |
| ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); |
| ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos); |
| mDataCapacity = desired; |
| } |
| |
| return NO_ERROR; |
| } |
| |
| void Parcel::initState() |
| { |
| LOG_ALLOC("Parcel %p: initState", this); |
| mError = NO_ERROR; |
| mData = nullptr; |
| mDataSize = 0; |
| mDataCapacity = 0; |
| mDataPos = 0; |
| ALOGV("initState Setting data size of %p to %zu", this, mDataSize); |
| ALOGV("initState Setting data pos of %p to %zu", this, mDataPos); |
| mObjects = nullptr; |
| mObjectsSize = 0; |
| mObjectsCapacity = 0; |
| mNextObjectHint = 0; |
| mHasFds = false; |
| mFdsKnown = true; |
| mAllowFds = true; |
| mDeallocZero = false; |
| mOwner = nullptr; |
| clearCache(); |
| |
| // racing multiple init leads only to multiple identical write |
| if (gMaxFds == 0) { |
| struct rlimit result; |
| if (!getrlimit(RLIMIT_NOFILE, &result)) { |
| gMaxFds = (size_t)result.rlim_cur; |
| //ALOGI("parcel fd limit set to %zu", gMaxFds); |
| } else { |
| ALOGW("Unable to getrlimit: %s", strerror(errno)); |
| gMaxFds = 1024; |
| } |
| } |
| } |
| |
| void Parcel::scanForFds() const |
| { |
| bool hasFds = false; |
| for (size_t i=0; i<mObjectsSize; i++) { |
| const flat_binder_object* flat |
| = reinterpret_cast<const flat_binder_object*>(mData + mObjects[i]); |
| if (flat->hdr.type == BINDER_TYPE_FD) { |
| hasFds = true; |
| break; |
| } |
| } |
| mHasFds = hasFds; |
| mFdsKnown = true; |
| } |
| |
| } // namespace hardware |
| } // namespace android |