1 /* 2 * Copyright (C) 2005 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #define LOG_TAG "Parcel" 18 //#define LOG_NDEBUG 0 19 20 #include <binder/Parcel.h> 21 22 #include <binder/IPCThreadState.h> 23 #include <binder/Binder.h> 24 #include <binder/BpBinder.h> 25 #include <binder/ProcessState.h> 26 #include <binder/TextOutput.h> 27 28 #include <errno.h> 29 #include <utils/Debug.h> 30 #include <utils/Log.h> 31 #include <utils/String8.h> 32 #include <utils/String16.h> 33 #include <utils/misc.h> 34 #include <utils/Flattenable.h> 35 #include <cutils/ashmem.h> 36 37 #include <private/binder/binder_module.h> 38 #include <private/binder/Static.h> 39 40 #include <inttypes.h> 41 #include <stdio.h> 42 #include <stdlib.h> 43 #include <stdint.h> 44 #include <sys/mman.h> 45 46 #ifndef INT32_MAX 47 #define INT32_MAX ((int32_t)(2147483647)) 48 #endif 49 50 #define LOG_REFS(...) 51 //#define LOG_REFS(...) ALOG(LOG_DEBUG, "Parcel", __VA_ARGS__) 52 #define LOG_ALLOC(...) 53 //#define LOG_ALLOC(...) ALOG(LOG_DEBUG, "Parcel", __VA_ARGS__) 54 55 // --------------------------------------------------------------------------- 56 57 #define PAD_SIZE(s) (((s)+3)&~3) 58 59 // Note: must be kept in sync with android/os/StrictMode.java's PENALTY_GATHER 60 #define STRICT_MODE_PENALTY_GATHER 0x100 61 62 // Note: must be kept in sync with android/os/Parcel.java's EX_HAS_REPLY_HEADER 63 #define EX_HAS_REPLY_HEADER -128 64 65 // Maximum size of a blob to transfer in-place. 66 static const size_t IN_PLACE_BLOB_LIMIT = 40 * 1024; 67 68 // XXX This can be made public if we want to provide 69 // support for typed data. 70 struct small_flat_data 71 { 72 uint32_t type; 73 uint32_t data; 74 }; 75 76 namespace android { 77 78 static pthread_mutex_t gParcelGlobalAllocSizeLock = PTHREAD_MUTEX_INITIALIZER; 79 static size_t gParcelGlobalAllocSize = 0; 80 static size_t gParcelGlobalAllocCount = 0; 81 82 void acquire_object(const sp<ProcessState>& proc, 83 const flat_binder_object& obj, const void* who) 84 { 85 switch (obj.type) { 86 case BINDER_TYPE_BINDER: 87 if (obj.binder) { 88 LOG_REFS("Parcel %p acquiring reference on local %p", who, obj.cookie); 89 reinterpret_cast<IBinder*>(obj.cookie)->incStrong(who); 90 } 91 return; 92 case BINDER_TYPE_WEAK_BINDER: 93 if (obj.binder) 94 reinterpret_cast<RefBase::weakref_type*>(obj.binder)->incWeak(who); 95 return; 96 case BINDER_TYPE_HANDLE: { 97 const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle); 98 if (b != NULL) { 99 LOG_REFS("Parcel %p acquiring reference on remote %p", who, b.get()); 100 b->incStrong(who); 101 } 102 return; 103 } 104 case BINDER_TYPE_WEAK_HANDLE: { 105 const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle); 106 if (b != NULL) b.get_refs()->incWeak(who); 107 return; 108 } 109 case BINDER_TYPE_FD: { 110 // intentionally blank -- nothing to do to acquire this, but we do 111 // recognize it as a legitimate object type. 112 return; 113 } 114 } 115 116 ALOGD("Invalid object type 0x%08x", obj.type); 117 } 118 119 void release_object(const sp<ProcessState>& proc, 120 const flat_binder_object& obj, const void* who) 121 { 122 switch (obj.type) { 123 case BINDER_TYPE_BINDER: 124 if (obj.binder) { 125 LOG_REFS("Parcel %p releasing reference on local %p", who, obj.cookie); 126 reinterpret_cast<IBinder*>(obj.cookie)->decStrong(who); 127 } 128 return; 129 case BINDER_TYPE_WEAK_BINDER: 130 if (obj.binder) 131 reinterpret_cast<RefBase::weakref_type*>(obj.binder)->decWeak(who); 132 return; 133 case BINDER_TYPE_HANDLE: { 134 const sp<IBinder> b = proc->getStrongProxyForHandle(obj.handle); 135 if (b != NULL) { 136 LOG_REFS("Parcel %p releasing reference on remote %p", who, b.get()); 137 b->decStrong(who); 138 } 139 return; 140 } 141 case BINDER_TYPE_WEAK_HANDLE: { 142 const wp<IBinder> b = proc->getWeakProxyForHandle(obj.handle); 143 if (b != NULL) b.get_refs()->decWeak(who); 144 return; 145 } 146 case BINDER_TYPE_FD: { 147 if (obj.cookie != 0) close(obj.handle); 148 return; 149 } 150 } 151 152 ALOGE("Invalid object type 0x%08x", obj.type); 153 } 154 155 inline static status_t finish_flatten_binder( 156 const sp<IBinder>& /*binder*/, const flat_binder_object& flat, Parcel* out) 157 { 158 return out->writeObject(flat, false); 159 } 160 161 status_t flatten_binder(const sp<ProcessState>& /*proc*/, 162 const sp<IBinder>& binder, Parcel* out) 163 { 164 flat_binder_object obj; 165 166 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS; 167 if (binder != NULL) { 168 IBinder *local = binder->localBinder(); 169 if (!local) { 170 BpBinder *proxy = binder->remoteBinder(); 171 if (proxy == NULL) { 172 ALOGE("null proxy"); 173 } 174 const int32_t handle = proxy ? proxy->handle() : 0; 175 obj.type = BINDER_TYPE_HANDLE; 176 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */ 177 obj.handle = handle; 178 obj.cookie = 0; 179 } else { 180 obj.type = BINDER_TYPE_BINDER; 181 obj.binder = reinterpret_cast<uintptr_t>(local->getWeakRefs()); 182 obj.cookie = reinterpret_cast<uintptr_t>(local); 183 } 184 } else { 185 obj.type = BINDER_TYPE_BINDER; 186 obj.binder = 0; 187 obj.cookie = 0; 188 } 189 190 return finish_flatten_binder(binder, obj, out); 191 } 192 193 status_t flatten_binder(const sp<ProcessState>& /*proc*/, 194 const wp<IBinder>& binder, Parcel* out) 195 { 196 flat_binder_object obj; 197 198 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS; 199 if (binder != NULL) { 200 sp<IBinder> real = binder.promote(); 201 if (real != NULL) { 202 IBinder *local = real->localBinder(); 203 if (!local) { 204 BpBinder *proxy = real->remoteBinder(); 205 if (proxy == NULL) { 206 ALOGE("null proxy"); 207 } 208 const int32_t handle = proxy ? proxy->handle() : 0; 209 obj.type = BINDER_TYPE_WEAK_HANDLE; 210 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */ 211 obj.handle = handle; 212 obj.cookie = 0; 213 } else { 214 obj.type = BINDER_TYPE_WEAK_BINDER; 215 obj.binder = reinterpret_cast<uintptr_t>(binder.get_refs()); 216 obj.cookie = reinterpret_cast<uintptr_t>(binder.unsafe_get()); 217 } 218 return finish_flatten_binder(real, obj, out); 219 } 220 221 // XXX How to deal? In order to flatten the given binder, 222 // we need to probe it for information, which requires a primary 223 // reference... but we don't have one. 224 // 225 // The OpenBinder implementation uses a dynamic_cast<> here, 226 // but we can't do that with the different reference counting 227 // implementation we are using. 228 ALOGE("Unable to unflatten Binder weak reference!"); 229 obj.type = BINDER_TYPE_BINDER; 230 obj.binder = 0; 231 obj.cookie = 0; 232 return finish_flatten_binder(NULL, obj, out); 233 234 } else { 235 obj.type = BINDER_TYPE_BINDER; 236 obj.binder = 0; 237 obj.cookie = 0; 238 return finish_flatten_binder(NULL, obj, out); 239 } 240 } 241 242 inline static status_t finish_unflatten_binder( 243 BpBinder* /*proxy*/, const flat_binder_object& /*flat*/, 244 const Parcel& /*in*/) 245 { 246 return NO_ERROR; 247 } 248 249 status_t unflatten_binder(const sp<ProcessState>& proc, 250 const Parcel& in, sp<IBinder>* out) 251 { 252 const flat_binder_object* flat = in.readObject(false); 253 254 if (flat) { 255 switch (flat->type) { 256 case BINDER_TYPE_BINDER: 257 *out = reinterpret_cast<IBinder*>(flat->cookie); 258 return finish_unflatten_binder(NULL, *flat, in); 259 case BINDER_TYPE_HANDLE: 260 *out = proc->getStrongProxyForHandle(flat->handle); 261 return finish_unflatten_binder( 262 static_cast<BpBinder*>(out->get()), *flat, in); 263 } 264 } 265 return BAD_TYPE; 266 } 267 268 status_t unflatten_binder(const sp<ProcessState>& proc, 269 const Parcel& in, wp<IBinder>* out) 270 { 271 const flat_binder_object* flat = in.readObject(false); 272 273 if (flat) { 274 switch (flat->type) { 275 case BINDER_TYPE_BINDER: 276 *out = reinterpret_cast<IBinder*>(flat->cookie); 277 return finish_unflatten_binder(NULL, *flat, in); 278 case BINDER_TYPE_WEAK_BINDER: 279 if (flat->binder != 0) { 280 out->set_object_and_refs( 281 reinterpret_cast<IBinder*>(flat->cookie), 282 reinterpret_cast<RefBase::weakref_type*>(flat->binder)); 283 } else { 284 *out = NULL; 285 } 286 return finish_unflatten_binder(NULL, *flat, in); 287 case BINDER_TYPE_HANDLE: 288 case BINDER_TYPE_WEAK_HANDLE: 289 *out = proc->getWeakProxyForHandle(flat->handle); 290 return finish_unflatten_binder( 291 static_cast<BpBinder*>(out->unsafe_get()), *flat, in); 292 } 293 } 294 return BAD_TYPE; 295 } 296 297 // --------------------------------------------------------------------------- 298 299 Parcel::Parcel() 300 { 301 LOG_ALLOC("Parcel %p: constructing", this); 302 initState(); 303 } 304 305 Parcel::~Parcel() 306 { 307 freeDataNoInit(); 308 LOG_ALLOC("Parcel %p: destroyed", this); 309 } 310 311 size_t Parcel::getGlobalAllocSize() { 312 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 313 size_t size = gParcelGlobalAllocSize; 314 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 315 return size; 316 } 317 318 size_t Parcel::getGlobalAllocCount() { 319 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 320 size_t count = gParcelGlobalAllocCount; 321 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 322 return count; 323 } 324 325 const uint8_t* Parcel::data() const 326 { 327 return mData; 328 } 329 330 size_t Parcel::dataSize() const 331 { 332 return (mDataSize > mDataPos ? mDataSize : mDataPos); 333 } 334 335 size_t Parcel::dataAvail() const 336 { 337 // TODO: decide what to do about the possibility that this can 338 // report an available-data size that exceeds a Java int's max 339 // positive value, causing havoc. Fortunately this will only 340 // happen if someone constructs a Parcel containing more than two 341 // gigabytes of data, which on typical phone hardware is simply 342 // not possible. 343 return dataSize() - dataPosition(); 344 } 345 346 size_t Parcel::dataPosition() const 347 { 348 return mDataPos; 349 } 350 351 size_t Parcel::dataCapacity() const 352 { 353 return mDataCapacity; 354 } 355 356 status_t Parcel::setDataSize(size_t size) 357 { 358 status_t err; 359 err = continueWrite(size); 360 if (err == NO_ERROR) { 361 mDataSize = size; 362 ALOGV("setDataSize Setting data size of %p to %zu", this, mDataSize); 363 } 364 return err; 365 } 366 367 void Parcel::setDataPosition(size_t pos) const 368 { 369 mDataPos = pos; 370 mNextObjectHint = 0; 371 } 372 373 status_t Parcel::setDataCapacity(size_t size) 374 { 375 if (size > mDataCapacity) return continueWrite(size); 376 return NO_ERROR; 377 } 378 379 status_t Parcel::setData(const uint8_t* buffer, size_t len) 380 { 381 status_t err = restartWrite(len); 382 if (err == NO_ERROR) { 383 memcpy(const_cast<uint8_t*>(data()), buffer, len); 384 mDataSize = len; 385 mFdsKnown = false; 386 } 387 return err; 388 } 389 390 status_t Parcel::appendFrom(const Parcel *parcel, size_t offset, size_t len) 391 { 392 const sp<ProcessState> proc(ProcessState::self()); 393 status_t err; 394 const uint8_t *data = parcel->mData; 395 const binder_size_t *objects = parcel->mObjects; 396 size_t size = parcel->mObjectsSize; 397 int startPos = mDataPos; 398 int firstIndex = -1, lastIndex = -2; 399 400 if (len == 0) { 401 return NO_ERROR; 402 } 403 404 // range checks against the source parcel size 405 if ((offset > parcel->mDataSize) 406 || (len > parcel->mDataSize) 407 || (offset + len > parcel->mDataSize)) { 408 return BAD_VALUE; 409 } 410 411 // Count objects in range 412 for (int i = 0; i < (int) size; i++) { 413 size_t off = objects[i]; 414 if ((off >= offset) && (off < offset + len)) { 415 if (firstIndex == -1) { 416 firstIndex = i; 417 } 418 lastIndex = i; 419 } 420 } 421 int numObjects = lastIndex - firstIndex + 1; 422 423 if ((mDataSize+len) > mDataCapacity) { 424 // grow data 425 err = growData(len); 426 if (err != NO_ERROR) { 427 return err; 428 } 429 } 430 431 // append data 432 memcpy(mData + mDataPos, data + offset, len); 433 mDataPos += len; 434 mDataSize += len; 435 436 err = NO_ERROR; 437 438 if (numObjects > 0) { 439 // grow objects 440 if (mObjectsCapacity < mObjectsSize + numObjects) { 441 int newSize = ((mObjectsSize + numObjects)*3)/2; 442 binder_size_t *objects = 443 (binder_size_t*)realloc(mObjects, newSize*sizeof(binder_size_t)); 444 if (objects == (binder_size_t*)0) { 445 return NO_MEMORY; 446 } 447 mObjects = objects; 448 mObjectsCapacity = newSize; 449 } 450 451 // append and acquire objects 452 int idx = mObjectsSize; 453 for (int i = firstIndex; i <= lastIndex; i++) { 454 size_t off = objects[i] - offset + startPos; 455 mObjects[idx++] = off; 456 mObjectsSize++; 457 458 flat_binder_object* flat 459 = reinterpret_cast<flat_binder_object*>(mData + off); 460 acquire_object(proc, *flat, this); 461 462 if (flat->type == BINDER_TYPE_FD) { 463 // If this is a file descriptor, we need to dup it so the 464 // new Parcel now owns its own fd, and can declare that we 465 // officially know we have fds. 466 flat->handle = dup(flat->handle); 467 flat->cookie = 1; 468 mHasFds = mFdsKnown = true; 469 if (!mAllowFds) { 470 err = FDS_NOT_ALLOWED; 471 } 472 } 473 } 474 } 475 476 return err; 477 } 478 479 bool Parcel::pushAllowFds(bool allowFds) 480 { 481 const bool origValue = mAllowFds; 482 if (!allowFds) { 483 mAllowFds = false; 484 } 485 return origValue; 486 } 487 488 void Parcel::restoreAllowFds(bool lastValue) 489 { 490 mAllowFds = lastValue; 491 } 492 493 bool Parcel::hasFileDescriptors() const 494 { 495 if (!mFdsKnown) { 496 scanForFds(); 497 } 498 return mHasFds; 499 } 500 501 // Write RPC headers. (previously just the interface token) 502 status_t Parcel::writeInterfaceToken(const String16& interface) 503 { 504 writeInt32(IPCThreadState::self()->getStrictModePolicy() | 505 STRICT_MODE_PENALTY_GATHER); 506 // currently the interface identification token is just its name as a string 507 return writeString16(interface); 508 } 509 510 bool Parcel::checkInterface(IBinder* binder) const 511 { 512 return enforceInterface(binder->getInterfaceDescriptor()); 513 } 514 515 bool Parcel::enforceInterface(const String16& interface, 516 IPCThreadState* threadState) const 517 { 518 int32_t strictPolicy = readInt32(); 519 if (threadState == NULL) { 520 threadState = IPCThreadState::self(); 521 } 522 if ((threadState->getLastTransactionBinderFlags() & 523 IBinder::FLAG_ONEWAY) != 0) { 524 // For one-way calls, the callee is running entirely 525 // disconnected from the caller, so disable StrictMode entirely. 526 // Not only does disk/network usage not impact the caller, but 527 // there's no way to commuicate back any violations anyway. 528 threadState->setStrictModePolicy(0); 529 } else { 530 threadState->setStrictModePolicy(strictPolicy); 531 } 532 const String16 str(readString16()); 533 if (str == interface) { 534 return true; 535 } else { 536 ALOGW("**** enforceInterface() expected '%s' but read '%s'", 537 String8(interface).string(), String8(str).string()); 538 return false; 539 } 540 } 541 542 const binder_size_t* Parcel::objects() const 543 { 544 return mObjects; 545 } 546 547 size_t Parcel::objectsCount() const 548 { 549 return mObjectsSize; 550 } 551 552 status_t Parcel::errorCheck() const 553 { 554 return mError; 555 } 556 557 void Parcel::setError(status_t err) 558 { 559 mError = err; 560 } 561 562 status_t Parcel::finishWrite(size_t len) 563 { 564 //printf("Finish write of %d\n", len); 565 mDataPos += len; 566 ALOGV("finishWrite Setting data pos of %p to %zu", this, mDataPos); 567 if (mDataPos > mDataSize) { 568 mDataSize = mDataPos; 569 ALOGV("finishWrite Setting data size of %p to %zu", this, mDataSize); 570 } 571 //printf("New pos=%d, size=%d\n", mDataPos, mDataSize); 572 return NO_ERROR; 573 } 574 575 status_t Parcel::writeUnpadded(const void* data, size_t len) 576 { 577 size_t end = mDataPos + len; 578 if (end < mDataPos) { 579 // integer overflow 580 return BAD_VALUE; 581 } 582 583 if (end <= mDataCapacity) { 584 restart_write: 585 memcpy(mData+mDataPos, data, len); 586 return finishWrite(len); 587 } 588 589 status_t err = growData(len); 590 if (err == NO_ERROR) goto restart_write; 591 return err; 592 } 593 594 status_t Parcel::write(const void* data, size_t len) 595 { 596 void* const d = writeInplace(len); 597 if (d) { 598 memcpy(d, data, len); 599 return NO_ERROR; 600 } 601 return mError; 602 } 603 604 void* Parcel::writeInplace(size_t len) 605 { 606 const size_t padded = PAD_SIZE(len); 607 608 // sanity check for integer overflow 609 if (mDataPos+padded < mDataPos) { 610 return NULL; 611 } 612 613 if ((mDataPos+padded) <= mDataCapacity) { 614 restart_write: 615 //printf("Writing %ld bytes, padded to %ld\n", len, padded); 616 uint8_t* const data = mData+mDataPos; 617 618 // Need to pad at end? 619 if (padded != len) { 620 #if BYTE_ORDER == BIG_ENDIAN 621 static const uint32_t mask[4] = { 622 0x00000000, 0xffffff00, 0xffff0000, 0xff000000 623 }; 624 #endif 625 #if BYTE_ORDER == LITTLE_ENDIAN 626 static const uint32_t mask[4] = { 627 0x00000000, 0x00ffffff, 0x0000ffff, 0x000000ff 628 }; 629 #endif 630 //printf("Applying pad mask: %p to %p\n", (void*)mask[padded-len], 631 // *reinterpret_cast<void**>(data+padded-4)); 632 *reinterpret_cast<uint32_t*>(data+padded-4) &= mask[padded-len]; 633 } 634 635 finishWrite(padded); 636 return data; 637 } 638 639 status_t err = growData(padded); 640 if (err == NO_ERROR) goto restart_write; 641 return NULL; 642 } 643 644 status_t Parcel::writeInt32(int32_t val) 645 { 646 return writeAligned(val); 647 } 648 status_t Parcel::writeInt32Array(size_t len, const int32_t *val) { 649 if (!val) { 650 return writeAligned(-1); 651 } 652 status_t ret = writeAligned(len); 653 if (ret == NO_ERROR) { 654 ret = write(val, len * sizeof(*val)); 655 } 656 return ret; 657 } 658 status_t Parcel::writeByteArray(size_t len, const uint8_t *val) { 659 if (!val) { 660 return writeAligned(-1); 661 } 662 status_t ret = writeAligned(len); 663 if (ret == NO_ERROR) { 664 ret = write(val, len * sizeof(*val)); 665 } 666 return ret; 667 } 668 669 status_t Parcel::writeInt64(int64_t val) 670 { 671 return writeAligned(val); 672 } 673 674 status_t Parcel::writePointer(uintptr_t val) 675 { 676 return writeAligned<binder_uintptr_t>(val); 677 } 678 679 status_t Parcel::writeFloat(float val) 680 { 681 return writeAligned(val); 682 } 683 684 #if defined(__mips__) && defined(__mips_hard_float) 685 686 status_t Parcel::writeDouble(double val) 687 { 688 union { 689 double d; 690 unsigned long long ll; 691 } u; 692 u.d = val; 693 return writeAligned(u.ll); 694 } 695 696 #else 697 698 status_t Parcel::writeDouble(double val) 699 { 700 return writeAligned(val); 701 } 702 703 #endif 704 705 status_t Parcel::writeCString(const char* str) 706 { 707 return write(str, strlen(str)+1); 708 } 709 710 status_t Parcel::writeString8(const String8& str) 711 { 712 status_t err = writeInt32(str.bytes()); 713 // only write string if its length is more than zero characters, 714 // as readString8 will only read if the length field is non-zero. 715 // this is slightly different from how writeString16 works. 716 if (str.bytes() > 0 && err == NO_ERROR) { 717 err = write(str.string(), str.bytes()+1); 718 } 719 return err; 720 } 721 722 status_t Parcel::writeString16(const String16& str) 723 { 724 return writeString16(str.string(), str.size()); 725 } 726 727 status_t Parcel::writeString16(const char16_t* str, size_t len) 728 { 729 if (str == NULL) return writeInt32(-1); 730 731 status_t err = writeInt32(len); 732 if (err == NO_ERROR) { 733 len *= sizeof(char16_t); 734 uint8_t* data = (uint8_t*)writeInplace(len+sizeof(char16_t)); 735 if (data) { 736 memcpy(data, str, len); 737 *reinterpret_cast<char16_t*>(data+len) = 0; 738 return NO_ERROR; 739 } 740 err = mError; 741 } 742 return err; 743 } 744 745 status_t Parcel::writeStrongBinder(const sp<IBinder>& val) 746 { 747 return flatten_binder(ProcessState::self(), val, this); 748 } 749 750 status_t Parcel::writeWeakBinder(const wp<IBinder>& val) 751 { 752 return flatten_binder(ProcessState::self(), val, this); 753 } 754 755 status_t Parcel::writeNativeHandle(const native_handle* handle) 756 { 757 if (!handle || handle->version != sizeof(native_handle)) 758 return BAD_TYPE; 759 760 status_t err; 761 err = writeInt32(handle->numFds); 762 if (err != NO_ERROR) return err; 763 764 err = writeInt32(handle->numInts); 765 if (err != NO_ERROR) return err; 766 767 for (int i=0 ; err==NO_ERROR && i<handle->numFds ; i++) 768 err = writeDupFileDescriptor(handle->data[i]); 769 770 if (err != NO_ERROR) { 771 ALOGD("write native handle, write dup fd failed"); 772 return err; 773 } 774 err = write(handle->data + handle->numFds, sizeof(int)*handle->numInts); 775 return err; 776 } 777 778 status_t Parcel::writeFileDescriptor(int fd, bool takeOwnership) 779 { 780 flat_binder_object obj; 781 obj.type = BINDER_TYPE_FD; 782 obj.flags = 0x7f | FLAT_BINDER_FLAG_ACCEPTS_FDS; 783 obj.binder = 0; /* Don't pass uninitialized stack data to a remote process */ 784 obj.handle = fd; 785 obj.cookie = takeOwnership ? 1 : 0; 786 return writeObject(obj, true); 787 } 788 789 status_t Parcel::writeDupFileDescriptor(int fd) 790 { 791 int dupFd = dup(fd); 792 if (dupFd < 0) { 793 return -errno; 794 } 795 status_t err = writeFileDescriptor(dupFd, true /*takeOwnership*/); 796 if (err) { 797 close(dupFd); 798 } 799 return err; 800 } 801 802 // WARNING: This method must stay in sync with 803 // Parcelable.Creator<ParcelFileDescriptor> CREATOR 804 // in frameworks/base/core/java/android/os/ParcelFileDescriptor.java 805 status_t Parcel::writeParcelFileDescriptor(int fd, int commChannel) { 806 status_t status; 807 808 if (fd < 0) { 809 status = writeInt32(0); // ParcelFileDescriptor is null 810 if (status) return status; 811 } else { 812 status = writeInt32(1); // ParcelFileDescriptor is not null 813 if (status) return status; 814 status = writeDupFileDescriptor(fd); 815 if (status) return status; 816 if (commChannel < 0) { 817 status = writeInt32(0); // commChannel is null 818 if (status) return status; 819 } else { 820 status = writeInt32(1); // commChannel is not null 821 if (status) return status; 822 status = writeDupFileDescriptor(commChannel); 823 } 824 } 825 return status; 826 } 827 828 status_t Parcel::writeBlob(size_t len, WritableBlob* outBlob) 829 { 830 status_t status; 831 832 if (!mAllowFds || len <= IN_PLACE_BLOB_LIMIT) { 833 ALOGV("writeBlob: write in place"); 834 status = writeInt32(0); 835 if (status) return status; 836 837 void* ptr = writeInplace(len); 838 if (!ptr) return NO_MEMORY; 839 840 outBlob->init(false /*mapped*/, ptr, len); 841 return NO_ERROR; 842 } 843 844 ALOGV("writeBlob: write to ashmem"); 845 int fd = ashmem_create_region("Parcel Blob", len); 846 if (fd < 0) return NO_MEMORY; 847 848 int result = ashmem_set_prot_region(fd, PROT_READ | PROT_WRITE); 849 if (result < 0) { 850 status = result; 851 } else { 852 void* ptr = ::mmap(NULL, len, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0); 853 if (ptr == MAP_FAILED) { 854 status = -errno; 855 } else { 856 result = ashmem_set_prot_region(fd, PROT_READ); 857 if (result < 0) { 858 status = result; 859 } else { 860 status = writeInt32(1); 861 if (!status) { 862 status = writeFileDescriptor(fd, true /*takeOwnership*/); 863 if (!status) { 864 outBlob->init(true /*mapped*/, ptr, len); 865 return NO_ERROR; 866 } 867 } 868 } 869 } 870 ::munmap(ptr, len); 871 } 872 ::close(fd); 873 return status; 874 } 875 876 status_t Parcel::write(const FlattenableHelperInterface& val) 877 { 878 status_t err; 879 880 // size if needed 881 const size_t len = val.getFlattenedSize(); 882 const size_t fd_count = val.getFdCount(); 883 884 err = this->writeInt32(len); 885 if (err) return err; 886 887 err = this->writeInt32(fd_count); 888 if (err) return err; 889 890 // payload 891 void* const buf = this->writeInplace(PAD_SIZE(len)); 892 if (buf == NULL) 893 return BAD_VALUE; 894 895 int* fds = NULL; 896 if (fd_count) { 897 fds = new int[fd_count]; 898 } 899 900 err = val.flatten(buf, len, fds, fd_count); 901 for (size_t i=0 ; i<fd_count && err==NO_ERROR ; i++) { 902 err = this->writeDupFileDescriptor( fds[i] ); 903 } 904 905 if (fd_count) { 906 delete [] fds; 907 } 908 909 return err; 910 } 911 912 status_t Parcel::writeObject(const flat_binder_object& val, bool nullMetaData) 913 { 914 const bool enoughData = (mDataPos+sizeof(val)) <= mDataCapacity; 915 const bool enoughObjects = mObjectsSize < mObjectsCapacity; 916 if (enoughData && enoughObjects) { 917 restart_write: 918 *reinterpret_cast<flat_binder_object*>(mData+mDataPos) = val; 919 920 // Need to write meta-data? 921 if (nullMetaData || val.binder != 0) { 922 mObjects[mObjectsSize] = mDataPos; 923 acquire_object(ProcessState::self(), val, this); 924 mObjectsSize++; 925 } 926 927 // remember if it's a file descriptor 928 if (val.type == BINDER_TYPE_FD) { 929 if (!mAllowFds) { 930 return FDS_NOT_ALLOWED; 931 } 932 mHasFds = mFdsKnown = true; 933 } 934 935 return finishWrite(sizeof(flat_binder_object)); 936 } 937 938 if (!enoughData) { 939 const status_t err = growData(sizeof(val)); 940 if (err != NO_ERROR) return err; 941 } 942 if (!enoughObjects) { 943 size_t newSize = ((mObjectsSize+2)*3)/2; 944 binder_size_t* objects = (binder_size_t*)realloc(mObjects, newSize*sizeof(binder_size_t)); 945 if (objects == NULL) return NO_MEMORY; 946 mObjects = objects; 947 mObjectsCapacity = newSize; 948 } 949 950 goto restart_write; 951 } 952 953 status_t Parcel::writeNoException() 954 { 955 return writeInt32(0); 956 } 957 958 void Parcel::remove(size_t /*start*/, size_t /*amt*/) 959 { 960 LOG_ALWAYS_FATAL("Parcel::remove() not yet implemented!"); 961 } 962 963 status_t Parcel::read(void* outData, size_t len) const 964 { 965 if ((mDataPos+PAD_SIZE(len)) >= mDataPos && (mDataPos+PAD_SIZE(len)) <= mDataSize 966 && len <= PAD_SIZE(len)) { 967 memcpy(outData, mData+mDataPos, len); 968 mDataPos += PAD_SIZE(len); 969 ALOGV("read Setting data pos of %p to %zu", this, mDataPos); 970 return NO_ERROR; 971 } 972 return NOT_ENOUGH_DATA; 973 } 974 975 const void* Parcel::readInplace(size_t len) const 976 { 977 if ((mDataPos+PAD_SIZE(len)) >= mDataPos && (mDataPos+PAD_SIZE(len)) <= mDataSize 978 && len <= PAD_SIZE(len)) { 979 const void* data = mData+mDataPos; 980 mDataPos += PAD_SIZE(len); 981 ALOGV("readInplace Setting data pos of %p to %zu", this, mDataPos); 982 return data; 983 } 984 return NULL; 985 } 986 987 template<class T> 988 status_t Parcel::readAligned(T *pArg) const { 989 COMPILE_TIME_ASSERT_FUNCTION_SCOPE(PAD_SIZE(sizeof(T)) == sizeof(T)); 990 991 if ((mDataPos+sizeof(T)) <= mDataSize) { 992 const void* data = mData+mDataPos; 993 mDataPos += sizeof(T); 994 *pArg = *reinterpret_cast<const T*>(data); 995 return NO_ERROR; 996 } else { 997 return NOT_ENOUGH_DATA; 998 } 999 } 1000 1001 template<class T> 1002 T Parcel::readAligned() const { 1003 T result; 1004 if (readAligned(&result) != NO_ERROR) { 1005 result = 0; 1006 } 1007 1008 return result; 1009 } 1010 1011 template<class T> 1012 status_t Parcel::writeAligned(T val) { 1013 COMPILE_TIME_ASSERT_FUNCTION_SCOPE(PAD_SIZE(sizeof(T)) == sizeof(T)); 1014 1015 if ((mDataPos+sizeof(val)) <= mDataCapacity) { 1016 restart_write: 1017 *reinterpret_cast<T*>(mData+mDataPos) = val; 1018 return finishWrite(sizeof(val)); 1019 } 1020 1021 status_t err = growData(sizeof(val)); 1022 if (err == NO_ERROR) goto restart_write; 1023 return err; 1024 } 1025 1026 status_t Parcel::readInt32(int32_t *pArg) const 1027 { 1028 return readAligned(pArg); 1029 } 1030 1031 int32_t Parcel::readInt32() const 1032 { 1033 return readAligned<int32_t>(); 1034 } 1035 1036 1037 status_t Parcel::readInt64(int64_t *pArg) const 1038 { 1039 return readAligned(pArg); 1040 } 1041 1042 1043 int64_t Parcel::readInt64() const 1044 { 1045 return readAligned<int64_t>(); 1046 } 1047 1048 status_t Parcel::readPointer(uintptr_t *pArg) const 1049 { 1050 status_t ret; 1051 binder_uintptr_t ptr; 1052 ret = readAligned(&ptr); 1053 if (!ret) 1054 *pArg = ptr; 1055 return ret; 1056 } 1057 1058 uintptr_t Parcel::readPointer() const 1059 { 1060 return readAligned<binder_uintptr_t>(); 1061 } 1062 1063 1064 status_t Parcel::readFloat(float *pArg) const 1065 { 1066 return readAligned(pArg); 1067 } 1068 1069 1070 float Parcel::readFloat() const 1071 { 1072 return readAligned<float>(); 1073 } 1074 1075 #if defined(__mips__) && defined(__mips_hard_float) 1076 1077 status_t Parcel::readDouble(double *pArg) const 1078 { 1079 union { 1080 double d; 1081 unsigned long long ll; 1082 } u; 1083 u.d = 0; 1084 status_t status; 1085 status = readAligned(&u.ll); 1086 *pArg = u.d; 1087 return status; 1088 } 1089 1090 double Parcel::readDouble() const 1091 { 1092 union { 1093 double d; 1094 unsigned long long ll; 1095 } u; 1096 u.ll = readAligned<unsigned long long>(); 1097 return u.d; 1098 } 1099 1100 #else 1101 1102 status_t Parcel::readDouble(double *pArg) const 1103 { 1104 return readAligned(pArg); 1105 } 1106 1107 double Parcel::readDouble() const 1108 { 1109 return readAligned<double>(); 1110 } 1111 1112 #endif 1113 1114 status_t Parcel::readIntPtr(intptr_t *pArg) const 1115 { 1116 return readAligned(pArg); 1117 } 1118 1119 1120 intptr_t Parcel::readIntPtr() const 1121 { 1122 return readAligned<intptr_t>(); 1123 } 1124 1125 1126 const char* Parcel::readCString() const 1127 { 1128 const size_t avail = mDataSize-mDataPos; 1129 if (avail > 0) { 1130 const char* str = reinterpret_cast<const char*>(mData+mDataPos); 1131 // is the string's trailing NUL within the parcel's valid bounds? 1132 const char* eos = reinterpret_cast<const char*>(memchr(str, 0, avail)); 1133 if (eos) { 1134 const size_t len = eos - str; 1135 mDataPos += PAD_SIZE(len+1); 1136 ALOGV("readCString Setting data pos of %p to %zu", this, mDataPos); 1137 return str; 1138 } 1139 } 1140 return NULL; 1141 } 1142 1143 String8 Parcel::readString8() const 1144 { 1145 int32_t size = readInt32(); 1146 // watch for potential int overflow adding 1 for trailing NUL 1147 if (size > 0 && size < INT32_MAX) { 1148 const char* str = (const char*)readInplace(size+1); 1149 if (str) return String8(str, size); 1150 } 1151 return String8(); 1152 } 1153 1154 String16 Parcel::readString16() const 1155 { 1156 size_t len; 1157 const char16_t* str = readString16Inplace(&len); 1158 if (str) return String16(str, len); 1159 ALOGE("Reading a NULL string not supported here."); 1160 return String16(); 1161 } 1162 1163 const char16_t* Parcel::readString16Inplace(size_t* outLen) const 1164 { 1165 int32_t size = readInt32(); 1166 // watch for potential int overflow from size+1 1167 if (size >= 0 && size < INT32_MAX) { 1168 *outLen = size; 1169 const char16_t* str = (const char16_t*)readInplace((size+1)*sizeof(char16_t)); 1170 if (str != NULL) { 1171 return str; 1172 } 1173 } 1174 *outLen = 0; 1175 return NULL; 1176 } 1177 1178 sp<IBinder> Parcel::readStrongBinder() const 1179 { 1180 sp<IBinder> val; 1181 unflatten_binder(ProcessState::self(), *this, &val); 1182 return val; 1183 } 1184 1185 wp<IBinder> Parcel::readWeakBinder() const 1186 { 1187 wp<IBinder> val; 1188 unflatten_binder(ProcessState::self(), *this, &val); 1189 return val; 1190 } 1191 1192 int32_t Parcel::readExceptionCode() const 1193 { 1194 int32_t exception_code = readAligned<int32_t>(); 1195 if (exception_code == EX_HAS_REPLY_HEADER) { 1196 int32_t header_start = dataPosition(); 1197 int32_t header_size = readAligned<int32_t>(); 1198 // Skip over fat responses headers. Not used (or propagated) in 1199 // native code 1200 setDataPosition(header_start + header_size); 1201 // And fat response headers are currently only used when there are no 1202 // exceptions, so return no error: 1203 return 0; 1204 } 1205 return exception_code; 1206 } 1207 1208 native_handle* Parcel::readNativeHandle() const 1209 { 1210 int numFds, numInts; 1211 status_t err; 1212 err = readInt32(&numFds); 1213 if (err != NO_ERROR) return 0; 1214 err = readInt32(&numInts); 1215 if (err != NO_ERROR) return 0; 1216 1217 native_handle* h = native_handle_create(numFds, numInts); 1218 for (int i=0 ; err==NO_ERROR && i<numFds ; i++) { 1219 h->data[i] = dup(readFileDescriptor()); 1220 if (h->data[i] < 0) err = BAD_VALUE; 1221 } 1222 err = read(h->data + numFds, sizeof(int)*numInts); 1223 if (err != NO_ERROR) { 1224 native_handle_close(h); 1225 native_handle_delete(h); 1226 h = 0; 1227 } 1228 return h; 1229 } 1230 1231 1232 int Parcel::readFileDescriptor() const 1233 { 1234 const flat_binder_object* flat = readObject(true); 1235 if (flat) { 1236 switch (flat->type) { 1237 case BINDER_TYPE_FD: 1238 //ALOGI("Returning file descriptor %ld from parcel %p", flat->handle, this); 1239 return flat->handle; 1240 } 1241 } 1242 return BAD_TYPE; 1243 } 1244 1245 // WARNING: This method must stay in sync with writeToParcel() 1246 // in frameworks/base/core/java/android/os/ParcelFileDescriptor.java 1247 int Parcel::readParcelFileDescriptor(int& outCommChannel) const { 1248 int fd; 1249 outCommChannel = -1; 1250 1251 if (readInt32() == 0) { 1252 fd = -1; 1253 } else { 1254 fd = readFileDescriptor(); 1255 if (fd >= 0 && readInt32() != 0) { 1256 outCommChannel = readFileDescriptor(); 1257 } 1258 } 1259 return fd; 1260 } 1261 1262 status_t Parcel::readBlob(size_t len, ReadableBlob* outBlob) const 1263 { 1264 int32_t useAshmem; 1265 status_t status = readInt32(&useAshmem); 1266 if (status) return status; 1267 1268 if (!useAshmem) { 1269 ALOGV("readBlob: read in place"); 1270 const void* ptr = readInplace(len); 1271 if (!ptr) return BAD_VALUE; 1272 1273 outBlob->init(false /*mapped*/, const_cast<void*>(ptr), len); 1274 return NO_ERROR; 1275 } 1276 1277 ALOGV("readBlob: read from ashmem"); 1278 int fd = readFileDescriptor(); 1279 if (fd == int(BAD_TYPE)) return BAD_VALUE; 1280 1281 void* ptr = ::mmap(NULL, len, PROT_READ, MAP_SHARED, fd, 0); 1282 if (ptr == MAP_FAILED) return NO_MEMORY; 1283 1284 outBlob->init(true /*mapped*/, ptr, len); 1285 return NO_ERROR; 1286 } 1287 1288 status_t Parcel::read(FlattenableHelperInterface& val) const 1289 { 1290 // size 1291 const size_t len = this->readInt32(); 1292 const size_t fd_count = this->readInt32(); 1293 1294 // payload 1295 void const* const buf = this->readInplace(PAD_SIZE(len)); 1296 if (buf == NULL) 1297 return BAD_VALUE; 1298 1299 int* fds = NULL; 1300 if (fd_count) { 1301 fds = new int[fd_count]; 1302 } 1303 1304 status_t err = NO_ERROR; 1305 for (size_t i=0 ; i<fd_count && err==NO_ERROR ; i++) { 1306 fds[i] = dup(this->readFileDescriptor()); 1307 if (fds[i] < 0) { 1308 err = BAD_VALUE; 1309 ALOGE("dup() failed in Parcel::read, i is %zu, fds[i] is %d, fd_count is %zu, error: %s", 1310 i, fds[i], fd_count, strerror(errno)); 1311 } 1312 } 1313 1314 if (err == NO_ERROR) { 1315 err = val.unflatten(buf, len, fds, fd_count); 1316 } 1317 1318 if (fd_count) { 1319 delete [] fds; 1320 } 1321 1322 return err; 1323 } 1324 const flat_binder_object* Parcel::readObject(bool nullMetaData) const 1325 { 1326 const size_t DPOS = mDataPos; 1327 if ((DPOS+sizeof(flat_binder_object)) <= mDataSize) { 1328 const flat_binder_object* obj 1329 = reinterpret_cast<const flat_binder_object*>(mData+DPOS); 1330 mDataPos = DPOS + sizeof(flat_binder_object); 1331 if (!nullMetaData && (obj->cookie == 0 && obj->binder == 0)) { 1332 // When transferring a NULL object, we don't write it into 1333 // the object list, so we don't want to check for it when 1334 // reading. 1335 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); 1336 return obj; 1337 } 1338 1339 // Ensure that this object is valid... 1340 binder_size_t* const OBJS = mObjects; 1341 const size_t N = mObjectsSize; 1342 size_t opos = mNextObjectHint; 1343 1344 if (N > 0) { 1345 ALOGV("Parcel %p looking for obj at %zu, hint=%zu", 1346 this, DPOS, opos); 1347 1348 // Start at the current hint position, looking for an object at 1349 // the current data position. 1350 if (opos < N) { 1351 while (opos < (N-1) && OBJS[opos] < DPOS) { 1352 opos++; 1353 } 1354 } else { 1355 opos = N-1; 1356 } 1357 if (OBJS[opos] == DPOS) { 1358 // Found it! 1359 ALOGV("Parcel %p found obj %zu at index %zu with forward search", 1360 this, DPOS, opos); 1361 mNextObjectHint = opos+1; 1362 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); 1363 return obj; 1364 } 1365 1366 // Look backwards for it... 1367 while (opos > 0 && OBJS[opos] > DPOS) { 1368 opos--; 1369 } 1370 if (OBJS[opos] == DPOS) { 1371 // Found it! 1372 ALOGV("Parcel %p found obj %zu at index %zu with backward search", 1373 this, DPOS, opos); 1374 mNextObjectHint = opos+1; 1375 ALOGV("readObject Setting data pos of %p to %zu", this, mDataPos); 1376 return obj; 1377 } 1378 } 1379 ALOGW("Attempt to read object from Parcel %p at offset %zu that is not in the object list", 1380 this, DPOS); 1381 } 1382 return NULL; 1383 } 1384 1385 void Parcel::closeFileDescriptors() 1386 { 1387 size_t i = mObjectsSize; 1388 if (i > 0) { 1389 //ALOGI("Closing file descriptors for %zu objects...", i); 1390 } 1391 while (i > 0) { 1392 i--; 1393 const flat_binder_object* flat 1394 = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]); 1395 if (flat->type == BINDER_TYPE_FD) { 1396 //ALOGI("Closing fd: %ld", flat->handle); 1397 close(flat->handle); 1398 } 1399 } 1400 } 1401 1402 uintptr_t Parcel::ipcData() const 1403 { 1404 return reinterpret_cast<uintptr_t>(mData); 1405 } 1406 1407 size_t Parcel::ipcDataSize() const 1408 { 1409 return (mDataSize > mDataPos ? mDataSize : mDataPos); 1410 } 1411 1412 uintptr_t Parcel::ipcObjects() const 1413 { 1414 return reinterpret_cast<uintptr_t>(mObjects); 1415 } 1416 1417 size_t Parcel::ipcObjectsCount() const 1418 { 1419 return mObjectsSize; 1420 } 1421 1422 void Parcel::ipcSetDataReference(const uint8_t* data, size_t dataSize, 1423 const binder_size_t* objects, size_t objectsCount, release_func relFunc, void* relCookie) 1424 { 1425 binder_size_t minOffset = 0; 1426 freeDataNoInit(); 1427 mError = NO_ERROR; 1428 mData = const_cast<uint8_t*>(data); 1429 mDataSize = mDataCapacity = dataSize; 1430 //ALOGI("setDataReference Setting data size of %p to %lu (pid=%d)", this, mDataSize, getpid()); 1431 mDataPos = 0; 1432 ALOGV("setDataReference Setting data pos of %p to %zu", this, mDataPos); 1433 mObjects = const_cast<binder_size_t*>(objects); 1434 mObjectsSize = mObjectsCapacity = objectsCount; 1435 mNextObjectHint = 0; 1436 mOwner = relFunc; 1437 mOwnerCookie = relCookie; 1438 for (size_t i = 0; i < mObjectsSize; i++) { 1439 binder_size_t offset = mObjects[i]; 1440 if (offset < minOffset) { 1441 ALOGE("%s: bad object offset %"PRIu64" < %"PRIu64"\n", 1442 __func__, (uint64_t)offset, (uint64_t)minOffset); 1443 mObjectsSize = 0; 1444 break; 1445 } 1446 minOffset = offset + sizeof(flat_binder_object); 1447 } 1448 scanForFds(); 1449 } 1450 1451 void Parcel::print(TextOutput& to, uint32_t /*flags*/) const 1452 { 1453 to << "Parcel("; 1454 1455 if (errorCheck() != NO_ERROR) { 1456 const status_t err = errorCheck(); 1457 to << "Error: " << (void*)(intptr_t)err << " \"" << strerror(-err) << "\""; 1458 } else if (dataSize() > 0) { 1459 const uint8_t* DATA = data(); 1460 to << indent << HexDump(DATA, dataSize()) << dedent; 1461 const binder_size_t* OBJS = objects(); 1462 const size_t N = objectsCount(); 1463 for (size_t i=0; i<N; i++) { 1464 const flat_binder_object* flat 1465 = reinterpret_cast<const flat_binder_object*>(DATA+OBJS[i]); 1466 to << endl << "Object #" << i << " @ " << (void*)OBJS[i] << ": " 1467 << TypeCode(flat->type & 0x7f7f7f00) 1468 << " = " << flat->binder; 1469 } 1470 } else { 1471 to << "NULL"; 1472 } 1473 1474 to << ")"; 1475 } 1476 1477 void Parcel::releaseObjects() 1478 { 1479 const sp<ProcessState> proc(ProcessState::self()); 1480 size_t i = mObjectsSize; 1481 uint8_t* const data = mData; 1482 binder_size_t* const objects = mObjects; 1483 while (i > 0) { 1484 i--; 1485 const flat_binder_object* flat 1486 = reinterpret_cast<flat_binder_object*>(data+objects[i]); 1487 release_object(proc, *flat, this); 1488 } 1489 } 1490 1491 void Parcel::acquireObjects() 1492 { 1493 const sp<ProcessState> proc(ProcessState::self()); 1494 size_t i = mObjectsSize; 1495 uint8_t* const data = mData; 1496 binder_size_t* const objects = mObjects; 1497 while (i > 0) { 1498 i--; 1499 const flat_binder_object* flat 1500 = reinterpret_cast<flat_binder_object*>(data+objects[i]); 1501 acquire_object(proc, *flat, this); 1502 } 1503 } 1504 1505 void Parcel::freeData() 1506 { 1507 freeDataNoInit(); 1508 initState(); 1509 } 1510 1511 void Parcel::freeDataNoInit() 1512 { 1513 if (mOwner) { 1514 LOG_ALLOC("Parcel %p: freeing other owner data", this); 1515 //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid()); 1516 mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie); 1517 } else { 1518 LOG_ALLOC("Parcel %p: freeing allocated data", this); 1519 releaseObjects(); 1520 if (mData) { 1521 LOG_ALLOC("Parcel %p: freeing with %zu capacity", this, mDataCapacity); 1522 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1523 gParcelGlobalAllocSize -= mDataCapacity; 1524 gParcelGlobalAllocCount--; 1525 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1526 free(mData); 1527 } 1528 if (mObjects) free(mObjects); 1529 } 1530 } 1531 1532 status_t Parcel::growData(size_t len) 1533 { 1534 size_t newSize = ((mDataSize+len)*3)/2; 1535 return (newSize <= mDataSize) 1536 ? (status_t) NO_MEMORY 1537 : continueWrite(newSize); 1538 } 1539 1540 status_t Parcel::restartWrite(size_t desired) 1541 { 1542 if (mOwner) { 1543 freeData(); 1544 return continueWrite(desired); 1545 } 1546 1547 uint8_t* data = (uint8_t*)realloc(mData, desired); 1548 if (!data && desired > mDataCapacity) { 1549 mError = NO_MEMORY; 1550 return NO_MEMORY; 1551 } 1552 1553 releaseObjects(); 1554 1555 if (data) { 1556 LOG_ALLOC("Parcel %p: restart from %zu to %zu capacity", this, mDataCapacity, desired); 1557 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1558 gParcelGlobalAllocSize += desired; 1559 gParcelGlobalAllocSize -= mDataCapacity; 1560 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1561 mData = data; 1562 mDataCapacity = desired; 1563 } 1564 1565 mDataSize = mDataPos = 0; 1566 ALOGV("restartWrite Setting data size of %p to %zu", this, mDataSize); 1567 ALOGV("restartWrite Setting data pos of %p to %zu", this, mDataPos); 1568 1569 free(mObjects); 1570 mObjects = NULL; 1571 mObjectsSize = mObjectsCapacity = 0; 1572 mNextObjectHint = 0; 1573 mHasFds = false; 1574 mFdsKnown = true; 1575 mAllowFds = true; 1576 1577 return NO_ERROR; 1578 } 1579 1580 status_t Parcel::continueWrite(size_t desired) 1581 { 1582 // If shrinking, first adjust for any objects that appear 1583 // after the new data size. 1584 size_t objectsSize = mObjectsSize; 1585 if (desired < mDataSize) { 1586 if (desired == 0) { 1587 objectsSize = 0; 1588 } else { 1589 while (objectsSize > 0) { 1590 if (mObjects[objectsSize-1] < desired) 1591 break; 1592 objectsSize--; 1593 } 1594 } 1595 } 1596 1597 if (mOwner) { 1598 // If the size is going to zero, just release the owner's data. 1599 if (desired == 0) { 1600 freeData(); 1601 return NO_ERROR; 1602 } 1603 1604 // If there is a different owner, we need to take 1605 // posession. 1606 uint8_t* data = (uint8_t*)malloc(desired); 1607 if (!data) { 1608 mError = NO_MEMORY; 1609 return NO_MEMORY; 1610 } 1611 binder_size_t* objects = NULL; 1612 1613 if (objectsSize) { 1614 objects = (binder_size_t*)malloc(objectsSize*sizeof(binder_size_t)); 1615 if (!objects) { 1616 free(data); 1617 1618 mError = NO_MEMORY; 1619 return NO_MEMORY; 1620 } 1621 1622 // Little hack to only acquire references on objects 1623 // we will be keeping. 1624 size_t oldObjectsSize = mObjectsSize; 1625 mObjectsSize = objectsSize; 1626 acquireObjects(); 1627 mObjectsSize = oldObjectsSize; 1628 } 1629 1630 if (mData) { 1631 memcpy(data, mData, mDataSize < desired ? mDataSize : desired); 1632 } 1633 if (objects && mObjects) { 1634 memcpy(objects, mObjects, objectsSize*sizeof(binder_size_t)); 1635 } 1636 //ALOGI("Freeing data ref of %p (pid=%d)", this, getpid()); 1637 mOwner(this, mData, mDataSize, mObjects, mObjectsSize, mOwnerCookie); 1638 mOwner = NULL; 1639 1640 LOG_ALLOC("Parcel %p: taking ownership of %zu capacity", this, desired); 1641 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1642 gParcelGlobalAllocSize += desired; 1643 gParcelGlobalAllocCount++; 1644 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1645 1646 mData = data; 1647 mObjects = objects; 1648 mDataSize = (mDataSize < desired) ? mDataSize : desired; 1649 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); 1650 mDataCapacity = desired; 1651 mObjectsSize = mObjectsCapacity = objectsSize; 1652 mNextObjectHint = 0; 1653 1654 } else if (mData) { 1655 if (objectsSize < mObjectsSize) { 1656 // Need to release refs on any objects we are dropping. 1657 const sp<ProcessState> proc(ProcessState::self()); 1658 for (size_t i=objectsSize; i<mObjectsSize; i++) { 1659 const flat_binder_object* flat 1660 = reinterpret_cast<flat_binder_object*>(mData+mObjects[i]); 1661 if (flat->type == BINDER_TYPE_FD) { 1662 // will need to rescan because we may have lopped off the only FDs 1663 mFdsKnown = false; 1664 } 1665 release_object(proc, *flat, this); 1666 } 1667 binder_size_t* objects = 1668 (binder_size_t*)realloc(mObjects, objectsSize*sizeof(binder_size_t)); 1669 if (objects) { 1670 mObjects = objects; 1671 } 1672 mObjectsSize = objectsSize; 1673 mNextObjectHint = 0; 1674 } 1675 1676 // We own the data, so we can just do a realloc(). 1677 if (desired > mDataCapacity) { 1678 uint8_t* data = (uint8_t*)realloc(mData, desired); 1679 if (data) { 1680 LOG_ALLOC("Parcel %p: continue from %zu to %zu capacity", this, mDataCapacity, 1681 desired); 1682 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1683 gParcelGlobalAllocSize += desired; 1684 gParcelGlobalAllocSize -= mDataCapacity; 1685 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1686 mData = data; 1687 mDataCapacity = desired; 1688 } else if (desired > mDataCapacity) { 1689 mError = NO_MEMORY; 1690 return NO_MEMORY; 1691 } 1692 } else { 1693 if (mDataSize > desired) { 1694 mDataSize = desired; 1695 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); 1696 } 1697 if (mDataPos > desired) { 1698 mDataPos = desired; 1699 ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos); 1700 } 1701 } 1702 1703 } else { 1704 // This is the first data. Easy! 1705 uint8_t* data = (uint8_t*)malloc(desired); 1706 if (!data) { 1707 mError = NO_MEMORY; 1708 return NO_MEMORY; 1709 } 1710 1711 if(!(mDataCapacity == 0 && mObjects == NULL 1712 && mObjectsCapacity == 0)) { 1713 ALOGE("continueWrite: %zu/%p/%zu/%zu", mDataCapacity, mObjects, mObjectsCapacity, desired); 1714 } 1715 1716 LOG_ALLOC("Parcel %p: allocating with %zu capacity", this, desired); 1717 pthread_mutex_lock(&gParcelGlobalAllocSizeLock); 1718 gParcelGlobalAllocSize += desired; 1719 gParcelGlobalAllocCount++; 1720 pthread_mutex_unlock(&gParcelGlobalAllocSizeLock); 1721 1722 mData = data; 1723 mDataSize = mDataPos = 0; 1724 ALOGV("continueWrite Setting data size of %p to %zu", this, mDataSize); 1725 ALOGV("continueWrite Setting data pos of %p to %zu", this, mDataPos); 1726 mDataCapacity = desired; 1727 } 1728 1729 return NO_ERROR; 1730 } 1731 1732 void Parcel::initState() 1733 { 1734 LOG_ALLOC("Parcel %p: initState", this); 1735 mError = NO_ERROR; 1736 mData = 0; 1737 mDataSize = 0; 1738 mDataCapacity = 0; 1739 mDataPos = 0; 1740 ALOGV("initState Setting data size of %p to %zu", this, mDataSize); 1741 ALOGV("initState Setting data pos of %p to %zu", this, mDataPos); 1742 mObjects = NULL; 1743 mObjectsSize = 0; 1744 mObjectsCapacity = 0; 1745 mNextObjectHint = 0; 1746 mHasFds = false; 1747 mFdsKnown = true; 1748 mAllowFds = true; 1749 mOwner = NULL; 1750 } 1751 1752 void Parcel::scanForFds() const 1753 { 1754 bool hasFds = false; 1755 for (size_t i=0; i<mObjectsSize; i++) { 1756 const flat_binder_object* flat 1757 = reinterpret_cast<const flat_binder_object*>(mData + mObjects[i]); 1758 if (flat->type == BINDER_TYPE_FD) { 1759 hasFds = true; 1760 break; 1761 } 1762 } 1763 mHasFds = hasFds; 1764 mFdsKnown = true; 1765 } 1766 1767 // --- Parcel::Blob --- 1768 1769 Parcel::Blob::Blob() : 1770 mMapped(false), mData(NULL), mSize(0) { 1771 } 1772 1773 Parcel::Blob::~Blob() { 1774 release(); 1775 } 1776 1777 void Parcel::Blob::release() { 1778 if (mMapped && mData) { 1779 ::munmap(mData, mSize); 1780 } 1781 clear(); 1782 } 1783 1784 void Parcel::Blob::init(bool mapped, void* data, size_t size) { 1785 mMapped = mapped; 1786 mData = data; 1787 mSize = size; 1788 } 1789 1790 void Parcel::Blob::clear() { 1791 mMapped = false; 1792 mData = NULL; 1793 mSize = 0; 1794 } 1795 1796 }; // namespace android 1797