Home | History | Annotate | Download | only in space
      1 /*
      2  * Copyright (C) 2013 The Android Open Source Project
      3  *
      4  * Licensed under the Apache License, Version 2.0 (the "License");
      5  * you may not use this file except in compliance with the License.
      6  * You may obtain a copy of the License at
      7  *
      8  *      http://www.apache.org/licenses/LICENSE-2.0
      9  *
     10  * Unless required by applicable law or agreed to in writing, software
     11  * distributed under the License is distributed on an "AS IS" BASIS,
     12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
     13  * See the License for the specific language governing permissions and
     14  * limitations under the License.
     15  */
     16 
     17 #ifndef ART_RUNTIME_GC_SPACE_BUMP_POINTER_SPACE_INL_H_
     18 #define ART_RUNTIME_GC_SPACE_BUMP_POINTER_SPACE_INL_H_
     19 
     20 #include "base/bit_utils.h"
     21 #include "bump_pointer_space.h"
     22 
     23 namespace art {
     24 namespace gc {
     25 namespace space {
     26 
     27 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated,
     28                                                size_t* usable_size,
     29                                                size_t* bytes_tl_bulk_allocated) {
     30   num_bytes = RoundUp(num_bytes, kAlignment);
     31   mirror::Object* ret = AllocNonvirtual(num_bytes);
     32   if (LIKELY(ret != nullptr)) {
     33     *bytes_allocated = num_bytes;
     34     if (usable_size != nullptr) {
     35       *usable_size = num_bytes;
     36     }
     37     *bytes_tl_bulk_allocated = num_bytes;
     38   }
     39   return ret;
     40 }
     41 
     42 inline mirror::Object* BumpPointerSpace::AllocThreadUnsafe(Thread* self, size_t num_bytes,
     43                                                            size_t* bytes_allocated,
     44                                                            size_t* usable_size,
     45                                                            size_t* bytes_tl_bulk_allocated) {
     46   Locks::mutator_lock_->AssertExclusiveHeld(self);
     47   num_bytes = RoundUp(num_bytes, kAlignment);
     48   uint8_t* end = end_.LoadRelaxed();
     49   if (end + num_bytes > growth_end_) {
     50     return nullptr;
     51   }
     52   mirror::Object* obj = reinterpret_cast<mirror::Object*>(end);
     53   end_.StoreRelaxed(end + num_bytes);
     54   *bytes_allocated = num_bytes;
     55   // Use the CAS free versions as an optimization.
     56   objects_allocated_.StoreRelaxed(objects_allocated_.LoadRelaxed() + 1);
     57   bytes_allocated_.StoreRelaxed(bytes_allocated_.LoadRelaxed() + num_bytes);
     58   if (UNLIKELY(usable_size != nullptr)) {
     59     *usable_size = num_bytes;
     60   }
     61   *bytes_tl_bulk_allocated = num_bytes;
     62   return obj;
     63 }
     64 
     65 inline mirror::Object* BumpPointerSpace::AllocNonvirtualWithoutAccounting(size_t num_bytes) {
     66   DCHECK(IsAligned<kAlignment>(num_bytes));
     67   uint8_t* old_end;
     68   uint8_t* new_end;
     69   do {
     70     old_end = end_.LoadRelaxed();
     71     new_end = old_end + num_bytes;
     72     // If there is no more room in the region, we are out of memory.
     73     if (UNLIKELY(new_end > growth_end_)) {
     74       return nullptr;
     75     }
     76   } while (!end_.CompareExchangeWeakSequentiallyConsistent(old_end, new_end));
     77   return reinterpret_cast<mirror::Object*>(old_end);
     78 }
     79 
     80 inline mirror::Object* BumpPointerSpace::AllocNonvirtual(size_t num_bytes) {
     81   mirror::Object* ret = AllocNonvirtualWithoutAccounting(num_bytes);
     82   if (ret != nullptr) {
     83     objects_allocated_.FetchAndAddSequentiallyConsistent(1);
     84     bytes_allocated_.FetchAndAddSequentiallyConsistent(num_bytes);
     85   }
     86   return ret;
     87 }
     88 
     89 inline size_t BumpPointerSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
     90     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
     91   size_t num_bytes = obj->SizeOf();
     92   if (usable_size != nullptr) {
     93     *usable_size = RoundUp(num_bytes, kAlignment);
     94   }
     95   return num_bytes;
     96 }
     97 
     98 }  // namespace space
     99 }  // namespace gc
    100 }  // namespace art
    101 
    102 #endif  // ART_RUNTIME_GC_SPACE_BUMP_POINTER_SPACE_INL_H_
    103