Lines Matching defs:code
6 // * Redistributions of source code must retain the above copyright
250 Code* lazy_compile = isolate_->builtins()->builtin(Builtins::kLazyCompile);
259 Code* code = shared->unchecked_code();
260 if (!code->IsMarked()) {
275 Code* lazy_compile = isolate_->builtins()->builtin(Builtins::kLazyCompile);
283 Code* code = candidate->unchecked_code();
284 if (!code->IsMarked()) {
308 STATIC_ASSERT(kPointerSize <= Code::kHeaderSize - Code::kHeaderPaddingStart);
312 Code* code = candidate->unchecked_code();
314 code->address() + Code::kHeaderPaddingStart);
461 Code* code = Code
462 if (FLAG_cleanup_ics_at_gc && code->is_inline_cache_stub()) {
467 heap->mark_compact_collector()->MarkObject(code);
486 HeapObject* code = Code::GetCodeFromTargetAddress(rinfo->call_address());
487 heap->mark_compact_collector()->MarkObject(code);
557 reinterpret_cast<Code*>(object)->CodeIterateBody<StaticMarkingVisitor>(
561 // Code flushing support.
563 // How many collections newly compiled code object will survive before being
587 // Code is either on stack, in compilation cache or referenced
594 // We do not flush code for optimized functions.
595 if (function->code() != shared_info->unchecked_code()) {
603 // Code is either on stack, in compilation cache or referenced
610 // The function must be compiled and have the source code available,
616 // We never flush code for Api functions.
624 // Only flush code for functions.
625 if (shared_info->code()->kind() != Code::FUNCTION) return false;
630 // If this is a full script wrapped in a function we do no flush the code.
646 // This function's code looks flushable. But we have to postpone the
649 // That would make the nonoptimized version of the code nonflushable,
650 // because it is required for bailing out from optimized code.
732 Object* code = Code::GetObjectFromEntryAddress(entry_address);
733 Object* old_code = code;
734 VisitPointer(heap, &code);
735 if (code != old_code) {
737 reinterpret_cast<Code*>(code)->entry();
760 if (jsfunction->unchecked_code()->kind() == Code::OPTIMIZED_FUNCTION) {
762 // of it's code and non-optimized version of all inlined functions.
763 // This is required to support bailing out from inlined code.
809 // Don't visit code object.
948 // Make sure we are not referencing the code from the stack.
955 // the code is referenced.
1447 // Flush code from collected candidates.
1648 // strings are promoted to the code space, large objects to large object space,
1911 Object* target = Code::GetCodeFromTargetAddress(rinfo->target_address());
1913 rinfo->set_target_address(Code::cast(target)->instruction_start());
1921 Object* target = Code::GetCodeFromTargetAddress(rinfo->call_address());
1923 rinfo->set_call_address(Code::cast(target)->instruction_start());
2266 // Compute new space next to last after the old and code spaces have been
2267 // compacted. Objects in new space can be promoted to old or code space.
2622 Object* target = Code::GetCodeFromTargetAddress(rinfo->target_address());
2625 reinterpret_cast<Code*>(target)->instruction_start());
2633 Object* target = Code::GetCodeFromTargetAddress(rinfo->call_address());
2636 reinterpret_cast<Code*>(target)->instruction_start());
3044 Code::cast(copied_to)->Relocate(new_addr - old_addr);
3045 // Notify the logger that compiled code has moved.
3116 GDBJITInterface::RemoveCode(reinterpret_cast<Code*>(obj));