AddIncrementCounter(counters->fast_new_closure_total());
- IfBuilder optimize_now(this);
- HInstruction* compile_hint = Add<HLoadNamedField>(
- shared_info, static_cast<HValue*>(NULL), HObjectAccess::ForCompileHint());
- HValue* hint_mask = Add<HConstant>(
- static_cast<int32_t>(1 << SharedFunctionInfo::kOptimizeNextClosure));
- HInstruction* optimize =
- AddUncasted<HBitwise>(Token::BIT_AND, compile_hint, hint_mask);
- optimize_now.If<HCompareNumericAndBranch>(optimize, hint_mask, Token::EQ);
- optimize_now.Then();
- {
- Add<HPushArguments>(context(), shared_info, graph()->GetConstantFalse());
- Push(Add<HCallRuntime>(isolate()->factory()->empty_string(),
- Runtime::FunctionForId(Runtime::kNewClosure), 3));
- }
- optimize_now.Else();
- {
- // Create a new closure from the given function info in new space
- HValue* size = Add<HConstant>(JSFunction::kSize);
- HInstruction* js_function =
- Add<HAllocate>(size, HType::JSObject(), NOT_TENURED, JS_FUNCTION_TYPE);
-
- int map_index = Context::FunctionMapIndex(casted_stub()->strict_mode(),
- casted_stub()->kind());
-
- // Compute the function map in the current native context and set that
- // as the map of the allocated object.
- HInstruction* native_context = BuildGetNativeContext();
- HInstruction* map_slot_value =
- Add<HLoadNamedField>(native_context, static_cast<HValue*>(NULL),
- HObjectAccess::ForContextSlot(map_index));
- Add<HStoreNamedField>(js_function, HObjectAccess::ForMap(), map_slot_value);
-
- // Initialize the rest of the function.
- Add<HStoreNamedField>(js_function, HObjectAccess::ForPropertiesPointer(),
- empty_fixed_array);
- Add<HStoreNamedField>(js_function, HObjectAccess::ForElementsPointer(),
- empty_fixed_array);
- Add<HStoreNamedField>(js_function, HObjectAccess::ForLiteralsPointer(),
- empty_fixed_array);
- Add<HStoreNamedField>(js_function,
- HObjectAccess::ForPrototypeOrInitialMap(),
- graph()->GetConstantHole());
- Add<HStoreNamedField>(js_function,
- HObjectAccess::ForSharedFunctionInfoPointer(),
- shared_info);
- Add<HStoreNamedField>(
- js_function, HObjectAccess::ForFunctionContextPointer(), context());
-
- // Initialize the code pointer in the function to be the one
- // found in the shared function info object.
- // But first check if there is an optimized version for our context.
- if (FLAG_cache_optimized_code) {
- BuildInstallFromOptimizedCodeMap(js_function, shared_info,
- native_context);
- } else {
- BuildInstallCode(js_function, shared_info);
- }
- Push(js_function);
+ // Create a new closure from the given function info in new space
+ HValue* size = Add<HConstant>(JSFunction::kSize);
+ HInstruction* js_function = Add<HAllocate>(size, HType::JSObject(),
+ NOT_TENURED, JS_FUNCTION_TYPE);
+
+ int map_index = Context::FunctionMapIndex(casted_stub()->strict_mode(),
+ casted_stub()->kind());
+
+ // Compute the function map in the current native context and set that
+ // as the map of the allocated object.
+ HInstruction* native_context = BuildGetNativeContext();
+ HInstruction* map_slot_value = Add<HLoadNamedField>(
+ native_context, static_cast<HValue*>(NULL),
+ HObjectAccess::ForContextSlot(map_index));
+ Add<HStoreNamedField>(js_function, HObjectAccess::ForMap(), map_slot_value);
+
+ // Initialize the rest of the function.
+ Add<HStoreNamedField>(js_function, HObjectAccess::ForPropertiesPointer(),
+ empty_fixed_array);
+ Add<HStoreNamedField>(js_function, HObjectAccess::ForElementsPointer(),
+ empty_fixed_array);
+ Add<HStoreNamedField>(js_function, HObjectAccess::ForLiteralsPointer(),
+ empty_fixed_array);
+ Add<HStoreNamedField>(js_function, HObjectAccess::ForPrototypeOrInitialMap(),
+ graph()->GetConstantHole());
+ Add<HStoreNamedField>(js_function,
+ HObjectAccess::ForSharedFunctionInfoPointer(),
+ shared_info);
+ Add<HStoreNamedField>(js_function, HObjectAccess::ForFunctionContextPointer(),
+ context());
+
+ // Initialize the code pointer in the function to be the one
+ // found in the shared function info object.
+ // But first check if there is an optimized version for our context.
+ if (FLAG_cache_optimized_code) {
+ BuildInstallFromOptimizedCodeMap(js_function, shared_info, native_context);
+ } else {
+ BuildInstallCode(js_function, shared_info);
}
- optimize_now.End();
- return Pop();
+
+ return js_function;
}
}
-static bool ShouldOptimizeNewClosure(Isolate* isolate,
- Handle<SharedFunctionInfo> info) {
- return isolate->use_crankshaft() && !info->is_toplevel() &&
- info->is_compiled() && info->allows_lazy_compilation() &&
- !info->optimization_disabled() && !isolate->DebuggerHasBreakPoints();
-}
-
-
Handle<JSFunction> Factory::NewFunctionFromSharedFunctionInfo(
Handle<SharedFunctionInfo> info,
Handle<Context> context,
return result;
}
- if (FLAG_always_opt && ShouldOptimizeNewClosure(isolate(), info)) {
+ if (isolate()->use_crankshaft() &&
+ FLAG_always_opt &&
+ result->is_compiled() &&
+ !info->is_toplevel() &&
+ info->allows_lazy_compilation() &&
+ !info->optimization_disabled() &&
+ !isolate()->DebuggerHasBreakPoints()) {
result->MarkForOptimization();
- } else if (info->optimize_next_closure() &&
- ShouldOptimizeNewClosure(isolate(), info)) {
- result->AttemptConcurrentOptimization();
}
return result;
}
}
-void JSFunction::AttemptConcurrentOptimization() {
- Isolate* isolate = GetIsolate();
- if (!isolate->concurrent_recompilation_enabled() ||
- isolate->bootstrapper()->IsActive()) {
- MarkForOptimization();
- return;
- }
- if (isolate->concurrent_osr_enabled() &&
- isolate->optimizing_compiler_thread()->IsQueuedForOSR(this)) {
- // Do not attempt regular recompilation if we already queued this for OSR.
- // TODO(yangguo): This is necessary so that we don't install optimized
- // code on a function that is already optimized, since OSR and regular
- // recompilation race. This goes away as soon as OSR becomes one-shot.
- return;
- }
- DCHECK(!IsInOptimizationQueue());
- DCHECK(is_compiled() || isolate->DebuggerHasBreakPoints());
+void JSFunction::MarkForConcurrentOptimization() {
+ DCHECK(is_compiled() || GetIsolate()->DebuggerHasBreakPoints());
DCHECK(!IsOptimized());
DCHECK(shared()->allows_lazy_compilation() || code()->optimizable());
DCHECK(!shared()->is_generator());
- DCHECK(isolate->concurrent_recompilation_enabled());
+ DCHECK(GetIsolate()->concurrent_recompilation_enabled());
if (FLAG_trace_concurrent_recompilation) {
PrintF(" ** Marking ");
ShortPrint();