Revert "Reland "Optimize function across closures.""
authoryangguo@chromium.org <yangguo@chromium.org>
Wed, 5 Nov 2014 11:37:02 +0000 (11:37 +0000)
committeryangguo@chromium.org <yangguo@chromium.org>
Wed, 5 Nov 2014 11:37:42 +0000 (11:37 +0000)
This reverts commit r25142.

TBR=ishell@chromium.org

Review URL: https://codereview.chromium.org/702853002

Cr-Commit-Position: refs/heads/master@{#25145}
git-svn-id: https://v8.googlecode.com/svn/branches/bleeding_edge@25145 ce2b1a6d-e550-0410-aec6-3dcde31c8c00

src/code-stubs-hydrogen.cc
src/compiler.cc
src/factory.cc
src/hydrogen-instructions.h
src/objects-inl.h
src/objects.cc
src/objects.h
src/runtime-profiler.cc
src/runtime/runtime-test.cc

index af4ee50..5579e05 100644 (file)
@@ -1549,67 +1549,47 @@ HValue* CodeStubGraphBuilder<FastNewClosureStub>::BuildCodeStub() {
 
   AddIncrementCounter(counters->fast_new_closure_total());
 
-  IfBuilder optimize_now(this);
-  HInstruction* compile_hint = Add<HLoadNamedField>(
-      shared_info, static_cast<HValue*>(NULL), HObjectAccess::ForCompileHint());
-  HValue* hint_mask = Add<HConstant>(
-      static_cast<int32_t>(1 << SharedFunctionInfo::kOptimizeNextClosure));
-  HInstruction* optimize =
-      AddUncasted<HBitwise>(Token::BIT_AND, compile_hint, hint_mask);
-  optimize_now.If<HCompareNumericAndBranch>(optimize, hint_mask, Token::EQ);
-  optimize_now.Then();
-  {
-    Add<HPushArguments>(context(), shared_info, graph()->GetConstantFalse());
-    Push(Add<HCallRuntime>(isolate()->factory()->empty_string(),
-                           Runtime::FunctionForId(Runtime::kNewClosure), 3));
-  }
-  optimize_now.Else();
-  {
-    // Create a new closure from the given function info in new space
-    HValue* size = Add<HConstant>(JSFunction::kSize);
-    HInstruction* js_function =
-        Add<HAllocate>(size, HType::JSObject(), NOT_TENURED, JS_FUNCTION_TYPE);
-
-    int map_index = Context::FunctionMapIndex(casted_stub()->strict_mode(),
-                                              casted_stub()->kind());
-
-    // Compute the function map in the current native context and set that
-    // as the map of the allocated object.
-    HInstruction* native_context = BuildGetNativeContext();
-    HInstruction* map_slot_value =
-        Add<HLoadNamedField>(native_context, static_cast<HValue*>(NULL),
-                             HObjectAccess::ForContextSlot(map_index));
-    Add<HStoreNamedField>(js_function, HObjectAccess::ForMap(), map_slot_value);
-
-    // Initialize the rest of the function.
-    Add<HStoreNamedField>(js_function, HObjectAccess::ForPropertiesPointer(),
-                          empty_fixed_array);
-    Add<HStoreNamedField>(js_function, HObjectAccess::ForElementsPointer(),
-                          empty_fixed_array);
-    Add<HStoreNamedField>(js_function, HObjectAccess::ForLiteralsPointer(),
-                          empty_fixed_array);
-    Add<HStoreNamedField>(js_function,
-                          HObjectAccess::ForPrototypeOrInitialMap(),
-                          graph()->GetConstantHole());
-    Add<HStoreNamedField>(js_function,
-                          HObjectAccess::ForSharedFunctionInfoPointer(),
-                          shared_info);
-    Add<HStoreNamedField>(
-        js_function, HObjectAccess::ForFunctionContextPointer(), context());
-
-    // Initialize the code pointer in the function to be the one
-    // found in the shared function info object.
-    // But first check if there is an optimized version for our context.
-    if (FLAG_cache_optimized_code) {
-      BuildInstallFromOptimizedCodeMap(js_function, shared_info,
-                                       native_context);
-    } else {
-      BuildInstallCode(js_function, shared_info);
-    }
-    Push(js_function);
+  // Create a new closure from the given function info in new space
+  HValue* size = Add<HConstant>(JSFunction::kSize);
+  HInstruction* js_function = Add<HAllocate>(size, HType::JSObject(),
+                                             NOT_TENURED, JS_FUNCTION_TYPE);
+
+  int map_index = Context::FunctionMapIndex(casted_stub()->strict_mode(),
+                                            casted_stub()->kind());
+
+  // Compute the function map in the current native context and set that
+  // as the map of the allocated object.
+  HInstruction* native_context = BuildGetNativeContext();
+  HInstruction* map_slot_value = Add<HLoadNamedField>(
+      native_context, static_cast<HValue*>(NULL),
+      HObjectAccess::ForContextSlot(map_index));
+  Add<HStoreNamedField>(js_function, HObjectAccess::ForMap(), map_slot_value);
+
+  // Initialize the rest of the function.
+  Add<HStoreNamedField>(js_function, HObjectAccess::ForPropertiesPointer(),
+                        empty_fixed_array);
+  Add<HStoreNamedField>(js_function, HObjectAccess::ForElementsPointer(),
+                        empty_fixed_array);
+  Add<HStoreNamedField>(js_function, HObjectAccess::ForLiteralsPointer(),
+                        empty_fixed_array);
+  Add<HStoreNamedField>(js_function, HObjectAccess::ForPrototypeOrInitialMap(),
+                        graph()->GetConstantHole());
+  Add<HStoreNamedField>(js_function,
+                        HObjectAccess::ForSharedFunctionInfoPointer(),
+                        shared_info);
+  Add<HStoreNamedField>(js_function, HObjectAccess::ForFunctionContextPointer(),
+                        context());
+
+  // Initialize the code pointer in the function to be the one
+  // found in the shared function info object.
+  // But first check if there is an optimized version for our context.
+  if (FLAG_cache_optimized_code) {
+    BuildInstallFromOptimizedCodeMap(js_function, shared_info, native_context);
+  } else {
+    BuildInstallCode(js_function, shared_info);
   }
-  optimize_now.End();
-  return Pop();
+
+  return js_function;
 }
 
 
index 2349640..ba5f3fd 100644 (file)
@@ -1361,7 +1361,6 @@ MaybeHandle<Code> Compiler::GetOptimizedCode(Handle<JSFunction> function,
   PostponeInterruptsScope postpone(isolate);
 
   Handle<SharedFunctionInfo> shared = info->shared_info();
-  shared->set_optimize_next_closure(false);
   if (shared->code()->kind() != Code::FUNCTION ||
       ScopeInfo::Empty(isolate) == shared->scope_info()) {
     // The function was never compiled. Compile it unoptimized first.
index c500212..72974a3 100644 (file)
@@ -1358,14 +1358,6 @@ Handle<JSObject> Factory::NewFunctionPrototype(Handle<JSFunction> function) {
 }
 
 
-static bool ShouldOptimizeNewClosure(Isolate* isolate,
-                                     Handle<SharedFunctionInfo> info) {
-  return isolate->use_crankshaft() && !info->is_toplevel() &&
-         info->is_compiled() && info->allows_lazy_compilation() &&
-         !info->optimization_disabled() && !isolate->DebuggerHasBreakPoints();
-}
-
-
 Handle<JSFunction> Factory::NewFunctionFromSharedFunctionInfo(
     Handle<SharedFunctionInfo> info,
     Handle<Context> context,
@@ -1403,11 +1395,14 @@ Handle<JSFunction> Factory::NewFunctionFromSharedFunctionInfo(
     return result;
   }
 
-  if (FLAG_always_opt && ShouldOptimizeNewClosure(isolate(), info)) {
+  if (isolate()->use_crankshaft() &&
+      FLAG_always_opt &&
+      result->is_compiled() &&
+      !info->is_toplevel() &&
+      info->allows_lazy_compilation() &&
+      !info->optimization_disabled() &&
+      !isolate()->DebuggerHasBreakPoints()) {
     result->MarkForOptimization();
-  } else if (info->optimize_next_closure() &&
-             ShouldOptimizeNewClosure(isolate(), info)) {
-    result->AttemptConcurrentOptimization();
   }
   return result;
 }
index f590381..810a2fd 100644 (file)
@@ -6121,11 +6121,6 @@ class HObjectAccess FINAL {
                          SharedFunctionInfo::kOptimizedCodeMapOffset);
   }
 
-  static HObjectAccess ForCompileHint() {
-    return HObjectAccess(kInobject, SharedFunctionInfo::kCompilerHintsOffset,
-                         Representation::Smi());
-  }
-
   static HObjectAccess ForFunctionContextPointer() {
     return HObjectAccess(kInobject, JSFunction::kContextOffset);
   }
index ff81d9d..85584c4 100644 (file)
@@ -5516,9 +5516,9 @@ BOOL_ACCESSORS(SharedFunctionInfo, start_position_and_type, is_expression,
 BOOL_ACCESSORS(SharedFunctionInfo, start_position_and_type, is_toplevel,
                kIsTopLevelBit)
 
-BOOL_ACCESSORS(SharedFunctionInfo, compiler_hints, optimize_next_closure,
-               kOptimizeNextClosure)
-BOOL_ACCESSORS(SharedFunctionInfo, compiler_hints, allows_lazy_compilation,
+BOOL_ACCESSORS(SharedFunctionInfo,
+               compiler_hints,
+               allows_lazy_compilation,
                kAllowLazyCompilation)
 BOOL_ACCESSORS(SharedFunctionInfo,
                compiler_hints,
index 6e68c36..258390c 100644 (file)
@@ -9204,27 +9204,12 @@ void JSFunction::MarkForOptimization() {
 }
 
 
-void JSFunction::AttemptConcurrentOptimization() {
-  Isolate* isolate = GetIsolate();
-  if (!isolate->concurrent_recompilation_enabled() ||
-      isolate->bootstrapper()->IsActive()) {
-    MarkForOptimization();
-    return;
-  }
-  if (isolate->concurrent_osr_enabled() &&
-      isolate->optimizing_compiler_thread()->IsQueuedForOSR(this)) {
-    // Do not attempt regular recompilation if we already queued this for OSR.
-    // TODO(yangguo): This is necessary so that we don't install optimized
-    // code on a function that is already optimized, since OSR and regular
-    // recompilation race.  This goes away as soon as OSR becomes one-shot.
-    return;
-  }
-  DCHECK(!IsInOptimizationQueue());
-  DCHECK(is_compiled() || isolate->DebuggerHasBreakPoints());
+void JSFunction::MarkForConcurrentOptimization() {
+  DCHECK(is_compiled() || GetIsolate()->DebuggerHasBreakPoints());
   DCHECK(!IsOptimized());
   DCHECK(shared()->allows_lazy_compilation() || code()->optimizable());
   DCHECK(!shared()->is_generator());
-  DCHECK(isolate->concurrent_recompilation_enabled());
+  DCHECK(GetIsolate()->concurrent_recompilation_enabled());
   if (FLAG_trace_concurrent_recompilation) {
     PrintF("  ** Marking ");
     ShortPrint();
index d412005..9333e9e 100644 (file)
@@ -6787,8 +6787,6 @@ class SharedFunctionInfo: public HeapObject {
   inline int ic_age();
   inline void set_ic_age(int age);
 
-  DECL_BOOLEAN_ACCESSORS(optimize_next_closure)
-
   // Indicates if this function can be lazy compiled.
   // This is used to determine if we can safely flush code from a function
   // when doing GC if we expect that the function will no longer be used.
@@ -7079,7 +7077,6 @@ class SharedFunctionInfo: public HeapObject {
   enum CompilerHints {
     kAllowLazyCompilation,
     kAllowLazyCompilationWithoutContext,
-    kOptimizeNextClosure,
     kOptimizationDisabled,
     kStrictModeFunction,
     kUsesArguments,
@@ -7305,7 +7302,7 @@ class JSFunction: public JSObject {
   // Mark this function for lazy recompilation. The function will be
   // recompiled the next time it is executed.
   void MarkForOptimization();
-  void AttemptConcurrentOptimization();
+  void MarkForConcurrentOptimization();
   void MarkInOptimizationQueue();
 
   // Tells whether or not the function is already marked for lazy
index 7330262..f786846 100644 (file)
@@ -106,8 +106,23 @@ void RuntimeProfiler::Optimize(JSFunction* function, const char* reason) {
     PrintF("]\n");
   }
 
-  function->shared()->set_optimize_next_closure(true);
-  function->AttemptConcurrentOptimization();
+
+  if (isolate_->concurrent_recompilation_enabled() &&
+      !isolate_->bootstrapper()->IsActive()) {
+    if (isolate_->concurrent_osr_enabled() &&
+        isolate_->optimizing_compiler_thread()->IsQueuedForOSR(function)) {
+      // Do not attempt regular recompilation if we already queued this for OSR.
+      // TODO(yangguo): This is necessary so that we don't install optimized
+      // code on a function that is already optimized, since OSR and regular
+      // recompilation race.  This goes away as soon as OSR becomes one-shot.
+      return;
+    }
+    DCHECK(!function->IsInOptimizationQueue());
+    function->MarkForConcurrentOptimization();
+  } else {
+    // The next call to the function will trigger optimization.
+    function->MarkForOptimization();
+  }
 }
 
 
index 0a72213..4995984 100644 (file)
@@ -75,7 +75,7 @@ RUNTIME_FUNCTION(Runtime_OptimizeFunctionOnNextCall) {
           *function, Code::kMaxLoopNestingMarker);
     } else if (type->IsOneByteEqualTo(STATIC_CHAR_VECTOR("concurrent")) &&
                isolate->concurrent_recompilation_enabled()) {
-      function->AttemptConcurrentOptimization();
+      function->MarkForConcurrentOptimization();
     }
   }