CallDescriptor* Linkage::GetStubCallDescriptor(
- CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count) {
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize, Zone* zone) {
return LinkageHelper::GetStubCallDescriptor<LinkageHelperTraits>(
- this->info_->zone(), descriptor, stack_parameter_count);
+ zone, descriptor, stack_parameter_count, can_deoptimize);
}
CallDescriptor* Linkage::GetStubCallDescriptor(
- CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count) {
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize, Zone* zone) {
return LinkageHelper::GetStubCallDescriptor<LinkageHelperTraits>(
- this->info_->zone(), descriptor, stack_parameter_count);
+ zone, descriptor, stack_parameter_count, can_deoptimize);
}
AstGraphBuilder::AstContext::AstContext(AstGraphBuilder* own,
- Expression::Context kind)
- : kind_(kind), owner_(own), outer_(own->ast_context()) {
+ Expression::Context kind,
+ BailoutId bailout_id)
+ : bailout_id_(bailout_id),
+ kind_(kind),
+ owner_(own),
+ outer_(own->ast_context()) {
owner()->set_ast_context(this); // Push.
#ifdef DEBUG
original_height_ = environment()->stack_height();
}
+void AstGraphBuilder::AstEffectContext::ProduceValueWithLazyBailout(
+ Node* value) {
+ ProduceValue(value);
+ owner()->BuildLazyBailout(value, bailout_id_);
+}
+
+
+void AstGraphBuilder::AstValueContext::ProduceValueWithLazyBailout(
+ Node* value) {
+ ProduceValue(value);
+ owner()->BuildLazyBailout(value, bailout_id_);
+}
+
+
+void AstGraphBuilder::AstTestContext::ProduceValueWithLazyBailout(Node* value) {
+ environment()->Push(value);
+ owner()->BuildLazyBailout(value, bailout_id_);
+ environment()->Pop();
+ ProduceValue(value);
+}
+
+
void AstGraphBuilder::AstEffectContext::ProduceValue(Node* value) {
// The value is ignored.
}
void AstGraphBuilder::VisitForValue(Expression* expr) {
- AstValueContext for_value(this);
+ AstValueContext for_value(this, expr->id());
if (!HasStackOverflow()) {
expr->Accept(this);
}
void AstGraphBuilder::VisitForEffect(Expression* expr) {
- AstEffectContext for_effect(this);
+ AstEffectContext for_effect(this, expr->id());
if (!HasStackOverflow()) {
expr->Accept(this);
}
void AstGraphBuilder::VisitForTest(Expression* expr) {
- AstTestContext for_condition(this);
+ AstTestContext for_condition(this, expr->id());
if (!HasStackOverflow()) {
expr->Accept(this);
}
Node* index = environment()->Peek(0);
Node* exit_cond =
NewNode(javascript()->LessThan(), index, cache_length);
+ // TODO(jarin): provide real bailout id.
+ BuildLazyBailout(exit_cond, BailoutId::None());
for_loop.BreakUnless(exit_cond);
// TODO(dcarney): this runtime call should be a handful of
// simplified instructions that
// is gone.
Node* res = ProcessArguments(
javascript()->Call(3, NO_CALL_FUNCTION_FLAGS), 3);
+ // TODO(jarin): provide real bailout id.
+ BuildLazyBailout(res, BailoutId::None());
Node* property_missing = NewNode(javascript()->StrictEqual(), res,
jsgraph()->ZeroConstant());
{
void AstGraphBuilder::VisitVariableProxy(VariableProxy* expr) {
- Node* value = BuildVariableLoad(expr->var());
+ Node* value = BuildVariableLoad(expr->var(), expr->id());
ast_context()->ProduceValue(value);
}
VisitForValue(property->value());
Node* value = environment()->Pop();
PrintableUnique<Name> name = MakeUnique(key->AsPropertyName());
- NewNode(javascript()->StoreNamed(name), literal, value);
+ Node* store =
+ NewNode(javascript()->StoreNamed(name), literal, value);
+ BuildLazyBailout(store, key->id());
} else {
VisitForEffect(property->value());
}
VisitForValue(subexpr);
Node* value = environment()->Pop();
Node* index = jsgraph()->Constant(i);
- NewNode(javascript()->StoreProperty(), literal, index, value);
+ Node* store = NewNode(javascript()->StoreProperty(), literal, index, value);
+ BuildLazyBailout(store, expr->GetIdForElement(i));
}
environment()->Pop(); // Array literal index.
switch (assign_type) {
case VARIABLE: {
Variable* var = expr->AsVariableProxy()->var();
- BuildVariableAssignment(var, value, Token::ASSIGN);
+ // TODO(jarin) Fill in the correct bailout id.
+ BuildVariableAssignment(var, value, Token::ASSIGN, BailoutId::None());
break;
}
case NAMED_PROPERTY: {
value = environment()->Pop();
PrintableUnique<Name> name =
MakeUnique(property->key()->AsLiteral()->AsPropertyName());
- NewNode(javascript()->StoreNamed(name), object, value);
+ Node* store = NewNode(javascript()->StoreNamed(name), object, value);
+ // TODO(jarin) Fill in the correct bailout id.
+ BuildLazyBailout(store, BailoutId::None());
break;
}
case KEYED_PROPERTY: {
Node* key = environment()->Pop();
Node* object = environment()->Pop();
value = environment()->Pop();
- NewNode(javascript()->StoreProperty(), object, key, value);
+ Node* store = NewNode(javascript()->StoreProperty(), object, key, value);
+ // TODO(jarin) Fill in the correct bailout id.
+ BuildLazyBailout(store, BailoutId::None());
break;
}
}
switch (assign_type) {
case VARIABLE: {
Variable* variable = expr->target()->AsVariableProxy()->var();
- old_value = BuildVariableLoad(variable);
+ old_value = BuildVariableLoad(variable, expr->target()->id());
break;
}
case NAMED_PROPERTY: {
PrintableUnique<Name> name =
MakeUnique(property->key()->AsLiteral()->AsPropertyName());
old_value = NewNode(javascript()->LoadNamed(name), object);
+ BuildLazyBailoutWithPushedNode(old_value, property->LoadId());
break;
}
case KEYED_PROPERTY: {
Node* key = environment()->Top();
Node* object = environment()->Peek(1);
old_value = NewNode(javascript()->LoadProperty(), object, key);
+ BuildLazyBailoutWithPushedNode(old_value, property->LoadId());
break;
}
}
Node* left = environment()->Pop();
Node* value = BuildBinaryOp(left, right, expr->binary_op());
environment()->Push(value);
+ BuildLazyBailout(value, expr->binary_operation()->id());
} else {
VisitForValue(expr->value());
}
switch (assign_type) {
case VARIABLE: {
Variable* variable = expr->target()->AsVariableProxy()->var();
- BuildVariableAssignment(variable, value, expr->op());
+ BuildVariableAssignment(variable, value, expr->op(),
+ expr->AssignmentId());
break;
}
case NAMED_PROPERTY: {
Node* object = environment()->Pop();
PrintableUnique<Name> name =
MakeUnique(property->key()->AsLiteral()->AsPropertyName());
- NewNode(javascript()->StoreNamed(name), object, value);
+ Node* store = NewNode(javascript()->StoreNamed(name), object, value);
+ BuildLazyBailout(store, expr->AssignmentId());
break;
}
case KEYED_PROPERTY: {
Node* key = environment()->Pop();
Node* object = environment()->Pop();
- NewNode(javascript()->StoreProperty(), object, key, value);
+ Node* store = NewNode(javascript()->StoreProperty(), object, key, value);
+ BuildLazyBailout(store, expr->AssignmentId());
break;
}
}
Node* object = environment()->Pop();
value = NewNode(javascript()->LoadProperty(), object, key);
}
- ast_context()->ProduceValue(value);
+ ast_context()->ProduceValueWithLazyBailout(value);
}
switch (call_type) {
case Call::GLOBAL_CALL: {
Variable* variable = callee->AsVariableProxy()->var();
- callee_value = BuildVariableLoad(variable);
+ callee_value = BuildVariableLoad(variable, expr->expression()->id());
receiver_value = jsgraph()->UndefinedConstant();
break;
}
Node* key = environment()->Pop();
callee_value = NewNode(javascript()->LoadProperty(), object, key);
}
+ BuildLazyBailoutWithPushedNode(callee_value, property->LoadId());
receiver_value = environment()->Pop();
// Note that a PROPERTY_CALL requires the receiver to be wrapped into an
// object for sloppy callees. This could also be modeled explicitly here,
// Create node to perform the function call.
Operator* call = javascript()->Call(args->length() + 2, flags);
Node* value = ProcessArguments(call, args->length() + 2);
- ast_context()->ProduceValue(value);
+ ast_context()->ProduceValueWithLazyBailout(value);
}
// Create node to perform the construct call.
Operator* call = javascript()->CallNew(args->length() + 1);
Node* value = ProcessArguments(call, args->length() + 1);
- ast_context()->ProduceValue(value);
+ ast_context()->ProduceValueWithLazyBailout(value);
}
PrintableUnique<String> unique = MakeUnique(name);
Node* callee_value = NewNode(javascript()->LoadNamed(unique), receiver_value);
environment()->Push(callee_value);
+ // TODO(jarin): Find/create a bailout id to deoptimize to (crankshaft
+ // refuses to optimize functions with jsruntime calls).
+ BuildLazyBailout(callee_value, BailoutId::None());
environment()->Push(receiver_value);
// Evaluate all arguments to the JS runtime call.
// Create node to perform the JS runtime call.
Operator* call = javascript()->Call(args->length() + 2, flags);
Node* value = ProcessArguments(call, args->length() + 2);
- ast_context()->ProduceValue(value);
+ ast_context()->ProduceValueWithLazyBailout(value);
}
Runtime::FunctionId functionId = function->function_id;
Operator* call = javascript()->Runtime(functionId, args->length());
Node* value = ProcessArguments(call, args->length());
- ast_context()->ProduceValue(value);
-
- BuildLazyBailout(value, expr->id());
+ ast_context()->ProduceValueWithLazyBailout(value);
}
switch (assign_type) {
case VARIABLE: {
Variable* variable = expr->expression()->AsVariableProxy()->var();
- old_value = BuildVariableLoad(variable);
+ old_value = BuildVariableLoad(variable, expr->expression()->id());
stack_depth = 0;
break;
}
PrintableUnique<Name> name =
MakeUnique(property->key()->AsLiteral()->AsPropertyName());
old_value = NewNode(javascript()->LoadNamed(name), object);
+ BuildLazyBailoutWithPushedNode(old_value, property->LoadId());
stack_depth = 1;
break;
}
Node* key = environment()->Top();
Node* object = environment()->Peek(1);
old_value = NewNode(javascript()->LoadProperty(), object, key);
+ BuildLazyBailoutWithPushedNode(old_value, property->LoadId());
stack_depth = 2;
break;
}
// Create node to perform +1/-1 operation.
Node* value =
BuildBinaryOp(old_value, jsgraph()->OneConstant(), expr->binary_op());
+ // TODO(jarin) Insert proper bailout id here (will need to change
+ // full code generator).
+ BuildLazyBailout(value, BailoutId::None());
// Store the value.
switch (assign_type) {
case VARIABLE: {
Variable* variable = expr->expression()->AsVariableProxy()->var();
- BuildVariableAssignment(variable, value, expr->op());
+ BuildVariableAssignment(variable, value, expr->op(),
+ expr->AssignmentId());
break;
}
case NAMED_PROPERTY: {
Node* object = environment()->Pop();
PrintableUnique<Name> name =
MakeUnique(property->key()->AsLiteral()->AsPropertyName());
- NewNode(javascript()->StoreNamed(name), object, value);
+ Node* store = NewNode(javascript()->StoreNamed(name), object, value);
+ BuildLazyBailout(store, expr->AssignmentId());
break;
}
case KEYED_PROPERTY: {
Node* key = environment()->Pop();
Node* object = environment()->Pop();
- NewNode(javascript()->StoreProperty(), object, key, value);
+ Node* store = NewNode(javascript()->StoreProperty(), object, key, value);
+ BuildLazyBailout(store, expr->AssignmentId());
break;
}
}
Node* right = environment()->Pop();
Node* left = environment()->Pop();
Node* value = BuildBinaryOp(left, right, expr->op());
- ast_context()->ProduceValue(value);
+ ast_context()->ProduceValueWithLazyBailout(value);
}
}
}
Node* left = environment()->Pop();
Node* value = NewNode(op, left, right);
ast_context()->ProduceValue(value);
+
+ BuildLazyBailout(value, expr->id());
}
// Typeof does not throw a reference error on global variables, hence we
// perform a non-contextual load in case the operand is a variable proxy.
Variable* variable = expr->expression()->AsVariableProxy()->var();
- operand = BuildVariableLoad(variable, NOT_CONTEXTUAL);
+ operand =
+ BuildVariableLoad(variable, expr->expression()->id(), NOT_CONTEXTUAL);
} else {
VisitForValue(expr->expression());
operand = environment()->Pop();
// Assign the object to the arguments variable.
DCHECK(arguments->IsContextSlot() || arguments->IsStackAllocated());
- BuildVariableAssignment(arguments, object, Token::ASSIGN);
+ // This should never lazy deopt, so it is fine to send invalid bailout id.
+ BuildVariableAssignment(arguments, object, Token::ASSIGN, BailoutId::None());
return object;
}
Node* AstGraphBuilder::BuildVariableLoad(Variable* variable,
+ BailoutId bailout_id,
ContextualMode contextual_mode) {
Node* the_hole = jsgraph()->TheHoleConstant();
VariableMode mode = variable->mode();
Node* global = BuildLoadGlobalObject();
PrintableUnique<Name> name = MakeUnique(variable->name());
Operator* op = javascript()->LoadNamed(name, contextual_mode);
- return NewNode(op, global);
+ Node* node = NewNode(op, global);
+ BuildLazyBailoutWithPushedNode(node, bailout_id);
+ return node;
}
case Variable::PARAMETER:
case Variable::LOCAL: {
Node* AstGraphBuilder::BuildVariableAssignment(Variable* variable, Node* value,
- Token::Value op) {
+ Token::Value op,
+ BailoutId bailout_id) {
Node* the_hole = jsgraph()->TheHoleConstant();
VariableMode mode = variable->mode();
switch (variable->location()) {
Node* global = BuildLoadGlobalObject();
PrintableUnique<Name> name = MakeUnique(variable->name());
Operator* op = javascript()->StoreNamed(name);
- return NewNode(op, global, value);
+ Node* store = NewNode(op, global, value);
+ BuildLazyBailout(store, bailout_id);
+ return store;
}
case Variable::PARAMETER:
case Variable::LOCAL:
NewNode(common()->LazyDeoptimization());
+ // TODO(jarin) If ast_id.IsNone(), perhaps we should generate an empty
+ // deopt block and make sure there is no patch entry for this (so
+ // that the deoptimizer dies when trying to deoptimize here).
+
Node* state_node = environment()->Checkpoint(ast_id);
Node* deoptimize_node = NewNode(common()->Deoptimize(), state_node);
NewNode(common()->Continuation());
}
}
+
+
+void AstGraphBuilder::BuildLazyBailoutWithPushedNode(Node* node,
+ BailoutId ast_id) {
+ environment()->Push(node);
+ BuildLazyBailout(node, ast_id);
+ environment()->Pop();
+}
}
}
} // namespace v8::internal::compiler
Node* BuildArgumentsObject(Variable* arguments);
// Builders for variable load and assignment.
- Node* BuildVariableAssignment(Variable* var, Node* value, Token::Value op);
+ Node* BuildVariableAssignment(Variable* var, Node* value, Token::Value op,
+ BailoutId bailout_id);
Node* BuildVariableDelete(Variable* var);
- Node* BuildVariableLoad(Variable* var, ContextualMode mode = CONTEXTUAL);
+ Node* BuildVariableLoad(Variable* var, BailoutId bailout_id,
+ ContextualMode mode = CONTEXTUAL);
// Builders for accessing the function context.
Node* BuildLoadBuiltinsObject();
void VisitForInAssignment(Expression* expr, Node* value);
void BuildLazyBailout(Node* node, BailoutId ast_id);
+ void BuildLazyBailoutWithPushedNode(Node* node, BailoutId ast_id);
DEFINE_AST_VISITOR_SUBCLASS_MEMBERS();
DISALLOW_COPY_AND_ASSIGN(AstGraphBuilder);
// Plug a node into this expression context. Call this function in tail
// position in the Visit functions for expressions.
virtual void ProduceValue(Node* value) = 0;
+ virtual void ProduceValueWithLazyBailout(Node* value) = 0;
// Unplugs a node from this expression context. Call this to retrieve the
// result of another Visit function that already plugged the context.
void ReplaceValue() { ProduceValue(ConsumeValue()); }
protected:
- AstContext(AstGraphBuilder* owner, Expression::Context kind);
+ AstContext(AstGraphBuilder* owner, Expression::Context kind,
+ BailoutId bailout_id);
virtual ~AstContext();
AstGraphBuilder* owner() const { return owner_; }
int original_height_;
#endif
+ BailoutId bailout_id_;
+
private:
Expression::Context kind_;
AstGraphBuilder* owner_;
// Context to evaluate expression for its side effects only.
class AstGraphBuilder::AstEffectContext V8_FINAL : public AstContext {
public:
- explicit AstEffectContext(AstGraphBuilder* owner)
- : AstContext(owner, Expression::kEffect) {}
+ explicit AstEffectContext(AstGraphBuilder* owner, BailoutId bailout_id)
+ : AstContext(owner, Expression::kEffect, bailout_id) {}
virtual ~AstEffectContext();
virtual void ProduceValue(Node* value) V8_OVERRIDE;
+ virtual void ProduceValueWithLazyBailout(Node* value) V8_OVERRIDE;
virtual Node* ConsumeValue() V8_OVERRIDE;
};
// Context to evaluate expression for its value (and side effects).
class AstGraphBuilder::AstValueContext V8_FINAL : public AstContext {
public:
- explicit AstValueContext(AstGraphBuilder* owner)
- : AstContext(owner, Expression::kValue) {}
+ explicit AstValueContext(AstGraphBuilder* owner, BailoutId bailout_id)
+ : AstContext(owner, Expression::kValue, bailout_id) {}
virtual ~AstValueContext();
virtual void ProduceValue(Node* value) V8_OVERRIDE;
+ virtual void ProduceValueWithLazyBailout(Node* value) V8_OVERRIDE;
virtual Node* ConsumeValue() V8_OVERRIDE;
};
// Context to evaluate expression for a condition value (and side effects).
class AstGraphBuilder::AstTestContext V8_FINAL : public AstContext {
public:
- explicit AstTestContext(AstGraphBuilder* owner)
- : AstContext(owner, Expression::kTest) {}
+ explicit AstTestContext(AstGraphBuilder* owner, BailoutId bailout_id)
+ : AstContext(owner, Expression::kTest, bailout_id) {}
virtual ~AstTestContext();
virtual void ProduceValue(Node* value) V8_OVERRIDE;
+ virtual void ProduceValueWithLazyBailout(Node* value) V8_OVERRIDE;
virtual Node* ConsumeValue() V8_OVERRIDE;
};
for (int i = 0; i < deopt_count; i++) {
FrameStateDescriptor* descriptor = code()->GetDeoptimizationEntry(i);
data->SetAstId(i, descriptor->bailout_id());
- data->SetTranslationIndex(i, Smi::FromInt(0));
+ CHECK_NE(NULL, deoptimization_states_[i]);
+ data->SetTranslationIndex(
+ i, Smi::FromInt(deoptimization_states_[i]->translation_id_));
data->SetArgumentsStackHeight(i, Smi::FromInt(0));
data->SetPc(i, Smi::FromInt(-1));
}
CallDescriptor* Linkage::GetStubCallDescriptor(
- CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count) {
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize, Zone* zone) {
return LinkageHelper::GetStubCallDescriptor<LinkageHelperTraits>(
- this->info_->zone(), descriptor, stack_parameter_count);
+ zone, descriptor, stack_parameter_count, can_deoptimize);
}
#undef REPLACE_UNIMPLEMENTED
+static CallDescriptor::DeoptimizationSupport DeoptimizationSupportForNode(
+ Node* node) {
+ return OperatorProperties::CanLazilyDeoptimize(node->op())
+ ? CallDescriptor::kCanDeoptimize
+ : CallDescriptor::kCannotDeoptimize;
+}
+
+
void JSGenericLowering::ReplaceWithCompareIC(Node* node, Token::Value token,
bool pure) {
BinaryOpICStub stub(isolate(), Token::ADD); // TODO(mstarzinger): Hack.
void JSGenericLowering::ReplaceWithICStubCall(Node* node,
HydrogenCodeStub* stub) {
CodeStubInterfaceDescriptor* d = stub->GetInterfaceDescriptor();
- CallDescriptor* desc = linkage()->GetStubCallDescriptor(d);
+ CallDescriptor* desc = linkage()->GetStubCallDescriptor(
+ d, 0, DeoptimizationSupportForNode(node));
Node* stub_code = CodeConstant(stub->GetCode());
PatchInsertInput(node, 0, stub_code);
PatchOperator(node, common()->Call(desc));
Operator::Property props = node->op()->properties();
const Runtime::Function* fun = Runtime::FunctionForId(f);
int nargs = (nargs_override < 0) ? fun->nargs : nargs_override;
- CallDescriptor::DeoptimizationSupport deopt =
- OperatorProperties::CanLazilyDeoptimize(node->op())
- ? CallDescriptor::kCanDeoptimize
- : CallDescriptor::kCannotDeoptimize;
- CallDescriptor* desc =
- linkage()->GetRuntimeCallDescriptor(f, nargs, props, deopt);
+ CallDescriptor* desc = linkage()->GetRuntimeCallDescriptor(
+ f, nargs, props, DeoptimizationSupportForNode(node));
Node* ref = ExternalConstant(ExternalReference(f, isolate()));
Node* arity = Int32Constant(nargs);
if (!centrystub_constant_.is_set()) {
int arity = OpParameter<int>(node);
CallConstructStub stub(isolate(), NO_CALL_CONSTRUCTOR_FLAGS);
CodeStubInterfaceDescriptor* d = GetInterfaceDescriptor(isolate(), &stub);
- CallDescriptor* desc = linkage()->GetStubCallDescriptor(d, arity);
+ CallDescriptor* desc = linkage()->GetStubCallDescriptor(
+ d, arity, DeoptimizationSupportForNode(node));
Node* stub_code = CodeConstant(stub.GetCode());
Node* construct = NodeProperties::GetValueInput(node, 0);
PatchInsertInput(node, 0, stub_code);
CallParameters p = OpParameter<CallParameters>(node);
CallFunctionStub stub(isolate(), p.arity - 2, p.flags);
CodeStubInterfaceDescriptor* d = GetInterfaceDescriptor(isolate(), &stub);
- CallDescriptor* desc = linkage()->GetStubCallDescriptor(d, p.arity - 1);
+ CallDescriptor* desc = linkage()->GetStubCallDescriptor(
+ d, p.arity - 1, DeoptimizationSupportForNode(node));
Node* stub_code = CodeConstant(stub.GetCode());
PatchInsertInput(node, 0, stub_code);
PatchOperator(node, common()->Call(desc));
#ifndef V8_COMPILER_JS_OPERATOR_H_
#define V8_COMPILER_JS_OPERATOR_H_
+#include "src/compiler/linkage.h"
#include "src/compiler/opcodes.h"
#include "src/compiler/operator.h"
#include "src/unique.h"
template <typename LinkageTraits>
static CallDescriptor* GetStubCallDescriptor(
Zone* zone, CodeStubInterfaceDescriptor* descriptor,
- int stack_parameter_count) {
+ int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize) {
int register_parameter_count = descriptor->GetEnvironmentParameterCount();
int parameter_count = register_parameter_count + stack_parameter_count;
const int code_count = 1;
locations, // locations
Operator::kNoProperties, // properties
kNoCalleeSaved, // callee-saved registers
- CallDescriptor::kCannotDeoptimize, // deoptimization
+ can_deoptimize, // deoptimization
CodeStub::MajorName(descriptor->MajorKey(), false));
- // TODO(jarin) should deoptimize!
}
}
+CallDescriptor* Linkage::GetStubCallDescriptor(
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize) {
+ return GetStubCallDescriptor(descriptor, stack_parameter_count,
+ can_deoptimize, this->info_->zone());
+}
+
+
//==============================================================================
// Provide unimplemented methods on unsupported architectures, to at least link.
//==============================================================================
CallDescriptor* Linkage::GetStubCallDescriptor(
- CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count) {
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize, Zone* zone) {
UNIMPLEMENTED();
return NULL;
}
Operator::Property properties,
CallDescriptor::DeoptimizationSupport can_deoptimize, Zone* zone);
- CallDescriptor* GetStubCallDescriptor(CodeStubInterfaceDescriptor* descriptor,
- int stack_parameter_count = 0);
+ CallDescriptor* GetStubCallDescriptor(
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count = 0,
+ CallDescriptor::DeoptimizationSupport can_deoptimize =
+ CallDescriptor::kCannotDeoptimize);
+ static CallDescriptor* GetStubCallDescriptor(
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize, Zone* zone);
// Creates a call descriptor for simplified C calls that is appropriate
// for the host platform. This simplified calling convention only supports
#include "src/v8.h"
+#include "src/compiler/js-operator.h"
#include "src/compiler/opcodes.h"
#include "src/compiler/operator-properties.h"
#undef OPCODE_CASE
return static_cast<ControlOperator*>(op)->ControlInputCount();
default:
+ // If a node can lazily deoptimize, it needs control dependency.
+ if (CanLazilyDeoptimize(op)) {
+ return 1;
+ }
// Operators that have write effects must have a control
// dependency. Effect dependencies only ensure the correct order of
// write/read operations without consideration of control flow. Without an
}
inline bool OperatorProperties::CanLazilyDeoptimize(Operator* op) {
- if (op->opcode() == IrOpcode::kCall) {
- CallOperator* call_op = reinterpret_cast<CallOperator*>(op);
- CallDescriptor* descriptor = call_op->parameter();
- return descriptor->CanLazilyDeoptimize();
+ // TODO(jarin) This function allows turning on lazy deoptimization
+ // incrementally. It will change as we turn on lazy deopt for
+ // more nodes.
+
+ if (!FLAG_turbo_deoptimization) {
+ return false;
}
- if (op->opcode() == IrOpcode::kJSCallRuntime) {
- // TODO(jarin) At the moment, we only support lazy deoptimization for
- // the %DeoptimizeFunction runtime function.
- Runtime::FunctionId function =
- reinterpret_cast<Operator1<Runtime::FunctionId>*>(op)->parameter();
- return function == Runtime::kDeoptimizeFunction;
+
+ switch (op->opcode()) {
+ case IrOpcode::kCall: {
+ CallOperator* call_op = reinterpret_cast<CallOperator*>(op);
+ CallDescriptor* descriptor = call_op->parameter();
+ return descriptor->CanLazilyDeoptimize();
+ }
+ case IrOpcode::kJSCallRuntime: {
+ Runtime::FunctionId function =
+ reinterpret_cast<Operator1<Runtime::FunctionId>*>(op)->parameter();
+ // TODO(jarin) At the moment, we only support lazy deoptimization for
+ // the %DeoptimizeFunction runtime function.
+ return function == Runtime::kDeoptimizeFunction;
+ }
+
+ // JS function calls
+ case IrOpcode::kJSCallFunction:
+ case IrOpcode::kJSCallConstruct:
+
+ // Binary operations
+ case IrOpcode::kJSBitwiseOr:
+ case IrOpcode::kJSBitwiseXor:
+ case IrOpcode::kJSBitwiseAnd:
+ case IrOpcode::kJSShiftLeft:
+ case IrOpcode::kJSShiftRight:
+ case IrOpcode::kJSShiftRightLogical:
+ case IrOpcode::kJSAdd:
+ case IrOpcode::kJSSubtract:
+ case IrOpcode::kJSMultiply:
+ case IrOpcode::kJSDivide:
+ case IrOpcode::kJSModulus:
+ case IrOpcode::kJSLoadProperty:
+ case IrOpcode::kJSStoreProperty:
+ case IrOpcode::kJSLoadNamed:
+ case IrOpcode::kJSStoreNamed:
+ return true;
+
+ default:
+ return false;
}
return false;
}
if (!AllocationOk()) return;
}
}
+
+ // Meet register constraints for the instruction in the end.
+ if (!code()->IsGapAt(end)) {
+ MeetRegisterConstraintsForLastInstructionInBlock(block);
+ }
+}
+
+
+void RegisterAllocator::MeetRegisterConstraintsForLastInstructionInBlock(
+ BasicBlock* block) {
+ int end = block->last_instruction_index();
+ Instruction* last_instruction = InstructionAt(end);
+ for (size_t i = 0; i < last_instruction->OutputCount(); i++) {
+ InstructionOperand* output_operand = last_instruction->OutputAt(i);
+ DCHECK(!output_operand->IsConstant());
+ UnallocatedOperand* output = UnallocatedOperand::cast(output_operand);
+ int output_vreg = output->virtual_register();
+ LiveRange* range = LiveRangeFor(output_vreg);
+ bool assigned = false;
+ if (output->HasFixedPolicy()) {
+ AllocateFixed(output, -1, false);
+ // This value is produced on the stack, we never need to spill it.
+ if (output->IsStackSlot()) {
+ range->SetSpillOperand(output);
+ range->SetSpillStartIndex(end);
+ assigned = true;
+ }
+
+ BasicBlock::Successors successors = block->successors();
+ for (BasicBlock::Successors::iterator succ = successors.begin();
+ succ != successors.end(); ++succ) {
+ DCHECK((*succ)->PredecessorCount() == 1);
+ int gap_index = (*succ)->first_instruction_index() + 1;
+ DCHECK(code()->IsGapAt(gap_index));
+
+ // Create an unconstrained operand for the same virtual register
+ // and insert a gap move from the fixed output to the operand.
+ UnallocatedOperand* output_copy =
+ new (code_zone()) UnallocatedOperand(UnallocatedOperand::ANY);
+ output_copy->set_virtual_register(output_vreg);
+
+ code()->AddGapMove(gap_index, output, output_copy);
+ }
+ }
+
+ if (!assigned) {
+ BasicBlock::Successors successors = block->successors();
+ for (BasicBlock::Successors::iterator succ = successors.begin();
+ succ != successors.end(); ++succ) {
+ DCHECK((*succ)->PredecessorCount() == 1);
+ int gap_index = (*succ)->first_instruction_index() + 1;
+ range->SetSpillStartIndex(gap_index);
+
+ // This move to spill operand is not a real use. Liveness analysis
+ // and splitting of live ranges do not account for it.
+ // Thus it should be inserted to a lifetime position corresponding to
+ // the instruction end.
+ GapInstruction* gap = code()->GapAt(gap_index);
+ ParallelMove* move =
+ gap->GetOrCreateParallelMove(GapInstruction::BEFORE, code_zone());
+ move->AddMove(output, range->GetSpillOperand(), code_zone());
+ }
+ }
+ }
}
code()->AddGapMove(gap_index, first_output, output_copy);
}
+ // Make sure we add a gap move for spilling (if we have not done
+ // so already).
if (!assigned) {
range->SetSpillStartIndex(gap_index);
void MeetRegisterConstraints(BasicBlock* block);
void MeetConstraintsBetween(Instruction* first, Instruction* second,
int gap_index);
+ void MeetRegisterConstraintsForLastInstructionInBlock(BasicBlock* block);
void ResolvePhis(BasicBlock* block);
// Helper methods for building intervals.
CallDescriptor* Linkage::GetStubCallDescriptor(
- CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count) {
+ CodeStubInterfaceDescriptor* descriptor, int stack_parameter_count,
+ CallDescriptor::DeoptimizationSupport can_deoptimize, Zone* zone) {
return LinkageHelper::GetStubCallDescriptor<LinkageHelperTraits>(
- this->info_->zone(), descriptor, stack_parameter_count);
+ zone, descriptor, stack_parameter_count, can_deoptimize);
}
int patch_count = input_data->ReturnAddressPatchCount();
for (int i = 0; i < patch_count; i++) {
int return_pc = input_data->ReturnAddressPc(i)->value();
- if (pc == return_pc) {
- return input_data->PatchedAddressPc(i)->value();
+ int patch_pc = input_data->PatchedAddressPc(i)->value();
+ // If the supplied pc matches the return pc or if the address
+ // has been already patched, return the patch pc.
+ if (pc == return_pc || pc == patch_pc) {
+ return patch_pc;
}
}
return -1;
SafepointEntry Code::GetSafepointEntry(Address pc) {
SafepointTable table(this);
- return table.FindEntry(pc);
+ SafepointEntry entry = table.FindEntry(pc);
+ if (entry.is_valid() || !is_turbofanned()) {
+ return entry;
+ }
+
+ // If the code is turbofanned, we might be looking for
+ // an address that was patched by lazy deoptimization.
+ // In that case look through the patch table, try to
+ // lookup the original address there, and then use this
+ // to find the safepoint entry.
+ DeoptimizationInputData* deopt_data =
+ DeoptimizationInputData::cast(deoptimization_data());
+ intptr_t offset = pc - instruction_start();
+ for (int i = 0; i < deopt_data->ReturnAddressPatchCount(); i++) {
+ if (deopt_data->PatchedAddressPc(i)->value() == offset) {
+ int original_offset = deopt_data->ReturnAddressPc(i)->value();
+ return table.FindEntry(instruction_start() + original_offset);
+ }
+ }
+ return SafepointEntry();
}
os << "Deoptimization Input Data (deopt points = " << deopt_count << ")\n";
if (0 != deopt_count) {
os << " index ast id argc pc";
- if (FLAG_print_code_verbose) os << "commands";
+ if (FLAG_print_code_verbose) os << " commands";
os << "\n";
}
for (int i = 0; i < deopt_count; i++) {
Translation::BEGIN !=
(opcode = static_cast<Translation::Opcode>(iterator.Next()))) {
Vector<char> buf2 = Vector<char>::New(128);
- SNPrintF(buf2, "%24s %s ", "", Translation::StringFor(opcode));
+ SNPrintF(buf2, "%27s %s ", "", Translation::StringFor(opcode));
os << buf2.start();
switch (opcode) {
if (return_address_patch_count != 0) {
os << "Return address patch data (count = " << return_address_patch_count
<< ")\n";
- os << "index pc patched_pc\n";
+ os << " index pc patched_pc\n";
}
for (int i = 0; i < return_address_patch_count; i++) {
Vector<char> buf = Vector<char>::New(128);
- SNPrintF(buf, "%6d %6d %10d", i, ReturnAddressPc(i)->value(),
+ SNPrintF(buf, "%6d %6d %12d\n", i, ReturnAddressPc(i)->value(),
PatchedAddressPc(i)->value());
os << buf.start();
}
'test-debug/DebugBreakLoop': [PASS, NO_VARIANTS],
# Support for lazy deoptimization is missing.
- 'test-deoptimization/DeoptimizeSimple': [PASS, NO_VARIANTS],
- 'test-deoptimization/DeoptimizeSimpleNested': [PASS, NO_VARIANTS],
- 'test-deoptimization/DeoptimizeSimpleWithArguments': [PASS, NO_VARIANTS],
- 'test-deoptimization/DeoptimizeBinaryOperation*': [PASS, NO_VARIANTS],
'test-deoptimization/DeoptimizeCompare': [PASS, NO_VARIANTS],
- 'test-deoptimization/DeoptimizeLoadICStoreIC': [PASS, NO_VARIANTS],
- 'test-deoptimization/DeoptimizeLoadICStoreICNested': [PASS, NO_VARIANTS],
# Support for breakpoints requires using LoadICs and StoreICs.
'test-debug/BreakPointICStore': [PASS, NO_VARIANTS],
TEST(BuildScheduleTrivialLazyDeoptCall) {
+ FLAG_turbo_deoptimization = true;
+
HandleAndZoneScope scope;
Isolate* isolate = scope.main_isolate();
Graph graph(scope.main_zone());
TEST(DeoptimizeSimple) {
+ i::FLAG_turbo_deoptimization = true;
+
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeSimpleWithArguments) {
+ i::FLAG_turbo_deoptimization = true;
+
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeSimpleNested) {
+ i::FLAG_turbo_deoptimization = true;
+
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeRecursive) {
+ i::FLAG_turbo_deoptimization = true;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeMultiple) {
+ i::FLAG_turbo_deoptimization = true;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeConstructor) {
+ i::FLAG_turbo_deoptimization = true;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeConstructorMultiple) {
+ i::FLAG_turbo_deoptimization = true;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeBinaryOperationADDString) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
AllowNativesSyntaxNoInlining options;
LocalContext env;
TEST(DeoptimizeBinaryOperationADD) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeBinaryOperationSUB) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeBinaryOperationMUL) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeBinaryOperationDIV) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeBinaryOperationMOD) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeCompare) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeLoadICStoreIC) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());
TEST(DeoptimizeLoadICStoreICNested) {
+ i::FLAG_turbo_deoptimization = true;
i::FLAG_concurrent_recompilation = false;
LocalContext env;
v8::HandleScope scope(env->GetIsolate());