summaryrefslogtreecommitdiff
path: root/Source/JavaScriptCore/dfg/DFGNode.h
diff options
context:
space:
mode:
Diffstat (limited to 'Source/JavaScriptCore/dfg/DFGNode.h')
-rw-r--r--Source/JavaScriptCore/dfg/DFGNode.h1718
1 files changed, 1362 insertions, 356 deletions
diff --git a/Source/JavaScriptCore/dfg/DFGNode.h b/Source/JavaScriptCore/dfg/DFGNode.h
index 28eccc07a..e29374f04 100644
--- a/Source/JavaScriptCore/dfg/DFGNode.h
+++ b/Source/JavaScriptCore/dfg/DFGNode.h
@@ -1,5 +1,5 @@
/*
- * Copyright (C) 2011, 2012, 2013 Apple Inc. All rights reserved.
+ * Copyright (C) 2011-2016 Apple Inc. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
@@ -26,39 +26,57 @@
#ifndef DFGNode_h
#define DFGNode_h
-#include <wtf/Platform.h>
-
#if ENABLE(DFG_JIT)
+#include "BasicBlockLocation.h"
#include "CodeBlock.h"
-#include "CodeOrigin.h"
#include "DFGAbstractValue.h"
#include "DFGAdjacencyList.h"
+#include "DFGArithMode.h"
#include "DFGArrayMode.h"
#include "DFGCommon.h"
+#include "DFGEpoch.h"
+#include "DFGLazyJSValue.h"
+#include "DFGMultiGetByOffsetData.h"
#include "DFGNodeFlags.h"
+#include "DFGNodeOrigin.h"
#include "DFGNodeType.h"
+#include "DFGObjectMaterializationData.h"
+#include "DFGTransition.h"
+#include "DFGUseKind.h"
#include "DFGVariableAccessData.h"
-#include "JSCJSValueInlines.h"
+#include "GetByIdVariant.h"
#include "JSCJSValue.h"
#include "Operands.h"
+#include "PutByIdVariant.h"
#include "SpeculatedType.h"
#include "StructureSet.h"
+#include "TypeLocation.h"
#include "ValueProfile.h"
+#include <wtf/ListDump.h>
namespace JSC { namespace DFG {
-struct StructureTransitionData {
- Structure* previousStructure;
- Structure* newStructure;
-
- StructureTransitionData() { }
+class Graph;
+class PromotedLocationDescriptor;
+struct BasicBlock;
+
+struct StorageAccessData {
+ PropertyOffset offset;
+ unsigned identifierNumber;
+
+ // This needs to know the inferred type. For puts, this is necessary because we need to remember
+ // what check is needed. For gets, this is necessary because otherwise AI might forget what type is
+ // guaranteed.
+ InferredType::Descriptor inferredType;
+};
+
+struct MultiPutByOffsetData {
+ unsigned identifierNumber;
+ Vector<PutByIdVariant, 2> variants;
- StructureTransitionData(Structure* previousStructure, Structure* newStructure)
- : previousStructure(previousStructure)
- , newStructure(newStructure)
- {
- }
+ bool writesStructures() const;
+ bool reallocatesStorage() const;
};
struct NewArrayBufferData {
@@ -67,10 +85,145 @@ struct NewArrayBufferData {
IndexingType indexingType;
};
+struct BranchTarget {
+ BranchTarget()
+ : block(0)
+ , count(PNaN)
+ {
+ }
+
+ explicit BranchTarget(BasicBlock* block)
+ : block(block)
+ , count(PNaN)
+ {
+ }
+
+ void setBytecodeIndex(unsigned bytecodeIndex)
+ {
+ block = bitwise_cast<BasicBlock*>(static_cast<uintptr_t>(bytecodeIndex));
+ }
+ unsigned bytecodeIndex() const { return bitwise_cast<uintptr_t>(block); }
+
+ void dump(PrintStream&) const;
+
+ BasicBlock* block;
+ float count;
+};
+
+struct BranchData {
+ static BranchData withBytecodeIndices(
+ unsigned takenBytecodeIndex, unsigned notTakenBytecodeIndex)
+ {
+ BranchData result;
+ result.taken.block = bitwise_cast<BasicBlock*>(static_cast<uintptr_t>(takenBytecodeIndex));
+ result.notTaken.block = bitwise_cast<BasicBlock*>(static_cast<uintptr_t>(notTakenBytecodeIndex));
+ return result;
+ }
+
+ unsigned takenBytecodeIndex() const { return taken.bytecodeIndex(); }
+ unsigned notTakenBytecodeIndex() const { return notTaken.bytecodeIndex(); }
+
+ BasicBlock*& forCondition(bool condition)
+ {
+ if (condition)
+ return taken.block;
+ return notTaken.block;
+ }
+
+ BranchTarget taken;
+ BranchTarget notTaken;
+};
+
+// The SwitchData and associated data structures duplicate the information in
+// JumpTable. The DFG may ultimately end up using the JumpTable, though it may
+// instead decide to do something different - this is entirely up to the DFG.
+// These data structures give the DFG a higher-level semantic description of
+// what is going on, which will allow it to make the right decision.
+//
+// Note that there will never be multiple SwitchCases in SwitchData::cases that
+// have the same SwitchCase::value, since the bytecode's JumpTables never have
+// duplicates - since the JumpTable maps a value to a target. It's a
+// one-to-many mapping. So we may have duplicate targets, but never duplicate
+// values.
+struct SwitchCase {
+ SwitchCase()
+ {
+ }
+
+ SwitchCase(LazyJSValue value, BasicBlock* target)
+ : value(value)
+ , target(target)
+ {
+ }
+
+ static SwitchCase withBytecodeIndex(LazyJSValue value, unsigned bytecodeIndex)
+ {
+ SwitchCase result;
+ result.value = value;
+ result.target.setBytecodeIndex(bytecodeIndex);
+ return result;
+ }
+
+ LazyJSValue value;
+ BranchTarget target;
+};
+
+struct SwitchData {
+ // Initializes most fields to obviously invalid values. Anyone
+ // constructing this should make sure to initialize everything they
+ // care about manually.
+ SwitchData()
+ : kind(static_cast<SwitchKind>(-1))
+ , switchTableIndex(UINT_MAX)
+ , didUseJumpTable(false)
+ {
+ }
+
+ Vector<SwitchCase> cases;
+ BranchTarget fallThrough;
+ SwitchKind kind;
+ unsigned switchTableIndex;
+ bool didUseJumpTable;
+};
+
+struct CallVarargsData {
+ int firstVarArgOffset;
+};
+
+struct LoadVarargsData {
+ VirtualRegister start; // Local for the first element. This is the first actual argument, not this.
+ VirtualRegister count; // Local for the count.
+ VirtualRegister machineStart;
+ VirtualRegister machineCount;
+ unsigned offset; // Which array element to start with. Usually this is 0.
+ unsigned mandatoryMinimum; // The number of elements on the stack that must be initialized; if the array is too short then the missing elements must get undefined. Does not include "this".
+ unsigned limit; // Maximum number of elements to load. Includes "this".
+};
+
+struct StackAccessData {
+ StackAccessData()
+ : format(DeadFlush)
+ {
+ }
+
+ StackAccessData(VirtualRegister local, FlushFormat format)
+ : local(local)
+ , format(format)
+ {
+ }
+
+ VirtualRegister local;
+ VirtualRegister machineLocal;
+ FlushFormat format;
+
+ FlushedAt flushedAt() { return FlushedAt(format, machineLocal); }
+};
+
// This type used in passing an immediate argument to Node constructor;
// distinguishes an immediate value (typically an index into a CodeBlock data structure -
// a constant index, argument, or identifier) from a Node*.
struct OpInfo {
+ OpInfo() : m_value(0) { }
explicit OpInfo(int32_t value) : m_value(static_cast<uintptr_t>(value)) { }
explicit OpInfo(uint32_t value) : m_value(static_cast<uintptr_t>(value)) { }
#if OS(DARWIN) || USE(JSVALUE64)
@@ -88,65 +241,112 @@ struct Node {
Node() { }
- Node(NodeType op, CodeOrigin codeOrigin, const AdjacencyList& children)
- : codeOrigin(codeOrigin)
+ Node(NodeType op, NodeOrigin nodeOrigin, const AdjacencyList& children)
+ : origin(nodeOrigin)
, children(children)
- , m_virtualRegister(InvalidVirtualRegister)
+ , m_virtualRegister(VirtualRegister())
, m_refCount(1)
, m_prediction(SpecNone)
+ , owner(nullptr)
{
+ m_misc.replacement = nullptr;
setOpAndDefaultFlags(op);
}
// Construct a node with up to 3 children, no immediate value.
- Node(NodeType op, CodeOrigin codeOrigin, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
- : codeOrigin(codeOrigin)
+ Node(NodeType op, NodeOrigin nodeOrigin, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
+ : origin(nodeOrigin)
, children(AdjacencyList::Fixed, child1, child2, child3)
- , m_virtualRegister(InvalidVirtualRegister)
+ , m_virtualRegister(VirtualRegister())
, m_refCount(1)
, m_prediction(SpecNone)
+ , m_opInfo(0)
+ , m_opInfo2(0)
+ , owner(nullptr)
{
+ m_misc.replacement = nullptr;
setOpAndDefaultFlags(op);
ASSERT(!(m_flags & NodeHasVarArgs));
}
+ // Construct a node with up to 3 children, no immediate value.
+ Node(NodeFlags result, NodeType op, NodeOrigin nodeOrigin, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
+ : origin(nodeOrigin)
+ , children(AdjacencyList::Fixed, child1, child2, child3)
+ , m_virtualRegister(VirtualRegister())
+ , m_refCount(1)
+ , m_prediction(SpecNone)
+ , m_opInfo(0)
+ , m_opInfo2(0)
+ , owner(nullptr)
+ {
+ m_misc.replacement = nullptr;
+ setOpAndDefaultFlags(op);
+ setResult(result);
+ ASSERT(!(m_flags & NodeHasVarArgs));
+ }
+
// Construct a node with up to 3 children and an immediate value.
- Node(NodeType op, CodeOrigin codeOrigin, OpInfo imm, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
- : codeOrigin(codeOrigin)
+ Node(NodeType op, NodeOrigin nodeOrigin, OpInfo imm, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
+ : origin(nodeOrigin)
, children(AdjacencyList::Fixed, child1, child2, child3)
- , m_virtualRegister(InvalidVirtualRegister)
+ , m_virtualRegister(VirtualRegister())
, m_refCount(1)
+ , m_prediction(SpecNone)
, m_opInfo(imm.m_value)
+ , m_opInfo2(0)
+ , owner(nullptr)
+ {
+ m_misc.replacement = nullptr;
+ setOpAndDefaultFlags(op);
+ ASSERT(!(m_flags & NodeHasVarArgs));
+ }
+
+ // Construct a node with up to 3 children and an immediate value.
+ Node(NodeFlags result, NodeType op, NodeOrigin nodeOrigin, OpInfo imm, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
+ : origin(nodeOrigin)
+ , children(AdjacencyList::Fixed, child1, child2, child3)
+ , m_virtualRegister(VirtualRegister())
+ , m_refCount(1)
, m_prediction(SpecNone)
+ , m_opInfo(imm.m_value)
+ , m_opInfo2(0)
+ , owner(nullptr)
{
+ m_misc.replacement = nullptr;
setOpAndDefaultFlags(op);
+ setResult(result);
ASSERT(!(m_flags & NodeHasVarArgs));
}
// Construct a node with up to 3 children and two immediate values.
- Node(NodeType op, CodeOrigin codeOrigin, OpInfo imm1, OpInfo imm2, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
- : codeOrigin(codeOrigin)
+ Node(NodeType op, NodeOrigin nodeOrigin, OpInfo imm1, OpInfo imm2, Edge child1 = Edge(), Edge child2 = Edge(), Edge child3 = Edge())
+ : origin(nodeOrigin)
, children(AdjacencyList::Fixed, child1, child2, child3)
- , m_virtualRegister(InvalidVirtualRegister)
+ , m_virtualRegister(VirtualRegister())
, m_refCount(1)
- , m_opInfo(imm1.m_value)
- , m_opInfo2(safeCast<unsigned>(imm2.m_value))
, m_prediction(SpecNone)
+ , m_opInfo(imm1.m_value)
+ , m_opInfo2(imm2.m_value)
+ , owner(nullptr)
{
+ m_misc.replacement = nullptr;
setOpAndDefaultFlags(op);
ASSERT(!(m_flags & NodeHasVarArgs));
}
// Construct a node with a variable number of children and two immediate values.
- Node(VarArgTag, NodeType op, CodeOrigin codeOrigin, OpInfo imm1, OpInfo imm2, unsigned firstChild, unsigned numChildren)
- : codeOrigin(codeOrigin)
+ Node(VarArgTag, NodeType op, NodeOrigin nodeOrigin, OpInfo imm1, OpInfo imm2, unsigned firstChild, unsigned numChildren)
+ : origin(nodeOrigin)
, children(AdjacencyList::Variable, firstChild, numChildren)
- , m_virtualRegister(InvalidVirtualRegister)
+ , m_virtualRegister(VirtualRegister())
, m_refCount(1)
- , m_opInfo(imm1.m_value)
- , m_opInfo2(safeCast<unsigned>(imm2.m_value))
, m_prediction(SpecNone)
+ , m_opInfo(imm1.m_value)
+ , m_opInfo2(imm2.m_value)
+ , owner(nullptr)
{
+ m_misc.replacement = nullptr;
setOpAndDefaultFlags(op);
ASSERT(m_flags & NodeHasVarArgs);
}
@@ -169,7 +369,6 @@ struct Node {
bool mergeFlags(NodeFlags flags)
{
- ASSERT(!(flags & NodeDoesNotExit));
NodeFlags newFlags = m_flags | flags;
if (newFlags == m_flags)
return false;
@@ -179,7 +378,6 @@ struct Node {
bool filterFlags(NodeFlags flags)
{
- ASSERT(flags & NodeDoesNotExit);
NodeFlags newFlags = m_flags & flags;
if (newFlags == m_flags)
return false;
@@ -192,162 +390,231 @@ struct Node {
return filterFlags(~flags);
}
- void setOpAndDefaultFlags(NodeType op)
+ void setResult(NodeFlags result)
{
- m_op = op;
- m_flags = defaultFlags(op);
+ ASSERT(!(result & ~NodeResultMask));
+ clearFlags(NodeResultMask);
+ mergeFlags(result);
}
-
- void setOpAndDefaultNonExitFlags(NodeType op)
+
+ NodeFlags result() const
{
- ASSERT(!(m_flags & NodeHasVarArgs));
- setOpAndDefaultNonExitFlagsUnchecked(op);
+ return flags() & NodeResultMask;
}
-
- void setOpAndDefaultNonExitFlagsUnchecked(NodeType op)
+
+ void setOpAndDefaultFlags(NodeType op)
{
m_op = op;
- m_flags = (defaultFlags(op) & ~NodeExitsForward) | (m_flags & NodeExitsForward);
+ m_flags = defaultFlags(op);
}
- void convertToPhantom()
+ void remove();
+
+ void convertToCheckStructure(StructureSet* set)
{
- setOpAndDefaultNonExitFlags(Phantom);
+ setOpAndDefaultFlags(CheckStructure);
+ m_opInfo = bitwise_cast<uintptr_t>(set);
}
- void convertToPhantomUnchecked()
+ void convertToCheckStructureImmediate(Node* structure)
{
- setOpAndDefaultNonExitFlagsUnchecked(Phantom);
+ ASSERT(op() == CheckStructure);
+ m_op = CheckStructureImmediate;
+ children.setChild1(Edge(structure, CellUse));
}
-
- void convertToIdentity()
+
+ void replaceWith(Node* other)
{
- RELEASE_ASSERT(child1());
- RELEASE_ASSERT(!child2());
- setOpAndDefaultNonExitFlags(Identity);
+ remove();
+ setReplacement(other);
}
+ void convertToIdentity();
+ void convertToIdentityOn(Node*);
+
bool mustGenerate()
{
return m_flags & NodeMustGenerate;
}
- void setCanExit(bool exits)
- {
- if (exits)
- m_flags &= ~NodeDoesNotExit;
- else
- m_flags |= NodeDoesNotExit;
- }
-
- bool canExit()
- {
- return !(m_flags & NodeDoesNotExit);
- }
-
bool isConstant()
{
- return op() == JSConstant;
- }
-
- bool isWeakConstant()
- {
- return op() == WeakJSConstant;
- }
-
- bool isStronglyProvedConstantIn(InlineCallFrame* inlineCallFrame)
- {
- return isConstant() && codeOrigin.inlineCallFrame == inlineCallFrame;
- }
-
- bool isStronglyProvedConstantIn(const CodeOrigin& codeOrigin)
- {
- return isStronglyProvedConstantIn(codeOrigin.inlineCallFrame);
- }
-
- bool isPhantomArguments()
- {
- return op() == PhantomArguments;
+ switch (op()) {
+ case JSConstant:
+ case DoubleConstant:
+ case Int52Constant:
+ return true;
+ default:
+ return false;
+ }
}
bool hasConstant()
{
switch (op()) {
case JSConstant:
- case WeakJSConstant:
- case PhantomArguments:
+ case DoubleConstant:
+ case Int52Constant:
+ return true;
+
+ case PhantomDirectArguments:
+ case PhantomClonedArguments:
+ // These pretend to be the empty value constant for the benefit of the DFG backend, which
+ // otherwise wouldn't take kindly to a node that doesn't compute a value.
return true;
+
default:
return false;
}
}
- unsigned constantNumber()
+ FrozenValue* constant()
{
- ASSERT(isConstant());
- return m_opInfo;
+ ASSERT(hasConstant());
+
+ if (op() == PhantomDirectArguments || op() == PhantomClonedArguments) {
+ // These pretend to be the empty value constant for the benefit of the DFG backend, which
+ // otherwise wouldn't take kindly to a node that doesn't compute a value.
+ return FrozenValue::emptySingleton();
+ }
+
+ return bitwise_cast<FrozenValue*>(m_opInfo);
}
- void convertToConstant(unsigned constantNumber)
+ // Don't call this directly - use Graph::convertToConstant() instead!
+ void convertToConstant(FrozenValue* value)
{
- m_op = JSConstant;
- m_flags &= ~(NodeMustGenerate | NodeMightClobber | NodeClobbersWorld);
- m_opInfo = constantNumber;
+ if (hasDoubleResult())
+ m_op = DoubleConstant;
+ else if (hasInt52Result())
+ m_op = Int52Constant;
+ else
+ m_op = JSConstant;
+ m_flags &= ~NodeMustGenerate;
+ m_opInfo = bitwise_cast<uintptr_t>(value);
children.reset();
}
- void convertToWeakConstant(JSCell* cell)
+ void convertToConstantStoragePointer(void* pointer)
{
- m_op = WeakJSConstant;
- m_flags &= ~(NodeMustGenerate | NodeMightClobber | NodeClobbersWorld);
- m_opInfo = bitwise_cast<uintptr_t>(cell);
+ ASSERT(op() == GetIndexedPropertyStorage);
+ m_op = ConstantStoragePointer;
+ m_opInfo = bitwise_cast<uintptr_t>(pointer);
children.reset();
}
void convertToGetLocalUnlinked(VirtualRegister local)
{
m_op = GetLocalUnlinked;
- m_flags &= ~(NodeMustGenerate | NodeMightClobber | NodeClobbersWorld);
- m_opInfo = local;
+ m_flags &= ~NodeMustGenerate;
+ m_opInfo = local.offset();
+ m_opInfo2 = VirtualRegister().offset();
children.reset();
}
- void convertToStructureTransitionWatchpoint(Structure* structure)
+ void convertToPutStack(StackAccessData* data)
{
- ASSERT(m_op == CheckStructure || m_op == ForwardCheckStructure || m_op == ArrayifyToStructure);
- m_opInfo = bitwise_cast<uintptr_t>(structure);
- if (m_op == CheckStructure || m_op == ArrayifyToStructure)
- m_op = StructureTransitionWatchpoint;
- else
- m_op = ForwardStructureTransitionWatchpoint;
+ m_op = PutStack;
+ m_flags |= NodeMustGenerate;
+ m_opInfo = bitwise_cast<uintptr_t>(data);
+ m_opInfo2 = 0;
}
- void convertToStructureTransitionWatchpoint()
+ void convertToGetStack(StackAccessData* data)
{
- convertToStructureTransitionWatchpoint(structureSet().singletonStructure());
+ m_op = GetStack;
+ m_flags &= ~NodeMustGenerate;
+ m_opInfo = bitwise_cast<uintptr_t>(data);
+ m_opInfo2 = 0;
+ children.reset();
}
- void convertToGetByOffset(unsigned storageAccessDataIndex, Edge storage)
+ void convertToGetByOffset(StorageAccessData& data, Edge storage)
{
- ASSERT(m_op == GetById || m_op == GetByIdFlush);
- m_opInfo = storageAccessDataIndex;
+ ASSERT(m_op == GetById || m_op == GetByIdFlush || m_op == MultiGetByOffset);
+ m_opInfo = bitwise_cast<uintptr_t>(&data);
+ children.setChild2(children.child1());
+ children.child2().setUseKind(KnownCellUse);
children.setChild1(storage);
m_op = GetByOffset;
- m_flags &= ~NodeClobbersWorld;
+ m_flags &= ~NodeMustGenerate;
+ }
+
+ void convertToMultiGetByOffset(MultiGetByOffsetData* data)
+ {
+ ASSERT(m_op == GetById || m_op == GetByIdFlush);
+ m_opInfo = bitwise_cast<intptr_t>(data);
+ child1().setUseKind(CellUse);
+ m_op = MultiGetByOffset;
+ ASSERT(m_flags & NodeMustGenerate);
}
- void convertToPutByOffset(unsigned storageAccessDataIndex, Edge storage)
+ void convertToPutByOffset(StorageAccessData& data, Edge storage)
{
- ASSERT(m_op == PutById || m_op == PutByIdDirect);
- m_opInfo = storageAccessDataIndex;
+ ASSERT(m_op == PutById || m_op == PutByIdDirect || m_op == PutByIdFlush || m_op == MultiPutByOffset);
+ m_opInfo = bitwise_cast<uintptr_t>(&data);
children.setChild3(children.child2());
children.setChild2(children.child1());
children.setChild1(storage);
m_op = PutByOffset;
- m_flags &= ~NodeClobbersWorld;
}
- void convertToPhantomLocal()
+ void convertToMultiPutByOffset(MultiPutByOffsetData* data)
+ {
+ ASSERT(m_op == PutById || m_op == PutByIdDirect || m_op == PutByIdFlush);
+ m_opInfo = bitwise_cast<intptr_t>(data);
+ m_op = MultiPutByOffset;
+ }
+
+ void convertToPutHint(const PromotedLocationDescriptor&, Node* base, Node* value);
+
+ void convertToPutByOffsetHint();
+ void convertToPutStructureHint(Node* structure);
+ void convertToPutClosureVarHint();
+
+ void convertToPhantomNewObject()
+ {
+ ASSERT(m_op == NewObject || m_op == MaterializeNewObject);
+ m_op = PhantomNewObject;
+ m_flags &= ~NodeHasVarArgs;
+ m_flags |= NodeMustGenerate;
+ m_opInfo = 0;
+ m_opInfo2 = 0;
+ children = AdjacencyList();
+ }
+
+ void convertToPhantomNewFunction()
+ {
+ ASSERT(m_op == NewFunction || m_op == NewArrowFunction || m_op == NewGeneratorFunction);
+ m_op = PhantomNewFunction;
+ m_flags |= NodeMustGenerate;
+ m_opInfo = 0;
+ m_opInfo2 = 0;
+ children = AdjacencyList();
+ }
+
+ void convertToPhantomNewGeneratorFunction()
+ {
+ ASSERT(m_op == NewGeneratorFunction);
+ m_op = PhantomNewGeneratorFunction;
+ m_flags |= NodeMustGenerate;
+ m_opInfo = 0;
+ m_opInfo2 = 0;
+ children = AdjacencyList();
+ }
+
+ void convertToPhantomCreateActivation()
+ {
+ ASSERT(m_op == CreateActivation || m_op == MaterializeCreateActivation);
+ m_op = PhantomCreateActivation;
+ m_flags &= ~NodeHasVarArgs;
+ m_flags |= NodeMustGenerate;
+ m_opInfo = 0;
+ m_opInfo2 = 0;
+ children = AdjacencyList();
+ }
+
+ void convertPhantomToPhantomLocal()
{
ASSERT(m_op == Phantom && (child1()->op() == Phi || child1()->op() == SetLocal || child1()->op() == SetArgument));
m_op = PhantomLocal;
@@ -355,11 +622,19 @@ struct Node {
children.setChild1(Edge());
}
+ void convertFlushToPhantomLocal()
+ {
+ ASSERT(m_op == Flush);
+ m_op = PhantomLocal;
+ children = AdjacencyList();
+ }
+
void convertToGetLocal(VariableAccessData* variable, Node* phi)
{
ASSERT(m_op == GetLocalUnlinked);
m_op = GetLocal;
m_opInfo = bitwise_cast<uintptr_t>(variable);
+ m_opInfo2 = 0;
children.setChild1(Edge(phi));
}
@@ -368,59 +643,116 @@ struct Node {
ASSERT(m_op == ToPrimitive);
m_op = ToString;
}
+
+ void convertToArithSqrt()
+ {
+ ASSERT(m_op == ArithPow);
+ child2() = Edge();
+ m_op = ArithSqrt;
+ }
+
+ void convertToArithNegate()
+ {
+ ASSERT(m_op == ArithAbs && child1().useKind() == Int32Use);
+ m_op = ArithNegate;
+ }
- JSCell* weakConstant()
+ JSValue asJSValue()
+ {
+ return constant()->value();
+ }
+
+ bool isInt32Constant()
+ {
+ return isConstant() && constant()->value().isInt32();
+ }
+
+ int32_t asInt32()
+ {
+ return asJSValue().asInt32();
+ }
+
+ uint32_t asUInt32()
+ {
+ return asInt32();
+ }
+
+ bool isDoubleConstant()
+ {
+ return isConstant() && constant()->value().isDouble();
+ }
+
+ bool isNumberConstant()
{
- ASSERT(op() == WeakJSConstant);
- return bitwise_cast<JSCell*>(m_opInfo);
+ return isConstant() && constant()->value().isNumber();
}
- JSValue valueOfJSConstant(CodeBlock* codeBlock)
+ double asNumber()
{
- switch (op()) {
- case WeakJSConstant:
- return JSValue(weakConstant());
- case JSConstant:
- return codeBlock->constantRegister(FirstConstantRegisterIndex + constantNumber()).get();
- case PhantomArguments:
- return JSValue();
- default:
- RELEASE_ASSERT_NOT_REACHED();
- return JSValue(); // Have to return something in release mode.
- }
+ return asJSValue().asNumber();
+ }
+
+ bool isMachineIntConstant()
+ {
+ return isConstant() && constant()->value().isMachineInt();
+ }
+
+ int64_t asMachineInt()
+ {
+ return asJSValue().asMachineInt();
+ }
+
+ bool isBooleanConstant()
+ {
+ return isConstant() && constant()->value().isBoolean();
+ }
+
+ bool asBoolean()
+ {
+ return constant()->value().asBoolean();
}
- bool isInt32Constant(CodeBlock* codeBlock)
+ bool isUndefinedOrNullConstant()
{
- return isConstant() && valueOfJSConstant(codeBlock).isInt32();
+ return isConstant() && constant()->value().isUndefinedOrNull();
}
-
- bool isDoubleConstant(CodeBlock* codeBlock)
+
+ bool isCellConstant()
{
- bool result = isConstant() && valueOfJSConstant(codeBlock).isDouble();
- if (result)
- ASSERT(!isInt32Constant(codeBlock));
- return result;
+ return isConstant() && constant()->value() && constant()->value().isCell();
+ }
+
+ JSCell* asCell()
+ {
+ return constant()->value().asCell();
+ }
+
+ template<typename T>
+ T dynamicCastConstant()
+ {
+ if (!isCellConstant())
+ return nullptr;
+ return jsDynamicCast<T>(asCell());
}
- bool isNumberConstant(CodeBlock* codeBlock)
+ template<typename T>
+ T castConstant()
{
- bool result = isConstant() && valueOfJSConstant(codeBlock).isNumber();
- ASSERT(result == (isInt32Constant(codeBlock) || isDoubleConstant(codeBlock)));
+ T result = dynamicCastConstant<T>();
+ RELEASE_ASSERT(result);
return result;
}
-
- bool isBooleanConstant(CodeBlock* codeBlock)
+
+ JSValue initializationValueForActivation() const
{
- return isConstant() && valueOfJSConstant(codeBlock).isBoolean();
+ ASSERT(op() == CreateActivation);
+ return bitwise_cast<FrozenValue*>(m_opInfo2)->value();
}
-
+
bool containsMovHint()
{
switch (op()) {
- case SetLocal:
case MovHint:
- case MovHintAndCheck:
case ZombieHint:
return true;
default:
@@ -428,53 +760,118 @@ struct Node {
}
}
- bool hasVariableAccessData()
+ bool hasVariableAccessData(Graph&);
+ bool hasLocal(Graph& graph)
+ {
+ return hasVariableAccessData(graph);
+ }
+
+ // This is useful for debugging code, where a node that should have a variable
+ // access data doesn't have one because it hasn't been initialized yet.
+ VariableAccessData* tryGetVariableAccessData()
+ {
+ VariableAccessData* result = reinterpret_cast<VariableAccessData*>(m_opInfo);
+ if (!result)
+ return 0;
+ return result->find();
+ }
+
+ VariableAccessData* variableAccessData()
+ {
+ return reinterpret_cast<VariableAccessData*>(m_opInfo)->find();
+ }
+
+ VirtualRegister local()
+ {
+ return variableAccessData()->local();
+ }
+
+ VirtualRegister machineLocal()
+ {
+ return variableAccessData()->machineLocal();
+ }
+
+ bool hasUnlinkedLocal()
{
switch (op()) {
- case GetLocal:
- case SetLocal:
+ case GetLocalUnlinked:
+ case ExtractOSREntryLocal:
case MovHint:
- case MovHintAndCheck:
case ZombieHint:
- case Phi:
- case SetArgument:
- case Flush:
- case PhantomLocal:
+ case KillStack:
return true;
default:
return false;
}
}
- bool hasLocal()
+ VirtualRegister unlinkedLocal()
{
- return hasVariableAccessData();
+ ASSERT(hasUnlinkedLocal());
+ return static_cast<VirtualRegister>(m_opInfo);
}
- VariableAccessData* variableAccessData()
+ bool hasUnlinkedMachineLocal()
{
- ASSERT(hasVariableAccessData());
- return reinterpret_cast<VariableAccessData*>(m_opInfo)->find();
+ return op() == GetLocalUnlinked;
}
- VirtualRegister local()
+ void setUnlinkedMachineLocal(VirtualRegister reg)
{
- return variableAccessData()->local();
+ ASSERT(hasUnlinkedMachineLocal());
+ m_opInfo2 = reg.offset();
}
- VirtualRegister unlinkedLocal()
+ VirtualRegister unlinkedMachineLocal()
{
- ASSERT(op() == GetLocalUnlinked);
- return static_cast<VirtualRegister>(m_opInfo);
+ ASSERT(hasUnlinkedMachineLocal());
+ return VirtualRegister(m_opInfo2);
}
+ bool hasStackAccessData()
+ {
+ switch (op()) {
+ case PutStack:
+ case GetStack:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ StackAccessData* stackAccessData()
+ {
+ ASSERT(hasStackAccessData());
+ return bitwise_cast<StackAccessData*>(m_opInfo);
+ }
+
+ bool hasPhi()
+ {
+ return op() == Upsilon;
+ }
+
+ Node* phi()
+ {
+ ASSERT(hasPhi());
+ return bitwise_cast<Node*>(m_opInfo);
+ }
+
+ bool isStoreBarrier()
+ {
+ return op() == StoreBarrier;
+ }
+
bool hasIdentifier()
{
switch (op()) {
case GetById:
case GetByIdFlush:
case PutById:
+ case PutByIdFlush:
case PutByIdDirect:
+ case PutGetterById:
+ case PutSetterById:
+ case PutGetterSetterById:
return true;
default:
return false;
@@ -486,50 +883,61 @@ struct Node {
ASSERT(hasIdentifier());
return m_opInfo;
}
-
- unsigned resolveGlobalDataIndex()
- {
- ASSERT(op() == ResolveGlobal);
- return m_opInfo;
- }
- unsigned resolveOperationsDataIndex()
+ bool hasAccessorAttributes()
{
- ASSERT(op() == Resolve || op() == ResolveBase || op() == ResolveBaseStrictPut);
- return m_opInfo;
+ switch (op()) {
+ case PutGetterById:
+ case PutSetterById:
+ case PutGetterSetterById:
+ case PutGetterByVal:
+ case PutSetterByVal:
+ return true;
+ default:
+ return false;
+ }
}
- bool hasArithNodeFlags()
+ int32_t accessorAttributes()
{
+ ASSERT(hasAccessorAttributes());
switch (op()) {
- case UInt32ToNumber:
- case ArithAdd:
- case ArithSub:
- case ArithNegate:
- case ArithMul:
- case ArithAbs:
- case ArithMin:
- case ArithMax:
- case ArithMod:
- case ArithDiv:
- case ValueAdd:
- return true;
+ case PutGetterById:
+ case PutSetterById:
+ case PutGetterSetterById:
+ return m_opInfo2;
+ case PutGetterByVal:
+ case PutSetterByVal:
+ return m_opInfo;
default:
- return false;
+ RELEASE_ASSERT_NOT_REACHED();
+ return 0;
}
}
+ bool hasPromotedLocationDescriptor()
+ {
+ return op() == PutHint;
+ }
+
+ PromotedLocationDescriptor promotedLocationDescriptor();
+
// This corrects the arithmetic node flags, so that irrelevant bits are
// ignored. In particular, anything other than ArithMul does not need
// to know if it can speculate on negative zero.
NodeFlags arithNodeFlags()
{
NodeFlags result = m_flags & NodeArithFlagsMask;
- if (op() == ArithMul || op() == ArithDiv || op() == ArithMod || op() == ArithNegate || op() == DoubleAsInt32)
+ if (op() == ArithMul || op() == ArithDiv || op() == ArithMod || op() == ArithNegate || op() == ArithPow || op() == ArithRound || op() == ArithFloor || op() == ArithCeil || op() == DoubleAsInt32)
return result;
- return result & ~NodeNeedsNegZero;
+ return result & ~NodeBytecodeNeedsNegZero;
}
-
+
+ bool mayHaveNonIntResult()
+ {
+ return m_flags & NodeMayHaveNonIntResult;
+ }
+
bool hasConstantBuffer()
{
return op() == NewArrayBuffer;
@@ -562,7 +970,15 @@ struct Node {
return false;
}
}
-
+
+ // Return the indexing type that an array allocation *wants* to use. It may end up using a different
+ // type if we're having a bad time. You can determine the actual indexing type by asking the global
+ // object:
+ //
+ // m_graph.globalObjectFor(node->origin.semantic)->arrayStructureForIndexingTypeDuringAllocation(node->indexingType())
+ //
+ // This will give you a Structure*, and that will have some indexing type that may be different from
+ // the this one.
IndexingType indexingType()
{
ASSERT(hasIndexingType());
@@ -571,6 +987,24 @@ struct Node {
return m_opInfo;
}
+ bool hasTypedArrayType()
+ {
+ switch (op()) {
+ case NewTypedArray:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ TypedArrayType typedArrayType()
+ {
+ ASSERT(hasTypedArrayType());
+ TypedArrayType result = static_cast<TypedArrayType>(m_opInfo);
+ ASSERT(isTypedView(result));
+ return result;
+ }
+
bool hasInlineCapacity()
{
return op() == CreateThis;
@@ -588,6 +1022,9 @@ struct Node {
m_opInfo = indexingType;
}
+ // FIXME: We really should be able to inline code that uses NewRegexp. That means
+ // using something other than the index into the CodeBlock here.
+ // https://bugs.webkit.org/show_bug.cgi?id=154808
bool hasRegexpIndex()
{
return op() == NewRegexp;
@@ -599,66 +1036,120 @@ struct Node {
return m_opInfo;
}
- bool hasVarNumber()
+ bool hasScopeOffset()
{
- return op() == GetScopedVar || op() == PutScopedVar;
+ return op() == GetClosureVar || op() == PutClosureVar;
}
- unsigned varNumber()
+ ScopeOffset scopeOffset()
{
- ASSERT(hasVarNumber());
- return m_opInfo;
+ ASSERT(hasScopeOffset());
+ return ScopeOffset(m_opInfo);
}
- bool hasIdentifierNumberForCheck()
+ bool hasDirectArgumentsOffset()
{
- return op() == GlobalVarWatchpoint || op() == PutGlobalVarCheck;
+ return op() == GetFromArguments || op() == PutToArguments;
}
- unsigned identifierNumberForCheck()
+ DirectArgumentsOffset capturedArgumentsOffset()
{
- ASSERT(hasIdentifierNumberForCheck());
- return m_opInfo2;
+ ASSERT(hasDirectArgumentsOffset());
+ return DirectArgumentsOffset(m_opInfo);
}
bool hasRegisterPointer()
{
- return op() == GetGlobalVar || op() == PutGlobalVar || op() == GlobalVarWatchpoint || op() == PutGlobalVarCheck;
+ return op() == GetGlobalVar || op() == GetGlobalLexicalVariable || op() == PutGlobalVariable;
}
- WriteBarrier<Unknown>* registerPointer()
+ WriteBarrier<Unknown>* variablePointer()
{
return bitwise_cast<WriteBarrier<Unknown>*>(m_opInfo);
}
-
+
+ bool hasCallVarargsData()
+ {
+ switch (op()) {
+ case CallVarargs:
+ case CallForwardVarargs:
+ case TailCallVarargs:
+ case TailCallForwardVarargs:
+ case TailCallVarargsInlinedCaller:
+ case TailCallForwardVarargsInlinedCaller:
+ case ConstructVarargs:
+ case ConstructForwardVarargs:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ CallVarargsData* callVarargsData()
+ {
+ ASSERT(hasCallVarargsData());
+ return bitwise_cast<CallVarargsData*>(m_opInfo);
+ }
+
+ bool hasLoadVarargsData()
+ {
+ return op() == LoadVarargs || op() == ForwardVarargs;
+ }
+
+ LoadVarargsData* loadVarargsData()
+ {
+ ASSERT(hasLoadVarargsData());
+ return bitwise_cast<LoadVarargsData*>(m_opInfo);
+ }
+
bool hasResult()
{
- return m_flags & NodeResultMask;
+ return !!result();
}
bool hasInt32Result()
{
- return (m_flags & NodeResultMask) == NodeResultInt32;
+ return result() == NodeResultInt32;
+ }
+
+ bool hasInt52Result()
+ {
+ return result() == NodeResultInt52;
}
bool hasNumberResult()
{
- return (m_flags & NodeResultMask) == NodeResultNumber;
+ return result() == NodeResultNumber;
+ }
+
+ bool hasDoubleResult()
+ {
+ return result() == NodeResultDouble;
}
bool hasJSResult()
{
- return (m_flags & NodeResultMask) == NodeResultJS;
+ return result() == NodeResultJS;
}
bool hasBooleanResult()
{
- return (m_flags & NodeResultMask) == NodeResultBoolean;
+ return result() == NodeResultBoolean;
}
bool hasStorageResult()
{
- return (m_flags & NodeResultMask) == NodeResultStorage;
+ return result() == NodeResultStorage;
+ }
+
+ UseKind defaultUseKind()
+ {
+ return useKindForResult(result());
+ }
+
+ Edge defaultEdge()
+ {
+ return Edge(this, defaultUseKind());
}
bool isJump()
@@ -670,13 +1161,22 @@ struct Node {
{
return op() == Branch;
}
+
+ bool isSwitch()
+ {
+ return op() == Switch;
+ }
bool isTerminal()
{
switch (op()) {
case Jump:
case Branch:
+ case Switch:
case Return:
+ case TailCall:
+ case TailCallVarargs:
+ case TailCallForwardVarargs:
case Unreachable:
return true;
default:
@@ -684,40 +1184,36 @@ struct Node {
}
}
- unsigned takenBytecodeOffsetDuringParsing()
+ bool isFunctionTerminal()
{
- ASSERT(isBranch() || isJump());
- return m_opInfo;
+ if (isTerminal() && !numSuccessors())
+ return true;
+
+ return false;
}
- unsigned notTakenBytecodeOffsetDuringParsing()
+ unsigned targetBytecodeOffsetDuringParsing()
{
- ASSERT(isBranch());
- return m_opInfo2;
+ ASSERT(isJump());
+ return m_opInfo;
}
-
- void setTakenBlockIndex(BlockIndex blockIndex)
+
+ BasicBlock*& targetBlock()
{
- ASSERT(isBranch() || isJump());
- m_opInfo = blockIndex;
+ ASSERT(isJump());
+ return *bitwise_cast<BasicBlock**>(&m_opInfo);
}
- void setNotTakenBlockIndex(BlockIndex blockIndex)
+ BranchData* branchData()
{
ASSERT(isBranch());
- m_opInfo2 = blockIndex;
- }
-
- BlockIndex takenBlockIndex()
- {
- ASSERT(isBranch() || isJump());
- return m_opInfo;
+ return bitwise_cast<BranchData*>(m_opInfo);
}
- BlockIndex notTakenBlockIndex()
+ SwitchData* switchData()
{
- ASSERT(isBranch());
- return m_opInfo2;
+ ASSERT(isSwitch());
+ return bitwise_cast<SwitchData*>(m_opInfo);
}
unsigned numSuccessors()
@@ -727,51 +1223,142 @@ struct Node {
return 1;
case Branch:
return 2;
+ case Switch:
+ return switchData()->cases.size() + 1;
default:
return 0;
}
}
- BlockIndex successor(unsigned index)
+ BasicBlock*& successor(unsigned index)
{
+ if (isSwitch()) {
+ if (index < switchData()->cases.size())
+ return switchData()->cases[index].target.block;
+ RELEASE_ASSERT(index == switchData()->cases.size());
+ return switchData()->fallThrough.block;
+ }
switch (index) {
case 0:
- return takenBlockIndex();
+ if (isJump())
+ return targetBlock();
+ return branchData()->taken.block;
case 1:
- return notTakenBlockIndex();
+ return branchData()->notTaken.block;
default:
RELEASE_ASSERT_NOT_REACHED();
- return NoBlock;
+ return targetBlock();
+ }
+ }
+
+ class SuccessorsIterable {
+ public:
+ SuccessorsIterable()
+ : m_terminal(nullptr)
+ {
+ }
+
+ SuccessorsIterable(Node* terminal)
+ : m_terminal(terminal)
+ {
+ }
+
+ class iterator {
+ public:
+ iterator()
+ : m_terminal(nullptr)
+ , m_index(UINT_MAX)
+ {
+ }
+
+ iterator(Node* terminal, unsigned index)
+ : m_terminal(terminal)
+ , m_index(index)
+ {
+ }
+
+ BasicBlock* operator*()
+ {
+ return m_terminal->successor(m_index);
+ }
+
+ iterator& operator++()
+ {
+ m_index++;
+ return *this;
+ }
+
+ bool operator==(const iterator& other) const
+ {
+ return m_index == other.m_index;
+ }
+
+ bool operator!=(const iterator& other) const
+ {
+ return !(*this == other);
+ }
+ private:
+ Node* m_terminal;
+ unsigned m_index;
+ };
+
+ iterator begin()
+ {
+ return iterator(m_terminal, 0);
}
+
+ iterator end()
+ {
+ return iterator(m_terminal, m_terminal->numSuccessors());
+ }
+
+ size_t size() const { return m_terminal->numSuccessors(); }
+ BasicBlock* at(size_t index) const { return m_terminal->successor(index); }
+ BasicBlock* operator[](size_t index) const { return at(index); }
+
+ private:
+ Node* m_terminal;
+ };
+
+ SuccessorsIterable successors()
+ {
+ return SuccessorsIterable(this);
}
- BlockIndex successorForCondition(bool condition)
+ BasicBlock*& successorForCondition(bool condition)
{
- ASSERT(isBranch());
- return condition ? takenBlockIndex() : notTakenBlockIndex();
+ return branchData()->forCondition(condition);
}
bool hasHeapPrediction()
{
switch (op()) {
+ case ArithRound:
+ case ArithFloor:
+ case ArithCeil:
+ case GetDirectPname:
case GetById:
case GetByIdFlush:
case GetByVal:
- case GetMyArgumentByVal:
- case GetMyArgumentByValSafe:
case Call:
+ case TailCallInlinedCaller:
case Construct:
+ case CallVarargs:
+ case TailCallVarargsInlinedCaller:
+ case ConstructVarargs:
+ case CallForwardVarargs:
+ case TailCallForwardVarargsInlinedCaller:
case GetByOffset:
- case GetScopedVar:
- case Resolve:
- case ResolveBase:
- case ResolveBaseStrictPut:
- case ResolveGlobal:
+ case MultiGetByOffset:
+ case GetClosureVar:
+ case GetFromArguments:
case ArrayPop:
case ArrayPush:
case RegExpExec:
case RegExpTest:
case GetGlobalVar:
+ case GetGlobalLexicalVariable:
+ case StringReplace:
return true;
default:
return false;
@@ -783,48 +1370,95 @@ struct Node {
ASSERT(hasHeapPrediction());
return static_cast<SpeculatedType>(m_opInfo2);
}
-
- bool predictHeap(SpeculatedType prediction)
+
+ void setHeapPrediction(SpeculatedType prediction)
{
ASSERT(hasHeapPrediction());
-
- return mergeSpeculation(m_opInfo2, prediction);
+ m_opInfo2 = prediction;
}
- bool hasFunction()
+ bool hasCellOperand()
{
switch (op()) {
- case CheckFunction:
- case AllocationProfileWatchpoint:
+ case CheckCell:
+ case OverridesHasInstance:
+ case NewFunction:
+ case NewArrowFunction:
+ case NewGeneratorFunction:
+ case CreateActivation:
+ case MaterializeCreateActivation:
return true;
default:
return false;
}
}
- JSCell* function()
+ FrozenValue* cellOperand()
{
- ASSERT(hasFunction());
- JSCell* result = reinterpret_cast<JSFunction*>(m_opInfo);
- ASSERT(JSValue(result).isFunction());
- return result;
+ ASSERT(hasCellOperand());
+ return reinterpret_cast<FrozenValue*>(m_opInfo);
}
- bool hasExecutable()
+ template<typename T>
+ T castOperand()
{
- return op() == CheckExecutable;
+ return cellOperand()->cast<T>();
}
- ExecutableBase* executable()
+ void setCellOperand(FrozenValue* value)
{
- return jsCast<ExecutableBase*>(reinterpret_cast<JSCell*>(m_opInfo));
+ ASSERT(hasCellOperand());
+ m_opInfo = bitwise_cast<uintptr_t>(value);
+ }
+
+ bool hasWatchpointSet()
+ {
+ return op() == NotifyWrite;
+ }
+
+ WatchpointSet* watchpointSet()
+ {
+ ASSERT(hasWatchpointSet());
+ return reinterpret_cast<WatchpointSet*>(m_opInfo);
+ }
+
+ bool hasStoragePointer()
+ {
+ return op() == ConstantStoragePointer;
+ }
+
+ void* storagePointer()
+ {
+ ASSERT(hasStoragePointer());
+ return reinterpret_cast<void*>(m_opInfo);
}
- bool hasStructureTransitionData()
+ bool hasUidOperand()
+ {
+ return op() == CheckIdent;
+ }
+
+ UniquedStringImpl* uidOperand()
+ {
+ ASSERT(hasUidOperand());
+ return reinterpret_cast<UniquedStringImpl*>(m_opInfo);
+ }
+
+ bool hasTypeInfoOperand()
+ {
+ return op() == CheckTypeInfoFlags;
+ }
+
+ unsigned typeInfoOperand()
+ {
+ ASSERT(hasTypeInfoOperand() && m_opInfo <= UCHAR_MAX);
+ return static_cast<unsigned>(m_opInfo);
+ }
+
+ bool hasTransition()
{
switch (op()) {
case PutStructure:
- case PhantomPutStructure:
case AllocatePropertyStorage:
case ReallocatePropertyStorage:
return true;
@@ -833,17 +1467,18 @@ struct Node {
}
}
- StructureTransitionData& structureTransitionData()
+ Transition* transition()
{
- ASSERT(hasStructureTransitionData());
- return *reinterpret_cast<StructureTransitionData*>(m_opInfo);
+ ASSERT(hasTransition());
+ return reinterpret_cast<Transition*>(m_opInfo);
}
bool hasStructureSet()
{
switch (op()) {
case CheckStructure:
- case ForwardCheckStructure:
+ case CheckStructureImmediate:
+ case MaterializeNewObject:
return true;
default:
return false;
@@ -859,8 +1494,6 @@ struct Node {
bool hasStructure()
{
switch (op()) {
- case StructureTransitionWatchpoint:
- case ForwardStructureTransitionWatchpoint:
case ArrayifyToStructure:
case NewObject:
case NewStringObject:
@@ -878,36 +1511,140 @@ struct Node {
bool hasStorageAccessData()
{
- return op() == GetByOffset || op() == PutByOffset;
+ switch (op()) {
+ case GetByOffset:
+ case PutByOffset:
+ case GetGetterSetterByOffset:
+ return true;
+ default:
+ return false;
+ }
}
- unsigned storageAccessDataIndex()
+ StorageAccessData& storageAccessData()
{
ASSERT(hasStorageAccessData());
- return m_opInfo;
+ return *bitwise_cast<StorageAccessData*>(m_opInfo);
}
- bool hasFunctionDeclIndex()
+ bool hasMultiGetByOffsetData()
{
- return op() == NewFunction
- || op() == NewFunctionNoCheck;
+ return op() == MultiGetByOffset;
}
- unsigned functionDeclIndex()
+ MultiGetByOffsetData& multiGetByOffsetData()
{
- ASSERT(hasFunctionDeclIndex());
- return m_opInfo;
+ ASSERT(hasMultiGetByOffsetData());
+ return *reinterpret_cast<MultiGetByOffsetData*>(m_opInfo);
}
- bool hasFunctionExprIndex()
+ bool hasMultiPutByOffsetData()
{
- return op() == NewFunctionExpression;
+ return op() == MultiPutByOffset;
}
- unsigned functionExprIndex()
+ MultiPutByOffsetData& multiPutByOffsetData()
{
- ASSERT(hasFunctionExprIndex());
- return m_opInfo;
+ ASSERT(hasMultiPutByOffsetData());
+ return *reinterpret_cast<MultiPutByOffsetData*>(m_opInfo);
+ }
+
+ bool hasObjectMaterializationData()
+ {
+ switch (op()) {
+ case MaterializeNewObject:
+ case MaterializeCreateActivation:
+ return true;
+
+ default:
+ return false;
+ }
+ }
+
+ ObjectMaterializationData& objectMaterializationData()
+ {
+ ASSERT(hasObjectMaterializationData());
+ return *reinterpret_cast<ObjectMaterializationData*>(m_opInfo2);
+ }
+
+ bool isObjectAllocation()
+ {
+ switch (op()) {
+ case NewObject:
+ case MaterializeNewObject:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ bool isPhantomObjectAllocation()
+ {
+ switch (op()) {
+ case PhantomNewObject:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ bool isActivationAllocation()
+ {
+ switch (op()) {
+ case CreateActivation:
+ case MaterializeCreateActivation:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ bool isPhantomActivationAllocation()
+ {
+ switch (op()) {
+ case PhantomCreateActivation:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ bool isFunctionAllocation()
+ {
+ switch (op()) {
+ case NewArrowFunction:
+ case NewFunction:
+ case NewGeneratorFunction:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ bool isPhantomFunctionAllocation()
+ {
+ switch (op()) {
+ case PhantomNewFunction:
+ case PhantomNewGeneratorFunction:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ bool isPhantomAllocation()
+ {
+ switch (op()) {
+ case PhantomNewObject:
+ case PhantomDirectArguments:
+ case PhantomClonedArguments:
+ case PhantomNewFunction:
+ case PhantomNewGeneratorFunction:
+ case PhantomCreateActivation:
+ return true;
+ default:
+ return false;
+ }
}
bool hasArrayMode()
@@ -915,6 +1652,7 @@ struct Node {
switch (op()) {
case GetIndexedPropertyStorage:
case GetArrayLength:
+ case PutByValDirect:
case PutByVal:
case PutByValAlias:
case GetByVal:
@@ -925,6 +1663,7 @@ struct Node {
case ArrayifyToStructure:
case ArrayPush:
case ArrayPop:
+ case HasIndexedProperty:
return true;
default:
return false;
@@ -948,36 +1687,71 @@ struct Node {
return true;
}
+ bool hasArithMode()
+ {
+ switch (op()) {
+ case ArithAbs:
+ case ArithAdd:
+ case ArithSub:
+ case ArithNegate:
+ case ArithMul:
+ case ArithDiv:
+ case ArithMod:
+ case UInt32ToNumber:
+ case DoubleAsInt32:
+ return true;
+ default:
+ return false;
+ }
+ }
+
+ Arith::Mode arithMode()
+ {
+ ASSERT(hasArithMode());
+ return static_cast<Arith::Mode>(m_opInfo);
+ }
+
+ void setArithMode(Arith::Mode mode)
+ {
+ m_opInfo = mode;
+ }
+
+ bool hasArithRoundingMode()
+ {
+ return op() == ArithRound || op() == ArithFloor || op() == ArithCeil;
+ }
+
+ Arith::RoundingMode arithRoundingMode()
+ {
+ ASSERT(hasArithRoundingMode());
+ return static_cast<Arith::RoundingMode>(m_opInfo);
+ }
+
+ void setArithRoundingMode(Arith::RoundingMode mode)
+ {
+ ASSERT(hasArithRoundingMode());
+ m_opInfo = static_cast<uintptr_t>(mode);
+ }
+
bool hasVirtualRegister()
{
- return m_virtualRegister != InvalidVirtualRegister;
+ return m_virtualRegister.isValid();
}
VirtualRegister virtualRegister()
{
ASSERT(hasResult());
- ASSERT(m_virtualRegister != InvalidVirtualRegister);
+ ASSERT(m_virtualRegister.isValid());
return m_virtualRegister;
}
void setVirtualRegister(VirtualRegister virtualRegister)
{
ASSERT(hasResult());
- ASSERT(m_virtualRegister == InvalidVirtualRegister);
+ ASSERT(!m_virtualRegister.isValid());
m_virtualRegister = virtualRegister;
}
- bool hasArgumentPositionStart()
- {
- return op() == InlineStart;
- }
-
- unsigned argumentPositionStart()
- {
- ASSERT(hasArgumentPositionStart());
- return m_opInfo;
- }
-
bool hasExecutionCounter()
{
return op() == CountExecution;
@@ -993,27 +1767,9 @@ struct Node {
return m_refCount;
}
- bool willHaveCodeGenOrOSR()
+ bool isSemanticallySkippable()
{
- switch (op()) {
- case SetLocal:
- case MovHint:
- case ZombieHint:
- case MovHintAndCheck:
- case Int32ToDouble:
- case ForwardInt32ToDouble:
- case ValueToInt32:
- case UInt32ToNumber:
- case DoubleAsInt32:
- case PhantomArguments:
- return true;
- case Nop:
- return false;
- case Phantom:
- return child1().useKindUnchecked() != UntypedUse || child2().useKindUnchecked() != UntypedUse || child3().useKindUnchecked() != UntypedUse;
- default:
- return shouldGenerate();
- }
+ return op() == CountExecution;
}
unsigned refCount()
@@ -1080,9 +1836,25 @@ struct Node {
return child1().useKind();
}
+ bool isBinaryUseKind(UseKind left, UseKind right)
+ {
+ return child1().useKind() == left && child2().useKind() == right;
+ }
+
bool isBinaryUseKind(UseKind useKind)
{
- return child1().useKind() == useKind && child2().useKind() == useKind;
+ return isBinaryUseKind(useKind, useKind);
+ }
+
+ Edge childFor(UseKind useKind)
+ {
+ if (child1().useKind() == useKind)
+ return child1();
+ if (child2().useKind() == useKind)
+ return child2();
+ if (child3().useKind() == useKind)
+ return child3();
+ return Edge();
}
SpeculatedType prediction()
@@ -1095,19 +1867,39 @@ struct Node {
return mergeSpeculation(m_prediction, prediction);
}
- bool shouldSpeculateInteger()
+ bool shouldSpeculateInt32()
{
return isInt32Speculation(prediction());
}
- bool shouldSpeculateIntegerForArithmetic()
+ bool sawBooleans()
+ {
+ return !!(prediction() & SpecBoolean);
+ }
+
+ bool shouldSpeculateInt32OrBoolean()
+ {
+ return isInt32OrBooleanSpeculation(prediction());
+ }
+
+ bool shouldSpeculateInt32ForArithmetic()
{
return isInt32SpeculationForArithmetic(prediction());
}
- bool shouldSpeculateIntegerExpectingDefined()
+ bool shouldSpeculateInt32OrBooleanForArithmetic()
+ {
+ return isInt32OrBooleanSpeculationForArithmetic(prediction());
+ }
+
+ bool shouldSpeculateInt32OrBooleanExpectingDefined()
{
- return isInt32SpeculationExpectingDefined(prediction());
+ return isInt32OrBooleanSpeculationExpectingDefined(prediction());
+ }
+
+ bool shouldSpeculateMachineInt()
+ {
+ return isMachineIntSpeculation(prediction());
}
bool shouldSpeculateDouble()
@@ -1115,31 +1907,61 @@ struct Node {
return isDoubleSpeculation(prediction());
}
- bool shouldSpeculateDoubleForArithmetic()
+ bool shouldSpeculateDoubleReal()
{
- return isDoubleSpeculationForArithmetic(prediction());
+ return isDoubleRealSpeculation(prediction());
}
bool shouldSpeculateNumber()
{
- return isNumberSpeculation(prediction());
+ return isFullNumberSpeculation(prediction());
+ }
+
+ bool shouldSpeculateNumberOrBoolean()
+ {
+ return isFullNumberOrBooleanSpeculation(prediction());
}
- bool shouldSpeculateNumberExpectingDefined()
+ bool shouldSpeculateNumberOrBooleanExpectingDefined()
{
- return isNumberSpeculationExpectingDefined(prediction());
+ return isFullNumberOrBooleanSpeculationExpectingDefined(prediction());
}
bool shouldSpeculateBoolean()
{
return isBooleanSpeculation(prediction());
}
+
+ bool shouldSpeculateOther()
+ {
+ return isOtherSpeculation(prediction());
+ }
+
+ bool shouldSpeculateMisc()
+ {
+ return isMiscSpeculation(prediction());
+ }
+ bool shouldSpeculateStringIdent()
+ {
+ return isStringIdentSpeculation(prediction());
+ }
+
+ bool shouldSpeculateNotStringVar()
+ {
+ return isNotStringVarSpeculation(prediction());
+ }
+
bool shouldSpeculateString()
{
return isStringSpeculation(prediction());
}
+ bool shouldSpeculateStringOrOther()
+ {
+ return isStringOrOtherSpeculation(prediction());
+ }
+
bool shouldSpeculateStringObject()
{
return isStringObjectSpeculation(prediction());
@@ -1149,6 +1971,16 @@ struct Node {
{
return isStringOrStringObjectSpeculation(prediction());
}
+
+ bool shouldSpeculateRegExpObject()
+ {
+ return isRegExpObjectSpeculation(prediction());
+ }
+
+ bool shouldSpeculateSymbol()
+ {
+ return isSymbolSpeculation(prediction());
+ }
bool shouldSpeculateFinalObject()
{
@@ -1165,9 +1997,14 @@ struct Node {
return isArraySpeculation(prediction());
}
- bool shouldSpeculateArguments()
+ bool shouldSpeculateDirectArguments()
{
- return isArgumentsSpeculation(prediction());
+ return isDirectArgumentsSpeculation(prediction());
+ }
+
+ bool shouldSpeculateScopedArguments()
+ {
+ return isScopedArgumentsSpeculation(prediction());
}
bool shouldSpeculateInt8Array()
@@ -1235,29 +2072,67 @@ struct Node {
return isCellSpeculation(prediction());
}
+ bool shouldSpeculateCellOrOther()
+ {
+ return isCellOrOtherSpeculation(prediction());
+ }
+
+ bool shouldSpeculateNotCell()
+ {
+ return isNotCellSpeculation(prediction());
+ }
+
+ bool shouldSpeculateUntypedForArithmetic()
+ {
+ return isUntypedSpeculationForArithmetic(prediction());
+ }
+
+ static bool shouldSpeculateUntypedForArithmetic(Node* op1, Node* op2)
+ {
+ return op1->shouldSpeculateUntypedForArithmetic() || op2->shouldSpeculateUntypedForArithmetic();
+ }
+
+ bool shouldSpeculateUntypedForBitOps()
+ {
+ return isUntypedSpeculationForBitOps(prediction());
+ }
+
+ static bool shouldSpeculateUntypedForBitOps(Node* op1, Node* op2)
+ {
+ return op1->shouldSpeculateUntypedForBitOps() || op2->shouldSpeculateUntypedForBitOps();
+ }
+
static bool shouldSpeculateBoolean(Node* op1, Node* op2)
{
return op1->shouldSpeculateBoolean() && op2->shouldSpeculateBoolean();
}
- static bool shouldSpeculateInteger(Node* op1, Node* op2)
+ static bool shouldSpeculateInt32(Node* op1, Node* op2)
{
- return op1->shouldSpeculateInteger() && op2->shouldSpeculateInteger();
+ return op1->shouldSpeculateInt32() && op2->shouldSpeculateInt32();
}
- static bool shouldSpeculateIntegerForArithmetic(Node* op1, Node* op2)
+ static bool shouldSpeculateInt32OrBoolean(Node* op1, Node* op2)
{
- return op1->shouldSpeculateIntegerForArithmetic() && op2->shouldSpeculateIntegerForArithmetic();
+ return op1->shouldSpeculateInt32OrBoolean()
+ && op2->shouldSpeculateInt32OrBoolean();
}
- static bool shouldSpeculateIntegerExpectingDefined(Node* op1, Node* op2)
+ static bool shouldSpeculateInt32OrBooleanForArithmetic(Node* op1, Node* op2)
{
- return op1->shouldSpeculateIntegerExpectingDefined() && op2->shouldSpeculateIntegerExpectingDefined();
+ return op1->shouldSpeculateInt32OrBooleanForArithmetic()
+ && op2->shouldSpeculateInt32OrBooleanForArithmetic();
}
- static bool shouldSpeculateDoubleForArithmetic(Node* op1, Node* op2)
+ static bool shouldSpeculateInt32OrBooleanExpectingDefined(Node* op1, Node* op2)
{
- return op1->shouldSpeculateDoubleForArithmetic() && op2->shouldSpeculateDoubleForArithmetic();
+ return op1->shouldSpeculateInt32OrBooleanExpectingDefined()
+ && op2->shouldSpeculateInt32OrBooleanExpectingDefined();
+ }
+
+ static bool shouldSpeculateMachineInt(Node* op1, Node* op2)
+ {
+ return op1->shouldSpeculateMachineInt() && op2->shouldSpeculateMachineInt();
}
static bool shouldSpeculateNumber(Node* op1, Node* op2)
@@ -1265,9 +2140,21 @@ struct Node {
return op1->shouldSpeculateNumber() && op2->shouldSpeculateNumber();
}
- static bool shouldSpeculateNumberExpectingDefined(Node* op1, Node* op2)
+ static bool shouldSpeculateNumberOrBoolean(Node* op1, Node* op2)
{
- return op1->shouldSpeculateNumberExpectingDefined() && op2->shouldSpeculateNumberExpectingDefined();
+ return op1->shouldSpeculateNumberOrBoolean()
+ && op2->shouldSpeculateNumberOrBoolean();
+ }
+
+ static bool shouldSpeculateNumberOrBooleanExpectingDefined(Node* op1, Node* op2)
+ {
+ return op1->shouldSpeculateNumberOrBooleanExpectingDefined()
+ && op2->shouldSpeculateNumberOrBooleanExpectingDefined();
+ }
+
+ static bool shouldSpeculateSymbol(Node* op1, Node* op2)
+ {
+ return op1->shouldSpeculateSymbol() && op2->shouldSpeculateSymbol();
}
static bool shouldSpeculateFinalObject(Node* op1, Node* op2)
@@ -1280,11 +2167,81 @@ struct Node {
return op1->shouldSpeculateArray() && op2->shouldSpeculateArray();
}
- bool canSpeculateInteger()
+ bool canSpeculateInt32(RareCaseProfilingSource source)
+ {
+ return nodeCanSpeculateInt32(arithNodeFlags(), source);
+ }
+
+ bool canSpeculateInt52(RareCaseProfilingSource source)
+ {
+ return nodeCanSpeculateInt52(arithNodeFlags(), source);
+ }
+
+ RareCaseProfilingSource sourceFor(PredictionPass pass)
+ {
+ if (pass == PrimaryPass || child1()->sawBooleans() || (child2() && child2()->sawBooleans()))
+ return DFGRareCase;
+ return AllRareCases;
+ }
+
+ bool canSpeculateInt32(PredictionPass pass)
+ {
+ return canSpeculateInt32(sourceFor(pass));
+ }
+
+ bool canSpeculateInt52(PredictionPass pass)
+ {
+ return canSpeculateInt52(sourceFor(pass));
+ }
+
+ bool hasTypeLocation()
+ {
+ return op() == ProfileType;
+ }
+
+ TypeLocation* typeLocation()
+ {
+ ASSERT(hasTypeLocation());
+ return reinterpret_cast<TypeLocation*>(m_opInfo);
+ }
+
+ bool hasBasicBlockLocation()
+ {
+ return op() == ProfileControlFlow;
+ }
+
+ BasicBlockLocation* basicBlockLocation()
+ {
+ ASSERT(hasBasicBlockLocation());
+ return reinterpret_cast<BasicBlockLocation*>(m_opInfo);
+ }
+
+ Node* replacement() const
+ {
+ return m_misc.replacement;
+ }
+
+ void setReplacement(Node* replacement)
{
- return nodeCanSpeculateInteger(arithNodeFlags());
+ m_misc.replacement = replacement;
}
+ Epoch epoch() const
+ {
+ return Epoch::fromUnsigned(m_misc.epoch);
+ }
+
+ void setEpoch(Epoch epoch)
+ {
+ m_misc.epoch = epoch.toUnsigned();
+ }
+
+ unsigned numberOfArgumentsToSkip()
+ {
+ ASSERT(op() == CopyRest || op() == GetRestLength);
+ return static_cast<unsigned>(m_opInfo);
+ }
+
void dumpChildren(PrintStream& out)
{
if (!child1())
@@ -1299,9 +2256,9 @@ struct Node {
}
// NB. This class must have a trivial destructor.
-
- // Used to look up exception handling information (currently implemented as a bytecode index).
- CodeOrigin codeOrigin;
+
+ NodeOrigin origin;
+
// References to up to 3 children, or links to a variable length set of children.
AdjacencyList children;
@@ -1312,26 +2269,75 @@ private:
VirtualRegister m_virtualRegister;
// The number of uses of the result of this operation (+1 for 'must generate' nodes, which have side-effects).
unsigned m_refCount;
+ // The prediction ascribed to this node after propagation.
+ SpeculatedType m_prediction;
// Immediate values, accesses type-checked via accessors above. The first one is
// big enough to store a pointer.
uintptr_t m_opInfo;
- unsigned m_opInfo2;
- // The prediction ascribed to this node after propagation.
- SpeculatedType m_prediction;
+ uintptr_t m_opInfo2;
public:
// Fields used by various analyses.
AbstractValue value;
- Node* replacement;
+
+ // Miscellaneous data that is usually meaningless, but can hold some analysis results
+ // if you ask right. For example, if you do Graph::initializeNodeOwners(), Node::owner
+ // will tell you which basic block a node belongs to. You cannot rely on this persisting
+ // across transformations unless you do the maintenance work yourself. Other phases use
+ // Node::replacement, but they do so manually: first you do Graph::clearReplacements()
+ // and then you set, and use, replacement's yourself. Same thing for epoch.
+ //
+ // Bottom line: don't use these fields unless you initialize them yourself, or by
+ // calling some appropriate methods that initialize them the way you want. Otherwise,
+ // these fields are meaningless.
+private:
+ union {
+ Node* replacement;
+ unsigned epoch;
+ } m_misc;
+public:
+ BasicBlock* owner;
};
+inline bool nodeComparator(Node* a, Node* b)
+{
+ return a->index() < b->index();
+}
+
+template<typename T>
+CString nodeListDump(const T& nodeList)
+{
+ return sortedListDump(nodeList, nodeComparator);
+}
+
+template<typename T>
+CString nodeMapDump(const T& nodeMap, DumpContext* context = 0)
+{
+ Vector<typename T::KeyType> keys;
+ for (
+ typename T::const_iterator iter = nodeMap.begin();
+ iter != nodeMap.end(); ++iter)
+ keys.append(iter->key);
+ std::sort(keys.begin(), keys.end(), nodeComparator);
+ StringPrintStream out;
+ CommaPrinter comma;
+ for(unsigned i = 0; i < keys.size(); ++i)
+ out.print(comma, keys[i], "=>", inContext(nodeMap.get(keys[i]), context));
+ return out.toCString();
+}
+
} } // namespace JSC::DFG
namespace WTF {
+void printInternal(PrintStream&, JSC::DFG::SwitchKind);
void printInternal(PrintStream&, JSC::DFG::Node*);
+inline JSC::DFG::Node* inContext(JSC::DFG::Node* node, JSC::DumpContext*) { return node; }
+
} // namespace WTF
+using WTF::inContext;
+
#endif
#endif