Source code

Revision control

Copy as Markdown

Other Tools

/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*-
* vim: set ts=8 sts=2 et sw=2 tw=80:
*
* Copyright 2016 Mozilla Foundation
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef wasm_op_iter_h
#define wasm_op_iter_h
#include "mozilla/CompactPair.h"
#include "mozilla/Poison.h"
#include <type_traits>
#include "js/Printf.h"
#include "wasm/WasmBuiltinModule.h"
#include "wasm/WasmUtility.h"
#include "wasm/WasmValidate.h"
namespace js {
namespace wasm {
// The kind of a control-flow stack item.
enum class LabelKind : uint8_t {
Body,
Block,
Loop,
Then,
Else,
Try,
Catch,
CatchAll,
TryTable,
};
// The type of values on the operand stack during validation. This is either a
// ValType or the special type "Bottom".
class StackType {
PackedTypeCode tc_;
explicit StackType(PackedTypeCode tc) : tc_(tc) {}
public:
StackType() : tc_(PackedTypeCode::invalid()) {}
explicit StackType(const ValType& t) : tc_(t.packed()) {
MOZ_ASSERT(tc_.isValid());
MOZ_ASSERT(!isStackBottom());
}
static StackType bottom() {
return StackType(PackedTypeCode::pack(TypeCode::Limit));
}
bool isStackBottom() const {
MOZ_ASSERT(tc_.isValid());
return tc_.typeCode() == TypeCode::Limit;
}
// Returns whether this input is nullable when interpreted as an operand.
// When the type is bottom for unreachable code, this returns false as that
// is the most permissive option.
bool isNullableAsOperand() const {
MOZ_ASSERT(tc_.isValid());
return isStackBottom() ? false : tc_.isNullable();
}
ValType valType() const {
MOZ_ASSERT(tc_.isValid());
MOZ_ASSERT(!isStackBottom());
return ValType(tc_);
}
ValType valTypeOr(ValType ifBottom) const {
MOZ_ASSERT(tc_.isValid());
if (isStackBottom()) {
return ifBottom;
}
return valType();
}
ValType asNonNullable() const {
MOZ_ASSERT(tc_.isValid());
MOZ_ASSERT(!isStackBottom());
return ValType(tc_.withIsNullable(false));
}
bool isValidForUntypedSelect() const {
MOZ_ASSERT(tc_.isValid());
if (isStackBottom()) {
return true;
}
switch (valType().kind()) {
case ValType::I32:
case ValType::F32:
case ValType::I64:
case ValType::F64:
#ifdef ENABLE_WASM_SIMD
case ValType::V128:
#endif
return true;
default:
return false;
}
}
bool operator==(const StackType& that) const {
MOZ_ASSERT(tc_.isValid() && that.tc_.isValid());
return tc_ == that.tc_;
}
bool operator!=(const StackType& that) const {
MOZ_ASSERT(tc_.isValid() && that.tc_.isValid());
return tc_ != that.tc_;
}
};
#ifdef DEBUG
// Families of opcodes that share a signature and validation logic.
enum class OpKind {
Block,
Loop,
Unreachable,
Drop,
I32,
I64,
F32,
F64,
V128,
Br,
BrIf,
BrTable,
Nop,
Unary,
Binary,
Ternary,
Comparison,
Conversion,
Load,
Store,
TeeStore,
MemorySize,
MemoryGrow,
Select,
GetLocal,
SetLocal,
TeeLocal,
GetGlobal,
SetGlobal,
TeeGlobal,
Call,
ReturnCall,
CallIndirect,
ReturnCallIndirect,
# ifdef ENABLE_WASM_GC
CallRef,
ReturnCallRef,
# endif
OldCallDirect,
OldCallIndirect,
Return,
If,
Else,
End,
Wait,
Wake,
Fence,
AtomicLoad,
AtomicStore,
AtomicBinOp,
AtomicCompareExchange,
MemOrTableCopy,
DataOrElemDrop,
MemFill,
MemOrTableInit,
TableFill,
MemDiscard,
TableGet,
TableGrow,
TableSet,
TableSize,
RefNull,
RefFunc,
RefAsNonNull,
BrOnNull,
BrOnNonNull,
StructNew,
StructNewDefault,
StructGet,
StructSet,
ArrayNew,
ArrayNewFixed,
ArrayNewDefault,
ArrayNewData,
ArrayNewElem,
ArrayInitData,
ArrayInitElem,
ArrayGet,
ArraySet,
ArrayLen,
ArrayCopy,
ArrayFill,
RefTest,
RefCast,
BrOnCast,
RefConversion,
# ifdef ENABLE_WASM_SIMD
ExtractLane,
ReplaceLane,
LoadLane,
StoreLane,
VectorShift,
VectorShuffle,
# endif
Catch,
CatchAll,
Delegate,
Throw,
ThrowRef,
Rethrow,
Try,
TryTable,
CallBuiltinModuleFunc,
};
// Return the OpKind for a given Op. This is used for sanity-checking that
// API users use the correct read function for a given Op.
OpKind Classify(OpBytes op);
#endif
// Common fields for linear memory access.
template <typename Value>
struct LinearMemoryAddress {
Value base;
uint32_t memoryIndex;
uint64_t offset;
uint32_t align;
LinearMemoryAddress() : memoryIndex(0), offset(0), align(0) {}
LinearMemoryAddress(Value base, uint32_t memoryIndex, uint64_t offset,
uint32_t align)
: base(base), memoryIndex(memoryIndex), offset(offset), align(align) {}
};
template <typename ControlItem>
class ControlStackEntry {
// Use a pair to optimize away empty ControlItem.
mozilla::CompactPair<BlockType, ControlItem> typeAndItem_;
// The "base" of a control stack entry is valueStack_.length() minus
// type().params().length(), i.e., the size of the value stack "below"
// this block.
uint32_t valueStackBase_;
bool polymorphicBase_;
LabelKind kind_;
public:
ControlStackEntry(LabelKind kind, BlockType type, uint32_t valueStackBase)
: typeAndItem_(type, ControlItem()),
valueStackBase_(valueStackBase),
polymorphicBase_(false),
kind_(kind) {
MOZ_ASSERT(type != BlockType());
}
LabelKind kind() const { return kind_; }
BlockType type() const { return typeAndItem_.first(); }
ResultType resultType() const { return type().results(); }
ResultType branchTargetType() const {
return kind_ == LabelKind::Loop ? type().params() : type().results();
}
uint32_t valueStackBase() const { return valueStackBase_; }
ControlItem& controlItem() { return typeAndItem_.second(); }
void setPolymorphicBase() { polymorphicBase_ = true; }
bool polymorphicBase() const { return polymorphicBase_; }
void switchToElse() {
MOZ_ASSERT(kind() == LabelKind::Then);
kind_ = LabelKind::Else;
polymorphicBase_ = false;
}
void switchToCatch() {
MOZ_ASSERT(kind() == LabelKind::Try || kind() == LabelKind::Catch);
kind_ = LabelKind::Catch;
polymorphicBase_ = false;
}
void switchToCatchAll() {
MOZ_ASSERT(kind() == LabelKind::Try || kind() == LabelKind::Catch);
kind_ = LabelKind::CatchAll;
polymorphicBase_ = false;
}
};
// Track state of the non-defaultable locals. Every time such local is
// initialized, the stack will record at what depth and which local was set.
// On a block end, the "unset" state will be rolled back to how it was before
// the block started.
//
// It is very likely only a few functions will have non-defaultable locals and
// very few locals will be non-defaultable. This class is optimized to be fast
// for this common case.
class UnsetLocalsState {
struct SetLocalEntry {
uint32_t depth;
uint32_t localUnsetIndex;
SetLocalEntry(uint32_t depth_, uint32_t localUnsetIndex_)
: depth(depth_), localUnsetIndex(localUnsetIndex_) {}
};
using SetLocalsStack = Vector<SetLocalEntry, 16, SystemAllocPolicy>;
using UnsetLocals = Vector<uint32_t, 16, SystemAllocPolicy>;
static constexpr size_t WordSize = 4;
static constexpr size_t WordBits = WordSize * 8;
// Bit array of "unset" function locals. Stores only unset states of the
// locals that are declared after the first non-defaultable local.
UnsetLocals unsetLocals_;
// Stack of "set" operations. Contains pair where the first field is a depth,
// and the second field is local id (offset by firstNonDefaultLocal_).
SetLocalsStack setLocalsStack_;
uint32_t firstNonDefaultLocal_;
public:
UnsetLocalsState() : firstNonDefaultLocal_(UINT32_MAX) {}
[[nodiscard]] bool init(const ValTypeVector& locals, size_t numParams);
inline bool isUnset(uint32_t id) const {
if (MOZ_LIKELY(id < firstNonDefaultLocal_)) {
return false;
}
uint32_t localUnsetIndex = id - firstNonDefaultLocal_;
return unsetLocals_[localUnsetIndex / WordBits] &
(1 << (localUnsetIndex % WordBits));
}
inline void set(uint32_t id, uint32_t depth) {
MOZ_ASSERT(isUnset(id));
MOZ_ASSERT(id >= firstNonDefaultLocal_ &&
(id - firstNonDefaultLocal_) / WordBits < unsetLocals_.length());
uint32_t localUnsetIndex = id - firstNonDefaultLocal_;
unsetLocals_[localUnsetIndex / WordBits] ^= 1
<< (localUnsetIndex % WordBits);
// The setLocalsStack_ is reserved upfront in the UnsetLocalsState::init.
// A SetLocalEntry will be pushed only once per local.
setLocalsStack_.infallibleEmplaceBack(depth, localUnsetIndex);
}
inline void resetToBlock(uint32_t controlDepth) {
while (MOZ_UNLIKELY(setLocalsStack_.length() > 0) &&
setLocalsStack_.back().depth > controlDepth) {
uint32_t localUnsetIndex = setLocalsStack_.back().localUnsetIndex;
MOZ_ASSERT(!(unsetLocals_[localUnsetIndex / WordBits] &
(1 << (localUnsetIndex % WordBits))));
unsetLocals_[localUnsetIndex / WordBits] |=
1 << (localUnsetIndex % WordBits);
setLocalsStack_.popBack();
}
}
int empty() const { return setLocalsStack_.empty(); }
};
template <typename Value>
class TypeAndValueT {
// Use a Pair to optimize away empty Value.
mozilla::CompactPair<StackType, Value> tv_;
public:
TypeAndValueT() : tv_(StackType::bottom(), Value()) {}
explicit TypeAndValueT(StackType type) : tv_(type, Value()) {}
explicit TypeAndValueT(ValType type) : tv_(StackType(type), Value()) {}
TypeAndValueT(StackType type, Value value) : tv_(type, value) {}
TypeAndValueT(ValType type, Value value) : tv_(StackType(type), value) {}
StackType type() const { return tv_.first(); }
void setType(StackType type) { tv_.first() = type; }
Value value() const { return tv_.second(); }
void setValue(Value value) { tv_.second() = value; }
};
// An iterator over the bytes of a function body. It performs validation
// and unpacks the data into a usable form.
//
// The MOZ_STACK_CLASS attribute here is because of the use of DebugOnly.
// There's otherwise nothing inherent in this class which would require
// it to be used on the stack.
template <typename Policy>
class MOZ_STACK_CLASS OpIter : private Policy {
public:
using Value = typename Policy::Value;
using ValueVector = typename Policy::ValueVector;
using TypeAndValue = TypeAndValueT<Value>;
using TypeAndValueStack = Vector<TypeAndValue, 32, SystemAllocPolicy>;
using ControlItem = typename Policy::ControlItem;
using Control = ControlStackEntry<ControlItem>;
using ControlStack = Vector<Control, 16, SystemAllocPolicy>;
enum Kind {
Func,
InitExpr,
};
private:
Kind kind_;
Decoder& d_;
const ModuleEnvironment& env_;
TypeAndValueStack valueStack_;
TypeAndValueStack elseParamStack_;
ControlStack controlStack_;
UnsetLocalsState unsetLocals_;
// The exclusive max index of a global that can be accessed by global.get in
// this expression. When GC is enabled, this is any previously defined
// immutable global. Otherwise this is always set to zero, and only imported
// immutable globals are allowed.
uint32_t maxInitializedGlobalsIndexPlus1_;
FeatureUsage featureUsage_;
uint32_t lastBranchHintIndex_;
BranchHintVector* branchHintVector_;
#ifdef DEBUG
OpBytes op_;
#endif
size_t offsetOfLastReadOp_;
[[nodiscard]] bool readFixedU8(uint8_t* out) { return d_.readFixedU8(out); }
[[nodiscard]] bool readFixedU32(uint32_t* out) {
return d_.readFixedU32(out);
}
[[nodiscard]] bool readVarS32(int32_t* out) { return d_.readVarS32(out); }
[[nodiscard]] bool readVarU32(uint32_t* out) { return d_.readVarU32(out); }
[[nodiscard]] bool readVarS64(int64_t* out) { return d_.readVarS64(out); }
[[nodiscard]] bool readVarU64(uint64_t* out) { return d_.readVarU64(out); }
[[nodiscard]] bool readFixedF32(float* out) { return d_.readFixedF32(out); }
[[nodiscard]] bool readFixedF64(double* out) { return d_.readFixedF64(out); }
[[nodiscard]] bool readLinearMemoryAddress(uint32_t byteSize,
LinearMemoryAddress<Value>* addr);
[[nodiscard]] bool readLinearMemoryAddressAligned(
uint32_t byteSize, LinearMemoryAddress<Value>* addr);
[[nodiscard]] bool readBlockType(BlockType* type);
[[nodiscard]] bool readGcTypeIndex(uint32_t* typeIndex);
[[nodiscard]] bool readStructTypeIndex(uint32_t* typeIndex);
[[nodiscard]] bool readArrayTypeIndex(uint32_t* typeIndex);
[[nodiscard]] bool readFuncTypeIndex(uint32_t* typeIndex);
[[nodiscard]] bool readFieldIndex(uint32_t* fieldIndex,
const StructType& structType);
[[nodiscard]] bool popCallArgs(const ValTypeVector& expectedTypes,
ValueVector* values);
[[nodiscard]] bool failEmptyStack();
[[nodiscard]] bool popStackType(StackType* type, Value* value);
[[nodiscard]] bool popWithType(ValType expected, Value* value,
StackType* stackType);
[[nodiscard]] bool popWithType(ValType expected, Value* value);
[[nodiscard]] bool popWithType(ResultType expected, ValueVector* values);
template <typename ValTypeSpanT>
[[nodiscard]] bool popWithTypes(ValTypeSpanT expected, ValueVector* values);
[[nodiscard]] bool popWithRefType(Value* value, StackType* type);
// Check that the top of the value stack has type `expected`, bearing in
// mind that it may be a block type, hence involving multiple values.
//
// If the block's stack contains polymorphic values at its base (because we
// are in unreachable code) then suitable extra values are inserted into the
// value stack, as controlled by `rewriteStackTypes`: if this is true,
// polymorphic values have their types created/updated from `expected`. If
// it is false, such values are left as `StackType::bottom()`.
//
// If `values` is non-null, it is filled in with Value components of the
// relevant stack entries, including those of any new entries created.
[[nodiscard]] bool checkTopTypeMatches(ResultType expected,
ValueVector* values,
bool rewriteStackTypes);
[[nodiscard]] bool pushControl(LabelKind kind, BlockType type);
[[nodiscard]] bool checkStackAtEndOfBlock(ResultType* type,
ValueVector* values);
[[nodiscard]] bool getControl(uint32_t relativeDepth, Control** controlEntry);
[[nodiscard]] bool checkBranchValueAndPush(uint32_t relativeDepth,
ResultType* type,
ValueVector* values,
bool rewriteStackTypes);
[[nodiscard]] bool checkBrTableEntryAndPush(uint32_t* relativeDepth,
ResultType prevBranchType,
ResultType* branchType,
ValueVector* branchValues);
[[nodiscard]] bool push(StackType t) { return valueStack_.emplaceBack(t); }
[[nodiscard]] bool push(ValType t) { return valueStack_.emplaceBack(t); }
[[nodiscard]] bool push(TypeAndValue tv) { return valueStack_.append(tv); }
[[nodiscard]] bool push(ResultType t) {
for (size_t i = 0; i < t.length(); i++) {
if (!push(t[i])) {
return false;
}
}
return true;
}
void infalliblePush(StackType t) { valueStack_.infallibleEmplaceBack(t); }
void infalliblePush(ValType t) {
valueStack_.infallibleEmplaceBack(StackType(t));
}
void infalliblePush(TypeAndValue tv) { valueStack_.infallibleAppend(tv); }
void afterUnconditionalBranch() {
valueStack_.shrinkTo(controlStack_.back().valueStackBase());
controlStack_.back().setPolymorphicBase();
}
inline bool checkIsSubtypeOf(StorageType actual, StorageType expected);
inline bool checkIsSubtypeOf(RefType actual, RefType expected) {
return checkIsSubtypeOf(ValType(actual).storageType(),
ValType(expected).storageType());
}
inline bool checkIsSubtypeOf(ValType actual, ValType expected) {
return checkIsSubtypeOf(actual.storageType(), expected.storageType());
}
inline bool checkIsSubtypeOf(ResultType params, ResultType results);
#ifdef ENABLE_WASM_GC
inline bool checkIsSubtypeOf(uint32_t actualTypeIndex,
uint32_t expectedTypeIndex);
#endif
public:
#ifdef DEBUG
explicit OpIter(const ModuleEnvironment& env, Decoder& decoder,
Kind kind = OpIter::Func)
: kind_(kind),
d_(decoder),
env_(env),
maxInitializedGlobalsIndexPlus1_(0),
featureUsage_(FeatureUsage::None),
branchHintVector_(nullptr),
op_(OpBytes(Op::Limit)),
offsetOfLastReadOp_(0) {}
#else
explicit OpIter(const ModuleEnvironment& env, Decoder& decoder,
Kind kind = OpIter::Func)
: kind_(kind),
d_(decoder),
env_(env),
maxInitializedGlobalsIndexPlus1_(0),
featureUsage_(FeatureUsage::None),
offsetOfLastReadOp_(0) {}
#endif
FeatureUsage featureUsage() const { return featureUsage_; }
// Return the decoding byte offset.
uint32_t currentOffset() const { return d_.currentOffset(); }
// Return the offset within the entire module of the last-read op.
size_t lastOpcodeOffset() const {
return offsetOfLastReadOp_ ? offsetOfLastReadOp_ : d_.currentOffset();
}
// Return a BytecodeOffset describing where the current op should be reported
// to trap/call.
BytecodeOffset bytecodeOffset() const {
return BytecodeOffset(lastOpcodeOffset());
}
// Test whether the iterator has reached the end of the buffer.
bool done() const { return d_.done(); }
// Return a pointer to the end of the buffer being decoded by this iterator.
const uint8_t* end() const { return d_.end(); }
// Report a general failure.
[[nodiscard]] bool fail(const char* msg) MOZ_COLD;
// Report a general failure with a context
[[nodiscard]] bool fail_ctx(const char* fmt, const char* context) MOZ_COLD;
// Report an unrecognized opcode.
[[nodiscard]] bool unrecognizedOpcode(const OpBytes* expr) MOZ_COLD;
// Return whether the innermost block has a polymorphic base of its stack.
// Ideally this accessor would be removed; consider using something else.
bool currentBlockHasPolymorphicBase() const {
return !controlStack_.empty() && controlStack_.back().polymorphicBase();
}
// If it exists, return the BranchHint value from a function index and a
// branch offset.
// Branch hints are stored in a sorted vector. Because code in compiled in
// order, we keep track of the most recently accessed index.
// Retrieving branch hints is also done in order inside a function.
BranchHint getBranchHint(uint32_t funcIndex, uint32_t branchOffset) {
if (!env_.branchHintingEnabled()) {
return BranchHint::Invalid;
}
// Get the next hint in the collection
while (lastBranchHintIndex_ < branchHintVector_->length() &&
(*branchHintVector_)[lastBranchHintIndex_].branchOffset <
branchOffset) {
lastBranchHintIndex_++;
}
// No hint found for this branch.
if (lastBranchHintIndex_ >= branchHintVector_->length()) {
return BranchHint::Invalid;
}
// The last index is saved, now return the hint.
return (*branchHintVector_)[lastBranchHintIndex_].value;
}
// ------------------------------------------------------------------------
// Decoding and validation interface.
// Initialization and termination
[[nodiscard]] bool startFunction(uint32_t funcIndex,
const ValTypeVector& locals);
[[nodiscard]] bool endFunction(const uint8_t* bodyEnd);
[[nodiscard]] bool startInitExpr(ValType expected);
[[nodiscard]] bool endInitExpr();
// Value and reference types
[[nodiscard]] bool readValType(ValType* type);
[[nodiscard]] bool readHeapType(bool nullable, RefType* type);
// Instructions
[[nodiscard]] bool readOp(OpBytes* op);
[[nodiscard]] bool readReturn(ValueVector* values);
[[nodiscard]] bool readBlock(ResultType* paramType);
[[nodiscard]] bool readLoop(ResultType* paramType);
[[nodiscard]] bool readIf(ResultType* paramType, Value* condition);
[[nodiscard]] bool readElse(ResultType* paramType, ResultType* resultType,
ValueVector* thenResults);
[[nodiscard]] bool readEnd(LabelKind* kind, ResultType* type,
ValueVector* results,
ValueVector* resultsForEmptyElse);
void popEnd();
[[nodiscard]] bool readBr(uint32_t* relativeDepth, ResultType* type,
ValueVector* values);
[[nodiscard]] bool readBrIf(uint32_t* relativeDepth, ResultType* type,
ValueVector* values, Value* condition);
[[nodiscard]] bool readBrTable(Uint32Vector* depths, uint32_t* defaultDepth,
ResultType* defaultBranchType,
ValueVector* branchValues, Value* index);
[[nodiscard]] bool readTry(ResultType* type);
[[nodiscard]] bool readTryTable(ResultType* type,
TryTableCatchVector* catches);
[[nodiscard]] bool readCatch(LabelKind* kind, uint32_t* tagIndex,
ResultType* paramType, ResultType* resultType,
ValueVector* tryResults);
[[nodiscard]] bool readCatchAll(LabelKind* kind, ResultType* paramType,
ResultType* resultType,
ValueVector* tryResults);
[[nodiscard]] bool readDelegate(uint32_t* relativeDepth,
ResultType* resultType,
ValueVector* tryResults);
void popDelegate();
[[nodiscard]] bool readThrow(uint32_t* tagIndex, ValueVector* argValues);
[[nodiscard]] bool readThrowRef(Value* exnRef);
[[nodiscard]] bool readRethrow(uint32_t* relativeDepth);
[[nodiscard]] bool readUnreachable();
[[nodiscard]] bool readDrop();
[[nodiscard]] bool readUnary(ValType operandType, Value* input);
[[nodiscard]] bool readConversion(ValType operandType, ValType resultType,
Value* input);
[[nodiscard]] bool readBinary(ValType operandType, Value* lhs, Value* rhs);
[[nodiscard]] bool readComparison(ValType operandType, Value* lhs,
Value* rhs);
[[nodiscard]] bool readTernary(ValType operandType, Value* v0, Value* v1,
Value* v2);
[[nodiscard]] bool readLoad(ValType resultType, uint32_t byteSize,
LinearMemoryAddress<Value>* addr);
[[nodiscard]] bool readStore(ValType resultType, uint32_t byteSize,
LinearMemoryAddress<Value>* addr, Value* value);
[[nodiscard]] bool readTeeStore(ValType resultType, uint32_t byteSize,
LinearMemoryAddress<Value>* addr,
Value* value);
[[nodiscard]] bool readNop();
[[nodiscard]] bool readMemorySize(uint32_t* memoryIndex);
[[nodiscard]] bool readMemoryGrow(uint32_t* memoryIndex, Value* input);
[[nodiscard]] bool readSelect(bool typed, StackType* type, Value* trueValue,
Value* falseValue, Value* condition);
[[nodiscard]] bool readGetLocal(const ValTypeVector& locals, uint32_t* id);
[[nodiscard]] bool readSetLocal(const ValTypeVector& locals, uint32_t* id,
Value* value);
[[nodiscard]] bool readTeeLocal(const ValTypeVector& locals, uint32_t* id,
Value* value);
[[nodiscard]] bool readGetGlobal(uint32_t* id);
[[nodiscard]] bool readSetGlobal(uint32_t* id, Value* value);
[[nodiscard]] bool readTeeGlobal(uint32_t* id, Value* value);
[[nodiscard]] bool readI32Const(int32_t* i32);
[[nodiscard]] bool readI64Const(int64_t* i64);
[[nodiscard]] bool readF32Const(float* f32);
[[nodiscard]] bool readF64Const(double* f64);
[[nodiscard]] bool readRefFunc(uint32_t* funcIndex);
[[nodiscard]] bool readRefNull(RefType* type);
[[nodiscard]] bool readRefIsNull(Value* input);
[[nodiscard]] bool readRefAsNonNull(Value* input);
[[nodiscard]] bool readBrOnNull(uint32_t* relativeDepth, ResultType* type,
ValueVector* values, Value* condition);
[[nodiscard]] bool readBrOnNonNull(uint32_t* relativeDepth, ResultType* type,
ValueVector* values, Value* condition);
[[nodiscard]] bool readCall(uint32_t* funcTypeIndex, ValueVector* argValues);
[[nodiscard]] bool readCallIndirect(uint32_t* funcTypeIndex,
uint32_t* tableIndex, Value* callee,
ValueVector* argValues);
#ifdef ENABLE_WASM_TAIL_CALLS
[[nodiscard]] bool readReturnCall(uint32_t* funcTypeIndex,
ValueVector* argValues);
[[nodiscard]] bool readReturnCallIndirect(uint32_t* funcTypeIndex,
uint32_t* tableIndex, Value* callee,
ValueVector* argValues);
#endif
#ifdef ENABLE_WASM_GC
[[nodiscard]] bool readCallRef(const FuncType** funcType, Value* callee,
ValueVector* argValues);
# ifdef ENABLE_WASM_TAIL_CALLS
[[nodiscard]] bool readReturnCallRef(const FuncType** funcType, Value* callee,
ValueVector* argValues);
# endif
#endif
[[nodiscard]] bool readOldCallDirect(uint32_t numFuncImports,
uint32_t* funcTypeIndex,
ValueVector* argValues);
[[nodiscard]] bool readOldCallIndirect(uint32_t* funcTypeIndex, Value* callee,
ValueVector* argValues);
[[nodiscard]] bool readWake(LinearMemoryAddress<Value>* addr, Value* count);
[[nodiscard]] bool readWait(LinearMemoryAddress<Value>* addr,
ValType valueType, uint32_t byteSize,
Value* value, Value* timeout);
[[nodiscard]] bool readFence();
[[nodiscard]] bool readAtomicLoad(LinearMemoryAddress<Value>* addr,
ValType resultType, uint32_t byteSize);
[[nodiscard]] bool readAtomicStore(LinearMemoryAddress<Value>* addr,
ValType resultType, uint32_t byteSize,
Value* value);
[[nodiscard]] bool readAtomicRMW(LinearMemoryAddress<Value>* addr,
ValType resultType, uint32_t byteSize,
Value* value);
[[nodiscard]] bool readAtomicCmpXchg(LinearMemoryAddress<Value>* addr,
ValType resultType, uint32_t byteSize,
Value* oldValue, Value* newValue);
[[nodiscard]] bool readMemOrTableCopy(bool isMem,
uint32_t* dstMemOrTableIndex,
Value* dst,
uint32_t* srcMemOrTableIndex,
Value* src, Value* len);
[[nodiscard]] bool readDataOrElemDrop(bool isData, uint32_t* segIndex);
[[nodiscard]] bool readMemFill(uint32_t* memoryIndex, Value* start,
Value* val, Value* len);
[[nodiscard]] bool readMemOrTableInit(bool isMem, uint32_t* segIndex,
uint32_t* dstMemOrTableIndex,
Value* dst, Value* src, Value* len);
[[nodiscard]] bool readTableFill(uint32_t* tableIndex, Value* start,
Value* val, Value* len);
[[nodiscard]] bool readMemDiscard(uint32_t* memoryIndex, Value* start,
Value* len);
[[nodiscard]] bool readTableGet(uint32_t* tableIndex, Value* index);
[[nodiscard]] bool readTableGrow(uint32_t* tableIndex, Value* initValue,
Value* delta);
[[nodiscard]] bool readTableSet(uint32_t* tableIndex, Value* index,
Value* value);
[[nodiscard]] bool readTableSize(uint32_t* tableIndex);
#ifdef ENABLE_WASM_GC
[[nodiscard]] bool readStructNew(uint32_t* typeIndex, ValueVector* argValues);
[[nodiscard]] bool readStructNewDefault(uint32_t* typeIndex);
[[nodiscard]] bool readStructGet(uint32_t* typeIndex, uint32_t* fieldIndex,
FieldWideningOp wideningOp, Value* ptr);
[[nodiscard]] bool readStructSet(uint32_t* typeIndex, uint32_t* fieldIndex,
Value* ptr, Value* val);
[[nodiscard]] bool readArrayNew(uint32_t* typeIndex, Value* numElements,
Value* argValue);
[[nodiscard]] bool readArrayNewFixed(uint32_t* typeIndex,
uint32_t* numElements,
ValueVector* values);
[[nodiscard]] bool readArrayNewDefault(uint32_t* typeIndex,
Value* numElements);
[[nodiscard]] bool readArrayNewData(uint32_t* typeIndex, uint32_t* segIndex,
Value* offset, Value* numElements);
[[nodiscard]] bool readArrayNewElem(uint32_t* typeIndex, uint32_t* segIndex,
Value* offset, Value* numElements);
[[nodiscard]] bool readArrayInitData(uint32_t* typeIndex, uint32_t* segIndex,
Value* array, Value* arrayIndex,
Value* segOffset, Value* length);
[[nodiscard]] bool readArrayInitElem(uint32_t* typeIndex, uint32_t* segIndex,
Value* array, Value* arrayIndex,
Value* segOffset, Value* length);
[[nodiscard]] bool readArrayGet(uint32_t* typeIndex,
FieldWideningOp wideningOp, Value* index,
Value* ptr);
[[nodiscard]] bool readArraySet(uint32_t* typeIndex, Value* val, Value* index,
Value* ptr);
[[nodiscard]] bool readArrayLen(Value* ptr);
[[nodiscard]] bool readArrayCopy(int32_t* elemSize, bool* elemsAreRefTyped,
Value* dstArray, Value* dstIndex,
Value* srcArray, Value* srcIndex,
Value* numElements);
[[nodiscard]] bool readArrayFill(uint32_t* typeIndex, Value* array,
Value* index, Value* val, Value* length);
[[nodiscard]] bool readRefTest(bool nullable, RefType* sourceType,
RefType* destType, Value* ref);
[[nodiscard]] bool readRefCast(bool nullable, RefType* sourceType,
RefType* destType, Value* ref);
[[nodiscard]] bool readBrOnCast(bool onSuccess, uint32_t* labelRelativeDepth,
RefType* sourceType, RefType* destType,
ResultType* labelType, ValueVector* values);
[[nodiscard]] bool readRefConversion(RefType operandType, RefType resultType,
Value* operandValue);
#endif
#ifdef ENABLE_WASM_SIMD
[[nodiscard]] bool readLaneIndex(uint32_t inputLanes, uint32_t* laneIndex);
[[nodiscard]] bool readExtractLane(ValType resultType, uint32_t inputLanes,
uint32_t* laneIndex, Value* input);
[[nodiscard]] bool readReplaceLane(ValType operandType, uint32_t inputLanes,
uint32_t* laneIndex, Value* baseValue,
Value* operand);
[[nodiscard]] bool readVectorShift(Value* baseValue, Value* shift);
[[nodiscard]] bool readVectorShuffle(Value* v1, Value* v2, V128* selectMask);
[[nodiscard]] bool readV128Const(V128* value);
[[nodiscard]] bool readLoadSplat(uint32_t byteSize,
LinearMemoryAddress<Value>* addr);
[[nodiscard]] bool readLoadExtend(LinearMemoryAddress<Value>* addr);
[[nodiscard]] bool readLoadLane(uint32_t byteSize,
LinearMemoryAddress<Value>* addr,
uint32_t* laneIndex, Value* input);
[[nodiscard]] bool readStoreLane(uint32_t byteSize,
LinearMemoryAddress<Value>* addr,
uint32_t* laneIndex, Value* input);
#endif
[[nodiscard]] bool readCallBuiltinModuleFunc(
const BuiltinModuleFunc** builtinModuleFunc, ValueVector* params);
// At a location where readOp is allowed, peek at the next opcode
// without consuming it or updating any internal state.
// Never fails: returns uint16_t(Op::Limit) in op->b0 if it can't read.
void peekOp(OpBytes* op);
// ------------------------------------------------------------------------
// Stack management.
// Set the top N result values.
void setResults(size_t count, const ValueVector& values) {
MOZ_ASSERT(valueStack_.length() >= count);
size_t base = valueStack_.length() - count;
for (size_t i = 0; i < count; i++) {
valueStack_[base + i].setValue(values[i]);
}
}
bool getResults(size_t count, ValueVector* values) {
MOZ_ASSERT(valueStack_.length() >= count);
if (!values->resize(count)) {
return false;
}
size_t base = valueStack_.length() - count;
for (size_t i = 0; i < count; i++) {
(*values)[i] = valueStack_[base + i].value();
}
return true;
}
// Set the result value of the current top-of-value-stack expression.
void setResult(Value value) { valueStack_.back().setValue(value); }
// Return the result value of the current top-of-value-stack expression.
Value getResult() { return valueStack_.back().value(); }
// Return a reference to the top of the control stack.
ControlItem& controlItem() { return controlStack_.back().controlItem(); }
// Return a reference to an element in the control stack.
ControlItem& controlItem(uint32_t relativeDepth) {
return controlStack_[controlStack_.length() - 1 - relativeDepth]
.controlItem();
}
// Return the LabelKind of an element in the control stack.
LabelKind controlKind(uint32_t relativeDepth) {
return controlStack_[controlStack_.length() - 1 - relativeDepth].kind();
}
// Return a reference to the outermost element on the control stack.
ControlItem& controlOutermost() { return controlStack_[0].controlItem(); }
// Test whether the control-stack is empty, meaning we've consumed the final
// end of the function body.
bool controlStackEmpty() const { return controlStack_.empty(); }
// Return the depth of the control stack.
size_t controlStackDepth() const { return controlStack_.length(); }
// Find the innermost control item matching a predicate, starting to search
// from a certain relative depth, and returning true if such innermost
// control item is found. The relative depth of the found item is returned
// via a parameter.
template <typename Predicate>
bool controlFindInnermostFrom(Predicate predicate, uint32_t fromRelativeDepth,
uint32_t* foundRelativeDepth) {
int32_t fromAbsoluteDepth = controlStack_.length() - fromRelativeDepth - 1;
for (int32_t i = fromAbsoluteDepth; i >= 0; i--) {
if (predicate(controlStack_[i].kind(), controlStack_[i].controlItem())) {
*foundRelativeDepth = controlStack_.length() - 1 - i;
return true;
}
}
return false;
}
};
template <typename Policy>
inline bool OpIter<Policy>::checkIsSubtypeOf(StorageType subType,
StorageType superType) {
return CheckIsSubtypeOf(d_, env_, lastOpcodeOffset(), subType, superType);
}
template <typename Policy>
inline bool OpIter<Policy>::checkIsSubtypeOf(ResultType params,
ResultType results) {
if (params.length() != results.length()) {
UniqueChars error(
JS_smprintf("type mismatch: expected %zu values, got %zu values",
results.length(), params.length()));
if (!error) {
return false;
}
return fail(error.get());
}
for (uint32_t i = 0; i < params.length(); i++) {
ValType param = params[i];
ValType result = results[i];
if (!checkIsSubtypeOf(param, result)) {
return false;
}
}
return true;
}
#ifdef ENABLE_WASM_GC
template <typename Policy>
inline bool OpIter<Policy>::checkIsSubtypeOf(uint32_t actualTypeIndex,
uint32_t expectedTypeIndex) {
const TypeDef& actualTypeDef = env_.types->type(actualTypeIndex);
const TypeDef& expectedTypeDef = env_.types->type(expectedTypeIndex);
return CheckIsSubtypeOf(
d_, env_, lastOpcodeOffset(),
ValType(RefType::fromTypeDef(&actualTypeDef, true)),
ValType(RefType::fromTypeDef(&expectedTypeDef, true)));
}
#endif
template <typename Policy>
inline bool OpIter<Policy>::unrecognizedOpcode(const OpBytes* expr) {
UniqueChars error(JS_smprintf("unrecognized opcode: %x %x", expr->b0,
IsPrefixByte(expr->b0) ? expr->b1 : 0));
if (!error) {
return false;
}
return fail(error.get());
}
template <typename Policy>
inline bool OpIter<Policy>::fail(const char* msg) {
return d_.fail(lastOpcodeOffset(), msg);
}
template <typename Policy>
inline bool OpIter<Policy>::fail_ctx(const char* fmt, const char* context) {
UniqueChars error(JS_smprintf(fmt, context));
if (!error) {
return false;
}
return fail(error.get());
}
template <typename Policy>
inline bool OpIter<Policy>::failEmptyStack() {
return valueStack_.empty() ? fail("popping value from empty stack")
: fail("popping value from outside block");
}
// This function pops exactly one value from the stack, yielding Bottom types in
// various cases and therefore making it the caller's responsibility to do the
// right thing for StackType::Bottom. Prefer (pop|top)WithType. This is an
// optimization for the super-common case where the caller is statically
// expecting the resulttype `[valtype]`.
template <typename Policy>
inline bool OpIter<Policy>::popStackType(StackType* type, Value* value) {
Control& block = controlStack_.back();
MOZ_ASSERT(valueStack_.length() >= block.valueStackBase());
if (MOZ_UNLIKELY(valueStack_.length() == block.valueStackBase())) {
// If the base of this block's stack is polymorphic, then we can pop a
// dummy value of the bottom type; it won't be used since we're in
// unreachable code.
if (block.polymorphicBase()) {
*type = StackType::bottom();
*value = Value();
// Maintain the invariant that, after a pop, there is always memory
// reserved to push a value infallibly.
return valueStack_.reserve(valueStack_.length() + 1);
}
return failEmptyStack();
}
TypeAndValue& tv = valueStack_.back();
*type = tv.type();
*value = tv.value();
valueStack_.popBack();
return true;
}
// This function pops exactly one value from the stack, checking that it has the
// expected type which can either be a specific value type or the bottom type.
template <typename Policy>
inline bool OpIter<Policy>::popWithType(ValType expectedType, Value* value,
StackType* stackType) {
if (!popStackType(stackType, value)) {
return false;
}
return stackType->isStackBottom() ||
checkIsSubtypeOf(stackType->valType(), expectedType);
}
// This function pops exactly one value from the stack, checking that it has the
// expected type which can either be a specific value type or the bottom type.
template <typename Policy>
inline bool OpIter<Policy>::popWithType(ValType expectedType, Value* value) {
StackType stackType;
return popWithType(expectedType, value, &stackType);
}
template <typename Policy>
inline bool OpIter<Policy>::popWithType(ResultType expected,
ValueVector* values) {
return popWithTypes(expected, values);
}
// Pops each of the given expected types (in reverse, because it's a stack).
template <typename Policy>
template <typename ValTypeSpanT>
inline bool OpIter<Policy>::popWithTypes(ValTypeSpanT expected,
ValueVector* values) {
size_t expectedLength = expected.size();
if (!values->resize(expectedLength)) {
return false;
}
for (size_t i = 0; i < expectedLength; i++) {
size_t reverseIndex = expectedLength - i - 1;
ValType expectedType = expected[reverseIndex];
Value* value = &(*values)[reverseIndex];
if (!popWithType(expectedType, value)) {
return false;
}
}
return true;
}
// This function pops exactly one value from the stack, checking that it is a
// reference type.
template <typename Policy>
inline bool OpIter<Policy>::popWithRefType(Value* value, StackType* type) {
if (!popStackType(type, value)) {
return false;
}
if (type->isStackBottom() || type->valType().isRefType()) {
return true;
}
UniqueChars actualText = ToString(type->valType(), env_.types);
if (!actualText) {
return false;
}
UniqueChars error(JS_smprintf(
"type mismatch: expression has type %s but expected a reference type",
actualText.get()));
if (!error) {
return false;
}
return fail(error.get());
}
template <typename Policy>
inline bool OpIter<Policy>::checkTopTypeMatches(ResultType expected,
ValueVector* values,
bool rewriteStackTypes) {
if (expected.empty()) {
return true;
}
Control& block = controlStack_.back();
size_t expectedLength = expected.length();
if (values && !values->resize(expectedLength)) {
return false;
}
for (size_t i = 0; i != expectedLength; i++) {
// We're iterating as-if we were popping each expected/actual type one by
// one, which means iterating the array of expected results backwards.
// The "current" value stack length refers to what the value stack length
// would have been if we were popping it.
size_t reverseIndex = expectedLength - i - 1;
ValType expectedType = expected[reverseIndex];
auto collectValue = [&](const Value& v) {
if (values) {
(*values)[reverseIndex] = v;
}
};
size_t currentValueStackLength = valueStack_.length() - i;
MOZ_ASSERT(currentValueStackLength >= block.valueStackBase());
if (currentValueStackLength == block.valueStackBase()) {
if (!block.polymorphicBase()) {
return failEmptyStack();
}
// If the base of this block's stack is polymorphic, then we can just
// pull out as many fake values as we need to validate, and create dummy
// stack entries accordingly; they won't be used since we're in
// unreachable code. However, if `rewriteStackTypes` is true, we must
// set the types on these new entries to whatever `expected` requires
// them to be.
TypeAndValue newTandV =
rewriteStackTypes ? TypeAndValue(expectedType) : TypeAndValue();
if (!valueStack_.insert(valueStack_.begin() + currentValueStackLength,
newTandV)) {
return false;
}
collectValue(Value());
} else {
TypeAndValue& observed = valueStack_[currentValueStackLength - 1];
if (observed.type().isStackBottom()) {
collectValue(Value());
} else {
if (!checkIsSubtypeOf(observed.type().valType(), expectedType)) {
return false;
}
collectValue(observed.value());
}
if (rewriteStackTypes) {
observed.setType(StackType(expectedType));
}
}
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::pushControl(LabelKind kind, BlockType type) {
ResultType paramType = type.params();
ValueVector values;
if (!checkTopTypeMatches(paramType, &values, /*rewriteStackTypes=*/true)) {
return false;
}
MOZ_ASSERT(valueStack_.length() >= paramType.length());
uint32_t valueStackBase = valueStack_.length() - paramType.length();
return controlStack_.emplaceBack(kind, type, valueStackBase);
}
template <typename Policy>
inline bool OpIter<Policy>::checkStackAtEndOfBlock(ResultType* expectedType,
ValueVector* values) {
Control& block = controlStack_.back();
*expectedType = block.type().results();
MOZ_ASSERT(valueStack_.length() >= block.valueStackBase());
if (expectedType->length() < valueStack_.length() - block.valueStackBase()) {
return fail("unused values not explicitly dropped by end of block");
}
return checkTopTypeMatches(*expectedType, values,
/*rewriteStackTypes=*/true);
}
template <typename Policy>
inline bool OpIter<Policy>::getControl(uint32_t relativeDepth,
Control** controlEntry) {
if (relativeDepth >= controlStack_.length()) {
return fail("branch depth exceeds current nesting level");
}
*controlEntry = &controlStack_[controlStack_.length() - 1 - relativeDepth];
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readBlockType(BlockType* type) {
uint8_t nextByte;
if (!d_.peekByte(&nextByte)) {
return fail("unable to read block type");
}
if (nextByte == uint8_t(TypeCode::BlockVoid)) {
d_.uncheckedReadFixedU8();
*type = BlockType::VoidToVoid();
return true;
}
if ((nextByte & SLEB128SignMask) == SLEB128SignBit) {
ValType v;
if (!readValType(&v)) {
return false;
}
*type = BlockType::VoidToSingle(v);
return true;
}
int32_t x;
if (!d_.readVarS32(&x) || x < 0 || uint32_t(x) >= env_.types->length()) {
return fail("invalid block type type index");
}
const TypeDef* typeDef = &env_.types->type(x);
if (!typeDef->isFuncType()) {
return fail("block type type index must be func type");
}
*type = BlockType::Func(typeDef->funcType());
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readOp(OpBytes* op) {
MOZ_ASSERT(!controlStack_.empty());
offsetOfLastReadOp_ = d_.currentOffset();
if (MOZ_UNLIKELY(!d_.readOp(op))) {
return fail("unable to read opcode");
}
#ifdef DEBUG
op_ = *op;
#endif
return true;
}
template <typename Policy>
inline void OpIter<Policy>::peekOp(OpBytes* op) {
const uint8_t* pos = d_.currentPosition();
if (MOZ_UNLIKELY(!d_.readOp(op))) {
op->b0 = uint16_t(Op::Limit);
}
d_.rollbackPosition(pos);
}
template <typename Policy>
inline bool OpIter<Policy>::startFunction(uint32_t funcIndex,
const ValTypeVector& locals) {
MOZ_ASSERT(kind_ == OpIter::Func);
MOZ_ASSERT(elseParamStack_.empty());
MOZ_ASSERT(valueStack_.empty());
MOZ_ASSERT(controlStack_.empty());
MOZ_ASSERT(op_.b0 == uint16_t(Op::Limit));
MOZ_ASSERT(maxInitializedGlobalsIndexPlus1_ == 0);
BlockType type = BlockType::FuncResults(*env_.funcs[funcIndex].type);
// Initialize information related to branch hinting.
lastBranchHintIndex_ = 0;
if (env_.branchHintingEnabled()) {
branchHintVector_ = &env_.branchHints.getHintVector(funcIndex);
}
size_t numArgs = env_.funcs[funcIndex].type->args().length();
if (!unsetLocals_.init(locals, numArgs)) {
return false;
}
return pushControl(LabelKind::Body, type);
}
template <typename Policy>
inline bool OpIter<Policy>::endFunction(const uint8_t* bodyEnd) {
if (d_.currentPosition() != bodyEnd) {
return fail("function body length mismatch");
}
if (!controlStack_.empty()) {
return fail("unbalanced function body control flow");
}
MOZ_ASSERT(elseParamStack_.empty());
MOZ_ASSERT(unsetLocals_.empty());
#ifdef DEBUG
op_ = OpBytes(Op::Limit);
#endif
valueStack_.clear();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::startInitExpr(ValType expected) {
MOZ_ASSERT(kind_ == OpIter::InitExpr);
MOZ_ASSERT(elseParamStack_.empty());
MOZ_ASSERT(valueStack_.empty());
MOZ_ASSERT(controlStack_.empty());
MOZ_ASSERT(op_.b0 == uint16_t(Op::Limit));
lastBranchHintIndex_ = 0;
// GC allows accessing any previously defined global, not just those that are
// imported and immutable.
if (env_.features.gc) {
maxInitializedGlobalsIndexPlus1_ = env_.globals.length();
} else {
maxInitializedGlobalsIndexPlus1_ = env_.numGlobalImports;
}
BlockType type = BlockType::VoidToSingle(expected);
return pushControl(LabelKind::Body, type);
}
template <typename Policy>
inline bool OpIter<Policy>::endInitExpr() {
MOZ_ASSERT(controlStack_.empty());
MOZ_ASSERT(elseParamStack_.empty());
#ifdef DEBUG
op_ = OpBytes(Op::Limit);
#endif
valueStack_.clear();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readValType(ValType* type) {
return d_.readValType(*env_.types, env_.features, type);
}
template <typename Policy>
inline bool OpIter<Policy>::readHeapType(bool nullable, RefType* type) {
return d_.readHeapType(*env_.types, env_.features, nullable, type);
}
template <typename Policy>
inline bool OpIter<Policy>::readReturn(ValueVector* values) {
MOZ_ASSERT(Classify(op_) == OpKind::Return);
Control& body = controlStack_[0];
MOZ_ASSERT(body.kind() == LabelKind::Body);
if (!popWithType(body.resultType(), values)) {
return false;
}
afterUnconditionalBranch();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readBlock(ResultType* paramType) {
MOZ_ASSERT(Classify(op_) == OpKind::Block);
BlockType type;
if (!readBlockType(&type)) {
return false;
}
*paramType = type.params();
return pushControl(LabelKind::Block, type);
}
template <typename Policy>
inline bool OpIter<Policy>::readLoop(ResultType* paramType) {
MOZ_ASSERT(Classify(op_) == OpKind::Loop);
BlockType type;
if (!readBlockType(&type)) {
return false;
}
*paramType = type.params();
return pushControl(LabelKind::Loop, type);
}
template <typename Policy>
inline bool OpIter<Policy>::readIf(ResultType* paramType, Value* condition) {
MOZ_ASSERT(Classify(op_) == OpKind::If);
BlockType type;
if (!readBlockType(&type)) {
return false;
}
if (!popWithType(ValType::I32, condition)) {
return false;
}
if (!pushControl(LabelKind::Then, type)) {
return false;
}
*paramType = type.params();
size_t paramsLength = type.params().length();
return elseParamStack_.append(valueStack_.end() - paramsLength, paramsLength);
}
template <typename Policy>
inline bool OpIter<Policy>::readElse(ResultType* paramType,
ResultType* resultType,
ValueVector* thenResults) {
MOZ_ASSERT(Classify(op_) == OpKind::Else);
Control& block = controlStack_.back();
if (block.kind() != LabelKind::Then) {
return fail("else can only be used within an if");
}
*paramType = block.type().params();
if (!checkStackAtEndOfBlock(resultType, thenResults)) {
return false;
}
valueStack_.shrinkTo(block.valueStackBase());
size_t nparams = block.type().params().length();
MOZ_ASSERT(elseParamStack_.length() >= nparams);
valueStack_.infallibleAppend(elseParamStack_.end() - nparams, nparams);
elseParamStack_.shrinkBy(nparams);
// Reset local state to the beginning of the 'if' block for the new block
// started by 'else'.
unsetLocals_.resetToBlock(controlStack_.length() - 1);
block.switchToElse();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readEnd(LabelKind* kind, ResultType* type,
ValueVector* results,
ValueVector* resultsForEmptyElse) {
MOZ_ASSERT(Classify(op_) == OpKind::End);
Control& block = controlStack_.back();
if (!checkStackAtEndOfBlock(type, results)) {
return false;
}
if (block.kind() == LabelKind::Then) {
ResultType params = block.type().params();
// If an `if` block ends with `end` instead of `else`, then the `else` block
// implicitly passes the `if` parameters as the `else` results. In that
// case, assert that the `if`'s param type matches the result type.
if (params != block.type().results()) {
return fail("if without else with a result value");
}
size_t nparams = params.length();
MOZ_ASSERT(elseParamStack_.length() >= nparams);
if (!resultsForEmptyElse->resize(nparams)) {
return false;
}
const TypeAndValue* elseParams = elseParamStack_.end() - nparams;
for (size_t i = 0; i < nparams; i++) {
(*resultsForEmptyElse)[i] = elseParams[i].value();
}
elseParamStack_.shrinkBy(nparams);
}
*kind = block.kind();
return true;
}
template <typename Policy>
inline void OpIter<Policy>::popEnd() {
MOZ_ASSERT(Classify(op_) == OpKind::End);
controlStack_.popBack();
unsetLocals_.resetToBlock(controlStack_.length());
}
template <typename Policy>
inline bool OpIter<Policy>::checkBranchValueAndPush(uint32_t relativeDepth,
ResultType* type,
ValueVector* values,
bool rewriteStackTypes) {
Control* block = nullptr;
if (!getControl(relativeDepth, &block)) {
return false;
}
*type = block->branchTargetType();
return checkTopTypeMatches(*type, values, rewriteStackTypes);
}
template <typename Policy>
inline bool OpIter<Policy>::readBr(uint32_t* relativeDepth, ResultType* type,
ValueVector* values) {
MOZ_ASSERT(Classify(op_) == OpKind::Br);
if (!readVarU32(relativeDepth)) {
return fail("unable to read br depth");
}
if (!checkBranchValueAndPush(*relativeDepth, type, values,
/*rewriteStackTypes=*/false)) {
return false;
}
afterUnconditionalBranch();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readBrIf(uint32_t* relativeDepth, ResultType* type,
ValueVector* values, Value* condition) {
MOZ_ASSERT(Classify(op_) == OpKind::BrIf);
if (!readVarU32(relativeDepth)) {
return fail("unable to read br_if depth");
}
if (!popWithType(ValType::I32, condition)) {
return false;
}
return checkBranchValueAndPush(*relativeDepth, type, values,
/*rewriteStackTypes=*/true);
}
#define UNKNOWN_ARITY UINT32_MAX
template <typename Policy>
inline bool OpIter<Policy>::checkBrTableEntryAndPush(
uint32_t* relativeDepth, ResultType prevBranchType, ResultType* type,
ValueVector* branchValues) {
if (!readVarU32(relativeDepth)) {
return fail("unable to read br_table depth");
}
Control* block = nullptr;
if (!getControl(*relativeDepth, &block)) {
return false;
}
*type = block->branchTargetType();
if (prevBranchType.valid()) {
if (prevBranchType.length() != type->length()) {
return fail("br_table targets must all have the same arity");
}
// Avoid re-collecting the same values for subsequent branch targets.
branchValues = nullptr;
}
return checkTopTypeMatches(*type, branchValues, /*rewriteStackTypes=*/false);
}
template <typename Policy>
inline bool OpIter<Policy>::readBrTable(Uint32Vector* depths,
uint32_t* defaultDepth,
ResultType* defaultBranchType,
ValueVector* branchValues,
Value* index) {
MOZ_ASSERT(Classify(op_) == OpKind::BrTable);
uint32_t tableLength;
if (!readVarU32(&tableLength)) {
return fail("unable to read br_table table length");
}
if (tableLength > MaxBrTableElems) {
return fail("br_table too big");
}
if (!popWithType(ValType::I32, index)) {
return false;
}
if (!depths->resize(tableLength)) {
return false;
}
ResultType prevBranchType;
for (uint32_t i = 0; i < tableLength; i++) {
ResultType branchType;
if (!checkBrTableEntryAndPush(&(*depths)[i], prevBranchType, &branchType,
branchValues)) {
return false;
}
prevBranchType = branchType;
}
if (!checkBrTableEntryAndPush(defaultDepth, prevBranchType, defaultBranchType,
branchValues)) {
return false;
}
MOZ_ASSERT(defaultBranchType->valid());
afterUnconditionalBranch();
return true;
}
#undef UNKNOWN_ARITY
template <typename Policy>
inline bool OpIter<Policy>::readTry(ResultType* paramType) {
MOZ_ASSERT(Classify(op_) == OpKind::Try);
featureUsage_ |= FeatureUsage::LegacyExceptions;
BlockType type;
if (!readBlockType(&type)) {
return false;
}
*paramType = type.params();
return pushControl(LabelKind::Try, type);
}
enum class TryTableCatchFlags : uint8_t {
CaptureExnRef = 0x1,
CatchAll = 0x1 << 1,
AllowedMask = uint8_t(CaptureExnRef) | uint8_t(CatchAll),
};
template <typename Policy>
inline bool OpIter<Policy>::readTryTable(ResultType* paramType,
TryTableCatchVector* catches) {
MOZ_ASSERT(Classify(op_) == OpKind::TryTable);
BlockType type;
if (!readBlockType(&type)) {
return false;
}
*paramType = type.params();
if (!pushControl(LabelKind::TryTable, type)) {
return false;
}
uint32_t catchesLength;
if (!readVarU32(&catchesLength)) {
return fail("failed to read catches length");
}
if (catchesLength > MaxTryTableCatches) {
return fail("too many catches");
}
if (!catches->reserve(catchesLength)) {
return false;
}
for (uint32_t i = 0; i < catchesLength; i++) {
TryTableCatch tryTableCatch;
// Decode the flags
uint8_t flags;
if (!readFixedU8(&flags)) {
return fail("expected flags");
}
if ((flags & ~uint8_t(TryTableCatchFlags::AllowedMask)) != 0) {
return fail("invalid try_table catch flags");
}
// Decode if this catch wants to capture an exnref
tryTableCatch.captureExnRef =
(flags & uint8_t(TryTableCatchFlags::CaptureExnRef)) != 0;
// Decode the tag, if any
if ((flags & uint8_t(TryTableCatchFlags::CatchAll)) != 0) {
tryTableCatch.tagIndex = CatchAllIndex;
} else {
if (!readVarU32(&tryTableCatch.tagIndex)) {
return fail("expected tag index");
}
if (tryTableCatch.tagIndex >= env_.tags.length()) {
return fail("tag index out of range");
}
}
// Decode the target branch and construct the type we need to compare
// against the branch
if (!readVarU32(&tryTableCatch.labelRelativeDepth)) {
return fail("unable to read catch depth");
}
// The target branch depth is relative to the control labels outside of
// this try_table. e.g. `0` is a branch to the control outside of this
// try_table, not to the try_table itself. However, we've already pushed
// the control block for the try_table, and users will read it after we've
// returned, so we need to return the relative depth adjusted by 1 to
// account for our own control block.
if (tryTableCatch.labelRelativeDepth == UINT32_MAX) {
return fail("catch depth out of range");
}
tryTableCatch.labelRelativeDepth += 1;
// Tagged catches will unpack the exception package and pass it to the
// branch
if (tryTableCatch.tagIndex != CatchAllIndex) {
const TagType& tagType = *env_.tags[tryTableCatch.tagIndex].type;
ResultType tagResult = tagType.resultType();
if (!tagResult.cloneToVector(&tryTableCatch.labelType)) {
return false;
}
}
// Any captured exnref is the final parameter
if (tryTableCatch.captureExnRef &&
!tryTableCatch.labelType.append(ValType(RefType::exn()))) {
return false;
}
Control* block;
if (!getControl(tryTableCatch.labelRelativeDepth, &block)) {
return false;
}
ResultType blockTargetType = block->branchTargetType();
if (!checkIsSubtypeOf(ResultType::Vector(tryTableCatch.labelType),
blockTargetType)) {
return false;
}
catches->infallibleAppend(std::move(tryTableCatch));
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readCatch(LabelKind* kind, uint32_t* tagIndex,
ResultType* paramType,
ResultType* resultType,
ValueVector* tryResults) {
MOZ_ASSERT(Classify(op_) == OpKind::Catch);
if (!readVarU32(tagIndex)) {
return fail("expected tag index");
}
if (*tagIndex >= env_.tags.length()) {
return fail("tag index out of range");
}
Control& block = controlStack_.back();
if (block.kind() == LabelKind::CatchAll) {
return fail("catch cannot follow a catch_all");
}
if (block.kind() != LabelKind::Try && block.kind() != LabelKind::Catch) {
return fail("catch can only be used within a try-catch");
}
*kind = block.kind();
*paramType = block.type().params();
if (!checkStackAtEndOfBlock(resultType, tryResults)) {
return false;
}
valueStack_.shrinkTo(block.valueStackBase());
block.switchToCatch();
// Reset local state to the beginning of the 'try' block.
unsetLocals_.resetToBlock(controlStack_.length() - 1);
return push(env_.tags[*tagIndex].type->resultType());
}
template <typename Policy>
inline bool OpIter<Policy>::readCatchAll(LabelKind* kind, ResultType* paramType,
ResultType* resultType,
ValueVector* tryResults) {
MOZ_ASSERT(Classify(op_) == OpKind::CatchAll);
Control& block = controlStack_.back();
if (block.kind() != LabelKind::Try && block.kind() != LabelKind::Catch) {
return fail("catch_all can only be used within a try-catch");
}
*kind = block.kind();
*paramType = block.type().params();
if (!checkStackAtEndOfBlock(resultType, tryResults)) {
return false;
}
valueStack_.shrinkTo(block.valueStackBase());
block.switchToCatchAll();
// Reset local state to the beginning of the 'try' block.
unsetLocals_.resetToBlock(controlStack_.length() - 1);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readDelegate(uint32_t* relativeDepth,
ResultType* resultType,
ValueVector* tryResults) {
MOZ_ASSERT(Classify(op_) == OpKind::Delegate);
Control& block = controlStack_.back();
if (block.kind() != LabelKind::Try) {
return fail("delegate can only be used within a try");
}
uint32_t delegateDepth;
if (!readVarU32(&delegateDepth)) {
return fail("unable to read delegate depth");
}
// Depths for delegate start counting in the surrounding block.
if (delegateDepth >= controlStack_.length() - 1) {
return fail("delegate depth exceeds current nesting level");
}
*relativeDepth = delegateDepth + 1;
// Because `delegate` acts like `end` and ends the block, we will check
// the stack here.
return checkStackAtEndOfBlock(resultType, tryResults);
}
// We need popDelegate because readDelegate cannot pop the control stack
// itself, as its caller may need to use the control item for delegate.
template <typename Policy>
inline void OpIter<Policy>::popDelegate() {
MOZ_ASSERT(Classify(op_) == OpKind::Delegate);
controlStack_.popBack();
unsetLocals_.resetToBlock(controlStack_.length());
}
template <typename Policy>
inline bool OpIter<Policy>::readThrow(uint32_t* tagIndex,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::Throw);
if (!readVarU32(tagIndex)) {
return fail("expected tag index");
}
if (*tagIndex >= env_.tags.length()) {
return fail("tag index out of range");
}
if (!popWithType(env_.tags[*tagIndex].type->resultType(), argValues)) {
return false;
}
afterUnconditionalBranch();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readThrowRef(Value* exnRef) {
MOZ_ASSERT(Classify(op_) == OpKind::ThrowRef);
if (!popWithType(ValType(RefType::exn()), exnRef)) {
return false;
}
afterUnconditionalBranch();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readRethrow(uint32_t* relativeDepth) {
MOZ_ASSERT(Classify(op_) == OpKind::Rethrow);
if (!readVarU32(relativeDepth)) {
return fail("unable to read rethrow depth");
}
if (*relativeDepth >= controlStack_.length()) {
return fail("rethrow depth exceeds current nesting level");
}
LabelKind kind = controlKind(*relativeDepth);
if (kind != LabelKind::Catch && kind != LabelKind::CatchAll) {
return fail("rethrow target was not a catch block");
}
afterUnconditionalBranch();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readUnreachable() {
MOZ_ASSERT(Classify(op_) == OpKind::Unreachable);
afterUnconditionalBranch();
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readDrop() {
MOZ_ASSERT(Classify(op_) == OpKind::Drop);
StackType type;
Value value;
return popStackType(&type, &value);
}
template <typename Policy>
inline bool OpIter<Policy>::readUnary(ValType operandType, Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::Unary);
if (!popWithType(operandType, input)) {
return false;
}
infalliblePush(operandType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readConversion(ValType operandType,
ValType resultType, Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::Conversion);
if (!popWithType(operandType, input)) {
return false;
}
infalliblePush(resultType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readBinary(ValType operandType, Value* lhs,
Value* rhs) {
MOZ_ASSERT(Classify(op_) == OpKind::Binary);
if (!popWithType(operandType, rhs)) {
return false;
}
if (!popWithType(operandType, lhs)) {
return false;
}
infalliblePush(operandType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readComparison(ValType operandType, Value* lhs,
Value* rhs) {
MOZ_ASSERT(Classify(op_) == OpKind::Comparison);
if (!popWithType(operandType, rhs)) {
return false;
}
if (!popWithType(operandType, lhs)) {
return false;
}
infalliblePush(ValType::I32);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readTernary(ValType operandType, Value* v0,
Value* v1, Value* v2) {
MOZ_ASSERT(Classify(op_) == OpKind::Ternary);
if (!popWithType(operandType, v2)) {
return false;
}
if (!popWithType(operandType, v1)) {
return false;
}
if (!popWithType(operandType, v0)) {
return false;
}
infalliblePush(operandType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readLinearMemoryAddress(
uint32_t byteSize, LinearMemoryAddress<Value>* addr) {
uint32_t flags;
if (!readVarU32(&flags)) {
return fail("unable to read load alignment");
}
uint8_t alignLog2 = flags & ((1 << 6) - 1);
uint8_t hasMemoryIndex = flags & (1 << 6);
uint8_t undefinedBits = flags & ~((1 << 7) - 1);
if (undefinedBits != 0) {
return fail("invalid memory flags");
}
if (hasMemoryIndex != 0) {
if (!readVarU32(&addr->memoryIndex)) {
return fail("unable to read memory index");
}
} else {
addr->memoryIndex = 0;
}
if (addr->memoryIndex >= env_.numMemories()) {
return fail("memory index out of range");
}
if (!readVarU64(&addr->offset)) {
return fail("unable to read load offset");
}
IndexType it = env_.memories[addr->memoryIndex].indexType();
if (it == IndexType::I32 && addr->offset > UINT32_MAX) {
return fail("offset too large for memory type");
}
if (alignLog2 >= 32 || (uint32_t(1) << alignLog2) > byteSize) {
return fail("greater than natural alignment");
}
if (!popWithType(ToValType(it), &addr->base)) {
return false;
}
addr->align = uint32_t(1) << alignLog2;
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readLinearMemoryAddressAligned(
uint32_t byteSize, LinearMemoryAddress<Value>* addr) {
if (!readLinearMemoryAddress(byteSize, addr)) {
return false;
}
if (addr->align != byteSize) {
return fail("not natural alignment");
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readLoad(ValType resultType, uint32_t byteSize,
LinearMemoryAddress<Value>* addr) {
MOZ_ASSERT(Classify(op_) == OpKind::Load);
if (!readLinearMemoryAddress(byteSize, addr)) {
return false;
}
infalliblePush(resultType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readStore(ValType resultType, uint32_t byteSize,
LinearMemoryAddress<Value>* addr,
Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::Store);
if (!popWithType(resultType, value)) {
return false;
}
return readLinearMemoryAddress(byteSize, addr);
}
template <typename Policy>
inline bool OpIter<Policy>::readTeeStore(ValType resultType, uint32_t byteSize,
LinearMemoryAddress<Value>* addr,
Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::TeeStore);
if (!popWithType(resultType, value)) {
return false;
}
if (!readLinearMemoryAddress(byteSize, addr)) {
return false;
}
infalliblePush(TypeAndValue(resultType, *value));
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readNop() {
MOZ_ASSERT(Classify(op_) == OpKind::Nop);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readMemorySize(uint32_t* memoryIndex) {
MOZ_ASSERT(Classify(op_) == OpKind::MemorySize);
if (!readVarU32(memoryIndex)) {
return fail("failed to read memory flags");
}
if (*memoryIndex >= env_.numMemories()) {
return fail("memory index out of range for memory.size");
}
ValType ptrType = ToValType(env_.memories[*memoryIndex].indexType());
return push(ptrType);
}
template <typename Policy>
inline bool OpIter<Policy>::readMemoryGrow(uint32_t* memoryIndex,
Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::MemoryGrow);
if (!readVarU32(memoryIndex)) {
return fail("failed to read memory flags");
}
if (*memoryIndex >= env_.numMemories()) {
return fail("memory index out of range for memory.grow");
}
ValType ptrType = ToValType(env_.memories[*memoryIndex].indexType());
if (!popWithType(ptrType, input)) {
return false;
}
infalliblePush(ptrType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readSelect(bool typed, StackType* type,
Value* trueValue, Value* falseValue,
Value* condition) {
MOZ_ASSERT(Classify(op_) == OpKind::Select);
if (typed) {
uint32_t length;
if (!readVarU32(&length)) {
return fail("unable to read select result length");
}
if (length != 1) {
return fail("bad number of results");
}
ValType result;
if (!readValType(&result)) {
return fail("invalid result type for select");
}
if (!popWithType(ValType::I32, condition)) {
return false;
}
if (!popWithType(result, falseValue)) {
return false;
}
if (!popWithType(result, trueValue)) {
return false;
}
*type = StackType(result);
infalliblePush(*type);
return true;
}
if (!popWithType(ValType::I32, condition)) {
return false;
}
StackType falseType;
if (!popStackType(&falseType, falseValue)) {
return false;
}
StackType trueType;
if (!popStackType(&trueType, trueValue)) {
return false;
}
if (!falseType.isValidForUntypedSelect() ||
!trueType.isValidForUntypedSelect()) {
return fail("invalid types for untyped select");
}
if (falseType.isStackBottom()) {
*type = trueType;
} else if (trueType.isStackBottom() || falseType == trueType) {
*type = falseType;
} else {
return fail("select operand types must match");
}
infalliblePush(*type);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readGetLocal(const ValTypeVector& locals,
uint32_t* id) {
MOZ_ASSERT(Classify(op_) == OpKind::GetLocal);
if (!readVarU32(id)) {
return fail("unable to read local index");
}
if (*id >= locals.length()) {
return fail("local.get index out of range");
}
if (unsetLocals_.isUnset(*id)) {
return fail("local.get read from unset local");
}
return push(locals[*id]);
}
template <typename Policy>
inline bool OpIter<Policy>::readSetLocal(const ValTypeVector& locals,
uint32_t* id, Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::SetLocal);
if (!readVarU32(id)) {
return fail("unable to read local index");
}
if (*id >= locals.length()) {
return fail("local.set index out of range");
}
if (unsetLocals_.isUnset(*id)) {
unsetLocals_.set(*id, controlStackDepth());
}
return popWithType(locals[*id], value);
}
template <typename Policy>
inline bool OpIter<Policy>::readTeeLocal(const ValTypeVector& locals,
uint32_t* id, Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::TeeLocal);
if (!readVarU32(id)) {
return fail("unable to read local index");
}
if (*id >= locals.length()) {
return fail("local.set index out of range");
}
if (unsetLocals_.isUnset(*id)) {
unsetLocals_.set(*id, controlStackDepth());
}
ValueVector single;
if (!checkTopTypeMatches(ResultType::Single(locals[*id]), &single,
/*rewriteStackTypes=*/true)) {
return false;
}
*value = single[0];
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readGetGlobal(uint32_t* id) {
MOZ_ASSERT(Classify(op_) == OpKind::GetGlobal);
if (!d_.readGlobalIndex(id)) {
return false;
}
if (*id >= env_.globals.length()) {
return fail("global.get index out of range");
}
// Initializer expressions can access immutable imported globals, or any
// previously defined immutable global with GC enabled.
if (kind_ == OpIter::InitExpr && (env_.globals[*id].isMutable() ||
*id >= maxInitializedGlobalsIndexPlus1_)) {
return fail(
"global.get in initializer expression must reference a global "
"immutable import");
}
return push(env_.globals[*id].type());
}
template <typename Policy>
inline bool OpIter<Policy>::readSetGlobal(uint32_t* id, Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::SetGlobal);
if (!d_.readGlobalIndex(id)) {
return false;
}
if (*id >= env_.globals.length()) {
return fail("global.set index out of range");
}
if (!env_.globals[*id].isMutable()) {
return fail("can't write an immutable global");
}
return popWithType(env_.globals[*id].type(), value);
}
template <typename Policy>
inline bool OpIter<Policy>::readTeeGlobal(uint32_t* id, Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::TeeGlobal);
if (!d_.readGlobalIndex(id)) {
return false;
}
if (*id >= env_.globals.length()) {
return fail("global.set index out of range");
}
if (!env_.globals[*id].isMutable()) {
return fail("can't write an immutable global");
}
ValueVector single;
if (!checkTopTypeMatches(ResultType::Single(env_.globals[*id].type()),
&single,
/*rewriteStackTypes=*/true)) {
return false;
}
MOZ_ASSERT(single.length() == 1);
*value = single[0];
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readI32Const(int32_t* i32) {
MOZ_ASSERT(Classify(op_) == OpKind::I32);
if (!d_.readI32Const(i32)) {
return false;
}
return push(ValType::I32);
}
template <typename Policy>
inline bool OpIter<Policy>::readI64Const(int64_t* i64) {
MOZ_ASSERT(Classify(op_) == OpKind::I64);
if (!d_.readI64Const(i64)) {
return false;
}
return push(ValType::I64);
}
template <typename Policy>
inline bool OpIter<Policy>::readF32Const(float* f32) {
MOZ_ASSERT(Classify(op_) == OpKind::F32);
if (!d_.readF32Const(f32)) {
return false;
}
return push(ValType::F32);
}
template <typename Policy>
inline bool OpIter<Policy>::readF64Const(double* f64) {
MOZ_ASSERT(Classify(op_) == OpKind::F64);
if (!d_.readF64Const(f64)) {
return false;
}
return push(ValType::F64);
}
template <typename Policy>
inline bool OpIter<Policy>::readRefFunc(uint32_t* funcIndex) {
MOZ_ASSERT(Classify(op_) == OpKind::RefFunc);
if (!d_.readFuncIndex(funcIndex)) {
return false;
}
if (*funcIndex >= env_.funcs.length()) {
return fail("function index out of range");
}
if (kind_ == OpIter::Func && !env_.funcs[*funcIndex].canRefFunc()) {
return fail(
"function index is not declared in a section before the code section");
}
#ifdef ENABLE_WASM_GC
// When function references enabled, push type index on the stack, e.g. for
// validation of the call_ref instruction.
if (env_.gcEnabled()) {
const uint32_t typeIndex = env_.funcs[*funcIndex].typeIndex;
const TypeDef& typeDef = env_.types->type(typeIndex);
return push(RefType::fromTypeDef(&typeDef, false));
}
#endif
return push(RefType::func());
}
template <typename Policy>
inline bool OpIter<Policy>::readRefNull(RefType* type) {
MOZ_ASSERT(Classify(op_) == OpKind::RefNull);
if (!d_.readRefNull(*env_.types, env_.features, type)) {
return false;
}
return push(*type);
}
template <typename Policy>
inline bool OpIter<Policy>::readRefIsNull(Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::Conversion);
StackType type;
if (!popWithRefType(input, &type)) {
return false;
}
return push(ValType::I32);
}
template <typename Policy>
inline bool OpIter<Policy>::readRefAsNonNull(Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::RefAsNonNull);
StackType type;
if (!popWithRefType(input, &type)) {
return false;
}
if (type.isStackBottom()) {
infalliblePush(type);
} else {
infalliblePush(TypeAndValue(type.asNonNullable(), *input));
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readBrOnNull(uint32_t* relativeDepth,
ResultType* type, ValueVector* values,
Value* condition) {
MOZ_ASSERT(Classify(op_) == OpKind::BrOnNull);
if (!readVarU32(relativeDepth)) {
return fail("unable to read br_on_null depth");
}
StackType refType;
if (!popWithRefType(condition, &refType)) {
return false;
}
if (!checkBranchValueAndPush(*relativeDepth, type, values,
/*rewriteStackTypes=*/true)) {
return false;
}
if (refType.isStackBottom()) {
return push(refType);
}
return push(TypeAndValue(refType.asNonNullable(), *condition));
}
template <typename Policy>
inline bool OpIter<Policy>::readBrOnNonNull(uint32_t* relativeDepth,
ResultType* type,
ValueVector* values,
Value* condition) {
MOZ_ASSERT(Classify(op_) == OpKind::BrOnNonNull);
if (!readVarU32(relativeDepth)) {
return fail("unable to read br_on_non_null depth");
}
Control* block = nullptr;
if (!getControl(*relativeDepth, &block)) {
return false;
}
*type = block->branchTargetType();
// Check we at least have one type in the branch target type.
if (type->length() < 1) {
return fail("type mismatch: target block type expected to be [_, ref]");
}
// Pop the condition reference.
StackType refType;
if (!popWithRefType(condition, &refType)) {
return false;
}
// Push non-nullable version of condition reference on the stack, prior
// checking the target type below.
if (!(refType.isStackBottom()
? push(refType)
: push(TypeAndValue(refType.asNonNullable(), *condition)))) {
return false;
}
// Check if the type stack matches the branch target type.
if (!checkTopTypeMatches(*type, values, /*rewriteStackTypes=*/true)) {
return false;
}
// Pop the condition reference -- the null-branch does not receive the value.
StackType unusedType;
Value unusedValue;
return popStackType(&unusedType, &unusedValue);
}
template <typename Policy>
inline bool OpIter<Policy>::popCallArgs(const ValTypeVector& expectedTypes,
ValueVector* values) {
// Iterate through the argument types backward so that pops occur in the
// right order.
if (!values->resize(expectedTypes.length())) {
return false;
}
for (int32_t i = int32_t(expectedTypes.length()) - 1; i >= 0; i--) {
if (!popWithType(expectedTypes[i], &(*values)[i])) {
return false;
}
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readCall(uint32_t* funcTypeIndex,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::Call);
if (!readVarU32(funcTypeIndex)) {
return fail("unable to read call function index");
}
if (*funcTypeIndex >= env_.funcs.length()) {
return fail("callee index out of range");
}
const FuncType& funcType = *env_.funcs[*funcTypeIndex].type;
if (!popCallArgs(funcType.args(), argValues)) {
return false;
}
return push(ResultType::Vector(funcType.results()));
}
#ifdef ENABLE_WASM_TAIL_CALLS
template <typename Policy>
inline bool OpIter<Policy>::readReturnCall(uint32_t* funcTypeIndex,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::ReturnCall);
if (!readVarU32(funcTypeIndex)) {
return fail("unable to read call function index");
}
if (*funcTypeIndex >= env_.funcs.length()) {
return fail("callee index out of range");
}
const FuncType& funcType = *env_.funcs[*funcTypeIndex].type;
if (!popCallArgs(funcType.args(), argValues)) {
return false;
}
// Check if callee results are subtypes of caller's.
Control& body = controlStack_[0];
MOZ_ASSERT(body.kind() == LabelKind::Body);
if (!checkIsSubtypeOf(ResultType::Vector(funcType.results()),
body.resultType())) {
return false;
}
afterUnconditionalBranch();
return true;
}
#endif
template <typename Policy>
inline bool OpIter<Policy>::readCallIndirect(uint32_t* funcTypeIndex,
uint32_t* tableIndex,
Value* callee,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::CallIndirect);
MOZ_ASSERT(funcTypeIndex != tableIndex);
if (!readVarU32(funcTypeIndex)) {
return fail("unable to read call_indirect signature index");
}
if (*funcTypeIndex >= env_.numTypes()) {
return fail("signature index out of range");
}
if (!readVarU32(tableIndex)) {
return fail("unable to read call_indirect table index");
}
if (*tableIndex >= env_.tables.length()) {
// Special case this for improved user experience.
if (!env_.tables.length()) {
return fail("can't call_indirect without a table");
}
return fail("table index out of range for call_indirect");
}
if (!env_.tables[*tableIndex].elemType.isFuncHierarchy()) {
return fail("indirect calls must go through a table of 'funcref'");
}
if (!popWithType(ValType::I32, callee)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*funcTypeIndex);
if (!typeDef.isFuncType()) {
return fail("expected signature type");
}
const FuncType& funcType = typeDef.funcType();
if (!popCallArgs(funcType.args(), argValues)) {
return false;
}
return push(ResultType::Vector(funcType.results()));
}
#ifdef ENABLE_WASM_TAIL_CALLS
template <typename Policy>
inline bool OpIter<Policy>::readReturnCallIndirect(uint32_t* funcTypeIndex,
uint32_t* tableIndex,
Value* callee,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::ReturnCallIndirect);
MOZ_ASSERT(funcTypeIndex != tableIndex);
if (!readVarU32(funcTypeIndex)) {
return fail("unable to read return_call_indirect signature index");
}
if (*funcTypeIndex >= env_.numTypes()) {
return fail("signature index out of range");
}
if (!readVarU32(tableIndex)) {
return fail("unable to read return_call_indirect table index");
}
if (*tableIndex >= env_.tables.length()) {
// Special case this for improved user experience.
if (!env_.tables.length()) {
return fail("can't return_call_indirect without a table");
}
return fail("table index out of range for return_call_indirect");
}
if (!env_.tables[*tableIndex].elemType.isFuncHierarchy()) {
return fail("indirect calls must go through a table of 'funcref'");
}
if (!popWithType(ValType::I32, callee)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*funcTypeIndex);
if (!typeDef.isFuncType()) {
return fail("expected signature type");
}
const FuncType& funcType = typeDef.funcType();
if (!popCallArgs(funcType.args(), argValues)) {
return false;
}
// Check if callee results are subtypes of caller's.
Control& body = controlStack_[0];
MOZ_ASSERT(body.kind() == LabelKind::Body);
if (!checkIsSubtypeOf(ResultType::Vector(funcType.results()),
body.resultType())) {
return false;
}
afterUnconditionalBranch();
return true;
}
#endif
#ifdef ENABLE_WASM_GC
template <typename Policy>
inline bool OpIter<Policy>::readCallRef(const FuncType** funcType,
Value* callee, ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::CallRef);
uint32_t funcTypeIndex;
if (!readFuncTypeIndex(&funcTypeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(funcTypeIndex);
*funcType = &typeDef.funcType();
if (!popWithType(ValType(RefType::fromTypeDef(&typeDef, true)), callee)) {
return false;
}
if (!popCallArgs((*funcType)->args(), argValues)) {
return false;
}
return push(ResultType::Vector((*funcType)->results()));
}
#endif
#if defined(ENABLE_WASM_TAIL_CALLS) && defined(ENABLE_WASM_GC)
template <typename Policy>
inline bool OpIter<Policy>::readReturnCallRef(const FuncType** funcType,
Value* callee,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::ReturnCallRef);
uint32_t funcTypeIndex;
if (!readFuncTypeIndex(&funcTypeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(funcTypeIndex);
*funcType = &typeDef.funcType();
if (!popWithType(ValType(RefType::fromTypeDef(&typeDef, true)), callee)) {
return false;
}
if (!popCallArgs((*funcType)->args(), argValues)) {
return false;
}
// Check if callee results are subtypes of caller's.
Control& body = controlStack_[0];
MOZ_ASSERT(body.kind() == LabelKind::Body);
if (!checkIsSubtypeOf(ResultType::Vector((*funcType)->results()),
body.resultType())) {
return false;
}
afterUnconditionalBranch();
return true;
}
#endif
template <typename Policy>
inline bool OpIter<Policy>::readOldCallDirect(uint32_t numFuncImports,
uint32_t* funcTypeIndex,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::OldCallDirect);
uint32_t funcDefIndex;
if (!readVarU32(&funcDefIndex)) {
return fail("unable to read call function index");
}
if (UINT32_MAX - funcDefIndex < numFuncImports) {
return fail("callee index out of range");
}
*funcTypeIndex = numFuncImports + funcDefIndex;
if (*funcTypeIndex >= env_.funcs.length()) {
return fail("callee index out of range");
}
const FuncType& funcType = *env_.funcs[*funcTypeIndex].type;
if (!popCallArgs(funcType.args(), argValues)) {
return false;
}
return push(ResultType::Vector(funcType.results()));
}
template <typename Policy>
inline bool OpIter<Policy>::readOldCallIndirect(uint32_t* funcTypeIndex,
Value* callee,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::OldCallIndirect);
if (!readVarU32(funcTypeIndex)) {
return fail("unable to read call_indirect signature index");
}
if (*funcTypeIndex >= env_.numTypes()) {
return fail("signature index out of range");
}
const TypeDef& typeDef = env_.types->type(*funcTypeIndex);
if (!typeDef.isFuncType()) {
return fail("expected signature type");
}
const FuncType& funcType = typeDef.funcType();
if (!popCallArgs(funcType.args(), argValues)) {
return false;
}
if (!popWithType(ValType::I32, callee)) {
return false;
}
return push(ResultType::Vector(funcType.results()));
}
template <typename Policy>
inline bool OpIter<Policy>::readWake(LinearMemoryAddress<Value>* addr,
Value* count) {
MOZ_ASSERT(Classify(op_) == OpKind::Wake);
if (!popWithType(ValType::I32, count)) {
return false;
}
uint32_t byteSize = 4; // Per spec; smallest WAIT is i32.
if (!readLinearMemoryAddressAligned(byteSize, addr)) {
return false;
}
infalliblePush(ValType::I32);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readWait(LinearMemoryAddress<Value>* addr,
ValType valueType, uint32_t byteSize,
Value* value, Value* timeout) {
MOZ_ASSERT(Classify(op_) == OpKind::Wait);
if (!popWithType(ValType::I64, timeout)) {
return false;
}
if (!popWithType(valueType, value)) {
return false;
}
if (!readLinearMemoryAddressAligned(byteSize, addr)) {
return false;
}
infalliblePush(ValType::I32);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readFence() {
MOZ_ASSERT(Classify(op_) == OpKind::Fence);
uint8_t flags;
if (!readFixedU8(&flags)) {
return fail("expected memory order after fence");
}
if (flags != 0) {
return fail("non-zero memory order not supported yet");
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readAtomicLoad(LinearMemoryAddress<Value>* addr,
ValType resultType,
uint32_t byteSize) {
MOZ_ASSERT(Classify(op_) == OpKind::AtomicLoad);
if (!readLinearMemoryAddressAligned(byteSize, addr)) {
return false;
}
infalliblePush(resultType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readAtomicStore(LinearMemoryAddress<Value>* addr,
ValType resultType,
uint32_t byteSize, Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::AtomicStore);
if (!popWithType(resultType, value)) {
return false;
}
return readLinearMemoryAddressAligned(byteSize, addr);
}
template <typename Policy>
inline bool OpIter<Policy>::readAtomicRMW(LinearMemoryAddress<Value>* addr,
ValType resultType, uint32_t byteSize,
Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::AtomicBinOp);
if (!popWithType(resultType, value)) {
return false;
}
if (!readLinearMemoryAddressAligned(byteSize, addr)) {
return false;
}
infalliblePush(resultType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readAtomicCmpXchg(LinearMemoryAddress<Value>* addr,
ValType resultType,
uint32_t byteSize,
Value* oldValue,
Value* newValue) {
MOZ_ASSERT(Classify(op_) == OpKind::AtomicCompareExchange);
if (!popWithType(resultType, newValue)) {
return false;
}
if (!popWithType(resultType, oldValue)) {
return false;
}
if (!readLinearMemoryAddressAligned(byteSize, addr)) {
return false;
}
infalliblePush(resultType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readMemOrTableCopy(bool isMem,
uint32_t* dstMemOrTableIndex,
Value* dst,
uint32_t* srcMemOrTableIndex,
Value* src, Value* len) {
MOZ_ASSERT(Classify(op_) == OpKind::MemOrTableCopy);
MOZ_ASSERT(dstMemOrTableIndex != srcMemOrTableIndex);
// Spec requires (dest, src) as of 2019-10-04.
if (!readVarU32(dstMemOrTableIndex)) {
return false;
}
if (!readVarU32(srcMemOrTableIndex)) {
return false;
}
if (isMem) {
if (*srcMemOrTableIndex >= env_.memories.length() ||
*dstMemOrTableIndex >= env_.memories.length()) {
return fail("memory index out of range for memory.copy");
}
} else {
if (*dstMemOrTableIndex >= env_.tables.length() ||
*srcMemOrTableIndex >= env_.tables.length()) {
return fail("table index out of range for table.copy");
}
ValType dstElemType = env_.tables[*dstMemOrTableIndex].elemType;
ValType srcElemType = env_.tables[*srcMemOrTableIndex].elemType;
if (!checkIsSubtypeOf(srcElemType, dstElemType)) {
return false;
}
}
ValType dstPtrType;
ValType srcPtrType;
ValType lenType;
if (isMem) {
dstPtrType = ToValType(env_.memories[*dstMemOrTableIndex].indexType());
srcPtrType = ToValType(env_.memories[*srcMemOrTableIndex].indexType());
if (dstPtrType == ValType::I64 && srcPtrType == ValType::I64) {
lenType = ValType::I64;
} else {
lenType = ValType::I32;
}
} else {
dstPtrType = srcPtrType = lenType = ValType::I32;
}
if (!popWithType(lenType, len)) {
return false;
}
if (!popWithType(srcPtrType, src)) {
return false;
}
return popWithType(dstPtrType, dst);
}
template <typename Policy>
inline bool OpIter<Policy>::readDataOrElemDrop(bool isData,
uint32_t* segIndex) {
MOZ_ASSERT(Classify(op_) == OpKind::DataOrElemDrop);
if (!readVarU32(segIndex)) {
return fail("unable to read segment index");
}
if (isData) {
if (env_.dataCount.isNothing()) {
return fail("data.drop requires a DataCount section");
}
if (*segIndex >= *env_.dataCount) {
return fail("data.drop segment index out of range");
}
} else {
if (*segIndex >= env_.elemSegments.length()) {
return fail("element segment index out of range for elem.drop");
}
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readMemFill(uint32_t* memoryIndex, Value* start,
Value* val, Value* len) {
MOZ_ASSERT(Classify(op_) == OpKind::MemFill);
if (!readVarU32(memoryIndex)) {
return fail("failed to read memory index");
}
if (*memoryIndex >= env_.numMemories()) {
return fail("memory index out of range for memory.fill");
}
ValType ptrType = ToValType(env_.memories[*memoryIndex].indexType());
if (!popWithType(ptrType, len)) {
return false;
}
if (!popWithType(ValType::I32, val)) {
return false;
}
return popWithType(ptrType, start);
}
template <typename Policy>
inline bool OpIter<Policy>::readMemOrTableInit(bool isMem, uint32_t* segIndex,
uint32_t* dstMemOrTableIndex,
Value* dst, Value* src,
Value* len) {
MOZ_ASSERT(Classify(op_) == OpKind::MemOrTableInit);
MOZ_ASSERT(segIndex != dstMemOrTableIndex);
if (!readVarU32(segIndex)) {
return fail("unable to read segment index");
}
uint32_t memOrTableIndex = 0;
if (!readVarU32(&memOrTableIndex)) {
return false;
}
if (isMem) {
if (memOrTableIndex >= env_.memories.length()) {
return fail("memory index out of range for memory.init");
}
*dstMemOrTableIndex = memOrTableIndex;
if (env_.dataCount.isNothing()) {
return fail("memory.init requires a DataCount section");
}
if (*segIndex >= *env_.dataCount) {
return fail("memory.init segment index out of range");
}
} else {
if (memOrTableIndex >= env_.tables.length()) {
return fail("table index out of range for table.init");
}
*dstMemOrTableIndex = memOrTableIndex;
if (*segIndex >= env_.elemSegments.length()) {
return fail("table.init segment index out of range");
}
if (!checkIsSubtypeOf(env_.elemSegments[*segIndex].elemType,
env_.tables[*dstMemOrTableIndex].elemType)) {
return false;
}
}
if (!popWithType(ValType::I32, len)) {
return false;
}
if (!popWithType(ValType::I32, src)) {
return false;
}
ValType ptrType =
isMem ? ToValType(env_.memories[*dstMemOrTableIndex].indexType())
: ValType::I32;
return popWithType(ptrType, dst);
}
template <typename Policy>
inline bool OpIter<Policy>::readTableFill(uint32_t* tableIndex, Value* start,
Value* val, Value* len) {
MOZ_ASSERT(Classify(op_) == OpKind::TableFill);
if (!readVarU32(tableIndex)) {
return fail("unable to read table index");
}
if (*tableIndex >= env_.tables.length()) {
return fail("table index out of range for table.fill");
}
if (!popWithType(ValType::I32, len)) {
return false;
}
if (!popWithType(env_.tables[*tableIndex].elemType, val)) {
return false;
}
return popWithType(ValType::I32, start);
}
template <typename Policy>
inline bool OpIter<Policy>::readMemDiscard(uint32_t* memoryIndex, Value* start,
Value* len) {
MOZ_ASSERT(Classify(op_) == OpKind::MemDiscard);
if (!readVarU32(memoryIndex)) {
return fail("failed to read memory index");
}
if (*memoryIndex >= env_.memories.length()) {
return fail("memory index out of range for memory.discard");
}
ValType ptrType = ToValType(env_.memories[*memoryIndex].indexType());
if (!popWithType(ptrType, len)) {
return false;
}
return popWithType(ptrType, start);
}
template <typename Policy>
inline bool OpIter<Policy>::readTableGet(uint32_t* tableIndex, Value* index) {
MOZ_ASSERT(Classify(op_) == OpKind::TableGet);
if (!readVarU32(tableIndex)) {
return fail("unable to read table index");
}
if (*tableIndex >= env_.tables.length()) {
return fail("table index out of range for table.get");
}
if (!popWithType(ValType::I32, index)) {
return false;
}
infalliblePush(env_.tables[*tableIndex].elemType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readTableGrow(uint32_t* tableIndex,
Value* initValue, Value* delta) {
MOZ_ASSERT(Classify(op_) == OpKind::TableGrow);
if (!readVarU32(tableIndex)) {
return fail("unable to read table index");
}
if (*tableIndex >= env_.tables.length()) {
return fail("table index out of range for table.grow");
}
if (!popWithType(ValType::I32, delta)) {
return false;
}
if (!popWithType(env_.tables[*tableIndex].elemType, initValue)) {
return false;
}
infalliblePush(ValType::I32);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readTableSet(uint32_t* tableIndex, Value* index,
Value* value) {
MOZ_ASSERT(Classify(op_) == OpKind::TableSet);
if (!readVarU32(tableIndex)) {
return fail("unable to read table index");
}
if (*tableIndex >= env_.tables.length()) {
return fail("table index out of range for table.set");
}
if (!popWithType(env_.tables[*tableIndex].elemType, value)) {
return false;
}
return popWithType(ValType::I32, index);
}
template <typename Policy>
inline bool OpIter<Policy>::readTableSize(uint32_t* tableIndex) {
MOZ_ASSERT(Classify(op_) == OpKind::TableSize);
*tableIndex = 0;
if (!readVarU32(tableIndex)) {
return fail("unable to read table index");
}
if (*tableIndex >= env_.tables.length()) {
return fail("table index out of range for table.size");
}
return push(ValType::I32);
}
template <typename Policy>
inline bool OpIter<Policy>::readGcTypeIndex(uint32_t* typeIndex) {
if (!d_.readTypeIndex(typeIndex)) {
return false;
}
if (*typeIndex >= env_.types->length()) {
return fail("type index out of range");
}
if (!env_.types->type(*typeIndex).isStructType() &&
!env_.types->type(*typeIndex).isArrayType()) {
return fail("not a gc type");
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readStructTypeIndex(uint32_t* typeIndex) {
if (!readVarU32(typeIndex)) {
return fail("unable to read type index");
}
if (*typeIndex >= env_.types->length()) {
return fail("type index out of range");
}
if (!env_.types->type(*typeIndex).isStructType()) {
return fail("not a struct type");
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayTypeIndex(uint32_t* typeIndex) {
if (!readVarU32(typeIndex)) {
return fail("unable to read type index");
}
if (*typeIndex >= env_.types->length()) {
return fail("type index out of range");
}
if (!env_.types->type(*typeIndex).isArrayType()) {
return fail("not an array type");
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readFuncTypeIndex(uint32_t* typeIndex) {
if (!readVarU32(typeIndex)) {
return fail("unable to read type index");
}
if (*typeIndex >= env_.types->length()) {
return fail("type index out of range");
}
if (!env_.types->type(*typeIndex).isFuncType()) {
return fail("not an func type");
}
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readFieldIndex(uint32_t* fieldIndex,
const StructType& structType) {
if (!readVarU32(fieldIndex)) {
return fail("unable to read field index");
}
if (structType.fields_.length() <= *fieldIndex) {
return fail("field index out of range");
}
return true;
}
#ifdef ENABLE_WASM_GC
template <typename Policy>
inline bool OpIter<Policy>::readStructNew(uint32_t* typeIndex,
ValueVector* argValues) {
MOZ_ASSERT(Classify(op_) == OpKind::StructNew);
if (!readStructTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const StructType& structType = typeDef.structType();
if (!argValues->resize(structType.fields_.length())) {
return false;
}
static_assert(MaxStructFields <= INT32_MAX, "Or we iloop below");
for (int32_t i = structType.fields_.length() - 1; i >= 0; i--) {
if (!popWithType(structType.fields_[i].type.widenToValType(),
&(*argValues)[i])) {
return false;
}
}
return push(RefType::fromTypeDef(&typeDef, false));
}
template <typename Policy>
inline bool OpIter<Policy>::readStructNewDefault(uint32_t* typeIndex) {
MOZ_ASSERT(Classify(op_) == OpKind::StructNewDefault);
if (!readStructTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const StructType& structType = typeDef.structType();
if (!structType.isDefaultable()) {
return fail("struct must be defaultable");
}
return push(RefType::fromTypeDef(&typeDef, false));
}
template <typename Policy>
inline bool OpIter<Policy>::readStructGet(uint32_t* typeIndex,
uint32_t* fieldIndex,
FieldWideningOp wideningOp,
Value* ptr) {
MOZ_ASSERT(typeIndex != fieldIndex);
MOZ_ASSERT(Classify(op_) == OpKind::StructGet);
if (!readStructTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const StructType& structType = typeDef.structType();
if (!readFieldIndex(fieldIndex, structType)) {
return false;
}
if (!popWithType(RefType::fromTypeDef(&typeDef, true), ptr)) {
return false;
}
StorageType StorageType = structType.fields_[*fieldIndex].type;
if (StorageType.isValType() && wideningOp != FieldWideningOp::None) {
return fail("must not specify signedness for unpacked field type");
}
if (!StorageType.isValType() && wideningOp == FieldWideningOp::None) {
return fail("must specify signedness for packed field type");
}
return push(StorageType.widenToValType());
}
template <typename Policy>
inline bool OpIter<Policy>::readStructSet(uint32_t* typeIndex,
uint32_t* fieldIndex, Value* ptr,
Value* val) {
MOZ_ASSERT(typeIndex != fieldIndex);
MOZ_ASSERT(Classify(op_) == OpKind::StructSet);
if (!readStructTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const StructType& structType = typeDef.structType();
if (!readFieldIndex(fieldIndex, structType)) {
return false;
}
if (!popWithType(structType.fields_[*fieldIndex].type.widenToValType(),
val)) {
return false;
}
if (!structType.fields_[*fieldIndex].isMutable) {
return fail("field is not mutable");
}
return popWithType(RefType::fromTypeDef(&typeDef, true), ptr);
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayNew(uint32_t* typeIndex,
Value* numElements, Value* argValue) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayNew);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
if (!popWithType(ValType::I32, numElements)) {
return false;
}
if (!popWithType(arrayType.elementType_.widenToValType(), argValue)) {
return false;
}
return push(RefType::fromTypeDef(&typeDef, false));
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayNewFixed(uint32_t* typeIndex,
uint32_t* numElements,
ValueVector* values) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayNewFixed);
MOZ_ASSERT(values->length() == 0);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
if (!readVarU32(numElements)) {
return false;
}
if (*numElements > MaxArrayNewFixedElements) {
return fail("too many array.new_fixed elements");
}
if (!values->reserve(*numElements)) {
return false;
}
ValType widenedElementType = arrayType.elementType_.widenToValType();
for (uint32_t i = 0; i < *numElements; i++) {
Value v;
if (!popWithType(widenedElementType, &v)) {
return false;
}
values->infallibleAppend(v);
}
return push(RefType::fromTypeDef(&typeDef, false));
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayNewDefault(uint32_t* typeIndex,
Value* numElements) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayNewDefault);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
if (!popWithType(ValType::I32, numElements)) {
return false;
}
if (!arrayType.elementType_.isDefaultable()) {
return fail("array must be defaultable");
}
return push(RefType::fromTypeDef(&typeDef, false));
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayNewData(uint32_t* typeIndex,
uint32_t* segIndex, Value* offset,
Value* numElements) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayNewData);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
if (!readVarU32(segIndex)) {
return fail("unable to read segment index");
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
StorageType elemType = arrayType.elementType_;
if (!elemType.isNumber() && !elemType.isPacked() && !elemType.isVector()) {
return fail("element type must be i8/i16/i32/i64/f32/f64/v128");
}
if (env_.dataCount.isNothing()) {
return fail("datacount section missing");
}
if (*segIndex >= *env_.dataCount) {
return fail("segment index is out of range");
}
if (!popWithType(ValType::I32, numElements)) {
return false;
}
if (!popWithType(ValType::I32, offset)) {
return false;
}
return push(RefType::fromTypeDef(&typeDef, false));
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayNewElem(uint32_t* typeIndex,
uint32_t* segIndex, Value* offset,
Value* numElements) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayNewElem);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
if (!readVarU32(segIndex)) {
return fail("unable to read segment index");
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
StorageType dstElemType = arrayType.elementType_;
if (!dstElemType.isRefType()) {
return fail("element type is not a reftype");
}
if (*segIndex >= env_.elemSegments.length()) {
return fail("segment index is out of range");
}
const ModuleElemSegment& elemSeg = env_.elemSegments[*segIndex];
RefType srcElemType = elemSeg.elemType;
// srcElemType needs to be a subtype (child) of dstElemType
if (!checkIsSubtypeOf(srcElemType, dstElemType.refType())) {
return fail("incompatible element types");
}
if (!popWithType(ValType::I32, numElements)) {
return false;
}
if (!popWithType(ValType::I32, offset)) {
return false;
}
return push(RefType::fromTypeDef(&typeDef, false));
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayInitData(uint32_t* typeIndex,
uint32_t* segIndex, Value* array,
Value* arrayIndex,
Value* segOffset, Value* length) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayInitData);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
if (!readVarU32(segIndex)) {
return fail("unable to read segment index");
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
StorageType elemType = arrayType.elementType_;
if (!elemType.isNumber() && !elemType.isPacked() && !elemType.isVector()) {
return fail("element type must be i8/i16/i32/i64/f32/f64/v128");
}
if (!arrayType.isMutable_) {
return fail("destination array is not mutable");
}
if (env_.dataCount.isNothing()) {
return fail("datacount section missing");
}
if (*segIndex >= *env_.dataCount) {
return fail("segment index is out of range");
}
if (!popWithType(ValType::I32, length)) {
return false;
}
if (!popWithType(ValType::I32, segOffset)) {
return false;
}
if (!popWithType(ValType::I32, arrayIndex)) {
return false;
}
return popWithType(RefType::fromTypeDef(&typeDef, true), array);
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayInitElem(uint32_t* typeIndex,
uint32_t* segIndex, Value* array,
Value* arrayIndex,
Value* segOffset, Value* length) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayInitElem);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
if (!readVarU32(segIndex)) {
return fail("unable to read segment index");
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
StorageType dstElemType = arrayType.elementType_;
if (!arrayType.isMutable_) {
return fail("destination array is not mutable");
}
if (!dstElemType.isRefType()) {
return fail("element type is not a reftype");
}
if (*segIndex >= env_.elemSegments.length()) {
return fail("segment index is out of range");
}
const ModuleElemSegment& elemSeg = env_.elemSegments[*segIndex];
RefType srcElemType = elemSeg.elemType;
// srcElemType needs to be a subtype (child) of dstElemType
if (!checkIsSubtypeOf(srcElemType, dstElemType.refType())) {
return fail("incompatible element types");
}
if (!popWithType(ValType::I32, length)) {
return false;
}
if (!popWithType(ValType::I32, segOffset)) {
return false;
}
if (!popWithType(ValType::I32, arrayIndex)) {
return false;
}
return popWithType(RefType::fromTypeDef(&typeDef, true), array);
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayGet(uint32_t* typeIndex,
FieldWideningOp wideningOp,
Value* index, Value* ptr) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayGet);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
if (!popWithType(ValType::I32, index)) {
return false;
}
if (!popWithType(RefType::fromTypeDef(&typeDef, true), ptr)) {
return false;
}
StorageType elementType = arrayType.elementType_;
if (elementType.isValType() && wideningOp != FieldWideningOp::None) {
return fail("must not specify signedness for unpacked element type");
}
if (!elementType.isValType() && wideningOp == FieldWideningOp::None) {
return fail("must specify signedness for packed element type");
}
return push(elementType.widenToValType());
}
template <typename Policy>
inline bool OpIter<Policy>::readArraySet(uint32_t* typeIndex, Value* val,
Value* index, Value* ptr) {
MOZ_ASSERT(Classify(op_) == OpKind::ArraySet);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
if (!arrayType.isMutable_) {
return fail("array is not mutable");
}
if (!popWithType(arrayType.elementType_.widenToValType(), val)) {
return false;
}
if (!popWithType(ValType::I32, index)) {
return false;
}
return popWithType(RefType::fromTypeDef(&typeDef, true), ptr);
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayLen(Value* ptr) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayLen);
if (!popWithType(RefType::array(), ptr)) {
return false;
}
return push(ValType::I32);
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayCopy(int32_t* elemSize,
bool* elemsAreRefTyped,
Value* dstArray, Value* dstIndex,
Value* srcArray, Value* srcIndex,
Value* numElements) {
// *elemSize is set to 1/2/4/8/16, and *elemsAreRefTyped is set to indicate
// *ref-typeness of elements.
MOZ_ASSERT(Classify(op_) == OpKind::ArrayCopy);
uint32_t dstTypeIndex, srcTypeIndex;
if (!readArrayTypeIndex(&dstTypeIndex)) {
return false;
}
if (!readArrayTypeIndex(&srcTypeIndex)) {
return false;
}
// `dstTypeIndex`/`srcTypeIndex` are ensured by the above to both be array
// types. Reject if:
// * the dst array is not of mutable type
// * the element types are incompatible
const TypeDef& dstTypeDef = env_.types->type(dstTypeIndex);
const ArrayType& dstArrayType = dstTypeDef.arrayType();
const TypeDef& srcTypeDef = env_.types->type(srcTypeIndex);
const ArrayType& srcArrayType = srcTypeDef.arrayType();
StorageType dstElemType = dstArrayType.elementType_;
StorageType srcElemType = srcArrayType.elementType_;
if (!dstArrayType.isMutable_) {
return fail("destination array is not mutable");
}
if (!checkIsSubtypeOf(srcElemType, dstElemType)) {
return fail("incompatible element types");
}
bool dstIsRefType = dstElemType.isRefType();
MOZ_ASSERT(dstIsRefType == srcElemType.isRefType());
*elemSize = int32_t(dstElemType.size());
*elemsAreRefTyped = dstIsRefType;
MOZ_ASSERT(*elemSize >= 1 && *elemSize <= 16);
MOZ_ASSERT_IF(*elemsAreRefTyped, *elemSize == 4 || *elemSize == 8);
if (!popWithType(ValType::I32, numElements)) {
return false;
}
if (!popWithType(ValType::I32, srcIndex)) {
return false;
}
if (!popWithType(RefType::fromTypeDef(&srcTypeDef, true), srcArray)) {
return false;
}
if (!popWithType(ValType::I32, dstIndex)) {
return false;
}
return popWithType(RefType::fromTypeDef(&dstTypeDef, true), dstArray);
}
template <typename Policy>
inline bool OpIter<Policy>::readArrayFill(uint32_t* typeIndex, Value* array,
Value* index, Value* val,
Value* length) {
MOZ_ASSERT(Classify(op_) == OpKind::ArrayFill);
if (!readArrayTypeIndex(typeIndex)) {
return false;
}
const TypeDef& typeDef = env_.types->type(*typeIndex);
const ArrayType& arrayType = typeDef.arrayType();
if (!arrayType.isMutable_) {
return fail("destination array is not mutable");
}
if (!popWithType(ValType::I32, length)) {
return false;
}
if (!popWithType(arrayType.elementType_.widenToValType(), val)) {
return false;
}
if (!popWithType(ValType::I32, index)) {
return false;
}
return popWithType(RefType::fromTypeDef(&typeDef, true), array);
}
template <typename Policy>
inline bool OpIter<Policy>::readRefTest(bool nullable, RefType* sourceType,
RefType* destType, Value* ref) {
MOZ_ASSERT(Classify(op_) == OpKind::RefTest);
if (!readHeapType(nullable, destType)) {
return false;
}
StackType inputType;
if (!popWithType(destType->topType(), ref, &inputType)) {
return false;
}
*sourceType = inputType.valTypeOr(RefType::any()).refType();
return push(ValType(ValType::I32));
}
template <typename Policy>
inline bool OpIter<Policy>::readRefCast(bool nullable, RefType* sourceType,
RefType* destType, Value* ref) {
MOZ_ASSERT(Classify(op_) == OpKind::RefCast);
if (!readHeapType(nullable, destType)) {
return false;
}
StackType inputType;
if (!popWithType(destType->topType(), ref, &inputType)) {
return false;
}
*sourceType = inputType.valTypeOr(RefType::any()).refType();
return push(*destType);
}
// `br_on_cast <flags> <labelRelativeDepth> <rt1> <rt2>`
// branches if a reference has a given heap type.
//
// V6 spec text follows - note that br_on_cast and br_on_cast_fail are both
// handled by this function (disambiguated by a flag).
//
// * `br_on_cast <labelidx> <reftype> <reftype>` branches if a reference has a
// given type
// - `br_on_cast $l rt1 rt2 : [t0* rt1] -> [t0* rt1\rt2]`
// - iff `$l : [t0* rt2]`
// - and `rt2 <: rt1`
// - passes operand along with branch under target type, plus possible extra
// args
// - if `rt2` contains `null`, branches on null, otherwise does not
// * `br_on_cast_fail <labelidx> <reftype> <reftype>` branches if a reference
// does not have a given type
// - `br_on_cast_fail $l rt1 rt2 : [t0* rt1] -> [t0* rt2]`
// - iff `$l : [t0* rt1\rt2]`
// - and `rt2 <: rt1`
// - passes operand along with branch, plus possible extra args
// - if `rt2` contains `null`, does not branch on null, otherwise does
// where:
// - `(ref null1? ht1)\(ref null ht2) = (ref ht1)`
// - `(ref null1? ht1)\(ref ht2) = (ref null1? ht1)`
//
// The `rt1\rt2` syntax is a "diff" - it is basically rt1 minus rt2, because a
// successful cast to rt2 will branch away. So if rt2 allows null, the result
// after a non-branch will be non-null; on the other hand, if rt2 is
// non-nullable, the cast will have nothing to say about nullability and the
// nullability of rt1 will be preserved.
//
// `values` will be nonempty after the call, and its last entry will be the
// type that causes a branch (rt1\rt2 or rt2, depending).
enum class BrOnCastFlags : uint8_t {
SourceNullable = 0x1,
DestNullable = 0x1 << 1,
AllowedMask = uint8_t(SourceNullable) | uint8_t(DestNullable),
};
template <typename Policy>
inline bool OpIter<Policy>::readBrOnCast(bool onSuccess,
uint32_t* labelRelativeDepth,
RefType* sourceType, RefType* destType,
ResultType* labelType,
ValueVector* values) {
MOZ_ASSERT(Classify(op_) == OpKind::BrOnCast);
uint8_t flags;
if (!readFixedU8(&flags)) {
return fail("unable to read br_on_cast flags");
}
if ((flags & ~uint8_t(BrOnCastFlags::AllowedMask)) != 0) {
return fail("invalid br_on_cast flags");
}
bool sourceNullable = flags & uint8_t(BrOnCastFlags::SourceNullable);
bool destNullable = flags & uint8_t(BrOnCastFlags::DestNullable);
if (!readVarU32(labelRelativeDepth)) {
return fail("unable to read br_on_cast depth");
}
// This is distinct from the actual source type we pop from the stack, which
// can be more specific and allow for better optimizations.
RefType immediateSourceType;
if (!readHeapType(sourceNullable, &immediateSourceType)) {
return fail("unable to read br_on_cast source type");
}
if (!readHeapType(destNullable, destType)) {
return fail("unable to read br_on_cast dest type");
}
// Check that source and destination types are compatible
if (!checkIsSubtypeOf(*destType, immediateSourceType)) {
return fail(
"type mismatch: source and destination types for cast are "
"incompatible");
}
RefType typeOnSuccess = *destType;
// This is rt1\rt2
RefType typeOnFail =
destNullable ? immediateSourceType.asNonNullable() : immediateSourceType;
RefType typeOnBranch = onSuccess ? typeOnSuccess : typeOnFail;
RefType typeOnFallthrough = onSuccess ? typeOnFail : typeOnSuccess;
// Get the branch target type, which will also determine the type of extra
// values that are passed along on branch.
Control* block = nullptr;
if (!getControl(*labelRelativeDepth, &block)) {
return false;
}
*labelType = block->branchTargetType();
// Check we have at least one value slot in the branch target type, so as to
// receive the casted or non-casted type when we branch.
const size_t labelTypeNumValues = labelType->length();
if (labelTypeNumValues < 1) {
return fail("type mismatch: branch target type has no value types");
}
// The last value slot in the branch target type is what is being cast.
// This slot is guaranteed to exist by the above check.
// Check that the branch target type can accept typeOnBranch.
if (!checkIsSubtypeOf(typeOnBranch, (*labelType)[labelTypeNumValues - 1])) {
return false;
}
// Replace the top operand with the result of falling through. Even branching
// on success can change the type on top of the stack on fallthrough.
Value inputValue;
StackType inputType;
if (!popWithType(immediateSourceType, &inputValue, &inputType)) {
return false;
}
*sourceType = inputType.valTypeOr(immediateSourceType).refType();
infalliblePush(TypeAndValue(typeOnFallthrough, inputValue));
// Create a copy of the branch target type, with the relevant value slot
// replaced by typeOnFallthrough.
ValTypeVector fallthroughTypes;
if (!labelType->cloneToVector(&fallthroughTypes)) {
return false;
}
fallthroughTypes[labelTypeNumValues - 1] = typeOnFallthrough;
return checkTopTypeMatches(ResultType::Vector(fallthroughTypes), values,
/*rewriteStackTypes=*/true);
}
template <typename Policy>
inline bool OpIter<Policy>::readRefConversion(RefType operandType,
RefType resultType,
Value* operandValue) {
MOZ_ASSERT(Classify(op_) == OpKind::RefConversion);
StackType actualOperandType;
if (!popWithType(ValType(operandType), operandValue, &actualOperandType)) {
return false;
}
// The result nullability is the same as the operand nullability
bool outputNullable = actualOperandType.isNullableAsOperand();
infalliblePush(ValType(resultType.withIsNullable(outputNullable)));
return true;
}
#endif // ENABLE_WASM_GC
#ifdef ENABLE_WASM_SIMD
template <typename Policy>
inline bool OpIter<Policy>::readLaneIndex(uint32_t inputLanes,
uint32_t* laneIndex) {
uint8_t tmp;
if (!readFixedU8(&tmp)) {
return false; // Caller signals error
}
if (tmp >= inputLanes) {
return false;
}
*laneIndex = tmp;
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readExtractLane(ValType resultType,
uint32_t inputLanes,
uint32_t* laneIndex, Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::ExtractLane);
if (!readLaneIndex(inputLanes, laneIndex)) {
return fail("missing or invalid extract_lane lane index");
}
if (!popWithType(ValType::V128, input)) {
return false;
}
infalliblePush(resultType);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readReplaceLane(ValType operandType,
uint32_t inputLanes,
uint32_t* laneIndex,
Value* baseValue, Value* operand) {
MOZ_ASSERT(Classify(op_) == OpKind::ReplaceLane);
if (!readLaneIndex(inputLanes, laneIndex)) {
return fail("missing or invalid replace_lane lane index");
}
if (!popWithType(operandType, operand)) {
return false;
}
if (!popWithType(ValType::V128, baseValue)) {
return false;
}
infalliblePush(ValType::V128);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readVectorShift(Value* baseValue, Value* shift) {
MOZ_ASSERT(Classify(op_) == OpKind::VectorShift);
if (!popWithType(ValType::I32, shift)) {
return false;
}
if (!popWithType(ValType::V128, baseValue)) {
return false;
}
infalliblePush(ValType::V128);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readVectorShuffle(Value* v1, Value* v2,
V128* selectMask) {
MOZ_ASSERT(Classify(op_) == OpKind::VectorShuffle);
for (unsigned char& byte : selectMask->bytes) {
uint8_t tmp;
if (!readFixedU8(&tmp)) {
return fail("unable to read shuffle index");
}
if (tmp > 31) {
return fail("shuffle index out of range");
}
byte = tmp;
}
if (!popWithType(ValType::V128, v2)) {
return false;
}
if (!popWithType(ValType::V128, v1)) {
return false;
}
infalliblePush(ValType::V128);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readV128Const(V128* value) {
MOZ_ASSERT(Classify(op_) == OpKind::V128);
if (!d_.readV128Const(value)) {
return false;
}
return push(ValType::V128);
}
template <typename Policy>
inline bool OpIter<Policy>::readLoadSplat(uint32_t byteSize,
LinearMemoryAddress<Value>* addr) {
MOZ_ASSERT(Classify(op_) == OpKind::Load);
if (!readLinearMemoryAddress(byteSize, addr)) {
return false;
}
infalliblePush(ValType::V128);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readLoadExtend(LinearMemoryAddress<Value>* addr) {
MOZ_ASSERT(Classify(op_) == OpKind::Load);
if (!readLinearMemoryAddress(/*byteSize=*/8, addr)) {
return false;
}
infalliblePush(ValType::V128);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readLoadLane(uint32_t byteSize,
LinearMemoryAddress<Value>* addr,
uint32_t* laneIndex, Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::LoadLane);
if (!popWithType(ValType::V128, input)) {
return false;
}
if (!readLinearMemoryAddress(byteSize, addr)) {
return false;
}
uint32_t inputLanes = 16 / byteSize;
if (!readLaneIndex(inputLanes, laneIndex)) {
return fail("missing or invalid load_lane lane index");
}
infalliblePush(ValType::V128);
return true;
}
template <typename Policy>
inline bool OpIter<Policy>::readStoreLane(uint32_t byteSize,
LinearMemoryAddress<Value>* addr,
uint32_t* laneIndex, Value* input) {
MOZ_ASSERT(Classify(op_) == OpKind::StoreLane);
if (!popWithType(ValType::V128, input)) {
return false;
}
if (!readLinearMemoryAddress(byteSize, addr)) {
return false;
}
uint32_t inputLanes = 16 / byteSize;
if (!readLaneIndex(inputLanes, laneIndex)) {
return fail("missing or invalid store_lane lane index");
}
return true;
}
#endif // ENABLE_WASM_SIMD
template <typename Policy>
inline bool OpIter<Policy>::readCallBuiltinModuleFunc(
const BuiltinModuleFunc** builtinModuleFunc, ValueVector* params) {
MOZ_ASSERT(Classify(op_) == OpKind::CallBuiltinModuleFunc);
uint32_t id;
if (!d_.readVarU32(&id)) {
return false;
}
if (id >= uint32_t(BuiltinModuleFuncId::Limit)) {
return fail("index out of range");
}
*builtinModuleFunc = &BuiltinModuleFuncs::getFromId(BuiltinModuleFuncId(id));
if ((*builtinModuleFunc)->usesMemory() && env_.numMemories() == 0) {
return fail("can't touch memory without memory");
}
const FuncType& funcType = *(*builtinModuleFunc)->funcType();
if (!popCallArgs(funcType.args(), params)) {
return false;
}
return push(ResultType::Vector(funcType.results()));
}
} // namespace wasm
} // namespace js
static_assert(std::is_trivially_copyable<
js::wasm::TypeAndValueT<mozilla::Nothing>>::value,
"Must be trivially copyable");
static_assert(std::is_trivially_destructible<
js::wasm::TypeAndValueT<mozilla::Nothing>>::value,
"Must be trivially destructible");
static_assert(std::is_trivially_copyable<
js::wasm::ControlStackEntry<mozilla::Nothing>>::value,
"Must be trivially copyable");
static_assert(std::is_trivially_destructible<
js::wasm::ControlStackEntry<mozilla::Nothing>>::value,
"Must be trivially destructible");
#endif // wasm_op_iter_h