author Sebastian Hengst <archaeopteryx@coole-files.de>
Tue, 04 Sep 2018 20:07:59 +0300
changeset 434558 034c5ef24e98b0ce85fa849face079f568eb397c
parent 434476 c91d997687bf454984288ff865e05aa6a4e680ea
child 434746 295b49332ca1079fc66cef1e8e5ee674491d0e85
permissions -rw-r--r--
Backed out changeset c91d997687bf (bug 1487327) for busting beta. a=backout

/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*-
 * vim: set ts=8 sts=4 et sw=4 tw=99:
 * Copyright 2016 Mozilla Foundation
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *     http://www.apache.org/licenses/LICENSE-2.0
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * See the License for the specific language governing permissions and
 * limitations under the License.

#ifndef wasm_code_h
#define wasm_code_h

#include "js/HashTable.h"
#include "threading/ExclusiveData.h"
#include "vm/MutexIDs.h"
#include "wasm/WasmTypes.h"

namespace js {

struct AsmJSMetadata;
class WasmInstanceObject;

namespace wasm {

struct LinkDataTier;
struct MetadataTier;
struct Metadata;
class LinkData;

// ShareableBytes is a reference-counted Vector of bytes.

struct ShareableBytes : ShareableBase<ShareableBytes>
    // Vector is 'final', so instead make Vector a member and add boilerplate.
    Bytes bytes;
    ShareableBytes() = default;
    explicit ShareableBytes(Bytes&& bytes) : bytes(std::move(bytes)) {}
    size_t sizeOfExcludingThis(MallocSizeOf m) const { return bytes.sizeOfExcludingThis(m); }
    const uint8_t* begin() const { return bytes.begin(); }
    const uint8_t* end() const { return bytes.end(); }
    size_t length() const { return bytes.length(); }
    bool append(const uint8_t *p, uint32_t ct) { return bytes.append(p, ct); }

typedef RefPtr<ShareableBytes> MutableBytes;
typedef RefPtr<const ShareableBytes> SharedBytes;

// Executable code must be deallocated specially.

struct FreeCode {
    uint32_t codeLength;
    FreeCode() : codeLength(0) {}
    explicit FreeCode(uint32_t codeLength) : codeLength(codeLength) {}
    void operator()(uint8_t* codeBytes);

using UniqueCodeBytes = UniquePtr<uint8_t, FreeCode>;

class Code;
class CodeTier;
class ModuleSegment;
class LazyStubSegment;

// CodeSegment contains common helpers for determining the base and length of a
// code segment and if a pc belongs to this segment. It is inherited by:
// - ModuleSegment, i.e. the code segment of a Module, generated
// eagerly when a Module is instanciated.
// - LazyStubSegment, i.e. the code segment of entry stubs that are lazily
// generated.

class CodeSegment
    static UniqueCodeBytes AllocateCodeBytes(uint32_t codeLength);

    enum class Kind {

    CodeSegment(UniqueCodeBytes bytes, uint32_t length, Kind kind)
      : bytes_(std::move(bytes)),

    bool initialize(const CodeTier& codeTier);

    const UniqueCodeBytes bytes_;
    const uint32_t        length_;
    const Kind            kind_;
    const CodeTier*       codeTier_;
    bool                  unregisterOnDestroy_;

    bool initialized() const { return !!codeTier_; }

    bool isLazyStubs() const { return kind_ == Kind::LazyStubs; }
    bool isModule() const { return kind_ == Kind::Module; }
    const ModuleSegment* asModule() const {
        return (ModuleSegment*) this;
    const LazyStubSegment* asLazyStub() const {
        return (LazyStubSegment*) this;

    uint8_t* base() const { return bytes_.get(); }
    uint32_t length() const { MOZ_ASSERT(length_ != UINT32_MAX); return length_; }

    bool containsCodePC(const void* pc) const {
        return pc >= base() && pc < (base() + length_);

    const CodeTier& codeTier() const { MOZ_ASSERT(initialized()); return *codeTier_; }
    const Code& code() const;

    void addSizeOfMisc(MallocSizeOf mallocSizeOf, size_t* code) const;

// A wasm ModuleSegment owns the allocated executable code for a wasm module.

typedef UniquePtr<ModuleSegment> UniqueModuleSegment;

class ModuleSegment : public CodeSegment
    const Tier      tier_;
    uint8_t* const  trapCode_;

    ModuleSegment(Tier tier,
                  UniqueCodeBytes codeBytes,
                  uint32_t codeLength,
                  const LinkDataTier& linkData);

    static UniqueModuleSegment create(Tier tier,
                                      jit::MacroAssembler& masm,
                                      const LinkDataTier& linkData);
    static UniqueModuleSegment create(Tier tier,
                                      const Bytes& unlinkedBytes,
                                      const LinkDataTier& linkData);

    bool initialize(const CodeTier& codeTier,
                    const ShareableBytes& bytecode,
                    const LinkDataTier& linkData,
                    const Metadata& metadata,
                    const MetadataTier& metadataTier);

    Tier tier() const { return tier_; }

    // Pointers to stubs to which PC is redirected from the signal-handler.

    uint8_t* trapCode() const { return trapCode_; }

    // Structured clone support:

    size_t serializedSize() const;
    uint8_t* serialize(uint8_t* cursor, const LinkDataTier& linkData) const;
    static const uint8_t* deserialize(const uint8_t* cursor, const LinkDataTier& linkData,
                                      UniqueModuleSegment* segment);

    const CodeRange* lookupRange(const void* pc) const;

    void addSizeOfMisc(mozilla::MallocSizeOf mallocSizeOf, size_t* code, size_t* data) const;

// A FuncExport represents a single function definition inside a wasm Module
// that has been exported one or more times. A FuncExport represents an
// internal entry point that can be called via function definition index by
// Instance::callExport(). To allow O(log(n)) lookup of a FuncExport by
// function definition index, the FuncExportVector is stored sorted by
// function definition index.

class FuncExport
    FuncType funcType_;
    MOZ_INIT_OUTSIDE_CTOR struct CacheablePod {
        uint32_t funcIndex_;
        uint32_t funcCodeRangeIndex_;
        uint32_t eagerInterpEntryOffset_; // Machine code offset
        bool     hasEagerStubs_;
    } pod;

    FuncExport() = default;
    explicit FuncExport(FuncType&& funcType, uint32_t funcIndex, bool hasEagerStubs)
      : funcType_(std::move(funcType))
        pod.funcIndex_ = funcIndex;
        pod.funcCodeRangeIndex_ = UINT32_MAX;
        pod.eagerInterpEntryOffset_ = UINT32_MAX;
        pod.hasEagerStubs_ = hasEagerStubs;
    void initEagerInterpEntryOffset(uint32_t entryOffset) {
        MOZ_ASSERT(pod.eagerInterpEntryOffset_ == UINT32_MAX);
        pod.eagerInterpEntryOffset_ = entryOffset;
    void initFuncCodeRangeIndex(uint32_t codeRangeIndex) {
        MOZ_ASSERT(pod.funcCodeRangeIndex_ == UINT32_MAX);
        pod.funcCodeRangeIndex_ = codeRangeIndex;

    bool hasEagerStubs() const {
        return pod.hasEagerStubs_;
    const FuncType& funcType() const {
        return funcType_;
    uint32_t funcIndex() const {
        return pod.funcIndex_;
    uint32_t funcCodeRangeIndex() const {
        MOZ_ASSERT(pod.funcCodeRangeIndex_ != UINT32_MAX);
        return pod.funcCodeRangeIndex_;
    uint32_t eagerInterpEntryOffset() const {
        MOZ_ASSERT(pod.eagerInterpEntryOffset_ != UINT32_MAX);
        return pod.eagerInterpEntryOffset_;

    bool clone(const FuncExport& src) {
        mozilla::PodAssign(&pod, &src.pod);
        return funcType_.clone(src.funcType_);


typedef Vector<FuncExport, 0, SystemAllocPolicy> FuncExportVector;

// An FuncImport contains the runtime metadata needed to implement a call to an
// imported function. Each function import has two call stubs: an optimized path
// into JIT code and a slow path into the generic C++ js::Invoke and these
// offsets of these stubs are stored so that function-import callsites can be
// dynamically patched at runtime.

class FuncImport
    FuncType funcType_;
    struct CacheablePod {
        uint32_t tlsDataOffset_;
        uint32_t interpExitCodeOffset_; // Machine code offset
        uint32_t jitExitCodeOffset_;    // Machine code offset
    } pod;

    FuncImport() {
        memset(&pod, 0, sizeof(CacheablePod));

    FuncImport(FuncType&& funcType, uint32_t tlsDataOffset)
      : funcType_(std::move(funcType))
        pod.tlsDataOffset_ = tlsDataOffset;
        pod.interpExitCodeOffset_ = 0;
        pod.jitExitCodeOffset_ = 0;

    void initInterpExitOffset(uint32_t off) {
        pod.interpExitCodeOffset_ = off;
    void initJitExitOffset(uint32_t off) {
        pod.jitExitCodeOffset_ = off;

    const FuncType& funcType() const {
        return funcType_;
    uint32_t tlsDataOffset() const {
        return pod.tlsDataOffset_;
    uint32_t interpExitCodeOffset() const {
        return pod.interpExitCodeOffset_;
    uint32_t jitExitCodeOffset() const {
        return pod.jitExitCodeOffset_;

    bool clone(const FuncImport& src) {
        mozilla::PodAssign(&pod, &src.pod);
        return funcType_.clone(src.funcType_);


typedef Vector<FuncImport, 0, SystemAllocPolicy> FuncImportVector;

// A wasm module can either use no memory, a unshared memory (ArrayBuffer) or
// shared memory (SharedArrayBuffer).

enum class MemoryUsage
    None = false,
    Unshared = 1,
    Shared = 2

// NameInBytecode represents a name that is embedded in the wasm bytecode.
// The presence of NameInBytecode implies that bytecode has been kept.

struct NameInBytecode
    uint32_t offset;
    uint32_t length;

      : offset(UINT32_MAX), length(0)
    NameInBytecode(uint32_t offset, uint32_t length)
      : offset(offset), length(length)

typedef Vector<NameInBytecode, 0, SystemAllocPolicy> NameInBytecodeVector;

// CustomSection represents a custom section in the bytecode which can be
// extracted via Module.customSections. The (offset, length) pair does not
// include the custom section name.

struct CustomSection
    NameInBytecode name;
    uint32_t offset;
    uint32_t length;

    CustomSection() = default;
    CustomSection(NameInBytecode name, uint32_t offset, uint32_t length)
      : name(name), offset(offset), length(length)

typedef Vector<CustomSection, 0, SystemAllocPolicy> CustomSectionVector;
typedef Vector<ValTypeVector, 0, SystemAllocPolicy> FuncArgTypesVector;
typedef Vector<ExprType, 0, SystemAllocPolicy> FuncReturnTypesVector;

// Metadata holds all the data that is needed to describe compiled wasm code
// at runtime (as opposed to data that is only used to statically link or
// instantiate a module).
// Metadata is built incrementally by ModuleGenerator and then shared immutably
// between modules.
// The Metadata structure is split into tier-invariant and tier-variant parts;
// the former points to instances of the latter.  Additionally, the asm.js
// subsystem subclasses the Metadata, adding more tier-invariant data, some of
// which is serialized.  See AsmJS.cpp.

struct MetadataCacheablePod
    ModuleKind            kind;
    MemoryUsage           memoryUsage;
    HasGcTypes            temporaryHasGcTypes;
    uint32_t              minMemoryLength;
    uint32_t              globalDataLength;
    Maybe<uint32_t>       maxMemoryLength;
    Maybe<uint32_t>       startFuncIndex;
    Maybe<NameInBytecode> moduleName;
    bool                  filenameIsURL;

    explicit MetadataCacheablePod(ModuleKind kind)
      : kind(kind),

typedef uint8_t ModuleHash[8];

struct Metadata : public ShareableBase<Metadata>, public MetadataCacheablePod
    FuncTypeWithIdVector  funcTypeIds;
    GlobalDescVector      globals;
    TableDescVector       tables;
    NameInBytecodeVector  funcNames;
    CustomSectionVector   customSections;
    CacheableChars        filename;
    CacheableChars        sourceMapURL;

    // Debug-enabled code is not serialized.
    bool                  debugEnabled;
    FuncArgTypesVector    debugFuncArgTypes;
    FuncReturnTypesVector debugFuncReturnTypes;
    ModuleHash            debugHash;

    explicit Metadata(ModuleKind kind = ModuleKind::Wasm)
      : MetadataCacheablePod(kind),
    virtual ~Metadata() {}

    MetadataCacheablePod& pod() { return *this; }
    const MetadataCacheablePod& pod() const { return *this; }

    bool usesMemory() const { return memoryUsage != MemoryUsage::None; }
    bool usesSharedMemory() const { return memoryUsage == MemoryUsage::Shared; }

    // AsmJSMetadata derives Metadata iff isAsmJS(). Mostly this distinction is
    // encapsulated within AsmJS.cpp, but the additional virtual functions allow
    // asm.js to override wasm behavior in the handful of cases that can't be
    // easily encapsulated by AsmJS.cpp.

    bool isAsmJS() const {
        return kind == ModuleKind::AsmJS;
    const AsmJSMetadata& asAsmJS() const {
        return *(const AsmJSMetadata*)this;
    virtual bool mutedErrors() const {
        return false;
    virtual const char16_t* displayURL() const {
        return nullptr;
    virtual ScriptSource* maybeScriptSource() const {
        return nullptr;

    // The Developer-Facing Display Conventions section of the WebAssembly Web
    // API spec defines two cases for displaying a wasm function name:
    //  1. the function name stands alone
    //  2. the function name precedes the location

    enum NameContext { Standalone, BeforeLocation };

    virtual bool getFuncName(NameContext ctx, const Bytes* maybeBytecode, uint32_t funcIndex,
                             UTF8Bytes* name) const;

    bool getFuncNameStandalone(const Bytes* maybeBytecode, uint32_t funcIndex, UTF8Bytes* name) const {
        return getFuncName(NameContext::Standalone, maybeBytecode, funcIndex, name);
    bool getFuncNameBeforeLocation(const Bytes* maybeBytecode, uint32_t funcIndex, UTF8Bytes* name) const {
        return getFuncName(NameContext::BeforeLocation, maybeBytecode, funcIndex, name);


typedef RefPtr<Metadata> MutableMetadata;
typedef RefPtr<const Metadata> SharedMetadata;

struct MetadataTier
    explicit MetadataTier(Tier tier) : tier(tier) {}

    const Tier            tier;

    CodeRangeVector       codeRanges;
    CallSiteVector        callSites;
    TrapSiteVectorArray   trapSites;
    FuncImportVector      funcImports;
    FuncExportVector      funcExports;

    // Debug information, not serialized.
    Uint32Vector          debugTrapFarJumpOffsets;
    Uint32Vector          debugFuncToCodeRange;

    FuncExport& lookupFuncExport(uint32_t funcIndex, size_t* funcExportIndex = nullptr);
    const FuncExport& lookupFuncExport(uint32_t funcIndex, size_t* funcExportIndex = nullptr) const;

    bool clone(const MetadataTier& src);


using UniqueMetadataTier = UniquePtr<MetadataTier>;

// LazyStubSegment is a code segment lazily generated for function entry stubs
// (both interpreter and jit ones).
// Because a stub is usually small (a few KiB) and an executable code segment
// isn't (64KiB), a given stub segment can contain entry stubs of many
// functions.

using UniqueLazyStubSegment = UniquePtr<LazyStubSegment>;
using LazyStubSegmentVector = Vector<UniqueLazyStubSegment, 0, SystemAllocPolicy>;

class LazyStubSegment : public CodeSegment
    CodeRangeVector codeRanges_;
    size_t usedBytes_;

    LazyStubSegment(UniqueCodeBytes bytes, size_t length)
      : CodeSegment(std::move(bytes), length, CodeSegment::Kind::LazyStubs),

    static UniqueLazyStubSegment create(const CodeTier& codeTier, size_t codeLength);

    static size_t AlignBytesNeeded(size_t bytes) { return AlignBytes(bytes, gc::SystemPageSize()); }

    bool hasSpace(size_t bytes) const;
    bool addStubs(size_t codeLength, const Uint32Vector& funcExportIndices,
                  const FuncExportVector& funcExports, const CodeRangeVector& codeRanges,
                  uint8_t** codePtr, size_t* indexFirstInsertedCodeRange);

    const CodeRangeVector& codeRanges() const { return codeRanges_; }
    const CodeRange* lookupRange(const void* pc) const;

    void addSizeOfMisc(MallocSizeOf mallocSizeOf, size_t* code, size_t* data) const;

// LazyFuncExport helps to efficiently lookup a CodeRange from a given function
// index. It is inserted in a vector sorted by function index, to perform
// binary search on it later.

struct LazyFuncExport
    size_t funcIndex;
    size_t lazyStubSegmentIndex;
    size_t funcCodeRangeIndex;
    LazyFuncExport(size_t funcIndex, size_t lazyStubSegmentIndex, size_t funcCodeRangeIndex)
      : funcIndex(funcIndex),

using LazyFuncExportVector = Vector<LazyFuncExport, 0, SystemAllocPolicy>;

// LazyStubTier contains all the necessary information for lazy function entry
// stubs that are generated at runtime. None of its data is ever serialized.
// It must be protected by a lock, because the main thread can both read and
// write lazy stubs at any time while a background thread can regenerate lazy
// stubs for tier2 at any time.

class LazyStubTier
    LazyStubSegmentVector stubSegments_;
    LazyFuncExportVector exports_;
    size_t lastStubSegmentIndex_;

    bool createMany(HasGcTypes gcTypesEnabled, const Uint32Vector& funcExportIndices,
                    const CodeTier& codeTier, size_t* stubSegmentIndex);

    LazyStubTier() : lastStubSegmentIndex_(0) {}

    bool empty() const { return stubSegments_.empty(); }
    bool hasStub(uint32_t funcIndex) const;

    // Returns a pointer to the raw interpreter entry of a given function which
    // stubs have been lazily generated.
    void* lookupInterpEntry(uint32_t funcIndex) const;

    // Creates one lazy stub for the exported function, for which the jit entry
    // will be set to the lazily-generated one.
    bool createOne(uint32_t funcExportIndex, const CodeTier& codeTier);

    // Create one lazy stub for all the functions in funcExportIndices, putting
    // them in a single stub. Jit entries won't be used until
    // setJitEntries() is actually called, after the Code owner has committed
    // tier2.
    bool createTier2(HasGcTypes gcTypesEnabled, const Uint32Vector& funcExportIndices,
                     const CodeTier& codeTier, Maybe<size_t>* stubSegmentIndex);
    void setJitEntries(const Maybe<size_t>& stubSegmentIndex, const Code& code);

    void addSizeOfMisc(MallocSizeOf mallocSizeOf, size_t* code, size_t* data) const;

// CodeTier contains all the data related to a given compilation tier. It is
// built during module generation and then immutably stored in a Code.

typedef UniquePtr<CodeTier> UniqueCodeTier;
typedef UniquePtr<const CodeTier> UniqueConstCodeTier;

class CodeTier
    const Code* code_;

    // Serialized information.
    const UniqueMetadataTier metadata_;
    const UniqueModuleSegment segment_;

    // Lazy stubs, not serialized.
    ExclusiveData<LazyStubTier> lazyStubs_;

    static const MutexId& mutexForTier(Tier tier) {
        if (tier == Tier::Baseline)
            return mutexid::WasmLazyStubsTier1;
        MOZ_ASSERT(tier == Tier::Ion);
        return mutexid::WasmLazyStubsTier2;

    CodeTier(UniqueMetadataTier metadata, UniqueModuleSegment segment)
      : code_(nullptr),

    bool initialized() const { return !!code_ && segment_->initialized(); }

    bool initialize(const Code& code,
                    const ShareableBytes& bytecode,
                    const LinkDataTier& linkData,
                    const Metadata& metadata);

    Tier tier() const { return segment_->tier(); }
    const ExclusiveData<LazyStubTier>& lazyStubs() const { return lazyStubs_; }
    const MetadataTier& metadata() const { return *metadata_.get(); }
    const ModuleSegment& segment() const { return *segment_.get(); }
    const Code& code() const { MOZ_ASSERT(initialized()); return *code_; }

    const CodeRange* lookupRange(const void* pc) const;

    size_t serializedSize() const;
    uint8_t* serialize(uint8_t* cursor, const LinkDataTier& linkData) const;
    static const uint8_t* deserialize(const uint8_t* cursor, const LinkDataTier& linkData,
                                      UniqueCodeTier* codeTier);
    void addSizeOfMisc(MallocSizeOf mallocSizeOf, size_t* code, size_t* data) const;

// Jump tables to take tiering into account, when calling either from wasm to
// wasm (through rabaldr) or from jit to wasm (jit entry).

class JumpTables
    using TablePointer = mozilla::UniquePtr<void*[], JS::FreePolicy>;

    CompileMode mode_;
    TablePointer tiering_;
    TablePointer jit_;
    size_t numFuncs_;

    bool init(CompileMode mode, const ModuleSegment& ms, const CodeRangeVector& codeRanges);

    void setJitEntry(size_t i, void* target) const {
        // See comment in wasm::Module::finishTier2 and JumpTables::init.
        MOZ_ASSERT(i < numFuncs_);
        jit_.get()[2 * i] = target;
        jit_.get()[2 * i + 1] = target;
    void** getAddressOfJitEntry(size_t i) const {
        MOZ_ASSERT(i < numFuncs_);
        MOZ_ASSERT(jit_.get()[2 * i]);
        return &jit_.get()[2 * i];
    size_t funcIndexFromJitEntry(void** target) const {
        MOZ_ASSERT(target >= &jit_.get()[0]);
        MOZ_ASSERT(target <= &(jit_.get()[2 * numFuncs_ - 1]));
        size_t index = (intptr_t*)target - (intptr_t*)&jit_.get()[0];
        MOZ_ASSERT(index % 2 == 0);
        return index / 2;

    void setTieringEntry(size_t i, void* target) const {
        MOZ_ASSERT(i < numFuncs_);
        // See comment in wasm::Module::finishTier2.
        if (mode_ == CompileMode::Tier1)
            tiering_.get()[i] = target;
    void** tiering() const {
        return tiering_.get();

    size_t sizeOfMiscExcludingThis() const {
        // 2 words per function for the jit entry table, plus maybe 1 per
        // function if we're tiering.
        return sizeof(void*) * (2 + (tiering_ ? 1 : 0)) * numFuncs_;

// Code objects own executable code and the metadata that describe it. A single
// Code object is normally shared between a module and all its instances.
// profilingLabels_ is lazily initialized, but behind a lock.

typedef RefPtr<const Code> SharedCode;
typedef RefPtr<Code> MutableCode;

class Code : public ShareableBase<Code>
    UniqueCodeTier                      tier1_;
    mutable UniqueConstCodeTier         tier2_; // Access only when hasTier2() is true
    mutable Atomic<bool>                hasTier2_;
    SharedMetadata                      metadata_;
    ExclusiveData<CacheableCharsVector> profilingLabels_;
    JumpTables                          jumpTables_;

    Code(UniqueCodeTier tier1, const Metadata& metadata, JumpTables&& maybeJumpTables);
    bool initialized() const { return tier1_->initialized(); }

    bool initialize(const ShareableBytes& bytecode, const LinkDataTier& linkData);

    void setTieringEntry(size_t i, void* target) const { jumpTables_.setTieringEntry(i, target); }
    void** tieringJumpTable() const { return jumpTables_.tiering(); }

    void setJitEntry(size_t i, void* target) const { jumpTables_.setJitEntry(i, target); }
    void** getAddressOfJitEntry(size_t i) const { return jumpTables_.getAddressOfJitEntry(i); }
    uint32_t getFuncIndex(JSFunction* fun) const;

    bool setTier2(UniqueCodeTier tier2, const ShareableBytes& bytecode,
                  const LinkDataTier& linkData) const;
    void commitTier2() const;

    bool hasTier2() const { return hasTier2_; }
    Tiers tiers() const;
    bool hasTier(Tier t) const;

    Tier stableTier() const;    // This is stable during a run
    Tier bestTier() const;      // This may transition from Baseline -> Ion at any time

    const CodeTier& codeTier(Tier tier) const;
    const Metadata& metadata() const { return *metadata_; }

    const ModuleSegment& segment(Tier iter) const {
        return codeTier(iter).segment();
    const MetadataTier& metadata(Tier iter) const {
        return codeTier(iter).metadata();

    // Metadata lookup functions:

    const CallSite* lookupCallSite(void* returnAddress) const;
    const CodeRange* lookupFuncRange(void* pc) const;
    bool containsCodePC(const void* pc) const;
    bool lookupTrap(void* pc, Trap* trap, BytecodeOffset* bytecode) const;

    // To save memory, profilingLabels_ are generated lazily when profiling mode
    // is enabled.

    void ensureProfilingLabels(const Bytes* maybeBytecode, bool profilingEnabled) const;
    const char* profilingLabel(uint32_t funcIndex) const;

    // about:memory reporting:

    void addSizeOfMiscIfNotSeen(MallocSizeOf mallocSizeOf,
                                Metadata::SeenSet* seenMetadata,
                                Code::SeenSet* seenCode,
                                size_t* code,
                                size_t* data) const;

    // A Code object is serialized as the length and bytes of the machine code
    // after statically unlinking it; the Code is then later recreated from the
    // machine code and other parts.

    size_t serializedSize() const;
    uint8_t* serialize(uint8_t* cursor, const LinkData& linkData) const;
    static const uint8_t* deserialize(const uint8_t* cursor,
                                      const ShareableBytes& bytecode,
                                      const LinkData& linkData,
                                      Metadata& metadata,
                                      SharedCode* code);

} // namespace wasm
} // namespace js

#endif // wasm_code_h