/* * Copyright (C) 2014 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef ART_RUNTIME_STACK_MAP_H_ #define ART_RUNTIME_STACK_MAP_H_ #include "base/bit_vector.h" #include "base/bit_utils.h" #include "memory_region.h" #include "leb128.h" namespace art { class VariableIndentationOutputStream; // Size of a frame slot, in bytes. This constant is a signed value, // to please the compiler in arithmetic operations involving int32_t // (signed) values. static constexpr ssize_t kFrameSlotSize = 4; // Size of Dex virtual registers. static constexpr size_t kVRegSize = 4; class CodeInfo; class StackMapEncoding; struct CodeInfoEncoding; /** * Classes in the following file are wrapper on stack map information backed * by a MemoryRegion. As such they read and write to the region, they don't have * their own fields. */ // Dex register location container used by DexRegisterMap and StackMapStream. class DexRegisterLocation { public: /* * The location kind used to populate the Dex register information in a * StackMapStream can either be: * - kStack: vreg stored on the stack, value holds the stack offset; * - kInRegister: vreg stored in low 32 bits of a core physical register, * value holds the register number; * - kInRegisterHigh: vreg stored in high 32 bits of a core physical register, * value holds the register number; * - kInFpuRegister: vreg stored in low 32 bits of an FPU register, * value holds the register number; * - kInFpuRegisterHigh: vreg stored in high 32 bits of an FPU register, * value holds the register number; * - kConstant: value holds the constant; * * In addition, DexRegisterMap also uses these values: * - kInStackLargeOffset: value holds a "large" stack offset (greater than * or equal to 128 bytes); * - kConstantLargeValue: value holds a "large" constant (lower than 0, or * or greater than or equal to 32); * - kNone: the register has no location, meaning it has not been set. */ enum class Kind : uint8_t { // Short location kinds, for entries fitting on one byte (3 bits // for the kind, 5 bits for the value) in a DexRegisterMap. kInStack = 0, // 0b000 kInRegister = 1, // 0b001 kInRegisterHigh = 2, // 0b010 kInFpuRegister = 3, // 0b011 kInFpuRegisterHigh = 4, // 0b100 kConstant = 5, // 0b101 // Large location kinds, requiring a 5-byte encoding (1 byte for the // kind, 4 bytes for the value). // Stack location at a large offset, meaning that the offset value // divided by the stack frame slot size (4 bytes) cannot fit on a // 5-bit unsigned integer (i.e., this offset value is greater than // or equal to 2^5 * 4 = 128 bytes). kInStackLargeOffset = 6, // 0b110 // Large constant, that cannot fit on a 5-bit signed integer (i.e., // lower than 0, or greater than or equal to 2^5 = 32). kConstantLargeValue = 7, // 0b111 // Entries with no location are not stored and do not need own marker. kNone = static_cast<uint8_t>(-1), kLastLocationKind = kConstantLargeValue }; static_assert( sizeof(Kind) == 1u, "art::DexRegisterLocation::Kind has a size different from one byte."); static bool IsShortLocationKind(Kind kind) { switch (kind) { case Kind::kInStack: case Kind::kInRegister: case Kind::kInRegisterHigh: case Kind::kInFpuRegister: case Kind::kInFpuRegisterHigh: case Kind::kConstant: return true; case Kind::kInStackLargeOffset: case Kind::kConstantLargeValue: return false; case Kind::kNone: LOG(FATAL) << "Unexpected location kind"; } UNREACHABLE(); } // Convert `kind` to a "surface" kind, i.e. one that doesn't include // any value with a "large" qualifier. // TODO: Introduce another enum type for the surface kind? static Kind ConvertToSurfaceKind(Kind kind) { switch (kind) { case Kind::kInStack: case Kind::kInRegister: case Kind::kInRegisterHigh: case Kind::kInFpuRegister: case Kind::kInFpuRegisterHigh: case Kind::kConstant: return kind; case Kind::kInStackLargeOffset: return Kind::kInStack; case Kind::kConstantLargeValue: return Kind::kConstant; case Kind::kNone: return kind; } UNREACHABLE(); } // Required by art::StackMapStream::LocationCatalogEntriesIndices. DexRegisterLocation() : kind_(Kind::kNone), value_(0) {} DexRegisterLocation(Kind kind, int32_t value) : kind_(kind), value_(value) {} static DexRegisterLocation None() { return DexRegisterLocation(Kind::kNone, 0); } // Get the "surface" kind of the location, i.e., the one that doesn't // include any value with a "large" qualifier. Kind GetKind() const { return ConvertToSurfaceKind(kind_); } // Get the value of the location. int32_t GetValue() const { return value_; } // Get the actual kind of the location. Kind GetInternalKind() const { return kind_; } bool operator==(DexRegisterLocation other) const { return kind_ == other.kind_ && value_ == other.value_; } bool operator!=(DexRegisterLocation other) const { return !(*this == other); } private: Kind kind_; int32_t value_; friend class DexRegisterLocationHashFn; }; std::ostream& operator<<(std::ostream& stream, const DexRegisterLocation::Kind& kind); /** * Store information on unique Dex register locations used in a method. * The information is of the form: * * [DexRegisterLocation+]. * * DexRegisterLocations are either 1- or 5-byte wide (see art::DexRegisterLocation::Kind). */ class DexRegisterLocationCatalog { public: explicit DexRegisterLocationCatalog(MemoryRegion region) : region_(region) {} // Short (compressed) location, fitting on one byte. typedef uint8_t ShortLocation; void SetRegisterInfo(size_t offset, const DexRegisterLocation& dex_register_location) { DexRegisterLocation::Kind kind = ComputeCompressedKind(dex_register_location); int32_t value = dex_register_location.GetValue(); if (DexRegisterLocation::IsShortLocationKind(kind)) { // Short location. Compress the kind and the value as a single byte. if (kind == DexRegisterLocation::Kind::kInStack) { // Instead of storing stack offsets expressed in bytes for // short stack locations, store slot offsets. A stack offset // is a multiple of 4 (kFrameSlotSize). This means that by // dividing it by 4, we can fit values from the [0, 128) // interval in a short stack location, and not just values // from the [0, 32) interval. DCHECK_EQ(value % kFrameSlotSize, 0); value /= kFrameSlotSize; } DCHECK(IsShortValue(value)) << value; region_.StoreUnaligned<ShortLocation>(offset, MakeShortLocation(kind, value)); } else { // Large location. Write the location on one byte and the value // on 4 bytes. DCHECK(!IsShortValue(value)) << value; if (kind == DexRegisterLocation::Kind::kInStackLargeOffset) { // Also divide large stack offsets by 4 for the sake of consistency. DCHECK_EQ(value % kFrameSlotSize, 0); value /= kFrameSlotSize; } // Data can be unaligned as the written Dex register locations can // either be 1-byte or 5-byte wide. Use // art::MemoryRegion::StoreUnaligned instead of // art::MemoryRegion::Store to prevent unligned word accesses on ARM. region_.StoreUnaligned<DexRegisterLocation::Kind>(offset, kind); region_.StoreUnaligned<int32_t>(offset + sizeof(DexRegisterLocation::Kind), value); } } // Find the offset of the location catalog entry number `location_catalog_entry_index`. size_t FindLocationOffset(size_t location_catalog_entry_index) const { size_t offset = kFixedSize; // Skip the first `location_catalog_entry_index - 1` entries. for (uint16_t i = 0; i < location_catalog_entry_index; ++i) { // Read the first next byte and inspect its first 3 bits to decide // whether it is a short or a large location. DexRegisterLocation::Kind kind = ExtractKindAtOffset(offset); if (DexRegisterLocation::IsShortLocationKind(kind)) { // Short location. Skip the current byte. offset += SingleShortEntrySize(); } else { // Large location. Skip the 5 next bytes. offset += SingleLargeEntrySize(); } } return offset; } // Get the internal kind of entry at `location_catalog_entry_index`. DexRegisterLocation::Kind GetLocationInternalKind(size_t location_catalog_entry_index) const { if (location_catalog_entry_index == kNoLocationEntryIndex) { return DexRegisterLocation::Kind::kNone; } return ExtractKindAtOffset(FindLocationOffset(location_catalog_entry_index)); } // Get the (surface) kind and value of entry at `location_catalog_entry_index`. DexRegisterLocation GetDexRegisterLocation(size_t location_catalog_entry_index) const { if (location_catalog_entry_index == kNoLocationEntryIndex) { return DexRegisterLocation::None(); } size_t offset = FindLocationOffset(location_catalog_entry_index); // Read the first byte and inspect its first 3 bits to get the location. ShortLocation first_byte = region_.LoadUnaligned<ShortLocation>(offset); DexRegisterLocation::Kind kind = ExtractKindFromShortLocation(first_byte); if (DexRegisterLocation::IsShortLocationKind(kind)) { // Short location. Extract the value from the remaining 5 bits. int32_t value = ExtractValueFromShortLocation(first_byte); if (kind == DexRegisterLocation::Kind::kInStack) { // Convert the stack slot (short) offset to a byte offset value. value *= kFrameSlotSize; } return DexRegisterLocation(kind, value); } else { // Large location. Read the four next bytes to get the value. int32_t value = region_.LoadUnaligned<int32_t>(offset + sizeof(DexRegisterLocation::Kind)); if (kind == DexRegisterLocation::Kind::kInStackLargeOffset) { // Convert the stack slot (large) offset to a byte offset value. value *= kFrameSlotSize; } return DexRegisterLocation(kind, value); } } // Compute the compressed kind of `location`. static DexRegisterLocation::Kind ComputeCompressedKind(const DexRegisterLocation& location) { DexRegisterLocation::Kind kind = location.GetInternalKind(); switch (kind) { case DexRegisterLocation::Kind::kInStack: return IsShortStackOffsetValue(location.GetValue()) ? DexRegisterLocation::Kind::kInStack : DexRegisterLocation::Kind::kInStackLargeOffset; case DexRegisterLocation::Kind::kInRegister: case DexRegisterLocation::Kind::kInRegisterHigh: DCHECK_GE(location.GetValue(), 0); DCHECK_LT(location.GetValue(), 1 << kValueBits); return kind; case DexRegisterLocation::Kind::kInFpuRegister: case DexRegisterLocation::Kind::kInFpuRegisterHigh: DCHECK_GE(location.GetValue(), 0); DCHECK_LT(location.GetValue(), 1 << kValueBits); return kind; case DexRegisterLocation::Kind::kConstant: return IsShortConstantValue(location.GetValue()) ? DexRegisterLocation::Kind::kConstant : DexRegisterLocation::Kind::kConstantLargeValue; case DexRegisterLocation::Kind::kConstantLargeValue: case DexRegisterLocation::Kind::kInStackLargeOffset: case DexRegisterLocation::Kind::kNone: LOG(FATAL) << "Unexpected location kind " << kind; } UNREACHABLE(); } // Can `location` be turned into a short location? static bool CanBeEncodedAsShortLocation(const DexRegisterLocation& location) { DexRegisterLocation::Kind kind = location.GetInternalKind(); switch (kind) { case DexRegisterLocation::Kind::kInStack: return IsShortStackOffsetValue(location.GetValue()); case DexRegisterLocation::Kind::kInRegister: case DexRegisterLocation::Kind::kInRegisterHigh: case DexRegisterLocation::Kind::kInFpuRegister: case DexRegisterLocation::Kind::kInFpuRegisterHigh: return true; case DexRegisterLocation::Kind::kConstant: return IsShortConstantValue(location.GetValue()); case DexRegisterLocation::Kind::kConstantLargeValue: case DexRegisterLocation::Kind::kInStackLargeOffset: case DexRegisterLocation::Kind::kNone: LOG(FATAL) << "Unexpected location kind " << kind; } UNREACHABLE(); } static size_t EntrySize(const DexRegisterLocation& location) { return CanBeEncodedAsShortLocation(location) ? SingleShortEntrySize() : SingleLargeEntrySize(); } static size_t SingleShortEntrySize() { return sizeof(ShortLocation); } static size_t SingleLargeEntrySize() { return sizeof(DexRegisterLocation::Kind) + sizeof(int32_t); } size_t Size() const { return region_.size(); } void Dump(VariableIndentationOutputStream* vios, const CodeInfo& code_info); // Special (invalid) Dex register location catalog entry index meaning // that there is no location for a given Dex register (i.e., it is // mapped to a DexRegisterLocation::Kind::kNone location). static constexpr size_t kNoLocationEntryIndex = -1; private: static constexpr int kFixedSize = 0; // Width of the kind "field" in a short location, in bits. static constexpr size_t kKindBits = 3; // Width of the value "field" in a short location, in bits. static constexpr size_t kValueBits = 5; static constexpr uint8_t kKindMask = (1 << kKindBits) - 1; static constexpr int32_t kValueMask = (1 << kValueBits) - 1; static constexpr size_t kKindOffset = 0; static constexpr size_t kValueOffset = kKindBits; static bool IsShortStackOffsetValue(int32_t value) { DCHECK_EQ(value % kFrameSlotSize, 0); return IsShortValue(value / kFrameSlotSize); } static bool IsShortConstantValue(int32_t value) { return IsShortValue(value); } static bool IsShortValue(int32_t value) { return IsUint<kValueBits>(value); } static ShortLocation MakeShortLocation(DexRegisterLocation::Kind kind, int32_t value) { uint8_t kind_integer_value = static_cast<uint8_t>(kind); DCHECK(IsUint<kKindBits>(kind_integer_value)) << kind_integer_value; DCHECK(IsShortValue(value)) << value; return (kind_integer_value & kKindMask) << kKindOffset | (value & kValueMask) << kValueOffset; } static DexRegisterLocation::Kind ExtractKindFromShortLocation(ShortLocation location) { uint8_t kind = (location >> kKindOffset) & kKindMask; DCHECK_LE(kind, static_cast<uint8_t>(DexRegisterLocation::Kind::kLastLocationKind)); // We do not encode kNone locations in the stack map. DCHECK_NE(kind, static_cast<uint8_t>(DexRegisterLocation::Kind::kNone)); return static_cast<DexRegisterLocation::Kind>(kind); } static int32_t ExtractValueFromShortLocation(ShortLocation location) { return (location >> kValueOffset) & kValueMask; } // Extract a location kind from the byte at position `offset`. DexRegisterLocation::Kind ExtractKindAtOffset(size_t offset) const { ShortLocation first_byte = region_.LoadUnaligned<ShortLocation>(offset); return ExtractKindFromShortLocation(first_byte); } MemoryRegion region_; friend class CodeInfo; friend class StackMapStream; }; /* Information on Dex register locations for a specific PC, mapping a * stack map's Dex register to a location entry in a DexRegisterLocationCatalog. * The information is of the form: * * [live_bit_mask, entries*] * * where entries are concatenated unsigned integer values encoded on a number * of bits (fixed per DexRegisterMap instances of a CodeInfo object) depending * on the number of entries in the Dex register location catalog * (see DexRegisterMap::SingleEntrySizeInBits). The map is 1-byte aligned. */ class DexRegisterMap { public: explicit DexRegisterMap(MemoryRegion region) : region_(region) {} DexRegisterMap() {} bool IsValid() const { return region_.pointer() != nullptr; } // Get the surface kind of Dex register `dex_register_number`. DexRegisterLocation::Kind GetLocationKind(uint16_t dex_register_number, uint16_t number_of_dex_registers, const CodeInfo& code_info, const CodeInfoEncoding& enc) const { return DexRegisterLocation::ConvertToSurfaceKind( GetLocationInternalKind(dex_register_number, number_of_dex_registers, code_info, enc)); } // Get the internal kind of Dex register `dex_register_number`. DexRegisterLocation::Kind GetLocationInternalKind(uint16_t dex_register_number, uint16_t number_of_dex_registers, const CodeInfo& code_info, const CodeInfoEncoding& enc) const; // Get the Dex register location `dex_register_number`. DexRegisterLocation GetDexRegisterLocation(uint16_t dex_register_number, uint16_t number_of_dex_registers, const CodeInfo& code_info, const CodeInfoEncoding& enc) const; int32_t GetStackOffsetInBytes(uint16_t dex_register_number, uint16_t number_of_dex_registers, const CodeInfo& code_info, const CodeInfoEncoding& enc) const { DexRegisterLocation location = GetDexRegisterLocation(dex_register_number, number_of_dex_registers, code_info, enc); DCHECK(location.GetKind() == DexRegisterLocation::Kind::kInStack); // GetDexRegisterLocation returns the offset in bytes. return location.GetValue(); } int32_t GetConstant(uint16_t dex_register_number, uint16_t number_of_dex_registers, const CodeInfo& code_info, const CodeInfoEncoding& enc) const { DexRegisterLocation location = GetDexRegisterLocation(dex_register_number, number_of_dex_registers, code_info, enc); DCHECK_EQ(location.GetKind(), DexRegisterLocation::Kind::kConstant); return location.GetValue(); } int32_t GetMachineRegister(uint16_t dex_register_number, uint16_t number_of_dex_registers, const CodeInfo& code_info, const CodeInfoEncoding& enc) const { DexRegisterLocation location = GetDexRegisterLocation(dex_register_number, number_of_dex_registers, code_info, enc); DCHECK(location.GetInternalKind() == DexRegisterLocation::Kind::kInRegister || location.GetInternalKind() == DexRegisterLocation::Kind::kInRegisterHigh || location.GetInternalKind() == DexRegisterLocation::Kind::kInFpuRegister || location.GetInternalKind() == DexRegisterLocation::Kind::kInFpuRegisterHigh) << location.GetInternalKind(); return location.GetValue(); } // Get the index of the entry in the Dex register location catalog // corresponding to `dex_register_number`. size_t GetLocationCatalogEntryIndex(uint16_t dex_register_number, uint16_t number_of_dex_registers, size_t number_of_location_catalog_entries) const { if (!IsDexRegisterLive(dex_register_number)) { return DexRegisterLocationCatalog::kNoLocationEntryIndex; } if (number_of_location_catalog_entries == 1) { // We do not allocate space for location maps in the case of a // single-entry location catalog, as it is useless. The only valid // entry index is 0; return 0; } // The bit offset of the beginning of the map locations. size_t map_locations_offset_in_bits = GetLocationMappingDataOffset(number_of_dex_registers) * kBitsPerByte; size_t index_in_dex_register_map = GetIndexInDexRegisterMap(dex_register_number); DCHECK_LT(index_in_dex_register_map, GetNumberOfLiveDexRegisters(number_of_dex_registers)); // The bit size of an entry. size_t map_entry_size_in_bits = SingleEntrySizeInBits(number_of_location_catalog_entries); // The bit offset where `index_in_dex_register_map` is located. size_t entry_offset_in_bits = map_locations_offset_in_bits + index_in_dex_register_map * map_entry_size_in_bits; size_t location_catalog_entry_index = region_.LoadBits(entry_offset_in_bits, map_entry_size_in_bits); DCHECK_LT(location_catalog_entry_index, number_of_location_catalog_entries); return location_catalog_entry_index; } // Map entry at `index_in_dex_register_map` to `location_catalog_entry_index`. void SetLocationCatalogEntryIndex(size_t index_in_dex_register_map, size_t location_catalog_entry_index, uint16_t number_of_dex_registers, size_t number_of_location_catalog_entries) { DCHECK_LT(index_in_dex_register_map, GetNumberOfLiveDexRegisters(number_of_dex_registers)); DCHECK_LT(location_catalog_entry_index, number_of_location_catalog_entries); if (number_of_location_catalog_entries == 1) { // We do not allocate space for location maps in the case of a // single-entry location catalog, as it is useless. return; } // The bit offset of the beginning of the map locations. size_t map_locations_offset_in_bits = GetLocationMappingDataOffset(number_of_dex_registers) * kBitsPerByte; // The bit size of an entry. size_t map_entry_size_in_bits = SingleEntrySizeInBits(number_of_location_catalog_entries); // The bit offset where `index_in_dex_register_map` is located. size_t entry_offset_in_bits = map_locations_offset_in_bits + index_in_dex_register_map * map_entry_size_in_bits; region_.StoreBits(entry_offset_in_bits, location_catalog_entry_index, map_entry_size_in_bits); } void SetLiveBitMask(uint16_t number_of_dex_registers, const BitVector& live_dex_registers_mask) { size_t live_bit_mask_offset_in_bits = GetLiveBitMaskOffset() * kBitsPerByte; for (uint16_t i = 0; i < number_of_dex_registers; ++i) { region_.StoreBit(live_bit_mask_offset_in_bits + i, live_dex_registers_mask.IsBitSet(i)); } } bool IsDexRegisterLive(uint16_t dex_register_number) const { size_t live_bit_mask_offset_in_bits = GetLiveBitMaskOffset() * kBitsPerByte; return region_.LoadBit(live_bit_mask_offset_in_bits + dex_register_number); } size_t GetNumberOfLiveDexRegisters(uint16_t number_of_dex_registers) const { size_t number_of_live_dex_registers = 0; for (size_t i = 0; i < number_of_dex_registers; ++i) { if (IsDexRegisterLive(i)) { ++number_of_live_dex_registers; } } return number_of_live_dex_registers; } static size_t GetLiveBitMaskOffset() { return kFixedSize; } // Compute the size of the live register bit mask (in bytes), for a // method having `number_of_dex_registers` Dex registers. static size_t GetLiveBitMaskSize(uint16_t number_of_dex_registers) { return RoundUp(number_of_dex_registers, kBitsPerByte) / kBitsPerByte; } static size_t GetLocationMappingDataOffset(uint16_t number_of_dex_registers) { return GetLiveBitMaskOffset() + GetLiveBitMaskSize(number_of_dex_registers); } size_t GetLocationMappingDataSize(uint16_t number_of_dex_registers, size_t number_of_location_catalog_entries) const { size_t location_mapping_data_size_in_bits = GetNumberOfLiveDexRegisters(number_of_dex_registers) * SingleEntrySizeInBits(number_of_location_catalog_entries); return RoundUp(location_mapping_data_size_in_bits, kBitsPerByte) / kBitsPerByte; } // Return the size of a map entry in bits. Note that if // `number_of_location_catalog_entries` equals 1, this function returns 0, // which is fine, as there is no need to allocate a map for a // single-entry location catalog; the only valid location catalog entry index // for a live register in this case is 0 and there is no need to // store it. static size_t SingleEntrySizeInBits(size_t number_of_location_catalog_entries) { // Handle the case of 0, as we cannot pass 0 to art::WhichPowerOf2. return number_of_location_catalog_entries == 0 ? 0u : WhichPowerOf2(RoundUpToPowerOfTwo(number_of_location_catalog_entries)); } // Return the size of the DexRegisterMap object, in bytes. size_t Size() const { return region_.size(); } void Dump(VariableIndentationOutputStream* vios, const CodeInfo& code_info, uint16_t number_of_dex_registers) const; private: // Return the index in the Dex register map corresponding to the Dex // register number `dex_register_number`. size_t GetIndexInDexRegisterMap(uint16_t dex_register_number) const { if (!IsDexRegisterLive(dex_register_number)) { return kInvalidIndexInDexRegisterMap; } return GetNumberOfLiveDexRegisters(dex_register_number); } // Special (invalid) Dex register map entry index meaning that there // is no index in the map for a given Dex register (i.e., it must // have been mapped to a DexRegisterLocation::Kind::kNone location). static constexpr size_t kInvalidIndexInDexRegisterMap = -1; static constexpr int kFixedSize = 0; MemoryRegion region_; friend class CodeInfo; friend class StackMapStream; }; // Represents bit range of bit-packed integer field. // We reuse the idea from ULEB128p1 to support encoding of -1 (aka 0xFFFFFFFF). // If min_value is set to -1, we implicitly subtract one from any loaded value, // and add one to any stored value. This is generalized to any negative values. // In other words, min_value acts as a base and the stored value is added to it. struct FieldEncoding { FieldEncoding(size_t start_offset, size_t end_offset, int32_t min_value = 0) : start_offset_(start_offset), end_offset_(end_offset), min_value_(min_value) { DCHECK_LE(start_offset_, end_offset_); DCHECK_LE(BitSize(), 32u); } ALWAYS_INLINE size_t BitSize() const { return end_offset_ - start_offset_; } ALWAYS_INLINE int32_t Load(const MemoryRegion& region) const { DCHECK_LE(end_offset_, region.size_in_bits()); const size_t bit_count = BitSize(); if (bit_count == 0) { // Do not touch any memory if the range is empty. return min_value_; } uint8_t* address = region.start() + start_offset_ / kBitsPerByte; const uint32_t shift = start_offset_ & (kBitsPerByte - 1); // Load the value (reading only the strictly needed bytes). const uint32_t load_bit_count = shift + bit_count; uint32_t value = *address++ >> shift; if (load_bit_count > 8) { value |= static_cast<uint32_t>(*address++) << (8 - shift); if (load_bit_count > 16) { value |= static_cast<uint32_t>(*address++) << (16 - shift); if (load_bit_count > 24) { value |= static_cast<uint32_t>(*address++) << (24 - shift); if (load_bit_count > 32) { value |= static_cast<uint32_t>(*address++) << (32 - shift); } } } } // Clear unwanted most significant bits. uint32_t clear_bit_count = 32 - bit_count; value = (value << clear_bit_count) >> clear_bit_count; return value + min_value_; } ALWAYS_INLINE void Store(MemoryRegion region, int32_t value) const { region.StoreBits(start_offset_, value - min_value_, BitSize()); DCHECK_EQ(Load(region), value); } private: size_t start_offset_; size_t end_offset_; int32_t min_value_; }; class StackMapEncoding { public: StackMapEncoding() {} // Set stack map bit layout based on given sizes. // Returns the size of stack map in bytes. size_t SetFromSizes(size_t native_pc_max, size_t dex_pc_max, size_t dex_register_map_size, size_t inline_info_size, size_t register_mask_max, size_t stack_mask_bit_size) { size_t bit_offset = 0; DCHECK_EQ(kNativePcBitOffset, bit_offset); bit_offset += MinimumBitsToStore(native_pc_max); dex_pc_bit_offset_ = dchecked_integral_cast<uint8_t>(bit_offset); bit_offset += MinimumBitsToStore(1 /* kNoDexPc */ + dex_pc_max); // We also need +1 for kNoDexRegisterMap, but since the size is strictly // greater than any offset we might try to encode, we already implicitly have it. dex_register_map_bit_offset_ = dchecked_integral_cast<uint8_t>(bit_offset); bit_offset += MinimumBitsToStore(dex_register_map_size); // We also need +1 for kNoInlineInfo, but since the inline_info_size is strictly // greater than the offset we might try to encode, we already implicitly have it. // If inline_info_size is zero, we can encode only kNoInlineInfo (in zero bits). inline_info_bit_offset_ = dchecked_integral_cast<uint8_t>(bit_offset); if (inline_info_size != 0) { bit_offset += MinimumBitsToStore(dex_register_map_size + inline_info_size); } register_mask_bit_offset_ = dchecked_integral_cast<uint8_t>(bit_offset); bit_offset += MinimumBitsToStore(register_mask_max); stack_mask_bit_offset_ = dchecked_integral_cast<uint8_t>(bit_offset); bit_offset += stack_mask_bit_size; return RoundUp(bit_offset, kBitsPerByte) / kBitsPerByte; } ALWAYS_INLINE FieldEncoding GetNativePcEncoding() const { return FieldEncoding(kNativePcBitOffset, dex_pc_bit_offset_); } ALWAYS_INLINE FieldEncoding GetDexPcEncoding() const { return FieldEncoding(dex_pc_bit_offset_, dex_register_map_bit_offset_, -1 /* min_value */); } ALWAYS_INLINE FieldEncoding GetDexRegisterMapEncoding() const { return FieldEncoding(dex_register_map_bit_offset_, inline_info_bit_offset_, -1 /* min_value */); } ALWAYS_INLINE FieldEncoding GetInlineInfoEncoding() const { return FieldEncoding(inline_info_bit_offset_, register_mask_bit_offset_, -1 /* min_value */); } ALWAYS_INLINE FieldEncoding GetRegisterMaskEncoding() const { return FieldEncoding(register_mask_bit_offset_, stack_mask_bit_offset_); } ALWAYS_INLINE size_t GetStackMaskBitOffset() const { // The end offset is not encoded. It is implicitly the end of stack map entry. return stack_mask_bit_offset_; } void Dump(VariableIndentationOutputStream* vios) const; private: static constexpr size_t kNativePcBitOffset = 0; uint8_t dex_pc_bit_offset_; uint8_t dex_register_map_bit_offset_; uint8_t inline_info_bit_offset_; uint8_t register_mask_bit_offset_; uint8_t stack_mask_bit_offset_; }; /** * A Stack Map holds compilation information for a specific PC necessary for: * - Mapping it to a dex PC, * - Knowing which stack entries are objects, * - Knowing which registers hold objects, * - Knowing the inlining information, * - Knowing the values of dex registers. * * The information is of the form: * * [native_pc_offset, dex_pc, dex_register_map_offset, inlining_info_offset, register_mask, * stack_mask]. */ class StackMap { public: StackMap() {} explicit StackMap(MemoryRegion region) : region_(region) {} ALWAYS_INLINE bool IsValid() const { return region_.pointer() != nullptr; } ALWAYS_INLINE uint32_t GetDexPc(const StackMapEncoding& encoding) const { return encoding.GetDexPcEncoding().Load(region_); } ALWAYS_INLINE void SetDexPc(const StackMapEncoding& encoding, uint32_t dex_pc) { encoding.GetDexPcEncoding().Store(region_, dex_pc); } ALWAYS_INLINE uint32_t GetNativePcOffset(const StackMapEncoding& encoding) const { return encoding.GetNativePcEncoding().Load(region_); } ALWAYS_INLINE void SetNativePcOffset(const StackMapEncoding& encoding, uint32_t native_pc_offset) { encoding.GetNativePcEncoding().Store(region_, native_pc_offset); } ALWAYS_INLINE uint32_t GetDexRegisterMapOffset(const StackMapEncoding& encoding) const { return encoding.GetDexRegisterMapEncoding().Load(region_); } ALWAYS_INLINE void SetDexRegisterMapOffset(const StackMapEncoding& encoding, uint32_t offset) { encoding.GetDexRegisterMapEncoding().Store(region_, offset); } ALWAYS_INLINE uint32_t GetInlineDescriptorOffset(const StackMapEncoding& encoding) const { return encoding.GetInlineInfoEncoding().Load(region_); } ALWAYS_INLINE void SetInlineDescriptorOffset(const StackMapEncoding& encoding, uint32_t offset) { encoding.GetInlineInfoEncoding().Store(region_, offset); } ALWAYS_INLINE uint32_t GetRegisterMask(const StackMapEncoding& encoding) const { return encoding.GetRegisterMaskEncoding().Load(region_); } ALWAYS_INLINE void SetRegisterMask(const StackMapEncoding& encoding, uint32_t mask) { encoding.GetRegisterMaskEncoding().Store(region_, mask); } ALWAYS_INLINE size_t GetNumberOfStackMaskBits(const StackMapEncoding& encoding) const { return region_.size_in_bits() - encoding.GetStackMaskBitOffset(); } ALWAYS_INLINE bool GetStackMaskBit(const StackMapEncoding& encoding, size_t index) const { return region_.LoadBit(encoding.GetStackMaskBitOffset() + index); } ALWAYS_INLINE void SetStackMaskBit(const StackMapEncoding& encoding, size_t index, bool value) { region_.StoreBit(encoding.GetStackMaskBitOffset() + index, value); } ALWAYS_INLINE bool HasDexRegisterMap(const StackMapEncoding& encoding) const { return GetDexRegisterMapOffset(encoding) != kNoDexRegisterMap; } ALWAYS_INLINE bool HasInlineInfo(const StackMapEncoding& encoding) const { return GetInlineDescriptorOffset(encoding) != kNoInlineInfo; } ALWAYS_INLINE bool Equals(const StackMap& other) const { return region_.pointer() == other.region_.pointer() && region_.size() == other.region_.size(); } void Dump(VariableIndentationOutputStream* vios, const CodeInfo& code_info, const CodeInfoEncoding& encoding, uint32_t code_offset, uint16_t number_of_dex_registers, const std::string& header_suffix = "") const; // Special (invalid) offset for the DexRegisterMapOffset field meaning // that there is no Dex register map for this stack map. static constexpr uint32_t kNoDexRegisterMap = -1; // Special (invalid) offset for the InlineDescriptorOffset field meaning // that there is no inline info for this stack map. static constexpr uint32_t kNoInlineInfo = -1; private: static constexpr int kFixedSize = 0; MemoryRegion region_; friend class StackMapStream; }; class InlineInfoEncoding { public: void SetFromSizes(size_t method_index_max, size_t dex_pc_max, size_t invoke_type_max, size_t dex_register_map_size) { total_bit_size_ = kMethodIndexBitOffset; total_bit_size_ += MinimumBitsToStore(method_index_max); dex_pc_bit_offset_ = dchecked_integral_cast<uint8_t>(total_bit_size_); total_bit_size_ += MinimumBitsToStore(1 /* kNoDexPc */ + dex_pc_max); invoke_type_bit_offset_ = dchecked_integral_cast<uint8_t>(total_bit_size_); total_bit_size_ += MinimumBitsToStore(invoke_type_max); // We also need +1 for kNoDexRegisterMap, but since the size is strictly // greater than any offset we might try to encode, we already implicitly have it. dex_register_map_bit_offset_ = dchecked_integral_cast<uint8_t>(total_bit_size_); total_bit_size_ += MinimumBitsToStore(dex_register_map_size); } ALWAYS_INLINE FieldEncoding GetMethodIndexEncoding() const { return FieldEncoding(kMethodIndexBitOffset, dex_pc_bit_offset_); } ALWAYS_INLINE FieldEncoding GetDexPcEncoding() const { return FieldEncoding(dex_pc_bit_offset_, invoke_type_bit_offset_, -1 /* min_value */); } ALWAYS_INLINE FieldEncoding GetInvokeTypeEncoding() const { return FieldEncoding(invoke_type_bit_offset_, dex_register_map_bit_offset_); } ALWAYS_INLINE FieldEncoding GetDexRegisterMapEncoding() const { return FieldEncoding(dex_register_map_bit_offset_, total_bit_size_, -1 /* min_value */); } ALWAYS_INLINE size_t GetEntrySize() const { return RoundUp(total_bit_size_, kBitsPerByte) / kBitsPerByte; } void Dump(VariableIndentationOutputStream* vios) const; private: static constexpr uint8_t kIsLastBitOffset = 0; static constexpr uint8_t kMethodIndexBitOffset = 1; uint8_t dex_pc_bit_offset_; uint8_t invoke_type_bit_offset_; uint8_t dex_register_map_bit_offset_; uint8_t total_bit_size_; }; /** * Inline information for a specific PC. The information is of the form: * * [is_last, method_index, dex_pc, invoke_type, dex_register_map_offset]+. */ class InlineInfo { public: explicit InlineInfo(MemoryRegion region) : region_(region) { } ALWAYS_INLINE uint32_t GetDepth(const InlineInfoEncoding& encoding) const { size_t depth = 0; while (!GetRegionAtDepth(encoding, depth++).LoadBit(0)) { } // Check is_last bit. return depth; } ALWAYS_INLINE void SetDepth(const InlineInfoEncoding& encoding, uint32_t depth) { DCHECK_GT(depth, 0u); for (size_t d = 0; d < depth; ++d) { GetRegionAtDepth(encoding, d).StoreBit(0, d == depth - 1); // Set is_last bit. } } ALWAYS_INLINE uint32_t GetMethodIndexAtDepth(const InlineInfoEncoding& encoding, uint32_t depth) const { return encoding.GetMethodIndexEncoding().Load(GetRegionAtDepth(encoding, depth)); } ALWAYS_INLINE void SetMethodIndexAtDepth(const InlineInfoEncoding& encoding, uint32_t depth, uint32_t index) { encoding.GetMethodIndexEncoding().Store(GetRegionAtDepth(encoding, depth), index); } ALWAYS_INLINE uint32_t GetDexPcAtDepth(const InlineInfoEncoding& encoding, uint32_t depth) const { return encoding.GetDexPcEncoding().Load(GetRegionAtDepth(encoding, depth)); } ALWAYS_INLINE void SetDexPcAtDepth(const InlineInfoEncoding& encoding, uint32_t depth, uint32_t dex_pc) { encoding.GetDexPcEncoding().Store(GetRegionAtDepth(encoding, depth), dex_pc); } ALWAYS_INLINE uint32_t GetInvokeTypeAtDepth(const InlineInfoEncoding& encoding, uint32_t depth) const { return encoding.GetInvokeTypeEncoding().Load(GetRegionAtDepth(encoding, depth)); } ALWAYS_INLINE void SetInvokeTypeAtDepth(const InlineInfoEncoding& encoding, uint32_t depth, uint32_t invoke_type) { encoding.GetInvokeTypeEncoding().Store(GetRegionAtDepth(encoding, depth), invoke_type); } ALWAYS_INLINE uint32_t GetDexRegisterMapOffsetAtDepth(const InlineInfoEncoding& encoding, uint32_t depth) const { return encoding.GetDexRegisterMapEncoding().Load(GetRegionAtDepth(encoding, depth)); } ALWAYS_INLINE void SetDexRegisterMapOffsetAtDepth(const InlineInfoEncoding& encoding, uint32_t depth, uint32_t offset) { encoding.GetDexRegisterMapEncoding().Store(GetRegionAtDepth(encoding, depth), offset); } ALWAYS_INLINE bool HasDexRegisterMapAtDepth(const InlineInfoEncoding& encoding, uint32_t depth) const { return GetDexRegisterMapOffsetAtDepth(encoding, depth) != StackMap::kNoDexRegisterMap; } void Dump(VariableIndentationOutputStream* vios, const CodeInfo& info, uint16_t* number_of_dex_registers) const; private: ALWAYS_INLINE MemoryRegion GetRegionAtDepth(const InlineInfoEncoding& encoding, uint32_t depth) const { size_t entry_size = encoding.GetEntrySize(); DCHECK_GT(entry_size, 0u); return region_.Subregion(depth * entry_size, entry_size); } MemoryRegion region_; }; // Most of the fields are encoded as ULEB128 to save space. struct CodeInfoEncoding { uint32_t non_header_size; uint32_t number_of_stack_maps; uint32_t stack_map_size_in_bytes; uint32_t number_of_location_catalog_entries; StackMapEncoding stack_map_encoding; InlineInfoEncoding inline_info_encoding; uint8_t header_size; CodeInfoEncoding() { } explicit CodeInfoEncoding(const void* data) { const uint8_t* ptr = reinterpret_cast<const uint8_t*>(data); non_header_size = DecodeUnsignedLeb128(&ptr); number_of_stack_maps = DecodeUnsignedLeb128(&ptr); stack_map_size_in_bytes = DecodeUnsignedLeb128(&ptr); number_of_location_catalog_entries = DecodeUnsignedLeb128(&ptr); static_assert(alignof(StackMapEncoding) == 1, "StackMapEncoding should not require alignment"); stack_map_encoding = *reinterpret_cast<const StackMapEncoding*>(ptr); ptr += sizeof(StackMapEncoding); if (stack_map_encoding.GetInlineInfoEncoding().BitSize() > 0) { static_assert(alignof(InlineInfoEncoding) == 1, "InlineInfoEncoding should not require alignment"); inline_info_encoding = *reinterpret_cast<const InlineInfoEncoding*>(ptr); ptr += sizeof(InlineInfoEncoding); } else { inline_info_encoding = InlineInfoEncoding{}; // NOLINT. } header_size = dchecked_integral_cast<uint8_t>(ptr - reinterpret_cast<const uint8_t*>(data)); } template<typename Vector> void Compress(Vector* dest) const { EncodeUnsignedLeb128(dest, non_header_size); EncodeUnsignedLeb128(dest, number_of_stack_maps); EncodeUnsignedLeb128(dest, stack_map_size_in_bytes); EncodeUnsignedLeb128(dest, number_of_location_catalog_entries); const uint8_t* stack_map_ptr = reinterpret_cast<const uint8_t*>(&stack_map_encoding); dest->insert(dest->end(), stack_map_ptr, stack_map_ptr + sizeof(StackMapEncoding)); if (stack_map_encoding.GetInlineInfoEncoding().BitSize() > 0) { const uint8_t* inline_info_ptr = reinterpret_cast<const uint8_t*>(&inline_info_encoding); dest->insert(dest->end(), inline_info_ptr, inline_info_ptr + sizeof(InlineInfoEncoding)); } } }; /** * Wrapper around all compiler information collected for a method. * The information is of the form: * * [CodeInfoEncoding, StackMap+, DexRegisterLocationCatalog+, DexRegisterMap+, InlineInfo*] * * where CodeInfoEncoding is of the form: * * [non_header_size, number_of_stack_maps, stack_map_size_in_bytes, * number_of_location_catalog_entries, StackMapEncoding] */ class CodeInfo { public: explicit CodeInfo(MemoryRegion region) : region_(region) { } explicit CodeInfo(const void* data) { CodeInfoEncoding encoding = CodeInfoEncoding(data); region_ = MemoryRegion(const_cast<void*>(data), encoding.header_size + encoding.non_header_size); } CodeInfoEncoding ExtractEncoding() const { return CodeInfoEncoding(region_.start()); } bool HasInlineInfo(const CodeInfoEncoding& encoding) const { return encoding.stack_map_encoding.GetInlineInfoEncoding().BitSize() > 0; } DexRegisterLocationCatalog GetDexRegisterLocationCatalog(const CodeInfoEncoding& encoding) const { return DexRegisterLocationCatalog(region_.Subregion( GetDexRegisterLocationCatalogOffset(encoding), GetDexRegisterLocationCatalogSize(encoding))); } StackMap GetStackMapAt(size_t i, const CodeInfoEncoding& encoding) const { size_t stack_map_size = encoding.stack_map_size_in_bytes; return StackMap(GetStackMaps(encoding).Subregion(i * stack_map_size, stack_map_size)); } uint32_t GetNumberOfLocationCatalogEntries(const CodeInfoEncoding& encoding) const { return encoding.number_of_location_catalog_entries; } uint32_t GetDexRegisterLocationCatalogSize(const CodeInfoEncoding& encoding) const { return ComputeDexRegisterLocationCatalogSize(GetDexRegisterLocationCatalogOffset(encoding), GetNumberOfLocationCatalogEntries(encoding)); } uint32_t GetNumberOfStackMaps(const CodeInfoEncoding& encoding) const { return encoding.number_of_stack_maps; } // Get the size of all the stack maps of this CodeInfo object, in bytes. size_t GetStackMapsSize(const CodeInfoEncoding& encoding) const { return encoding.stack_map_size_in_bytes * GetNumberOfStackMaps(encoding); } uint32_t GetDexRegisterLocationCatalogOffset(const CodeInfoEncoding& encoding) const { return GetStackMapsOffset(encoding) + GetStackMapsSize(encoding); } size_t GetDexRegisterMapsOffset(const CodeInfoEncoding& encoding) const { return GetDexRegisterLocationCatalogOffset(encoding) + GetDexRegisterLocationCatalogSize(encoding); } uint32_t GetStackMapsOffset(const CodeInfoEncoding& encoding) const { return encoding.header_size; } DexRegisterMap GetDexRegisterMapOf(StackMap stack_map, const CodeInfoEncoding& encoding, uint32_t number_of_dex_registers) const { if (!stack_map.HasDexRegisterMap(encoding.stack_map_encoding)) { return DexRegisterMap(); } else { uint32_t offset = GetDexRegisterMapsOffset(encoding) + stack_map.GetDexRegisterMapOffset(encoding.stack_map_encoding); size_t size = ComputeDexRegisterMapSizeOf(encoding, offset, number_of_dex_registers); return DexRegisterMap(region_.Subregion(offset, size)); } } // Return the `DexRegisterMap` pointed by `inline_info` at depth `depth`. DexRegisterMap GetDexRegisterMapAtDepth(uint8_t depth, InlineInfo inline_info, const CodeInfoEncoding& encoding, uint32_t number_of_dex_registers) const { if (!inline_info.HasDexRegisterMapAtDepth(encoding.inline_info_encoding, depth)) { return DexRegisterMap(); } else { uint32_t offset = GetDexRegisterMapsOffset(encoding) + inline_info.GetDexRegisterMapOffsetAtDepth(encoding.inline_info_encoding, depth); size_t size = ComputeDexRegisterMapSizeOf(encoding, offset, number_of_dex_registers); return DexRegisterMap(region_.Subregion(offset, size)); } } InlineInfo GetInlineInfoOf(StackMap stack_map, const CodeInfoEncoding& encoding) const { DCHECK(stack_map.HasInlineInfo(encoding.stack_map_encoding)); uint32_t offset = stack_map.GetInlineDescriptorOffset(encoding.stack_map_encoding) + GetDexRegisterMapsOffset(encoding); return InlineInfo(region_.Subregion(offset, region_.size() - offset)); } StackMap GetStackMapForDexPc(uint32_t dex_pc, const CodeInfoEncoding& encoding) const { for (size_t i = 0, e = GetNumberOfStackMaps(encoding); i < e; ++i) { StackMap stack_map = GetStackMapAt(i, encoding); if (stack_map.GetDexPc(encoding.stack_map_encoding) == dex_pc) { return stack_map; } } return StackMap(); } // Searches the stack map list backwards because catch stack maps are stored // at the end. StackMap GetCatchStackMapForDexPc(uint32_t dex_pc, const CodeInfoEncoding& encoding) const { for (size_t i = GetNumberOfStackMaps(encoding); i > 0; --i) { StackMap stack_map = GetStackMapAt(i - 1, encoding); if (stack_map.GetDexPc(encoding.stack_map_encoding) == dex_pc) { return stack_map; } } return StackMap(); } StackMap GetOsrStackMapForDexPc(uint32_t dex_pc, const CodeInfoEncoding& encoding) const { size_t e = GetNumberOfStackMaps(encoding); if (e == 0) { // There cannot be OSR stack map if there is no stack map. return StackMap(); } // Walk over all stack maps. If two consecutive stack maps are identical, then we // have found a stack map suitable for OSR. const StackMapEncoding& stack_map_encoding = encoding.stack_map_encoding; for (size_t i = 0; i < e - 1; ++i) { StackMap stack_map = GetStackMapAt(i, encoding); if (stack_map.GetDexPc(stack_map_encoding) == dex_pc) { StackMap other = GetStackMapAt(i + 1, encoding); if (other.GetDexPc(stack_map_encoding) == dex_pc && other.GetNativePcOffset(stack_map_encoding) == stack_map.GetNativePcOffset(stack_map_encoding)) { DCHECK_EQ(other.GetDexRegisterMapOffset(stack_map_encoding), stack_map.GetDexRegisterMapOffset(stack_map_encoding)); DCHECK(!stack_map.HasInlineInfo(stack_map_encoding)); if (i < e - 2) { // Make sure there are not three identical stack maps following each other. DCHECK_NE(stack_map.GetNativePcOffset(stack_map_encoding), GetStackMapAt(i + 2, encoding).GetNativePcOffset(stack_map_encoding)); } return stack_map; } } } return StackMap(); } StackMap GetStackMapForNativePcOffset(uint32_t native_pc_offset, const CodeInfoEncoding& encoding) const { // TODO: Safepoint stack maps are sorted by native_pc_offset but catch stack // maps are not. If we knew that the method does not have try/catch, // we could do binary search. for (size_t i = 0, e = GetNumberOfStackMaps(encoding); i < e; ++i) { StackMap stack_map = GetStackMapAt(i, encoding); if (stack_map.GetNativePcOffset(encoding.stack_map_encoding) == native_pc_offset) { return stack_map; } } return StackMap(); } // Dump this CodeInfo object on `os`. `code_offset` is the (absolute) // native PC of the compiled method and `number_of_dex_registers` the // number of Dex virtual registers used in this method. If // `dump_stack_maps` is true, also dump the stack maps and the // associated Dex register maps. void Dump(VariableIndentationOutputStream* vios, uint32_t code_offset, uint16_t number_of_dex_registers, bool dump_stack_maps) const; private: MemoryRegion GetStackMaps(const CodeInfoEncoding& encoding) const { return region_.size() == 0 ? MemoryRegion() : region_.Subregion(GetStackMapsOffset(encoding), GetStackMapsSize(encoding)); } // Compute the size of the Dex register map associated to the stack map at // `dex_register_map_offset_in_code_info`. size_t ComputeDexRegisterMapSizeOf(const CodeInfoEncoding& encoding, uint32_t dex_register_map_offset_in_code_info, uint16_t number_of_dex_registers) const { // Offset where the actual mapping data starts within art::DexRegisterMap. size_t location_mapping_data_offset_in_dex_register_map = DexRegisterMap::GetLocationMappingDataOffset(number_of_dex_registers); // Create a temporary art::DexRegisterMap to be able to call // art::DexRegisterMap::GetNumberOfLiveDexRegisters and DexRegisterMap dex_register_map_without_locations( MemoryRegion(region_.Subregion(dex_register_map_offset_in_code_info, location_mapping_data_offset_in_dex_register_map))); size_t number_of_live_dex_registers = dex_register_map_without_locations.GetNumberOfLiveDexRegisters(number_of_dex_registers); size_t location_mapping_data_size_in_bits = DexRegisterMap::SingleEntrySizeInBits(GetNumberOfLocationCatalogEntries(encoding)) * number_of_live_dex_registers; size_t location_mapping_data_size_in_bytes = RoundUp(location_mapping_data_size_in_bits, kBitsPerByte) / kBitsPerByte; size_t dex_register_map_size = location_mapping_data_offset_in_dex_register_map + location_mapping_data_size_in_bytes; return dex_register_map_size; } // Compute the size of a Dex register location catalog starting at offset `origin` // in `region_` and containing `number_of_dex_locations` entries. size_t ComputeDexRegisterLocationCatalogSize(uint32_t origin, uint32_t number_of_dex_locations) const { // TODO: Ideally, we would like to use art::DexRegisterLocationCatalog::Size or // art::DexRegisterLocationCatalog::FindLocationOffset, but the // DexRegisterLocationCatalog is not yet built. Try to factor common code. size_t offset = origin + DexRegisterLocationCatalog::kFixedSize; // Skip the first `number_of_dex_locations - 1` entries. for (uint16_t i = 0; i < number_of_dex_locations; ++i) { // Read the first next byte and inspect its first 3 bits to decide // whether it is a short or a large location. DexRegisterLocationCatalog::ShortLocation first_byte = region_.LoadUnaligned<DexRegisterLocationCatalog::ShortLocation>(offset); DexRegisterLocation::Kind kind = DexRegisterLocationCatalog::ExtractKindFromShortLocation(first_byte); if (DexRegisterLocation::IsShortLocationKind(kind)) { // Short location. Skip the current byte. offset += DexRegisterLocationCatalog::SingleShortEntrySize(); } else { // Large location. Skip the 5 next bytes. offset += DexRegisterLocationCatalog::SingleLargeEntrySize(); } } size_t size = offset - origin; return size; } MemoryRegion region_; friend class StackMapStream; }; #undef ELEMENT_BYTE_OFFSET_AFTER #undef ELEMENT_BIT_OFFSET_AFTER } // namespace art #endif // ART_RUNTIME_STACK_MAP_H_