%PDF- <> %âãÏÓ endobj 2 0 obj <> endobj 3 0 obj <>/ExtGState<>/ProcSet[/PDF/Text/ImageB/ImageC/ImageI] >>/Annots[ 28 0 R 29 0 R] /MediaBox[ 0 0 595.5 842.25] /Contents 4 0 R/Group<>/Tabs/S>> endobj ºaâÚÎΞ-ÌE1ÍØÄ÷{òò2ÿ ÛÖ^ÔÀá TÎ{¦?§®¥kuµùÕ5sLOšuY>endobj 2 0 obj<>endobj 2 0 obj<>endobj 2 0 obj<>endobj 2 0 obj<> endobj 2 0 obj<>endobj 2 0 obj<>es 3 0 R>> endobj 2 0 obj<> ox[ 0.000000 0.000000 609.600000 935.600000]/Fi endobj 3 0 obj<> endobj 7 1 obj<>/ProcSet[/PDF/Text/ImageB/ImageC/ImageI]>>/Subtype/Form>> stream
// Copyright 2018 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef V8_OBJECTS_JS_ARRAY_BUFFER_INL_H_ #define V8_OBJECTS_JS_ARRAY_BUFFER_INL_H_ #include "src/common/external-pointer.h" #include "src/objects/js-array-buffer.h" #include "src/common/external-pointer-inl.h" #include "src/heap/heap-write-barrier-inl.h" #include "src/objects/js-objects-inl.h" #include "src/objects/objects-inl.h" // Has to be the last include (doesn't have include guards): #include "src/objects/object-macros.h" namespace v8 { namespace internal { #include "torque-generated/src/objects/js-array-buffer-tq-inl.inc" TQ_OBJECT_CONSTRUCTORS_IMPL(JSArrayBuffer) TQ_OBJECT_CONSTRUCTORS_IMPL(JSArrayBufferView) TQ_OBJECT_CONSTRUCTORS_IMPL(JSTypedArray) TQ_OBJECT_CONSTRUCTORS_IMPL(JSDataView) ACCESSORS(JSTypedArray, base_pointer, Object, kBasePointerOffset) RELEASE_ACQUIRE_ACCESSORS(JSTypedArray, base_pointer, Object, kBasePointerOffset) void JSArrayBuffer::AllocateExternalPointerEntries(Isolate* isolate) { InitExternalPointerField(kBackingStoreOffset, isolate); } size_t JSArrayBuffer::byte_length() const { return ReadField<size_t>(kByteLengthOffset); } void JSArrayBuffer::set_byte_length(size_t value) { WriteField<size_t>(kByteLengthOffset, value); } DEF_GETTER(JSArrayBuffer, backing_store, void*) { Isolate* isolate = GetIsolateForHeapSandbox(*this); Address value = ReadExternalPointerField(kBackingStoreOffset, isolate, kArrayBufferBackingStoreTag); return reinterpret_cast<void*>(value); } void JSArrayBuffer::set_backing_store(Isolate* isolate, void* value) { WriteExternalPointerField(kBackingStoreOffset, isolate, reinterpret_cast<Address>(value), kArrayBufferBackingStoreTag); } uint32_t JSArrayBuffer::GetBackingStoreRefForDeserialization() const { return static_cast<uint32_t>( ReadField<ExternalPointer_t>(kBackingStoreOffset)); } void JSArrayBuffer::SetBackingStoreRefForSerialization(uint32_t ref) { WriteField<ExternalPointer_t>(kBackingStoreOffset, static_cast<ExternalPointer_t>(ref)); } ArrayBufferExtension* JSArrayBuffer::extension() const { #if V8_COMPRESS_POINTERS // With pointer compression the extension-field might not be // pointer-aligned. However on ARM64 this field needs to be aligned to // perform atomic operations on it. Therefore we split the pointer into two // 32-bit words that we update atomically. We don't have an ABA problem here // since there can never be an Attach() after Detach() (transitions only // from NULL --> some ptr --> NULL). // Synchronize with publishing release store of non-null extension uint32_t lo = base::AsAtomic32::Acquire_Load(extension_lo()); if (lo & kUninitializedTagMask) return nullptr; // Synchronize with release store of null extension uint32_t hi = base::AsAtomic32::Acquire_Load(extension_hi()); uint32_t verify_lo = base::AsAtomic32::Relaxed_Load(extension_lo()); if (lo != verify_lo) return nullptr; uintptr_t address = static_cast<uintptr_t>(lo); address |= static_cast<uintptr_t>(hi) << 32; return reinterpret_cast<ArrayBufferExtension*>(address); #else return base::AsAtomicPointer::Acquire_Load(extension_location()); #endif } void JSArrayBuffer::set_extension(ArrayBufferExtension* extension) { #if V8_COMPRESS_POINTERS if (extension != nullptr) { uintptr_t address = reinterpret_cast<uintptr_t>(extension); base::AsAtomic32::Relaxed_Store(extension_hi(), static_cast<uint32_t>(address >> 32)); base::AsAtomic32::Release_Store(extension_lo(), static_cast<uint32_t>(address)); } else { base::AsAtomic32::Relaxed_Store(extension_lo(), 0 | kUninitializedTagMask); base::AsAtomic32::Release_Store(extension_hi(), 0); } #else base::AsAtomicPointer::Release_Store(extension_location(), extension); #endif WriteBarrier::Marking(*this, extension); } ArrayBufferExtension** JSArrayBuffer::extension_location() const { Address location = field_address(kExtensionOffset); return reinterpret_cast<ArrayBufferExtension**>(location); } #if V8_COMPRESS_POINTERS uint32_t* JSArrayBuffer::extension_lo() const { Address location = field_address(kExtensionOffset); return reinterpret_cast<uint32_t*>(location); } uint32_t* JSArrayBuffer::extension_hi() const { Address location = field_address(kExtensionOffset) + sizeof(uint32_t); return reinterpret_cast<uint32_t*>(location); } #endif size_t JSArrayBuffer::allocation_length() const { if (backing_store() == nullptr) { return 0; } return byte_length(); } void* JSArrayBuffer::allocation_base() const { if (backing_store() == nullptr) { return nullptr; } return backing_store(); } void JSArrayBuffer::clear_padding() { if (FIELD_SIZE(kOptionalPaddingOffset) != 0) { DCHECK_EQ(4, FIELD_SIZE(kOptionalPaddingOffset)); memset(reinterpret_cast<void*>(address() + kOptionalPaddingOffset), 0, FIELD_SIZE(kOptionalPaddingOffset)); } } void JSArrayBuffer::set_bit_field(uint32_t bits) { RELAXED_WRITE_UINT32_FIELD(*this, kBitFieldOffset, bits); } uint32_t JSArrayBuffer::bit_field() const { return RELAXED_READ_UINT32_FIELD(*this, kBitFieldOffset); } // |bit_field| fields. BIT_FIELD_ACCESSORS(JSArrayBuffer, bit_field, is_external, JSArrayBuffer::IsExternalBit) BIT_FIELD_ACCESSORS(JSArrayBuffer, bit_field, is_detachable, JSArrayBuffer::IsDetachableBit) BIT_FIELD_ACCESSORS(JSArrayBuffer, bit_field, was_detached, JSArrayBuffer::WasDetachedBit) BIT_FIELD_ACCESSORS(JSArrayBuffer, bit_field, is_asmjs_memory, JSArrayBuffer::IsAsmJsMemoryBit) BIT_FIELD_ACCESSORS(JSArrayBuffer, bit_field, is_shared, JSArrayBuffer::IsSharedBit) BIT_FIELD_ACCESSORS(JSArrayBuffer, bit_field, is_resizable, JSArrayBuffer::IsResizableBit) size_t JSArrayBufferView::byte_offset() const { return ReadField<size_t>(kByteOffsetOffset); } void JSArrayBufferView::set_byte_offset(size_t value) { WriteField<size_t>(kByteOffsetOffset, value); } size_t JSArrayBufferView::byte_length() const { return ReadField<size_t>(kByteLengthOffset); } void JSArrayBufferView::set_byte_length(size_t value) { WriteField<size_t>(kByteLengthOffset, value); } bool JSArrayBufferView::WasDetached() const { return JSArrayBuffer::cast(buffer()).was_detached(); } BIT_FIELD_ACCESSORS(JSTypedArray, bit_field, is_length_tracking, JSTypedArray::IsLengthTrackingBit) BIT_FIELD_ACCESSORS(JSTypedArray, bit_field, is_backed_by_rab, JSTypedArray::IsBackedByRabBit) bool JSTypedArray::IsVariableLength() const { return is_length_tracking() || is_backed_by_rab(); } size_t JSTypedArray::GetLengthOrOutOfBounds(bool& out_of_bounds) const { DCHECK(!out_of_bounds); if (WasDetached()) return 0; if (is_length_tracking()) { if (is_backed_by_rab()) { if (byte_offset() >= buffer().byte_length()) { out_of_bounds = true; return 0; } return (buffer().byte_length() - byte_offset()) / element_size(); } if (byte_offset() >= buffer().GetBackingStore()->byte_length(std::memory_order_seq_cst)) { out_of_bounds = true; return 0; } return (buffer().GetBackingStore()->byte_length(std::memory_order_seq_cst) - byte_offset()) / element_size(); } size_t array_length = LengthUnchecked(); if (is_backed_by_rab()) { // The sum can't overflow, since we have managed to allocate the // JSTypedArray. if (byte_offset() + array_length * element_size() > buffer().byte_length()) { out_of_bounds = true; return 0; } } return array_length; } size_t JSTypedArray::GetLength() const { bool out_of_bounds = false; return GetLengthOrOutOfBounds(out_of_bounds); } void JSTypedArray::AllocateExternalPointerEntries(Isolate* isolate) { InitExternalPointerField(kExternalPointerOffset, isolate); } size_t JSTypedArray::length() const { DCHECK(!is_length_tracking()); DCHECK(!is_backed_by_rab()); return ReadField<size_t>(kLengthOffset); } size_t JSTypedArray::LengthUnchecked() const { return ReadField<size_t>(kLengthOffset); } void JSTypedArray::set_length(size_t value) { WriteField<size_t>(kLengthOffset, value); } DEF_GETTER(JSTypedArray, external_pointer, Address) { Isolate* isolate = GetIsolateForHeapSandbox(*this); return ReadExternalPointerField(kExternalPointerOffset, isolate, kTypedArrayExternalPointerTag); } DEF_GETTER(JSTypedArray, external_pointer_raw, ExternalPointer_t) { return ReadField<ExternalPointer_t>(kExternalPointerOffset); } void JSTypedArray::set_external_pointer(Isolate* isolate, Address value) { WriteExternalPointerField(kExternalPointerOffset, isolate, value, kTypedArrayExternalPointerTag); } Address JSTypedArray::ExternalPointerCompensationForOnHeapArray( PtrComprCageBase cage_base) { #ifdef V8_COMPRESS_POINTERS return cage_base.address(); #else return 0; #endif } uint32_t JSTypedArray::GetExternalBackingStoreRefForDeserialization() const { DCHECK(!is_on_heap()); return static_cast<uint32_t>( ReadField<ExternalPointer_t>(kExternalPointerOffset)); } void JSTypedArray::SetExternalBackingStoreRefForSerialization(uint32_t ref) { DCHECK(!is_on_heap()); WriteField<ExternalPointer_t>(kExternalPointerOffset, static_cast<ExternalPointer_t>(ref)); } void JSTypedArray::RemoveExternalPointerCompensationForSerialization( Isolate* isolate) { DCHECK(is_on_heap()); // TODO(v8:10391): once we have an external table, avoid the need for // compensation by replacing external_pointer and base_pointer fields // with one data_pointer field which can point to either external data // backing store or into on-heap backing store. Address offset = external_pointer() - ExternalPointerCompensationForOnHeapArray(isolate); #ifdef V8_HEAP_SANDBOX // Write decompensated offset directly to the external pointer field, thus // allowing the offset to be propagated through serialization-deserialization. WriteField<ExternalPointer_t>(kExternalPointerOffset, offset); #else set_external_pointer(isolate, offset); #endif } void* JSTypedArray::DataPtr() { // Zero-extend Tagged_t to Address according to current compression scheme // so that the addition with |external_pointer| (which already contains // compensated offset value) will decompress the tagged value. // See JSTypedArray::ExternalPointerCompensationForOnHeapArray() for details. STATIC_ASSERT(kOffHeapDataPtrEqualsExternalPointer); return reinterpret_cast<void*>(external_pointer() + static_cast<Tagged_t>(base_pointer().ptr())); } void JSTypedArray::SetOffHeapDataPtr(Isolate* isolate, void* base, Address offset) { Address address = reinterpret_cast<Address>(base) + offset; set_external_pointer(isolate, address); // This is the only spot in which the `base_pointer` field can be mutated // after object initialization. Note this can happen at most once, when // `JSTypedArray::GetBuffer` transitions from an on- to off-heap // representation. // To play well with Turbofan concurrency requirements, `base_pointer` is set // with a release store, after external_pointer has been set. set_base_pointer(Smi::zero(), kReleaseStore, SKIP_WRITE_BARRIER); DCHECK_EQ(address, reinterpret_cast<Address>(DataPtr())); } void JSTypedArray::SetOnHeapDataPtr(Isolate* isolate, HeapObject base, Address offset) { set_base_pointer(base); set_external_pointer( isolate, offset + ExternalPointerCompensationForOnHeapArray(isolate)); DCHECK_EQ(base.ptr() + offset, reinterpret_cast<Address>(DataPtr())); } bool JSTypedArray::is_on_heap() const { // Keep synced with `is_on_heap(AcquireLoadTag)`. DisallowGarbageCollection no_gc; return base_pointer() != Smi::zero(); } bool JSTypedArray::is_on_heap(AcquireLoadTag tag) const { // Keep synced with `is_on_heap()`. // Note: For Turbofan concurrency requirements, it's important that this // function reads only `base_pointer`. DisallowGarbageCollection no_gc; return base_pointer(tag) != Smi::zero(); } // static MaybeHandle<JSTypedArray> JSTypedArray::Validate(Isolate* isolate, Handle<Object> receiver, const char* method_name) { if (V8_UNLIKELY(!receiver->IsJSTypedArray())) { const MessageTemplate message = MessageTemplate::kNotTypedArray; THROW_NEW_ERROR(isolate, NewTypeError(message), JSTypedArray); } Handle<JSTypedArray> array = Handle<JSTypedArray>::cast(receiver); if (V8_UNLIKELY(array->WasDetached())) { const MessageTemplate message = MessageTemplate::kDetachedOperation; Handle<String> operation = isolate->factory()->NewStringFromAsciiChecked(method_name); THROW_NEW_ERROR(isolate, NewTypeError(message, operation), JSTypedArray); } if (V8_UNLIKELY(array->IsVariableLength())) { bool out_of_bounds = false; array->GetLengthOrOutOfBounds(out_of_bounds); if (out_of_bounds) { const MessageTemplate message = MessageTemplate::kDetachedOperation; Handle<String> operation = isolate->factory()->NewStringFromAsciiChecked(method_name); THROW_NEW_ERROR(isolate, NewTypeError(message, operation), JSTypedArray); } } // spec describes to return `buffer`, but it may disrupt current // implementations, and it's much useful to return array for now. return array; } DEF_GETTER(JSDataView, data_pointer, void*) { Isolate* isolate = GetIsolateForHeapSandbox(*this); return reinterpret_cast<void*>(ReadExternalPointerField( kDataPointerOffset, isolate, kDataViewDataPointerTag)); } void JSDataView::AllocateExternalPointerEntries(Isolate* isolate) { InitExternalPointerField(kDataPointerOffset, isolate); } void JSDataView::set_data_pointer(Isolate* isolate, void* value) { WriteExternalPointerField(kDataPointerOffset, isolate, reinterpret_cast<Address>(value), kDataViewDataPointerTag); } } // namespace internal } // namespace v8 #include "src/objects/object-macros-undef.h" #endif // V8_OBJECTS_JS_ARRAY_BUFFER_INL_H_