/* * Created by Brett on 08/02/23. * Licensed under GNU General Public License V3.0 * See LICENSE file for license detail */ #ifndef BLT_TESTS_MEMORY_H #define BLT_TESTS_MEMORY_H #include #include #include #include "queue.h" #include "utility.h" #include #include #include #include #include #include #include #include #include #if defined(__clang__) || defined(__llvm__) || defined(__GNUC__) || defined(__GNUG__) #include #define SWAP16(val) bswap_16(val) #define SWAP32(val) bswap_32(val) #define SWAP64(val) bswap_64(val) #if __cplusplus >= 202002L #include #define ENDIAN_LOOKUP(little_endian) (std::endian::native == std::endian::little && !little_endian) || \ (std::endian::native == std::endian::big && little_endian) #else #define ENDIAN_LOOKUP(little_endian) !little_endian #endif #elif defined(_MSC_VER) #include #define SWAP16(val) _byteswap_ushort(val) #define SWAP32(val) _byteswap_ulong(val) #define SWAP64(val) _byteswap_uint64(val) #define ENDIAN_LOOKUP(little_endian) !little_endian #endif namespace blt { namespace mem { // Used to grab the byte-data of any T element. Defaults to Big Endian, however can be configured to use little endian template inline static int toBytes(const T& in, BYTE_TYPE* out) { if constexpr (!(std::is_same_v || std::is_same_v)) static_assert("Must provide a signed/unsigned int8 type"); std::memcpy(out, (void*) &in, sizeof(T)); if constexpr (ENDIAN_LOOKUP(little_endian)) { // TODO: this but better. for (size_t i = 0; i < sizeof(T) / 2; i++) std::swap(out[i], out[sizeof(T) - 1 - i]); } return 0; } // Used to cast the binary data of any T object, into a T object. Assumes data is in big ending (configurable) template inline static int fromBytes(const BYTE_TYPE* in, T& out) { if constexpr (!(std::is_same_v || std::is_same_v)) static_assert("Must provide a signed/unsigned int8 type"); std::array data; std::memcpy(data.data(), in, sizeof(T)); if constexpr (ENDIAN_LOOKUP(little_endian)) { // if we need to swap find the best way to do so if constexpr (std::is_same_v || std::is_same_v) out = SWAP16(*reinterpret_cast(data.data())); else if constexpr (std::is_same_v || std::is_same_v) out = SWAP32(*reinterpret_cast(data.data())); else if constexpr (std::is_same_v || std::is_same_v) out = SWAP64(*reinterpret_cast(data.data())); else { std::reverse(data.begin(), data.end()); out = *reinterpret_cast(data.data()); } } return 0; } template inline static int fromBytes(const BYTE_TYPE* in, T* out) { return fromBytes(in, *out); } } template struct ptr_iterator { public: using iterator_category = std::random_access_iterator_tag; using difference_type = std::ptrdiff_t; using value_type = V; using pointer = value_type*; using reference = value_type&; explicit ptr_iterator(V* v): _v(v) {} reference operator*() const { return *_v; } pointer operator->() { return _v; } ptr_iterator& operator++() { _v++; return *this; } ptr_iterator& operator--() { _v--; return *this; } ptr_iterator operator++(int) { auto tmp = *this; ++(*this); return tmp; } ptr_iterator operator--(int) { auto tmp = *this; --(*this); return tmp; } friend bool operator==(const ptr_iterator& a, const ptr_iterator& b) { return a._v == b._v; } friend bool operator!=(const ptr_iterator& a, const ptr_iterator& b) { return a._v != b._v; } private: V* _v; }; /** * Creates an encapsulation of a T array which will be automatically deleted when this object goes out of scope. * This is a simple buffer meant to be used only inside of a function and not copied around. * The internal buffer is allocated on the heap. * The operator * has been overloaded to return the internal buffer. * @tparam T type that is stored in buffer eg char */ template || std::is_copy_assignable_v> class scoped_buffer { private: T* buffer_ = nullptr; size_t size_; public: scoped_buffer(): buffer_(nullptr), size_(0) {} explicit scoped_buffer(size_t size): size_(size) { if (size > 0) buffer_ = new T[size]; else buffer_ = nullptr; } scoped_buffer(const scoped_buffer& copy) { if (copy.size() == 0) { buffer_ = nullptr; size_ = 0; return; } buffer_ = new T[copy.size()]; size_ = copy.size_; if constexpr (std::is_trivially_copyable_v) { std::memcpy(buffer_, copy.buffer_, copy.size() * sizeof(T)); } else { if constexpr (std::is_copy_constructible_v && !std::is_copy_assignable_v) { for (size_t i = 0; i < this->size_; i++) buffer_[i] = T(copy[i]); } else for (size_t i = 0; i < this->size_; i++) buffer_[i] = copy[i]; } } scoped_buffer& operator=(const scoped_buffer& copy) { if (© == this) return *this; if (copy.size() == 0) { buffer_ = nullptr; size_ = 0; return *this; } delete[] this->buffer_; buffer_ = new T[copy.size()]; size_ = copy.size_; if constexpr (std::is_trivially_copyable_v) { std::memcpy(buffer_, copy.buffer_, copy.size() * sizeof(T)); } else { if constexpr (std::is_copy_constructible_v && !std::is_copy_assignable_v) { for (size_t i = 0; i < this->size_; i++) buffer_[i] = T(copy[i]); } else for (size_t i = 0; i < this->size_; i++) buffer_[i] = copy[i]; } return *this; } scoped_buffer(scoped_buffer&& move) noexcept { delete[] buffer_; buffer_ = move.buffer_; size_ = move.size(); move.buffer_ = nullptr; } scoped_buffer& operator=(scoped_buffer&& moveAssignment) noexcept { delete[] buffer_; buffer_ = moveAssignment.buffer_; size_ = moveAssignment.size(); moveAssignment.buffer_ = nullptr; return *this; } inline T& operator[](size_t index) { return buffer_[index]; } inline const T& operator[](size_t index) const { return buffer_[index]; } inline T* operator*() { return buffer_; } [[nodiscard]] inline size_t size() const { return size_; } inline T*& ptr() { return buffer_; } inline const T* const& ptr() const { return buffer_; } inline const T* const& data() const { return buffer_; } inline T*& data() { return buffer_; } inline ptr_iterator begin() { return ptr_iterator{buffer_}; } inline ptr_iterator end() { return ptr_iterator{&buffer_[size_]}; } ~scoped_buffer() { delete[] buffer_; } }; template class static_vector { private: T buffer_[MAX_SIZE]; size_t size_ = 0; public: static_vector() = default; inline bool push_back(const T& copy) { if (size_ >= MAX_SIZE) return false; buffer_[size_++] = copy; return true; } inline bool push_back(T&& move) { if (size_ >= MAX_SIZE) return false; buffer_[size_++] = std::move(move); return true; } inline T& at(size_t index) { if (index >= MAX_SIZE) throw std::runtime_error("Array index " + std::to_string(index) + " out of bounds! (Max size: " + std::to_string(MAX_SIZE) + ')'); } inline T& operator[](size_t index) { return buffer_[index]; } inline const T& operator[](size_t index) const { return buffer_[index]; } inline void reserve(size_t size) { if (size > MAX_SIZE) size = MAX_SIZE; size_ = size; } [[nodiscard]] inline size_t size() const { return size_; } [[nodiscard]] inline size_t capacity() const { return MAX_SIZE; } inline T* data() { return buffer_; } inline T* operator*() { return buffer_; } inline T* data() const { return buffer_; } inline T* begin() { return buffer_; } inline T* end() { return &buffer_[size_]; } }; template class scoped_buffer : scoped_buffer { using scoped_buffer::scoped_buffer; public: scoped_buffer(const scoped_buffer& copy) = delete; scoped_buffer operator=(scoped_buffer& copyAssignment) = delete; }; template struct nullptr_initializer { private: T* m_ptr = nullptr; public: nullptr_initializer() = default; explicit nullptr_initializer(T* ptr): m_ptr(ptr) {} nullptr_initializer(const nullptr_initializer& ptr): m_ptr(ptr.m_ptr) {} nullptr_initializer(nullptr_initializer&& ptr) noexcept: m_ptr(ptr.m_ptr) {} nullptr_initializer& operator=(const nullptr_initializer& ptr) { if (&ptr == this) return *this; this->m_ptr = ptr.m_ptr; return *this; } nullptr_initializer& operator=(nullptr_initializer&& ptr) noexcept { if (&ptr == this) return *this; this->m_ptr = ptr.m_ptr; return *this; } inline T* operator->() { return m_ptr; } ~nullptr_initializer() = default; }; /** * Creates a hash-map like association between an enum key and any arbitrary value. * The storage is backed by a contiguous array for faster access. * @tparam K enum value * @tparam V associated value */ template class enum_storage { private: V* m_values; size_t m_size = 0; public: enum_storage(std::initializer_list> init) { for (auto& i : init) m_size = std::max((size_t) i.first, m_size); m_values = new V[m_size]; for (auto& v : init) m_values[(size_t) v.first] = v.second; } inline V& operator[](size_t index) { return m_values[index]; } inline const V& operator[](size_t index) const { return m_values[index]; } [[nodiscard]] inline size_t size() const { return m_size; } ptr_iterator begin() { return ptr_iterator{m_values}; } ptr_iterator end() { return ptr_iterator{&m_values[m_size]}; } ~enum_storage() { delete[] m_values; } }; template class area_allocator { public: typedef T value_type; typedef T* pointer; typedef const T* const_pointer; typedef void* void_pointer; typedef const void* const_void_pointer; private: /** * Stores a view to a region of memory that has been deallocated * This is a non-owning reference to the memory block * * pointer p is the pointer to the beginning of the block of memory * size_t n is the number of elements that this block can hold */ struct pointer_view { pointer p; size_t n; }; /** * Stores the actual data for allocated blocks. Since we would like to be able to allocate an arbitrary number of items * we need a way of storing that data. The block storage holds an owning pointer to a region of memory with used elements * Only up to used has to have their destructors called, which should be handled by the deallocate function * it is UB to not deallocate memory allocated by this allocator * * an internal vector is used to store the regions of memory which have been deallocated. the allocate function will search for * free blocks with sufficient size in order to maximize memory usage. In the future more advanced methods should be used * for both faster access to deallocated blocks of sufficient size and to ensure coherent memory. */ struct block_storage { pointer data; size_t used = 0; // TODO: b-tree? std::vector unallocated_blocks; }; /** * Stores an index to a pointer_view along with the amount of memory leftover after the allocation * it also stores the block being allocated to in question. The new inserted leftover should start at old_ptr + size */ struct block_view { block_storage* blk; size_t index; size_t leftover; block_view(block_storage* blk, size_t index, size_t leftover): blk(blk), index(index), leftover(leftover) {} }; /** * Allocate a new block of memory and push it to the back of blocks. */ inline void allocate_block() { //BLT_INFO("Allocating a new block of size %d", BLOCK_SIZE); auto* blk = new block_storage(); blk->data = static_cast(malloc(sizeof(T) * BLOCK_SIZE)); blocks.push_back(blk); } /** * Searches for a free block inside the block storage with sufficient space and returns an optional view to it * The optional will be empty if no open block can be found. */ inline std::optional search_for_block(block_storage* blk, size_t n) { for (auto kv : blt::enumerate(blk->unallocated_blocks)) { if (kv.second.n >= n) return block_view{blk, kv.first, kv.second.n - n}; } return {}; } /** * removes the block of memory from the unallocated_blocks storage in the underlying block, inserting a new unallocated block if * there was any leftover. Returns a pointer to the beginning of the new block. */ inline pointer swap_pop_resize_if(const block_view& view, size_t n) { pointer_view ptr = view.blk->unallocated_blocks[view.index]; std::iter_swap(view.blk->unallocated_blocks.begin() + view.index, view.blk->unallocated_blocks.end() - 1); view.blk->unallocated_blocks.pop_back(); if (view.leftover > 0) view.blk->unallocated_blocks.push_back({ptr.p + n, view.leftover}); return ptr.p; } /** * Finds the next available unallocated block of memory, or empty if there is none which meet size requirements */ inline std::optional find_available_block(size_t n) { for (auto* blk : blocks) { if (auto view = search_for_block(blk, n)) return swap_pop_resize_if(view.value(), n); } return {}; } /** * returns a pointer to a block of memory along with an offset into that pointer that the requested block can be found at */ inline std::pair getBlock(size_t n) { if (auto blk = find_available_block(n)) return {blk.value(), 0}; if (blocks.back()->used + n > BLOCK_SIZE) allocate_block(); auto ptr = std::pair{blocks.back()->data, blocks.back()->used}; blocks.back()->used += n; return ptr; } /** * Calls the constructor on elements if they require construction, otherwise constructor will not be called and this function is useless */ inline void allocate_in_block(pointer begin, size_t n) { if constexpr (std::is_default_constructible_v && !std::is_trivially_default_constructible_v) { for (size_t i = 0; i < n; i++) new(&begin[i]) T(); } } public: area_allocator() { allocate_block(); } [[nodiscard]] pointer allocate(size_t n) { if (n > BLOCK_SIZE) throw std::runtime_error("Requested allocation is too large!"); auto block_info = getBlock(n); auto* ptr = &block_info.first[block_info.second]; // call constructors on the objects if they require it allocate_in_block(ptr, n); return ptr; } void deallocate(pointer p, size_t n) noexcept { for (size_t i = 0; i < n; i++) p[i].~T(); for (auto*& blk : blocks) { if (p >= blk->data && p <= (blk->data + BLOCK_SIZE)) { blk->unallocated_blocks.push_back({p, n}); break; } } } ~area_allocator() { for (auto*& blk : blocks) { free(blk->data); delete blk; } } private: std::vector blocks; }; } #endif //BLT_TESTS_MEMORY_H