AuroraRuntime/Source/Memory/Heap.cpp

292 lines
8.0 KiB
C++
Raw Normal View History

2021-06-27 21:25:29 +00:00
/***
Copyright (C) 2021 J Reece Wilson (a/k/a "Reece"). All rights reserved.
File: Heap.cpp
Date: 2021-6-12
Author: Reece
***/
2021-09-30 14:57:41 +00:00
#include <Source/RuntimeInternal.hpp>
2021-06-27 21:25:29 +00:00
#include "Memory.hpp"
#include "Heap.hpp"
#include "mimalloc.h"
#include "o1heap.hpp"
namespace Aurora::Memory
{
static AuUInt32 RoundPageUp(AuUInt32 value)
{
auto pageMask = HWInfo::GetPageSize() - 1;
return (value + pageMask) & ~(pageMask);
}
static void *HeapLargeAllocate(AuUInt length)
{
length = RoundPageUp(length);
#if defined(AURORA_IS_MODERNNT_DERIVED)
return VirtualAlloc(nullptr, length, MEM_COMMIT | MEM_RESERVE, PAGE_READWRITE);
#elif defined(AURORA_IS_POSIX_DERIVED)
return mmap(0, length, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
#else
// ideally we should page align.
// i think mimalloc has fast paths with warnings for overly large passthrough allocations. unsure.
// 32 alignment in the fastest way mimalloc can provide us memory seems adequate
return Memory::FAlloc<void *>(length, 32);
#endif
}
static void HeapLargeFree(void *buffer, AuUInt length)
{
length = RoundPageUp(length);
#if defined(AURORA_IS_MODERNNT_DERIVED)
VirtualFree(buffer, 0, MEM_RELEASE);
#elif defined(AURORA_IS_POSIX_DERIVED)
munmap(buffer, length);
#else
Memory::Free(buffer);
mi_collect(false);
#endif
}
class InternalHeap : public Heap, std::enable_shared_from_this<InternalHeap>
2021-06-27 21:25:29 +00:00
{
public:
InternalHeap() : base_(nullptr), mutex_(nullptr), heap_(nullptr), count_(0)
2021-07-12 12:17:16 +00:00
{ }
2021-06-27 21:25:29 +00:00
~InternalHeap();
bool Init(AuUInt length);
typedef struct FragmentHeader
{
void *next;
void *prev;
size_t size;
bool used;
} FragmentHeader;
static AuUInt GetHeapSize(const void *ptr)
{
return reinterpret_cast<const FragmentHeader *>(ptr)[-1].size;
}
Types::size_t GetChunkSize(const void *head) override;
void *_FAlloc(Types::size_t length) override;
void *_FAlloc(Types::size_t length, Types::size_t align) override;
void *_ZAlloc(Types::size_t length) override;
void *_ZAlloc(Types::size_t length, Types::size_t align) override;
void *_ZRealloc(void *buffer, Types::size_t length) override;
void *_ZRealloc(void *buffer, Types::size_t length, Types::size_t align) override;
void *_FRealloc(void *buffer, Types::size_t length) override;
void *_FRealloc(void *buffer, Types::size_t length, Types::size_t align) override;
void _Free(void *buffer) override;
AuSPtr<Heap> GetSelfReference() override;
2021-07-12 12:17:16 +00:00
2021-06-27 21:25:29 +00:00
void TryRelease();
void DecrementUsers();
2021-06-27 21:25:29 +00:00
void RequestTermination();
private:
AuThreadPrimitives::MutexUnique_t mutex_;
2021-07-12 12:17:16 +00:00
void *base_ {};
O1HeapInstance *heap_ {};
int count_ {};
AuUInt length_ {};
bool isDangling_ {};
2021-06-27 21:25:29 +00:00
};
2021-07-12 12:17:16 +00:00
2021-06-27 21:25:29 +00:00
InternalHeap::~InternalHeap()
{
SysAssertDbgExp(count_ == 0);
2021-06-27 21:25:29 +00:00
if (base_)
{
o1HeapReleaseCpp(heap_);// ->~O1HeapInstance(); // TODO: make free func
HeapLargeFree(base_, length_);
2021-06-27 21:25:29 +00:00
base_ = nullptr;
}
2021-07-12 12:17:16 +00:00
mutex_.reset();
2021-06-27 21:25:29 +00:00
}
bool InternalHeap::Init(AuUInt length)
{
SysAssert(!base_, "heap already initialized");
SysAssert(!mutex_, "heap already initialized");
SysAssert(length, "invalid heap allocation");
length_ = length;
2021-06-27 21:25:29 +00:00
mutex_ = AuThreadPrimitives::MutexUnique();
2021-06-27 21:25:29 +00:00
if (!mutex_) return false;
base_ = Memory::FAlloc<void *>(length);
if (!base_) return false;
heap_ = o1heapInit(base_, length,
[this](const O1HeapInstance *const handle) -> void
{
SysAssertDbg(this->mutex_, "missing mutex");
this->mutex_->Lock();
},
[this](const O1HeapInstance *const handle) -> void
{
SysAssertDbg(this->mutex_, "missing mutex");
this->mutex_->Unlock();
}
);
2021-06-27 21:25:29 +00:00
return true;
}
Types::size_t InternalHeap::GetChunkSize(const void *head)
{
return InternalHeap::GetHeapSize(head);
}
void *InternalHeap::_FAlloc(Types::size_t length)
{
if (!heap_) return nullptr;
auto ret = o1heapAllocate(heap_, length);
if (ret) count_++;
2021-06-27 21:25:29 +00:00
return ret;
}
void *InternalHeap::_FAlloc(Types::size_t length, Types::size_t align)
{
SysAssert(align < O1HEAP_ALIGNMENT, "heap wrapping is unsupported, alignment past the supported 2^x alignment is not possible");
return _FAlloc(length);
}
void *InternalHeap::_ZAlloc(Types::size_t length)
{
if (!heap_) return nullptr;
auto ptr = _FAlloc(length);
if (!ptr) return nullptr;
std::memset(ptr, 0, length);
return ptr;
}
void *InternalHeap::_ZAlloc(Types::size_t length, Types::size_t align)
{
SysAssert(align < O1HEAP_ALIGNMENT, "heap wrapping is unsupported, alignment past the supported 2^x alignment is not possible");
return _ZAlloc(length);
}
void *InternalHeap::_ZRealloc(void *buffer, Types::size_t length)
{
auto prevLength = GetHeapSize(buffer);
auto alloc = _ZAlloc(length);
if (!alloc) return nullptr;
std::memcpy(alloc, buffer, std::min(prevLength, length));
_Free(buffer);
return alloc;
}
void *InternalHeap::_ZRealloc(void *buffer, Types::size_t length, Types::size_t align)
{
SysAssert(align < O1HEAP_ALIGNMENT, "heap wrapping is unsupported, alignment past the supported 2^x alignment is not possible");
return _ZRealloc(buffer, length);
}
void *InternalHeap::_FRealloc(void *buffer, Types::size_t length)
{
auto prevLength = GetHeapSize(buffer);
auto alloc = _FAlloc(length);
if (!alloc) return nullptr;
std::memcpy(alloc, buffer, std::min(prevLength, length));
_Free(buffer);
return alloc;
}
void *InternalHeap::_FRealloc(void *buffer, Types::size_t length, Types::size_t align)
{
SysAssert(align < O1HEAP_ALIGNMENT, "heap wrapping is unsupported, alignment past the supported 2^x alignment is not possible");
return _FRealloc(buffer, length);
}
void InternalHeap::_Free(void *buffer)
{
if (buffer == nullptr) return;
o1heapFree(heap_, buffer);
DecrementUsers();
}
void InternalHeap::DecrementUsers()
{
if (--count_ == 0)
{
AU_LOCK_GUARD(this->mutex_);
TryRelease();
}
2021-06-27 21:25:29 +00:00
}
void InternalHeap::TryRelease()
{
if (!isDangling_)
{
return;
}
2021-06-27 21:25:29 +00:00
if (count_ == 0)
2021-06-27 21:25:29 +00:00
{
delete this;
}
}
void InternalHeap::RequestTermination()
{
AU_LOCK_GUARD(this->mutex_);
if (count_)
2021-06-27 21:25:29 +00:00
{
LogWarn("Heap life was less than its allocations, waiting for final free");
LogWarn("Reporting using mayday!");
Telemetry::Mayday();
isDangling_ = true;
2021-06-27 21:25:29 +00:00
TryRelease();
}
else
{
delete this;
}
}
AuSPtr<Heap> InternalHeap::GetSelfReference()
{
try
{
return shared_from_this();
}
catch (...)
{
return {};
}
}
2021-06-27 21:25:29 +00:00
AUKN_SYM Heap *AllocHeapNew(AuUInt size)
{
auto heap = _new InternalHeap();
if (!heap)
{
return nullptr;
}
if (!heap->Init(size))
{
delete heap;
return nullptr;
}
return heap;
}
2021-07-12 12:17:16 +00:00
AUKN_SYM void AllocHeapRelease(Heap *heap)
2021-06-27 21:25:29 +00:00
{
2021-09-06 10:58:08 +00:00
static_cast<InternalHeap *>(heap)->RequestTermination();
2021-06-27 21:25:29 +00:00
}
}