2020-01-29 09:49:04 +00:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2018-2020 Atmosphère-NX
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
|
|
* under the terms and conditions of the GNU General Public License,
|
|
|
|
* version 2, as published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
|
|
* more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
*/
|
|
|
|
#include <mesosphere.hpp>
|
|
|
|
|
|
|
|
namespace ams::kern {
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
|
2020-02-07 01:40:57 +00:00
|
|
|
constexpr KMemoryManager::Pool GetPoolFromMemoryRegionType(u32 type) {
|
|
|
|
switch (type) {
|
|
|
|
case KMemoryRegionType_VirtualDramApplicationPool: return KMemoryManager::Pool_Application;
|
|
|
|
case KMemoryRegionType_VirtualDramAppletPool: return KMemoryManager::Pool_Applet;
|
|
|
|
case KMemoryRegionType_VirtualDramSystemPool: return KMemoryManager::Pool_System;
|
|
|
|
case KMemoryRegionType_VirtualDramSystemNonSecurePool: return KMemoryManager::Pool_SystemNonSecure;
|
|
|
|
MESOSPHERE_UNREACHABLE_DEFAULT_CASE();
|
|
|
|
}
|
|
|
|
}
|
2020-01-29 09:49:04 +00:00
|
|
|
|
|
|
|
}
|
|
|
|
|
2020-02-07 01:40:57 +00:00
|
|
|
void KMemoryManager::Initialize(KVirtualAddress metadata_region, size_t metadata_region_size) {
|
|
|
|
/* Clear the metadata region to zero. */
|
|
|
|
const KVirtualAddress metadata_region_end = metadata_region + metadata_region_size;
|
|
|
|
std::memset(GetVoidPointer(metadata_region), 0, metadata_region_size);
|
|
|
|
|
|
|
|
/* Traverse the virtual memory layout tree, initializing each manager as appropriate. */
|
|
|
|
while (true) {
|
|
|
|
/* Locate the region that should initialize the current manager. */
|
|
|
|
const KMemoryRegion *region = nullptr;
|
|
|
|
for (const auto &it : KMemoryLayout::GetVirtualMemoryRegionTree()) {
|
|
|
|
/* We only care about regions that we need to create managers for. */
|
|
|
|
if (!it.IsDerivedFrom(KMemoryRegionType_VirtualDramManagedPool)) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* We want to initialize the managers in order. */
|
|
|
|
if (it.GetAttributes() != this->num_managers) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
region = std::addressof(it);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If we didn't find a region, then we're done initializing managers. */
|
|
|
|
if (region == nullptr) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Ensure that the region is correct. */
|
|
|
|
MESOSPHERE_ASSERT(region->GetAddress() != Null<decltype(region->GetAddress())>);
|
|
|
|
MESOSPHERE_ASSERT(region->GetSize() > 0);
|
|
|
|
MESOSPHERE_ASSERT(region->GetEndAddress() >= region->GetAddress());
|
|
|
|
MESOSPHERE_ASSERT(region->IsDerivedFrom(KMemoryRegionType_VirtualDramManagedPool));
|
|
|
|
MESOSPHERE_ASSERT(region->GetAttributes() == this->num_managers);
|
|
|
|
|
|
|
|
/* Initialize a new manager for the region. */
|
|
|
|
const Pool pool = GetPoolFromMemoryRegionType(region->GetType());
|
|
|
|
Impl *manager = std::addressof(this->managers[this->num_managers++]);
|
|
|
|
MESOSPHERE_ABORT_UNLESS(this->num_managers <= util::size(this->managers));
|
|
|
|
|
|
|
|
const size_t cur_size = manager->Initialize(region, pool, metadata_region, metadata_region_end);
|
|
|
|
metadata_region += cur_size;
|
|
|
|
MESOSPHERE_ABORT_UNLESS(metadata_region <= metadata_region_end);
|
|
|
|
|
|
|
|
/* Insert the manager into the pool list. */
|
|
|
|
if (this->pool_managers_tail[pool] == nullptr) {
|
|
|
|
this->pool_managers_head[pool] = manager;
|
|
|
|
} else {
|
|
|
|
this->pool_managers_tail[pool]->SetNext(manager);
|
|
|
|
manager->SetPrev(this->pool_managers_tail[pool]);
|
|
|
|
}
|
|
|
|
this->pool_managers_tail[pool] = manager;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-22 05:13:16 +00:00
|
|
|
Result KMemoryManager::InitializeOptimizedMemory(u64 process_id, Pool pool) {
|
|
|
|
/* Lock the pool. */
|
|
|
|
KScopedLightLock lk(this->pool_locks[pool]);
|
|
|
|
|
|
|
|
/* Check that we don't already have an optimized process. */
|
|
|
|
R_UNLESS(!this->has_optimized_process[pool], svc::ResultBusy());
|
|
|
|
|
|
|
|
/* Set the optimized process id. */
|
|
|
|
this->optimized_process_ids[pool] = process_id;
|
|
|
|
this->has_optimized_process[pool] = true;
|
|
|
|
|
|
|
|
/* Clear the management area for the optimized process. */
|
|
|
|
for (auto *manager = this->GetFirstManager(pool, Direction_FromFront); manager != nullptr; manager = this->GetNextManager(manager, Direction_FromFront)) {
|
|
|
|
manager->InitializeOptimizedMemory();
|
|
|
|
}
|
|
|
|
|
|
|
|
return ResultSuccess();
|
|
|
|
}
|
|
|
|
|
2020-07-25 03:44:15 +00:00
|
|
|
void KMemoryManager::FinalizeOptimizedMemory(u64 process_id, Pool pool) {
|
|
|
|
/* Lock the pool. */
|
|
|
|
KScopedLightLock lk(this->pool_locks[pool]);
|
|
|
|
|
|
|
|
/* If the process was optimized, clear it. */
|
|
|
|
if (this->has_optimized_process[pool] && this->optimized_process_ids[pool] == process_id) {
|
|
|
|
this->has_optimized_process[pool] = false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-22 05:13:16 +00:00
|
|
|
|
2020-02-07 04:36:26 +00:00
|
|
|
KVirtualAddress KMemoryManager::AllocateContinuous(size_t num_pages, size_t align_pages, u32 option) {
|
|
|
|
/* Early return if we're allocating no pages. */
|
|
|
|
if (num_pages == 0) {
|
|
|
|
return Null<KVirtualAddress>;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Lock the pool that we're allocating from. */
|
|
|
|
const auto [pool, dir] = DecodeOption(option);
|
|
|
|
KScopedLightLock lk(this->pool_locks[pool]);
|
|
|
|
|
|
|
|
/* Choose a heap based on our page size request. */
|
|
|
|
const s32 heap_index = KPageHeap::GetAlignedBlockIndex(num_pages, align_pages);
|
|
|
|
|
|
|
|
/* Loop, trying to iterate from each block. */
|
|
|
|
Impl *chosen_manager = nullptr;
|
|
|
|
KVirtualAddress allocated_block = Null<KVirtualAddress>;
|
2020-02-14 01:38:56 +00:00
|
|
|
for (chosen_manager = this->GetFirstManager(pool, dir); chosen_manager != nullptr; chosen_manager = this->GetNextManager(chosen_manager, dir)) {
|
2020-04-19 00:10:26 +00:00
|
|
|
allocated_block = chosen_manager->AllocateBlock(heap_index, true);
|
2020-02-14 01:38:56 +00:00
|
|
|
if (allocated_block != Null<KVirtualAddress>) {
|
|
|
|
break;
|
2020-02-07 04:36:26 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If we failed to allocate, quit now. */
|
|
|
|
if (allocated_block == Null<KVirtualAddress>) {
|
|
|
|
return Null<KVirtualAddress>;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If we allocated more than we need, free some. */
|
|
|
|
const size_t allocated_pages = KPageHeap::GetBlockNumPages(heap_index);
|
|
|
|
if (allocated_pages > num_pages) {
|
|
|
|
chosen_manager->Free(allocated_block + num_pages * PageSize, allocated_pages - num_pages);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Maintain the optimized memory bitmap, if we should. */
|
|
|
|
if (this->has_optimized_process[pool]) {
|
2020-07-24 15:07:34 +00:00
|
|
|
chosen_manager->TrackUnoptimizedAllocation(allocated_block, num_pages);
|
2020-02-07 04:36:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return allocated_block;
|
|
|
|
}
|
|
|
|
|
2020-07-24 15:07:34 +00:00
|
|
|
Result KMemoryManager::AllocatePageGroupImpl(KPageGroup *out, size_t num_pages, Pool pool, Direction dir, bool unoptimized, bool random) {
|
2020-02-14 01:38:56 +00:00
|
|
|
/* Choose a heap based on our page size request. */
|
|
|
|
const s32 heap_index = KPageHeap::GetBlockIndex(num_pages);
|
|
|
|
R_UNLESS(0 <= heap_index, svc::ResultOutOfMemory());
|
|
|
|
|
|
|
|
/* Ensure that we don't leave anything un-freed. */
|
|
|
|
auto group_guard = SCOPE_GUARD {
|
|
|
|
for (const auto &it : *out) {
|
|
|
|
auto &manager = this->GetManager(it.GetAddress());
|
|
|
|
const size_t num_pages = std::min(it.GetNumPages(), (manager.GetEndAddress() - it.GetAddress()) / PageSize);
|
|
|
|
manager.Free(it.GetAddress(), num_pages);
|
|
|
|
}
|
|
|
|
out->Finalize();
|
|
|
|
};
|
|
|
|
|
|
|
|
/* Keep allocating until we've allocated all our pages. */
|
|
|
|
for (s32 index = heap_index; index >= 0 && num_pages > 0; index--) {
|
|
|
|
const size_t pages_per_alloc = KPageHeap::GetBlockNumPages(index);
|
|
|
|
for (Impl *cur_manager = this->GetFirstManager(pool, dir); cur_manager != nullptr; cur_manager = this->GetNextManager(cur_manager, dir)) {
|
|
|
|
while (num_pages >= pages_per_alloc) {
|
|
|
|
/* Allocate a block. */
|
2020-04-19 00:10:26 +00:00
|
|
|
KVirtualAddress allocated_block = cur_manager->AllocateBlock(index, random);
|
2020-02-14 01:38:56 +00:00
|
|
|
if (allocated_block == Null<KVirtualAddress>) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Safely add it to our group. */
|
|
|
|
{
|
|
|
|
auto block_guard = SCOPE_GUARD { cur_manager->Free(allocated_block, pages_per_alloc); };
|
|
|
|
R_TRY(out->AddBlock(allocated_block, pages_per_alloc));
|
|
|
|
block_guard.Cancel();
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Maintain the optimized memory bitmap, if we should. */
|
2020-07-24 15:07:34 +00:00
|
|
|
if (unoptimized) {
|
|
|
|
cur_manager->TrackUnoptimizedAllocation(allocated_block, pages_per_alloc);
|
2020-02-14 01:38:56 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
num_pages -= pages_per_alloc;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Only succeed if we allocated as many pages as we wanted. */
|
|
|
|
R_UNLESS(num_pages == 0, svc::ResultOutOfMemory());
|
|
|
|
|
|
|
|
/* We succeeded! */
|
|
|
|
group_guard.Cancel();
|
|
|
|
return ResultSuccess();
|
|
|
|
}
|
|
|
|
|
2020-04-19 00:10:26 +00:00
|
|
|
Result KMemoryManager::Allocate(KPageGroup *out, size_t num_pages, u32 option) {
|
|
|
|
MESOSPHERE_ASSERT(out != nullptr);
|
|
|
|
MESOSPHERE_ASSERT(out->GetNumPages() == 0);
|
|
|
|
|
|
|
|
/* Early return if we're allocating no pages. */
|
|
|
|
R_SUCCEED_IF(num_pages == 0);
|
|
|
|
|
|
|
|
/* Lock the pool that we're allocating from. */
|
|
|
|
const auto [pool, dir] = DecodeOption(option);
|
|
|
|
KScopedLightLock lk(this->pool_locks[pool]);
|
|
|
|
|
|
|
|
/* Allocate the page group. */
|
|
|
|
return this->AllocatePageGroupImpl(out, num_pages, pool, dir, this->has_optimized_process[pool], true);
|
|
|
|
}
|
|
|
|
|
2020-07-24 15:07:34 +00:00
|
|
|
Result KMemoryManager::AllocateForProcess(KPageGroup *out, size_t num_pages, u32 option, u64 process_id, u8 fill_pattern) {
|
|
|
|
MESOSPHERE_ASSERT(out != nullptr);
|
|
|
|
MESOSPHERE_ASSERT(out->GetNumPages() == 0);
|
|
|
|
|
|
|
|
/* Decode the option. */
|
|
|
|
const auto [pool, dir] = DecodeOption(option);
|
|
|
|
|
|
|
|
/* Allocate the memory. */
|
|
|
|
bool has_optimized, is_optimized;
|
|
|
|
{
|
|
|
|
/* Lock the pool that we're allocating from. */
|
|
|
|
KScopedLightLock lk(this->pool_locks[pool]);
|
|
|
|
|
|
|
|
/* Check if we have an optimized process. */
|
|
|
|
has_optimized = this->has_optimized_process[pool];
|
|
|
|
is_optimized = this->optimized_process_ids[pool] == process_id;
|
|
|
|
|
|
|
|
/* Allocate the page group. */
|
|
|
|
R_TRY(this->AllocatePageGroupImpl(out, num_pages, pool, dir, has_optimized && !is_optimized, false));
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Perform optimized memory tracking, if we should. */
|
|
|
|
if (has_optimized && is_optimized) {
|
|
|
|
/* Iterate over the allocated blocks. */
|
|
|
|
for (const auto &block : *out) {
|
|
|
|
/* Get the block extents. */
|
|
|
|
const KVirtualAddress block_address = block.GetAddress();
|
|
|
|
const size_t block_pages = block.GetNumPages();
|
|
|
|
|
|
|
|
/* If it has no pages, we don't need to do anything. */
|
|
|
|
if (block_pages == 0) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Fill all the pages that we need to fill. */
|
|
|
|
bool any_new = false;
|
|
|
|
{
|
|
|
|
KVirtualAddress cur_address = block_address;
|
|
|
|
size_t cur_pages = block_pages;
|
|
|
|
while (cur_pages > 0) {
|
|
|
|
/* Get the manager for the current address. */
|
|
|
|
auto &manager = this->GetManager(cur_address);
|
|
|
|
|
|
|
|
/* Process part or all of the block. */
|
|
|
|
const size_t processed_pages = manager.ProcessOptimizedAllocation(std::addressof(any_new), cur_address, cur_pages, fill_pattern);
|
|
|
|
|
|
|
|
/* Advance. */
|
|
|
|
cur_address += processed_pages * PageSize;
|
|
|
|
cur_pages -= processed_pages;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If there are no new pages, move on to the next block. */
|
|
|
|
if (!any_new) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Update tracking for the allocation. */
|
|
|
|
KVirtualAddress cur_address = block_address;
|
|
|
|
size_t cur_pages = block_pages;
|
|
|
|
while (cur_pages > 0) {
|
|
|
|
/* Get the manager for the current address. */
|
|
|
|
auto &manager = this->GetManager(cur_address);
|
|
|
|
|
|
|
|
/* Lock the pool for the manager. */
|
|
|
|
KScopedLightLock lk(this->pool_locks[manager.GetPool()]);
|
|
|
|
|
|
|
|
/* Track some or all of the current pages. */
|
|
|
|
const size_t processed_pages = manager.TrackOptimizedAllocation(cur_address, cur_pages);
|
|
|
|
|
|
|
|
/* Advance. */
|
|
|
|
cur_address += processed_pages * PageSize;
|
|
|
|
cur_pages -= processed_pages;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/* Set all the allocated memory. */
|
|
|
|
for (const auto &block : *out) {
|
|
|
|
std::memset(GetVoidPointer(block.GetAddress()), fill_pattern, block.GetSize());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return ResultSuccess();
|
|
|
|
}
|
|
|
|
|
2020-02-07 01:40:57 +00:00
|
|
|
size_t KMemoryManager::Impl::Initialize(const KMemoryRegion *region, Pool p, KVirtualAddress metadata, KVirtualAddress metadata_end) {
|
|
|
|
/* Calculate metadata sizes. */
|
|
|
|
const size_t ref_count_size = (region->GetSize() / PageSize) * sizeof(u16);
|
2020-07-22 05:13:16 +00:00
|
|
|
const size_t optimize_map_size = CalculateOptimizedProcessOverheadSize(region->GetSize());
|
2020-02-07 01:40:57 +00:00
|
|
|
const size_t manager_size = util::AlignUp(optimize_map_size + ref_count_size, PageSize);
|
|
|
|
const size_t page_heap_size = KPageHeap::CalculateMetadataOverheadSize(region->GetSize());
|
|
|
|
const size_t total_metadata_size = manager_size + page_heap_size;
|
|
|
|
MESOSPHERE_ABORT_UNLESS(manager_size <= total_metadata_size);
|
|
|
|
MESOSPHERE_ABORT_UNLESS(metadata + total_metadata_size <= metadata_end);
|
|
|
|
MESOSPHERE_ABORT_UNLESS(util::IsAligned(total_metadata_size, PageSize));
|
|
|
|
|
|
|
|
/* Setup region. */
|
|
|
|
this->pool = p;
|
|
|
|
this->metadata_region = metadata;
|
|
|
|
this->page_reference_counts = GetPointer<RefCount>(metadata + optimize_map_size);
|
|
|
|
MESOSPHERE_ABORT_UNLESS(util::IsAligned(GetInteger(this->metadata_region), PageSize));
|
|
|
|
|
|
|
|
/* Initialize the manager's KPageHeap. */
|
|
|
|
this->heap.Initialize(region->GetAddress(), region->GetSize(), metadata + manager_size, page_heap_size);
|
|
|
|
|
2020-02-07 04:36:26 +00:00
|
|
|
/* Free the memory to the heap. */
|
|
|
|
this->heap.Free(region->GetAddress(), region->GetSize() / PageSize);
|
|
|
|
|
|
|
|
/* Update the heap's used size. */
|
|
|
|
this->heap.UpdateUsedSize();
|
|
|
|
|
2020-02-07 01:40:57 +00:00
|
|
|
return total_metadata_size;
|
|
|
|
}
|
|
|
|
|
2020-07-24 15:07:34 +00:00
|
|
|
void KMemoryManager::Impl::TrackUnoptimizedAllocation(KVirtualAddress block, size_t num_pages) {
|
2020-02-07 04:36:26 +00:00
|
|
|
size_t offset = this->heap.GetPageOffset(block);
|
|
|
|
const size_t last = offset + num_pages - 1;
|
|
|
|
u64 *optimize_map = GetPointer<u64>(this->metadata_region);
|
|
|
|
while (offset <= last) {
|
|
|
|
optimize_map[offset / BITSIZEOF(u64)] &= ~(u64(1) << (offset % BITSIZEOF(u64)));
|
|
|
|
offset++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-24 15:07:34 +00:00
|
|
|
size_t KMemoryManager::Impl::TrackOptimizedAllocation(KVirtualAddress block, size_t num_pages) {
|
|
|
|
/* Get the number of tracking pages. */
|
|
|
|
const size_t cur_pages = std::min(num_pages, this->heap.GetPageOffsetToEnd(block));
|
|
|
|
|
|
|
|
/* Get the range we're tracking. */
|
|
|
|
size_t offset = this->heap.GetPageOffset(block);
|
|
|
|
const size_t last = offset + cur_pages - 1;
|
|
|
|
|
|
|
|
/* Track. */
|
|
|
|
u64 *optimize_map = GetPointer<u64>(this->metadata_region);
|
|
|
|
while (offset <= last) {
|
|
|
|
/* Mark the page as being optimized-allocated. */
|
|
|
|
optimize_map[offset / BITSIZEOF(u64)] |= (u64(1) << (offset % BITSIZEOF(u64)));
|
|
|
|
|
|
|
|
offset++;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Return the number of pages we tracked. */
|
|
|
|
return cur_pages;
|
|
|
|
}
|
|
|
|
|
|
|
|
size_t KMemoryManager::Impl::ProcessOptimizedAllocation(bool *out_any_new, KVirtualAddress block, size_t num_pages, u8 fill_pattern) {
|
|
|
|
/* Get the number of processable pages. */
|
|
|
|
const size_t cur_pages = std::min(num_pages, this->heap.GetPageOffsetToEnd(block));
|
|
|
|
|
|
|
|
/* Clear any new. */
|
|
|
|
*out_any_new = false;
|
|
|
|
|
|
|
|
/* Get the range we're processing. */
|
|
|
|
size_t offset = this->heap.GetPageOffset(block);
|
|
|
|
const size_t last = offset + cur_pages - 1;
|
|
|
|
|
|
|
|
/* Process. */
|
|
|
|
u64 *optimize_map = GetPointer<u64>(this->metadata_region);
|
|
|
|
while (offset <= last) {
|
|
|
|
/* Check if the page has been optimized-allocated before. */
|
|
|
|
if ((optimize_map[offset / BITSIZEOF(u64)] & (u64(1) << (offset % BITSIZEOF(u64)))) == 0) {
|
|
|
|
/* If not, it's new. */
|
|
|
|
*out_any_new = true;
|
|
|
|
|
|
|
|
/* Fill the page. */
|
|
|
|
std::memset(GetVoidPointer(this->heap.GetAddress() + offset * PageSize), fill_pattern, PageSize);
|
|
|
|
}
|
|
|
|
|
|
|
|
offset++;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Return the number of pages we processed. */
|
|
|
|
return cur_pages;
|
|
|
|
}
|
|
|
|
|
2020-01-29 09:49:04 +00:00
|
|
|
size_t KMemoryManager::Impl::CalculateMetadataOverheadSize(size_t region_size) {
|
2020-02-07 01:40:57 +00:00
|
|
|
const size_t ref_count_size = (region_size / PageSize) * sizeof(u16);
|
|
|
|
const size_t optimize_map_size = (util::AlignUp((region_size / PageSize), BITSIZEOF(u64)) / BITSIZEOF(u64)) * sizeof(u64);
|
|
|
|
const size_t manager_meta_size = util::AlignUp(optimize_map_size + ref_count_size, PageSize);
|
|
|
|
const size_t page_heap_size = KPageHeap::CalculateMetadataOverheadSize(region_size);
|
|
|
|
return manager_meta_size + page_heap_size;
|
2020-01-29 09:49:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
}
|