Get rid of this for now...
parent
1d0fef067f
commit
21e53083da
|
@ -176,7 +176,7 @@ u32 _heap_contract(T_HEAP *Heap, u64 Size)
|
||||||
return (Heap->End - Heap->Start);
|
return (Heap->End - Heap->Start);
|
||||||
}
|
}
|
||||||
|
|
||||||
void *heap_alloc_p(T_HEAP *Heap, u64 Size, u8 Aligned, u64 *Physical)
|
/*void *heap_alloc_p(T_HEAP *Heap, u64 Size, u8 Aligned, u64 *Physical)
|
||||||
{
|
{
|
||||||
void *Address = heap_alloc(Heap, Size, Aligned);
|
void *Address = heap_alloc(Heap, Size, Aligned);
|
||||||
|
|
||||||
|
@ -184,7 +184,7 @@ void *heap_alloc_p(T_HEAP *Heap, u64 Size, u8 Aligned, u64 *Physical)
|
||||||
paging_get_physical((u64)Address, Physical);
|
paging_get_physical((u64)Address, Physical);
|
||||||
|
|
||||||
return Address;
|
return Address;
|
||||||
}
|
}*/
|
||||||
|
|
||||||
void *heap_alloc(T_HEAP *Heap, u64 Size, u8 Aligned)
|
void *heap_alloc(T_HEAP *Heap, u64 Size, u8 Aligned)
|
||||||
{
|
{
|
||||||
|
@ -391,10 +391,10 @@ void *kmalloc(u64 Size)
|
||||||
return heap_alloc(g_Heap, Size, 0);
|
return heap_alloc(g_Heap, Size, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
void *kmalloc_p(u64 Size, u8 Aligned, u64 *Physical)
|
/*void *kmalloc_p(u64 Size, u8 Aligned, u64 *Physical)
|
||||||
{
|
{
|
||||||
return heap_alloc_p(g_Heap, Size, Aligned, Physical);
|
return heap_alloc_p(g_Heap, Size, Aligned, Physical);
|
||||||
}
|
}*/
|
||||||
|
|
||||||
void kfree(void *Data)
|
void kfree(void *Data)
|
||||||
{
|
{
|
||||||
|
|
|
@ -85,7 +85,7 @@ void paging_kernel_initialize(u64 KernelVirtualStart, u64 KernelPhysicalStart, u
|
||||||
g_KernelPaging.KernelSize = KernelSize;
|
g_KernelPaging.KernelSize = KernelSize;
|
||||||
}
|
}
|
||||||
|
|
||||||
u8 paging_get_physical_ex(u64 Virtual, u64 *Physical, T_PAGING_ML4 *ML4)
|
/*u8 paging_get_physical_ex(u64 Virtual, u64 *Physical, T_PAGING_ML4 *ML4)
|
||||||
{
|
{
|
||||||
if (Virtual < g_KernelPaging.KernelVirtualStart || Virtual > g_KernelPaging.KernelVirtualStart + g_KernelPaging.KernelSize)
|
if (Virtual < g_KernelPaging.KernelVirtualStart || Virtual > g_KernelPaging.KernelVirtualStart + g_KernelPaging.KernelSize)
|
||||||
{
|
{
|
||||||
|
@ -101,7 +101,7 @@ u8 paging_get_physical(u64 Virtual, u64 *Physical)
|
||||||
{
|
{
|
||||||
T_PAGING_ML4 *ml4 = paging_get_ml4();
|
T_PAGING_ML4 *ml4 = paging_get_ml4();
|
||||||
return paging_get_physical_ex(Virtual, Physical, ml4);
|
return paging_get_physical_ex(Virtual, Physical, ml4);
|
||||||
}
|
}*/
|
||||||
|
|
||||||
void paging_set_ml4(u64 ML4Physical)
|
void paging_set_ml4(u64 ML4Physical)
|
||||||
{
|
{
|
||||||
|
|
Loading…
Reference in New Issue