197 lines
6.6 KiB
C++
197 lines
6.6 KiB
C++
//===- Win32/Memory.cpp - Win32 Memory Implementation -----------*- C++ -*-===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file provides the Win32 specific implementation of various Memory
|
|
// management utilities
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "llvm/Support/DataTypes.h"
|
|
#include "llvm/Support/ErrorHandling.h"
|
|
#include "llvm/Support/Process.h"
|
|
#include "llvm/Support/WindowsError.h"
|
|
|
|
// The Windows.h header must be the last one included.
|
|
#include "llvm/Support/Windows/WindowsSupport.h"
|
|
|
|
static DWORD getWindowsProtectionFlags(unsigned Flags) {
|
|
switch (Flags & llvm::sys::Memory::MF_RWE_MASK) {
|
|
// Contrary to what you might expect, the Windows page protection flags
|
|
// are not a bitwise combination of RWX values
|
|
case llvm::sys::Memory::MF_READ:
|
|
return PAGE_READONLY;
|
|
case llvm::sys::Memory::MF_WRITE:
|
|
// Note: PAGE_WRITE is not supported by VirtualProtect
|
|
return PAGE_READWRITE;
|
|
case llvm::sys::Memory::MF_READ|llvm::sys::Memory::MF_WRITE:
|
|
return PAGE_READWRITE;
|
|
case llvm::sys::Memory::MF_READ|llvm::sys::Memory::MF_EXEC:
|
|
return PAGE_EXECUTE_READ;
|
|
case llvm::sys::Memory::MF_READ |
|
|
llvm::sys::Memory::MF_WRITE |
|
|
llvm::sys::Memory::MF_EXEC:
|
|
return PAGE_EXECUTE_READWRITE;
|
|
case llvm::sys::Memory::MF_EXEC:
|
|
return PAGE_EXECUTE;
|
|
default:
|
|
llvm_unreachable("Illegal memory protection flag specified!");
|
|
}
|
|
// Provide a default return value as required by some compilers.
|
|
return PAGE_NOACCESS;
|
|
}
|
|
|
|
// While we'd be happy to allocate single pages, the Windows allocation
|
|
// granularity may be larger than a single page (in practice, it is 64K)
|
|
// so mapping less than that will create an unreachable fragment of memory.
|
|
static size_t getAllocationGranularity() {
|
|
SYSTEM_INFO Info;
|
|
::GetSystemInfo(&Info);
|
|
if (Info.dwPageSize > Info.dwAllocationGranularity)
|
|
return Info.dwPageSize;
|
|
else
|
|
return Info.dwAllocationGranularity;
|
|
}
|
|
|
|
// Large/huge memory pages need explicit process permissions in order to be
|
|
// used. See https://blogs.msdn.microsoft.com/oldnewthing/20110128-00/?p=11643
|
|
// Also large pages need to be manually enabled on your OS. If all this is
|
|
// sucessfull, we return the minimal large memory page size.
|
|
static size_t enableProcessLargePages() {
|
|
HANDLE Token = 0;
|
|
size_t LargePageMin = GetLargePageMinimum();
|
|
if (LargePageMin)
|
|
OpenProcessToken(GetCurrentProcess(), TOKEN_ADJUST_PRIVILEGES | TOKEN_QUERY,
|
|
&Token);
|
|
if (!Token)
|
|
return 0;
|
|
LUID Luid;
|
|
if (!LookupPrivilegeValue(0, SE_LOCK_MEMORY_NAME, &Luid)) {
|
|
CloseHandle(Token);
|
|
return 0;
|
|
}
|
|
TOKEN_PRIVILEGES TP{};
|
|
TP.PrivilegeCount = 1;
|
|
TP.Privileges[0].Luid = Luid;
|
|
TP.Privileges[0].Attributes = SE_PRIVILEGE_ENABLED;
|
|
if (!AdjustTokenPrivileges(Token, FALSE, &TP, 0, 0, 0)) {
|
|
CloseHandle(Token);
|
|
return 0;
|
|
}
|
|
DWORD E = GetLastError();
|
|
CloseHandle(Token);
|
|
if (E == ERROR_SUCCESS)
|
|
return LargePageMin;
|
|
return 0;
|
|
}
|
|
|
|
namespace llvm {
|
|
namespace sys {
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
//=== WARNING: Implementation here must contain only Win32 specific code
|
|
//=== and must not be UNIX code
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
MemoryBlock Memory::allocateMappedMemory(size_t NumBytes,
|
|
const MemoryBlock *const NearBlock,
|
|
unsigned Flags,
|
|
std::error_code &EC) {
|
|
EC = std::error_code();
|
|
if (NumBytes == 0)
|
|
return MemoryBlock();
|
|
|
|
static size_t DefaultGranularity = getAllocationGranularity();
|
|
static size_t LargePageGranularity = enableProcessLargePages();
|
|
|
|
DWORD AllocType = MEM_RESERVE | MEM_COMMIT;
|
|
bool HugePages = false;
|
|
size_t Granularity = DefaultGranularity;
|
|
|
|
if ((Flags & MF_HUGE_HINT) && LargePageGranularity > 0) {
|
|
AllocType |= MEM_LARGE_PAGES;
|
|
HugePages = true;
|
|
Granularity = LargePageGranularity;
|
|
}
|
|
|
|
size_t NumBlocks = (NumBytes + Granularity - 1) / Granularity;
|
|
|
|
uintptr_t Start = NearBlock ? reinterpret_cast<uintptr_t>(NearBlock->base()) +
|
|
NearBlock->allocatedSize()
|
|
: 0;
|
|
|
|
// If the requested address is not aligned to the allocation granularity,
|
|
// round up to get beyond NearBlock. VirtualAlloc would have rounded down.
|
|
if (Start && Start % Granularity != 0)
|
|
Start += Granularity - Start % Granularity;
|
|
|
|
DWORD Protect = getWindowsProtectionFlags(Flags);
|
|
|
|
size_t AllocSize = NumBlocks * Granularity;
|
|
void *PA = ::VirtualAlloc(reinterpret_cast<void *>(Start),
|
|
AllocSize, AllocType, Protect);
|
|
if (PA == NULL) {
|
|
if (NearBlock || HugePages) {
|
|
// Try again without the NearBlock hint and without large memory pages
|
|
return allocateMappedMemory(NumBytes, NULL, Flags & ~MF_HUGE_HINT, EC);
|
|
}
|
|
EC = mapWindowsError(::GetLastError());
|
|
return MemoryBlock();
|
|
}
|
|
|
|
MemoryBlock Result;
|
|
Result.Address = PA;
|
|
Result.AllocatedSize = AllocSize;
|
|
Result.Flags = (Flags & ~MF_HUGE_HINT) | (HugePages ? MF_HUGE_HINT : 0);
|
|
|
|
if (Flags & MF_EXEC)
|
|
Memory::InvalidateInstructionCache(Result.Address, AllocSize);
|
|
|
|
return Result;
|
|
}
|
|
|
|
std::error_code Memory::releaseMappedMemory(MemoryBlock &M) {
|
|
if (M.Address == 0 || M.AllocatedSize == 0)
|
|
return std::error_code();
|
|
|
|
if (!VirtualFree(M.Address, 0, MEM_RELEASE))
|
|
return mapWindowsError(::GetLastError());
|
|
|
|
M.Address = 0;
|
|
M.AllocatedSize = 0;
|
|
|
|
return std::error_code();
|
|
}
|
|
|
|
std::error_code Memory::protectMappedMemory(const MemoryBlock &M,
|
|
unsigned Flags) {
|
|
if (M.Address == 0 || M.AllocatedSize == 0)
|
|
return std::error_code();
|
|
|
|
DWORD Protect = getWindowsProtectionFlags(Flags);
|
|
|
|
DWORD OldFlags;
|
|
if (!VirtualProtect(M.Address, M.AllocatedSize, Protect, &OldFlags))
|
|
return mapWindowsError(::GetLastError());
|
|
|
|
if (Flags & MF_EXEC)
|
|
Memory::InvalidateInstructionCache(M.Address, M.AllocatedSize);
|
|
|
|
return std::error_code();
|
|
}
|
|
|
|
/// InvalidateInstructionCache - Before the JIT can run a block of code
|
|
/// that has been emitted it must invalidate the instruction cache on some
|
|
/// platforms.
|
|
void Memory::InvalidateInstructionCache(
|
|
const void *Addr, size_t Len) {
|
|
FlushInstructionCache(GetCurrentProcess(), Addr, Len);
|
|
}
|
|
|
|
} // namespace sys
|
|
} // namespace llvm
|