acrn-hypervisor/hypervisor/arch/x86/e820.c

321 lines
9.4 KiB
C

/*
* Copyright (C) 2018-2022 Intel Corporation.
*
* SPDX-License-Identifier: BSD-3-Clause
*/
#include <types.h>
#include <acrn_hv_defs.h>
#include <asm/page.h>
#include <asm/e820.h>
#include <asm/mmu.h>
#include <boot.h>
#include <efi_mmap.h>
#include <logmsg.h>
#include <asm/guest/ept.h>
/*
* e820.c contains the related e820 operations; like HV to get memory info for its MMU setup;
* and hide HV memory from Service VM...
*/
static uint32_t hv_e820_entries_nr;
static uint64_t hv_e820_ram_size;
/* Describe the memory layout the hypervisor uses */
static struct e820_entry hv_e820[E820_MAX_ENTRIES];
#define DBG_LEVEL_E820 6U
/*
* @brief reserve some RAM, hide it from Service VM, return its start address
*
* e820_alloc_memory requires 4k alignment, so size_arg will be converted
* in the function.
*
* @param size_arg Amount of memory to be found and marked reserved
* @param max_addr Maximum address below which memory is to be identified
*
* @pre hv_e820_entries_nr > 0U
* @return base address of the memory region
*/
uint64_t e820_alloc_memory(uint64_t size_arg, uint64_t max_addr)
{
int32_t i;
uint64_t size = round_page_up(size_arg);
uint64_t ret = INVALID_HPA;
struct e820_entry *entry, *new_entry;
for (i = (int32_t)hv_e820_entries_nr - 1; i >= 0; i--) {
entry = &hv_e820[i];
uint64_t start, end, length;
start = round_page_up(entry->baseaddr);
end = round_page_down(entry->baseaddr + entry->length);
length = (end > start) ? (end - start) : 0UL;
if ((entry->type == E820_TYPE_RAM) && (length >= size) && ((start + size) <= max_addr)) {
/* found exact size of e820 entry */
if (length == size) {
entry->type = E820_TYPE_RESERVED;
ret = start;
} else {
/*
* found entry with available memory larger than requested (length > size)
* Reserve memory if
* 1) hv_e820_entries_nr < E820_MAX_ENTRIES
* 2) if end of this "entry" is <= max_addr
* use memory from end of this e820 "entry".
*/
if ((hv_e820_entries_nr < E820_MAX_ENTRIES) && (end <= max_addr)) {
new_entry = &hv_e820[hv_e820_entries_nr];
new_entry->type = E820_TYPE_RESERVED;
new_entry->baseaddr = end - size;
new_entry->length = (entry->baseaddr + entry->length) - new_entry->baseaddr;
/* Shrink the existing entry and total available memory */
entry->length -= new_entry->length;
hv_e820_entries_nr++;
ret = new_entry->baseaddr;
}
}
if (ret != INVALID_HPA) {
break;
}
}
}
if ((ret == INVALID_HPA) || (ret == 0UL)) {
/* current memory allocation algorithm is to find the available address from the highest
* possible address below max_addr. if ret == 0, means all memory is used up and we have to
* put the resource at address 0, this is dangerous.
* Also ret == 0 would make code logic very complicated, since memcpy_s() doesn't support
* address 0 copy.
*/
panic("Requested memory from E820 cannot be reserved!!");
}
return ret;
}
static void insert_e820_entry(uint32_t index, uint64_t addr, uint64_t length, uint64_t type)
{
uint32_t i;
hv_e820_entries_nr++;
ASSERT(hv_e820_entries_nr <= E820_MAX_ENTRIES, "e820 entry overflow");
for (i = hv_e820_entries_nr - 1; i > index; i--) {
hv_e820[i] = hv_e820[i-1];
}
hv_e820[index].baseaddr = addr;
hv_e820[index].length = length;
hv_e820[index].type = type;
}
static void init_e820_from_efi_mmap(void)
{
uint32_t i, e820_idx = 0U;
const struct efi_memory_desc *efi_mmap_entry = get_efi_mmap_entry();
for (i = 0U; i < get_efi_mmap_entries_count(); i++) {
if (e820_idx >= E820_MAX_ENTRIES) {
pr_err("Too many efi memmap entries !");
break;
}
hv_e820[e820_idx].baseaddr = efi_mmap_entry[i].phys_addr;
hv_e820[e820_idx].length = efi_mmap_entry[i].num_pages * PAGE_SIZE;
/* The EFI BOOT Service releated regions need to be set to reserved and avoid being touched by
* hypervisor, because at least below software modules rely on them:
* 1. EFI ESRT(The EFI System Resource Table) which used for UEFI firmware upgrade;
* 2. Image resource in ACPI BGRT(Boottime Graphics Resource Table) which used for boot time logo;
*/
switch (efi_mmap_entry[i].type) {
case EFI_LOADER_CODE:
case EFI_LOADER_DATA:
case EFI_CONVENTIONAL_MEMORY:
if ((efi_mmap_entry[i].attribute & EFI_MEMORY_WB) != 0UL) {
hv_e820[e820_idx].type = E820_TYPE_RAM;
} else {
hv_e820[e820_idx].type = E820_TYPE_RESERVED;
}
break;
case EFI_UNUSABLE_MEMORY:
hv_e820[e820_idx].type = E820_TYPE_UNUSABLE;
break;
case EFI_ACPI_RECLAIM_MEMORY:
hv_e820[e820_idx].type = E820_TYPE_ACPI_RECLAIM;
break;
case EFI_ACPI_MEMORY_NVS:
hv_e820[e820_idx].type = E820_TYPE_ACPI_NVS;
break;
/* case EFI_RESERVED_MEMORYTYPE:
* case EFI_BOOT_SERVICES_CODE:
* case EFI_BOOT_SERVICES_DATA:
* case EFI_RUNTIME_SERVICES_CODE:
* case EFI_RUNTIME_SERVICES_DATA:
* case EFI_MEMORYMAPPED_IO:
* case EFI_MEMORYMAPPED_IOPORTSPACE:
* case EFI_PALCODE:
* case EFI_PERSISTENT_MEMORY:
*/
default:
hv_e820[e820_idx].type = E820_TYPE_RESERVED;
break;
}
/* Given the efi memmap has been sorted, the hv_e820[] is also sorted.
* Then the algorithm is very simple, just merge with previous mmap entry
* if type is same and base addr is continuous.
*/
if ((e820_idx > 0U) && (hv_e820[e820_idx].type == hv_e820[e820_idx - 1U].type)
&& (hv_e820[e820_idx].baseaddr ==
(hv_e820[e820_idx - 1U].baseaddr
+ hv_e820[e820_idx - 1U].length))) {
hv_e820[e820_idx - 1U].length += hv_e820[e820_idx].length;
} else {
dev_dbg(DBG_LEVEL_E820, "efi mmap hv_e820[%d]: type: 0x%x Base: 0x%016lx length: 0x%016lx",
e820_idx, hv_e820[e820_idx].type, hv_e820[e820_idx].baseaddr, hv_e820[e820_idx].length);
e820_idx ++;
}
}
hv_e820_entries_nr = e820_idx;
}
/* HV read multiboot header to get e820 entries info and calc total RAM info */
static void init_e820_from_mmap(struct acrn_boot_info *abi)
{
uint32_t i;
struct abi_mmap *mmap = abi->mmap_entry;
hv_e820_entries_nr = abi->mmap_entries;
dev_dbg(DBG_LEVEL_E820, "mmap addr 0x%x entries %d\n",
abi->mmap_entry, hv_e820_entries_nr);
for (i = 0U; i < hv_e820_entries_nr; i++) {
hv_e820[i].baseaddr = mmap[i].baseaddr;
hv_e820[i].length = mmap[i].length;
hv_e820[i].type = mmap[i].type;
dev_dbg(DBG_LEVEL_E820, "mmap hv_e820[%d]: type: 0x%x Base: 0x%016lx length: 0x%016lx", i,
mmap[i].type, mmap[i].baseaddr, mmap[i].length);
}
}
static void calculate_e820_ram_size(void)
{
uint32_t i;
for(i = 0; i < hv_e820_entries_nr; i++){
dev_dbg(DBG_LEVEL_E820, "hv_e820[%d]:type: 0x%x Base: 0x%016lx length: 0x%016lx", i,
hv_e820[i].type, hv_e820[i].baseaddr, hv_e820[i].length);
if (hv_e820[i].type == E820_TYPE_RAM) {
hv_e820_ram_size += hv_e820[i].baseaddr + hv_e820[i].length;
}
}
dev_dbg(DBG_LEVEL_E820, "ram size: 0x%016lx ",hv_e820_ram_size);
}
static void alloc_mods_memory(void)
{
uint32_t mod_index, e820_index, target_index;
uint64_t mod_start, mod_end;
uint64_t entry_start, entry_end;
struct acrn_boot_info *abi = get_acrn_boot_info();
/* 1st pass: remove the exact region */
for (mod_index = 0; mod_index < abi->mods_count; mod_index++) {
mod_start = hva2hpa(abi->mods[mod_index].start);
mod_end = hva2hpa(abi->mods[mod_index].start) + abi->mods[mod_index].size;
for (e820_index = 0; e820_index < hv_e820_entries_nr; e820_index++) {
entry_start = hv_e820[e820_index].baseaddr;
entry_end = hv_e820[e820_index].baseaddr + hv_e820[e820_index].length;
/* No need handle in these cases*/
if ((hv_e820[e820_index].type != E820_TYPE_RAM) || (entry_end <= mod_start) || (entry_start >= mod_end)) {
continue;
}
if ((entry_start <= mod_start) && (entry_end >= mod_end)) {
hv_e820[e820_index].length = mod_start - entry_start;
if (mod_end < entry_end) {
/*
* .......|start_pa... ....................end_pa|.....
* |entry_start..............................entry_end|
*/
insert_e820_entry(e820_index + 1, mod_end, entry_end - mod_end, E820_TYPE_RAM);
}
} else {
panic("%s: region 0x%016x-0x%016x crosses multiple e820 entries, check your bootloader!",
__func__, entry_start, entry_end);
}
}
}
/* 2nd pass: shrink the entries to page boundary */
target_index = 0;
for (e820_index = 0; e820_index < hv_e820_entries_nr; e820_index++) {
entry_start = round_page_up(hv_e820[e820_index].baseaddr);
entry_end = round_page_down(hv_e820[e820_index].baseaddr + hv_e820[e820_index].length);
if (entry_start < entry_end) {
hv_e820[target_index].baseaddr = entry_start;
hv_e820[target_index].length = entry_end - entry_start;
hv_e820[target_index].type = hv_e820[e820_index].type;
target_index++;
}
}
memset(&hv_e820[target_index], 0, (hv_e820_entries_nr - target_index) * sizeof(struct e820_entry));
hv_e820_entries_nr = target_index;
}
void init_e820(void)
{
struct acrn_boot_info *abi = get_acrn_boot_info();
if (boot_from_uefi(abi)) {
init_efi_mmap_entries(&abi->uefi_info);
init_e820_from_efi_mmap();
} else {
init_e820_from_mmap(abi);
}
calculate_e820_ram_size();
/* reserve multiboot modules memory */
alloc_mods_memory();
}
uint64_t get_e820_ram_size(void)
{
return hv_e820_ram_size;
}
uint32_t get_e820_entries_count(void)
{
return hv_e820_entries_nr;
}
const struct e820_entry *get_e820_entry(void)
{
return hv_e820;
}