mirror of
https://github.com/mii443/nel_os.git
synced 2025-08-22 16:15:38 +00:00
dynamic memory allocation
This commit is contained in:
@ -14,7 +14,7 @@ const ALIGN_2MIB_MASK: u64 = SIZE_2MIB - 1;
|
||||
pub struct BootInfoFrameAllocator {
|
||||
memory_map: &'static MemoryMap,
|
||||
next: usize,
|
||||
aligned_regions: [(u64, u64); 32],
|
||||
aligned_regions: [(u64, u64); 256],
|
||||
aligned_count: usize,
|
||||
}
|
||||
|
||||
@ -23,7 +23,7 @@ impl BootInfoFrameAllocator {
|
||||
let mut allocator = Self {
|
||||
memory_map,
|
||||
next: 0,
|
||||
aligned_regions: [(0, 0); 32],
|
||||
aligned_regions: [(0, 0); 256],
|
||||
aligned_count: 0,
|
||||
};
|
||||
|
||||
|
117
src/vmm/ept.rs
117
src/vmm/ept.rs
@ -107,32 +107,135 @@ impl EPT {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub fn map_4k(
|
||||
&mut self,
|
||||
gpa: u64,
|
||||
hpa: u64,
|
||||
allocator: &mut impl FrameAllocator<Size4KiB>,
|
||||
) -> Result<(), &'static str> {
|
||||
let lv4_index = (gpa >> 39) & 0x1FF;
|
||||
let lv3_index = (gpa >> 30) & 0x1FF;
|
||||
let lv2_index = (gpa >> 21) & 0x1FF;
|
||||
let lv1_index = (gpa >> 12) & 0x1FF;
|
||||
|
||||
let lv4_table = Self::frame_to_table_ptr(&self.root_table);
|
||||
let lv4_entry = &mut lv4_table[lv4_index as usize];
|
||||
|
||||
let lv3_table = if !lv4_entry.present() {
|
||||
let frame = allocator
|
||||
.allocate_frame()
|
||||
.ok_or("Failed to allocate frame for L3 table")?;
|
||||
let table_ptr = Self::init_table(&frame);
|
||||
lv4_entry.set_phys(frame.start_address().as_u64() >> 12);
|
||||
lv4_entry.set_map_memory(false);
|
||||
lv4_entry.set_typ(0);
|
||||
lv4_entry.set_read(true);
|
||||
lv4_entry.set_write(true);
|
||||
lv4_entry.set_exec_super(true);
|
||||
table_ptr
|
||||
} else {
|
||||
let frame =
|
||||
PhysFrame::from_start_address(PhysAddr::new(lv4_entry.phys() << 12)).unwrap();
|
||||
Self::frame_to_table_ptr(&frame)
|
||||
};
|
||||
|
||||
let lv3_entry = &mut lv3_table[lv3_index as usize];
|
||||
|
||||
let lv2_table = if !lv3_entry.present() {
|
||||
let frame = allocator
|
||||
.allocate_frame()
|
||||
.ok_or("Failed to allocate frame for L2 table")?;
|
||||
let table_ptr = Self::init_table(&frame);
|
||||
lv3_entry.set_phys(frame.start_address().as_u64() >> 12);
|
||||
lv3_entry.set_map_memory(false);
|
||||
lv3_entry.set_typ(0);
|
||||
lv3_entry.set_read(true);
|
||||
lv3_entry.set_write(true);
|
||||
lv3_entry.set_exec_super(true);
|
||||
table_ptr
|
||||
} else {
|
||||
let frame =
|
||||
PhysFrame::from_start_address(PhysAddr::new(lv3_entry.phys() << 12)).unwrap();
|
||||
Self::frame_to_table_ptr(&frame)
|
||||
};
|
||||
|
||||
let lv2_entry = &mut lv2_table[lv2_index as usize];
|
||||
|
||||
let lv1_table = if !lv2_entry.present() || lv2_entry.map_memory() {
|
||||
let frame = allocator
|
||||
.allocate_frame()
|
||||
.ok_or("Failed to allocate frame for L1 table")?;
|
||||
let table_ptr = Self::init_table(&frame);
|
||||
lv2_entry.set_phys(frame.start_address().as_u64() >> 12);
|
||||
lv2_entry.set_map_memory(false);
|
||||
lv2_entry.set_typ(0);
|
||||
lv2_entry.set_read(true);
|
||||
lv2_entry.set_write(true);
|
||||
lv2_entry.set_exec_super(true);
|
||||
table_ptr
|
||||
} else {
|
||||
let frame =
|
||||
PhysFrame::from_start_address(PhysAddr::new(lv2_entry.phys() << 12)).unwrap();
|
||||
Self::frame_to_table_ptr(&frame)
|
||||
};
|
||||
|
||||
let lv1_entry = &mut lv1_table[lv1_index as usize];
|
||||
lv1_entry.set_phys(hpa >> 12);
|
||||
lv1_entry.set_map_memory(true);
|
||||
lv1_entry.set_typ(0);
|
||||
lv1_entry.set_read(true);
|
||||
lv1_entry.set_write(true);
|
||||
lv1_entry.set_exec_super(true);
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub fn get_phys_addr(&self, gpa: u64) -> Option<u64> {
|
||||
let lv4_index = (gpa >> 39) & 0x1FF;
|
||||
let lv3_index = (gpa >> 30) & 0x1FF;
|
||||
let lv2_index = (gpa >> 21) & 0x1FF;
|
||||
let page_offset = gpa & 0x1FFFFF;
|
||||
let lv1_index = (gpa >> 12) & 0x1FF;
|
||||
|
||||
let lv4_table = Self::frame_to_table_ptr(&self.root_table);
|
||||
let lv4_entry = &lv4_table[lv4_index as usize];
|
||||
|
||||
if !lv4_entry.present() {
|
||||
return None;
|
||||
}
|
||||
|
||||
let frame = PhysFrame::from_start_address(PhysAddr::new(lv4_entry.phys() << 12)).unwrap();
|
||||
let lv3_table = Self::frame_to_table_ptr(&frame);
|
||||
let lv3_entry = &lv3_table[lv3_index as usize];
|
||||
|
||||
if !lv3_entry.present() {
|
||||
return None;
|
||||
}
|
||||
|
||||
let frame = PhysFrame::from_start_address(PhysAddr::new(lv3_entry.phys() << 12)).unwrap();
|
||||
let lv2_table = Self::frame_to_table_ptr(&frame);
|
||||
let lv2_entry = &lv2_table[lv2_index as usize];
|
||||
|
||||
if !lv2_entry.map_memory() {
|
||||
info!("EPT: No mapping found for GPA: {:#x}", gpa);
|
||||
info!("{:#x}", lv2_entry.address().as_u64());
|
||||
info!("{:#x}", lv2_entry as *const _ as u64);
|
||||
if !lv2_entry.present() {
|
||||
return None;
|
||||
}
|
||||
|
||||
let phys_addr_base = lv2_entry.address().as_u64();
|
||||
Some(phys_addr_base | page_offset)
|
||||
if lv2_entry.map_memory() {
|
||||
let page_offset = gpa & 0x1FFFFF;
|
||||
let phys_addr_base = lv2_entry.address().as_u64();
|
||||
Some(phys_addr_base | page_offset)
|
||||
} else {
|
||||
let frame = PhysFrame::from_start_address(PhysAddr::new(lv2_entry.phys() << 12)).unwrap();
|
||||
let lv1_table = Self::frame_to_table_ptr(&frame);
|
||||
let lv1_entry = &lv1_table[lv1_index as usize];
|
||||
|
||||
if !lv1_entry.present() || !lv1_entry.map_memory() {
|
||||
return None;
|
||||
}
|
||||
|
||||
let page_offset = gpa & 0xFFF;
|
||||
let phys_addr_base = lv1_entry.address().as_u64();
|
||||
Some(phys_addr_base | page_offset)
|
||||
}
|
||||
}
|
||||
|
||||
pub fn get(&mut self, gpa: u64) -> Result<u8, &'static str> {
|
||||
|
@ -6,7 +6,7 @@ pub const INITRD: &'static [u8] = include_bytes!("../../rootfs-n.cpio.gz");
|
||||
pub const LAYOUT_BOOTPARAM: u64 = 0x0001_0000;
|
||||
pub const LAYOUT_CMDLINE: u64 = 0x0002_0000;
|
||||
pub const LAYOUT_KERNEL_BASE: u64 = 0x0010_0000;
|
||||
pub const LAYOUT_INITRD: u64 = 0x0600_0000;
|
||||
pub const LAYOUT_INITRD: u64 = 0x0800_0000;
|
||||
|
||||
#[repr(C, packed)]
|
||||
#[derive(Debug, Clone, Copy)]
|
||||
|
@ -132,6 +132,7 @@ impl ShadowMsr {
|
||||
Self::set_ret_val(vcpu, unsafe { vmread(vmcs::guest::GS_BASE).unwrap() })
|
||||
}
|
||||
x86::msr::IA32_KERNEL_GSBASE => Self::shadow_read(vcpu, msr_kind),
|
||||
0x1b => Self::shadow_read(vcpu, msr_kind),
|
||||
_ => {
|
||||
panic!("Unhandled RDMSR: {:#x}", msr_kind);
|
||||
}
|
||||
@ -170,6 +171,7 @@ impl ShadowMsr {
|
||||
x86::msr::IA32_EFER => unsafe { vmwrite(vmcs::guest::IA32_EFER_FULL, value).unwrap() },
|
||||
x86::msr::IA32_FS_BASE => unsafe { vmwrite(vmcs::guest::FS_BASE, value).unwrap() },
|
||||
x86::msr::IA32_GS_BASE => unsafe { vmwrite(vmcs::guest::GS_BASE, value).unwrap() },
|
||||
0x1b => Self::shadow_write(vcpu, msr_kind),
|
||||
|
||||
_ => {
|
||||
panic!("Unhandled WRMSR: {:#x}", msr_kind);
|
||||
|
144
src/vmm/vcpu.rs
144
src/vmm/vcpu.rs
@ -2,6 +2,7 @@ use core::{
|
||||
arch::asm,
|
||||
arch::x86_64::{_xgetbv, _xsetbv},
|
||||
convert::TryInto,
|
||||
sync::atomic::{AtomicPtr, Ordering},
|
||||
u64, u8,
|
||||
};
|
||||
|
||||
@ -47,6 +48,10 @@ use super::{
|
||||
vmxon::Vmxon,
|
||||
};
|
||||
|
||||
const SIZE_2MIB: u64 = 2 * 1024 * 1024;
|
||||
|
||||
static EPT_FRAME_ALLOCATOR: AtomicPtr<BootInfoFrameAllocator> = AtomicPtr::new(core::ptr::null_mut());
|
||||
|
||||
#[repr(C)]
|
||||
pub struct VCpu {
|
||||
pub guest_registers: GuestRegisters,
|
||||
@ -137,7 +142,6 @@ impl VCpu {
|
||||
Ok(page_base | page_offset)
|
||||
}
|
||||
|
||||
/// Read 8 bytes from guest physical address
|
||||
fn read_guest_phys_u64(&mut self, gpa: u64) -> Result<u64, &'static str> {
|
||||
let mut result_bytes = [0u8; 8];
|
||||
|
||||
@ -188,6 +192,8 @@ impl VCpu {
|
||||
frame_allocator: &mut BootInfoFrameAllocator,
|
||||
mapper: &OffsetPageTable<'static>,
|
||||
) {
|
||||
EPT_FRAME_ALLOCATOR.store(frame_allocator as *mut _, Ordering::Release);
|
||||
|
||||
self.vmxon.activate_vmxon().unwrap();
|
||||
|
||||
let revision_id = unsafe { rdmsr(x86::msr::IA32_VMX_BASIC) } as u32;
|
||||
@ -200,13 +206,12 @@ impl VCpu {
|
||||
self.setup_host_state().unwrap();
|
||||
self.setup_guest_state().unwrap();
|
||||
self.setup_io_bitmaps();
|
||||
self.setup_guest_memory(frame_allocator);
|
||||
let _ = self.setup_guest_memory(frame_allocator);
|
||||
self.register_msrs(&mapper);
|
||||
}
|
||||
|
||||
pub fn load_kernel(&mut self, kernel: &[u8]) {
|
||||
pub fn load_kernel(&mut self, kernel: &[u8], guest_mem_size: u64) {
|
||||
info!("Loading kernel into guest memory");
|
||||
let guest_mem_size = 100 * 1024 * 1024;
|
||||
let mut bp = BootParams::from_bytes(kernel).unwrap();
|
||||
bp.e820_entries = 0;
|
||||
|
||||
@ -236,12 +241,9 @@ impl VCpu {
|
||||
|
||||
let cmdline_start = linux::LAYOUT_CMDLINE as u64;
|
||||
let cmdline_end = cmdline_start + cmdline_max_size as u64;
|
||||
self.ept.set_range(cmdline_start, cmdline_end, 0).unwrap();
|
||||
let cmdline_val = "console=ttyS0 earlyprintk=serial nokaslr";
|
||||
let cmdline_bytes = cmdline_val.as_bytes();
|
||||
for (i, &byte) in cmdline_bytes.iter().enumerate() {
|
||||
self.ept.set(cmdline_start + i as u64, byte).unwrap();
|
||||
}
|
||||
|
||||
let cmdline_bytes = b"console=ttyS0 earlyprintk=serial nokaslr\0";
|
||||
self.load_image(cmdline_bytes, cmdline_start as usize);
|
||||
|
||||
let bp_bytes = unsafe {
|
||||
core::slice::from_raw_parts(
|
||||
@ -257,39 +259,61 @@ impl VCpu {
|
||||
&kernel[code_offset..code_offset + code_size],
|
||||
linux::LAYOUT_KERNEL_BASE as usize,
|
||||
);
|
||||
|
||||
info!(
|
||||
"Loading initrd at {:#x}, size: {} bytes",
|
||||
linux::LAYOUT_INITRD,
|
||||
linux::INITRD.len()
|
||||
);
|
||||
self.load_image(linux::INITRD, linux::LAYOUT_INITRD as usize);
|
||||
|
||||
info!("Kernel loaded into guest memory");
|
||||
}
|
||||
|
||||
pub fn load_image(&mut self, image: &[u8], addr: usize) {
|
||||
info!("Loading image at {:#x}, size: {} bytes", addr, image.len());
|
||||
|
||||
let start_page = addr & !0xFFF;
|
||||
let end_page = ((addr + image.len() - 1) & !0xFFF) + 0x1000;
|
||||
|
||||
unsafe {
|
||||
let frame_allocator_ptr = EPT_FRAME_ALLOCATOR.load(Ordering::Acquire);
|
||||
if !frame_allocator_ptr.is_null() {
|
||||
let frame_allocator = &mut *(frame_allocator_ptr as *mut BootInfoFrameAllocator);
|
||||
|
||||
let mut current_page = start_page;
|
||||
while current_page < end_page {
|
||||
if self.ept.get_phys_addr(current_page as u64).is_none() {
|
||||
if let Some(frame) = frame_allocator.allocate_frame() {
|
||||
let hpa = frame.start_address().as_u64();
|
||||
self.ept.map_4k(current_page as u64, hpa, frame_allocator).unwrap();
|
||||
} else {
|
||||
panic!("Failed to allocate frame for image at {:#x}", current_page);
|
||||
}
|
||||
}
|
||||
current_page += 0x1000;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
for (i, &byte) in image.iter().enumerate() {
|
||||
let gpa = addr + i;
|
||||
self.ept.set(gpa as u64, byte).unwrap();
|
||||
}
|
||||
}
|
||||
|
||||
pub fn setup_guest_memory(&mut self, frame_allocator: &mut BootInfoFrameAllocator) {
|
||||
let mut pages = 1000;
|
||||
let mut gpa = 0;
|
||||
pub fn setup_guest_memory(&mut self, frame_allocator: &mut BootInfoFrameAllocator) -> u64 {
|
||||
let guest_memory_size = 2 * 1024 * 1024 * 1024;
|
||||
|
||||
info!("Setting up guest memory...");
|
||||
while pages > 0 {
|
||||
let frame = frame_allocator
|
||||
.allocate_2mib_frame()
|
||||
.expect("Failed to allocate frame");
|
||||
let hpa = frame.start_address().as_u64();
|
||||
info!("Setting up guest memory with on-demand allocation (reported size: {}MB)",
|
||||
guest_memory_size / (1024 * 1024));
|
||||
|
||||
self.ept.map_2m(gpa, hpa, frame_allocator).unwrap();
|
||||
gpa += (4 * 1024) << 9;
|
||||
pages -= 1;
|
||||
}
|
||||
info!("Guest memory setup complete");
|
||||
|
||||
self.load_kernel(linux::BZIMAGE);
|
||||
self.load_kernel(linux::BZIMAGE, guest_memory_size);
|
||||
|
||||
let eptp = EPTP::new(&self.ept.root_table);
|
||||
unsafe { vmwrite(vmcs::control::EPTP_FULL, eptp.0).unwrap() };
|
||||
|
||||
guest_memory_size
|
||||
}
|
||||
|
||||
pub fn register_msrs(&mut self, mapper: &OffsetPageTable<'static>) {
|
||||
@ -329,6 +353,7 @@ impl VCpu {
|
||||
self.guest_msr
|
||||
.set(x86::msr::MSR_C5_PMON_BOX_CTRL, 0)
|
||||
.unwrap();
|
||||
self.guest_msr.set(0x1b, 0).unwrap();
|
||||
|
||||
vmwrite(
|
||||
vmcs::control::VMEXIT_MSR_LOAD_ADDR_FULL,
|
||||
@ -617,13 +642,18 @@ impl VCpu {
|
||||
info!("Setting up guest state");
|
||||
|
||||
unsafe {
|
||||
let cr0 = (Cr0::empty() | Cr0::CR0_PROTECTED_MODE | Cr0::CR0_NUMERIC_ERROR)
|
||||
let cr0 = (Cr0::empty()
|
||||
| Cr0::CR0_PROTECTED_MODE
|
||||
| Cr0::CR0_NUMERIC_ERROR
|
||||
| Cr0::CR0_EXTENSION_TYPE)
|
||||
& !Cr0::CR0_ENABLE_PAGING;
|
||||
vmwrite(vmcs::guest::CR0, cr0.bits() as u64)?;
|
||||
vmwrite(vmcs::guest::CR3, cr3())?;
|
||||
vmwrite(vmcs::guest::CR3, 0)?;
|
||||
vmwrite(
|
||||
vmcs::guest::CR4,
|
||||
vmread(vmcs::guest::CR4)? | Cr4Flags::VIRTUAL_MACHINE_EXTENSIONS.bits(),
|
||||
vmread(vmcs::guest::CR4)?
|
||||
| Cr4Flags::VIRTUAL_MACHINE_EXTENSIONS.bits()
|
||||
& !Cr4Flags::PHYSICAL_ADDRESS_EXTENSION.bits(),
|
||||
)?;
|
||||
|
||||
vmwrite(vmcs::guest::CS_BASE, 0)?;
|
||||
@ -949,6 +979,34 @@ impl VCpu {
|
||||
}
|
||||
}
|
||||
|
||||
fn handle_ept_violation(&mut self, gpa: u64) {
|
||||
if gpa >= 2 * 1024 * 1024 * 1024 {
|
||||
panic!("EPT Violation: Guest tried to access memory beyond 2GB at {:#x}", gpa);
|
||||
}
|
||||
|
||||
unsafe {
|
||||
let frame_allocator_ptr = EPT_FRAME_ALLOCATOR.load(Ordering::Acquire);
|
||||
if frame_allocator_ptr.is_null() {
|
||||
panic!("EPT Violation: Frame allocator not initialized!");
|
||||
}
|
||||
|
||||
let frame_allocator = &mut *(frame_allocator_ptr as *mut BootInfoFrameAllocator);
|
||||
|
||||
match frame_allocator.allocate_frame() {
|
||||
Some(frame) => {
|
||||
let hpa = frame.start_address().as_u64();
|
||||
|
||||
if let Err(e) = self.ept.map_4k(gpa, hpa, frame_allocator) {
|
||||
panic!("Failed to map page at GPA {:#x}: {}", gpa, e);
|
||||
}
|
||||
}
|
||||
None => {
|
||||
panic!("EPT Violation: Out of memory! Cannot allocate frame for GPA {:#x}", gpa);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn vmexit_handler(&mut self) {
|
||||
let exit_reason_raw = unsafe { vmread(vmcs::ro::EXIT_REASON).unwrap() as u32 };
|
||||
|
||||
@ -1037,7 +1095,6 @@ impl VCpu {
|
||||
self.step_next_inst().unwrap();
|
||||
}
|
||||
VmxExitReason::EXCEPTION => {
|
||||
|
||||
// Get exception information
|
||||
let vmexit_intr_info =
|
||||
unsafe { vmread(vmcs::ro::VMEXIT_INTERRUPTION_INFO).unwrap() };
|
||||
@ -1096,19 +1153,23 @@ impl VCpu {
|
||||
0x01 => match instruction_bytes[2] {
|
||||
0xCA => {
|
||||
self.step_next_inst().unwrap();
|
||||
},
|
||||
}
|
||||
0xCB => {
|
||||
self.step_next_inst().unwrap();
|
||||
},
|
||||
}
|
||||
_ => {
|
||||
self.inject_exception(vector, error_code).unwrap();
|
||||
}
|
||||
},
|
||||
_ => {self.inject_exception(vector, error_code).unwrap();},
|
||||
_ => {
|
||||
self.inject_exception(vector, error_code).unwrap();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
_ => {self.inject_exception(vector, error_code).unwrap();},
|
||||
_ => {
|
||||
self.inject_exception(vector, error_code).unwrap();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1135,8 +1196,19 @@ impl VCpu {
|
||||
VmxExitReason::EPT_VIOLATION => {
|
||||
let guest_address =
|
||||
unsafe { vmread(vmcs::ro::GUEST_PHYSICAL_ADDR_FULL).unwrap() };
|
||||
info!("EPT Violation at address: {:#x}", guest_address);
|
||||
self.step_next_inst().unwrap();
|
||||
let exit_qualification =
|
||||
unsafe { vmread(vmcs::ro::EXIT_QUALIFICATION).unwrap() };
|
||||
let guest_rip = unsafe { vmread(vmcs::guest::RIP).unwrap() };
|
||||
|
||||
let read_access = (exit_qualification & 0x1) != 0;
|
||||
let write_access = (exit_qualification & 0x2) != 0;
|
||||
let execute_access = (exit_qualification & 0x4) != 0;
|
||||
let gpa_valid = (exit_qualification & 0x80) != 0;
|
||||
let translation_valid = (exit_qualification & 0x100) != 0;
|
||||
|
||||
let page_addr = guest_address & !0xFFF;
|
||||
|
||||
self.handle_ept_violation(page_addr);
|
||||
}
|
||||
_ => {
|
||||
panic!("VMExit reason: {:?}", exit_reason);
|
||||
|
Reference in New Issue
Block a user