mirror of
https://github.com/DragonOS-Community/DragonOS.git
synced 2025-06-08 14:16:47 +00:00
1884 lines
62 KiB
Rust
1884 lines
62 KiB
Rust
// 进程的用户空间内存管理
|
||
|
||
use core::{
|
||
cmp,
|
||
hash::Hasher,
|
||
intrinsics::unlikely,
|
||
ops::Add,
|
||
sync::atomic::{compiler_fence, Ordering},
|
||
};
|
||
|
||
use alloc::{
|
||
collections::BTreeMap,
|
||
sync::{Arc, Weak},
|
||
vec::Vec,
|
||
};
|
||
use hashbrown::HashSet;
|
||
use ida::IdAllocator;
|
||
use system_error::SystemError;
|
||
|
||
use crate::{
|
||
arch::{mm::PageMapper, CurrentIrqArch, MMArch},
|
||
exception::InterruptArch,
|
||
filesystem::vfs::file::File,
|
||
ipc::shm::{shm_manager_lock, ShmFlags},
|
||
libs::{
|
||
align::page_align_up,
|
||
rwlock::RwLock,
|
||
spinlock::{SpinLock, SpinLockGuard},
|
||
},
|
||
mm::page::page_manager_lock_irqsave,
|
||
process::ProcessManager,
|
||
syscall::user_access::{UserBufferReader, UserBufferWriter},
|
||
};
|
||
|
||
use super::{
|
||
allocator::page_frame::{
|
||
deallocate_page_frames, PageFrameCount, PhysPageFrame, VirtPageFrame, VirtPageFrameIter,
|
||
},
|
||
page::{EntryFlags, Flusher, InactiveFlusher, PageFlushAll, PageType},
|
||
syscall::{MadvFlags, MapFlags, MremapFlags, ProtFlags},
|
||
MemoryManagementArch, PageTableKind, VirtAddr, VirtRegion, VmFlags,
|
||
};
|
||
|
||
/// MMAP_MIN_ADDR的默认值
|
||
/// 以下内容来自linux-5.19:
|
||
/// This is the portion of low virtual memory which should be protected
|
||
// from userspace allocation. Keeping a user from writing to low pages
|
||
// can help reduce the impact of kernel NULL pointer bugs.
|
||
// For most ia64, ppc64 and x86 users with lots of address space
|
||
// a value of 65536 is reasonable and should cause no problems.
|
||
// On arm and other archs it should not be higher than 32768.
|
||
// Programs which use vm86 functionality or have some need to map
|
||
// this low address space will need CAP_SYS_RAWIO or disable this
|
||
// protection by setting the value to 0.
|
||
pub const DEFAULT_MMAP_MIN_ADDR: usize = 65536;
|
||
|
||
/// LockedVMA的id分配器
|
||
static LOCKEDVMA_ID_ALLOCATOR: SpinLock<IdAllocator> =
|
||
SpinLock::new(IdAllocator::new(0, usize::MAX).unwrap());
|
||
|
||
#[derive(Debug)]
|
||
pub struct AddressSpace {
|
||
inner: RwLock<InnerAddressSpace>,
|
||
}
|
||
|
||
impl AddressSpace {
|
||
pub fn new(create_stack: bool) -> Result<Arc<Self>, SystemError> {
|
||
let inner = InnerAddressSpace::new(create_stack)?;
|
||
let result = Self {
|
||
inner: RwLock::new(inner),
|
||
};
|
||
return Ok(Arc::new(result));
|
||
}
|
||
|
||
/// 从pcb中获取当前进程的地址空间结构体的Arc指针
|
||
pub fn current() -> Result<Arc<AddressSpace>, SystemError> {
|
||
let vm = ProcessManager::current_pcb()
|
||
.basic()
|
||
.user_vm()
|
||
.expect("Current process has no address space");
|
||
|
||
return Ok(vm);
|
||
}
|
||
|
||
/// 判断某个地址空间是否为当前进程的地址空间
|
||
pub fn is_current(self: &Arc<Self>) -> bool {
|
||
let current = Self::current();
|
||
if let Ok(current) = current {
|
||
return Arc::ptr_eq(¤t, self);
|
||
}
|
||
return false;
|
||
}
|
||
}
|
||
|
||
impl core::ops::Deref for AddressSpace {
|
||
type Target = RwLock<InnerAddressSpace>;
|
||
|
||
fn deref(&self) -> &Self::Target {
|
||
&self.inner
|
||
}
|
||
}
|
||
|
||
impl core::ops::DerefMut for AddressSpace {
|
||
fn deref_mut(&mut self) -> &mut Self::Target {
|
||
&mut self.inner
|
||
}
|
||
}
|
||
|
||
/// @brief 用户地址空间结构体(每个进程都有一个)
|
||
#[derive(Debug)]
|
||
pub struct InnerAddressSpace {
|
||
pub user_mapper: UserMapper,
|
||
pub mappings: UserMappings,
|
||
pub mmap_min: VirtAddr,
|
||
/// 用户栈信息结构体
|
||
pub user_stack: Option<UserStack>,
|
||
|
||
pub elf_brk_start: VirtAddr,
|
||
pub elf_brk: VirtAddr,
|
||
|
||
/// 当前进程的堆空间的起始地址
|
||
pub brk_start: VirtAddr,
|
||
/// 当前进程的堆空间的结束地址(不包含)
|
||
pub brk: VirtAddr,
|
||
|
||
pub start_code: VirtAddr,
|
||
pub end_code: VirtAddr,
|
||
pub start_data: VirtAddr,
|
||
pub end_data: VirtAddr,
|
||
}
|
||
|
||
impl InnerAddressSpace {
|
||
pub fn new(create_stack: bool) -> Result<Self, SystemError> {
|
||
let mut result = Self {
|
||
user_mapper: MMArch::setup_new_usermapper()?,
|
||
mappings: UserMappings::new(),
|
||
mmap_min: VirtAddr(DEFAULT_MMAP_MIN_ADDR),
|
||
elf_brk_start: VirtAddr::new(0),
|
||
elf_brk: VirtAddr::new(0),
|
||
brk_start: MMArch::USER_BRK_START,
|
||
brk: MMArch::USER_BRK_START,
|
||
user_stack: None,
|
||
start_code: VirtAddr(0),
|
||
end_code: VirtAddr(0),
|
||
start_data: VirtAddr(0),
|
||
end_data: VirtAddr(0),
|
||
};
|
||
if create_stack {
|
||
// debug!("to create user stack.");
|
||
result.new_user_stack(UserStack::DEFAULT_USER_STACK_SIZE)?;
|
||
}
|
||
|
||
return Ok(result);
|
||
}
|
||
|
||
/// 尝试克隆当前进程的地址空间,包括这些映射都会被克隆
|
||
///
|
||
/// # Returns
|
||
///
|
||
/// 返回克隆后的,新的地址空间的Arc指针
|
||
#[inline(never)]
|
||
pub fn try_clone(&mut self) -> Result<Arc<AddressSpace>, SystemError> {
|
||
let irq_guard = unsafe { CurrentIrqArch::save_and_disable_irq() };
|
||
let new_addr_space = AddressSpace::new(false)?;
|
||
let mut new_guard = new_addr_space.write();
|
||
unsafe {
|
||
new_guard
|
||
.user_mapper
|
||
.clone_from(&mut self.user_mapper, MMArch::PAGE_FAULT_ENABLED)
|
||
};
|
||
|
||
// 拷贝用户栈的结构体信息,但是不拷贝用户栈的内容(因为后面VMA的拷贝会拷贝用户栈的内容)
|
||
unsafe {
|
||
new_guard.user_stack = Some(self.user_stack.as_ref().unwrap().clone_info_only());
|
||
}
|
||
let _current_stack_size = self.user_stack.as_ref().unwrap().stack_size();
|
||
|
||
// 拷贝空洞
|
||
new_guard.mappings.vm_holes = self.mappings.vm_holes.clone();
|
||
|
||
for vma in self.mappings.vmas.iter() {
|
||
// TODO: 增加对VMA是否为文件映射的判断,如果是的话,就跳过
|
||
|
||
let vma_guard: SpinLockGuard<'_, VMA> = vma.lock_irqsave();
|
||
|
||
// 仅拷贝VMA信息并添加反向映射,因为UserMapper克隆时已经分配了新的物理页
|
||
let new_vma = LockedVMA::new(vma_guard.clone_info_only());
|
||
new_guard.mappings.vmas.insert(new_vma.clone());
|
||
// debug!("new vma: {:x?}", new_vma);
|
||
let new_vma_guard = new_vma.lock_irqsave();
|
||
let new_mapper = &new_guard.user_mapper.utable;
|
||
let mut page_manager_guard = page_manager_lock_irqsave();
|
||
for page in new_vma_guard.pages().map(|p| p.virt_address()) {
|
||
if let Some((paddr, _)) = new_mapper.translate(page) {
|
||
let page = page_manager_guard.get_unwrap(&paddr);
|
||
page.write_irqsave().insert_vma(new_vma.clone());
|
||
}
|
||
}
|
||
|
||
drop(page_manager_guard);
|
||
drop(vma_guard);
|
||
drop(new_vma_guard);
|
||
}
|
||
drop(new_guard);
|
||
drop(irq_guard);
|
||
return Ok(new_addr_space);
|
||
}
|
||
|
||
/// 拓展用户栈
|
||
/// ## 参数
|
||
///
|
||
/// - `bytes`: 拓展大小
|
||
#[allow(dead_code)]
|
||
pub fn extend_stack(&mut self, mut bytes: usize) -> Result<(), SystemError> {
|
||
// debug!("extend user stack");
|
||
let prot_flags = ProtFlags::PROT_READ | ProtFlags::PROT_WRITE | ProtFlags::PROT_EXEC;
|
||
let map_flags = MapFlags::MAP_PRIVATE | MapFlags::MAP_ANONYMOUS | MapFlags::MAP_GROWSDOWN;
|
||
let stack = self.user_stack.as_mut().unwrap();
|
||
|
||
bytes = page_align_up(bytes);
|
||
stack.mapped_size += bytes;
|
||
let len = stack.stack_bottom - stack.mapped_size;
|
||
self.map_anonymous(len, bytes, prot_flags, map_flags, false, false)?;
|
||
return Ok(());
|
||
}
|
||
|
||
/// 判断当前的地址空间是否是当前进程的地址空间
|
||
#[inline]
|
||
pub fn is_current(&self) -> bool {
|
||
return self.user_mapper.utable.is_current();
|
||
}
|
||
|
||
/// 进行匿名页映射
|
||
///
|
||
/// ## 参数
|
||
///
|
||
/// - `start_vaddr`:映射的起始地址
|
||
/// - `len`:映射的长度
|
||
/// - `prot_flags`:保护标志
|
||
/// - `map_flags`:映射标志
|
||
/// - `round_to_min`:是否将`start_vaddr`对齐到`mmap_min`,如果为`true`,则当`start_vaddr`不为0时,会对齐到`mmap_min`,否则仅向下对齐到页边界
|
||
/// - `allocate_at_once`:是否立即分配物理空间
|
||
///
|
||
/// ## 返回
|
||
///
|
||
/// 返回映射的起始虚拟页帧
|
||
pub fn map_anonymous(
|
||
&mut self,
|
||
start_vaddr: VirtAddr,
|
||
len: usize,
|
||
prot_flags: ProtFlags,
|
||
map_flags: MapFlags,
|
||
round_to_min: bool,
|
||
allocate_at_once: bool,
|
||
) -> Result<VirtPageFrame, SystemError> {
|
||
let allocate_at_once = if MMArch::PAGE_FAULT_ENABLED {
|
||
allocate_at_once
|
||
} else {
|
||
true
|
||
};
|
||
// 用于对齐hint的函数
|
||
let round_hint_to_min = |hint: VirtAddr| {
|
||
// 先把hint向下对齐到页边界
|
||
let addr = hint.data() & (!MMArch::PAGE_OFFSET_MASK);
|
||
// debug!("map_anonymous: hint = {:?}, addr = {addr:#x}", hint);
|
||
// 如果hint不是0,且hint小于DEFAULT_MMAP_MIN_ADDR,则对齐到DEFAULT_MMAP_MIN_ADDR
|
||
if (addr != 0) && round_to_min && (addr < DEFAULT_MMAP_MIN_ADDR) {
|
||
Some(VirtAddr::new(page_align_up(DEFAULT_MMAP_MIN_ADDR)))
|
||
} else if addr == 0 {
|
||
None
|
||
} else {
|
||
Some(VirtAddr::new(addr))
|
||
}
|
||
};
|
||
// debug!("map_anonymous: start_vaddr = {:?}", start_vaddr);
|
||
// debug!("map_anonymous: len(no align) = {}", len);
|
||
|
||
let len = page_align_up(len);
|
||
|
||
// debug!("map_anonymous: len = {}", len);
|
||
|
||
let start_page: VirtPageFrame = self.mmap(
|
||
round_hint_to_min(start_vaddr),
|
||
PageFrameCount::from_bytes(len).unwrap(),
|
||
prot_flags,
|
||
map_flags,
|
||
move |page, count, vm_flags, flags, mapper, flusher| {
|
||
if allocate_at_once {
|
||
VMA::zeroed(page, count, vm_flags, flags, mapper, flusher, None, None)
|
||
} else {
|
||
Ok(LockedVMA::new(VMA::new(
|
||
VirtRegion::new(page.virt_address(), count.data() * MMArch::PAGE_SIZE),
|
||
vm_flags,
|
||
flags,
|
||
None,
|
||
None,
|
||
false,
|
||
)))
|
||
}
|
||
},
|
||
)?;
|
||
|
||
return Ok(start_page);
|
||
}
|
||
|
||
/// 进行文件页映射
|
||
///
|
||
/// ## 参数
|
||
///
|
||
/// - `start_vaddr`:映射的起始地址
|
||
/// - `len`:映射的长度
|
||
/// - `prot_flags`:保护标志
|
||
/// - `map_flags`:映射标志
|
||
/// - `fd`:文件描述符
|
||
/// - `offset`:映射偏移量
|
||
/// - `round_to_min`:是否将`start_vaddr`对齐到`mmap_min`,如果为`true`,则当`start_vaddr`不为0时,会对齐到`mmap_min`,否则仅向下对齐到页边界
|
||
/// - `allocate_at_once`:是否立即分配物理空间
|
||
///
|
||
/// ## 返回
|
||
///
|
||
/// 返回映射的起始虚拟页帧
|
||
#[allow(clippy::too_many_arguments)]
|
||
pub fn file_mapping(
|
||
&mut self,
|
||
start_vaddr: VirtAddr,
|
||
len: usize,
|
||
prot_flags: ProtFlags,
|
||
map_flags: MapFlags,
|
||
fd: i32,
|
||
offset: usize,
|
||
round_to_min: bool,
|
||
allocate_at_once: bool,
|
||
) -> Result<VirtPageFrame, SystemError> {
|
||
let allocate_at_once = if MMArch::PAGE_FAULT_ENABLED {
|
||
allocate_at_once
|
||
} else {
|
||
true
|
||
};
|
||
// 用于对齐hint的函数
|
||
let round_hint_to_min = |hint: VirtAddr| {
|
||
// 先把hint向下对齐到页边界
|
||
let addr = hint.data() & (!MMArch::PAGE_OFFSET_MASK);
|
||
// debug!("map_anonymous: hint = {:?}, addr = {addr:#x}", hint);
|
||
// 如果hint不是0,且hint小于DEFAULT_MMAP_MIN_ADDR,则对齐到DEFAULT_MMAP_MIN_ADDR
|
||
if (addr != 0) && round_to_min && (addr < DEFAULT_MMAP_MIN_ADDR) {
|
||
Some(VirtAddr::new(page_align_up(DEFAULT_MMAP_MIN_ADDR)))
|
||
} else if addr == 0 {
|
||
None
|
||
} else {
|
||
Some(VirtAddr::new(addr))
|
||
}
|
||
};
|
||
// debug!("map_anonymous: start_vaddr = {:?}", start_vaddr);
|
||
// debug!("map_anonymous: len(no align) = {}", len);
|
||
|
||
let len = page_align_up(len);
|
||
|
||
// debug!("map_anonymous: len = {}", len);
|
||
|
||
let binding = ProcessManager::current_pcb().fd_table();
|
||
let fd_table_guard = binding.read();
|
||
|
||
let file = fd_table_guard.get_file_by_fd(fd);
|
||
if file.is_none() {
|
||
return Err(SystemError::EBADF);
|
||
}
|
||
// drop guard 以避免无法调度的问题
|
||
drop(fd_table_guard);
|
||
|
||
// offset需要4K对齐
|
||
if !offset & (MMArch::PAGE_SIZE - 1) == 0 {
|
||
return Err(SystemError::EINVAL);
|
||
}
|
||
let pgoff = offset >> MMArch::PAGE_SHIFT;
|
||
|
||
let start_page: VirtPageFrame = self.mmap(
|
||
round_hint_to_min(start_vaddr),
|
||
PageFrameCount::from_bytes(len).unwrap(),
|
||
prot_flags,
|
||
map_flags,
|
||
|page, count, vm_flags, flags, mapper, flusher| {
|
||
if allocate_at_once {
|
||
VMA::zeroed(
|
||
page,
|
||
count,
|
||
vm_flags,
|
||
flags,
|
||
mapper,
|
||
flusher,
|
||
file.clone(),
|
||
Some(pgoff),
|
||
)
|
||
} else {
|
||
Ok(LockedVMA::new(VMA::new(
|
||
VirtRegion::new(page.virt_address(), count.data() * MMArch::PAGE_SIZE),
|
||
vm_flags,
|
||
flags,
|
||
file.clone(),
|
||
Some(pgoff),
|
||
false,
|
||
)))
|
||
}
|
||
},
|
||
)?;
|
||
// todo!(impl mmap for other file)
|
||
// https://github.com/DragonOS-Community/DragonOS/pull/912#discussion_r1765334272
|
||
let file = file.unwrap();
|
||
let _ = file.inode().mmap(start_vaddr.data(), len, offset);
|
||
return Ok(start_page);
|
||
}
|
||
|
||
/// 向进程的地址空间映射页面
|
||
///
|
||
/// # 参数
|
||
///
|
||
/// - `addr`:映射的起始地址,如果为`None`,则由内核自动分配
|
||
/// - `page_count`:映射的页面数量
|
||
/// - `prot_flags`:保护标志
|
||
/// - `map_flags`:映射标志
|
||
/// - `map_func`:映射函数,用于创建VMA
|
||
///
|
||
/// # Returns
|
||
///
|
||
/// 返回映射的起始虚拟页帧
|
||
///
|
||
/// # Errors
|
||
///
|
||
/// - `EINVAL`:参数错误
|
||
pub fn mmap<
|
||
F: FnOnce(
|
||
VirtPageFrame,
|
||
PageFrameCount,
|
||
VmFlags,
|
||
EntryFlags<MMArch>,
|
||
&mut PageMapper,
|
||
&mut dyn Flusher<MMArch>,
|
||
) -> Result<Arc<LockedVMA>, SystemError>,
|
||
>(
|
||
&mut self,
|
||
addr: Option<VirtAddr>,
|
||
page_count: PageFrameCount,
|
||
prot_flags: ProtFlags,
|
||
map_flags: MapFlags,
|
||
map_func: F,
|
||
) -> Result<VirtPageFrame, SystemError> {
|
||
if page_count == PageFrameCount::new(0) {
|
||
return Err(SystemError::EINVAL);
|
||
}
|
||
// debug!("mmap: addr: {addr:?}, page_count: {page_count:?}, prot_flags: {prot_flags:?}, map_flags: {map_flags:?}");
|
||
|
||
// 找到未使用的区域
|
||
let region = match addr {
|
||
Some(vaddr) => {
|
||
self.mappings
|
||
.find_free_at(self.mmap_min, vaddr, page_count.bytes(), map_flags)?
|
||
}
|
||
None => self
|
||
.mappings
|
||
.find_free(self.mmap_min, page_count.bytes())
|
||
.ok_or(SystemError::ENOMEM)?,
|
||
};
|
||
|
||
let page = VirtPageFrame::new(region.start());
|
||
|
||
let vm_flags = VmFlags::from(prot_flags)
|
||
| VmFlags::from(map_flags)
|
||
| VmFlags::VM_MAYREAD
|
||
| VmFlags::VM_MAYWRITE
|
||
| VmFlags::VM_MAYEXEC;
|
||
|
||
// debug!("mmap: page: {:?}, region={region:?}", page.virt_address());
|
||
|
||
compiler_fence(Ordering::SeqCst);
|
||
let (mut active, mut inactive);
|
||
let flusher = if self.is_current() {
|
||
active = PageFlushAll::new();
|
||
&mut active as &mut dyn Flusher<MMArch>
|
||
} else {
|
||
inactive = InactiveFlusher::new();
|
||
&mut inactive as &mut dyn Flusher<MMArch>
|
||
};
|
||
compiler_fence(Ordering::SeqCst);
|
||
// 映射页面,并将VMA插入到地址空间的VMA列表中
|
||
self.mappings.insert_vma(map_func(
|
||
page,
|
||
page_count,
|
||
vm_flags,
|
||
EntryFlags::from_prot_flags(prot_flags, true),
|
||
&mut self.user_mapper.utable,
|
||
flusher,
|
||
)?);
|
||
|
||
return Ok(page);
|
||
}
|
||
|
||
/// 重映射内存区域
|
||
///
|
||
/// # 参数
|
||
///
|
||
/// - `old_vaddr`:原映射的起始地址
|
||
/// - `old_len`:原映射的长度
|
||
/// - `new_len`:重新映射的长度
|
||
/// - `mremap_flags`:重映射标志
|
||
/// - `new_vaddr`:重新映射的起始地址
|
||
/// - `vm_flags`:旧内存区域标志
|
||
///
|
||
/// # Returns
|
||
///
|
||
/// 返回重映射的起始虚拟页帧地址
|
||
///
|
||
/// # Errors
|
||
///
|
||
/// - `EINVAL`:参数错误
|
||
pub fn mremap(
|
||
&mut self,
|
||
old_vaddr: VirtAddr,
|
||
old_len: usize,
|
||
new_len: usize,
|
||
mremap_flags: MremapFlags,
|
||
new_vaddr: VirtAddr,
|
||
vm_flags: VmFlags,
|
||
) -> Result<VirtAddr, SystemError> {
|
||
// 检查新内存地址是否对齐
|
||
if !new_vaddr.check_aligned(MMArch::PAGE_SIZE) {
|
||
return Err(SystemError::EINVAL);
|
||
}
|
||
|
||
// 检查新、旧内存区域是否冲突
|
||
let old_region = VirtRegion::new(old_vaddr, old_len);
|
||
let new_region = VirtRegion::new(new_vaddr, new_len);
|
||
if old_region.collide(&new_region) {
|
||
return Err(SystemError::EINVAL);
|
||
}
|
||
|
||
// 初始化映射标志
|
||
let mut map_flags: MapFlags = vm_flags.into();
|
||
// 初始化内存区域保护标志
|
||
let prot_flags: ProtFlags = vm_flags.into();
|
||
|
||
// 取消新内存区域的原映射
|
||
if mremap_flags.contains(MremapFlags::MREMAP_FIXED) {
|
||
map_flags |= MapFlags::MAP_FIXED;
|
||
let start_page = VirtPageFrame::new(new_vaddr);
|
||
let page_count = PageFrameCount::from_bytes(new_len).unwrap();
|
||
self.munmap(start_page, page_count)?;
|
||
}
|
||
|
||
// 获取映射后的新内存页面
|
||
let new_page = self.map_anonymous(new_vaddr, new_len, prot_flags, map_flags, true, true)?;
|
||
let new_page_vaddr = new_page.virt_address();
|
||
|
||
// 拷贝旧内存区域内容到新内存区域
|
||
let old_buffer_reader =
|
||
UserBufferReader::new(old_vaddr.data() as *const u8, old_len, true)?;
|
||
let old_buf: &[u8] = old_buffer_reader.read_from_user(0)?;
|
||
let mut new_buffer_writer =
|
||
UserBufferWriter::new(new_page_vaddr.data() as *mut u8, new_len, true)?;
|
||
let new_buf: &mut [u8] = new_buffer_writer.buffer(0)?;
|
||
let len = old_buf.len().min(new_buf.len());
|
||
new_buf[..len].copy_from_slice(&old_buf[..len]);
|
||
|
||
return Ok(new_page_vaddr);
|
||
}
|
||
|
||
/// 取消进程的地址空间中的映射
|
||
///
|
||
/// # 参数
|
||
///
|
||
/// - `start_page`:起始页帧
|
||
/// - `page_count`:取消映射的页帧数量
|
||
///
|
||
/// # Errors
|
||
///
|
||
/// - `EINVAL`:参数错误
|
||
/// - `ENOMEM`:内存不足
|
||
pub fn munmap(
|
||
&mut self,
|
||
start_page: VirtPageFrame,
|
||
page_count: PageFrameCount,
|
||
) -> Result<(), SystemError> {
|
||
let to_unmap = VirtRegion::new(start_page.virt_address(), page_count.bytes());
|
||
let mut flusher: PageFlushAll<MMArch> = PageFlushAll::new();
|
||
|
||
let regions: Vec<Arc<LockedVMA>> = self.mappings.conflicts(to_unmap).collect::<Vec<_>>();
|
||
|
||
for r in regions {
|
||
let r = r.lock_irqsave().region;
|
||
let r = self.mappings.remove_vma(&r).unwrap();
|
||
let intersection = r.lock_irqsave().region().intersect(&to_unmap).unwrap();
|
||
let split_result = r.extract(intersection, &self.user_mapper.utable).unwrap();
|
||
|
||
// TODO: 当引入后备页映射后,这里需要增加通知文件的逻辑
|
||
|
||
if let Some(before) = split_result.prev {
|
||
// 如果前面有VMA,则需要将前面的VMA重新插入到地址空间的VMA列表中
|
||
self.mappings.insert_vma(before);
|
||
}
|
||
|
||
if let Some(after) = split_result.after {
|
||
// 如果后面有VMA,则需要将后面的VMA重新插入到地址空间的VMA列表中
|
||
self.mappings.insert_vma(after);
|
||
}
|
||
|
||
r.unmap(&mut self.user_mapper.utable, &mut flusher);
|
||
}
|
||
|
||
// TODO: 当引入后备页映射后,这里需要增加通知文件的逻辑
|
||
|
||
return Ok(());
|
||
}
|
||
|
||
pub fn mprotect(
|
||
&mut self,
|
||
start_page: VirtPageFrame,
|
||
page_count: PageFrameCount,
|
||
prot_flags: ProtFlags,
|
||
) -> Result<(), SystemError> {
|
||
// debug!(
|
||
// "mprotect: start_page: {:?}, page_count: {:?}, prot_flags:{prot_flags:?}",
|
||
// start_page,
|
||
// page_count
|
||
// );
|
||
let (mut active, mut inactive);
|
||
let flusher = if self.is_current() {
|
||
active = PageFlushAll::new();
|
||
&mut active as &mut dyn Flusher<MMArch>
|
||
} else {
|
||
inactive = InactiveFlusher::new();
|
||
&mut inactive as &mut dyn Flusher<MMArch>
|
||
};
|
||
|
||
let mapper = &mut self.user_mapper.utable;
|
||
let region = VirtRegion::new(start_page.virt_address(), page_count.bytes());
|
||
// debug!("mprotect: region: {:?}", region);
|
||
|
||
let regions = self.mappings.conflicts(region).collect::<Vec<_>>();
|
||
// debug!("mprotect: regions: {:?}", regions);
|
||
|
||
for r in regions {
|
||
// debug!("mprotect: r: {:?}", r);
|
||
let r = *r.lock_irqsave().region();
|
||
let r = self.mappings.remove_vma(&r).unwrap();
|
||
|
||
let intersection = r.lock_irqsave().region().intersect(®ion).unwrap();
|
||
let split_result = r
|
||
.extract(intersection, mapper)
|
||
.expect("Failed to extract VMA");
|
||
|
||
if let Some(before) = split_result.prev {
|
||
self.mappings.insert_vma(before);
|
||
}
|
||
if let Some(after) = split_result.after {
|
||
self.mappings.insert_vma(after);
|
||
}
|
||
|
||
let mut r_guard = r.lock_irqsave();
|
||
// 如果VMA的保护标志不允许指定的修改,则返回错误
|
||
if !r_guard.can_have_flags(prot_flags) {
|
||
drop(r_guard);
|
||
self.mappings.insert_vma(r.clone());
|
||
return Err(SystemError::EACCES);
|
||
}
|
||
r_guard.set_vm_flags(VmFlags::from(prot_flags));
|
||
|
||
let new_flags: EntryFlags<MMArch> = r_guard
|
||
.flags()
|
||
.set_execute(prot_flags.contains(ProtFlags::PROT_EXEC))
|
||
.set_write(prot_flags.contains(ProtFlags::PROT_WRITE));
|
||
|
||
r_guard.remap(new_flags, mapper, &mut *flusher)?;
|
||
drop(r_guard);
|
||
self.mappings.insert_vma(r);
|
||
}
|
||
|
||
return Ok(());
|
||
}
|
||
|
||
pub fn madvise(
|
||
&mut self,
|
||
start_page: VirtPageFrame,
|
||
page_count: PageFrameCount,
|
||
behavior: MadvFlags,
|
||
) -> Result<(), SystemError> {
|
||
let (mut active, mut inactive);
|
||
let flusher = if self.is_current() {
|
||
active = PageFlushAll::new();
|
||
&mut active as &mut dyn Flusher<MMArch>
|
||
} else {
|
||
inactive = InactiveFlusher::new();
|
||
&mut inactive as &mut dyn Flusher<MMArch>
|
||
};
|
||
|
||
let mapper = &mut self.user_mapper.utable;
|
||
|
||
let region = VirtRegion::new(start_page.virt_address(), page_count.bytes());
|
||
let regions = self.mappings.conflicts(region).collect::<Vec<_>>();
|
||
|
||
for r in regions {
|
||
let r = *r.lock_irqsave().region();
|
||
let r = self.mappings.remove_vma(&r).unwrap();
|
||
|
||
let intersection = r.lock_irqsave().region().intersect(®ion).unwrap();
|
||
let split_result = r
|
||
.extract(intersection, mapper)
|
||
.expect("Failed to extract VMA");
|
||
|
||
if let Some(before) = split_result.prev {
|
||
self.mappings.insert_vma(before);
|
||
}
|
||
if let Some(after) = split_result.after {
|
||
self.mappings.insert_vma(after);
|
||
}
|
||
r.do_madvise(behavior, mapper, &mut *flusher)?;
|
||
self.mappings.insert_vma(r);
|
||
}
|
||
Ok(())
|
||
}
|
||
|
||
/// 创建新的用户栈
|
||
///
|
||
/// ## 参数
|
||
///
|
||
/// - `size`:栈的大小
|
||
pub fn new_user_stack(&mut self, size: usize) -> Result<(), SystemError> {
|
||
assert!(self.user_stack.is_none(), "User stack already exists");
|
||
let stack = UserStack::new(self, None, size)?;
|
||
self.user_stack = Some(stack);
|
||
return Ok(());
|
||
}
|
||
|
||
#[inline(always)]
|
||
pub fn user_stack_mut(&mut self) -> Option<&mut UserStack> {
|
||
return self.user_stack.as_mut();
|
||
}
|
||
|
||
/// 取消用户空间内的所有映射
|
||
pub unsafe fn unmap_all(&mut self) {
|
||
let mut flusher: PageFlushAll<MMArch> = PageFlushAll::new();
|
||
for vma in self.mappings.iter_vmas() {
|
||
if vma.mapped() {
|
||
vma.unmap(&mut self.user_mapper.utable, &mut flusher);
|
||
}
|
||
}
|
||
}
|
||
|
||
/// 设置进程的堆的内存空间
|
||
///
|
||
/// ## 参数
|
||
///
|
||
/// - `new_brk`:新的堆的结束地址。需要满足页对齐要求,并且是用户空间地址,且大于等于当前的堆的起始地址
|
||
///
|
||
/// ## 返回值
|
||
///
|
||
/// 返回旧的堆的结束地址
|
||
pub unsafe fn set_brk(&mut self, new_brk: VirtAddr) -> Result<VirtAddr, SystemError> {
|
||
assert!(new_brk.check_aligned(MMArch::PAGE_SIZE));
|
||
|
||
if !new_brk.check_user() || new_brk < self.brk_start {
|
||
return Err(SystemError::EFAULT);
|
||
}
|
||
|
||
let old_brk = self.brk;
|
||
|
||
if new_brk > self.brk {
|
||
let len = new_brk - self.brk;
|
||
let prot_flags = ProtFlags::PROT_READ | ProtFlags::PROT_WRITE | ProtFlags::PROT_EXEC;
|
||
let map_flags = MapFlags::MAP_PRIVATE | MapFlags::MAP_ANONYMOUS | MapFlags::MAP_FIXED;
|
||
self.map_anonymous(old_brk, len, prot_flags, map_flags, true, false)?;
|
||
|
||
self.brk = new_brk;
|
||
return Ok(old_brk);
|
||
} else {
|
||
let unmap_len = self.brk - new_brk;
|
||
let unmap_start = new_brk;
|
||
if unmap_len == 0 {
|
||
return Ok(old_brk);
|
||
}
|
||
self.munmap(
|
||
VirtPageFrame::new(unmap_start),
|
||
PageFrameCount::from_bytes(unmap_len).unwrap(),
|
||
)?;
|
||
self.brk = new_brk;
|
||
return Ok(old_brk);
|
||
}
|
||
}
|
||
|
||
pub unsafe fn sbrk(&mut self, incr: isize) -> Result<VirtAddr, SystemError> {
|
||
if incr == 0 {
|
||
return Ok(self.brk);
|
||
}
|
||
|
||
let new_brk = if incr > 0 {
|
||
self.brk + incr as usize
|
||
} else {
|
||
self.brk - incr.unsigned_abs()
|
||
};
|
||
|
||
let new_brk = VirtAddr::new(page_align_up(new_brk.data()));
|
||
|
||
return self.set_brk(new_brk);
|
||
}
|
||
}
|
||
|
||
impl Drop for InnerAddressSpace {
|
||
fn drop(&mut self) {
|
||
unsafe {
|
||
self.unmap_all();
|
||
}
|
||
}
|
||
}
|
||
|
||
#[derive(Debug, Hash)]
|
||
pub struct UserMapper {
|
||
pub utable: PageMapper,
|
||
}
|
||
|
||
impl UserMapper {
|
||
pub fn new(utable: PageMapper) -> Self {
|
||
return Self { utable };
|
||
}
|
||
|
||
/// 拷贝用户空间映射
|
||
/// ## 参数
|
||
///
|
||
/// - `umapper`: 要拷贝的用户空间
|
||
/// - `copy_on_write`: 是否写时复制
|
||
pub unsafe fn clone_from(&mut self, umapper: &mut Self, copy_on_write: bool) {
|
||
self.utable
|
||
.clone_user_mapping(&mut umapper.utable, copy_on_write);
|
||
}
|
||
}
|
||
|
||
impl Drop for UserMapper {
|
||
fn drop(&mut self) {
|
||
if self.utable.is_current() {
|
||
// 如果当前要被销毁的用户空间的页表是当前进程的页表,那么就切换回初始内核页表
|
||
unsafe { MMArch::set_table(PageTableKind::User, MMArch::initial_page_table()) }
|
||
}
|
||
// 释放用户空间顶层页表占用的页帧
|
||
// 请注意,在释放这个页帧之前,用户页表应该已经被完全释放,否则会产生内存泄露
|
||
unsafe {
|
||
deallocate_page_frames(
|
||
PhysPageFrame::new(self.utable.table().phys()),
|
||
PageFrameCount::new(1),
|
||
)
|
||
};
|
||
}
|
||
}
|
||
|
||
/// 用户空间映射信息
|
||
#[derive(Debug)]
|
||
pub struct UserMappings {
|
||
/// 当前用户空间的虚拟内存区域
|
||
vmas: HashSet<Arc<LockedVMA>>,
|
||
/// 当前用户空间的VMA空洞
|
||
vm_holes: BTreeMap<VirtAddr, usize>,
|
||
}
|
||
|
||
impl UserMappings {
|
||
pub fn new() -> Self {
|
||
return Self {
|
||
vmas: HashSet::new(),
|
||
vm_holes: core::iter::once((VirtAddr::new(0), MMArch::USER_END_VADDR.data()))
|
||
.collect::<BTreeMap<_, _>>(),
|
||
};
|
||
}
|
||
|
||
/// 判断当前进程的VMA内,是否有包含指定的虚拟地址的VMA。
|
||
///
|
||
/// 如果有,返回包含指定虚拟地址的VMA的Arc指针,否则返回None。
|
||
#[allow(dead_code)]
|
||
pub fn contains(&self, vaddr: VirtAddr) -> Option<Arc<LockedVMA>> {
|
||
for v in self.vmas.iter() {
|
||
let guard = v.lock_irqsave();
|
||
if guard.region.contains(vaddr) {
|
||
return Some(v.clone());
|
||
}
|
||
}
|
||
return None;
|
||
}
|
||
|
||
/// 向下寻找距离虚拟地址最近的VMA
|
||
/// ## 参数
|
||
///
|
||
/// - `vaddr`: 虚拟地址
|
||
///
|
||
/// ## 返回值
|
||
/// - Some(Arc<LockedVMA>): 虚拟地址所在的或最近的下一个VMA
|
||
/// - None: 未找到VMA
|
||
#[allow(dead_code)]
|
||
pub fn find_nearest(&self, vaddr: VirtAddr) -> Option<Arc<LockedVMA>> {
|
||
let mut nearest: Option<Arc<LockedVMA>> = None;
|
||
for v in self.vmas.iter() {
|
||
let guard = v.lock_irqsave();
|
||
if guard.region.contains(vaddr) {
|
||
return Some(v.clone());
|
||
}
|
||
if guard.region.start >= vaddr
|
||
&& if let Some(ref nearest) = nearest {
|
||
guard.region.start < nearest.lock_irqsave().region.start
|
||
} else {
|
||
true
|
||
}
|
||
{
|
||
nearest = Some(v.clone());
|
||
}
|
||
}
|
||
return nearest;
|
||
}
|
||
|
||
/// 获取当前进程的地址空间中,与给定虚拟地址范围有重叠的VMA的迭代器。
|
||
pub fn conflicts(&self, request: VirtRegion) -> impl Iterator<Item = Arc<LockedVMA>> + '_ {
|
||
let r = self
|
||
.vmas
|
||
.iter()
|
||
.filter(move |v| v.lock_irqsave().region.intersect(&request).is_some())
|
||
.cloned();
|
||
return r;
|
||
}
|
||
|
||
/// 在当前进程的地址空间中,寻找第一个符合条件的空闲的虚拟内存范围。
|
||
///
|
||
/// @param min_vaddr 最小的起始地址
|
||
/// @param size 请求的大小
|
||
///
|
||
/// @return 如果找到了,返回虚拟内存范围,否则返回None
|
||
pub fn find_free(&self, min_vaddr: VirtAddr, size: usize) -> Option<VirtRegion> {
|
||
let _vaddr = min_vaddr;
|
||
let mut iter = self
|
||
.vm_holes
|
||
.iter()
|
||
.skip_while(|(hole_vaddr, hole_size)| hole_vaddr.add(**hole_size) <= min_vaddr);
|
||
|
||
let (hole_vaddr, size) = iter.find(|(hole_vaddr, hole_size)| {
|
||
// 计算当前空洞的可用大小
|
||
let available_size: usize =
|
||
if hole_vaddr <= &&min_vaddr && min_vaddr <= hole_vaddr.add(**hole_size) {
|
||
**hole_size - (min_vaddr - **hole_vaddr)
|
||
} else {
|
||
**hole_size
|
||
};
|
||
|
||
size <= available_size
|
||
})?;
|
||
|
||
// 创建一个新的虚拟内存范围。
|
||
let region = VirtRegion::new(cmp::max(*hole_vaddr, min_vaddr), *size);
|
||
|
||
return Some(region);
|
||
}
|
||
|
||
pub fn find_free_at(
|
||
&self,
|
||
min_vaddr: VirtAddr,
|
||
vaddr: VirtAddr,
|
||
size: usize,
|
||
flags: MapFlags,
|
||
) -> Result<VirtRegion, SystemError> {
|
||
// 如果没有指定地址,那么就在当前进程的地址空间中寻找一个空闲的虚拟内存范围。
|
||
if vaddr == VirtAddr::new(0) {
|
||
return self.find_free(min_vaddr, size).ok_or(SystemError::ENOMEM);
|
||
}
|
||
|
||
// 如果指定了地址,那么就检查指定的地址是否可用。
|
||
|
||
let requested = VirtRegion::new(vaddr, size);
|
||
|
||
if requested.end() >= MMArch::USER_END_VADDR || !vaddr.check_aligned(MMArch::PAGE_SIZE) {
|
||
return Err(SystemError::EINVAL);
|
||
}
|
||
|
||
if let Some(_x) = self.conflicts(requested).next() {
|
||
if flags.contains(MapFlags::MAP_FIXED_NOREPLACE) {
|
||
// 如果指定了 MAP_FIXED_NOREPLACE 标志,由于所指定的地址无法成功建立映射,则放弃映射,不对地址做修正
|
||
return Err(SystemError::EEXIST);
|
||
}
|
||
|
||
if flags.contains(MapFlags::MAP_FIXED) {
|
||
// todo: 支持MAP_FIXED标志对已有的VMA进行覆盖
|
||
return Err(SystemError::ENOSYS);
|
||
}
|
||
|
||
// 如果没有指定MAP_FIXED标志,那么就对地址做修正
|
||
let requested = self.find_free(min_vaddr, size).ok_or(SystemError::ENOMEM)?;
|
||
return Ok(requested);
|
||
}
|
||
|
||
return Ok(requested);
|
||
}
|
||
|
||
/// 在当前进程的地址空间中,保留一个指定大小的区域,使得该区域不在空洞中。
|
||
/// 该函数会修改vm_holes中的空洞信息。
|
||
///
|
||
/// @param region 要保留的区域
|
||
///
|
||
/// 请注意,在调用本函数之前,必须先确定region所在范围内没有VMA。
|
||
fn reserve_hole(&mut self, region: &VirtRegion) {
|
||
let prev_hole: Option<(&VirtAddr, &mut usize)> =
|
||
self.vm_holes.range_mut(..=region.start()).next_back();
|
||
|
||
if let Some((prev_hole_vaddr, prev_hole_size)) = prev_hole {
|
||
let prev_hole_end = prev_hole_vaddr.add(*prev_hole_size);
|
||
|
||
if prev_hole_end > region.start() {
|
||
// 如果前一个空洞的结束地址大于当前空洞的起始地址,那么就需要调整前一个空洞的大小。
|
||
*prev_hole_size = region.start().data() - prev_hole_vaddr.data();
|
||
}
|
||
|
||
if prev_hole_end > region.end() {
|
||
// 如果前一个空洞的结束地址大于当前空洞的结束地址,那么就需要增加一个新的空洞。
|
||
self.vm_holes
|
||
.insert(region.end(), prev_hole_end - region.end());
|
||
}
|
||
}
|
||
}
|
||
|
||
/// 在当前进程的地址空间中,释放一个指定大小的区域,使得该区域成为一个空洞。
|
||
/// 该函数会修改vm_holes中的空洞信息。
|
||
fn unreserve_hole(&mut self, region: &VirtRegion) {
|
||
// 如果将要插入的空洞与后一个空洞相邻,那么就需要合并。
|
||
let next_hole_size: Option<usize> = self.vm_holes.remove(®ion.end());
|
||
|
||
if let Some((_prev_hole_vaddr, prev_hole_size)) = self
|
||
.vm_holes
|
||
.range_mut(..region.start())
|
||
.next_back()
|
||
.filter(|(offset, size)| offset.data() + **size == region.start().data())
|
||
{
|
||
*prev_hole_size += region.size() + next_hole_size.unwrap_or(0);
|
||
} else {
|
||
self.vm_holes
|
||
.insert(region.start(), region.size() + next_hole_size.unwrap_or(0));
|
||
}
|
||
}
|
||
|
||
/// 在当前进程的映射关系中,插入一个新的VMA。
|
||
pub fn insert_vma(&mut self, vma: Arc<LockedVMA>) {
|
||
let region = vma.lock_irqsave().region;
|
||
// 要求插入的地址范围必须是空闲的,也就是说,当前进程的地址空间中,不能有任何与之重叠的VMA。
|
||
assert!(self.conflicts(region).next().is_none());
|
||
self.reserve_hole(®ion);
|
||
|
||
self.vmas.insert(vma);
|
||
}
|
||
|
||
/// @brief 删除一个VMA,并把对应的地址空间加入空洞中。
|
||
///
|
||
/// 这里不会取消VMA对应的地址的映射
|
||
///
|
||
/// @param region 要删除的VMA所在的地址范围
|
||
///
|
||
/// @return 如果成功删除了VMA,则返回被删除的VMA,否则返回None
|
||
/// 如果没有可以删除的VMA,则不会执行删除操作,并报告失败。
|
||
pub fn remove_vma(&mut self, region: &VirtRegion) -> Option<Arc<LockedVMA>> {
|
||
// 请注意,由于这里会对每个VMA加锁,因此性能很低
|
||
let vma: Arc<LockedVMA> = self
|
||
.vmas
|
||
.drain_filter(|vma| vma.lock_irqsave().region == *region)
|
||
.next()?;
|
||
self.unreserve_hole(region);
|
||
|
||
return Some(vma);
|
||
}
|
||
|
||
/// @brief Get the iterator of all VMAs in this process.
|
||
pub fn iter_vmas(&self) -> hashbrown::hash_set::Iter<Arc<LockedVMA>> {
|
||
return self.vmas.iter();
|
||
}
|
||
}
|
||
|
||
impl Default for UserMappings {
|
||
fn default() -> Self {
|
||
return Self::new();
|
||
}
|
||
}
|
||
|
||
/// 加了锁的VMA
|
||
///
|
||
/// 备注:进行性能测试,看看SpinLock和RwLock哪个更快。
|
||
#[derive(Debug)]
|
||
pub struct LockedVMA {
|
||
/// 用于计算哈希值,避免总是获取vma锁来计算哈希值
|
||
id: usize,
|
||
vma: SpinLock<VMA>,
|
||
}
|
||
|
||
impl core::hash::Hash for LockedVMA {
|
||
fn hash<H: Hasher>(&self, state: &mut H) {
|
||
self.id.hash(state);
|
||
}
|
||
}
|
||
|
||
impl PartialEq for LockedVMA {
|
||
fn eq(&self, other: &Self) -> bool {
|
||
self.id.eq(&other.id)
|
||
}
|
||
}
|
||
|
||
impl Eq for LockedVMA {}
|
||
|
||
#[allow(dead_code)]
|
||
impl LockedVMA {
|
||
pub fn new(vma: VMA) -> Arc<Self> {
|
||
let r = Arc::new(Self {
|
||
id: LOCKEDVMA_ID_ALLOCATOR.lock().alloc().unwrap(),
|
||
vma: SpinLock::new(vma),
|
||
});
|
||
r.vma.lock_irqsave().self_ref = Arc::downgrade(&r);
|
||
return r;
|
||
}
|
||
|
||
pub fn id(&self) -> usize {
|
||
self.id
|
||
}
|
||
|
||
pub fn lock(&self) -> SpinLockGuard<VMA> {
|
||
return self.vma.lock();
|
||
}
|
||
|
||
pub fn lock_irqsave(&self) -> SpinLockGuard<VMA> {
|
||
return self.vma.lock_irqsave();
|
||
}
|
||
|
||
/// 调整当前VMA的页面的标志位
|
||
///
|
||
/// TODO:增加调整虚拟页映射的物理地址的功能
|
||
///
|
||
/// @param flags 新的标志位
|
||
/// @param mapper 页表映射器
|
||
/// @param flusher 页表项刷新器
|
||
///
|
||
pub fn remap(
|
||
&self,
|
||
flags: EntryFlags<MMArch>,
|
||
mapper: &mut PageMapper,
|
||
mut flusher: impl Flusher<MMArch>,
|
||
) -> Result<(), SystemError> {
|
||
let mut guard = self.lock_irqsave();
|
||
for page in guard.region.pages() {
|
||
// 暂时要求所有的页帧都已经映射到页表
|
||
// TODO: 引入Lazy Mapping, 通过缺页中断来映射页帧,这里就不必要求所有的页帧都已经映射到页表了
|
||
let r = unsafe {
|
||
mapper
|
||
.remap(page.virt_address(), flags)
|
||
.expect("Failed to remap, beacuse of some page is not mapped")
|
||
};
|
||
flusher.consume(r);
|
||
}
|
||
guard.flags = flags;
|
||
return Ok(());
|
||
}
|
||
|
||
pub fn unmap(&self, mapper: &mut PageMapper, mut flusher: impl Flusher<MMArch>) {
|
||
// todo: 如果当前vma与文件相关,完善文件相关的逻辑
|
||
let mut guard = self.lock_irqsave();
|
||
|
||
// 获取物理页的anon_vma的守卫
|
||
let mut page_manager_guard: SpinLockGuard<'_, crate::mm::page::PageManager> =
|
||
page_manager_lock_irqsave();
|
||
|
||
// 获取映射的物理地址
|
||
if let Some((paddr, _flags)) = mapper.translate(guard.region().start()) {
|
||
// 如果是共享页,执行释放操作
|
||
let page = page_manager_guard.get(&paddr).unwrap();
|
||
let page_guard = page.read_irqsave();
|
||
if let PageType::Shm(shm_id) = page_guard.page_type() {
|
||
let mut shm_manager_guard = shm_manager_lock();
|
||
if let Some(kernel_shm) = shm_manager_guard.get_mut(shm_id) {
|
||
// 更新最后一次断开连接时间
|
||
kernel_shm.update_dtim();
|
||
|
||
// 映射计数减少
|
||
kernel_shm.decrease_count();
|
||
|
||
// 释放shm_id
|
||
if kernel_shm.map_count() == 0 && kernel_shm.mode().contains(ShmFlags::SHM_DEST)
|
||
{
|
||
shm_manager_guard.free_id(shm_id);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
for page in guard.region.pages() {
|
||
if mapper.translate(page.virt_address()).is_none() {
|
||
continue;
|
||
}
|
||
let (paddr, _, flush) = unsafe { mapper.unmap_phys(page.virt_address(), true) }
|
||
.expect("Failed to unmap, beacuse of some page is not mapped");
|
||
|
||
// 从anon_vma中删除当前VMA
|
||
let page = page_manager_guard.get_unwrap(&paddr);
|
||
let mut page_guard = page.write_irqsave();
|
||
page_guard.remove_vma(self);
|
||
|
||
// 如果物理页的vma链表长度为0并且未标记为不可回收,则释放物理页.
|
||
// TODO 后续由lru释放物理页面
|
||
if page_guard.can_deallocate() {
|
||
page_manager_guard.remove_page(&paddr);
|
||
}
|
||
|
||
flusher.consume(flush);
|
||
}
|
||
guard.mapped = false;
|
||
|
||
// 当vma对应共享文件的写映射时,唤醒脏页回写线程
|
||
if guard.vm_file().is_some()
|
||
&& guard
|
||
.vm_flags()
|
||
.contains(VmFlags::VM_SHARED | VmFlags::VM_WRITE)
|
||
{
|
||
crate::mm::page::PageReclaimer::wakeup_claim_thread();
|
||
}
|
||
}
|
||
|
||
pub fn mapped(&self) -> bool {
|
||
return self.vma.lock_irqsave().mapped;
|
||
}
|
||
|
||
/// 将当前VMA进行切分,切分成3个VMA,分别是:
|
||
///
|
||
/// 1. 前面的VMA,如果没有则为None
|
||
/// 2. 中间的VMA,也就是传入的Region
|
||
/// 3. 后面的VMA,如果没有则为None
|
||
pub fn extract(&self, region: VirtRegion, utable: &PageMapper) -> Option<VMASplitResult> {
|
||
assert!(region.start().check_aligned(MMArch::PAGE_SIZE));
|
||
assert!(region.end().check_aligned(MMArch::PAGE_SIZE));
|
||
|
||
let mut guard = self.lock_irqsave();
|
||
{
|
||
// 如果传入的region不在当前VMA的范围内,则直接返回None
|
||
if unlikely(region.start() < guard.region.start() || region.end() > guard.region.end())
|
||
{
|
||
return None;
|
||
}
|
||
|
||
let intersect: Option<VirtRegion> = guard.region.intersect(®ion);
|
||
// 如果当前VMA不包含region,则直接返回None
|
||
if unlikely(intersect.is_none()) {
|
||
return None;
|
||
}
|
||
let intersect: VirtRegion = intersect.unwrap();
|
||
if unlikely(intersect == guard.region) {
|
||
// 如果当前VMA完全包含region,则直接返回当前VMA
|
||
return Some(VMASplitResult::new(
|
||
None,
|
||
guard.self_ref.upgrade().unwrap(),
|
||
None,
|
||
));
|
||
}
|
||
}
|
||
|
||
let before: Option<Arc<LockedVMA>> = guard.region.before(®ion).map(|virt_region| {
|
||
let mut vma: VMA = unsafe { guard.clone() };
|
||
vma.region = virt_region;
|
||
vma.mapped = false;
|
||
let vma: Arc<LockedVMA> = LockedVMA::new(vma);
|
||
vma
|
||
});
|
||
|
||
let after: Option<Arc<LockedVMA>> = guard.region.after(®ion).map(|virt_region| {
|
||
let mut vma: VMA = unsafe { guard.clone() };
|
||
vma.region = virt_region;
|
||
vma.mapped = false;
|
||
let vma: Arc<LockedVMA> = LockedVMA::new(vma);
|
||
vma
|
||
});
|
||
|
||
// 重新设置before、after这两个VMA里面的物理页的anon_vma
|
||
let mut page_manager_guard = page_manager_lock_irqsave();
|
||
if let Some(before) = before.clone() {
|
||
let virt_iter = before.lock_irqsave().region.iter_pages();
|
||
for frame in virt_iter {
|
||
if let Some((paddr, _)) = utable.translate(frame.virt_address()) {
|
||
let page = page_manager_guard.get_unwrap(&paddr);
|
||
let mut page_guard = page.write_irqsave();
|
||
page_guard.insert_vma(before.clone());
|
||
page_guard.remove_vma(self);
|
||
before.lock_irqsave().mapped = true;
|
||
}
|
||
}
|
||
}
|
||
|
||
if let Some(after) = after.clone() {
|
||
let virt_iter = after.lock_irqsave().region.iter_pages();
|
||
for frame in virt_iter {
|
||
if let Some((paddr, _)) = utable.translate(frame.virt_address()) {
|
||
let page = page_manager_guard.get_unwrap(&paddr);
|
||
let mut page_guard = page.write_irqsave();
|
||
page_guard.insert_vma(after.clone());
|
||
page_guard.remove_vma(self);
|
||
after.lock_irqsave().mapped = true;
|
||
}
|
||
}
|
||
}
|
||
|
||
guard.region = region;
|
||
|
||
return Some(VMASplitResult::new(
|
||
before,
|
||
guard.self_ref.upgrade().unwrap(),
|
||
after,
|
||
));
|
||
}
|
||
|
||
/// 判断VMA是否为外部(非当前进程空间)的VMA
|
||
pub fn is_foreign(&self) -> bool {
|
||
let guard = self.lock_irqsave();
|
||
if let Some(space) = guard.user_address_space.clone() {
|
||
if let Some(space) = space.upgrade() {
|
||
return AddressSpace::is_current(&space);
|
||
} else {
|
||
return true;
|
||
}
|
||
} else {
|
||
return true;
|
||
}
|
||
}
|
||
|
||
/// 判断VMA是否可访问
|
||
pub fn is_accessible(&self) -> bool {
|
||
let guard = self.lock_irqsave();
|
||
let vm_access_flags: VmFlags = VmFlags::VM_READ | VmFlags::VM_WRITE | VmFlags::VM_EXEC;
|
||
guard.vm_flags().intersects(vm_access_flags)
|
||
}
|
||
|
||
/// 判断VMA是否为匿名映射
|
||
pub fn is_anonymous(&self) -> bool {
|
||
let guard = self.lock_irqsave();
|
||
guard.vm_file.is_none()
|
||
}
|
||
|
||
/// 判断VMA是否为大页映射
|
||
pub fn is_hugepage(&self) -> bool {
|
||
//TODO: 实现巨页映射判断逻辑,目前不支持巨页映射
|
||
false
|
||
}
|
||
}
|
||
|
||
impl Drop for LockedVMA {
|
||
fn drop(&mut self) {
|
||
LOCKEDVMA_ID_ALLOCATOR.lock().free(self.id);
|
||
}
|
||
}
|
||
|
||
/// VMA切分结果
|
||
#[allow(dead_code)]
|
||
pub struct VMASplitResult {
|
||
pub prev: Option<Arc<LockedVMA>>,
|
||
pub middle: Arc<LockedVMA>,
|
||
pub after: Option<Arc<LockedVMA>>,
|
||
}
|
||
|
||
impl VMASplitResult {
|
||
pub fn new(
|
||
prev: Option<Arc<LockedVMA>>,
|
||
middle: Arc<LockedVMA>,
|
||
post: Option<Arc<LockedVMA>>,
|
||
) -> Self {
|
||
Self {
|
||
prev,
|
||
middle,
|
||
after: post,
|
||
}
|
||
}
|
||
}
|
||
|
||
/// @brief 虚拟内存区域
|
||
#[derive(Debug)]
|
||
pub struct VMA {
|
||
/// 虚拟内存区域对应的虚拟地址范围
|
||
region: VirtRegion,
|
||
/// 虚拟内存区域标志
|
||
vm_flags: VmFlags,
|
||
/// VMA内的页帧的标志
|
||
flags: EntryFlags<MMArch>,
|
||
/// VMA内的页帧是否已经映射到页表
|
||
mapped: bool,
|
||
/// VMA所属的用户地址空间
|
||
user_address_space: Option<Weak<AddressSpace>>,
|
||
self_ref: Weak<LockedVMA>,
|
||
|
||
vm_file: Option<Arc<File>>,
|
||
/// VMA映射的文件部分相对于整个文件的偏移页数
|
||
file_pgoff: Option<usize>,
|
||
|
||
provider: Provider,
|
||
}
|
||
|
||
impl core::hash::Hash for VMA {
|
||
fn hash<H: Hasher>(&self, state: &mut H) {
|
||
self.region.hash(state);
|
||
self.flags.hash(state);
|
||
self.mapped.hash(state);
|
||
}
|
||
}
|
||
|
||
/// 描述不同类型的内存提供者或资源
|
||
#[derive(Debug)]
|
||
pub enum Provider {
|
||
Allocated, // TODO:其他
|
||
}
|
||
|
||
#[allow(dead_code)]
|
||
impl VMA {
|
||
pub fn new(
|
||
region: VirtRegion,
|
||
vm_flags: VmFlags,
|
||
flags: EntryFlags<MMArch>,
|
||
file: Option<Arc<File>>,
|
||
pgoff: Option<usize>,
|
||
mapped: bool,
|
||
) -> Self {
|
||
VMA {
|
||
region,
|
||
vm_flags,
|
||
flags,
|
||
mapped,
|
||
user_address_space: None,
|
||
self_ref: Weak::default(),
|
||
provider: Provider::Allocated,
|
||
vm_file: file,
|
||
file_pgoff: pgoff,
|
||
}
|
||
}
|
||
|
||
pub fn region(&self) -> &VirtRegion {
|
||
return &self.region;
|
||
}
|
||
|
||
pub fn vm_flags(&self) -> &VmFlags {
|
||
return &self.vm_flags;
|
||
}
|
||
|
||
pub fn vm_file(&self) -> Option<Arc<File>> {
|
||
return self.vm_file.clone();
|
||
}
|
||
|
||
pub fn address_space(&self) -> Option<Weak<AddressSpace>> {
|
||
return self.user_address_space.clone();
|
||
}
|
||
|
||
pub fn set_vm_flags(&mut self, vm_flags: VmFlags) {
|
||
self.vm_flags = vm_flags;
|
||
}
|
||
|
||
pub fn set_region_size(&mut self, new_region_size: usize) {
|
||
self.region.set_size(new_region_size);
|
||
}
|
||
|
||
pub fn set_mapped(&mut self, mapped: bool) {
|
||
self.mapped = mapped;
|
||
}
|
||
|
||
pub fn set_flags(&mut self) {
|
||
self.flags = MMArch::vm_get_page_prot(self.vm_flags);
|
||
}
|
||
|
||
/// # 拷贝当前VMA的内容
|
||
///
|
||
/// ### 安全性
|
||
///
|
||
/// 由于这样操作可能由于错误的拷贝,导致内存泄露、内存重复释放等问题,所以需要小心使用。
|
||
pub unsafe fn clone(&self) -> Self {
|
||
return Self {
|
||
region: self.region,
|
||
vm_flags: self.vm_flags,
|
||
flags: self.flags,
|
||
mapped: self.mapped,
|
||
user_address_space: self.user_address_space.clone(),
|
||
self_ref: self.self_ref.clone(),
|
||
provider: Provider::Allocated,
|
||
file_pgoff: self.file_pgoff,
|
||
vm_file: self.vm_file.clone(),
|
||
};
|
||
}
|
||
|
||
pub fn clone_info_only(&self) -> Self {
|
||
return Self {
|
||
region: self.region,
|
||
vm_flags: self.vm_flags,
|
||
flags: self.flags,
|
||
mapped: self.mapped,
|
||
user_address_space: None,
|
||
self_ref: Weak::default(),
|
||
provider: Provider::Allocated,
|
||
file_pgoff: self.file_pgoff,
|
||
vm_file: self.vm_file.clone(),
|
||
};
|
||
}
|
||
|
||
#[inline(always)]
|
||
pub fn flags(&self) -> EntryFlags<MMArch> {
|
||
return self.flags;
|
||
}
|
||
|
||
#[inline(always)]
|
||
pub fn file_page_offset(&self) -> Option<usize> {
|
||
return self.file_pgoff;
|
||
}
|
||
|
||
pub fn pages(&self) -> VirtPageFrameIter {
|
||
return VirtPageFrameIter::new(
|
||
VirtPageFrame::new(self.region.start()),
|
||
VirtPageFrame::new(self.region.end()),
|
||
);
|
||
}
|
||
|
||
pub fn remap(
|
||
&mut self,
|
||
flags: EntryFlags<MMArch>,
|
||
mapper: &mut PageMapper,
|
||
mut flusher: impl Flusher<MMArch>,
|
||
) -> Result<(), SystemError> {
|
||
for page in self.region.pages() {
|
||
// debug!("remap page {:?}", page.virt_address());
|
||
if mapper.translate(page.virt_address()).is_some() {
|
||
let r = unsafe {
|
||
mapper
|
||
.remap(page.virt_address(), flags)
|
||
.expect("Failed to remap")
|
||
};
|
||
flusher.consume(r);
|
||
}
|
||
// debug!("consume page {:?}", page.virt_address());
|
||
// debug!("remap page {:?} done", page.virt_address());
|
||
}
|
||
self.flags = flags;
|
||
return Ok(());
|
||
}
|
||
|
||
/// 检查当前VMA是否可以拥有指定的标志位
|
||
///
|
||
/// ## 参数
|
||
///
|
||
/// - `prot_flags` 要检查的标志位
|
||
pub fn can_have_flags(&self, prot_flags: ProtFlags) -> bool {
|
||
let is_downgrade = (self.flags.has_write() || !prot_flags.contains(ProtFlags::PROT_WRITE))
|
||
&& (self.flags.has_execute() || !prot_flags.contains(ProtFlags::PROT_EXEC));
|
||
|
||
match self.provider {
|
||
Provider::Allocated { .. } => true,
|
||
|
||
#[allow(unreachable_patterns)]
|
||
_ => is_downgrade,
|
||
}
|
||
}
|
||
|
||
/// 把物理地址映射到虚拟地址
|
||
///
|
||
/// @param phys 要映射的物理地址
|
||
/// @param destination 要映射到的虚拟地址
|
||
/// @param count 要映射的页帧数量
|
||
/// @param flags 页面标志位
|
||
/// @param mapper 页表映射器
|
||
/// @param flusher 页表项刷新器
|
||
///
|
||
/// @return 返回映射后的虚拟内存区域
|
||
pub fn physmap(
|
||
phys: PhysPageFrame,
|
||
destination: VirtPageFrame,
|
||
count: PageFrameCount,
|
||
vm_flags: VmFlags,
|
||
flags: EntryFlags<MMArch>,
|
||
mapper: &mut PageMapper,
|
||
mut flusher: impl Flusher<MMArch>,
|
||
) -> Result<Arc<LockedVMA>, SystemError> {
|
||
let mut cur_phy = phys;
|
||
let mut cur_dest = destination;
|
||
|
||
for _ in 0..count.data() {
|
||
// 将物理页帧映射到虚拟页帧
|
||
let r =
|
||
unsafe { mapper.map_phys(cur_dest.virt_address(), cur_phy.phys_address(), flags) }
|
||
.expect("Failed to map phys, may be OOM error");
|
||
|
||
// todo: 增加OOM处理
|
||
|
||
// 刷新TLB
|
||
flusher.consume(r);
|
||
|
||
cur_phy = cur_phy.next();
|
||
cur_dest = cur_dest.next();
|
||
}
|
||
|
||
let r: Arc<LockedVMA> = LockedVMA::new(VMA::new(
|
||
VirtRegion::new(destination.virt_address(), count.data() * MMArch::PAGE_SIZE),
|
||
vm_flags,
|
||
flags,
|
||
None,
|
||
None,
|
||
true,
|
||
));
|
||
|
||
// 将VMA加入到anon_vma中
|
||
let mut page_manager_guard = page_manager_lock_irqsave();
|
||
cur_phy = phys;
|
||
for _ in 0..count.data() {
|
||
let paddr = cur_phy.phys_address();
|
||
let page = page_manager_guard.get_unwrap(&paddr);
|
||
page.write_irqsave().insert_vma(r.clone());
|
||
cur_phy = cur_phy.next();
|
||
}
|
||
|
||
return Ok(r);
|
||
}
|
||
|
||
/// 从页分配器中分配一些物理页,并把它们映射到指定的虚拟地址,然后创建VMA
|
||
/// ## 参数
|
||
///
|
||
/// - `destination`: 要映射到的虚拟地址
|
||
/// - `page_count`: 要映射的页帧数量
|
||
/// - `vm_flags`: VMA标志位
|
||
/// - `flags`: 页面标志位
|
||
/// - `mapper`: 页表映射器
|
||
/// - `flusher`: 页表项刷新器
|
||
/// - `file`: 映射文件
|
||
/// - `pgoff`: 返回映射后的虚拟内存区域
|
||
///
|
||
/// ## 返回值
|
||
/// - 页面错误处理信息标志
|
||
#[allow(clippy::too_many_arguments)]
|
||
pub fn zeroed(
|
||
destination: VirtPageFrame,
|
||
page_count: PageFrameCount,
|
||
vm_flags: VmFlags,
|
||
flags: EntryFlags<MMArch>,
|
||
mapper: &mut PageMapper,
|
||
mut flusher: impl Flusher<MMArch>,
|
||
file: Option<Arc<File>>,
|
||
pgoff: Option<usize>,
|
||
) -> Result<Arc<LockedVMA>, SystemError> {
|
||
let mut cur_dest: VirtPageFrame = destination;
|
||
// debug!(
|
||
// "VMA::zeroed: page_count = {:?}, destination={destination:?}",
|
||
// page_count
|
||
// );
|
||
for _ in 0..page_count.data() {
|
||
// debug!(
|
||
// "VMA::zeroed: cur_dest={cur_dest:?}, vaddr = {:?}",
|
||
// cur_dest.virt_address()
|
||
// );
|
||
let r = unsafe { mapper.map(cur_dest.virt_address(), flags) }
|
||
.expect("Failed to map zero, may be OOM error");
|
||
// todo: 增加OOM处理
|
||
|
||
// 稍后再刷新TLB,这里取消刷新
|
||
flusher.consume(r);
|
||
cur_dest = cur_dest.next();
|
||
}
|
||
let r = LockedVMA::new(VMA::new(
|
||
VirtRegion::new(
|
||
destination.virt_address(),
|
||
page_count.data() * MMArch::PAGE_SIZE,
|
||
),
|
||
vm_flags,
|
||
flags,
|
||
file,
|
||
pgoff,
|
||
true,
|
||
));
|
||
drop(flusher);
|
||
// debug!("VMA::zeroed: flusher dropped");
|
||
|
||
// 清空这些内存并将VMA加入到anon_vma中
|
||
let mut page_manager_guard = page_manager_lock_irqsave();
|
||
let virt_iter: VirtPageFrameIter =
|
||
VirtPageFrameIter::new(destination, destination.add(page_count));
|
||
for frame in virt_iter {
|
||
let paddr = mapper.translate(frame.virt_address()).unwrap().0;
|
||
|
||
// 将VMA加入到anon_vma
|
||
let page = page_manager_guard.get_unwrap(&paddr);
|
||
page.write_irqsave().insert_vma(r.clone());
|
||
}
|
||
// debug!("VMA::zeroed: done");
|
||
return Ok(r);
|
||
}
|
||
|
||
pub fn page_address(&self, index: usize) -> Result<VirtAddr, SystemError> {
|
||
if index >= self.file_pgoff.unwrap() {
|
||
let address =
|
||
self.region.start + ((index - self.file_pgoff.unwrap()) << MMArch::PAGE_SHIFT);
|
||
if address <= self.region.end() {
|
||
return Ok(address);
|
||
}
|
||
}
|
||
return Err(SystemError::EFAULT);
|
||
}
|
||
}
|
||
|
||
impl Drop for VMA {
|
||
fn drop(&mut self) {
|
||
// 当VMA被释放时,需要确保它已经被从页表中解除映射
|
||
assert!(!self.mapped, "VMA is still mapped");
|
||
}
|
||
}
|
||
|
||
impl PartialEq for VMA {
|
||
fn eq(&self, other: &Self) -> bool {
|
||
return self.region == other.region;
|
||
}
|
||
}
|
||
|
||
impl Eq for VMA {}
|
||
|
||
impl PartialOrd for VMA {
|
||
fn partial_cmp(&self, other: &Self) -> Option<cmp::Ordering> {
|
||
Some(self.cmp(other))
|
||
}
|
||
}
|
||
|
||
impl Ord for VMA {
|
||
fn cmp(&self, other: &Self) -> cmp::Ordering {
|
||
return self.region.cmp(&other.region);
|
||
}
|
||
}
|
||
|
||
#[derive(Debug)]
|
||
pub struct UserStack {
|
||
// 栈底地址
|
||
stack_bottom: VirtAddr,
|
||
// 当前已映射的大小
|
||
mapped_size: usize,
|
||
/// 栈顶地址(这个值需要仔细确定!因为它可能不会实时与用户栈的真实栈顶保持一致!要小心!)
|
||
current_sp: VirtAddr,
|
||
}
|
||
|
||
impl UserStack {
|
||
/// 默认的用户栈底地址
|
||
pub const DEFAULT_USER_STACK_BOTTOM: VirtAddr = MMArch::USER_STACK_START;
|
||
/// 默认的用户栈大小为8MB
|
||
pub const DEFAULT_USER_STACK_SIZE: usize = 8 * 1024 * 1024;
|
||
/// 用户栈的保护页数量
|
||
pub const GUARD_PAGES_NUM: usize = 4;
|
||
|
||
/// 创建一个用户栈
|
||
pub fn new(
|
||
vm: &mut InnerAddressSpace,
|
||
stack_bottom: Option<VirtAddr>,
|
||
stack_size: usize,
|
||
) -> Result<Self, SystemError> {
|
||
let stack_bottom = stack_bottom.unwrap_or(Self::DEFAULT_USER_STACK_BOTTOM);
|
||
assert!(stack_bottom.check_aligned(MMArch::PAGE_SIZE));
|
||
|
||
// 分配用户栈的保护页
|
||
let guard_size = Self::GUARD_PAGES_NUM * MMArch::PAGE_SIZE;
|
||
let actual_stack_bottom = stack_bottom - guard_size;
|
||
|
||
let mut prot_flags = ProtFlags::PROT_READ | ProtFlags::PROT_WRITE;
|
||
let map_flags = MapFlags::MAP_PRIVATE
|
||
| MapFlags::MAP_ANONYMOUS
|
||
| MapFlags::MAP_FIXED_NOREPLACE
|
||
| MapFlags::MAP_GROWSDOWN;
|
||
// debug!(
|
||
// "map anonymous stack: {:?} {}",
|
||
// actual_stack_bottom,
|
||
// guard_size
|
||
// );
|
||
vm.map_anonymous(
|
||
actual_stack_bottom,
|
||
guard_size,
|
||
prot_flags,
|
||
map_flags,
|
||
false,
|
||
false,
|
||
)?;
|
||
// test_buddy();
|
||
// 设置保护页只读
|
||
prot_flags.remove(ProtFlags::PROT_WRITE);
|
||
// debug!(
|
||
// "to mprotect stack guard pages: {:?} {}",
|
||
// actual_stack_bottom,
|
||
// guard_size
|
||
// );
|
||
vm.mprotect(
|
||
VirtPageFrame::new(actual_stack_bottom),
|
||
PageFrameCount::new(Self::GUARD_PAGES_NUM),
|
||
prot_flags,
|
||
)?;
|
||
|
||
// debug!(
|
||
// "mprotect stack guard pages done: {:?} {}",
|
||
// actual_stack_bottom,
|
||
// guard_size
|
||
// );
|
||
|
||
let mut user_stack = UserStack {
|
||
stack_bottom: actual_stack_bottom,
|
||
mapped_size: guard_size,
|
||
current_sp: actual_stack_bottom - guard_size,
|
||
};
|
||
|
||
// debug!("extend user stack: {:?} {}", stack_bottom, stack_size);
|
||
// 分配用户栈
|
||
user_stack.initial_extend(vm, stack_size)?;
|
||
// debug!("user stack created: {:?} {}", stack_bottom, stack_size);
|
||
return Ok(user_stack);
|
||
}
|
||
|
||
fn initial_extend(
|
||
&mut self,
|
||
vm: &mut InnerAddressSpace,
|
||
mut bytes: usize,
|
||
) -> Result<(), SystemError> {
|
||
let prot_flags = ProtFlags::PROT_READ | ProtFlags::PROT_WRITE | ProtFlags::PROT_EXEC;
|
||
let map_flags = MapFlags::MAP_PRIVATE | MapFlags::MAP_ANONYMOUS | MapFlags::MAP_GROWSDOWN;
|
||
|
||
bytes = page_align_up(bytes);
|
||
self.mapped_size += bytes;
|
||
|
||
vm.map_anonymous(
|
||
self.stack_bottom - self.mapped_size,
|
||
bytes,
|
||
prot_flags,
|
||
map_flags,
|
||
false,
|
||
false,
|
||
)?;
|
||
|
||
return Ok(());
|
||
}
|
||
|
||
/// 扩展用户栈
|
||
///
|
||
/// ## 参数
|
||
///
|
||
/// - `vm` 用户地址空间结构体
|
||
/// - `bytes` 要扩展的字节数
|
||
///
|
||
/// ## 返回值
|
||
///
|
||
/// - **Ok(())** 扩展成功
|
||
/// - **Err(SystemError)** 扩展失败
|
||
#[allow(dead_code)]
|
||
pub fn extend(
|
||
&mut self,
|
||
vm: &mut InnerAddressSpace,
|
||
mut bytes: usize,
|
||
) -> Result<(), SystemError> {
|
||
let prot_flags = ProtFlags::PROT_READ | ProtFlags::PROT_WRITE | ProtFlags::PROT_EXEC;
|
||
let map_flags = MapFlags::MAP_PRIVATE | MapFlags::MAP_ANONYMOUS;
|
||
|
||
bytes = page_align_up(bytes);
|
||
self.mapped_size += bytes;
|
||
|
||
vm.map_anonymous(
|
||
self.stack_bottom - self.mapped_size,
|
||
bytes,
|
||
prot_flags,
|
||
map_flags,
|
||
false,
|
||
false,
|
||
)?;
|
||
|
||
return Ok(());
|
||
}
|
||
|
||
/// 获取栈顶地址
|
||
///
|
||
/// 请注意,如果用户栈的栈顶地址发生变化,这个值可能不会实时更新!
|
||
pub fn sp(&self) -> VirtAddr {
|
||
return self.current_sp;
|
||
}
|
||
|
||
pub unsafe fn set_sp(&mut self, sp: VirtAddr) {
|
||
self.current_sp = sp;
|
||
}
|
||
|
||
/// 仅仅克隆用户栈的信息,不会克隆用户栈的内容/映射
|
||
pub unsafe fn clone_info_only(&self) -> Self {
|
||
return Self {
|
||
stack_bottom: self.stack_bottom,
|
||
mapped_size: self.mapped_size,
|
||
current_sp: self.current_sp,
|
||
};
|
||
}
|
||
|
||
/// 获取当前用户栈的大小(不包括保护页)
|
||
pub fn stack_size(&self) -> usize {
|
||
return self.mapped_size - Self::GUARD_PAGES_NUM * MMArch::PAGE_SIZE;
|
||
}
|
||
}
|