Allow cloning of files and memory

This commit is contained in:
Jeremy Soller 2016-09-16 18:27:54 -06:00
parent 8ee9f1d7da
commit e680a84a57
2 changed files with 99 additions and 55 deletions

View file

@ -1,5 +1,7 @@
use alloc::arc::Arc;
use alloc::boxed::Box; use alloc::boxed::Box;
use collections::Vec; use collections::Vec;
use spin::Mutex;
use arch; use arch;
use super::file::File; use super::file::File;
@ -32,7 +34,7 @@ pub struct Context {
/// User stack /// User stack
pub stack: Option<Memory>, pub stack: Option<Memory>,
/// The open files in the scheme /// The open files in the scheme
pub files: Vec<Option<File>> pub files: Arc<Mutex<Vec<Option<File>>>>
} }
impl Context { impl Context {
@ -47,22 +49,23 @@ impl Context {
image: Vec::new(), image: Vec::new(),
heap: None, heap: None,
stack: None, stack: None,
files: Vec::new() files: Arc::new(Mutex::new(Vec::new()))
} }
} }
/// Add a file to the lowest available slot. /// Add a file to the lowest available slot.
/// Return the file descriptor number or None if no slot was found /// Return the file descriptor number or None if no slot was found
pub fn add_file(&mut self, file: File) -> Option<usize> { pub fn add_file(&mut self, file: File) -> Option<usize> {
for (i, mut file_option) in self.files.iter_mut().enumerate() { let mut files = self.files.lock();
for (i, mut file_option) in files.iter_mut().enumerate() {
if file_option.is_none() { if file_option.is_none() {
*file_option = Some(file); *file_option = Some(file);
return Some(i); return Some(i);
} }
} }
let len = self.files.len(); let len = files.len();
if len < super::CONTEXT_MAX_FILES { if len < super::CONTEXT_MAX_FILES {
self.files.push(Some(file)); files.push(Some(file));
Some(len) Some(len)
} else { } else {
None None
@ -71,8 +74,9 @@ impl Context {
/// Get a file /// Get a file
pub fn get_file(&self, i: usize) -> Option<File> { pub fn get_file(&self, i: usize) -> Option<File> {
if i < self.files.len() { let files = self.files.lock();
self.files[i] if i < files.len() {
files[i]
} else { } else {
None None
} }
@ -81,8 +85,9 @@ impl Context {
/// Remove a file /// Remove a file
// TODO: adjust files vector to smaller size if possible // TODO: adjust files vector to smaller size if possible
pub fn remove_file(&mut self, i: usize) -> Option<File> { pub fn remove_file(&mut self, i: usize) -> Option<File> {
if i < self.files.len() { let mut files = self.files.lock();
self.files[i].take() if i < files.len() {
files[i].take()
} else { } else {
None None
} }

View file

@ -1,7 +1,9 @@
///! Process syscalls ///! Process syscalls
use alloc::arc::Arc;
use core::mem; use core::mem;
use core::str; use core::str;
use spin::Mutex;
use arch; use arch;
use arch::memory::allocate_frame; use arch::memory::allocate_frame;
@ -62,7 +64,7 @@ pub fn clone(flags: usize, stack_base: usize) -> Result<usize> {
let mut image = vec![]; let mut image = vec![];
let mut heap_option = None; let mut heap_option = None;
let mut stack_option = None; let mut stack_option = None;
let mut files = vec![]; let mut files = Arc::new(Mutex::new(vec![]));
// Copy from old process // Copy from old process
{ {
@ -74,15 +76,23 @@ pub fn clone(flags: usize, stack_base: usize) -> Result<usize> {
if let Some(ref stack) = context.kstack { if let Some(ref stack) = context.kstack {
offset = stack_base - stack.as_ptr() as usize - mem::size_of::<usize>(); // Add clone ret offset = stack_base - stack.as_ptr() as usize - mem::size_of::<usize>(); // Add clone ret
let mut new_stack = stack.clone(); let mut new_stack = stack.clone();
unsafe { unsafe {
let func_ptr = new_stack.as_mut_ptr().offset(offset as isize); let func_ptr = new_stack.as_mut_ptr().offset(offset as isize);
*(func_ptr as *mut usize) = arch::interrupt::syscall::clone_ret as usize; *(func_ptr as *mut usize) = arch::interrupt::syscall::clone_ret as usize;
} }
kstack_option = Some(new_stack); kstack_option = Some(new_stack);
} }
if flags & CLONE_VM == CLONE_VM { if flags & CLONE_VM == CLONE_VM {
panic!("unimplemented: CLONE_VM"); for memory_shared in context.image.iter() {
image.push(memory_shared.borrow());
}
if let Some(ref heap_shared) = context.heap {
heap_option = Some(heap_shared.borrow());
}
} else { } else {
for memory_shared in context.image.iter() { for memory_shared in context.image.iter() {
memory_shared.with(|memory| { memory_shared.with(|memory| {
@ -135,19 +145,21 @@ pub fn clone(flags: usize, stack_base: usize) -> Result<usize> {
true, true,
false false
); );
unsafe { unsafe {
arch::externs::memcpy(new_stack.start_address().get() as *mut u8, arch::externs::memcpy(new_stack.start_address().get() as *mut u8,
stack.start_address().get() as *const u8, stack.start_address().get() as *const u8,
stack.size()); stack.size());
} }
new_stack.remap(stack.flags(), true); new_stack.remap(stack.flags(), true);
stack_option = Some(new_stack); stack_option = Some(new_stack);
} }
if flags & CLONE_FILES == CLONE_FILES { if flags & CLONE_FILES == CLONE_FILES {
panic!("unimplemented: CLONE_FILES"); files = context.files.clone();
} else { } else {
for (fd, file_option) in context.files.iter().enumerate() { for (fd, file_option) in context.files.lock().iter().enumerate() {
if let Some(file) = *file_option { if let Some(file) = *file_option {
let result = { let result = {
let schemes = scheme::schemes(); let schemes = scheme::schemes();
@ -157,14 +169,14 @@ pub fn clone(flags: usize, stack_base: usize) -> Result<usize> {
}; };
match result { match result {
Ok(new_number) => { Ok(new_number) => {
files.push(Some(context::file::File { scheme: file.scheme, number: new_number })); files.lock().push(Some(context::file::File { scheme: file.scheme, number: new_number }));
}, },
Err(err) => { Err(err) => {
println!("clone: failed to dup {}: {:?}", fd, err); println!("clone: failed to dup {}: {:?}", fd, err);
} }
} }
} else { } else {
files.push(None); files.lock().push(None);
} }
} }
} }
@ -191,58 +203,85 @@ pub fn clone(flags: usize, stack_base: usize) -> Result<usize> {
// Copy kernel mapping // Copy kernel mapping
{ {
let kernel_frame = active_table.p4()[510].pointed_frame().expect("kernel table not mapped"); let frame = active_table.p4()[510].pointed_frame().expect("kernel table not mapped");
let flags = active_table.p4()[510].flags();
active_table.with(&mut new_table, &mut temporary_page, |mapper| { active_table.with(&mut new_table, &mut temporary_page, |mapper| {
mapper.p4_mut()[510].set(kernel_frame, entry::PRESENT | entry::WRITABLE); mapper.p4_mut()[510].set(frame, flags);
}); });
} }
// Copy percpu mapping // Set kernel stack
{
extern {
/// The starting byte of the thread data segment
static mut __tdata_start: u8;
/// The ending byte of the thread BSS segment
static mut __tbss_end: u8;
}
let size = unsafe { & __tbss_end as *const _ as usize - & __tdata_start as *const _ as usize };
let start = arch::KERNEL_PERCPU_OFFSET + arch::KERNEL_PERCPU_SIZE * ::cpu_id();
let end = start + size;
let start_page = Page::containing_address(VirtualAddress::new(start));
let end_page = Page::containing_address(VirtualAddress::new(end - 1));
for page in Page::range_inclusive(start_page, end_page) {
let frame = active_table.translate_page(page).expect("kernel percpu not mapped");
active_table.with(&mut new_table, &mut temporary_page, |mapper| {
mapper.map_to(page, frame, entry::PRESENT | entry::NO_EXECUTE | entry::WRITABLE);
});
}
}
if let Some(stack) = kstack_option.take() { if let Some(stack) = kstack_option.take() {
context.arch.set_stack(stack.as_ptr() as usize + offset); context.arch.set_stack(stack.as_ptr() as usize + offset);
context.kstack = Some(stack); context.kstack = Some(stack);
} }
for memory_shared in image.iter_mut() { // Setup heap
memory_shared.with(|memory| { if flags & CLONE_VM == CLONE_VM {
let start = VirtualAddress::new(memory.start_address().get() - arch::USER_TMP_OFFSET + arch::USER_OFFSET); // Copy user image mapping, if found
memory.move_to(start, &mut new_table, &mut temporary_page, true); if ! image.is_empty() {
}); let frame = active_table.p4()[0].pointed_frame().expect("user image not mapped");
} let flags = active_table.p4()[0].flags();
context.image = image; active_table.with(&mut new_table, &mut temporary_page, |mapper| {
mapper.p4_mut()[0].set(frame, flags);
});
}
context.image = image;
if let Some(heap_shared) = heap_option.take() { // Copy user heap mapping, if found
heap_shared.with(|heap| { if let Some(heap_shared) = heap_option {
heap.move_to(VirtualAddress::new(arch::USER_HEAP_OFFSET), &mut new_table, &mut temporary_page, true); let frame = active_table.p4()[1].pointed_frame().expect("user heap not mapped");
}); let flags = active_table.p4()[0].flags();
context.heap = Some(heap_shared); active_table.with(&mut new_table, &mut temporary_page, |mapper| {
mapper.p4_mut()[1].set(frame, flags);
});
context.heap = Some(heap_shared);
}
} else {
// Copy percpu mapping
{
extern {
/// The starting byte of the thread data segment
static mut __tdata_start: u8;
/// The ending byte of the thread BSS segment
static mut __tbss_end: u8;
}
let size = unsafe { & __tbss_end as *const _ as usize - & __tdata_start as *const _ as usize };
let start = arch::KERNEL_PERCPU_OFFSET + arch::KERNEL_PERCPU_SIZE * ::cpu_id();
let end = start + size;
let start_page = Page::containing_address(VirtualAddress::new(start));
let end_page = Page::containing_address(VirtualAddress::new(end - 1));
for page in Page::range_inclusive(start_page, end_page) {
let frame = active_table.translate_page(page).expect("kernel percpu not mapped");
active_table.with(&mut new_table, &mut temporary_page, |mapper| {
mapper.map_to(page, frame, entry::PRESENT | entry::NO_EXECUTE | entry::WRITABLE);
});
}
}
// Move copy of image
for memory_shared in image.iter_mut() {
memory_shared.with(|memory| {
let start = VirtualAddress::new(memory.start_address().get() - arch::USER_TMP_OFFSET + arch::USER_OFFSET);
memory.move_to(start, &mut new_table, &mut temporary_page, true);
});
}
context.image = image;
// Move copy of heap
if let Some(heap_shared) = heap_option {
heap_shared.with(|heap| {
heap.move_to(VirtualAddress::new(arch::USER_HEAP_OFFSET), &mut new_table, &mut temporary_page, true);
});
context.heap = Some(heap_shared);
}
} }
if let Some(mut stack) = stack_option.take() { // Setup user stack
if let Some(mut stack) = stack_option {
stack.move_to(VirtualAddress::new(arch::USER_STACK_OFFSET), &mut new_table, &mut temporary_page, true); stack.move_to(VirtualAddress::new(arch::USER_STACK_OFFSET), &mut new_table, &mut temporary_page, true);
context.stack = Some(stack); context.stack = Some(stack);
} }