redox/arch/x86_64/src/context.rs

185 lines
5.1 KiB
Rust
Raw Normal View History

2016-11-17 20:12:02 +01:00
use core::mem;
use core::sync::atomic::{AtomicBool, ATOMIC_BOOL_INIT};
2016-08-29 02:38:53 +02:00
/// This must be used by the kernel to ensure that context switches are done atomically
/// Compare and exchange this to true when beginning a context switch on any CPU
/// The Context::switch_to function will set it back to false, allowing other CPU's to switch
/// This must be done, as no locks can be held on the stack during switch
pub static CONTEXT_SWITCH_LOCK: AtomicBool = ATOMIC_BOOL_INIT;
#[derive(Clone, Debug)]
2016-08-23 03:56:35 +02:00
pub struct Context {
2016-09-23 00:13:05 +02:00
/// FX valid?
loadable: bool,
/// FX location
fx: usize,
2016-09-12 20:21:34 +02:00
/// Page table pointer
cr3: usize,
2016-08-24 18:35:42 +02:00
/// RFLAGS register
rflags: usize,
/// RBX register
rbx: usize,
/// R12 register
2016-08-23 03:56:35 +02:00
r12: usize,
2016-08-24 18:35:42 +02:00
/// R13 register
2016-08-23 03:56:35 +02:00
r13: usize,
2016-08-24 18:35:42 +02:00
/// R14 register
2016-08-23 03:56:35 +02:00
r14: usize,
2016-08-24 18:35:42 +02:00
/// R15 register
2016-08-23 03:56:35 +02:00
r15: usize,
2016-08-24 18:35:42 +02:00
/// Base pointer
rbp: usize,
/// Stack pointer
2016-09-12 20:21:34 +02:00
rsp: usize
2016-08-23 03:56:35 +02:00
}
impl Context {
pub fn new() -> Context {
Context {
2016-09-23 00:13:05 +02:00
loadable: false,
fx: 0,
2016-09-12 20:21:34 +02:00
cr3: 0,
2016-08-24 18:35:42 +02:00
rflags: 0,
rbx: 0,
2016-08-23 03:56:35 +02:00
r12: 0,
r13: 0,
r14: 0,
r15: 0,
2016-08-24 18:35:42 +02:00
rbp: 0,
2016-09-12 20:21:34 +02:00
rsp: 0
2016-08-23 03:56:35 +02:00
}
}
pub fn get_page_table(&self) -> usize {
self.cr3
}
2016-09-23 00:13:05 +02:00
pub fn set_fx(&mut self, address: usize) {
self.fx = address;
}
2016-09-12 20:21:34 +02:00
pub fn set_page_table(&mut self, address: usize) {
self.cr3 = address;
}
2016-08-29 02:38:53 +02:00
pub fn set_stack(&mut self, address: usize) {
self.rsp = address;
}
2016-11-17 20:12:02 +01:00
pub unsafe fn signal_stack(&mut self, handler: extern fn(usize), sig: u8) {
self.push_stack(sig as usize);
self.push_stack(handler as usize);
self.push_stack(signal_handler_wrapper as usize);
}
pub unsafe fn push_stack(&mut self, value: usize) {
self.rsp -= mem::size_of::<usize>();
*(self.rsp as *mut usize) = value;
}
pub unsafe fn pop_stack(&mut self) -> usize {
let value = *(self.rsp as *const usize);
self.rsp += mem::size_of::<usize>();
value
}
2016-08-29 02:38:53 +02:00
/// Switch to the next context by restoring its stack and registers
#[cold]
2016-08-23 03:56:35 +02:00
#[inline(never)]
#[naked]
pub unsafe fn switch_to(&mut self, next: &mut Context) {
asm!("fxsave [$0]" : : "r"(self.fx) : "memory" : "intel", "volatile");
self.loadable = true;
if next.loadable {
asm!("fxrstor [$0]" : : "r"(next.fx) : "memory" : "intel", "volatile");
}else{
asm!("fninit" : : : "memory" : "intel", "volatile");
}
2016-09-12 20:21:34 +02:00
asm!("mov $0, cr3" : "=r"(self.cr3) : : "memory" : "intel", "volatile");
if next.cr3 != self.cr3 {
asm!("mov cr3, $0" : : "r"(next.cr3) : "memory" : "intel", "volatile");
}
2016-08-24 18:35:42 +02:00
asm!("pushfq ; pop $0" : "=r"(self.rflags) : : "memory" : "intel", "volatile");
asm!("push $0 ; popfq" : : "r"(next.rflags) : "memory" : "intel", "volatile");
2016-08-23 03:56:35 +02:00
2016-08-24 18:35:42 +02:00
asm!("mov $0, rbx" : "=r"(self.rbx) : : "memory" : "intel", "volatile");
asm!("mov rbx, $0" : : "r"(next.rbx) : "memory" : "intel", "volatile");
2016-08-23 03:56:35 +02:00
asm!("mov $0, r12" : "=r"(self.r12) : : "memory" : "intel", "volatile");
asm!("mov r12, $0" : : "r"(next.r12) : "memory" : "intel", "volatile");
asm!("mov $0, r13" : "=r"(self.r13) : : "memory" : "intel", "volatile");
asm!("mov r13, $0" : : "r"(next.r13) : "memory" : "intel", "volatile");
asm!("mov $0, r14" : "=r"(self.r14) : : "memory" : "intel", "volatile");
asm!("mov r14, $0" : : "r"(next.r14) : "memory" : "intel", "volatile");
asm!("mov $0, r15" : "=r"(self.r15) : : "memory" : "intel", "volatile");
asm!("mov r15, $0" : : "r"(next.r15) : "memory" : "intel", "volatile");
2016-08-24 18:35:42 +02:00
asm!("mov $0, rsp" : "=r"(self.rsp) : : "memory" : "intel", "volatile");
asm!("mov rsp, $0" : : "r"(next.rsp) : "memory" : "intel", "volatile");
2016-08-23 03:56:35 +02:00
2016-08-29 02:38:53 +02:00
asm!("mov $0, rbp" : "=r"(self.rbp) : : "memory" : "intel", "volatile");
asm!("mov rbp, $0" : : "r"(next.rbp) : "memory" : "intel", "volatile");
2016-08-23 03:56:35 +02:00
}
}
2016-11-17 20:12:02 +01:00
#[repr(packed)]
pub struct SignalHandlerStack {
r11: usize,
r10: usize,
r9: usize,
r8: usize,
rsi: usize,
rdi: usize,
rdx: usize,
rcx: usize,
rax: usize,
handler: extern fn(usize),
sig: usize,
rip: usize,
}
#[naked]
unsafe extern fn signal_handler_wrapper() {
#[inline(never)]
unsafe fn inner(stack: &SignalHandlerStack) {
(stack.handler)(stack.sig);
}
// Push scratch registers
2016-11-20 17:06:55 +01:00
asm!("push rax
2016-11-17 20:12:02 +01:00
push rcx
push rdx
push rdi
push rsi
push r8
push r9
push r10
push r11"
: : : : "intel", "volatile");
// Get reference to stack variables
let rsp: usize;
asm!("" : "={rsp}"(rsp) : : : "intel", "volatile");
// Call inner rust function
inner(&*(rsp as *const SignalHandlerStack));
// Pop scratch registers, error code, and return
2016-11-20 17:06:55 +01:00
asm!("pop r11
2016-11-17 20:12:02 +01:00
pop r10
pop r9
pop r8
pop rsi
pop rdi
pop rdx
pop rcx
pop rax
add rsp, 16"
: : : : "intel", "volatile");
}