Add thread support and thread spawn syscall

This commit is contained in:
pjht 2024-11-27 17:52:16 -06:00
parent 98f665298c
commit 5f1274dd39
Signed by: pjht
GPG Key ID: CA239FC6934E6F3A
3 changed files with 199 additions and 77 deletions

View File

@ -248,7 +248,11 @@ pub fn send_ipc_to(
clippy::unwrap_used,
reason = "The PID is known valid due to using it in process_mut above"
)]
TASKING.wake(pid, SleepReason::WaitingForIPC).unwrap();
TASKING.process(pid, |process| {
for (tid, _) in &*process.threads().read() {
TASKING.wake(pid, tid, SleepReason::WaitingForIPC).unwrap();
}
}).unwrap();
Ok(())
}
@ -759,7 +763,7 @@ extern "C" fn syscall_handler() {
}
24 => {
let pid = usize(regs.rcx);
if TASKING.wake(pid, SleepReason::NewProcess).is_err() {
if TASKING.wake(pid, 0,SleepReason::NewProcess).is_err() {
retval = 1;
} else {
retval = 0;
@ -773,6 +777,15 @@ extern "C" fn syscall_handler() {
retval = 0;
}
}
26 => {
let res = TASKING.new_thread(ptr::with_exposed_provenance(usize(regs.rcx)), usize(regs.rdx));
if let Ok(tid) = res {
retval = 0;
retval2 = u64(tid);
} else {
retval = 1;
}
}
_ => (),
};
unsafe { SYSCALL_REGS = regs };

View File

@ -290,5 +290,5 @@ pub fn main() {
let init_pid = TASKING
.new_process(ptr::with_exposed_provenance(usize(init.ehdr.e_entry)), init_addr_space)
.expect("Failed to create init process");
TASKING.wake(init_pid, SleepReason::NewProcess).expect("Failed to wake new init process");
TASKING.wake(init_pid, 0,SleepReason::NewProcess).expect("Failed to wake new init process");
}

View File

@ -64,6 +64,7 @@ extern "C" fn task_init() {
naked_asm!(
"pop rcx", // Get the user stack pointer
"pop rbx", // Get the entry point
"pop rdi", // Get argument to the entry point
"push 43", // Push the stack segment selector - same as data
"push rcx", // Push the stack pointer
"pushfq", // Get the flags into RAX
@ -87,13 +88,103 @@ pub enum SleepReason {
#[derive(Debug)]
pub struct Process {
address_space: Option<AddressSpace>,
kernel_stack: Box<[usize], &'static ASpaceMutex>,
kernel_esp: *mut usize,
kernel_esp_top: VirtAddr,
address_spaces: Mutex<Slab<AddressSpace>>,
data_buffers: Mutex<Slab<*mut [u8]>>,
message_queue: Mutex<SegQueue<IpcMessage>>,
threads: RwLock<Slab<Thread>>,
main_thread: usize,
num_threads: usize,
}
impl Process {
fn new(
address_space: AddressSpace,
entry_point: *const extern "C" fn(usize) -> !,
) -> Result<Self, PagingError> {
let mut proc = Self {
address_space: Some(address_space),
address_spaces: Mutex::new(Slab::new()),
data_buffers: Mutex::new(Slab::new()),
message_queue: Mutex::new(SegQueue::new()),
threads: RwLock::new(Slab::new()),
main_thread: 0,
num_threads: 0,
};
let main_thread = proc.new_thread(entry_point, 0)?;
proc.main_thread = main_thread;
Ok(proc)
}
fn new_thread(
&mut self,
entry_point: *const extern "C" fn(usize) -> !,
argument: usize,
) -> Result<usize, PagingError> {
let stack_start = 0xFFF_FF80_0000 - (u64(self.num_threads) * (16*4096));
let mut kernel_stack = Vec::with_capacity_in(KSTACK_SIZE, &*KERNEL_SPACE);
kernel_stack.resize(KSTACK_SIZE - 0x4, 0);
#[expect(clippy::as_conversions, reason = "Needed to get address of function")]
kernel_stack.push(task_init as usize);
kernel_stack.push(stack_start as usize + (16 * 4096));
kernel_stack.push(entry_point.expose_provenance());
kernel_stack.push(argument);
let mut kernel_stack = kernel_stack.into_boxed_slice();
if let Some(address_space) = self.address_space.as_mut() {
address_space.map_assert_unused(
#[expect(
clippy::unwrap_used,
reason = "from_start_address requires the address to be page aligned, which it is."
)]
Page::from_start_address(VirtAddr::new(stack_start)).unwrap(),
16,
PageTableFlags::USER_ACCESSIBLE,
)?;
} else {
ACTIVE_SPACE.lock().map_assert_unused(
#[expect(
clippy::unwrap_used,
reason = "from_start_address requires the address to be page aligned, which it is."
)]
Page::from_start_address(VirtAddr::new(stack_start)).unwrap(),
16,
PageTableFlags::USER_ACCESSIBLE,
)?;
}
let sleeping =
if self.threads.read().is_empty() { Some(SleepReason::NewProcess) } else { None };
let thread = Thread {
#[expect(
clippy::indexing_slicing,
reason = "Stack length is 0x1_0000, this cannot panic"
)]
kernel_esp: &mut kernel_stack[KSTACK_SIZE - 10],
#[expect(
clippy::indexing_slicing,
reason = "Stack length is 0x1_0000, this cannot panic"
)]
kernel_esp_top: VirtAddr::from_ptr(
addr_of!(kernel_stack[KSTACK_SIZE - 1]).wrapping_add(1),
),
kernel_stack,
sleeping: RwLock::new(sleeping),
};
let idx = self.threads.write().insert(thread);
self.num_threads += 1;
Ok(idx)
}
pub fn threads(&self) -> &RwLock<Slab<Thread>> {
&self.threads
}
}
#[derive(Debug)]
pub struct Thread {
sleeping: RwLock<Option<SleepReason>>,
kernel_stack: Box<[usize], &'static ASpaceMutex>,
kernel_esp: *mut usize,
kernel_esp_top: VirtAddr,
}
pub struct IpcMessage {
@ -130,6 +221,7 @@ pub static TASKING: Lazy<Tasking> = Lazy::new(|| Tasking {
processes: RwLock::new(Slab::new()),
ready_to_run: Mutex::new(VecDeque::new()),
current_pid: RwLock::new(None),
current_tid: RwLock::new(None),
freeable_kstacks: Mutex::new(Vec::new()),
wfi_loop: AtomicBool::new(false),
});
@ -137,8 +229,9 @@ pub static TASKING: Lazy<Tasking> = Lazy::new(|| Tasking {
#[derive(Debug)]
pub struct Tasking {
processes: RwLock<Slab<Process>>,
ready_to_run: Mutex<VecDeque<usize>>,
ready_to_run: Mutex<VecDeque<(usize, usize)>>,
current_pid: RwLock<Option<usize>>,
current_tid: RwLock<Option<usize>>,
freeable_kstacks: Mutex<Vec<Box<[usize], &'static ASpaceMutex>>>,
wfi_loop: AtomicBool,
}
@ -148,45 +241,11 @@ pub const KSTACK_SIZE: usize = (4 * 4096) / 8;
impl Tasking {
pub fn new_process(
&self,
entry_point: *const extern "C" fn() -> !,
mut address_space: AddressSpace,
entry_point: *const extern "C" fn(usize) -> !,
address_space: AddressSpace,
) -> Result<usize, PagingError> {
let mut kernel_stack = Vec::with_capacity_in(KSTACK_SIZE, &*KERNEL_SPACE);
kernel_stack.resize(KSTACK_SIZE - 0x3, 0);
#[expect(clippy::as_conversions, reason = "Needed to get address of function")]
kernel_stack.push(task_init as usize);
kernel_stack.push(0xFFF_FF80_0000 + (16 * 4096));
kernel_stack.push(entry_point.expose_provenance());
let mut kernel_stack = kernel_stack.into_boxed_slice();
address_space.map_assert_unused(
#[expect(
clippy::unwrap_used,
reason = "from_start_address requires the address to be page aligned, which it is."
)]
Page::from_start_address(VirtAddr::new(0xFFF_FF80_0000)).unwrap(),
16,
PageTableFlags::USER_ACCESSIBLE,
)?;
let pid = self.processes.write().insert(Process {
#[expect(
clippy::indexing_slicing,
reason = "Stack length is 0x1_0000, this cannot panic"
)]
kernel_esp: &mut kernel_stack[KSTACK_SIZE - 9],
#[expect(
clippy::indexing_slicing,
reason = "Stack length is 0x1_0000, this cannot panic"
)]
kernel_esp_top: VirtAddr::from_ptr(
addr_of!(kernel_stack[KSTACK_SIZE - 1]).wrapping_add(1),
),
kernel_stack,
address_space: Some(address_space),
address_spaces: Mutex::new(Slab::new()),
data_buffers: Mutex::new(Slab::new()),
message_queue: Mutex::new(SegQueue::new()),
sleeping: RwLock::new(Some(SleepReason::NewProcess)),
});
let process = Process::new(address_space, entry_point)?;
let pid = self.processes.write().insert(process);
if let Some(&proc_man_pid) = REGISTERD_PIDS.read().get(&3) {
let mut varint_buf = unsigned_varint::encode::u64_buffer();
let mut buffer = Vec::new_in(&*ACTIVE_SPACE);
@ -220,10 +279,21 @@ impl Tasking {
Ok(pid)
}
pub fn new_thread(&self, entry_point: *const extern "C" fn(usize) -> !, argument: usize) -> Result<usize, PagingError> {
let current_pid = self.current_pid().unwrap();
let new_tid = self.current_process_mut(|process| {
process.new_thread(entry_point, argument)
})?;
self.ready_to_run.lock().push_back((current_pid, new_tid));
Ok(new_tid)
}
pub fn ok_to_yield(&self) -> bool {
!(self.freeable_kstacks.is_locked()
|| (self.current_pid.reader_count() > 0)
|| (self.current_pid.writer_count() > 0)
|| (self.current_tid.reader_count() > 0)
|| (self.current_tid.writer_count() > 0)
|| self.ready_to_run.is_locked()
|| (self.processes.reader_count() > 0)
|| (self.processes.writer_count() > 0)
@ -238,15 +308,18 @@ impl Tasking {
self.wfi_loop.store(false, Ordering::Relaxed);
break;
};
let next_process_pid = self.ready_to_run.lock().pop_front();
if let Some(next_process_pid) = next_process_pid {
let rtr_head = self.ready_to_run.lock().pop_front();
if let Some((next_process_pid, next_process_tid)) = rtr_head {
self.wfi_loop.store(false, Ordering::Relaxed);
if Some(next_process_pid) == self.current_pid() {
println!("Yielding to current process! Returning");
if Some(next_process_pid) == self.current_pid()
&& Some(next_process_tid) == self.current_tid()
{
println!("Yielding to current thread! Returning");
break;
}
#[warn(clippy::unwrap_used, reason = "FIXME(?)")]
let current_address_space = self
if current_pid != next_process_pid {
let current_address_space = self
.process_mut(next_process_pid, |process| {
#[expect(
clippy::expect_used,
@ -259,22 +332,23 @@ impl Tasking {
.activate()
})
.unwrap();
self.current_process_mut(|process| {
process.address_space = Some(current_address_space);
});
if self.current_process(|process| process.sleeping.read().is_none()) {
self.ready_to_run.lock().push_back(current_pid);
self.current_process_mut(|process| {
process.address_space = Some(current_address_space);
});
}
let curr_stack =
self.current_process_mut(|process| addr_of_mut!(process.kernel_esp));
if self.current_thread(|thread| thread.sleeping.read().is_none()) {
self.ready_to_run.lock().push_back((current_pid, self.current_tid().unwrap()));
}
let curr_stack = self.current_thread_mut(|thread| addr_of_mut!(thread.kernel_esp));
*self.current_pid.write() = Some(next_process_pid);
let kernel_esp = self.current_process(|process| {
gdt::set_tss_stack(process.kernel_esp_top);
process.kernel_esp
*self.current_tid.write() = Some(next_process_tid);
let kernel_esp = self.current_thread(|thread| {
gdt::set_tss_stack(thread.kernel_esp_top);
thread.kernel_esp
});
switch_to_asm(curr_stack, kernel_esp);
break;
} else if self.current_process(|process| process.sleeping.read().is_some()) {
} else if self.current_thread(|thread| thread.sleeping.read().is_some()) {
self.wfi_loop.store(true, Ordering::Relaxed);
x86_64::instructions::interrupts::enable_and_hlt();
x86_64::instructions::interrupts::disable();
@ -289,8 +363,13 @@ impl Tasking {
*self.current_pid.read()
}
pub fn current_tid(&self) -> Option<usize> {
*self.current_tid.read()
}
pub fn exit(&self, code: u8) -> ! {
if let Some(current_pid) = self.current_pid() {
self.ready_to_run.lock().retain(|&(ent_pid, _ent_tid)| ent_pid != current_pid);
if let Some(&proc_man_pid) = REGISTERD_PIDS.read().get(&3) {
let mut varint_buf = unsigned_varint::encode::u64_buffer();
let mut buffer = Vec::new_in(&*ACTIVE_SPACE);
@ -321,17 +400,18 @@ impl Tasking {
}
}
loop {
let next_process_pid = self.ready_to_run.lock().pop_front();
if let Some(next_process_pid) = next_process_pid {
let rtr_head = self.ready_to_run.lock().pop_front();
if let Some((next_process_pid, next_process_tid)) = rtr_head {
self.wfi_loop.store(false, Ordering::Relaxed);
#[warn(clippy::indexing_slicing, reason = "FIXME(?)")]
if self.current_pid.read().is_some() {
self.current_process(|process| {
self.current_thread(|process| {
*process.sleeping.write() = Some(SleepReason::Exited);
});
}
*self.current_pid.write() = Some(next_process_pid);
let kernel_esp = self
*self.current_tid.write() = Some(next_process_tid);
self
.current_process_mut(|process| {
#[expect(
clippy::expect_used,
@ -342,9 +422,11 @@ impl Tasking {
.take()
.expect("Non-current process has active page table")
.activate();
gdt::set_tss_stack(process.kernel_esp_top);
process.kernel_esp
});
let kernel_esp = self.current_thread_mut(|thread| {
gdt::set_tss_stack(thread.kernel_esp_top);
thread.kernel_esp
});
switch_to_asm_exit(kernel_esp);
unreachable!()
} else {
@ -360,14 +442,29 @@ impl Tasking {
return Err(());
}
let mut processes = self.processes.write();
let process = processes.get(pid).ok_or(())?;
if process.sleeping.read().is_none() {
self.ready_to_run.lock().retain(|e| *e != pid);
let process = processes.remove(pid);
self.ready_to_run.lock().retain(|&(ent_pid, _ent_tid)| ent_pid != pid);
let threads = process.threads.into_inner();
for (_, thread) in threads {
self.freeable_kstacks.lock().push(thread.kernel_stack);
}
self.freeable_kstacks.lock().push(processes.remove(pid).kernel_stack);
Ok(())
}
pub fn current_thread<F: FnOnce(&Thread) -> T, T>(&self, func: F) -> T {
self.current_process(|process| {
let threads = process.threads.read();
func(&threads[self.current_tid().unwrap()])
})
}
pub fn current_thread_mut<F: FnOnce(&mut Thread) -> T, T>(&self, func: F) -> T {
self.current_process_mut(|process| {
let mut threads = process.threads.write();
func(&mut threads[self.current_tid().unwrap()])
})
}
pub fn current_process<F: FnOnce(&Process) -> T, T>(&self, func: F) -> T {
let processes = self.processes.read();
#[warn(clippy::unwrap_used, reason = "FIXME")]
@ -382,6 +479,16 @@ impl Tasking {
func(&mut processes[self.current_pid().unwrap()])
}
pub fn process<F: FnOnce(&Process) -> T, T>(
&self,
pid: usize,
func: F,
) -> Result<T, InvalidPid> {
let processes = self.processes.read();
#[warn(clippy::unwrap_used, reason = "FIXME")]
Ok(func(processes.get(pid).ok_or(InvalidPid)?))
}
pub fn process_mut<F: FnOnce(&mut Process) -> T, T>(
&self,
pid: usize,
@ -395,19 +502,21 @@ impl Tasking {
pub fn sleep(&self, reason: SleepReason) {
#[warn(clippy::unwrap_used, reason = "FIXME")]
#[warn(clippy::indexing_slicing, reason = "FIXME(?)")]
self.current_process(|process| {
*process.sleeping.write() = Some(reason);
self.current_thread(|thread| {
*thread.sleeping.write() = Some(reason);
});
self.task_yield();
}
pub fn wake(&self, pid: usize, reason: SleepReason) -> Result<(), InvalidPid> {
pub fn wake(&self, pid: usize, tid: usize, reason: SleepReason) -> Result<(), InvalidPid> {
let processes = self.processes.read();
let process = processes.get(pid).ok_or(InvalidPid)?;
let mut sleeping = process.sleeping.write();
let threads = process.threads.read();
let thread = threads.get(tid).ok_or(InvalidPid)?;
let mut sleeping = thread.sleeping.write();
if *sleeping == Some(reason) {
if Some(pid) != self.current_pid() {
self.ready_to_run.lock().push_back(pid);
if Some(pid) != self.current_pid() || Some(tid) != self.current_tid() {
self.ready_to_run.lock().push_back((pid, tid));
}
*sleeping = None;
}