// Copyright 2013-2014 The Rust Project Developers. See the COPYRIGHT // file at the top-level directory of this distribution and at // http://rust-lang.org/COPYRIGHT. // // Licensed under the Apache License, Version 2.0 or the MIT license // , at your // option. This file may not be copied, modified, or distributed // except according to those terms. //! Language-level runtime services that should reasonably expected //! to be available 'everywhere'. Local heaps, GC, unwinding, //! local storage, and logging. Even a 'freestanding' Rust would likely want //! to implement this. use core::prelude::*; use alloc::arc::Arc; use alloc::owned::{AnyOwnExt, Box}; use core::any::Any; use core::atomics::{AtomicUint, SeqCst}; use core::finally::Finally; use core::iter::Take; use core::mem; use core::raw; use local_data; use Runtime; use local::Local; use local_heap::LocalHeap; use rtio::LocalIo; use unwind::Unwinder; use collections::str::SendStr; /// The Task struct represents all state associated with a rust /// task. There are at this point two primary "subtypes" of task, /// however instead of using a subtype we just have a "task_type" field /// in the struct. This contains a pointer to another struct that holds /// the type-specific state. pub struct Task { pub heap: LocalHeap, pub gc: GarbageCollector, pub storage: LocalStorage, pub unwinder: Unwinder, pub death: Death, pub destroyed: bool, pub name: Option, imp: Option>, } pub struct TaskOpts { /// Invoke this procedure with the result of the task when it finishes. pub on_exit: Option, /// A name for the task-to-be, for identification in failure messages pub name: Option, /// The size of the stack for the spawned task pub stack_size: Option, } /// Indicates the manner in which a task exited. /// /// A task that completes without failing is considered to exit successfully. /// /// If you wish for this result's delivery to block until all /// children tasks complete, recommend using a result future. pub type Result = ::core::result::Result<(), Box>; pub struct GarbageCollector; pub struct LocalStorage(pub Option); /// A handle to a blocked task. Usually this means having the Box /// pointer by ownership, but if the task is killable, a killer can steal it /// at any time. pub enum BlockedTask { Owned(Box), Shared(Arc), } /// Per-task state related to task death, killing, failure, etc. pub struct Death { pub on_exit: Option, } pub struct BlockedTasks { inner: Arc, } impl Task { pub fn new() -> Task { Task { heap: LocalHeap::new(), gc: GarbageCollector, storage: LocalStorage(None), unwinder: Unwinder::new(), death: Death::new(), destroyed: false, name: None, imp: None, } } /// Executes the given closure as if it's running inside this task. The task /// is consumed upon entry, and the destroyed task is returned from this /// function in order for the caller to free. This function is guaranteed to /// not unwind because the closure specified is run inside of a `rust_try` /// block. (this is the only try/catch block in the world). /// /// This function is *not* meant to be abused as a "try/catch" block. This /// is meant to be used at the absolute boundaries of a task's lifetime, and /// only for that purpose. pub fn run(~self, mut f: ||) -> Box { // Need to put ourselves into TLS, but also need access to the unwinder. // Unsafely get a handle to the task so we can continue to use it after // putting it in tls (so we can invoke the unwinder). let handle: *mut Task = unsafe { *mem::transmute::<&Box, &*mut Task>(&self) }; Local::put(self); // The only try/catch block in the world. Attempt to run the task's // client-specified code and catch any failures. let try_block = || { // Run the task main function, then do some cleanup. f.finally(|| { // First, destroy task-local storage. This may run user dtors. // // FIXME #8302: Dear diary. I'm so tired and confused. // There's some interaction in rustc between the box // annihilator and the TLS dtor by which TLS is // accessed from annihilated box dtors *after* TLS is // destroyed. Somehow setting TLS back to null, as the // old runtime did, makes this work, but I don't currently // understand how. I would expect that, if the annihilator // reinvokes TLS while TLS is uninitialized, that // TLS would be reinitialized but never destroyed, // but somehow this works. I have no idea what's going // on but this seems to make things magically work. FML. // // (added after initial comment) A possible interaction here is // that the destructors for the objects in TLS themselves invoke // TLS, or possibly some destructors for those objects being // annihilated invoke TLS. Sadly these two operations seemed to // be intertwined, and miraculously work for now... drop({ let mut task = Local::borrow(None::); let &LocalStorage(ref mut optmap) = &mut task.storage; optmap.take() }); // Destroy remaining boxes. Also may run user dtors. let mut heap = { let mut task = Local::borrow(None::); mem::replace(&mut task.heap, LocalHeap::new()) }; unsafe { heap.annihilate() } drop(heap); }) }; unsafe { (*handle).unwinder.try(try_block); } // Here we must unsafely borrow the task in order to not remove it from // TLS. When collecting failure, we may attempt to send on a channel (or // just run arbitrary code), so we must be sure to still have a local // task in TLS. unsafe { let me: *mut Task = Local::unsafe_borrow(); (*me).death.collect_failure((*me).unwinder.result()); } let mut me: Box = Local::take(); me.destroyed = true; return me; } /// Inserts a runtime object into this task, transferring ownership to the /// task. It is illegal to replace a previous runtime object in this task /// with this argument. pub fn put_runtime(&mut self, ops: Box) { assert!(self.imp.is_none()); self.imp = Some(ops); } /// Attempts to extract the runtime as a specific type. If the runtime does /// not have the provided type, then the runtime is not removed. If the /// runtime does have the specified type, then it is removed and returned /// (transfer of ownership). /// /// It is recommended to only use this method when *absolutely necessary*. /// This function may not be available in the future. pub fn maybe_take_runtime(&mut self) -> Option> { // This is a terrible, terrible function. The general idea here is to // take the runtime, cast it to Box, check if it has the right // type, and then re-cast it back if necessary. The method of doing // this is pretty sketchy and involves shuffling vtables of trait // objects around, but it gets the job done. // // FIXME: This function is a serious code smell and should be avoided at // all costs. I have yet to think of a method to avoid this // function, and I would be saddened if more usage of the function // crops up. unsafe { let imp = self.imp.take_unwrap(); let vtable = mem::transmute::<_, &raw::TraitObject>(&imp).vtable; match imp.wrap().move::() { Ok(t) => Some(t), Err(t) => { let data = mem::transmute::<_, raw::TraitObject>(t).data; let obj: Box = mem::transmute(raw::TraitObject { vtable: vtable, data: data, }); self.put_runtime(obj); None } } } } /// Spawns a sibling to this task. The newly spawned task is configured with /// the `opts` structure and will run `f` as the body of its code. pub fn spawn_sibling(mut ~self, opts: TaskOpts, f: proc(): Send) { let ops = self.imp.take_unwrap(); ops.spawn_sibling(self, opts, f) } /// Deschedules the current task, invoking `f` `amt` times. It is not /// recommended to use this function directly, but rather communication /// primitives in `std::comm` should be used. pub fn deschedule(mut ~self, amt: uint, f: |BlockedTask| -> ::core::result::Result<(), BlockedTask>) { let ops = self.imp.take_unwrap(); ops.deschedule(amt, self, f) } /// Wakes up a previously blocked task, optionally specifying whether the /// current task can accept a change in scheduling. This function can only /// be called on tasks that were previously blocked in `deschedule`. pub fn reawaken(mut ~self) { let ops = self.imp.take_unwrap(); ops.reawaken(self); } /// Yields control of this task to another task. This function will /// eventually return, but possibly not immediately. This is used as an /// opportunity to allow other tasks a chance to run. pub fn yield_now(mut ~self) { let ops = self.imp.take_unwrap(); ops.yield_now(self); } /// Similar to `yield_now`, except that this function may immediately return /// without yielding (depending on what the runtime decides to do). pub fn maybe_yield(mut ~self) { let ops = self.imp.take_unwrap(); ops.maybe_yield(self); } /// Acquires a handle to the I/O factory that this task contains, normally /// stored in the task's runtime. This factory may not always be available, /// which is why the return type is `Option` pub fn local_io<'a>(&'a mut self) -> Option> { self.imp.get_mut_ref().local_io() } /// Returns the stack bounds for this task in (lo, hi) format. The stack /// bounds may not be known for all tasks, so the return value may be /// `None`. pub fn stack_bounds(&self) -> (uint, uint) { self.imp.get_ref().stack_bounds() } /// Returns whether it is legal for this task to block the OS thread that it /// is running on. pub fn can_block(&self) -> bool { self.imp.get_ref().can_block() } } impl Drop for Task { fn drop(&mut self) { rtdebug!("called drop for a task: {}", self as *mut Task as uint); rtassert!(self.destroyed); } } impl TaskOpts { pub fn new() -> TaskOpts { TaskOpts { on_exit: None, name: None, stack_size: None } } } impl Iterator for BlockedTasks { fn next(&mut self) -> Option { Some(Shared(self.inner.clone())) } } impl BlockedTask { /// Returns Some if the task was successfully woken; None if already killed. pub fn wake(self) -> Option> { match self { Owned(task) => Some(task), Shared(arc) => { match arc.swap(0, SeqCst) { 0 => None, n => Some(unsafe { mem::transmute(n) }), } } } } /// Reawakens this task if ownership is acquired. If finer-grained control /// is desired, use `wake` instead. pub fn reawaken(self) { self.wake().map(|t| t.reawaken()); } // This assertion has two flavours because the wake involves an atomic op. // In the faster version, destructors will fail dramatically instead. #[cfg(not(test))] pub fn trash(self) { } #[cfg(test)] pub fn trash(self) { assert!(self.wake().is_none()); } /// Create a blocked task, unless the task was already killed. pub fn block(task: Box) -> BlockedTask { Owned(task) } /// Converts one blocked task handle to a list of many handles to the same. pub fn make_selectable(self, num_handles: uint) -> Take { let arc = match self { Owned(task) => { let flag = unsafe { AtomicUint::new(mem::transmute(task)) }; Arc::new(flag) } Shared(arc) => arc.clone(), }; BlockedTasks{ inner: arc }.take(num_handles) } /// Convert to an unsafe uint value. Useful for storing in a pipe's state /// flag. #[inline] pub unsafe fn cast_to_uint(self) -> uint { match self { Owned(task) => { let blocked_task_ptr: uint = mem::transmute(task); rtassert!(blocked_task_ptr & 0x1 == 0); blocked_task_ptr } Shared(arc) => { let blocked_task_ptr: uint = mem::transmute(box arc); rtassert!(blocked_task_ptr & 0x1 == 0); blocked_task_ptr | 0x1 } } } /// Convert from an unsafe uint value. Useful for retrieving a pipe's state /// flag. #[inline] pub unsafe fn cast_from_uint(blocked_task_ptr: uint) -> BlockedTask { if blocked_task_ptr & 0x1 == 0 { Owned(mem::transmute(blocked_task_ptr)) } else { let ptr: Box> = mem::transmute(blocked_task_ptr & !1); Shared(*ptr) } } } impl Death { pub fn new() -> Death { Death { on_exit: None, } } /// Collect failure exit codes from children and propagate them to a parent. pub fn collect_failure(&mut self, result: Result) { match self.on_exit.take() { Some(f) => f(result), None => {} } } } impl Drop for Death { fn drop(&mut self) { // make this type noncopyable } } #[cfg(test)] mod test { use super::*; use std::prelude::*; use std::task; use std::gc::{Gc, GC}; #[test] fn local_heap() { let a = box(GC) 5; let b = a; assert!(*a == 5); assert!(*b == 5); } #[test] fn tls() { local_data_key!(key: Gc) key.replace(Some(box(GC) "data".to_string())); assert_eq!(key.get().unwrap().as_slice(), "data"); local_data_key!(key2: Gc) key2.replace(Some(box(GC) "data".to_string())); assert_eq!(key2.get().unwrap().as_slice(), "data"); } #[test] fn unwind() { let result = task::try(proc()()); rtdebug!("trying first assert"); assert!(result.is_ok()); let result = task::try::<()>(proc() fail!()); rtdebug!("trying second assert"); assert!(result.is_err()); } #[test] fn rng() { use std::rand::{StdRng, Rng}; let mut r = StdRng::new().ok().unwrap(); let _ = r.next_u32(); } #[test] fn comm_stream() { let (tx, rx) = channel(); tx.send(10); assert!(rx.recv() == 10); } #[test] fn comm_shared_chan() { let (tx, rx) = channel(); tx.send(10); assert!(rx.recv() == 10); } #[test] fn heap_cycles() { use std::cell::RefCell; struct List { next: Option>>, } let a = box(GC) RefCell::new(List { next: None }); let b = box(GC) RefCell::new(List { next: Some(a) }); { let mut a = a.borrow_mut(); a.next = Some(b); } } #[test] #[should_fail] fn test_begin_unwind() { use std::rt::unwind::begin_unwind; begin_unwind("cause", file!(), line!()) } // Task blocking tests #[test] fn block_and_wake() { let task = box Task::new(); let mut task = BlockedTask::block(task).wake().unwrap(); task.destroyed = true; } }