rust/src/rt/rust_kernel.cpp

234 lines
6.1 KiB
C++
Raw Normal View History

#include "rust_internal.h"
rust_kernel::rust_kernel(rust_srv *srv) :
_region(&srv->local_region),
_log(srv, NULL),
_srv(srv),
_interrupt_kernel_loop(FALSE),
domains(&srv->local_region),
message_queues(&srv->local_region) {
// Nop.
}
rust_handle<rust_dom> *
rust_kernel::create_domain(const rust_crate *crate, const char *name) {
2010-09-08 17:16:14 -05:00
LOCK(_kernel_lock);
rust_message_queue *message_queue =
new (this) rust_message_queue(_srv, this);
rust_srv *srv = _srv->clone();
rust_dom *dom =
new (this) rust_dom(this, message_queue, srv, crate, name);
2010-09-08 17:16:14 -05:00
rust_handle<rust_dom> *handle = internal_get_dom_handle(dom);
message_queue->associate(handle);
domains.append(dom);
message_queues.append(message_queue);
2010-09-08 17:16:14 -05:00
UNLOCK(_kernel_lock);
return handle;
}
void
rust_kernel::destroy_domain(rust_dom *dom) {
2010-09-08 17:16:14 -05:00
LOCK(_kernel_lock);
log(rust_log::KERN, "deleting domain: " PTR ", index: %d, domains %d",
dom, dom->list_index, domains.length());
domains.remove(dom);
dom->message_queue->disassociate();
rust_srv *srv = dom->srv;
delete dom;
delete srv;
2010-09-08 17:16:14 -05:00
UNLOCK(_kernel_lock);
}
rust_handle<rust_dom> *
2010-09-08 17:16:14 -05:00
rust_kernel::internal_get_dom_handle(rust_dom *dom) {
rust_handle<rust_dom> *handle = NULL;
2010-09-08 17:16:14 -05:00
if (_dom_handles.get(dom, &handle) == false) {
handle =
new (this) rust_handle<rust_dom>(this, dom->message_queue, dom);
_dom_handles.put(dom, handle);
}
2010-09-08 17:16:14 -05:00
return handle;
}
rust_handle<rust_dom> *
rust_kernel::get_dom_handle(rust_dom *dom) {
LOCK(_kernel_lock);
rust_handle<rust_dom> *handle = internal_get_dom_handle(dom);
UNLOCK(_kernel_lock);
return handle;
}
rust_handle<rust_task> *
rust_kernel::get_task_handle(rust_task *task) {
2010-09-08 17:16:14 -05:00
LOCK(_kernel_lock);
rust_handle<rust_task> *handle = NULL;
2010-09-08 17:16:14 -05:00
if (_task_handles.get(task, &handle) == false) {
handle =
new (this) rust_handle<rust_task>(this, task->dom->message_queue,
task);
_task_handles.put(task, handle);
}
2010-09-08 17:16:14 -05:00
UNLOCK(_kernel_lock);
return handle;
}
rust_handle<rust_port> *
rust_kernel::get_port_handle(rust_port *port) {
2010-09-08 17:16:14 -05:00
PLOCK(_kernel_lock);
rust_handle<rust_port> *handle = NULL;
2010-09-08 17:16:14 -05:00
if (_port_handles.get(port, &handle) == false) {
handle =
new (this) rust_handle<rust_port>(this,
port->task->dom->message_queue,
port);
_port_handles.put(port, handle);
}
2010-09-08 17:16:14 -05:00
PUNLOCK(_kernel_lock);
return handle;
}
void
rust_kernel::join_all_domains() {
// TODO: Perhaps we can do this a little smarter. Just spin wait for now.
while (domains.length() > 0) {
sync::yield();
}
2010-09-08 17:16:14 -05:00
log(rust_log::KERN, "joined domains");
}
void
rust_kernel::log_all_domain_state() {
log(rust_log::KERN, "log_all_domain_state: %d domains", domains.length());
for (uint32_t i = 0; i < domains.length(); i++) {
domains[i]->log_state();
}
}
/**
* Checks for simple deadlocks.
*/
bool
rust_kernel::is_deadlocked() {
return false;
}
void
rust_kernel::log(uint32_t type_bits, char const *fmt, ...) {
char buf[256];
if (_log.is_tracing(type_bits)) {
va_list args;
va_start(args, fmt);
vsnprintf(buf, sizeof(buf), fmt, args);
_log.trace_ln(NULL, type_bits, buf);
va_end(args);
}
}
void
2010-09-08 01:37:51 -05:00
rust_kernel::pump_message_queues() {
2010-09-08 17:16:14 -05:00
LOCK(_kernel_lock);
2010-09-08 01:37:51 -05:00
for (size_t i = 0; i < message_queues.length(); i++) {
rust_message_queue *queue = message_queues[i];
if (queue->is_associated() == false) {
rust_message *message = NULL;
while (queue->dequeue(&message)) {
message->kernel_process();
delete message;
}
}
2010-09-08 01:37:51 -05:00
}
2010-09-08 17:16:14 -05:00
UNLOCK(_kernel_lock);
2010-09-08 01:37:51 -05:00
}
void
rust_kernel::start_kernel_loop() {
while (_interrupt_kernel_loop == false) {
pump_message_queues();
// FIXME: this is a complete hack to make the testsuite finish in a
// sane time when executing under valgrind. The whole message-loop
// system here needs replacement with an OS-level event-queue such
// that actually wait on inter-thread notices, rather than
// busy-waiting.
size_t ms = TIME_SLICE_IN_MS;
#if defined(__WIN32__)
Sleep(ms);
#else
usleep(ms * 1000);
#endif
}
}
void
rust_kernel::run() {
log(rust_log::KERN, "started kernel loop");
start_kernel_loop();
log(rust_log::KERN, "finished kernel loop");
}
2010-09-08 01:37:51 -05:00
void
rust_kernel::terminate_kernel_loop() {
2010-09-08 17:16:14 -05:00
log(rust_log::KERN, "terminating kernel loop");
2010-09-08 01:37:51 -05:00
_interrupt_kernel_loop = true;
join();
}
rust_kernel::~rust_kernel() {
K(_srv, domains.length() == 0,
"Kernel has %d live domain(s), join all domains before killing "
"the kernel.", domains.length());
2010-09-08 01:37:51 -05:00
terminate_kernel_loop();
// It's possible that the message pump misses some messages because
// of races, so pump any remaining messages here. By now all domain
// threads should have been joined, so we shouldn't miss any more
// messages.
pump_message_queues();
2010-09-08 17:16:14 -05:00
log(rust_log::KERN, "freeing handles");
free_handles(_task_handles);
free_handles(_port_handles);
free_handles(_dom_handles);
2010-09-08 17:16:14 -05:00
log(rust_log::KERN, "freeing queues");
rust_message_queue *queue = NULL;
while (message_queues.pop(&queue)) {
K(_srv, queue->is_empty(), "Kernel message queue should be empty "
"before killing the kernel.");
delete queue;
}
}
void *
rust_kernel::malloc(size_t size) {
return _region->malloc(size);
}
void rust_kernel::free(void *mem) {
_region->free(mem);
}
template<class T> void
rust_kernel::free_handles(hash_map<T*, rust_handle<T>* > &map) {
T* key;
rust_handle<T> *value;
while (map.pop(&key, &value)) {
delete value;
}
}
//
// Local Variables:
// mode: C++
// fill-column: 78;
// indent-tabs-mode: nil
// c-basic-offset: 4
// buffer-file-coding-system: utf-8-unix
// compile-command: "make -k -C .. 2>&1 | sed -e 's/\\/x\\//x:\\//g'";
// End:
//