133 lines
5.8 KiB
Rust
133 lines
5.8 KiB
Rust
//! Job management on Windows for bootstrapping
|
|
//!
|
|
//! Most of the time when you're running a build system (e.g., make) you expect
|
|
//! Ctrl-C or abnormal termination to actually terminate the entire tree of
|
|
//! process in play, not just the one at the top. This currently works "by
|
|
//! default" on Unix platforms because Ctrl-C actually sends a signal to the
|
|
//! *process group* rather than the parent process, so everything will get torn
|
|
//! down. On Windows, however, this does not happen and Ctrl-C just kills the
|
|
//! parent process.
|
|
//!
|
|
//! To achieve the same semantics on Windows we use Job Objects to ensure that
|
|
//! all processes die at the same time. Job objects have a mode of operation
|
|
//! where when all handles to the object are closed it causes all child
|
|
//! processes associated with the object to be terminated immediately.
|
|
//! Conveniently whenever a process in the job object spawns a new process the
|
|
//! child will be associated with the job object as well. This means if we add
|
|
//! ourselves to the job object we create then everything will get torn down!
|
|
//!
|
|
//! Unfortunately most of the time the build system is actually called from a
|
|
//! python wrapper (which manages things like building the build system) so this
|
|
//! all doesn't quite cut it so far. To go the last mile we duplicate the job
|
|
//! object handle into our parent process (a python process probably) and then
|
|
//! close our own handle. This means that the only handle to the job object
|
|
//! resides in the parent python process, so when python dies the whole build
|
|
//! system dies (as one would probably expect!).
|
|
//!
|
|
//! Note that this module has a #[cfg(windows)] above it as none of this logic
|
|
//! is required on Unix.
|
|
|
|
#![allow(nonstandard_style, dead_code)]
|
|
|
|
use crate::Build;
|
|
use std::env;
|
|
use std::io;
|
|
use std::mem;
|
|
use std::ptr;
|
|
|
|
use winapi::shared::minwindef::{DWORD, FALSE, LPVOID};
|
|
use winapi::um::errhandlingapi::SetErrorMode;
|
|
use winapi::um::handleapi::{CloseHandle, DuplicateHandle};
|
|
use winapi::um::jobapi2::{AssignProcessToJobObject, CreateJobObjectW, SetInformationJobObject};
|
|
use winapi::um::processthreadsapi::{GetCurrentProcess, OpenProcess};
|
|
use winapi::um::winbase::{BELOW_NORMAL_PRIORITY_CLASS, SEM_NOGPFAULTERRORBOX};
|
|
use winapi::um::winnt::{
|
|
JobObjectExtendedLimitInformation, DUPLICATE_SAME_ACCESS, JOBOBJECT_EXTENDED_LIMIT_INFORMATION,
|
|
JOB_OBJECT_LIMIT_KILL_ON_JOB_CLOSE, JOB_OBJECT_LIMIT_PRIORITY_CLASS, PROCESS_DUP_HANDLE,
|
|
};
|
|
|
|
pub unsafe fn setup(build: &mut Build) {
|
|
// Enable the Windows Error Reporting dialog which msys disables,
|
|
// so we can JIT debug rustc
|
|
let mode = SetErrorMode(0);
|
|
SetErrorMode(mode & !SEM_NOGPFAULTERRORBOX);
|
|
|
|
// Create a new job object for us to use
|
|
let job = CreateJobObjectW(ptr::null_mut(), ptr::null());
|
|
assert!(!job.is_null(), "{}", io::Error::last_os_error());
|
|
|
|
// Indicate that when all handles to the job object are gone that all
|
|
// process in the object should be killed. Note that this includes our
|
|
// entire process tree by default because we've added ourselves and our
|
|
// children will reside in the job by default.
|
|
let mut info = mem::zeroed::<JOBOBJECT_EXTENDED_LIMIT_INFORMATION>();
|
|
info.BasicLimitInformation.LimitFlags = JOB_OBJECT_LIMIT_KILL_ON_JOB_CLOSE;
|
|
if build.config.low_priority {
|
|
info.BasicLimitInformation.LimitFlags |= JOB_OBJECT_LIMIT_PRIORITY_CLASS;
|
|
info.BasicLimitInformation.PriorityClass = BELOW_NORMAL_PRIORITY_CLASS;
|
|
}
|
|
let r = SetInformationJobObject(
|
|
job,
|
|
JobObjectExtendedLimitInformation,
|
|
&mut info as *mut _ as LPVOID,
|
|
mem::size_of_val(&info) as DWORD,
|
|
);
|
|
assert!(r != 0, "{}", io::Error::last_os_error());
|
|
|
|
// Assign our process to this job object. Note that if this fails, one very
|
|
// likely reason is that we are ourselves already in a job object! This can
|
|
// happen on the build bots that we've got for Windows, or if just anyone
|
|
// else is instrumenting the build. In this case we just bail out
|
|
// immediately and assume that they take care of it.
|
|
//
|
|
// Also note that nested jobs (why this might fail) are supported in recent
|
|
// versions of Windows, but the version of Windows that our bots are running
|
|
// at least don't support nested job objects.
|
|
let r = AssignProcessToJobObject(job, GetCurrentProcess());
|
|
if r == 0 {
|
|
CloseHandle(job);
|
|
return;
|
|
}
|
|
|
|
// If we've got a parent process (e.g., the python script that called us)
|
|
// then move ownership of this job object up to them. That way if the python
|
|
// script is killed (e.g., via ctrl-c) then we'll all be torn down.
|
|
//
|
|
// If we don't have a parent (e.g., this was run directly) then we
|
|
// intentionally leak the job object handle. When our process exits
|
|
// (normally or abnormally) it will close the handle implicitly, causing all
|
|
// processes in the job to be cleaned up.
|
|
let pid = match env::var("BOOTSTRAP_PARENT_ID") {
|
|
Ok(s) => s,
|
|
Err(..) => return,
|
|
};
|
|
|
|
let parent = OpenProcess(PROCESS_DUP_HANDLE, FALSE, pid.parse().unwrap());
|
|
assert!(
|
|
!parent.is_null(),
|
|
"PID `{}` doesn't seem to exist: {}",
|
|
pid,
|
|
io::Error::last_os_error()
|
|
);
|
|
let mut parent_handle = ptr::null_mut();
|
|
let r = DuplicateHandle(
|
|
GetCurrentProcess(),
|
|
job,
|
|
parent,
|
|
&mut parent_handle,
|
|
0,
|
|
FALSE,
|
|
DUPLICATE_SAME_ACCESS,
|
|
);
|
|
|
|
// If this failed, well at least we tried! An example of DuplicateHandle
|
|
// failing in the past has been when the wrong python2 package spawned this
|
|
// build system (e.g., the `python2` package in MSYS instead of
|
|
// `mingw-w64-x86_64-python2`. Not sure why it failed, but the "failure
|
|
// mode" here is that we only clean everything up when the build system
|
|
// dies, not when the python parent does, so not too bad.
|
|
if r != 0 {
|
|
CloseHandle(job);
|
|
}
|
|
}
|