// Copyright 2012-2013 The Rust Project Developers. See the COPYRIGHT // file at the top-level directory of this distribution and at // http://rust-lang.org/COPYRIGHT. // // Licensed under the Apache License, Version 2.0 or the MIT license // , at your // option. This file may not be copied, modified, or distributed // except according to those terms. #[doc(hidden)]; // Support code for rustc's built in test runner generator. Currently, // none of this is meant for users. It is intended to support the // simplest interface possible for representing and running tests // while providing a base that other test frameworks may build off of. use getopts; use getopts::groups; use json::ToJson; use json; use serialize::Decodable; use sort; use stats::Stats; use stats; use term; use time::precise_time_ns; use treemap::TreeMap; use std::clone::Clone; use std::comm::{stream, SharedChan}; use std::libc; use std::either; use std::io; use std::result; use std::task; use std::to_str::ToStr; use std::u64; use std::f64; use std::os; // The name of a test. By convention this follows the rules for rust // paths; i.e. it should be a series of identifiers separated by double // colons. This way if some test runner wants to arrange the tests // hierarchically it may. #[deriving(Clone)] pub enum TestName { StaticTestName(&'static str), DynTestName(~str) } impl ToStr for TestName { fn to_str(&self) -> ~str { match (*self).clone() { StaticTestName(s) => s.to_str(), DynTestName(s) => s.to_str() } } } // A function that runs a test. If the function returns successfully, // the test succeeds; if the function fails then the test fails. We // may need to come up with a more clever definition of test in order // to support isolation of tests into tasks. pub enum TestFn { StaticTestFn(extern fn()), StaticBenchFn(extern fn(&mut BenchHarness)), StaticMetricFn(~fn(&mut MetricMap)), DynTestFn(~fn()), DynMetricFn(~fn(&mut MetricMap)), DynBenchFn(~fn(&mut BenchHarness)) } // Structure passed to BenchFns pub struct BenchHarness { iterations: u64, ns_start: u64, ns_end: u64, bytes: u64 } // The definition of a single test. A test runner will run a list of // these. #[deriving(Clone)] pub struct TestDesc { name: TestName, ignore: bool, should_fail: bool } pub struct TestDescAndFn { desc: TestDesc, testfn: TestFn, } #[deriving(Clone, Encodable, Decodable, Eq)] pub struct Metric { value: f64, noise: f64 } #[deriving(Eq)] pub struct MetricMap(TreeMap<~str,Metric>); impl Clone for MetricMap { pub fn clone(&self) -> MetricMap { MetricMap((**self).clone()) } } /// Analysis of a single change in metric #[deriving(Eq)] pub enum MetricChange { LikelyNoise, MetricAdded, MetricRemoved, Improvement(f64), Regression(f64) } pub type MetricDiff = TreeMap<~str,MetricChange>; // The default console test runner. It accepts the command line // arguments and a vector of test_descs. pub fn test_main(args: &[~str], tests: ~[TestDescAndFn]) { let opts = match parse_opts(args) { either::Left(o) => o, either::Right(m) => fail!(m) }; if !run_tests_console(&opts, tests) { fail!("Some tests failed"); } } // A variant optimized for invocation with a static test vector. // This will fail (intentionally) when fed any dynamic tests, because // it is copying the static values out into a dynamic vector and cannot // copy dynamic values. It is doing this because from this point on // a ~[TestDescAndFn] is used in order to effect ownership-transfer // semantics into parallel test runners, which in turn requires a ~[] // rather than a &[]. pub fn test_main_static(args: &[~str], tests: &[TestDescAndFn]) { let owned_tests = do tests.map |t| { match t.testfn { StaticTestFn(f) => TestDescAndFn { testfn: StaticTestFn(f), desc: t.desc.clone() }, StaticBenchFn(f) => TestDescAndFn { testfn: StaticBenchFn(f), desc: t.desc.clone() }, _ => { fail!("non-static tests passed to test::test_main_static"); } } }; test_main(args, owned_tests) } pub struct TestOpts { filter: Option<~str>, run_ignored: bool, run_tests: bool, run_benchmarks: bool, ratchet_metrics: Option, ratchet_noise_percent: Option, save_metrics: Option, logfile: Option } type OptRes = Either; fn optgroups() -> ~[getopts::groups::OptGroup] { ~[groups::optflag("", "ignored", "Run ignored tests"), groups::optflag("", "test", "Run tests and not benchmarks"), groups::optflag("", "bench", "Run benchmarks instead of tests"), groups::optflag("h", "help", "Display this message (longer with --help)"), groups::optopt("", "save-metrics", "Location to save bench metrics", "PATH"), groups::optopt("", "ratchet-metrics", "Location to load and save metrics from. The metrics \ loaded are cause benchmarks to fail if they run too \ slowly", "PATH"), groups::optopt("", "ratchet-noise-percent", "Tests within N% of the recorded metrics will be \ considered as passing", "PERCENTAGE"), groups::optopt("", "logfile", "Write logs to the specified file instead \ of stdout", "PATH")] } fn usage(binary: &str, helpstr: &str) -> ! { let message = fmt!("Usage: %s [OPTIONS] [FILTER]", binary); println(groups::usage(message, optgroups())); if helpstr == "help" { println("\ The FILTER is matched against the name of all tests to run, and if any tests have a substring match, only those tests are run. By default, all tests are run in parallel. This can be altered with the RUST_THREADS environment variable when running tests (set it to 1). Test Attributes: #[test] - Indicates a function is a test to be run. This function takes no arguments. #[bench] - Indicates a function is a benchmark to be run. This function takes one argument (extra::test::BenchHarness). #[should_fail] - This function (also labeled with #[test]) will only pass if the code causes a failure (an assertion failure or fail!) #[ignore] - When applied to a function which is already attributed as a test, then the test runner will ignore these tests during normal test runs. Running with --ignored will run these tests. This may also be written as #[ignore(cfg(...))] to ignore the test on certain configurations."); } unsafe { libc::exit(0) } } // Parses command line arguments into test options pub fn parse_opts(args: &[~str]) -> OptRes { let args_ = args.tail(); let matches = match groups::getopts(args_, optgroups()) { Ok(m) => m, Err(f) => return either::Right(getopts::fail_str(f)) }; if getopts::opt_present(&matches, "h") { usage(args[0], "h"); } if getopts::opt_present(&matches, "help") { usage(args[0], "help"); } let filter = if matches.free.len() > 0 { Some((matches).free[0].clone()) } else { None }; let run_ignored = getopts::opt_present(&matches, "ignored"); let logfile = getopts::opt_maybe_str(&matches, "logfile"); let logfile = logfile.map(|s| Path(*s)); let run_benchmarks = getopts::opt_present(&matches, "bench"); let run_tests = ! run_benchmarks || getopts::opt_present(&matches, "test"); let ratchet_metrics = getopts::opt_maybe_str(&matches, "ratchet-metrics"); let ratchet_metrics = ratchet_metrics.map(|s| Path(*s)); let ratchet_noise_percent = getopts::opt_maybe_str(&matches, "ratchet-noise-percent"); let ratchet_noise_percent = ratchet_noise_percent.map(|s| f64::from_str(*s).get()); let save_metrics = getopts::opt_maybe_str(&matches, "save-metrics"); let save_metrics = save_metrics.map(|s| Path(*s)); let test_opts = TestOpts { filter: filter, run_ignored: run_ignored, run_tests: run_tests, run_benchmarks: run_benchmarks, ratchet_metrics: ratchet_metrics, ratchet_noise_percent: ratchet_noise_percent, save_metrics: save_metrics, logfile: logfile }; either::Left(test_opts) } #[deriving(Clone, Eq)] pub struct BenchSamples { ns_iter_summ: stats::Summary, mb_s: uint } #[deriving(Clone, Eq)] pub enum TestResult { TrOk, TrFailed, TrIgnored, TrMetrics(MetricMap), TrBench(BenchSamples), } struct ConsoleTestState { out: @io::Writer, log_out: Option<@io::Writer>, term: Option, use_color: bool, total: uint, passed: uint, failed: uint, ignored: uint, measured: uint, metrics: MetricMap, failures: ~[TestDesc] } impl ConsoleTestState { pub fn new(opts: &TestOpts) -> ConsoleTestState { let log_out = match opts.logfile { Some(ref path) => match io::file_writer(path, [io::Create, io::Truncate]) { result::Ok(w) => Some(w), result::Err(ref s) => { fail!("can't open output file: %s", *s) } }, None => None }; let out = io::stdout(); let term = match term::Terminal::new(out) { Err(_) => None, Ok(t) => Some(t) }; ConsoleTestState { out: out, log_out: log_out, use_color: use_color(), term: term, total: 0u, passed: 0u, failed: 0u, ignored: 0u, measured: 0u, metrics: MetricMap::new(), failures: ~[] } } pub fn write_ok(&self) { self.write_pretty("ok", term::color::GREEN); } pub fn write_failed(&self) { self.write_pretty("FAILED", term::color::RED); } pub fn write_ignored(&self) { self.write_pretty("ignored", term::color::YELLOW); } pub fn write_metric(&self) { self.write_pretty("metric", term::color::CYAN); } pub fn write_bench(&self) { self.write_pretty("bench", term::color::CYAN); } pub fn write_added(&self) { self.write_pretty("added", term::color::GREEN); } pub fn write_improved(&self) { self.write_pretty("improved", term::color::GREEN); } pub fn write_removed(&self) { self.write_pretty("removed", term::color::YELLOW); } pub fn write_regressed(&self) { self.write_pretty("regressed", term::color::RED); } pub fn write_pretty(&self, word: &str, color: term::color::Color) { match self.term { None => self.out.write_str(word), Some(ref t) => { if self.use_color { t.fg(color); } self.out.write_str(word); if self.use_color { t.reset(); } } } } pub fn write_run_start(&mut self, len: uint) { self.total = len; let noun = if len != 1 { &"tests" } else { &"test" }; self.out.write_line(fmt!("\nrunning %u %s", len, noun)); } pub fn write_test_start(&self, test: &TestDesc) { self.out.write_str(fmt!("test %s ... ", test.name.to_str())); } pub fn write_result(&self, result: &TestResult) { match *result { TrOk => self.write_ok(), TrFailed => self.write_failed(), TrIgnored => self.write_ignored(), TrMetrics(ref mm) => { self.write_metric(); self.out.write_str(": " + fmt_metrics(mm)); } TrBench(ref bs) => { self.write_bench(); self.out.write_str(": " + fmt_bench_samples(bs)) } } self.out.write_str(&"\n"); } pub fn write_log(&self, test: &TestDesc, result: &TestResult) { match self.log_out { None => (), Some(out) => { out.write_line(fmt!("%s %s", match *result { TrOk => ~"ok", TrFailed => ~"failed", TrIgnored => ~"ignored", TrMetrics(ref mm) => fmt_metrics(mm), TrBench(ref bs) => fmt_bench_samples(bs) }, test.name.to_str())); } } } pub fn write_failures(&self) { self.out.write_line("\nfailures:"); let mut failures = ~[]; for self.failures.iter().advance() |f| { failures.push(f.name.to_str()); } sort::tim_sort(failures); for failures.iter().advance |name| { self.out.write_line(fmt!(" %s", name.to_str())); } } pub fn write_metric_diff(&self, diff: &MetricDiff) { let mut noise = 0; let mut improved = 0; let mut regressed = 0; let mut added = 0; let mut removed = 0; for diff.iter().advance() |(k, v)| { match *v { LikelyNoise => noise += 1, MetricAdded => { added += 1; self.write_added(); self.out.write_line(fmt!(": %s", *k)); } MetricRemoved => { removed += 1; self.write_removed(); self.out.write_line(fmt!(": %s", *k)); } Improvement(pct) => { improved += 1; self.out.write_str(*k); self.out.write_str(": "); self.write_improved(); self.out.write_line(fmt!(" by %.2f%%", pct as float)) } Regression(pct) => { regressed += 1; self.out.write_str(*k); self.out.write_str(": "); self.write_regressed(); self.out.write_line(fmt!(" by %.2f%%", pct as float)) } } } self.out.write_line(fmt!("result of ratchet: %u matrics added, %u removed, \ %u improved, %u regressed, %u noise", added, removed, improved, regressed, noise)); if regressed == 0 { self.out.write_line("updated ratchet file") } else { self.out.write_line("left ratchet file untouched") } } pub fn write_run_finish(&self, ratchet_metrics: &Option, ratchet_pct: Option) -> bool { assert!(self.passed + self.failed + self.ignored + self.measured == self.total); let ratchet_success = match *ratchet_metrics { None => true, Some(ref pth) => { self.out.write_str(fmt!("\nusing metrics ratchet: %s\n", pth.to_str())); match ratchet_pct { None => (), Some(pct) => self.out.write_str(fmt!("with noise-tolerance forced to: %f%%\n", pct as float)) } let (diff, ok) = self.metrics.ratchet(pth, ratchet_pct); self.write_metric_diff(&diff); ok } }; let test_success = self.failed == 0u; if !test_success { self.write_failures(); } let success = ratchet_success && test_success; self.out.write_str("\ntest result: "); if success { // There's no parallelism at this point so it's safe to use color self.write_ok(); } else { self.write_failed(); } self.out.write_str(fmt!(". %u passed; %u failed; %u ignored; %u measured\n\n", self.passed, self.failed, self.ignored, self.measured)); return success; } } pub fn fmt_metrics(mm: &MetricMap) -> ~str { use std::iterator::IteratorUtil; let v : ~[~str] = mm.iter() .transform(|(k,v)| fmt!("%s: %f (+/- %f)", *k, v.value as float, v.noise as float)) .collect(); v.connect(", ") } pub fn fmt_bench_samples(bs: &BenchSamples) -> ~str { if bs.mb_s != 0 { fmt!("%u ns/iter (+/- %u) = %u MB/s", bs.ns_iter_summ.median as uint, (bs.ns_iter_summ.max - bs.ns_iter_summ.min) as uint, bs.mb_s) } else { fmt!("%u ns/iter (+/- %u)", bs.ns_iter_summ.median as uint, (bs.ns_iter_summ.max - bs.ns_iter_summ.min) as uint) } } // A simple console test runner pub fn run_tests_console(opts: &TestOpts, tests: ~[TestDescAndFn]) -> bool { fn callback(event: &TestEvent, st: &mut ConsoleTestState) { debug!("callback(event=%?)", event); match (*event).clone() { TeFiltered(ref filtered_tests) => st.write_run_start(filtered_tests.len()), TeWait(ref test) => st.write_test_start(test), TeResult(test, result) => { st.write_log(&test, &result); st.write_result(&result); match result { TrOk => st.passed += 1, TrIgnored => st.ignored += 1, TrMetrics(mm) => { let tname = test.name.to_str(); for mm.iter().advance() |(k,v)| { st.metrics.insert_metric(tname + "." + *k, v.value, v.noise); } st.measured += 1 } TrBench(bs) => { st.metrics.insert_metric(test.name.to_str(), bs.ns_iter_summ.median, bs.ns_iter_summ.max - bs.ns_iter_summ.min); st.measured += 1 } TrFailed => { st.failed += 1; st.failures.push(test); } } } } } let st = @mut ConsoleTestState::new(opts); run_tests(opts, tests, |x| callback(&x, st)); match opts.save_metrics { None => (), Some(ref pth) => { st.metrics.save(pth); st.out.write_str(fmt!("\nmetrics saved to: %s", pth.to_str())); } } return st.write_run_finish(&opts.ratchet_metrics, opts.ratchet_noise_percent); } #[test] fn should_sort_failures_before_printing_them() { fn dummy() {} let s = do io::with_str_writer |wr| { let test_a = TestDesc { name: StaticTestName("a"), ignore: false, should_fail: false }; let test_b = TestDesc { name: StaticTestName("b"), ignore: false, should_fail: false }; let st = @ConsoleTestState { out: wr, log_out: None, term: None, use_color: false, total: 0u, passed: 0u, failed: 0u, ignored: 0u, measured: 0u, metrics: MetricMap::new(), failures: ~[test_b, test_a] }; st.write_failures(); }; let apos = s.find_str("a").get(); let bpos = s.find_str("b").get(); assert!(apos < bpos); } fn use_color() -> bool { return get_concurrency() == 1; } #[deriving(Clone)] enum TestEvent { TeFiltered(~[TestDesc]), TeWait(TestDesc), TeResult(TestDesc, TestResult), } type MonitorMsg = (TestDesc, TestResult); fn run_tests(opts: &TestOpts, tests: ~[TestDescAndFn], callback: &fn(e: TestEvent)) { let filtered_tests = filter_tests(opts, tests); let filtered_descs = filtered_tests.map(|t| t.desc.clone()); callback(TeFiltered(filtered_descs)); let (filtered_tests, filtered_benchs_and_metrics) = do filtered_tests.partition |e| { match e.testfn { StaticTestFn(_) | DynTestFn(_) => true, _ => false } }; // It's tempting to just spawn all the tests at once, but since we have // many tests that run in other processes we would be making a big mess. let concurrency = get_concurrency(); debug!("using %u test tasks", concurrency); let mut remaining = filtered_tests; remaining.reverse(); let mut pending = 0; let (p, ch) = stream(); let ch = SharedChan::new(ch); while pending > 0 || !remaining.is_empty() { while pending < concurrency && !remaining.is_empty() { let test = remaining.pop(); if concurrency == 1 { // We are doing one test at a time so we can print the name // of the test before we run it. Useful for debugging tests // that hang forever. callback(TeWait(test.desc.clone())); } run_test(!opts.run_tests, test, ch.clone()); pending += 1; } let (desc, result) = p.recv(); if concurrency != 1 { callback(TeWait(desc.clone())); } callback(TeResult(desc, result)); pending -= 1; } // All benchmarks run at the end, in serial. // (this includes metric fns) for filtered_benchs_and_metrics.consume_iter().advance |b| { callback(TeWait(b.desc.clone())); run_test(!opts.run_benchmarks, b, ch.clone()); let (test, result) = p.recv(); callback(TeResult(test, result)); } } // Windows tends to dislike being overloaded with threads. #[cfg(windows)] static SCHED_OVERCOMMIT : uint = 1; #[cfg(unix)] static SCHED_OVERCOMMIT : uint = 4u; fn get_concurrency() -> uint { use std::rt; let threads = rt::util::default_sched_threads(); if threads == 1 { 1 } else { threads * SCHED_OVERCOMMIT } } pub fn filter_tests( opts: &TestOpts, tests: ~[TestDescAndFn]) -> ~[TestDescAndFn] { let mut filtered = tests; // Remove tests that don't match the test filter filtered = if opts.filter.is_none() { filtered } else { let filter_str = match opts.filter { Some(ref f) => (*f).clone(), None => ~"" }; fn filter_fn(test: TestDescAndFn, filter_str: &str) -> Option { if test.desc.name.to_str().contains(filter_str) { return Some(test); } else { return None; } } filtered.consume_iter().filter_map(|x| filter_fn(x, filter_str)).collect() }; // Maybe pull out the ignored test and unignore them filtered = if !opts.run_ignored { filtered } else { fn filter(test: TestDescAndFn) -> Option { if test.desc.ignore { let TestDescAndFn {desc, testfn} = test; Some(TestDescAndFn { desc: TestDesc {ignore: false, ..desc}, testfn: testfn }) } else { None } }; filtered.consume_iter().filter_map(|x| filter(x)).collect() }; // Sort the tests alphabetically fn lteq(t1: &TestDescAndFn, t2: &TestDescAndFn) -> bool { t1.desc.name.to_str() < t2.desc.name.to_str() } sort::quick_sort(filtered, lteq); filtered } struct TestFuture { test: TestDesc, wait: @fn() -> TestResult, } pub fn run_test(force_ignore: bool, test: TestDescAndFn, monitor_ch: SharedChan) { let TestDescAndFn {desc, testfn} = test; if force_ignore || desc.ignore { monitor_ch.send((desc, TrIgnored)); return; } fn run_test_inner(desc: TestDesc, monitor_ch: SharedChan, testfn: ~fn()) { let testfn_cell = ::std::cell::Cell::new(testfn); do task::spawn { let mut result_future = None; // task::future_result(builder); let mut task = task::task(); task.unlinked(); task.future_result(|r| { result_future = Some(r) }); task.spawn(testfn_cell.take()); let task_result = result_future.unwrap().recv(); let test_result = calc_result(&desc, task_result == task::Success); monitor_ch.send((desc.clone(), test_result)); } } match testfn { DynBenchFn(benchfn) => { let bs = ::test::bench::benchmark(benchfn); monitor_ch.send((desc, TrBench(bs))); return; } StaticBenchFn(benchfn) => { let bs = ::test::bench::benchmark(benchfn); monitor_ch.send((desc, TrBench(bs))); return; } DynMetricFn(f) => { let mut mm = MetricMap::new(); f(&mut mm); monitor_ch.send((desc, TrMetrics(mm))); return; } StaticMetricFn(f) => { let mut mm = MetricMap::new(); f(&mut mm); monitor_ch.send((desc, TrMetrics(mm))); return; } DynTestFn(f) => run_test_inner(desc, monitor_ch, f), StaticTestFn(f) => run_test_inner(desc, monitor_ch, || f()) } } fn calc_result(desc: &TestDesc, task_succeeded: bool) -> TestResult { if task_succeeded { if desc.should_fail { TrFailed } else { TrOk } } else { if desc.should_fail { TrOk } else { TrFailed } } } impl ToJson for Metric { fn to_json(&self) -> json::Json { let mut map = ~TreeMap::new(); map.insert(~"value", json::Number(self.value as float)); map.insert(~"noise", json::Number(self.noise as float)); json::Object(map) } } impl MetricMap { pub fn new() -> MetricMap { MetricMap(TreeMap::new()) } /// Load MetricDiff from a file. pub fn load(p: &Path) -> MetricMap { assert!(os::path_exists(p)); let f = io::file_reader(p).unwrap(); let mut decoder = json::Decoder(json::from_reader(f).get()); MetricMap(Decodable::decode(&mut decoder)) } /// Write MetricDiff to a file. pub fn save(&self, p: &Path) { let f = io::file_writer(p, [io::Create, io::Truncate]).unwrap(); json::to_pretty_writer(f, &self.to_json()); } /// Compare against another MetricMap. Optionally compare all /// measurements in the maps using the provided `noise_pct` as a /// percentage of each value to consider noise. If `None`, each /// measurement's noise threshold is independently chosen as the /// maximum of that measurement's recorded noise quantity in either /// map. pub fn compare_to_old(&self, old: &MetricMap, noise_pct: Option) -> MetricDiff { let mut diff : MetricDiff = TreeMap::new(); for old.iter().advance |(k, vold)| { let r = match self.find(k) { None => MetricRemoved, Some(v) => { let delta = (v.value - vold.value); let noise = match noise_pct { None => f64::max(vold.noise.abs(), v.noise.abs()), Some(pct) => vold.value * pct / 100.0 }; if delta.abs() <= noise { LikelyNoise } else { let pct = delta.abs() / (vold.value).max(&f64::epsilon) * 100.0; if vold.noise < 0.0 { // When 'noise' is negative, it means we want // to see deltas that go up over time, and can // only tolerate slight negative movement. if delta < 0.0 { Regression(pct) } else { Improvement(pct) } } else { // When 'noise' is positive, it means we want // to see deltas that go down over time, and // can only tolerate slight positive movements. if delta < 0.0 { Improvement(pct) } else { Regression(pct) } } } } }; diff.insert((*k).clone(), r); } for self.iter().advance |(k, _)| { if !diff.contains_key(k) { diff.insert((*k).clone(), MetricAdded); } } diff } /// Insert a named `value` (+/- `noise`) metric into the map. The value /// must be non-negative. The `noise` indicates the uncertainty of the /// metric, which doubles as the "noise range" of acceptable /// pairwise-regressions on this named value, when comparing from one /// metric to the next using `compare_to_old`. /// /// If `noise` is positive, then it means this metric is of a value /// you want to see grow smaller, so a change larger than `noise` in the /// positive direction represents a regression. /// /// If `noise` is negative, then it means this metric is of a value /// you want to see grow larger, so a change larger than `noise` in the /// negative direction represents a regression. pub fn insert_metric(&mut self, name: &str, value: f64, noise: f64) { let m = Metric { value: value, noise: noise }; self.insert(name.to_owned(), m); } /// Attempt to "ratchet" an external metric file. This involves loading /// metrics from a metric file (if it exists), comparing against /// the metrics in `self` using `compare_to_old`, and rewriting the /// file to contain the metrics in `self` if none of the /// `MetricChange`s are `Regression`. Returns the diff as well /// as a boolean indicating whether the ratchet succeeded. pub fn ratchet(&self, p: &Path, pct: Option) -> (MetricDiff, bool) { let old = if os::path_exists(p) { MetricMap::load(p) } else { MetricMap::new() }; let diff : MetricDiff = self.compare_to_old(&old, pct); let ok = do diff.iter().all() |(_, v)| { match *v { Regression(_) => false, _ => true } }; if ok { debug!("rewriting file '%s' with updated metrics"); self.save(p); } return (diff, ok) } } // Benchmarking impl BenchHarness { /// Callback for benchmark functions to run in their body. pub fn iter(&mut self, inner:&fn()) { self.ns_start = precise_time_ns(); let k = self.iterations; for u64::range(0, k) |_| { inner(); } self.ns_end = precise_time_ns(); } pub fn ns_elapsed(&mut self) -> u64 { if self.ns_start == 0 || self.ns_end == 0 { 0 } else { self.ns_end - self.ns_start } } pub fn ns_per_iter(&mut self) -> u64 { if self.iterations == 0 { 0 } else { self.ns_elapsed() / self.iterations.max(&1) } } pub fn bench_n(&mut self, n: u64, f: &fn(&mut BenchHarness)) { self.iterations = n; debug!("running benchmark for %u iterations", n as uint); f(self); } // This is a more statistics-driven benchmark algorithm pub fn auto_bench(&mut self, f: &fn(&mut BenchHarness)) -> stats::Summary { // Initial bench run to get ballpark figure. let mut n = 1_u64; self.bench_n(n, |x| f(x)); // Try to estimate iter count for 1ms falling back to 1m // iterations if first run took < 1ns. if self.ns_per_iter() == 0 { n = 1_000_000; } else { n = 1_000_000 / self.ns_per_iter().max(&1); } let mut total_run = 0; let samples : &mut [f64] = [0.0_f64, ..50]; loop { let loop_start = precise_time_ns(); for samples.mut_iter().advance() |p| { self.bench_n(n as u64, |x| f(x)); *p = self.ns_per_iter() as f64; }; stats::winsorize(samples, 5.0); let summ = stats::Summary::new(samples); for samples.mut_iter().advance() |p| { self.bench_n(5 * n as u64, |x| f(x)); *p = self.ns_per_iter() as f64; }; stats::winsorize(samples, 5.0); let summ5 = stats::Summary::new(samples); debug!("%u samples, median %f, MAD=%f, MADP=%f", samples.len(), summ.median as float, summ.median_abs_dev as float, summ.median_abs_dev_pct as float); let now = precise_time_ns(); let loop_run = now - loop_start; // If we've run for 100ms an seem to have converged to a // stable median. if loop_run > 100_000_000 && summ.median_abs_dev_pct < 1.0 && summ.median - summ5.median < summ5.median_abs_dev { return summ5; } total_run += loop_run; // Longest we ever run for is 3s. if total_run > 3_000_000_000 { return summ5; } n *= 2; } } } pub mod bench { use test::{BenchHarness, BenchSamples}; pub fn benchmark(f: &fn(&mut BenchHarness)) -> BenchSamples { let mut bs = BenchHarness { iterations: 0, ns_start: 0, ns_end: 0, bytes: 0 }; let ns_iter_summ = bs.auto_bench(f); let ns_iter = (ns_iter_summ.median as u64).max(&1); let iter_s = 1_000_000_000 / ns_iter; let mb_s = (bs.bytes * iter_s) / 1_000_000; BenchSamples { ns_iter_summ: ns_iter_summ, mb_s: mb_s as uint } } } #[cfg(test)] mod tests { use test::{TrFailed, TrIgnored, TrOk, filter_tests, parse_opts, TestDesc, TestDescAndFn, Metric, MetricMap, MetricAdded, MetricRemoved, Improvement, Regression, LikelyNoise, StaticTestName, DynTestName, DynTestFn}; use test::{TestOpts, run_test}; use std::either; use std::comm::{stream, SharedChan}; use std::vec; use tempfile; use std::os; #[test] pub fn do_not_run_ignored_tests() { fn f() { fail!(); } let desc = TestDescAndFn { desc: TestDesc { name: StaticTestName("whatever"), ignore: true, should_fail: false }, testfn: DynTestFn(|| f()), }; let (p, ch) = stream(); let ch = SharedChan::new(ch); run_test(false, desc, ch); let (_, res) = p.recv(); assert!(res != TrOk); } #[test] pub fn ignored_tests_result_in_ignored() { fn f() { } let desc = TestDescAndFn { desc: TestDesc { name: StaticTestName("whatever"), ignore: true, should_fail: false }, testfn: DynTestFn(|| f()), }; let (p, ch) = stream(); let ch = SharedChan::new(ch); run_test(false, desc, ch); let (_, res) = p.recv(); assert_eq!(res, TrIgnored); } #[test] #[ignore(cfg(windows))] fn test_should_fail() { fn f() { fail!(); } let desc = TestDescAndFn { desc: TestDesc { name: StaticTestName("whatever"), ignore: false, should_fail: true }, testfn: DynTestFn(|| f()), }; let (p, ch) = stream(); let ch = SharedChan::new(ch); run_test(false, desc, ch); let (_, res) = p.recv(); assert_eq!(res, TrOk); } #[test] fn test_should_fail_but_succeeds() { fn f() { } let desc = TestDescAndFn { desc: TestDesc { name: StaticTestName("whatever"), ignore: false, should_fail: true }, testfn: DynTestFn(|| f()), }; let (p, ch) = stream(); let ch = SharedChan::new(ch); run_test(false, desc, ch); let (_, res) = p.recv(); assert_eq!(res, TrFailed); } #[test] fn first_free_arg_should_be_a_filter() { let args = ~[~"progname", ~"filter"]; let opts = match parse_opts(args) { either::Left(o) => o, _ => fail!("Malformed arg in first_free_arg_should_be_a_filter") }; assert!("filter" == opts.filter.clone().get()); } #[test] fn parse_ignored_flag() { let args = ~[~"progname", ~"filter", ~"--ignored"]; let opts = match parse_opts(args) { either::Left(o) => o, _ => fail!("Malformed arg in parse_ignored_flag") }; assert!((opts.run_ignored)); } #[test] pub fn filter_for_ignored_option() { fn dummy() {} // When we run ignored tests the test filter should filter out all the // unignored tests and flip the ignore flag on the rest to false let opts = TestOpts { filter: None, run_ignored: true, logfile: None, run_tests: true, run_benchmarks: false, ratchet_noise_percent: None, ratchet_metrics: None, save_metrics: None, }; let tests = ~[ TestDescAndFn { desc: TestDesc { name: StaticTestName("1"), ignore: true, should_fail: false, }, testfn: DynTestFn(|| {}), }, TestDescAndFn { desc: TestDesc { name: StaticTestName("2"), ignore: false, should_fail: false }, testfn: DynTestFn(|| {}), }, ]; let filtered = filter_tests(&opts, tests); assert_eq!(filtered.len(), 1); assert_eq!(filtered[0].desc.name.to_str(), ~"1"); assert!(filtered[0].desc.ignore == false); } #[test] pub fn sort_tests() { let opts = TestOpts { filter: None, run_ignored: false, logfile: None, run_tests: true, run_benchmarks: false, ratchet_noise_percent: None, ratchet_metrics: None, save_metrics: None, }; let names = ~[~"sha1::test", ~"int::test_to_str", ~"int::test_pow", ~"test::do_not_run_ignored_tests", ~"test::ignored_tests_result_in_ignored", ~"test::first_free_arg_should_be_a_filter", ~"test::parse_ignored_flag", ~"test::filter_for_ignored_option", ~"test::sort_tests"]; let tests = { fn testfn() { } let mut tests = ~[]; for names.iter().advance |name| { let test = TestDescAndFn { desc: TestDesc { name: DynTestName((*name).clone()), ignore: false, should_fail: false }, testfn: DynTestFn(testfn), }; tests.push(test); } tests }; let filtered = filter_tests(&opts, tests); let expected = ~[~"int::test_pow", ~"int::test_to_str", ~"sha1::test", ~"test::do_not_run_ignored_tests", ~"test::filter_for_ignored_option", ~"test::first_free_arg_should_be_a_filter", ~"test::ignored_tests_result_in_ignored", ~"test::parse_ignored_flag", ~"test::sort_tests"]; let pairs = vec::zip(expected, filtered); for pairs.iter().advance |p| { match *p { (ref a, ref b) => { assert!(*a == b.desc.name.to_str()); } } } } #[test] pub fn test_metricmap_compare() { let mut m1 = MetricMap::new(); let mut m2 = MetricMap::new(); m1.insert_metric("in-both-noise", 1000.0, 200.0); m2.insert_metric("in-both-noise", 1100.0, 200.0); m1.insert_metric("in-first-noise", 1000.0, 2.0); m2.insert_metric("in-second-noise", 1000.0, 2.0); m1.insert_metric("in-both-want-downwards-but-regressed", 1000.0, 10.0); m2.insert_metric("in-both-want-downwards-but-regressed", 2000.0, 10.0); m1.insert_metric("in-both-want-downwards-and-improved", 2000.0, 10.0); m2.insert_metric("in-both-want-downwards-and-improved", 1000.0, 10.0); m1.insert_metric("in-both-want-upwards-but-regressed", 2000.0, -10.0); m2.insert_metric("in-both-want-upwards-but-regressed", 1000.0, -10.0); m1.insert_metric("in-both-want-upwards-and-improved", 1000.0, -10.0); m2.insert_metric("in-both-want-upwards-and-improved", 2000.0, -10.0); let diff1 = m2.compare_to_old(&m1, None); assert_eq!(*(diff1.find(&~"in-both-noise").get()), LikelyNoise); assert_eq!(*(diff1.find(&~"in-first-noise").get()), MetricRemoved); assert_eq!(*(diff1.find(&~"in-second-noise").get()), MetricAdded); assert_eq!(*(diff1.find(&~"in-both-want-downwards-but-regressed").get()), Regression(100.0)); assert_eq!(*(diff1.find(&~"in-both-want-downwards-and-improved").get()), Improvement(50.0)); assert_eq!(*(diff1.find(&~"in-both-want-upwards-but-regressed").get()), Regression(50.0)); assert_eq!(*(diff1.find(&~"in-both-want-upwards-and-improved").get()), Improvement(100.0)); assert_eq!(diff1.len(), 7); let diff2 = m2.compare_to_old(&m1, Some(200.0)); assert_eq!(*(diff2.find(&~"in-both-noise").get()), LikelyNoise); assert_eq!(*(diff2.find(&~"in-first-noise").get()), MetricRemoved); assert_eq!(*(diff2.find(&~"in-second-noise").get()), MetricAdded); assert_eq!(*(diff2.find(&~"in-both-want-downwards-but-regressed").get()), LikelyNoise); assert_eq!(*(diff2.find(&~"in-both-want-downwards-and-improved").get()), LikelyNoise); assert_eq!(*(diff2.find(&~"in-both-want-upwards-but-regressed").get()), LikelyNoise); assert_eq!(*(diff2.find(&~"in-both-want-upwards-and-improved").get()), LikelyNoise); assert_eq!(diff2.len(), 7); } pub fn ratchet_test() { let dpth = tempfile::mkdtemp(&os::tmpdir(), "test-ratchet").expect("missing test for ratchet"); let pth = dpth.push("ratchet.json"); let mut m1 = MetricMap::new(); m1.insert_metric("runtime", 1000.0, 2.0); m1.insert_metric("throughput", 50.0, 2.0); let mut m2 = MetricMap::new(); m2.insert_metric("runtime", 1100.0, 2.0); m2.insert_metric("throughput", 50.0, 2.0); m1.save(&pth); // Ask for a ratchet that should fail to advance. let (diff1, ok1) = m2.ratchet(&pth, None); assert_eq!(ok1, false); assert_eq!(diff1.len(), 2); assert_eq!(*(diff1.find(&~"runtime").get()), Regression(10.0)); assert_eq!(*(diff1.find(&~"throughput").get()), LikelyNoise); // Check that it was not rewritten. let m3 = MetricMap::load(&pth); assert_eq!(m3.len(), 2); assert_eq!(*(m3.find(&~"runtime").get()), Metric { value: 1000.0, noise: 2.0 }); assert_eq!(*(m3.find(&~"throughput").get()), Metric { value: 50.0, noise: 2.0 }); // Ask for a ratchet with an explicit noise-percentage override, // that should advance. let (diff2, ok2) = m2.ratchet(&pth, Some(10.0)); assert_eq!(ok2, true); assert_eq!(diff2.len(), 2); assert_eq!(*(diff2.find(&~"runtime").get()), LikelyNoise); assert_eq!(*(diff2.find(&~"throughput").get()), LikelyNoise); // Check that it was rewritten. let m4 = MetricMap::load(&pth); assert_eq!(m4.len(), 2); assert_eq!(*(m4.find(&~"runtime").get()), Metric { value: 1100.0, noise: 2.0 }); assert_eq!(*(m4.find(&~"throughput").get()), Metric { value: 50.0, noise: 2.0 }); os::remove_dir_recursive(&dpth); } }