Use CmdArgs pattern for bench & analysis stats
This commit is contained in:
parent
be49547b44
commit
cd9f863b01
@ -8,7 +8,7 @@
|
||||
use anyhow::{bail, Result};
|
||||
use pico_args::Arguments;
|
||||
use ra_ssr::{SsrPattern, SsrRule};
|
||||
use rust_analyzer::cli::{BenchWhat, Position, Verbosity};
|
||||
use rust_analyzer::cli::{AnalysisStatsCmd, BenchCmd, BenchWhat, Position, Verbosity};
|
||||
use vfs::AbsPathBuf;
|
||||
|
||||
pub(crate) struct Args {
|
||||
@ -24,23 +24,8 @@ pub(crate) enum Command {
|
||||
Highlight {
|
||||
rainbow: bool,
|
||||
},
|
||||
Stats {
|
||||
randomize: bool,
|
||||
parallel: bool,
|
||||
memory_usage: bool,
|
||||
only: Option<String>,
|
||||
with_deps: bool,
|
||||
path: PathBuf,
|
||||
load_output_dirs: bool,
|
||||
with_proc_macro: bool,
|
||||
},
|
||||
Bench {
|
||||
memory_usage: bool,
|
||||
path: PathBuf,
|
||||
what: BenchWhat,
|
||||
load_output_dirs: bool,
|
||||
with_proc_macro: bool,
|
||||
},
|
||||
AnalysisStats(AnalysisStatsCmd),
|
||||
Bench(BenchCmd),
|
||||
Diagnostics {
|
||||
path: PathBuf,
|
||||
load_output_dirs: bool,
|
||||
@ -199,7 +184,7 @@ pub(crate) fn parse() -> Result<Result<Args, HelpPrinted>> {
|
||||
trailing.pop().unwrap().into()
|
||||
};
|
||||
|
||||
Command::Stats {
|
||||
Command::AnalysisStats(AnalysisStatsCmd {
|
||||
randomize,
|
||||
parallel,
|
||||
memory_usage,
|
||||
@ -208,7 +193,7 @@ pub(crate) fn parse() -> Result<Result<Args, HelpPrinted>> {
|
||||
path,
|
||||
load_output_dirs,
|
||||
with_proc_macro,
|
||||
}
|
||||
})
|
||||
}
|
||||
"analysis-bench" => {
|
||||
if matches.contains(["-h", "--help"]) {
|
||||
@ -256,7 +241,13 @@ pub(crate) fn parse() -> Result<Result<Args, HelpPrinted>> {
|
||||
let memory_usage = matches.contains("--memory-usage");
|
||||
let load_output_dirs = matches.contains("--load-output-dirs");
|
||||
let with_proc_macro = matches.contains("--with-proc-macro");
|
||||
Command::Bench { memory_usage, path, what, load_output_dirs, with_proc_macro }
|
||||
Command::Bench(BenchCmd {
|
||||
memory_usage,
|
||||
path,
|
||||
what,
|
||||
load_output_dirs,
|
||||
with_proc_macro,
|
||||
})
|
||||
}
|
||||
"diagnostics" => {
|
||||
if matches.contains(["-h", "--help"]) {
|
||||
|
@ -33,36 +33,8 @@ fn main() -> Result<()> {
|
||||
args::Command::Parse { no_dump } => cli::parse(no_dump)?,
|
||||
args::Command::Symbols => cli::symbols()?,
|
||||
args::Command::Highlight { rainbow } => cli::highlight(rainbow)?,
|
||||
args::Command::Stats {
|
||||
randomize,
|
||||
parallel,
|
||||
memory_usage,
|
||||
only,
|
||||
with_deps,
|
||||
path,
|
||||
load_output_dirs,
|
||||
with_proc_macro,
|
||||
} => cli::analysis_stats(
|
||||
args.verbosity,
|
||||
memory_usage,
|
||||
path.as_ref(),
|
||||
only.as_ref().map(String::as_ref),
|
||||
with_deps,
|
||||
randomize,
|
||||
parallel,
|
||||
load_output_dirs,
|
||||
with_proc_macro,
|
||||
)?,
|
||||
args::Command::Bench { memory_usage, path, what, load_output_dirs, with_proc_macro } => {
|
||||
cli::analysis_bench(
|
||||
args.verbosity,
|
||||
path.as_ref(),
|
||||
what,
|
||||
memory_usage,
|
||||
load_output_dirs,
|
||||
with_proc_macro,
|
||||
)?
|
||||
}
|
||||
args::Command::AnalysisStats(cmd) => cmd.run(args.verbosity)?,
|
||||
args::Command::Bench(cmd) => cmd.run(args.verbosity)?,
|
||||
args::Command::Diagnostics { path, load_output_dirs, with_proc_macro, all } => {
|
||||
cli::diagnostics(path.as_ref(), load_output_dirs, with_proc_macro, all)?
|
||||
}
|
||||
|
@ -14,8 +14,8 @@
|
||||
use ra_prof::profile;
|
||||
use ra_syntax::{AstNode, SourceFile};
|
||||
|
||||
pub use analysis_bench::{analysis_bench, BenchWhat, Position};
|
||||
pub use analysis_stats::analysis_stats;
|
||||
pub use analysis_bench::{BenchCmd, BenchWhat, Position};
|
||||
pub use analysis_stats::AnalysisStatsCmd;
|
||||
pub use diagnostics::diagnostics;
|
||||
pub use load_cargo::load_cargo;
|
||||
pub use ssr::{apply_ssr_rules, search_for_patterns};
|
||||
|
@ -1,6 +1,6 @@
|
||||
//! Benchmark operations like highlighting or goto definition.
|
||||
|
||||
use std::{env, path::Path, str::FromStr, sync::Arc, time::Instant};
|
||||
use std::{env, path::PathBuf, str::FromStr, sync::Arc, time::Instant};
|
||||
|
||||
use anyhow::{bail, format_err, Result};
|
||||
use ra_db::{
|
||||
@ -15,6 +15,14 @@
|
||||
print_memory_usage,
|
||||
};
|
||||
|
||||
pub struct BenchCmd {
|
||||
pub path: PathBuf,
|
||||
pub what: BenchWhat,
|
||||
pub memory_usage: bool,
|
||||
pub load_output_dirs: bool,
|
||||
pub with_proc_macro: bool,
|
||||
}
|
||||
|
||||
pub enum BenchWhat {
|
||||
Highlight { path: AbsPathBuf },
|
||||
Complete(Position),
|
||||
@ -42,72 +50,68 @@ fn from_str(s: &str) -> Result<Self> {
|
||||
}
|
||||
}
|
||||
|
||||
pub fn analysis_bench(
|
||||
verbosity: Verbosity,
|
||||
path: &Path,
|
||||
what: BenchWhat,
|
||||
memory_usage: bool,
|
||||
load_output_dirs: bool,
|
||||
with_proc_macro: bool,
|
||||
) -> Result<()> {
|
||||
ra_prof::init();
|
||||
impl BenchCmd {
|
||||
pub fn run(self, verbosity: Verbosity) -> Result<()> {
|
||||
ra_prof::init();
|
||||
|
||||
let start = Instant::now();
|
||||
eprint!("loading: ");
|
||||
let (mut host, vfs) = load_cargo(path, load_output_dirs, with_proc_macro)?;
|
||||
eprintln!("{:?}\n", start.elapsed());
|
||||
let start = Instant::now();
|
||||
eprint!("loading: ");
|
||||
let (mut host, vfs) = load_cargo(&self.path, self.load_output_dirs, self.with_proc_macro)?;
|
||||
eprintln!("{:?}\n", start.elapsed());
|
||||
|
||||
let file_id = {
|
||||
let path = match &what {
|
||||
BenchWhat::Highlight { path } => path,
|
||||
BenchWhat::Complete(pos) | BenchWhat::GotoDef(pos) => &pos.path,
|
||||
let file_id = {
|
||||
let path = match &self.what {
|
||||
BenchWhat::Highlight { path } => path,
|
||||
BenchWhat::Complete(pos) | BenchWhat::GotoDef(pos) => &pos.path,
|
||||
};
|
||||
let path = path.clone().into();
|
||||
vfs.file_id(&path).ok_or_else(|| format_err!("Can't find {}", path))?
|
||||
};
|
||||
let path = path.clone().into();
|
||||
vfs.file_id(&path).ok_or_else(|| format_err!("Can't find {}", path))?
|
||||
};
|
||||
|
||||
match &what {
|
||||
BenchWhat::Highlight { .. } => {
|
||||
let res = do_work(&mut host, file_id, |analysis| {
|
||||
analysis.diagnostics(file_id, true).unwrap();
|
||||
analysis.highlight_as_html(file_id, false).unwrap()
|
||||
});
|
||||
if verbosity.is_verbose() {
|
||||
println!("\n{}", res);
|
||||
}
|
||||
}
|
||||
BenchWhat::Complete(pos) | BenchWhat::GotoDef(pos) => {
|
||||
let is_completion = matches!(what, BenchWhat::Complete(..));
|
||||
|
||||
let offset = host
|
||||
.analysis()
|
||||
.file_line_index(file_id)?
|
||||
.offset(LineCol { line: pos.line - 1, col_utf16: pos.column });
|
||||
let file_position = FilePosition { file_id, offset };
|
||||
|
||||
if is_completion {
|
||||
let options = CompletionConfig::default();
|
||||
match &self.what {
|
||||
BenchWhat::Highlight { .. } => {
|
||||
let res = do_work(&mut host, file_id, |analysis| {
|
||||
analysis.completions(&options, file_position)
|
||||
analysis.diagnostics(file_id, true).unwrap();
|
||||
analysis.highlight_as_html(file_id, false).unwrap()
|
||||
});
|
||||
if verbosity.is_verbose() {
|
||||
println!("\n{:#?}", res);
|
||||
println!("\n{}", res);
|
||||
}
|
||||
} else {
|
||||
let res =
|
||||
do_work(&mut host, file_id, |analysis| analysis.goto_definition(file_position));
|
||||
if verbosity.is_verbose() {
|
||||
println!("\n{:#?}", res);
|
||||
}
|
||||
BenchWhat::Complete(pos) | BenchWhat::GotoDef(pos) => {
|
||||
let is_completion = matches!(self.what, BenchWhat::Complete(..));
|
||||
|
||||
let offset = host
|
||||
.analysis()
|
||||
.file_line_index(file_id)?
|
||||
.offset(LineCol { line: pos.line - 1, col_utf16: pos.column });
|
||||
let file_position = FilePosition { file_id, offset };
|
||||
|
||||
if is_completion {
|
||||
let options = CompletionConfig::default();
|
||||
let res = do_work(&mut host, file_id, |analysis| {
|
||||
analysis.completions(&options, file_position)
|
||||
});
|
||||
if verbosity.is_verbose() {
|
||||
println!("\n{:#?}", res);
|
||||
}
|
||||
} else {
|
||||
let res = do_work(&mut host, file_id, |analysis| {
|
||||
analysis.goto_definition(file_position)
|
||||
});
|
||||
if verbosity.is_verbose() {
|
||||
println!("\n{:#?}", res);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if memory_usage {
|
||||
print_memory_usage(host, vfs);
|
||||
}
|
||||
if self.memory_usage {
|
||||
print_memory_usage(host, vfs);
|
||||
}
|
||||
|
||||
Ok(())
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
fn do_work<F: Fn(&Analysis) -> T, T>(host: &mut AnalysisHost, file_id: FileId, work: F) -> T {
|
||||
|
@ -2,7 +2,7 @@
|
||||
//! errors.
|
||||
|
||||
use std::{
|
||||
path::Path,
|
||||
path::PathBuf,
|
||||
time::{SystemTime, UNIX_EPOCH},
|
||||
};
|
||||
|
||||
@ -39,273 +39,280 @@ fn clone(&self) -> Snap<salsa::Snapshot<DB>> {
|
||||
}
|
||||
}
|
||||
|
||||
pub fn analysis_stats(
|
||||
verbosity: Verbosity,
|
||||
memory_usage: bool,
|
||||
path: &Path,
|
||||
only: Option<&str>,
|
||||
with_deps: bool,
|
||||
randomize: bool,
|
||||
parallel: bool,
|
||||
load_output_dirs: bool,
|
||||
with_proc_macro: bool,
|
||||
) -> Result<()> {
|
||||
let mut rng = {
|
||||
let seed = SystemTime::now().duration_since(UNIX_EPOCH).unwrap().as_millis() as u64;
|
||||
Rand32::new(seed)
|
||||
};
|
||||
pub struct AnalysisStatsCmd {
|
||||
pub randomize: bool,
|
||||
pub parallel: bool,
|
||||
pub memory_usage: bool,
|
||||
pub only: Option<String>,
|
||||
pub with_deps: bool,
|
||||
pub path: PathBuf,
|
||||
pub load_output_dirs: bool,
|
||||
pub with_proc_macro: bool,
|
||||
}
|
||||
|
||||
let mut db_load_sw = StopWatch::start().memory(memory_usage);
|
||||
let (host, vfs) = load_cargo(path, load_output_dirs, with_proc_macro)?;
|
||||
let db = host.raw_database();
|
||||
eprintln!("Database loaded {}", db_load_sw.elapsed());
|
||||
impl AnalysisStatsCmd {
|
||||
pub fn run(self, verbosity: Verbosity) -> Result<()> {
|
||||
let mut rng = {
|
||||
let seed = SystemTime::now().duration_since(UNIX_EPOCH).unwrap().as_millis() as u64;
|
||||
Rand32::new(seed)
|
||||
};
|
||||
|
||||
let mut analysis_sw = StopWatch::start().memory(memory_usage);
|
||||
let mut num_crates = 0;
|
||||
let mut visited_modules = FxHashSet::default();
|
||||
let mut visit_queue = Vec::new();
|
||||
let mut db_load_sw = self.stop_watch();
|
||||
let (host, vfs) = load_cargo(&self.path, self.load_output_dirs, self.with_proc_macro)?;
|
||||
let db = host.raw_database();
|
||||
eprintln!("Database loaded {}", db_load_sw.elapsed());
|
||||
|
||||
let mut krates = Crate::all(db);
|
||||
if randomize {
|
||||
shuffle(&mut rng, &mut krates);
|
||||
}
|
||||
for krate in krates {
|
||||
let module = krate.root_module(db).expect("crate without root module");
|
||||
let file_id = module.definition_source(db).file_id;
|
||||
let file_id = file_id.original_file(db);
|
||||
let source_root = db.file_source_root(file_id);
|
||||
let source_root = db.source_root(source_root);
|
||||
if !source_root.is_library || with_deps {
|
||||
num_crates += 1;
|
||||
visit_queue.push(module);
|
||||
let mut analysis_sw = self.stop_watch();
|
||||
let mut num_crates = 0;
|
||||
let mut visited_modules = FxHashSet::default();
|
||||
let mut visit_queue = Vec::new();
|
||||
|
||||
let mut krates = Crate::all(db);
|
||||
if self.randomize {
|
||||
shuffle(&mut rng, &mut krates);
|
||||
}
|
||||
}
|
||||
|
||||
if randomize {
|
||||
shuffle(&mut rng, &mut visit_queue);
|
||||
}
|
||||
|
||||
eprintln!("Crates in this dir: {}", num_crates);
|
||||
let mut num_decls = 0;
|
||||
let mut funcs = Vec::new();
|
||||
while let Some(module) = visit_queue.pop() {
|
||||
if visited_modules.insert(module) {
|
||||
visit_queue.extend(module.children(db));
|
||||
|
||||
for decl in module.declarations(db) {
|
||||
num_decls += 1;
|
||||
if let ModuleDef::Function(f) = decl {
|
||||
funcs.push(f);
|
||||
}
|
||||
for krate in krates {
|
||||
let module = krate.root_module(db).expect("crate without root module");
|
||||
let file_id = module.definition_source(db).file_id;
|
||||
let file_id = file_id.original_file(db);
|
||||
let source_root = db.file_source_root(file_id);
|
||||
let source_root = db.source_root(source_root);
|
||||
if !source_root.is_library || self.with_deps {
|
||||
num_crates += 1;
|
||||
visit_queue.push(module);
|
||||
}
|
||||
}
|
||||
|
||||
for impl_def in module.impl_defs(db) {
|
||||
for item in impl_def.items(db) {
|
||||
if self.randomize {
|
||||
shuffle(&mut rng, &mut visit_queue);
|
||||
}
|
||||
|
||||
eprintln!("Crates in this dir: {}", num_crates);
|
||||
let mut num_decls = 0;
|
||||
let mut funcs = Vec::new();
|
||||
while let Some(module) = visit_queue.pop() {
|
||||
if visited_modules.insert(module) {
|
||||
visit_queue.extend(module.children(db));
|
||||
|
||||
for decl in module.declarations(db) {
|
||||
num_decls += 1;
|
||||
if let AssocItem::Function(f) = item {
|
||||
if let ModuleDef::Function(f) = decl {
|
||||
funcs.push(f);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
eprintln!("Total modules found: {}", visited_modules.len());
|
||||
eprintln!("Total declarations: {}", num_decls);
|
||||
eprintln!("Total functions: {}", funcs.len());
|
||||
eprintln!("Item Collection: {}", analysis_sw.elapsed());
|
||||
|
||||
if randomize {
|
||||
shuffle(&mut rng, &mut funcs);
|
||||
}
|
||||
|
||||
let mut bar = match verbosity {
|
||||
Verbosity::Quiet | Verbosity::Spammy => ProgressReport::hidden(),
|
||||
_ if parallel => ProgressReport::hidden(),
|
||||
_ => ProgressReport::new(funcs.len() as u64),
|
||||
};
|
||||
|
||||
if parallel {
|
||||
let mut inference_sw = StopWatch::start().memory(memory_usage);
|
||||
let snap = Snap(db.snapshot());
|
||||
funcs
|
||||
.par_iter()
|
||||
.map_with(snap, |snap, &f| {
|
||||
let f_id = FunctionId::from(f);
|
||||
snap.0.body(f_id.into());
|
||||
snap.0.infer(f_id.into());
|
||||
})
|
||||
.count();
|
||||
eprintln!("Parallel Inference: {}", inference_sw.elapsed());
|
||||
}
|
||||
|
||||
let mut inference_sw = StopWatch::start().memory(memory_usage);
|
||||
bar.tick();
|
||||
let mut num_exprs = 0;
|
||||
let mut num_exprs_unknown = 0;
|
||||
let mut num_exprs_partially_unknown = 0;
|
||||
let mut num_type_mismatches = 0;
|
||||
for f in funcs {
|
||||
let name = f.name(db);
|
||||
let full_name = f
|
||||
.module(db)
|
||||
.path_to_root(db)
|
||||
.into_iter()
|
||||
.rev()
|
||||
.filter_map(|it| it.name(db))
|
||||
.chain(Some(f.name(db)))
|
||||
.join("::");
|
||||
if let Some(only_name) = only {
|
||||
if name.to_string() != only_name && full_name != only_name {
|
||||
continue;
|
||||
}
|
||||
}
|
||||
let mut msg = format!("processing: {}", full_name);
|
||||
if verbosity.is_verbose() {
|
||||
let src = f.source(db);
|
||||
let original_file = src.file_id.original_file(db);
|
||||
let path = vfs.file_path(original_file);
|
||||
let syntax_range = src.value.syntax().text_range();
|
||||
format_to!(msg, " ({} {:?})", path, syntax_range);
|
||||
}
|
||||
if verbosity.is_spammy() {
|
||||
bar.println(msg.to_string());
|
||||
}
|
||||
bar.set_message(&msg);
|
||||
let f_id = FunctionId::from(f);
|
||||
let body = db.body(f_id.into());
|
||||
let inference_result = db.infer(f_id.into());
|
||||
let (previous_exprs, previous_unknown, previous_partially_unknown) =
|
||||
(num_exprs, num_exprs_unknown, num_exprs_partially_unknown);
|
||||
for (expr_id, _) in body.exprs.iter() {
|
||||
let ty = &inference_result[expr_id];
|
||||
num_exprs += 1;
|
||||
if let Ty::Unknown = ty {
|
||||
num_exprs_unknown += 1;
|
||||
} else {
|
||||
let mut is_partially_unknown = false;
|
||||
ty.walk(&mut |ty| {
|
||||
if let Ty::Unknown = ty {
|
||||
is_partially_unknown = true;
|
||||
for impl_def in module.impl_defs(db) {
|
||||
for item in impl_def.items(db) {
|
||||
num_decls += 1;
|
||||
if let AssocItem::Function(f) = item {
|
||||
funcs.push(f);
|
||||
}
|
||||
}
|
||||
});
|
||||
if is_partially_unknown {
|
||||
num_exprs_partially_unknown += 1;
|
||||
}
|
||||
}
|
||||
if only.is_some() && verbosity.is_spammy() {
|
||||
// in super-verbose mode for just one function, we print every single expression
|
||||
let (_, sm) = db.body_with_source_map(f_id.into());
|
||||
let src = sm.expr_syntax(expr_id);
|
||||
if let Ok(src) = src {
|
||||
let node = {
|
||||
let root = db.parse_or_expand(src.file_id).unwrap();
|
||||
src.value.to_node(&root)
|
||||
};
|
||||
let original_file = src.file_id.original_file(db);
|
||||
let line_index = host.analysis().file_line_index(original_file).unwrap();
|
||||
let text_range = node.syntax().text_range();
|
||||
let (start, end) = (
|
||||
line_index.line_col(text_range.start()),
|
||||
line_index.line_col(text_range.end()),
|
||||
);
|
||||
bar.println(format!(
|
||||
"{}:{}-{}:{}: {}",
|
||||
start.line + 1,
|
||||
start.col_utf16,
|
||||
end.line + 1,
|
||||
end.col_utf16,
|
||||
ty.display(db)
|
||||
));
|
||||
}
|
||||
eprintln!("Total modules found: {}", visited_modules.len());
|
||||
eprintln!("Total declarations: {}", num_decls);
|
||||
eprintln!("Total functions: {}", funcs.len());
|
||||
eprintln!("Item Collection: {}", analysis_sw.elapsed());
|
||||
|
||||
if self.randomize {
|
||||
shuffle(&mut rng, &mut funcs);
|
||||
}
|
||||
|
||||
let mut bar = match verbosity {
|
||||
Verbosity::Quiet | Verbosity::Spammy => ProgressReport::hidden(),
|
||||
_ if self.parallel => ProgressReport::hidden(),
|
||||
_ => ProgressReport::new(funcs.len() as u64),
|
||||
};
|
||||
|
||||
if self.parallel {
|
||||
let mut inference_sw = self.stop_watch();
|
||||
let snap = Snap(db.snapshot());
|
||||
funcs
|
||||
.par_iter()
|
||||
.map_with(snap, |snap, &f| {
|
||||
let f_id = FunctionId::from(f);
|
||||
snap.0.body(f_id.into());
|
||||
snap.0.infer(f_id.into());
|
||||
})
|
||||
.count();
|
||||
eprintln!("Parallel Inference: {}", inference_sw.elapsed());
|
||||
}
|
||||
|
||||
let mut inference_sw = self.stop_watch();
|
||||
bar.tick();
|
||||
let mut num_exprs = 0;
|
||||
let mut num_exprs_unknown = 0;
|
||||
let mut num_exprs_partially_unknown = 0;
|
||||
let mut num_type_mismatches = 0;
|
||||
for f in funcs {
|
||||
let name = f.name(db);
|
||||
let full_name = f
|
||||
.module(db)
|
||||
.path_to_root(db)
|
||||
.into_iter()
|
||||
.rev()
|
||||
.filter_map(|it| it.name(db))
|
||||
.chain(Some(f.name(db)))
|
||||
.join("::");
|
||||
if let Some(only_name) = self.only.as_deref() {
|
||||
if name.to_string() != only_name && full_name != only_name {
|
||||
continue;
|
||||
}
|
||||
}
|
||||
let mut msg = format!("processing: {}", full_name);
|
||||
if verbosity.is_verbose() {
|
||||
let src = f.source(db);
|
||||
let original_file = src.file_id.original_file(db);
|
||||
let path = vfs.file_path(original_file);
|
||||
let syntax_range = src.value.syntax().text_range();
|
||||
format_to!(msg, " ({} {:?})", path, syntax_range);
|
||||
}
|
||||
if verbosity.is_spammy() {
|
||||
bar.println(msg.to_string());
|
||||
}
|
||||
bar.set_message(&msg);
|
||||
let f_id = FunctionId::from(f);
|
||||
let body = db.body(f_id.into());
|
||||
let inference_result = db.infer(f_id.into());
|
||||
let (previous_exprs, previous_unknown, previous_partially_unknown) =
|
||||
(num_exprs, num_exprs_unknown, num_exprs_partially_unknown);
|
||||
for (expr_id, _) in body.exprs.iter() {
|
||||
let ty = &inference_result[expr_id];
|
||||
num_exprs += 1;
|
||||
if let Ty::Unknown = ty {
|
||||
num_exprs_unknown += 1;
|
||||
} else {
|
||||
bar.println(format!("unknown location: {}", ty.display(db)));
|
||||
let mut is_partially_unknown = false;
|
||||
ty.walk(&mut |ty| {
|
||||
if let Ty::Unknown = ty {
|
||||
is_partially_unknown = true;
|
||||
}
|
||||
});
|
||||
if is_partially_unknown {
|
||||
num_exprs_partially_unknown += 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
if let Some(mismatch) = inference_result.type_mismatch_for_expr(expr_id) {
|
||||
num_type_mismatches += 1;
|
||||
if verbosity.is_verbose() {
|
||||
if self.only.is_some() && verbosity.is_spammy() {
|
||||
// in super-verbose mode for just one function, we print every single expression
|
||||
let (_, sm) = db.body_with_source_map(f_id.into());
|
||||
let src = sm.expr_syntax(expr_id);
|
||||
if let Ok(src) = src {
|
||||
// FIXME: it might be nice to have a function (on Analysis?) that goes from Source<T> -> (LineCol, LineCol) directly
|
||||
// But also, we should just turn the type mismatches into diagnostics and provide these
|
||||
let root = db.parse_or_expand(src.file_id).unwrap();
|
||||
let node = src.map(|e| e.to_node(&root).syntax().clone());
|
||||
let original_range = original_range(db, node.as_ref());
|
||||
let path = vfs.file_path(original_range.file_id);
|
||||
let line_index =
|
||||
host.analysis().file_line_index(original_range.file_id).unwrap();
|
||||
let text_range = original_range.range;
|
||||
let node = {
|
||||
let root = db.parse_or_expand(src.file_id).unwrap();
|
||||
src.value.to_node(&root)
|
||||
};
|
||||
let original_file = src.file_id.original_file(db);
|
||||
let line_index = host.analysis().file_line_index(original_file).unwrap();
|
||||
let text_range = node.syntax().text_range();
|
||||
let (start, end) = (
|
||||
line_index.line_col(text_range.start()),
|
||||
line_index.line_col(text_range.end()),
|
||||
);
|
||||
bar.println(format!(
|
||||
"{} {}:{}-{}:{}: Expected {}, got {}",
|
||||
path,
|
||||
"{}:{}-{}:{}: {}",
|
||||
start.line + 1,
|
||||
start.col_utf16,
|
||||
end.line + 1,
|
||||
end.col_utf16,
|
||||
mismatch.expected.display(db),
|
||||
mismatch.actual.display(db)
|
||||
ty.display(db)
|
||||
));
|
||||
} else {
|
||||
bar.println(format!(
|
||||
"{}: Expected {}, got {}",
|
||||
name,
|
||||
mismatch.expected.display(db),
|
||||
mismatch.actual.display(db)
|
||||
));
|
||||
bar.println(format!("unknown location: {}", ty.display(db)));
|
||||
}
|
||||
}
|
||||
if let Some(mismatch) = inference_result.type_mismatch_for_expr(expr_id) {
|
||||
num_type_mismatches += 1;
|
||||
if verbosity.is_verbose() {
|
||||
let (_, sm) = db.body_with_source_map(f_id.into());
|
||||
let src = sm.expr_syntax(expr_id);
|
||||
if let Ok(src) = src {
|
||||
// FIXME: it might be nice to have a function (on Analysis?) that goes from Source<T> -> (LineCol, LineCol) directly
|
||||
// But also, we should just turn the type mismatches into diagnostics and provide these
|
||||
let root = db.parse_or_expand(src.file_id).unwrap();
|
||||
let node = src.map(|e| e.to_node(&root).syntax().clone());
|
||||
let original_range = original_range(db, node.as_ref());
|
||||
let path = vfs.file_path(original_range.file_id);
|
||||
let line_index =
|
||||
host.analysis().file_line_index(original_range.file_id).unwrap();
|
||||
let text_range = original_range.range;
|
||||
let (start, end) = (
|
||||
line_index.line_col(text_range.start()),
|
||||
line_index.line_col(text_range.end()),
|
||||
);
|
||||
bar.println(format!(
|
||||
"{} {}:{}-{}:{}: Expected {}, got {}",
|
||||
path,
|
||||
start.line + 1,
|
||||
start.col_utf16,
|
||||
end.line + 1,
|
||||
end.col_utf16,
|
||||
mismatch.expected.display(db),
|
||||
mismatch.actual.display(db)
|
||||
));
|
||||
} else {
|
||||
bar.println(format!(
|
||||
"{}: Expected {}, got {}",
|
||||
name,
|
||||
mismatch.expected.display(db),
|
||||
mismatch.actual.display(db)
|
||||
));
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
if verbosity.is_spammy() {
|
||||
bar.println(format!(
|
||||
"In {}: {} exprs, {} unknown, {} partial",
|
||||
full_name,
|
||||
num_exprs - previous_exprs,
|
||||
num_exprs_unknown - previous_unknown,
|
||||
num_exprs_partially_unknown - previous_partially_unknown
|
||||
));
|
||||
}
|
||||
bar.inc(1);
|
||||
}
|
||||
if verbosity.is_spammy() {
|
||||
bar.println(format!(
|
||||
"In {}: {} exprs, {} unknown, {} partial",
|
||||
full_name,
|
||||
num_exprs - previous_exprs,
|
||||
num_exprs_unknown - previous_unknown,
|
||||
num_exprs_partially_unknown - previous_partially_unknown
|
||||
));
|
||||
bar.finish_and_clear();
|
||||
eprintln!("Total expressions: {}", num_exprs);
|
||||
eprintln!(
|
||||
"Expressions of unknown type: {} ({}%)",
|
||||
num_exprs_unknown,
|
||||
if num_exprs > 0 { num_exprs_unknown * 100 / num_exprs } else { 100 }
|
||||
);
|
||||
report_metric("unknown type", num_exprs_unknown, "#");
|
||||
|
||||
eprintln!(
|
||||
"Expressions of partially unknown type: {} ({}%)",
|
||||
num_exprs_partially_unknown,
|
||||
if num_exprs > 0 { num_exprs_partially_unknown * 100 / num_exprs } else { 100 }
|
||||
);
|
||||
|
||||
eprintln!("Type mismatches: {}", num_type_mismatches);
|
||||
report_metric("type mismatches", num_type_mismatches, "#");
|
||||
|
||||
eprintln!("Inference: {}", inference_sw.elapsed());
|
||||
|
||||
let total_span = analysis_sw.elapsed();
|
||||
eprintln!("Total: {}", total_span);
|
||||
report_metric("total time", total_span.time.as_millis() as u64, "ms");
|
||||
if let Some(instructions) = total_span.instructions {
|
||||
report_metric("total instructions", instructions, "#instr");
|
||||
}
|
||||
if let Some(memory) = total_span.memory {
|
||||
report_metric("total memory", memory.allocated.megabytes() as u64, "MB");
|
||||
}
|
||||
bar.inc(1);
|
||||
}
|
||||
bar.finish_and_clear();
|
||||
eprintln!("Total expressions: {}", num_exprs);
|
||||
eprintln!(
|
||||
"Expressions of unknown type: {} ({}%)",
|
||||
num_exprs_unknown,
|
||||
if num_exprs > 0 { num_exprs_unknown * 100 / num_exprs } else { 100 }
|
||||
);
|
||||
report_metric("unknown type", num_exprs_unknown, "#");
|
||||
|
||||
eprintln!(
|
||||
"Expressions of partially unknown type: {} ({}%)",
|
||||
num_exprs_partially_unknown,
|
||||
if num_exprs > 0 { num_exprs_partially_unknown * 100 / num_exprs } else { 100 }
|
||||
);
|
||||
if self.memory_usage {
|
||||
print_memory_usage(host, vfs);
|
||||
}
|
||||
|
||||
eprintln!("Type mismatches: {}", num_type_mismatches);
|
||||
report_metric("type mismatches", num_type_mismatches, "#");
|
||||
|
||||
eprintln!("Inference: {}", inference_sw.elapsed());
|
||||
|
||||
let total_span = analysis_sw.elapsed();
|
||||
eprintln!("Total: {}", total_span);
|
||||
report_metric("total time", total_span.time.as_millis() as u64, "ms");
|
||||
if let Some(instructions) = total_span.instructions {
|
||||
report_metric("total instructions", instructions, "#instr");
|
||||
}
|
||||
if let Some(memory) = total_span.memory {
|
||||
report_metric("total memory", memory.allocated.megabytes() as u64, "MB");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
if memory_usage {
|
||||
print_memory_usage(host, vfs);
|
||||
fn stop_watch(&self) -> StopWatch {
|
||||
StopWatch::start().memory(self.memory_usage)
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn shuffle<T>(rng: &mut Rand32, slice: &mut [T]) {
|
||||
|
Loading…
Reference in New Issue
Block a user