rust/crates/ide/src/prime_caches.rs

175 lines
6.0 KiB
Rust
Raw Normal View History

2021-07-15 14:28:30 -05:00
//! rust-analyzer is lazy and doesn't compute anything unless asked. This
2020-03-05 05:42:04 -06:00
//! sometimes is counter productive when, for example, the first goto definition
2021-07-15 14:28:30 -05:00
//! request takes longer to compute. This modules implemented prepopulation of
2020-03-05 05:42:04 -06:00
//! various caches, it's not really advanced at the moment.
2022-01-14 03:11:47 -06:00
mod topologic_sort;
2020-03-05 05:42:04 -06:00
use hir::db::DefDatabase;
2022-01-14 03:35:41 -06:00
use ide_db::{
base_db::{
salsa::{Database, ParallelDatabase, Snapshot},
CrateGraph, CrateId, SourceDatabase, SourceDatabaseExt,
},
FxIndexMap,
2022-01-14 03:11:47 -06:00
};
use rustc_hash::{FxHashMap, FxHashSet};
2020-03-05 05:42:04 -06:00
use crate::RootDatabase;
/// We started indexing a crate.
#[derive(Debug)]
pub struct PrimeCachesProgress {
pub on_crate: String,
pub n_done: usize,
pub n_total: usize,
}
pub(crate) fn prime_caches(db: &RootDatabase, cb: &(dyn Fn(PrimeCachesProgress) + Sync)) {
let _p = profile::span("prime_caches");
let graph = db.crate_graph();
2022-01-14 03:11:47 -06:00
let to_prime = compute_crates_to_prime(db, &graph);
let n_total = to_prime.len();
for (n_done, &crate_id) in to_prime.iter().enumerate() {
let crate_name = graph[crate_id].display_name.as_deref().unwrap_or_default().to_string();
cb(PrimeCachesProgress { on_crate: crate_name, n_done, n_total });
// This also computes the DefMap
db.import_map(crate_id);
}
}
/// We're indexing many crates.
#[derive(Debug)]
pub struct ParallelPrimeCachesProgress {
/// the crates that we are currently priming.
pub crates_currently_indexing: Vec<String>,
/// the total number of crates we want to prime.
pub crates_total: usize,
/// the total number of crates that have finished priming
pub crates_done: usize,
}
2022-01-14 03:33:40 -06:00
pub(crate) fn parallel_prime_caches(
db: &RootDatabase,
num_worker_threads: u8,
cb: &(dyn Fn(ParallelPrimeCachesProgress) + Sync),
) {
2022-01-14 03:11:47 -06:00
let _p = profile::span("prime_caches");
let graph = db.crate_graph();
let mut crates_to_prime = {
let crate_ids = compute_crates_to_prime(db, &graph);
let mut builder = topologic_sort::TopologicalSortIter::builder();
for &crate_id in &crate_ids {
let crate_data = &graph[crate_id];
let dependencies = crate_data
.dependencies
.iter()
.map(|d| d.crate_id)
.filter(|i| crate_ids.contains(i));
builder.add(crate_id, dependencies);
}
builder.build()
};
crossbeam_utils::thread::scope(move |s| {
let (work_sender, work_receiver) = crossbeam_channel::unbounded();
let (progress_sender, progress_receiver) = crossbeam_channel::unbounded();
enum ParallelPrimeCacheWorkerProgress {
BeginCrate { crate_id: CrateId, crate_name: String },
2022-01-14 03:21:25 -06:00
EndCrate { crate_id: CrateId },
2022-01-14 03:11:47 -06:00
}
let prime_caches_worker = move |db: Snapshot<RootDatabase>| {
while let Ok((crate_id, crate_name)) = work_receiver.recv() {
progress_sender
.send(ParallelPrimeCacheWorkerProgress::BeginCrate { crate_id, crate_name })?;
2022-01-14 03:21:25 -06:00
// This also computes the DefMap
db.import_map(crate_id);
2022-01-14 03:11:47 -06:00
2022-01-14 03:21:25 -06:00
progress_sender.send(ParallelPrimeCacheWorkerProgress::EndCrate { crate_id })?;
2022-01-14 03:11:47 -06:00
}
Ok::<_, crossbeam_channel::SendError<_>>(())
};
for _ in 0..num_worker_threads {
let worker = prime_caches_worker.clone();
let db = db.snapshot();
s.spawn(move |_| worker(db));
}
let crates_total = crates_to_prime.len();
let mut crates_done = 0;
2022-01-14 03:35:41 -06:00
// an index map is used to preserve ordering so we can sort the progress report in order of
// "longest crate to index" first
2022-01-14 03:11:47 -06:00
let mut crates_currently_indexing =
2022-01-14 03:35:41 -06:00
FxIndexMap::with_capacity_and_hasher(num_worker_threads as _, Default::default());
2022-01-14 03:11:47 -06:00
2022-01-14 03:21:25 -06:00
while !crates_to_prime.is_empty() {
db.unwind_if_cancelled();
2022-01-14 03:11:47 -06:00
for crate_id in &mut crates_to_prime {
work_sender
.send((
crate_id,
graph[crate_id].display_name.as_deref().unwrap_or_default().to_string(),
))
.ok();
}
let worker_progress = match progress_receiver.recv() {
Ok(p) => p,
2022-01-14 03:21:25 -06:00
Err(_) => {
// our workers may have died from a cancelled task, so we'll check and re-raise here.
db.unwind_if_cancelled();
break;
}
2022-01-14 03:11:47 -06:00
};
match worker_progress {
ParallelPrimeCacheWorkerProgress::BeginCrate { crate_id, crate_name } => {
crates_currently_indexing.insert(crate_id, crate_name);
}
2022-01-14 03:21:25 -06:00
ParallelPrimeCacheWorkerProgress::EndCrate { crate_id } => {
2022-01-14 03:11:47 -06:00
crates_currently_indexing.remove(&crate_id);
crates_to_prime.mark_done(crate_id);
crates_done += 1;
}
};
let progress = ParallelPrimeCachesProgress {
crates_currently_indexing: crates_currently_indexing.values().cloned().collect(),
crates_done,
crates_total,
};
cb(progress);
}
})
.unwrap();
}
fn compute_crates_to_prime(db: &RootDatabase, graph: &CrateGraph) -> FxHashSet<CrateId> {
// We're only interested in the workspace crates and the `ImportMap`s of their direct
// dependencies, though in practice the latter also compute the `DefMap`s.
// We don't prime transitive dependencies because they're generally not visible in
// the current workspace.
2022-01-14 03:11:47 -06:00
graph
2021-11-11 07:39:20 -06:00
.iter()
.filter(|&id| {
let file_id = graph[id].root_file_id;
let root_id = db.file_source_root(file_id);
!db.source_root(root_id).is_library
})
.flat_map(|id| graph[id].dependencies.iter().map(|krate| krate.crate_id))
2022-01-14 03:11:47 -06:00
.collect()
2020-03-05 05:42:04 -06:00
}