Switch to fully dynamically dispatched salsa
This improves compile times quite a bit
This commit is contained in:
parent
d4bc2f25de
commit
4bbc385277
8
Cargo.lock
generated
8
Cargo.lock
generated
@ -1490,9 +1490,8 @@ checksum = "71d301d4193d031abdd79ff7e3dd721168a9572ef3fe51a1517aba235bd8f86e"
|
||||
|
||||
[[package]]
|
||||
name = "salsa"
|
||||
version = "0.14.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "d4ca1c656054666a642affbbc86ab95ed7541125a89f032483d34ee56c0f5390"
|
||||
version = "0.14.3"
|
||||
source = "git+https://github.com/nikomatsakis/salsa?branch=dynamic-databases-rfc#fd036a4f154c46253443b3a79b6f4400c40e87b1"
|
||||
dependencies = [
|
||||
"crossbeam-utils",
|
||||
"indexmap",
|
||||
@ -1508,8 +1507,7 @@ dependencies = [
|
||||
[[package]]
|
||||
name = "salsa-macros"
|
||||
version = "0.14.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "038a09b6271446f1123f142fe7e5bef6d4687c4cf82e6986be574c2af3745530"
|
||||
source = "git+https://github.com/nikomatsakis/salsa?branch=dynamic-databases-rfc#fd036a4f154c46253443b3a79b6f4400c40e87b1"
|
||||
dependencies = [
|
||||
"heck",
|
||||
"proc-macro2",
|
||||
|
@ -8,7 +8,7 @@ authors = ["rust-analyzer developers"]
|
||||
doctest = false
|
||||
|
||||
[dependencies]
|
||||
salsa = "0.14.1"
|
||||
salsa = { git = "https://github.com/nikomatsakis/salsa", branch = "dynamic-databases-rfc" }
|
||||
relative-path = "1.0.0"
|
||||
rustc-hash = "1.1.0"
|
||||
|
||||
|
@ -113,7 +113,7 @@ pub trait SourceDatabase: CheckCanceled + FileLoader + std::fmt::Debug {
|
||||
fn crate_graph(&self) -> Arc<CrateGraph>;
|
||||
}
|
||||
|
||||
fn parse_query(db: &impl SourceDatabase, file_id: FileId) -> Parse<ast::SourceFile> {
|
||||
fn parse_query(db: &dyn SourceDatabase, file_id: FileId) -> Parse<ast::SourceFile> {
|
||||
let _p = profile("parse_query").detail(|| format!("{:?}", file_id));
|
||||
let text = db.file_text(file_id);
|
||||
SourceFile::parse(&*text)
|
||||
@ -136,10 +136,7 @@ pub trait SourceDatabaseExt: SourceDatabase {
|
||||
fn source_root_crates(&self, id: SourceRootId) -> Arc<FxHashSet<CrateId>>;
|
||||
}
|
||||
|
||||
fn source_root_crates(
|
||||
db: &(impl SourceDatabaseExt + SourceDatabase),
|
||||
id: SourceRootId,
|
||||
) -> Arc<FxHashSet<CrateId>> {
|
||||
fn source_root_crates(db: &dyn SourceDatabaseExt, id: SourceRootId) -> Arc<FxHashSet<CrateId>> {
|
||||
let graph = db.crate_graph();
|
||||
let res = graph
|
||||
.iter()
|
||||
|
@ -31,7 +31,7 @@ pub struct FunctionData {
|
||||
}
|
||||
|
||||
impl FunctionData {
|
||||
pub(crate) fn fn_data_query(db: &impl DefDatabase, func: FunctionId) -> Arc<FunctionData> {
|
||||
pub(crate) fn fn_data_query(db: &dyn DefDatabase, func: FunctionId) -> Arc<FunctionData> {
|
||||
let loc = func.lookup(db);
|
||||
let item_tree = db.item_tree(loc.id.file_id);
|
||||
let func = &item_tree[loc.id.value];
|
||||
|
@ -1,7 +1,7 @@
|
||||
//! Database used for testing `hir_def`.
|
||||
|
||||
use std::{
|
||||
panic,
|
||||
fmt, panic,
|
||||
sync::{Arc, Mutex},
|
||||
};
|
||||
|
||||
@ -18,10 +18,10 @@
|
||||
crate::db::InternDatabaseStorage,
|
||||
crate::db::DefDatabaseStorage
|
||||
)]
|
||||
#[derive(Debug, Default)]
|
||||
#[derive(Default)]
|
||||
pub struct TestDB {
|
||||
runtime: salsa::Runtime<TestDB>,
|
||||
events: Mutex<Option<Vec<salsa::Event<TestDB>>>>,
|
||||
storage: salsa::Storage<TestDB>,
|
||||
events: Mutex<Option<Vec<salsa::Event>>>,
|
||||
}
|
||||
|
||||
impl Upcast<dyn AstDatabase> for TestDB {
|
||||
@ -37,20 +37,20 @@ fn upcast(&self) -> &(dyn DefDatabase + 'static) {
|
||||
}
|
||||
|
||||
impl salsa::Database for TestDB {
|
||||
fn salsa_runtime(&self) -> &salsa::Runtime<Self> {
|
||||
&self.runtime
|
||||
}
|
||||
fn salsa_runtime_mut(&mut self) -> &mut salsa::Runtime<Self> {
|
||||
&mut self.runtime
|
||||
}
|
||||
fn salsa_event(&self, event: impl Fn() -> salsa::Event<TestDB>) {
|
||||
fn salsa_event(&self, event: salsa::Event) {
|
||||
let mut events = self.events.lock().unwrap();
|
||||
if let Some(events) = &mut *events {
|
||||
events.push(event());
|
||||
events.push(event);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl fmt::Debug for TestDB {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct("TestDB").finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl panic::RefUnwindSafe for TestDB {}
|
||||
|
||||
impl FileLoader for TestDB {
|
||||
@ -78,7 +78,7 @@ pub fn module_for_file(&self, file_id: FileId) -> crate::ModuleId {
|
||||
panic!("Can't find module for file")
|
||||
}
|
||||
|
||||
pub fn log(&self, f: impl FnOnce()) -> Vec<salsa::Event<TestDB>> {
|
||||
pub fn log(&self, f: impl FnOnce()) -> Vec<salsa::Event> {
|
||||
*self.events.lock().unwrap() = Some(Vec::new());
|
||||
f();
|
||||
self.events.lock().unwrap().take().unwrap()
|
||||
@ -92,7 +92,7 @@ pub fn log_executed(&self, f: impl FnOnce()) -> Vec<String> {
|
||||
// This pretty horrible, but `Debug` is the only way to inspect
|
||||
// QueryDescriptor at the moment.
|
||||
salsa::EventKind::WillExecute { database_key } => {
|
||||
Some(format!("{:?}", database_key))
|
||||
Some(format!("{:?}", database_key.debug(self)))
|
||||
}
|
||||
_ => None,
|
||||
})
|
||||
|
@ -1,7 +1,7 @@
|
||||
//! Database used for testing `hir_expand`.
|
||||
|
||||
use std::{
|
||||
panic,
|
||||
fmt, panic,
|
||||
sync::{Arc, Mutex},
|
||||
};
|
||||
|
||||
@ -13,25 +13,23 @@
|
||||
ra_db::SourceDatabaseStorage,
|
||||
crate::db::AstDatabaseStorage
|
||||
)]
|
||||
#[derive(Debug, Default)]
|
||||
#[derive(Default)]
|
||||
pub struct TestDB {
|
||||
runtime: salsa::Runtime<TestDB>,
|
||||
events: Mutex<Option<Vec<salsa::Event<TestDB>>>>,
|
||||
storage: salsa::Storage<TestDB>,
|
||||
events: Mutex<Option<Vec<salsa::Event>>>,
|
||||
}
|
||||
|
||||
impl fmt::Debug for TestDB {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct("TestDB").finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl salsa::Database for TestDB {
|
||||
fn salsa_runtime(&self) -> &salsa::Runtime<Self> {
|
||||
&self.runtime
|
||||
}
|
||||
|
||||
fn salsa_runtime_mut(&mut self) -> &mut salsa::Runtime<Self> {
|
||||
&mut self.runtime
|
||||
}
|
||||
|
||||
fn salsa_event(&self, event: impl Fn() -> salsa::Event<TestDB>) {
|
||||
fn salsa_event(&self, event: salsa::Event) {
|
||||
let mut events = self.events.lock().unwrap();
|
||||
if let Some(events) = &mut *events {
|
||||
events.push(event());
|
||||
events.push(event);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -19,7 +19,6 @@
|
||||
use hir_expand::name::Name;
|
||||
|
||||
#[salsa::query_group(HirDatabaseStorage)]
|
||||
#[salsa::requires(salsa::Database)]
|
||||
pub trait HirDatabase: DefDatabase + Upcast<dyn DefDatabase> {
|
||||
#[salsa::invoke(infer_wait)]
|
||||
#[salsa::transparent]
|
||||
|
@ -1216,7 +1216,7 @@ pub(crate) fn impl_trait_query(db: &dyn HirDatabase, impl_id: ImplId) -> Option<
|
||||
}
|
||||
|
||||
pub(crate) fn return_type_impl_traits(
|
||||
db: &impl HirDatabase,
|
||||
db: &dyn HirDatabase,
|
||||
def: hir_def::FunctionId,
|
||||
) -> Option<Arc<Binders<ReturnTypeImplTraits>>> {
|
||||
// FIXME unify with fn_sig_for_fn instead of doing lowering twice, maybe
|
||||
|
@ -1,7 +1,7 @@
|
||||
//! Database used for testing `hir`.
|
||||
|
||||
use std::{
|
||||
panic,
|
||||
fmt, panic,
|
||||
sync::{Arc, Mutex},
|
||||
};
|
||||
|
||||
@ -26,10 +26,15 @@
|
||||
hir_def::db::DefDatabaseStorage,
|
||||
crate::db::HirDatabaseStorage
|
||||
)]
|
||||
#[derive(Debug, Default)]
|
||||
#[derive(Default)]
|
||||
pub struct TestDB {
|
||||
events: Mutex<Option<Vec<salsa::Event<TestDB>>>>,
|
||||
runtime: salsa::Runtime<TestDB>,
|
||||
storage: salsa::Storage<TestDB>,
|
||||
events: Mutex<Option<Vec<salsa::Event>>>,
|
||||
}
|
||||
impl fmt::Debug for TestDB {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct("TestDB").finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl Upcast<dyn AstDatabase> for TestDB {
|
||||
@ -45,18 +50,10 @@ fn upcast(&self) -> &(dyn DefDatabase + 'static) {
|
||||
}
|
||||
|
||||
impl salsa::Database for TestDB {
|
||||
fn salsa_runtime(&self) -> &salsa::Runtime<TestDB> {
|
||||
&self.runtime
|
||||
}
|
||||
|
||||
fn salsa_runtime_mut(&mut self) -> &mut salsa::Runtime<Self> {
|
||||
&mut self.runtime
|
||||
}
|
||||
|
||||
fn salsa_event(&self, event: impl Fn() -> salsa::Event<TestDB>) {
|
||||
fn salsa_event(&self, event: salsa::Event) {
|
||||
let mut events = self.events.lock().unwrap();
|
||||
if let Some(events) = &mut *events {
|
||||
events.push(event());
|
||||
events.push(event);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -64,8 +61,8 @@ fn salsa_event(&self, event: impl Fn() -> salsa::Event<TestDB>) {
|
||||
impl salsa::ParallelDatabase for TestDB {
|
||||
fn snapshot(&self) -> salsa::Snapshot<TestDB> {
|
||||
salsa::Snapshot::new(TestDB {
|
||||
storage: self.storage.snapshot(),
|
||||
events: Default::default(),
|
||||
runtime: self.runtime.snapshot(self),
|
||||
})
|
||||
}
|
||||
}
|
||||
@ -182,7 +179,7 @@ pub fn extract_annotations(&self) -> FxHashMap<FileId, Vec<(TextRange, String)>>
|
||||
}
|
||||
|
||||
impl TestDB {
|
||||
pub fn log(&self, f: impl FnOnce()) -> Vec<salsa::Event<TestDB>> {
|
||||
pub fn log(&self, f: impl FnOnce()) -> Vec<salsa::Event> {
|
||||
*self.events.lock().unwrap() = Some(Vec::new());
|
||||
f();
|
||||
self.events.lock().unwrap().take().unwrap()
|
||||
@ -196,7 +193,7 @@ pub fn log_executed(&self, f: impl FnOnce()) -> Vec<String> {
|
||||
// This pretty horrible, but `Debug` is the only way to inspect
|
||||
// QueryDescriptor at the moment.
|
||||
salsa::EventKind::WillExecute { database_key } => {
|
||||
Some(format!("{:?}", database_key))
|
||||
Some(format!("{:?}", database_key.debug(self)))
|
||||
}
|
||||
_ => None,
|
||||
})
|
||||
|
@ -21,7 +21,7 @@
|
||||
};
|
||||
use hir_expand::{db::AstDatabase, InFile};
|
||||
use insta::assert_snapshot;
|
||||
use ra_db::{fixture::WithFixture, salsa::Database, FileRange, SourceDatabase};
|
||||
use ra_db::{fixture::WithFixture, FileRange, SourceDatabase, SourceDatabaseExt};
|
||||
use ra_syntax::{
|
||||
algo,
|
||||
ast::{self, AstNode},
|
||||
@ -317,7 +317,7 @@ fn foo() -> i32 {
|
||||
"
|
||||
.to_string();
|
||||
|
||||
db.query_mut(ra_db::FileTextQuery).set(pos.file_id, Arc::new(new_text));
|
||||
db.set_file_text(pos.file_id, Arc::new(new_text));
|
||||
|
||||
{
|
||||
let events = db.log_executed(|| {
|
||||
|
@ -10,7 +10,7 @@
|
||||
|
||||
pub use unsafe_tls::{set_current_program, with_current_program};
|
||||
|
||||
pub struct DebugContext<'a>(&'a (dyn HirDatabase + 'a));
|
||||
pub struct DebugContext<'a>(&'a dyn HirDatabase);
|
||||
|
||||
impl DebugContext<'_> {
|
||||
pub fn debug_struct_id(
|
||||
|
@ -2,10 +2,7 @@
|
||||
|
||||
use hir::MacroFile;
|
||||
use ra_db::{
|
||||
salsa::{
|
||||
debug::{DebugQueryTable, TableEntry},
|
||||
Database,
|
||||
},
|
||||
salsa::debug::{DebugQueryTable, TableEntry},
|
||||
FileTextQuery, SourceRootId,
|
||||
};
|
||||
use ra_ide_db::{
|
||||
@ -14,15 +11,15 @@
|
||||
};
|
||||
use ra_prof::{memory_usage, Bytes};
|
||||
use ra_syntax::{ast, Parse, SyntaxNode};
|
||||
|
||||
use crate::FileId;
|
||||
use rustc_hash::FxHashMap;
|
||||
|
||||
use crate::FileId;
|
||||
|
||||
fn syntax_tree_stats(db: &RootDatabase) -> SyntaxTreeStats {
|
||||
db.query(ra_db::ParseQuery).entries::<SyntaxTreeStats>()
|
||||
ra_db::ParseQuery.in_db(db).entries::<SyntaxTreeStats>()
|
||||
}
|
||||
fn macro_syntax_tree_stats(db: &RootDatabase) -> SyntaxTreeStats {
|
||||
db.query(hir::db::ParseMacroQuery).entries::<SyntaxTreeStats>()
|
||||
hir::db::ParseMacroQuery.in_db(db).entries::<SyntaxTreeStats>()
|
||||
}
|
||||
|
||||
// Feature: Status
|
||||
@ -35,10 +32,10 @@ fn macro_syntax_tree_stats(db: &RootDatabase) -> SyntaxTreeStats {
|
||||
// | VS Code | **Rust Analyzer: Status**
|
||||
// |===
|
||||
pub(crate) fn status(db: &RootDatabase) -> String {
|
||||
let files_stats = db.query(FileTextQuery).entries::<FilesStats>();
|
||||
let files_stats = FileTextQuery.in_db(db).entries::<FilesStats>();
|
||||
let syntax_tree_stats = syntax_tree_stats(db);
|
||||
let macro_syntax_tree_stats = macro_syntax_tree_stats(db);
|
||||
let symbols_stats = db.query(LibrarySymbolsQuery).entries::<LibrarySymbolsStats>();
|
||||
let symbols_stats = LibrarySymbolsQuery.in_db(db).entries::<LibrarySymbolsStats>();
|
||||
format!(
|
||||
"{}\n{}\n{}\n{} (macros)\n\n\nmemory:\n{}\ngc {:?} seconds ago",
|
||||
files_stats,
|
||||
|
@ -147,21 +147,21 @@ pub fn collect_garbage(&mut self) {
|
||||
|
||||
let sweep = SweepStrategy::default().discard_values().sweep_all_revisions();
|
||||
|
||||
self.query(ra_db::ParseQuery).sweep(sweep);
|
||||
self.query(hir::db::ParseMacroQuery).sweep(sweep);
|
||||
ra_db::ParseQuery.in_db(self).sweep(sweep);
|
||||
hir::db::ParseMacroQuery.in_db(self).sweep(sweep);
|
||||
|
||||
// Macros do take significant space, but less then the syntax trees
|
||||
// self.query(hir::db::MacroDefQuery).sweep(sweep);
|
||||
// self.query(hir::db::MacroArgQuery).sweep(sweep);
|
||||
// self.query(hir::db::MacroExpandQuery).sweep(sweep);
|
||||
|
||||
self.query(hir::db::AstIdMapQuery).sweep(sweep);
|
||||
hir::db::AstIdMapQuery.in_db(self).sweep(sweep);
|
||||
|
||||
self.query(hir::db::BodyWithSourceMapQuery).sweep(sweep);
|
||||
hir::db::BodyWithSourceMapQuery.in_db(self).sweep(sweep);
|
||||
|
||||
self.query(hir::db::ExprScopesQuery).sweep(sweep);
|
||||
self.query(hir::db::InferQueryQuery).sweep(sweep);
|
||||
self.query(hir::db::BodyQuery).sweep(sweep);
|
||||
hir::db::ExprScopesQuery.in_db(self).sweep(sweep);
|
||||
hir::db::InferQueryQuery.in_db(self).sweep(sweep);
|
||||
hir::db::BodyQuery.in_db(self).sweep(sweep);
|
||||
}
|
||||
|
||||
pub fn per_query_memory_usage(&mut self) -> Vec<(String, Bytes)> {
|
||||
@ -170,14 +170,14 @@ pub fn per_query_memory_usage(&mut self) -> Vec<(String, Bytes)> {
|
||||
macro_rules! sweep_each_query {
|
||||
($($q:path)*) => {$(
|
||||
let before = memory_usage().allocated;
|
||||
self.query($q).sweep(sweep);
|
||||
$q.in_db(self).sweep(sweep);
|
||||
let after = memory_usage().allocated;
|
||||
let q: $q = Default::default();
|
||||
let name = format!("{:?}", q);
|
||||
acc.push((name, before - after));
|
||||
|
||||
let before = memory_usage().allocated;
|
||||
self.query($q).sweep(sweep.discard_everything());
|
||||
$q.in_db(self).sweep(sweep.discard_everything());
|
||||
let after = memory_usage().allocated;
|
||||
let q: $q = Default::default();
|
||||
let name = format!("{:?} (deps)", q);
|
||||
@ -252,7 +252,7 @@ macro_rules! sweep_each_query {
|
||||
// write.
|
||||
// We do this after collecting the non-interned queries to correctly attribute memory used
|
||||
// by interned data.
|
||||
self.runtime.synthetic_write(Durability::HIGH);
|
||||
self.salsa_runtime_mut().synthetic_write(Durability::HIGH);
|
||||
|
||||
sweep_each_query![
|
||||
// AstDatabase
|
||||
|
@ -11,11 +11,11 @@
|
||||
pub mod source_change;
|
||||
mod wasm_shims;
|
||||
|
||||
use std::sync::Arc;
|
||||
use std::{fmt, sync::Arc};
|
||||
|
||||
use hir::db::{AstDatabase, DefDatabase, HirDatabase};
|
||||
use ra_db::{
|
||||
salsa::{self, Database, Durability},
|
||||
salsa::{self, Durability},
|
||||
Canceled, CheckCanceled, CrateId, FileId, FileLoader, FileLoaderDelegate, SourceDatabase,
|
||||
Upcast,
|
||||
};
|
||||
@ -33,13 +33,18 @@
|
||||
hir::db::DefDatabaseStorage,
|
||||
hir::db::HirDatabaseStorage
|
||||
)]
|
||||
#[derive(Debug)]
|
||||
pub struct RootDatabase {
|
||||
runtime: salsa::Runtime<RootDatabase>,
|
||||
storage: salsa::Storage<RootDatabase>,
|
||||
pub last_gc: crate::wasm_shims::Instant,
|
||||
pub last_gc_check: crate::wasm_shims::Instant,
|
||||
}
|
||||
|
||||
impl fmt::Debug for RootDatabase {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_struct("RootDatabase").finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl Upcast<dyn AstDatabase> for RootDatabase {
|
||||
fn upcast(&self) -> &(dyn AstDatabase + 'static) {
|
||||
&*self
|
||||
@ -71,17 +76,11 @@ fn relevant_crates(&self, file_id: FileId) -> Arc<FxHashSet<CrateId>> {
|
||||
}
|
||||
|
||||
impl salsa::Database for RootDatabase {
|
||||
fn salsa_runtime(&self) -> &salsa::Runtime<RootDatabase> {
|
||||
&self.runtime
|
||||
}
|
||||
fn salsa_runtime_mut(&mut self) -> &mut salsa::Runtime<Self> {
|
||||
&mut self.runtime
|
||||
}
|
||||
fn on_propagated_panic(&self) -> ! {
|
||||
Canceled::throw()
|
||||
}
|
||||
fn salsa_event(&self, event: impl Fn() -> salsa::Event<RootDatabase>) {
|
||||
match event().kind {
|
||||
fn salsa_event(&self, event: salsa::Event) {
|
||||
match event.kind {
|
||||
salsa::EventKind::DidValidateMemoizedValue { .. }
|
||||
| salsa::EventKind::WillExecute { .. } => {
|
||||
self.check_canceled();
|
||||
@ -100,7 +99,7 @@ fn default() -> RootDatabase {
|
||||
impl RootDatabase {
|
||||
pub fn new(lru_capacity: Option<usize>) -> RootDatabase {
|
||||
let mut db = RootDatabase {
|
||||
runtime: salsa::Runtime::default(),
|
||||
storage: salsa::Storage::default(),
|
||||
last_gc: crate::wasm_shims::Instant::now(),
|
||||
last_gc_check: crate::wasm_shims::Instant::now(),
|
||||
};
|
||||
@ -113,16 +112,16 @@ pub fn new(lru_capacity: Option<usize>) -> RootDatabase {
|
||||
|
||||
pub fn update_lru_capacity(&mut self, lru_capacity: Option<usize>) {
|
||||
let lru_capacity = lru_capacity.unwrap_or(ra_db::DEFAULT_LRU_CAP);
|
||||
self.query_mut(ra_db::ParseQuery).set_lru_capacity(lru_capacity);
|
||||
self.query_mut(hir::db::ParseMacroQuery).set_lru_capacity(lru_capacity);
|
||||
self.query_mut(hir::db::MacroExpandQuery).set_lru_capacity(lru_capacity);
|
||||
ra_db::ParseQuery.in_db_mut(self).set_lru_capacity(lru_capacity);
|
||||
hir::db::ParseMacroQuery.in_db_mut(self).set_lru_capacity(lru_capacity);
|
||||
hir::db::MacroExpandQuery.in_db_mut(self).set_lru_capacity(lru_capacity);
|
||||
}
|
||||
}
|
||||
|
||||
impl salsa::ParallelDatabase for RootDatabase {
|
||||
fn snapshot(&self) -> salsa::Snapshot<RootDatabase> {
|
||||
salsa::Snapshot::new(RootDatabase {
|
||||
runtime: self.runtime.snapshot(self),
|
||||
storage: self.storage.snapshot(),
|
||||
last_gc: self.last_gc,
|
||||
last_gc_check: self.last_gc_check,
|
||||
})
|
||||
@ -134,7 +133,7 @@ pub trait LineIndexDatabase: ra_db::SourceDatabase + CheckCanceled {
|
||||
fn line_index(&self, file_id: FileId) -> Arc<LineIndex>;
|
||||
}
|
||||
|
||||
fn line_index(db: &impl LineIndexDatabase, file_id: FileId) -> Arc<LineIndex> {
|
||||
fn line_index(db: &dyn LineIndexDatabase, file_id: FileId) -> Arc<LineIndex> {
|
||||
let text = db.file_text(file_id);
|
||||
Arc::new(LineIndex::new(&*text))
|
||||
}
|
||||
|
@ -87,7 +87,7 @@ pub fn limit(&mut self, limit: usize) {
|
||||
}
|
||||
|
||||
#[salsa::query_group(SymbolsDatabaseStorage)]
|
||||
pub trait SymbolsDatabase: hir::db::HirDatabase + SourceDatabaseExt + ParallelDatabase {
|
||||
pub trait SymbolsDatabase: hir::db::HirDatabase + SourceDatabaseExt {
|
||||
fn file_symbols(&self, file_id: FileId) -> Arc<SymbolIndex>;
|
||||
fn library_symbols(&self) -> Arc<FxHashMap<SourceRootId, SymbolIndex>>;
|
||||
/// The set of "local" (that is, from the current workspace) roots.
|
||||
@ -100,9 +100,7 @@ pub trait SymbolsDatabase: hir::db::HirDatabase + SourceDatabaseExt + ParallelDa
|
||||
fn library_roots(&self) -> Arc<FxHashSet<SourceRootId>>;
|
||||
}
|
||||
|
||||
fn library_symbols(
|
||||
db: &(impl SymbolsDatabase + ParallelDatabase),
|
||||
) -> Arc<FxHashMap<SourceRootId, SymbolIndex>> {
|
||||
fn library_symbols(db: &dyn SymbolsDatabase) -> Arc<FxHashMap<SourceRootId, SymbolIndex>> {
|
||||
let _p = profile("library_symbols");
|
||||
|
||||
let roots = db.library_roots();
|
||||
@ -123,7 +121,7 @@ fn library_symbols(
|
||||
Arc::new(res)
|
||||
}
|
||||
|
||||
fn file_symbols(db: &impl SymbolsDatabase, file_id: FileId) -> Arc<SymbolIndex> {
|
||||
fn file_symbols(db: &dyn SymbolsDatabase, file_id: FileId) -> Arc<SymbolIndex> {
|
||||
db.check_canceled();
|
||||
let parse = db.parse(file_id);
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user