//! This module provides `StaticIndex` which is used for powering //! read-only code browsers and emitting LSIF use std::collections::HashMap; use hir::{db::HirDatabase, Crate, Module, Semantics}; use ide_db::{ base_db::{FileId, FileRange, SourceDatabaseExt}, defs::{Definition, IdentClass}, FxHashSet, RootDatabase, }; use syntax::{AstNode, SyntaxKind::*, SyntaxToken, TextRange, T}; use crate::{ hover::hover_for_definition, inlay_hints::AdjustmentHintsMode, moniker::{def_to_moniker, MonikerResult}, parent_module::crates_for, Analysis, Fold, HoverConfig, HoverResult, InlayHint, InlayHintsConfig, TryToNav, }; /// A static representation of fully analyzed source code. /// /// The intended use-case is powering read-only code browsers and emitting LSIF #[derive(Debug)] pub struct StaticIndex<'a> { pub files: Vec, pub tokens: TokenStore, analysis: &'a Analysis, db: &'a RootDatabase, def_map: HashMap, } #[derive(Debug)] pub struct ReferenceData { pub range: FileRange, pub is_definition: bool, } #[derive(Debug)] pub struct TokenStaticData { pub hover: Option, pub definition: Option, pub references: Vec, pub moniker: Option, } #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)] pub struct TokenId(usize); impl TokenId { pub fn raw(self) -> usize { self.0 } } #[derive(Default, Debug)] pub struct TokenStore(Vec); impl TokenStore { pub fn insert(&mut self, data: TokenStaticData) -> TokenId { let id = TokenId(self.0.len()); self.0.push(data); id } pub fn get_mut(&mut self, id: TokenId) -> Option<&mut TokenStaticData> { self.0.get_mut(id.0) } pub fn get(&self, id: TokenId) -> Option<&TokenStaticData> { self.0.get(id.0) } pub fn iter(self) -> impl Iterator { self.0.into_iter().enumerate().map(|(i, x)| (TokenId(i), x)) } } #[derive(Debug)] pub struct StaticIndexedFile { pub file_id: FileId, pub folds: Vec, pub inlay_hints: Vec, pub tokens: Vec<(TextRange, TokenId)>, } fn all_modules(db: &dyn HirDatabase) -> Vec { let mut worklist: Vec<_> = Crate::all(db).into_iter().map(|krate| krate.root_module(db)).collect(); let mut modules = Vec::new(); while let Some(module) = worklist.pop() { modules.push(module); worklist.extend(module.children(db)); } modules } impl StaticIndex<'_> { fn add_file(&mut self, file_id: FileId) { let current_crate = crates_for(self.db, file_id).pop().map(Into::into); let folds = self.analysis.folding_ranges(file_id).unwrap(); let inlay_hints = self .analysis .inlay_hints( &InlayHintsConfig { render_colons: true, discriminant_hints: crate::DiscriminantHints::Fieldless, type_hints: true, parameter_hints: true, chaining_hints: true, closure_return_type_hints: crate::ClosureReturnTypeHints::WithBlock, lifetime_elision_hints: crate::LifetimeElisionHints::Never, adjustment_hints: crate::AdjustmentHints::Never, adjustment_hints_mode: AdjustmentHintsMode::Prefix, adjustment_hints_hide_outside_unsafe: false, hide_named_constructor_hints: false, hide_closure_initialization_hints: false, closure_style: hir::ClosureStyle::ImplFn, param_names_for_lifetime_elision_hints: false, binding_mode_hints: false, max_length: Some(25), closing_brace_hints_min_lines: Some(25), }, file_id, None, ) .unwrap(); // hovers let sema = hir::Semantics::new(self.db); let tokens_or_nodes = sema.parse(file_id).syntax().clone(); let tokens = tokens_or_nodes.descendants_with_tokens().filter_map(|x| match x { syntax::NodeOrToken::Node(_) => None, syntax::NodeOrToken::Token(x) => Some(x), }); let hover_config = HoverConfig { links_in_hover: true, documentation: true, keywords: true, format: crate::HoverDocFormat::Markdown, }; let tokens = tokens.filter(|token| { matches!( token.kind(), IDENT | INT_NUMBER | LIFETIME_IDENT | T![self] | T![super] | T![crate] | T![Self] ) }); let mut result = StaticIndexedFile { file_id, inlay_hints, folds, tokens: vec![] }; for token in tokens { let range = token.text_range(); let node = token.parent().unwrap(); let def = match get_definition(&sema, token.clone()) { Some(x) => x, None => continue, }; let id = if let Some(x) = self.def_map.get(&def) { *x } else { let x = self.tokens.insert(TokenStaticData { hover: hover_for_definition(&sema, file_id, def, &node, &hover_config), definition: def .try_to_nav(self.db) .map(|x| FileRange { file_id: x.file_id, range: x.focus_or_full_range() }), references: vec![], moniker: current_crate.and_then(|cc| def_to_moniker(self.db, def, cc)), }); self.def_map.insert(def, x); x }; let token = self.tokens.get_mut(id).unwrap(); token.references.push(ReferenceData { range: FileRange { range, file_id }, is_definition: match def.try_to_nav(self.db) { Some(x) => x.file_id == file_id && x.focus_or_full_range() == range, None => false, }, }); result.tokens.push((range, id)); } self.files.push(result); } pub fn compute(analysis: &Analysis) -> StaticIndex<'_> { let db = &*analysis.db; let work = all_modules(db).into_iter().filter(|module| { let file_id = module.definition_source(db).file_id.original_file(db); let source_root = db.file_source_root(file_id); let source_root = db.source_root(source_root); !source_root.is_library }); let mut this = StaticIndex { files: vec![], tokens: Default::default(), analysis, db, def_map: Default::default(), }; let mut visited_files = FxHashSet::default(); for module in work { let file_id = module.definition_source(db).file_id.original_file(db); if visited_files.contains(&file_id) { continue; } this.add_file(file_id); // mark the file visited_files.insert(file_id); } this } } fn get_definition(sema: &Semantics<'_, RootDatabase>, token: SyntaxToken) -> Option { for token in sema.descend_into_macros(token) { let def = IdentClass::classify_token(sema, &token).map(IdentClass::definitions_no_ops); if let Some(&[x]) = def.as_deref() { return Some(x); } } None } #[cfg(test)] mod tests { use crate::{fixture, StaticIndex}; use ide_db::base_db::FileRange; use std::collections::HashSet; use syntax::TextSize; fn check_all_ranges(ra_fixture: &str) { let (analysis, ranges) = fixture::annotations_without_marker(ra_fixture); let s = StaticIndex::compute(&analysis); let mut range_set: HashSet<_> = ranges.iter().map(|x| x.0).collect(); for f in s.files { for (range, _) in f.tokens { let x = FileRange { file_id: f.file_id, range }; if !range_set.contains(&x) { panic!("additional range {x:?}"); } range_set.remove(&x); } } if !range_set.is_empty() { panic!("unfound ranges {range_set:?}"); } } fn check_definitions(ra_fixture: &str) { let (analysis, ranges) = fixture::annotations_without_marker(ra_fixture); let s = StaticIndex::compute(&analysis); let mut range_set: HashSet<_> = ranges.iter().map(|x| x.0).collect(); for (_, t) in s.tokens.iter() { if let Some(x) = t.definition { if x.range.start() == TextSize::from(0) { // ignore definitions that are whole of file continue; } if !range_set.contains(&x) { panic!("additional definition {x:?}"); } range_set.remove(&x); } } if !range_set.is_empty() { panic!("unfound definitions {range_set:?}"); } } #[test] fn struct_and_enum() { check_all_ranges( r#" struct Foo; //^^^ enum E { X(Foo) } //^ ^ ^^^ "#, ); check_definitions( r#" struct Foo; //^^^ enum E { X(Foo) } //^ ^ "#, ); } #[test] fn multi_crate() { check_definitions( r#" //- /main.rs crate:main deps:foo use foo::func; fn main() { //^^^^ func(); } //- /foo/lib.rs crate:foo pub func() { } "#, ); } #[test] fn derives() { check_all_ranges( r#" //- minicore:derive #[rustc_builtin_macro] //^^^^^^^^^^^^^^^^^^^ pub macro Copy {} //^^^^ #[derive(Copy)] //^^^^^^ ^^^^ struct Hello(i32); //^^^^^ ^^^ "#, ); } }