2019-01-08 13:33:36 -06:00
|
|
|
//! This module handles fuzzy-searching of functions, structs and other symbols
|
|
|
|
//! by name across the whole workspace and dependencies.
|
|
|
|
//!
|
|
|
|
//! It works by building an incrementally-updated text-search index of all
|
|
|
|
//! symbols. The backbone of the index is the **awesome** `fst` crate by
|
|
|
|
//! @BurntSushi.
|
|
|
|
//!
|
2019-02-11 10:18:27 -06:00
|
|
|
//! In a nutshell, you give a set of strings to `fst`, and it builds a
|
2019-01-20 20:35:18 -06:00
|
|
|
//! finite state machine describing this set of strings. The strings which
|
2019-01-08 13:33:36 -06:00
|
|
|
//! could fuzzy-match a pattern can also be described by a finite state machine.
|
2019-02-11 10:18:27 -06:00
|
|
|
//! What is freaking cool is that you can now traverse both state machines in
|
2019-01-08 13:33:36 -06:00
|
|
|
//! lock-step to enumerate the strings which are both in the input set and
|
2019-01-20 20:35:18 -06:00
|
|
|
//! fuzz-match the query. Or, more formally, given two languages described by
|
2019-02-11 10:18:27 -06:00
|
|
|
//! FSTs, one can build a product FST which describes the intersection of the
|
2019-01-08 13:33:36 -06:00
|
|
|
//! languages.
|
|
|
|
//!
|
|
|
|
//! `fst` does not support cheap updating of the index, but it supports unioning
|
2019-02-11 10:18:27 -06:00
|
|
|
//! of state machines. So, to account for changing source code, we build an FST
|
|
|
|
//! for each library (which is assumed to never change) and an FST for each Rust
|
2019-01-20 20:35:18 -06:00
|
|
|
//! file in the current workspace, and run a query against the union of all
|
2019-02-11 10:18:27 -06:00
|
|
|
//! those FSTs.
|
2019-01-08 13:33:36 -06:00
|
|
|
use std::{
|
|
|
|
cmp::Ordering,
|
|
|
|
hash::{Hash, Hasher},
|
|
|
|
sync::Arc,
|
2019-01-25 08:20:52 -06:00
|
|
|
mem,
|
2019-03-05 08:09:48 -06:00
|
|
|
fmt,
|
2019-01-08 13:33:36 -06:00
|
|
|
};
|
|
|
|
|
|
|
|
use fst::{self, Streamer};
|
|
|
|
use ra_syntax::{
|
2019-01-23 08:37:10 -06:00
|
|
|
SyntaxNode, SyntaxNodePtr, SourceFile, SmolStr, TreeArc, AstNode,
|
2019-03-14 05:14:54 -05:00
|
|
|
algo::{visit::{visitor, Visitor}},
|
2019-01-08 13:33:36 -06:00
|
|
|
SyntaxKind::{self, *},
|
|
|
|
ast::{self, NameOwner},
|
2019-02-12 13:47:51 -06:00
|
|
|
WalkEvent,
|
2019-02-23 05:05:45 -06:00
|
|
|
TextRange,
|
2019-01-08 13:33:36 -06:00
|
|
|
};
|
2019-01-17 05:11:00 -06:00
|
|
|
use ra_db::{
|
2019-01-26 02:20:30 -06:00
|
|
|
SourceRootId, SourceDatabase,
|
2019-01-17 05:11:00 -06:00
|
|
|
salsa::{self, ParallelDatabase},
|
|
|
|
};
|
2019-01-08 13:33:36 -06:00
|
|
|
use rayon::prelude::*;
|
|
|
|
|
|
|
|
use crate::{
|
2019-01-15 09:19:09 -06:00
|
|
|
FileId, Query,
|
2019-01-08 13:33:36 -06:00
|
|
|
db::RootDatabase,
|
|
|
|
};
|
|
|
|
|
2019-01-25 14:27:16 -06:00
|
|
|
#[salsa::query_group(SymbolsDatabaseStorage)]
|
2019-01-17 05:11:00 -06:00
|
|
|
pub(crate) trait SymbolsDatabase: hir::db::HirDatabase {
|
|
|
|
fn file_symbols(&self, file_id: FileId) -> Arc<SymbolIndex>;
|
|
|
|
#[salsa::input]
|
|
|
|
fn library_symbols(&self, id: SourceRootId) -> Arc<SymbolIndex>;
|
2019-01-26 02:17:05 -06:00
|
|
|
/// The set of "local" (that is, from the current workspace) roots.
|
|
|
|
/// Files in local roots are assumed to change frequently.
|
|
|
|
#[salsa::input]
|
|
|
|
fn local_roots(&self) -> Arc<Vec<SourceRootId>>;
|
|
|
|
/// The set of roots for crates.io libraries.
|
|
|
|
/// Files in libraries are assumed to never change.
|
|
|
|
#[salsa::input]
|
|
|
|
fn library_roots(&self) -> Arc<Vec<SourceRootId>>;
|
2019-01-08 13:33:36 -06:00
|
|
|
}
|
|
|
|
|
2019-01-15 09:19:09 -06:00
|
|
|
fn file_symbols(db: &impl SymbolsDatabase, file_id: FileId) -> Arc<SymbolIndex> {
|
2019-01-15 06:45:48 -06:00
|
|
|
db.check_canceled();
|
2019-01-26 02:51:36 -06:00
|
|
|
let source_file = db.parse(file_id);
|
2019-02-12 13:47:51 -06:00
|
|
|
|
2019-03-14 05:14:54 -05:00
|
|
|
let symbols = source_file_to_file_symbols(&source_file, file_id);
|
2019-01-08 13:33:36 -06:00
|
|
|
|
2019-03-14 05:14:54 -05:00
|
|
|
// TODO: add macros here
|
2019-01-08 13:33:36 -06:00
|
|
|
|
2019-01-15 09:19:09 -06:00
|
|
|
Arc::new(SymbolIndex::new(symbols))
|
2019-01-08 13:33:36 -06:00
|
|
|
}
|
|
|
|
|
2019-01-15 09:19:09 -06:00
|
|
|
pub(crate) fn world_symbols(db: &RootDatabase, query: Query) -> Vec<FileSymbol> {
|
2019-01-08 13:33:36 -06:00
|
|
|
/// Need to wrap Snapshot to provide `Clone` impl for `map_with`
|
|
|
|
struct Snap(salsa::Snapshot<RootDatabase>);
|
|
|
|
impl Clone for Snap {
|
|
|
|
fn clone(&self) -> Snap {
|
|
|
|
Snap(self.0.snapshot())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
let buf: Vec<Arc<SymbolIndex>> = if query.libs {
|
|
|
|
let snap = Snap(db.snapshot());
|
|
|
|
db.library_roots()
|
|
|
|
.par_iter()
|
|
|
|
.map_with(snap, |db, &lib_id| db.0.library_symbols(lib_id))
|
|
|
|
.collect()
|
|
|
|
} else {
|
|
|
|
let mut files = Vec::new();
|
|
|
|
for &root in db.local_roots().iter() {
|
|
|
|
let sr = db.source_root(root);
|
|
|
|
files.extend(sr.files.values().map(|&it| it))
|
|
|
|
}
|
|
|
|
|
|
|
|
let snap = Snap(db.snapshot());
|
2019-02-08 05:49:43 -06:00
|
|
|
files.par_iter().map_with(snap, |db, &file_id| db.0.file_symbols(file_id)).collect()
|
2019-01-08 13:33:36 -06:00
|
|
|
};
|
2019-01-15 09:19:09 -06:00
|
|
|
query.search(&buf)
|
2019-01-08 13:33:36 -06:00
|
|
|
}
|
|
|
|
|
2019-02-08 05:09:57 -06:00
|
|
|
pub(crate) fn index_resolve(db: &RootDatabase, name_ref: &ast::NameRef) -> Vec<FileSymbol> {
|
|
|
|
let name = name_ref.text();
|
|
|
|
let mut query = Query::new(name.to_string());
|
|
|
|
query.exact();
|
|
|
|
query.limit(4);
|
|
|
|
crate::symbol_index::world_symbols(db, query)
|
|
|
|
}
|
|
|
|
|
2019-03-05 08:09:48 -06:00
|
|
|
#[derive(Default)]
|
2019-01-08 13:33:36 -06:00
|
|
|
pub(crate) struct SymbolIndex {
|
|
|
|
symbols: Vec<FileSymbol>,
|
|
|
|
map: fst::Map,
|
|
|
|
}
|
|
|
|
|
2019-03-05 08:09:48 -06:00
|
|
|
impl fmt::Debug for SymbolIndex {
|
|
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
|
|
f.debug_struct("SymbolIndex").field("n_symbols", &self.symbols.len()).finish()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-08 13:33:36 -06:00
|
|
|
impl PartialEq for SymbolIndex {
|
|
|
|
fn eq(&self, other: &SymbolIndex) -> bool {
|
|
|
|
self.symbols == other.symbols
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl Eq for SymbolIndex {}
|
|
|
|
|
|
|
|
impl Hash for SymbolIndex {
|
|
|
|
fn hash<H: Hasher>(&self, hasher: &mut H) {
|
|
|
|
self.symbols.hash(hasher)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl SymbolIndex {
|
|
|
|
fn new(mut symbols: Vec<FileSymbol>) -> SymbolIndex {
|
|
|
|
fn cmp(s1: &FileSymbol, s2: &FileSymbol) -> Ordering {
|
|
|
|
unicase::Ascii::new(s1.name.as_str()).cmp(&unicase::Ascii::new(s2.name.as_str()))
|
|
|
|
}
|
|
|
|
symbols.par_sort_by(cmp);
|
|
|
|
symbols.dedup_by(|s1, s2| cmp(s1, s2) == Ordering::Equal);
|
|
|
|
let names = symbols.iter().map(|it| it.name.as_str().to_lowercase());
|
2019-02-06 14:50:26 -06:00
|
|
|
let map = fst::Map::from_iter(names.zip(0u64..)).unwrap();
|
2019-01-08 13:33:36 -06:00
|
|
|
SymbolIndex { symbols, map }
|
|
|
|
}
|
|
|
|
|
|
|
|
pub(crate) fn len(&self) -> usize {
|
|
|
|
self.symbols.len()
|
|
|
|
}
|
|
|
|
|
2019-01-25 12:10:28 -06:00
|
|
|
pub(crate) fn memory_size(&self) -> usize {
|
2019-01-25 12:10:28 -06:00
|
|
|
self.map.as_fst().size() + self.symbols.len() * mem::size_of::<FileSymbol>()
|
2019-01-25 08:20:52 -06:00
|
|
|
}
|
|
|
|
|
2019-01-08 13:33:36 -06:00
|
|
|
pub(crate) fn for_files(
|
2019-01-11 10:59:06 -06:00
|
|
|
files: impl ParallelIterator<Item = (FileId, TreeArc<SourceFile>)>,
|
2019-01-08 13:33:36 -06:00
|
|
|
) -> SymbolIndex {
|
|
|
|
let symbols = files
|
2019-02-12 13:47:51 -06:00
|
|
|
.flat_map(|(file_id, file)| source_file_to_file_symbols(&file, file_id))
|
2019-01-08 13:33:36 -06:00
|
|
|
.collect::<Vec<_>>();
|
|
|
|
SymbolIndex::new(symbols)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl Query {
|
|
|
|
pub(crate) fn search(self, indices: &[Arc<SymbolIndex>]) -> Vec<FileSymbol> {
|
|
|
|
let mut op = fst::map::OpBuilder::new();
|
|
|
|
for file_symbols in indices.iter() {
|
|
|
|
let automaton = fst::automaton::Subsequence::new(&self.lowercased);
|
|
|
|
op = op.add(file_symbols.map.search(automaton))
|
|
|
|
}
|
|
|
|
let mut stream = op.union();
|
|
|
|
let mut res = Vec::new();
|
|
|
|
while let Some((_, indexed_values)) = stream.next() {
|
|
|
|
if res.len() >= self.limit {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
for indexed_value in indexed_values {
|
|
|
|
let file_symbols = &indices[indexed_value.index];
|
|
|
|
let idx = indexed_value.value as usize;
|
|
|
|
|
|
|
|
let symbol = &file_symbols.symbols[idx];
|
|
|
|
if self.only_types && !is_type(symbol.ptr.kind()) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if self.exact && symbol.name != self.query {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
res.push(symbol.clone());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
res
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn is_type(kind: SyntaxKind) -> bool {
|
|
|
|
match kind {
|
2019-02-25 04:38:52 -06:00
|
|
|
STRUCT_DEF | ENUM_DEF | TRAIT_DEF | TYPE_ALIAS_DEF => true,
|
2019-01-08 13:33:36 -06:00
|
|
|
_ => false,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/// The actual data that is stored in the index. It should be as compact as
|
|
|
|
/// possible.
|
|
|
|
#[derive(Debug, Clone, PartialEq, Eq, Hash)]
|
|
|
|
pub(crate) struct FileSymbol {
|
|
|
|
pub(crate) file_id: FileId,
|
|
|
|
pub(crate) name: SmolStr,
|
2019-01-23 08:37:10 -06:00
|
|
|
pub(crate) ptr: SyntaxNodePtr,
|
2019-02-23 05:05:45 -06:00
|
|
|
pub(crate) name_range: Option<TextRange>,
|
2019-02-12 13:47:51 -06:00
|
|
|
pub(crate) container_name: Option<SmolStr>,
|
|
|
|
}
|
|
|
|
|
|
|
|
fn source_file_to_file_symbols(source_file: &SourceFile, file_id: FileId) -> Vec<FileSymbol> {
|
|
|
|
let mut symbols = Vec::new();
|
|
|
|
let mut stack = Vec::new();
|
|
|
|
|
|
|
|
for event in source_file.syntax().preorder() {
|
|
|
|
match event {
|
|
|
|
WalkEvent::Enter(node) => {
|
|
|
|
if let Some(mut symbol) = to_file_symbol(node, file_id) {
|
2019-02-13 09:42:15 -06:00
|
|
|
symbol.container_name = stack.last().cloned();
|
2019-02-12 13:47:51 -06:00
|
|
|
|
|
|
|
stack.push(symbol.name.clone());
|
|
|
|
symbols.push(symbol);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
WalkEvent::Leave(node) => {
|
2019-02-13 10:02:18 -06:00
|
|
|
if to_symbol(node).is_some() {
|
2019-02-12 13:47:51 -06:00
|
|
|
stack.pop();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
symbols
|
2019-01-08 13:33:36 -06:00
|
|
|
}
|
|
|
|
|
2019-02-23 05:05:45 -06:00
|
|
|
fn to_symbol(node: &SyntaxNode) -> Option<(SmolStr, SyntaxNodePtr, TextRange)> {
|
|
|
|
fn decl<N: NameOwner>(node: &N) -> Option<(SmolStr, SyntaxNodePtr, TextRange)> {
|
|
|
|
let name = node.name()?;
|
|
|
|
let name_range = name.syntax().range();
|
|
|
|
let name = name.text().clone();
|
2019-01-23 08:37:10 -06:00
|
|
|
let ptr = SyntaxNodePtr::new(node.syntax());
|
2019-02-12 13:47:51 -06:00
|
|
|
|
2019-02-23 05:05:45 -06:00
|
|
|
Some((name, ptr, name_range))
|
2019-01-08 13:33:36 -06:00
|
|
|
}
|
|
|
|
visitor()
|
|
|
|
.visit(decl::<ast::FnDef>)
|
|
|
|
.visit(decl::<ast::StructDef>)
|
|
|
|
.visit(decl::<ast::EnumDef>)
|
|
|
|
.visit(decl::<ast::TraitDef>)
|
|
|
|
.visit(decl::<ast::Module>)
|
2019-02-25 04:38:52 -06:00
|
|
|
.visit(decl::<ast::TypeAliasDef>)
|
2019-01-08 13:33:36 -06:00
|
|
|
.visit(decl::<ast::ConstDef>)
|
|
|
|
.visit(decl::<ast::StaticDef>)
|
|
|
|
.accept(node)?
|
|
|
|
}
|
2019-02-12 13:47:51 -06:00
|
|
|
|
|
|
|
fn to_file_symbol(node: &SyntaxNode, file_id: FileId) -> Option<FileSymbol> {
|
2019-02-23 05:05:45 -06:00
|
|
|
to_symbol(node).map(move |(name, ptr, name_range)| FileSymbol {
|
|
|
|
name,
|
|
|
|
ptr,
|
|
|
|
file_id,
|
|
|
|
name_range: Some(name_range),
|
|
|
|
container_name: None,
|
|
|
|
})
|
2019-02-12 13:47:51 -06:00
|
|
|
}
|