//! The context or environment in which the language server functions. In our //! server implementation this is know as the `WorldState`. //! //! Each tick provides an immutable snapshot of the state as `WorldSnapshot`. use std::{collections::hash_map::Entry, time::Instant}; use crossbeam_channel::{unbounded, Receiver, Sender}; use flycheck::FlycheckHandle; use hir::ChangeWithProcMacros; use ide::{Analysis, AnalysisHost, Cancellable, FileId, SourceRootId}; use ide_db::base_db::{CrateId, ProcMacroPaths}; use load_cargo::SourceRootConfig; use lsp_types::{SemanticTokens, Url}; use nohash_hasher::IntMap; use parking_lot::{ MappedRwLockReadGuard, Mutex, RwLock, RwLockReadGuard, RwLockUpgradableReadGuard, RwLockWriteGuard, }; use proc_macro_api::ProcMacroServer; use project_model::{CargoWorkspace, ProjectWorkspace, Target, WorkspaceBuildScripts}; use rustc_hash::{FxHashMap, FxHashSet}; use triomphe::Arc; use vfs::{AnchoredPathBuf, ChangedFile, Vfs}; use crate::{ config::{Config, ConfigError}, diagnostics::{CheckFixes, DiagnosticCollection}, line_index::{LineEndings, LineIndex}, lsp::{from_proto, to_proto::url_from_abs_path}, lsp_ext, main_loop::Task, mem_docs::MemDocs, op_queue::OpQueue, reload, task_pool::{TaskPool, TaskQueue}, }; // Enforces drop order pub(crate) struct Handle { pub(crate) handle: H, pub(crate) receiver: C, } pub(crate) type ReqHandler = fn(&mut GlobalState, lsp_server::Response); type ReqQueue = lsp_server::ReqQueue<(String, Instant), ReqHandler>; /// `GlobalState` is the primary mutable state of the language server /// /// The most interesting components are `vfs`, which stores a consistent /// snapshot of the file systems, and `analysis_host`, which stores our /// incremental salsa database. /// /// Note that this struct has more than one impl in various modules! #[doc(alias = "GlobalMess")] pub(crate) struct GlobalState { sender: Sender, req_queue: ReqQueue, pub(crate) task_pool: Handle, Receiver>, pub(crate) fmt_pool: Handle, Receiver>, pub(crate) config: Arc, pub(crate) config_errors: Option, pub(crate) analysis_host: AnalysisHost, pub(crate) diagnostics: DiagnosticCollection, pub(crate) mem_docs: MemDocs, pub(crate) source_root_config: SourceRootConfig, /// A mapping that maps a local source root's `SourceRootId` to it parent's `SourceRootId`, if it has one. pub(crate) local_roots_parent_map: FxHashMap, pub(crate) semantic_tokens_cache: Arc>>, // status pub(crate) shutdown_requested: bool, pub(crate) send_hint_refresh_query: bool, pub(crate) last_reported_status: Option, // proc macros pub(crate) proc_macro_clients: Arc<[anyhow::Result]>, pub(crate) build_deps_changed: bool, // Flycheck pub(crate) flycheck: Arc<[FlycheckHandle]>, pub(crate) flycheck_sender: Sender, pub(crate) flycheck_receiver: Receiver, pub(crate) last_flycheck_error: Option, // Test explorer pub(crate) test_run_session: Option, // VFS pub(crate) loader: Handle, Receiver>, pub(crate) vfs: Arc)>>, pub(crate) vfs_config_version: u32, pub(crate) vfs_progress_config_version: u32, pub(crate) vfs_progress_n_total: usize, pub(crate) vfs_progress_n_done: usize, /// `workspaces` field stores the data we actually use, while the `OpQueue` /// stores the result of the last fetch. /// /// If the fetch (partially) fails, we do not update the current value. /// /// The handling of build data is subtle. We fetch workspace in two phases: /// /// *First*, we run `cargo metadata`, which gives us fast results for /// initial analysis. /// /// *Second*, we run `cargo check` which runs build scripts and compiles /// proc macros. /// /// We need both for the precise analysis, but we want rust-analyzer to be /// at least partially available just after the first phase. That's because /// first phase is much faster, and is much less likely to fail. /// /// This creates a complication -- by the time the second phase completes, /// the results of the first phase could be invalid. That is, while we run /// `cargo check`, the user edits `Cargo.toml`, we notice this, and the new /// `cargo metadata` completes before `cargo check`. /// /// An additional complication is that we want to avoid needless work. When /// the user just adds comments or whitespace to Cargo.toml, we do not want /// to invalidate any salsa caches. pub(crate) workspaces: Arc>, pub(crate) crate_graph_file_dependencies: FxHashSet, // op queues pub(crate) fetch_workspaces_queue: OpQueue>, bool)>>, pub(crate) fetch_build_data_queue: OpQueue<(), (Arc>, Vec>)>, pub(crate) fetch_proc_macros_queue: OpQueue, bool>, pub(crate) prime_caches_queue: OpQueue, /// A deferred task queue. /// /// This queue is used for doing database-dependent work inside of sync /// handlers, as accessing the database may block latency-sensitive /// interactions and should be moved away from the main thread. /// /// For certain features, such as [`lsp_ext::UnindexedProjectParams`], /// this queue should run only *after* [`GlobalState::process_changes`] has /// been called. pub(crate) deferred_task_queue: TaskQueue, } /// An immutable snapshot of the world's state at a point in time. pub(crate) struct GlobalStateSnapshot { pub(crate) config: Arc, pub(crate) analysis: Analysis, pub(crate) check_fixes: CheckFixes, mem_docs: MemDocs, pub(crate) semantic_tokens_cache: Arc>>, vfs: Arc)>>, pub(crate) workspaces: Arc>, // used to signal semantic highlighting to fall back to syntax based highlighting until proc-macros have been loaded pub(crate) proc_macros_loaded: bool, pub(crate) flycheck: Arc<[FlycheckHandle]>, } impl std::panic::UnwindSafe for GlobalStateSnapshot {} impl GlobalState { pub(crate) fn new(sender: Sender, config: Config) -> GlobalState { let loader = { let (sender, receiver) = unbounded::(); let handle: vfs_notify::NotifyHandle = vfs::loader::Handle::spawn(Box::new(move |msg| sender.send(msg).unwrap())); let handle = Box::new(handle) as Box; Handle { handle, receiver } }; let task_pool = { let (sender, receiver) = unbounded(); let handle = TaskPool::new_with_threads(sender, config.main_loop_num_threads()); Handle { handle, receiver } }; let fmt_pool = { let (sender, receiver) = unbounded(); let handle = TaskPool::new_with_threads(sender, 1); Handle { handle, receiver } }; let task_queue = { let (sender, receiver) = unbounded(); TaskQueue { sender, receiver } }; let mut analysis_host = AnalysisHost::new(config.lru_parse_query_capacity()); if let Some(capacities) = config.lru_query_capacities() { analysis_host.update_lru_capacities(capacities); } let (flycheck_sender, flycheck_receiver) = unbounded(); let mut this = GlobalState { sender, req_queue: ReqQueue::default(), task_pool, fmt_pool, loader, config: Arc::new(config.clone()), analysis_host, diagnostics: Default::default(), mem_docs: MemDocs::default(), semantic_tokens_cache: Arc::new(Default::default()), shutdown_requested: false, send_hint_refresh_query: false, last_reported_status: None, source_root_config: SourceRootConfig::default(), local_roots_parent_map: FxHashMap::default(), config_errors: Default::default(), proc_macro_clients: Arc::from_iter([]), build_deps_changed: false, flycheck: Arc::from_iter([]), flycheck_sender, flycheck_receiver, last_flycheck_error: None, test_run_session: None, vfs: Arc::new(RwLock::new((vfs::Vfs::default(), IntMap::default()))), vfs_config_version: 0, vfs_progress_config_version: 0, vfs_progress_n_total: 0, vfs_progress_n_done: 0, workspaces: Arc::from(Vec::new()), crate_graph_file_dependencies: FxHashSet::default(), fetch_workspaces_queue: OpQueue::default(), fetch_build_data_queue: OpQueue::default(), fetch_proc_macros_queue: OpQueue::default(), prime_caches_queue: OpQueue::default(), deferred_task_queue: task_queue, }; // Apply any required database inputs from the config. this.update_configuration(config); this } pub(crate) fn process_changes(&mut self) -> bool { let _p = tracing::span!(tracing::Level::INFO, "GlobalState::process_changes").entered(); let mut file_changes = FxHashMap::<_, (bool, ChangedFile)>::default(); let (change, modified_rust_files, workspace_structure_change) = { let mut change = ChangeWithProcMacros::new(); let mut guard = self.vfs.write(); let changed_files = guard.0.take_changes(); if changed_files.is_empty() { return false; } // downgrade to read lock to allow more readers while we are normalizing text let guard = RwLockWriteGuard::downgrade_to_upgradable(guard); let vfs: &Vfs = &guard.0; // We need to fix up the changed events a bit. If we have a create or modify for a file // id that is followed by a delete we actually skip observing the file text from the // earlier event, to avoid problems later on. for changed_file in changed_files { use vfs::Change::*; match file_changes.entry(changed_file.file_id) { Entry::Occupied(mut o) => { let (just_created, change) = o.get_mut(); match (&mut change.change, just_created, changed_file.change) { // latter `Delete` wins (change, _, Delete) => *change = Delete, // merge `Create` with `Create` or `Modify` (Create(prev), _, Create(new) | Modify(new)) => *prev = new, // collapse identical `Modify`es (Modify(prev), _, Modify(new)) => *prev = new, // equivalent to `Modify` (change @ Delete, just_created, Create(new)) => { *change = Modify(new); *just_created = true; } // shouldn't occur, but collapse into `Create` (change @ Delete, just_created, Modify(new)) => { *change = Create(new); *just_created = true; } // shouldn't occur, but keep the Create (prev @ Modify(_), _, new @ Create(_)) => *prev = new, } } Entry::Vacant(v) => { _ = v.insert((matches!(&changed_file.change, Create(_)), changed_file)) } } } let changed_files: Vec<_> = file_changes .into_iter() .filter(|(_, (just_created, change))| { !(*just_created && matches!(change.change, vfs::Change::Delete)) }) .map(|(file_id, (_, change))| vfs::ChangedFile { file_id, ..change }) .collect(); let mut workspace_structure_change = None; // A file was added or deleted let mut has_structure_changes = false; let mut bytes = vec![]; let mut modified_rust_files = vec![]; for file in changed_files { let vfs_path = vfs.file_path(file.file_id); if let Some(path) = vfs_path.as_path() { has_structure_changes = file.is_created_or_deleted(); if file.is_modified() && path.extension() == Some("rs") { modified_rust_files.push(file.file_id); } let path = path.to_path_buf(); if file.is_created_or_deleted() { workspace_structure_change.get_or_insert((path, false)).1 |= self.crate_graph_file_dependencies.contains(vfs_path); } else if reload::should_refresh_for_change(&path, file.kind()) { workspace_structure_change.get_or_insert((path.clone(), false)); } } // Clear native diagnostics when their file gets deleted if !file.exists() { self.diagnostics.clear_native_for(file.file_id); } let text = if let vfs::Change::Create(v) | vfs::Change::Modify(v) = file.change { String::from_utf8(v).ok().map(|text| { // FIXME: Consider doing normalization in the `vfs` instead? That allows // getting rid of some locking let (text, line_endings) = LineEndings::normalize(text); (text, line_endings) }) } else { None }; // delay `line_endings_map` changes until we are done normalizing the text // this allows delaying the re-acquisition of the write lock bytes.push((file.file_id, text)); } let (vfs, line_endings_map) = &mut *RwLockUpgradableReadGuard::upgrade(guard); bytes.into_iter().for_each(|(file_id, text)| match text { None => change.change_file(file_id, None), Some((text, line_endings)) => { line_endings_map.insert(file_id, line_endings); change.change_file(file_id, Some(text)); } }); if has_structure_changes { let roots = self.source_root_config.partition(vfs); change.set_roots(roots); } (change, modified_rust_files, workspace_structure_change) }; self.analysis_host.apply_change(change); { if !matches!(&workspace_structure_change, Some((.., true))) { _ = self .deferred_task_queue .sender .send(crate::main_loop::QueuedTask::CheckProcMacroSources(modified_rust_files)); } // FIXME: ideally we should only trigger a workspace fetch for non-library changes // but something's going wrong with the source root business when we add a new local // crate see https://github.com/rust-lang/rust-analyzer/issues/13029 if let Some((path, force_crate_graph_reload)) = workspace_structure_change { self.fetch_workspaces_queue.request_op( format!("workspace vfs file change: {path}"), force_crate_graph_reload, ); } } true } pub(crate) fn snapshot(&self) -> GlobalStateSnapshot { GlobalStateSnapshot { config: Arc::clone(&self.config), workspaces: Arc::clone(&self.workspaces), analysis: self.analysis_host.analysis(), vfs: Arc::clone(&self.vfs), check_fixes: Arc::clone(&self.diagnostics.check_fixes), mem_docs: self.mem_docs.clone(), semantic_tokens_cache: Arc::clone(&self.semantic_tokens_cache), proc_macros_loaded: !self.config.expand_proc_macros() || *self.fetch_proc_macros_queue.last_op_result(), flycheck: self.flycheck.clone(), } } pub(crate) fn send_request( &mut self, params: R::Params, handler: ReqHandler, ) { let request = self.req_queue.outgoing.register(R::METHOD.to_owned(), params, handler); self.send(request.into()); } pub(crate) fn complete_request(&mut self, response: lsp_server::Response) { let handler = self .req_queue .outgoing .complete(response.id.clone()) .expect("received response for unknown request"); handler(self, response) } pub(crate) fn send_notification( &self, params: N::Params, ) { let not = lsp_server::Notification::new(N::METHOD.to_owned(), params); self.send(not.into()); } pub(crate) fn register_request( &mut self, request: &lsp_server::Request, request_received: Instant, ) { self.req_queue .incoming .register(request.id.clone(), (request.method.clone(), request_received)); } pub(crate) fn respond(&mut self, response: lsp_server::Response) { if let Some((method, start)) = self.req_queue.incoming.complete(response.id.clone()) { if let Some(err) = &response.error { if err.message.starts_with("server panicked") { self.poke_rust_analyzer_developer(format!("{}, check the log", err.message)) } } let duration = start.elapsed(); tracing::debug!("handled {} - ({}) in {:0.2?}", method, response.id, duration); self.send(response.into()); } } pub(crate) fn cancel(&mut self, request_id: lsp_server::RequestId) { if let Some(response) = self.req_queue.incoming.cancel(request_id) { self.send(response.into()); } } pub(crate) fn is_completed(&self, request: &lsp_server::Request) -> bool { self.req_queue.incoming.is_completed(&request.id) } fn send(&self, message: lsp_server::Message) { self.sender.send(message).unwrap() } } impl Drop for GlobalState { fn drop(&mut self) { self.analysis_host.request_cancellation(); } } impl GlobalStateSnapshot { fn vfs_read(&self) -> MappedRwLockReadGuard<'_, vfs::Vfs> { RwLockReadGuard::map(self.vfs.read(), |(it, _)| it) } pub(crate) fn url_to_file_id(&self, url: &Url) -> anyhow::Result { url_to_file_id(&self.vfs_read(), url) } pub(crate) fn file_id_to_url(&self, id: FileId) -> Url { file_id_to_url(&self.vfs_read(), id) } pub(crate) fn file_line_index(&self, file_id: FileId) -> Cancellable { let endings = self.vfs.read().1[&file_id]; let index = self.analysis.file_line_index(file_id)?; let res = LineIndex { index, endings, encoding: self.config.position_encoding() }; Ok(res) } pub(crate) fn url_file_version(&self, url: &Url) -> Option { let path = from_proto::vfs_path(url).ok()?; Some(self.mem_docs.get(&path)?.version) } pub(crate) fn anchored_path(&self, path: &AnchoredPathBuf) -> Url { let mut base = self.vfs_read().file_path(path.anchor).clone(); base.pop(); let path = base.join(&path.path).unwrap(); let path = path.as_path().unwrap(); url_from_abs_path(path) } pub(crate) fn file_id_to_file_path(&self, file_id: FileId) -> vfs::VfsPath { self.vfs_read().file_path(file_id).clone() } pub(crate) fn cargo_target_for_crate_root( &self, crate_id: CrateId, ) -> Option<(&CargoWorkspace, Target)> { let file_id = self.analysis.crate_root(crate_id).ok()?; let path = self.vfs_read().file_path(file_id).clone(); let path = path.as_path()?; self.workspaces.iter().find_map(|ws| match ws { ProjectWorkspace::Cargo { cargo, .. } => { cargo.target_by_root(path).map(|it| (cargo, it)) } ProjectWorkspace::Json { .. } => None, ProjectWorkspace::DetachedFiles { .. } => None, }) } pub(crate) fn file_exists(&self, file_id: FileId) -> bool { self.vfs.read().0.exists(file_id) } } pub(crate) fn file_id_to_url(vfs: &vfs::Vfs, id: FileId) -> Url { let path = vfs.file_path(id); let path = path.as_path().unwrap(); url_from_abs_path(path) } pub(crate) fn url_to_file_id(vfs: &vfs::Vfs, url: &Url) -> anyhow::Result { let path = from_proto::vfs_path(url)?; let res = vfs.file_id(&path).ok_or_else(|| anyhow::format_err!("file not found: {path}"))?; Ok(res) }