Make query keys Copy

This commit is contained in:
John Kåre Alsaker 2023-02-17 17:33:05 +01:00
parent f4f5fc3e5c
commit 056c5b3b57
3 changed files with 18 additions and 17 deletions

View File

@ -21,7 +21,7 @@ pub trait QueryStorage {
}
pub trait QueryCache: QueryStorage + Sized {
type Key: Hash + Eq + Clone + Debug;
type Key: Hash + Eq + Copy + Debug;
/// Checks if the query is already computed and in the cache.
/// It returns the shard index and a lock guard to the shard,
@ -61,7 +61,7 @@ impl<K: Eq + Hash, V: Copy + Debug> QueryStorage for DefaultCache<K, V> {
impl<K, V> QueryCache for DefaultCache<K, V>
where
K: Eq + Hash + Clone + Debug,
K: Eq + Hash + Copy + Debug,
V: Copy + Debug,
{
type Key = K;
@ -179,7 +179,7 @@ impl<K: Eq + Idx, V: Copy + Debug> QueryStorage for VecCache<K, V> {
impl<K, V> QueryCache for VecCache<K, V>
where
K: Eq + Idx + Clone + Debug,
K: Eq + Idx + Copy + Debug,
V: Copy + Debug,
{
type Key = K;

View File

@ -19,7 +19,9 @@
pub trait QueryConfig<Qcx: QueryContext> {
const NAME: &'static str;
type Key: DepNodeParams<Qcx::DepContext> + Eq + Hash + Clone + Debug;
// `Key` and `Value` are `Copy` instead of `Clone` to ensure copying them stays cheap,
// but it isn't necessary.
type Key: DepNodeParams<Qcx::DepContext> + Eq + Hash + Copy + Debug;
type Value: Debug + Copy;
type Cache: QueryCache<Key = Self::Key, Value = Self::Value>;

View File

@ -49,7 +49,7 @@ enum QueryResult<D: DepKind> {
impl<K, D> QueryState<K, D>
where
K: Eq + Hash + Clone + Debug,
K: Eq + Hash + Copy + Debug,
D: DepKind,
{
pub fn all_inactive(&self) -> bool {
@ -78,7 +78,7 @@ pub fn try_collect_active_jobs<Qcx: Copy>(
for shard in shards.iter() {
for (k, v) in shard.iter() {
if let QueryResult::Started(ref job) = *v {
let query = make_query(qcx, k.clone());
let query = make_query(qcx, *k);
jobs.insert(job.id, QueryJobInfo { query, job: job.clone() });
}
}
@ -92,7 +92,7 @@ pub fn try_collect_active_jobs<Qcx: Copy>(
// really hurt much.)
for (k, v) in self.active.try_lock()?.iter() {
if let QueryResult::Started(ref job) = *v {
let query = make_query(qcx, k.clone());
let query = make_query(qcx, *k);
jobs.insert(job.id, QueryJobInfo { query, job: job.clone() });
}
}
@ -112,7 +112,7 @@ fn default() -> QueryState<K, D> {
/// This will poison the relevant query if dropped.
struct JobOwner<'tcx, K, D: DepKind>
where
K: Eq + Hash + Clone,
K: Eq + Hash + Copy,
{
state: &'tcx QueryState<K, D>,
key: K,
@ -164,7 +164,7 @@ fn handle_cycle_error<Tcx, V>(
impl<'tcx, K, D: DepKind> JobOwner<'tcx, K, D>
where
K: Eq + Hash + Clone,
K: Eq + Hash + Copy,
{
/// Either gets a `JobOwner` corresponding the query, allowing us to
/// start executing the query, or returns with the result of the query.
@ -196,7 +196,7 @@ fn try_start<'b, Qcx>(
let job = qcx.current_query_job();
let job = QueryJob::new(id, span, job);
let key = entry.key().clone();
let key = *entry.key();
entry.insert(QueryResult::Started(job));
let owner = JobOwner { state, id, key };
@ -275,7 +275,7 @@ fn complete<C>(self, cache: &C, result: C::Value, dep_node_index: DepNodeIndex)
impl<'tcx, K, D> Drop for JobOwner<'tcx, K, D>
where
K: Eq + Hash + Clone,
K: Eq + Hash + Copy,
D: DepKind,
{
#[inline(never)]
@ -292,7 +292,7 @@ fn drop(&mut self) {
QueryResult::Started(job) => job,
QueryResult::Poisoned => panic!(),
};
shard.insert(self.key.clone(), QueryResult::Poisoned);
shard.insert(self.key, QueryResult::Poisoned);
job
};
// Also signal the completion of the job, so waiters
@ -311,7 +311,7 @@ pub(crate) struct CycleError<D: DepKind> {
/// The result of `try_start`.
enum TryGetJob<'tcx, K, D>
where
K: Eq + Hash + Clone,
K: Eq + Hash + Copy,
D: DepKind,
{
/// The query is not yet started. Contains a guard to the cache eventually used to start it.
@ -359,10 +359,9 @@ fn try_execute_query<Q, Qcx>(
Q: QueryConfig<Qcx>,
Qcx: QueryContext,
{
match JobOwner::<'_, Q::Key, Qcx::DepKind>::try_start(&qcx, state, span, key.clone()) {
match JobOwner::<'_, Q::Key, Qcx::DepKind>::try_start(&qcx, state, span, key) {
TryGetJob::NotYetStarted(job) => {
let (result, dep_node_index) =
execute_job::<Q, Qcx>(qcx, key.clone(), dep_node, job.id);
let (result, dep_node_index) = execute_job::<Q, Qcx>(qcx, key, dep_node, job.id);
if Q::FEEDABLE {
// We may have put a value inside the cache from inside the execution.
// Verify that it has the same hash as what we have now, to ensure consistency.
@ -547,7 +546,7 @@ fn try_load_from_disk_and_cache_in_memory<Q, Qcx>(
let prof_timer = qcx.dep_context().profiler().query_provider();
// The dep-graph for this computation is already in-place.
let result = dep_graph.with_ignore(|| Q::compute(qcx, key.clone()));
let result = dep_graph.with_ignore(|| Q::compute(qcx, *key));
prof_timer.finish_with_query_invocation_id(dep_node_index.into());