Add optimized lock methods for Sharded
This commit is contained in:
parent
3d249706aa
commit
8fc160b742
5 changed files with 133 additions and 33 deletions
|
@ -629,12 +629,7 @@ impl<K: DepKind> DepGraphData<K> {
|
|||
if let Some(prev_index) = self.previous.node_to_index_opt(dep_node) {
|
||||
self.current.prev_index_to_index.lock()[prev_index]
|
||||
} else {
|
||||
self.current
|
||||
.new_node_to_index
|
||||
.get_shard_by_value(dep_node)
|
||||
.lock()
|
||||
.get(dep_node)
|
||||
.copied()
|
||||
self.current.new_node_to_index.lock_shard_by_value(dep_node).get(dep_node).copied()
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -1201,8 +1196,7 @@ impl<K: DepKind> CurrentDepGraph<K> {
|
|||
edges: EdgesVec,
|
||||
current_fingerprint: Fingerprint,
|
||||
) -> DepNodeIndex {
|
||||
let dep_node_index = match self.new_node_to_index.get_shard_by_value(&key).lock().entry(key)
|
||||
{
|
||||
let dep_node_index = match self.new_node_to_index.lock_shard_by_value(&key).entry(key) {
|
||||
Entry::Occupied(entry) => *entry.get(),
|
||||
Entry::Vacant(entry) => {
|
||||
let dep_node_index =
|
||||
|
@ -1328,7 +1322,7 @@ impl<K: DepKind> CurrentDepGraph<K> {
|
|||
) {
|
||||
let node = &prev_graph.index_to_node(prev_index);
|
||||
debug_assert!(
|
||||
!self.new_node_to_index.get_shard_by_value(node).lock().contains_key(node),
|
||||
!self.new_node_to_index.lock_shard_by_value(node).contains_key(node),
|
||||
"node from previous graph present in new node collection"
|
||||
);
|
||||
}
|
||||
|
|
|
@ -55,7 +55,7 @@ where
|
|||
#[inline(always)]
|
||||
fn lookup(&self, key: &K) -> Option<(V, DepNodeIndex)> {
|
||||
let key_hash = sharded::make_hash(key);
|
||||
let lock = self.cache.get_shard_by_hash(key_hash).lock();
|
||||
let lock = self.cache.lock_shard_by_hash(key_hash);
|
||||
let result = lock.raw_entry().from_key_hashed_nocheck(key_hash, key);
|
||||
|
||||
if let Some((_, value)) = result { Some(*value) } else { None }
|
||||
|
@ -63,7 +63,7 @@ where
|
|||
|
||||
#[inline]
|
||||
fn complete(&self, key: K, value: V, index: DepNodeIndex) {
|
||||
let mut lock = self.cache.get_shard_by_value(&key).lock();
|
||||
let mut lock = self.cache.lock_shard_by_value(&key);
|
||||
// We may be overwriting another value. This is all right, since the dep-graph
|
||||
// will check that the fingerprint matches.
|
||||
lock.insert(key, (value, index));
|
||||
|
@ -148,13 +148,13 @@ where
|
|||
|
||||
#[inline(always)]
|
||||
fn lookup(&self, key: &K) -> Option<(V, DepNodeIndex)> {
|
||||
let lock = self.cache.get_shard_by_hash(key.index() as u64).lock();
|
||||
let lock = self.cache.lock_shard_by_hash(key.index() as u64);
|
||||
if let Some(Some(value)) = lock.get(*key) { Some(*value) } else { None }
|
||||
}
|
||||
|
||||
#[inline]
|
||||
fn complete(&self, key: K, value: V, index: DepNodeIndex) {
|
||||
let mut lock = self.cache.get_shard_by_hash(key.index() as u64).lock();
|
||||
let mut lock = self.cache.lock_shard_by_hash(key.index() as u64);
|
||||
lock.insert(key, (value, index));
|
||||
}
|
||||
|
||||
|
|
|
@ -158,7 +158,7 @@ where
|
|||
cache.complete(key, result, dep_node_index);
|
||||
|
||||
let job = {
|
||||
let mut lock = state.active.get_shard_by_value(&key).lock();
|
||||
let mut lock = state.active.lock_shard_by_value(&key);
|
||||
match lock.remove(&key).unwrap() {
|
||||
QueryResult::Started(job) => job,
|
||||
QueryResult::Poisoned => panic!(),
|
||||
|
@ -180,7 +180,7 @@ where
|
|||
// Poison the query so jobs waiting on it panic.
|
||||
let state = self.state;
|
||||
let job = {
|
||||
let mut shard = state.active.get_shard_by_value(&self.key).lock();
|
||||
let mut shard = state.active.lock_shard_by_value(&self.key);
|
||||
let job = match shard.remove(&self.key).unwrap() {
|
||||
QueryResult::Started(job) => job,
|
||||
QueryResult::Poisoned => panic!(),
|
||||
|
@ -303,7 +303,7 @@ where
|
|||
Qcx: QueryContext,
|
||||
{
|
||||
let state = query.query_state(qcx);
|
||||
let mut state_lock = state.active.get_shard_by_value(&key).lock();
|
||||
let mut state_lock = state.active.lock_shard_by_value(&key);
|
||||
|
||||
// For the parallel compiler we need to check both the query cache and query state structures
|
||||
// while holding the state lock to ensure that 1) the query has not yet completed and 2) the
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue