mirror of
https://github.com/rust-lang/rust-analyzer
synced 2025-01-11 20:58:54 +00:00
more frugal map
This commit is contained in:
parent
4f3fc6fa1a
commit
278e7c3311
3 changed files with 29 additions and 16 deletions
2
Cargo.lock
generated
2
Cargo.lock
generated
|
@ -525,11 +525,11 @@ dependencies = [
|
||||||
"cfg",
|
"cfg",
|
||||||
"cov-mark",
|
"cov-mark",
|
||||||
"either",
|
"either",
|
||||||
|
"hashbrown",
|
||||||
"itertools",
|
"itertools",
|
||||||
"la-arena",
|
"la-arena",
|
||||||
"limit",
|
"limit",
|
||||||
"mbe",
|
"mbe",
|
||||||
"once_cell",
|
|
||||||
"profile",
|
"profile",
|
||||||
"rustc-hash",
|
"rustc-hash",
|
||||||
"syntax",
|
"syntax",
|
||||||
|
|
|
@ -16,7 +16,7 @@ either = "1.5.3"
|
||||||
rustc-hash = "1.0.0"
|
rustc-hash = "1.0.0"
|
||||||
la-arena = { version = "0.3.0", path = "../../lib/arena" }
|
la-arena = { version = "0.3.0", path = "../../lib/arena" }
|
||||||
itertools = "0.10.0"
|
itertools = "0.10.0"
|
||||||
once_cell = "1"
|
hashbrown = { version = "0.11", features = ["inline-more"], default-features = false }
|
||||||
|
|
||||||
base_db = { path = "../base_db", version = "0.0.0" }
|
base_db = { path = "../base_db", version = "0.0.0" }
|
||||||
cfg = { path = "../cfg", version = "0.0.0" }
|
cfg = { path = "../cfg", version = "0.0.0" }
|
||||||
|
|
|
@ -8,13 +8,13 @@
|
||||||
use std::{
|
use std::{
|
||||||
any::type_name,
|
any::type_name,
|
||||||
fmt,
|
fmt,
|
||||||
hash::{Hash, Hasher},
|
hash::{BuildHasher, BuildHasherDefault, Hash, Hasher},
|
||||||
marker::PhantomData,
|
marker::PhantomData,
|
||||||
};
|
};
|
||||||
|
|
||||||
use la_arena::{Arena, Idx};
|
use la_arena::{Arena, Idx};
|
||||||
use profile::Count;
|
use profile::Count;
|
||||||
use rustc_hash::FxHashMap;
|
use rustc_hash::FxHasher;
|
||||||
use syntax::{ast, match_ast, AstNode, AstPtr, SyntaxNode, SyntaxNodePtr};
|
use syntax::{ast, match_ast, AstNode, AstPtr, SyntaxNode, SyntaxNodePtr};
|
||||||
|
|
||||||
/// `AstId` points to an AST node in a specific file.
|
/// `AstId` points to an AST node in a specific file.
|
||||||
|
@ -63,11 +63,10 @@ type ErasedFileAstId = Idx<SyntaxNodePtr>;
|
||||||
/// Maps items' `SyntaxNode`s to `ErasedFileAstId`s and back.
|
/// Maps items' `SyntaxNode`s to `ErasedFileAstId`s and back.
|
||||||
#[derive(Default)]
|
#[derive(Default)]
|
||||||
pub struct AstIdMap {
|
pub struct AstIdMap {
|
||||||
|
/// Maps stable id to unstable ptr.
|
||||||
arena: Arena<SyntaxNodePtr>,
|
arena: Arena<SyntaxNodePtr>,
|
||||||
/// Reversed mapping lazily derived from [`self.arena`].
|
/// Reverse: map ptr to id.
|
||||||
///
|
map: hashbrown::HashMap<Idx<SyntaxNodePtr>, (), ()>,
|
||||||
/// FIXE: Do not store `SyntaxNodePtr` twice.
|
|
||||||
map: once_cell::sync::OnceCell<FxHashMap<SyntaxNodePtr, ErasedFileAstId>>,
|
|
||||||
_c: Count<Self>,
|
_c: Count<Self>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -107,26 +106,34 @@ impl AstIdMap {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
res.map = hashbrown::HashMap::with_capacity_and_hasher(res.arena.len(), ());
|
||||||
|
for (idx, ptr) in res.arena.iter() {
|
||||||
|
let hash = hash_ptr(ptr);
|
||||||
|
match res.map.raw_entry_mut().from_hash(hash, |idx2| *idx2 == idx) {
|
||||||
|
hashbrown::hash_map::RawEntryMut::Occupied(_) => unreachable!(),
|
||||||
|
hashbrown::hash_map::RawEntryMut::Vacant(entry) => {
|
||||||
|
entry.insert_with_hasher(hash, idx, (), |&idx| hash_ptr(&res.arena[idx]));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
res
|
res
|
||||||
}
|
}
|
||||||
fn map(&self) -> &FxHashMap<SyntaxNodePtr, ErasedFileAstId> {
|
|
||||||
self.map.get_or_init(|| self.arena.iter().map(|(idx, ptr)| (ptr.clone(), idx)).collect())
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn ast_id<N: AstNode>(&self, item: &N) -> FileAstId<N> {
|
pub fn ast_id<N: AstNode>(&self, item: &N) -> FileAstId<N> {
|
||||||
let raw = self.erased_ast_id(item.syntax());
|
let raw = self.erased_ast_id(item.syntax());
|
||||||
FileAstId { raw, _ty: PhantomData }
|
FileAstId { raw, _ty: PhantomData }
|
||||||
}
|
}
|
||||||
|
|
||||||
fn erased_ast_id(&self, item: &SyntaxNode) -> ErasedFileAstId {
|
fn erased_ast_id(&self, item: &SyntaxNode) -> ErasedFileAstId {
|
||||||
let ptr = SyntaxNodePtr::new(item);
|
let ptr = SyntaxNodePtr::new(item);
|
||||||
*self.map().get(&ptr).unwrap_or_else(|| {
|
let hash = hash_ptr(&ptr);
|
||||||
panic!(
|
match self.map.raw_entry().from_hash(hash, |&idx| self.arena[idx] == ptr) {
|
||||||
|
Some((&idx, &())) => idx,
|
||||||
|
None => panic!(
|
||||||
"Can't find {:?} in AstIdMap:\n{:?}",
|
"Can't find {:?} in AstIdMap:\n{:?}",
|
||||||
item,
|
item,
|
||||||
self.arena.iter().map(|(_id, i)| i).collect::<Vec<_>>(),
|
self.arena.iter().map(|(_id, i)| i).collect::<Vec<_>>(),
|
||||||
)
|
),
|
||||||
})
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get<N: AstNode>(&self, id: FileAstId<N>) -> AstPtr<N> {
|
pub fn get<N: AstNode>(&self, id: FileAstId<N>) -> AstPtr<N> {
|
||||||
|
@ -138,6 +145,12 @@ impl AstIdMap {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn hash_ptr(ptr: &SyntaxNodePtr) -> u64 {
|
||||||
|
let mut hasher = BuildHasherDefault::<FxHasher>::default().build_hasher();
|
||||||
|
ptr.hash(&mut hasher);
|
||||||
|
hasher.finish()
|
||||||
|
}
|
||||||
|
|
||||||
/// Walks the subtree in bdfs order, calling `f` for each node. What is bdfs
|
/// Walks the subtree in bdfs order, calling `f` for each node. What is bdfs
|
||||||
/// order? It is a mix of breadth-first and depth first orders. Nodes for which
|
/// order? It is a mix of breadth-first and depth first orders. Nodes for which
|
||||||
/// `f` returns true are visited breadth-first, all the other nodes are explored
|
/// `f` returns true are visited breadth-first, all the other nodes are explored
|
||||||
|
|
Loading…
Reference in a new issue