rust-analyzer/crates/hir-expand/src/db.rs

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

798 lines
31 KiB
Rust
Raw Normal View History

2019-10-29 13:03:29 +00:00
//! Defines database & queries for macro expansion.
use base_db::{
salsa::{self, debug::DebugQueryTable},
CrateId, Edition, FileId, SourceDatabase,
};
use either::Either;
2021-07-10 20:49:17 +00:00
use limit::Limit;
2023-11-24 15:38:48 +00:00
use mbe::{syntax_node_to_token_tree, ValueResult};
2023-11-29 14:48:40 +00:00
use rustc_hash::FxHashSet;
2023-12-18 12:30:41 +00:00
use span::SyntaxContextId;
use syntax::{
ast::{self, HasAttrs},
2023-11-24 15:38:48 +00:00
AstNode, Parse, SyntaxError, SyntaxNode, SyntaxToken, T,
};
2023-05-02 14:12:22 +00:00
use triomphe::Arc;
use crate::{
ast_id_map::AstIdMap,
attrs::{collect_attrs, RawAttrs},
builtin_attr_macro::pseudo_derive_attr_expansion,
builtin_fn_macro::EagerExpander,
fixup::{self, reverse_fixups, SyntaxFixupUndoInfo},
hygiene::{apply_mark, SyntaxContextData, Transparency},
proc_macro::ProcMacros,
2023-12-18 12:30:41 +00:00
span_map::{RealSpanMap, SpanMap, SpanMapRef},
tt, AstId, BuiltinAttrExpander, BuiltinDeriveExpander, BuiltinFnLikeExpander,
CustomProcMacroExpander, EagerCallInfo, ExpandError, ExpandResult, ExpandTo, ExpansionSpanMap,
HirFileId, HirFileIdRepr, MacroCallId, MacroCallKind, MacroCallLoc, MacroDefId, MacroDefKind,
MacroFileId,
};
/// Total limit on the number of tokens produced by any macro invocation.
///
/// If an invocation produces more tokens than this limit, it will not be stored in the database and
/// an error will be emitted.
///
/// Actual max for `analysis-stats .` at some point: 30672.
2022-12-10 04:09:18 +00:00
static TOKEN_LIMIT: Limit = Limit::new(1_048_576);
2023-07-10 14:23:29 +00:00
#[derive(Debug, Clone, Eq, PartialEq)]
/// Old-style `macro_rules` or the new macros 2.0
pub struct DeclarativeMacroExpander {
2023-12-18 12:30:41 +00:00
pub mac: mbe::DeclarativeMacro<span::Span>,
2023-11-17 18:07:31 +00:00
pub transparency: Transparency,
2023-07-10 14:23:29 +00:00
}
impl DeclarativeMacroExpander {
2023-11-17 18:07:31 +00:00
pub fn expand(
&self,
db: &dyn ExpandDatabase,
tt: tt::Subtree,
call_id: MacroCallId,
) -> ExpandResult<tt::Subtree> {
match self.mac.err() {
Some(e) => ExpandResult::new(
tt::Subtree::empty(tt::DelimSpan::DUMMY),
2023-11-17 18:07:31 +00:00
ExpandError::other(format!("invalid macro definition: {e}")),
),
None => self
.mac
.expand(&tt, |s| s.ctx = apply_mark(db, s.ctx, call_id, self.transparency))
2023-11-17 18:07:31 +00:00
.map_err(Into::into),
}
}
pub fn expand_unhygienic(&self, tt: tt::Subtree) -> ExpandResult<tt::Subtree> {
2023-07-10 14:23:29 +00:00
match self.mac.err() {
Some(e) => ExpandResult::new(
tt::Subtree::empty(tt::DelimSpan::DUMMY),
2023-07-10 14:23:29 +00:00
ExpandError::other(format!("invalid macro definition: {e}")),
),
2023-11-17 18:07:31 +00:00
None => self.mac.expand(&tt, |_| ()).map_err(Into::into),
2023-07-10 14:23:29 +00:00
}
}
}
2019-11-10 03:03:24 +00:00
#[derive(Debug, Clone, Eq, PartialEq)]
pub enum TokenExpander {
/// Old-style `macro_rules` or the new macros 2.0
2023-07-10 14:23:29 +00:00
DeclarativeMacro(Arc<DeclarativeMacroExpander>),
/// Stuff like `line!` and `file!`.
2023-07-10 14:23:29 +00:00
BuiltIn(BuiltinFnLikeExpander),
/// Built-in eagerly expanded fn-like macros (`include!`, `concat!`, etc.)
2023-07-10 14:23:29 +00:00
BuiltInEager(EagerExpander),
/// `global_allocator` and such.
2023-07-10 14:23:29 +00:00
BuiltInAttr(BuiltinAttrExpander),
/// `derive(Copy)` and such.
2023-07-10 14:23:29 +00:00
BuiltInDerive(BuiltinDeriveExpander),
/// The thing we love the most here in rust-analyzer -- procedural macros.
ProcMacro(CustomProcMacroExpander),
2019-11-10 03:03:24 +00:00
}
#[salsa::query_group(ExpandDatabaseStorage)]
pub trait ExpandDatabase: SourceDatabase {
/// The proc macros.
#[salsa::input]
fn proc_macros(&self) -> Arc<ProcMacros>;
fn ast_id_map(&self, file_id: HirFileId) -> Arc<AstIdMap>;
2021-06-08 19:51:28 +00:00
/// Main public API -- parses a hir file, not caring whether it's a real
2021-05-04 19:40:10 +00:00
/// file or a macro expansion.
#[salsa::transparent]
2023-04-16 17:20:48 +00:00
fn parse_or_expand(&self, file_id: HirFileId) -> SyntaxNode;
#[salsa::transparent]
2023-04-16 17:20:48 +00:00
fn parse_or_expand_with_err(&self, file_id: HirFileId) -> ExpandResult<Parse<SyntaxNode>>;
2021-05-04 19:40:10 +00:00
/// Implementation for the macro case.
// This query is LRU cached
fn parse_macro_expansion(
2020-11-24 18:00:23 +00:00
&self,
macro_file: MacroFileId,
2023-11-24 15:38:48 +00:00
) -> ExpandResult<(Parse<SyntaxNode>, Arc<ExpansionSpanMap>)>;
2023-11-17 18:07:31 +00:00
#[salsa::transparent]
2023-11-24 15:38:48 +00:00
fn span_map(&self, file_id: HirFileId) -> SpanMap;
fn real_span_map(&self, file_id: FileId) -> Arc<RealSpanMap>;
2021-05-04 19:40:10 +00:00
/// Macro ids. That's probably the tricksiest bit in rust-analyzer, and the
/// reason why we use salsa at all.
///
/// We encode macro definitions into ids of macro calls, this what allows us
/// to be incremental.
2021-05-04 18:49:00 +00:00
#[salsa::interned]
fn intern_macro_call(&self, macro_call: MacroCallLoc) -> MacroCallId;
#[salsa::interned]
fn intern_syntax_context(&self, ctx: SyntaxContextData) -> SyntaxContextId;
#[salsa::transparent]
fn setup_syntax_context_root(&self) -> ();
#[salsa::transparent]
fn dump_syntax_contexts(&self) -> String;
2020-05-14 09:57:37 +00:00
/// Lowers syntactic macro call to a token tree representation. That's a firewall
2021-05-04 19:40:10 +00:00
/// query, only typing in the macro call itself changes the returned
/// subtree.
fn macro_arg(
&self,
id: MacroCallId,
2023-11-29 14:48:40 +00:00
) -> ValueResult<Option<(Arc<tt::Subtree>, SyntaxFixupUndoInfo)>, Arc<Box<[SyntaxError]>>>;
/// Fetches the expander for this macro.
2023-07-10 14:23:29 +00:00
#[salsa::transparent]
fn macro_expander(&self, id: MacroDefId) -> TokenExpander;
/// Fetches (and compiles) the expander of this decl macro.
2023-07-10 14:23:29 +00:00
fn decl_macro_expander(
&self,
def_crate: CrateId,
id: AstId<ast::Macro>,
) -> Arc<DeclarativeMacroExpander>;
2021-05-04 19:40:10 +00:00
/// Special case of the previous query for procedural macros. We can't LRU
/// proc macros, since they are not deterministic in general, and
/// non-determinism breaks salsa in a very, very, very bad way.
/// @edwin0cheng heroically debugged this once! See #4315 for details
fn expand_proc_macro(&self, call: MacroCallId) -> ExpandResult<Arc<tt::Subtree>>;
2023-04-16 16:29:42 +00:00
/// Firewall query that returns the errors from the `parse_macro_expansion` query.
fn parse_macro_expansion_error(
&self,
macro_call: MacroCallId,
2023-04-16 17:20:42 +00:00
) -> ExpandResult<Box<[SyntaxError]>>;
2023-11-17 18:07:31 +00:00
}
2021-01-04 02:53:31 +00:00
2023-11-24 15:38:48 +00:00
#[inline]
pub fn span_map(db: &dyn ExpandDatabase, file_id: HirFileId) -> SpanMap {
2023-11-17 18:07:31 +00:00
match file_id.repr() {
2023-11-24 15:38:48 +00:00
HirFileIdRepr::FileId(file_id) => SpanMap::RealSpanMap(db.real_span_map(file_id)),
HirFileIdRepr::MacroFile(m) => {
SpanMap::ExpansionSpanMap(db.parse_macro_expansion(m).value.1)
}
2023-11-17 18:07:31 +00:00
}
}
2023-11-24 15:38:48 +00:00
pub fn real_span_map(db: &dyn ExpandDatabase, file_id: FileId) -> Arc<RealSpanMap> {
Arc::new(RealSpanMap::from_file(db, file_id))
}
/// This expands the given macro call, but with different arguments. This is
/// used for completion, where we want to see what 'would happen' if we insert a
/// token. The `token_to_map` mapped down into the expansion, with the mapped
/// token returned.
pub fn expand_speculative(
db: &dyn ExpandDatabase,
actual_macro_call: MacroCallId,
speculative_args: &SyntaxNode,
2021-05-04 17:36:48 +00:00
token_to_map: SyntaxToken,
) -> Option<(SyntaxNode, SyntaxToken)> {
let loc = db.lookup_intern_macro_call(actual_macro_call);
let span_map = RealSpanMap::absolute(FileId::BOGUS);
2023-11-25 16:10:18 +00:00
let span_map = SpanMapRef::RealSpanMap(&span_map);
2023-11-29 14:48:40 +00:00
// Build the subtree and token mapping for the speculative args
let (mut tt, undo_info) = match loc.kind {
MacroCallKind::FnLike { .. } => {
(mbe::syntax_node_to_token_tree(speculative_args, span_map), SyntaxFixupUndoInfo::NONE)
}
MacroCallKind::Derive { .. } | MacroCallKind::Attr { .. } => {
let censor = censor_for_macro_input(&loc, speculative_args);
let mut fixups = fixup::fixup_syntax(span_map, speculative_args);
fixups.append.retain(|it, _| match it {
syntax::NodeOrToken::Node(it) => !censor.contains(it),
syntax::NodeOrToken::Token(_) => true,
});
fixups.remove.extend(censor);
(
mbe::syntax_node_to_token_tree_modified(
speculative_args,
span_map,
fixups.append,
fixups.remove,
),
fixups.undo_info,
)
}
};
let attr_arg = match loc.kind {
MacroCallKind::Attr { invoc_attr_index, .. } => {
let attr = if loc.def.is_attribute_derive() {
// for pseudo-derive expansion we actually pass the attribute itself only
ast::Attr::cast(speculative_args.clone())
} else {
// Attributes may have an input token tree, build the subtree and map for this as well
// then try finding a token id for our token if it is inside this input subtree.
let item = ast::Item::cast(speculative_args.clone())?;
collect_attrs(&item)
.nth(invoc_attr_index.ast_index())
.and_then(|x| Either::left(x.1))
}?;
match attr.token_tree() {
Some(token_tree) => {
2023-11-25 16:10:18 +00:00
let mut tree = syntax_node_to_token_tree(token_tree.syntax(), span_map);
tree.delimiter = tt::Delimiter::DUMMY_INVISIBLE;
Some(tree)
}
_ => None,
}
}
_ => None,
};
// Do the actual expansion, we need to directly expand the proc macro due to the attribute args
// Otherwise the expand query will fetch the non speculative attribute args and pass those instead.
2023-11-29 14:48:40 +00:00
let mut speculative_expansion = match loc.def.kind {
MacroDefKind::ProcMacro(expander, ..) => {
tt.delimiter = tt::Delimiter::DUMMY_INVISIBLE;
2023-11-28 15:28:51 +00:00
let call_site = loc.span(db);
expander.expand(
db,
loc.def.krate,
loc.krate,
&tt,
attr_arg.as_ref(),
call_site,
call_site,
call_site,
)
}
MacroDefKind::BuiltInAttr(BuiltinAttrExpander::Derive, _) => {
pseudo_derive_attr_expansion(&tt, attr_arg.as_ref()?, loc.call_site)
}
MacroDefKind::BuiltInDerive(expander, ..) => {
// this cast is a bit sus, can we avoid losing the typedness here?
let adt = ast::Adt::cast(speculative_args.clone()).unwrap();
2023-11-25 16:10:18 +00:00
expander.expand(db, actual_macro_call, &adt, span_map)
}
2023-11-17 18:07:31 +00:00
MacroDefKind::Declarative(it) => {
db.decl_macro_expander(loc.krate, it).expand_unhygienic(tt)
}
2023-07-10 14:23:29 +00:00
MacroDefKind::BuiltIn(it, _) => it.expand(db, actual_macro_call, &tt).map_err(Into::into),
MacroDefKind::BuiltInEager(it, _) => {
it.expand(db, actual_macro_call, &tt).map_err(Into::into)
}
MacroDefKind::BuiltInAttr(it, _) => it.expand(db, actual_macro_call, &tt),
};
let expand_to = loc.expand_to();
2023-11-29 14:48:40 +00:00
fixup::reverse_fixups(&mut speculative_expansion.value, &undo_info);
2023-11-28 09:55:21 +00:00
let (node, rev_tmap) = token_tree_to_syntax_node(&speculative_expansion.value, expand_to);
2021-05-04 18:49:00 +00:00
let syntax_node = node.syntax_node();
let token = rev_tmap
.ranges_with_span(span_map.span_for_range(token_to_map.text_range()))
.filter_map(|range| syntax_node.covering_element(range).into_token())
.min_by_key(|t| {
// prefer tokens of the same kind and text
// Note the inversion of the score here, as we want to prefer the first token in case
// of all tokens having the same score
(t.kind() != token_to_map.kind()) as u8 + (t.text() != token_to_map.text()) as u8
})?;
Some((node.syntax_node(), token))
}
fn ast_id_map(db: &dyn ExpandDatabase, file_id: HirFileId) -> Arc<AstIdMap> {
2023-04-16 17:20:48 +00:00
Arc::new(AstIdMap::from_source(&db.parse_or_expand(file_id)))
}
2023-04-16 17:20:48 +00:00
fn parse_or_expand(db: &dyn ExpandDatabase, file_id: HirFileId) -> SyntaxNode {
match file_id.repr() {
HirFileIdRepr::FileId(file_id) => db.parse(file_id).syntax_node(),
2021-05-04 18:49:00 +00:00
HirFileIdRepr::MacroFile(macro_file) => {
2023-04-16 17:20:42 +00:00
db.parse_macro_expansion(macro_file).value.0.syntax_node()
2021-05-04 18:49:00 +00:00
}
2023-04-16 17:20:48 +00:00
}
2021-05-04 18:49:00 +00:00
}
fn parse_or_expand_with_err(
db: &dyn ExpandDatabase,
file_id: HirFileId,
2023-04-16 17:20:48 +00:00
) -> ExpandResult<Parse<SyntaxNode>> {
match file_id.repr() {
2023-04-16 17:20:48 +00:00
HirFileIdRepr::FileId(file_id) => ExpandResult::ok(db.parse(file_id).to_syntax()),
HirFileIdRepr::MacroFile(macro_file) => {
2023-04-16 17:20:48 +00:00
db.parse_macro_expansion(macro_file).map(|(it, _)| it)
}
}
}
2023-11-25 16:10:18 +00:00
// FIXME: We should verify that the parsed node is one of the many macro node variants we expect
// instead of having it be untyped
2021-05-04 18:49:00 +00:00
fn parse_macro_expansion(
db: &dyn ExpandDatabase,
macro_file: MacroFileId,
2023-11-24 15:38:48 +00:00
) -> ExpandResult<(Parse<SyntaxNode>, Arc<ExpansionSpanMap>)> {
2021-05-04 18:49:00 +00:00
let _p = profile::span("parse_macro_expansion");
let loc = db.lookup_intern_macro_call(macro_file.macro_call_id);
let expand_to = loc.expand_to();
let mbe::ValueResult { value: tt, err } = macro_expand(db, macro_file.macro_call_id, loc);
2021-05-04 18:49:00 +00:00
2023-11-28 09:55:21 +00:00
let (parse, rev_token_map) = token_tree_to_syntax_node(&tt, expand_to);
2021-05-04 18:49:00 +00:00
2023-04-16 17:20:42 +00:00
ExpandResult { value: (parse, Arc::new(rev_token_map)), err }
2021-05-04 18:49:00 +00:00
}
fn parse_macro_expansion_error(
db: &dyn ExpandDatabase,
macro_call_id: MacroCallId,
) -> ExpandResult<Box<[SyntaxError]>> {
db.parse_macro_expansion(MacroFileId { macro_call_id })
.map(|it| it.0.errors().to_vec().into_boxed_slice())
}
2023-11-25 16:10:18 +00:00
fn parse_with_map(db: &dyn ExpandDatabase, file_id: HirFileId) -> (Parse<SyntaxNode>, SpanMap) {
match file_id.repr() {
HirFileIdRepr::FileId(file_id) => {
(db.parse(file_id).to_syntax(), SpanMap::RealSpanMap(db.real_span_map(file_id)))
}
HirFileIdRepr::MacroFile(macro_file) => {
let (parse, map) = db.parse_macro_expansion(macro_file).value;
(parse, SpanMap::ExpansionSpanMap(map))
}
}
}
fn macro_arg(
db: &dyn ExpandDatabase,
id: MacroCallId,
2023-11-29 14:48:40 +00:00
// FIXME: consider the following by putting fixup info into eager call info args
// ) -> ValueResult<Option<Arc<(tt::Subtree, SyntaxFixupUndoInfo)>>, Arc<Box<[SyntaxError]>>> {
) -> ValueResult<Option<(Arc<tt::Subtree>, SyntaxFixupUndoInfo)>, Arc<Box<[SyntaxError]>>> {
let mismatched_delimiters = |arg: &SyntaxNode| {
let first = arg.first_child_or_token().map_or(T![.], |it| it.kind());
let last = arg.last_child_or_token().map_or(T![.], |it| it.kind());
let well_formed_tt =
matches!((first, last), (T!['('], T![')']) | (T!['['], T![']']) | (T!['{'], T!['}']));
if !well_formed_tt {
// Don't expand malformed (unbalanced) macro invocations. This is
// less than ideal, but trying to expand unbalanced macro calls
// sometimes produces pathological, deeply nested code which breaks
// all kinds of things.
//
// Some day, we'll have explicit recursion counters for all
// recursive things, at which point this code might be removed.
cov_mark::hit!(issue9358_bad_macro_stack_overflow);
Some(Arc::new(Box::new([SyntaxError::new(
"unbalanced token tree".to_owned(),
arg.text_range(),
)]) as Box<[_]>))
} else {
None
}
};
let loc = db.lookup_intern_macro_call(id);
if let Some(EagerCallInfo { arg, .. }) = matches!(loc.def.kind, MacroDefKind::BuiltInEager(..))
.then(|| loc.eager.as_deref())
.flatten()
{
2023-11-29 14:48:40 +00:00
ValueResult::ok(Some((arg.clone(), SyntaxFixupUndoInfo::NONE)))
} else {
2023-11-25 16:10:18 +00:00
let (parse, map) = parse_with_map(db, loc.kind.file_id());
let root = parse.syntax_node();
2023-11-24 15:38:48 +00:00
let syntax = match loc.kind {
MacroCallKind::FnLike { ast_id, .. } => {
let node = &ast_id.to_ptr(db).to_node(&root);
let offset = node.syntax().text_range().start();
2023-11-25 16:10:18 +00:00
match node.token_tree() {
Some(tt) => {
2023-11-25 16:10:18 +00:00
let tt = tt.syntax();
if let Some(e) = mismatched_delimiters(tt) {
return ValueResult::only_err(e);
}
2023-11-25 16:10:18 +00:00
tt.clone()
}
None => {
return ValueResult::only_err(Arc::new(Box::new([
SyntaxError::new_at_offset("missing token tree".to_owned(), offset),
])));
}
}
}
MacroCallKind::Derive { ast_id, .. } => {
2023-11-24 15:38:48 +00:00
ast_id.to_ptr(db).to_node(&root).syntax().clone()
}
2023-11-24 15:38:48 +00:00
MacroCallKind::Attr { ast_id, .. } => ast_id.to_ptr(db).to_node(&root).syntax().clone(),
};
2023-11-29 14:48:40 +00:00
let (mut tt, undo_info) = match loc.kind {
2023-11-28 15:28:51 +00:00
MacroCallKind::FnLike { .. } => {
2023-11-29 14:48:40 +00:00
(mbe::syntax_node_to_token_tree(&syntax, map.as_ref()), SyntaxFixupUndoInfo::NONE)
2023-11-28 15:28:51 +00:00
}
MacroCallKind::Derive { .. } | MacroCallKind::Attr { .. } => {
2023-11-29 14:48:40 +00:00
let censor = censor_for_macro_input(&loc, &syntax);
let mut fixups = fixup::fixup_syntax(map.as_ref(), &syntax);
fixups.append.retain(|it, _| match it {
syntax::NodeOrToken::Node(it) => !censor.contains(it),
syntax::NodeOrToken::Token(_) => true,
});
fixups.remove.extend(censor);
{
let mut tt = mbe::syntax_node_to_token_tree_modified(
&syntax,
map.as_ref(),
fixups.append.clone(),
fixups.remove.clone(),
);
reverse_fixups(&mut tt, &fixups.undo_info);
}
2023-11-29 14:48:40 +00:00
(
mbe::syntax_node_to_token_tree_modified(
&syntax,
map,
fixups.append,
fixups.remove,
),
fixups.undo_info,
)
2023-11-28 15:28:51 +00:00
}
};
if loc.def.is_proc_macro() {
// proc macros expect their inputs without parentheses, MBEs expect it with them included
tt.delimiter = tt::Delimiter::DUMMY_INVISIBLE;
}
2023-06-07 09:20:10 +00:00
if matches!(loc.def.kind, MacroDefKind::BuiltInEager(..)) {
match parse.errors() {
2023-11-29 14:48:40 +00:00
[] => ValueResult::ok(Some((Arc::new(tt), undo_info))),
errors => ValueResult::new(
2023-11-29 14:48:40 +00:00
Some((Arc::new(tt), undo_info)),
// Box::<[_]>::from(res.errors()), not stable yet
Arc::new(errors.to_vec().into_boxed_slice()),
),
}
} else {
2023-11-29 14:48:40 +00:00
ValueResult::ok(Some((Arc::new(tt), undo_info)))
}
}
}
// FIXME: Censoring info should be calculated by the caller! Namely by name resolution
/// Certain macro calls expect some nodes in the input to be preprocessed away, namely:
/// - derives expect all `#[derive(..)]` invocations up to the currently invoked one to be stripped
/// - attributes expect the invoking attribute to be stripped
2023-11-29 14:48:40 +00:00
fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<SyntaxNode> {
// FIXME: handle `cfg_attr`
(|| {
let censor = match loc.kind {
MacroCallKind::FnLike { .. } => return None,
MacroCallKind::Derive { derive_attr_index, .. } => {
cov_mark::hit!(derive_censoring);
ast::Item::cast(node.clone())?
.attrs()
.take(derive_attr_index.ast_index() + 1)
2022-08-15 14:16:59 +00:00
// FIXME, this resolution should not be done syntactically
// derive is a proper macro now, no longer builtin
// But we do not have resolution at this stage, this means
// we need to know about all macro calls for the given ast item here
// so we require some kind of mapping...
.filter(|attr| attr.simple_name().as_deref() == Some("derive"))
.map(|it| it.syntax().clone())
.collect()
}
MacroCallKind::Attr { .. } if loc.def.is_attribute_derive() => return None,
MacroCallKind::Attr { invoc_attr_index, .. } => {
cov_mark::hit!(attribute_macro_attr_censoring);
collect_attrs(&ast::Item::cast(node.clone())?)
.nth(invoc_attr_index.ast_index())
.and_then(|x| Either::left(x.1))
.map(|attr| attr.syntax().clone())
.into_iter()
.collect()
}
};
Some(censor)
})()
.unwrap_or_default()
2021-05-04 18:49:00 +00:00
}
2023-07-10 14:23:29 +00:00
fn decl_macro_expander(
db: &dyn ExpandDatabase,
2023-07-10 14:23:29 +00:00
def_crate: CrateId,
id: AstId<ast::Macro>,
) -> Arc<DeclarativeMacroExpander> {
let is_2021 = db.crate_graph()[def_crate].edition >= Edition::Edition2021;
2023-11-25 16:10:18 +00:00
let (root, map) = parse_with_map(db, id.file_id);
let root = root.syntax_node();
2023-11-17 18:07:31 +00:00
let transparency = |node| {
// ... would be nice to have the item tree here
2023-11-24 15:38:48 +00:00
let attrs = RawAttrs::new(db, node, map.as_ref()).filter(db, def_crate);
2023-11-17 18:07:31 +00:00
match &*attrs
.iter()
.find(|it| {
it.path.as_ident().and_then(|it| it.as_str()) == Some("rustc_macro_transparency")
})?
.token_tree_value()?
.token_trees
{
[tt::TokenTree::Leaf(tt::Leaf::Ident(i)), ..] => match &*i.text {
"transparent" => Some(Transparency::Transparent),
"semitransparent" => Some(Transparency::SemiTransparent),
"opaque" => Some(Transparency::Opaque),
_ => None,
},
_ => None,
}
};
let (mac, transparency) = match id.to_ptr(db).to_node(&root) {
ast::Macro::MacroRules(macro_rules) => (
match macro_rules.token_tree() {
Some(arg) => {
2023-11-24 15:38:48 +00:00
let tt = mbe::syntax_node_to_token_tree(arg.syntax(), map.as_ref());
2023-11-17 18:07:31 +00:00
let mac = mbe::DeclarativeMacro::parse_macro_rules(&tt, is_2021);
mac
}
None => mbe::DeclarativeMacro::from_err(
mbe::ParseError::Expected("expected a token tree".into()),
is_2021,
),
},
transparency(&macro_rules).unwrap_or(Transparency::SemiTransparent),
),
ast::Macro::MacroDef(macro_def) => (
match macro_def.body() {
Some(arg) => {
2023-11-24 15:38:48 +00:00
let tt = mbe::syntax_node_to_token_tree(arg.syntax(), map.as_ref());
2023-11-17 18:07:31 +00:00
let mac = mbe::DeclarativeMacro::parse_macro2(&tt, is_2021);
mac
}
None => mbe::DeclarativeMacro::from_err(
mbe::ParseError::Expected("expected a token tree".into()),
is_2021,
),
},
transparency(&macro_def).unwrap_or(Transparency::Opaque),
),
2023-07-10 14:23:29 +00:00
};
2023-11-17 18:07:31 +00:00
Arc::new(DeclarativeMacroExpander { mac, transparency })
2023-07-10 14:23:29 +00:00
}
fn macro_expander(db: &dyn ExpandDatabase, id: MacroDefId) -> TokenExpander {
2019-11-11 10:45:55 +00:00
match id.kind {
2021-10-10 18:07:43 +00:00
MacroDefKind::Declarative(ast_id) => {
2023-07-10 14:23:29 +00:00
TokenExpander::DeclarativeMacro(db.decl_macro_expander(id.krate, ast_id))
}
2023-07-10 14:23:29 +00:00
MacroDefKind::BuiltIn(expander, _) => TokenExpander::BuiltIn(expander),
MacroDefKind::BuiltInAttr(expander, _) => TokenExpander::BuiltInAttr(expander),
MacroDefKind::BuiltInDerive(expander, _) => TokenExpander::BuiltInDerive(expander),
MacroDefKind::BuiltInEager(expander, ..) => TokenExpander::BuiltInEager(expander),
MacroDefKind::ProcMacro(expander, ..) => TokenExpander::ProcMacro(expander),
2019-11-10 03:03:24 +00:00
}
2019-10-29 12:11:42 +00:00
}
2023-11-17 18:07:31 +00:00
fn macro_expand(
db: &dyn ExpandDatabase,
macro_call_id: MacroCallId,
loc: MacroCallLoc,
2023-11-17 18:07:31 +00:00
) -> ExpandResult<Arc<tt::Subtree>> {
2020-11-27 17:07:16 +00:00
let _p = profile::span("macro_expand");
2023-07-10 09:22:29 +00:00
let ExpandResult { value: tt, mut err } = match loc.def.kind {
2023-11-17 18:07:31 +00:00
MacroDefKind::ProcMacro(..) => return db.expand_proc_macro(macro_call_id),
MacroDefKind::BuiltInDerive(expander, ..) => {
2023-11-25 16:10:18 +00:00
let (root, map) = parse_with_map(db, loc.kind.file_id());
let root = root.syntax_node();
let MacroCallKind::Derive { ast_id, .. } = loc.kind else { unreachable!() };
let node = ast_id.to_ptr(db).to_node(&root);
2023-11-29 14:48:40 +00:00
// FIXME: Use censoring
let _censor = censor_for_macro_input(&loc, node.syntax());
2023-11-24 15:38:48 +00:00
expander.expand(db, macro_call_id, &node, map.as_ref())
}
_ => {
2023-11-17 18:07:31 +00:00
let ValueResult { value, err } = db.macro_arg(macro_call_id);
2023-11-29 14:48:40 +00:00
let Some((macro_arg, undo_info)) = value else {
return ExpandResult {
value: Arc::new(tt::Subtree {
delimiter: tt::Delimiter::DUMMY_INVISIBLE,
token_trees: Vec::new(),
}),
// FIXME: We should make sure to enforce an invariant that invalid macro
// calls do not reach this call path!
err: Some(ExpandError::other("invalid token tree")),
};
};
let arg = &*macro_arg;
match loc.def.kind {
2023-07-10 14:23:29 +00:00
MacroDefKind::Declarative(id) => {
2023-11-17 18:07:31 +00:00
db.decl_macro_expander(loc.def.krate, id).expand(db, arg.clone(), macro_call_id)
}
MacroDefKind::BuiltIn(it, _) => {
it.expand(db, macro_call_id, &arg).map_err(Into::into)
2023-07-10 14:23:29 +00:00
}
// This might look a bit odd, but we do not expand the inputs to eager macros here.
// Eager macros inputs are expanded, well, eagerly when we collect the macro calls.
// That kind of expansion uses the ast id map of an eager macros input though which goes through
// the HirFileId machinery. As eager macro inputs are assigned a macro file id that query
// will end up going through here again, whereas we want to just want to inspect the raw input.
// As such we just return the input subtree here.
MacroDefKind::BuiltInEager(..) if loc.eager.is_none() => {
return ExpandResult {
2023-11-29 14:48:40 +00:00
value: macro_arg.clone(),
err: err.map(|err| {
let mut buf = String::new();
for err in &**err {
use std::fmt::Write;
_ = write!(buf, "{}, ", err);
}
buf.pop();
buf.pop();
ExpandError::other(buf)
}),
};
}
2023-11-17 18:07:31 +00:00
MacroDefKind::BuiltInEager(it, _) => {
it.expand(db, macro_call_id, &arg).map_err(Into::into)
}
2023-11-29 14:48:40 +00:00
MacroDefKind::BuiltInAttr(it, _) => {
let mut res = it.expand(db, macro_call_id, &arg);
fixup::reverse_fixups(&mut res.value, &undo_info);
res
}
2023-07-10 14:23:29 +00:00
_ => unreachable!(),
}
}
};
2023-06-07 09:20:10 +00:00
if let Some(EagerCallInfo { error, .. }) = loc.eager.as_deref() {
// FIXME: We should report both errors!
err = error.clone().or(err);
}
// Skip checking token tree limit for include! macro call
2023-11-10 14:48:47 +00:00
if !loc.def.is_include() {
// Set a hard limit for the expanded tt
if let Err(value) = check_tt_count(&tt) {
return value;
}
2019-10-29 12:11:42 +00:00
}
2020-11-24 18:00:23 +00:00
2023-04-16 17:20:42 +00:00
ExpandResult { value: Arc::new(tt), err }
2019-10-29 12:11:42 +00:00
}
fn expand_proc_macro(db: &dyn ExpandDatabase, id: MacroCallId) -> ExpandResult<Arc<tt::Subtree>> {
let loc = db.lookup_intern_macro_call(id);
2023-11-29 14:48:40 +00:00
let Some((macro_arg, undo_info)) = db.macro_arg(id).value else {
return ExpandResult {
value: Arc::new(tt::Subtree {
delimiter: tt::Delimiter::DUMMY_INVISIBLE,
token_trees: Vec::new(),
}),
// FIXME: We should make sure to enforce an invariant that invalid macro
// calls do not reach this call path!
2023-07-03 18:34:09 +00:00
err: Some(ExpandError::other("invalid token tree")),
};
};
2020-05-14 09:57:37 +00:00
let expander = match loc.def.kind {
2021-03-18 15:11:18 +00:00
MacroDefKind::ProcMacro(expander, ..) => expander,
2020-05-14 09:57:37 +00:00
_ => unreachable!(),
};
2021-05-31 11:37:11 +00:00
let attr_arg = match &loc.kind {
MacroCallKind::Attr { attr_args: Some(attr_args), .. } => Some(&**attr_args),
2021-05-31 11:37:11 +00:00
_ => None,
};
2023-11-28 15:28:51 +00:00
let call_site = loc.span(db);
2023-11-29 14:48:40 +00:00
let ExpandResult { value: mut tt, err } = expander.expand(
2023-11-28 15:28:51 +00:00
db,
loc.def.krate,
loc.krate,
&macro_arg,
attr_arg,
// FIXME
call_site,
call_site,
// FIXME
call_site,
);
// Set a hard limit for the expanded tt
if let Err(value) = check_tt_count(&tt) {
return value;
}
2023-11-29 14:48:40 +00:00
fixup::reverse_fixups(&mut tt, &undo_info);
ExpandResult { value: Arc::new(tt), err }
2020-05-14 09:57:37 +00:00
}
fn token_tree_to_syntax_node(
tt: &tt::Subtree,
expand_to: ExpandTo,
2023-11-24 15:38:48 +00:00
) -> (Parse<SyntaxNode>, ExpansionSpanMap) {
let entry_point = match expand_to {
2021-12-27 14:54:51 +00:00
ExpandTo::Statements => mbe::TopEntryPoint::MacroStmts,
ExpandTo::Items => mbe::TopEntryPoint::MacroItems,
ExpandTo::Pattern => mbe::TopEntryPoint::Pattern,
ExpandTo::Type => mbe::TopEntryPoint::Type,
ExpandTo::Expr => mbe::TopEntryPoint::Expr,
};
2023-11-28 09:55:21 +00:00
mbe::token_tree_to_syntax_node(tt, entry_point)
2019-12-08 08:16:52 +00:00
}
fn check_tt_count(tt: &tt::Subtree) -> Result<(), ExpandResult<Arc<tt::Subtree>>> {
let count = tt.count();
if TOKEN_LIMIT.check(count).is_err() {
Err(ExpandResult {
value: Arc::new(tt::Subtree {
delimiter: tt::Delimiter::DUMMY_INVISIBLE,
token_trees: vec![],
}),
err: Some(ExpandError::other(format!(
"macro invocation exceeds token limit: produced {} tokens, limit is {}",
count,
TOKEN_LIMIT.inner(),
))),
})
} else {
Ok(())
}
}
fn setup_syntax_context_root(db: &dyn ExpandDatabase) {
db.intern_syntax_context(SyntaxContextData::root());
}
fn dump_syntax_contexts(db: &dyn ExpandDatabase) -> String {
let mut s = String::from("Expansions:");
let mut entries = InternMacroCallLookupQuery.in_db(db).entries::<Vec<_>>();
entries.sort_by_key(|e| e.key);
for e in entries {
let id = e.key;
let expn_data = e.value.as_ref().unwrap();
s.push_str(&format!(
"\n{:?}: parent: {:?}, call_site_ctxt: {:?}, def_site_ctxt: {:?}, kind: {:?}",
id,
expn_data.kind.file_id(),
expn_data.call_site,
SyntaxContextId::ROOT, // FIXME expn_data.def_site,
expn_data.kind.descr(),
));
}
s.push_str("\n\nSyntaxContexts:\n");
let mut entries = InternSyntaxContextLookupQuery.in_db(db).entries::<Vec<_>>();
entries.sort_by_key(|e| e.key);
for e in entries {
struct SyntaxContextDebug<'a>(
&'a dyn ExpandDatabase,
SyntaxContextId,
&'a SyntaxContextData,
);
impl<'a> std::fmt::Debug for SyntaxContextDebug<'a> {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
self.2.fancy_debug(self.1, self.0, f)
}
}
stdx::format_to!(s, "{:?}\n", SyntaxContextDebug(db, e.key, &e.value.unwrap()));
}
s
}