Replace expressions with errors in them

This commit is contained in:
Florian Diebold 2022-02-09 11:58:52 +01:00
parent 30287e6051
commit ecf3cff4a6
7 changed files with 84 additions and 33 deletions

View file

@ -186,7 +186,7 @@ pub fn identity_when_valid(_attr: TokenStream, item: TokenStream) -> TokenStream
let range: Range<usize> = range.into(); let range: Range<usize> = range.into();
if show_token_ids { if show_token_ids {
if let Some((tree, map)) = arg.as_deref() { if let Some((tree, map, _)) = arg.as_deref() {
let tt_range = call.token_tree().unwrap().syntax().text_range(); let tt_range = call.token_tree().unwrap().syntax().text_range();
let mut ranges = Vec::new(); let mut ranges = Vec::new();
extract_id_ranges(&mut ranges, &map, &tree); extract_id_ranges(&mut ranges, &map, &tree);

View file

@ -108,7 +108,10 @@ pub trait AstDatabase: SourceDatabase {
/// Lowers syntactic macro call to a token tree representation. /// Lowers syntactic macro call to a token tree representation.
#[salsa::transparent] #[salsa::transparent]
fn macro_arg(&self, id: MacroCallId) -> Option<Arc<(tt::Subtree, mbe::TokenMap)>>; fn macro_arg(
&self,
id: MacroCallId,
) -> Option<Arc<(tt::Subtree, mbe::TokenMap, fixup::SyntaxFixupMap)>>;
/// Extracts syntax node, corresponding to a macro call. That's a firewall /// Extracts syntax node, corresponding to a macro call. That's a firewall
/// query, only typing in the macro call itself changes the returned /// query, only typing in the macro call itself changes the returned
/// subtree. /// subtree.
@ -291,29 +294,27 @@ fn parse_macro_expansion(
} }
} }
fn macro_arg(db: &dyn AstDatabase, id: MacroCallId) -> Option<Arc<(tt::Subtree, mbe::TokenMap)>> { fn macro_arg(
db: &dyn AstDatabase,
id: MacroCallId,
) -> Option<Arc<(tt::Subtree, mbe::TokenMap, fixup::SyntaxFixupMap)>> {
let arg = db.macro_arg_text(id)?; let arg = db.macro_arg_text(id)?;
let loc = db.lookup_intern_macro_call(id); let loc = db.lookup_intern_macro_call(id);
let node = SyntaxNode::new_root(arg); let node = SyntaxNode::new_root(arg);
eprintln!("input text:\n{node}");
eprintln!("input syntax:\n{node:#?}");
let censor = censor_for_macro_input(&loc, &node); let censor = censor_for_macro_input(&loc, &node);
// TODO only fixup for attribute macro input // TODO only fixup for attribute macro input
let mut fixups = fixup::fixup_syntax(&node); let mut fixups = fixup::fixup_syntax(&node);
fixups.replace.extend(censor.into_iter().map(|node| (node, Vec::new()))); fixups.replace.extend(censor.into_iter().map(|node| (node, Vec::new())));
eprintln!("fixups: {fixups:?}");
let (mut tt, tmap) = let (mut tt, tmap) =
mbe::syntax_node_to_token_tree_censored(&node, fixups.replace, fixups.append); mbe::syntax_node_to_token_tree_censored(&node, fixups.replace, fixups.append);
eprintln!("fixed-up input: {}", tt);
if loc.def.is_proc_macro() { if loc.def.is_proc_macro() {
// proc macros expect their inputs without parentheses, MBEs expect it with them included // proc macros expect their inputs without parentheses, MBEs expect it with them included
tt.delimiter = None; tt.delimiter = None;
} }
Some(Arc::new((tt, tmap))) Some(Arc::new((tt, tmap, fixups.map)))
} }
fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<SyntaxNode> { fn censor_for_macro_input(loc: &MacroCallLoc, node: &SyntaxNode) -> FxHashSet<SyntaxNode> {
@ -433,7 +434,6 @@ fn macro_expand(db: &dyn AstDatabase, id: MacroCallId) -> ExpandResult<Option<Ar
let ExpandResult { value: mut tt, err } = expander.expand(db, id, &macro_arg.0); let ExpandResult { value: mut tt, err } = expander.expand(db, id, &macro_arg.0);
// Set a hard limit for the expanded tt // Set a hard limit for the expanded tt
let count = tt.count(); let count = tt.count();
// XXX: Make ExpandResult a real error and use .map_err instead?
if TOKEN_LIMIT.check(count).is_err() { if TOKEN_LIMIT.check(count).is_err() {
return ExpandResult::str_err(format!( return ExpandResult::str_err(format!(
"macro invocation exceeds token limit: produced {} tokens, limit is {}", "macro invocation exceeds token limit: produced {} tokens, limit is {}",
@ -442,7 +442,7 @@ fn macro_expand(db: &dyn AstDatabase, id: MacroCallId) -> ExpandResult<Option<Ar
)); ));
} }
fixup::reverse_fixups(&mut tt, &macro_arg.1); fixup::reverse_fixups(&mut tt, &macro_arg.1, &macro_arg.2);
ExpandResult { value: Some(Arc::new(tt)), err } ExpandResult { value: Some(Arc::new(tt)), err }
} }

View file

@ -10,21 +10,39 @@ use tt::Subtree;
pub struct SyntaxFixups { pub struct SyntaxFixups {
pub append: FxHashMap<SyntaxNode, Vec<SyntheticToken>>, pub append: FxHashMap<SyntaxNode, Vec<SyntheticToken>>,
pub replace: FxHashMap<SyntaxNode, Vec<SyntheticToken>>, pub replace: FxHashMap<SyntaxNode, Vec<SyntheticToken>>,
pub map: SyntaxFixupMap,
} }
#[derive(Debug, PartialEq, Eq)]
pub struct SyntaxFixupMap {
original: Vec<(Subtree, TokenMap)>,
}
const EMPTY_ID: SyntheticTokenId = SyntheticTokenId(!0);
pub fn fixup_syntax(node: &SyntaxNode) -> SyntaxFixups { pub fn fixup_syntax(node: &SyntaxNode) -> SyntaxFixups {
let mut append = FxHashMap::default(); let mut append = FxHashMap::default();
let mut replace = FxHashMap::default(); let mut replace = FxHashMap::default();
let mut preorder = node.preorder(); let mut preorder = node.preorder();
let empty_id = SyntheticTokenId(0); let mut original = Vec::new();
while let Some(event) = preorder.next() { while let Some(event) = preorder.next() {
let node = match event { let node = match event {
syntax::WalkEvent::Enter(node) => node, syntax::WalkEvent::Enter(node) => node,
syntax::WalkEvent::Leave(_) => continue, syntax::WalkEvent::Leave(_) => continue,
}; };
if node.kind() == SyntaxKind::ERROR { if can_handle_error(&node) && has_error_to_handle(&node) {
// TODO this might not be helpful // the node contains an error node, we have to completely replace it by something valid
replace.insert(node, Vec::new()); let original_tree = mbe::syntax_node_to_token_tree(&node);
// TODO handle token ids / token map
let idx = original.len() as u32;
original.push(original_tree);
let replacement = SyntheticToken {
kind: SyntaxKind::IDENT,
text: "__ra_fixup".into(),
range: node.text_range(),
id: SyntheticTokenId(idx),
};
replace.insert(node.clone(), vec![replacement]);
preorder.skip_subtree(); preorder.skip_subtree();
continue; continue;
} }
@ -39,7 +57,7 @@ pub fn fixup_syntax(node: &SyntaxNode) -> SyntaxFixups {
kind: SyntaxKind::IDENT, kind: SyntaxKind::IDENT,
text: "__ra_fixup".into(), text: "__ra_fixup".into(),
range: end_range, range: end_range,
id: empty_id, id: EMPTY_ID,
}, },
]); ]);
} }
@ -51,7 +69,7 @@ pub fn fixup_syntax(node: &SyntaxNode) -> SyntaxFixups {
kind: SyntaxKind::SEMICOLON, kind: SyntaxKind::SEMICOLON,
text: ";".into(), text: ";".into(),
range: end_range, range: end_range,
id: empty_id, id: EMPTY_ID,
}, },
]); ]);
} }
@ -60,18 +78,37 @@ pub fn fixup_syntax(node: &SyntaxNode) -> SyntaxFixups {
} }
} }
} }
SyntaxFixups { append, replace } SyntaxFixups { append, replace, map: SyntaxFixupMap { original } }
} }
pub fn reverse_fixups(tt: &mut Subtree, token_map: &TokenMap) { fn has_error(node: &SyntaxNode) -> bool {
eprintln!("token_map: {:?}", token_map); node.children().any(|c| c.kind() == SyntaxKind::ERROR)
}
fn can_handle_error(node: &SyntaxNode) -> bool {
ast::Expr::can_cast(node.kind())
}
fn has_error_to_handle(node: &SyntaxNode) -> bool {
has_error(node) || node.children().any(|c| !can_handle_error(&c) && has_error_to_handle(&c))
}
pub fn reverse_fixups(tt: &mut Subtree, token_map: &TokenMap, fixup_map: &SyntaxFixupMap) {
tt.token_trees.retain(|tt| match tt { tt.token_trees.retain(|tt| match tt {
tt::TokenTree::Leaf(leaf) => token_map.synthetic_token_id(leaf.id()).is_none(), tt::TokenTree::Leaf(leaf) => {
token_map.synthetic_token_id(leaf.id()).is_none()
|| token_map.synthetic_token_id(leaf.id()) != Some(EMPTY_ID)
}
_ => true, _ => true,
}); });
tt.token_trees.iter_mut().for_each(|tt| match tt { tt.token_trees.iter_mut().for_each(|tt| match tt {
tt::TokenTree::Subtree(tt) => reverse_fixups(tt, token_map), tt::TokenTree::Subtree(tt) => reverse_fixups(tt, token_map, fixup_map),
_ => {} tt::TokenTree::Leaf(leaf) => {
if let Some(id) = token_map.synthetic_token_id(leaf.id()) {
let (original, _original_tmap) = &fixup_map.original[id.0 as usize];
*tt = tt::TokenTree::Subtree(original.clone());
}
}
}); });
} }
@ -84,6 +121,7 @@ mod tests {
#[track_caller] #[track_caller]
fn check(ra_fixture: &str, mut expect: Expect) { fn check(ra_fixture: &str, mut expect: Expect) {
let parsed = syntax::SourceFile::parse(ra_fixture); let parsed = syntax::SourceFile::parse(ra_fixture);
eprintln!("parse: {:#?}", parsed.syntax_node());
let fixups = super::fixup_syntax(&parsed.syntax_node()); let fixups = super::fixup_syntax(&parsed.syntax_node());
let (mut tt, tmap) = mbe::syntax_node_to_token_tree_censored( let (mut tt, tmap) = mbe::syntax_node_to_token_tree_censored(
&parsed.syntax_node(), &parsed.syntax_node(),
@ -106,7 +144,7 @@ mod tests {
parse.syntax_node() parse.syntax_node()
); );
reverse_fixups(&mut tt, &tmap); reverse_fixups(&mut tt, &tmap, &fixups.map);
// the fixed-up + reversed version should be equivalent to the original input // the fixed-up + reversed version should be equivalent to the original input
// (but token IDs don't matter) // (but token IDs don't matter)
@ -169,6 +207,20 @@ fn foo() {
"#, "#,
expect![[r#" expect![[r#"
fn foo () {a . b ; bar () ;} fn foo () {a . b ; bar () ;}
"#]],
)
}
#[test]
fn extraneous_comma() {
check(
r#"
fn foo() {
bar(,);
}
"#,
expect![[r#"
fn foo () {__ra_fixup ;}
"#]], "#]],
) )
} }

View file

@ -15,6 +15,7 @@ use syntax::{
use crate::{ use crate::{
db::{self, AstDatabase}, db::{self, AstDatabase},
fixup,
name::{AsName, Name}, name::{AsName, Name},
HirFileId, HirFileIdRepr, InFile, MacroCallKind, MacroCallLoc, MacroDefKind, MacroFile, HirFileId, HirFileIdRepr, InFile, MacroCallKind, MacroCallLoc, MacroDefKind, MacroFile,
}; };
@ -127,7 +128,7 @@ struct HygieneInfo {
attr_input_or_mac_def_start: Option<InFile<TextSize>>, attr_input_or_mac_def_start: Option<InFile<TextSize>>,
macro_def: Arc<TokenExpander>, macro_def: Arc<TokenExpander>,
macro_arg: Arc<(tt::Subtree, mbe::TokenMap)>, macro_arg: Arc<(tt::Subtree, mbe::TokenMap, fixup::SyntaxFixupMap)>,
macro_arg_shift: mbe::Shift, macro_arg_shift: mbe::Shift,
exp_map: Arc<mbe::TokenMap>, exp_map: Arc<mbe::TokenMap>,
} }

View file

@ -427,7 +427,7 @@ pub struct ExpansionInfo {
attr_input_or_mac_def: Option<InFile<ast::TokenTree>>, attr_input_or_mac_def: Option<InFile<ast::TokenTree>>,
macro_def: Arc<TokenExpander>, macro_def: Arc<TokenExpander>,
macro_arg: Arc<(tt::Subtree, mbe::TokenMap)>, macro_arg: Arc<(tt::Subtree, mbe::TokenMap, fixup::SyntaxFixupMap)>,
/// A shift built from `macro_arg`'s subtree, relevant for attributes as the item is the macro arg /// A shift built from `macro_arg`'s subtree, relevant for attributes as the item is the macro arg
/// and as such we need to shift tokens if they are part of an attributes input instead of their item. /// and as such we need to shift tokens if they are part of an attributes input instead of their item.
macro_arg_shift: mbe::Shift, macro_arg_shift: mbe::Shift,

View file

@ -30,8 +30,8 @@ pub fn syntax_node_to_token_tree_censored(
let mut c = Convertor::new(node, global_offset, replace, append); let mut c = Convertor::new(node, global_offset, replace, append);
let subtree = convert_tokens(&mut c); let subtree = convert_tokens(&mut c);
c.id_alloc.map.shrink_to_fit(); c.id_alloc.map.shrink_to_fit();
always!(c.replace.is_empty()); always!(c.replace.is_empty(), "replace: {:?}", c.replace);
always!(c.append.is_empty()); always!(c.append.is_empty(), "append: {:?}", c.append);
(subtree, c.id_alloc.map) (subtree, c.id_alloc.map)
} }
@ -539,7 +539,6 @@ impl Convertor {
WalkEvent::Enter(ele) => ele, WalkEvent::Enter(ele) => ele,
WalkEvent::Leave(SyntaxElement::Node(node)) => { WalkEvent::Leave(SyntaxElement::Node(node)) => {
if let Some(mut v) = append.remove(&node) { if let Some(mut v) = append.remove(&node) {
eprintln!("after {:?}, appending {:?}", node, v);
if !v.is_empty() { if !v.is_empty() {
v.reverse(); v.reverse();
return (None, v); return (None, v);
@ -554,7 +553,6 @@ impl Convertor {
SyntaxElement::Node(node) => { SyntaxElement::Node(node) => {
if let Some(mut v) = replace.remove(&node) { if let Some(mut v) = replace.remove(&node) {
preorder.skip_subtree(); preorder.skip_subtree();
eprintln!("replacing {:?} by {:?}", node, v);
if !v.is_empty() { if !v.is_empty() {
v.reverse(); v.reverse();
return (None, v); return (None, v);
@ -640,8 +638,8 @@ impl TokenConvertor for Convertor {
self.current = new_current; self.current = new_current;
self.current_synthetic = new_synth; self.current_synthetic = new_synth;
} }
// TODO fix range? let range = synth_token.range;
return Some((SynToken::Synthetic(synth_token), self.range)); return Some((SynToken::Synthetic(synth_token), range));
} }
let curr = self.current.clone()?; let curr = self.current.clone()?;
@ -675,7 +673,6 @@ impl TokenConvertor for Convertor {
} }
if let Some(synth_token) = self.current_synthetic.last() { if let Some(synth_token) = self.current_synthetic.last() {
// TODO fix range?
return Some(SynToken::Synthetic(synth_token.clone())); return Some(SynToken::Synthetic(synth_token.clone()));
} }

View file

@ -74,6 +74,7 @@ impl TokenMap {
pub(crate) fn shrink_to_fit(&mut self) { pub(crate) fn shrink_to_fit(&mut self) {
self.entries.shrink_to_fit(); self.entries.shrink_to_fit();
self.synthetic_entries.shrink_to_fit();
} }
pub(crate) fn insert(&mut self, token_id: tt::TokenId, relative_range: TextRange) { pub(crate) fn insert(&mut self, token_id: tt::TokenId, relative_range: TextRange) {