mirror of
https://github.com/rust-lang/rust-analyzer
synced 2024-12-25 20:43:21 +00:00
G: shebang
This commit is contained in:
parent
9e4052cc2e
commit
7c6f0f9128
8 changed files with 64 additions and 47 deletions
|
@ -6,7 +6,7 @@ use syntax_kinds::*;
|
|||
|
||||
pub fn file(p: &mut Parser) {
|
||||
node(p, FILE, |p| {
|
||||
shebang(p);
|
||||
p.optional(SHEBANG);
|
||||
inner_attributes(p);
|
||||
many(p, |p| {
|
||||
skip_to_first(
|
||||
|
@ -17,11 +17,6 @@ pub fn file(p: &mut Parser) {
|
|||
})
|
||||
}
|
||||
|
||||
|
||||
fn shebang(_: &mut Parser) {
|
||||
//TODO
|
||||
}
|
||||
|
||||
fn inner_attributes(_: &mut Parser) {
|
||||
//TODO
|
||||
}
|
||||
|
@ -136,4 +131,10 @@ impl<'p> Parser<'p> {
|
|||
false
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn optional(&mut self, kind: SyntaxKind) {
|
||||
if self.current_is(kind) {
|
||||
self.bump();
|
||||
}
|
||||
}
|
||||
}
|
2
tests/data/parser/err/0002_duplicate_shebang.rs
Normal file
2
tests/data/parser/err/0002_duplicate_shebang.rs
Normal file
|
@ -0,0 +1,2 @@
|
|||
#!/use/bin/env rusti
|
||||
#!/use/bin/env rusti
|
7
tests/data/parser/err/0002_duplicate_shebang.txt
Normal file
7
tests/data/parser/err/0002_duplicate_shebang.txt
Normal file
|
@ -0,0 +1,7 @@
|
|||
FILE@[0; 42)
|
||||
SHEBANG@[0; 20)
|
||||
ERROR@[20; 42)
|
||||
err: `expected item`
|
||||
WHITESPACE@[20; 21)
|
||||
SHEBANG@[21; 41)
|
||||
WHITESPACE@[41; 42)
|
1
tests/data/parser/ok/0004_file_shebang.rs
Normal file
1
tests/data/parser/ok/0004_file_shebang.rs
Normal file
|
@ -0,0 +1 @@
|
|||
#!/use/bin/env rusti
|
2
tests/data/parser/ok/0004_file_shebang.txt
Normal file
2
tests/data/parser/ok/0004_file_shebang.txt
Normal file
|
@ -0,0 +1,2 @@
|
|||
FILE@[0; 20)
|
||||
SHEBANG@[0; 20)
|
|
@ -2,30 +2,20 @@ extern crate file;
|
|||
extern crate libsyntax2;
|
||||
extern crate testutils;
|
||||
|
||||
use std::path::{Path};
|
||||
use std::fmt::Write;
|
||||
|
||||
use libsyntax2::{Token, tokenize};
|
||||
use testutils::{assert_equal_text, collect_tests};
|
||||
use testutils::dir_tests;
|
||||
|
||||
#[test]
|
||||
fn lexer_tests() {
|
||||
for test_case in collect_tests(&["lexer"]) {
|
||||
lexer_test_case(&test_case);
|
||||
}
|
||||
}
|
||||
|
||||
fn lexer_test_case(path: &Path) {
|
||||
let actual = {
|
||||
let text = file::get_text(path).unwrap();
|
||||
let tokens = tokenize(&text);
|
||||
dump_tokens(&tokens, &text)
|
||||
};
|
||||
let path = path.with_extension("txt");
|
||||
let expected = file::get_text(&path).unwrap();
|
||||
let expected = expected.as_str();
|
||||
let actual = actual.as_str();
|
||||
assert_equal_text(expected, actual, &path)
|
||||
dir_tests(
|
||||
&["lexer"],
|
||||
|text| {
|
||||
let tokens = tokenize(text);
|
||||
dump_tokens(&tokens, text)
|
||||
}
|
||||
)
|
||||
}
|
||||
|
||||
fn dump_tokens(tokens: &[Token], text: &str) -> String {
|
||||
|
|
|
@ -2,33 +2,21 @@ extern crate file;
|
|||
extern crate libsyntax2;
|
||||
extern crate testutils;
|
||||
|
||||
use std::path::{Path};
|
||||
use std::fmt::Write;
|
||||
|
||||
use libsyntax2::{tokenize, parse, Node, File};
|
||||
use testutils::{collect_tests, assert_equal_text};
|
||||
use testutils::dir_tests;
|
||||
|
||||
#[test]
|
||||
fn parser_tests() {
|
||||
for test_case in collect_tests(&["parser/ok", "parser/err"]) {
|
||||
parser_test_case(&test_case);
|
||||
}
|
||||
}
|
||||
|
||||
fn parser_test_case(path: &Path) {
|
||||
let actual = {
|
||||
let text = file::get_text(path).unwrap();
|
||||
let tokens = tokenize(&text);
|
||||
let file = parse(text, &tokens);
|
||||
dump_tree(&file)
|
||||
};
|
||||
let expected_path = path.with_extension("txt");
|
||||
let expected = file::get_text(&expected_path).expect(
|
||||
&format!("Can't read {}", expected_path.display())
|
||||
);
|
||||
let expected = expected.as_str();
|
||||
let actual = actual.as_str();
|
||||
assert_equal_text(expected, actual, &expected_path);
|
||||
dir_tests(
|
||||
&["parser/ok", "parser/err"],
|
||||
|text| {
|
||||
let tokens = tokenize(text);
|
||||
let file = parse(text.to_string(), &tokens);
|
||||
dump_tree(&file)
|
||||
}
|
||||
)
|
||||
}
|
||||
|
||||
fn dump_tree(file: &File) -> String {
|
||||
|
|
|
@ -6,7 +6,33 @@ use std::fs::read_dir;
|
|||
|
||||
use difference::Changeset;
|
||||
|
||||
pub fn assert_equal_text(
|
||||
pub fn dir_tests<F>(
|
||||
paths: &[&str],
|
||||
f: F
|
||||
)
|
||||
where
|
||||
F: Fn(&str) -> String
|
||||
{
|
||||
for path in collect_tests(paths) {
|
||||
let actual = {
|
||||
let text = file::get_text(&path).unwrap();
|
||||
f(&text)
|
||||
};
|
||||
let path = path.with_extension("txt");
|
||||
if !path.exists() {
|
||||
println!("\nfile: {}", path.display());
|
||||
println!("No .txt file with expected result, creating...");
|
||||
file::put_text(&path, actual).unwrap();
|
||||
panic!("No expected result")
|
||||
}
|
||||
let expected = file::get_text(&path).unwrap();
|
||||
let expected = expected.as_str();
|
||||
let actual = actual.as_str();
|
||||
assert_equal_text(expected, actual, &path);
|
||||
}
|
||||
}
|
||||
|
||||
fn assert_equal_text(
|
||||
expected: &str,
|
||||
actual: &str,
|
||||
path: &Path
|
||||
|
@ -16,7 +42,7 @@ pub fn assert_equal_text(
|
|||
}
|
||||
}
|
||||
|
||||
pub fn collect_tests(paths: &[&str]) -> Vec<PathBuf> {
|
||||
fn collect_tests(paths: &[&str]) -> Vec<PathBuf> {
|
||||
paths.iter().flat_map(|path| {
|
||||
let path = test_data_dir().join(path);
|
||||
test_from_dir(&path).into_iter()
|
||||
|
|
Loading…
Reference in a new issue