mirror of
https://github.com/rust-lang/rust-analyzer
synced 2024-12-26 13:03:31 +00:00
Test utils
This commit is contained in:
parent
18f9e50b2d
commit
9e4052cc2e
9 changed files with 131 additions and 82 deletions
|
@ -12,4 +12,4 @@ file = "1.1.1"
|
||||||
ron = "0.1.5"
|
ron = "0.1.5"
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
difference = "1.0.0"
|
testutils = { path = "./tests/testutils" }
|
|
@ -8,7 +8,12 @@ pub fn file(p: &mut Parser) {
|
||||||
node(p, FILE, |p| {
|
node(p, FILE, |p| {
|
||||||
shebang(p);
|
shebang(p);
|
||||||
inner_attributes(p);
|
inner_attributes(p);
|
||||||
many(p, |p| skip_to_first(p, item_first, item));
|
many(p, |p| {
|
||||||
|
skip_to_first(
|
||||||
|
p, item_first, item,
|
||||||
|
"expected item",
|
||||||
|
)
|
||||||
|
});
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -84,19 +89,34 @@ fn comma_list<F: Fn(&mut Parser) -> bool>(p: &mut Parser, f: F) {
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
fn skip_to_first<C, F>(p: &mut Parser, cond: C, f: F) -> bool
|
fn skip_to_first<C, F>(p: &mut Parser, cond: C, f: F, message: &str) -> bool
|
||||||
where
|
where
|
||||||
C: Fn(&Parser) -> bool,
|
C: Fn(&Parser) -> bool,
|
||||||
F: FnOnce(&mut Parser),
|
F: FnOnce(&mut Parser),
|
||||||
{
|
{
|
||||||
|
let mut skipped = false;
|
||||||
loop {
|
loop {
|
||||||
if cond(p) {
|
if cond(p) {
|
||||||
|
if skipped {
|
||||||
|
p.finish();
|
||||||
|
}
|
||||||
f(p);
|
f(p);
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
if p.bump().is_none() {
|
if p.is_eof() {
|
||||||
|
if skipped {
|
||||||
|
p.finish();
|
||||||
|
}
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
if !skipped {
|
||||||
|
p.start(ERROR);
|
||||||
|
p.error()
|
||||||
|
.message(message)
|
||||||
|
.emit();
|
||||||
|
}
|
||||||
|
p.bump().unwrap();
|
||||||
|
skipped = true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -73,7 +73,9 @@ impl FileBuilder {
|
||||||
pub fn finish(self) -> File {
|
pub fn finish(self) -> File {
|
||||||
assert!(
|
assert!(
|
||||||
self.in_progress.is_empty(),
|
self.in_progress.is_empty(),
|
||||||
"some nodes in FileBuilder are unfinished"
|
"some nodes in FileBuilder are unfinished: {:?}",
|
||||||
|
self.in_progress.iter().map(|&(idx, _)| self.nodes[idx].kind)
|
||||||
|
.collect::<Vec<_>>()
|
||||||
);
|
);
|
||||||
assert!(
|
assert!(
|
||||||
self.pos == (self.text.len() as u32).into(),
|
self.pos == (self.text.len() as u32).into(),
|
||||||
|
@ -122,11 +124,6 @@ impl FileBuilder {
|
||||||
let idx = self.current_id();
|
let idx = self.current_id();
|
||||||
&mut self.nodes[idx]
|
&mut self.nodes[idx]
|
||||||
}
|
}
|
||||||
|
|
||||||
fn current_sibling(&mut self) -> Option<&mut NodeData> {
|
|
||||||
let idx = self.in_progress.last().unwrap().1?;
|
|
||||||
Some(&mut self.nodes[idx])
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn fill<T>(slot: &mut Option<T>, value: T) {
|
fn fill<T>(slot: &mut Option<T>, value: T) {
|
||||||
|
|
3
tests/data/parser/err/0001_item_recovery_in_file.rs
Normal file
3
tests/data/parser/err/0001_item_recovery_in_file.rs
Normal file
|
@ -0,0 +1,3 @@
|
||||||
|
if match
|
||||||
|
|
||||||
|
struct S {}
|
14
tests/data/parser/err/0001_item_recovery_in_file.txt
Normal file
14
tests/data/parser/err/0001_item_recovery_in_file.txt
Normal file
|
@ -0,0 +1,14 @@
|
||||||
|
FILE@[0; 21)
|
||||||
|
ERROR@[0; 10)
|
||||||
|
err: `expected item`
|
||||||
|
IDENT@[0; 2)
|
||||||
|
WHITESPACE@[2; 3)
|
||||||
|
IDENT@[3; 8)
|
||||||
|
WHITESPACE@[8; 10)
|
||||||
|
STRUCT_ITEM@[10; 21)
|
||||||
|
STRUCT_KW@[10; 16)
|
||||||
|
WHITESPACE@[16; 17)
|
||||||
|
IDENT@[17; 18)
|
||||||
|
WHITESPACE@[18; 19)
|
||||||
|
L_CURLY@[19; 20)
|
||||||
|
R_CURLY@[20; 21)
|
|
@ -1,56 +1,31 @@
|
||||||
extern crate file;
|
extern crate file;
|
||||||
#[macro_use(assert_diff)]
|
|
||||||
extern crate difference;
|
|
||||||
extern crate libsyntax2;
|
extern crate libsyntax2;
|
||||||
|
extern crate testutils;
|
||||||
|
|
||||||
use std::path::{PathBuf, Path};
|
use std::path::{Path};
|
||||||
use std::fs::read_dir;
|
|
||||||
use std::fmt::Write;
|
use std::fmt::Write;
|
||||||
|
|
||||||
use libsyntax2::{Token, tokenize};
|
use libsyntax2::{Token, tokenize};
|
||||||
|
use testutils::{assert_equal_text, collect_tests};
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn lexer_tests() {
|
fn lexer_tests() {
|
||||||
for test_case in lexer_test_cases() {
|
for test_case in collect_tests(&["lexer"]) {
|
||||||
lexer_test_case(&test_case);
|
lexer_test_case(&test_case);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn lexer_test_dir() -> PathBuf {
|
|
||||||
let dir = env!("CARGO_MANIFEST_DIR");
|
|
||||||
PathBuf::from(dir).join("tests/data/lexer")
|
|
||||||
}
|
|
||||||
|
|
||||||
fn lexer_test_cases() -> Vec<PathBuf> {
|
|
||||||
let mut acc = Vec::new();
|
|
||||||
let dir = lexer_test_dir();
|
|
||||||
for file in read_dir(&dir).unwrap() {
|
|
||||||
let file = file.unwrap();
|
|
||||||
let path = file.path();
|
|
||||||
if path.extension().unwrap_or_default() == "rs" {
|
|
||||||
acc.push(path);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
acc.sort();
|
|
||||||
acc
|
|
||||||
}
|
|
||||||
|
|
||||||
fn lexer_test_case(path: &Path) {
|
fn lexer_test_case(path: &Path) {
|
||||||
let actual = {
|
let actual = {
|
||||||
let text = file::get_text(path).unwrap();
|
let text = file::get_text(path).unwrap();
|
||||||
let tokens = tokenize(&text);
|
let tokens = tokenize(&text);
|
||||||
dump_tokens(&tokens, &text)
|
dump_tokens(&tokens, &text)
|
||||||
};
|
};
|
||||||
let expected = file::get_text(&path.with_extension("txt")).unwrap();
|
let path = path.with_extension("txt");
|
||||||
|
let expected = file::get_text(&path).unwrap();
|
||||||
let expected = expected.as_str();
|
let expected = expected.as_str();
|
||||||
let actual = actual.as_str();
|
let actual = actual.as_str();
|
||||||
if expected == actual {
|
assert_equal_text(expected, actual, &path)
|
||||||
return
|
|
||||||
}
|
|
||||||
if expected.trim() == actual.trim() {
|
|
||||||
panic!("Whitespace difference!")
|
|
||||||
}
|
|
||||||
assert_diff!(expected, actual, "\n", 0)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn dump_tokens(tokens: &[Token], text: &str) -> String {
|
fn dump_tokens(tokens: &[Token], text: &str) -> String {
|
||||||
|
|
|
@ -1,46 +1,20 @@
|
||||||
extern crate file;
|
extern crate file;
|
||||||
#[macro_use(assert_diff)]
|
|
||||||
extern crate difference;
|
|
||||||
extern crate libsyntax2;
|
extern crate libsyntax2;
|
||||||
|
extern crate testutils;
|
||||||
|
|
||||||
use std::path::{PathBuf, Path};
|
use std::path::{Path};
|
||||||
use std::fs::read_dir;
|
|
||||||
use std::fmt::Write;
|
use std::fmt::Write;
|
||||||
|
|
||||||
use libsyntax2::{tokenize, parse, Node, File};
|
use libsyntax2::{tokenize, parse, Node, File};
|
||||||
|
use testutils::{collect_tests, assert_equal_text};
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn parser_tests() {
|
fn parser_tests() {
|
||||||
for test_case in parser_test_cases() {
|
for test_case in collect_tests(&["parser/ok", "parser/err"]) {
|
||||||
parser_test_case(&test_case);
|
parser_test_case(&test_case);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn parser_test_dir() -> PathBuf {
|
|
||||||
let dir = env!("CARGO_MANIFEST_DIR");
|
|
||||||
PathBuf::from(dir).join("tests/data/parser")
|
|
||||||
}
|
|
||||||
|
|
||||||
fn test_from_dir(dir: &Path) -> Vec<PathBuf> {
|
|
||||||
let mut acc = Vec::new();
|
|
||||||
for file in read_dir(&dir).unwrap() {
|
|
||||||
let file = file.unwrap();
|
|
||||||
let path = file.path();
|
|
||||||
if path.extension().unwrap_or_default() == "rs" {
|
|
||||||
acc.push(path);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
acc.sort();
|
|
||||||
acc
|
|
||||||
}
|
|
||||||
|
|
||||||
fn parser_test_cases() -> Vec<PathBuf> {
|
|
||||||
let mut acc = Vec::new();
|
|
||||||
acc.extend(test_from_dir(&parser_test_dir().join("ok")));
|
|
||||||
acc.extend(test_from_dir(&parser_test_dir().join("err")));
|
|
||||||
acc
|
|
||||||
}
|
|
||||||
|
|
||||||
fn parser_test_case(path: &Path) {
|
fn parser_test_case(path: &Path) {
|
||||||
let actual = {
|
let actual = {
|
||||||
let text = file::get_text(path).unwrap();
|
let text = file::get_text(path).unwrap();
|
||||||
|
@ -48,19 +22,13 @@ fn parser_test_case(path: &Path) {
|
||||||
let file = parse(text, &tokens);
|
let file = parse(text, &tokens);
|
||||||
dump_tree(&file)
|
dump_tree(&file)
|
||||||
};
|
};
|
||||||
let expected = path.with_extension("txt");
|
let expected_path = path.with_extension("txt");
|
||||||
let expected = file::get_text(&expected).expect(
|
let expected = file::get_text(&expected_path).expect(
|
||||||
&format!("Can't read {}", expected.display())
|
&format!("Can't read {}", expected_path.display())
|
||||||
);
|
);
|
||||||
let expected = expected.as_str();
|
let expected = expected.as_str();
|
||||||
let actual = actual.as_str();
|
let actual = actual.as_str();
|
||||||
if expected == actual {
|
assert_equal_text(expected, actual, &expected_path);
|
||||||
return
|
|
||||||
}
|
|
||||||
if expected.trim() == actual.trim() {
|
|
||||||
panic!("Whitespace difference! {}", path.display())
|
|
||||||
}
|
|
||||||
assert_diff!(expected, actual, "\n", 0)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn dump_tree(file: &File) -> String {
|
fn dump_tree(file: &File) -> String {
|
||||||
|
|
8
tests/testutils/Cargo.toml
Normal file
8
tests/testutils/Cargo.toml
Normal file
|
@ -0,0 +1,8 @@
|
||||||
|
[package]
|
||||||
|
name = "testutils"
|
||||||
|
version = "0.1.0"
|
||||||
|
authors = ["Aleksey Kladov <aleksey.kladov@gmail.com>"]
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
file = "1.0"
|
||||||
|
difference = "1.0.0"
|
64
tests/testutils/src/lib.rs
Normal file
64
tests/testutils/src/lib.rs
Normal file
|
@ -0,0 +1,64 @@
|
||||||
|
extern crate difference;
|
||||||
|
extern crate file;
|
||||||
|
|
||||||
|
use std::path::{PathBuf, Path};
|
||||||
|
use std::fs::read_dir;
|
||||||
|
|
||||||
|
use difference::Changeset;
|
||||||
|
|
||||||
|
pub fn assert_equal_text(
|
||||||
|
expected: &str,
|
||||||
|
actual: &str,
|
||||||
|
path: &Path
|
||||||
|
) {
|
||||||
|
if expected != actual {
|
||||||
|
print_difference(expected, actual, path)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn collect_tests(paths: &[&str]) -> Vec<PathBuf> {
|
||||||
|
paths.iter().flat_map(|path| {
|
||||||
|
let path = test_data_dir().join(path);
|
||||||
|
test_from_dir(&path).into_iter()
|
||||||
|
}).collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn test_from_dir(dir: &Path) -> Vec<PathBuf> {
|
||||||
|
let mut acc = Vec::new();
|
||||||
|
for file in read_dir(&dir).unwrap() {
|
||||||
|
let file = file.unwrap();
|
||||||
|
let path = file.path();
|
||||||
|
if path.extension().unwrap_or_default() == "rs" {
|
||||||
|
acc.push(path);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
acc.sort();
|
||||||
|
acc
|
||||||
|
}
|
||||||
|
|
||||||
|
fn print_difference(expected: &str, actual: &str, path: &Path) {
|
||||||
|
let dir = project_dir();
|
||||||
|
let path = path.strip_prefix(&dir).unwrap_or_else(|_| path);
|
||||||
|
println!("\nfile: {}", path.display());
|
||||||
|
if expected.trim() == actual.trim() {
|
||||||
|
println!("whitespace difference");
|
||||||
|
println!("rewriting the file");
|
||||||
|
file::put_text(path, actual).unwrap();
|
||||||
|
} else {
|
||||||
|
let changeset = Changeset::new(actual, expected, "\n");
|
||||||
|
println!("{}", changeset);
|
||||||
|
}
|
||||||
|
panic!("Comparison failed")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn project_dir() -> PathBuf {
|
||||||
|
let dir = env!("CARGO_MANIFEST_DIR");
|
||||||
|
PathBuf::from(dir)
|
||||||
|
.parent().unwrap()
|
||||||
|
.parent().unwrap()
|
||||||
|
.to_owned()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn test_data_dir() -> PathBuf {
|
||||||
|
project_dir().join("tests/data")
|
||||||
|
}
|
Loading…
Reference in a new issue