mirror of
https://github.com/rust-lang/rust-analyzer
synced 2024-12-25 20:43:21 +00:00
Cleanup tools
This commit is contained in:
parent
9a4957d143
commit
6983091d6d
9 changed files with 290 additions and 300 deletions
|
@ -1,4 +1,4 @@
|
||||||
[alias]
|
[alias]
|
||||||
parse = "run --package tools --bin parse"
|
parse = "run --package tools --bin parse"
|
||||||
gen = "run --package tools --bin gen"
|
gen-kinds = "run --package tools -- gen-kinds"
|
||||||
collect-tests = "run --package tools --bin collect-tests --"
|
gen-tests = "run --package tools -- gen-tests"
|
||||||
|
|
|
@ -8,6 +8,8 @@ matrix:
|
||||||
script:
|
script:
|
||||||
- cargo fmt --all -- --write-mode=diff
|
- cargo fmt --all -- --write-mode=diff
|
||||||
- cargo test
|
- cargo test
|
||||||
|
- cargo gen-kinds --verify
|
||||||
|
- cargo gen-tests --verify
|
||||||
- rust: nightly
|
- rust: nightly
|
||||||
before_script:
|
before_script:
|
||||||
- rustup component add clippy-preview
|
- rustup component add clippy-preview
|
||||||
|
|
|
@ -2,8 +2,9 @@ extern crate libsyntax2;
|
||||||
|
|
||||||
use std::io::Read;
|
use std::io::Read;
|
||||||
|
|
||||||
use libsyntax2::{parse};
|
use libsyntax2::{
|
||||||
use libsyntax2::utils::dump_tree_green;
|
parse, utils::dump_tree_green
|
||||||
|
};
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
let text = read_input();
|
let text = read_input();
|
|
@ -1,6 +1,5 @@
|
||||||
#![allow(bad_style, missing_docs, unreachable_pub)]
|
#![allow(bad_style, missing_docs, unreachable_pub)]
|
||||||
#![cfg_attr(rustfmt, rustfmt_skip)]
|
#![cfg_attr(rustfmt, rustfmt_skip)]
|
||||||
//! Generated from grammar.ron
|
|
||||||
use super::SyntaxInfo;
|
use super::SyntaxInfo;
|
||||||
|
|
||||||
/// The kind of syntax node, e.g. `IDENT`, `USE_KW`, or `STRUCT_DEF`.
|
/// The kind of syntax node, e.g. `IDENT`, `USE_KW`, or `STRUCT_DEF`.
|
||||||
|
@ -138,7 +137,6 @@ pub enum SyntaxKind {
|
||||||
VALUE_PARAMETER,
|
VALUE_PARAMETER,
|
||||||
BLOCK,
|
BLOCK,
|
||||||
LET_STMT,
|
LET_STMT,
|
||||||
|
|
||||||
// Technical SyntaxKinds: they appear temporally during parsing,
|
// Technical SyntaxKinds: they appear temporally during parsing,
|
||||||
// but never end up in the final tree
|
// but never end up in the final tree
|
||||||
#[doc(hidden)]
|
#[doc(hidden)]
|
||||||
|
@ -146,7 +144,7 @@ pub enum SyntaxKind {
|
||||||
#[doc(hidden)]
|
#[doc(hidden)]
|
||||||
EOF,
|
EOF,
|
||||||
}
|
}
|
||||||
pub(crate) use self::SyntaxKind::*;
|
use self::SyntaxKind::*;
|
||||||
|
|
||||||
impl SyntaxKind {
|
impl SyntaxKind {
|
||||||
pub(crate) fn info(self) -> &'static SyntaxInfo {
|
pub(crate) fn info(self) -> &'static SyntaxInfo {
|
||||||
|
@ -289,38 +287,39 @@ impl SyntaxKind {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pub(crate) fn from_keyword(ident: &str) -> Option<SyntaxKind> {
|
pub(crate) fn from_keyword(ident: &str) -> Option<SyntaxKind> {
|
||||||
match ident {
|
let kw = match ident {
|
||||||
"use" => Some(USE_KW),
|
"use" => USE_KW,
|
||||||
"fn" => Some(FN_KW),
|
"fn" => FN_KW,
|
||||||
"struct" => Some(STRUCT_KW),
|
"struct" => STRUCT_KW,
|
||||||
"enum" => Some(ENUM_KW),
|
"enum" => ENUM_KW,
|
||||||
"trait" => Some(TRAIT_KW),
|
"trait" => TRAIT_KW,
|
||||||
"impl" => Some(IMPL_KW),
|
"impl" => IMPL_KW,
|
||||||
"true" => Some(TRUE_KW),
|
"true" => TRUE_KW,
|
||||||
"false" => Some(FALSE_KW),
|
"false" => FALSE_KW,
|
||||||
"as" => Some(AS_KW),
|
"as" => AS_KW,
|
||||||
"extern" => Some(EXTERN_KW),
|
"extern" => EXTERN_KW,
|
||||||
"crate" => Some(CRATE_KW),
|
"crate" => CRATE_KW,
|
||||||
"mod" => Some(MOD_KW),
|
"mod" => MOD_KW,
|
||||||
"pub" => Some(PUB_KW),
|
"pub" => PUB_KW,
|
||||||
"self" => Some(SELF_KW),
|
"self" => SELF_KW,
|
||||||
"super" => Some(SUPER_KW),
|
"super" => SUPER_KW,
|
||||||
"in" => Some(IN_KW),
|
"in" => IN_KW,
|
||||||
"where" => Some(WHERE_KW),
|
"where" => WHERE_KW,
|
||||||
"for" => Some(FOR_KW),
|
"for" => FOR_KW,
|
||||||
"loop" => Some(LOOP_KW),
|
"loop" => LOOP_KW,
|
||||||
"while" => Some(WHILE_KW),
|
"while" => WHILE_KW,
|
||||||
"if" => Some(IF_KW),
|
"if" => IF_KW,
|
||||||
"match" => Some(MATCH_KW),
|
"match" => MATCH_KW,
|
||||||
"const" => Some(CONST_KW),
|
"const" => CONST_KW,
|
||||||
"static" => Some(STATIC_KW),
|
"static" => STATIC_KW,
|
||||||
"mut" => Some(MUT_KW),
|
"mut" => MUT_KW,
|
||||||
"unsafe" => Some(UNSAFE_KW),
|
"unsafe" => UNSAFE_KW,
|
||||||
"type" => Some(TYPE_KW),
|
"type" => TYPE_KW,
|
||||||
"ref" => Some(REF_KW),
|
"ref" => REF_KW,
|
||||||
"let" => Some(LET_KW),
|
"let" => LET_KW,
|
||||||
_ => None,
|
_ => return None,
|
||||||
}
|
};
|
||||||
|
Some(kw)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
59
src/syntax_kinds/generated.rs.tera
Normal file
59
src/syntax_kinds/generated.rs.tera
Normal file
|
@ -0,0 +1,59 @@
|
||||||
|
#![allow(bad_style, missing_docs, unreachable_pub)]
|
||||||
|
#![cfg_attr(rustfmt, rustfmt_skip)]
|
||||||
|
use super::SyntaxInfo;
|
||||||
|
|
||||||
|
/// The kind of syntax node, e.g. `IDENT`, `USE_KW`, or `STRUCT_DEF`.
|
||||||
|
#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, Hash)]
|
||||||
|
pub enum SyntaxKind {
|
||||||
|
{%- for t in tokens %}
|
||||||
|
{{t}},
|
||||||
|
{%- endfor -%}
|
||||||
|
{% for kw in keywords %}
|
||||||
|
{{kw | upper}}_KW,
|
||||||
|
{%- endfor -%}
|
||||||
|
{% for kw in contextual_keywords %}
|
||||||
|
{{kw | upper}}_KW,
|
||||||
|
{%- endfor -%}
|
||||||
|
{% for node in nodes %}
|
||||||
|
{{node}},
|
||||||
|
{%- endfor %}
|
||||||
|
// Technical SyntaxKinds: they appear temporally during parsing,
|
||||||
|
// but never end up in the final tree
|
||||||
|
#[doc(hidden)]
|
||||||
|
TOMBSTONE,
|
||||||
|
#[doc(hidden)]
|
||||||
|
EOF,
|
||||||
|
}
|
||||||
|
use self::SyntaxKind::*;
|
||||||
|
|
||||||
|
impl SyntaxKind {
|
||||||
|
pub(crate) fn info(self) -> &'static SyntaxInfo {
|
||||||
|
match self {
|
||||||
|
{%- for t in tokens %}
|
||||||
|
{{t}} => &SyntaxInfo { name: "{{t}}" },
|
||||||
|
{%- endfor -%}
|
||||||
|
{% for kw in keywords %}
|
||||||
|
{{kw | upper}}_KW => &SyntaxInfo { name: "{{kw | upper}}_KW" },
|
||||||
|
{%- endfor -%}
|
||||||
|
{% for kw in contextual_keywords %}
|
||||||
|
{{kw | upper}}_KW => &SyntaxInfo { name: "{{kw | upper}}_KW" },
|
||||||
|
{%- endfor -%}
|
||||||
|
{% for node in nodes %}
|
||||||
|
{{node}} => &SyntaxInfo { name: "{{node}}" },
|
||||||
|
{%- endfor %}
|
||||||
|
|
||||||
|
TOMBSTONE => &SyntaxInfo { name: "TOMBSTONE" },
|
||||||
|
EOF => &SyntaxInfo { name: "EOF" },
|
||||||
|
}
|
||||||
|
}
|
||||||
|
pub(crate) fn from_keyword(ident: &str) -> Option<SyntaxKind> {
|
||||||
|
let kw = match ident {
|
||||||
|
{%- for kw in keywords %}
|
||||||
|
"{{kw}}" => {{kw | upper}}_KW,
|
||||||
|
{%- endfor %}
|
||||||
|
_ => return None,
|
||||||
|
};
|
||||||
|
Some(kw)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
@ -5,10 +5,9 @@ authors = ["Aleksey Kladov <aleksey.kladov@gmail.com>"]
|
||||||
publish = false
|
publish = false
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
serde = "1.0.26"
|
ron = "0.1.7"
|
||||||
serde_derive = "1.0.26"
|
walkdir = "2.1.3"
|
||||||
file = "1.1.1"
|
itertools = "0.7.8"
|
||||||
ron = "0.1.5"
|
tera = "0.11"
|
||||||
walkdir = "2"
|
clap = "2.32.0"
|
||||||
itertools = "0.7"
|
failure = "0.1.1"
|
||||||
libsyntax2 = { path = "../" }
|
|
||||||
|
|
|
@ -1,133 +0,0 @@
|
||||||
extern crate file;
|
|
||||||
extern crate itertools;
|
|
||||||
extern crate walkdir;
|
|
||||||
|
|
||||||
use walkdir::WalkDir;
|
|
||||||
use itertools::Itertools;
|
|
||||||
|
|
||||||
use std::path::{Path, PathBuf};
|
|
||||||
use std::collections::HashSet;
|
|
||||||
use std::fs;
|
|
||||||
|
|
||||||
fn main() {
|
|
||||||
let verify = ::std::env::args().any(|arg| arg == "--verify");
|
|
||||||
|
|
||||||
let d = grammar_dir();
|
|
||||||
let tests = tests_from_dir(&d);
|
|
||||||
let existing = existing_tests();
|
|
||||||
|
|
||||||
for t in existing.difference(&tests) {
|
|
||||||
panic!("Test is deleted: {}\n{}", t.name, t.text);
|
|
||||||
}
|
|
||||||
|
|
||||||
let new_tests = tests.difference(&existing);
|
|
||||||
for (i, t) in new_tests.enumerate() {
|
|
||||||
if verify {
|
|
||||||
panic!("Inline test is not recorded: {}", t.name);
|
|
||||||
}
|
|
||||||
|
|
||||||
let name = format!("{:04}_{}.rs", existing.len() + i + 1, t.name);
|
|
||||||
println!("Creating {}", name);
|
|
||||||
let path = inline_tests_dir().join(name);
|
|
||||||
file::put_text(&path, &t.text).unwrap();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Debug, Eq)]
|
|
||||||
struct Test {
|
|
||||||
name: String,
|
|
||||||
text: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl PartialEq for Test {
|
|
||||||
fn eq(&self, other: &Test) -> bool {
|
|
||||||
self.name.eq(&other.name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl ::std::hash::Hash for Test {
|
|
||||||
fn hash<H: ::std::hash::Hasher>(&self, state: &mut H) {
|
|
||||||
self.name.hash(state)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn tests_from_dir(dir: &Path) -> HashSet<Test> {
|
|
||||||
let mut res = HashSet::new();
|
|
||||||
for entry in WalkDir::new(dir) {
|
|
||||||
let entry = entry.unwrap();
|
|
||||||
if !entry.file_type().is_file() {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
if entry.path().extension().unwrap_or_default() != "rs" {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
let text = file::get_text(entry.path()).unwrap();
|
|
||||||
|
|
||||||
for test in collect_tests(&text) {
|
|
||||||
if let Some(old_test) = res.replace(test) {
|
|
||||||
panic!("Duplicate test: {}", old_test.name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
fn collect_tests(s: &str) -> Vec<Test> {
|
|
||||||
let mut res = vec![];
|
|
||||||
let prefix = "// ";
|
|
||||||
let comment_blocks = s.lines()
|
|
||||||
.map(str::trim_left)
|
|
||||||
.group_by(|line| line.starts_with(prefix));
|
|
||||||
|
|
||||||
'outer: for (is_comment, block) in comment_blocks.into_iter() {
|
|
||||||
if !is_comment {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
let mut block = block.map(|line| &line[prefix.len()..]);
|
|
||||||
|
|
||||||
let name = loop {
|
|
||||||
match block.next() {
|
|
||||||
Some(line) if line.starts_with("test ") => break line["test ".len()..].to_string(),
|
|
||||||
Some(_) => (),
|
|
||||||
None => continue 'outer,
|
|
||||||
}
|
|
||||||
};
|
|
||||||
let text: String = itertools::join(block.chain(::std::iter::once("")), "\n");
|
|
||||||
assert!(!text.trim().is_empty() && text.ends_with("\n"));
|
|
||||||
res.push(Test { name, text })
|
|
||||||
}
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
fn existing_tests() -> HashSet<Test> {
|
|
||||||
let mut res = HashSet::new();
|
|
||||||
for file in fs::read_dir(&inline_tests_dir()).unwrap() {
|
|
||||||
let file = file.unwrap();
|
|
||||||
let path = file.path();
|
|
||||||
if path.extension().unwrap_or_default() != "rs" {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
let name = path.file_name().unwrap().to_str().unwrap();
|
|
||||||
let name = name["0000_".len()..name.len() - 3].to_string();
|
|
||||||
let text = file::get_text(&path).unwrap();
|
|
||||||
res.insert(Test { name, text });
|
|
||||||
}
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
fn inline_tests_dir() -> PathBuf {
|
|
||||||
let res = base_dir().join("tests/data/parser/inline");
|
|
||||||
if !res.is_dir() {
|
|
||||||
fs::create_dir_all(&res).unwrap();
|
|
||||||
}
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
fn grammar_dir() -> PathBuf {
|
|
||||||
base_dir().join("src/parser/grammar")
|
|
||||||
}
|
|
||||||
|
|
||||||
fn base_dir() -> PathBuf {
|
|
||||||
let dir = env!("CARGO_MANIFEST_DIR");
|
|
||||||
PathBuf::from(dir).parent().unwrap().to_owned()
|
|
||||||
}
|
|
|
@ -1,121 +0,0 @@
|
||||||
extern crate serde;
|
|
||||||
#[macro_use]
|
|
||||||
extern crate serde_derive;
|
|
||||||
|
|
||||||
extern crate file;
|
|
||||||
extern crate ron;
|
|
||||||
|
|
||||||
use std::path::PathBuf;
|
|
||||||
use std::fmt::Write;
|
|
||||||
|
|
||||||
fn main() {
|
|
||||||
let grammar = Grammar::read();
|
|
||||||
let text = grammar.to_syntax_kinds();
|
|
||||||
let target = generated_file();
|
|
||||||
if text != file::get_text(&target).unwrap_or_default() {
|
|
||||||
file::put_text(&target, &text).unwrap();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Deserialize)]
|
|
||||||
struct Grammar {
|
|
||||||
keywords: Vec<String>,
|
|
||||||
contextual_keywords: Vec<String>,
|
|
||||||
tokens: Vec<String>,
|
|
||||||
nodes: Vec<String>,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Grammar {
|
|
||||||
fn read() -> Grammar {
|
|
||||||
let text = file::get_text(&grammar_file()).unwrap();
|
|
||||||
ron::de::from_str(&text).unwrap()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn to_syntax_kinds(&self) -> String {
|
|
||||||
let mut acc = String::new();
|
|
||||||
acc.push_str("#![allow(bad_style, missing_docs, unreachable_pub)]\n");
|
|
||||||
acc.push_str("#![cfg_attr(rustfmt, rustfmt_skip)]\n");
|
|
||||||
acc.push_str("//! Generated from grammar.ron\n");
|
|
||||||
acc.push_str("use super::SyntaxInfo;\n");
|
|
||||||
acc.push_str("\n");
|
|
||||||
|
|
||||||
let syntax_kinds: Vec<String> = self.tokens
|
|
||||||
.iter()
|
|
||||||
.cloned()
|
|
||||||
.chain(self.keywords.iter().map(|kw| kw_token(kw)))
|
|
||||||
.chain(self.contextual_keywords.iter().map(|kw| kw_token(kw)))
|
|
||||||
.chain(self.nodes.iter().cloned())
|
|
||||||
.collect();
|
|
||||||
|
|
||||||
// enum SyntaxKind
|
|
||||||
acc.push_str("/// The kind of syntax node, e.g. `IDENT`, `USE_KW`, or `STRUCT_DEF`.\n");
|
|
||||||
acc.push_str("#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, Hash)]\n");
|
|
||||||
acc.push_str("pub enum SyntaxKind {\n");
|
|
||||||
for kind in syntax_kinds.iter() {
|
|
||||||
write!(acc, " {},\n", scream(kind)).unwrap();
|
|
||||||
}
|
|
||||||
acc.push_str("\n");
|
|
||||||
acc.push_str(" // Technical SyntaxKinds: they appear temporally during parsing,\n");
|
|
||||||
acc.push_str(" // but never end up in the final tree\n");
|
|
||||||
acc.push_str(" #[doc(hidden)]\n");
|
|
||||||
acc.push_str(" TOMBSTONE,\n");
|
|
||||||
acc.push_str(" #[doc(hidden)]\n");
|
|
||||||
acc.push_str(" EOF,\n");
|
|
||||||
acc.push_str("}\n");
|
|
||||||
acc.push_str("pub(crate) use self::SyntaxKind::*;\n");
|
|
||||||
acc.push_str("\n");
|
|
||||||
|
|
||||||
// fn info
|
|
||||||
acc.push_str("impl SyntaxKind {\n");
|
|
||||||
acc.push_str(" pub(crate) fn info(self) -> &'static SyntaxInfo {\n");
|
|
||||||
acc.push_str(" match self {\n");
|
|
||||||
for kind in syntax_kinds.iter() {
|
|
||||||
let sname = scream(kind);
|
|
||||||
write!(
|
|
||||||
acc,
|
|
||||||
" {sname} => &SyntaxInfo {{ name: \"{sname}\" }},\n",
|
|
||||||
sname = sname
|
|
||||||
).unwrap();
|
|
||||||
}
|
|
||||||
acc.push_str("\n");
|
|
||||||
acc.push_str(" TOMBSTONE => &SyntaxInfo { name: \"TOMBSTONE\" },\n");
|
|
||||||
acc.push_str(" EOF => &SyntaxInfo { name: \"EOF\" },\n");
|
|
||||||
acc.push_str(" }\n");
|
|
||||||
acc.push_str(" }\n");
|
|
||||||
|
|
||||||
// fn from_keyword
|
|
||||||
acc.push_str(" pub(crate) fn from_keyword(ident: &str) -> Option<SyntaxKind> {\n");
|
|
||||||
acc.push_str(" match ident {\n");
|
|
||||||
// NB: no contextual_keywords here!
|
|
||||||
for kw in self.keywords.iter() {
|
|
||||||
write!(acc, " {:?} => Some({}),\n", kw, kw_token(kw)).unwrap();
|
|
||||||
}
|
|
||||||
acc.push_str(" _ => None,\n");
|
|
||||||
acc.push_str(" }\n");
|
|
||||||
acc.push_str(" }\n");
|
|
||||||
acc.push_str("}\n");
|
|
||||||
acc.push_str("\n");
|
|
||||||
acc
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn grammar_file() -> PathBuf {
|
|
||||||
base_dir().join("src/grammar.ron")
|
|
||||||
}
|
|
||||||
|
|
||||||
fn generated_file() -> PathBuf {
|
|
||||||
base_dir().join("src/syntax_kinds/generated.rs")
|
|
||||||
}
|
|
||||||
|
|
||||||
fn scream(word: &str) -> String {
|
|
||||||
word.chars().map(|c| c.to_ascii_uppercase()).collect()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn kw_token(keyword: &str) -> String {
|
|
||||||
format!("{}_KW", scream(keyword))
|
|
||||||
}
|
|
||||||
|
|
||||||
fn base_dir() -> PathBuf {
|
|
||||||
let dir = env!("CARGO_MANIFEST_DIR");
|
|
||||||
PathBuf::from(dir).parent().unwrap().to_owned()
|
|
||||||
}
|
|
184
tools/src/bin/main.rs
Normal file
184
tools/src/bin/main.rs
Normal file
|
@ -0,0 +1,184 @@
|
||||||
|
extern crate clap;
|
||||||
|
#[macro_use]
|
||||||
|
extern crate failure;
|
||||||
|
extern crate tera;
|
||||||
|
extern crate ron;
|
||||||
|
extern crate walkdir;
|
||||||
|
extern crate itertools;
|
||||||
|
|
||||||
|
use std::{
|
||||||
|
fs,
|
||||||
|
path::{Path},
|
||||||
|
collections::HashSet,
|
||||||
|
};
|
||||||
|
use clap::{App, Arg, SubCommand};
|
||||||
|
use itertools::Itertools;
|
||||||
|
|
||||||
|
type Result<T> = ::std::result::Result<T, failure::Error>;
|
||||||
|
|
||||||
|
const GRAMMAR_DIR: &str = "./src/parser/grammar";
|
||||||
|
const INLINE_TESTS_DIR: &str = "tests/data/parser/inline";
|
||||||
|
const GRAMMAR: &str = "./src/grammar.ron";
|
||||||
|
const SYNTAX_KINDS: &str = "./src/syntax_kinds/generated.rs";
|
||||||
|
const SYNTAX_KINDS_TEMPLATE: &str = "./src/syntax_kinds/generated.rs.tera";
|
||||||
|
|
||||||
|
fn main() -> Result<()> {
|
||||||
|
let matches = App::new("tasks")
|
||||||
|
.setting(clap::AppSettings::SubcommandRequiredElseHelp)
|
||||||
|
.arg(
|
||||||
|
Arg::with_name("verify")
|
||||||
|
.long("--verify")
|
||||||
|
.help("Verify that generated code is up-to-date")
|
||||||
|
.global(true)
|
||||||
|
)
|
||||||
|
.subcommand(SubCommand::with_name("gen-kinds"))
|
||||||
|
.subcommand(SubCommand::with_name("gen-tests"))
|
||||||
|
.get_matches();
|
||||||
|
match matches.subcommand() {
|
||||||
|
(name, Some(matches)) => run_gen_command(name, matches.is_present("verify"))?,
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn run_gen_command(name: &str, verify: bool) -> Result<()> {
|
||||||
|
match name {
|
||||||
|
"gen-kinds" => update(Path::new(SYNTAX_KINDS), &get_kinds()?, verify),
|
||||||
|
"gen-tests" => gen_tests(verify),
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn update(path: &Path, contents: &str, verify: bool) -> Result<()> {
|
||||||
|
match fs::read_to_string(path) {
|
||||||
|
Ok(ref old_contents) if old_contents == contents => {
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
_ => (),
|
||||||
|
}
|
||||||
|
if verify {
|
||||||
|
bail!("`{}` is not up-to-date", path.display());
|
||||||
|
}
|
||||||
|
fs::write(path, contents)?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_kinds() -> Result<String> {
|
||||||
|
let grammar = grammar()?;
|
||||||
|
let template = fs::read_to_string(SYNTAX_KINDS_TEMPLATE)?;
|
||||||
|
let ret = tera::Tera::one_off(&template, &grammar, false).map_err(|e| {
|
||||||
|
format_err!("template error: {}", e)
|
||||||
|
})?;
|
||||||
|
Ok(ret)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn grammar() -> Result<ron::value::Value> {
|
||||||
|
let text = fs::read_to_string(GRAMMAR)?;
|
||||||
|
let ret = ron::de::from_str(&text)?;
|
||||||
|
Ok(ret)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn gen_tests(verify: bool) -> Result<()> {
|
||||||
|
let tests = tests_from_dir(Path::new(GRAMMAR_DIR))?;
|
||||||
|
|
||||||
|
let inline_tests_dir = Path::new(INLINE_TESTS_DIR);
|
||||||
|
if !inline_tests_dir.is_dir() {
|
||||||
|
fs::create_dir_all(inline_tests_dir)?;
|
||||||
|
}
|
||||||
|
let existing = existing_tests(inline_tests_dir)?;
|
||||||
|
|
||||||
|
for t in existing.difference(&tests) {
|
||||||
|
panic!("Test is deleted: {}\n{}", t.name, t.text);
|
||||||
|
}
|
||||||
|
|
||||||
|
let new_tests = tests.difference(&existing);
|
||||||
|
for (i, t) in new_tests.enumerate() {
|
||||||
|
let name = format!("{:04}_{}.rs", existing.len() + i + 1, t.name);
|
||||||
|
let path = inline_tests_dir.join(name);
|
||||||
|
update(&path, &t.text, verify)?;
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Eq)]
|
||||||
|
struct Test {
|
||||||
|
name: String,
|
||||||
|
text: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl PartialEq for Test {
|
||||||
|
fn eq(&self, other: &Test) -> bool {
|
||||||
|
self.name.eq(&other.name)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ::std::hash::Hash for Test {
|
||||||
|
fn hash<H: ::std::hash::Hasher>(&self, state: &mut H) {
|
||||||
|
self.name.hash(state)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn tests_from_dir(dir: &Path) -> Result<HashSet<Test>> {
|
||||||
|
let mut res = HashSet::new();
|
||||||
|
for entry in ::walkdir::WalkDir::new(dir) {
|
||||||
|
let entry = entry.unwrap();
|
||||||
|
if !entry.file_type().is_file() {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if entry.path().extension().unwrap_or_default() != "rs" {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
let text = fs::read_to_string(entry.path())?;
|
||||||
|
|
||||||
|
for test in collect_tests(&text) {
|
||||||
|
if let Some(old_test) = res.replace(test) {
|
||||||
|
bail!("Duplicate test: {}", old_test.name)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(res)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn collect_tests(s: &str) -> Vec<Test> {
|
||||||
|
let mut res = vec![];
|
||||||
|
let prefix = "// ";
|
||||||
|
let comment_blocks = s.lines()
|
||||||
|
.map(str::trim_left)
|
||||||
|
.group_by(|line| line.starts_with(prefix));
|
||||||
|
|
||||||
|
'outer: for (is_comment, block) in comment_blocks.into_iter() {
|
||||||
|
if !is_comment {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
let mut block = block.map(|line| &line[prefix.len()..]);
|
||||||
|
|
||||||
|
let name = loop {
|
||||||
|
match block.next() {
|
||||||
|
Some(line) if line.starts_with("test ") => break line["test ".len()..].to_string(),
|
||||||
|
Some(_) => (),
|
||||||
|
None => continue 'outer,
|
||||||
|
}
|
||||||
|
};
|
||||||
|
let text: String = itertools::join(block.chain(::std::iter::once("")), "\n");
|
||||||
|
assert!(!text.trim().is_empty() && text.ends_with("\n"));
|
||||||
|
res.push(Test { name, text })
|
||||||
|
}
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
||||||
|
fn existing_tests(dir: &Path) -> Result<HashSet<Test>> {
|
||||||
|
let mut res = HashSet::new();
|
||||||
|
for file in fs::read_dir(dir)? {
|
||||||
|
let file = file?;
|
||||||
|
let path = file.path();
|
||||||
|
if path.extension().unwrap_or_default() != "rs" {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
let name = path.file_name().unwrap().to_str().unwrap();
|
||||||
|
let name = name["0000_".len()..name.len() - 3].to_string();
|
||||||
|
let text = fs::read_to_string(&path)?;
|
||||||
|
res.insert(Test { name, text });
|
||||||
|
}
|
||||||
|
Ok(res)
|
||||||
|
}
|
||||||
|
|
Loading…
Reference in a new issue