2020-12-23 12:02:02 +00:00
|
|
|
// Run clippy on a fixed set of crates and collect the warnings.
|
|
|
|
// This helps observing the impact clippy changs have on a set of real-world code.
|
|
|
|
//
|
|
|
|
// When a new lint is introduced, we can search the results for new warnings and check for false
|
|
|
|
// positives.
|
|
|
|
|
2021-01-22 23:25:29 +00:00
|
|
|
#![cfg(feature = "lintcheck")]
|
2020-12-18 21:53:45 +00:00
|
|
|
#![allow(clippy::filter_map)]
|
|
|
|
|
2020-12-18 21:08:18 +00:00
|
|
|
use crate::clippy_project_root;
|
2020-12-23 12:02:02 +00:00
|
|
|
|
2020-12-22 12:07:55 +00:00
|
|
|
use std::collections::HashMap;
|
2020-12-18 12:21:13 +00:00
|
|
|
use std::process::Command;
|
2021-02-18 18:09:12 +00:00
|
|
|
use std::sync::atomic::{AtomicUsize, Ordering};
|
2021-02-12 23:39:19 +00:00
|
|
|
use std::{env, fmt, fs::write, path::PathBuf};
|
2020-12-18 21:53:45 +00:00
|
|
|
|
2020-12-27 15:13:42 +00:00
|
|
|
use clap::ArgMatches;
|
2021-02-18 18:09:12 +00:00
|
|
|
use rayon::prelude::*;
|
2020-12-23 12:02:02 +00:00
|
|
|
use serde::{Deserialize, Serialize};
|
2020-12-23 14:00:51 +00:00
|
|
|
use serde_json::Value;
|
2020-12-23 12:02:02 +00:00
|
|
|
|
2021-02-27 11:05:27 +00:00
|
|
|
const CLIPPY_DRIVER_PATH: &str = "target/debug/clippy-driver";
|
|
|
|
const CARGO_CLIPPY_PATH: &str = "target/debug/cargo-clippy";
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// List of sources to check, loaded from a .toml file
|
2020-12-23 12:03:19 +00:00
|
|
|
#[derive(Debug, Serialize, Deserialize)]
|
2021-02-14 17:51:53 +00:00
|
|
|
struct SourceList {
|
2021-02-05 22:13:59 +00:00
|
|
|
crates: HashMap<String, TomlCrate>,
|
2020-12-23 12:03:19 +00:00
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// A crate source stored inside the .toml
|
|
|
|
/// will be translated into on one of the `CrateSource` variants
|
2021-02-05 22:13:59 +00:00
|
|
|
#[derive(Debug, Serialize, Deserialize)]
|
2020-12-23 12:02:02 +00:00
|
|
|
struct TomlCrate {
|
2020-12-23 00:21:31 +00:00
|
|
|
name: String,
|
2021-02-05 22:13:59 +00:00
|
|
|
versions: Option<Vec<String>>,
|
|
|
|
git_url: Option<String>,
|
|
|
|
git_hash: Option<String>,
|
2021-02-10 10:32:10 +00:00
|
|
|
path: Option<String>,
|
2021-02-16 12:38:01 +00:00
|
|
|
options: Option<Vec<String>>,
|
2020-12-23 00:21:31 +00:00
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// Represents an archive we download from crates.io, or a git repo, or a local repo/folder
|
|
|
|
/// Once processed (downloaded/extracted/cloned/copied...), this will be translated into a `Crate`
|
2021-02-19 21:16:53 +00:00
|
|
|
#[derive(Debug, Serialize, Deserialize, Eq, Hash, PartialEq, Ord, PartialOrd)]
|
2021-02-06 11:02:42 +00:00
|
|
|
enum CrateSource {
|
2021-02-16 12:38:01 +00:00
|
|
|
CratesIo {
|
|
|
|
name: String,
|
|
|
|
version: String,
|
|
|
|
options: Option<Vec<String>>,
|
|
|
|
},
|
|
|
|
Git {
|
|
|
|
name: String,
|
|
|
|
url: String,
|
|
|
|
commit: String,
|
|
|
|
options: Option<Vec<String>>,
|
|
|
|
},
|
|
|
|
Path {
|
|
|
|
name: String,
|
|
|
|
path: PathBuf,
|
|
|
|
options: Option<Vec<String>>,
|
|
|
|
},
|
2020-12-18 12:21:13 +00:00
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// Represents the actual source code of a crate that we ran "cargo clippy" on
|
2020-12-18 13:14:15 +00:00
|
|
|
#[derive(Debug)]
|
2020-12-23 12:02:02 +00:00
|
|
|
struct Crate {
|
2020-12-18 12:21:13 +00:00
|
|
|
version: String,
|
|
|
|
name: String,
|
2020-12-18 19:58:46 +00:00
|
|
|
// path to the extracted sources that clippy can check
|
2020-12-18 13:14:15 +00:00
|
|
|
path: PathBuf,
|
2021-02-16 12:38:01 +00:00
|
|
|
options: Option<Vec<String>>,
|
2020-12-18 12:21:13 +00:00
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// A single warning that clippy issued while checking a `Crate`
|
2020-12-23 14:00:51 +00:00
|
|
|
#[derive(Debug)]
|
|
|
|
struct ClippyWarning {
|
|
|
|
crate_name: String,
|
|
|
|
crate_version: String,
|
|
|
|
file: String,
|
|
|
|
line: String,
|
|
|
|
column: String,
|
|
|
|
linttype: String,
|
|
|
|
message: String,
|
2021-02-14 17:37:08 +00:00
|
|
|
is_ice: bool,
|
2020-12-23 14:00:51 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
impl std::fmt::Display for ClippyWarning {
|
|
|
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
|
|
writeln!(
|
|
|
|
f,
|
2021-02-27 11:29:13 +00:00
|
|
|
r#"target/lintcheck/sources/{}-{}/{}:{}:{} {} "{}""#,
|
2020-12-23 14:00:51 +00:00
|
|
|
&self.crate_name, &self.crate_version, &self.file, &self.line, &self.column, &self.linttype, &self.message
|
|
|
|
)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-12-23 12:02:02 +00:00
|
|
|
impl CrateSource {
|
2021-02-14 17:51:53 +00:00
|
|
|
/// Makes the sources available on the disk for clippy to check.
|
|
|
|
/// Clones a git repo and checks out the specified commit or downloads a crate from crates.io or
|
|
|
|
/// copies a local folder
|
2020-12-23 12:02:02 +00:00
|
|
|
fn download_and_extract(&self) -> Crate {
|
2021-02-06 11:02:42 +00:00
|
|
|
match self {
|
2021-02-16 12:38:01 +00:00
|
|
|
CrateSource::CratesIo { name, version, options } => {
|
2021-02-06 11:02:42 +00:00
|
|
|
let extract_dir = PathBuf::from("target/lintcheck/crates");
|
|
|
|
let krate_download_dir = PathBuf::from("target/lintcheck/downloads");
|
|
|
|
|
|
|
|
// url to download the crate from crates.io
|
|
|
|
let url = format!("https://crates.io/api/v1/crates/{}/{}/download", name, version);
|
|
|
|
println!("Downloading and extracting {} {} from {}", name, version, url);
|
|
|
|
let _ = std::fs::create_dir("target/lintcheck/");
|
|
|
|
let _ = std::fs::create_dir(&krate_download_dir);
|
|
|
|
let _ = std::fs::create_dir(&extract_dir);
|
|
|
|
|
|
|
|
let krate_file_path = krate_download_dir.join(format!("{}-{}.crate.tar.gz", name, version));
|
|
|
|
// don't download/extract if we already have done so
|
|
|
|
if !krate_file_path.is_file() {
|
|
|
|
// create a file path to download and write the crate data into
|
|
|
|
let mut krate_dest = std::fs::File::create(&krate_file_path).unwrap();
|
|
|
|
let mut krate_req = ureq::get(&url).call().unwrap().into_reader();
|
|
|
|
// copy the crate into the file
|
|
|
|
std::io::copy(&mut krate_req, &mut krate_dest).unwrap();
|
|
|
|
|
|
|
|
// unzip the tarball
|
|
|
|
let ungz_tar = flate2::read::GzDecoder::new(std::fs::File::open(&krate_file_path).unwrap());
|
|
|
|
// extract the tar archive
|
|
|
|
let mut archive = tar::Archive::new(ungz_tar);
|
|
|
|
archive.unpack(&extract_dir).expect("Failed to extract!");
|
|
|
|
}
|
|
|
|
// crate is extracted, return a new Krate object which contains the path to the extracted
|
|
|
|
// sources that clippy can check
|
|
|
|
Crate {
|
|
|
|
version: version.clone(),
|
|
|
|
name: name.clone(),
|
|
|
|
path: extract_dir.join(format!("{}-{}/", name, version)),
|
2021-02-16 12:38:01 +00:00
|
|
|
options: options.clone(),
|
2021-02-06 11:02:42 +00:00
|
|
|
}
|
|
|
|
},
|
2021-02-16 12:38:01 +00:00
|
|
|
CrateSource::Git {
|
|
|
|
name,
|
|
|
|
url,
|
|
|
|
commit,
|
|
|
|
options,
|
|
|
|
} => {
|
2021-02-06 11:02:42 +00:00
|
|
|
let repo_path = {
|
2021-02-10 10:32:10 +00:00
|
|
|
let mut repo_path = PathBuf::from("target/lintcheck/crates");
|
2021-02-06 11:02:42 +00:00
|
|
|
// add a -git suffix in case we have the same crate from crates.io and a git repo
|
|
|
|
repo_path.push(format!("{}-git", name));
|
|
|
|
repo_path
|
|
|
|
};
|
|
|
|
// clone the repo if we have not done so
|
|
|
|
if !repo_path.is_dir() {
|
2021-02-06 10:36:06 +00:00
|
|
|
println!("Cloning {} and checking out {}", url, commit);
|
2021-02-15 21:46:58 +00:00
|
|
|
if !Command::new("git")
|
2021-02-06 11:02:42 +00:00
|
|
|
.arg("clone")
|
|
|
|
.arg(url)
|
|
|
|
.arg(&repo_path)
|
2021-02-15 21:46:58 +00:00
|
|
|
.status()
|
|
|
|
.expect("Failed to clone git repo!")
|
|
|
|
.success()
|
|
|
|
{
|
|
|
|
eprintln!("Failed to clone {} into {}", url, repo_path.display())
|
|
|
|
}
|
2021-02-06 11:02:42 +00:00
|
|
|
}
|
|
|
|
// check out the commit/branch/whatever
|
2021-02-15 21:46:58 +00:00
|
|
|
if !Command::new("git")
|
2021-02-06 11:02:42 +00:00
|
|
|
.arg("checkout")
|
|
|
|
.arg(commit)
|
2021-02-15 21:36:49 +00:00
|
|
|
.current_dir(&repo_path)
|
2021-02-15 21:46:58 +00:00
|
|
|
.status()
|
|
|
|
.expect("Failed to check out commit")
|
|
|
|
.success()
|
|
|
|
{
|
|
|
|
eprintln!("Failed to checkout {} of repo at {}", commit, repo_path.display())
|
|
|
|
}
|
2021-02-06 11:02:42 +00:00
|
|
|
|
|
|
|
Crate {
|
|
|
|
version: commit.clone(),
|
|
|
|
name: name.clone(),
|
|
|
|
path: repo_path,
|
2021-02-16 12:38:01 +00:00
|
|
|
options: options.clone(),
|
2021-02-06 11:02:42 +00:00
|
|
|
}
|
|
|
|
},
|
2021-02-16 12:38:01 +00:00
|
|
|
CrateSource::Path { name, path, options } => {
|
2021-02-10 10:32:10 +00:00
|
|
|
use fs_extra::dir;
|
|
|
|
|
|
|
|
// simply copy the entire directory into our target dir
|
|
|
|
let copy_dest = PathBuf::from("target/lintcheck/crates/");
|
|
|
|
|
|
|
|
// the source path of the crate we copied, ${copy_dest}/crate_name
|
|
|
|
let crate_root = copy_dest.join(name); // .../crates/local_crate
|
|
|
|
|
|
|
|
if !crate_root.exists() {
|
|
|
|
println!("Copying {} to {}", path.display(), copy_dest.display());
|
|
|
|
|
|
|
|
dir::copy(path, ©_dest, &dir::CopyOptions::new()).expect(&format!(
|
|
|
|
"Failed to copy from {}, to {}",
|
|
|
|
path.display(),
|
|
|
|
crate_root.display()
|
|
|
|
));
|
|
|
|
} else {
|
|
|
|
println!(
|
|
|
|
"Not copying {} to {}, destination already exists",
|
|
|
|
path.display(),
|
|
|
|
crate_root.display()
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
Crate {
|
|
|
|
version: String::from("local"),
|
|
|
|
name: name.clone(),
|
|
|
|
path: crate_root,
|
2021-02-16 12:38:01 +00:00
|
|
|
options: options.clone(),
|
2021-02-10 10:32:10 +00:00
|
|
|
}
|
|
|
|
},
|
2020-12-18 12:21:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-12-23 12:02:02 +00:00
|
|
|
impl Crate {
|
2021-02-14 17:51:53 +00:00
|
|
|
/// Run `cargo clippy` on the `Crate` and collect and return all the lint warnings that clippy
|
|
|
|
/// issued
|
2021-02-18 18:09:12 +00:00
|
|
|
fn run_clippy_lints(
|
|
|
|
&self,
|
|
|
|
cargo_clippy_path: &PathBuf,
|
|
|
|
target_dir_index: &AtomicUsize,
|
|
|
|
thread_limit: usize,
|
2021-02-19 21:06:50 +00:00
|
|
|
total_crates_to_lint: usize,
|
2021-02-18 18:09:12 +00:00
|
|
|
) -> Vec<ClippyWarning> {
|
|
|
|
// advance the atomic index by one
|
2021-02-19 21:06:50 +00:00
|
|
|
let index = target_dir_index.fetch_add(1, Ordering::SeqCst);
|
2021-02-18 18:09:12 +00:00
|
|
|
// "loop" the index within 0..thread_limit
|
2021-02-19 21:06:50 +00:00
|
|
|
let target_dir_index = index % thread_limit;
|
|
|
|
let perc = ((index * 100) as f32 / total_crates_to_lint as f32) as u8;
|
2021-02-19 22:20:05 +00:00
|
|
|
|
|
|
|
if thread_limit == 1 {
|
|
|
|
println!(
|
|
|
|
"{}/{} {}% Linting {} {}",
|
|
|
|
index, total_crates_to_lint, perc, &self.name, &self.version
|
|
|
|
);
|
|
|
|
} else {
|
|
|
|
println!(
|
|
|
|
"{}/{} {}% Linting {} {} in target dir {:?}",
|
|
|
|
index, total_crates_to_lint, perc, &self.name, &self.version, target_dir_index
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
2020-12-18 16:25:07 +00:00
|
|
|
let cargo_clippy_path = std::fs::canonicalize(cargo_clippy_path).unwrap();
|
|
|
|
|
2021-02-18 18:09:12 +00:00
|
|
|
let shared_target_dir = clippy_project_root().join("target/lintcheck/shared_target_dir");
|
2020-12-18 21:08:18 +00:00
|
|
|
|
2021-02-16 12:38:01 +00:00
|
|
|
let mut args = vec!["--", "--message-format=json", "--", "--cap-lints=warn"];
|
|
|
|
|
|
|
|
if let Some(options) = &self.options {
|
|
|
|
for opt in options {
|
|
|
|
args.push(opt);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
args.extend(&["-Wclippy::pedantic", "-Wclippy::cargo"])
|
|
|
|
}
|
|
|
|
|
2021-02-05 22:13:59 +00:00
|
|
|
let all_output = std::process::Command::new(&cargo_clippy_path)
|
2021-02-18 18:09:12 +00:00
|
|
|
// use the looping index to create individual target dirs
|
2021-02-19 21:06:50 +00:00
|
|
|
.env(
|
|
|
|
"CARGO_TARGET_DIR",
|
|
|
|
shared_target_dir.join(format!("_{:?}", target_dir_index)),
|
|
|
|
)
|
2020-12-18 19:58:46 +00:00
|
|
|
// lint warnings will look like this:
|
|
|
|
// src/cargo/ops/cargo_compile.rs:127:35: warning: usage of `FromIterator::from_iter`
|
2021-02-16 12:38:01 +00:00
|
|
|
.args(&args)
|
2020-12-18 19:58:46 +00:00
|
|
|
.current_dir(&self.path)
|
2020-12-18 17:34:09 +00:00
|
|
|
.output()
|
2021-02-05 22:13:59 +00:00
|
|
|
.unwrap_or_else(|error| {
|
2021-02-06 18:12:28 +00:00
|
|
|
panic!(
|
|
|
|
"Encountered error:\n{:?}\ncargo_clippy_path: {}\ncrate path:{}\n",
|
|
|
|
error,
|
|
|
|
&cargo_clippy_path.display(),
|
|
|
|
&self.path.display()
|
|
|
|
);
|
2021-02-05 22:13:59 +00:00
|
|
|
});
|
2020-12-23 14:00:51 +00:00
|
|
|
let stdout = String::from_utf8_lossy(&all_output.stdout);
|
|
|
|
let output_lines = stdout.lines();
|
|
|
|
let warnings: Vec<ClippyWarning> = output_lines
|
2020-12-18 17:34:09 +00:00
|
|
|
.into_iter()
|
2021-02-10 11:50:36 +00:00
|
|
|
// get all clippy warnings and ICEs
|
2021-02-15 22:13:41 +00:00
|
|
|
.filter(|line| filter_clippy_warnings(&line))
|
2020-12-23 14:00:51 +00:00
|
|
|
.map(|json_msg| parse_json_message(json_msg, &self))
|
2020-12-18 17:34:09 +00:00
|
|
|
.collect();
|
2020-12-23 14:31:18 +00:00
|
|
|
warnings
|
2020-12-18 12:21:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-26 23:29:42 +00:00
|
|
|
#[derive(Debug)]
|
|
|
|
struct LintcheckConfig {
|
|
|
|
// max number of jobs to spawn (default 1)
|
|
|
|
max_jobs: usize,
|
|
|
|
// we read the sources to check from here
|
|
|
|
sources_toml_path: PathBuf,
|
|
|
|
// we save the clippy lint results here
|
|
|
|
lintcheck_results_path: PathBuf,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl LintcheckConfig {
|
|
|
|
fn from_clap(clap_config: &ArgMatches) -> Self {
|
|
|
|
// first, check if we got anything passed via the LINTCHECK_TOML env var,
|
|
|
|
// if not, ask clap if we got any value for --crates-toml <foo>
|
|
|
|
// if not, use the default "clippy_dev/lintcheck_crates.toml"
|
|
|
|
let sources_toml = env::var("LINTCHECK_TOML").unwrap_or(
|
|
|
|
clap_config
|
|
|
|
.value_of("crates-toml")
|
|
|
|
.clone()
|
|
|
|
.unwrap_or("clippy_dev/lintcheck_crates.toml")
|
|
|
|
.to_string(),
|
|
|
|
);
|
|
|
|
|
|
|
|
let sources_toml_path = PathBuf::from(sources_toml);
|
|
|
|
|
|
|
|
// for the path where we save the lint results, get the filename without extenstion ( so for
|
|
|
|
// wasd.toml, use "wasd"....)
|
|
|
|
let filename: PathBuf = sources_toml_path.file_stem().unwrap().into();
|
|
|
|
let lintcheck_results_path = PathBuf::from(format!("lintcheck-logs/{}_logs.txt", filename.display()));
|
|
|
|
|
2021-02-27 11:05:27 +00:00
|
|
|
// look at the --threads arg, if 0 is passed, ask rayon rayon how many threads it would spawn and
|
|
|
|
// use half of that for the physical core count
|
|
|
|
// by default use a single thread
|
2021-02-26 23:29:42 +00:00
|
|
|
let max_jobs = match clap_config.value_of("threads") {
|
|
|
|
Some(threads) => {
|
|
|
|
let threads: usize = threads
|
|
|
|
.parse()
|
|
|
|
.expect(&format!("Failed to parse '{}' to a digit", threads));
|
|
|
|
if threads == 0 {
|
|
|
|
// automatic choice
|
|
|
|
// Rayon seems to return thread count so half that for core count
|
|
|
|
(rayon::current_num_threads() / 2) as usize
|
|
|
|
} else {
|
|
|
|
threads
|
|
|
|
}
|
|
|
|
},
|
|
|
|
// no -j passed, use a single thread
|
|
|
|
None => 1,
|
|
|
|
};
|
|
|
|
|
|
|
|
LintcheckConfig {
|
|
|
|
max_jobs,
|
|
|
|
sources_toml_path,
|
|
|
|
lintcheck_results_path,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-15 22:13:41 +00:00
|
|
|
/// takes a single json-formatted clippy warnings and returns true (we are interested in that line)
|
|
|
|
/// or false (we aren't)
|
|
|
|
fn filter_clippy_warnings(line: &str) -> bool {
|
|
|
|
// we want to collect ICEs because clippy might have crashed.
|
|
|
|
// these are summarized later
|
|
|
|
if line.contains("internal compiler error: ") {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
// in general, we want all clippy warnings
|
|
|
|
// however due to some kind of bug, sometimes there are absolute paths
|
|
|
|
// to libcore files inside the message
|
|
|
|
// or we end up with cargo-metadata output (https://github.com/rust-lang/rust-clippy/issues/6508)
|
|
|
|
|
|
|
|
// filter out these message to avoid unnecessary noise in the logs
|
|
|
|
if line.contains("clippy::")
|
|
|
|
&& !(line.contains("could not read cargo metadata")
|
|
|
|
|| (line.contains(".rustup") && line.contains("toolchains")))
|
|
|
|
{
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
false
|
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// Builds clippy inside the repo to make sure we have a clippy executable we can use.
|
2020-12-18 12:21:13 +00:00
|
|
|
fn build_clippy() {
|
2021-02-19 20:52:34 +00:00
|
|
|
let status = Command::new("cargo")
|
2020-12-18 12:21:13 +00:00
|
|
|
.arg("build")
|
2021-02-19 20:52:34 +00:00
|
|
|
.status()
|
2020-12-18 12:21:13 +00:00
|
|
|
.expect("Failed to build clippy!");
|
2021-02-19 20:52:34 +00:00
|
|
|
if !status.success() {
|
|
|
|
eprintln!("Error: Failed to compile Clippy!");
|
|
|
|
std::process::exit(1);
|
2021-02-16 15:58:00 +00:00
|
|
|
}
|
2020-12-18 12:21:13 +00:00
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// Read a `toml` file and return a list of `CrateSources` that we want to check with clippy
|
2021-02-26 23:29:42 +00:00
|
|
|
fn read_crates(toml_path: &PathBuf) -> Vec<CrateSource> {
|
2020-12-22 12:07:55 +00:00
|
|
|
let toml_content: String =
|
|
|
|
std::fs::read_to_string(&toml_path).unwrap_or_else(|_| panic!("Failed to read {}", toml_path.display()));
|
2021-02-14 17:51:53 +00:00
|
|
|
let crate_list: SourceList =
|
2020-12-22 12:07:55 +00:00
|
|
|
toml::from_str(&toml_content).unwrap_or_else(|e| panic!("Failed to parse {}: \n{}", toml_path.display(), e));
|
|
|
|
// parse the hashmap of the toml file into a list of crates
|
2020-12-23 12:02:02 +00:00
|
|
|
let tomlcrates: Vec<TomlCrate> = crate_list
|
2020-12-22 12:07:55 +00:00
|
|
|
.crates
|
2020-12-23 00:21:31 +00:00
|
|
|
.into_iter()
|
2021-02-05 22:13:59 +00:00
|
|
|
.map(|(_cratename, tomlcrate)| tomlcrate)
|
2020-12-23 00:21:31 +00:00
|
|
|
.collect();
|
|
|
|
|
2020-12-23 12:02:02 +00:00
|
|
|
// flatten TomlCrates into CrateSources (one TomlCrates may represent several versions of a crate =>
|
|
|
|
// multiple Cratesources)
|
|
|
|
let mut crate_sources = Vec::new();
|
|
|
|
tomlcrates.into_iter().for_each(|tk| {
|
2021-02-10 10:32:10 +00:00
|
|
|
if let Some(ref path) = tk.path {
|
|
|
|
crate_sources.push(CrateSource::Path {
|
|
|
|
name: tk.name.clone(),
|
|
|
|
path: PathBuf::from(path),
|
2021-02-16 12:38:01 +00:00
|
|
|
options: tk.options.clone(),
|
2021-02-10 10:32:10 +00:00
|
|
|
});
|
|
|
|
}
|
|
|
|
|
2021-02-06 11:02:42 +00:00
|
|
|
// if we have multiple versions, save each one
|
2021-02-05 22:13:59 +00:00
|
|
|
if let Some(ref versions) = tk.versions {
|
|
|
|
versions.iter().for_each(|ver| {
|
2021-02-06 11:02:42 +00:00
|
|
|
crate_sources.push(CrateSource::CratesIo {
|
2021-02-05 22:13:59 +00:00
|
|
|
name: tk.name.clone(),
|
|
|
|
version: ver.to_string(),
|
2021-02-16 12:38:01 +00:00
|
|
|
options: tk.options.clone(),
|
2021-02-05 22:13:59 +00:00
|
|
|
});
|
|
|
|
})
|
|
|
|
}
|
2021-02-06 11:02:42 +00:00
|
|
|
// otherwise, we should have a git source
|
|
|
|
if tk.git_url.is_some() && tk.git_hash.is_some() {
|
|
|
|
crate_sources.push(CrateSource::Git {
|
|
|
|
name: tk.name.clone(),
|
|
|
|
url: tk.git_url.clone().unwrap(),
|
|
|
|
commit: tk.git_hash.clone().unwrap(),
|
2021-02-16 12:38:01 +00:00
|
|
|
options: tk.options.clone(),
|
2021-02-06 11:02:42 +00:00
|
|
|
});
|
|
|
|
}
|
|
|
|
// if we have a version as well as a git data OR only one git data, something is funky
|
|
|
|
if tk.versions.is_some() && (tk.git_url.is_some() || tk.git_hash.is_some())
|
|
|
|
|| tk.git_hash.is_some() != tk.git_url.is_some()
|
|
|
|
{
|
2021-02-06 18:12:28 +00:00
|
|
|
eprintln!("tomlkrate: {:?}", tk);
|
2021-02-06 11:04:31 +00:00
|
|
|
if tk.git_hash.is_some() != tk.git_url.is_some() {
|
2021-02-10 10:32:10 +00:00
|
|
|
panic!("Error: Encountered TomlCrate with only one of git_hash and git_url!");
|
|
|
|
}
|
|
|
|
if tk.path.is_some() && (tk.git_hash.is_some() || tk.versions.is_some()) {
|
|
|
|
panic!("Error: TomlCrate can only have one of 'git_.*', 'version' or 'path' fields");
|
2021-02-06 11:04:31 +00:00
|
|
|
}
|
2021-02-06 11:02:42 +00:00
|
|
|
unreachable!("Failed to translate TomlCrate into CrateSource!");
|
|
|
|
}
|
2020-12-23 00:21:31 +00:00
|
|
|
});
|
2021-02-19 21:16:53 +00:00
|
|
|
// sort the crates
|
|
|
|
crate_sources.sort();
|
|
|
|
|
2021-02-26 23:29:42 +00:00
|
|
|
crate_sources
|
2020-12-22 12:07:55 +00:00
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// Parse the json output of clippy and return a `ClippyWarning`
|
2020-12-23 14:00:51 +00:00
|
|
|
fn parse_json_message(json_message: &str, krate: &Crate) -> ClippyWarning {
|
|
|
|
let jmsg: Value = serde_json::from_str(&json_message).unwrap_or_else(|e| panic!("Failed to parse json:\n{:?}", e));
|
|
|
|
|
|
|
|
ClippyWarning {
|
|
|
|
crate_name: krate.name.to_string(),
|
|
|
|
crate_version: krate.version.to_string(),
|
|
|
|
file: jmsg["message"]["spans"][0]["file_name"]
|
|
|
|
.to_string()
|
|
|
|
.trim_matches('"')
|
|
|
|
.into(),
|
|
|
|
line: jmsg["message"]["spans"][0]["line_start"]
|
|
|
|
.to_string()
|
|
|
|
.trim_matches('"')
|
|
|
|
.into(),
|
|
|
|
column: jmsg["message"]["spans"][0]["text"][0]["highlight_start"]
|
|
|
|
.to_string()
|
|
|
|
.trim_matches('"')
|
|
|
|
.into(),
|
|
|
|
linttype: jmsg["message"]["code"]["code"].to_string().trim_matches('"').into(),
|
|
|
|
message: jmsg["message"]["message"].to_string().trim_matches('"').into(),
|
2021-02-14 17:37:08 +00:00
|
|
|
is_ice: json_message.contains("internal compiler error: "),
|
2020-12-23 14:00:51 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-14 17:59:32 +00:00
|
|
|
/// Generate a short list of occuring lints-types and their count
|
2021-02-23 20:23:36 +00:00
|
|
|
fn gather_stats(clippy_warnings: &[ClippyWarning]) -> (String, HashMap<&String, usize>) {
|
2021-02-14 17:59:32 +00:00
|
|
|
// count lint type occurrences
|
|
|
|
let mut counter: HashMap<&String, usize> = HashMap::new();
|
|
|
|
clippy_warnings
|
|
|
|
.iter()
|
|
|
|
.for_each(|wrn| *counter.entry(&wrn.linttype).or_insert(0) += 1);
|
|
|
|
|
|
|
|
// collect into a tupled list for sorting
|
|
|
|
let mut stats: Vec<(&&String, &usize)> = counter.iter().map(|(lint, count)| (lint, count)).collect();
|
|
|
|
// sort by "000{count} {clippy::lintname}"
|
|
|
|
// to not have a lint with 200 and 2 warnings take the same spot
|
|
|
|
stats.sort_by_key(|(lint, count)| format!("{:0>4}, {}", count, lint));
|
|
|
|
|
2021-02-23 20:23:36 +00:00
|
|
|
let stats_string = stats
|
2021-02-14 17:59:32 +00:00
|
|
|
.iter()
|
|
|
|
.map(|(lint, count)| format!("{} {}\n", lint, count))
|
2021-02-23 20:23:36 +00:00
|
|
|
.collect::<String>();
|
|
|
|
|
|
|
|
(stats_string, counter)
|
2021-02-14 17:59:32 +00:00
|
|
|
}
|
|
|
|
|
2021-02-22 23:40:50 +00:00
|
|
|
/// check if the latest modification of the logfile is older than the modification date of the
|
|
|
|
/// clippy binary, if this is true, we should clean the lintchec shared target directory and recheck
|
2021-02-28 01:07:01 +00:00
|
|
|
fn lintcheck_needs_rerun(lintcheck_logs_path: &PathBuf) -> bool {
|
2021-02-22 23:40:50 +00:00
|
|
|
let clippy_modified: std::time::SystemTime = {
|
2021-02-27 11:05:27 +00:00
|
|
|
let mut times = [CLIPPY_DRIVER_PATH, CARGO_CLIPPY_PATH].iter().map(|p| {
|
|
|
|
std::fs::metadata(p)
|
|
|
|
.expect("failed to get metadata of file")
|
|
|
|
.modified()
|
|
|
|
.expect("failed to get modification date")
|
|
|
|
});
|
2021-02-23 20:23:36 +00:00
|
|
|
// the oldest modification of either of the binaries
|
2021-02-28 01:07:01 +00:00
|
|
|
std::cmp::max(times.next().unwrap(), times.next().unwrap())
|
2021-02-22 23:40:50 +00:00
|
|
|
};
|
|
|
|
|
2021-02-28 01:07:01 +00:00
|
|
|
let logs_modified: std::time::SystemTime = std::fs::metadata(lintcheck_logs_path)
|
2021-02-22 23:40:50 +00:00
|
|
|
.expect("failed to get metadata of file")
|
|
|
|
.modified()
|
|
|
|
.expect("failed to get modification date");
|
|
|
|
|
2021-02-28 01:07:01 +00:00
|
|
|
// time is represented in seconds since X
|
|
|
|
// logs_modified 2 and clippy_modified 5 means clippy binary is older and we need to recheck
|
|
|
|
logs_modified < clippy_modified
|
2021-02-22 23:40:50 +00:00
|
|
|
}
|
|
|
|
|
2021-02-14 17:51:53 +00:00
|
|
|
/// lintchecks `main()` function
|
2020-12-27 15:13:42 +00:00
|
|
|
pub fn run(clap_config: &ArgMatches) {
|
2021-02-26 23:29:42 +00:00
|
|
|
let config = LintcheckConfig::from_clap(clap_config);
|
|
|
|
|
2020-12-18 15:17:53 +00:00
|
|
|
println!("Compiling clippy...");
|
2020-12-18 12:21:13 +00:00
|
|
|
build_clippy();
|
2020-12-18 15:17:53 +00:00
|
|
|
println!("Done compiling");
|
|
|
|
|
2021-02-22 23:40:50 +00:00
|
|
|
// if the clippy bin is newer than our logs, throw away target dirs to force clippy to
|
|
|
|
// refresh the logs
|
2021-02-28 01:07:01 +00:00
|
|
|
if lintcheck_needs_rerun(&config.lintcheck_results_path) {
|
2021-02-22 23:40:50 +00:00
|
|
|
let shared_target_dir = "target/lintcheck/shared_target_dir";
|
2021-02-23 11:58:12 +00:00
|
|
|
match std::fs::metadata(&shared_target_dir) {
|
|
|
|
Ok(metadata) => {
|
|
|
|
if metadata.is_dir() {
|
|
|
|
println!("Clippy is newer than lint check logs, clearing lintcheck shared target dir...");
|
|
|
|
std::fs::remove_dir_all(&shared_target_dir)
|
|
|
|
.expect("failed to remove target/lintcheck/shared_target_dir");
|
|
|
|
}
|
|
|
|
},
|
2021-02-28 01:07:01 +00:00
|
|
|
Err(_) => { /* dir probably does not exist, don't remove anything */ },
|
2021-02-22 23:40:50 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-27 11:05:27 +00:00
|
|
|
let cargo_clippy_path: PathBuf = PathBuf::from(CARGO_CLIPPY_PATH)
|
2021-02-16 15:58:00 +00:00
|
|
|
.canonicalize()
|
|
|
|
.expect("failed to canonicalize path to clippy binary");
|
|
|
|
|
2020-12-18 12:21:13 +00:00
|
|
|
// assert that clippy is found
|
|
|
|
assert!(
|
|
|
|
cargo_clippy_path.is_file(),
|
2020-12-18 15:53:18 +00:00
|
|
|
"target/debug/cargo-clippy binary not found! {}",
|
|
|
|
cargo_clippy_path.display()
|
2020-12-18 12:21:13 +00:00
|
|
|
);
|
|
|
|
|
2021-02-27 11:05:27 +00:00
|
|
|
let clippy_ver = std::process::Command::new(CARGO_CLIPPY_PATH)
|
2020-12-29 15:18:31 +00:00
|
|
|
.arg("--version")
|
|
|
|
.output()
|
|
|
|
.map(|o| String::from_utf8_lossy(&o.stdout).into_owned())
|
|
|
|
.expect("could not get clippy version!");
|
|
|
|
|
2020-12-18 13:14:15 +00:00
|
|
|
// download and extract the crates, then run clippy on them and collect clippys warnings
|
2020-12-23 14:31:18 +00:00
|
|
|
// flatten into one big list of warnings
|
2020-12-22 12:07:55 +00:00
|
|
|
|
2021-02-26 23:29:42 +00:00
|
|
|
let crates = read_crates(&config.sources_toml_path);
|
|
|
|
let old_stats = read_stats_from_file(&config.lintcheck_results_path);
|
2020-12-27 15:20:32 +00:00
|
|
|
|
2021-02-28 01:07:01 +00:00
|
|
|
let counter = AtomicUsize::new(1);
|
|
|
|
|
2020-12-27 15:13:42 +00:00
|
|
|
let clippy_warnings: Vec<ClippyWarning> = if let Some(only_one_crate) = clap_config.value_of("only") {
|
2021-02-06 11:04:31 +00:00
|
|
|
// if we don't have the specified crate in the .toml, throw an error
|
|
|
|
if !crates.iter().any(|krate| {
|
|
|
|
let name = match krate {
|
|
|
|
CrateSource::CratesIo { name, .. } => name,
|
|
|
|
CrateSource::Git { name, .. } => name,
|
2021-02-10 10:32:10 +00:00
|
|
|
CrateSource::Path { name, .. } => name,
|
2021-02-06 11:04:31 +00:00
|
|
|
};
|
|
|
|
name == only_one_crate
|
|
|
|
}) {
|
2020-12-27 15:20:32 +00:00
|
|
|
eprintln!(
|
2021-01-22 23:25:29 +00:00
|
|
|
"ERROR: could not find crate '{}' in clippy_dev/lintcheck_crates.toml",
|
2020-12-27 15:20:32 +00:00
|
|
|
only_one_crate
|
|
|
|
);
|
|
|
|
std::process::exit(1);
|
2021-02-06 11:04:31 +00:00
|
|
|
}
|
2020-12-27 15:20:32 +00:00
|
|
|
|
|
|
|
// only check a single crate that was passed via cmdline
|
|
|
|
crates
|
2020-12-27 15:13:42 +00:00
|
|
|
.into_iter()
|
|
|
|
.map(|krate| krate.download_and_extract())
|
|
|
|
.filter(|krate| krate.name == only_one_crate)
|
2021-02-19 21:06:50 +00:00
|
|
|
.map(|krate| krate.run_clippy_lints(&cargo_clippy_path, &AtomicUsize::new(0), 1, 1))
|
2020-12-27 15:13:42 +00:00
|
|
|
.flatten()
|
|
|
|
.collect()
|
|
|
|
} else {
|
2021-02-27 00:34:45 +00:00
|
|
|
if config.max_jobs > 1 {
|
|
|
|
// run parallel with rayon
|
|
|
|
|
|
|
|
// Ask rayon for thread count. Assume that half of that is the number of physical cores
|
|
|
|
// Use one target dir for each core so that we can run N clippys in parallel.
|
|
|
|
// We need to use different target dirs because cargo would lock them for a single build otherwise,
|
|
|
|
// killing the parallelism. However this also means that deps will only be reused half/a
|
|
|
|
// quarter of the time which might result in a longer wall clock runtime
|
|
|
|
|
|
|
|
// This helps when we check many small crates with dep-trees that don't have a lot of branches in
|
|
|
|
// order to achive some kind of parallelism
|
|
|
|
|
|
|
|
// by default, use a single thread
|
|
|
|
let num_cpus = config.max_jobs;
|
|
|
|
let num_crates = crates.len();
|
|
|
|
|
|
|
|
// check all crates (default)
|
|
|
|
crates
|
|
|
|
.into_par_iter()
|
|
|
|
.map(|krate| krate.download_and_extract())
|
|
|
|
.map(|krate| krate.run_clippy_lints(&cargo_clippy_path, &counter, num_cpus, num_crates))
|
|
|
|
.flatten()
|
|
|
|
.collect()
|
|
|
|
} else {
|
|
|
|
// run sequential
|
|
|
|
let num_crates = crates.len();
|
|
|
|
crates
|
|
|
|
.into_iter()
|
|
|
|
.map(|krate| krate.download_and_extract())
|
2021-02-28 01:07:01 +00:00
|
|
|
.map(|krate| krate.run_clippy_lints(&cargo_clippy_path, &counter, 1, num_crates))
|
2021-02-27 00:34:45 +00:00
|
|
|
.flatten()
|
|
|
|
.collect()
|
|
|
|
}
|
2020-12-27 15:13:42 +00:00
|
|
|
};
|
2020-12-18 17:34:09 +00:00
|
|
|
|
2021-02-14 17:59:32 +00:00
|
|
|
// generate some stats
|
2021-02-23 20:23:36 +00:00
|
|
|
let (stats_formatted, new_stats) = gather_stats(&clippy_warnings);
|
2020-12-23 14:59:16 +00:00
|
|
|
|
2021-02-10 11:50:36 +00:00
|
|
|
// grab crashes/ICEs, save the crate name and the ice message
|
|
|
|
let ices: Vec<(&String, &String)> = clippy_warnings
|
|
|
|
.iter()
|
2021-02-14 17:37:08 +00:00
|
|
|
.filter(|warning| warning.is_ice)
|
2021-02-10 11:50:36 +00:00
|
|
|
.map(|w| (&w.crate_name, &w.message))
|
|
|
|
.collect();
|
|
|
|
|
2020-12-23 14:59:16 +00:00
|
|
|
let mut all_msgs: Vec<String> = clippy_warnings.iter().map(|warning| warning.to_string()).collect();
|
|
|
|
all_msgs.sort();
|
2021-02-23 20:23:36 +00:00
|
|
|
all_msgs.push("\n\n\n\nStats:\n".into());
|
2020-12-23 14:59:16 +00:00
|
|
|
all_msgs.push(stats_formatted);
|
2020-12-18 17:34:09 +00:00
|
|
|
|
2021-01-22 23:25:29 +00:00
|
|
|
// save the text into lintcheck-logs/logs.txt
|
2020-12-29 15:18:31 +00:00
|
|
|
let mut text = clippy_ver; // clippy version number on top
|
|
|
|
text.push_str(&format!("\n{}", all_msgs.join("")));
|
2021-02-10 11:50:36 +00:00
|
|
|
text.push_str("ICEs:\n");
|
|
|
|
ices.iter()
|
|
|
|
.for_each(|(cratename, msg)| text.push_str(&format!("{}: '{}'", cratename, msg)));
|
|
|
|
|
2021-02-26 23:29:42 +00:00
|
|
|
println!("Writing logs to {}", config.lintcheck_results_path.display());
|
|
|
|
write(&config.lintcheck_results_path, text).unwrap();
|
2021-02-23 20:23:36 +00:00
|
|
|
|
|
|
|
print_stats(old_stats, new_stats);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// read the previous stats from the lintcheck-log file
|
2021-02-26 23:29:42 +00:00
|
|
|
fn read_stats_from_file(file_path: &PathBuf) -> HashMap<String, usize> {
|
2021-02-23 20:23:36 +00:00
|
|
|
let file_content: String = match std::fs::read_to_string(file_path).ok() {
|
|
|
|
Some(content) => content,
|
|
|
|
None => {
|
|
|
|
eprintln!("RETURND");
|
|
|
|
return HashMap::new();
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
let lines: Vec<String> = file_content.lines().map(|l| l.to_string()).collect();
|
|
|
|
|
|
|
|
// search for the beginning "Stats:" and the end "ICEs:" of the section we want
|
|
|
|
let start = lines.iter().position(|line| line == "Stats:").unwrap();
|
|
|
|
let end = lines.iter().position(|line| line == "ICEs:").unwrap();
|
|
|
|
|
|
|
|
let stats_lines = &lines[start + 1..=end - 1];
|
|
|
|
|
|
|
|
stats_lines
|
|
|
|
.into_iter()
|
|
|
|
.map(|line| {
|
|
|
|
let mut spl = line.split(" ").into_iter();
|
|
|
|
(
|
|
|
|
spl.next().unwrap().to_string(),
|
|
|
|
spl.next().unwrap().parse::<usize>().unwrap(),
|
|
|
|
)
|
|
|
|
})
|
|
|
|
.collect::<HashMap<String, usize>>()
|
|
|
|
}
|
|
|
|
|
|
|
|
/// print how lint counts changed between runs
|
|
|
|
fn print_stats(old_stats: HashMap<String, usize>, new_stats: HashMap<&String, usize>) {
|
|
|
|
let same_in_both_hashmaps = old_stats
|
|
|
|
.iter()
|
|
|
|
.filter(|(old_key, old_val)| new_stats.get::<&String>(&old_key) == Some(old_val))
|
|
|
|
.map(|(k, v)| (k.to_string(), *v))
|
|
|
|
.collect::<Vec<(String, usize)>>();
|
|
|
|
|
|
|
|
let mut old_stats_deduped = old_stats;
|
|
|
|
let mut new_stats_deduped = new_stats;
|
|
|
|
|
|
|
|
// remove duplicates from both hashmaps
|
|
|
|
same_in_both_hashmaps.iter().for_each(|(k, v)| {
|
|
|
|
assert!(old_stats_deduped.remove(k) == Some(*v));
|
|
|
|
assert!(new_stats_deduped.remove(k) == Some(*v));
|
|
|
|
});
|
|
|
|
|
|
|
|
println!("\nStats:");
|
|
|
|
|
|
|
|
// list all new counts (key is in new stats but not in old stats)
|
|
|
|
new_stats_deduped
|
|
|
|
.iter()
|
|
|
|
.filter(|(new_key, _)| old_stats_deduped.get::<str>(&new_key).is_none())
|
|
|
|
.for_each(|(new_key, new_value)| {
|
|
|
|
println!("{} 0 => {}", new_key, new_value);
|
|
|
|
});
|
|
|
|
|
|
|
|
// list all changed counts (key is in both maps but value differs)
|
|
|
|
new_stats_deduped
|
|
|
|
.iter()
|
|
|
|
.filter(|(new_key, _new_val)| old_stats_deduped.get::<str>(&new_key).is_some())
|
|
|
|
.for_each(|(new_key, new_val)| {
|
|
|
|
let old_val = old_stats_deduped.get::<str>(&new_key).unwrap();
|
|
|
|
println!("{} {} => {}", new_key, old_val, new_val);
|
|
|
|
});
|
|
|
|
|
|
|
|
// list all gone counts (key is in old status but not in new stats)
|
|
|
|
old_stats_deduped
|
|
|
|
.iter()
|
|
|
|
.filter(|(old_key, _)| new_stats_deduped.get::<&String>(&old_key).is_none())
|
|
|
|
.for_each(|(old_key, old_value)| {
|
|
|
|
println!("{} {} => 0", old_key, old_value);
|
|
|
|
});
|
2020-12-18 12:21:13 +00:00
|
|
|
}
|