use crate::utils::span_lint; use itertools::Itertools; use pulldown_cmark; use rustc::lint::{EarlyContext, EarlyLintPass, LintArray, LintPass}; use rustc::{declare_tool_lint, lint_array}; use rustc_data_structures::fx::FxHashSet; use syntax::ast; use syntax::source_map::{BytePos, Span}; use syntax_pos::Pos; use url::Url; /// **What it does:** Checks for the presence of `_`, `::` or camel-case words /// outside ticks in documentation. /// /// **Why is this bad?** *Rustdoc* supports markdown formatting, `_`, `::` and /// camel-case probably indicates some code which should be included between /// ticks. `_` can also be used for emphasis in markdown, this lint tries to /// consider that. /// /// **Known problems:** Lots of bad docs won’t be fixed, what the lint checks /// for is limited, and there are still false positives. /// /// **Examples:** /// ```rust /// /// Do something with the foo_bar parameter. See also /// /// that::other::module::foo. /// // ^ `foo_bar` and `that::other::module::foo` should be ticked. /// fn doit(foo_bar) { .. } /// ``` declare_clippy_lint! { pub DOC_MARKDOWN, pedantic, "presence of `_`, `::` or camel-case outside backticks in documentation" } #[derive(Clone)] pub struct Doc { valid_idents: FxHashSet, } impl Doc { pub fn new(valid_idents: FxHashSet) -> Self { Self { valid_idents } } } impl LintPass for Doc { fn get_lints(&self) -> LintArray { lint_array![DOC_MARKDOWN] } } impl EarlyLintPass for Doc { fn check_crate(&mut self, cx: &EarlyContext<'_>, krate: &ast::Crate) { check_attrs(cx, &self.valid_idents, &krate.attrs); } fn check_item(&mut self, cx: &EarlyContext<'_>, item: &ast::Item) { check_attrs(cx, &self.valid_idents, &item.attrs); } } struct Parser<'a> { parser: pulldown_cmark::Parser<'a>, } impl<'a> Parser<'a> { fn new(parser: pulldown_cmark::Parser<'a>) -> Self { Self { parser } } } impl<'a> Iterator for Parser<'a> { type Item = (usize, pulldown_cmark::Event<'a>); fn next(&mut self) -> Option { let offset = self.parser.get_offset(); self.parser.next().map(|event| (offset, event)) } } /// Cleanup documentation decoration (`///` and such). /// /// We can't use `syntax::attr::AttributeMethods::with_desugared_doc` or /// `syntax::parse::lexer::comments::strip_doc_comment_decoration` because we /// need to keep track of /// the spans but this function is inspired from the later. #[allow(clippy::cast_possible_truncation)] pub fn strip_doc_comment_decoration(comment: &str, span: Span) -> (String, Vec<(usize, Span)>) { // one-line comments lose their prefix const ONELINERS: &[&str] = &["///!", "///", "//!", "//"]; for prefix in ONELINERS { if comment.starts_with(*prefix) { let doc = &comment[prefix.len()..]; let mut doc = doc.to_owned(); doc.push('\n'); return ( doc.to_owned(), vec![(doc.len(), span.with_lo(span.lo() + BytePos(prefix.len() as u32)))], ); } } if comment.starts_with("/*") { let doc = &comment[3..comment.len() - 2]; let mut sizes = vec![]; let mut contains_initial_stars = false; for line in doc.lines() { let offset = line.as_ptr() as usize - comment.as_ptr() as usize; debug_assert_eq!(offset as u32 as usize, offset); contains_initial_stars |= line.trim_start().starts_with('*'); // +1 for the newline sizes.push((line.len() + 1, span.with_lo(span.lo() + BytePos(offset as u32)))); } if !contains_initial_stars { return (doc.to_string(), sizes); } // remove the initial '*'s if any let mut no_stars = String::with_capacity(doc.len()); for line in doc.lines() { let mut chars = line.chars(); while let Some(c) = chars.next() { if c.is_whitespace() { no_stars.push(c); } else { no_stars.push(if c == '*' { ' ' } else { c }); break; } } no_stars.push_str(chars.as_str()); no_stars.push('\n'); } return (no_stars, sizes); } panic!("not a doc-comment: {}", comment); } pub fn check_attrs<'a>(cx: &EarlyContext<'_>, valid_idents: &FxHashSet, attrs: &'a [ast::Attribute]) { let mut doc = String::new(); let mut spans = vec![]; for attr in attrs { if attr.is_sugared_doc { if let Some(ref current) = attr.value_str() { let current = current.to_string(); let (current, current_spans) = strip_doc_comment_decoration(¤t, attr.span); spans.extend_from_slice(¤t_spans); doc.push_str(¤t); } } else if attr.name() == "doc" { // ignore mix of sugared and non-sugared doc return; } } let mut current = 0; for &mut (ref mut offset, _) in &mut spans { let offset_copy = *offset; *offset = current; current += offset_copy; } if !doc.is_empty() { let parser = Parser::new(pulldown_cmark::Parser::new(&doc)); let parser = parser.coalesce(|x, y| { use pulldown_cmark::Event::*; let x_offset = x.0; let y_offset = y.0; match (x.1, y.1) { (Text(x), Text(y)) => { let mut x = x.into_owned(); x.push_str(&y); Ok((x_offset, Text(x.into()))) }, (x, y) => Err(((x_offset, x), (y_offset, y))), } }); check_doc(cx, valid_idents, parser, &spans); } } fn check_doc<'a, Events: Iterator)>>( cx: &EarlyContext<'_>, valid_idents: &FxHashSet, docs: Events, spans: &[(usize, Span)], ) { use pulldown_cmark::Event::*; use pulldown_cmark::Tag::*; let mut in_code = false; let mut in_link = None; for (offset, event) in docs { match event { Start(CodeBlock(_)) | Start(Code) => in_code = true, End(CodeBlock(_)) | End(Code) => in_code = false, Start(Link(link, _)) => in_link = Some(link), End(Link(_, _)) => in_link = None, Start(_tag) | End(_tag) => (), // We don't care about other tags Html(_html) | InlineHtml(_html) => (), // HTML is weird, just ignore it SoftBreak | HardBreak => (), FootnoteReference(text) | Text(text) => { if Some(&text) == in_link.as_ref() { // Probably a link of the form `` // Which are represented as a link to "http://example.com" with // text "http://example.com" by pulldown-cmark continue; } if !in_code { let index = match spans.binary_search_by(|c| c.0.cmp(&offset)) { Ok(o) => o, Err(e) => e - 1, }; let (begin, span) = spans[index]; // Adjust for the beginning of the current `Event` let span = span.with_lo(span.lo() + BytePos::from_usize(offset - begin)); check_text(cx, valid_idents, &text, span); } }, } } } fn check_text(cx: &EarlyContext<'_>, valid_idents: &FxHashSet, text: &str, span: Span) { for word in text.split(|c: char| c.is_whitespace() || c == '\'') { // Trim punctuation as in `some comment (see foo::bar).` // ^^ // Or even as in `_foo bar_` which is emphasized. let word = word.trim_matches(|c: char| !c.is_alphanumeric()); if valid_idents.contains(word) { continue; } // Adjust for the current word let offset = word.as_ptr() as usize - text.as_ptr() as usize; let span = Span::new( span.lo() + BytePos::from_usize(offset), span.lo() + BytePos::from_usize(offset + word.len()), span.ctxt(), ); check_word(cx, word, span); } } fn check_word(cx: &EarlyContext<'_>, word: &str, span: Span) { /// Checks if a string is camel-case, ie. contains at least two uppercase /// letter (`Clippy` is /// ok) and one lower-case letter (`NASA` is ok). Plural are also excluded /// (`IDs` is ok). fn is_camel_case(s: &str) -> bool { if s.starts_with(|c: char| c.is_digit(10)) { return false; } let s = if s.ends_with('s') { &s[..s.len() - 1] } else { s }; s.chars().all(char::is_alphanumeric) && s.chars().filter(|&c| c.is_uppercase()).take(2).count() > 1 && s.chars().filter(|&c| c.is_lowercase()).take(1).count() > 0 } fn has_underscore(s: &str) -> bool { s != "_" && !s.contains("\\_") && s.contains('_') } fn has_hyphen(s: &str) -> bool { s != "-" && s.contains('-') } if let Ok(url) = Url::parse(word) { // try to get around the fact that `foo::bar` parses as a valid URL if !url.cannot_be_a_base() { span_lint( cx, DOC_MARKDOWN, span, "you should put bare URLs between `<`/`>` or make a proper Markdown link", ); return; } } // We assume that mixed-case words are not meant to be put inside bacticks. (Issue #2343) if has_underscore(word) && has_hyphen(word) { return; } if has_underscore(word) || word.contains("::") || is_camel_case(word) { span_lint( cx, DOC_MARKDOWN, span, &format!("you should put `{}` between ticks in the documentation", word), ); } }