mirror of
https://github.com/rust-lang/rust-clippy
synced 2024-11-28 07:30:57 +00:00
870 lines
31 KiB
Rust
870 lines
31 KiB
Rust
use clippy_utils::attrs::is_doc_hidden;
|
||
use clippy_utils::diagnostics::{span_lint, span_lint_and_help, span_lint_and_note, span_lint_and_then};
|
||
use clippy_utils::macros::{is_panic, root_macro_call_first_node};
|
||
use clippy_utils::source::snippet_with_applicability;
|
||
use clippy_utils::ty::{implements_trait, is_type_diagnostic_item};
|
||
use clippy_utils::{is_entrypoint_fn, method_chain_args, return_ty};
|
||
use if_chain::if_chain;
|
||
use pulldown_cmark::Event::{
|
||
Code, End, FootnoteReference, HardBreak, Html, Rule, SoftBreak, Start, TaskListMarker, Text,
|
||
};
|
||
use pulldown_cmark::Tag::{CodeBlock, Heading, Item, Link, Paragraph};
|
||
use pulldown_cmark::{BrokenLink, CodeBlockKind, CowStr, Options};
|
||
use rustc_ast::ast::{Async, Attribute, Fn, FnRetTy, ItemKind};
|
||
use rustc_data_structures::fx::FxHashSet;
|
||
use rustc_data_structures::sync::Lrc;
|
||
use rustc_errors::emitter::EmitterWriter;
|
||
use rustc_errors::{Applicability, Handler, SuggestionStyle};
|
||
use rustc_hir as hir;
|
||
use rustc_hir::intravisit::{self, Visitor};
|
||
use rustc_hir::{AnonConst, Expr};
|
||
use rustc_lint::{LateContext, LateLintPass};
|
||
use rustc_middle::hir::nested_filter;
|
||
use rustc_middle::lint::in_external_macro;
|
||
use rustc_middle::ty;
|
||
use rustc_parse::maybe_new_parser_from_source_str;
|
||
use rustc_parse::parser::ForceCollect;
|
||
use rustc_resolve::rustdoc::{
|
||
add_doc_fragment, attrs_to_doc_fragments, main_body_opts, source_span_for_markdown_range, DocFragment,
|
||
};
|
||
use rustc_session::parse::ParseSess;
|
||
use rustc_session::{declare_tool_lint, impl_lint_pass};
|
||
use rustc_span::edition::Edition;
|
||
use rustc_span::source_map::{BytePos, FilePathMapping, SourceMap, Span};
|
||
use rustc_span::{sym, FileName, Pos};
|
||
use std::ops::Range;
|
||
use std::{io, thread};
|
||
use url::Url;
|
||
|
||
declare_clippy_lint! {
|
||
/// ### What it does
|
||
/// Checks for the presence of `_`, `::` or camel-case words
|
||
/// outside ticks in documentation.
|
||
///
|
||
/// ### Why is this bad?
|
||
/// *Rustdoc* supports markdown formatting, `_`, `::` and
|
||
/// camel-case probably indicates some code which should be included between
|
||
/// ticks. `_` can also be used for emphasis in markdown, this lint tries to
|
||
/// consider that.
|
||
///
|
||
/// ### Known problems
|
||
/// Lots of bad docs won’t be fixed, what the lint checks
|
||
/// for is limited, and there are still false positives. HTML elements and their
|
||
/// content are not linted.
|
||
///
|
||
/// In addition, when writing documentation comments, including `[]` brackets
|
||
/// inside a link text would trip the parser. Therefore, documenting link with
|
||
/// `[`SmallVec<[T; INLINE_CAPACITY]>`]` and then [`SmallVec<[T; INLINE_CAPACITY]>`]: SmallVec
|
||
/// would fail.
|
||
///
|
||
/// ### Examples
|
||
/// ```rust
|
||
/// /// Do something with the foo_bar parameter. See also
|
||
/// /// that::other::module::foo.
|
||
/// // ^ `foo_bar` and `that::other::module::foo` should be ticked.
|
||
/// fn doit(foo_bar: usize) {}
|
||
/// ```
|
||
///
|
||
/// ```rust
|
||
/// // Link text with `[]` brackets should be written as following:
|
||
/// /// Consume the array and return the inner
|
||
/// /// [`SmallVec<[T; INLINE_CAPACITY]>`][SmallVec].
|
||
/// /// [SmallVec]: SmallVec
|
||
/// fn main() {}
|
||
/// ```
|
||
#[clippy::version = "pre 1.29.0"]
|
||
pub DOC_MARKDOWN,
|
||
pedantic,
|
||
"presence of `_`, `::` or camel-case outside backticks in documentation"
|
||
}
|
||
|
||
declare_clippy_lint! {
|
||
/// ### What it does
|
||
/// Checks for the doc comments of publicly visible
|
||
/// unsafe functions and warns if there is no `# Safety` section.
|
||
///
|
||
/// ### Why is this bad?
|
||
/// Unsafe functions should document their safety
|
||
/// preconditions, so that users can be sure they are using them safely.
|
||
///
|
||
/// ### Examples
|
||
/// ```rust
|
||
///# type Universe = ();
|
||
/// /// This function should really be documented
|
||
/// pub unsafe fn start_apocalypse(u: &mut Universe) {
|
||
/// unimplemented!();
|
||
/// }
|
||
/// ```
|
||
///
|
||
/// At least write a line about safety:
|
||
///
|
||
/// ```rust
|
||
///# type Universe = ();
|
||
/// /// # Safety
|
||
/// ///
|
||
/// /// This function should not be called before the horsemen are ready.
|
||
/// pub unsafe fn start_apocalypse(u: &mut Universe) {
|
||
/// unimplemented!();
|
||
/// }
|
||
/// ```
|
||
#[clippy::version = "1.39.0"]
|
||
pub MISSING_SAFETY_DOC,
|
||
style,
|
||
"`pub unsafe fn` without `# Safety` docs"
|
||
}
|
||
|
||
declare_clippy_lint! {
|
||
/// ### What it does
|
||
/// Checks the doc comments of publicly visible functions that
|
||
/// return a `Result` type and warns if there is no `# Errors` section.
|
||
///
|
||
/// ### Why is this bad?
|
||
/// Documenting the type of errors that can be returned from a
|
||
/// function can help callers write code to handle the errors appropriately.
|
||
///
|
||
/// ### Examples
|
||
/// Since the following function returns a `Result` it has an `# Errors` section in
|
||
/// its doc comment:
|
||
///
|
||
/// ```rust
|
||
///# use std::io;
|
||
/// /// # Errors
|
||
/// ///
|
||
/// /// Will return `Err` if `filename` does not exist or the user does not have
|
||
/// /// permission to read it.
|
||
/// pub fn read(filename: String) -> io::Result<String> {
|
||
/// unimplemented!();
|
||
/// }
|
||
/// ```
|
||
#[clippy::version = "1.41.0"]
|
||
pub MISSING_ERRORS_DOC,
|
||
pedantic,
|
||
"`pub fn` returns `Result` without `# Errors` in doc comment"
|
||
}
|
||
|
||
declare_clippy_lint! {
|
||
/// ### What it does
|
||
/// Checks the doc comments of publicly visible functions that
|
||
/// may panic and warns if there is no `# Panics` section.
|
||
///
|
||
/// ### Why is this bad?
|
||
/// Documenting the scenarios in which panicking occurs
|
||
/// can help callers who do not want to panic to avoid those situations.
|
||
///
|
||
/// ### Examples
|
||
/// Since the following function may panic it has a `# Panics` section in
|
||
/// its doc comment:
|
||
///
|
||
/// ```rust
|
||
/// /// # Panics
|
||
/// ///
|
||
/// /// Will panic if y is 0
|
||
/// pub fn divide_by(x: i32, y: i32) -> i32 {
|
||
/// if y == 0 {
|
||
/// panic!("Cannot divide by 0")
|
||
/// } else {
|
||
/// x / y
|
||
/// }
|
||
/// }
|
||
/// ```
|
||
#[clippy::version = "1.51.0"]
|
||
pub MISSING_PANICS_DOC,
|
||
pedantic,
|
||
"`pub fn` may panic without `# Panics` in doc comment"
|
||
}
|
||
|
||
declare_clippy_lint! {
|
||
/// ### What it does
|
||
/// Checks for `fn main() { .. }` in doctests
|
||
///
|
||
/// ### Why is this bad?
|
||
/// The test can be shorter (and likely more readable)
|
||
/// if the `fn main()` is left implicit.
|
||
///
|
||
/// ### Examples
|
||
/// ```rust
|
||
/// /// An example of a doctest with a `main()` function
|
||
/// ///
|
||
/// /// # Examples
|
||
/// ///
|
||
/// /// ```
|
||
/// /// fn main() {
|
||
/// /// // this needs not be in an `fn`
|
||
/// /// }
|
||
/// /// ```
|
||
/// fn needless_main() {
|
||
/// unimplemented!();
|
||
/// }
|
||
/// ```
|
||
#[clippy::version = "1.40.0"]
|
||
pub NEEDLESS_DOCTEST_MAIN,
|
||
style,
|
||
"presence of `fn main() {` in code examples"
|
||
}
|
||
|
||
declare_clippy_lint! {
|
||
/// ### What it does
|
||
/// Detects the syntax `['foo']` in documentation comments (notice quotes instead of backticks)
|
||
/// outside of code blocks
|
||
/// ### Why is this bad?
|
||
/// It is likely a typo when defining an intra-doc link
|
||
///
|
||
/// ### Example
|
||
/// ```rust
|
||
/// /// See also: ['foo']
|
||
/// fn bar() {}
|
||
/// ```
|
||
/// Use instead:
|
||
/// ```rust
|
||
/// /// See also: [`foo`]
|
||
/// fn bar() {}
|
||
/// ```
|
||
#[clippy::version = "1.63.0"]
|
||
pub DOC_LINK_WITH_QUOTES,
|
||
pedantic,
|
||
"possible typo for an intra-doc link"
|
||
}
|
||
|
||
declare_clippy_lint! {
|
||
/// ### What it does
|
||
/// Checks for the doc comments of publicly visible
|
||
/// safe functions and traits and warns if there is a `# Safety` section.
|
||
///
|
||
/// ### Why is this bad?
|
||
/// Safe functions and traits are safe to implement and therefore do not
|
||
/// need to describe safety preconditions that users are required to uphold.
|
||
///
|
||
/// ### Examples
|
||
/// ```rust
|
||
///# type Universe = ();
|
||
/// /// # Safety
|
||
/// ///
|
||
/// /// This function should not be called before the horsemen are ready.
|
||
/// pub fn start_apocalypse_but_safely(u: &mut Universe) {
|
||
/// unimplemented!();
|
||
/// }
|
||
/// ```
|
||
///
|
||
/// The function is safe, so there shouldn't be any preconditions
|
||
/// that have to be explained for safety reasons.
|
||
///
|
||
/// ```rust
|
||
///# type Universe = ();
|
||
/// /// This function should really be documented
|
||
/// pub fn start_apocalypse(u: &mut Universe) {
|
||
/// unimplemented!();
|
||
/// }
|
||
/// ```
|
||
#[clippy::version = "1.67.0"]
|
||
pub UNNECESSARY_SAFETY_DOC,
|
||
restriction,
|
||
"`pub fn` or `pub trait` with `# Safety` docs"
|
||
}
|
||
|
||
#[expect(clippy::module_name_repetitions)]
|
||
#[derive(Clone)]
|
||
pub struct DocMarkdown {
|
||
valid_idents: FxHashSet<String>,
|
||
in_trait_impl: bool,
|
||
}
|
||
|
||
impl DocMarkdown {
|
||
pub fn new(valid_idents: FxHashSet<String>) -> Self {
|
||
Self {
|
||
valid_idents,
|
||
in_trait_impl: false,
|
||
}
|
||
}
|
||
}
|
||
|
||
impl_lint_pass!(DocMarkdown => [
|
||
DOC_LINK_WITH_QUOTES,
|
||
DOC_MARKDOWN,
|
||
MISSING_SAFETY_DOC,
|
||
MISSING_ERRORS_DOC,
|
||
MISSING_PANICS_DOC,
|
||
NEEDLESS_DOCTEST_MAIN,
|
||
UNNECESSARY_SAFETY_DOC,
|
||
]);
|
||
|
||
impl<'tcx> LateLintPass<'tcx> for DocMarkdown {
|
||
fn check_crate(&mut self, cx: &LateContext<'tcx>) {
|
||
let attrs = cx.tcx.hir().attrs(hir::CRATE_HIR_ID);
|
||
check_attrs(cx, &self.valid_idents, attrs);
|
||
}
|
||
|
||
fn check_item(&mut self, cx: &LateContext<'tcx>, item: &'tcx hir::Item<'_>) {
|
||
let attrs = cx.tcx.hir().attrs(item.hir_id());
|
||
let Some(headers) = check_attrs(cx, &self.valid_idents, attrs) else {
|
||
return;
|
||
};
|
||
match item.kind {
|
||
hir::ItemKind::Fn(ref sig, _, body_id) => {
|
||
if !(is_entrypoint_fn(cx, item.owner_id.to_def_id()) || in_external_macro(cx.tcx.sess, item.span)) {
|
||
let body = cx.tcx.hir().body(body_id);
|
||
let mut fpu = FindPanicUnwrap {
|
||
cx,
|
||
typeck_results: cx.tcx.typeck(item.owner_id.def_id),
|
||
panic_span: None,
|
||
};
|
||
fpu.visit_expr(body.value);
|
||
lint_for_missing_headers(cx, item.owner_id, sig, headers, Some(body_id), fpu.panic_span);
|
||
}
|
||
},
|
||
hir::ItemKind::Impl(impl_) => {
|
||
self.in_trait_impl = impl_.of_trait.is_some();
|
||
},
|
||
hir::ItemKind::Trait(_, unsafety, ..) => match (headers.safety, unsafety) {
|
||
(false, hir::Unsafety::Unsafe) => span_lint(
|
||
cx,
|
||
MISSING_SAFETY_DOC,
|
||
cx.tcx.def_span(item.owner_id),
|
||
"docs for unsafe trait missing `# Safety` section",
|
||
),
|
||
(true, hir::Unsafety::Normal) => span_lint(
|
||
cx,
|
||
UNNECESSARY_SAFETY_DOC,
|
||
cx.tcx.def_span(item.owner_id),
|
||
"docs for safe trait have unnecessary `# Safety` section",
|
||
),
|
||
_ => (),
|
||
},
|
||
_ => (),
|
||
}
|
||
}
|
||
|
||
fn check_item_post(&mut self, _cx: &LateContext<'tcx>, item: &'tcx hir::Item<'_>) {
|
||
if let hir::ItemKind::Impl { .. } = item.kind {
|
||
self.in_trait_impl = false;
|
||
}
|
||
}
|
||
|
||
fn check_trait_item(&mut self, cx: &LateContext<'tcx>, item: &'tcx hir::TraitItem<'_>) {
|
||
let attrs = cx.tcx.hir().attrs(item.hir_id());
|
||
let Some(headers) = check_attrs(cx, &self.valid_idents, attrs) else {
|
||
return;
|
||
};
|
||
if let hir::TraitItemKind::Fn(ref sig, ..) = item.kind {
|
||
if !in_external_macro(cx.tcx.sess, item.span) {
|
||
lint_for_missing_headers(cx, item.owner_id, sig, headers, None, None);
|
||
}
|
||
}
|
||
}
|
||
|
||
fn check_impl_item(&mut self, cx: &LateContext<'tcx>, item: &'tcx hir::ImplItem<'_>) {
|
||
let attrs = cx.tcx.hir().attrs(item.hir_id());
|
||
let Some(headers) = check_attrs(cx, &self.valid_idents, attrs) else {
|
||
return;
|
||
};
|
||
if self.in_trait_impl || in_external_macro(cx.tcx.sess, item.span) {
|
||
return;
|
||
}
|
||
if let hir::ImplItemKind::Fn(ref sig, body_id) = item.kind {
|
||
let body = cx.tcx.hir().body(body_id);
|
||
let mut fpu = FindPanicUnwrap {
|
||
cx,
|
||
typeck_results: cx.tcx.typeck(item.owner_id.def_id),
|
||
panic_span: None,
|
||
};
|
||
fpu.visit_expr(body.value);
|
||
lint_for_missing_headers(cx, item.owner_id, sig, headers, Some(body_id), fpu.panic_span);
|
||
}
|
||
}
|
||
}
|
||
|
||
fn lint_for_missing_headers(
|
||
cx: &LateContext<'_>,
|
||
owner_id: hir::OwnerId,
|
||
sig: &hir::FnSig<'_>,
|
||
headers: DocHeaders,
|
||
body_id: Option<hir::BodyId>,
|
||
panic_span: Option<Span>,
|
||
) {
|
||
if !cx.effective_visibilities.is_exported(owner_id.def_id) {
|
||
return; // Private functions do not require doc comments
|
||
}
|
||
|
||
// do not lint if any parent has `#[doc(hidden)]` attribute (#7347)
|
||
if cx
|
||
.tcx
|
||
.hir()
|
||
.parent_iter(owner_id.into())
|
||
.any(|(id, _node)| is_doc_hidden(cx.tcx.hir().attrs(id)))
|
||
{
|
||
return;
|
||
}
|
||
|
||
let span = cx.tcx.def_span(owner_id);
|
||
match (headers.safety, sig.header.unsafety) {
|
||
(false, hir::Unsafety::Unsafe) => span_lint(
|
||
cx,
|
||
MISSING_SAFETY_DOC,
|
||
span,
|
||
"unsafe function's docs miss `# Safety` section",
|
||
),
|
||
(true, hir::Unsafety::Normal) => span_lint(
|
||
cx,
|
||
UNNECESSARY_SAFETY_DOC,
|
||
span,
|
||
"safe function's docs have unnecessary `# Safety` section",
|
||
),
|
||
_ => (),
|
||
}
|
||
if !headers.panics && panic_span.is_some() {
|
||
span_lint_and_note(
|
||
cx,
|
||
MISSING_PANICS_DOC,
|
||
span,
|
||
"docs for function which may panic missing `# Panics` section",
|
||
panic_span,
|
||
"first possible panic found here",
|
||
);
|
||
}
|
||
if !headers.errors {
|
||
if is_type_diagnostic_item(cx, return_ty(cx, owner_id), sym::Result) {
|
||
span_lint(
|
||
cx,
|
||
MISSING_ERRORS_DOC,
|
||
span,
|
||
"docs for function returning `Result` missing `# Errors` section",
|
||
);
|
||
} else {
|
||
if_chain! {
|
||
if let Some(body_id) = body_id;
|
||
if let Some(future) = cx.tcx.lang_items().future_trait();
|
||
let typeck = cx.tcx.typeck_body(body_id);
|
||
let body = cx.tcx.hir().body(body_id);
|
||
let ret_ty = typeck.expr_ty(body.value);
|
||
if implements_trait(cx, ret_ty, future, &[]);
|
||
if let ty::Generator(_, subs, _) = ret_ty.kind();
|
||
if is_type_diagnostic_item(cx, subs.as_generator().return_ty(), sym::Result);
|
||
then {
|
||
span_lint(
|
||
cx,
|
||
MISSING_ERRORS_DOC,
|
||
span,
|
||
"docs for function returning `Result` missing `# Errors` section",
|
||
);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
#[derive(Copy, Clone)]
|
||
struct Fragments<'a> {
|
||
doc: &'a str,
|
||
fragments: &'a [DocFragment],
|
||
}
|
||
|
||
impl Fragments<'_> {
|
||
fn span(self, cx: &LateContext<'_>, range: Range<usize>) -> Option<Span> {
|
||
source_span_for_markdown_range(cx.tcx, self.doc, &range, self.fragments)
|
||
}
|
||
}
|
||
|
||
#[derive(Copy, Clone, Default)]
|
||
struct DocHeaders {
|
||
safety: bool,
|
||
errors: bool,
|
||
panics: bool,
|
||
}
|
||
|
||
fn check_attrs(cx: &LateContext<'_>, valid_idents: &FxHashSet<String>, attrs: &[Attribute]) -> Option<DocHeaders> {
|
||
/// We don't want the parser to choke on intra doc links. Since we don't
|
||
/// actually care about rendering them, just pretend that all broken links
|
||
/// point to a fake address.
|
||
#[expect(clippy::unnecessary_wraps)] // we're following a type signature
|
||
fn fake_broken_link_callback<'a>(_: BrokenLink<'_>) -> Option<(CowStr<'a>, CowStr<'a>)> {
|
||
Some(("fake".into(), "fake".into()))
|
||
}
|
||
|
||
if is_doc_hidden(attrs) {
|
||
return None;
|
||
}
|
||
|
||
let (fragments, _) = attrs_to_doc_fragments(attrs.iter().map(|attr| (attr, None)), true);
|
||
let mut doc = String::new();
|
||
for fragment in &fragments {
|
||
add_doc_fragment(&mut doc, fragment);
|
||
}
|
||
doc.pop();
|
||
|
||
if doc.is_empty() {
|
||
return Some(DocHeaders::default());
|
||
}
|
||
|
||
let mut cb = fake_broken_link_callback;
|
||
|
||
// disable smart punctuation to pick up ['link'] more easily
|
||
let opts = main_body_opts() - Options::ENABLE_SMART_PUNCTUATION;
|
||
let parser = pulldown_cmark::Parser::new_with_broken_link_callback(&doc, opts, Some(&mut cb));
|
||
|
||
Some(check_doc(
|
||
cx,
|
||
valid_idents,
|
||
parser.into_offset_iter(),
|
||
Fragments {
|
||
fragments: &fragments,
|
||
doc: &doc,
|
||
},
|
||
))
|
||
}
|
||
|
||
const RUST_CODE: &[&str] = &["rust", "no_run", "should_panic", "compile_fail"];
|
||
|
||
#[allow(clippy::too_many_lines)] // Only a big match statement
|
||
fn check_doc<'a, Events: Iterator<Item = (pulldown_cmark::Event<'a>, Range<usize>)>>(
|
||
cx: &LateContext<'_>,
|
||
valid_idents: &FxHashSet<String>,
|
||
events: Events,
|
||
fragments: Fragments<'_>,
|
||
) -> DocHeaders {
|
||
// true if a safety header was found
|
||
let mut headers = DocHeaders::default();
|
||
let mut in_code = false;
|
||
let mut in_link = None;
|
||
let mut in_heading = false;
|
||
let mut is_rust = false;
|
||
let mut no_test = false;
|
||
let mut edition = None;
|
||
let mut ticks_unbalanced = false;
|
||
let mut text_to_check: Vec<(CowStr<'_>, Range<usize>)> = Vec::new();
|
||
let mut paragraph_range = 0..0;
|
||
for (event, range) in events {
|
||
match event {
|
||
Start(CodeBlock(ref kind)) => {
|
||
in_code = true;
|
||
if let CodeBlockKind::Fenced(lang) = kind {
|
||
for item in lang.split(',') {
|
||
if item == "ignore" {
|
||
is_rust = false;
|
||
break;
|
||
} else if item == "no_test" {
|
||
no_test = true;
|
||
}
|
||
if let Some(stripped) = item.strip_prefix("edition") {
|
||
is_rust = true;
|
||
edition = stripped.parse::<Edition>().ok();
|
||
} else if item.is_empty() || RUST_CODE.contains(&item) {
|
||
is_rust = true;
|
||
}
|
||
}
|
||
}
|
||
},
|
||
End(CodeBlock(_)) => {
|
||
in_code = false;
|
||
is_rust = false;
|
||
},
|
||
Start(Link(_, url, _)) => in_link = Some(url),
|
||
End(Link(..)) => in_link = None,
|
||
Start(Heading(_, _, _) | Paragraph | Item) => {
|
||
if let Start(Heading(_, _, _)) = event {
|
||
in_heading = true;
|
||
}
|
||
ticks_unbalanced = false;
|
||
paragraph_range = range;
|
||
},
|
||
End(Heading(_, _, _) | Paragraph | Item) => {
|
||
if let End(Heading(_, _, _)) = event {
|
||
in_heading = false;
|
||
}
|
||
if ticks_unbalanced
|
||
&& let Some(span) = fragments.span(cx, paragraph_range.clone())
|
||
{
|
||
span_lint_and_help(
|
||
cx,
|
||
DOC_MARKDOWN,
|
||
span,
|
||
"backticks are unbalanced",
|
||
None,
|
||
"a backtick may be missing a pair",
|
||
);
|
||
} else {
|
||
for (text, range) in text_to_check {
|
||
if let Some(span) = fragments.span(cx, range) {
|
||
check_text(cx, valid_idents, &text, span);
|
||
}
|
||
}
|
||
}
|
||
text_to_check = Vec::new();
|
||
},
|
||
Start(_tag) | End(_tag) => (), // We don't care about other tags
|
||
Html(_html) => (), // HTML is weird, just ignore it
|
||
SoftBreak | HardBreak | TaskListMarker(_) | Code(_) | Rule => (),
|
||
FootnoteReference(text) | Text(text) => {
|
||
paragraph_range.end = range.end;
|
||
ticks_unbalanced |= text.contains('`') && !in_code;
|
||
if Some(&text) == in_link.as_ref() || ticks_unbalanced {
|
||
// Probably a link of the form `<http://example.com>`
|
||
// Which are represented as a link to "http://example.com" with
|
||
// text "http://example.com" by pulldown-cmark
|
||
continue;
|
||
}
|
||
let trimmed_text = text.trim();
|
||
headers.safety |= in_heading && trimmed_text == "Safety";
|
||
headers.safety |= in_heading && trimmed_text == "Implementation safety";
|
||
headers.safety |= in_heading && trimmed_text == "Implementation Safety";
|
||
headers.errors |= in_heading && trimmed_text == "Errors";
|
||
headers.panics |= in_heading && trimmed_text == "Panics";
|
||
if in_code {
|
||
if is_rust && !no_test {
|
||
let edition = edition.unwrap_or_else(|| cx.tcx.sess.edition());
|
||
check_code(cx, &text, edition, range.clone(), fragments);
|
||
}
|
||
} else {
|
||
if in_link.is_some() {
|
||
check_link_quotes(cx, trimmed_text, range.clone(), fragments);
|
||
}
|
||
if let Some(link) = in_link.as_ref()
|
||
&& let Ok(url) = Url::parse(link)
|
||
&& (url.scheme() == "https" || url.scheme() == "http") {
|
||
// Don't check the text associated with external URLs
|
||
continue;
|
||
}
|
||
text_to_check.push((text, range));
|
||
}
|
||
},
|
||
}
|
||
}
|
||
headers
|
||
}
|
||
|
||
fn check_link_quotes(cx: &LateContext<'_>, trimmed_text: &str, range: Range<usize>, fragments: Fragments<'_>) {
|
||
if trimmed_text.starts_with('\'')
|
||
&& trimmed_text.ends_with('\'')
|
||
&& let Some(span) = fragments.span(cx, range)
|
||
{
|
||
span_lint(
|
||
cx,
|
||
DOC_LINK_WITH_QUOTES,
|
||
span,
|
||
"possible intra-doc link using quotes instead of backticks",
|
||
);
|
||
}
|
||
}
|
||
|
||
fn check_code(cx: &LateContext<'_>, text: &str, edition: Edition, range: Range<usize>, fragments: Fragments<'_>) {
|
||
fn has_needless_main(code: String, edition: Edition) -> bool {
|
||
rustc_driver::catch_fatal_errors(|| {
|
||
rustc_span::create_session_globals_then(edition, || {
|
||
let filename = FileName::anon_source_code(&code);
|
||
|
||
let sm = Lrc::new(SourceMap::new(FilePathMapping::empty()));
|
||
let fallback_bundle =
|
||
rustc_errors::fallback_fluent_bundle(rustc_driver::DEFAULT_LOCALE_RESOURCES.to_vec(), false);
|
||
let emitter = EmitterWriter::new(Box::new(io::sink()), fallback_bundle);
|
||
let handler = Handler::with_emitter(Box::new(emitter)).disable_warnings();
|
||
let sess = ParseSess::with_span_handler(handler, sm);
|
||
|
||
let mut parser = match maybe_new_parser_from_source_str(&sess, filename, code) {
|
||
Ok(p) => p,
|
||
Err(errs) => {
|
||
drop(errs);
|
||
return false;
|
||
},
|
||
};
|
||
|
||
let mut relevant_main_found = false;
|
||
loop {
|
||
match parser.parse_item(ForceCollect::No) {
|
||
Ok(Some(item)) => match &item.kind {
|
||
ItemKind::Fn(box Fn {
|
||
sig, body: Some(block), ..
|
||
}) if item.ident.name == sym::main => {
|
||
let is_async = matches!(sig.header.asyncness, Async::Yes { .. });
|
||
let returns_nothing = match &sig.decl.output {
|
||
FnRetTy::Default(..) => true,
|
||
FnRetTy::Ty(ty) if ty.kind.is_unit() => true,
|
||
FnRetTy::Ty(_) => false,
|
||
};
|
||
|
||
if returns_nothing && !is_async && !block.stmts.is_empty() {
|
||
// This main function should be linted, but only if there are no other functions
|
||
relevant_main_found = true;
|
||
} else {
|
||
// This main function should not be linted, we're done
|
||
return false;
|
||
}
|
||
},
|
||
// Tests with one of these items are ignored
|
||
ItemKind::Static(..)
|
||
| ItemKind::Const(..)
|
||
| ItemKind::ExternCrate(..)
|
||
| ItemKind::ForeignMod(..)
|
||
// Another function was found; this case is ignored
|
||
| ItemKind::Fn(..) => return false,
|
||
_ => {},
|
||
},
|
||
Ok(None) => break,
|
||
Err(e) => {
|
||
e.cancel();
|
||
return false;
|
||
},
|
||
}
|
||
}
|
||
|
||
relevant_main_found
|
||
})
|
||
})
|
||
.ok()
|
||
.unwrap_or_default()
|
||
}
|
||
|
||
let trailing_whitespace = text.len() - text.trim_end().len();
|
||
|
||
// Because of the global session, we need to create a new session in a different thread with
|
||
// the edition we need.
|
||
let text = text.to_owned();
|
||
if thread::spawn(move || has_needless_main(text, edition)).join().expect("thread::spawn failed")
|
||
&& let Some(span) = fragments.span(cx, range.start..range.end - trailing_whitespace)
|
||
{
|
||
span_lint(cx, NEEDLESS_DOCTEST_MAIN, span, "needless `fn main` in doctest");
|
||
}
|
||
}
|
||
|
||
fn check_text(cx: &LateContext<'_>, valid_idents: &FxHashSet<String>, text: &str, span: Span) {
|
||
for word in text.split(|c: char| c.is_whitespace() || c == '\'') {
|
||
// Trim punctuation as in `some comment (see foo::bar).`
|
||
// ^^
|
||
// Or even as in `_foo bar_` which is emphasized. Also preserve `::` as a prefix/suffix.
|
||
let mut word = word.trim_matches(|c: char| !c.is_alphanumeric() && c != ':');
|
||
|
||
// Remove leading or trailing single `:` which may be part of a sentence.
|
||
if word.starts_with(':') && !word.starts_with("::") {
|
||
word = word.trim_start_matches(':');
|
||
}
|
||
if word.ends_with(':') && !word.ends_with("::") {
|
||
word = word.trim_end_matches(':');
|
||
}
|
||
|
||
if valid_idents.contains(word) || word.chars().all(|c| c == ':') {
|
||
continue;
|
||
}
|
||
|
||
// Adjust for the current word
|
||
let offset = word.as_ptr() as usize - text.as_ptr() as usize;
|
||
let span = Span::new(
|
||
span.lo() + BytePos::from_usize(offset),
|
||
span.lo() + BytePos::from_usize(offset + word.len()),
|
||
span.ctxt(),
|
||
span.parent(),
|
||
);
|
||
|
||
check_word(cx, word, span);
|
||
}
|
||
}
|
||
|
||
fn check_word(cx: &LateContext<'_>, word: &str, span: Span) {
|
||
/// Checks if a string is camel-case, i.e., contains at least two uppercase
|
||
/// letters (`Clippy` is ok) and one lower-case letter (`NASA` is ok).
|
||
/// Plurals are also excluded (`IDs` is ok).
|
||
fn is_camel_case(s: &str) -> bool {
|
||
if s.starts_with(|c: char| c.is_ascii_digit()) {
|
||
return false;
|
||
}
|
||
|
||
let s = s.strip_suffix('s').unwrap_or(s);
|
||
|
||
s.chars().all(char::is_alphanumeric)
|
||
&& s.chars().filter(|&c| c.is_uppercase()).take(2).count() > 1
|
||
&& s.chars().filter(|&c| c.is_lowercase()).take(1).count() > 0
|
||
}
|
||
|
||
fn has_underscore(s: &str) -> bool {
|
||
s != "_" && !s.contains("\\_") && s.contains('_')
|
||
}
|
||
|
||
fn has_hyphen(s: &str) -> bool {
|
||
s != "-" && s.contains('-')
|
||
}
|
||
|
||
if let Ok(url) = Url::parse(word) {
|
||
// try to get around the fact that `foo::bar` parses as a valid URL
|
||
if !url.cannot_be_a_base() {
|
||
span_lint(
|
||
cx,
|
||
DOC_MARKDOWN,
|
||
span,
|
||
"you should put bare URLs between `<`/`>` or make a proper Markdown link",
|
||
);
|
||
|
||
return;
|
||
}
|
||
}
|
||
|
||
// We assume that mixed-case words are not meant to be put inside backticks. (Issue #2343)
|
||
if has_underscore(word) && has_hyphen(word) {
|
||
return;
|
||
}
|
||
|
||
if has_underscore(word) || word.contains("::") || is_camel_case(word) {
|
||
let mut applicability = Applicability::MachineApplicable;
|
||
|
||
span_lint_and_then(
|
||
cx,
|
||
DOC_MARKDOWN,
|
||
span,
|
||
"item in documentation is missing backticks",
|
||
|diag| {
|
||
let snippet = snippet_with_applicability(cx, span, "..", &mut applicability);
|
||
diag.span_suggestion_with_style(
|
||
span,
|
||
"try",
|
||
format!("`{snippet}`"),
|
||
applicability,
|
||
// always show the suggestion in a separate line, since the
|
||
// inline presentation adds another pair of backticks
|
||
SuggestionStyle::ShowAlways,
|
||
);
|
||
},
|
||
);
|
||
}
|
||
}
|
||
|
||
struct FindPanicUnwrap<'a, 'tcx> {
|
||
cx: &'a LateContext<'tcx>,
|
||
panic_span: Option<Span>,
|
||
typeck_results: &'tcx ty::TypeckResults<'tcx>,
|
||
}
|
||
|
||
impl<'a, 'tcx> Visitor<'tcx> for FindPanicUnwrap<'a, 'tcx> {
|
||
type NestedFilter = nested_filter::OnlyBodies;
|
||
|
||
fn visit_expr(&mut self, expr: &'tcx Expr<'_>) {
|
||
if self.panic_span.is_some() {
|
||
return;
|
||
}
|
||
|
||
if let Some(macro_call) = root_macro_call_first_node(self.cx, expr) {
|
||
if is_panic(self.cx, macro_call.def_id)
|
||
|| matches!(
|
||
self.cx.tcx.item_name(macro_call.def_id).as_str(),
|
||
"assert" | "assert_eq" | "assert_ne"
|
||
)
|
||
{
|
||
self.panic_span = Some(macro_call.span);
|
||
}
|
||
}
|
||
|
||
// check for `unwrap` and `expect` for both `Option` and `Result`
|
||
if let Some(arglists) = method_chain_args(expr, &["unwrap"]).or(method_chain_args(expr, &["expect"])) {
|
||
let receiver_ty = self.typeck_results.expr_ty(arglists[0].0).peel_refs();
|
||
if is_type_diagnostic_item(self.cx, receiver_ty, sym::Option)
|
||
|| is_type_diagnostic_item(self.cx, receiver_ty, sym::Result)
|
||
{
|
||
self.panic_span = Some(expr.span);
|
||
}
|
||
}
|
||
|
||
// and check sub-expressions
|
||
intravisit::walk_expr(self, expr);
|
||
}
|
||
|
||
// Panics in const blocks will cause compilation to fail.
|
||
fn visit_anon_const(&mut self, _: &'tcx AnonConst) {}
|
||
|
||
fn nested_visit_map(&mut self) -> Self::Map {
|
||
self.cx.tcx.hir()
|
||
}
|
||
}
|