2017-05-28 17:12:43 -05:00
|
|
|
|
use itertools::Itertools;
|
|
|
|
|
use pulldown_cmark;
|
2016-03-19 11:59:12 -05:00
|
|
|
|
use rustc::lint::*;
|
|
|
|
|
use syntax::ast;
|
2016-05-02 07:36:33 -05:00
|
|
|
|
use syntax::codemap::{Span, BytePos};
|
2017-05-29 17:11:08 -05:00
|
|
|
|
use syntax_pos::Pos;
|
2017-05-30 12:28:44 -05:00
|
|
|
|
use utils::span_lint;
|
2016-03-19 11:59:12 -05:00
|
|
|
|
|
2016-08-06 02:55:04 -05:00
|
|
|
|
/// **What it does:** Checks for the presence of `_`, `::` or camel-case words
|
|
|
|
|
/// outside ticks in documentation.
|
2016-03-19 11:59:12 -05:00
|
|
|
|
///
|
2016-08-06 02:55:04 -05:00
|
|
|
|
/// **Why is this bad?** *Rustdoc* supports markdown formatting, `_`, `::` and
|
|
|
|
|
/// camel-case probably indicates some code which should be included between
|
|
|
|
|
/// ticks. `_` can also be used for empasis in markdown, this lint tries to
|
|
|
|
|
/// consider that.
|
2016-03-19 11:59:12 -05:00
|
|
|
|
///
|
2016-08-06 02:55:04 -05:00
|
|
|
|
/// **Known problems:** Lots of bad docs won’t be fixed, what the lint checks
|
|
|
|
|
/// for is limited, and there are still false positives.
|
2016-03-19 11:59:12 -05:00
|
|
|
|
///
|
|
|
|
|
/// **Examples:**
|
|
|
|
|
/// ```rust
|
2016-03-28 11:00:24 -05:00
|
|
|
|
/// /// Do something with the foo_bar parameter. See also that::other::module::foo.
|
|
|
|
|
/// // ^ `foo_bar` and `that::other::module::foo` should be ticked.
|
2016-03-19 11:59:12 -05:00
|
|
|
|
/// fn doit(foo_bar) { .. }
|
|
|
|
|
/// ```
|
|
|
|
|
declare_lint! {
|
2016-08-06 03:18:36 -05:00
|
|
|
|
pub DOC_MARKDOWN,
|
|
|
|
|
Warn,
|
|
|
|
|
"presence of `_`, `::` or camel-case outside backticks in documentation"
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
|
|
|
|
|
2016-04-04 13:18:17 -05:00
|
|
|
|
#[derive(Clone)]
|
|
|
|
|
pub struct Doc {
|
|
|
|
|
valid_idents: Vec<String>,
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
impl Doc {
|
|
|
|
|
pub fn new(valid_idents: Vec<String>) -> Self {
|
|
|
|
|
Doc { valid_idents: valid_idents }
|
|
|
|
|
}
|
|
|
|
|
}
|
2016-03-19 11:59:12 -05:00
|
|
|
|
|
|
|
|
|
impl LintPass for Doc {
|
|
|
|
|
fn get_lints(&self) -> LintArray {
|
|
|
|
|
lint_array![DOC_MARKDOWN]
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
impl EarlyLintPass for Doc {
|
|
|
|
|
fn check_crate(&mut self, cx: &EarlyContext, krate: &ast::Crate) {
|
2016-05-02 07:36:33 -05:00
|
|
|
|
check_attrs(cx, &self.valid_idents, &krate.attrs);
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
fn check_item(&mut self, cx: &EarlyContext, item: &ast::Item) {
|
2016-05-02 07:36:33 -05:00
|
|
|
|
check_attrs(cx, &self.valid_idents, &item.attrs);
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
struct Parser<'a> {
|
|
|
|
|
parser: pulldown_cmark::Parser<'a>,
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
impl<'a> Parser<'a> {
|
|
|
|
|
fn new(parser: pulldown_cmark::Parser<'a>) -> Parser<'a> {
|
|
|
|
|
Self { parser }
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
impl<'a> Iterator for Parser<'a> {
|
|
|
|
|
type Item = (usize, pulldown_cmark::Event<'a>);
|
|
|
|
|
|
|
|
|
|
fn next(&mut self) -> Option<Self::Item> {
|
|
|
|
|
let offset = self.parser.get_offset();
|
|
|
|
|
self.parser.next().map(|event| (offset, event))
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2016-07-08 11:18:45 -05:00
|
|
|
|
/// Cleanup documentation decoration (`///` and such).
|
|
|
|
|
///
|
|
|
|
|
/// We can't use `syntax::attr::AttributeMethods::with_desugared_doc` or
|
|
|
|
|
/// `syntax::parse::lexer::comments::strip_doc_comment_decoration` because we need to keep track of
|
2017-05-28 17:12:43 -05:00
|
|
|
|
/// the spans but this function is inspired from the later.
|
2016-07-08 11:18:45 -05:00
|
|
|
|
#[allow(cast_possible_truncation)]
|
2017-05-30 12:28:44 -05:00
|
|
|
|
pub fn strip_doc_comment_decoration(comment: &str, span: Span) -> (String, Vec<(usize, Span)>) {
|
2016-07-08 11:18:45 -05:00
|
|
|
|
// one-line comments lose their prefix
|
|
|
|
|
const ONELINERS: &'static [&'static str] = &["///!", "///", "//!", "//"];
|
|
|
|
|
for prefix in ONELINERS {
|
|
|
|
|
if comment.starts_with(*prefix) {
|
2017-05-28 17:12:43 -05:00
|
|
|
|
let doc = &comment[prefix.len()..];
|
|
|
|
|
let mut doc = doc.to_owned();
|
|
|
|
|
doc.push('\n');
|
|
|
|
|
return (
|
|
|
|
|
doc.to_owned(),
|
|
|
|
|
vec![(doc.len(), Span { lo: span.lo + BytePos(prefix.len() as u32), ..span })]
|
|
|
|
|
);
|
2016-07-08 11:18:45 -05:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if comment.starts_with("/*") {
|
2017-05-28 17:12:43 -05:00
|
|
|
|
let doc = &comment[3..comment.len() - 2];
|
|
|
|
|
let mut sizes = vec![];
|
|
|
|
|
|
|
|
|
|
for line in doc.lines() {
|
|
|
|
|
let offset = line.as_ptr() as usize - comment.as_ptr() as usize;
|
|
|
|
|
debug_assert_eq!(offset as u32 as usize, offset);
|
|
|
|
|
|
2017-05-30 12:28:44 -05:00
|
|
|
|
// +1 for the newline
|
|
|
|
|
sizes.push((line.len()+1, Span { lo: span.lo + BytePos(offset as u32), ..span }));
|
2017-05-28 17:12:43 -05:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return (doc.to_string(), sizes);
|
2016-07-08 11:18:45 -05:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
panic!("not a doc-comment: {}", comment);
|
|
|
|
|
}
|
|
|
|
|
|
2016-05-02 07:36:33 -05:00
|
|
|
|
pub fn check_attrs<'a>(cx: &EarlyContext, valid_idents: &[String], attrs: &'a [ast::Attribute]) {
|
2017-05-28 17:12:43 -05:00
|
|
|
|
let mut doc = String::new();
|
|
|
|
|
let mut spans = vec![];
|
2016-05-26 15:53:38 -05:00
|
|
|
|
|
2016-05-02 07:36:33 -05:00
|
|
|
|
for attr in attrs {
|
2016-11-23 14:19:03 -06:00
|
|
|
|
if attr.is_sugared_doc {
|
2017-05-28 17:12:43 -05:00
|
|
|
|
if let Some(ref current) = attr.value_str() {
|
|
|
|
|
let current = current.to_string();
|
2017-05-30 12:28:44 -05:00
|
|
|
|
let (current, current_spans) = strip_doc_comment_decoration(¤t, attr.span);
|
2017-05-28 17:12:43 -05:00
|
|
|
|
spans.extend_from_slice(¤t_spans);
|
|
|
|
|
doc.push_str(¤t);
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
2017-05-30 12:50:07 -05:00
|
|
|
|
} else if let Some(name) = attr.name() {
|
|
|
|
|
// ignore mix of sugared and non-sugared doc
|
|
|
|
|
if name == "doc" {
|
|
|
|
|
return;
|
|
|
|
|
}
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
|
|
|
|
}
|
2016-04-11 16:22:30 -05:00
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
let mut current = 0;
|
|
|
|
|
for &mut (ref mut offset, _) in &mut spans {
|
|
|
|
|
let offset_copy = *offset;
|
|
|
|
|
*offset = current;
|
|
|
|
|
current += offset_copy;
|
2016-05-26 15:53:38 -05:00
|
|
|
|
}
|
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
if !doc.is_empty() {
|
|
|
|
|
let parser = Parser::new(pulldown_cmark::Parser::new(&doc));
|
|
|
|
|
let parser = parser.coalesce(|x, y| {
|
|
|
|
|
use pulldown_cmark::Event::*;
|
2016-05-27 20:18:52 -05:00
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
let x_offset = x.0;
|
|
|
|
|
let y_offset = y.0;
|
2016-05-26 15:53:38 -05:00
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
match (x.1, y.1) {
|
2017-05-30 12:28:44 -05:00
|
|
|
|
(Text(x), Text(y)) => {
|
|
|
|
|
let mut x = x.into_owned();
|
|
|
|
|
x.push_str(&y);
|
|
|
|
|
Ok((x_offset, Text(x.into())))
|
|
|
|
|
}
|
2017-05-28 17:12:43 -05:00
|
|
|
|
(x, y) => Err(((x_offset, x), (y_offset, y))),
|
2016-05-26 15:53:38 -05:00
|
|
|
|
}
|
2017-05-28 17:12:43 -05:00
|
|
|
|
});
|
|
|
|
|
check_doc(cx, valid_idents, parser, &spans);
|
2016-05-26 15:53:38 -05:00
|
|
|
|
}
|
2017-05-28 17:12:43 -05:00
|
|
|
|
}
|
2016-05-26 15:53:38 -05:00
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
fn check_doc<'a, Events: Iterator<Item=(usize, pulldown_cmark::Event<'a>)>>(
|
|
|
|
|
cx: &EarlyContext,
|
|
|
|
|
valid_idents: &[String],
|
|
|
|
|
docs: Events,
|
|
|
|
|
spans: &[(usize, Span)]
|
|
|
|
|
) {
|
|
|
|
|
use pulldown_cmark::Event::*;
|
|
|
|
|
use pulldown_cmark::Tag::*;
|
|
|
|
|
|
|
|
|
|
let mut in_code = false;
|
|
|
|
|
|
|
|
|
|
for (offset, event) in docs {
|
|
|
|
|
match event {
|
|
|
|
|
Start(CodeBlock(_)) | Start(Code) => in_code = true,
|
|
|
|
|
End(CodeBlock(_)) | End(Code) => in_code = false,
|
|
|
|
|
Start(_tag) | End(_tag) => (), // We don't care about other tags
|
|
|
|
|
Html(_html) | InlineHtml(_html) => (), // HTML is weird, just ignore it
|
|
|
|
|
SoftBreak => (),
|
|
|
|
|
HardBreak => (),
|
2017-05-30 12:28:44 -05:00
|
|
|
|
FootnoteReference(text) | Text(text) => {
|
2017-05-28 17:12:43 -05:00
|
|
|
|
if !in_code {
|
|
|
|
|
let index = match spans.binary_search_by(|c| c.0.cmp(&offset)) {
|
|
|
|
|
Ok(o) => o,
|
|
|
|
|
Err(e) => e-1,
|
|
|
|
|
};
|
|
|
|
|
|
2017-05-29 17:11:08 -05:00
|
|
|
|
let (begin, span) = spans[index];
|
|
|
|
|
|
|
|
|
|
// Adjust for the begining of the current `Event`
|
|
|
|
|
let span = Span {
|
|
|
|
|
lo: span.lo + BytePos::from_usize(offset - begin),
|
|
|
|
|
..span
|
|
|
|
|
};
|
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
check_text(cx, valid_idents, &text, span);
|
2016-05-27 20:18:52 -05:00
|
|
|
|
}
|
2017-05-28 17:12:43 -05:00
|
|
|
|
},
|
2016-05-26 15:53:38 -05:00
|
|
|
|
}
|
2016-05-02 07:36:48 -05:00
|
|
|
|
}
|
2017-05-28 17:12:43 -05:00
|
|
|
|
}
|
2016-05-02 07:36:48 -05:00
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
fn check_text(cx: &EarlyContext, valid_idents: &[String], text: &str, span: Span) {
|
|
|
|
|
for word in text.split_whitespace() {
|
|
|
|
|
// Trim punctuation as in `some comment (see foo::bar).`
|
|
|
|
|
// ^^
|
|
|
|
|
// Or even as in `_foo bar_` which is emphasized.
|
|
|
|
|
let word = word.trim_matches(|c: char| !c.is_alphanumeric());
|
2016-05-05 14:42:59 -05:00
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
if valid_idents.iter().any(|i| i == word) {
|
|
|
|
|
continue;
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
2016-05-26 15:53:38 -05:00
|
|
|
|
|
2017-05-29 17:11:08 -05:00
|
|
|
|
// Adjust for the current word
|
|
|
|
|
let offset = word.as_ptr() as usize - text.as_ptr() as usize;
|
|
|
|
|
let span = Span {
|
|
|
|
|
lo: span.lo + BytePos::from_usize(offset),
|
|
|
|
|
hi: span.lo + BytePos::from_usize(offset + word.len()),
|
|
|
|
|
..span
|
|
|
|
|
};
|
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
check_word(cx, word, span);
|
|
|
|
|
}
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
|
|
|
|
|
2017-05-28 17:12:43 -05:00
|
|
|
|
fn check_word(cx: &EarlyContext, word: &str, span: Span) {
|
|
|
|
|
/// Checks if a string is camel-case, ie. contains at least two uppercase letter (`Clippy` is
|
2016-03-19 11:59:12 -05:00
|
|
|
|
/// ok) and one lower-case letter (`NASA` is ok). Plural are also excluded (`IDs` is ok).
|
|
|
|
|
fn is_camel_case(s: &str) -> bool {
|
2016-04-04 13:18:17 -05:00
|
|
|
|
if s.starts_with(|c: char| c.is_digit(10)) {
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
2016-03-19 11:59:12 -05:00
|
|
|
|
let s = if s.ends_with('s') {
|
2016-04-14 13:14:03 -05:00
|
|
|
|
&s[..s.len() - 1]
|
2016-03-19 11:59:12 -05:00
|
|
|
|
} else {
|
|
|
|
|
s
|
|
|
|
|
};
|
|
|
|
|
|
2016-12-20 11:21:30 -06:00
|
|
|
|
s.chars().all(char::is_alphanumeric) && s.chars().filter(|&c| c.is_uppercase()).take(2).count() > 1 &&
|
2016-03-19 11:59:12 -05:00
|
|
|
|
s.chars().filter(|&c| c.is_lowercase()).take(1).count() > 0
|
|
|
|
|
}
|
|
|
|
|
|
2016-03-28 11:00:24 -05:00
|
|
|
|
fn has_underscore(s: &str) -> bool {
|
|
|
|
|
s != "_" && !s.contains("\\_") && s.contains('_')
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if has_underscore(word) || word.contains("::") || is_camel_case(word) {
|
2016-04-14 13:14:03 -05:00
|
|
|
|
span_lint(cx,
|
|
|
|
|
DOC_MARKDOWN,
|
|
|
|
|
span,
|
|
|
|
|
&format!("you should put `{}` between ticks in the documentation", word));
|
2016-03-19 11:59:12 -05:00
|
|
|
|
}
|
|
|
|
|
}
|