Fix clippy::redundant_field_names
This commit is contained in:
parent
dedcd97e92
commit
12806b7050
@ -8,7 +8,7 @@ where
|
||||
Ls: Links,
|
||||
{
|
||||
VecLinkedListIterator {
|
||||
links: links,
|
||||
links,
|
||||
current: first,
|
||||
}
|
||||
}
|
||||
|
@ -94,7 +94,7 @@ impl<'a> DiagnosticConverter<'a> {
|
||||
annotation_type: Self::annotation_type_for_level(self.level),
|
||||
}),
|
||||
footer: vec![],
|
||||
slices: slices,
|
||||
slices,
|
||||
})
|
||||
} else {
|
||||
// FIXME(#59346): Is it ok to return None if there's no source_map?
|
||||
|
@ -388,7 +388,7 @@ impl Diagnostic {
|
||||
}],
|
||||
msg: msg.to_owned(),
|
||||
style: SuggestionStyle::CompletelyHidden,
|
||||
applicability: applicability,
|
||||
applicability,
|
||||
});
|
||||
self
|
||||
}
|
||||
|
@ -19,7 +19,7 @@ pub fn opts() -> TargetOptions {
|
||||
is_like_fuchsia: true,
|
||||
linker_is_gnu: true,
|
||||
has_rpath: false,
|
||||
pre_link_args: pre_link_args,
|
||||
pre_link_args,
|
||||
pre_link_objects_exe: vec![
|
||||
"Scrt1.o".to_string()
|
||||
],
|
||||
|
@ -461,7 +461,7 @@ impl<'a> Encoder<'a> {
|
||||
/// Creates a new JSON encoder whose output will be written to the writer
|
||||
/// specified.
|
||||
pub fn new(writer: &'a mut dyn fmt::Write) -> Encoder<'a> {
|
||||
Encoder { writer: writer, is_emitting_map_key: false, }
|
||||
Encoder { writer, is_emitting_map_key: false, }
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -1832,7 +1832,7 @@ impl Arg {
|
||||
lt,
|
||||
MutTy {
|
||||
ty: infer_ty,
|
||||
mutbl: mutbl,
|
||||
mutbl,
|
||||
},
|
||||
),
|
||||
span,
|
||||
@ -2120,7 +2120,7 @@ impl PolyTraitRef {
|
||||
PolyTraitRef {
|
||||
bound_generic_params: generic_params,
|
||||
trait_ref: TraitRef {
|
||||
path: path,
|
||||
path,
|
||||
ref_id: DUMMY_NODE_ID,
|
||||
},
|
||||
span,
|
||||
|
@ -815,7 +815,7 @@ impl<'a> AstBuilder for ExtCtxt<'a> {
|
||||
|
||||
|
||||
fn pat(&self, span: Span, pat: PatKind) -> P<ast::Pat> {
|
||||
P(ast::Pat { id: ast::DUMMY_NODE_ID, node: pat, span: span })
|
||||
P(ast::Pat { id: ast::DUMMY_NODE_ID, node: pat, span })
|
||||
}
|
||||
fn pat_wild(&self, span: Span) -> P<ast::Pat> {
|
||||
self.pat(span, PatKind::Wild)
|
||||
|
@ -231,7 +231,7 @@ pub struct MacroExpander<'a, 'b> {
|
||||
|
||||
impl<'a, 'b> MacroExpander<'a, 'b> {
|
||||
pub fn new(cx: &'a mut ExtCtxt<'b>, monotonic: bool) -> Self {
|
||||
MacroExpander { cx: cx, monotonic: monotonic }
|
||||
MacroExpander { cx, monotonic }
|
||||
}
|
||||
|
||||
pub fn expand_crate(&mut self, mut krate: ast::Crate) -> ast::Crate {
|
||||
@ -377,7 +377,7 @@ impl<'a, 'b> MacroExpander<'a, 'b> {
|
||||
_ => item.clone(),
|
||||
};
|
||||
invocations.push(Invocation {
|
||||
kind: InvocationKind::Derive { path: path.clone(), item: item },
|
||||
kind: InvocationKind::Derive { path: path.clone(), item },
|
||||
fragment_kind: invoc.fragment_kind,
|
||||
expansion_data: ExpansionData {
|
||||
mark,
|
||||
@ -944,7 +944,7 @@ impl<'a, 'b> InvocationCollector<'a, 'b> {
|
||||
}
|
||||
|
||||
fn collect_bang(&mut self, mac: ast::Mac, span: Span, kind: AstFragmentKind) -> AstFragment {
|
||||
self.collect(kind, InvocationKind::Bang { mac: mac, ident: None, span: span })
|
||||
self.collect(kind, InvocationKind::Bang { mac, ident: None, span })
|
||||
}
|
||||
|
||||
fn collect_attr(&mut self,
|
||||
|
@ -319,7 +319,7 @@ fn parse_tree(
|
||||
tokenstream::TokenTree::Delimited(span, delim, tts) => TokenTree::Delimited(
|
||||
span,
|
||||
Lrc::new(Delimited {
|
||||
delim: delim,
|
||||
delim,
|
||||
tts: parse(
|
||||
tts.into(),
|
||||
expect_matchers,
|
||||
|
@ -23,8 +23,8 @@ enum Frame {
|
||||
impl Frame {
|
||||
/// Construct a new frame around the delimited set of tokens.
|
||||
fn new(tts: Vec<quoted::TokenTree>) -> Frame {
|
||||
let forest = Lrc::new(quoted::Delimited { delim: token::NoDelim, tts: tts });
|
||||
Frame::Delimited { forest: forest, idx: 0, span: DelimSpan::dummy() }
|
||||
let forest = Lrc::new(quoted::Delimited { delim: token::NoDelim, tts });
|
||||
Frame::Delimited { forest, idx: 0, span: DelimSpan::dummy() }
|
||||
}
|
||||
}
|
||||
|
||||
@ -248,7 +248,7 @@ pub fn transcribe(
|
||||
// the previous results (from outside the Delimited).
|
||||
quoted::TokenTree::Delimited(mut span, delimited) => {
|
||||
span = span.apply_mark(cx.current_expansion.mark);
|
||||
stack.push(Frame::Delimited { forest: delimited, idx: 0, span: span });
|
||||
stack.push(Frame::Delimited { forest: delimited, idx: 0, span });
|
||||
result_stack.push(mem::replace(&mut result, Vec::new()));
|
||||
}
|
||||
|
||||
|
@ -1665,7 +1665,7 @@ impl<'a> Context<'a> {
|
||||
}
|
||||
|
||||
pub fn check_attribute(attr: &ast::Attribute, parse_sess: &ParseSess, features: &Features) {
|
||||
let cx = Context { features: features, parse_sess: parse_sess, plugin_attributes: &[] };
|
||||
let cx = Context { features, parse_sess, plugin_attributes: &[] };
|
||||
cx.check_attribute(
|
||||
attr,
|
||||
attr.ident().and_then(|ident| BUILTIN_ATTRIBUTE_MAP.get(&ident.name).map(|a| *a)),
|
||||
|
@ -290,10 +290,10 @@ crate enum LastToken {
|
||||
}
|
||||
|
||||
impl TokenCursorFrame {
|
||||
fn new(sp: DelimSpan, delim: DelimToken, tts: &TokenStream) -> Self {
|
||||
fn new(span: DelimSpan, delim: DelimToken, tts: &TokenStream) -> Self {
|
||||
TokenCursorFrame {
|
||||
delim: delim,
|
||||
span: sp,
|
||||
delim,
|
||||
span,
|
||||
open_delim: delim == token::NoDelim,
|
||||
tree_cursor: tts.clone().into_trees(),
|
||||
close_delim: delim == token::NoDelim,
|
||||
@ -1449,7 +1449,7 @@ impl<'a> Parser<'a> {
|
||||
let opt_lifetime = if self.check_lifetime() { Some(self.expect_lifetime()) } else { None };
|
||||
let mutbl = self.parse_mutability();
|
||||
let ty = self.parse_ty_no_plus()?;
|
||||
return Ok(TyKind::Rptr(opt_lifetime, MutTy { ty: ty, mutbl: mutbl }));
|
||||
return Ok(TyKind::Rptr(opt_lifetime, MutTy { ty, mutbl }));
|
||||
}
|
||||
|
||||
fn parse_ptr(&mut self) -> PResult<'a, MutTy> {
|
||||
@ -1467,7 +1467,7 @@ impl<'a> Parser<'a> {
|
||||
Mutability::Immutable
|
||||
};
|
||||
let t = self.parse_ty_no_plus()?;
|
||||
Ok(MutTy { ty: t, mutbl: mutbl })
|
||||
Ok(MutTy { ty: t, mutbl })
|
||||
}
|
||||
|
||||
fn is_named_argument(&self) -> bool {
|
||||
@ -4366,7 +4366,7 @@ impl<'a> Parser<'a> {
|
||||
self.report_invalid_macro_expansion_item();
|
||||
}
|
||||
|
||||
(ident, ast::MacroDef { tokens: tokens, legacy: true })
|
||||
(ident, ast::MacroDef { tokens, legacy: true })
|
||||
}
|
||||
_ => return Ok(None),
|
||||
};
|
||||
@ -6789,12 +6789,12 @@ impl<'a> Parser<'a> {
|
||||
let hi = self.token.span;
|
||||
self.expect(&token::Semi)?;
|
||||
Ok(ast::ForeignItem {
|
||||
ident: ident,
|
||||
attrs: attrs,
|
||||
ident,
|
||||
attrs,
|
||||
node: ForeignItemKind::Ty,
|
||||
id: ast::DUMMY_NODE_ID,
|
||||
span: lo.to(hi),
|
||||
vis: vis
|
||||
vis
|
||||
})
|
||||
}
|
||||
|
||||
|
@ -150,7 +150,7 @@ impl SourceMap {
|
||||
-> SourceMap {
|
||||
SourceMap {
|
||||
files: Default::default(),
|
||||
file_loader: file_loader,
|
||||
file_loader,
|
||||
path_mapping,
|
||||
}
|
||||
}
|
||||
@ -396,7 +396,7 @@ impl SourceMap {
|
||||
let f = (*self.files.borrow().source_files)[idx].clone();
|
||||
|
||||
match f.lookup_line(pos) {
|
||||
Some(line) => Ok(SourceFileAndLine { sf: f, line: line }),
|
||||
Some(line) => Ok(SourceFileAndLine { sf: f, line }),
|
||||
None => Err(f)
|
||||
}
|
||||
}
|
||||
@ -511,7 +511,7 @@ impl SourceMap {
|
||||
start_col,
|
||||
end_col: hi.col });
|
||||
|
||||
Ok(FileLines {file: lo.file, lines: lines})
|
||||
Ok(FileLines {file: lo.file, lines})
|
||||
}
|
||||
|
||||
/// Extracts the source surrounding the given `Span` using the `extract_source` function. The
|
||||
@ -820,7 +820,7 @@ impl SourceMap {
|
||||
let idx = self.lookup_source_file_idx(bpos);
|
||||
let sf = (*self.files.borrow().source_files)[idx].clone();
|
||||
let offset = bpos - sf.start_pos;
|
||||
SourceFileAndBytePos {sf: sf, pos: offset}
|
||||
SourceFileAndBytePos {sf, pos: offset}
|
||||
}
|
||||
|
||||
/// Converts an absolute BytePos to a CharPos relative to the source_file.
|
||||
|
Loading…
x
Reference in New Issue
Block a user