refactor: add chunks method to TokenStream to obviate rustdoc clones
This commit is contained in:
parent
2c41369acc
commit
00c3f7552e
@ -551,6 +551,10 @@ pub fn push_stream(&mut self, stream: TokenStream) {
|
||||
vec_mut.extend(stream_iter);
|
||||
}
|
||||
}
|
||||
|
||||
pub fn chunks(&self, chunk_size: usize) -> core::slice::Chunks<'_, TokenTree> {
|
||||
self.0.chunks(chunk_size)
|
||||
}
|
||||
}
|
||||
|
||||
/// By-reference iterator over a [`TokenStream`], that produces `&TokenTree`
|
||||
|
@ -594,9 +594,8 @@ pub(super) fn display_macro_source(
|
||||
def_id: DefId,
|
||||
vis: ty::Visibility<DefId>,
|
||||
) -> String {
|
||||
let tts: Vec<_> = def.body.tokens.clone().into_trees().collect();
|
||||
// Extract the spans of all matchers. They represent the "interface" of the macro.
|
||||
let matchers = tts.chunks(4).map(|arm| &arm[0]);
|
||||
let matchers = def.body.tokens.chunks(4).map(|arm| &arm[0]);
|
||||
|
||||
if def.macro_rules {
|
||||
format!("macro_rules! {} {{\n{}}}", name, render_macro_arms(cx.tcx, matchers, ";"))
|
||||
|
Loading…
Reference in New Issue
Block a user