]> git.lizzy.rs Git - rust.git/blobdiff - src/librustc_parse/lexer/tokentrees.rs
use find(x) instead of filter(x).next()
[rust.git] / src / librustc_parse / lexer / tokentrees.rs
index a28bff3babfbf53909ab982746a2e6b6beb216bb..20a7fcb650a2e17ba3fa8d62dd17b5aa37dc3265 100644 (file)
@@ -1,9 +1,9 @@
 use super::{StringReader, UnmatchedBrace};
 
+use rustc_ast_pretty::pprust::token_to_string;
 use rustc_data_structures::fx::FxHashMap;
 use rustc_errors::PResult;
 use rustc_span::Span;
-use syntax::print::pprust::token_to_string;
 use syntax::token::{self, Token};
 use syntax::tokenstream::{
     DelimSpan,
@@ -93,10 +93,8 @@ fn parse_token_tree(&mut self) -> PResult<'a, TreeAndJoint> {
                 }
 
                 if let Some((delim, _)) = self.open_braces.last() {
-                    if let Some((_, open_sp, close_sp)) = self
-                        .matching_delim_spans
-                        .iter()
-                        .filter(|(d, open_sp, close_sp)| {
+                    if let Some((_, open_sp, close_sp)) =
+                        self.matching_delim_spans.iter().find(|(d, open_sp, close_sp)| {
                             if let Some(close_padding) = sm.span_to_margin(*close_sp) {
                                 if let Some(open_padding) = sm.span_to_margin(*open_sp) {
                                     return delim == d && close_padding != open_padding;
@@ -104,7 +102,6 @@ fn parse_token_tree(&mut self) -> PResult<'a, TreeAndJoint> {
                             }
                             false
                         })
-                        .next()
                     // these are in reverse order as they get inserted on close, but
                     {
                         // we want the last open/first close