]> git.lizzy.rs Git - rust.git/blobdiff - src/librustc_parse/lexer/tokentrees.rs
use find(x) instead of filter(x).next()
[rust.git] / src / librustc_parse / lexer / tokentrees.rs
index c28b59a7908015572cce6e33c3d1858a8536cf88..20a7fcb650a2e17ba3fa8d62dd17b5aa37dc3265 100644 (file)
@@ -93,10 +93,8 @@ fn parse_token_tree(&mut self) -> PResult<'a, TreeAndJoint> {
                 }
 
                 if let Some((delim, _)) = self.open_braces.last() {
-                    if let Some((_, open_sp, close_sp)) = self
-                        .matching_delim_spans
-                        .iter()
-                        .filter(|(d, open_sp, close_sp)| {
+                    if let Some((_, open_sp, close_sp)) =
+                        self.matching_delim_spans.iter().find(|(d, open_sp, close_sp)| {
                             if let Some(close_padding) = sm.span_to_margin(*close_sp) {
                                 if let Some(open_padding) = sm.span_to_margin(*open_sp) {
                                     return delim == d && close_padding != open_padding;
@@ -104,7 +102,6 @@ fn parse_token_tree(&mut self) -> PResult<'a, TreeAndJoint> {
                             }
                             false
                         })
-                        .next()
                     // these are in reverse order as they get inserted on close, but
                     {
                         // we want the last open/first close