X-Git-Url: https://git.lizzy.rs/?a=blobdiff_plain;f=crates%2Fide_db%2Fsrc%2Fhelpers.rs;h=c355016c5dfd054dac27f949f37eeb4a6882bab6;hb=979b5b32bce12c1efb88d4b58874606eef0a4c6a;hp=173e55b33f6ac7b6579b96d4a0065639be65740e;hpb=dfa355b4313a7ea3eb5a262ee9f1da71f50884d5;p=rust.git diff --git a/crates/ide_db/src/helpers.rs b/crates/ide_db/src/helpers.rs index 173e55b33f6..c355016c5df 100644 --- a/crates/ide_db/src/helpers.rs +++ b/crates/ide_db/src/helpers.rs @@ -4,20 +4,23 @@ pub mod import_assets; pub mod insert_use; pub mod merge_imports; +pub mod insert_whitespace_into_node; pub mod node_ext; pub mod rust_doc; +pub mod format_string; -use std::collections::VecDeque; +use std::{collections::VecDeque, iter}; use base_db::FileId; -use either::Either; -use hir::{ItemInNs, MacroDef, ModuleDef, Name, Semantics}; +use hir::{ItemInNs, MacroDef, ModuleDef, Name, PathResolution, Semantics}; +use itertools::Itertools; use syntax::{ ast::{self, make, HasLoopBody}, - AstNode, Direction, SyntaxElement, SyntaxKind, SyntaxToken, TokenAtOffset, WalkEvent, T, + AstNode, AstToken, Direction, SyntaxElement, SyntaxKind, SyntaxToken, TokenAtOffset, WalkEvent, + T, }; -use crate::RootDatabase; +use crate::{defs::Definition, RootDatabase}; pub use self::famous_defs::FamousDefs; @@ -29,33 +32,47 @@ pub fn item_name(db: &RootDatabase, item: ItemInNs) -> Option { } } -/// Resolves the path at the cursor token as a derive macro if it inside a token tree of a derive attribute. -pub fn try_resolve_derive_input_at( +/// Parses and returns the derive path at the cursor position in the given attribute, if it is a derive. +/// This special case is required because the derive macro is a compiler builtin that discards the input derives. +/// +/// The returned path is synthesized from TokenTree tokens and as such cannot be used with the [`Semantics`]. +pub fn get_path_in_derive_attr( sema: &hir::Semantics, - derive_attr: &ast::Attr, - cursor: &SyntaxToken, -) -> Option { - use itertools::Itertools; - if cursor.kind() != T![ident] { + attr: &ast::Attr, + cursor: &ast::Ident, +) -> Option { + let path = attr.path()?; + let tt = attr.token_tree()?; + if !tt.syntax().text_range().contains_range(cursor.syntax().text_range()) { return None; } - let tt = match derive_attr.as_simple_call() { - Some((name, tt)) - if name == "derive" && tt.syntax().text_range().contains_range(cursor.text_range()) => - { - tt - } - _ => return None, - }; - let tokens: Vec<_> = cursor + let scope = sema.scope(attr.syntax()); + let resolved_attr = sema.resolve_path(&path)?; + let derive = FamousDefs(sema, scope.krate()).core_macros_builtin_derive()?; + if PathResolution::Macro(derive) != resolved_attr { + return None; + } + get_path_at_cursor_in_tt(cursor) +} + +/// Parses the path the identifier is part of inside a token tree. +pub fn get_path_at_cursor_in_tt(cursor: &ast::Ident) -> Option { + let cursor = cursor.syntax(); + let first = cursor .siblings_with_tokens(Direction::Prev) - .flat_map(SyntaxElement::into_token) + .filter_map(SyntaxElement::into_token) .take_while(|tok| tok.kind() != T!['('] && tok.kind() != T![,]) - .collect(); - let path = ast::Path::parse(&tokens.into_iter().rev().join("")).ok()?; - sema.scope(tt.syntax()) - .speculative_resolve_as_mac(&path) - .filter(|mac| mac.kind() == hir::MacroKind::Derive) + .last()?; + let path_tokens = first + .siblings_with_tokens(Direction::Next) + .filter_map(SyntaxElement::into_token) + .take_while(|tok| tok != cursor); + + syntax::hacks::parse_expr_from_str(&path_tokens.chain(iter::once(cursor.clone())).join("")) + .and_then(|expr| match expr { + ast::Expr::PathExpr(it) => it.path(), + _ => None, + }) } /// Picks the token with the highest rank returned by the passed in function. @@ -65,6 +82,9 @@ pub fn pick_best_token( ) -> Option { tokens.max_by_key(move |t| f(t.kind())) } +pub fn pick_token(mut tokens: TokenAtOffset) -> Option { + tokens.find_map(T::cast) +} /// Converts the mod path struct into its ast representation. pub fn mod_path_to_ast(path: &hir::ModPath) -> ast::Path { @@ -85,7 +105,7 @@ pub fn mod_path_to_ast(path: &hir::ModPath) -> ast::Path { segments.extend( path.segments() .iter() - .map(|segment| make::path_segment(make::name_ref(&segment.to_string()))), + .map(|segment| make::path_segment(make::name_ref(&segment.to_smol_str()))), ); make::path_from_segments(segments, is_abs) } @@ -94,7 +114,7 @@ pub fn mod_path_to_ast(path: &hir::ModPath) -> ast::Path { pub fn visit_file_defs( sema: &Semantics, file_id: FileId, - cb: &mut dyn FnMut(Either), + cb: &mut dyn FnMut(Definition), ) { let db = sema.db; let module = match sema.to_module_def(file_id) { @@ -106,12 +126,12 @@ pub fn visit_file_defs( if let ModuleDef::Module(submodule) = def { if let hir::ModuleSource::Module(_) = submodule.definition_source(db).value { defs.extend(submodule.declarations(db)); - submodule.impl_defs(db).into_iter().for_each(|impl_| cb(Either::Right(impl_))); + submodule.impl_defs(db).into_iter().for_each(|impl_| cb(impl_.into())); } } - cb(Either::Left(def)); + cb(def.into()); } - module.impl_defs(db).into_iter().for_each(|impl_| cb(Either::Right(impl_))); + module.impl_defs(db).into_iter().for_each(|impl_| cb(impl_.into())); } #[derive(Clone, Copy, Debug, PartialEq, Eq)] @@ -209,6 +229,7 @@ pub fn for_each_tail_expr(expr: &ast::Expr, cb: &mut dyn FnMut(&ast::Expr)) { | ast::Expr::TryExpr(_) | ast::Expr::TupleExpr(_) | ast::Expr::WhileExpr(_) + | ast::Expr::LetExpr(_) | ast::Expr::YieldExpr(_) => cb(expr), } } @@ -255,3 +276,48 @@ pub fn for_each_break_expr( } } } + +/// Checks if the given lint is equal or is contained by the other lint which may or may not be a group. +pub fn lint_eq_or_in_group(lint: &str, lint_is: &str) -> bool { + if lint == lint_is { + return true; + } + + if let Some(group) = generated_lints::DEFAULT_LINT_GROUPS + .iter() + .chain(generated_lints::CLIPPY_LINT_GROUPS.iter()) + .chain(generated_lints::RUSTDOC_LINT_GROUPS.iter()) + .find(|&check| check.lint.label == lint_is) + { + group.children.contains(&lint) + } else { + false + } +} + +/// Parses the input token tree as comma separated plain paths. +pub fn parse_tt_as_comma_sep_paths(input: ast::TokenTree) -> Option> { + let r_paren = input.r_paren_token(); + let tokens = + input.syntax().children_with_tokens().skip(1).map_while(|it| match it.into_token() { + // seeing a keyword means the attribute is unclosed so stop parsing here + Some(tok) if tok.kind().is_keyword() => None, + // don't include the right token tree parenthesis if it exists + tok @ Some(_) if tok == r_paren => None, + // only nodes that we can find are other TokenTrees, those are unexpected in this parse though + None => None, + Some(tok) => Some(tok), + }); + let input_expressions = tokens.into_iter().group_by(|tok| tok.kind() == T![,]); + let paths = input_expressions + .into_iter() + .filter_map(|(is_sep, group)| (!is_sep).then(|| group)) + .filter_map(|mut tokens| { + syntax::hacks::parse_expr_from_str(&tokens.join("")).and_then(|expr| match expr { + ast::Expr::PathExpr(it) => it.path(), + _ => None, + }) + }) + .collect(); + Some(paths) +}