1 // Copyright 2012-2014 The Rust Project Developers. See the COPYRIGHT
2 // file at the top-level directory of this distribution and at
3 // http://rust-lang.org/COPYRIGHT.
5 // Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
6 // http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
7 // <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
8 // option. This file may not be copied, modified, or distributed
9 // except according to those terms.
11 //! A classic liveness analysis based on dataflow over the AST. Computes,
12 //! for each local variable in a function, whether that variable is live
13 //! at a given point. Program execution points are identified by their
18 //! The basic model is that each local variable is assigned an index. We
19 //! represent sets of local variables using a vector indexed by this
20 //! index. The value in the vector is either 0, indicating the variable
21 //! is dead, or the id of an expression that uses the variable.
23 //! We conceptually walk over the AST in reverse execution order. If we
24 //! find a use of a variable, we add it to the set of live variables. If
25 //! we find an assignment to a variable, we remove it from the set of live
26 //! variables. When we have to merge two flows, we take the union of
27 //! those two flows---if the variable is live on both paths, we simply
28 //! pick one id. In the event of loops, we continue doing this until a
29 //! fixed point is reached.
31 //! ## Checking initialization
33 //! At the function entry point, all variables must be dead. If this is
34 //! not the case, we can report an error using the id found in the set of
35 //! live variables, which identifies a use of the variable which is not
36 //! dominated by an assignment.
40 //! After each explicit move, the variable must be dead.
42 //! ## Computing last uses
44 //! Any use of the variable where the variable is dead afterwards is a
47 //! # Implementation details
49 //! The actual implementation contains two (nested) walks over the AST.
50 //! The outer walk has the job of building up the ir_maps instance for the
51 //! enclosing function. On the way down the tree, it identifies those AST
52 //! nodes and variable IDs that will be needed for the liveness analysis
53 //! and assigns them contiguous IDs. The liveness id for an AST node is
54 //! called a `live_node` (it's a newtype'd u32) and the id for a variable
55 //! is called a `variable` (another newtype'd u32).
57 //! On the way back up the tree, as we are about to exit from a function
58 //! declaration we allocate a `liveness` instance. Now that we know
59 //! precisely how many nodes and variables we need, we can allocate all
60 //! the various arrays that we will need to precisely the right size. We then
61 //! perform the actual propagation on the `liveness` instance.
63 //! This propagation is encoded in the various `propagate_through_*()`
64 //! methods. It effectively does a reverse walk of the AST; whenever we
65 //! reach a loop node, we iterate until a fixed point is reached.
67 //! ## The `Users` struct
69 //! At each live node `N`, we track three pieces of information for each
70 //! variable `V` (these are encapsulated in the `Users` struct):
72 //! - `reader`: the `LiveNode` ID of some node which will read the value
73 //! that `V` holds on entry to `N`. Formally: a node `M` such
74 //! that there exists a path `P` from `N` to `M` where `P` does not
75 //! write `V`. If the `reader` is `invalid_node()`, then the current
76 //! value will never be read (the variable is dead, essentially).
78 //! - `writer`: the `LiveNode` ID of some node which will write the
79 //! variable `V` and which is reachable from `N`. Formally: a node `M`
80 //! such that there exists a path `P` from `N` to `M` and `M` writes
81 //! `V`. If the `writer` is `invalid_node()`, then there is no writer
82 //! of `V` that follows `N`.
84 //! - `used`: a boolean value indicating whether `V` is *used*. We
85 //! distinguish a *read* from a *use* in that a *use* is some read that
86 //! is not just used to generate a new value. For example, `x += 1` is
87 //! a read but not a use. This is used to generate better warnings.
89 //! ## Special Variables
91 //! We generate various special variables for various, well, special purposes.
92 //! These are described in the `specials` struct:
94 //! - `exit_ln`: a live node that is generated to represent every 'exit' from
95 //! the function, whether it be by explicit return, panic, or other means.
97 //! - `fallthrough_ln`: a live node that represents a fallthrough
99 //! - `clean_exit_var`: a synthetic variable that is only 'read' from the
100 //! fallthrough node. It is only live if the function could converge
101 //! via means other than an explicit `return` expression. That is, it is
102 //! only dead if the end of the function's block can never be reached.
103 //! It is the responsibility of typeck to ensure that there are no
104 //! `return` expressions in a function declared as diverging.
105 use self::LoopKind::*;
106 use self::LiveNodeKind::*;
107 use self::VarKind::*;
110 use ty::{self, TyCtxt};
112 use errors::Applicability;
113 use util::nodemap::{NodeMap, HirIdMap, HirIdSet};
115 use std::collections::VecDeque;
117 use std::io::prelude::*;
120 use syntax::ast::{self, NodeId};
122 use syntax::symbol::keywords;
123 use syntax_pos::Span;
125 use hir::{Expr, HirId};
127 use hir::intravisit::{self, Visitor, FnKind, NestedVisitorMap};
129 /// For use with `propagate_through_loop`.
131 /// An endless `loop` loop.
133 /// A `while` loop, with the given expression as condition.
137 #[derive(Copy, Clone, PartialEq)]
138 struct Variable(u32);
140 #[derive(Copy, Clone, PartialEq)]
141 struct LiveNode(u32);
144 fn get(&self) -> usize { self.0 as usize }
148 fn get(&self) -> usize { self.0 as usize }
151 #[derive(Copy, Clone, PartialEq, Debug)]
159 fn live_node_kind_to_string(lnk: LiveNodeKind, tcx: TyCtxt) -> String {
160 let cm = tcx.sess.source_map();
163 format!("Free var node [{}]", cm.span_to_string(s))
166 format!("Expr node [{}]", cm.span_to_string(s))
169 format!("Var def node [{}]", cm.span_to_string(s))
171 ExitNode => "Exit node".to_string(),
175 impl<'a, 'tcx> Visitor<'tcx> for IrMaps<'a, 'tcx> {
176 fn nested_visit_map<'this>(&'this mut self) -> NestedVisitorMap<'this, 'tcx> {
177 NestedVisitorMap::OnlyBodies(&self.tcx.hir)
180 fn visit_fn(&mut self, fk: FnKind<'tcx>, fd: &'tcx hir::FnDecl,
181 b: hir::BodyId, s: Span, id: NodeId) {
182 visit_fn(self, fk, fd, b, s, id);
185 fn visit_local(&mut self, l: &'tcx hir::Local) { visit_local(self, l); }
186 fn visit_expr(&mut self, ex: &'tcx Expr) { visit_expr(self, ex); }
187 fn visit_arm(&mut self, a: &'tcx hir::Arm) { visit_arm(self, a); }
190 pub fn check_crate<'a, 'tcx>(tcx: TyCtxt<'a, 'tcx, 'tcx>) {
191 tcx.hir.krate().visit_all_item_likes(&mut IrMaps::new(tcx).as_deep_visitor());
192 tcx.sess.abort_if_errors();
195 impl fmt::Debug for LiveNode {
196 fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
197 write!(f, "ln({})", self.get())
201 impl fmt::Debug for Variable {
202 fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
203 write!(f, "v({})", self.get())
207 // ______________________________________________________________________
210 // This is the first pass and the one that drives the main
211 // computation. It walks up and down the IR once. On the way down,
212 // we count for each function the number of variables as well as
213 // liveness nodes. A liveness node is basically an expression or
214 // capture clause that does something of interest: either it has
215 // interesting control flow or it uses/defines a local variable.
217 // On the way back up, at each function node we create liveness sets
218 // (we now know precisely how big to make our various vectors and so
219 // forth) and then do the data-flow propagation to compute the set
220 // of live variables at each program point.
222 // Finally, we run back over the IR one last time and, using the
223 // computed liveness, check various safety conditions. For example,
224 // there must be no live nodes at the definition site for a variable
225 // unless it has an initializer. Similarly, each non-mutable local
226 // variable must not be assigned if there is some successor
227 // assignment. And so forth.
230 fn is_valid(&self) -> bool {
235 fn invalid_node() -> LiveNode { LiveNode(u32::MAX) }
242 #[derive(Copy, Clone, Debug)]
249 #[derive(Copy, Clone, Debug)]
251 Arg(HirId, ast::Name),
256 struct IrMaps<'a, 'tcx: 'a> {
257 tcx: TyCtxt<'a, 'tcx, 'tcx>,
259 num_live_nodes: usize,
261 live_node_map: HirIdMap<LiveNode>,
262 variable_map: HirIdMap<Variable>,
263 capture_info_map: NodeMap<Rc<Vec<CaptureInfo>>>,
264 var_kinds: Vec<VarKind>,
265 lnks: Vec<LiveNodeKind>,
268 impl<'a, 'tcx> IrMaps<'a, 'tcx> {
269 fn new(tcx: TyCtxt<'a, 'tcx, 'tcx>) -> IrMaps<'a, 'tcx> {
274 live_node_map: HirIdMap(),
275 variable_map: HirIdMap(),
276 capture_info_map: NodeMap(),
277 var_kinds: Vec::new(),
282 fn add_live_node(&mut self, lnk: LiveNodeKind) -> LiveNode {
283 let ln = LiveNode(self.num_live_nodes as u32);
285 self.num_live_nodes += 1;
287 debug!("{:?} is of kind {}", ln,
288 live_node_kind_to_string(lnk, self.tcx));
293 fn add_live_node_for_node(&mut self, hir_id: HirId, lnk: LiveNodeKind) {
294 let ln = self.add_live_node(lnk);
295 self.live_node_map.insert(hir_id, ln);
297 debug!("{:?} is node {:?}", ln, hir_id);
300 fn add_variable(&mut self, vk: VarKind) -> Variable {
301 let v = Variable(self.num_vars as u32);
302 self.var_kinds.push(vk);
306 Local(LocalInfo { id: node_id, .. }) | Arg(node_id, _) => {
307 self.variable_map.insert(node_id, v);
312 debug!("{:?} is {:?}", v, vk);
317 fn variable(&self, hir_id: HirId, span: Span) -> Variable {
318 match self.variable_map.get(&hir_id) {
321 span_bug!(span, "no variable registered for id {:?}", hir_id);
326 fn variable_name(&self, var: Variable) -> String {
327 match self.var_kinds[var.get()] {
328 Local(LocalInfo { name, .. }) | Arg(_, name) => {
331 CleanExit => "<clean-exit>".to_string()
335 fn variable_is_shorthand(&self, var: Variable) -> bool {
336 match self.var_kinds[var.get()] {
337 Local(LocalInfo { is_shorthand, .. }) => is_shorthand,
338 Arg(..) | CleanExit => false
342 fn set_captures(&mut self, node_id: NodeId, cs: Vec<CaptureInfo>) {
343 self.capture_info_map.insert(node_id, Rc::new(cs));
346 fn lnk(&self, ln: LiveNode) -> LiveNodeKind {
351 fn visit_fn<'a, 'tcx: 'a>(ir: &mut IrMaps<'a, 'tcx>,
353 decl: &'tcx hir::FnDecl,
354 body_id: hir::BodyId,
359 // swap in a new set of IR maps for this function body:
360 let mut fn_maps = IrMaps::new(ir.tcx);
362 // Don't run unused pass for #[derive()]
363 if let FnKind::Method(..) = fk {
364 let parent = ir.tcx.hir.get_parent(id);
365 if let Some(hir::map::Node::NodeItem(i)) = ir.tcx.hir.find(parent) {
366 if i.attrs.iter().any(|a| a.check_name("automatically_derived")) {
372 debug!("creating fn_maps: {:?}", &fn_maps as *const IrMaps);
374 let body = ir.tcx.hir.body(body_id);
376 for arg in &body.arguments {
377 arg.pat.each_binding(|_bm, hir_id, _x, ident| {
378 debug!("adding argument {:?}", hir_id);
379 fn_maps.add_variable(Arg(hir_id, ident.name));
383 // gather up the various local variables, significant expressions,
385 intravisit::walk_fn(&mut fn_maps, fk, decl, body_id, sp, id);
388 let mut lsets = Liveness::new(&mut fn_maps, body_id);
389 let entry_ln = lsets.compute(&body.value);
391 // check for various error conditions
392 lsets.visit_body(body);
393 lsets.warn_about_unused_args(body, entry_ln);
396 fn add_from_pat<'a, 'tcx>(ir: &mut IrMaps<'a, 'tcx>, pat: &P<hir::Pat>) {
397 // For struct patterns, take note of which fields used shorthand
398 // (`x` rather than `x: x`).
399 let mut shorthand_field_ids = HirIdSet();
400 let mut pats = VecDeque::new();
402 while let Some(pat) = pats.pop_front() {
405 Binding(_, _, _, ref inner_pat) => {
406 pats.extend(inner_pat.iter());
408 Struct(_, ref fields, _) => {
409 for field in fields {
410 if field.node.is_shorthand {
411 shorthand_field_ids.insert(field.node.pat.hir_id);
415 Ref(ref inner_pat, _) |
416 Box(ref inner_pat) => {
417 pats.push_back(inner_pat);
419 TupleStruct(_, ref inner_pats, _) |
420 Tuple(ref inner_pats, _) => {
421 pats.extend(inner_pats.iter());
423 Slice(ref pre_pats, ref inner_pat, ref post_pats) => {
424 pats.extend(pre_pats.iter());
425 pats.extend(inner_pat.iter());
426 pats.extend(post_pats.iter());
432 pat.each_binding(|_bm, hir_id, _sp, ident| {
433 ir.add_live_node_for_node(hir_id, VarDefNode(ident.span));
434 ir.add_variable(Local(LocalInfo {
437 is_shorthand: shorthand_field_ids.contains(&hir_id)
442 fn visit_local<'a, 'tcx>(ir: &mut IrMaps<'a, 'tcx>, local: &'tcx hir::Local) {
443 add_from_pat(ir, &local.pat);
444 intravisit::walk_local(ir, local);
447 fn visit_arm<'a, 'tcx>(ir: &mut IrMaps<'a, 'tcx>, arm: &'tcx hir::Arm) {
448 for pat in &arm.pats {
449 add_from_pat(ir, pat);
451 intravisit::walk_arm(ir, arm);
454 fn visit_expr<'a, 'tcx>(ir: &mut IrMaps<'a, 'tcx>, expr: &'tcx Expr) {
456 // live nodes required for uses or definitions of variables:
457 hir::ExprKind::Path(hir::QPath::Resolved(_, ref path)) => {
458 debug!("expr {}: path that leads to {:?}", expr.id, path.def);
459 if let Def::Local(..) = path.def {
460 ir.add_live_node_for_node(expr.hir_id, ExprNode(expr.span));
462 intravisit::walk_expr(ir, expr);
464 hir::ExprKind::Closure(..) => {
465 // Interesting control flow (for loops can contain labeled
466 // breaks or continues)
467 ir.add_live_node_for_node(expr.hir_id, ExprNode(expr.span));
469 // Make a live_node for each captured variable, with the span
470 // being the location that the variable is used. This results
471 // in better error messages than just pointing at the closure
472 // construction site.
473 let mut call_caps = Vec::new();
474 ir.tcx.with_freevars(expr.id, |freevars| {
476 if let Def::Local(rv) = fv.def {
477 let fv_ln = ir.add_live_node(FreeVarNode(fv.span));
478 let var_hid = ir.tcx.hir.node_to_hir_id(rv);
479 call_caps.push(CaptureInfo { ln: fv_ln, var_hid });
483 ir.set_captures(expr.id, call_caps);
485 intravisit::walk_expr(ir, expr);
488 // live nodes required for interesting control flow:
489 hir::ExprKind::If(..) |
490 hir::ExprKind::Match(..) |
491 hir::ExprKind::While(..) |
492 hir::ExprKind::Loop(..) => {
493 ir.add_live_node_for_node(expr.hir_id, ExprNode(expr.span));
494 intravisit::walk_expr(ir, expr);
496 hir::ExprKind::Binary(op, ..) if op.node.is_lazy() => {
497 ir.add_live_node_for_node(expr.hir_id, ExprNode(expr.span));
498 intravisit::walk_expr(ir, expr);
501 // otherwise, live nodes are not required:
502 hir::ExprKind::Index(..) |
503 hir::ExprKind::Field(..) |
504 hir::ExprKind::Array(..) |
505 hir::ExprKind::Call(..) |
506 hir::ExprKind::MethodCall(..) |
507 hir::ExprKind::Tup(..) |
508 hir::ExprKind::Binary(..) |
509 hir::ExprKind::AddrOf(..) |
510 hir::ExprKind::Cast(..) |
511 hir::ExprKind::Unary(..) |
512 hir::ExprKind::Break(..) |
513 hir::ExprKind::Continue(_) |
514 hir::ExprKind::Lit(_) |
515 hir::ExprKind::Ret(..) |
516 hir::ExprKind::Block(..) |
517 hir::ExprKind::Assign(..) |
518 hir::ExprKind::AssignOp(..) |
519 hir::ExprKind::Struct(..) |
520 hir::ExprKind::Repeat(..) |
521 hir::ExprKind::InlineAsm(..) |
522 hir::ExprKind::Box(..) |
523 hir::ExprKind::Yield(..) |
524 hir::ExprKind::Type(..) |
525 hir::ExprKind::Path(hir::QPath::TypeRelative(..)) => {
526 intravisit::walk_expr(ir, expr);
531 // ______________________________________________________________________
532 // Computing liveness sets
534 // Actually we compute just a bit more than just liveness, but we use
535 // the same basic propagation framework in all cases.
537 #[derive(Clone, Copy)]
544 fn invalid_users() -> Users {
546 reader: invalid_node(),
547 writer: invalid_node(),
552 #[derive(Copy, Clone)]
555 fallthrough_ln: LiveNode,
556 clean_exit_var: Variable
559 const ACC_READ: u32 = 1;
560 const ACC_WRITE: u32 = 2;
561 const ACC_USE: u32 = 4;
563 struct Liveness<'a, 'tcx: 'a> {
564 ir: &'a mut IrMaps<'a, 'tcx>,
565 tables: &'a ty::TypeckTables<'tcx>,
567 successors: Vec<LiveNode>,
570 // mappings from loop node ID to LiveNode
571 // ("break" label should map to loop node ID,
572 // it probably doesn't now)
573 break_ln: NodeMap<LiveNode>,
574 cont_ln: NodeMap<LiveNode>,
577 impl<'a, 'tcx> Liveness<'a, 'tcx> {
578 fn new(ir: &'a mut IrMaps<'a, 'tcx>, body: hir::BodyId) -> Liveness<'a, 'tcx> {
579 // Special nodes and variables:
580 // - exit_ln represents the end of the fn, either by return or panic
581 // - implicit_ret_var is a pseudo-variable that represents
582 // an implicit return
583 let specials = Specials {
584 exit_ln: ir.add_live_node(ExitNode),
585 fallthrough_ln: ir.add_live_node(ExitNode),
586 clean_exit_var: ir.add_variable(CleanExit)
589 let tables = ir.tcx.body_tables(body);
591 let num_live_nodes = ir.num_live_nodes;
592 let num_vars = ir.num_vars;
598 successors: vec![invalid_node(); num_live_nodes],
599 users: vec![invalid_users(); num_live_nodes * num_vars],
605 fn live_node(&self, hir_id: HirId, span: Span) -> LiveNode {
606 match self.ir.live_node_map.get(&hir_id) {
609 // This must be a mismatch between the ir_map construction
610 // above and the propagation code below; the two sets of
611 // code have to agree about which AST nodes are worth
612 // creating liveness nodes for.
615 "no live node registered for node {:?}",
621 fn variable(&self, hir_id: HirId, span: Span) -> Variable {
622 self.ir.variable(hir_id, span)
625 fn pat_bindings<F>(&mut self, pat: &hir::Pat, mut f: F) where
626 F: FnMut(&mut Liveness<'a, 'tcx>, LiveNode, Variable, Span, HirId),
628 pat.each_binding(|_bm, hir_id, sp, n| {
629 let ln = self.live_node(hir_id, sp);
630 let var = self.variable(hir_id, n.span);
631 f(self, ln, var, n.span, hir_id);
635 fn arm_pats_bindings<F>(&mut self, pat: Option<&hir::Pat>, f: F) where
636 F: FnMut(&mut Liveness<'a, 'tcx>, LiveNode, Variable, Span, HirId),
638 if let Some(pat) = pat {
639 self.pat_bindings(pat, f);
643 fn define_bindings_in_pat(&mut self, pat: &hir::Pat, succ: LiveNode)
645 self.define_bindings_in_arm_pats(Some(pat), succ)
648 fn define_bindings_in_arm_pats(&mut self, pat: Option<&hir::Pat>, succ: LiveNode)
651 self.arm_pats_bindings(pat, |this, ln, var, _sp, _id| {
652 this.init_from_succ(ln, succ);
653 this.define(ln, var);
659 fn idx(&self, ln: LiveNode, var: Variable) -> usize {
660 ln.get() * self.ir.num_vars + var.get()
663 fn live_on_entry(&self, ln: LiveNode, var: Variable)
664 -> Option<LiveNodeKind> {
665 assert!(ln.is_valid());
666 let reader = self.users[self.idx(ln, var)].reader;
667 if reader.is_valid() {Some(self.ir.lnk(reader))} else {None}
671 Is this variable live on entry to any of its successor nodes?
673 fn live_on_exit(&self, ln: LiveNode, var: Variable)
674 -> Option<LiveNodeKind> {
675 let successor = self.successors[ln.get()];
676 self.live_on_entry(successor, var)
679 fn used_on_entry(&self, ln: LiveNode, var: Variable) -> bool {
680 assert!(ln.is_valid());
681 self.users[self.idx(ln, var)].used
684 fn assigned_on_entry(&self, ln: LiveNode, var: Variable)
685 -> Option<LiveNodeKind> {
686 assert!(ln.is_valid());
687 let writer = self.users[self.idx(ln, var)].writer;
688 if writer.is_valid() {Some(self.ir.lnk(writer))} else {None}
691 fn assigned_on_exit(&self, ln: LiveNode, var: Variable)
692 -> Option<LiveNodeKind> {
693 let successor = self.successors[ln.get()];
694 self.assigned_on_entry(successor, var)
697 fn indices2<F>(&mut self, ln: LiveNode, succ_ln: LiveNode, mut op: F) where
698 F: FnMut(&mut Liveness<'a, 'tcx>, usize, usize),
700 let node_base_idx = self.idx(ln, Variable(0));
701 let succ_base_idx = self.idx(succ_ln, Variable(0));
702 for var_idx in 0..self.ir.num_vars {
703 op(self, node_base_idx + var_idx, succ_base_idx + var_idx);
707 fn write_vars<F>(&self,
711 -> io::Result<()> where
712 F: FnMut(usize) -> LiveNode,
714 let node_base_idx = self.idx(ln, Variable(0));
715 for var_idx in 0..self.ir.num_vars {
716 let idx = node_base_idx + var_idx;
717 if test(idx).is_valid() {
718 write!(wr, " {:?}", Variable(var_idx as u32))?;
725 #[allow(unused_must_use)]
726 fn ln_str(&self, ln: LiveNode) -> String {
727 let mut wr = Vec::new();
729 let wr = &mut wr as &mut dyn Write;
730 write!(wr, "[ln({:?}) of kind {:?} reads", ln.get(), self.ir.lnk(ln));
731 self.write_vars(wr, ln, |idx| self.users[idx].reader);
732 write!(wr, " writes");
733 self.write_vars(wr, ln, |idx| self.users[idx].writer);
734 write!(wr, " precedes {:?}]", self.successors[ln.get()]);
736 String::from_utf8(wr).unwrap()
739 fn init_empty(&mut self, ln: LiveNode, succ_ln: LiveNode) {
740 self.successors[ln.get()] = succ_ln;
742 // It is not necessary to initialize the
743 // values to empty because this is the value
744 // they have when they are created, and the sets
745 // only grow during iterations.
747 // self.indices(ln) { |idx|
748 // self.users[idx] = invalid_users();
752 fn init_from_succ(&mut self, ln: LiveNode, succ_ln: LiveNode) {
753 // more efficient version of init_empty() / merge_from_succ()
754 self.successors[ln.get()] = succ_ln;
756 self.indices2(ln, succ_ln, |this, idx, succ_idx| {
757 this.users[idx] = this.users[succ_idx]
759 debug!("init_from_succ(ln={}, succ={})",
760 self.ln_str(ln), self.ln_str(succ_ln));
763 fn merge_from_succ(&mut self,
768 if ln == succ_ln { return false; }
770 let mut changed = false;
771 self.indices2(ln, succ_ln, |this, idx, succ_idx| {
772 changed |= copy_if_invalid(this.users[succ_idx].reader,
773 &mut this.users[idx].reader);
774 changed |= copy_if_invalid(this.users[succ_idx].writer,
775 &mut this.users[idx].writer);
776 if this.users[succ_idx].used && !this.users[idx].used {
777 this.users[idx].used = true;
782 debug!("merge_from_succ(ln={:?}, succ={}, first_merge={}, changed={})",
783 ln, self.ln_str(succ_ln), first_merge, changed);
786 fn copy_if_invalid(src: LiveNode, dst: &mut LiveNode) -> bool {
787 if src.is_valid() && !dst.is_valid() {
796 // Indicates that a local variable was *defined*; we know that no
797 // uses of the variable can precede the definition (resolve checks
798 // this) so we just clear out all the data.
799 fn define(&mut self, writer: LiveNode, var: Variable) {
800 let idx = self.idx(writer, var);
801 self.users[idx].reader = invalid_node();
802 self.users[idx].writer = invalid_node();
804 debug!("{:?} defines {:?} (idx={}): {}", writer, var,
805 idx, self.ln_str(writer));
808 // Either read, write, or both depending on the acc bitset
809 fn acc(&mut self, ln: LiveNode, var: Variable, acc: u32) {
810 debug!("{:?} accesses[{:x}] {:?}: {}",
811 ln, acc, var, self.ln_str(ln));
813 let idx = self.idx(ln, var);
814 let user = &mut self.users[idx];
816 if (acc & ACC_WRITE) != 0 {
817 user.reader = invalid_node();
821 // Important: if we both read/write, must do read second
822 // or else the write will override.
823 if (acc & ACC_READ) != 0 {
827 if (acc & ACC_USE) != 0 {
832 // _______________________________________________________________________
834 fn compute(&mut self, body: &hir::Expr) -> LiveNode {
835 // if there is a `break` or `again` at the top level, then it's
836 // effectively a return---this only occurs in `for` loops,
837 // where the body is really a closure.
839 debug!("compute: using id for body, {}", self.ir.tcx.hir.node_to_pretty_string(body.id));
841 let exit_ln = self.s.exit_ln;
843 self.break_ln.insert(body.id, exit_ln);
844 self.cont_ln.insert(body.id, exit_ln);
846 // the fallthrough exit is only for those cases where we do not
847 // explicitly return:
849 self.init_from_succ(s.fallthrough_ln, s.exit_ln);
850 self.acc(s.fallthrough_ln, s.clean_exit_var, ACC_READ);
852 let entry_ln = self.propagate_through_expr(body, s.fallthrough_ln);
854 // hack to skip the loop unless debug! is enabled:
855 debug!("^^ liveness computation results for body {} (entry={:?})",
857 for ln_idx in 0..self.ir.num_live_nodes {
858 debug!("{:?}", self.ln_str(LiveNode(ln_idx as u32)));
867 fn propagate_through_block(&mut self, blk: &hir::Block, succ: LiveNode)
869 if blk.targeted_by_break {
870 self.break_ln.insert(blk.id, succ);
872 let succ = self.propagate_through_opt_expr(blk.expr.as_ref().map(|e| &**e), succ);
873 blk.stmts.iter().rev().fold(succ, |succ, stmt| {
874 self.propagate_through_stmt(stmt, succ)
878 fn propagate_through_stmt(&mut self, stmt: &hir::Stmt, succ: LiveNode)
881 hir::StmtKind::Decl(ref decl, _) => {
882 self.propagate_through_decl(&decl, succ)
885 hir::StmtKind::Expr(ref expr, _) | hir::StmtKind::Semi(ref expr, _) => {
886 self.propagate_through_expr(&expr, succ)
891 fn propagate_through_decl(&mut self, decl: &hir::Decl, succ: LiveNode)
894 hir::DeclKind::Local(ref local) => {
895 self.propagate_through_local(&local, succ)
897 hir::DeclKind::Item(_) => succ,
901 fn propagate_through_local(&mut self, local: &hir::Local, succ: LiveNode)
903 // Note: we mark the variable as defined regardless of whether
904 // there is an initializer. Initially I had thought to only mark
905 // the live variable as defined if it was initialized, and then we
906 // could check for uninit variables just by scanning what is live
907 // at the start of the function. But that doesn't work so well for
908 // immutable variables defined in a loop:
909 // loop { let x; x = 5; }
910 // because the "assignment" loops back around and generates an error.
912 // So now we just check that variables defined w/o an
913 // initializer are not live at the point of their
914 // initialization, which is mildly more complex than checking
915 // once at the func header but otherwise equivalent.
917 let succ = self.propagate_through_opt_expr(local.init.as_ref().map(|e| &**e), succ);
918 self.define_bindings_in_pat(&local.pat, succ)
921 fn propagate_through_exprs(&mut self, exprs: &[Expr], succ: LiveNode)
923 exprs.iter().rev().fold(succ, |succ, expr| {
924 self.propagate_through_expr(&expr, succ)
928 fn propagate_through_opt_expr(&mut self,
929 opt_expr: Option<&Expr>,
932 opt_expr.map_or(succ, |expr| self.propagate_through_expr(expr, succ))
935 fn propagate_through_expr(&mut self, expr: &Expr, succ: LiveNode)
937 debug!("propagate_through_expr: {}", self.ir.tcx.hir.node_to_pretty_string(expr.id));
940 // Interesting cases with control flow or which gen/kill
941 hir::ExprKind::Path(hir::QPath::Resolved(_, ref path)) => {
942 self.access_path(expr.hir_id, path, succ, ACC_READ | ACC_USE)
945 hir::ExprKind::Field(ref e, _) => {
946 self.propagate_through_expr(&e, succ)
949 hir::ExprKind::Closure(.., blk_id, _, _) => {
950 debug!("{} is an ExprKind::Closure", self.ir.tcx.hir.node_to_pretty_string(expr.id));
952 // The next-node for a break is the successor of the entire
953 // loop. The next-node for a continue is the top of this loop.
954 let node = self.live_node(expr.hir_id, expr.span);
958 self.break_ln.insert(blk_id.node_id, break_ln);
959 self.cont_ln.insert(blk_id.node_id, cont_ln);
961 // the construction of a closure itself is not important,
962 // but we have to consider the closed over variables.
963 let caps = match self.ir.capture_info_map.get(&expr.id) {
964 Some(caps) => caps.clone(),
966 span_bug!(expr.span, "no registered caps");
969 caps.iter().rev().fold(succ, |succ, cap| {
970 self.init_from_succ(cap.ln, succ);
971 let var = self.variable(cap.var_hid, expr.span);
972 self.acc(cap.ln, var, ACC_READ | ACC_USE);
977 hir::ExprKind::If(ref cond, ref then, ref els) => {
991 let else_ln = self.propagate_through_opt_expr(els.as_ref().map(|e| &**e), succ);
992 let then_ln = self.propagate_through_expr(&then, succ);
993 let ln = self.live_node(expr.hir_id, expr.span);
994 self.init_from_succ(ln, else_ln);
995 self.merge_from_succ(ln, then_ln, false);
996 self.propagate_through_expr(&cond, ln)
999 hir::ExprKind::While(ref cond, ref blk, _) => {
1000 self.propagate_through_loop(expr, WhileLoop(&cond), &blk, succ)
1003 // Note that labels have been resolved, so we don't need to look
1004 // at the label ident
1005 hir::ExprKind::Loop(ref blk, _, _) => {
1006 self.propagate_through_loop(expr, LoopLoop, &blk, succ)
1009 hir::ExprKind::Match(ref e, ref arms, _) => {
1024 let ln = self.live_node(expr.hir_id, expr.span);
1025 self.init_empty(ln, succ);
1026 let mut first_merge = true;
1029 self.propagate_through_expr(&arm.body, succ);
1031 self.propagate_through_opt_expr(arm.guard.as_ref().map(|e| &**e), body_succ);
1032 // only consider the first pattern; any later patterns must have
1033 // the same bindings, and we also consider the first pattern to be
1034 // the "authoritative" set of ids
1036 self.define_bindings_in_arm_pats(arm.pats.first().map(|p| &**p),
1038 self.merge_from_succ(ln, arm_succ, first_merge);
1039 first_merge = false;
1041 self.propagate_through_expr(&e, ln)
1044 hir::ExprKind::Ret(ref o_e) => {
1045 // ignore succ and subst exit_ln:
1046 let exit_ln = self.s.exit_ln;
1047 self.propagate_through_opt_expr(o_e.as_ref().map(|e| &**e), exit_ln)
1050 hir::ExprKind::Break(label, ref opt_expr) => {
1051 // Find which label this break jumps to
1052 let target = match label.target_id {
1053 Ok(node_id) => self.break_ln.get(&node_id),
1054 Err(err) => span_bug!(expr.span, "loop scope error: {}", err),
1057 // Now that we know the label we're going to,
1058 // look it up in the break loop nodes table
1061 Some(b) => self.propagate_through_opt_expr(opt_expr.as_ref().map(|e| &**e), b),
1062 None => span_bug!(expr.span, "break to unknown label")
1066 hir::ExprKind::Continue(label) => {
1067 // Find which label this expr continues to
1068 let sc = match label.target_id {
1069 Ok(node_id) => node_id,
1070 Err(err) => span_bug!(expr.span, "loop scope error: {}", err),
1073 // Now that we know the label we're going to,
1074 // look it up in the continue loop nodes table
1076 match self.cont_ln.get(&sc) {
1078 None => span_bug!(expr.span, "continue to unknown label")
1082 hir::ExprKind::Assign(ref l, ref r) => {
1083 // see comment on places in
1084 // propagate_through_place_components()
1085 let succ = self.write_place(&l, succ, ACC_WRITE);
1086 let succ = self.propagate_through_place_components(&l, succ);
1087 self.propagate_through_expr(&r, succ)
1090 hir::ExprKind::AssignOp(_, ref l, ref r) => {
1091 // an overloaded assign op is like a method call
1092 if self.tables.is_method_call(expr) {
1093 let succ = self.propagate_through_expr(&l, succ);
1094 self.propagate_through_expr(&r, succ)
1096 // see comment on places in
1097 // propagate_through_place_components()
1098 let succ = self.write_place(&l, succ, ACC_WRITE|ACC_READ);
1099 let succ = self.propagate_through_expr(&r, succ);
1100 self.propagate_through_place_components(&l, succ)
1104 // Uninteresting cases: just propagate in rev exec order
1106 hir::ExprKind::Array(ref exprs) => {
1107 self.propagate_through_exprs(exprs, succ)
1110 hir::ExprKind::Struct(_, ref fields, ref with_expr) => {
1111 let succ = self.propagate_through_opt_expr(with_expr.as_ref().map(|e| &**e), succ);
1112 fields.iter().rev().fold(succ, |succ, field| {
1113 self.propagate_through_expr(&field.expr, succ)
1117 hir::ExprKind::Call(ref f, ref args) => {
1118 // FIXME(canndrew): This is_never should really be an is_uninhabited
1119 let succ = if self.tables.expr_ty(expr).is_never() {
1124 let succ = self.propagate_through_exprs(args, succ);
1125 self.propagate_through_expr(&f, succ)
1128 hir::ExprKind::MethodCall(.., ref args) => {
1129 // FIXME(canndrew): This is_never should really be an is_uninhabited
1130 let succ = if self.tables.expr_ty(expr).is_never() {
1135 self.propagate_through_exprs(args, succ)
1138 hir::ExprKind::Tup(ref exprs) => {
1139 self.propagate_through_exprs(exprs, succ)
1142 hir::ExprKind::Binary(op, ref l, ref r) if op.node.is_lazy() => {
1143 let r_succ = self.propagate_through_expr(&r, succ);
1145 let ln = self.live_node(expr.hir_id, expr.span);
1146 self.init_from_succ(ln, succ);
1147 self.merge_from_succ(ln, r_succ, false);
1149 self.propagate_through_expr(&l, ln)
1152 hir::ExprKind::Index(ref l, ref r) |
1153 hir::ExprKind::Binary(_, ref l, ref r) => {
1154 let r_succ = self.propagate_through_expr(&r, succ);
1155 self.propagate_through_expr(&l, r_succ)
1158 hir::ExprKind::Box(ref e) |
1159 hir::ExprKind::AddrOf(_, ref e) |
1160 hir::ExprKind::Cast(ref e, _) |
1161 hir::ExprKind::Type(ref e, _) |
1162 hir::ExprKind::Unary(_, ref e) |
1163 hir::ExprKind::Yield(ref e) |
1164 hir::ExprKind::Repeat(ref e, _) => {
1165 self.propagate_through_expr(&e, succ)
1168 hir::ExprKind::InlineAsm(ref ia, ref outputs, ref inputs) => {
1169 let succ = ia.outputs.iter().zip(outputs).rev().fold(succ, |succ, (o, output)| {
1170 // see comment on places
1171 // in propagate_through_place_components()
1173 self.propagate_through_expr(output, succ)
1175 let acc = if o.is_rw { ACC_WRITE|ACC_READ } else { ACC_WRITE };
1176 let succ = self.write_place(output, succ, acc);
1177 self.propagate_through_place_components(output, succ)
1181 // Inputs are executed first. Propagate last because of rev order
1182 self.propagate_through_exprs(inputs, succ)
1185 hir::ExprKind::Lit(..) | hir::ExprKind::Path(hir::QPath::TypeRelative(..)) => {
1189 // Note that labels have been resolved, so we don't need to look
1190 // at the label ident
1191 hir::ExprKind::Block(ref blk, _) => {
1192 self.propagate_through_block(&blk, succ)
1197 fn propagate_through_place_components(&mut self,
1203 // In general, the full flow graph structure for an
1204 // assignment/move/etc can be handled in one of two ways,
1205 // depending on whether what is being assigned is a "tracked
1206 // value" or not. A tracked value is basically a local
1207 // variable or argument.
1209 // The two kinds of graphs are:
1211 // Tracked place Untracked place
1212 // ----------------------++-----------------------
1216 // (rvalue) || (rvalue)
1219 // (write of place) || (place components)
1224 // ----------------------++-----------------------
1226 // I will cover the two cases in turn:
1230 // A tracked place is a local variable/argument `x`. In
1231 // these cases, the link_node where the write occurs is linked
1232 // to node id of `x`. The `write_place()` routine generates
1233 // the contents of this node. There are no subcomponents to
1236 // # Non-tracked places
1238 // These are places like `x[5]` or `x.f`. In that case, we
1239 // basically ignore the value which is written to but generate
1240 // reads for the components---`x` in these two examples. The
1241 // components reads are generated by
1242 // `propagate_through_place_components()` (this fn).
1246 // It is still possible to observe assignments to non-places;
1247 // these errors are detected in the later pass borrowck. We
1248 // just ignore such cases and treat them as reads.
1251 hir::ExprKind::Path(_) => succ,
1252 hir::ExprKind::Field(ref e, _) => self.propagate_through_expr(&e, succ),
1253 _ => self.propagate_through_expr(expr, succ)
1257 // see comment on propagate_through_place()
1258 fn write_place(&mut self, expr: &Expr, succ: LiveNode, acc: u32)
1261 hir::ExprKind::Path(hir::QPath::Resolved(_, ref path)) => {
1262 self.access_path(expr.hir_id, path, succ, acc)
1265 // We do not track other places, so just propagate through
1266 // to their subcomponents. Also, it may happen that
1267 // non-places occur here, because those are detected in the
1268 // later pass borrowck.
1273 fn access_var(&mut self, hir_id: HirId, nid: NodeId, succ: LiveNode, acc: u32, span: Span)
1275 let ln = self.live_node(hir_id, span);
1277 self.init_from_succ(ln, succ);
1278 let var_hid = self.ir.tcx.hir.node_to_hir_id(nid);
1279 let var = self.variable(var_hid, span);
1280 self.acc(ln, var, acc);
1285 fn access_path(&mut self, hir_id: HirId, path: &hir::Path, succ: LiveNode, acc: u32)
1288 Def::Local(nid) => {
1289 self.access_var(hir_id, nid, succ, acc, path.span)
1295 fn propagate_through_loop(&mut self,
1304 We model control flow like this:
1322 let mut first_merge = true;
1323 let ln = self.live_node(expr.hir_id, expr.span);
1324 self.init_empty(ln, succ);
1328 // If this is not a `loop` loop, then it's possible we bypass
1329 // the body altogether. Otherwise, the only way is via a `break`
1330 // in the loop body.
1331 self.merge_from_succ(ln, succ, first_merge);
1332 first_merge = false;
1335 debug!("propagate_through_loop: using id for loop body {} {}",
1336 expr.id, self.ir.tcx.hir.node_to_pretty_string(body.id));
1338 let break_ln = succ;
1340 self.break_ln.insert(expr.id, break_ln);
1341 self.cont_ln.insert(expr.id, cont_ln);
1343 let cond_ln = match kind {
1345 WhileLoop(ref cond) => self.propagate_through_expr(&cond, ln),
1347 let body_ln = self.propagate_through_block(body, cond_ln);
1349 // repeat until fixed point is reached:
1350 while self.merge_from_succ(ln, body_ln, first_merge) {
1351 first_merge = false;
1353 let new_cond_ln = match kind {
1355 WhileLoop(ref cond) => {
1356 self.propagate_through_expr(&cond, ln)
1359 assert!(cond_ln == new_cond_ln);
1360 assert!(body_ln == self.propagate_through_block(body, cond_ln));
1367 // _______________________________________________________________________
1368 // Checking for error conditions
1370 impl<'a, 'tcx> Visitor<'tcx> for Liveness<'a, 'tcx> {
1371 fn nested_visit_map<'this>(&'this mut self) -> NestedVisitorMap<'this, 'tcx> {
1372 NestedVisitorMap::None
1375 fn visit_local(&mut self, l: &'tcx hir::Local) {
1376 check_local(self, l);
1378 fn visit_expr(&mut self, ex: &'tcx Expr) {
1379 check_expr(self, ex);
1381 fn visit_arm(&mut self, a: &'tcx hir::Arm) {
1386 fn check_local<'a, 'tcx>(this: &mut Liveness<'a, 'tcx>, local: &'tcx hir::Local) {
1389 this.warn_about_unused_or_dead_vars_in_pat(&local.pat);
1392 this.pat_bindings(&local.pat, |this, ln, var, sp, id| {
1393 let span = local.pat.simple_ident().map_or(sp, |ident| ident.span);
1394 this.warn_about_unused(span, id, ln, var);
1399 intravisit::walk_local(this, local);
1402 fn check_arm<'a, 'tcx>(this: &mut Liveness<'a, 'tcx>, arm: &'tcx hir::Arm) {
1403 // only consider the first pattern; any later patterns must have
1404 // the same bindings, and we also consider the first pattern to be
1405 // the "authoritative" set of ids
1406 this.arm_pats_bindings(arm.pats.first().map(|p| &**p), |this, ln, var, sp, id| {
1407 this.warn_about_unused(sp, id, ln, var);
1409 intravisit::walk_arm(this, arm);
1412 fn check_expr<'a, 'tcx>(this: &mut Liveness<'a, 'tcx>, expr: &'tcx Expr) {
1414 hir::ExprKind::Assign(ref l, _) => {
1415 this.check_place(&l);
1417 intravisit::walk_expr(this, expr);
1420 hir::ExprKind::AssignOp(_, ref l, _) => {
1421 if !this.tables.is_method_call(expr) {
1422 this.check_place(&l);
1425 intravisit::walk_expr(this, expr);
1428 hir::ExprKind::InlineAsm(ref ia, ref outputs, ref inputs) => {
1429 for input in inputs {
1430 this.visit_expr(input);
1433 // Output operands must be places
1434 for (o, output) in ia.outputs.iter().zip(outputs) {
1436 this.check_place(output);
1438 this.visit_expr(output);
1441 intravisit::walk_expr(this, expr);
1444 // no correctness conditions related to liveness
1445 hir::ExprKind::Call(..) | hir::ExprKind::MethodCall(..) | hir::ExprKind::If(..) |
1446 hir::ExprKind::Match(..) | hir::ExprKind::While(..) | hir::ExprKind::Loop(..) |
1447 hir::ExprKind::Index(..) | hir::ExprKind::Field(..) |
1448 hir::ExprKind::Array(..) | hir::ExprKind::Tup(..) | hir::ExprKind::Binary(..) |
1449 hir::ExprKind::Cast(..) | hir::ExprKind::Unary(..) | hir::ExprKind::Ret(..) |
1450 hir::ExprKind::Break(..) | hir::ExprKind::Continue(..) | hir::ExprKind::Lit(_) |
1451 hir::ExprKind::Block(..) | hir::ExprKind::AddrOf(..) |
1452 hir::ExprKind::Struct(..) | hir::ExprKind::Repeat(..) |
1453 hir::ExprKind::Closure(..) | hir::ExprKind::Path(_) | hir::ExprKind::Yield(..) |
1454 hir::ExprKind::Box(..) | hir::ExprKind::Type(..) => {
1455 intravisit::walk_expr(this, expr);
1460 impl<'a, 'tcx> Liveness<'a, 'tcx> {
1461 fn check_place(&mut self, expr: &'tcx Expr) {
1463 hir::ExprKind::Path(hir::QPath::Resolved(_, ref path)) => {
1464 if let Def::Local(nid) = path.def {
1465 // Assignment to an immutable variable or argument: only legal
1466 // if there is no later assignment. If this local is actually
1467 // mutable, then check for a reassignment to flag the mutability
1469 let ln = self.live_node(expr.hir_id, expr.span);
1470 let var_hid = self.ir.tcx.hir.node_to_hir_id(nid);
1471 let var = self.variable(var_hid, expr.span);
1472 self.warn_about_dead_assign(expr.span, expr.hir_id, ln, var);
1476 // For other kinds of places, no checks are required,
1477 // and any embedded expressions are actually rvalues
1478 intravisit::walk_expr(self, expr);
1483 fn should_warn(&self, var: Variable) -> Option<String> {
1484 let name = self.ir.variable_name(var);
1485 if name.is_empty() || name.as_bytes()[0] == ('_' as u8) {
1492 fn warn_about_unused_args(&self, body: &hir::Body, entry_ln: LiveNode) {
1493 for arg in &body.arguments {
1494 arg.pat.each_binding(|_bm, hir_id, _, ident| {
1495 let sp = ident.span;
1496 let var = self.variable(hir_id, sp);
1497 // Ignore unused self.
1498 if ident.name != keywords::SelfValue.name() {
1499 if !self.warn_about_unused(sp, hir_id, entry_ln, var) {
1500 if self.live_on_entry(entry_ln, var).is_none() {
1501 self.report_dead_assign(hir_id, sp, var, true);
1509 fn warn_about_unused_or_dead_vars_in_pat(&mut self, pat: &hir::Pat) {
1510 self.pat_bindings(pat, |this, ln, var, sp, id| {
1511 if !this.warn_about_unused(sp, id, ln, var) {
1512 this.warn_about_dead_assign(sp, id, ln, var);
1517 fn warn_about_unused(&self,
1523 if !self.used_on_entry(ln, var) {
1524 let r = self.should_warn(var);
1525 if let Some(name) = r {
1527 // annoying: for parameters in funcs like `fn(x: i32)
1528 // {ret}`, there is only one node, so asking about
1529 // assigned_on_exit() is not meaningful.
1530 let is_assigned = if ln == self.s.exit_ln {
1533 self.assigned_on_exit(ln, var).is_some()
1536 let suggest_underscore_msg = format!("consider using `_{}` instead",
1541 .lint_hir_note(lint::builtin::UNUSED_VARIABLES, hir_id, sp,
1542 &format!("variable `{}` is assigned to, but never used",
1544 &suggest_underscore_msg);
1545 } else if name != "self" {
1546 let msg = format!("unused variable: `{}`", name);
1547 let mut err = self.ir.tcx
1548 .struct_span_lint_hir(lint::builtin::UNUSED_VARIABLES, hir_id, sp, &msg);
1549 if self.ir.variable_is_shorthand(var) {
1550 err.span_suggestion_with_applicability(sp, "try ignoring the field",
1551 format!("{}: _", name),
1552 Applicability::MachineApplicable);
1554 err.span_suggestion_short_with_applicability(
1555 sp, &suggest_underscore_msg,
1556 format!("_{}", name),
1557 Applicability::MachineApplicable,
1569 fn warn_about_dead_assign(&self,
1574 if self.live_on_exit(ln, var).is_none() {
1575 self.report_dead_assign(hir_id, sp, var, false);
1579 fn report_dead_assign(&self, hir_id: HirId, sp: Span, var: Variable, is_argument: bool) {
1580 if let Some(name) = self.should_warn(var) {
1582 self.ir.tcx.lint_hir(lint::builtin::UNUSED_ASSIGNMENTS, hir_id, sp,
1583 &format!("value passed to `{}` is never read", name));
1585 self.ir.tcx.lint_hir(lint::builtin::UNUSED_ASSIGNMENTS, hir_id, sp,
1586 &format!("value assigned to `{}` is never read", name));