1 use rustc::middle::const_val;
2 use rustc::hir::def_id::DefId;
3 use rustc::mir::mir_map::MirMap;
4 use rustc::mir::repr as mir;
5 use rustc::traits::{self, ProjectionMode};
6 use rustc::ty::fold::TypeFoldable;
7 use rustc::ty::layout::{self, Layout, Size};
8 use rustc::ty::subst::{self, Subst, Substs};
9 use rustc::ty::{self, Ty, TyCtxt};
10 use rustc::util::nodemap::DefIdMap;
11 use std::cell::RefCell;
12 use std::ops::{Deref, DerefMut};
17 use syntax::codemap::{self, DUMMY_SP};
19 use error::{EvalError, EvalResult};
20 use memory::{Memory, Pointer};
21 use primval::{self, PrimVal};
23 const TRACE_EXECUTION: bool = true;
25 struct GlobalEvalContext<'a, 'tcx: 'a> {
26 /// The results of the type checker, from rustc.
27 tcx: TyCtxt<'a, 'tcx, 'tcx>,
29 /// A mapping from NodeIds to Mir, from rustc. Only contains MIR for crate-local items.
30 mir_map: &'a MirMap<'tcx>,
32 /// A local cache from DefIds to Mir for non-crate-local items.
33 mir_cache: RefCell<DefIdMap<Rc<mir::Mir<'tcx>>>>,
35 /// The virtual memory system.
38 /// Another stack containing the type substitutions for the current function invocation. It
39 /// exists separately from `stack` because it must contain the `Substs` for a function while
40 /// *creating* the `Frame` for that same function.
41 substs_stack: Vec<&'tcx Substs<'tcx>>,
43 // TODO(solson): Merge with `substs_stack`. Also try restructuring `Frame` to accomodate.
44 /// A stack of the things necessary to print good strack traces:
45 /// * Function DefIds and Substs to print proper substituted function names.
46 /// * Spans pointing to specific function calls in the source.
47 name_stack: Vec<(DefId, &'tcx Substs<'tcx>, codemap::Span)>,
50 struct FnEvalContext<'a, 'b: 'a + 'mir, 'mir, 'tcx: 'b> {
51 gecx: &'a mut GlobalEvalContext<'b, 'tcx>,
53 /// The virtual call stack.
54 stack: Vec<Frame<'mir, 'tcx>>,
57 impl<'a, 'b, 'mir, 'tcx> Deref for FnEvalContext<'a, 'b, 'mir, 'tcx> {
58 type Target = GlobalEvalContext<'b, 'tcx>;
59 fn deref(&self) -> &Self::Target {
64 impl<'a, 'b, 'mir, 'tcx> DerefMut for FnEvalContext<'a, 'b, 'mir, 'tcx> {
65 fn deref_mut(&mut self) -> &mut Self::Target {
71 struct Frame<'a, 'tcx: 'a> {
72 /// The MIR for the function called on this frame.
73 mir: CachedMir<'a, 'tcx>,
75 /// The block this frame will execute when a function call returns back to this frame.
76 next_block: mir::BasicBlock,
78 /// A pointer for writing the return value of the current call if it's not a diverging call.
79 return_ptr: Option<Pointer>,
81 /// The list of locals for the current function, stored in order as
82 /// `[arguments..., variables..., temporaries...]`. The variables begin at `self.var_offset`
83 /// and the temporaries at `self.temp_offset`.
86 /// The offset of the first variable in `self.locals`.
89 /// The offset of the first temporary in `self.locals`.
93 #[derive(Copy, Clone, Debug, Eq, PartialEq)]
99 #[derive(Copy, Clone, Debug, Eq, PartialEq)]
103 // TODO(solson): Vtable(memory::AllocId),
104 DowncastVariant(usize),
108 enum CachedMir<'mir, 'tcx: 'mir> {
109 Ref(&'mir mir::Mir<'tcx>),
110 Owned(Rc<mir::Mir<'tcx>>)
113 /// Represents the action to be taken in the main loop as a result of executing a terminator.
114 enum TerminatorTarget {
115 /// Make a local jump to the given block.
116 Block(mir::BasicBlock),
118 /// Start executing from the new current frame. (For function calls.)
121 /// Stop executing the current frame and resume the previous frame.
125 impl<'a, 'tcx> GlobalEvalContext<'a, 'tcx> {
126 fn new(tcx: TyCtxt<'a, 'tcx, 'tcx>, mir_map: &'a MirMap<'tcx>) -> Self {
130 mir_cache: RefCell::new(DefIdMap()),
131 memory: Memory::new(tcx.sess
135 .expect("Session::target::uint_type was usize")/8),
136 substs_stack: Vec::new(),
137 name_stack: Vec::new(),
142 impl<'a, 'b, 'mir, 'tcx> FnEvalContext<'a, 'b, 'mir, 'tcx> {
143 fn new(gecx: &'a mut GlobalEvalContext<'b, 'tcx>) -> Self {
150 fn maybe_report<T>(&self, span: codemap::Span, r: EvalResult<T>) -> EvalResult<T> {
151 if let Err(ref e) = r {
152 let mut err = self.tcx.sess.struct_span_err(span, &e.to_string());
153 for &(def_id, substs, span) in self.name_stack.iter().rev() {
154 // FIXME(solson): Find a way to do this without this Display impl hack.
155 use rustc::util::ppaux;
157 struct Instance<'tcx>(DefId, &'tcx Substs<'tcx>);
158 impl<'tcx> fmt::Display for Instance<'tcx> {
159 fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
160 ppaux::parameterized(f, self.1, self.0, ppaux::Ns::Value, &[],
161 |tcx| tcx.lookup_item_type(self.0).generics)
164 err.span_note(span, &format!("inside call to {}", Instance(def_id, substs)));
171 fn log<F>(&self, extra_indent: usize, f: F) where F: FnOnce() {
172 let indent = self.stack.len() + extra_indent;
173 if !TRACE_EXECUTION { return; }
174 for _ in 0..indent { print!(" "); }
179 fn run(&mut self) -> EvalResult<()> {
180 'outer: while !self.stack.is_empty() {
181 let mut current_block = self.frame().next_block;
184 self.log(0, || print!("// {:?}", current_block));
185 let current_mir = self.mir().clone(); // Cloning a reference.
186 let block_data = current_mir.basic_block_data(current_block);
188 for stmt in &block_data.statements {
189 self.log(0, || print!("{:?}", stmt));
190 let mir::StatementKind::Assign(ref lvalue, ref rvalue) = stmt.kind;
191 let result = self.eval_assignment(lvalue, rvalue);
192 self.maybe_report(stmt.span, result)?;
195 let terminator = block_data.terminator();
196 self.log(0, || print!("{:?}", terminator.kind));
198 let result = self.eval_terminator(terminator);
199 match self.maybe_report(terminator.span, result)? {
200 TerminatorTarget::Block(block) => current_block = block,
201 TerminatorTarget::Return => {
202 self.pop_stack_frame();
203 self.name_stack.pop();
206 TerminatorTarget::Call => continue 'outer,
214 fn call_nested(&mut self, mir: &mir::Mir<'tcx>) -> EvalResult<Option<Pointer>> {
215 let mut nested_fecx = FnEvalContext::new(self.gecx);
217 let return_ptr = match mir.return_ty {
218 ty::FnConverging(ty) => {
219 let size = nested_fecx.type_size(ty);
220 Some(nested_fecx.memory.allocate(size))
222 ty::FnDiverging => None,
225 let substs = nested_fecx.substs();
226 nested_fecx.push_stack_frame(CachedMir::Ref(mir), substs, return_ptr);
231 fn push_stack_frame(&mut self, mir: CachedMir<'mir, 'tcx>, substs: &'tcx Substs<'tcx>,
232 return_ptr: Option<Pointer>)
234 self.substs_stack.push(substs);
236 let arg_tys = mir.arg_decls.iter().map(|a| a.ty);
237 let var_tys = mir.var_decls.iter().map(|v| v.ty);
238 let temp_tys = mir.temp_decls.iter().map(|t| t.ty);
240 let locals: Vec<Pointer> = arg_tys.chain(var_tys).chain(temp_tys).map(|ty| {
241 let size = self.type_size(ty);
242 self.memory.allocate(size)
245 let num_args = mir.arg_decls.len();
246 let num_vars = mir.var_decls.len();
248 self.stack.push(Frame {
250 next_block: mir::START_BLOCK,
251 return_ptr: return_ptr,
253 var_offset: num_args,
254 temp_offset: num_args + num_vars,
258 fn pop_stack_frame(&mut self) {
259 let _frame = self.stack.pop().expect("tried to pop a stack frame, but there were none");
260 // TODO(solson): Deallocate local variables.
261 self.substs_stack.pop();
264 fn eval_terminator(&mut self, terminator: &mir::Terminator<'tcx>)
265 -> EvalResult<TerminatorTarget> {
266 use rustc::mir::repr::TerminatorKind::*;
267 let target = match terminator.kind {
268 Return => TerminatorTarget::Return,
270 Goto { target } => TerminatorTarget::Block(target),
272 If { ref cond, targets: (then_target, else_target) } => {
273 let cond_ptr = self.eval_operand(cond)?;
274 let cond_val = self.memory.read_bool(cond_ptr)?;
275 TerminatorTarget::Block(if cond_val { then_target } else { else_target })
278 SwitchInt { ref discr, ref values, ref targets, .. } => {
279 let discr_ptr = self.eval_lvalue(discr)?.to_ptr();
280 let discr_size = self
281 .type_layout(self.lvalue_ty(discr))
282 .size(&self.tcx.data_layout)
284 let discr_val = self.memory.read_uint(discr_ptr, discr_size)?;
286 // Branch to the `otherwise` case by default, if no match is found.
287 let mut target_block = targets[targets.len() - 1];
289 for (index, val_const) in values.iter().enumerate() {
290 let ptr = self.const_to_ptr(val_const)?;
291 let val = self.memory.read_uint(ptr, discr_size)?;
292 if discr_val == val {
293 target_block = targets[index];
298 TerminatorTarget::Block(target_block)
301 Switch { ref discr, ref targets, adt_def } => {
302 let adt_ptr = self.eval_lvalue(discr)?.to_ptr();
303 let adt_ty = self.lvalue_ty(discr);
304 let discr_val = self.read_discriminant_value(adt_ptr, adt_ty)?;
305 let matching = adt_def.variants.iter()
306 .position(|v| discr_val == v.disr_val.to_u64_unchecked());
309 Some(i) => TerminatorTarget::Block(targets[i]),
310 None => return Err(EvalError::InvalidDiscriminant),
314 Call { ref func, ref args, ref destination, .. } => {
315 let mut return_ptr = None;
316 if let Some((ref lv, target)) = *destination {
317 self.frame_mut().next_block = target;
318 return_ptr = Some(self.eval_lvalue(lv)?.to_ptr());
321 let func_ty = self.operand_ty(func);
323 ty::TyFnDef(def_id, substs, fn_ty) => {
324 use syntax::abi::Abi;
326 Abi::RustIntrinsic => {
327 let name = self.tcx.item_name(def_id).as_str();
328 match fn_ty.sig.0.output {
329 ty::FnConverging(ty) => {
330 let size = self.type_size(ty);
331 let ret = return_ptr.unwrap();
332 self.call_intrinsic(&name, substs, args, ret, size)?
334 ty::FnDiverging => unimplemented!(),
339 match fn_ty.sig.0.output {
340 ty::FnConverging(ty) => {
341 let size = self.type_size(ty);
342 self.call_c_abi(def_id, args, return_ptr.unwrap(), size)?
344 ty::FnDiverging => unimplemented!(),
348 Abi::Rust | Abi::RustCall => {
349 // TODO(solson): Adjust the first argument when calling a Fn or
350 // FnMut closure via FnOnce::call_once.
352 // Only trait methods can have a Self parameter.
353 let (resolved_def_id, resolved_substs) = if substs.self_ty().is_some() {
354 self.trait_method(def_id, substs)
359 let mut arg_srcs = Vec::new();
361 let src = self.eval_operand(arg)?;
362 let src_ty = self.operand_ty(arg);
363 arg_srcs.push((src, src_ty));
366 if fn_ty.abi == Abi::RustCall && !args.is_empty() {
368 let last_arg = args.last().unwrap();
369 let last = self.eval_operand(last_arg)?;
370 let last_ty = self.operand_ty(last_arg);
371 let last_layout = self.type_layout(last_ty);
372 match (&last_ty.sty, last_layout) {
373 (&ty::TyTuple(fields),
374 &Layout::Univariant { ref variant, .. }) => {
375 let offsets = iter::once(0)
376 .chain(variant.offset_after_field.iter()
377 .map(|s| s.bytes()));
378 for (offset, ty) in offsets.zip(fields) {
379 let src = last.offset(offset as isize);
380 arg_srcs.push((src, ty));
383 ty => panic!("expected tuple as last argument in function with 'rust-call' ABI, got {:?}", ty),
387 let mir = self.load_mir(resolved_def_id);
388 self.name_stack.push((def_id, substs, terminator.span));
389 self.push_stack_frame(mir, resolved_substs, return_ptr);
391 for (i, (src, src_ty)) in arg_srcs.into_iter().enumerate() {
392 let dest = self.frame().locals[i];
393 self.move_(src, dest, src_ty)?;
396 TerminatorTarget::Call
399 abi => return Err(EvalError::Unimplemented(format!("can't handle function with {:?} ABI", abi))),
403 _ => return Err(EvalError::Unimplemented(format!("can't handle callee of type {:?}", func_ty))),
407 Drop { ref value, target, .. } => {
408 let ptr = self.eval_lvalue(value)?.to_ptr();
409 let ty = self.lvalue_ty(value);
411 TerminatorTarget::Block(target)
414 Resume => unimplemented!(),
420 fn drop(&mut self, ptr: Pointer, ty: Ty<'tcx>) -> EvalResult<()> {
421 if !self.type_needs_drop(ty) {
422 self.log(1, || print!("no need to drop {:?}", ty));
425 self.log(1, || print!("need to drop {:?}", ty));
427 // TODO(solson): Call user-defined Drop::drop impls.
430 ty::TyBox(contents_ty) => {
431 match self.memory.read_ptr(ptr) {
432 Ok(contents_ptr) => {
433 self.drop(contents_ptr, contents_ty)?;
434 self.log(1, || print!("deallocating box"));
435 self.memory.deallocate(contents_ptr)?;
437 Err(EvalError::ReadBytesAsPointer) => {
438 let size = self.memory.pointer_size;
439 let possible_drop_fill = self.memory.read_bytes(ptr, size)?;
440 if possible_drop_fill.iter().all(|&b| b == mem::POST_DROP_U8) {
443 return Err(EvalError::ReadBytesAsPointer);
446 Err(e) => return Err(e),
450 // TODO(solson): Implement drop for other relevant types (e.g. aggregates).
455 // FIXME(solson): Trait objects (with no static size) probably get filled, too.
456 let size = self.type_size(ty);
457 self.memory.drop_fill(ptr, size)?;
462 fn read_discriminant_value(&self, adt_ptr: Pointer, adt_ty: Ty<'tcx>) -> EvalResult<u64> {
463 use rustc::ty::layout::Layout::*;
464 let adt_layout = self.type_layout(adt_ty);
466 let discr_val = match *adt_layout {
467 General { discr, .. } | CEnum { discr, .. } => {
468 let discr_size = discr.size().bytes();
469 self.memory.read_uint(adt_ptr, discr_size as usize)?
472 RawNullablePointer { nndiscr, .. } => {
473 self.read_nonnull_discriminant_value(adt_ptr, nndiscr)?
476 StructWrappedNullablePointer { nndiscr, ref discrfield, .. } => {
477 let offset = self.nonnull_offset(adt_ty, nndiscr, discrfield)?;
478 let nonnull = adt_ptr.offset(offset.bytes() as isize);
479 self.read_nonnull_discriminant_value(nonnull, nndiscr)?
482 // The discriminant_value intrinsic returns 0 for non-sum types.
483 Array { .. } | FatPointer { .. } | Scalar { .. } | Univariant { .. } |
490 fn read_nonnull_discriminant_value(&self, ptr: Pointer, nndiscr: u64) -> EvalResult<u64> {
491 let not_null = match self.memory.read_usize(ptr) {
493 Ok(_) | Err(EvalError::ReadPointerAsBytes) => true,
494 Err(e) => return Err(e),
496 assert!(nndiscr == 0 || nndiscr == 1);
497 Ok(if not_null { nndiscr } else { 1 - nndiscr })
503 substs: &'tcx Substs<'tcx>,
504 args: &[mir::Operand<'tcx>],
507 ) -> EvalResult<TerminatorTarget> {
508 let args_res: EvalResult<Vec<Pointer>> = args.iter()
509 .map(|arg| self.eval_operand(arg))
511 let args = args_res?;
514 // FIXME(solson): Handle different integer types correctly.
515 "add_with_overflow" => {
516 let ty = *substs.types.get(subst::FnSpace, 0);
517 let size = self.type_size(ty);
518 let left = self.memory.read_int(args[0], size)?;
519 let right = self.memory.read_int(args[1], size)?;
520 let (n, overflowed) = unsafe {
521 ::std::intrinsics::add_with_overflow::<i64>(left, right)
523 self.memory.write_int(dest, n, size)?;
524 self.memory.write_bool(dest.offset(size as isize), overflowed)?;
529 "copy_nonoverlapping" => {
530 let elem_ty = *substs.types.get(subst::FnSpace, 0);
531 let elem_size = self.type_size(elem_ty);
532 let src = self.memory.read_ptr(args[0])?;
533 let dest = self.memory.read_ptr(args[1])?;
534 let count = self.memory.read_isize(args[2])?;
535 self.memory.copy(src, dest, count as usize * elem_size)?;
538 "discriminant_value" => {
539 let ty = *substs.types.get(subst::FnSpace, 0);
540 let adt_ptr = self.memory.read_ptr(args[0])?;
541 let discr_val = self.read_discriminant_value(adt_ptr, ty)?;
542 self.memory.write_uint(dest, discr_val, dest_size)?;
546 let arg_ty = *substs.types.get(subst::FnSpace, 0);
547 let arg_size = self.type_size(arg_ty);
548 self.memory.drop_fill(args[0], arg_size)?;
551 "init" => self.memory.write_repeat(dest, 0, dest_size)?,
554 self.memory.write_int(dest, 1, dest_size)?;
558 let ty = *substs.types.get(subst::FnSpace, 0);
559 let ptr = self.memory.read_ptr(args[0])?;
560 self.move_(args[1], ptr, ty)?;
563 // FIXME(solson): Handle different integer types correctly.
564 "mul_with_overflow" => {
565 let ty = *substs.types.get(subst::FnSpace, 0);
566 let size = self.type_size(ty);
567 let left = self.memory.read_int(args[0], size)?;
568 let right = self.memory.read_int(args[1], size)?;
569 let (n, overflowed) = unsafe {
570 ::std::intrinsics::mul_with_overflow::<i64>(left, right)
572 self.memory.write_int(dest, n, size)?;
573 self.memory.write_bool(dest.offset(size as isize), overflowed)?;
577 let pointee_ty = *substs.types.get(subst::FnSpace, 0);
578 let pointee_size = self.type_size(pointee_ty) as isize;
579 let ptr_arg = args[0];
580 let offset = self.memory.read_isize(args[1])?;
582 match self.memory.read_ptr(ptr_arg) {
584 let result_ptr = ptr.offset(offset as isize * pointee_size);
585 self.memory.write_ptr(dest, result_ptr)?;
587 Err(EvalError::ReadBytesAsPointer) => {
588 let addr = self.memory.read_isize(ptr_arg)?;
589 let result_addr = addr + offset * pointee_size as i64;
590 self.memory.write_isize(dest, result_addr)?;
592 Err(e) => return Err(e),
596 // FIXME(solson): Handle different integer types correctly. Use primvals?
597 "overflowing_sub" => {
598 let ty = *substs.types.get(subst::FnSpace, 0);
599 let size = self.type_size(ty);
600 let left = self.memory.read_int(args[0], size)?;
601 let right = self.memory.read_int(args[1], size)?;
602 let n = left.wrapping_sub(right);
603 self.memory.write_int(dest, n, size)?;
607 let ty = *substs.types.get(subst::FnSpace, 0);
608 let size = self.type_size(ty) as u64;
609 self.memory.write_uint(dest, size, dest_size)?;
613 let ty = *substs.types.get(subst::FnSpace, 0);
614 if self.type_is_sized(ty) {
615 let size = self.type_size(ty) as u64;
616 self.memory.write_uint(dest, size, dest_size)?;
619 ty::TySlice(_) | ty::TyStr => {
620 let elem_ty = ty.sequence_element_type(self.tcx);
621 let elem_size = self.type_size(elem_ty) as u64;
622 let ptr_size = self.memory.pointer_size as isize;
623 let n = self.memory.read_usize(args[0].offset(ptr_size))?;
624 self.memory.write_uint(dest, n * elem_size, dest_size)?;
627 _ => return Err(EvalError::Unimplemented(format!("unimplemented: size_of_val::<{:?}>", ty))),
633 let ty = *substs.types.get(subst::FnSpace, 0);
634 self.move_(args[0], dest, ty)?;
636 "uninit" => self.memory.mark_definedness(dest, dest_size, false)?,
638 name => return Err(EvalError::Unimplemented(format!("unimplemented intrinsic: {}", name))),
641 // Since we pushed no stack frame, the main loop will act
642 // as if the call just completed and it's returning to the
644 Ok(TerminatorTarget::Call)
650 args: &[mir::Operand<'tcx>],
653 ) -> EvalResult<TerminatorTarget> {
654 let name = self.tcx.item_name(def_id);
655 let attrs = self.tcx.get_attrs(def_id);
656 let link_name = match attr::first_attr_value_str_by_name(&attrs, "link_name") {
657 Some(ln) => ln.clone(),
658 None => name.as_str(),
661 let args_res: EvalResult<Vec<Pointer>> = args.iter()
662 .map(|arg| self.eval_operand(arg))
664 let args = args_res?;
666 match &link_name[..] {
667 "__rust_allocate" => {
668 let size = self.memory.read_usize(args[0])?;
669 let ptr = self.memory.allocate(size as usize);
670 self.memory.write_ptr(dest, ptr)?;
673 "__rust_reallocate" => {
674 let ptr = self.memory.read_ptr(args[0])?;
675 let size = self.memory.read_usize(args[2])?;
676 self.memory.reallocate(ptr, size as usize)?;
677 self.memory.write_ptr(dest, ptr)?;
681 let left = self.memory.read_ptr(args[0])?;
682 let right = self.memory.read_ptr(args[1])?;
683 let n = self.memory.read_usize(args[2])? as usize;
686 let left_bytes = self.memory.read_bytes(left, n)?;
687 let right_bytes = self.memory.read_bytes(right, n)?;
689 use std::cmp::Ordering::*;
690 match left_bytes.cmp(right_bytes) {
697 self.memory.write_int(dest, result, dest_size)?;
700 _ => return Err(EvalError::Unimplemented(format!("can't call C ABI function: {}", link_name))),
703 // Since we pushed no stack frame, the main loop will act
704 // as if the call just completed and it's returning to the
706 Ok(TerminatorTarget::Call)
709 fn assign_fields<I: IntoIterator<Item = u64>>(
713 operands: &[mir::Operand<'tcx>],
714 ) -> EvalResult<()> {
715 for (offset, operand) in offsets.into_iter().zip(operands) {
716 let src = self.eval_operand(operand)?;
717 let src_ty = self.operand_ty(operand);
718 let field_dest = dest.offset(offset as isize);
719 self.move_(src, field_dest, src_ty)?;
724 fn eval_assignment(&mut self, lvalue: &mir::Lvalue<'tcx>, rvalue: &mir::Rvalue<'tcx>)
727 let dest = self.eval_lvalue(lvalue)?.to_ptr();
728 let dest_ty = self.lvalue_ty(lvalue);
729 let dest_layout = self.type_layout(dest_ty);
731 use rustc::mir::repr::Rvalue::*;
733 Use(ref operand) => {
734 let src = self.eval_operand(operand)?;
735 self.move_(src, dest, dest_ty)?;
738 BinaryOp(bin_op, ref left, ref right) => {
739 let left_ptr = self.eval_operand(left)?;
740 let left_ty = self.operand_ty(left);
741 let left_val = self.read_primval(left_ptr, left_ty)?;
743 let right_ptr = self.eval_operand(right)?;
744 let right_ty = self.operand_ty(right);
745 let right_val = self.read_primval(right_ptr, right_ty)?;
747 let val = primval::binary_op(bin_op, left_val, right_val)?;
748 self.memory.write_primval(dest, val)?;
751 UnaryOp(un_op, ref operand) => {
752 let ptr = self.eval_operand(operand)?;
753 let ty = self.operand_ty(operand);
754 let val = self.read_primval(ptr, ty)?;
755 self.memory.write_primval(dest, primval::unary_op(un_op, val)?)?;
758 Aggregate(ref kind, ref operands) => {
759 use rustc::ty::layout::Layout::*;
761 Univariant { ref variant, .. } => {
762 let offsets = iter::once(0)
763 .chain(variant.offset_after_field.iter().map(|s| s.bytes()));
764 self.assign_fields(dest, offsets, operands)?;
768 let elem_size = match dest_ty.sty {
769 ty::TyArray(elem_ty, _) => self.type_size(elem_ty) as u64,
770 _ => panic!("tried to assign {:?} to non-array type {:?}",
773 let offsets = (0..).map(|i| i * elem_size);
774 self.assign_fields(dest, offsets, operands)?;
777 General { discr, ref variants, .. } => {
778 if let mir::AggregateKind::Adt(adt_def, variant, _) = *kind {
779 let discr_val = adt_def.variants[variant].disr_val.to_u64_unchecked();
780 let discr_size = discr.size().bytes() as usize;
781 self.memory.write_uint(dest, discr_val, discr_size)?;
783 let offsets = variants[variant].offset_after_field.iter()
785 self.assign_fields(dest, offsets, operands)?;
787 panic!("tried to assign {:?} to Layout::General", kind);
791 RawNullablePointer { nndiscr, .. } => {
792 if let mir::AggregateKind::Adt(_, variant, _) = *kind {
793 if nndiscr == variant as u64 {
794 assert_eq!(operands.len(), 1);
795 let operand = &operands[0];
796 let src = self.eval_operand(operand)?;
797 let src_ty = self.operand_ty(operand);
798 self.move_(src, dest, src_ty)?;
800 assert_eq!(operands.len(), 0);
801 self.memory.write_isize(dest, 0)?;
804 panic!("tried to assign {:?} to Layout::RawNullablePointer", kind);
808 StructWrappedNullablePointer { nndiscr, ref nonnull, ref discrfield } => {
809 if let mir::AggregateKind::Adt(_, variant, _) = *kind {
810 if nndiscr == variant as u64 {
811 let offsets = iter::once(0)
812 .chain(nonnull.offset_after_field.iter().map(|s| s.bytes()));
813 try!(self.assign_fields(dest, offsets, operands));
815 assert_eq!(operands.len(), 0);
816 let offset = self.nonnull_offset(dest_ty, nndiscr, discrfield)?;
817 let dest = dest.offset(offset.bytes() as isize);
818 try!(self.memory.write_isize(dest, 0));
821 panic!("tried to assign {:?} to Layout::RawNullablePointer", kind);
825 CEnum { discr, signed, .. } => {
826 assert_eq!(operands.len(), 0);
827 if let mir::AggregateKind::Adt(adt_def, variant, _) = *kind {
828 let val = adt_def.variants[variant].disr_val.to_u64_unchecked();
829 let size = discr.size().bytes() as usize;
832 self.memory.write_int(dest, val as i64, size)?;
834 self.memory.write_uint(dest, val, size)?;
837 panic!("tried to assign {:?} to Layout::CEnum", kind);
841 _ => return Err(EvalError::Unimplemented(format!("can't handle destination layout {:?} when assigning {:?}", dest_layout, kind))),
845 Repeat(ref operand, _) => {
846 let (elem_size, length) = match dest_ty.sty {
847 ty::TyArray(elem_ty, n) => (self.type_size(elem_ty), n),
848 _ => panic!("tried to assign array-repeat to non-array type {:?}", dest_ty),
851 let src = self.eval_operand(operand)?;
853 let elem_dest = dest.offset((i * elem_size) as isize);
854 self.memory.copy(src, elem_dest, elem_size)?;
859 let src = self.eval_lvalue(lvalue)?;
860 let ty = self.lvalue_ty(lvalue);
861 let len = match ty.sty {
862 ty::TyArray(_, n) => n as u64,
863 ty::TySlice(_) => if let LvalueExtra::Length(n) = src.extra {
866 panic!("Rvalue::Len of a slice given non-slice pointer: {:?}", src);
868 _ => panic!("Rvalue::Len expected array or slice, got {:?}", ty),
870 self.memory.write_usize(dest, len)?;
873 Ref(_, _, ref lvalue) => {
874 let lv = self.eval_lvalue(lvalue)?;
875 self.memory.write_ptr(dest, lv.ptr)?;
877 LvalueExtra::None => {},
878 LvalueExtra::Length(len) => {
879 let len_ptr = dest.offset(self.memory.pointer_size as isize);
880 self.memory.write_usize(len_ptr, len)?;
882 LvalueExtra::DowncastVariant(..) =>
883 panic!("attempted to take a reference to an enum downcast lvalue"),
888 let size = self.type_size(ty);
889 let ptr = self.memory.allocate(size);
890 self.memory.write_ptr(dest, ptr)?;
893 Cast(kind, ref operand, dest_ty) => {
894 let src = self.eval_operand(operand)?;
895 let src_ty = self.operand_ty(operand);
897 use rustc::mir::repr::CastKind::*;
900 self.move_(src, dest, src_ty)?;
901 let src_pointee_ty = pointee_type(src_ty).unwrap();
902 let dest_pointee_ty = pointee_type(dest_ty).unwrap();
904 match (&src_pointee_ty.sty, &dest_pointee_ty.sty) {
905 (&ty::TyArray(_, length), &ty::TySlice(_)) => {
906 let len_ptr = dest.offset(self.memory.pointer_size as isize);
907 self.memory.write_usize(len_ptr, length as u64)?;
910 _ => return Err(EvalError::Unimplemented(format!("can't handle cast: {:?}", rvalue))),
915 // FIXME(solson): Wrong for almost everything.
916 let size = dest_layout.size(&self.tcx.data_layout).bytes() as usize;
917 self.memory.copy(src, dest, size)?;
920 _ => return Err(EvalError::Unimplemented(format!("can't handle cast: {:?}", rvalue))),
924 Slice { .. } => unimplemented!(),
925 InlineAsm { .. } => unimplemented!(),
931 fn nonnull_offset(&self, ty: Ty<'tcx>, nndiscr: u64, discrfield: &[u32]) -> EvalResult<Size> {
932 // Skip the constant 0 at the start meant for LLVM GEP.
933 let mut path = discrfield.iter().skip(1).map(|&i| i as usize);
935 // Handle the field index for the outer non-null variant.
936 let inner_ty = match ty.sty {
937 ty::TyEnum(adt_def, substs) => {
938 let variant = &adt_def.variants[nndiscr as usize];
939 let index = path.next().unwrap();
940 let field = &variant.fields[index];
941 field.ty(self.tcx, substs)
944 "non-enum for StructWrappedNullablePointer: {}",
949 self.field_path_offset(inner_ty, path)
952 fn field_path_offset<I: Iterator<Item = usize>>(&self, mut ty: Ty<'tcx>, path: I) -> EvalResult<Size> {
953 let mut offset = Size::from_bytes(0);
955 // Skip the initial 0 intended for LLVM GEP.
956 for field_index in path {
957 let field_offset = self.get_field_offset(ty, field_index)?;
958 ty = self.get_field_ty(ty, field_index)?;
959 offset = offset.checked_add(field_offset, &self.tcx.data_layout).unwrap();
965 fn get_field_ty(&self, ty: Ty<'tcx>, field_index: usize) -> EvalResult<Ty<'tcx>> {
967 ty::TyStruct(adt_def, substs) => {
968 Ok(adt_def.struct_variant().fields[field_index].ty(self.tcx, substs))
971 ty::TyRef(_, ty::TypeAndMut { ty, .. }) |
972 ty::TyRawPtr(ty::TypeAndMut { ty, .. }) |
974 assert_eq!(field_index, 0);
977 _ => Err(EvalError::Unimplemented(format!("can't handle type: {:?}", ty))),
981 fn get_field_offset(&self, ty: Ty<'tcx>, field_index: usize) -> EvalResult<Size> {
982 let layout = self.type_layout(ty);
984 use rustc::ty::layout::Layout::*;
986 Univariant { .. } => {
987 assert_eq!(field_index, 0);
988 Ok(Size::from_bytes(0))
990 FatPointer { .. } => {
991 let bytes = layout::FAT_PTR_ADDR * self.memory.pointer_size;
992 Ok(Size::from_bytes(bytes as u64))
994 _ => Err(EvalError::Unimplemented(format!("can't handle type: {:?}, with layout: {:?}", ty, layout))),
998 fn eval_operand(&mut self, op: &mir::Operand<'tcx>) -> EvalResult<Pointer> {
999 use rustc::mir::repr::Operand::*;
1001 Consume(ref lvalue) => Ok(self.eval_lvalue(lvalue)?.to_ptr()),
1002 Constant(mir::Constant { ref literal, .. }) => {
1003 use rustc::mir::repr::Literal::*;
1005 Value { ref value } => Ok(self.const_to_ptr(value)?),
1006 Item { .. } => unimplemented!(),
1007 Promoted { index } => {
1008 // TODO(solson): Mark constants and statics as read-only and cache their
1010 let current_mir = self.mir();
1011 let mir = ¤t_mir.promoted[index];
1012 self.call_nested(mir).map(Option::unwrap)
1019 fn eval_lvalue(&mut self, lvalue: &mir::Lvalue<'tcx>) -> EvalResult<Lvalue> {
1020 use rustc::mir::repr::Lvalue::*;
1021 let ptr = match *lvalue {
1022 ReturnPointer => self.frame().return_ptr
1023 .expect("ReturnPointer used in a function with no return value"),
1024 Arg(i) => self.frame().locals[i as usize],
1025 Var(i) => self.frame().locals[self.frame().var_offset + i as usize],
1026 Temp(i) => self.frame().locals[self.frame().temp_offset + i as usize],
1029 // TODO(solson): Mark constants and statics as read-only and cache their values.
1030 let mir = self.load_mir(def_id);
1031 self.call_nested(&mir)?.unwrap()
1034 Projection(ref proj) => {
1035 let base = self.eval_lvalue(&proj.base)?;
1036 let base_ty = self.lvalue_ty(&proj.base);
1037 let base_layout = self.type_layout(base_ty);
1039 use rustc::mir::repr::ProjectionElem::*;
1041 Field(field, _) => {
1042 use rustc::ty::layout::Layout::*;
1043 let variant = match *base_layout {
1044 Univariant { ref variant, .. } => variant,
1045 General { ref variants, .. } => {
1046 if let LvalueExtra::DowncastVariant(variant_idx) = base.extra {
1047 &variants[variant_idx]
1049 panic!("field access on enum had no variant index");
1052 RawNullablePointer { .. } => {
1053 assert_eq!(field.index(), 0);
1056 StructWrappedNullablePointer { ref nonnull, .. } => nonnull,
1057 _ => panic!("field access on non-product type: {:?}", base_layout),
1060 let offset = variant.field_offset(field.index()).bytes();
1061 base.ptr.offset(offset as isize)
1064 Downcast(_, variant) => {
1065 use rustc::ty::layout::Layout::*;
1066 match *base_layout {
1067 General { discr, .. } => {
1069 ptr: base.ptr.offset(discr.size().bytes() as isize),
1070 extra: LvalueExtra::DowncastVariant(variant),
1073 RawNullablePointer { .. } | StructWrappedNullablePointer { .. } => {
1076 _ => panic!("variant downcast on non-aggregate: {:?}", base_layout),
1081 let pointee_ty = pointee_type(base_ty).expect("Deref of non-pointer");
1082 let ptr = self.memory.read_ptr(base.ptr)?;
1083 let extra = match pointee_ty.sty {
1084 ty::TySlice(_) | ty::TyStr => {
1085 let len_ptr = base.ptr.offset(self.memory.pointer_size as isize);
1086 let len = self.memory.read_usize(len_ptr)?;
1087 LvalueExtra::Length(len)
1089 ty::TyTrait(_) => unimplemented!(),
1090 _ => LvalueExtra::None,
1092 return Ok(Lvalue { ptr: ptr, extra: extra });
1095 Index(ref operand) => {
1096 let elem_size = match base_ty.sty {
1097 ty::TyArray(elem_ty, _) |
1098 ty::TySlice(elem_ty) => self.type_size(elem_ty),
1099 _ => panic!("indexing expected an array or slice, got {:?}", base_ty),
1101 let n_ptr = self.eval_operand(operand)?;
1102 let n = self.memory.read_usize(n_ptr)?;
1103 base.ptr.offset(n as isize * elem_size as isize)
1106 ConstantIndex { .. } => unimplemented!(),
1111 Ok(Lvalue { ptr: ptr, extra: LvalueExtra::None })
1114 // TODO(solson): Try making const_to_primval instead.
1115 fn const_to_ptr(&mut self, const_val: &const_val::ConstVal) -> EvalResult<Pointer> {
1116 use rustc::middle::const_val::ConstVal::*;
1118 Float(_f) => unimplemented!(),
1120 // TODO(solson): Check int constant type.
1121 let ptr = self.memory.allocate(8);
1122 self.memory.write_uint(ptr, int.to_u64_unchecked(), 8)?;
1126 let psize = self.memory.pointer_size;
1127 let static_ptr = self.memory.allocate(s.len());
1128 let ptr = self.memory.allocate(psize * 2);
1129 self.memory.write_bytes(static_ptr, s.as_bytes())?;
1130 self.memory.write_ptr(ptr, static_ptr)?;
1131 self.memory.write_usize(ptr.offset(psize as isize), s.len() as u64)?;
1134 ByteStr(ref bs) => {
1135 let psize = self.memory.pointer_size;
1136 let static_ptr = self.memory.allocate(bs.len());
1137 let ptr = self.memory.allocate(psize);
1138 self.memory.write_bytes(static_ptr, bs)?;
1139 self.memory.write_ptr(ptr, static_ptr)?;
1143 let ptr = self.memory.allocate(1);
1144 self.memory.write_bool(ptr, b)?;
1147 Char(_c) => unimplemented!(),
1148 Struct(_node_id) => unimplemented!(),
1149 Tuple(_node_id) => unimplemented!(),
1150 Function(_def_id) => unimplemented!(),
1151 Array(_, _) => unimplemented!(),
1152 Repeat(_, _) => unimplemented!(),
1153 Dummy => unimplemented!(),
1157 fn lvalue_ty(&self, lvalue: &mir::Lvalue<'tcx>) -> Ty<'tcx> {
1158 self.monomorphize(self.mir().lvalue_ty(self.tcx, lvalue).to_ty(self.tcx))
1161 fn operand_ty(&self, operand: &mir::Operand<'tcx>) -> Ty<'tcx> {
1162 self.monomorphize(self.mir().operand_ty(self.tcx, operand))
1165 fn monomorphize(&self, ty: Ty<'tcx>) -> Ty<'tcx> {
1166 let substituted = ty.subst(self.tcx, self.substs());
1167 self.tcx.normalize_associated_type(&substituted)
1170 fn type_needs_drop(&self, ty: Ty<'tcx>) -> bool {
1171 self.tcx.type_needs_drop_given_env(ty, &self.tcx.empty_parameter_environment())
1174 fn move_(&mut self, src: Pointer, dest: Pointer, ty: Ty<'tcx>) -> EvalResult<()> {
1175 let size = self.type_size(ty);
1176 self.memory.copy(src, dest, size)?;
1177 if self.type_needs_drop(ty) {
1178 self.memory.drop_fill(src, size)?;
1183 fn type_is_sized(&self, ty: Ty<'tcx>) -> bool {
1184 ty.is_sized(self.tcx, &self.tcx.empty_parameter_environment(), DUMMY_SP)
1187 fn type_size(&self, ty: Ty<'tcx>) -> usize {
1188 self.type_layout(ty).size(&self.tcx.data_layout).bytes() as usize
1191 fn type_layout(&self, ty: Ty<'tcx>) -> &'tcx Layout {
1192 // TODO(solson): Is this inefficient? Needs investigation.
1193 let ty = self.monomorphize(ty);
1195 self.tcx.normalizing_infer_ctxt(ProjectionMode::Any).enter(|infcx| {
1196 // TODO(solson): Report this error properly.
1197 ty.layout(&infcx).unwrap()
1201 pub fn read_primval(&mut self, ptr: Pointer, ty: Ty<'tcx>) -> EvalResult<PrimVal> {
1202 use syntax::ast::{IntTy, UintTy};
1203 let val = match (self.memory.pointer_size, &ty.sty) {
1204 (_, &ty::TyBool) => PrimVal::Bool(self.memory.read_bool(ptr)?),
1205 (_, &ty::TyInt(IntTy::I8)) => PrimVal::I8(self.memory.read_int(ptr, 1)? as i8),
1206 (2, &ty::TyInt(IntTy::Is)) |
1207 (_, &ty::TyInt(IntTy::I16)) => PrimVal::I16(self.memory.read_int(ptr, 2)? as i16),
1208 (4, &ty::TyInt(IntTy::Is)) |
1209 (_, &ty::TyInt(IntTy::I32)) => PrimVal::I32(self.memory.read_int(ptr, 4)? as i32),
1210 (8, &ty::TyInt(IntTy::Is)) |
1211 (_, &ty::TyInt(IntTy::I64)) => PrimVal::I64(self.memory.read_int(ptr, 8)? as i64),
1212 (_, &ty::TyUint(UintTy::U8)) => PrimVal::U8(self.memory.read_uint(ptr, 1)? as u8),
1213 (2, &ty::TyUint(UintTy::Us)) |
1214 (_, &ty::TyUint(UintTy::U16)) => PrimVal::U16(self.memory.read_uint(ptr, 2)? as u16),
1215 (4, &ty::TyUint(UintTy::Us)) |
1216 (_, &ty::TyUint(UintTy::U32)) => PrimVal::U32(self.memory.read_uint(ptr, 4)? as u32),
1217 (8, &ty::TyUint(UintTy::Us)) |
1218 (_, &ty::TyUint(UintTy::U64)) => PrimVal::U64(self.memory.read_uint(ptr, 8)? as u64),
1220 (_, &ty::TyRef(_, ty::TypeAndMut { ty, .. })) |
1221 (_, &ty::TyRawPtr(ty::TypeAndMut { ty, .. })) => {
1222 if self.type_is_sized(ty) {
1223 match self.memory.read_ptr(ptr) {
1224 Ok(p) => PrimVal::AbstractPtr(p),
1225 Err(EvalError::ReadBytesAsPointer) => {
1226 PrimVal::IntegerPtr(self.memory.read_usize(ptr)?)
1228 Err(e) => return Err(e),
1231 return Err(EvalError::Unimplemented(format!("unimplemented: primitive read of fat pointer type: {:?}", ty)));
1235 _ => panic!("primitive read of non-primitive type: {:?}", ty),
1240 fn frame(&self) -> &Frame<'mir, 'tcx> {
1241 self.stack.last().expect("no call frames exist")
1244 fn frame_mut(&mut self) -> &mut Frame<'mir, 'tcx> {
1245 self.stack.last_mut().expect("no call frames exist")
1248 fn mir(&self) -> CachedMir<'mir, 'tcx> {
1249 self.frame().mir.clone()
1252 fn substs(&self) -> &'tcx Substs<'tcx> {
1253 self.substs_stack.last().cloned().unwrap_or_else(|| self.tcx.mk_substs(Substs::empty()))
1256 fn load_mir(&self, def_id: DefId) -> CachedMir<'mir, 'tcx> {
1257 match self.tcx.map.as_local_node_id(def_id) {
1258 Some(node_id) => CachedMir::Ref(self.mir_map.map.get(&node_id).unwrap()),
1260 let mut mir_cache = self.mir_cache.borrow_mut();
1261 if let Some(mir) = mir_cache.get(&def_id) {
1262 return CachedMir::Owned(mir.clone());
1265 let cs = &self.tcx.sess.cstore;
1266 let mir = cs.maybe_get_item_mir(self.tcx, def_id).unwrap_or_else(|| {
1267 panic!("no mir for {:?}", def_id);
1269 let cached = Rc::new(mir);
1270 mir_cache.insert(def_id, cached.clone());
1271 CachedMir::Owned(cached)
1276 fn fulfill_obligation(&self, trait_ref: ty::PolyTraitRef<'tcx>) -> traits::Vtable<'tcx, ()> {
1277 // Do the initial selection for the obligation. This yields the shallow result we are
1278 // looking for -- that is, what specific impl.
1279 self.tcx.normalizing_infer_ctxt(ProjectionMode::Any).enter(|infcx| {
1280 let mut selcx = traits::SelectionContext::new(&infcx);
1282 let obligation = traits::Obligation::new(
1283 traits::ObligationCause::misc(DUMMY_SP, ast::DUMMY_NODE_ID),
1284 trait_ref.to_poly_trait_predicate(),
1286 let selection = selcx.select(&obligation).unwrap().unwrap();
1288 // Currently, we use a fulfillment context to completely resolve all nested obligations.
1289 // This is because they can inform the inference of the impl's type parameters.
1290 let mut fulfill_cx = traits::FulfillmentContext::new();
1291 let vtable = selection.map(|predicate| {
1292 fulfill_cx.register_predicate_obligation(&infcx, predicate);
1294 infcx.drain_fulfillment_cx_or_panic(DUMMY_SP, &mut fulfill_cx, &vtable)
1298 /// Trait method, which has to be resolved to an impl method.
1299 pub fn trait_method(
1302 substs: &'tcx Substs<'tcx>
1303 ) -> (DefId, &'tcx Substs<'tcx>) {
1304 let method_item = self.tcx.impl_or_trait_item(def_id);
1305 let trait_id = method_item.container().id();
1306 let trait_ref = ty::Binder(substs.to_trait_ref(self.tcx, trait_id));
1307 match self.fulfill_obligation(trait_ref) {
1308 traits::VtableImpl(vtable_impl) => {
1309 let impl_did = vtable_impl.impl_def_id;
1310 let mname = self.tcx.item_name(def_id);
1311 // Create a concatenated set of substitutions which includes those from the impl
1312 // and those from the method:
1313 let impl_substs = vtable_impl.substs.with_method_from(substs);
1314 let substs = self.tcx.mk_substs(impl_substs);
1315 let mth = get_impl_method(self.tcx, impl_did, substs, mname);
1317 (mth.method.def_id, mth.substs)
1320 traits::VtableClosure(vtable_closure) =>
1321 (vtable_closure.closure_def_id, vtable_closure.substs.func_substs),
1323 traits::VtableFnPointer(_fn_ty) => {
1324 let _trait_closure_kind = self.tcx.lang_items.fn_trait_kind(trait_id).unwrap();
1326 // let llfn = trans_fn_pointer_shim(ccx, trait_closure_kind, fn_ty);
1328 // let method_ty = def_ty(tcx, def_id, substs);
1329 // let fn_ptr_ty = match method_ty.sty {
1330 // ty::TyFnDef(_, _, fty) => tcx.mk_ty(ty::TyFnPtr(fty)),
1331 // _ => unreachable!("expected fn item type, found {}",
1334 // Callee::ptr(immediate_rvalue(llfn, fn_ptr_ty))
1337 traits::VtableObject(ref _data) => {
1340 // data: Virtual(traits::get_vtable_index_of_object_method(
1341 // tcx, data, def_id)),
1342 // ty: def_ty(tcx, def_id, substs)
1345 vtable => unreachable!("resolved vtable bad vtable {:?} in trans", vtable),
1350 fn pointee_type(ptr_ty: ty::Ty) -> Option<ty::Ty> {
1352 ty::TyRef(_, ty::TypeAndMut { ty, .. }) |
1353 ty::TyRawPtr(ty::TypeAndMut { ty, .. }) |
1362 fn to_ptr(self) -> Pointer {
1363 assert_eq!(self.extra, LvalueExtra::None);
1368 impl<'mir, 'tcx: 'mir> Deref for CachedMir<'mir, 'tcx> {
1369 type Target = mir::Mir<'tcx>;
1370 fn deref(&self) -> &mir::Mir<'tcx> {
1372 CachedMir::Ref(r) => r,
1373 CachedMir::Owned(ref rc) => rc,
1379 pub struct ImplMethod<'tcx> {
1380 pub method: Rc<ty::Method<'tcx>>,
1381 pub substs: &'tcx Substs<'tcx>,
1382 pub is_provided: bool,
1385 /// Locates the applicable definition of a method, given its name.
1386 pub fn get_impl_method<'a, 'tcx>(
1387 tcx: TyCtxt<'a, 'tcx, 'tcx>,
1389 substs: &'tcx Substs<'tcx>,
1391 ) -> ImplMethod<'tcx> {
1392 assert!(!substs.types.needs_infer());
1394 let trait_def_id = tcx.trait_id_of_impl(impl_def_id).unwrap();
1395 let trait_def = tcx.lookup_trait_def(trait_def_id);
1397 match trait_def.ancestors(impl_def_id).fn_defs(tcx, name).next() {
1398 Some(node_item) => {
1399 let substs = tcx.normalizing_infer_ctxt(ProjectionMode::Any).enter(|infcx| {
1400 let substs = traits::translate_substs(&infcx, impl_def_id,
1401 substs, node_item.node);
1402 tcx.lift(&substs).unwrap_or_else(|| {
1403 bug!("trans::meth::get_impl_method: translate_substs \
1404 returned {:?} which contains inference types/regions",
1409 method: node_item.item,
1411 is_provided: node_item.node.is_from_trait(),
1415 bug!("method {:?} not found in {:?}", name, impl_def_id)
1420 pub fn interpret_start_points<'a, 'tcx>(
1421 tcx: TyCtxt<'a, 'tcx, 'tcx>,
1422 mir_map: &MirMap<'tcx>,
1424 for (&id, mir) in &mir_map.map {
1425 for attr in tcx.map.attrs(id) {
1426 use syntax::attr::AttrMetaMethods;
1427 if attr.check_name("miri_run") {
1428 let item = tcx.map.expect_item(id);
1430 if TRACE_EXECUTION {
1431 println!("Interpreting: {}", item.name);
1434 let mut gecx = GlobalEvalContext::new(tcx, mir_map);
1435 let mut fecx = FnEvalContext::new(&mut gecx);
1436 match fecx.call_nested(mir) {
1437 Ok(Some(return_ptr)) => if TRACE_EXECUTION {
1438 fecx.memory.dump(return_ptr.alloc_id);
1440 Ok(None) => println!("(diverging function returned)"),
1442 // TODO(solson): Detect whether the error was already reported or not.
1443 // tcx.sess.err(&e.to_string());
1447 if TRACE_EXECUTION {
1455 // TODO(solson): Upstream these methods into rustc::ty::layout.
1458 fn size(self) -> Size;
1461 impl IntegerExt for layout::Integer {
1462 fn size(self) -> Size {
1463 use rustc::ty::layout::Integer::*;
1465 I1 | I8 => Size::from_bits(8),
1466 I16 => Size::from_bits(16),
1467 I32 => Size::from_bits(32),
1468 I64 => Size::from_bits(64),
1474 fn field_offset(&self, index: usize) -> Size;
1477 impl StructExt for layout::Struct {
1478 fn field_offset(&self, index: usize) -> Size {
1482 self.offset_after_field[index - 1]