1000 lines
39 KiB
Rust
Raw Normal View History

2016-03-31 22:34:07 -06:00
use rustc::middle::const_val;
2016-04-14 00:01:00 +02:00
use rustc::hir::def_id::DefId;
2016-02-18 19:06:22 -06:00
use rustc::mir::mir_map::MirMap;
2016-03-07 10:14:47 -06:00
use rustc::mir::repr as mir;
use rustc::traits::ProjectionMode;
use rustc::ty::layout::{self, Layout, Size};
2016-03-28 17:43:23 -06:00
use rustc::ty::subst::{self, Subst, Substs};
use rustc::ty::{self, Ty, TyCtxt};
use rustc::util::nodemap::DefIdMap;
2016-06-11 12:38:28 -06:00
use rustc_data_structures::indexed_vec::Idx;
use std::cell::RefCell;
use std::ops::Deref;
use std::rc::Rc;
use std::iter;
use syntax::ast;
use syntax::codemap::{self, DUMMY_SP};
use error::{EvalError, EvalResult};
use memory::{Memory, Pointer};
use primval::{self, PrimVal};
use std::collections::HashMap;
2016-06-23 00:04:10 -06:00
mod step;
mod terminator;
2016-06-01 17:05:20 +02:00
pub struct EvalContext<'a, 'tcx: 'a> {
/// The results of the type checker, from rustc.
2016-05-13 22:34:50 -06:00
tcx: TyCtxt<'a, 'tcx, 'tcx>,
/// A mapping from NodeIds to Mir, from rustc. Only contains MIR for crate-local items.
mir_map: &'a MirMap<'tcx>,
/// A local cache from DefIds to Mir for non-crate-local items.
mir_cache: RefCell<DefIdMap<Rc<mir::Mir<'tcx>>>>,
/// The virtual memory system.
memory: Memory<'a, 'tcx>,
/// Precomputed statics, constants and promoteds.
2016-06-03 17:41:36 +02:00
statics: HashMap<ConstantId<'tcx>, Pointer>,
/// The virtual call stack.
stack: Vec<Frame<'a, 'tcx>>,
}
/// A stack frame.
pub struct Frame<'a, 'tcx: 'a> {
////////////////////////////////////////////////////////////////////////////////
// Function and callsite information
////////////////////////////////////////////////////////////////////////////////
/// The MIR for the function called on this frame.
pub mir: CachedMir<'a, 'tcx>,
/// The def_id of the current function.
pub def_id: DefId,
/// type substitutions for the current function invocation.
pub substs: &'tcx Substs<'tcx>,
/// The span of the call site.
pub span: codemap::Span,
////////////////////////////////////////////////////////////////////////////////
// Return pointer and local allocations
////////////////////////////////////////////////////////////////////////////////
/// A pointer for writing the return value of the current call if it's not a diverging call.
pub return_ptr: Option<Pointer>,
/// The list of locals for the current function, stored in order as
/// `[arguments..., variables..., temporaries...]`. The variables begin at `self.var_offset`
/// and the temporaries at `self.temp_offset`.
pub locals: Vec<Pointer>,
/// The offset of the first variable in `self.locals`.
pub var_offset: usize,
/// The offset of the first temporary in `self.locals`.
pub temp_offset: usize,
////////////////////////////////////////////////////////////////////////////////
// Current position within the function
////////////////////////////////////////////////////////////////////////////////
/// The block that is currently executed (or will be executed after the above call stacks
/// return).
pub block: mir::BasicBlock,
/// The index of the currently evaluated statment.
pub stmt: usize,
}
#[derive(Copy, Clone, Debug, Eq, PartialEq)]
struct Lvalue {
ptr: Pointer,
extra: LvalueExtra,
}
#[derive(Copy, Clone, Debug, Eq, PartialEq)]
enum LvalueExtra {
None,
Length(u64),
2016-05-09 20:08:37 -06:00
// TODO(solson): Vtable(memory::AllocId),
DowncastVariant(usize),
}
2016-03-14 21:48:00 -06:00
#[derive(Clone)]
pub enum CachedMir<'mir, 'tcx: 'mir> {
2016-03-14 21:48:00 -06:00
Ref(&'mir mir::Mir<'tcx>),
Owned(Rc<mir::Mir<'tcx>>)
}
#[derive(Clone, Debug, Eq, PartialEq, Hash)]
2016-06-08 12:35:15 +02:00
/// Uniquely identifies a specific constant or static
struct ConstantId<'tcx> {
2016-06-08 12:35:15 +02:00
/// the def id of the constant/static or in case of promoteds, the def id of the function they belong to
def_id: DefId,
2016-06-08 12:35:15 +02:00
/// In case of statics and constants this is `Substs::empty()`, so only promoteds and associated
/// constants actually have something useful here. We could special case statics and constants,
/// but that would only require more branching when working with constants, and not bring any
/// real benefits.
substs: &'tcx Substs<'tcx>,
kind: ConstantKind,
}
#[derive(Clone, Debug, Eq, PartialEq, Hash)]
enum ConstantKind {
2016-06-11 12:38:28 -06:00
Promoted(mir::Promoted),
2016-06-08 12:35:15 +02:00
/// Statics, constants and associated constants
Global,
2016-06-03 17:41:36 +02:00
}
impl<'a, 'tcx> EvalContext<'a, 'tcx> {
pub fn new(tcx: TyCtxt<'a, 'tcx, 'tcx>, mir_map: &'a MirMap<'tcx>) -> Self {
EvalContext {
2015-11-19 07:07:47 -06:00
tcx: tcx,
mir_map: mir_map,
mir_cache: RefCell::new(DefIdMap()),
memory: Memory::new(&tcx.data_layout),
2016-06-03 17:41:36 +02:00
statics: HashMap::new(),
stack: Vec::new(),
}
}
pub fn alloc_ret_ptr(&mut self, output_ty: ty::FnOutput<'tcx>, substs: &'tcx Substs<'tcx>) -> Option<Pointer> {
2016-06-08 12:30:25 +02:00
match output_ty {
ty::FnConverging(ty) => {
let size = self.type_size_with_substs(ty, substs);
Some(self.memory.allocate(size))
}
ty::FnDiverging => None,
}
}
2016-06-08 11:11:08 +02:00
pub fn memory(&self) -> &Memory {
&self.memory
}
pub fn memory_mut(&mut self) -> &mut Memory<'a, 'tcx> {
2016-06-15 12:55:04 +02:00
&mut self.memory
}
pub fn stack(&self) -> &[Frame] {
&self.stack
}
// TODO(solson): Try making const_to_primval instead.
fn const_to_ptr(&mut self, const_val: &const_val::ConstVal) -> EvalResult<'tcx, Pointer> {
use rustc::middle::const_val::ConstVal::*;
match *const_val {
Float(_f) => unimplemented!(),
Integral(int) => {
// TODO(solson): Check int constant type.
let ptr = self.memory.allocate(8);
self.memory.write_uint(ptr, int.to_u64_unchecked(), 8)?;
Ok(ptr)
}
Str(ref s) => {
let psize = self.memory.pointer_size();
let static_ptr = self.memory.allocate(s.len());
let ptr = self.memory.allocate(psize * 2);
self.memory.write_bytes(static_ptr, s.as_bytes())?;
self.memory.write_ptr(ptr, static_ptr)?;
self.memory.write_usize(ptr.offset(psize as isize), s.len() as u64)?;
Ok(ptr)
}
ByteStr(ref bs) => {
let psize = self.memory.pointer_size();
let static_ptr = self.memory.allocate(bs.len());
let ptr = self.memory.allocate(psize);
self.memory.write_bytes(static_ptr, bs)?;
self.memory.write_ptr(ptr, static_ptr)?;
Ok(ptr)
}
Bool(b) => {
let ptr = self.memory.allocate(1);
self.memory.write_bool(ptr, b)?;
Ok(ptr)
}
2016-06-20 12:29:45 +02:00
Char(c) => {
let ptr = self.memory.allocate(4);
self.memory.write_uint(ptr, c as u64, 4)?;
2016-06-20 12:29:45 +02:00
Ok(ptr)
},
Struct(_node_id) => unimplemented!(),
Tuple(_node_id) => unimplemented!(),
Function(_def_id) => unimplemented!(),
Array(_, _) => unimplemented!(),
Repeat(_, _) => unimplemented!(),
Dummy => unimplemented!(),
}
}
fn type_is_sized(&self, ty: Ty<'tcx>) -> bool {
ty.is_sized(self.tcx, &self.tcx.empty_parameter_environment(), DUMMY_SP)
}
fn load_mir(&self, def_id: DefId) -> CachedMir<'a, 'tcx> {
use rustc_trans::back::symbol_names::def_id_to_string;
match self.tcx.map.as_local_node_id(def_id) {
Some(node_id) => CachedMir::Ref(self.mir_map.map.get(&node_id).unwrap()),
None => {
let mut mir_cache = self.mir_cache.borrow_mut();
if let Some(mir) = mir_cache.get(&def_id) {
return CachedMir::Owned(mir.clone());
}
let cs = &self.tcx.sess.cstore;
let mir = cs.maybe_get_item_mir(self.tcx, def_id).unwrap_or_else(|| {
panic!("no mir for `{}`", def_id_to_string(self.tcx, def_id));
});
let cached = Rc::new(mir);
mir_cache.insert(def_id, cached.clone());
CachedMir::Owned(cached)
}
}
}
2016-06-08 11:11:08 +02:00
fn monomorphize(&self, ty: Ty<'tcx>, substs: &'tcx Substs<'tcx>) -> Ty<'tcx> {
let substituted = ty.subst(self.tcx, substs);
self.tcx.normalize_associated_type(&substituted)
}
fn type_size(&self, ty: Ty<'tcx>) -> usize {
self.type_size_with_substs(ty, self.substs())
2016-06-08 11:11:08 +02:00
}
fn type_size_with_substs(&self, ty: Ty<'tcx>, substs: &'tcx Substs<'tcx>) -> usize {
self.type_layout_with_substs(ty, substs).size(&self.tcx.data_layout).bytes() as usize
}
fn type_layout(&self, ty: Ty<'tcx>) -> &'tcx Layout {
self.type_layout_with_substs(ty, self.substs())
}
fn type_layout_with_substs(&self, ty: Ty<'tcx>, substs: &'tcx Substs<'tcx>) -> &'tcx Layout {
2016-06-08 11:11:08 +02:00
// TODO(solson): Is this inefficient? Needs investigation.
let ty = self.monomorphize(ty, substs);
self.tcx.normalizing_infer_ctxt(ProjectionMode::Any).enter(|infcx| {
// TODO(solson): Report this error properly.
ty.layout(&infcx).unwrap()
})
}
pub fn push_stack_frame(&mut self, def_id: DefId, span: codemap::Span, mir: CachedMir<'a, 'tcx>, substs: &'tcx Substs<'tcx>,
return_ptr: Option<Pointer>)
{
2016-03-18 10:48:31 -06:00
let arg_tys = mir.arg_decls.iter().map(|a| a.ty);
let var_tys = mir.var_decls.iter().map(|v| v.ty);
let temp_tys = mir.temp_decls.iter().map(|t| t.ty);
2016-03-18 10:48:31 -06:00
let num_args = mir.arg_decls.len();
let num_vars = mir.var_decls.len();
::log_settings::settings().indentation += 1;
let locals: Vec<Pointer> = arg_tys.chain(var_tys).chain(temp_tys).map(|ty| {
let size = self.type_size_with_substs(ty, substs);
self.memory.allocate(size)
}).collect();
self.stack.push(Frame {
mir: mir.clone(),
block: mir::START_BLOCK,
return_ptr: return_ptr,
locals: locals,
var_offset: num_args,
temp_offset: num_args + num_vars,
span: span,
def_id: def_id,
substs: substs,
stmt: 0,
});
}
fn pop_stack_frame(&mut self) {
::log_settings::settings().indentation -= 1;
let _frame = self.stack.pop().expect("tried to pop a stack frame, but there were none");
2016-05-09 20:08:37 -06:00
// TODO(solson): Deallocate local variables.
}
/// Applies the binary operation `op` to the two operands and writes a tuple of the result
/// and a boolean signifying the potential overflow to the destination.
fn intrinsic_with_overflow(
&mut self,
op: mir::BinOp,
left: &mir::Operand<'tcx>,
right: &mir::Operand<'tcx>,
dest: Pointer,
dest_layout: &'tcx Layout,
) -> EvalResult<'tcx, ()> {
use rustc::ty::layout::Layout::*;
let tup_layout = match *dest_layout {
Univariant { ref variant, .. } => variant,
_ => panic!("checked bin op returns something other than a tuple"),
};
let overflowed = self.intrinsic_overflowing(op, left, right, dest)?;
let offset = tup_layout.field_offset(1).bytes() as isize;
self.memory.write_bool(dest.offset(offset), overflowed)
}
2016-06-20 16:57:36 +02:00
/// Applies the binary operation `op` to the arguments and writes the result to the destination.
/// Returns `true` if the operation overflowed.
fn intrinsic_overflowing(
&mut self,
op: mir::BinOp,
left: &mir::Operand<'tcx>,
right: &mir::Operand<'tcx>,
2016-06-20 16:57:36 +02:00
dest: Pointer,
) -> EvalResult<'tcx, bool> {
let left_ptr = self.eval_operand(left)?;
let left_ty = self.operand_ty(left);
let left_val = self.read_primval(left_ptr, left_ty)?;
let right_ptr = self.eval_operand(right)?;
let right_ty = self.operand_ty(right);
let right_val = self.read_primval(right_ptr, right_ty)?;
2016-06-20 16:57:36 +02:00
let (val, overflow) = primval::binary_op(op, left_val, right_val)?;
self.memory.write_primval(dest, val)?;
Ok(overflow)
}
fn assign_fields<I: IntoIterator<Item = u64>>(
&mut self,
dest: Pointer,
offsets: I,
operands: &[mir::Operand<'tcx>],
) -> EvalResult<'tcx, ()> {
for (offset, operand) in offsets.into_iter().zip(operands) {
2016-05-09 18:52:44 -06:00
let src = self.eval_operand(operand)?;
let src_ty = self.operand_ty(operand);
let field_dest = dest.offset(offset as isize);
2016-05-09 18:52:44 -06:00
self.move_(src, field_dest, src_ty)?;
}
Ok(())
}
fn eval_assignment(&mut self, lvalue: &mir::Lvalue<'tcx>, rvalue: &mir::Rvalue<'tcx>)
-> EvalResult<'tcx, ()>
{
2016-05-09 18:52:44 -06:00
let dest = self.eval_lvalue(lvalue)?.to_ptr();
2016-04-07 05:56:07 -06:00
let dest_ty = self.lvalue_ty(lvalue);
let dest_layout = self.type_layout(dest_ty);
use rustc::mir::repr::Rvalue::*;
match *rvalue {
Use(ref operand) => {
2016-05-09 18:52:44 -06:00
let src = self.eval_operand(operand)?;
self.move_(src, dest, dest_ty)?;
}
BinaryOp(bin_op, ref left, ref right) => {
// ignore overflow bit, rustc inserts check branches for us
2016-06-20 16:57:36 +02:00
self.intrinsic_overflowing(bin_op, left, right, dest)?;
}
CheckedBinaryOp(bin_op, ref left, ref right) => {
self.intrinsic_with_overflow(bin_op, left, right, dest, dest_layout)?;
}
2016-06-11 12:38:28 -06:00
2016-03-07 07:57:08 -06:00
UnaryOp(un_op, ref operand) => {
2016-05-09 18:52:44 -06:00
let ptr = self.eval_operand(operand)?;
2016-03-17 02:53:03 -06:00
let ty = self.operand_ty(operand);
2016-05-09 18:52:44 -06:00
let val = self.read_primval(ptr, ty)?;
2016-05-30 15:27:52 +02:00
self.memory.write_primval(dest, primval::unary_op(un_op, val)?)?;
2016-03-07 07:57:08 -06:00
}
Aggregate(ref kind, ref operands) => {
use rustc::ty::layout::Layout::*;
match *dest_layout {
Univariant { ref variant, .. } => {
let offsets = iter::once(0)
.chain(variant.offset_after_field.iter().map(|s| s.bytes()));
2016-05-09 18:52:44 -06:00
self.assign_fields(dest, offsets, operands)?;
}
Array { .. } => {
let elem_size = match dest_ty.sty {
ty::TyArray(elem_ty, _) => self.type_size(elem_ty) as u64,
_ => panic!("tried to assign {:?} to non-array type {:?}",
kind, dest_ty),
};
let offsets = (0..).map(|i| i * elem_size);
2016-05-09 18:52:44 -06:00
self.assign_fields(dest, offsets, operands)?;
}
General { discr, ref variants, .. } => {
if let mir::AggregateKind::Adt(adt_def, variant, _) = *kind {
let discr_val = adt_def.variants[variant].disr_val.to_u64_unchecked();
let discr_size = discr.size().bytes() as usize;
2016-05-09 18:52:44 -06:00
self.memory.write_uint(dest, discr_val, discr_size)?;
let offsets = variants[variant].offset_after_field.iter()
.map(|s| s.bytes());
2016-05-09 18:52:44 -06:00
self.assign_fields(dest, offsets, operands)?;
} else {
panic!("tried to assign {:?} to Layout::General", kind);
2016-03-15 05:50:53 -06:00
}
}
RawNullablePointer { nndiscr, .. } => {
if let mir::AggregateKind::Adt(_, variant, _) = *kind {
if nndiscr == variant as u64 {
assert_eq!(operands.len(), 1);
let operand = &operands[0];
2016-05-09 18:52:44 -06:00
let src = self.eval_operand(operand)?;
let src_ty = self.operand_ty(operand);
2016-05-09 18:52:44 -06:00
self.move_(src, dest, src_ty)?;
} else {
assert_eq!(operands.len(), 0);
2016-05-09 18:52:44 -06:00
self.memory.write_isize(dest, 0)?;
}
} else {
panic!("tried to assign {:?} to Layout::RawNullablePointer", kind);
2016-03-15 05:50:53 -06:00
}
}
StructWrappedNullablePointer { nndiscr, ref nonnull, ref discrfield } => {
if let mir::AggregateKind::Adt(_, variant, _) = *kind {
if nndiscr == variant as u64 {
let offsets = iter::once(0)
.chain(nonnull.offset_after_field.iter().map(|s| s.bytes()));
try!(self.assign_fields(dest, offsets, operands));
} else {
assert_eq!(operands.len(), 0);
2016-05-30 15:27:52 +02:00
let offset = self.nonnull_offset(dest_ty, nndiscr, discrfield)?;
let dest = dest.offset(offset.bytes() as isize);
try!(self.memory.write_isize(dest, 0));
}
} else {
panic!("tried to assign {:?} to Layout::RawNullablePointer", kind);
}
}
CEnum { discr, signed, .. } => {
assert_eq!(operands.len(), 0);
if let mir::AggregateKind::Adt(adt_def, variant, _) = *kind {
let val = adt_def.variants[variant].disr_val.to_u64_unchecked();
let size = discr.size().bytes() as usize;
if signed {
2016-05-09 18:52:44 -06:00
self.memory.write_int(dest, val as i64, size)?;
} else {
2016-05-09 18:52:44 -06:00
self.memory.write_uint(dest, val, size)?;
}
} else {
panic!("tried to assign {:?} to Layout::CEnum", kind);
}
}
2016-05-30 15:27:52 +02:00
_ => return Err(EvalError::Unimplemented(format!("can't handle destination layout {:?} when assigning {:?}", dest_layout, kind))),
}
}
2016-03-21 03:34:24 -06:00
Repeat(ref operand, _) => {
let (elem_size, length) = match dest_ty.sty {
ty::TyArray(elem_ty, n) => (self.type_size(elem_ty), n),
_ => panic!("tried to assign array-repeat to non-array type {:?}", dest_ty),
};
2016-05-09 18:52:44 -06:00
let src = self.eval_operand(operand)?;
for i in 0..length {
let elem_dest = dest.offset((i * elem_size) as isize);
2016-05-09 18:52:44 -06:00
self.memory.copy(src, elem_dest, elem_size)?;
2016-03-21 03:34:24 -06:00
}
}
Len(ref lvalue) => {
2016-05-09 18:52:44 -06:00
let src = self.eval_lvalue(lvalue)?;
let ty = self.lvalue_ty(lvalue);
let len = match ty.sty {
ty::TyArray(_, n) => n as u64,
ty::TySlice(_) => if let LvalueExtra::Length(n) = src.extra {
n
} else {
panic!("Rvalue::Len of a slice given non-slice pointer: {:?}", src);
},
_ => panic!("Rvalue::Len expected array or slice, got {:?}", ty),
};
2016-05-09 18:52:44 -06:00
self.memory.write_usize(dest, len)?;
}
2016-03-13 14:36:25 -06:00
Ref(_, _, ref lvalue) => {
2016-05-09 18:52:44 -06:00
let lv = self.eval_lvalue(lvalue)?;
self.memory.write_ptr(dest, lv.ptr)?;
match lv.extra {
LvalueExtra::None => {},
LvalueExtra::Length(len) => {
let len_ptr = dest.offset(self.memory.pointer_size() as isize);
2016-05-09 18:52:44 -06:00
self.memory.write_usize(len_ptr, len)?;
}
LvalueExtra::DowncastVariant(..) =>
panic!("attempted to take a reference to an enum downcast lvalue"),
}
2016-03-13 14:36:25 -06:00
}
2015-12-28 22:24:05 -06:00
2016-03-14 22:05:50 -06:00
Box(ty) => {
let size = self.type_size(ty);
2016-03-17 03:19:13 -06:00
let ptr = self.memory.allocate(size);
2016-05-09 18:52:44 -06:00
self.memory.write_ptr(dest, ptr)?;
2016-03-14 22:05:50 -06:00
}
Cast(kind, ref operand, dest_ty) => {
use rustc::mir::repr::CastKind::*;
match kind {
Unsize => {
2016-06-08 13:43:34 +02:00
let src = self.eval_operand(operand)?;
let src_ty = self.operand_ty(operand);
2016-05-09 18:52:44 -06:00
self.move_(src, dest, src_ty)?;
let src_pointee_ty = pointee_type(src_ty).unwrap();
let dest_pointee_ty = pointee_type(dest_ty).unwrap();
match (&src_pointee_ty.sty, &dest_pointee_ty.sty) {
2016-03-17 08:01:34 -06:00
(&ty::TyArray(_, length), &ty::TySlice(_)) => {
let len_ptr = dest.offset(self.memory.pointer_size() as isize);
2016-05-09 18:52:44 -06:00
self.memory.write_usize(len_ptr, length as u64)?;
2016-03-17 08:01:34 -06:00
}
2016-05-30 15:27:52 +02:00
_ => return Err(EvalError::Unimplemented(format!("can't handle cast: {:?}", rvalue))),
}
}
Misc => {
2016-06-08 13:43:34 +02:00
let src = self.eval_operand(operand)?;
let src_ty = self.operand_ty(operand);
2016-05-09 20:08:37 -06:00
// FIXME(solson): Wrong for almost everything.
warn!("misc cast from {:?} to {:?}", src_ty, dest_ty);
let dest_size = self.type_size(dest_ty);
let src_size = self.type_size(src_ty);
// Hack to support fat pointer -> thin pointer casts to keep tests for
// other things passing for now.
2016-06-17 16:03:11 +02:00
let is_fat_ptr_cast = pointee_type(src_ty).map_or(false, |ty| !self.type_is_sized(ty));
if dest_size == src_size || is_fat_ptr_cast {
self.memory.copy(src, dest, dest_size)?;
} else {
return Err(EvalError::Unimplemented(format!("can't handle cast: {:?}", rvalue)));
}
}
2016-06-08 13:43:34 +02:00
ReifyFnPointer => match self.operand_ty(operand).sty {
ty::TyFnDef(def_id, substs, fn_ty) => {
let fn_ptr = self.memory.create_fn_ptr(def_id, substs, fn_ty);
2016-06-08 13:43:34 +02:00
self.memory.write_ptr(dest, fn_ptr)?;
},
ref other => panic!("reify fn pointer on {:?}", other),
},
UnsafeFnPointer => match dest_ty.sty {
ty::TyFnPtr(unsafe_fn_ty) => {
let src = self.eval_operand(operand)?;
let ptr = self.memory.read_ptr(src)?;
let fn_def = self.memory.get_fn(ptr.alloc_id)?;
let fn_ptr = self.memory.create_fn_ptr(fn_def.def_id, fn_def.substs, unsafe_fn_ty);
self.memory.write_ptr(dest, fn_ptr)?;
},
ref other => panic!("fn to unsafe fn cast on {:?}", other),
},
}
}
2016-03-28 17:43:23 -06:00
InlineAsm { .. } => unimplemented!(),
}
Ok(())
}
fn nonnull_offset(&self, ty: Ty<'tcx>, nndiscr: u64, discrfield: &[u32]) -> EvalResult<'tcx, Size> {
// Skip the constant 0 at the start meant for LLVM GEP.
let mut path = discrfield.iter().skip(1).map(|&i| i as usize);
// Handle the field index for the outer non-null variant.
let inner_ty = match ty.sty {
ty::TyEnum(adt_def, substs) => {
let variant = &adt_def.variants[nndiscr as usize];
let index = path.next().unwrap();
let field = &variant.fields[index];
field.ty(self.tcx, substs)
}
_ => panic!(
"non-enum for StructWrappedNullablePointer: {}",
ty,
),
};
self.field_path_offset(inner_ty, path)
}
fn field_path_offset<I: Iterator<Item = usize>>(&self, mut ty: Ty<'tcx>, path: I) -> EvalResult<'tcx, Size> {
let mut offset = Size::from_bytes(0);
// Skip the initial 0 intended for LLVM GEP.
for field_index in path {
2016-05-30 15:27:52 +02:00
let field_offset = self.get_field_offset(ty, field_index)?;
ty = self.get_field_ty(ty, field_index)?;
offset = offset.checked_add(field_offset, &self.tcx.data_layout).unwrap();
}
2016-05-30 15:27:52 +02:00
Ok(offset)
}
fn get_field_ty(&self, ty: Ty<'tcx>, field_index: usize) -> EvalResult<'tcx, Ty<'tcx>> {
match ty.sty {
ty::TyStruct(adt_def, substs) => {
2016-05-30 15:27:52 +02:00
Ok(adt_def.struct_variant().fields[field_index].ty(self.tcx, substs))
}
ty::TyRef(_, ty::TypeAndMut { ty, .. }) |
ty::TyRawPtr(ty::TypeAndMut { ty, .. }) |
ty::TyBox(ty) => {
assert_eq!(field_index, 0);
2016-05-30 15:27:52 +02:00
Ok(ty)
}
2016-05-30 15:27:52 +02:00
_ => Err(EvalError::Unimplemented(format!("can't handle type: {:?}", ty))),
}
}
fn get_field_offset(&self, ty: Ty<'tcx>, field_index: usize) -> EvalResult<'tcx, Size> {
let layout = self.type_layout(ty);
use rustc::ty::layout::Layout::*;
match *layout {
Univariant { .. } => {
assert_eq!(field_index, 0);
2016-05-30 15:27:52 +02:00
Ok(Size::from_bytes(0))
}
FatPointer { .. } => {
let bytes = layout::FAT_PTR_ADDR * self.memory.pointer_size();
2016-05-30 15:27:52 +02:00
Ok(Size::from_bytes(bytes as u64))
}
2016-05-30 15:27:52 +02:00
_ => Err(EvalError::Unimplemented(format!("can't handle type: {:?}, with layout: {:?}", ty, layout))),
}
}
fn eval_operand(&mut self, op: &mir::Operand<'tcx>) -> EvalResult<'tcx, Pointer> {
use rustc::mir::repr::Operand::*;
match *op {
2016-05-09 18:52:44 -06:00
Consume(ref lvalue) => Ok(self.eval_lvalue(lvalue)?.to_ptr()),
Constant(mir::Constant { ref literal, ty, .. }) => {
use rustc::mir::repr::Literal::*;
2016-03-13 01:14:20 -06:00
match *literal {
2016-05-09 18:52:44 -06:00
Value { ref value } => Ok(self.const_to_ptr(value)?),
2016-06-03 15:48:56 +02:00
Item { def_id, substs } => {
if let ty::TyFnDef(..) = ty.sty {
2016-06-08 13:43:34 +02:00
// function items are zero sized
Ok(self.memory.allocate(0))
2016-06-03 15:48:56 +02:00
} else {
let cid = ConstantId {
def_id: def_id,
substs: substs,
2016-06-08 12:35:15 +02:00
kind: ConstantKind::Global,
};
2016-06-03 17:41:36 +02:00
Ok(*self.statics.get(&cid).expect("static should have been cached (rvalue)"))
2016-06-03 15:48:56 +02:00
}
},
2016-06-03 17:41:36 +02:00
Promoted { index } => {
let cid = ConstantId {
2016-06-03 17:41:36 +02:00
def_id: self.frame().def_id,
substs: self.substs(),
kind: ConstantKind::Promoted(index),
2016-06-03 17:41:36 +02:00
};
Ok(*self.statics.get(&cid).expect("a promoted constant hasn't been precomputed"))
},
}
}
}
}
2015-11-12 17:44:29 -06:00
fn eval_lvalue(&mut self, lvalue: &mir::Lvalue<'tcx>) -> EvalResult<'tcx, Lvalue> {
2016-03-07 07:48:38 -06:00
use rustc::mir::repr::Lvalue::*;
let ptr = match *lvalue {
2016-03-20 22:07:25 -06:00
ReturnPointer => self.frame().return_ptr
.expect("ReturnPointer used in a function with no return value"),
2016-06-11 12:38:28 -06:00
Arg(i) => self.frame().locals[i.index()],
Var(i) => self.frame().locals[self.frame().var_offset + i.index()],
Temp(i) => self.frame().locals[self.frame().temp_offset + i.index()],
2016-03-13 06:48:04 -06:00
2016-06-03 17:41:36 +02:00
Static(def_id) => {
let substs = self.tcx.mk_substs(subst::Substs::empty());
let cid = ConstantId {
def_id: def_id,
substs: substs,
2016-06-08 12:35:15 +02:00
kind: ConstantKind::Global,
};
*self.statics.get(&cid).expect("static should have been cached (lvalue)")
2016-06-03 17:41:36 +02:00
},
2016-03-13 06:48:04 -06:00
Projection(ref proj) => {
2016-05-09 18:52:44 -06:00
let base = self.eval_lvalue(&proj.base)?;
let base_ty = self.lvalue_ty(&proj.base);
let base_layout = self.type_layout(base_ty);
2016-03-13 06:48:04 -06:00
use rustc::mir::repr::ProjectionElem::*;
match proj.elem {
Field(field, _) => {
use rustc::ty::layout::Layout::*;
let variant = match *base_layout {
Univariant { ref variant, .. } => variant,
General { ref variants, .. } => {
if let LvalueExtra::DowncastVariant(variant_idx) = base.extra {
&variants[variant_idx]
} else {
panic!("field access on enum had no variant index");
}
}
RawNullablePointer { .. } => {
assert_eq!(field.index(), 0);
return Ok(base);
}
StructWrappedNullablePointer { ref nonnull, .. } => nonnull,
_ => panic!("field access on non-product type: {:?}", base_layout),
};
let offset = variant.field_offset(field.index()).bytes();
base.ptr.offset(offset as isize)
},
Downcast(_, variant) => {
use rustc::ty::layout::Layout::*;
match *base_layout {
General { discr, .. } => {
return Ok(Lvalue {
ptr: base.ptr.offset(discr.size().bytes() as isize),
extra: LvalueExtra::DowncastVariant(variant),
});
}
RawNullablePointer { .. } | StructWrappedNullablePointer { .. } => {
return Ok(base);
}
_ => panic!("variant downcast on non-aggregate: {:?}", base_layout),
}
2016-03-13 06:48:04 -06:00
},
Deref => {
let pointee_ty = pointee_type(base_ty).expect("Deref of non-pointer");
2016-05-09 18:52:44 -06:00
let ptr = self.memory.read_ptr(base.ptr)?;
let extra = match pointee_ty.sty {
ty::TySlice(_) | ty::TyStr => {
let len_ptr = base.ptr.offset(self.memory.pointer_size() as isize);
2016-05-09 18:52:44 -06:00
let len = self.memory.read_usize(len_ptr)?;
LvalueExtra::Length(len)
}
ty::TyTrait(_) => unimplemented!(),
_ => LvalueExtra::None,
};
return Ok(Lvalue { ptr: ptr, extra: extra });
}
2016-03-13 14:36:25 -06:00
Index(ref operand) => {
let elem_size = match base_ty.sty {
2016-04-29 06:01:17 +02:00
ty::TyArray(elem_ty, _) |
ty::TySlice(elem_ty) => self.type_size(elem_ty),
_ => panic!("indexing expected an array or slice, got {:?}", base_ty),
};
2016-05-09 18:52:44 -06:00
let n_ptr = self.eval_operand(operand)?;
let n = self.memory.read_usize(n_ptr)?;
base.ptr.offset(n as isize * elem_size as isize)
}
ConstantIndex { .. } => unimplemented!(),
2016-06-11 12:38:28 -06:00
Subslice { .. } => unimplemented!(),
2016-03-13 06:48:04 -06:00
}
}
2016-03-07 07:48:38 -06:00
};
Ok(Lvalue { ptr: ptr, extra: LvalueExtra::None })
2016-03-07 07:48:38 -06:00
}
fn lvalue_ty(&self, lvalue: &mir::Lvalue<'tcx>) -> Ty<'tcx> {
self.monomorphize(self.mir().lvalue_ty(self.tcx, lvalue).to_ty(self.tcx), self.substs())
2016-03-20 23:11:06 -06:00
}
fn operand_ty(&self, operand: &mir::Operand<'tcx>) -> Ty<'tcx> {
self.monomorphize(self.mir().operand_ty(self.tcx, operand), self.substs())
2016-04-07 03:02:02 -06:00
}
fn move_(&mut self, src: Pointer, dest: Pointer, ty: Ty<'tcx>) -> EvalResult<'tcx, ()> {
let size = self.type_size(ty);
2016-05-09 18:52:44 -06:00
self.memory.copy(src, dest, size)?;
2016-04-07 05:56:07 -06:00
if self.type_needs_drop(ty) {
2016-05-09 18:52:44 -06:00
self.memory.drop_fill(src, size)?;
2016-04-07 05:56:07 -06:00
}
Ok(())
}
pub fn read_primval(&mut self, ptr: Pointer, ty: Ty<'tcx>) -> EvalResult<'tcx, PrimVal> {
use syntax::ast::{IntTy, UintTy};
let val = match (self.memory.pointer_size(), &ty.sty) {
2016-05-31 12:05:25 +02:00
(_, &ty::TyBool) => PrimVal::Bool(self.memory.read_bool(ptr)?),
2016-06-20 12:29:45 +02:00
(_, &ty::TyChar) => {
let c = self.memory.read_uint(ptr, 4)? as u32;
match ::std::char::from_u32(c) {
Some(ch) => PrimVal::Char(ch),
None => return Err(EvalError::InvalidChar(c)),
}
}
2016-05-31 12:05:25 +02:00
(_, &ty::TyInt(IntTy::I8)) => PrimVal::I8(self.memory.read_int(ptr, 1)? as i8),
(2, &ty::TyInt(IntTy::Is)) |
(_, &ty::TyInt(IntTy::I16)) => PrimVal::I16(self.memory.read_int(ptr, 2)? as i16),
(4, &ty::TyInt(IntTy::Is)) |
(_, &ty::TyInt(IntTy::I32)) => PrimVal::I32(self.memory.read_int(ptr, 4)? as i32),
(8, &ty::TyInt(IntTy::Is)) |
(_, &ty::TyInt(IntTy::I64)) => PrimVal::I64(self.memory.read_int(ptr, 8)? as i64),
(_, &ty::TyUint(UintTy::U8)) => PrimVal::U8(self.memory.read_uint(ptr, 1)? as u8),
(2, &ty::TyUint(UintTy::Us)) |
(_, &ty::TyUint(UintTy::U16)) => PrimVal::U16(self.memory.read_uint(ptr, 2)? as u16),
(4, &ty::TyUint(UintTy::Us)) |
(_, &ty::TyUint(UintTy::U32)) => PrimVal::U32(self.memory.read_uint(ptr, 4)? as u32),
(8, &ty::TyUint(UintTy::Us)) |
(_, &ty::TyUint(UintTy::U64)) => PrimVal::U64(self.memory.read_uint(ptr, 8)? as u64),
2016-06-20 10:35:15 +02:00
(_, &ty::TyFnDef(def_id, substs, fn_ty)) => {
PrimVal::FnPtr(self.memory.create_fn_ptr(def_id, substs, fn_ty))
},
(_, &ty::TyFnPtr(_)) => self.memory.read_ptr(ptr).map(PrimVal::FnPtr)?,
2016-05-31 12:05:25 +02:00
(_, &ty::TyRef(_, ty::TypeAndMut { ty, .. })) |
(_, &ty::TyRawPtr(ty::TypeAndMut { ty, .. })) => {
if self.type_is_sized(ty) {
match self.memory.read_ptr(ptr) {
Ok(p) => PrimVal::AbstractPtr(p),
Err(EvalError::ReadBytesAsPointer) => {
2016-05-09 18:52:44 -06:00
PrimVal::IntegerPtr(self.memory.read_usize(ptr)?)
}
Err(e) => return Err(e),
}
} else {
2016-05-30 15:27:52 +02:00
return Err(EvalError::Unimplemented(format!("unimplemented: primitive read of fat pointer type: {:?}", ty)));
}
}
_ => panic!("primitive read of non-primitive type: {:?}", ty),
};
Ok(val)
}
fn frame(&self) -> &Frame<'a, 'tcx> {
self.stack.last().expect("no call frames exist")
}
2016-06-15 12:55:04 +02:00
pub fn frame_mut(&mut self) -> &mut Frame<'a, 'tcx> {
self.stack.last_mut().expect("no call frames exist")
}
fn mir(&self) -> CachedMir<'a, 'tcx> {
self.frame().mir.clone()
2016-03-20 22:07:25 -06:00
}
2016-06-08 11:11:08 +02:00
fn substs(&self) -> &'tcx Substs<'tcx> {
self.frame().substs
}
2016-03-14 21:48:00 -06:00
}
2016-04-29 06:01:17 +02:00
fn pointee_type(ptr_ty: ty::Ty) -> Option<ty::Ty> {
match ptr_ty.sty {
ty::TyRef(_, ty::TypeAndMut { ty, .. }) |
ty::TyRawPtr(ty::TypeAndMut { ty, .. }) |
ty::TyBox(ty) => {
Some(ty)
}
_ => None,
}
}
impl Lvalue {
fn to_ptr(self) -> Pointer {
assert_eq!(self.extra, LvalueExtra::None);
self.ptr
}
}
2016-03-14 21:48:00 -06:00
impl<'mir, 'tcx: 'mir> Deref for CachedMir<'mir, 'tcx> {
type Target = mir::Mir<'tcx>;
fn deref(&self) -> &mir::Mir<'tcx> {
match *self {
CachedMir::Ref(r) => r,
2016-05-27 16:12:17 +02:00
CachedMir::Owned(ref rc) => rc,
2016-03-14 21:48:00 -06:00
}
}
}
pub fn eval_main<'a, 'tcx: 'a>(
tcx: TyCtxt<'a, 'tcx, 'tcx>,
mir_map: &'a MirMap<'tcx>,
node_id: ast::NodeId,
) {
let mir = mir_map.map.get(&node_id).expect("no mir for main function");
let def_id = tcx.map.local_def_id(node_id);
let mut ecx = EvalContext::new(tcx, mir_map);
let substs = tcx.mk_substs(subst::Substs::empty());
let return_ptr = ecx.alloc_ret_ptr(mir.return_ty, substs).expect("main function should not be diverging");
ecx.push_stack_frame(def_id, mir.span, CachedMir::Ref(mir), substs, Some(return_ptr));
if mir.arg_decls.len() == 2 {
// start function
let ptr_size = ecx.memory().pointer_size();
let nargs = ecx.memory_mut().allocate(ptr_size);
ecx.memory_mut().write_usize(nargs, 0).unwrap();
let args = ecx.memory_mut().allocate(ptr_size);
ecx.memory_mut().write_usize(args, 0).unwrap();
ecx.frame_mut().locals[0] = nargs;
ecx.frame_mut().locals[1] = args;
}
loop {
match ecx.step() {
Ok(true) => {}
Ok(false) => break,
// FIXME: diverging functions can end up here in some future miri
Err(e) => {
report(tcx, &ecx, e);
break;
}
}
}
}
fn report(tcx: TyCtxt, ecx: &EvalContext, e: EvalError) {
let frame = ecx.stack().last().expect("stackframe was empty");
let block = &frame.mir.basic_blocks()[frame.block];
let span = if frame.stmt < block.statements.len() {
block.statements[frame.stmt].source_info.span
} else {
block.terminator().source_info.span
};
let mut err = tcx.sess.struct_span_err(span, &e.to_string());
for &Frame { def_id, substs, span, .. } in ecx.stack().iter().rev() {
// FIXME(solson): Find a way to do this without this Display impl hack.
use rustc::util::ppaux;
use std::fmt;
struct Instance<'tcx>(DefId, &'tcx subst::Substs<'tcx>);
impl<'tcx> fmt::Display for Instance<'tcx> {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
ppaux::parameterized(f, self.1, self.0, ppaux::Ns::Value, &[],
|tcx| Some(tcx.lookup_item_type(self.0).generics))
}
}
err.span_note(span, &format!("inside call to {}", Instance(def_id, substs)));
}
err.emit();
}
// TODO(solson): Upstream these methods into rustc::ty::layout.
trait IntegerExt {
fn size(self) -> Size;
}
impl IntegerExt for layout::Integer {
fn size(self) -> Size {
use rustc::ty::layout::Integer::*;
match self {
I1 | I8 => Size::from_bits(8),
I16 => Size::from_bits(16),
I32 => Size::from_bits(32),
I64 => Size::from_bits(64),
}
}
}
trait StructExt {
fn field_offset(&self, index: usize) -> Size;
}
impl StructExt for layout::Struct {
fn field_offset(&self, index: usize) -> Size {
if index == 0 {
Size::from_bytes(0)
} else {
self.offset_after_field[index - 1]
}
}
}