1
Fork 0

Auto merge of #82124 - tmiasko:op-ty-ref, r=oli-obk

Pass large interpreter types by reference, not value

r? `@ghost`
This commit is contained in:
bors 2021-02-20 10:20:42 +00:00
commit e7c23ab933
17 changed files with 401 additions and 358 deletions

View file

@ -56,7 +56,7 @@ fn eval_body_using_ecx<'mir, 'tcx>(
ecx.push_stack_frame( ecx.push_stack_frame(
cid.instance, cid.instance,
body, body,
Some(ret.into()), Some(&ret.into()),
StackPopCleanup::None { cleanup: false }, StackPopCleanup::None { cleanup: false },
)?; )?;
@ -72,7 +72,7 @@ fn eval_body_using_ecx<'mir, 'tcx>(
None => InternKind::Constant, None => InternKind::Constant,
} }
}; };
intern_const_alloc_recursive(ecx, intern_kind, ret)?; intern_const_alloc_recursive(ecx, intern_kind, &ret)?;
debug!("eval_body_using_ecx done: {:?}", *ret); debug!("eval_body_using_ecx done: {:?}", *ret);
Ok(ret) Ok(ret)
@ -105,7 +105,7 @@ pub(super) fn mk_eval_cx<'mir, 'tcx>(
/// type system. /// type system.
pub(super) fn op_to_const<'tcx>( pub(super) fn op_to_const<'tcx>(
ecx: &CompileTimeEvalContext<'_, 'tcx>, ecx: &CompileTimeEvalContext<'_, 'tcx>,
op: OpTy<'tcx>, op: &OpTy<'tcx>,
) -> ConstValue<'tcx> { ) -> ConstValue<'tcx> {
// We do not have value optimizations for everything. // We do not have value optimizations for everything.
// Only scalars and slices, since they are very common. // Only scalars and slices, since they are very common.
@ -137,7 +137,7 @@ pub(super) fn op_to_const<'tcx>(
op.try_as_mplace(ecx) op.try_as_mplace(ecx)
}; };
let to_const_value = |mplace: MPlaceTy<'_>| match mplace.ptr { let to_const_value = |mplace: &MPlaceTy<'_>| match mplace.ptr {
Scalar::Ptr(ptr) => { Scalar::Ptr(ptr) => {
let alloc = ecx.tcx.global_alloc(ptr.alloc_id).unwrap_memory(); let alloc = ecx.tcx.global_alloc(ptr.alloc_id).unwrap_memory();
ConstValue::ByRef { alloc, offset: ptr.offset } ConstValue::ByRef { alloc, offset: ptr.offset }
@ -155,12 +155,12 @@ pub(super) fn op_to_const<'tcx>(
} }
}; };
match immediate { match immediate {
Ok(mplace) => to_const_value(mplace), Ok(ref mplace) => to_const_value(mplace),
// see comment on `let try_as_immediate` above // see comment on `let try_as_immediate` above
Err(imm) => match *imm { Err(imm) => match *imm {
Immediate::Scalar(x) => match x { Immediate::Scalar(x) => match x {
ScalarMaybeUninit::Scalar(s) => ConstValue::Scalar(s), ScalarMaybeUninit::Scalar(s) => ConstValue::Scalar(s),
ScalarMaybeUninit::Uninit => to_const_value(op.assert_mem_place(ecx)), ScalarMaybeUninit::Uninit => to_const_value(&op.assert_mem_place(ecx)),
}, },
Immediate::ScalarPair(a, b) => { Immediate::ScalarPair(a, b) => {
let (data, start) = match a.check_init().unwrap() { let (data, start) = match a.check_init().unwrap() {
@ -201,7 +201,7 @@ fn turn_into_const_value<'tcx>(
"the `eval_to_const_value_raw` query should not be used for statics, use `eval_to_allocation` instead" "the `eval_to_const_value_raw` query should not be used for statics, use `eval_to_allocation` instead"
); );
// Turn this into a proper constant. // Turn this into a proper constant.
op_to_const(&ecx, mplace.into()) op_to_const(&ecx, &mplace.into())
} }
pub fn eval_to_const_value_raw_provider<'tcx>( pub fn eval_to_const_value_raw_provider<'tcx>(
@ -348,7 +348,7 @@ pub fn eval_to_allocation_raw_provider<'tcx>(
Some(_) => CtfeValidationMode::Regular, // a `static` Some(_) => CtfeValidationMode::Regular, // a `static`
None => CtfeValidationMode::Const { inner, allow_static_ptrs: false }, None => CtfeValidationMode::Const { inner, allow_static_ptrs: false },
}; };
ecx.const_validate_operand(mplace.into(), path, &mut ref_tracking, mode)?; ecx.const_validate_operand(&mplace.into(), path, &mut ref_tracking, mode)?;
inner = true; inner = true;
} }
}; };

View file

@ -39,8 +39,8 @@ impl<'mir, 'tcx> InterpCx<'mir, 'tcx, CompileTimeInterpreter<'mir, 'tcx>> {
// &str // &str
assert!(args.len() == 1); assert!(args.len() == 1);
let msg_place = self.deref_operand(args[0])?; let msg_place = self.deref_operand(&args[0])?;
let msg = Symbol::intern(self.read_str(msg_place)?); let msg = Symbol::intern(self.read_str(&msg_place)?);
let span = self.find_closest_untracked_caller_location(); let span = self.find_closest_untracked_caller_location();
let (file, line, col) = self.location_triple_for_span(span); let (file, line, col) = self.location_triple_for_span(span);
Err(ConstEvalErrKind::Panic { msg, file, line, col }.into()) Err(ConstEvalErrKind::Panic { msg, file, line, col }.into())
@ -222,7 +222,7 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for CompileTimeInterpreter<'mir,
instance: ty::Instance<'tcx>, instance: ty::Instance<'tcx>,
_abi: Abi, _abi: Abi,
args: &[OpTy<'tcx>], args: &[OpTy<'tcx>],
_ret: Option<(PlaceTy<'tcx>, mir::BasicBlock)>, _ret: Option<(&PlaceTy<'tcx>, mir::BasicBlock)>,
_unwind: Option<mir::BasicBlock>, // unwinding is not supported in consts _unwind: Option<mir::BasicBlock>, // unwinding is not supported in consts
) -> InterpResult<'tcx, Option<&'mir mir::Body<'tcx>>> { ) -> InterpResult<'tcx, Option<&'mir mir::Body<'tcx>>> {
debug!("find_mir_or_eval_fn: {:?}", instance); debug!("find_mir_or_eval_fn: {:?}", instance);
@ -262,7 +262,7 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for CompileTimeInterpreter<'mir,
ecx: &mut InterpCx<'mir, 'tcx, Self>, ecx: &mut InterpCx<'mir, 'tcx, Self>,
instance: ty::Instance<'tcx>, instance: ty::Instance<'tcx>,
args: &[OpTy<'tcx>], args: &[OpTy<'tcx>],
ret: Option<(PlaceTy<'tcx>, mir::BasicBlock)>, ret: Option<(&PlaceTy<'tcx>, mir::BasicBlock)>,
_unwind: Option<mir::BasicBlock>, _unwind: Option<mir::BasicBlock>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// Shared intrinsics. // Shared intrinsics.
@ -284,8 +284,8 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for CompileTimeInterpreter<'mir,
}; };
match intrinsic_name { match intrinsic_name {
sym::ptr_guaranteed_eq | sym::ptr_guaranteed_ne => { sym::ptr_guaranteed_eq | sym::ptr_guaranteed_ne => {
let a = ecx.read_immediate(args[0])?.to_scalar()?; let a = ecx.read_immediate(&args[0])?.to_scalar()?;
let b = ecx.read_immediate(args[1])?.to_scalar()?; let b = ecx.read_immediate(&args[1])?.to_scalar()?;
let cmp = if intrinsic_name == sym::ptr_guaranteed_eq { let cmp = if intrinsic_name == sym::ptr_guaranteed_eq {
ecx.guaranteed_eq(a, b) ecx.guaranteed_eq(a, b)
} else { } else {
@ -294,8 +294,8 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for CompileTimeInterpreter<'mir,
ecx.write_scalar(Scalar::from_bool(cmp), dest)?; ecx.write_scalar(Scalar::from_bool(cmp), dest)?;
} }
sym::const_allocate => { sym::const_allocate => {
let size = ecx.read_scalar(args[0])?.to_machine_usize(ecx)?; let size = ecx.read_scalar(&args[0])?.to_machine_usize(ecx)?;
let align = ecx.read_scalar(args[1])?.to_machine_usize(ecx)?; let align = ecx.read_scalar(&args[1])?.to_machine_usize(ecx)?;
let align = match Align::from_bytes(align) { let align = match Align::from_bytes(align) {
Ok(a) => a, Ok(a) => a,
@ -330,7 +330,7 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for CompileTimeInterpreter<'mir,
use rustc_middle::mir::AssertKind::*; use rustc_middle::mir::AssertKind::*;
// Convert `AssertKind<Operand>` to `AssertKind<Scalar>`. // Convert `AssertKind<Operand>` to `AssertKind<Scalar>`.
let eval_to_int = let eval_to_int =
|op| ecx.read_immediate(ecx.eval_operand(op, None)?).map(|x| x.to_const_int()); |op| ecx.read_immediate(&ecx.eval_operand(op, None)?).map(|x| x.to_const_int());
let err = match msg { let err = match msg {
BoundsCheck { ref len, ref index } => { BoundsCheck { ref len, ref index } => {
let len = eval_to_int(len)?; let len = eval_to_int(len)?;
@ -358,15 +358,15 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for CompileTimeInterpreter<'mir,
fn binary_ptr_op( fn binary_ptr_op(
_ecx: &InterpCx<'mir, 'tcx, Self>, _ecx: &InterpCx<'mir, 'tcx, Self>,
_bin_op: mir::BinOp, _bin_op: mir::BinOp,
_left: ImmTy<'tcx>, _left: &ImmTy<'tcx>,
_right: ImmTy<'tcx>, _right: &ImmTy<'tcx>,
) -> InterpResult<'tcx, (Scalar, bool, Ty<'tcx>)> { ) -> InterpResult<'tcx, (Scalar, bool, Ty<'tcx>)> {
Err(ConstEvalErrKind::NeedsRfc("pointer arithmetic or comparison".to_string()).into()) Err(ConstEvalErrKind::NeedsRfc("pointer arithmetic or comparison".to_string()).into())
} }
fn box_alloc( fn box_alloc(
_ecx: &mut InterpCx<'mir, 'tcx, Self>, _ecx: &mut InterpCx<'mir, 'tcx, Self>,
_dest: PlaceTy<'tcx>, _dest: &PlaceTy<'tcx>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
Err(ConstEvalErrKind::NeedsRfc("heap allocations via `box` keyword".to_string()).into()) Err(ConstEvalErrKind::NeedsRfc("heap allocations via `box` keyword".to_string()).into())
} }

View file

@ -29,7 +29,7 @@ pub(crate) fn const_caller_location(
let mut ecx = mk_eval_cx(tcx, DUMMY_SP, ty::ParamEnv::reveal_all(), false); let mut ecx = mk_eval_cx(tcx, DUMMY_SP, ty::ParamEnv::reveal_all(), false);
let loc_place = ecx.alloc_caller_location(file, line, col); let loc_place = ecx.alloc_caller_location(file, line, col);
if intern_const_alloc_recursive(&mut ecx, InternKind::Constant, loc_place).is_err() { if intern_const_alloc_recursive(&mut ecx, InternKind::Constant, &loc_place).is_err() {
bug!("intern_const_alloc_recursive should not error in this case") bug!("intern_const_alloc_recursive should not error in this case")
} }
ConstValue::Scalar(loc_place.ptr) ConstValue::Scalar(loc_place.ptr)
@ -55,8 +55,8 @@ pub(crate) fn destructure_const<'tcx>(
return mir::DestructuredConst { variant: None, fields: &[] }; return mir::DestructuredConst { variant: None, fields: &[] };
} }
ty::Adt(def, _) => { ty::Adt(def, _) => {
let variant = ecx.read_discriminant(op).unwrap().1; let variant = ecx.read_discriminant(&op).unwrap().1;
let down = ecx.operand_downcast(op, variant).unwrap(); let down = ecx.operand_downcast(&op, variant).unwrap();
(def.variants[variant].fields.len(), Some(variant), down) (def.variants[variant].fields.len(), Some(variant), down)
} }
ty::Tuple(substs) => (substs.len(), None, op), ty::Tuple(substs) => (substs.len(), None, op),
@ -64,8 +64,8 @@ pub(crate) fn destructure_const<'tcx>(
}; };
let fields_iter = (0..field_count).map(|i| { let fields_iter = (0..field_count).map(|i| {
let field_op = ecx.operand_field(down, i).unwrap(); let field_op = ecx.operand_field(&down, i).unwrap();
let val = op_to_const(&ecx, field_op); let val = op_to_const(&ecx, &field_op);
ty::Const::from_value(tcx, val, field_op.layout.ty) ty::Const::from_value(tcx, val, field_op.layout.ty)
}); });
let fields = tcx.arena.alloc_from_iter(fields_iter); let fields = tcx.arena.alloc_from_iter(fields_iter);
@ -81,7 +81,7 @@ pub(crate) fn deref_const<'tcx>(
trace!("deref_const: {:?}", val); trace!("deref_const: {:?}", val);
let ecx = mk_eval_cx(tcx, DUMMY_SP, param_env, false); let ecx = mk_eval_cx(tcx, DUMMY_SP, param_env, false);
let op = ecx.const_to_op(val, None).unwrap(); let op = ecx.const_to_op(val, None).unwrap();
let mplace = ecx.deref_operand(op).unwrap(); let mplace = ecx.deref_operand(&op).unwrap();
if let Scalar::Ptr(ptr) = mplace.ptr { if let Scalar::Ptr(ptr) = mplace.ptr {
assert_eq!( assert_eq!(
ecx.memory.get_raw(ptr.alloc_id).unwrap().mutability, ecx.memory.get_raw(ptr.alloc_id).unwrap().mutability,
@ -106,5 +106,5 @@ pub(crate) fn deref_const<'tcx>(
}, },
}; };
tcx.mk_const(ty::Const { val: ty::ConstKind::Value(op_to_const(&ecx, mplace.into())), ty }) tcx.mk_const(ty::Const { val: ty::ConstKind::Value(op_to_const(&ecx, &mplace.into())), ty })
} }

View file

@ -17,10 +17,10 @@ use super::{
impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> { impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn cast( pub fn cast(
&mut self, &mut self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
cast_kind: CastKind, cast_kind: CastKind,
cast_ty: Ty<'tcx>, cast_ty: Ty<'tcx>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
use rustc_middle::mir::CastKind::*; use rustc_middle::mir::CastKind::*;
// FIXME: In which cases should we trigger UB when the source is uninit? // FIXME: In which cases should we trigger UB when the source is uninit?
@ -32,7 +32,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
Misc => { Misc => {
let src = self.read_immediate(src)?; let src = self.read_immediate(src)?;
let res = self.misc_cast(src, cast_ty)?; let res = self.misc_cast(&src, cast_ty)?;
self.write_immediate(res, dest)?; self.write_immediate(res, dest)?;
} }
@ -107,7 +107,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
fn misc_cast( fn misc_cast(
&self, &self,
src: ImmTy<'tcx, M::PointerTag>, src: &ImmTy<'tcx, M::PointerTag>,
cast_ty: Ty<'tcx>, cast_ty: Ty<'tcx>,
) -> InterpResult<'tcx, Immediate<M::PointerTag>> { ) -> InterpResult<'tcx, Immediate<M::PointerTag>> {
use rustc_middle::ty::TyKind::*; use rustc_middle::ty::TyKind::*;
@ -158,13 +158,13 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let dest_layout = self.layout_of(cast_ty)?; let dest_layout = self.layout_of(cast_ty)?;
if dest_layout.size == src.layout.size { if dest_layout.size == src.layout.size {
// Thin or fat pointer that just hast the ptr kind of target type changed. // Thin or fat pointer that just hast the ptr kind of target type changed.
return Ok(*src); return Ok(**src);
} else { } else {
// Casting the metadata away from a fat ptr. // Casting the metadata away from a fat ptr.
assert_eq!(src.layout.size, 2 * self.memory.pointer_size()); assert_eq!(src.layout.size, 2 * self.memory.pointer_size());
assert_eq!(dest_layout.size, self.memory.pointer_size()); assert_eq!(dest_layout.size, self.memory.pointer_size());
assert!(src.layout.ty.is_unsafe_ptr()); assert!(src.layout.ty.is_unsafe_ptr());
return match *src { return match **src {
Immediate::ScalarPair(data, _) => Ok(data.into()), Immediate::ScalarPair(data, _) => Ok(data.into()),
Immediate::Scalar(..) => span_bug!( Immediate::Scalar(..) => span_bug!(
self.cur_span(), self.cur_span(),
@ -259,8 +259,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
fn unsize_into_ptr( fn unsize_into_ptr(
&mut self, &mut self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
// The pointee types // The pointee types
source_ty: Ty<'tcx>, source_ty: Ty<'tcx>,
cast_ty: Ty<'tcx>, cast_ty: Ty<'tcx>,
@ -300,9 +300,9 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
fn unsize_into( fn unsize_into(
&mut self, &mut self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
cast_ty: TyAndLayout<'tcx>, cast_ty: TyAndLayout<'tcx>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
trace!("Unsizing {:?} of type {} into {:?}", *src, src.layout.ty, cast_ty.ty); trace!("Unsizing {:?} of type {} into {:?}", *src, src.layout.ty, cast_ty.ty);
match (&src.layout.ty.kind(), &cast_ty.ty.kind()) { match (&src.layout.ty.kind(), &cast_ty.ty.kind()) {
@ -340,9 +340,9 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let src_field = self.operand_field(src, i)?; let src_field = self.operand_field(src, i)?;
let dst_field = self.place_field(dest, i)?; let dst_field = self.place_field(dest, i)?;
if src_field.layout.ty == cast_ty_field.ty { if src_field.layout.ty == cast_ty_field.ty {
self.copy_op(src_field, dst_field)?; self.copy_op(&src_field, &dst_field)?;
} else { } else {
self.unsize_into(src_field, cast_ty_field, dst_field)?; self.unsize_into(&src_field, cast_ty_field, &dst_field)?;
} }
} }
Ok(()) Ok(())

View file

@ -548,8 +548,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
/// This can fail to provide an answer for extern types. /// This can fail to provide an answer for extern types.
pub(super) fn size_and_align_of( pub(super) fn size_and_align_of(
&self, &self,
metadata: MemPlaceMeta<M::PointerTag>, metadata: &MemPlaceMeta<M::PointerTag>,
layout: TyAndLayout<'tcx>, layout: &TyAndLayout<'tcx>,
) -> InterpResult<'tcx, Option<(Size, Align)>> { ) -> InterpResult<'tcx, Option<(Size, Align)>> {
if !layout.is_unsized() { if !layout.is_unsized() {
return Ok(Some((layout.size, layout.align.abi))); return Ok(Some((layout.size, layout.align.abi)));
@ -577,7 +577,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
// the last field). Can't have foreign types here, how would we // the last field). Can't have foreign types here, how would we
// adjust alignment and size for them? // adjust alignment and size for them?
let field = layout.field(self, layout.fields.count() - 1)?; let field = layout.field(self, layout.fields.count() - 1)?;
let (unsized_size, unsized_align) = match self.size_and_align_of(metadata, field)? { let (unsized_size, unsized_align) =
match self.size_and_align_of(metadata, &field)? {
Some(size_and_align) => size_and_align, Some(size_and_align) => size_and_align,
None => { None => {
// A field with extern type. If this field is at offset 0, we behave // A field with extern type. If this field is at offset 0, we behave
@ -645,16 +646,16 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
#[inline] #[inline]
pub fn size_and_align_of_mplace( pub fn size_and_align_of_mplace(
&self, &self,
mplace: MPlaceTy<'tcx, M::PointerTag>, mplace: &MPlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, Option<(Size, Align)>> { ) -> InterpResult<'tcx, Option<(Size, Align)>> {
self.size_and_align_of(mplace.meta, mplace.layout) self.size_and_align_of(&mplace.meta, &mplace.layout)
} }
pub fn push_stack_frame( pub fn push_stack_frame(
&mut self, &mut self,
instance: ty::Instance<'tcx>, instance: ty::Instance<'tcx>,
body: &'mir mir::Body<'tcx>, body: &'mir mir::Body<'tcx>,
return_place: Option<PlaceTy<'tcx, M::PointerTag>>, return_place: Option<&PlaceTy<'tcx, M::PointerTag>>,
return_to_block: StackPopCleanup, return_to_block: StackPopCleanup,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// first push a stack frame so we have access to the local substs // first push a stack frame so we have access to the local substs
@ -662,7 +663,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
body, body,
loc: Err(body.span), // Span used for errors caused during preamble. loc: Err(body.span), // Span used for errors caused during preamble.
return_to_block, return_to_block,
return_place, return_place: return_place.copied(),
// empty local array, we fill it in below, after we are inside the stack frame and // empty local array, we fill it in below, after we are inside the stack frame and
// all methods actually know about the frame // all methods actually know about the frame
locals: IndexVec::new(), locals: IndexVec::new(),
@ -777,10 +778,10 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
if !unwinding { if !unwinding {
// Copy the return value to the caller's stack frame. // Copy the return value to the caller's stack frame.
if let Some(return_place) = frame.return_place { if let Some(ref return_place) = frame.return_place {
let op = self.access_local(&frame, mir::RETURN_PLACE, None)?; let op = self.access_local(&frame, mir::RETURN_PLACE, None)?;
self.copy_op_transmute(op, return_place)?; self.copy_op_transmute(&op, return_place)?;
trace!("{:?}", self.dump_place(*return_place)); trace!("{:?}", self.dump_place(**return_place));
} else { } else {
throw_ub!(Unreachable); throw_ub!(Unreachable);
} }

View file

@ -167,7 +167,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: CompileTimeMachine<'mir, 'tcx, const_eval::Memory
fn visit_aggregate( fn visit_aggregate(
&mut self, &mut self,
mplace: MPlaceTy<'tcx>, mplace: &MPlaceTy<'tcx>,
fields: impl Iterator<Item = InterpResult<'tcx, Self::V>>, fields: impl Iterator<Item = InterpResult<'tcx, Self::V>>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// ZSTs cannot contain pointers, so we can skip them. // ZSTs cannot contain pointers, so we can skip them.
@ -191,14 +191,14 @@ impl<'rt, 'mir, 'tcx: 'mir, M: CompileTimeMachine<'mir, 'tcx, const_eval::Memory
self.walk_aggregate(mplace, fields) self.walk_aggregate(mplace, fields)
} }
fn visit_value(&mut self, mplace: MPlaceTy<'tcx>) -> InterpResult<'tcx> { fn visit_value(&mut self, mplace: &MPlaceTy<'tcx>) -> InterpResult<'tcx> {
// Handle Reference types, as these are the only relocations supported by const eval. // Handle Reference types, as these are the only relocations supported by const eval.
// Raw pointers (and boxes) are handled by the `leftover_relocations` logic. // Raw pointers (and boxes) are handled by the `leftover_relocations` logic.
let tcx = self.ecx.tcx; let tcx = self.ecx.tcx;
let ty = mplace.layout.ty; let ty = mplace.layout.ty;
if let ty::Ref(_, referenced_ty, ref_mutability) = *ty.kind() { if let ty::Ref(_, referenced_ty, ref_mutability) = *ty.kind() {
let value = self.ecx.read_immediate(mplace.into())?; let value = self.ecx.read_immediate(&(*mplace).into())?;
let mplace = self.ecx.ref_to_mplace(value)?; let mplace = self.ecx.ref_to_mplace(&value)?;
assert_eq!(mplace.layout.ty, referenced_ty); assert_eq!(mplace.layout.ty, referenced_ty);
// Handle trait object vtables. // Handle trait object vtables.
if let ty::Dynamic(..) = if let ty::Dynamic(..) =
@ -296,7 +296,7 @@ pub enum InternKind {
pub fn intern_const_alloc_recursive<M: CompileTimeMachine<'mir, 'tcx, const_eval::MemoryKind>>( pub fn intern_const_alloc_recursive<M: CompileTimeMachine<'mir, 'tcx, const_eval::MemoryKind>>(
ecx: &mut InterpCx<'mir, 'tcx, M>, ecx: &mut InterpCx<'mir, 'tcx, M>,
intern_kind: InternKind, intern_kind: InternKind,
ret: MPlaceTy<'tcx>, ret: &MPlaceTy<'tcx>,
) -> Result<(), ErrorReported> ) -> Result<(), ErrorReported>
where where
'tcx: 'mir, 'tcx: 'mir,
@ -328,7 +328,7 @@ where
Some(ret.layout.ty), Some(ret.layout.ty),
); );
ref_tracking.track((ret, base_intern_mode), || ()); ref_tracking.track((*ret, base_intern_mode), || ());
while let Some(((mplace, mode), _)) = ref_tracking.todo.pop() { while let Some(((mplace, mode), _)) = ref_tracking.todo.pop() {
let res = InternVisitor { let res = InternVisitor {
@ -338,7 +338,7 @@ where
leftover_allocations, leftover_allocations,
inside_unsafe_cell: false, inside_unsafe_cell: false,
} }
.visit_value(mplace); .visit_value(&mplace);
// We deliberately *ignore* interpreter errors here. When there is a problem, the remaining // We deliberately *ignore* interpreter errors here. When there is a problem, the remaining
// references are "leftover"-interned, and later validation will show a proper error // references are "leftover"-interned, and later validation will show a proper error
// and point at the right part of the value causing the problem. // and point at the right part of the value causing the problem.
@ -435,11 +435,11 @@ impl<'mir, 'tcx: 'mir, M: super::intern::CompileTimeMachine<'mir, 'tcx, !>>
layout: TyAndLayout<'tcx>, layout: TyAndLayout<'tcx>,
f: impl FnOnce( f: impl FnOnce(
&mut InterpCx<'mir, 'tcx, M>, &mut InterpCx<'mir, 'tcx, M>,
MPlaceTy<'tcx, M::PointerTag>, &MPlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, ()>, ) -> InterpResult<'tcx, ()>,
) -> InterpResult<'tcx, &'tcx Allocation> { ) -> InterpResult<'tcx, &'tcx Allocation> {
let dest = self.allocate(layout, MemoryKind::Stack); let dest = self.allocate(layout, MemoryKind::Stack);
f(self, dest)?; f(self, &dest)?;
let ptr = dest.ptr.assert_ptr(); let ptr = dest.ptr.assert_ptr();
assert_eq!(ptr.offset, Size::ZERO); assert_eq!(ptr.offset, Size::ZERO);
let mut alloc = self.memory.alloc_map.remove(&ptr.alloc_id).unwrap().1; let mut alloc = self.memory.alloc_map.remove(&ptr.alloc_id).unwrap().1;

View file

@ -115,7 +115,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
&mut self, &mut self,
instance: ty::Instance<'tcx>, instance: ty::Instance<'tcx>,
args: &[OpTy<'tcx, M::PointerTag>], args: &[OpTy<'tcx, M::PointerTag>],
ret: Option<(PlaceTy<'tcx, M::PointerTag>, mir::BasicBlock)>, ret: Option<(&PlaceTy<'tcx, M::PointerTag>, mir::BasicBlock)>,
) -> InterpResult<'tcx, bool> { ) -> InterpResult<'tcx, bool> {
let substs = instance.substs; let substs = instance.substs;
let intrinsic_name = self.tcx.item_name(instance.def_id()); let intrinsic_name = self.tcx.item_name(instance.def_id());
@ -143,9 +143,9 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
sym::min_align_of_val | sym::size_of_val => { sym::min_align_of_val | sym::size_of_val => {
// Avoid `deref_operand` -- this is not a deref, the ptr does not have to be // Avoid `deref_operand` -- this is not a deref, the ptr does not have to be
// dereferencable! // dereferencable!
let place = self.ref_to_mplace(self.read_immediate(args[0])?)?; let place = self.ref_to_mplace(&self.read_immediate(&args[0])?)?;
let (size, align) = self let (size, align) = self
.size_and_align_of_mplace(place)? .size_and_align_of_mplace(&place)?
.ok_or_else(|| err_unsup_format!("`extern type` does not have known layout"))?; .ok_or_else(|| err_unsup_format!("`extern type` does not have known layout"))?;
let result = match intrinsic_name { let result = match intrinsic_name {
@ -177,7 +177,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
self.tcx.const_eval_global_id(self.param_env, gid, Some(self.tcx.span))?; self.tcx.const_eval_global_id(self.param_env, gid, Some(self.tcx.span))?;
let const_ = ty::Const { val: ty::ConstKind::Value(val), ty }; let const_ = ty::Const { val: ty::ConstKind::Value(val), ty };
let val = self.const_to_op(&const_, None)?; let val = self.const_to_op(&const_, None)?;
self.copy_op(val, dest)?; self.copy_op(&val, dest)?;
} }
sym::ctpop sym::ctpop
@ -189,7 +189,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
| sym::bitreverse => { | sym::bitreverse => {
let ty = substs.type_at(0); let ty = substs.type_at(0);
let layout_of = self.layout_of(ty)?; let layout_of = self.layout_of(ty)?;
let val = self.read_scalar(args[0])?.check_init()?; let val = self.read_scalar(&args[0])?.check_init()?;
let bits = self.force_bits(val, layout_of.size)?; let bits = self.force_bits(val, layout_of.size)?;
let kind = match layout_of.abi { let kind = match layout_of.abi {
Abi::Scalar(ref scalar) => scalar.value, Abi::Scalar(ref scalar) => scalar.value,
@ -212,22 +212,25 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
self.write_scalar(out_val, dest)?; self.write_scalar(out_val, dest)?;
} }
sym::add_with_overflow | sym::sub_with_overflow | sym::mul_with_overflow => { sym::add_with_overflow | sym::sub_with_overflow | sym::mul_with_overflow => {
let lhs = self.read_immediate(args[0])?; let lhs = self.read_immediate(&args[0])?;
let rhs = self.read_immediate(args[1])?; let rhs = self.read_immediate(&args[1])?;
let bin_op = match intrinsic_name { let bin_op = match intrinsic_name {
sym::add_with_overflow => BinOp::Add, sym::add_with_overflow => BinOp::Add,
sym::sub_with_overflow => BinOp::Sub, sym::sub_with_overflow => BinOp::Sub,
sym::mul_with_overflow => BinOp::Mul, sym::mul_with_overflow => BinOp::Mul,
_ => bug!("Already checked for int ops"), _ => bug!("Already checked for int ops"),
}; };
self.binop_with_overflow(bin_op, lhs, rhs, dest)?; self.binop_with_overflow(bin_op, &lhs, &rhs, dest)?;
} }
sym::saturating_add | sym::saturating_sub => { sym::saturating_add | sym::saturating_sub => {
let l = self.read_immediate(args[0])?; let l = self.read_immediate(&args[0])?;
let r = self.read_immediate(args[1])?; let r = self.read_immediate(&args[1])?;
let is_add = intrinsic_name == sym::saturating_add; let is_add = intrinsic_name == sym::saturating_add;
let (val, overflowed, _ty) = let (val, overflowed, _ty) = self.overflowing_binary_op(
self.overflowing_binary_op(if is_add { BinOp::Add } else { BinOp::Sub }, l, r)?; if is_add { BinOp::Add } else { BinOp::Sub },
&l,
&r,
)?;
let val = if overflowed { let val = if overflowed {
let num_bits = l.layout.size.bits(); let num_bits = l.layout.size.bits();
if l.layout.abi.is_signed() { if l.layout.abi.is_signed() {
@ -269,8 +272,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
self.write_scalar(val, dest)?; self.write_scalar(val, dest)?;
} }
sym::discriminant_value => { sym::discriminant_value => {
let place = self.deref_operand(args[0])?; let place = self.deref_operand(&args[0])?;
let discr_val = self.read_discriminant(place.into())?.0; let discr_val = self.read_discriminant(&place.into())?.0;
self.write_scalar(discr_val, dest)?; self.write_scalar(discr_val, dest)?;
} }
sym::unchecked_shl sym::unchecked_shl
@ -280,8 +283,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
| sym::unchecked_mul | sym::unchecked_mul
| sym::unchecked_div | sym::unchecked_div
| sym::unchecked_rem => { | sym::unchecked_rem => {
let l = self.read_immediate(args[0])?; let l = self.read_immediate(&args[0])?;
let r = self.read_immediate(args[1])?; let r = self.read_immediate(&args[1])?;
let bin_op = match intrinsic_name { let bin_op = match intrinsic_name {
sym::unchecked_shl => BinOp::Shl, sym::unchecked_shl => BinOp::Shl,
sym::unchecked_shr => BinOp::Shr, sym::unchecked_shr => BinOp::Shr,
@ -292,7 +295,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
sym::unchecked_rem => BinOp::Rem, sym::unchecked_rem => BinOp::Rem,
_ => bug!("Already checked for int ops"), _ => bug!("Already checked for int ops"),
}; };
let (val, overflowed, _ty) = self.overflowing_binary_op(bin_op, l, r)?; let (val, overflowed, _ty) = self.overflowing_binary_op(bin_op, &l, &r)?;
if overflowed { if overflowed {
let layout = self.layout_of(substs.type_at(0))?; let layout = self.layout_of(substs.type_at(0))?;
let r_val = self.force_bits(r.to_scalar()?, layout.size)?; let r_val = self.force_bits(r.to_scalar()?, layout.size)?;
@ -308,9 +311,9 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
// rotate_left: (X << (S % BW)) | (X >> ((BW - S) % BW)) // rotate_left: (X << (S % BW)) | (X >> ((BW - S) % BW))
// rotate_right: (X << ((BW - S) % BW)) | (X >> (S % BW)) // rotate_right: (X << ((BW - S) % BW)) | (X >> (S % BW))
let layout = self.layout_of(substs.type_at(0))?; let layout = self.layout_of(substs.type_at(0))?;
let val = self.read_scalar(args[0])?.check_init()?; let val = self.read_scalar(&args[0])?.check_init()?;
let val_bits = self.force_bits(val, layout.size)?; let val_bits = self.force_bits(val, layout.size)?;
let raw_shift = self.read_scalar(args[1])?.check_init()?; let raw_shift = self.read_scalar(&args[1])?.check_init()?;
let raw_shift_bits = self.force_bits(raw_shift, layout.size)?; let raw_shift_bits = self.force_bits(raw_shift, layout.size)?;
let width_bits = u128::from(layout.size.bits()); let width_bits = u128::from(layout.size.bits());
let shift_bits = raw_shift_bits % width_bits; let shift_bits = raw_shift_bits % width_bits;
@ -327,15 +330,15 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
sym::copy | sym::copy_nonoverlapping => { sym::copy | sym::copy_nonoverlapping => {
let elem_ty = instance.substs.type_at(0); let elem_ty = instance.substs.type_at(0);
let elem_layout = self.layout_of(elem_ty)?; let elem_layout = self.layout_of(elem_ty)?;
let count = self.read_scalar(args[2])?.to_machine_usize(self)?; let count = self.read_scalar(&args[2])?.to_machine_usize(self)?;
let elem_align = elem_layout.align.abi; let elem_align = elem_layout.align.abi;
let size = elem_layout.size.checked_mul(count, self).ok_or_else(|| { let size = elem_layout.size.checked_mul(count, self).ok_or_else(|| {
err_ub_format!("overflow computing total size of `{}`", intrinsic_name) err_ub_format!("overflow computing total size of `{}`", intrinsic_name)
})?; })?;
let src = self.read_scalar(args[0])?.check_init()?; let src = self.read_scalar(&args[0])?.check_init()?;
let src = self.memory.check_ptr_access(src, size, elem_align)?; let src = self.memory.check_ptr_access(src, size, elem_align)?;
let dest = self.read_scalar(args[1])?.check_init()?; let dest = self.read_scalar(&args[1])?.check_init()?;
let dest = self.memory.check_ptr_access(dest, size, elem_align)?; let dest = self.memory.check_ptr_access(dest, size, elem_align)?;
if let (Some(src), Some(dest)) = (src, dest) { if let (Some(src), Some(dest)) = (src, dest) {
@ -348,16 +351,16 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
} }
} }
sym::offset => { sym::offset => {
let ptr = self.read_scalar(args[0])?.check_init()?; let ptr = self.read_scalar(&args[0])?.check_init()?;
let offset_count = self.read_scalar(args[1])?.to_machine_isize(self)?; let offset_count = self.read_scalar(&args[1])?.to_machine_isize(self)?;
let pointee_ty = substs.type_at(0); let pointee_ty = substs.type_at(0);
let offset_ptr = self.ptr_offset_inbounds(ptr, pointee_ty, offset_count)?; let offset_ptr = self.ptr_offset_inbounds(ptr, pointee_ty, offset_count)?;
self.write_scalar(offset_ptr, dest)?; self.write_scalar(offset_ptr, dest)?;
} }
sym::arith_offset => { sym::arith_offset => {
let ptr = self.read_scalar(args[0])?.check_init()?; let ptr = self.read_scalar(&args[0])?.check_init()?;
let offset_count = self.read_scalar(args[1])?.to_machine_isize(self)?; let offset_count = self.read_scalar(&args[1])?.to_machine_isize(self)?;
let pointee_ty = substs.type_at(0); let pointee_ty = substs.type_at(0);
let pointee_size = i64::try_from(self.layout_of(pointee_ty)?.size.bytes()).unwrap(); let pointee_size = i64::try_from(self.layout_of(pointee_ty)?.size.bytes()).unwrap();
@ -366,8 +369,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
self.write_scalar(offset_ptr, dest)?; self.write_scalar(offset_ptr, dest)?;
} }
sym::ptr_offset_from => { sym::ptr_offset_from => {
let a = self.read_immediate(args[0])?.to_scalar()?; let a = self.read_immediate(&args[0])?.to_scalar()?;
let b = self.read_immediate(args[1])?.to_scalar()?; let b = self.read_immediate(&args[1])?.to_scalar()?;
// Special case: if both scalars are *equal integers* // Special case: if both scalars are *equal integers*
// and not NULL, we pretend there is an allocation of size 0 right there, // and not NULL, we pretend there is an allocation of size 0 right there,
@ -406,16 +409,16 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let a_offset = ImmTy::from_uint(a.offset.bytes(), usize_layout); let a_offset = ImmTy::from_uint(a.offset.bytes(), usize_layout);
let b_offset = ImmTy::from_uint(b.offset.bytes(), usize_layout); let b_offset = ImmTy::from_uint(b.offset.bytes(), usize_layout);
let (val, _overflowed, _ty) = let (val, _overflowed, _ty) =
self.overflowing_binary_op(BinOp::Sub, a_offset, b_offset)?; self.overflowing_binary_op(BinOp::Sub, &a_offset, &b_offset)?;
let pointee_layout = self.layout_of(substs.type_at(0))?; let pointee_layout = self.layout_of(substs.type_at(0))?;
let val = ImmTy::from_scalar(val, isize_layout); let val = ImmTy::from_scalar(val, isize_layout);
let size = ImmTy::from_int(pointee_layout.size.bytes(), isize_layout); let size = ImmTy::from_int(pointee_layout.size.bytes(), isize_layout);
self.exact_div(val, size, dest)?; self.exact_div(&val, &size, dest)?;
} }
} }
sym::transmute => { sym::transmute => {
self.copy_op_transmute(args[0], dest)?; self.copy_op_transmute(&args[0], dest)?;
} }
sym::assert_inhabited => { sym::assert_inhabited => {
let ty = instance.substs.type_at(0); let ty = instance.substs.type_at(0);
@ -434,9 +437,9 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
} }
} }
sym::simd_insert => { sym::simd_insert => {
let index = u64::from(self.read_scalar(args[1])?.to_u32()?); let index = u64::from(self.read_scalar(&args[1])?.to_u32()?);
let elem = args[2]; let elem = &args[2];
let input = args[0]; let input = &args[0];
let (len, e_ty) = input.layout.ty.simd_size_and_type(*self.tcx); let (len, e_ty) = input.layout.ty.simd_size_and_type(*self.tcx);
assert!( assert!(
index < len, index < len,
@ -458,12 +461,12 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
for i in 0..len { for i in 0..len {
let place = self.place_index(dest, i)?; let place = self.place_index(dest, i)?;
let value = if i == index { elem } else { self.operand_index(input, i)? }; let value = if i == index { *elem } else { self.operand_index(input, i)? };
self.copy_op(value, place)?; self.copy_op(&value, &place)?;
} }
} }
sym::simd_extract => { sym::simd_extract => {
let index = u64::from(self.read_scalar(args[1])?.to_u32()?); let index = u64::from(self.read_scalar(&args[1])?.to_u32()?);
let (len, e_ty) = args[0].layout.ty.simd_size_and_type(*self.tcx); let (len, e_ty) = args[0].layout.ty.simd_size_and_type(*self.tcx);
assert!( assert!(
index < len, index < len,
@ -477,14 +480,14 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
"Return type `{}` must match vector element type `{}`", "Return type `{}` must match vector element type `{}`",
dest.layout.ty, e_ty dest.layout.ty, e_ty
); );
self.copy_op(self.operand_index(args[0], index)?, dest)?; self.copy_op(&self.operand_index(&args[0], index)?, dest)?;
} }
sym::likely | sym::unlikely => { sym::likely | sym::unlikely => {
// These just return their argument // These just return their argument
self.copy_op(args[0], dest)?; self.copy_op(&args[0], dest)?;
} }
sym::assume => { sym::assume => {
let cond = self.read_scalar(args[0])?.check_init()?.to_bool()?; let cond = self.read_scalar(&args[0])?.check_init()?.to_bool()?;
if !cond { if !cond {
throw_ub_format!("`assume` intrinsic called with `false`"); throw_ub_format!("`assume` intrinsic called with `false`");
} }
@ -492,21 +495,21 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
_ => return Ok(false), _ => return Ok(false),
} }
trace!("{:?}", self.dump_place(*dest)); trace!("{:?}", self.dump_place(**dest));
self.go_to_block(ret); self.go_to_block(ret);
Ok(true) Ok(true)
} }
pub fn exact_div( pub fn exact_div(
&mut self, &mut self,
a: ImmTy<'tcx, M::PointerTag>, a: &ImmTy<'tcx, M::PointerTag>,
b: ImmTy<'tcx, M::PointerTag>, b: &ImmTy<'tcx, M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// Performs an exact division, resulting in undefined behavior where // Performs an exact division, resulting in undefined behavior where
// `x % y != 0` or `y == 0` or `x == T::MIN && y == -1`. // `x % y != 0` or `y == 0` or `x == T::MIN && y == -1`.
// First, check x % y != 0 (or if that computation overflows). // First, check x % y != 0 (or if that computation overflows).
let (res, overflow, _ty) = self.overflowing_binary_op(BinOp::Rem, a, b)?; let (res, overflow, _ty) = self.overflowing_binary_op(BinOp::Rem, &a, &b)?;
if overflow || res.assert_bits(a.layout.size) != 0 { if overflow || res.assert_bits(a.layout.size) != 0 {
// Then, check if `b` is -1, which is the "MIN / -1" case. // Then, check if `b` is -1, which is the "MIN / -1" case.
let minus1 = Scalar::from_int(-1, dest.layout.size); let minus1 = Scalar::from_int(-1, dest.layout.size);
@ -518,7 +521,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
} }
} }
// `Rem` says this is all right, so we can let `Div` do its job. // `Rem` says this is all right, so we can let `Div` do its job.
self.binop_ignore_overflow(BinOp::Div, a, b, dest) self.binop_ignore_overflow(BinOp::Div, &a, &b, dest)
} }
/// Offsets a pointer by some multiple of its type, returning an error if the pointer leaves its /// Offsets a pointer by some multiple of its type, returning an error if the pointer leaves its

View file

@ -92,11 +92,11 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let location = self.allocate(loc_layout, MemoryKind::CallerLocation); let location = self.allocate(loc_layout, MemoryKind::CallerLocation);
// Initialize fields. // Initialize fields.
self.write_immediate(file.to_ref(), self.mplace_field(location, 0).unwrap().into()) self.write_immediate(file.to_ref(), &self.mplace_field(&location, 0).unwrap().into())
.expect("writing to memory we just allocated cannot fail"); .expect("writing to memory we just allocated cannot fail");
self.write_scalar(line, self.mplace_field(location, 1).unwrap().into()) self.write_scalar(line, &self.mplace_field(&location, 1).unwrap().into())
.expect("writing to memory we just allocated cannot fail"); .expect("writing to memory we just allocated cannot fail");
self.write_scalar(col, self.mplace_field(location, 2).unwrap().into()) self.write_scalar(col, &self.mplace_field(&location, 2).unwrap().into())
.expect("writing to memory we just allocated cannot fail"); .expect("writing to memory we just allocated cannot fail");
location location

View file

@ -157,7 +157,7 @@ pub trait Machine<'mir, 'tcx>: Sized {
instance: ty::Instance<'tcx>, instance: ty::Instance<'tcx>,
abi: Abi, abi: Abi,
args: &[OpTy<'tcx, Self::PointerTag>], args: &[OpTy<'tcx, Self::PointerTag>],
ret: Option<(PlaceTy<'tcx, Self::PointerTag>, mir::BasicBlock)>, ret: Option<(&PlaceTy<'tcx, Self::PointerTag>, mir::BasicBlock)>,
unwind: Option<mir::BasicBlock>, unwind: Option<mir::BasicBlock>,
) -> InterpResult<'tcx, Option<&'mir mir::Body<'tcx>>>; ) -> InterpResult<'tcx, Option<&'mir mir::Body<'tcx>>>;
@ -168,7 +168,7 @@ pub trait Machine<'mir, 'tcx>: Sized {
fn_val: Self::ExtraFnVal, fn_val: Self::ExtraFnVal,
abi: Abi, abi: Abi,
args: &[OpTy<'tcx, Self::PointerTag>], args: &[OpTy<'tcx, Self::PointerTag>],
ret: Option<(PlaceTy<'tcx, Self::PointerTag>, mir::BasicBlock)>, ret: Option<(&PlaceTy<'tcx, Self::PointerTag>, mir::BasicBlock)>,
unwind: Option<mir::BasicBlock>, unwind: Option<mir::BasicBlock>,
) -> InterpResult<'tcx>; ) -> InterpResult<'tcx>;
@ -178,7 +178,7 @@ pub trait Machine<'mir, 'tcx>: Sized {
ecx: &mut InterpCx<'mir, 'tcx, Self>, ecx: &mut InterpCx<'mir, 'tcx, Self>,
instance: ty::Instance<'tcx>, instance: ty::Instance<'tcx>,
args: &[OpTy<'tcx, Self::PointerTag>], args: &[OpTy<'tcx, Self::PointerTag>],
ret: Option<(PlaceTy<'tcx, Self::PointerTag>, mir::BasicBlock)>, ret: Option<(&PlaceTy<'tcx, Self::PointerTag>, mir::BasicBlock)>,
unwind: Option<mir::BasicBlock>, unwind: Option<mir::BasicBlock>,
) -> InterpResult<'tcx>; ) -> InterpResult<'tcx>;
@ -200,14 +200,14 @@ pub trait Machine<'mir, 'tcx>: Sized {
fn binary_ptr_op( fn binary_ptr_op(
ecx: &InterpCx<'mir, 'tcx, Self>, ecx: &InterpCx<'mir, 'tcx, Self>,
bin_op: mir::BinOp, bin_op: mir::BinOp,
left: ImmTy<'tcx, Self::PointerTag>, left: &ImmTy<'tcx, Self::PointerTag>,
right: ImmTy<'tcx, Self::PointerTag>, right: &ImmTy<'tcx, Self::PointerTag>,
) -> InterpResult<'tcx, (Scalar<Self::PointerTag>, bool, Ty<'tcx>)>; ) -> InterpResult<'tcx, (Scalar<Self::PointerTag>, bool, Ty<'tcx>)>;
/// Heap allocations via the `box` keyword. /// Heap allocations via the `box` keyword.
fn box_alloc( fn box_alloc(
ecx: &mut InterpCx<'mir, 'tcx, Self>, ecx: &mut InterpCx<'mir, 'tcx, Self>,
dest: PlaceTy<'tcx, Self::PointerTag>, dest: &PlaceTy<'tcx, Self::PointerTag>,
) -> InterpResult<'tcx>; ) -> InterpResult<'tcx>;
/// Called to read the specified `local` from the `frame`. /// Called to read the specified `local` from the `frame`.
@ -327,7 +327,7 @@ pub trait Machine<'mir, 'tcx>: Sized {
fn retag( fn retag(
_ecx: &mut InterpCx<'mir, 'tcx, Self>, _ecx: &mut InterpCx<'mir, 'tcx, Self>,
_kind: mir::RetagKind, _kind: mir::RetagKind,
_place: PlaceTy<'tcx, Self::PointerTag>, _place: &PlaceTy<'tcx, Self::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
Ok(()) Ok(())
} }
@ -420,7 +420,7 @@ pub macro compile_time_machine(<$mir: lifetime, $tcx: lifetime>) {
fn_val: !, fn_val: !,
_abi: Abi, _abi: Abi,
_args: &[OpTy<$tcx>], _args: &[OpTy<$tcx>],
_ret: Option<(PlaceTy<$tcx>, mir::BasicBlock)>, _ret: Option<(&PlaceTy<$tcx>, mir::BasicBlock)>,
_unwind: Option<mir::BasicBlock>, _unwind: Option<mir::BasicBlock>,
) -> InterpResult<$tcx> { ) -> InterpResult<$tcx> {
match fn_val {} match fn_val {}

View file

@ -32,6 +32,9 @@ pub enum Immediate<Tag = ()> {
ScalarPair(ScalarMaybeUninit<Tag>, ScalarMaybeUninit<Tag>), ScalarPair(ScalarMaybeUninit<Tag>, ScalarMaybeUninit<Tag>),
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(Immediate, 56);
impl<Tag> From<ScalarMaybeUninit<Tag>> for Immediate<Tag> { impl<Tag> From<ScalarMaybeUninit<Tag>> for Immediate<Tag> {
#[inline(always)] #[inline(always)]
fn from(val: ScalarMaybeUninit<Tag>) -> Self { fn from(val: ScalarMaybeUninit<Tag>) -> Self {
@ -92,6 +95,9 @@ pub struct ImmTy<'tcx, Tag = ()> {
pub layout: TyAndLayout<'tcx>, pub layout: TyAndLayout<'tcx>,
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(ImmTy<'_>, 72);
impl<Tag: Copy> std::fmt::Display for ImmTy<'tcx, Tag> { impl<Tag: Copy> std::fmt::Display for ImmTy<'tcx, Tag> {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
/// Helper function for printing a scalar to a FmtPrinter /// Helper function for printing a scalar to a FmtPrinter
@ -156,6 +162,9 @@ pub struct OpTy<'tcx, Tag = ()> {
pub layout: TyAndLayout<'tcx>, pub layout: TyAndLayout<'tcx>,
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(OpTy<'_, ()>, 80);
impl<'tcx, Tag> std::ops::Deref for OpTy<'tcx, Tag> { impl<'tcx, Tag> std::ops::Deref for OpTy<'tcx, Tag> {
type Target = Operand<Tag>; type Target = Operand<Tag>;
#[inline(always)] #[inline(always)]
@ -171,6 +180,13 @@ impl<'tcx, Tag: Copy> From<MPlaceTy<'tcx, Tag>> for OpTy<'tcx, Tag> {
} }
} }
impl<'tcx, Tag: Copy> From<&'_ MPlaceTy<'tcx, Tag>> for OpTy<'tcx, Tag> {
#[inline(always)]
fn from(mplace: &MPlaceTy<'tcx, Tag>) -> Self {
OpTy { op: Operand::Indirect(**mplace), layout: mplace.layout }
}
}
impl<'tcx, Tag> From<ImmTy<'tcx, Tag>> for OpTy<'tcx, Tag> { impl<'tcx, Tag> From<ImmTy<'tcx, Tag>> for OpTy<'tcx, Tag> {
#[inline(always)] #[inline(always)]
fn from(val: ImmTy<'tcx, Tag>) -> Self { fn from(val: ImmTy<'tcx, Tag>) -> Self {
@ -222,7 +238,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
#[inline] #[inline]
pub fn force_op_ptr( pub fn force_op_ptr(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
match op.try_as_mplace(self) { match op.try_as_mplace(self) {
Ok(mplace) => Ok(self.force_mplace_ptr(mplace)?.into()), Ok(mplace) => Ok(self.force_mplace_ptr(mplace)?.into()),
@ -234,7 +250,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
/// Returns `None` if the layout does not permit loading this as a value. /// Returns `None` if the layout does not permit loading this as a value.
fn try_read_immediate_from_mplace( fn try_read_immediate_from_mplace(
&self, &self,
mplace: MPlaceTy<'tcx, M::PointerTag>, mplace: &MPlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, Option<ImmTy<'tcx, M::PointerTag>>> { ) -> InterpResult<'tcx, Option<ImmTy<'tcx, M::PointerTag>>> {
if mplace.layout.is_unsized() { if mplace.layout.is_unsized() {
// Don't touch unsized // Don't touch unsized
@ -295,14 +311,14 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
/// in a `Immediate`, not on which data is stored there currently. /// in a `Immediate`, not on which data is stored there currently.
pub(crate) fn try_read_immediate( pub(crate) fn try_read_immediate(
&self, &self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, Result<ImmTy<'tcx, M::PointerTag>, MPlaceTy<'tcx, M::PointerTag>>> { ) -> InterpResult<'tcx, Result<ImmTy<'tcx, M::PointerTag>, MPlaceTy<'tcx, M::PointerTag>>> {
Ok(match src.try_as_mplace(self) { Ok(match src.try_as_mplace(self) {
Ok(mplace) => { Ok(ref mplace) => {
if let Some(val) = self.try_read_immediate_from_mplace(mplace)? { if let Some(val) = self.try_read_immediate_from_mplace(mplace)? {
Ok(val) Ok(val)
} else { } else {
Err(mplace) Err(*mplace)
} }
} }
Err(val) => Ok(val), Err(val) => Ok(val),
@ -313,7 +329,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
#[inline(always)] #[inline(always)]
pub fn read_immediate( pub fn read_immediate(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, ImmTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, ImmTy<'tcx, M::PointerTag>> {
if let Ok(imm) = self.try_read_immediate(op)? { if let Ok(imm) = self.try_read_immediate(op)? {
Ok(imm) Ok(imm)
@ -325,13 +341,13 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
/// Read a scalar from a place /// Read a scalar from a place
pub fn read_scalar( pub fn read_scalar(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, ScalarMaybeUninit<M::PointerTag>> { ) -> InterpResult<'tcx, ScalarMaybeUninit<M::PointerTag>> {
Ok(self.read_immediate(op)?.to_scalar_or_uninit()) Ok(self.read_immediate(op)?.to_scalar_or_uninit())
} }
// Turn the wide MPlace into a string (must already be dereferenced!) // Turn the wide MPlace into a string (must already be dereferenced!)
pub fn read_str(&self, mplace: MPlaceTy<'tcx, M::PointerTag>) -> InterpResult<'tcx, &str> { pub fn read_str(&self, mplace: &MPlaceTy<'tcx, M::PointerTag>) -> InterpResult<'tcx, &str> {
let len = mplace.len(self)?; let len = mplace.len(self)?;
let bytes = self.memory.read_bytes(mplace.ptr, Size::from_bytes(len))?; let bytes = self.memory.read_bytes(mplace.ptr, Size::from_bytes(len))?;
let str = std::str::from_utf8(bytes).map_err(|err| err_ub!(InvalidStr(err)))?; let str = std::str::from_utf8(bytes).map_err(|err| err_ub!(InvalidStr(err)))?;
@ -341,11 +357,11 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
/// Projection functions /// Projection functions
pub fn operand_field( pub fn operand_field(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
field: usize, field: usize,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
let base = match op.try_as_mplace(self) { let base = match op.try_as_mplace(self) {
Ok(mplace) => { Ok(ref mplace) => {
// We can reuse the mplace field computation logic for indirect operands. // We can reuse the mplace field computation logic for indirect operands.
let field = self.mplace_field(mplace, field)?; let field = self.mplace_field(mplace, field)?;
return Ok(field.into()); return Ok(field.into());
@ -379,7 +395,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn operand_index( pub fn operand_index(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
index: u64, index: u64,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
if let Ok(index) = usize::try_from(index) { if let Ok(index) = usize::try_from(index) {
@ -388,28 +404,28 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
} else { } else {
// Indexing into a big array. This must be an mplace. // Indexing into a big array. This must be an mplace.
let mplace = op.assert_mem_place(self); let mplace = op.assert_mem_place(self);
Ok(self.mplace_index(mplace, index)?.into()) Ok(self.mplace_index(&mplace, index)?.into())
} }
} }
pub fn operand_downcast( pub fn operand_downcast(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
variant: VariantIdx, variant: VariantIdx,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
// Downcasts only change the layout // Downcasts only change the layout
Ok(match op.try_as_mplace(self) { Ok(match op.try_as_mplace(self) {
Ok(mplace) => self.mplace_downcast(mplace, variant)?.into(), Ok(ref mplace) => self.mplace_downcast(mplace, variant)?.into(),
Err(..) => { Err(..) => {
let layout = op.layout.for_variant(self, variant); let layout = op.layout.for_variant(self, variant);
OpTy { layout, ..op } OpTy { layout, ..*op }
} }
}) })
} }
pub fn operand_projection( pub fn operand_projection(
&self, &self,
base: OpTy<'tcx, M::PointerTag>, base: &OpTy<'tcx, M::PointerTag>,
proj_elem: mir::PlaceElem<'tcx>, proj_elem: mir::PlaceElem<'tcx>,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
use rustc_middle::mir::ProjectionElem::*; use rustc_middle::mir::ProjectionElem::*;
@ -421,7 +437,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
// The rest should only occur as mplace, we do not use Immediates for types // The rest should only occur as mplace, we do not use Immediates for types
// allowing such operations. This matches place_projection forcing an allocation. // allowing such operations. This matches place_projection forcing an allocation.
let mplace = base.assert_mem_place(self); let mplace = base.assert_mem_place(self);
self.mplace_projection(mplace, proj_elem)?.into() self.mplace_projection(&mplace, proj_elem)?.into()
} }
}) })
} }
@ -453,9 +469,9 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
#[inline(always)] #[inline(always)]
pub fn place_to_op( pub fn place_to_op(
&self, &self,
place: PlaceTy<'tcx, M::PointerTag>, place: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
let op = match *place { let op = match **place {
Place::Ptr(mplace) => Operand::Indirect(mplace), Place::Ptr(mplace) => Operand::Indirect(mplace),
Place::Local { frame, local } => { Place::Local { frame, local } => {
*self.access_local(&self.stack()[frame], local, None)? *self.access_local(&self.stack()[frame], local, None)?
@ -480,7 +496,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let op = place let op = place
.projection .projection
.iter() .iter()
.try_fold(base_op, |op, elem| self.operand_projection(op, elem))?; .try_fold(base_op, |op, elem| self.operand_projection(&op, elem))?;
trace!("eval_place_to_op: got {:?}", *op); trace!("eval_place_to_op: got {:?}", *op);
// Sanity-check the type we ended up with. // Sanity-check the type we ended up with.
@ -590,7 +606,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
/// Read discriminant, return the runtime value as well as the variant index. /// Read discriminant, return the runtime value as well as the variant index.
pub fn read_discriminant( pub fn read_discriminant(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, (Scalar<M::PointerTag>, VariantIdx)> { ) -> InterpResult<'tcx, (Scalar<M::PointerTag>, VariantIdx)> {
trace!("read_discriminant_value {:#?}", op.layout); trace!("read_discriminant_value {:#?}", op.layout);
// Get type and layout of the discriminant. // Get type and layout of the discriminant.
@ -636,7 +652,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let tag_layout = self.layout_of(tag_scalar_layout.value.to_int_ty(*self.tcx))?; let tag_layout = self.layout_of(tag_scalar_layout.value.to_int_ty(*self.tcx))?;
// Read tag and sanity-check `tag_layout`. // Read tag and sanity-check `tag_layout`.
let tag_val = self.read_immediate(self.operand_field(op, tag_field)?)?; let tag_val = self.read_immediate(&self.operand_field(op, tag_field)?)?;
assert_eq!(tag_layout.size, tag_val.layout.size); assert_eq!(tag_layout.size, tag_val.layout.size);
assert_eq!(tag_layout.abi.is_signed(), tag_val.layout.abi.is_signed()); assert_eq!(tag_layout.abi.is_signed(), tag_val.layout.abi.is_signed());
let tag_val = tag_val.to_scalar()?; let tag_val = tag_val.to_scalar()?;
@ -690,7 +706,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let tag_val = ImmTy::from_uint(tag_bits, tag_layout); let tag_val = ImmTy::from_uint(tag_bits, tag_layout);
let niche_start_val = ImmTy::from_uint(niche_start, tag_layout); let niche_start_val = ImmTy::from_uint(niche_start, tag_layout);
let variant_index_relative_val = let variant_index_relative_val =
self.binary_op(mir::BinOp::Sub, tag_val, niche_start_val)?; self.binary_op(mir::BinOp::Sub, &tag_val, &niche_start_val)?;
let variant_index_relative = variant_index_relative_val let variant_index_relative = variant_index_relative_val
.to_scalar()? .to_scalar()?
.assert_bits(tag_val.layout.size); .assert_bits(tag_val.layout.size);

View file

@ -14,11 +14,11 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn binop_with_overflow( pub fn binop_with_overflow(
&mut self, &mut self,
op: mir::BinOp, op: mir::BinOp,
left: ImmTy<'tcx, M::PointerTag>, left: &ImmTy<'tcx, M::PointerTag>,
right: ImmTy<'tcx, M::PointerTag>, right: &ImmTy<'tcx, M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
let (val, overflowed, ty) = self.overflowing_binary_op(op, left, right)?; let (val, overflowed, ty) = self.overflowing_binary_op(op, &left, &right)?;
debug_assert_eq!( debug_assert_eq!(
self.tcx.intern_tup(&[ty, self.tcx.types.bool]), self.tcx.intern_tup(&[ty, self.tcx.types.bool]),
dest.layout.ty, dest.layout.ty,
@ -34,9 +34,9 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn binop_ignore_overflow( pub fn binop_ignore_overflow(
&mut self, &mut self,
op: mir::BinOp, op: mir::BinOp,
left: ImmTy<'tcx, M::PointerTag>, left: &ImmTy<'tcx, M::PointerTag>,
right: ImmTy<'tcx, M::PointerTag>, right: &ImmTy<'tcx, M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
let (val, _overflowed, ty) = self.overflowing_binary_op(op, left, right)?; let (val, _overflowed, ty) = self.overflowing_binary_op(op, left, right)?;
assert_eq!(ty, dest.layout.ty, "type mismatch for result of {:?}", op); assert_eq!(ty, dest.layout.ty, "type mismatch for result of {:?}", op);
@ -269,8 +269,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn overflowing_binary_op( pub fn overflowing_binary_op(
&self, &self,
bin_op: mir::BinOp, bin_op: mir::BinOp,
left: ImmTy<'tcx, M::PointerTag>, left: &ImmTy<'tcx, M::PointerTag>,
right: ImmTy<'tcx, M::PointerTag>, right: &ImmTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, (Scalar<M::PointerTag>, bool, Ty<'tcx>)> { ) -> InterpResult<'tcx, (Scalar<M::PointerTag>, bool, Ty<'tcx>)> {
trace!( trace!(
"Running binary op {:?}: {:?} ({:?}), {:?} ({:?})", "Running binary op {:?}: {:?} ({:?}), {:?} ({:?})",
@ -347,8 +347,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn binary_op( pub fn binary_op(
&self, &self,
bin_op: mir::BinOp, bin_op: mir::BinOp,
left: ImmTy<'tcx, M::PointerTag>, left: &ImmTy<'tcx, M::PointerTag>,
right: ImmTy<'tcx, M::PointerTag>, right: &ImmTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, ImmTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, ImmTy<'tcx, M::PointerTag>> {
let (val, _overflow, ty) = self.overflowing_binary_op(bin_op, left, right)?; let (val, _overflow, ty) = self.overflowing_binary_op(bin_op, left, right)?;
Ok(ImmTy::from_scalar(val, self.layout_of(ty)?)) Ok(ImmTy::from_scalar(val, self.layout_of(ty)?))
@ -359,7 +359,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn overflowing_unary_op( pub fn overflowing_unary_op(
&self, &self,
un_op: mir::UnOp, un_op: mir::UnOp,
val: ImmTy<'tcx, M::PointerTag>, val: &ImmTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, (Scalar<M::PointerTag>, bool, Ty<'tcx>)> { ) -> InterpResult<'tcx, (Scalar<M::PointerTag>, bool, Ty<'tcx>)> {
use rustc_middle::mir::UnOp::*; use rustc_middle::mir::UnOp::*;
@ -409,7 +409,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
pub fn unary_op( pub fn unary_op(
&self, &self,
un_op: mir::UnOp, un_op: mir::UnOp,
val: ImmTy<'tcx, M::PointerTag>, val: &ImmTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, ImmTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, ImmTy<'tcx, M::PointerTag>> {
let (val, _overflow, ty) = self.overflowing_unary_op(un_op, val)?; let (val, _overflow, ty) = self.overflowing_unary_op(un_op, val)?;
Ok(ImmTy::from_scalar(val, self.layout_of(ty)?)) Ok(ImmTy::from_scalar(val, self.layout_of(ty)?))

View file

@ -33,6 +33,9 @@ pub enum MemPlaceMeta<Tag = ()> {
Poison, Poison,
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(MemPlaceMeta, 24);
impl<Tag> MemPlaceMeta<Tag> { impl<Tag> MemPlaceMeta<Tag> {
pub fn unwrap_meta(self) -> Scalar<Tag> { pub fn unwrap_meta(self) -> Scalar<Tag> {
match self { match self {
@ -71,6 +74,9 @@ pub struct MemPlace<Tag = ()> {
pub meta: MemPlaceMeta<Tag>, pub meta: MemPlaceMeta<Tag>,
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(MemPlace, 56);
#[derive(Copy, Clone, Debug, Hash, PartialEq, Eq, HashStable)] #[derive(Copy, Clone, Debug, Hash, PartialEq, Eq, HashStable)]
pub enum Place<Tag = ()> { pub enum Place<Tag = ()> {
/// A place referring to a value allocated in the `Memory` system. /// A place referring to a value allocated in the `Memory` system.
@ -81,12 +87,18 @@ pub enum Place<Tag = ()> {
Local { frame: usize, local: mir::Local }, Local { frame: usize, local: mir::Local },
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(Place, 64);
#[derive(Copy, Clone, Debug)] #[derive(Copy, Clone, Debug)]
pub struct PlaceTy<'tcx, Tag = ()> { pub struct PlaceTy<'tcx, Tag = ()> {
place: Place<Tag>, // Keep this private; it helps enforce invariants. place: Place<Tag>, // Keep this private; it helps enforce invariants.
pub layout: TyAndLayout<'tcx>, pub layout: TyAndLayout<'tcx>,
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(PlaceTy<'_>, 80);
impl<'tcx, Tag> std::ops::Deref for PlaceTy<'tcx, Tag> { impl<'tcx, Tag> std::ops::Deref for PlaceTy<'tcx, Tag> {
type Target = Place<Tag>; type Target = Place<Tag>;
#[inline(always)] #[inline(always)]
@ -102,6 +114,9 @@ pub struct MPlaceTy<'tcx, Tag = ()> {
pub layout: TyAndLayout<'tcx>, pub layout: TyAndLayout<'tcx>,
} }
#[cfg(target_arch = "x86_64")]
rustc_data_structures::static_assert_size!(MPlaceTy<'_>, 72);
impl<'tcx, Tag> std::ops::Deref for MPlaceTy<'tcx, Tag> { impl<'tcx, Tag> std::ops::Deref for MPlaceTy<'tcx, Tag> {
type Target = MemPlace<Tag>; type Target = MemPlace<Tag>;
#[inline(always)] #[inline(always)]
@ -168,7 +183,7 @@ impl<Tag> MemPlace<Tag> {
} }
} }
impl<'tcx, Tag> MPlaceTy<'tcx, Tag> { impl<'tcx, Tag: Copy> MPlaceTy<'tcx, Tag> {
/// Produces a MemPlace that works for ZST but nothing else /// Produces a MemPlace that works for ZST but nothing else
#[inline] #[inline]
pub fn dangling(layout: TyAndLayout<'tcx>, cx: &impl HasDataLayout) -> Self { pub fn dangling(layout: TyAndLayout<'tcx>, cx: &impl HasDataLayout) -> Self {
@ -180,13 +195,13 @@ impl<'tcx, Tag> MPlaceTy<'tcx, Tag> {
/// Replace ptr tag, maintain vtable tag (if any) /// Replace ptr tag, maintain vtable tag (if any)
#[inline] #[inline]
pub fn replace_tag(self, new_tag: Tag) -> Self { pub fn replace_tag(&self, new_tag: Tag) -> Self {
MPlaceTy { mplace: self.mplace.replace_tag(new_tag), layout: self.layout } MPlaceTy { mplace: self.mplace.replace_tag(new_tag), layout: self.layout }
} }
#[inline] #[inline]
pub fn offset( pub fn offset(
self, &self,
offset: Size, offset: Size,
meta: MemPlaceMeta<Tag>, meta: MemPlaceMeta<Tag>,
layout: TyAndLayout<'tcx>, layout: TyAndLayout<'tcx>,
@ -201,7 +216,7 @@ impl<'tcx, Tag> MPlaceTy<'tcx, Tag> {
} }
#[inline] #[inline]
pub(super) fn len(self, cx: &impl HasDataLayout) -> InterpResult<'tcx, u64> { pub(super) fn len(&self, cx: &impl HasDataLayout) -> InterpResult<'tcx, u64> {
if self.layout.is_unsized() { if self.layout.is_unsized() {
// We need to consult `meta` metadata // We need to consult `meta` metadata
match self.layout.ty.kind() { match self.layout.ty.kind() {
@ -219,7 +234,7 @@ impl<'tcx, Tag> MPlaceTy<'tcx, Tag> {
} }
#[inline] #[inline]
pub(super) fn vtable(self) -> Scalar<Tag> { pub(super) fn vtable(&self) -> Scalar<Tag> {
match self.layout.ty.kind() { match self.layout.ty.kind() {
ty::Dynamic(..) => self.mplace.meta.unwrap_meta(), ty::Dynamic(..) => self.mplace.meta.unwrap_meta(),
_ => bug!("vtable not supported on type {:?}", self.layout.ty), _ => bug!("vtable not supported on type {:?}", self.layout.ty),
@ -233,10 +248,10 @@ impl<'tcx, Tag: Debug + Copy> OpTy<'tcx, Tag> {
/// Note: do not call `as_ref` on the resulting place. This function should only be used to /// Note: do not call `as_ref` on the resulting place. This function should only be used to
/// read from the resulting mplace, not to get its address back. /// read from the resulting mplace, not to get its address back.
pub fn try_as_mplace( pub fn try_as_mplace(
self, &self,
cx: &impl HasDataLayout, cx: &impl HasDataLayout,
) -> Result<MPlaceTy<'tcx, Tag>, ImmTy<'tcx, Tag>> { ) -> Result<MPlaceTy<'tcx, Tag>, ImmTy<'tcx, Tag>> {
match *self { match **self {
Operand::Indirect(mplace) => Ok(MPlaceTy { mplace, layout: self.layout }), Operand::Indirect(mplace) => Ok(MPlaceTy { mplace, layout: self.layout }),
Operand::Immediate(_) if self.layout.is_zst() => { Operand::Immediate(_) if self.layout.is_zst() => {
Ok(MPlaceTy::dangling(self.layout, cx)) Ok(MPlaceTy::dangling(self.layout, cx))
@ -248,7 +263,7 @@ impl<'tcx, Tag: Debug + Copy> OpTy<'tcx, Tag> {
#[inline(always)] #[inline(always)]
/// Note: do not call `as_ref` on the resulting place. This function should only be used to /// Note: do not call `as_ref` on the resulting place. This function should only be used to
/// read from the resulting mplace, not to get its address back. /// read from the resulting mplace, not to get its address back.
pub fn assert_mem_place(self, cx: &impl HasDataLayout) -> MPlaceTy<'tcx, Tag> { pub fn assert_mem_place(&self, cx: &impl HasDataLayout) -> MPlaceTy<'tcx, Tag> {
self.try_as_mplace(cx).unwrap() self.try_as_mplace(cx).unwrap()
} }
} }
@ -288,12 +303,12 @@ where
/// Generally prefer `deref_operand`. /// Generally prefer `deref_operand`.
pub fn ref_to_mplace( pub fn ref_to_mplace(
&self, &self,
val: ImmTy<'tcx, M::PointerTag>, val: &ImmTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
let pointee_type = let pointee_type =
val.layout.ty.builtin_deref(true).expect("`ref_to_mplace` called on non-ptr type").ty; val.layout.ty.builtin_deref(true).expect("`ref_to_mplace` called on non-ptr type").ty;
let layout = self.layout_of(pointee_type)?; let layout = self.layout_of(pointee_type)?;
let (ptr, meta) = match *val { let (ptr, meta) = match **val {
Immediate::Scalar(ptr) => (ptr.check_init()?, MemPlaceMeta::None), Immediate::Scalar(ptr) => (ptr.check_init()?, MemPlaceMeta::None),
Immediate::ScalarPair(ptr, meta) => { Immediate::ScalarPair(ptr, meta) => {
(ptr.check_init()?, MemPlaceMeta::Meta(meta.check_init()?)) (ptr.check_init()?, MemPlaceMeta::Meta(meta.check_init()?))
@ -316,11 +331,11 @@ where
/// will always be a MemPlace. Lives in `place.rs` because it creates a place. /// will always be a MemPlace. Lives in `place.rs` because it creates a place.
pub fn deref_operand( pub fn deref_operand(
&self, &self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
let val = self.read_immediate(src)?; let val = self.read_immediate(src)?;
trace!("deref to {} on {:?}", val.layout.ty, *val); trace!("deref to {} on {:?}", val.layout.ty, *val);
let place = self.ref_to_mplace(val)?; let place = self.ref_to_mplace(&val)?;
self.mplace_access_checked(place, None) self.mplace_access_checked(place, None)
} }
@ -333,7 +348,7 @@ where
#[inline] #[inline]
pub(super) fn check_mplace_access( pub(super) fn check_mplace_access(
&self, &self,
place: MPlaceTy<'tcx, M::PointerTag>, place: &MPlaceTy<'tcx, M::PointerTag>,
size: Option<Size>, size: Option<Size>,
) -> InterpResult<'tcx, Option<Pointer<M::PointerTag>>> { ) -> InterpResult<'tcx, Option<Pointer<M::PointerTag>>> {
let size = size.unwrap_or_else(|| { let size = size.unwrap_or_else(|| {
@ -355,13 +370,13 @@ where
force_align: Option<Align>, force_align: Option<Align>,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
let (size, align) = self let (size, align) = self
.size_and_align_of_mplace(place)? .size_and_align_of_mplace(&place)?
.unwrap_or((place.layout.size, place.layout.align.abi)); .unwrap_or((place.layout.size, place.layout.align.abi));
assert!(place.mplace.align <= align, "dynamic alignment less strict than static one?"); assert!(place.mplace.align <= align, "dynamic alignment less strict than static one?");
// Check (stricter) dynamic alignment, unless forced otherwise. // Check (stricter) dynamic alignment, unless forced otherwise.
place.mplace.align = force_align.unwrap_or(align); place.mplace.align = force_align.unwrap_or(align);
// When dereferencing a pointer, it must be non-NULL, aligned, and live. // When dereferencing a pointer, it must be non-NULL, aligned, and live.
if let Some(ptr) = self.check_mplace_access(place, Some(size))? { if let Some(ptr) = self.check_mplace_access(&place, Some(size))? {
place.mplace.ptr = ptr.into(); place.mplace.ptr = ptr.into();
} }
Ok(place) Ok(place)
@ -386,7 +401,7 @@ where
#[inline(always)] #[inline(always)]
pub fn mplace_field( pub fn mplace_field(
&self, &self,
base: MPlaceTy<'tcx, M::PointerTag>, base: &MPlaceTy<'tcx, M::PointerTag>,
field: usize, field: usize,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
let offset = base.layout.fields.offset(field); let offset = base.layout.fields.offset(field);
@ -397,7 +412,7 @@ where
// Re-use parent metadata to determine dynamic field layout. // Re-use parent metadata to determine dynamic field layout.
// With custom DSTS, this *will* execute user-defined code, but the same // With custom DSTS, this *will* execute user-defined code, but the same
// happens at run-time so that's okay. // happens at run-time so that's okay.
let align = match self.size_and_align_of(base.meta, field_layout)? { let align = match self.size_and_align_of(&base.meta, &field_layout)? {
Some((_, align)) => align, Some((_, align)) => align,
None if offset == Size::ZERO => { None if offset == Size::ZERO => {
// An extern type at offset 0, we fall back to its static alignment. // An extern type at offset 0, we fall back to its static alignment.
@ -427,7 +442,7 @@ where
#[inline(always)] #[inline(always)]
pub fn mplace_index( pub fn mplace_index(
&self, &self,
base: MPlaceTy<'tcx, M::PointerTag>, base: &MPlaceTy<'tcx, M::PointerTag>,
index: u64, index: u64,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
// Not using the layout method because we want to compute on u64 // Not using the layout method because we want to compute on u64
@ -457,8 +472,8 @@ where
// same by repeatedly calling `mplace_array`. // same by repeatedly calling `mplace_array`.
pub(super) fn mplace_array_fields( pub(super) fn mplace_array_fields(
&self, &self,
base: MPlaceTy<'tcx, Tag>, base: &'a MPlaceTy<'tcx, Tag>,
) -> InterpResult<'tcx, impl Iterator<Item = InterpResult<'tcx, MPlaceTy<'tcx, Tag>>> + 'tcx> ) -> InterpResult<'tcx, impl Iterator<Item = InterpResult<'tcx, MPlaceTy<'tcx, Tag>>> + 'a>
{ {
let len = base.len(self)?; // also asserts that we have a type where this makes sense let len = base.len(self)?; // also asserts that we have a type where this makes sense
let stride = match base.layout.fields { let stride = match base.layout.fields {
@ -473,7 +488,7 @@ where
fn mplace_subslice( fn mplace_subslice(
&self, &self,
base: MPlaceTy<'tcx, M::PointerTag>, base: &MPlaceTy<'tcx, M::PointerTag>,
from: u64, from: u64,
to: u64, to: u64,
from_end: bool, from_end: bool,
@ -518,30 +533,30 @@ where
pub(super) fn mplace_downcast( pub(super) fn mplace_downcast(
&self, &self,
base: MPlaceTy<'tcx, M::PointerTag>, base: &MPlaceTy<'tcx, M::PointerTag>,
variant: VariantIdx, variant: VariantIdx,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
// Downcasts only change the layout // Downcasts only change the layout
assert!(!base.meta.has_meta()); assert!(!base.meta.has_meta());
Ok(MPlaceTy { layout: base.layout.for_variant(self, variant), ..base }) Ok(MPlaceTy { layout: base.layout.for_variant(self, variant), ..*base })
} }
/// Project into an mplace /// Project into an mplace
pub(super) fn mplace_projection( pub(super) fn mplace_projection(
&self, &self,
base: MPlaceTy<'tcx, M::PointerTag>, base: &MPlaceTy<'tcx, M::PointerTag>,
proj_elem: mir::PlaceElem<'tcx>, proj_elem: mir::PlaceElem<'tcx>,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
use rustc_middle::mir::ProjectionElem::*; use rustc_middle::mir::ProjectionElem::*;
Ok(match proj_elem { Ok(match proj_elem {
Field(field, _) => self.mplace_field(base, field.index())?, Field(field, _) => self.mplace_field(base, field.index())?,
Downcast(_, variant) => self.mplace_downcast(base, variant)?, Downcast(_, variant) => self.mplace_downcast(base, variant)?,
Deref => self.deref_operand(base.into())?, Deref => self.deref_operand(&base.into())?,
Index(local) => { Index(local) => {
let layout = self.layout_of(self.tcx.types.usize)?; let layout = self.layout_of(self.tcx.types.usize)?;
let n = self.access_local(self.frame(), local, Some(layout))?; let n = self.access_local(self.frame(), local, Some(layout))?;
let n = self.read_scalar(n)?; let n = self.read_scalar(&n)?;
let n = u64::try_from( let n = u64::try_from(
self.force_bits(n.check_init()?, self.tcx.data_layout.pointer_size)?, self.force_bits(n.check_init()?, self.tcx.data_layout.pointer_size)?,
) )
@ -577,37 +592,37 @@ where
/// into the field of a local `ScalarPair`, we have to first allocate it. /// into the field of a local `ScalarPair`, we have to first allocate it.
pub fn place_field( pub fn place_field(
&mut self, &mut self,
base: PlaceTy<'tcx, M::PointerTag>, base: &PlaceTy<'tcx, M::PointerTag>,
field: usize, field: usize,
) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> {
// FIXME: We could try to be smarter and avoid allocation for fields that span the // FIXME: We could try to be smarter and avoid allocation for fields that span the
// entire place. // entire place.
let mplace = self.force_allocation(base)?; let mplace = self.force_allocation(base)?;
Ok(self.mplace_field(mplace, field)?.into()) Ok(self.mplace_field(&mplace, field)?.into())
} }
pub fn place_index( pub fn place_index(
&mut self, &mut self,
base: PlaceTy<'tcx, M::PointerTag>, base: &PlaceTy<'tcx, M::PointerTag>,
index: u64, index: u64,
) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> {
let mplace = self.force_allocation(base)?; let mplace = self.force_allocation(base)?;
Ok(self.mplace_index(mplace, index)?.into()) Ok(self.mplace_index(&mplace, index)?.into())
} }
pub fn place_downcast( pub fn place_downcast(
&self, &self,
base: PlaceTy<'tcx, M::PointerTag>, base: &PlaceTy<'tcx, M::PointerTag>,
variant: VariantIdx, variant: VariantIdx,
) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> {
// Downcast just changes the layout // Downcast just changes the layout
Ok(match base.place { Ok(match base.place {
Place::Ptr(mplace) => { Place::Ptr(mplace) => {
self.mplace_downcast(MPlaceTy { mplace, layout: base.layout }, variant)?.into() self.mplace_downcast(&MPlaceTy { mplace, layout: base.layout }, variant)?.into()
} }
Place::Local { .. } => { Place::Local { .. } => {
let layout = base.layout.for_variant(self, variant); let layout = base.layout.for_variant(self, variant);
PlaceTy { layout, ..base } PlaceTy { layout, ..*base }
} }
}) })
} }
@ -615,19 +630,19 @@ where
/// Projects into a place. /// Projects into a place.
pub fn place_projection( pub fn place_projection(
&mut self, &mut self,
base: PlaceTy<'tcx, M::PointerTag>, base: &PlaceTy<'tcx, M::PointerTag>,
&proj_elem: &mir::ProjectionElem<mir::Local, Ty<'tcx>>, &proj_elem: &mir::ProjectionElem<mir::Local, Ty<'tcx>>,
) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, PlaceTy<'tcx, M::PointerTag>> {
use rustc_middle::mir::ProjectionElem::*; use rustc_middle::mir::ProjectionElem::*;
Ok(match proj_elem { Ok(match proj_elem {
Field(field, _) => self.place_field(base, field.index())?, Field(field, _) => self.place_field(base, field.index())?,
Downcast(_, variant) => self.place_downcast(base, variant)?, Downcast(_, variant) => self.place_downcast(base, variant)?,
Deref => self.deref_operand(self.place_to_op(base)?)?.into(), Deref => self.deref_operand(&self.place_to_op(base)?)?.into(),
// For the other variants, we have to force an allocation. // For the other variants, we have to force an allocation.
// This matches `operand_projection`. // This matches `operand_projection`.
Subslice { .. } | ConstantIndex { .. } | Index(_) => { Subslice { .. } | ConstantIndex { .. } | Index(_) => {
let mplace = self.force_allocation(base)?; let mplace = self.force_allocation(base)?;
self.mplace_projection(mplace, proj_elem)?.into() self.mplace_projection(&mplace, proj_elem)?.into()
} }
}) })
} }
@ -645,7 +660,7 @@ where
}; };
for elem in place.projection.iter() { for elem in place.projection.iter() {
place_ty = self.place_projection(place_ty, &elem)? place_ty = self.place_projection(&place_ty, &elem)?
} }
trace!("{:?}", self.dump_place(place_ty.place)); trace!("{:?}", self.dump_place(place_ty.place));
@ -666,7 +681,7 @@ where
pub fn write_scalar( pub fn write_scalar(
&mut self, &mut self,
val: impl Into<ScalarMaybeUninit<M::PointerTag>>, val: impl Into<ScalarMaybeUninit<M::PointerTag>>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
self.write_immediate(Immediate::Scalar(val.into()), dest) self.write_immediate(Immediate::Scalar(val.into()), dest)
} }
@ -676,13 +691,13 @@ where
pub fn write_immediate( pub fn write_immediate(
&mut self, &mut self,
src: Immediate<M::PointerTag>, src: Immediate<M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
self.write_immediate_no_validate(src, dest)?; self.write_immediate_no_validate(src, dest)?;
if M::enforce_validity(self) { if M::enforce_validity(self) {
// Data got changed, better make sure it matches the type! // Data got changed, better make sure it matches the type!
self.validate_operand(self.place_to_op(dest)?)?; self.validate_operand(&self.place_to_op(dest)?)?;
} }
Ok(()) Ok(())
@ -693,13 +708,13 @@ where
pub fn write_immediate_to_mplace( pub fn write_immediate_to_mplace(
&mut self, &mut self,
src: Immediate<M::PointerTag>, src: Immediate<M::PointerTag>,
dest: MPlaceTy<'tcx, M::PointerTag>, dest: &MPlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
self.write_immediate_to_mplace_no_validate(src, dest)?; self.write_immediate_to_mplace_no_validate(src, dest)?;
if M::enforce_validity(self) { if M::enforce_validity(self) {
// Data got changed, better make sure it matches the type! // Data got changed, better make sure it matches the type!
self.validate_operand(dest.into())?; self.validate_operand(&dest.into())?;
} }
Ok(()) Ok(())
@ -711,7 +726,7 @@ where
fn write_immediate_no_validate( fn write_immediate_no_validate(
&mut self, &mut self,
src: Immediate<M::PointerTag>, src: Immediate<M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
if cfg!(debug_assertions) { if cfg!(debug_assertions) {
// This is a very common path, avoid some checks in release mode // This is a very common path, avoid some checks in release mode
@ -754,7 +769,7 @@ where
let dest = MPlaceTy { mplace, layout: dest.layout }; let dest = MPlaceTy { mplace, layout: dest.layout };
// This is already in memory, write there. // This is already in memory, write there.
self.write_immediate_to_mplace_no_validate(src, dest) self.write_immediate_to_mplace_no_validate(src, &dest)
} }
/// Write an immediate to memory. /// Write an immediate to memory.
@ -763,7 +778,7 @@ where
fn write_immediate_to_mplace_no_validate( fn write_immediate_to_mplace_no_validate(
&mut self, &mut self,
value: Immediate<M::PointerTag>, value: Immediate<M::PointerTag>,
dest: MPlaceTy<'tcx, M::PointerTag>, dest: &MPlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// Note that it is really important that the type here is the right one, and matches the // Note that it is really important that the type here is the right one, and matches the
// type things are read at. In case `src_val` is a `ScalarPair`, we don't do any magic here // type things are read at. In case `src_val` is a `ScalarPair`, we don't do any magic here
@ -828,14 +843,14 @@ where
#[inline(always)] #[inline(always)]
pub fn copy_op( pub fn copy_op(
&mut self, &mut self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
self.copy_op_no_validate(src, dest)?; self.copy_op_no_validate(src, dest)?;
if M::enforce_validity(self) { if M::enforce_validity(self) {
// Data got changed, better make sure it matches the type! // Data got changed, better make sure it matches the type!
self.validate_operand(self.place_to_op(dest)?)?; self.validate_operand(&self.place_to_op(dest)?)?;
} }
Ok(()) Ok(())
@ -847,8 +862,8 @@ where
/// right type. /// right type.
fn copy_op_no_validate( fn copy_op_no_validate(
&mut self, &mut self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// We do NOT compare the types for equality, because well-typed code can // We do NOT compare the types for equality, because well-typed code can
// actually "transmute" `&mut T` to `&T` in an assignment without a cast. // actually "transmute" `&mut T` to `&T` in an assignment without a cast.
@ -888,10 +903,10 @@ where
assert_eq!(src.meta, dest.meta, "Can only copy between equally-sized instances"); assert_eq!(src.meta, dest.meta, "Can only copy between equally-sized instances");
let src = self let src = self
.check_mplace_access(src, Some(size)) .check_mplace_access(&src, Some(size))
.expect("places should be checked on creation"); .expect("places should be checked on creation");
let dest = self let dest = self
.check_mplace_access(dest, Some(size)) .check_mplace_access(&dest, Some(size))
.expect("places should be checked on creation"); .expect("places should be checked on creation");
let (src_ptr, dest_ptr) = match (src, dest) { let (src_ptr, dest_ptr) = match (src, dest) {
(Some(src_ptr), Some(dest_ptr)) => (src_ptr, dest_ptr), (Some(src_ptr), Some(dest_ptr)) => (src_ptr, dest_ptr),
@ -906,8 +921,8 @@ where
/// have the same size. /// have the same size.
pub fn copy_op_transmute( pub fn copy_op_transmute(
&mut self, &mut self,
src: OpTy<'tcx, M::PointerTag>, src: &OpTy<'tcx, M::PointerTag>,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
if mir_assign_valid_types(*self.tcx, self.param_env, src.layout, dest.layout) { if mir_assign_valid_types(*self.tcx, self.param_env, src.layout, dest.layout) {
// Fast path: Just use normal `copy_op` // Fast path: Just use normal `copy_op`
@ -944,12 +959,12 @@ where
let dest = self.force_allocation(dest)?; let dest = self.force_allocation(dest)?;
self.copy_op_no_validate( self.copy_op_no_validate(
src, src,
PlaceTy::from(MPlaceTy { mplace: *dest, layout: src.layout }), &PlaceTy::from(MPlaceTy { mplace: *dest, layout: src.layout }),
)?; )?;
if M::enforce_validity(self) { if M::enforce_validity(self) {
// Data got changed, better make sure it matches the type! // Data got changed, better make sure it matches the type!
self.validate_operand(dest.into())?; self.validate_operand(&dest.into())?;
} }
Ok(()) Ok(())
@ -965,7 +980,7 @@ where
/// version. /// version.
pub fn force_allocation_maybe_sized( pub fn force_allocation_maybe_sized(
&mut self, &mut self,
place: PlaceTy<'tcx, M::PointerTag>, place: &PlaceTy<'tcx, M::PointerTag>,
meta: MemPlaceMeta<M::PointerTag>, meta: MemPlaceMeta<M::PointerTag>,
) -> InterpResult<'tcx, (MPlaceTy<'tcx, M::PointerTag>, Option<Size>)> { ) -> InterpResult<'tcx, (MPlaceTy<'tcx, M::PointerTag>, Option<Size>)> {
let (mplace, size) = match place.place { let (mplace, size) = match place.place {
@ -981,7 +996,7 @@ where
self.layout_of_local(&self.stack()[frame], local, None)?; self.layout_of_local(&self.stack()[frame], local, None)?;
// We also need to support unsized types, and hence cannot use `allocate`. // We also need to support unsized types, and hence cannot use `allocate`.
let (size, align) = self let (size, align) = self
.size_and_align_of(meta, local_layout)? .size_and_align_of(&meta, &local_layout)?
.expect("Cannot allocate for non-dyn-sized type"); .expect("Cannot allocate for non-dyn-sized type");
let ptr = self.memory.allocate(size, align, MemoryKind::Stack); let ptr = self.memory.allocate(size, align, MemoryKind::Stack);
let mplace = MemPlace { ptr: ptr.into(), align, meta }; let mplace = MemPlace { ptr: ptr.into(), align, meta };
@ -990,7 +1005,7 @@ where
// We don't have to validate as we can assume the local // We don't have to validate as we can assume the local
// was already valid for its type. // was already valid for its type.
let mplace = MPlaceTy { mplace, layout: local_layout }; let mplace = MPlaceTy { mplace, layout: local_layout };
self.write_immediate_to_mplace_no_validate(value, mplace)?; self.write_immediate_to_mplace_no_validate(value, &mplace)?;
} }
// Now we can call `access_mut` again, asserting it goes well, // Now we can call `access_mut` again, asserting it goes well,
// and actually overwrite things. // and actually overwrite things.
@ -1010,7 +1025,7 @@ where
#[inline(always)] #[inline(always)]
pub fn force_allocation( pub fn force_allocation(
&mut self, &mut self,
place: PlaceTy<'tcx, M::PointerTag>, place: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, MPlaceTy<'tcx, M::PointerTag>> {
Ok(self.force_allocation_maybe_sized(place, MemPlaceMeta::None)?.0) Ok(self.force_allocation_maybe_sized(place, MemPlaceMeta::None)?.0)
} }
@ -1046,7 +1061,7 @@ where
pub fn write_discriminant( pub fn write_discriminant(
&mut self, &mut self,
variant_index: VariantIdx, variant_index: VariantIdx,
dest: PlaceTy<'tcx, M::PointerTag>, dest: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// Layout computation excludes uninhabited variants from consideration // Layout computation excludes uninhabited variants from consideration
// therefore there's no way to represent those variants in the given layout. // therefore there's no way to represent those variants in the given layout.
@ -1077,7 +1092,7 @@ where
let tag_val = size.truncate(discr_val); let tag_val = size.truncate(discr_val);
let tag_dest = self.place_field(dest, tag_field)?; let tag_dest = self.place_field(dest, tag_field)?;
self.write_scalar(Scalar::from_uint(tag_val, size), tag_dest)?; self.write_scalar(Scalar::from_uint(tag_val, size), &tag_dest)?;
} }
Variants::Multiple { Variants::Multiple {
tag_encoding: tag_encoding:
@ -1103,12 +1118,12 @@ where
ImmTy::from_uint(variant_index_relative, tag_layout); ImmTy::from_uint(variant_index_relative, tag_layout);
let tag_val = self.binary_op( let tag_val = self.binary_op(
mir::BinOp::Add, mir::BinOp::Add,
variant_index_relative_val, &variant_index_relative_val,
niche_start_val, &niche_start_val,
)?; )?;
// Write result. // Write result.
let niche_dest = self.place_field(dest, tag_field)?; let niche_dest = self.place_field(dest, tag_field)?;
self.write_immediate(*tag_val, niche_dest)?; self.write_immediate(*tag_val, &niche_dest)?;
} }
} }
} }
@ -1131,7 +1146,7 @@ where
/// Also return some more information so drop doesn't have to run the same code twice. /// Also return some more information so drop doesn't have to run the same code twice.
pub(super) fn unpack_dyn_trait( pub(super) fn unpack_dyn_trait(
&self, &self,
mplace: MPlaceTy<'tcx, M::PointerTag>, mplace: &MPlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, (ty::Instance<'tcx>, MPlaceTy<'tcx, M::PointerTag>)> { ) -> InterpResult<'tcx, (ty::Instance<'tcx>, MPlaceTy<'tcx, M::PointerTag>)> {
let vtable = mplace.vtable(); // also sanity checks the type let vtable = mplace.vtable(); // also sanity checks the type
let (instance, ty) = self.read_drop_type_from_vtable(vtable)?; let (instance, ty) = self.read_drop_type_from_vtable(vtable)?;
@ -1145,7 +1160,7 @@ where
assert_eq!(align, layout.align.abi); assert_eq!(align, layout.align.abi);
} }
let mplace = MPlaceTy { mplace: MemPlace { meta: MemPlaceMeta::None, ..*mplace }, layout }; let mplace = MPlaceTy { mplace: MemPlace { meta: MemPlaceMeta::None, ..**mplace }, layout };
Ok((instance, mplace)) Ok((instance, mplace))
} }
} }

View file

@ -90,7 +90,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
SetDiscriminant { place, variant_index } => { SetDiscriminant { place, variant_index } => {
let dest = self.eval_place(**place)?; let dest = self.eval_place(**place)?;
self.write_discriminant(*variant_index, dest)?; self.write_discriminant(*variant_index, &dest)?;
} }
// Mark locals as alive // Mark locals as alive
@ -110,7 +110,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
// Stacked Borrows. // Stacked Borrows.
Retag(kind, place) => { Retag(kind, place) => {
let dest = self.eval_place(**place)?; let dest = self.eval_place(**place)?;
M::retag(self, *kind, dest)?; M::retag(self, *kind, &dest)?;
} }
// Statements we do not track. // Statements we do not track.
@ -156,45 +156,45 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
ThreadLocalRef(did) => { ThreadLocalRef(did) => {
let id = M::thread_local_static_alloc_id(self, did)?; let id = M::thread_local_static_alloc_id(self, did)?;
let val = self.global_base_pointer(id.into())?; let val = self.global_base_pointer(id.into())?;
self.write_scalar(val, dest)?; self.write_scalar(val, &dest)?;
} }
Use(ref operand) => { Use(ref operand) => {
// Avoid recomputing the layout // Avoid recomputing the layout
let op = self.eval_operand(operand, Some(dest.layout))?; let op = self.eval_operand(operand, Some(dest.layout))?;
self.copy_op(op, dest)?; self.copy_op(&op, &dest)?;
} }
BinaryOp(bin_op, ref left, ref right) => { BinaryOp(bin_op, ref left, ref right) => {
let layout = binop_left_homogeneous(bin_op).then_some(dest.layout); let layout = binop_left_homogeneous(bin_op).then_some(dest.layout);
let left = self.read_immediate(self.eval_operand(left, layout)?)?; let left = self.read_immediate(&self.eval_operand(left, layout)?)?;
let layout = binop_right_homogeneous(bin_op).then_some(left.layout); let layout = binop_right_homogeneous(bin_op).then_some(left.layout);
let right = self.read_immediate(self.eval_operand(right, layout)?)?; let right = self.read_immediate(&self.eval_operand(right, layout)?)?;
self.binop_ignore_overflow(bin_op, left, right, dest)?; self.binop_ignore_overflow(bin_op, &left, &right, &dest)?;
} }
CheckedBinaryOp(bin_op, ref left, ref right) => { CheckedBinaryOp(bin_op, ref left, ref right) => {
// Due to the extra boolean in the result, we can never reuse the `dest.layout`. // Due to the extra boolean in the result, we can never reuse the `dest.layout`.
let left = self.read_immediate(self.eval_operand(left, None)?)?; let left = self.read_immediate(&self.eval_operand(left, None)?)?;
let layout = binop_right_homogeneous(bin_op).then_some(left.layout); let layout = binop_right_homogeneous(bin_op).then_some(left.layout);
let right = self.read_immediate(self.eval_operand(right, layout)?)?; let right = self.read_immediate(&self.eval_operand(right, layout)?)?;
self.binop_with_overflow(bin_op, left, right, dest)?; self.binop_with_overflow(bin_op, &left, &right, &dest)?;
} }
UnaryOp(un_op, ref operand) => { UnaryOp(un_op, ref operand) => {
// The operand always has the same type as the result. // The operand always has the same type as the result.
let val = self.read_immediate(self.eval_operand(operand, Some(dest.layout))?)?; let val = self.read_immediate(&self.eval_operand(operand, Some(dest.layout))?)?;
let val = self.unary_op(un_op, val)?; let val = self.unary_op(un_op, &val)?;
assert_eq!(val.layout, dest.layout, "layout mismatch for result of {:?}", un_op); assert_eq!(val.layout, dest.layout, "layout mismatch for result of {:?}", un_op);
self.write_immediate(*val, dest)?; self.write_immediate(*val, &dest)?;
} }
Aggregate(ref kind, ref operands) => { Aggregate(ref kind, ref operands) => {
let (dest, active_field_index) = match **kind { let (dest, active_field_index) = match **kind {
mir::AggregateKind::Adt(adt_def, variant_index, _, _, active_field_index) => { mir::AggregateKind::Adt(adt_def, variant_index, _, _, active_field_index) => {
self.write_discriminant(variant_index, dest)?; self.write_discriminant(variant_index, &dest)?;
if adt_def.is_enum() { if adt_def.is_enum() {
(self.place_downcast(dest, variant_index)?, active_field_index) (self.place_downcast(&dest, variant_index)?, active_field_index)
} else { } else {
(dest, active_field_index) (dest, active_field_index)
} }
@ -207,21 +207,21 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
// Ignore zero-sized fields. // Ignore zero-sized fields.
if !op.layout.is_zst() { if !op.layout.is_zst() {
let field_index = active_field_index.unwrap_or(i); let field_index = active_field_index.unwrap_or(i);
let field_dest = self.place_field(dest, field_index)?; let field_dest = self.place_field(&dest, field_index)?;
self.copy_op(op, field_dest)?; self.copy_op(&op, &field_dest)?;
} }
} }
} }
Repeat(ref operand, _) => { Repeat(ref operand, _) => {
let op = self.eval_operand(operand, None)?; let op = self.eval_operand(operand, None)?;
let dest = self.force_allocation(dest)?; let dest = self.force_allocation(&dest)?;
let length = dest.len(self)?; let length = dest.len(self)?;
if let Some(first_ptr) = self.check_mplace_access(dest, None)? { if let Some(first_ptr) = self.check_mplace_access(&dest, None)? {
// Write the first. // Write the first.
let first = self.mplace_field(dest, 0)?; let first = self.mplace_field(&dest, 0)?;
self.copy_op(op, first.into())?; self.copy_op(&op, &first.into())?;
if length > 1 { if length > 1 {
let elem_size = first.layout.size; let elem_size = first.layout.size;
@ -242,23 +242,23 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
Len(place) => { Len(place) => {
// FIXME(CTFE): don't allow computing the length of arrays in const eval // FIXME(CTFE): don't allow computing the length of arrays in const eval
let src = self.eval_place(place)?; let src = self.eval_place(place)?;
let mplace = self.force_allocation(src)?; let mplace = self.force_allocation(&src)?;
let len = mplace.len(self)?; let len = mplace.len(self)?;
self.write_scalar(Scalar::from_machine_usize(len, self), dest)?; self.write_scalar(Scalar::from_machine_usize(len, self), &dest)?;
} }
AddressOf(_, place) | Ref(_, _, place) => { AddressOf(_, place) | Ref(_, _, place) => {
let src = self.eval_place(place)?; let src = self.eval_place(place)?;
let place = self.force_allocation(src)?; let place = self.force_allocation(&src)?;
if place.layout.size.bytes() > 0 { if place.layout.size.bytes() > 0 {
// definitely not a ZST // definitely not a ZST
assert!(place.ptr.is_ptr(), "non-ZST places should be normalized to `Pointer`"); assert!(place.ptr.is_ptr(), "non-ZST places should be normalized to `Pointer`");
} }
self.write_immediate(place.to_ref(), dest)?; self.write_immediate(place.to_ref(), &dest)?;
} }
NullaryOp(mir::NullOp::Box, _) => { NullaryOp(mir::NullOp::Box, _) => {
M::box_alloc(self, dest)?; M::box_alloc(self, &dest)?;
} }
NullaryOp(mir::NullOp::SizeOf, ty) => { NullaryOp(mir::NullOp::SizeOf, ty) => {
@ -272,19 +272,19 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
); );
throw_inval!(SizeOfUnsizedType(ty)); throw_inval!(SizeOfUnsizedType(ty));
} }
self.write_scalar(Scalar::from_machine_usize(layout.size.bytes(), self), dest)?; self.write_scalar(Scalar::from_machine_usize(layout.size.bytes(), self), &dest)?;
} }
Cast(cast_kind, ref operand, cast_ty) => { Cast(cast_kind, ref operand, cast_ty) => {
let src = self.eval_operand(operand, None)?; let src = self.eval_operand(operand, None)?;
let cast_ty = self.subst_from_current_frame_and_normalize_erasing_regions(cast_ty); let cast_ty = self.subst_from_current_frame_and_normalize_erasing_regions(cast_ty);
self.cast(src, cast_kind, cast_ty, dest)?; self.cast(&src, cast_kind, cast_ty, &dest)?;
} }
Discriminant(place) => { Discriminant(place) => {
let op = self.eval_place_to_op(place, None)?; let op = self.eval_place_to_op(place, None)?;
let discr_val = self.read_discriminant(op)?.0; let discr_val = self.read_discriminant(&op)?.0;
self.write_scalar(discr_val, dest)?; self.write_scalar(discr_val, &dest)?;
} }
} }

View file

@ -25,7 +25,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
Goto { target } => self.go_to_block(target), Goto { target } => self.go_to_block(target),
SwitchInt { ref discr, ref targets, switch_ty } => { SwitchInt { ref discr, ref targets, switch_ty } => {
let discr = self.read_immediate(self.eval_operand(discr, None)?)?; let discr = self.read_immediate(&self.eval_operand(discr, None)?)?;
trace!("SwitchInt({:?})", *discr); trace!("SwitchInt({:?})", *discr);
assert_eq!(discr.layout.ty, switch_ty); assert_eq!(discr.layout.ty, switch_ty);
@ -38,8 +38,8 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let res = self let res = self
.overflowing_binary_op( .overflowing_binary_op(
mir::BinOp::Eq, mir::BinOp::Eq,
discr, &discr,
ImmTy::from_uint(const_int, discr.layout), &ImmTy::from_uint(const_int, discr.layout),
)? )?
.0; .0;
if res.to_bool()? { if res.to_bool()? {
@ -58,7 +58,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let (fn_val, abi) = match *func.layout.ty.kind() { let (fn_val, abi) = match *func.layout.ty.kind() {
ty::FnPtr(sig) => { ty::FnPtr(sig) => {
let caller_abi = sig.abi(); let caller_abi = sig.abi();
let fn_ptr = self.read_scalar(func)?.check_init()?; let fn_ptr = self.read_scalar(&func)?.check_init()?;
let fn_val = self.memory.get_fn(fn_ptr)?; let fn_val = self.memory.get_fn(fn_ptr)?;
(fn_val, caller_abi) (fn_val, caller_abi)
} }
@ -78,8 +78,12 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
), ),
}; };
let args = self.eval_operands(args)?; let args = self.eval_operands(args)?;
let dest_place;
let ret = match destination { let ret = match destination {
Some((dest, ret)) => Some((self.eval_place(dest)?, ret)), Some((dest, ret)) => {
dest_place = self.eval_place(dest)?;
Some((&dest_place, ret))
}
None => None, None => None,
}; };
self.eval_fn_call(fn_val, abi, &args[..], ret, *cleanup)?; self.eval_fn_call(fn_val, abi, &args[..], ret, *cleanup)?;
@ -96,12 +100,12 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
trace!("TerminatorKind::drop: {:?}, type {}", place, ty); trace!("TerminatorKind::drop: {:?}, type {}", place, ty);
let instance = Instance::resolve_drop_in_place(*self.tcx, ty); let instance = Instance::resolve_drop_in_place(*self.tcx, ty);
self.drop_in_place(place, instance, target, unwind)?; self.drop_in_place(&place, instance, target, unwind)?;
} }
Assert { ref cond, expected, ref msg, target, cleanup } => { Assert { ref cond, expected, ref msg, target, cleanup } => {
let cond_val = let cond_val =
self.read_immediate(self.eval_operand(cond, None)?)?.to_scalar()?.to_bool()?; self.read_immediate(&self.eval_operand(cond, None)?)?.to_scalar()?.to_bool()?;
if expected == cond_val { if expected == cond_val {
self.go_to_block(target); self.go_to_block(target);
} else { } else {
@ -180,7 +184,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
&mut self, &mut self,
rust_abi: bool, rust_abi: bool,
caller_arg: &mut impl Iterator<Item = OpTy<'tcx, M::PointerTag>>, caller_arg: &mut impl Iterator<Item = OpTy<'tcx, M::PointerTag>>,
callee_arg: PlaceTy<'tcx, M::PointerTag>, callee_arg: &PlaceTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
if rust_abi && callee_arg.layout.is_zst() { if rust_abi && callee_arg.layout.is_zst() {
// Nothing to do. // Nothing to do.
@ -202,7 +206,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
) )
} }
// We allow some transmutes here // We allow some transmutes here
self.copy_op_transmute(caller_arg, callee_arg) self.copy_op_transmute(&caller_arg, callee_arg)
} }
/// Call this function -- pushing the stack frame and initializing the arguments. /// Call this function -- pushing the stack frame and initializing the arguments.
@ -211,7 +215,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
fn_val: FnVal<'tcx, M::ExtraFnVal>, fn_val: FnVal<'tcx, M::ExtraFnVal>,
caller_abi: Abi, caller_abi: Abi,
args: &[OpTy<'tcx, M::PointerTag>], args: &[OpTy<'tcx, M::PointerTag>],
ret: Option<(PlaceTy<'tcx, M::PointerTag>, mir::BasicBlock)>, ret: Option<(&PlaceTy<'tcx, M::PointerTag>, mir::BasicBlock)>,
unwind: Option<mir::BasicBlock>, unwind: Option<mir::BasicBlock>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
trace!("eval_fn_call: {:#?}", fn_val); trace!("eval_fn_call: {:#?}", fn_val);
@ -314,7 +318,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let caller_args: Cow<'_, [OpTy<'tcx, M::PointerTag>]> = let caller_args: Cow<'_, [OpTy<'tcx, M::PointerTag>]> =
if caller_abi == Abi::RustCall && !args.is_empty() { if caller_abi == Abi::RustCall && !args.is_empty() {
// Untuple // Untuple
let (&untuple_arg, args) = args.split_last().unwrap(); let (untuple_arg, args) = args.split_last().unwrap();
trace!("eval_fn_call: Will pass last argument by untupling"); trace!("eval_fn_call: Will pass last argument by untupling");
Cow::from( Cow::from(
args.iter() args.iter()
@ -344,12 +348,12 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
if Some(local) == body.spread_arg { if Some(local) == body.spread_arg {
// Must be a tuple // Must be a tuple
for i in 0..dest.layout.fields.count() { for i in 0..dest.layout.fields.count() {
let dest = self.place_field(dest, i)?; let dest = self.place_field(&dest, i)?;
self.pass_argument(rust_abi, &mut caller_iter, dest)?; self.pass_argument(rust_abi, &mut caller_iter, &dest)?;
} }
} else { } else {
// Normal argument // Normal argument
self.pass_argument(rust_abi, &mut caller_iter, dest)?; self.pass_argument(rust_abi, &mut caller_iter, &dest)?;
} }
} }
// Now we should have no more caller args // Now we should have no more caller args
@ -397,7 +401,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let receiver_place = match args[0].layout.ty.builtin_deref(true) { let receiver_place = match args[0].layout.ty.builtin_deref(true) {
Some(_) => { Some(_) => {
// Built-in pointer. // Built-in pointer.
self.deref_operand(args[0])? self.deref_operand(&args[0])?
} }
None => { None => {
// Unsized self. // Unsized self.
@ -426,7 +430,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
fn drop_in_place( fn drop_in_place(
&mut self, &mut self,
place: PlaceTy<'tcx, M::PointerTag>, place: &PlaceTy<'tcx, M::PointerTag>,
instance: ty::Instance<'tcx>, instance: ty::Instance<'tcx>,
target: mir::BasicBlock, target: mir::BasicBlock,
unwind: Option<mir::BasicBlock>, unwind: Option<mir::BasicBlock>,
@ -440,7 +444,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let (instance, place) = match place.layout.ty.kind() { let (instance, place) = match place.layout.ty.kind() {
ty::Dynamic(..) => { ty::Dynamic(..) => {
// Dropping a trait object. // Dropping a trait object.
self.unpack_dyn_trait(place)? self.unpack_dyn_trait(&place)?
} }
_ => (instance, place), _ => (instance, place),
}; };
@ -457,7 +461,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
FnVal::Instance(instance), FnVal::Instance(instance),
Abi::Rust, Abi::Rust,
&[arg.into()], &[arg.into()],
Some((dest.into(), target)), Some((&dest.into(), target)),
unwind, unwind,
) )
} }

View file

@ -378,7 +378,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValidityVisitor<'rt, 'mir, '
/// Check a reference or `Box`. /// Check a reference or `Box`.
fn check_safe_pointer( fn check_safe_pointer(
&mut self, &mut self,
value: OpTy<'tcx, M::PointerTag>, value: &OpTy<'tcx, M::PointerTag>,
kind: &str, kind: &str,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
let value = try_validation!( let value = try_validation!(
@ -389,7 +389,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValidityVisitor<'rt, 'mir, '
// Handle wide pointers. // Handle wide pointers.
// Check metadata early, for better diagnostics // Check metadata early, for better diagnostics
let place = try_validation!( let place = try_validation!(
self.ecx.ref_to_mplace(value), self.ecx.ref_to_mplace(&value),
self.path, self.path,
err_ub!(InvalidUninitBytes(None)) => { "uninitialized {}", kind }, err_ub!(InvalidUninitBytes(None)) => { "uninitialized {}", kind },
); );
@ -398,7 +398,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValidityVisitor<'rt, 'mir, '
} }
// Make sure this is dereferenceable and all. // Make sure this is dereferenceable and all.
let size_and_align = try_validation!( let size_and_align = try_validation!(
self.ecx.size_and_align_of_mplace(place), self.ecx.size_and_align_of_mplace(&place),
self.path, self.path,
err_ub!(InvalidMeta(msg)) => { "invalid {} metadata: {}", kind, msg }, err_ub!(InvalidMeta(msg)) => { "invalid {} metadata: {}", kind, msg },
); );
@ -494,7 +494,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValidityVisitor<'rt, 'mir, '
fn read_scalar( fn read_scalar(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, ScalarMaybeUninit<M::PointerTag>> { ) -> InterpResult<'tcx, ScalarMaybeUninit<M::PointerTag>> {
Ok(try_validation!( Ok(try_validation!(
self.ecx.read_scalar(op), self.ecx.read_scalar(op),
@ -507,7 +507,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValidityVisitor<'rt, 'mir, '
/// at that type. Return `true` if the type is indeed primitive. /// at that type. Return `true` if the type is indeed primitive.
fn try_visit_primitive( fn try_visit_primitive(
&mut self, &mut self,
value: OpTy<'tcx, M::PointerTag>, value: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, bool> { ) -> InterpResult<'tcx, bool> {
// Go over all the primitive types // Go over all the primitive types
let ty = value.layout.ty; let ty = value.layout.ty;
@ -555,7 +555,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValidityVisitor<'rt, 'mir, '
// actually enforce the strict rules for raw pointers (mostly because // actually enforce the strict rules for raw pointers (mostly because
// that lets us re-use `ref_to_mplace`). // that lets us re-use `ref_to_mplace`).
let place = try_validation!( let place = try_validation!(
self.ecx.read_immediate(value).and_then(|i| self.ecx.ref_to_mplace(i)), self.ecx.read_immediate(value).and_then(|ref i| self.ecx.ref_to_mplace(i)),
self.path, self.path,
err_ub!(InvalidUninitBytes(None)) => { "uninitialized raw pointer" }, err_ub!(InvalidUninitBytes(None)) => { "uninitialized raw pointer" },
err_unsup!(ReadPointerAsBytes) => { "part of a pointer" } expected { "a proper pointer or integer value" }, err_unsup!(ReadPointerAsBytes) => { "part of a pointer" } expected { "a proper pointer or integer value" },
@ -634,7 +634,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValidityVisitor<'rt, 'mir, '
fn visit_scalar( fn visit_scalar(
&mut self, &mut self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
scalar_layout: &Scalar, scalar_layout: &Scalar,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
let value = self.read_scalar(op)?; let value = self.read_scalar(op)?;
@ -708,7 +708,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValueVisitor<'mir, 'tcx, M>
fn read_discriminant( fn read_discriminant(
&mut self, &mut self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, VariantIdx> { ) -> InterpResult<'tcx, VariantIdx> {
self.with_elem(PathElem::EnumTag, move |this| { self.with_elem(PathElem::EnumTag, move |this| {
Ok(try_validation!( Ok(try_validation!(
@ -728,9 +728,9 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValueVisitor<'mir, 'tcx, M>
#[inline] #[inline]
fn visit_field( fn visit_field(
&mut self, &mut self,
old_op: OpTy<'tcx, M::PointerTag>, old_op: &OpTy<'tcx, M::PointerTag>,
field: usize, field: usize,
new_op: OpTy<'tcx, M::PointerTag>, new_op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
let elem = self.aggregate_field_path_elem(old_op.layout, field); let elem = self.aggregate_field_path_elem(old_op.layout, field);
self.with_elem(elem, move |this| this.visit_value(new_op)) self.with_elem(elem, move |this| this.visit_value(new_op))
@ -739,9 +739,9 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValueVisitor<'mir, 'tcx, M>
#[inline] #[inline]
fn visit_variant( fn visit_variant(
&mut self, &mut self,
old_op: OpTy<'tcx, M::PointerTag>, old_op: &OpTy<'tcx, M::PointerTag>,
variant_id: VariantIdx, variant_id: VariantIdx,
new_op: OpTy<'tcx, M::PointerTag>, new_op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
let name = match old_op.layout.ty.kind() { let name = match old_op.layout.ty.kind() {
ty::Adt(adt, _) => PathElem::Variant(adt.variants[variant_id].ident.name), ty::Adt(adt, _) => PathElem::Variant(adt.variants[variant_id].ident.name),
@ -755,14 +755,14 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValueVisitor<'mir, 'tcx, M>
#[inline(always)] #[inline(always)]
fn visit_union( fn visit_union(
&mut self, &mut self,
_op: OpTy<'tcx, M::PointerTag>, _op: &OpTy<'tcx, M::PointerTag>,
_fields: NonZeroUsize, _fields: NonZeroUsize,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
Ok(()) Ok(())
} }
#[inline] #[inline]
fn visit_value(&mut self, op: OpTy<'tcx, M::PointerTag>) -> InterpResult<'tcx> { fn visit_value(&mut self, op: &OpTy<'tcx, M::PointerTag>) -> InterpResult<'tcx> {
trace!("visit_value: {:?}, {:?}", *op, op.layout); trace!("visit_value: {:?}, {:?}", *op, op.layout);
// Check primitive types -- the leafs of our recursive descend. // Check primitive types -- the leafs of our recursive descend.
@ -819,7 +819,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValueVisitor<'mir, 'tcx, M>
fn visit_aggregate( fn visit_aggregate(
&mut self, &mut self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
fields: impl Iterator<Item = InterpResult<'tcx, Self::V>>, fields: impl Iterator<Item = InterpResult<'tcx, Self::V>>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
match op.layout.ty.kind() { match op.layout.ty.kind() {
@ -921,7 +921,7 @@ impl<'rt, 'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> ValueVisitor<'mir, 'tcx, M>
impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> { impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
fn validate_operand_internal( fn validate_operand_internal(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
path: Vec<PathElem>, path: Vec<PathElem>,
ref_tracking: Option<&mut RefTracking<MPlaceTy<'tcx, M::PointerTag>, Vec<PathElem>>>, ref_tracking: Option<&mut RefTracking<MPlaceTy<'tcx, M::PointerTag>, Vec<PathElem>>>,
ctfe_mode: Option<CtfeValidationMode>, ctfe_mode: Option<CtfeValidationMode>,
@ -932,10 +932,10 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
let mut visitor = ValidityVisitor { path, ref_tracking, ctfe_mode, ecx: self }; let mut visitor = ValidityVisitor { path, ref_tracking, ctfe_mode, ecx: self };
// Try to cast to ptr *once* instead of all the time. // Try to cast to ptr *once* instead of all the time.
let op = self.force_op_ptr(op).unwrap_or(op); let op = self.force_op_ptr(&op).unwrap_or(*op);
// Run it. // Run it.
match visitor.visit_value(op) { match visitor.visit_value(&op) {
Ok(()) => Ok(()), Ok(()) => Ok(()),
// Pass through validation failures. // Pass through validation failures.
Err(err) if matches!(err.kind(), err_ub!(ValidationFailure { .. })) => Err(err), Err(err) if matches!(err.kind(), err_ub!(ValidationFailure { .. })) => Err(err),
@ -963,7 +963,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
#[inline(always)] #[inline(always)]
pub fn const_validate_operand( pub fn const_validate_operand(
&self, &self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
path: Vec<PathElem>, path: Vec<PathElem>,
ref_tracking: &mut RefTracking<MPlaceTy<'tcx, M::PointerTag>, Vec<PathElem>>, ref_tracking: &mut RefTracking<MPlaceTy<'tcx, M::PointerTag>, Vec<PathElem>>,
ctfe_mode: CtfeValidationMode, ctfe_mode: CtfeValidationMode,
@ -975,7 +975,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> InterpCx<'mir, 'tcx, M> {
/// `op` is assumed to cover valid memory if it is an indirect operand. /// `op` is assumed to cover valid memory if it is an indirect operand.
/// It will error if the bits at the destination do not match the ones described by the layout. /// It will error if the bits at the destination do not match the ones described by the layout.
#[inline(always)] #[inline(always)]
pub fn validate_operand(&self, op: OpTy<'tcx, M::PointerTag>) -> InterpResult<'tcx> { pub fn validate_operand(&self, op: &OpTy<'tcx, M::PointerTag>) -> InterpResult<'tcx> {
self.validate_operand_internal(op, vec![], None, None) self.validate_operand_internal(op, vec![], None, None)
} }
} }

View file

@ -18,21 +18,25 @@ pub trait Value<'mir, 'tcx, M: Machine<'mir, 'tcx>>: Copy {
fn layout(&self) -> TyAndLayout<'tcx>; fn layout(&self) -> TyAndLayout<'tcx>;
/// Makes this into an `OpTy`. /// Makes this into an `OpTy`.
fn to_op(self, ecx: &InterpCx<'mir, 'tcx, M>) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>>; fn to_op(&self, ecx: &InterpCx<'mir, 'tcx, M>)
-> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>>;
/// Creates this from an `MPlaceTy`. /// Creates this from an `MPlaceTy`.
fn from_mem_place(mplace: MPlaceTy<'tcx, M::PointerTag>) -> Self; fn from_mem_place(mplace: MPlaceTy<'tcx, M::PointerTag>) -> Self;
/// Projects to the given enum variant. /// Projects to the given enum variant.
fn project_downcast( fn project_downcast(
self, &self,
ecx: &InterpCx<'mir, 'tcx, M>, ecx: &InterpCx<'mir, 'tcx, M>,
variant: VariantIdx, variant: VariantIdx,
) -> InterpResult<'tcx, Self>; ) -> InterpResult<'tcx, Self>;
/// Projects to the n-th field. /// Projects to the n-th field.
fn project_field(self, ecx: &InterpCx<'mir, 'tcx, M>, field: usize) fn project_field(
-> InterpResult<'tcx, Self>; &self,
ecx: &InterpCx<'mir, 'tcx, M>,
field: usize,
) -> InterpResult<'tcx, Self>;
} }
// Operands and memory-places are both values. // Operands and memory-places are both values.
@ -45,10 +49,10 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> Value<'mir, 'tcx, M> for OpTy<'tc
#[inline(always)] #[inline(always)]
fn to_op( fn to_op(
self, &self,
_ecx: &InterpCx<'mir, 'tcx, M>, _ecx: &InterpCx<'mir, 'tcx, M>,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
Ok(self) Ok(*self)
} }
#[inline(always)] #[inline(always)]
@ -58,7 +62,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> Value<'mir, 'tcx, M> for OpTy<'tc
#[inline(always)] #[inline(always)]
fn project_downcast( fn project_downcast(
self, &self,
ecx: &InterpCx<'mir, 'tcx, M>, ecx: &InterpCx<'mir, 'tcx, M>,
variant: VariantIdx, variant: VariantIdx,
) -> InterpResult<'tcx, Self> { ) -> InterpResult<'tcx, Self> {
@ -67,7 +71,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> Value<'mir, 'tcx, M> for OpTy<'tc
#[inline(always)] #[inline(always)]
fn project_field( fn project_field(
self, &self,
ecx: &InterpCx<'mir, 'tcx, M>, ecx: &InterpCx<'mir, 'tcx, M>,
field: usize, field: usize,
) -> InterpResult<'tcx, Self> { ) -> InterpResult<'tcx, Self> {
@ -85,10 +89,10 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> Value<'mir, 'tcx, M>
#[inline(always)] #[inline(always)]
fn to_op( fn to_op(
self, &self,
_ecx: &InterpCx<'mir, 'tcx, M>, _ecx: &InterpCx<'mir, 'tcx, M>,
) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> { ) -> InterpResult<'tcx, OpTy<'tcx, M::PointerTag>> {
Ok(self.into()) Ok((*self).into())
} }
#[inline(always)] #[inline(always)]
@ -98,7 +102,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> Value<'mir, 'tcx, M>
#[inline(always)] #[inline(always)]
fn project_downcast( fn project_downcast(
self, &self,
ecx: &InterpCx<'mir, 'tcx, M>, ecx: &InterpCx<'mir, 'tcx, M>,
variant: VariantIdx, variant: VariantIdx,
) -> InterpResult<'tcx, Self> { ) -> InterpResult<'tcx, Self> {
@ -107,7 +111,7 @@ impl<'mir, 'tcx: 'mir, M: Machine<'mir, 'tcx>> Value<'mir, 'tcx, M>
#[inline(always)] #[inline(always)]
fn project_field( fn project_field(
self, &self,
ecx: &InterpCx<'mir, 'tcx, M>, ecx: &InterpCx<'mir, 'tcx, M>,
field: usize, field: usize,
) -> InterpResult<'tcx, Self> { ) -> InterpResult<'tcx, Self> {
@ -129,7 +133,7 @@ macro_rules! make_value_visitor {
#[inline(always)] #[inline(always)]
fn read_discriminant( fn read_discriminant(
&mut self, &mut self,
op: OpTy<'tcx, M::PointerTag>, op: &OpTy<'tcx, M::PointerTag>,
) -> InterpResult<'tcx, VariantIdx> { ) -> InterpResult<'tcx, VariantIdx> {
Ok(self.ecx().read_discriminant(op)?.1) Ok(self.ecx().read_discriminant(op)?.1)
} }
@ -137,13 +141,13 @@ macro_rules! make_value_visitor {
// Recursive actions, ready to be overloaded. // Recursive actions, ready to be overloaded.
/// Visits the given value, dispatching as appropriate to more specialized visitors. /// Visits the given value, dispatching as appropriate to more specialized visitors.
#[inline(always)] #[inline(always)]
fn visit_value(&mut self, v: Self::V) -> InterpResult<'tcx> fn visit_value(&mut self, v: &Self::V) -> InterpResult<'tcx>
{ {
self.walk_value(v) self.walk_value(v)
} }
/// Visits the given value as a union. No automatic recursion can happen here. /// Visits the given value as a union. No automatic recursion can happen here.
#[inline(always)] #[inline(always)]
fn visit_union(&mut self, _v: Self::V, _fields: NonZeroUsize) -> InterpResult<'tcx> fn visit_union(&mut self, _v: &Self::V, _fields: NonZeroUsize) -> InterpResult<'tcx>
{ {
Ok(()) Ok(())
} }
@ -153,7 +157,7 @@ macro_rules! make_value_visitor {
#[inline(always)] #[inline(always)]
fn visit_aggregate( fn visit_aggregate(
&mut self, &mut self,
v: Self::V, v: &Self::V,
fields: impl Iterator<Item=InterpResult<'tcx, Self::V>>, fields: impl Iterator<Item=InterpResult<'tcx, Self::V>>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
self.walk_aggregate(v, fields) self.walk_aggregate(v, fields)
@ -167,9 +171,9 @@ macro_rules! make_value_visitor {
#[inline(always)] #[inline(always)]
fn visit_field( fn visit_field(
&mut self, &mut self,
_old_val: Self::V, _old_val: &Self::V,
_field: usize, _field: usize,
new_val: Self::V, new_val: &Self::V,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
self.visit_value(new_val) self.visit_value(new_val)
} }
@ -179,9 +183,9 @@ macro_rules! make_value_visitor {
#[inline(always)] #[inline(always)]
fn visit_variant( fn visit_variant(
&mut self, &mut self,
_old_val: Self::V, _old_val: &Self::V,
_variant: VariantIdx, _variant: VariantIdx,
new_val: Self::V, new_val: &Self::V,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
self.visit_value(new_val) self.visit_value(new_val)
} }
@ -189,16 +193,16 @@ macro_rules! make_value_visitor {
// Default recursors. Not meant to be overloaded. // Default recursors. Not meant to be overloaded.
fn walk_aggregate( fn walk_aggregate(
&mut self, &mut self,
v: Self::V, v: &Self::V,
fields: impl Iterator<Item=InterpResult<'tcx, Self::V>>, fields: impl Iterator<Item=InterpResult<'tcx, Self::V>>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
// Now iterate over it. // Now iterate over it.
for (idx, field_val) in fields.enumerate() { for (idx, field_val) in fields.enumerate() {
self.visit_field(v, idx, field_val?)?; self.visit_field(v, idx, &field_val?)?;
} }
Ok(()) Ok(())
} }
fn walk_value(&mut self, v: Self::V) -> InterpResult<'tcx> fn walk_value(&mut self, v: &Self::V) -> InterpResult<'tcx>
{ {
trace!("walk_value: type: {}", v.layout().ty); trace!("walk_value: type: {}", v.layout().ty);
@ -208,10 +212,10 @@ macro_rules! make_value_visitor {
ty::Dynamic(..) => { ty::Dynamic(..) => {
// immediate trait objects are not a thing // immediate trait objects are not a thing
let dest = v.to_op(self.ecx())?.assert_mem_place(self.ecx()); let dest = v.to_op(self.ecx())?.assert_mem_place(self.ecx());
let inner = self.ecx().unpack_dyn_trait(dest)?.1; let inner = self.ecx().unpack_dyn_trait(&dest)?.1;
trace!("walk_value: dyn object layout: {:#?}", inner.layout); trace!("walk_value: dyn object layout: {:#?}", inner.layout);
// recurse with the inner type // recurse with the inner type
return self.visit_field(v, 0, Value::from_mem_place(inner)); return self.visit_field(&v, 0, &Value::from_mem_place(inner));
}, },
// Slices do not need special handling here: they have `Array` field // Slices do not need special handling here: they have `Array` field
// placement with length 0, so we enter the `Array` case below which // placement with length 0, so we enter the `Array` case below which
@ -241,7 +245,7 @@ macro_rules! make_value_visitor {
// Now we can go over all the fields. // Now we can go over all the fields.
// This uses the *run-time length*, i.e., if we are a slice, // This uses the *run-time length*, i.e., if we are a slice,
// the dynamic info from the metadata is used. // the dynamic info from the metadata is used.
let iter = self.ecx().mplace_array_fields(mplace)? let iter = self.ecx().mplace_array_fields(&mplace)?
.map(|f| f.and_then(|f| { .map(|f| f.and_then(|f| {
Ok(Value::from_mem_place(f)) Ok(Value::from_mem_place(f))
})); }));
@ -254,11 +258,11 @@ macro_rules! make_value_visitor {
// with *its* fields. // with *its* fields.
Variants::Multiple { .. } => { Variants::Multiple { .. } => {
let op = v.to_op(self.ecx())?; let op = v.to_op(self.ecx())?;
let idx = self.read_discriminant(op)?; let idx = self.read_discriminant(&op)?;
let inner = v.project_downcast(self.ecx(), idx)?; let inner = v.project_downcast(self.ecx(), idx)?;
trace!("walk_value: variant layout: {:#?}", inner.layout()); trace!("walk_value: variant layout: {:#?}", inner.layout());
// recurse with the inner type // recurse with the inner type
self.visit_variant(v, idx, inner) self.visit_variant(v, idx, &inner)
} }
// For single-variant layouts, we already did anything there is to do. // For single-variant layouts, we already did anything there is to do.
Variants::Single { .. } => Ok(()) Variants::Single { .. } => Ok(())

View file

@ -197,7 +197,7 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for ConstPropMachine<'mir, 'tcx>
_instance: ty::Instance<'tcx>, _instance: ty::Instance<'tcx>,
_abi: Abi, _abi: Abi,
_args: &[OpTy<'tcx>], _args: &[OpTy<'tcx>],
_ret: Option<(PlaceTy<'tcx>, BasicBlock)>, _ret: Option<(&PlaceTy<'tcx>, BasicBlock)>,
_unwind: Option<BasicBlock>, _unwind: Option<BasicBlock>,
) -> InterpResult<'tcx, Option<&'mir Body<'tcx>>> { ) -> InterpResult<'tcx, Option<&'mir Body<'tcx>>> {
Ok(None) Ok(None)
@ -207,7 +207,7 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for ConstPropMachine<'mir, 'tcx>
_ecx: &mut InterpCx<'mir, 'tcx, Self>, _ecx: &mut InterpCx<'mir, 'tcx, Self>,
_instance: ty::Instance<'tcx>, _instance: ty::Instance<'tcx>,
_args: &[OpTy<'tcx>], _args: &[OpTy<'tcx>],
_ret: Option<(PlaceTy<'tcx>, BasicBlock)>, _ret: Option<(&PlaceTy<'tcx>, BasicBlock)>,
_unwind: Option<BasicBlock>, _unwind: Option<BasicBlock>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
throw_machine_stop_str!("calling intrinsics isn't supported in ConstProp") throw_machine_stop_str!("calling intrinsics isn't supported in ConstProp")
@ -228,8 +228,8 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for ConstPropMachine<'mir, 'tcx>
fn binary_ptr_op( fn binary_ptr_op(
_ecx: &InterpCx<'mir, 'tcx, Self>, _ecx: &InterpCx<'mir, 'tcx, Self>,
_bin_op: BinOp, _bin_op: BinOp,
_left: ImmTy<'tcx>, _left: &ImmTy<'tcx>,
_right: ImmTy<'tcx>, _right: &ImmTy<'tcx>,
) -> InterpResult<'tcx, (Scalar, bool, Ty<'tcx>)> { ) -> InterpResult<'tcx, (Scalar, bool, Ty<'tcx>)> {
// We can't do this because aliasing of memory can differ between const eval and llvm // We can't do this because aliasing of memory can differ between const eval and llvm
throw_machine_stop_str!("pointer arithmetic or comparisons aren't supported in ConstProp") throw_machine_stop_str!("pointer arithmetic or comparisons aren't supported in ConstProp")
@ -237,7 +237,7 @@ impl<'mir, 'tcx> interpret::Machine<'mir, 'tcx> for ConstPropMachine<'mir, 'tcx>
fn box_alloc( fn box_alloc(
_ecx: &mut InterpCx<'mir, 'tcx, Self>, _ecx: &mut InterpCx<'mir, 'tcx, Self>,
_dest: PlaceTy<'tcx>, _dest: &PlaceTy<'tcx>,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
throw_machine_stop_str!("can't const prop heap allocations") throw_machine_stop_str!("can't const prop heap allocations")
} }
@ -392,12 +392,12 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
.filter(|ret_layout| { .filter(|ret_layout| {
!ret_layout.is_zst() && ret_layout.size < Size::from_bytes(MAX_ALLOC_LIMIT) !ret_layout.is_zst() && ret_layout.size < Size::from_bytes(MAX_ALLOC_LIMIT)
}) })
.map(|ret_layout| ecx.allocate(ret_layout, MemoryKind::Stack)); .map(|ret_layout| ecx.allocate(ret_layout, MemoryKind::Stack).into());
ecx.push_stack_frame( ecx.push_stack_frame(
Instance::new(def_id, substs), Instance::new(def_id, substs),
dummy_body, dummy_body,
ret.map(Into::into), ret.as_ref(),
StackPopCleanup::None { cleanup: false }, StackPopCleanup::None { cleanup: false },
) )
.expect("failed to push initial stack frame"); .expect("failed to push initial stack frame");
@ -426,7 +426,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
// Try to read the local as an immediate so that if it is representable as a scalar, we can // Try to read the local as an immediate so that if it is representable as a scalar, we can
// handle it as such, but otherwise, just return the value as is. // handle it as such, but otherwise, just return the value as is.
Some(match self.ecx.try_read_immediate(op) { Some(match self.ecx.try_read_immediate(&op) {
Ok(Ok(imm)) => imm.into(), Ok(Ok(imm)) => imm.into(),
_ => op, _ => op,
}) })
@ -548,8 +548,8 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
source_info: SourceInfo, source_info: SourceInfo,
) -> Option<()> { ) -> Option<()> {
if let (val, true) = self.use_ecx(|this| { if let (val, true) = self.use_ecx(|this| {
let val = this.ecx.read_immediate(this.ecx.eval_operand(arg, None)?)?; let val = this.ecx.read_immediate(&this.ecx.eval_operand(arg, None)?)?;
let (_res, overflow, _ty) = this.ecx.overflowing_unary_op(op, val)?; let (_res, overflow, _ty) = this.ecx.overflowing_unary_op(op, &val)?;
Ok((val, overflow)) Ok((val, overflow))
})? { })? {
// `AssertKind` only has an `OverflowNeg` variant, so make sure that is // `AssertKind` only has an `OverflowNeg` variant, so make sure that is
@ -573,8 +573,8 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
right: &Operand<'tcx>, right: &Operand<'tcx>,
source_info: SourceInfo, source_info: SourceInfo,
) -> Option<()> { ) -> Option<()> {
let r = self.use_ecx(|this| this.ecx.read_immediate(this.ecx.eval_operand(right, None)?)); let r = self.use_ecx(|this| this.ecx.read_immediate(&this.ecx.eval_operand(right, None)?));
let l = self.use_ecx(|this| this.ecx.read_immediate(this.ecx.eval_operand(left, None)?)); let l = self.use_ecx(|this| this.ecx.read_immediate(&this.ecx.eval_operand(left, None)?));
// Check for exceeding shifts *even if* we cannot evaluate the LHS. // Check for exceeding shifts *even if* we cannot evaluate the LHS.
if op == BinOp::Shr || op == BinOp::Shl { if op == BinOp::Shr || op == BinOp::Shl {
let r = r?; let r = r?;
@ -609,7 +609,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
} }
} }
if let (Some(l), Some(r)) = (l, r) { if let (Some(l), Some(r)) = (&l, &r) {
// The remaining operators are handled through `overflowing_binary_op`. // The remaining operators are handled through `overflowing_binary_op`.
if self.use_ecx(|this| { if self.use_ecx(|this| {
let (_res, overflow, _ty) = this.ecx.overflowing_binary_op(op, l, r)?; let (_res, overflow, _ty) = this.ecx.overflowing_binary_op(op, l, r)?;
@ -630,7 +630,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
match *operand { match *operand {
Operand::Copy(l) | Operand::Move(l) => { Operand::Copy(l) | Operand::Move(l) => {
if let Some(value) = self.get_const(l) { if let Some(value) = self.get_const(l) {
if self.should_const_prop(value) { if self.should_const_prop(&value) {
// FIXME(felix91gr): this code only handles `Scalar` cases. // FIXME(felix91gr): this code only handles `Scalar` cases.
// For now, we're not handling `ScalarPair` cases because // For now, we're not handling `ScalarPair` cases because
// doing so here would require a lot of code duplication. // doing so here would require a lot of code duplication.
@ -745,7 +745,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
let r = this.ecx.eval_operand(right, None); let r = this.ecx.eval_operand(right, None);
let const_arg = match (l, r) { let const_arg = match (l, r) {
(Ok(x), Err(_)) | (Err(_), Ok(x)) => this.ecx.read_immediate(x)?, (Ok(ref x), Err(_)) | (Err(_), Ok(ref x)) => this.ecx.read_immediate(x)?,
(Err(e), Err(_)) => return Err(e), (Err(e), Err(_)) => return Err(e),
(Ok(_), Ok(_)) => { (Ok(_), Ok(_)) => {
this.ecx.eval_rvalue_into_place(rvalue, place)?; this.ecx.eval_rvalue_into_place(rvalue, place)?;
@ -760,14 +760,14 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
match op { match op {
BinOp::BitAnd => { BinOp::BitAnd => {
if arg_value == 0 { if arg_value == 0 {
this.ecx.write_immediate(*const_arg, dest)?; this.ecx.write_immediate(*const_arg, &dest)?;
} }
} }
BinOp::BitOr => { BinOp::BitOr => {
if arg_value == const_arg.layout.size.truncate(u128::MAX) if arg_value == const_arg.layout.size.truncate(u128::MAX)
|| (const_arg.layout.ty.is_bool() && arg_value == 1) || (const_arg.layout.ty.is_bool() && arg_value == 1)
{ {
this.ecx.write_immediate(*const_arg, dest)?; this.ecx.write_immediate(*const_arg, &dest)?;
} }
} }
BinOp::Mul => { BinOp::Mul => {
@ -777,9 +777,9 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
const_arg.to_scalar()?.into(), const_arg.to_scalar()?.into(),
Scalar::from_bool(false).into(), Scalar::from_bool(false).into(),
); );
this.ecx.write_immediate(val, dest)?; this.ecx.write_immediate(val, &dest)?;
} else { } else {
this.ecx.write_immediate(*const_arg, dest)?; this.ecx.write_immediate(*const_arg, &dest)?;
} }
} }
} }
@ -809,7 +809,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
fn replace_with_const( fn replace_with_const(
&mut self, &mut self,
rval: &mut Rvalue<'tcx>, rval: &mut Rvalue<'tcx>,
value: OpTy<'tcx>, value: &OpTy<'tcx>,
source_info: SourceInfo, source_info: SourceInfo,
) { ) {
if let Rvalue::Use(Operand::Constant(c)) = rval { if let Rvalue::Use(Operand::Constant(c)) = rval {
@ -902,7 +902,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
} }
/// Returns `true` if and only if this `op` should be const-propagated into. /// Returns `true` if and only if this `op` should be const-propagated into.
fn should_const_prop(&mut self, op: OpTy<'tcx>) -> bool { fn should_const_prop(&mut self, op: &OpTy<'tcx>) -> bool {
let mir_opt_level = self.tcx.sess.opts.debugging_opts.mir_opt_level; let mir_opt_level = self.tcx.sess.opts.debugging_opts.mir_opt_level;
if mir_opt_level == 0 { if mir_opt_level == 0 {
@ -913,7 +913,7 @@ impl<'mir, 'tcx> ConstPropagator<'mir, 'tcx> {
return false; return false;
} }
match *op { match **op {
interpret::Operand::Immediate(Immediate::Scalar(ScalarMaybeUninit::Scalar(s))) => { interpret::Operand::Immediate(Immediate::Scalar(ScalarMaybeUninit::Scalar(s))) => {
s.is_bits() s.is_bits()
} }
@ -1094,7 +1094,7 @@ impl<'mir, 'tcx> MutVisitor<'tcx> for ConstPropagator<'mir, 'tcx> {
// This will return None if the above `const_prop` invocation only "wrote" a // This will return None if the above `const_prop` invocation only "wrote" a
// type whose creation requires no write. E.g. a generator whose initial state // type whose creation requires no write. E.g. a generator whose initial state
// consists solely of uninitialized memory (so it doesn't capture any locals). // consists solely of uninitialized memory (so it doesn't capture any locals).
if let Some(value) = self.get_const(place) { if let Some(ref value) = self.get_const(place) {
if self.should_const_prop(value) { if self.should_const_prop(value) {
trace!("replacing {:?} with {:?}", rval, value); trace!("replacing {:?} with {:?}", rval, value);
self.replace_with_const(rval, value, source_info); self.replace_with_const(rval, value, source_info);
@ -1177,10 +1177,10 @@ impl<'mir, 'tcx> MutVisitor<'tcx> for ConstPropagator<'mir, 'tcx> {
self.super_terminator(terminator, location); self.super_terminator(terminator, location);
match &mut terminator.kind { match &mut terminator.kind {
TerminatorKind::Assert { expected, ref msg, ref mut cond, .. } => { TerminatorKind::Assert { expected, ref msg, ref mut cond, .. } => {
if let Some(value) = self.eval_operand(&cond, source_info) { if let Some(ref value) = self.eval_operand(&cond, source_info) {
trace!("assertion on {:?} should be {:?}", value, expected); trace!("assertion on {:?} should be {:?}", value, expected);
let expected = ScalarMaybeUninit::from(Scalar::from_bool(*expected)); let expected = ScalarMaybeUninit::from(Scalar::from_bool(*expected));
let value_const = self.ecx.read_scalar(value).unwrap(); let value_const = self.ecx.read_scalar(&value).unwrap();
if expected != value_const { if expected != value_const {
enum DbgVal<T> { enum DbgVal<T> {
Val(T), Val(T),
@ -1198,9 +1198,9 @@ impl<'mir, 'tcx> MutVisitor<'tcx> for ConstPropagator<'mir, 'tcx> {
// This can be `None` if the lhs wasn't const propagated and we just // This can be `None` if the lhs wasn't const propagated and we just
// triggered the assert on the value of the rhs. // triggered the assert on the value of the rhs.
match self.eval_operand(op, source_info) { match self.eval_operand(op, source_info) {
Some(op) => { Some(op) => DbgVal::Val(
DbgVal::Val(self.ecx.read_immediate(op).unwrap().to_const_int()) self.ecx.read_immediate(&op).unwrap().to_const_int(),
} ),
None => DbgVal::Underscore, None => DbgVal::Underscore,
} }
}; };