Use proper atomic rmw for {mutex, rwlock, cond, srwlock}_get_or_create_id

This commit is contained in:
Andy Wang 2022-05-07 21:30:15 +01:00
parent f84a976a77
commit 1a7f6d504a
No known key found for this signature in database
GPG Key ID: 181B49F9F38F3374
4 changed files with 98 additions and 36 deletions

View File

@ -441,6 +441,23 @@ impl MemoryCellClocks {
/// Evaluation context extensions. /// Evaluation context extensions.
impl<'mir, 'tcx: 'mir> EvalContextExt<'mir, 'tcx> for MiriEvalContext<'mir, 'tcx> {} impl<'mir, 'tcx: 'mir> EvalContextExt<'mir, 'tcx> for MiriEvalContext<'mir, 'tcx> {}
pub trait EvalContextExt<'mir, 'tcx: 'mir>: MiriEvalContextExt<'mir, 'tcx> { pub trait EvalContextExt<'mir, 'tcx: 'mir>: MiriEvalContextExt<'mir, 'tcx> {
/// Calculates the MPlaceTy given the offset and layout of an access on an operand
fn offset_and_layout_to_place(
&self,
op: &OpTy<'tcx, Tag>,
offset: u64,
layout: TyAndLayout<'tcx>,
) -> InterpResult<'tcx, MPlaceTy<'tcx, Tag>> {
let this = self.eval_context_ref();
let op_place = this.deref_operand(op)?;
let offset = Size::from_bytes(offset);
// Ensure that the access is within bounds.
assert!(op_place.layout.size >= offset + layout.size);
let value_place = op_place.offset(offset, MemPlaceMeta::None, layout, this)?;
Ok(value_place)
}
/// Atomic variant of read_scalar_at_offset. /// Atomic variant of read_scalar_at_offset.
fn read_scalar_at_offset_atomic( fn read_scalar_at_offset_atomic(
&self, &self,
@ -450,12 +467,7 @@ pub trait EvalContextExt<'mir, 'tcx: 'mir>: MiriEvalContextExt<'mir, 'tcx> {
atomic: AtomicReadOp, atomic: AtomicReadOp,
) -> InterpResult<'tcx, ScalarMaybeUninit<Tag>> { ) -> InterpResult<'tcx, ScalarMaybeUninit<Tag>> {
let this = self.eval_context_ref(); let this = self.eval_context_ref();
let op_place = this.deref_operand(op)?; let value_place = this.offset_and_layout_to_place(op, offset, layout)?;
let offset = Size::from_bytes(offset);
// Ensure that the following read at an offset is within bounds.
assert!(op_place.layout.size >= offset + layout.size);
let value_place = op_place.offset(offset, MemPlaceMeta::None, layout, this)?;
this.read_scalar_atomic(&value_place, atomic) this.read_scalar_atomic(&value_place, atomic)
} }
@ -469,12 +481,7 @@ pub trait EvalContextExt<'mir, 'tcx: 'mir>: MiriEvalContextExt<'mir, 'tcx> {
atomic: AtomicWriteOp, atomic: AtomicWriteOp,
) -> InterpResult<'tcx> { ) -> InterpResult<'tcx> {
let this = self.eval_context_mut(); let this = self.eval_context_mut();
let op_place = this.deref_operand(op)?; let value_place = this.offset_and_layout_to_place(op, offset, layout)?;
let offset = Size::from_bytes(offset);
// Ensure that the following read at an offset is within bounds.
assert!(op_place.layout.size >= offset + layout.size);
let value_place = op_place.offset(offset, MemPlaceMeta::None, layout, this)?;
this.write_scalar_atomic(value.into(), &value_place, atomic) this.write_scalar_atomic(value.into(), &value_place, atomic)
} }

View File

@ -112,15 +112,23 @@ fn mutex_get_or_create_id<'mir, 'tcx: 'mir>(
ecx: &mut MiriEvalContext<'mir, 'tcx>, ecx: &mut MiriEvalContext<'mir, 'tcx>,
mutex_op: &OpTy<'tcx, Tag>, mutex_op: &OpTy<'tcx, Tag>,
) -> InterpResult<'tcx, MutexId> { ) -> InterpResult<'tcx, MutexId> {
let id = mutex_get_id(ecx, mutex_op)?.to_u32()?; let value_place = ecx.offset_and_layout_to_place(mutex_op, 4, ecx.machine.layouts.u32)?;
if id == 0 { let (old, success) = ecx
// 0 is a default value and also not a valid mutex id. Need to allocate .atomic_compare_exchange_scalar(
// a new mutex. &value_place,
&ImmTy::from_uint(0u32, ecx.machine.layouts.u32),
ecx.mutex_next_id().to_u32_scalar().into(),
AtomicRwOp::Relaxed,
AtomicReadOp::Relaxed,
false,
)?
.to_scalar_pair()?;
if success.to_bool().expect("compare_exchange's second return value is a bool") {
let id = ecx.mutex_create(); let id = ecx.mutex_create();
mutex_set_id(ecx, mutex_op, id.to_u32_scalar())?;
Ok(id) Ok(id)
} else { } else {
Ok(MutexId::from_u32(id)) Ok(MutexId::from_u32(old.to_u32().expect("layout is u32")))
} }
} }
@ -156,15 +164,23 @@ fn rwlock_get_or_create_id<'mir, 'tcx: 'mir>(
ecx: &mut MiriEvalContext<'mir, 'tcx>, ecx: &mut MiriEvalContext<'mir, 'tcx>,
rwlock_op: &OpTy<'tcx, Tag>, rwlock_op: &OpTy<'tcx, Tag>,
) -> InterpResult<'tcx, RwLockId> { ) -> InterpResult<'tcx, RwLockId> {
let id = rwlock_get_id(ecx, rwlock_op)?.to_u32()?; let value_place = ecx.offset_and_layout_to_place(rwlock_op, 4, ecx.machine.layouts.u32)?;
if id == 0 { let (old, success) = ecx
// 0 is a default value and also not a valid rwlock id. Need to allocate .atomic_compare_exchange_scalar(
// a new read-write lock. &value_place,
&ImmTy::from_uint(0u32, ecx.machine.layouts.u32),
ecx.rwlock_next_id().to_u32_scalar().into(),
AtomicRwOp::Relaxed,
AtomicReadOp::Relaxed,
false,
)?
.to_scalar_pair()?;
if success.to_bool().expect("compare_exchange's second return value is a bool") {
let id = ecx.rwlock_create(); let id = ecx.rwlock_create();
rwlock_set_id(ecx, rwlock_op, id.to_u32_scalar())?;
Ok(id) Ok(id)
} else { } else {
Ok(RwLockId::from_u32(id)) Ok(RwLockId::from_u32(old.to_u32().expect("layout is u32")))
} }
} }
@ -228,15 +244,24 @@ fn cond_get_or_create_id<'mir, 'tcx: 'mir>(
ecx: &mut MiriEvalContext<'mir, 'tcx>, ecx: &mut MiriEvalContext<'mir, 'tcx>,
cond_op: &OpTy<'tcx, Tag>, cond_op: &OpTy<'tcx, Tag>,
) -> InterpResult<'tcx, CondvarId> { ) -> InterpResult<'tcx, CondvarId> {
let id = cond_get_id(ecx, cond_op)?.to_u32()?; let value_place = ecx.offset_and_layout_to_place(cond_op, 4, ecx.machine.layouts.u32)?;
if id == 0 {
// 0 is a default value and also not a valid conditional variable id. let (old, success) = ecx
// Need to allocate a new id. .atomic_compare_exchange_scalar(
&value_place,
&ImmTy::from_uint(0u32, ecx.machine.layouts.u32),
ecx.condvar_next_id().to_u32_scalar().into(),
AtomicRwOp::Relaxed,
AtomicReadOp::Relaxed,
false,
)?
.to_scalar_pair()?;
if success.to_bool().expect("compare_exchange's second return value is a bool") {
let id = ecx.condvar_create(); let id = ecx.condvar_create();
cond_set_id(ecx, cond_op, id.to_u32_scalar())?;
Ok(id) Ok(id)
} else { } else {
Ok(CondvarId::from_u32(id)) Ok(CondvarId::from_u32(old.to_u32().expect("layout is u32")))
} }
} }

View File

@ -7,15 +7,24 @@ fn srwlock_get_or_create_id<'mir, 'tcx: 'mir>(
ecx: &mut MiriEvalContext<'mir, 'tcx>, ecx: &mut MiriEvalContext<'mir, 'tcx>,
lock_op: &OpTy<'tcx, Tag>, lock_op: &OpTy<'tcx, Tag>,
) -> InterpResult<'tcx, RwLockId> { ) -> InterpResult<'tcx, RwLockId> {
let id = ecx.read_scalar_at_offset(lock_op, 0, ecx.machine.layouts.u32)?.to_u32()?; let value_place = ecx.offset_and_layout_to_place(lock_op, 0, ecx.machine.layouts.u32)?;
if id == 0 {
// 0 is a default value and also not a valid rwlock id. Need to allocate let (old, success) = ecx
// a new rwlock. .atomic_compare_exchange_scalar(
&value_place,
&ImmTy::from_uint(0u32, ecx.machine.layouts.u32),
ecx.rwlock_next_id().to_u32_scalar().into(),
AtomicRwOp::AcqRel,
AtomicReadOp::Acquire,
false,
)?
.to_scalar_pair()?;
if success.to_bool().expect("compare_exchange's second return value is a bool") {
let id = ecx.rwlock_create(); let id = ecx.rwlock_create();
ecx.write_scalar_at_offset(lock_op, 0, id.to_u32_scalar(), ecx.machine.layouts.u32)?;
Ok(id) Ok(id)
} else { } else {
Ok(RwLockId::from_u32(id)) Ok(RwLockId::from_u32(old.to_u32().expect("layout is u32")))
} }
} }

View File

@ -208,6 +208,13 @@ trait EvalContextExtPriv<'mir, 'tcx: 'mir>: crate::MiriEvalContextExt<'mir, 'tcx
// situations. // situations.
impl<'mir, 'tcx: 'mir> EvalContextExt<'mir, 'tcx> for crate::MiriEvalContext<'mir, 'tcx> {} impl<'mir, 'tcx: 'mir> EvalContextExt<'mir, 'tcx> for crate::MiriEvalContext<'mir, 'tcx> {}
pub trait EvalContextExt<'mir, 'tcx: 'mir>: crate::MiriEvalContextExt<'mir, 'tcx> { pub trait EvalContextExt<'mir, 'tcx: 'mir>: crate::MiriEvalContextExt<'mir, 'tcx> {
#[inline]
/// Peek the id of the next mutex
fn mutex_next_id(&self) -> MutexId {
let this = self.eval_context_ref();
this.machine.threads.sync.mutexes.next_index()
}
#[inline] #[inline]
/// Create state for a new mutex. /// Create state for a new mutex.
fn mutex_create(&mut self) -> MutexId { fn mutex_create(&mut self) -> MutexId {
@ -290,6 +297,13 @@ pub trait EvalContextExt<'mir, 'tcx: 'mir>: crate::MiriEvalContextExt<'mir, 'tcx
this.block_thread(thread); this.block_thread(thread);
} }
#[inline]
/// Peek the id of the next read write lock
fn rwlock_next_id(&self) -> RwLockId {
let this = self.eval_context_ref();
this.machine.threads.sync.rwlocks.next_index()
}
#[inline] #[inline]
/// Create state for a new read write lock. /// Create state for a new read write lock.
fn rwlock_create(&mut self) -> RwLockId { fn rwlock_create(&mut self) -> RwLockId {
@ -438,6 +452,13 @@ pub trait EvalContextExt<'mir, 'tcx: 'mir>: crate::MiriEvalContextExt<'mir, 'tcx
this.block_thread(writer); this.block_thread(writer);
} }
#[inline]
/// Peek the id of the next Condvar
fn condvar_next_id(&self) -> CondvarId {
let this = self.eval_context_ref();
this.machine.threads.sync.condvars.next_index()
}
#[inline] #[inline]
/// Create state for a new conditional variable. /// Create state for a new conditional variable.
fn condvar_create(&mut self) -> CondvarId { fn condvar_create(&mut self) -> CondvarId {