2022-06-26 20:14:18 -05:00
|
|
|
// compile-flags: -Zmiri-permissive-provenance
|
|
|
|
|
2022-06-27 07:13:55 -05:00
|
|
|
use std::mem;
|
|
|
|
|
2022-06-06 10:44:36 -05:00
|
|
|
// This strips provenance
|
2019-07-03 13:42:01 -05:00
|
|
|
fn transmute_ptr_to_int<T>(x: *const T) -> usize {
|
|
|
|
unsafe { std::mem::transmute(x) }
|
2019-06-29 13:17:15 -05:00
|
|
|
}
|
2019-06-22 09:25:16 -05:00
|
|
|
|
2019-07-23 16:25:06 -05:00
|
|
|
fn cast() {
|
2019-06-28 13:42:00 -05:00
|
|
|
// Some casting-to-int with arithmetic.
|
2019-07-03 13:42:01 -05:00
|
|
|
let x = &42 as *const i32 as usize;
|
2019-06-22 09:25:16 -05:00
|
|
|
let y = x * 2;
|
2019-06-24 21:47:37 -05:00
|
|
|
assert_eq!(y, x + x);
|
2019-06-22 09:25:16 -05:00
|
|
|
let z = y as u8 as usize;
|
|
|
|
assert_eq!(z, y % 256);
|
2019-07-23 16:25:06 -05:00
|
|
|
}
|
2019-06-28 13:42:00 -05:00
|
|
|
|
2020-07-11 13:56:47 -05:00
|
|
|
/// Test usize->ptr cast for dangling and OOB address.
|
|
|
|
/// That is safe, and thus has to work.
|
|
|
|
fn cast_dangling() {
|
|
|
|
let b = Box::new(0);
|
|
|
|
let x = &*b as *const i32 as usize;
|
|
|
|
drop(b);
|
|
|
|
let _val = x as *const i32;
|
|
|
|
|
|
|
|
let b = Box::new(0);
|
|
|
|
let mut x = &*b as *const i32 as usize;
|
|
|
|
x += 0x100;
|
|
|
|
let _val = x as *const i32;
|
|
|
|
}
|
|
|
|
|
2019-07-23 16:25:06 -05:00
|
|
|
fn format() {
|
2019-06-28 13:42:00 -05:00
|
|
|
// Pointer string formatting! We can't check the output as it changes when libstd changes,
|
|
|
|
// but we can make sure Miri does not error.
|
|
|
|
format!("{:?}", &mut 13 as *mut _);
|
2019-07-23 16:25:06 -05:00
|
|
|
}
|
2019-06-29 13:17:15 -05:00
|
|
|
|
2019-07-23 16:25:06 -05:00
|
|
|
fn transmute() {
|
2019-06-29 13:17:15 -05:00
|
|
|
// Check that intptrcast is triggered for explicit casts and that it is consistent with
|
|
|
|
// transmuting.
|
|
|
|
let a: *const i32 = &42;
|
|
|
|
let b = transmute_ptr_to_int(a) as u8;
|
2022-06-06 10:44:36 -05:00
|
|
|
let c = a as u8;
|
2019-06-29 13:17:15 -05:00
|
|
|
assert_eq!(b, c);
|
2019-06-22 09:25:16 -05:00
|
|
|
}
|
2019-07-23 16:25:06 -05:00
|
|
|
|
|
|
|
fn ptr_bitops1() {
|
|
|
|
let bytes = [0i8, 1, 2, 3, 4, 5, 6, 7, 8, 9];
|
|
|
|
let one = bytes.as_ptr().wrapping_offset(1);
|
|
|
|
let three = bytes.as_ptr().wrapping_offset(3);
|
|
|
|
let res = (one as usize) | (three as usize);
|
|
|
|
format!("{}", res);
|
|
|
|
}
|
|
|
|
|
|
|
|
fn ptr_bitops2() {
|
|
|
|
let val = 13usize;
|
|
|
|
let addr = &val as *const _ as usize;
|
|
|
|
let _val = addr & 13;
|
|
|
|
}
|
|
|
|
|
|
|
|
fn ptr_eq_dangling() {
|
|
|
|
let b = Box::new(0);
|
|
|
|
let x = &*b as *const i32; // soon-to-be dangling
|
|
|
|
drop(b);
|
|
|
|
let b = Box::new(0);
|
|
|
|
let y = &*b as *const i32; // different allocation
|
2020-07-11 13:56:47 -05:00
|
|
|
// They *could* be equal if memory was reused, but probably are not.
|
2019-07-23 16:25:06 -05:00
|
|
|
assert!(x != y);
|
|
|
|
}
|
|
|
|
|
|
|
|
fn ptr_eq_out_of_bounds() {
|
|
|
|
let b = Box::new(0);
|
|
|
|
let x = (&*b as *const i32).wrapping_sub(0x800); // out-of-bounds
|
|
|
|
let b = Box::new(0);
|
|
|
|
let y = &*b as *const i32; // different allocation
|
2020-07-11 13:56:47 -05:00
|
|
|
// They *could* be equal (with the right base addresses), but probably are not.
|
2019-07-23 16:25:06 -05:00
|
|
|
assert!(x != y);
|
|
|
|
}
|
|
|
|
|
|
|
|
fn ptr_eq_out_of_bounds_null() {
|
|
|
|
let b = Box::new(0);
|
|
|
|
let x = (&*b as *const i32).wrapping_sub(0x800); // out-of-bounds
|
2020-07-11 13:56:47 -05:00
|
|
|
// This *could* be NULL (with the right base address), but probably is not.
|
2019-07-23 16:25:06 -05:00
|
|
|
assert!(x != std::ptr::null());
|
|
|
|
}
|
|
|
|
|
|
|
|
fn ptr_eq_integer() {
|
|
|
|
let b = Box::new(0);
|
|
|
|
let x = &*b as *const i32;
|
2020-07-11 13:56:47 -05:00
|
|
|
// These *could* be equal (with the right base address), but probably are not.
|
2019-07-23 16:25:06 -05:00
|
|
|
assert!(x != 64 as *const i32);
|
|
|
|
}
|
|
|
|
|
2022-06-26 20:26:14 -05:00
|
|
|
fn zst_deref_of_dangling() {
|
|
|
|
let b = Box::new(0);
|
|
|
|
let addr = &*b as *const _ as usize;
|
|
|
|
drop(b);
|
|
|
|
// Now if we cast `addr` to a ptr it might pick up the dangling provenance.
|
|
|
|
// But if we only do a ZST deref there is no UB here!
|
|
|
|
let zst = addr as *const ();
|
|
|
|
let _val = unsafe { *zst };
|
|
|
|
}
|
|
|
|
|
2022-06-27 07:13:55 -05:00
|
|
|
fn functions() {
|
|
|
|
// Roundtrip a few functions through integers. Do this multiple times to make sure this does not
|
|
|
|
// work by chance. If we did not give unique addresses to ZST allocations -- which fn
|
|
|
|
// allocations are -- then we might be unable to cast back, or we might call the wrong function!
|
|
|
|
// Every function gets at most one address so doing a loop would not help...
|
|
|
|
fn fn0() -> i32 {
|
|
|
|
0
|
|
|
|
}
|
|
|
|
fn fn1() -> i32 {
|
|
|
|
1
|
|
|
|
}
|
|
|
|
fn fn2() -> i32 {
|
|
|
|
2
|
|
|
|
}
|
|
|
|
fn fn3() -> i32 {
|
|
|
|
3
|
|
|
|
}
|
|
|
|
fn fn4() -> i32 {
|
|
|
|
4
|
|
|
|
}
|
|
|
|
fn fn5() -> i32 {
|
|
|
|
5
|
|
|
|
}
|
|
|
|
fn fn6() -> i32 {
|
|
|
|
6
|
|
|
|
}
|
|
|
|
fn fn7() -> i32 {
|
|
|
|
7
|
|
|
|
}
|
|
|
|
let fns = [
|
|
|
|
fn0 as fn() -> i32 as *const () as usize,
|
|
|
|
fn1 as fn() -> i32 as *const () as usize,
|
|
|
|
fn2 as fn() -> i32 as *const () as usize,
|
|
|
|
fn3 as fn() -> i32 as *const () as usize,
|
|
|
|
fn4 as fn() -> i32 as *const () as usize,
|
|
|
|
fn5 as fn() -> i32 as *const () as usize,
|
|
|
|
fn6 as fn() -> i32 as *const () as usize,
|
|
|
|
fn7 as fn() -> i32 as *const () as usize,
|
|
|
|
];
|
|
|
|
for (idx, &addr) in fns.iter().enumerate() {
|
|
|
|
let fun: fn() -> i32 = unsafe { mem::transmute(addr as *const ()) };
|
|
|
|
assert_eq!(fun(), idx as i32);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-07-23 16:25:06 -05:00
|
|
|
fn main() {
|
|
|
|
cast();
|
2020-07-11 13:56:47 -05:00
|
|
|
cast_dangling();
|
2019-07-23 16:25:06 -05:00
|
|
|
format();
|
|
|
|
transmute();
|
|
|
|
ptr_bitops1();
|
|
|
|
ptr_bitops2();
|
|
|
|
ptr_eq_dangling();
|
|
|
|
ptr_eq_out_of_bounds();
|
|
|
|
ptr_eq_out_of_bounds_null();
|
|
|
|
ptr_eq_integer();
|
2022-06-26 20:26:14 -05:00
|
|
|
zst_deref_of_dangling();
|
2022-06-27 07:13:55 -05:00
|
|
|
functions();
|
2019-07-23 16:25:06 -05:00
|
|
|
}
|