use core::sync::atomic::Ordering::SeqCst; use core::sync::atomic::*; #[test] fn bool_() { let a = AtomicBool::new(false); assert_eq!(a.compare_exchange(false, true, SeqCst, SeqCst), Ok(false)); assert_eq!(a.compare_exchange(false, true, SeqCst, SeqCst), Err(true)); a.store(false, SeqCst); assert_eq!(a.compare_exchange(false, true, SeqCst, SeqCst), Ok(false)); } #[test] fn bool_and() { let a = AtomicBool::new(true); assert_eq!(a.fetch_and(false, SeqCst), true); assert_eq!(a.load(SeqCst), false); } #[test] fn bool_nand() { let a = AtomicBool::new(false); assert_eq!(a.fetch_nand(false, SeqCst), false); assert_eq!(a.load(SeqCst), true); assert_eq!(a.fetch_nand(false, SeqCst), true); assert_eq!(a.load(SeqCst), true); assert_eq!(a.fetch_nand(true, SeqCst), true); assert_eq!(a.load(SeqCst), false); assert_eq!(a.fetch_nand(true, SeqCst), false); assert_eq!(a.load(SeqCst), true); } #[test] fn uint_and() { let x = AtomicUsize::new(0xf731); assert_eq!(x.fetch_and(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731 & 0x137f); } #[test] fn uint_nand() { let x = AtomicUsize::new(0xf731); assert_eq!(x.fetch_nand(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), !(0xf731 & 0x137f)); } #[test] fn uint_or() { let x = AtomicUsize::new(0xf731); assert_eq!(x.fetch_or(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731 | 0x137f); } #[test] fn uint_xor() { let x = AtomicUsize::new(0xf731); assert_eq!(x.fetch_xor(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731 ^ 0x137f); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn uint_min() { let x = AtomicUsize::new(0xf731); assert_eq!(x.fetch_min(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0x137f); assert_eq!(x.fetch_min(0xf731, SeqCst), 0x137f); assert_eq!(x.load(SeqCst), 0x137f); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn uint_max() { let x = AtomicUsize::new(0x137f); assert_eq!(x.fetch_max(0xf731, SeqCst), 0x137f); assert_eq!(x.load(SeqCst), 0xf731); assert_eq!(x.fetch_max(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731); } #[test] fn int_and() { let x = AtomicIsize::new(0xf731); assert_eq!(x.fetch_and(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731 & 0x137f); } #[test] fn int_nand() { let x = AtomicIsize::new(0xf731); assert_eq!(x.fetch_nand(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), !(0xf731 & 0x137f)); } #[test] fn int_or() { let x = AtomicIsize::new(0xf731); assert_eq!(x.fetch_or(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731 | 0x137f); } #[test] fn int_xor() { let x = AtomicIsize::new(0xf731); assert_eq!(x.fetch_xor(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731 ^ 0x137f); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn int_min() { let x = AtomicIsize::new(0xf731); assert_eq!(x.fetch_min(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0x137f); assert_eq!(x.fetch_min(0xf731, SeqCst), 0x137f); assert_eq!(x.load(SeqCst), 0x137f); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn int_max() { let x = AtomicIsize::new(0x137f); assert_eq!(x.fetch_max(0xf731, SeqCst), 0x137f); assert_eq!(x.load(SeqCst), 0xf731); assert_eq!(x.fetch_max(0x137f, SeqCst), 0xf731); assert_eq!(x.load(SeqCst), 0xf731); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn ptr_add_null() { let atom = AtomicPtr::::new(core::ptr::null_mut()); assert_eq!(atom.fetch_ptr_add(1, SeqCst).addr(), 0); assert_eq!(atom.load(SeqCst).addr(), 8); assert_eq!(atom.fetch_byte_add(1, SeqCst).addr(), 8); assert_eq!(atom.load(SeqCst).addr(), 9); assert_eq!(atom.fetch_ptr_sub(1, SeqCst).addr(), 9); assert_eq!(atom.load(SeqCst).addr(), 1); assert_eq!(atom.fetch_byte_sub(1, SeqCst).addr(), 1); assert_eq!(atom.load(SeqCst).addr(), 0); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn ptr_add_data() { let num = 0i64; let n = &num as *const i64 as *mut _; let atom = AtomicPtr::::new(n); assert_eq!(atom.fetch_ptr_add(1, SeqCst), n); assert_eq!(atom.load(SeqCst), n.wrapping_add(1)); assert_eq!(atom.fetch_ptr_sub(1, SeqCst), n.wrapping_add(1)); assert_eq!(atom.load(SeqCst), n); let bytes_from_n = |b| n.wrapping_byte_add(b); assert_eq!(atom.fetch_byte_add(1, SeqCst), n); assert_eq!(atom.load(SeqCst), bytes_from_n(1)); assert_eq!(atom.fetch_byte_add(5, SeqCst), bytes_from_n(1)); assert_eq!(atom.load(SeqCst), bytes_from_n(6)); assert_eq!(atom.fetch_byte_sub(1, SeqCst), bytes_from_n(6)); assert_eq!(atom.load(SeqCst), bytes_from_n(5)); assert_eq!(atom.fetch_byte_sub(5, SeqCst), bytes_from_n(5)); assert_eq!(atom.load(SeqCst), n); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn ptr_bitops() { let atom = AtomicPtr::::new(core::ptr::null_mut()); assert_eq!(atom.fetch_or(0b0111, SeqCst).addr(), 0); assert_eq!(atom.load(SeqCst).addr(), 0b0111); assert_eq!(atom.fetch_and(0b1101, SeqCst).addr(), 0b0111); assert_eq!(atom.load(SeqCst).addr(), 0b0101); assert_eq!(atom.fetch_xor(0b1111, SeqCst).addr(), 0b0101); assert_eq!(atom.load(SeqCst).addr(), 0b1010); } #[test] #[cfg(any(not(target_arch = "arm"), target_os = "linux"))] // Missing intrinsic in compiler-builtins fn ptr_bitops_tagging() { #[repr(align(16))] struct Tagme(#[allow(dead_code)] u128); let tagme = Tagme(1000); let ptr = &tagme as *const Tagme as *mut Tagme; let atom: AtomicPtr = AtomicPtr::new(ptr); const MASK_TAG: usize = 0b1111; const MASK_PTR: usize = !MASK_TAG; assert_eq!(ptr.addr() & MASK_TAG, 0); assert_eq!(atom.fetch_or(0b0111, SeqCst), ptr); assert_eq!(atom.load(SeqCst), ptr.map_addr(|a| a | 0b111)); assert_eq!(atom.fetch_and(MASK_PTR | 0b0010, SeqCst), ptr.map_addr(|a| a | 0b111)); assert_eq!(atom.load(SeqCst), ptr.map_addr(|a| a | 0b0010)); assert_eq!(atom.fetch_xor(0b1011, SeqCst), ptr.map_addr(|a| a | 0b0010)); assert_eq!(atom.load(SeqCst), ptr.map_addr(|a| a | 0b1001)); assert_eq!(atom.fetch_and(MASK_PTR, SeqCst), ptr.map_addr(|a| a | 0b1001)); assert_eq!(atom.load(SeqCst), ptr); } static S_FALSE: AtomicBool = AtomicBool::new(false); static S_TRUE: AtomicBool = AtomicBool::new(true); static S_INT: AtomicIsize = AtomicIsize::new(0); static S_UINT: AtomicUsize = AtomicUsize::new(0); #[test] fn static_init() { // Note that we're not really testing the mutability here but it's important // on Android at the moment (#49775) assert!(!S_FALSE.swap(true, SeqCst)); assert!(S_TRUE.swap(false, SeqCst)); assert!(S_INT.fetch_add(1, SeqCst) == 0); assert!(S_UINT.fetch_add(1, SeqCst) == 0); } #[test] // FIXME(static_mut_refs): Do not allow `static_mut_refs` lint #[cfg_attr(not(bootstrap), allow(static_mut_refs))] fn atomic_access_bool() { static mut ATOMIC: AtomicBool = AtomicBool::new(false); unsafe { assert_eq!(*ATOMIC.get_mut(), false); ATOMIC.store(true, SeqCst); assert_eq!(*ATOMIC.get_mut(), true); ATOMIC.fetch_or(false, SeqCst); assert_eq!(*ATOMIC.get_mut(), true); ATOMIC.fetch_and(false, SeqCst); assert_eq!(*ATOMIC.get_mut(), false); ATOMIC.fetch_nand(true, SeqCst); assert_eq!(*ATOMIC.get_mut(), true); ATOMIC.fetch_xor(true, SeqCst); assert_eq!(*ATOMIC.get_mut(), false); } } #[test] fn atomic_alignment() { use std::mem::{align_of, size_of}; #[cfg(target_has_atomic = "8")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "ptr")] assert_eq!(align_of::>(), size_of::>()); #[cfg(target_has_atomic = "8")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "8")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "16")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "16")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "32")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "32")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "64")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "64")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "128")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "128")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "ptr")] assert_eq!(align_of::(), size_of::()); #[cfg(target_has_atomic = "ptr")] assert_eq!(align_of::(), size_of::()); } #[test] fn atomic_compare_exchange() { use Ordering::*; static ATOMIC: AtomicIsize = AtomicIsize::new(0); ATOMIC.compare_exchange(0, 1, Relaxed, Relaxed).ok(); ATOMIC.compare_exchange(0, 1, Acquire, Relaxed).ok(); ATOMIC.compare_exchange(0, 1, Release, Relaxed).ok(); ATOMIC.compare_exchange(0, 1, AcqRel, Relaxed).ok(); ATOMIC.compare_exchange(0, 1, SeqCst, Relaxed).ok(); ATOMIC.compare_exchange(0, 1, Acquire, Acquire).ok(); ATOMIC.compare_exchange(0, 1, AcqRel, Acquire).ok(); ATOMIC.compare_exchange(0, 1, SeqCst, Acquire).ok(); ATOMIC.compare_exchange(0, 1, SeqCst, SeqCst).ok(); ATOMIC.compare_exchange_weak(0, 1, Relaxed, Relaxed).ok(); ATOMIC.compare_exchange_weak(0, 1, Acquire, Relaxed).ok(); ATOMIC.compare_exchange_weak(0, 1, Release, Relaxed).ok(); ATOMIC.compare_exchange_weak(0, 1, AcqRel, Relaxed).ok(); ATOMIC.compare_exchange_weak(0, 1, SeqCst, Relaxed).ok(); ATOMIC.compare_exchange_weak(0, 1, Acquire, Acquire).ok(); ATOMIC.compare_exchange_weak(0, 1, AcqRel, Acquire).ok(); ATOMIC.compare_exchange_weak(0, 1, SeqCst, Acquire).ok(); ATOMIC.compare_exchange_weak(0, 1, SeqCst, SeqCst).ok(); } /* FIXME(#110395) #[test] fn atomic_const_from() { const _ATOMIC_U8: AtomicU8 = AtomicU8::from(1); const _ATOMIC_BOOL: AtomicBool = AtomicBool::from(true); const _ATOMIC_PTR: AtomicPtr = AtomicPtr::from(core::ptr::null_mut()); } */