/rust/registry/src/index.crates.io-6f17d22bba15001f/crossbeam-utils-0.8.19/src/atomic/consume.rs
Line | Count | Source (jump to first uncovered line) |
1 | | #[cfg(not(crossbeam_no_atomic))] |
2 | | use core::sync::atomic::Ordering; |
3 | | |
4 | | /// Trait which allows reading from primitive atomic types with "consume" ordering. |
5 | | pub trait AtomicConsume { |
6 | | /// Type returned by `load_consume`. |
7 | | type Val; |
8 | | |
9 | | /// Loads a value from the atomic using a "consume" memory ordering. |
10 | | /// |
11 | | /// This is similar to the "acquire" ordering, except that an ordering is |
12 | | /// only guaranteed with operations that "depend on" the result of the load. |
13 | | /// However consume loads are usually much faster than acquire loads on |
14 | | /// architectures with a weak memory model since they don't require memory |
15 | | /// fence instructions. |
16 | | /// |
17 | | /// The exact definition of "depend on" is a bit vague, but it works as you |
18 | | /// would expect in practice since a lot of software, especially the Linux |
19 | | /// kernel, rely on this behavior. |
20 | | /// |
21 | | /// This is currently only implemented on ARM and AArch64, where a fence |
22 | | /// can be avoided. On other architectures this will fall back to a simple |
23 | | /// `load(Ordering::Acquire)`. |
24 | | fn load_consume(&self) -> Self::Val; |
25 | | } |
26 | | |
27 | | #[cfg(not(crossbeam_no_atomic))] |
28 | | // Miri and Loom don't support "consume" ordering and ThreadSanitizer doesn't treat |
29 | | // load(Relaxed) + compiler_fence(Acquire) as "consume" load. |
30 | | // LLVM generates machine code equivalent to fence(Acquire) in compiler_fence(Acquire) |
31 | | // on PowerPC, MIPS, etc. (https://godbolt.org/z/hffvjvW7h), so for now the fence |
32 | | // can be actually avoided here only on ARM and AArch64. See also |
33 | | // https://github.com/rust-lang/rust/issues/62256. |
34 | | #[cfg(all( |
35 | | any(target_arch = "arm", target_arch = "aarch64"), |
36 | | not(any(miri, crossbeam_loom, crossbeam_sanitize_thread)), |
37 | | ))] |
38 | | macro_rules! impl_consume { |
39 | | () => { |
40 | | #[inline] |
41 | | fn load_consume(&self) -> Self::Val { |
42 | | use crate::primitive::sync::atomic::compiler_fence; |
43 | | let result = self.load(Ordering::Relaxed); |
44 | | compiler_fence(Ordering::Acquire); |
45 | | result |
46 | | } |
47 | | }; |
48 | | } |
49 | | |
50 | | #[cfg(not(crossbeam_no_atomic))] |
51 | | #[cfg(not(all( |
52 | | any(target_arch = "arm", target_arch = "aarch64"), |
53 | | not(any(miri, crossbeam_loom, crossbeam_sanitize_thread)), |
54 | | )))] |
55 | | macro_rules! impl_consume { |
56 | | () => { |
57 | | #[inline] |
58 | 0 | fn load_consume(&self) -> Self::Val { |
59 | 0 | self.load(Ordering::Acquire) |
60 | 0 | } Unexecuted instantiation: <core::sync::atomic::AtomicBool as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicUsize as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicIsize as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicU8 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicI8 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicU16 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicI16 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicU32 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicI32 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicU64 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicI64 as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume Unexecuted instantiation: <core::sync::atomic::AtomicPtr<_> as crossbeam_utils::atomic::consume::AtomicConsume>::load_consume |
61 | | }; |
62 | | } |
63 | | |
64 | | macro_rules! impl_atomic { |
65 | | ($atomic:ident, $val:ty) => { |
66 | | #[cfg(not(crossbeam_no_atomic))] |
67 | | impl AtomicConsume for core::sync::atomic::$atomic { |
68 | | type Val = $val; |
69 | | impl_consume!(); |
70 | | } |
71 | | #[cfg(crossbeam_loom)] |
72 | | impl AtomicConsume for loom::sync::atomic::$atomic { |
73 | | type Val = $val; |
74 | | impl_consume!(); |
75 | | } |
76 | | }; |
77 | | } |
78 | | |
79 | | impl_atomic!(AtomicBool, bool); |
80 | | impl_atomic!(AtomicUsize, usize); |
81 | | impl_atomic!(AtomicIsize, isize); |
82 | | impl_atomic!(AtomicU8, u8); |
83 | | impl_atomic!(AtomicI8, i8); |
84 | | impl_atomic!(AtomicU16, u16); |
85 | | impl_atomic!(AtomicI16, i16); |
86 | | #[cfg(any(target_has_atomic = "32", not(target_pointer_width = "16")))] |
87 | | impl_atomic!(AtomicU32, u32); |
88 | | #[cfg(any(target_has_atomic = "32", not(target_pointer_width = "16")))] |
89 | | impl_atomic!(AtomicI32, i32); |
90 | | #[cfg(any( |
91 | | target_has_atomic = "64", |
92 | | not(any(target_pointer_width = "16", target_pointer_width = "32")), |
93 | | ))] |
94 | | impl_atomic!(AtomicU64, u64); |
95 | | #[cfg(any( |
96 | | target_has_atomic = "64", |
97 | | not(any(target_pointer_width = "16", target_pointer_width = "32")), |
98 | | ))] |
99 | | impl_atomic!(AtomicI64, i64); |
100 | | |
101 | | #[cfg(not(crossbeam_no_atomic))] |
102 | | impl<T> AtomicConsume for core::sync::atomic::AtomicPtr<T> { |
103 | | type Val = *mut T; |
104 | | impl_consume!(); |
105 | | } |
106 | | |
107 | | #[cfg(crossbeam_loom)] |
108 | | impl<T> AtomicConsume for loom::sync::atomic::AtomicPtr<T> { |
109 | | type Val = *mut T; |
110 | | impl_consume!(); |
111 | | } |