2013-02-20 10:57:15 -06:00
|
|
|
// Copyright 2013 The Rust Project Developers. See the COPYRIGHT
|
|
|
|
// file at the top-level directory of this distribution and at
|
|
|
|
// http://rust-lang.org/COPYRIGHT.
|
|
|
|
//
|
|
|
|
// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
|
|
|
|
// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
|
|
|
|
// option. This file may not be copied, modified, or distributed
|
|
|
|
// except according to those terms.
|
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/*! rustc compiler intrinsics.
|
|
|
|
|
|
|
|
The corresponding definitions are in librustc/middle/trans/foreign.rs.
|
|
|
|
|
|
|
|
# Atomics
|
|
|
|
|
|
|
|
The atomic intrinsics provide common atomic operations on machine
|
|
|
|
words, with multiple possible memory orderings. They obey the same
|
2013-05-24 16:55:49 -05:00
|
|
|
semantics as C++11. See the LLVM documentation on [[atomics]].
|
2013-05-16 02:04:24 -05:00
|
|
|
|
|
|
|
[atomics]: http://llvm.org/docs/Atomics.html
|
|
|
|
|
|
|
|
A quick refresher on memory ordering:
|
|
|
|
|
|
|
|
* Acquire - a barrier for aquiring a lock. Subsequent reads and writes
|
|
|
|
take place after the barrier.
|
|
|
|
* Release - a barrier for releasing a lock. Preceding reads and writes
|
|
|
|
take place before the barrier.
|
|
|
|
* Sequentially consistent - sequentially consistent operations are
|
|
|
|
guaranteed to happen in order. This is the standard mode for working
|
|
|
|
with atomic types and is equivalent to Java's `volatile`.
|
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
*/
|
2013-05-24 17:05:27 -05:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
#[abi = "rust-intrinsic"]
|
2013-03-13 21:25:28 -05:00
|
|
|
pub extern "rust-intrinsic" {
|
2013-05-16 02:04:24 -05:00
|
|
|
|
|
|
|
/// Atomic compare and exchange, sequentially consistent.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_cxchg(dst: &mut int, old: int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic compare and exchange, acquire ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_cxchg_acq(dst: &mut int, old: int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic compare and exchange, release ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_cxchg_rel(dst: &mut int, old: int, src: int) -> int;
|
2013-05-13 06:33:34 -05:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic load, sequentially consistent.
|
2013-05-12 14:22:20 -05:00
|
|
|
pub fn atomic_load(src: &int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic load, acquire ordering.
|
2013-05-12 14:22:20 -05:00
|
|
|
pub fn atomic_load_acq(src: &int) -> int;
|
2013-05-13 06:33:34 -05:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic store, sequentially consistent.
|
2013-05-12 14:22:20 -05:00
|
|
|
pub fn atomic_store(dst: &mut int, val: int);
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic store, release ordering.
|
2013-05-12 14:22:20 -05:00
|
|
|
pub fn atomic_store_rel(dst: &mut int, val: int);
|
2013-02-20 10:57:15 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic exchange, sequentially consistent.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xchg(dst: &mut int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic exchange, acquire ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xchg_acq(dst: &mut int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic exchange, release ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xchg_rel(dst: &mut int, src: int) -> int;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic addition, sequentially consistent.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xadd(dst: &mut int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic addition, acquire ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xadd_acq(dst: &mut int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic addition, release ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xadd_rel(dst: &mut int, src: int) -> int;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic subtraction, sequentially consistent.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xsub(dst: &mut int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic subtraction, acquire ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xsub_acq(dst: &mut int, src: int) -> int;
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Atomic subtraction, release ordering.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn atomic_xsub_rel(dst: &mut int, src: int) -> int;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// The size of a type in bytes.
|
|
|
|
///
|
|
|
|
/// This is the exact number of bytes in memory taken up by a
|
|
|
|
/// value of the given type. In other words, a memset of this size
|
|
|
|
/// would *exactly* overwrite a value. When laid out in vectors
|
|
|
|
/// and structures there may be additional padding between
|
|
|
|
/// elements.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn size_of<T>() -> uint;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Move a value to a memory location containing a value.
|
|
|
|
///
|
|
|
|
/// Drop glue is run on the destination, which must contain a
|
|
|
|
/// valid Rust value.
|
2013-05-07 16:20:56 -05:00
|
|
|
pub fn move_val<T>(dst: &mut T, src: T);
|
2013-05-16 02:04:24 -05:00
|
|
|
|
|
|
|
/// Move a value to an uninitialized memory location.
|
|
|
|
///
|
|
|
|
/// Drop glue is not run on the destination.
|
2013-05-07 16:20:56 -05:00
|
|
|
pub fn move_val_init<T>(dst: &mut T, src: T);
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn min_align_of<T>() -> uint;
|
|
|
|
pub fn pref_align_of<T>() -> uint;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Get a static pointer to a type descriptor.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn get_tydesc<T>() -> *();
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Create a value initialized to zero.
|
|
|
|
///
|
|
|
|
/// `init` is unsafe because it returns a zeroed-out datum,
|
2013-05-09 13:44:52 -05:00
|
|
|
/// which is unsafe unless T is POD. We don't have a POD
|
2013-05-16 02:04:24 -05:00
|
|
|
/// kind yet. (See #4074).
|
2013-05-09 13:44:52 -05:00
|
|
|
pub unsafe fn init<T>() -> T;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Create an uninitialized value.
|
2013-05-09 05:23:38 -05:00
|
|
|
pub unsafe fn uninit<T>() -> T;
|
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Move a value out of scope without running drop glue.
|
|
|
|
///
|
|
|
|
/// `forget` is unsafe because the caller is responsible for
|
|
|
|
/// ensuring the argument is deallocated already.
|
2013-05-09 14:15:37 -05:00
|
|
|
pub unsafe fn forget<T>(_: T) -> ();
|
2013-05-09 14:49:14 -05:00
|
|
|
pub fn transmute<T,U>(e: T) -> U;
|
2013-02-20 10:57:15 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Returns `true` if a type requires drop glue.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn needs_drop<T>() -> bool;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
|
|
|
// XXX: intrinsic uses legacy modes and has reference to TyDesc
|
2013-02-20 10:57:15 -06:00
|
|
|
// and TyVisitor which are in librustc
|
|
|
|
//fn visit_tydesc(++td: *TyDesc, &&tv: TyVisitor) -> ();
|
2013-05-09 14:49:14 -05:00
|
|
|
|
|
|
|
pub fn frame_address(f: &once fn(*u8));
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Get the address of the `__morestack` stack growth function.
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn morestack_addr() -> *();
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-24 12:58:55 -05:00
|
|
|
/// Equivalent to the `llvm.memcpy.p0i8.0i8.i32` intrinsic, with a size of
|
|
|
|
/// `count` * `size_of::<T>()` and an alignment of `min_align_of::<T>()`
|
2013-05-23 21:15:31 -05:00
|
|
|
#[cfg(not(stage0))]
|
2013-05-24 12:58:55 -05:00
|
|
|
pub fn memcpy32<T>(dst: *mut T, src: *T, count: u32);
|
|
|
|
/// Equivalent to the `llvm.memcpy.p0i8.0i8.i64` intrinsic, with a size of
|
|
|
|
/// `count` * `size_of::<T>()` and an alignment of `min_align_of::<T>()`
|
2013-05-23 21:15:31 -05:00
|
|
|
#[cfg(not(stage0))]
|
2013-05-24 12:58:55 -05:00
|
|
|
pub fn memcpy64<T>(dst: *mut T, src: *T, count: u64);
|
2013-05-23 21:15:31 -05:00
|
|
|
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Equivalent to the `llvm.memmove.p0i8.0i8.i32` intrinsic.
|
2013-05-24 12:58:55 -05:00
|
|
|
#[cfg(stage0)]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn memmove32(dst: *mut u8, src: *u8, size: u32);
|
2013-05-16 02:04:24 -05:00
|
|
|
/// Equivalent to the `llvm.memmove.p0i8.0i8.i64` intrinsic.
|
2013-05-24 12:58:55 -05:00
|
|
|
#[cfg(stage0)]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn memmove64(dst: *mut u8, src: *u8, size: u64);
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-05-24 12:58:55 -05:00
|
|
|
/// Equivalent to the `llvm.memmove.p0i8.0i8.i32` intrinsic, with a size of
|
|
|
|
/// `count` * `size_of::<T>()` and an alignment of `min_align_of::<T>()`
|
|
|
|
#[cfg(not(stage0))]
|
|
|
|
pub fn memmove32<T>(dst: *mut T, src: *T, count: u32);
|
|
|
|
/// Equivalent to the `llvm.memmove.p0i8.0i8.i64` intrinsic, with a size of
|
|
|
|
/// `count` * `size_of::<T>()` and an alignment of `min_align_of::<T>()`
|
|
|
|
#[cfg(not(stage0))]
|
|
|
|
pub fn memmove64<T>(dst: *mut T, src: *T, count: u64);
|
|
|
|
|
2013-05-24 17:05:27 -05:00
|
|
|
/// Equivalent to the `llvm.memset.p0i8.i32` intrinsic, with a size of
|
|
|
|
/// `count` * `size_of::<T>()` and an alignment of `min_align_of::<T>()`
|
|
|
|
#[cfg(not(stage0))]
|
|
|
|
pub fn memset32<T>(dst: *mut T, val: u8, count: u32);
|
|
|
|
/// Equivalent to the `llvm.memset.p0i8.i64` intrinsic, with a size of
|
|
|
|
/// `count` * `size_of::<T>()` and an alignment of `min_align_of::<T>()`
|
|
|
|
#[cfg(not(stage0))]
|
|
|
|
pub fn memset64<T>(dst: *mut T, val: u8, count: u64);
|
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn sqrtf32(x: f32) -> f32;
|
|
|
|
pub fn sqrtf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn powif32(a: f32, x: i32) -> f32;
|
|
|
|
pub fn powif64(a: f64, x: i32) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-04-20 06:58:56 -05:00
|
|
|
// the following kill the stack canary without
|
|
|
|
// `fixed_stack_segment`. This possibly only affects the f64
|
|
|
|
// variants, but it's hard to be sure since it seems to only
|
|
|
|
// occur with fairly specific arguments.
|
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn sinf32(x: f32) -> f32;
|
2013-04-20 06:58:56 -05:00
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn sinf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-04-20 06:58:56 -05:00
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn cosf32(x: f32) -> f32;
|
2013-04-20 06:58:56 -05:00
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn cosf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-04-20 06:58:56 -05:00
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn powf32(a: f32, x: f32) -> f32;
|
2013-04-20 06:58:56 -05:00
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn powf64(a: f64, x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-04-20 06:58:56 -05:00
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn expf32(x: f32) -> f32;
|
2013-04-20 06:58:56 -05:00
|
|
|
#[fixed_stack_segment]
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn expf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn exp2f32(x: f32) -> f32;
|
|
|
|
pub fn exp2f64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn logf32(x: f32) -> f32;
|
|
|
|
pub fn logf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn log10f32(x: f32) -> f32;
|
|
|
|
pub fn log10f64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn log2f32(x: f32) -> f32;
|
|
|
|
pub fn log2f64(x: f64) -> f64;
|
|
|
|
|
|
|
|
pub fn fmaf32(a: f32, b: f32, c: f32) -> f32;
|
|
|
|
pub fn fmaf64(a: f64, b: f64, c: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn fabsf32(x: f32) -> f32;
|
|
|
|
pub fn fabsf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn floorf32(x: f32) -> f32;
|
|
|
|
pub fn floorf64(x: f64) -> f64;
|
|
|
|
|
|
|
|
pub fn ceilf32(x: f32) -> f32;
|
|
|
|
pub fn ceilf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn truncf32(x: f32) -> f32;
|
|
|
|
pub fn truncf64(x: f64) -> f64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn ctpop8(x: i8) -> i8;
|
|
|
|
pub fn ctpop16(x: i16) -> i16;
|
|
|
|
pub fn ctpop32(x: i32) -> i32;
|
|
|
|
pub fn ctpop64(x: i64) -> i64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn ctlz8(x: i8) -> i8;
|
|
|
|
pub fn ctlz16(x: i16) -> i16;
|
|
|
|
pub fn ctlz32(x: i32) -> i32;
|
|
|
|
pub fn ctlz64(x: i64) -> i64;
|
|
|
|
|
|
|
|
pub fn cttz8(x: i8) -> i8;
|
|
|
|
pub fn cttz16(x: i16) -> i16;
|
|
|
|
pub fn cttz32(x: i32) -> i32;
|
|
|
|
pub fn cttz64(x: i64) -> i64;
|
2013-02-20 13:41:24 -06:00
|
|
|
|
2013-02-20 10:57:15 -06:00
|
|
|
pub fn bswap16(x: i16) -> i16;
|
|
|
|
pub fn bswap32(x: i32) -> i32;
|
|
|
|
pub fn bswap64(x: i64) -> i64;
|
|
|
|
}
|