Run rustfmt over everything
This commit is contained in:
parent
ad1bcd2592
commit
c88c9502b7
101
build.rs
101
build.rs
|
@ -22,8 +22,9 @@ fn main() {
|
|||
|
||||
// Forcibly enable memory intrinsics on wasm32 & SGX as we don't have a libc to
|
||||
// provide them.
|
||||
if (target.contains("wasm32") && !target.contains("wasi")) ||
|
||||
(target.contains("sgx") && target.contains("fortanix")) {
|
||||
if (target.contains("wasm32") && !target.contains("wasi"))
|
||||
|| (target.contains("sgx") && target.contains("fortanix"))
|
||||
{
|
||||
println!("cargo:rustc-cfg=feature=\"mem\"");
|
||||
}
|
||||
|
||||
|
@ -85,7 +86,9 @@ mod c {
|
|||
|
||||
impl Sources {
|
||||
fn new() -> Sources {
|
||||
Sources { map: BTreeMap::new() }
|
||||
Sources {
|
||||
map: BTreeMap::new(),
|
||||
}
|
||||
}
|
||||
|
||||
fn extend(&mut self, sources: &[&'static str]) {
|
||||
|
@ -151,8 +154,7 @@ mod c {
|
|||
}
|
||||
|
||||
let mut sources = Sources::new();
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"absvdi2.c",
|
||||
"absvsi2.c",
|
||||
"addvdi3.c",
|
||||
|
@ -189,24 +191,20 @@ mod c {
|
|||
"truncdfsf2.c",
|
||||
"truncsfhf2.c",
|
||||
"ucmpdi2.c",
|
||||
],
|
||||
);
|
||||
]);
|
||||
|
||||
// When compiling in rustbuild (the rust-lang/rust repo) this library
|
||||
// also needs to satisfy intrinsics that jemalloc or C in general may
|
||||
// need, so include a few more that aren't typically needed by
|
||||
// LLVM/Rust.
|
||||
if cfg!(feature = "rustbuild") {
|
||||
sources.extend(&[
|
||||
"ffsdi2.c",
|
||||
]);
|
||||
sources.extend(&["ffsdi2.c"]);
|
||||
}
|
||||
|
||||
// On iOS and 32-bit OSX these are all just empty intrinsics, no need to
|
||||
// include them.
|
||||
if target_os != "ios" && (target_vendor != "apple" || target_arch != "x86") {
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"absvti2.c",
|
||||
"addvti3.c",
|
||||
"clzti2.c",
|
||||
|
@ -220,52 +218,41 @@ mod c {
|
|||
"popcountti2.c",
|
||||
"subvti3.c",
|
||||
"ucmpti2.c",
|
||||
],
|
||||
);
|
||||
]);
|
||||
}
|
||||
|
||||
if target_vendor == "apple" {
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"atomic_flag_clear.c",
|
||||
"atomic_flag_clear_explicit.c",
|
||||
"atomic_flag_test_and_set.c",
|
||||
"atomic_flag_test_and_set_explicit.c",
|
||||
"atomic_signal_fence.c",
|
||||
"atomic_thread_fence.c",
|
||||
],
|
||||
);
|
||||
]);
|
||||
}
|
||||
|
||||
if target_env == "msvc" {
|
||||
if target_arch == "x86_64" {
|
||||
sources.extend(
|
||||
&[
|
||||
"x86_64/floatdisf.c",
|
||||
"x86_64/floatdixf.c",
|
||||
],
|
||||
);
|
||||
sources.extend(&["x86_64/floatdisf.c", "x86_64/floatdixf.c"]);
|
||||
}
|
||||
} else {
|
||||
// None of these seem to be used on x86_64 windows, and they've all
|
||||
// got the wrong ABI anyway, so we want to avoid them.
|
||||
if target_os != "windows" {
|
||||
if target_arch == "x86_64" {
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"x86_64/floatdisf.c",
|
||||
"x86_64/floatdixf.c",
|
||||
"x86_64/floatundidf.S",
|
||||
"x86_64/floatundisf.S",
|
||||
"x86_64/floatundixf.S",
|
||||
],
|
||||
);
|
||||
]);
|
||||
}
|
||||
}
|
||||
|
||||
if target_arch == "x86" {
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"i386/ashldi3.S",
|
||||
"i386/ashrdi3.S",
|
||||
"i386/divdi3.S",
|
||||
|
@ -280,14 +267,12 @@ mod c {
|
|||
"i386/muldi3.S",
|
||||
"i386/udivdi3.S",
|
||||
"i386/umoddi3.S",
|
||||
],
|
||||
);
|
||||
]);
|
||||
}
|
||||
}
|
||||
|
||||
if target_arch == "arm" && target_os != "ios" && target_env != "msvc" {
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"arm/aeabi_div0.c",
|
||||
"arm/aeabi_drsub.c",
|
||||
"arm/aeabi_frsub.c",
|
||||
|
@ -306,8 +291,7 @@ mod c {
|
|||
"arm/udivmodsi4.S",
|
||||
"arm/udivsi3.S",
|
||||
"arm/umodsi3.S",
|
||||
],
|
||||
);
|
||||
]);
|
||||
|
||||
if target_os == "freebsd" {
|
||||
sources.extend(&["clear_cache.c"]);
|
||||
|
@ -316,22 +300,18 @@ mod c {
|
|||
// First of all aeabi_cdcmp and aeabi_cfcmp are never called by LLVM.
|
||||
// Second are little-endian only, so build fail on big-endian targets.
|
||||
// Temporally workaround: exclude these files for big-endian targets.
|
||||
if !llvm_target[0].starts_with("thumbeb") &&
|
||||
!llvm_target[0].starts_with("armeb") {
|
||||
sources.extend(
|
||||
&[
|
||||
if !llvm_target[0].starts_with("thumbeb") && !llvm_target[0].starts_with("armeb") {
|
||||
sources.extend(&[
|
||||
"arm/aeabi_cdcmp.S",
|
||||
"arm/aeabi_cdcmpeq_check_nan.c",
|
||||
"arm/aeabi_cfcmp.S",
|
||||
"arm/aeabi_cfcmpeq_check_nan.c",
|
||||
],
|
||||
);
|
||||
]);
|
||||
}
|
||||
}
|
||||
|
||||
if llvm_target[0] == "armv7" {
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"arm/sync_fetch_and_add_4.S",
|
||||
"arm/sync_fetch_and_add_8.S",
|
||||
"arm/sync_fetch_and_and_4.S",
|
||||
|
@ -352,30 +332,27 @@ mod c {
|
|||
"arm/sync_fetch_and_umin_8.S",
|
||||
"arm/sync_fetch_and_xor_4.S",
|
||||
"arm/sync_fetch_and_xor_8.S",
|
||||
],
|
||||
);
|
||||
]);
|
||||
}
|
||||
|
||||
if llvm_target.last().unwrap().ends_with("eabihf") {
|
||||
if !llvm_target[0].starts_with("thumbv7em") &&
|
||||
!llvm_target[0].starts_with("thumbv8m.main") {
|
||||
if !llvm_target[0].starts_with("thumbv7em")
|
||||
&& !llvm_target[0].starts_with("thumbv8m.main")
|
||||
{
|
||||
// The FPU option chosen for these architectures in cc-rs, ie:
|
||||
// -mfpu=fpv4-sp-d16 for thumbv7em
|
||||
// -mfpu=fpv5-sp-d16 for thumbv8m.main
|
||||
// do not support double precision floating points conversions so the files
|
||||
// that include such instructions are not included for these targets.
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"arm/fixdfsivfp.S",
|
||||
"arm/fixunsdfsivfp.S",
|
||||
"arm/floatsidfvfp.S",
|
||||
"arm/floatunssidfvfp.S",
|
||||
],
|
||||
);
|
||||
]);
|
||||
}
|
||||
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"arm/fixsfsivfp.S",
|
||||
"arm/fixunssfsivfp.S",
|
||||
"arm/floatsisfvfp.S",
|
||||
|
@ -385,14 +362,11 @@ mod c {
|
|||
"arm/save_vfp_d8_d15_regs.S",
|
||||
"arm/negdf2vfp.S",
|
||||
"arm/negsf2vfp.S",
|
||||
]
|
||||
);
|
||||
|
||||
]);
|
||||
}
|
||||
|
||||
if target_arch == "aarch64" {
|
||||
sources.extend(
|
||||
&[
|
||||
sources.extend(&[
|
||||
"comparetf2.c",
|
||||
"extenddftf2.c",
|
||||
"extendsftf2.c",
|
||||
|
@ -408,8 +382,7 @@ mod c {
|
|||
"floatunsitf.c",
|
||||
"trunctfdf2.c",
|
||||
"trunctfsf2.c",
|
||||
],
|
||||
);
|
||||
]);
|
||||
|
||||
if target_os != "windows" {
|
||||
sources.extend(&["multc3.c"]);
|
||||
|
@ -418,8 +391,7 @@ mod c {
|
|||
|
||||
// Remove the assembly implementations that won't compile for the target
|
||||
if llvm_target[0] == "thumbv6m" || llvm_target[0] == "thumbv8m.base" {
|
||||
sources.remove(
|
||||
&[
|
||||
sources.remove(&[
|
||||
"clzdi2",
|
||||
"clzsi2",
|
||||
"divmodsi4",
|
||||
|
@ -432,8 +404,7 @@ mod c {
|
|||
"udivmodsi4",
|
||||
"udivsi3",
|
||||
"umodsi3",
|
||||
],
|
||||
);
|
||||
]);
|
||||
|
||||
// But use some generic implementations where possible
|
||||
sources.extend(&["clzdi2.c", "clzsi2.c"])
|
||||
|
|
|
@ -17,7 +17,7 @@ extern crate panic_handler;
|
|||
|
||||
#[cfg(all(not(thumb), not(windows)))]
|
||||
#[link(name = "c")]
|
||||
extern {}
|
||||
extern "C" {}
|
||||
|
||||
// Every function in this module maps will be lowered to an intrinsic by LLVM, if the platform
|
||||
// doesn't have native support for the operation used in the function. ARM has a naming convention
|
||||
|
@ -340,11 +340,13 @@ fn run() {
|
|||
|
||||
something_with_a_dtor(&|| assert_eq!(bb(1), 1));
|
||||
|
||||
extern {
|
||||
extern "C" {
|
||||
fn rust_begin_unwind();
|
||||
}
|
||||
// if bb(false) {
|
||||
unsafe { rust_begin_unwind(); }
|
||||
unsafe {
|
||||
rust_begin_unwind();
|
||||
}
|
||||
// }
|
||||
}
|
||||
|
||||
|
@ -377,7 +379,7 @@ pub fn _start() -> ! {
|
|||
#[cfg(windows)]
|
||||
#[link(name = "kernel32")]
|
||||
#[link(name = "msvcrt")]
|
||||
extern {}
|
||||
extern "C" {}
|
||||
|
||||
// ARM targets need these symbols
|
||||
#[no_mangle]
|
||||
|
|
|
@ -94,7 +94,7 @@ macro_rules! atomic_rmw {
|
|||
pub unsafe extern "C" fn $name(ptr: *mut $ty, val: $ty) -> $ty {
|
||||
atomic_rmw(ptr, |x| $op(x as $ty, val) as u32) as $ty
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
macro_rules! atomic_cmpxchg {
|
||||
($name:ident, $ty:ty) => {
|
||||
|
@ -102,16 +102,20 @@ macro_rules! atomic_cmpxchg {
|
|||
pub unsafe extern "C" fn $name(ptr: *mut $ty, oldval: $ty, newval: $ty) -> $ty {
|
||||
atomic_cmpxchg(ptr, oldval as u32, newval as u32) as $ty
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
atomic_rmw!(__sync_fetch_and_add_1, u8, |a: u8, b: u8| a.wrapping_add(b));
|
||||
atomic_rmw!(__sync_fetch_and_add_2, u16, |a: u16, b: u16| a.wrapping_add(b));
|
||||
atomic_rmw!(__sync_fetch_and_add_4, u32, |a: u32, b: u32| a.wrapping_add(b));
|
||||
atomic_rmw!(__sync_fetch_and_add_2, u16, |a: u16, b: u16| a
|
||||
.wrapping_add(b));
|
||||
atomic_rmw!(__sync_fetch_and_add_4, u32, |a: u32, b: u32| a
|
||||
.wrapping_add(b));
|
||||
|
||||
atomic_rmw!(__sync_fetch_and_sub_1, u8, |a: u8, b: u8| a.wrapping_sub(b));
|
||||
atomic_rmw!(__sync_fetch_and_sub_2, u16, |a: u16, b: u16| a.wrapping_sub(b));
|
||||
atomic_rmw!(__sync_fetch_and_sub_4, u32, |a: u32, b: u32| a.wrapping_sub(b));
|
||||
atomic_rmw!(__sync_fetch_and_sub_2, u16, |a: u16, b: u16| a
|
||||
.wrapping_sub(b));
|
||||
atomic_rmw!(__sync_fetch_and_sub_4, u32, |a: u32, b: u32| a
|
||||
.wrapping_sub(b));
|
||||
|
||||
atomic_rmw!(__sync_fetch_and_and_1, u8, |a: u8, b: u8| a & b);
|
||||
atomic_rmw!(__sync_fetch_and_and_2, u16, |a: u16, b: u16| a & b);
|
||||
|
@ -129,21 +133,69 @@ atomic_rmw!(__sync_fetch_and_nand_1, u8, |a: u8, b: u8| !(a & b));
|
|||
atomic_rmw!(__sync_fetch_and_nand_2, u16, |a: u16, b: u16| !(a & b));
|
||||
atomic_rmw!(__sync_fetch_and_nand_4, u32, |a: u32, b: u32| !(a & b));
|
||||
|
||||
atomic_rmw!(__sync_fetch_and_max_1, i8, |a: i8, b: i8| if a > b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_max_2, i16, |a: i16, b: i16| if a > b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_max_4, i32, |a: i32, b: i32| if a > b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_max_1, i8, |a: i8, b: i8| if a > b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_max_2, i16, |a: i16, b: i16| if a > b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_max_4, i32, |a: i32, b: i32| if a > b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
|
||||
atomic_rmw!(__sync_fetch_and_umax_1, u8, |a: u8, b: u8| if a > b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_umax_2, u16, |a: u16, b: u16| if a > b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_umax_4, u32, |a: u32, b: u32| if a > b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_umax_1, u8, |a: u8, b: u8| if a > b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_umax_2, u16, |a: u16, b: u16| if a > b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_umax_4, u32, |a: u32, b: u32| if a > b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
|
||||
atomic_rmw!(__sync_fetch_and_min_1, i8, |a: i8, b: i8| if a < b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_min_2, i16, |a: i16, b: i16| if a < b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_min_4, i32, |a: i32, b: i32| if a < b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_min_1, i8, |a: i8, b: i8| if a < b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_min_2, i16, |a: i16, b: i16| if a < b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_min_4, i32, |a: i32, b: i32| if a < b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
|
||||
atomic_rmw!(__sync_fetch_and_umin_1, u8, |a: u8, b: u8| if a < b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_umin_2, u16, |a: u16, b: u16| if a < b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_umin_4, u32, |a: u32, b: u32| if a < b { a } else { b });
|
||||
atomic_rmw!(__sync_fetch_and_umin_1, u8, |a: u8, b: u8| if a < b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_umin_2, u16, |a: u16, b: u16| if a < b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
atomic_rmw!(__sync_fetch_and_umin_4, u32, |a: u32, b: u32| if a < b {
|
||||
a
|
||||
} else {
|
||||
b
|
||||
});
|
||||
|
||||
atomic_rmw!(__sync_lock_test_and_set_1, u8, |_: u8, b: u8| b);
|
||||
atomic_rmw!(__sync_lock_test_and_set_2, u16, |_: u16, b: u16| b);
|
||||
|
|
|
@ -1,8 +1,9 @@
|
|||
use int::{Int, CastInto};
|
||||
use float::Float;
|
||||
use int::{CastInto, Int};
|
||||
|
||||
/// Returns `a + b`
|
||||
fn add<F: Float>(a: F, b: F) -> F where
|
||||
fn add<F: Float>(a: F, b: F) -> F
|
||||
where
|
||||
u32: CastInto<F::Int>,
|
||||
F::Int: CastInto<u32>,
|
||||
i32: CastInto<F::Int>,
|
||||
|
@ -30,8 +31,7 @@ fn add<F: Float>(a: F, b: F) -> F where
|
|||
let b_abs = b_rep & abs_mask;
|
||||
|
||||
// Detect if a or b is zero, infinity, or NaN.
|
||||
if a_abs.wrapping_sub(one) >= inf_rep - one ||
|
||||
b_abs.wrapping_sub(one) >= inf_rep - one {
|
||||
if a_abs.wrapping_sub(one) >= inf_rep - one || b_abs.wrapping_sub(one) >= inf_rep - one {
|
||||
// NaN + anything = qNaN
|
||||
if a_abs > inf_rep {
|
||||
return F::from_repr(a_abs | quiet_bit);
|
||||
|
@ -115,7 +115,8 @@ fn add<F: Float>(a: F, b: F) -> F where
|
|||
let align = a_exponent.wrapping_sub(b_exponent).cast();
|
||||
if align != Int::ZERO {
|
||||
if align < bits {
|
||||
let sticky = F::Int::from_bool(b_significand << bits.wrapping_sub(align).cast() != Int::ZERO);
|
||||
let sticky =
|
||||
F::Int::from_bool(b_significand << bits.wrapping_sub(align).cast() != Int::ZERO);
|
||||
b_significand = (b_significand >> align.cast()) | sticky;
|
||||
} else {
|
||||
b_significand = one; // sticky; b is known to be non-zero.
|
||||
|
@ -131,12 +132,14 @@ fn add<F: Float>(a: F, b: F) -> F where
|
|||
// If partial cancellation occured, we need to left-shift the result
|
||||
// and adjust the exponent:
|
||||
if a_significand < implicit_bit << 3 {
|
||||
let shift = a_significand.leading_zeros() as i32
|
||||
- (implicit_bit << 3).leading_zeros() as i32;
|
||||
let shift =
|
||||
a_significand.leading_zeros() as i32 - (implicit_bit << 3).leading_zeros() as i32;
|
||||
a_significand <<= shift;
|
||||
a_exponent -= shift;
|
||||
}
|
||||
} else /* addition */ {
|
||||
} else
|
||||
/* addition */
|
||||
{
|
||||
a_significand += b_significand;
|
||||
|
||||
// If the addition carried up, we need to right-shift the result and
|
||||
|
@ -157,7 +160,8 @@ fn add<F: Float>(a: F, b: F) -> F where
|
|||
// Result is denormal before rounding; the exponent is zero and we
|
||||
// need to shift the significand.
|
||||
let shift = (1 - a_exponent).cast();
|
||||
let sticky = F::Int::from_bool((a_significand << bits.wrapping_sub(shift).cast()) != Int::ZERO);
|
||||
let sticky =
|
||||
F::Int::from_bool((a_significand << bits.wrapping_sub(shift).cast()) != Int::ZERO);
|
||||
a_significand = a_significand >> shift.cast() | sticky;
|
||||
a_exponent = 0;
|
||||
}
|
||||
|
@ -175,8 +179,12 @@ fn add<F: Float>(a: F, b: F) -> F where
|
|||
|
||||
// Final rounding. The result may overflow to infinity, but that is the
|
||||
// correct result in that case.
|
||||
if round_guard_sticky > 0x4 { result += one; }
|
||||
if round_guard_sticky == 0x4 { result += result & one; }
|
||||
if round_guard_sticky > 0x4 {
|
||||
result += one;
|
||||
}
|
||||
if round_guard_sticky == 0x4 {
|
||||
result += result & one;
|
||||
}
|
||||
|
||||
F::from_repr(result)
|
||||
}
|
||||
|
|
|
@ -1,14 +1,14 @@
|
|||
#![allow(unreachable_code)]
|
||||
|
||||
use int::{Int, CastInto};
|
||||
use float::Float;
|
||||
use int::{CastInto, Int};
|
||||
|
||||
#[derive(Clone, Copy)]
|
||||
enum Result {
|
||||
Less,
|
||||
Equal,
|
||||
Greater,
|
||||
Unordered
|
||||
Unordered,
|
||||
}
|
||||
|
||||
impl Result {
|
||||
|
@ -17,7 +17,7 @@ impl Result {
|
|||
Result::Less => -1,
|
||||
Result::Equal => 0,
|
||||
Result::Greater => 1,
|
||||
Result::Unordered => 1
|
||||
Result::Unordered => 1,
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -26,12 +26,13 @@ impl Result {
|
|||
Result::Less => -1,
|
||||
Result::Equal => 0,
|
||||
Result::Greater => 1,
|
||||
Result::Unordered => -1
|
||||
Result::Unordered => -1,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn cmp<F: Float>(a: F, b: F) -> Result where
|
||||
fn cmp<F: Float>(a: F, b: F) -> Result
|
||||
where
|
||||
u32: CastInto<F::Int>,
|
||||
F::Int: CastInto<u32>,
|
||||
i32: CastInto<F::Int>,
|
||||
|
@ -53,12 +54,12 @@ fn cmp<F: Float>(a: F, b: F) -> Result where
|
|||
|
||||
// If either a or b is NaN, they are unordered.
|
||||
if a_abs > inf_rep || b_abs > inf_rep {
|
||||
return Result::Unordered
|
||||
return Result::Unordered;
|
||||
}
|
||||
|
||||
// If a and b are both zeros, they are equal.
|
||||
if a_abs | b_abs == zero {
|
||||
return Result::Equal
|
||||
return Result::Equal;
|
||||
}
|
||||
|
||||
let a_srep = a.signed_repr();
|
||||
|
@ -68,29 +69,29 @@ fn cmp<F: Float>(a: F, b: F) -> Result where
|
|||
// a and b as signed integers as we would with a fp_ting-point compare.
|
||||
if a_srep & b_srep >= szero {
|
||||
if a_srep < b_srep {
|
||||
return Result::Less
|
||||
return Result::Less;
|
||||
} else if a_srep == b_srep {
|
||||
return Result::Equal
|
||||
return Result::Equal;
|
||||
} else {
|
||||
return Result::Greater
|
||||
return Result::Greater;
|
||||
}
|
||||
}
|
||||
|
||||
// Otherwise, both are negative, so we need to flip the sense of the
|
||||
// comparison to get the correct result. (This assumes a twos- or ones-
|
||||
// complement integer representation; if integers are represented in a
|
||||
// sign-magnitude representation, then this flip is incorrect).
|
||||
else {
|
||||
if a_srep > b_srep {
|
||||
return Result::Less
|
||||
return Result::Less;
|
||||
} else if a_srep == b_srep {
|
||||
return Result::Equal
|
||||
return Result::Equal;
|
||||
} else {
|
||||
return Result::Greater
|
||||
return Result::Greater;
|
||||
}
|
||||
}
|
||||
}
|
||||
fn unord<F: Float>(a: F, b: F) -> bool where
|
||||
fn unord<F: Float>(a: F, b: F) -> bool
|
||||
where
|
||||
u32: CastInto<F::Int>,
|
||||
F::Int: CastInto<u32>,
|
||||
i32: CastInto<F::Int>,
|
||||
|
|
|
@ -2,10 +2,10 @@ use float::Float;
|
|||
use int::Int;
|
||||
|
||||
macro_rules! int_to_float {
|
||||
($i:expr, $ity:ty, $fty:ty) => ({
|
||||
($i:expr, $ity:ty, $fty:ty) => {{
|
||||
let i = $i;
|
||||
if i == 0 {
|
||||
return 0.0
|
||||
return 0.0;
|
||||
}
|
||||
|
||||
let mant_dig = <$fty>::SIGNIFICAND_BITS + 1;
|
||||
|
@ -22,9 +22,11 @@ macro_rules! int_to_float {
|
|||
let mut e = sd - 1;
|
||||
|
||||
if <$ity>::BITS < mant_dig {
|
||||
return <$fty>::from_parts(s,
|
||||
return <$fty>::from_parts(
|
||||
s,
|
||||
(e + exponent_bias) as <$fty as Float>::Int,
|
||||
(a as <$fty as Float>::Int) << (mant_dig - e - 1))
|
||||
(a as <$fty as Float>::Int) << (mant_dig - e - 1),
|
||||
);
|
||||
}
|
||||
|
||||
a = if sd > mant_dig {
|
||||
|
@ -43,8 +45,10 @@ macro_rules! int_to_float {
|
|||
} else if sd == mant_dig_plus_two {
|
||||
a
|
||||
} else {
|
||||
(a >> (sd - mant_dig_plus_two)) as <$ity as Int>::UnsignedInt |
|
||||
((a & <$ity as Int>::UnsignedInt::max_value()).wrapping_shl((n + mant_dig_plus_two) - sd) != 0) as <$ity as Int>::UnsignedInt
|
||||
(a >> (sd - mant_dig_plus_two)) as <$ity as Int>::UnsignedInt
|
||||
| ((a & <$ity as Int>::UnsignedInt::max_value())
|
||||
.wrapping_shl((n + mant_dig_plus_two) - sd)
|
||||
!= 0) as <$ity as Int>::UnsignedInt
|
||||
};
|
||||
|
||||
/* finish: */
|
||||
|
@ -54,7 +58,8 @@ macro_rules! int_to_float {
|
|||
|
||||
/* a is now rounded to mant_dig or mant_dig+1 bits */
|
||||
if (a & (1 << mant_dig)) != 0 {
|
||||
a >>= 1; e += 1;
|
||||
a >>= 1;
|
||||
e += 1;
|
||||
}
|
||||
a
|
||||
/* a is now rounded to mant_dig bits */
|
||||
|
@ -63,10 +68,12 @@ macro_rules! int_to_float {
|
|||
/* a is now rounded to mant_dig bits */
|
||||
};
|
||||
|
||||
<$fty>::from_parts(s,
|
||||
<$fty>::from_parts(
|
||||
s,
|
||||
(e + exponent_bias) as <$fty as Float>::Int,
|
||||
a as <$fty as Float>::Int)
|
||||
})
|
||||
a as <$fty as Float>::Int,
|
||||
)
|
||||
}};
|
||||
}
|
||||
|
||||
intrinsics! {
|
||||
|
@ -160,11 +167,11 @@ intrinsics! {
|
|||
#[derive(PartialEq)]
|
||||
enum Sign {
|
||||
Positive,
|
||||
Negative
|
||||
Negative,
|
||||
}
|
||||
|
||||
macro_rules! float_to_int {
|
||||
($f:expr, $fty:ty, $ity:ty) => ({
|
||||
($f:expr, $fty:ty, $ity:ty) => {{
|
||||
let f = $f;
|
||||
let fixint_min = <$ity>::min_value();
|
||||
let fixint_max = <$ity>::max_value();
|
||||
|
@ -181,21 +188,34 @@ macro_rules! float_to_int {
|
|||
let a_abs = a_rep & !sign_bit;
|
||||
|
||||
// this is used to work around -1 not being available for unsigned
|
||||
let sign = if (a_rep & sign_bit) == 0 { Sign::Positive } else { Sign::Negative };
|
||||
let sign = if (a_rep & sign_bit) == 0 {
|
||||
Sign::Positive
|
||||
} else {
|
||||
Sign::Negative
|
||||
};
|
||||
let mut exponent = (a_abs >> significand_bits) as usize;
|
||||
let significand = (a_abs & <$fty>::SIGNIFICAND_MASK) | <$fty>::IMPLICIT_BIT;
|
||||
|
||||
// if < 1 or unsigned & negative
|
||||
if exponent < exponent_bias ||
|
||||
fixint_unsigned && sign == Sign::Negative {
|
||||
return 0
|
||||
if exponent < exponent_bias || fixint_unsigned && sign == Sign::Negative {
|
||||
return 0;
|
||||
}
|
||||
exponent -= exponent_bias;
|
||||
|
||||
// If the value is infinity, saturate.
|
||||
// If the value is too large for the integer type, 0.
|
||||
if exponent >= (if fixint_unsigned {fixint_bits} else {fixint_bits -1}) {
|
||||
return if sign == Sign::Positive {fixint_max} else {fixint_min}
|
||||
if exponent
|
||||
>= (if fixint_unsigned {
|
||||
fixint_bits
|
||||
} else {
|
||||
fixint_bits - 1
|
||||
})
|
||||
{
|
||||
return if sign == Sign::Positive {
|
||||
fixint_max
|
||||
} else {
|
||||
fixint_min
|
||||
};
|
||||
}
|
||||
// If 0 <= exponent < significand_bits, right shift to get the result.
|
||||
// Otherwise, shift left.
|
||||
|
@ -211,7 +231,7 @@ macro_rules! float_to_int {
|
|||
} else {
|
||||
r
|
||||
}
|
||||
})
|
||||
}};
|
||||
}
|
||||
|
||||
intrinsics! {
|
||||
|
|
|
@ -1,7 +1,5 @@
|
|||
use int::{CastInto, Int, WideInt};
|
||||
use float::Float;
|
||||
|
||||
|
||||
use int::{CastInto, Int, WideInt};
|
||||
|
||||
fn div32<F: Float>(a: F, b: F) -> F
|
||||
where
|
||||
|
@ -398,7 +396,6 @@ where
|
|||
// operation in C, so we need to be a little bit fussy.
|
||||
let (mut quotient, _) = <F::Int as WideInt>::wide_mul(a_significand << 2, reciprocal.cast());
|
||||
|
||||
|
||||
// Two cases: quotient is in [0.5, 1.0) or quotient is in [1.0, 2.0).
|
||||
// In either case, we are going to compute a residual of the form
|
||||
//
|
||||
|
@ -442,7 +439,6 @@ where
|
|||
}
|
||||
}
|
||||
|
||||
|
||||
intrinsics! {
|
||||
#[arm_aeabi_alias = __aeabi_fdiv]
|
||||
pub extern "C" fn __divsf3(a: f32, b: f32) -> f32 {
|
||||
|
|
|
@ -1,8 +1,9 @@
|
|||
use int::{CastInto, Int};
|
||||
use float::Float;
|
||||
use int::{CastInto, Int};
|
||||
|
||||
/// Generic conversion from a narrower to a wider IEEE-754 floating-point type
|
||||
fn extend<F: Float, R: Float>(a: F) -> R where
|
||||
fn extend<F: Float, R: Float>(a: F) -> R
|
||||
where
|
||||
F::Int: CastInto<u64>,
|
||||
u64: CastInto<F::Int>,
|
||||
u32: CastInto<R::Int>,
|
||||
|
|
|
@ -3,26 +3,26 @@ use core::ops;
|
|||
|
||||
use super::int::Int;
|
||||
|
||||
pub mod conv;
|
||||
pub mod cmp;
|
||||
pub mod add;
|
||||
pub mod pow;
|
||||
pub mod sub;
|
||||
pub mod mul;
|
||||
pub mod cmp;
|
||||
pub mod conv;
|
||||
pub mod div;
|
||||
pub mod extend;
|
||||
pub mod mul;
|
||||
pub mod pow;
|
||||
pub mod sub;
|
||||
|
||||
/// Trait for some basic operations on floats
|
||||
pub trait Float:
|
||||
Copy +
|
||||
PartialEq +
|
||||
PartialOrd +
|
||||
ops::AddAssign +
|
||||
ops::MulAssign +
|
||||
ops::Add<Output = Self> +
|
||||
ops::Sub<Output = Self> +
|
||||
ops::Div<Output = Self> +
|
||||
ops::Rem<Output = Self> +
|
||||
Copy
|
||||
+ PartialEq
|
||||
+ PartialOrd
|
||||
+ ops::AddAssign
|
||||
+ ops::MulAssign
|
||||
+ ops::Add<Output = Self>
|
||||
+ ops::Sub<Output = Self>
|
||||
+ ops::Div<Output = Self>
|
||||
+ ops::Rem<Output = Self>
|
||||
{
|
||||
/// A uint of the same with as the float
|
||||
type Int: Int;
|
||||
|
@ -118,17 +118,23 @@ macro_rules! float_impl {
|
|||
unsafe { mem::transmute(a) }
|
||||
}
|
||||
fn from_parts(sign: bool, exponent: Self::Int, significand: Self::Int) -> Self {
|
||||
Self::from_repr(((sign as Self::Int) << (Self::BITS - 1)) |
|
||||
((exponent << Self::SIGNIFICAND_BITS) & Self::EXPONENT_MASK) |
|
||||
(significand & Self::SIGNIFICAND_MASK))
|
||||
Self::from_repr(
|
||||
((sign as Self::Int) << (Self::BITS - 1))
|
||||
| ((exponent << Self::SIGNIFICAND_BITS) & Self::EXPONENT_MASK)
|
||||
| (significand & Self::SIGNIFICAND_MASK),
|
||||
)
|
||||
}
|
||||
fn normalize(significand: Self::Int) -> (i32, Self::Int) {
|
||||
let shift = significand.leading_zeros()
|
||||
let shift = significand
|
||||
.leading_zeros()
|
||||
.wrapping_sub((Self::Int::ONE << Self::SIGNIFICAND_BITS).leading_zeros());
|
||||
(1i32.wrapping_sub(shift as i32), significand << shift as Self::Int)
|
||||
}
|
||||
(
|
||||
1i32.wrapping_sub(shift as i32),
|
||||
significand << shift as Self::Int,
|
||||
)
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
float_impl!(f32, u32, i32, 32, 23);
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
use int::{CastInto, Int, WideInt};
|
||||
use float::Float;
|
||||
use int::{CastInto, Int, WideInt};
|
||||
|
||||
fn mul<F: Float>(a: F, b: F) -> F
|
||||
where
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
use int::Int;
|
||||
use float::Float;
|
||||
use int::Int;
|
||||
|
||||
trait Pow: Float {
|
||||
/// Returns `a` raised to the power `b`
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
use float::Float;
|
||||
use float::add::__addsf3;
|
||||
use float::add::__adddf3;
|
||||
use float::add::__addsf3;
|
||||
use float::Float;
|
||||
|
||||
intrinsics! {
|
||||
#[arm_aeabi_alias = __aeabi_fsub]
|
||||
|
|
|
@ -1,16 +1,24 @@
|
|||
use int::LargeInt;
|
||||
use int::Int;
|
||||
use int::LargeInt;
|
||||
|
||||
trait UAddSub: LargeInt {
|
||||
fn uadd(self, other: Self) -> Self {
|
||||
let (low, carry) = self.low().overflowing_add(other.low());
|
||||
let high = self.high().wrapping_add(other.high());
|
||||
let carry = if carry { Self::HighHalf::ONE } else { Self::HighHalf::ZERO };
|
||||
let carry = if carry {
|
||||
Self::HighHalf::ONE
|
||||
} else {
|
||||
Self::HighHalf::ZERO
|
||||
};
|
||||
Self::from_parts(low, high.wrapping_add(carry))
|
||||
}
|
||||
fn uadd_one(self) -> Self {
|
||||
let (low, carry) = self.low().overflowing_add(Self::LowHalf::ONE);
|
||||
let carry = if carry { Self::HighHalf::ONE } else { Self::HighHalf::ZERO };
|
||||
let carry = if carry {
|
||||
Self::HighHalf::ONE
|
||||
} else {
|
||||
Self::HighHalf::ZERO
|
||||
};
|
||||
Self::from_parts(low, self.high().wrapping_add(carry))
|
||||
}
|
||||
fn usub(self, other: Self) -> Self {
|
||||
|
@ -22,7 +30,8 @@ trait UAddSub: LargeInt {
|
|||
impl UAddSub for u128 {}
|
||||
|
||||
trait AddSub: Int
|
||||
where <Self as Int>::UnsignedInt: UAddSub
|
||||
where
|
||||
<Self as Int>::UnsignedInt: UAddSub,
|
||||
{
|
||||
fn add(self, other: Self) -> Self {
|
||||
Self::from_unsigned(self.unsigned().uadd(other.unsigned()))
|
||||
|
@ -36,7 +45,8 @@ impl AddSub for u128 {}
|
|||
impl AddSub for i128 {}
|
||||
|
||||
trait Addo: AddSub
|
||||
where <Self as Int>::UnsignedInt: UAddSub
|
||||
where
|
||||
<Self as Int>::UnsignedInt: UAddSub,
|
||||
{
|
||||
fn addo(self, other: Self, overflow: &mut i32) -> Self {
|
||||
*overflow = 0;
|
||||
|
@ -58,7 +68,8 @@ impl Addo for i128 {}
|
|||
impl Addo for u128 {}
|
||||
|
||||
trait Subo: AddSub
|
||||
where <Self as Int>::UnsignedInt: UAddSub
|
||||
where
|
||||
<Self as Int>::UnsignedInt: UAddSub,
|
||||
{
|
||||
fn subo(self, other: Self, overflow: &mut i32) -> Self {
|
||||
*overflow = 0;
|
||||
|
|
|
@ -3,13 +3,13 @@ use core::ops;
|
|||
macro_rules! hty {
|
||||
($ty:ty) => {
|
||||
<$ty as LargeInt>::HighHalf
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
macro_rules! os_ty {
|
||||
($ty:ty) => {
|
||||
<$ty as Int>::OtherSign
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
pub mod addsub;
|
||||
|
@ -20,23 +20,23 @@ pub mod udiv;
|
|||
|
||||
/// Trait for some basic operations on integers
|
||||
pub trait Int:
|
||||
Copy +
|
||||
PartialEq +
|
||||
PartialOrd +
|
||||
ops::AddAssign +
|
||||
ops::BitAndAssign +
|
||||
ops::BitOrAssign +
|
||||
ops::ShlAssign<i32> +
|
||||
ops::ShrAssign<u32> +
|
||||
ops::Add<Output = Self> +
|
||||
ops::Sub<Output = Self> +
|
||||
ops::Div<Output = Self> +
|
||||
ops::Shl<u32, Output = Self> +
|
||||
ops::Shr<u32, Output = Self> +
|
||||
ops::BitOr<Output = Self> +
|
||||
ops::BitXor<Output = Self> +
|
||||
ops::BitAnd<Output = Self> +
|
||||
ops::Not<Output = Self> +
|
||||
Copy
|
||||
+ PartialEq
|
||||
+ PartialOrd
|
||||
+ ops::AddAssign
|
||||
+ ops::BitAndAssign
|
||||
+ ops::BitOrAssign
|
||||
+ ops::ShlAssign<i32>
|
||||
+ ops::ShrAssign<u32>
|
||||
+ ops::Add<Output = Self>
|
||||
+ ops::Sub<Output = Self>
|
||||
+ ops::Div<Output = Self>
|
||||
+ ops::Shl<u32, Output = Self>
|
||||
+ ops::Shr<u32, Output = Self>
|
||||
+ ops::BitOr<Output = Self>
|
||||
+ ops::BitXor<Output = Self>
|
||||
+ ops::BitAnd<Output = Self>
|
||||
+ ops::Not<Output = Self>
|
||||
{
|
||||
/// Type with the same width but other signedness
|
||||
type OtherSign: Int;
|
||||
|
@ -182,7 +182,7 @@ macro_rules! int_impl {
|
|||
|
||||
int_impl_common!($ity, $bits);
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
int_impl!(i32, u32, 32);
|
||||
|
@ -223,7 +223,7 @@ macro_rules! large_int {
|
|||
low as $ty | ((high as $ty) << $halfbits)
|
||||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
large_int!(u64, u32, u32, 32);
|
||||
|
@ -295,7 +295,7 @@ macro_rules! impl_wide_int {
|
|||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
impl_wide_int!(u32, u64, 32);
|
||||
|
|
|
@ -1,7 +1,7 @@
|
|||
use core::ops;
|
||||
|
||||
use int::LargeInt;
|
||||
use int::Int;
|
||||
use int::LargeInt;
|
||||
|
||||
trait Mul: LargeInt {
|
||||
fn mul(self, other: Self) -> Self {
|
||||
|
@ -19,7 +19,8 @@ trait Mul: LargeInt {
|
|||
low += (t & lower_mask) << half_bits;
|
||||
high += Self::low_as_high(t >> half_bits);
|
||||
high += Self::low_as_high((self.low() >> half_bits).wrapping_mul(other.low() >> half_bits));
|
||||
high = high.wrapping_add(self.high().wrapping_mul(Self::low_as_high(other.low())))
|
||||
high = high
|
||||
.wrapping_add(self.high().wrapping_mul(Self::low_as_high(other.low())))
|
||||
.wrapping_add(Self::low_as_high(self.low()).wrapping_mul(other.high()));
|
||||
Self::from_parts(low, high)
|
||||
}
|
||||
|
|
|
@ -43,7 +43,8 @@ impl Mod for i128 {}
|
|||
trait Divmod: Int {
|
||||
/// Returns `a / b` and sets `*rem = n % d`
|
||||
fn divmod<F>(self, other: Self, rem: &mut Self, div: F) -> Self
|
||||
where F: Fn(Self, Self) -> Self,
|
||||
where
|
||||
F: Fn(Self, Self) -> Self,
|
||||
{
|
||||
let r = div(self, other);
|
||||
// NOTE won't overflow because it's using the result from the
|
||||
|
|
|
@ -3,7 +3,8 @@ use int::{Int, LargeInt};
|
|||
trait Ashl: Int + LargeInt {
|
||||
/// Returns `a << b`, requires `b < Self::BITS`
|
||||
fn ashl(self, offset: u32) -> Self
|
||||
where Self: LargeInt<HighHalf = <Self as LargeInt>::LowHalf>,
|
||||
where
|
||||
Self: LargeInt<HighHalf = <Self as LargeInt>::LowHalf>,
|
||||
{
|
||||
let half_bits = Self::BITS / 2;
|
||||
if offset & half_bits != 0 {
|
||||
|
@ -11,9 +12,10 @@ trait Ashl: Int + LargeInt {
|
|||
} else if offset == 0 {
|
||||
self
|
||||
} else {
|
||||
Self::from_parts(self.low() << offset,
|
||||
(self.high() << offset) |
|
||||
(self.low() >> (half_bits - offset)))
|
||||
Self::from_parts(
|
||||
self.low() << offset,
|
||||
(self.high() << offset) | (self.low() >> (half_bits - offset)),
|
||||
)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -24,18 +26,23 @@ impl Ashl for u128 {}
|
|||
trait Ashr: Int + LargeInt {
|
||||
/// Returns arithmetic `a >> b`, requires `b < Self::BITS`
|
||||
fn ashr(self, offset: u32) -> Self
|
||||
where Self: LargeInt<LowHalf = <<Self as LargeInt>::HighHalf as Int>::UnsignedInt>,
|
||||
where
|
||||
Self: LargeInt<LowHalf = <<Self as LargeInt>::HighHalf as Int>::UnsignedInt>,
|
||||
{
|
||||
let half_bits = Self::BITS / 2;
|
||||
if offset & half_bits != 0 {
|
||||
Self::from_parts((self.high() >> (offset - half_bits)).unsigned(),
|
||||
self.high() >> (half_bits - 1))
|
||||
Self::from_parts(
|
||||
(self.high() >> (offset - half_bits)).unsigned(),
|
||||
self.high() >> (half_bits - 1),
|
||||
)
|
||||
} else if offset == 0 {
|
||||
self
|
||||
} else {
|
||||
let high_unsigned = self.high().unsigned();
|
||||
Self::from_parts((high_unsigned << (half_bits - offset)) | (self.low() >> offset),
|
||||
self.high() >> offset)
|
||||
Self::from_parts(
|
||||
(high_unsigned << (half_bits - offset)) | (self.low() >> offset),
|
||||
self.high() >> offset,
|
||||
)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -46,7 +53,8 @@ impl Ashr for i128 {}
|
|||
trait Lshr: Int + LargeInt {
|
||||
/// Returns logical `a >> b`, requires `b < Self::BITS`
|
||||
fn lshr(self, offset: u32) -> Self
|
||||
where Self: LargeInt<HighHalf = <Self as LargeInt>::LowHalf>,
|
||||
where
|
||||
Self: LargeInt<HighHalf = <Self as LargeInt>::LowHalf>,
|
||||
{
|
||||
let half_bits = Self::BITS / 2;
|
||||
if offset & half_bits != 0 {
|
||||
|
@ -54,9 +62,10 @@ trait Lshr: Int + LargeInt {
|
|||
} else if offset == 0 {
|
||||
self
|
||||
} else {
|
||||
Self::from_parts((self.high() << (half_bits - offset)) |
|
||||
(self.low() >> offset),
|
||||
self.high() >> offset)
|
||||
Self::from_parts(
|
||||
(self.high() << (half_bits - offset)) | (self.low() >> offset),
|
||||
self.high() >> offset,
|
||||
)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
24
src/lib.rs
24
src/lib.rs
|
@ -3,11 +3,13 @@
|
|||
#![cfg_attr(feature = "compiler-builtins", compiler_builtins)]
|
||||
#![crate_name = "compiler_builtins"]
|
||||
#![crate_type = "rlib"]
|
||||
#![doc(html_logo_url = "https://www.rust-lang.org/logos/rust-logo-128x128-blk.png",
|
||||
#![doc(
|
||||
html_logo_url = "https://www.rust-lang.org/logos/rust-logo-128x128-blk.png",
|
||||
html_favicon_url = "https://doc.rust-lang.org/favicon.ico",
|
||||
html_root_url = "https://doc.rust-lang.org/nightly/",
|
||||
html_playground_url = "https://play.rust-lang.org/",
|
||||
test(attr(deny(warnings))))]
|
||||
test(attr(deny(warnings)))
|
||||
)]
|
||||
#![feature(asm)]
|
||||
#![feature(compiler_builtins)]
|
||||
#![feature(core_intrinsics)]
|
||||
|
@ -19,10 +21,14 @@
|
|||
#![allow(unused_features)]
|
||||
#![no_builtins]
|
||||
#![cfg_attr(feature = "compiler-builtins", feature(staged_api))]
|
||||
#![cfg_attr(feature = "compiler-builtins",
|
||||
unstable(feature = "compiler_builtins_lib",
|
||||
#![cfg_attr(
|
||||
feature = "compiler-builtins",
|
||||
unstable(
|
||||
feature = "compiler_builtins_lib",
|
||||
reason = "Compiler builtins. Will never become stable.",
|
||||
issue = "0"))]
|
||||
issue = "0"
|
||||
)
|
||||
)]
|
||||
|
||||
// We disable #[no_mangle] for tests so that we can verify the test results
|
||||
// against the native compiler-rt implementations of the builtins.
|
||||
|
@ -44,12 +50,14 @@ fn abort() -> ! {
|
|||
#[macro_use]
|
||||
mod macros;
|
||||
|
||||
pub mod int;
|
||||
pub mod float;
|
||||
pub mod int;
|
||||
|
||||
#[cfg(any(all(target_arch = "wasm32", target_os = "unknown"),
|
||||
#[cfg(any(
|
||||
all(target_arch = "wasm32", target_os = "unknown"),
|
||||
all(target_arch = "arm", target_os = "none"),
|
||||
all(target_vendor = "fortanix", target_env = "sgx")))]
|
||||
all(target_vendor = "fortanix", target_env = "sgx")
|
||||
))]
|
||||
pub mod math;
|
||||
pub mod mem;
|
||||
|
||||
|
|
10
src/mem.rs
10
src/mem.rs
|
@ -6,10 +6,7 @@ type c_int = i16;
|
|||
type c_int = i32;
|
||||
|
||||
#[cfg_attr(all(feature = "mem", not(feature = "mangled-names")), no_mangle)]
|
||||
pub unsafe extern "C" fn memcpy(dest: *mut u8,
|
||||
src: *const u8,
|
||||
n: usize)
|
||||
-> *mut u8 {
|
||||
pub unsafe extern "C" fn memcpy(dest: *mut u8, src: *const u8, n: usize) -> *mut u8 {
|
||||
let mut i = 0;
|
||||
while i < n {
|
||||
*dest.offset(i as isize) = *src.offset(i as isize);
|
||||
|
@ -19,10 +16,7 @@ pub unsafe extern "C" fn memcpy(dest: *mut u8,
|
|||
}
|
||||
|
||||
#[cfg_attr(all(feature = "mem", not(feature = "mangled-names")), no_mangle)]
|
||||
pub unsafe extern "C" fn memmove(dest: *mut u8,
|
||||
src: *const u8,
|
||||
n: usize)
|
||||
-> *mut u8 {
|
||||
pub unsafe extern "C" fn memmove(dest: *mut u8, src: *const u8, n: usize) -> *mut u8 {
|
||||
if src < dest as *const u8 {
|
||||
// copy from end
|
||||
let mut i = n;
|
||||
|
|
|
@ -46,7 +46,7 @@
|
|||
#[naked]
|
||||
#[no_mangle]
|
||||
#[cfg(all(target_arch = "x86_64", not(feature = "mangled-names")))]
|
||||
pub unsafe extern fn __rust_probestack() {
|
||||
pub unsafe extern "C" fn __rust_probestack() {
|
||||
// Our goal here is to touch each page between %rsp+8 and %rsp+8-%rax,
|
||||
// ensuring that if any pages are unmapped we'll make a page fault.
|
||||
//
|
||||
|
@ -97,7 +97,7 @@ pub unsafe extern fn __rust_probestack() {
|
|||
#[naked]
|
||||
#[no_mangle]
|
||||
#[cfg(all(target_arch = "x86", not(feature = "mangled-names")))]
|
||||
pub unsafe extern fn __rust_probestack() {
|
||||
pub unsafe extern "C" fn __rust_probestack() {
|
||||
// This is the same as x86_64 above, only translated for 32-bit sizes. Note
|
||||
// that on Unix we're expected to restore everything as it was, this
|
||||
// function basically can't tamper with anything.
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -1,7 +1,9 @@
|
|||
#![cfg(all(target_arch = "arm",
|
||||
#![cfg(all(
|
||||
target_arch = "arm",
|
||||
not(any(target_env = "gnu", target_env = "musl")),
|
||||
target_os = "linux",
|
||||
feature = "mem"))]
|
||||
feature = "mem"
|
||||
))]
|
||||
#![feature(compiler_builtins_lib)]
|
||||
#![feature(lang_items)]
|
||||
#![no_std]
|
||||
|
|
|
@ -1,7 +1,9 @@
|
|||
#![cfg(all(target_arch = "arm",
|
||||
#![cfg(all(
|
||||
target_arch = "arm",
|
||||
not(any(target_env = "gnu", target_env = "musl")),
|
||||
target_os = "linux",
|
||||
feature = "mem"))]
|
||||
feature = "mem"
|
||||
))]
|
||||
#![feature(compiler_builtins_lib)]
|
||||
#![feature(lang_items)]
|
||||
#![no_std]
|
||||
|
|
|
@ -1,7 +1,9 @@
|
|||
#![cfg(all(target_arch = "arm",
|
||||
#![cfg(all(
|
||||
target_arch = "arm",
|
||||
not(any(target_env = "gnu", target_env = "musl")),
|
||||
target_os = "linux",
|
||||
feature = "mem"))]
|
||||
feature = "mem"
|
||||
))]
|
||||
#![feature(compiler_builtins_lib)]
|
||||
#![feature(lang_items)]
|
||||
#![no_std]
|
||||
|
@ -48,9 +50,7 @@ fn zero() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), 0, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), 0, c) }
|
||||
|
||||
assert_eq!(*xs, [0; 8]);
|
||||
|
||||
|
@ -59,9 +59,7 @@ fn zero() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), 0, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), 0, c) }
|
||||
|
||||
assert_eq!(*xs, [1; 8]);
|
||||
}
|
||||
|
@ -74,9 +72,7 @@ fn one() {
|
|||
let n = 1;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0, 0, 0, 0, 0, 0, 0]);
|
||||
|
||||
|
@ -85,9 +81,7 @@ fn one() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 1, 1, 1, 1, 1, 1, 1]);
|
||||
}
|
||||
|
@ -100,9 +94,7 @@ fn two() {
|
|||
let n = 2;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0, 0, 0, 0, 0, 0]);
|
||||
|
||||
|
@ -111,9 +103,7 @@ fn two() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 1, 1, 1, 1, 1, 1]);
|
||||
}
|
||||
|
@ -126,9 +116,7 @@ fn three() {
|
|||
let n = 3;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0, 0, 0, 0, 0]);
|
||||
|
||||
|
@ -137,9 +125,7 @@ fn three() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 1, 1, 1, 1, 1]);
|
||||
}
|
||||
|
@ -152,9 +138,7 @@ fn four() {
|
|||
let n = 4;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0, 0, 0, 0]);
|
||||
|
||||
|
@ -163,9 +147,7 @@ fn four() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 1, 1, 1, 1]);
|
||||
}
|
||||
|
@ -178,9 +160,7 @@ fn five() {
|
|||
let n = 5;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 0, 0, 0]);
|
||||
|
||||
|
@ -189,9 +169,7 @@ fn five() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 1, 1, 1]);
|
||||
}
|
||||
|
@ -204,9 +182,7 @@ fn six() {
|
|||
let n = 6;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0, 0]);
|
||||
|
||||
|
@ -215,9 +191,7 @@ fn six() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 1, 1]);
|
||||
}
|
||||
|
@ -230,9 +204,7 @@ fn seven() {
|
|||
let n = 7;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0]);
|
||||
|
||||
|
@ -241,9 +213,7 @@ fn seven() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 1]);
|
||||
}
|
||||
|
@ -256,9 +226,7 @@ fn eight() {
|
|||
let n = 8;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0xef]);
|
||||
|
||||
|
@ -267,9 +235,7 @@ fn eight() {
|
|||
let xs = &mut aligned.array;
|
||||
let c = 0xdeadbeef;
|
||||
|
||||
unsafe {
|
||||
__aeabi_memset4(xs.as_mut_ptr(), n, c)
|
||||
}
|
||||
unsafe { __aeabi_memset4(xs.as_mut_ptr(), n, c) }
|
||||
|
||||
assert_eq!(*xs, [0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0xef, 0xef]);
|
||||
}
|
||||
|
|
|
@ -6,23 +6,29 @@
|
|||
|
||||
extern crate compiler_builtins as builtins;
|
||||
|
||||
#[cfg(all(target_arch = "arm",
|
||||
#[cfg(all(
|
||||
target_arch = "arm",
|
||||
not(any(target_env = "gnu", target_env = "musl")),
|
||||
target_os = "linux",
|
||||
test))]
|
||||
test
|
||||
))]
|
||||
extern crate utest_cortex_m_qemu;
|
||||
|
||||
#[cfg(all(target_arch = "arm",
|
||||
#[cfg(all(
|
||||
target_arch = "arm",
|
||||
not(any(target_env = "gnu", target_env = "musl")),
|
||||
target_os = "linux",
|
||||
test))]
|
||||
test
|
||||
))]
|
||||
#[macro_use]
|
||||
extern crate utest_macros;
|
||||
|
||||
#[cfg(all(target_arch = "arm",
|
||||
#[cfg(all(
|
||||
target_arch = "arm",
|
||||
not(any(target_env = "gnu", target_env = "musl")),
|
||||
target_os = "linux",
|
||||
test))]
|
||||
test
|
||||
))]
|
||||
macro_rules! panic { // overrides `panic!`
|
||||
($($tt:tt)*) => {
|
||||
upanic!($($tt)*);
|
||||
|
|
Loading…
Reference in New Issue