|
@@ -1,44 +1,26 @@
|
|
|
-/// Creates unsigned and signed division functions optimized for dividing integers with the same
|
|
|
+/// Creates an unsigned division function optimized for dividing integers with the same
|
|
|
/// bitwidth as the largest operand in an asymmetrically sized division. For example, x86-64 has an
|
|
|
/// assembly instruction that can divide a 128 bit integer by a 64 bit integer if the quotient fits
|
|
|
/// in 64 bits. The 128 bit version of this algorithm would use that fast hardware division to
|
|
|
/// construct a full 128 bit by 128 bit division.
|
|
|
+#[doc(hidden)]
|
|
|
#[macro_export]
|
|
|
macro_rules! impl_asymmetric {
|
|
|
(
|
|
|
- $unsigned_name:ident, // name of the unsigned division function
|
|
|
- $signed_name:ident, // name of the signed division function
|
|
|
+ $fn:ident, // name of the unsigned division function
|
|
|
$zero_div_fn:ident, // function called when division by zero is attempted
|
|
|
$half_division:ident, // function for division of a $uX by a $uX
|
|
|
$asymmetric_division:ident, // function for division of a $uD by a $uX
|
|
|
$n_h:expr, // the number of bits in a $iH or $uH
|
|
|
$uH:ident, // unsigned integer with half the bit width of $uX
|
|
|
$uX:ident, // unsigned integer with half the bit width of $uD
|
|
|
- $uD:ident, // unsigned integer type for the inputs and outputs of `$unsigned_name`
|
|
|
- $iD:ident, // signed integer type for the inputs and outputs of `$signed_name`
|
|
|
- $($unsigned_attr:meta),*; // attributes for the unsigned function
|
|
|
- $($signed_attr:meta),* // attributes for the signed function
|
|
|
+ $uD:ident // unsigned integer type for the inputs and outputs of `$fn`
|
|
|
) => {
|
|
|
/// Computes the quotient and remainder of `duo` divided by `div` and returns them as a
|
|
|
/// tuple.
|
|
|
- $(
|
|
|
- #[$unsigned_attr]
|
|
|
- )*
|
|
|
- pub fn $unsigned_name(duo: $uD, div: $uD) -> ($uD,$uD) {
|
|
|
- fn carrying_mul(lhs: $uX, rhs: $uX) -> ($uX, $uX) {
|
|
|
- let tmp = (lhs as $uD).wrapping_mul(rhs as $uD);
|
|
|
- (tmp as $uX, (tmp >> ($n_h * 2)) as $uX)
|
|
|
- }
|
|
|
- fn carrying_mul_add(lhs: $uX, mul: $uX, add: $uX) -> ($uX, $uX) {
|
|
|
- let tmp = (lhs as $uD).wrapping_mul(mul as $uD).wrapping_add(add as $uD);
|
|
|
- (tmp as $uX, (tmp >> ($n_h * 2)) as $uX)
|
|
|
- }
|
|
|
-
|
|
|
+ pub fn $fn(duo: $uD, div: $uD) -> ($uD, $uD) {
|
|
|
let n: u32 = $n_h * 2;
|
|
|
|
|
|
- // Many of these subalgorithms are taken from trifecta.rs, see that for better
|
|
|
- // documentation.
|
|
|
-
|
|
|
let duo_lo = duo as $uX;
|
|
|
let duo_hi = (duo >> n) as $uX;
|
|
|
let div_lo = div as $uX;
|
|
@@ -50,120 +32,39 @@ macro_rules! impl_asymmetric {
|
|
|
if duo_hi < div_lo {
|
|
|
// `$uD` by `$uX` division with a quotient that will fit into a `$uX`
|
|
|
let (quo, rem) = unsafe { $asymmetric_division(duo, div_lo) };
|
|
|
- return (quo as $uD, rem as $uD)
|
|
|
- } else if (div_lo >> $n_h) == 0 {
|
|
|
- // Short division of $uD by a $uH.
|
|
|
-
|
|
|
- // Some x86_64 CPUs have bad division implementations that make specializing
|
|
|
- // this case faster.
|
|
|
- let div_0 = div_lo as $uH as $uX;
|
|
|
- let (quo_hi, rem_3) = $half_division(duo_hi, div_0);
|
|
|
-
|
|
|
- let duo_mid =
|
|
|
- ((duo >> $n_h) as $uH as $uX)
|
|
|
- | (rem_3 << $n_h);
|
|
|
- let (quo_1, rem_2) = $half_division(duo_mid, div_0);
|
|
|
-
|
|
|
- let duo_lo =
|
|
|
- (duo as $uH as $uX)
|
|
|
- | (rem_2 << $n_h);
|
|
|
- let (quo_0, rem_1) = $half_division(duo_lo, div_0);
|
|
|
-
|
|
|
- return (
|
|
|
- (quo_0 as $uD)
|
|
|
- | ((quo_1 as $uD) << $n_h)
|
|
|
- | ((quo_hi as $uD) << n),
|
|
|
- rem_1 as $uD
|
|
|
- )
|
|
|
+ return (quo as $uD, rem as $uD);
|
|
|
} else {
|
|
|
// Short division using the $uD by $uX division
|
|
|
let (quo_hi, rem_hi) = $half_division(duo_hi, div_lo);
|
|
|
let tmp = unsafe {
|
|
|
$asymmetric_division((duo_lo as $uD) | ((rem_hi as $uD) << n), div_lo)
|
|
|
};
|
|
|
- return ((tmp.0 as $uD) | ((quo_hi as $uD) << n), tmp.1 as $uD)
|
|
|
+ return ((tmp.0 as $uD) | ((quo_hi as $uD) << n), tmp.1 as $uD);
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- let duo_lz = duo_hi.leading_zeros();
|
|
|
+ // This has been adapted from
|
|
|
+ // https://www.codeproject.com/tips/785014/uint-division-modulus which was in turn
|
|
|
+ // adapted from Hacker's Delight. This is similar to the two possibility algorithm
|
|
|
+ // in that it uses only more significant parts of `duo` and `div` to divide a large
|
|
|
+ // integer with a smaller division instruction.
|
|
|
let div_lz = div_hi.leading_zeros();
|
|
|
- let rel_leading_sb = div_lz.wrapping_sub(duo_lz);
|
|
|
- if rel_leading_sb < $n_h {
|
|
|
- // Some x86_64 CPUs have bad hardware division implementations that make putting
|
|
|
- // a two possibility algorithm here beneficial. We also avoid a full `$uD`
|
|
|
- // multiplication.
|
|
|
- let shift = n - duo_lz;
|
|
|
- let duo_sig_n = (duo >> shift) as $uX;
|
|
|
- let div_sig_n = (div >> shift) as $uX;
|
|
|
- let quo = $half_division(duo_sig_n, div_sig_n).0;
|
|
|
- let div_lo = div as $uX;
|
|
|
- let div_hi = (div >> n) as $uX;
|
|
|
- let (tmp_lo, carry) = carrying_mul(quo, div_lo);
|
|
|
- let (tmp_hi, overflow) = carrying_mul_add(quo, div_hi, carry);
|
|
|
- let tmp = (tmp_lo as $uD) | ((tmp_hi as $uD) << n);
|
|
|
- if (overflow != 0) || (duo < tmp) {
|
|
|
- return (
|
|
|
- (quo - 1) as $uD,
|
|
|
- duo.wrapping_add(div).wrapping_sub(tmp)
|
|
|
- )
|
|
|
- } else {
|
|
|
- return (
|
|
|
- quo as $uD,
|
|
|
- duo - tmp
|
|
|
- )
|
|
|
- }
|
|
|
- } else {
|
|
|
- // This has been adapted from
|
|
|
- // https://www.codeproject.com/tips/785014/uint-division-modulus which was in turn
|
|
|
- // adapted from Hacker's Delight. This is similar to the two possibility algorithm
|
|
|
- // in that it uses only more significant parts of `duo` and `div` to divide a large
|
|
|
- // integer with a smaller division instruction.
|
|
|
-
|
|
|
- let div_extra = n - div_lz;
|
|
|
- let div_sig_n = (div >> div_extra) as $uX;
|
|
|
- let tmp = unsafe {
|
|
|
- $asymmetric_division(duo >> 1, div_sig_n)
|
|
|
- };
|
|
|
-
|
|
|
- let mut quo = tmp.0 >> ((n - 1) - div_lz);
|
|
|
- if quo != 0 {
|
|
|
- quo -= 1;
|
|
|
- }
|
|
|
+ let div_extra = n - div_lz;
|
|
|
+ let div_sig_n = (div >> div_extra) as $uX;
|
|
|
+ let tmp = unsafe { $asymmetric_division(duo >> 1, div_sig_n) };
|
|
|
|
|
|
- // Note that this is a full `$uD` multiplication being used here
|
|
|
- let mut rem = duo - (quo as $uD).wrapping_mul(div);
|
|
|
- if div <= rem {
|
|
|
- quo += 1;
|
|
|
- rem -= div;
|
|
|
- }
|
|
|
- return (quo as $uD, rem)
|
|
|
+ let mut quo = tmp.0 >> ((n - 1) - div_lz);
|
|
|
+ if quo != 0 {
|
|
|
+ quo -= 1;
|
|
|
}
|
|
|
- }
|
|
|
|
|
|
- /// Computes the quotient and remainder of `duo` divided by `div` and returns them as a
|
|
|
- /// tuple.
|
|
|
- $(
|
|
|
- #[$signed_attr]
|
|
|
- )*
|
|
|
- pub fn $signed_name(duo: $iD, div: $iD) -> ($iD, $iD) {
|
|
|
- match (duo < 0, div < 0) {
|
|
|
- (false, false) => {
|
|
|
- let t = $unsigned_name(duo as $uD, div as $uD);
|
|
|
- (t.0 as $iD, t.1 as $iD)
|
|
|
- },
|
|
|
- (true, false) => {
|
|
|
- let t = $unsigned_name(duo.wrapping_neg() as $uD, div as $uD);
|
|
|
- ((t.0 as $iD).wrapping_neg(), (t.1 as $iD).wrapping_neg())
|
|
|
- },
|
|
|
- (false, true) => {
|
|
|
- let t = $unsigned_name(duo as $uD, div.wrapping_neg() as $uD);
|
|
|
- ((t.0 as $iD).wrapping_neg(), t.1 as $iD)
|
|
|
- },
|
|
|
- (true, true) => {
|
|
|
- let t = $unsigned_name(duo.wrapping_neg() as $uD, div.wrapping_neg() as $uD);
|
|
|
- (t.0 as $iD, (t.1 as $iD).wrapping_neg())
|
|
|
- },
|
|
|
+ // Note that this is a full `$uD` multiplication being used here
|
|
|
+ let mut rem = duo - (quo as $uD).wrapping_mul(div);
|
|
|
+ if div <= rem {
|
|
|
+ quo += 1;
|
|
|
+ rem -= div;
|
|
|
}
|
|
|
+ return (quo as $uD, rem);
|
|
|
}
|
|
|
- }
|
|
|
+ };
|
|
|
}
|