clippy: fix ptr_offset_with_cast warnings
i.e. use of `offset` with a `usize` casted to an `isize`
This commit is contained in:
parent
f1211b0d99
commit
74a09c8a5e
|
@ -22,7 +22,7 @@ impl<'a, N: Scalar, R: Dim, C: Dim, RStride: Dim, CStride: Dim>
|
||||||
cstride: CStride,
|
cstride: CStride,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
let data = SliceStorage::from_raw_parts(
|
let data = SliceStorage::from_raw_parts(
|
||||||
data.as_ptr().offset(start as isize),
|
data.as_ptr().add(start),
|
||||||
(nrows, ncols),
|
(nrows, ncols),
|
||||||
(rstride, cstride),
|
(rstride, cstride),
|
||||||
);
|
);
|
||||||
|
@ -156,7 +156,7 @@ impl<'a, N: Scalar, R: Dim, C: Dim, RStride: Dim, CStride: Dim>
|
||||||
cstride: CStride,
|
cstride: CStride,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
let data = SliceStorageMut::from_raw_parts(
|
let data = SliceStorageMut::from_raw_parts(
|
||||||
data.as_mut_ptr().offset(start as isize),
|
data.as_mut_ptr().add(start),
|
||||||
(nrows, ncols),
|
(nrows, ncols),
|
||||||
(rstride, cstride),
|
(rstride, cstride),
|
||||||
);
|
);
|
||||||
|
|
|
@ -341,11 +341,8 @@ impl<N: Scalar, R: Dim, C: Dim, S: Storage<N, R, C>> Matrix<N, R, C, S> {
|
||||||
offset += 1;
|
offset += 1;
|
||||||
} else {
|
} else {
|
||||||
unsafe {
|
unsafe {
|
||||||
let ptr_source = m
|
let ptr_source = m.data.ptr().add((target + offset) * nrows.value());
|
||||||
.data
|
let ptr_target = m.data.ptr_mut().add(target * nrows.value());
|
||||||
.ptr()
|
|
||||||
.offset(((target + offset) * nrows.value()) as isize);
|
|
||||||
let ptr_target = m.data.ptr_mut().offset((target * nrows.value()) as isize);
|
|
||||||
|
|
||||||
ptr::copy(ptr_source, ptr_target, nrows.value());
|
ptr::copy(ptr_source, ptr_target, nrows.value());
|
||||||
target += 1;
|
target += 1;
|
||||||
|
@ -378,8 +375,8 @@ impl<N: Scalar, R: Dim, C: Dim, S: Storage<N, R, C>> Matrix<N, R, C, S> {
|
||||||
offset += 1;
|
offset += 1;
|
||||||
} else {
|
} else {
|
||||||
unsafe {
|
unsafe {
|
||||||
let ptr_source = m.data.ptr().offset((target + offset) as isize);
|
let ptr_source = m.data.ptr().add(target + offset);
|
||||||
let ptr_target = m.data.ptr_mut().offset(target as isize);
|
let ptr_target = m.data.ptr_mut().add(target);
|
||||||
|
|
||||||
ptr::copy(ptr_source, ptr_target, 1);
|
ptr::copy(ptr_source, ptr_target, 1);
|
||||||
target += 1;
|
target += 1;
|
||||||
|
@ -442,11 +439,8 @@ impl<N: Scalar, R: Dim, C: Dim, S: Storage<N, R, C>> Matrix<N, R, C, S> {
|
||||||
let copied_value_start = i + nremove.value();
|
let copied_value_start = i + nremove.value();
|
||||||
|
|
||||||
unsafe {
|
unsafe {
|
||||||
let ptr_in = m
|
let ptr_in = m.data.ptr().add(copied_value_start * nrows.value());
|
||||||
.data
|
let ptr_out = m.data.ptr_mut().add(i * nrows.value());
|
||||||
.ptr()
|
|
||||||
.offset((copied_value_start * nrows.value()) as isize);
|
|
||||||
let ptr_out = m.data.ptr_mut().offset((i * nrows.value()) as isize);
|
|
||||||
|
|
||||||
ptr::copy(
|
ptr::copy(
|
||||||
ptr_in,
|
ptr_in,
|
||||||
|
@ -610,11 +604,11 @@ impl<N: Scalar, R: Dim, C: Dim, S: Storage<N, R, C>> Matrix<N, R, C, S> {
|
||||||
assert!(i <= ncols.value(), "Column insertion index out of range.");
|
assert!(i <= ncols.value(), "Column insertion index out of range.");
|
||||||
|
|
||||||
if ninsert.value() != 0 && i != ncols.value() {
|
if ninsert.value() != 0 && i != ncols.value() {
|
||||||
let ptr_in = res.data.ptr().offset((i * nrows.value()) as isize);
|
let ptr_in = res.data.ptr().add(i * nrows.value());
|
||||||
let ptr_out = res
|
let ptr_out = res
|
||||||
.data
|
.data
|
||||||
.ptr_mut()
|
.ptr_mut()
|
||||||
.offset(((i + ninsert.value()) * nrows.value()) as isize);
|
.add((i + ninsert.value()) * nrows.value());
|
||||||
|
|
||||||
ptr::copy(ptr_in, ptr_out, (ncols.value() - i) * nrows.value())
|
ptr::copy(ptr_in, ptr_out, (ncols.value() - i) * nrows.value())
|
||||||
}
|
}
|
||||||
|
@ -977,8 +971,8 @@ unsafe fn compress_rows<N: Scalar>(
|
||||||
|
|
||||||
for k in 0..ncols - 1 {
|
for k in 0..ncols - 1 {
|
||||||
ptr::copy(
|
ptr::copy(
|
||||||
ptr_in.offset((curr_i + (k + 1) * nremove) as isize),
|
ptr_in.add(curr_i + (k + 1) * nremove),
|
||||||
ptr_out.offset(curr_i as isize),
|
ptr_out.add(curr_i),
|
||||||
new_nrows,
|
new_nrows,
|
||||||
);
|
);
|
||||||
|
|
||||||
|
@ -988,8 +982,8 @@ unsafe fn compress_rows<N: Scalar>(
|
||||||
// Deal with the last column from which less values have to be copied.
|
// Deal with the last column from which less values have to be copied.
|
||||||
let remaining_len = nrows - i - nremove;
|
let remaining_len = nrows - i - nremove;
|
||||||
ptr::copy(
|
ptr::copy(
|
||||||
ptr_in.offset((nrows * ncols - remaining_len) as isize),
|
ptr_in.add(nrows * ncols - remaining_len),
|
||||||
ptr_out.offset(curr_i as isize),
|
ptr_out.add(curr_i),
|
||||||
remaining_len,
|
remaining_len,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
@ -1017,19 +1011,15 @@ unsafe fn extend_rows<N: Scalar>(
|
||||||
|
|
||||||
// Deal with the last column from which less values have to be copied.
|
// Deal with the last column from which less values have to be copied.
|
||||||
ptr::copy(
|
ptr::copy(
|
||||||
ptr_in.offset((nrows * ncols - remaining_len) as isize),
|
ptr_in.add(nrows * ncols - remaining_len),
|
||||||
ptr_out.offset(curr_i as isize),
|
ptr_out.add(curr_i),
|
||||||
remaining_len,
|
remaining_len,
|
||||||
);
|
);
|
||||||
|
|
||||||
for k in (0..ncols - 1).rev() {
|
for k in (0..ncols - 1).rev() {
|
||||||
curr_i -= new_nrows;
|
curr_i -= new_nrows;
|
||||||
|
|
||||||
ptr::copy(
|
ptr::copy(ptr_in.add(k * nrows + i), ptr_out.add(curr_i), nrows);
|
||||||
ptr_in.offset((k * nrows + i) as isize),
|
|
||||||
ptr_out.offset(curr_i as isize),
|
|
||||||
nrows,
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -44,7 +44,7 @@ macro_rules! iterator {
|
||||||
// If 'size' is non-zero, we know that 'ptr'
|
// If 'size' is non-zero, we know that 'ptr'
|
||||||
// is not dangling, and 'inner_offset' must lie
|
// is not dangling, and 'inner_offset' must lie
|
||||||
// within the allocation
|
// within the allocation
|
||||||
unsafe { ptr.offset(inner_offset as isize) }
|
unsafe { ptr.add(inner_offset) }
|
||||||
};
|
};
|
||||||
|
|
||||||
$Name {
|
$Name {
|
||||||
|
@ -87,13 +87,13 @@ macro_rules! iterator {
|
||||||
// Go to the next element.
|
// Go to the next element.
|
||||||
let old = self.ptr;
|
let old = self.ptr;
|
||||||
|
|
||||||
let stride = self.strides.0.value() as isize;
|
|
||||||
// Don't offset `self.ptr` for the last element,
|
// Don't offset `self.ptr` for the last element,
|
||||||
// as this will be out of bounds. Iteration is done
|
// as this will be out of bounds. Iteration is done
|
||||||
// at this point (the next call to `next` will return `None`)
|
// at this point (the next call to `next` will return `None`)
|
||||||
// so this is not observable.
|
// so this is not observable.
|
||||||
if self.size != 0 {
|
if self.size != 0 {
|
||||||
self.ptr = self.ptr.offset(stride);
|
let stride = self.strides.0.value();
|
||||||
|
self.ptr = self.ptr.add(stride);
|
||||||
}
|
}
|
||||||
Some(mem::transmute(old))
|
Some(mem::transmute(old))
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue