forked from M-Labs/nalgebra
302 lines
7.8 KiB
Rust
302 lines
7.8 KiB
Rust
use std::fmt::{self, Debug, Formatter};
|
|
// use std::hash::{Hash, Hasher};
|
|
#[cfg(feature = "abomonation-serialize")]
|
|
use std::io::{Result as IOResult, Write};
|
|
use std::ops::Mul;
|
|
|
|
#[cfg(feature = "serde-serialize")]
|
|
use serde::de::{Error, SeqAccess, Visitor};
|
|
#[cfg(feature = "serde-serialize")]
|
|
use serde::ser::SerializeSeq;
|
|
#[cfg(feature = "serde-serialize")]
|
|
use serde::{Deserialize, Deserializer, Serialize, Serializer};
|
|
#[cfg(feature = "serde-serialize")]
|
|
use std::marker::PhantomData;
|
|
#[cfg(feature = "serde-serialize")]
|
|
use std::mem;
|
|
|
|
#[cfg(feature = "abomonation-serialize")]
|
|
use abomonation::Abomonation;
|
|
|
|
use crate::base::allocator::Allocator;
|
|
use crate::base::default_allocator::DefaultAllocator;
|
|
use crate::base::dimension::{Const, ToTypenum};
|
|
use crate::base::storage::{
|
|
ContiguousStorage, ContiguousStorageMut, Owned, ReshapableStorage, Storage, StorageMut,
|
|
};
|
|
use crate::base::Scalar;
|
|
|
|
/*
|
|
*
|
|
* Static Storage.
|
|
*
|
|
*/
|
|
/// A array-based statically sized matrix data storage.
|
|
#[repr(C)]
|
|
#[derive(Copy, Clone, PartialEq, Eq, Hash)]
|
|
pub struct ArrayStorage<T, const R: usize, const C: usize>(pub [[T; R]; C]);
|
|
|
|
// TODO: remove this once the stdlib implements Default for arrays.
|
|
impl<T: Default, const R: usize, const C: usize> Default for ArrayStorage<T, R, C>
|
|
where
|
|
[[T; R]; C]: Default,
|
|
{
|
|
#[inline]
|
|
fn default() -> Self {
|
|
Self(Default::default())
|
|
}
|
|
}
|
|
|
|
impl<T: Debug, const R: usize, const C: usize> Debug for ArrayStorage<T, R, C> {
|
|
#[inline]
|
|
fn fmt(&self, fmt: &mut Formatter) -> fmt::Result {
|
|
self.0.fmt(fmt)
|
|
}
|
|
}
|
|
|
|
unsafe impl<T, const R: usize, const C: usize> Storage<T, Const<R>, Const<C>>
|
|
for ArrayStorage<T, R, C>
|
|
where
|
|
T: Scalar,
|
|
DefaultAllocator: Allocator<T, Const<R>, Const<C>, Buffer = Self>,
|
|
{
|
|
type RStride = Const<1>;
|
|
type CStride = Const<R>;
|
|
|
|
#[inline]
|
|
fn ptr(&self) -> *const T {
|
|
self.0.as_ptr() as *const T
|
|
}
|
|
|
|
#[inline]
|
|
fn shape(&self) -> (Const<R>, Const<C>) {
|
|
(Const, Const)
|
|
}
|
|
|
|
#[inline]
|
|
fn strides(&self) -> (Self::RStride, Self::CStride) {
|
|
(Const, Const)
|
|
}
|
|
|
|
#[inline]
|
|
fn is_contiguous(&self) -> bool {
|
|
true
|
|
}
|
|
|
|
#[inline]
|
|
fn into_owned(self) -> Owned<T, Const<R>, Const<C>>
|
|
where
|
|
DefaultAllocator: Allocator<T, Const<R>, Const<C>>,
|
|
{
|
|
self
|
|
}
|
|
|
|
#[inline]
|
|
fn clone_owned(&self) -> Owned<T, Const<R>, Const<C>>
|
|
where
|
|
DefaultAllocator: Allocator<T, Const<R>, Const<C>>,
|
|
{
|
|
let it = self.as_slice().iter().cloned();
|
|
DefaultAllocator::allocate_from_iterator(self.shape().0, self.shape().1, it)
|
|
}
|
|
|
|
#[inline]
|
|
fn as_slice(&self) -> &[T] {
|
|
unsafe { std::slice::from_raw_parts(self.ptr(), R * C) }
|
|
}
|
|
}
|
|
|
|
unsafe impl<T, const R: usize, const C: usize> StorageMut<T, Const<R>, Const<C>>
|
|
for ArrayStorage<T, R, C>
|
|
where
|
|
T: Scalar,
|
|
DefaultAllocator: Allocator<T, Const<R>, Const<C>, Buffer = Self>,
|
|
{
|
|
#[inline]
|
|
fn ptr_mut(&mut self) -> *mut T {
|
|
self.0.as_mut_ptr() as *mut T
|
|
}
|
|
|
|
#[inline]
|
|
fn as_mut_slice(&mut self) -> &mut [T] {
|
|
unsafe { std::slice::from_raw_parts_mut(self.ptr_mut(), R * C) }
|
|
}
|
|
}
|
|
|
|
unsafe impl<T, const R: usize, const C: usize> ContiguousStorage<T, Const<R>, Const<C>>
|
|
for ArrayStorage<T, R, C>
|
|
where
|
|
T: Scalar,
|
|
DefaultAllocator: Allocator<T, Const<R>, Const<C>, Buffer = Self>,
|
|
{
|
|
}
|
|
|
|
unsafe impl<T, const R: usize, const C: usize> ContiguousStorageMut<T, Const<R>, Const<C>>
|
|
for ArrayStorage<T, R, C>
|
|
where
|
|
T: Scalar,
|
|
DefaultAllocator: Allocator<T, Const<R>, Const<C>, Buffer = Self>,
|
|
{
|
|
}
|
|
|
|
impl<T, const R1: usize, const C1: usize, const R2: usize, const C2: usize>
|
|
ReshapableStorage<T, Const<R1>, Const<C1>, Const<R2>, Const<C2>> for ArrayStorage<T, R1, C1>
|
|
where
|
|
T: Scalar,
|
|
Const<R1>: ToTypenum,
|
|
Const<C1>: ToTypenum,
|
|
Const<R2>: ToTypenum,
|
|
Const<C2>: ToTypenum,
|
|
<Const<R1> as ToTypenum>::Typenum: Mul<<Const<C1> as ToTypenum>::Typenum>,
|
|
<Const<R2> as ToTypenum>::Typenum: Mul<
|
|
<Const<C2> as ToTypenum>::Typenum,
|
|
Output = typenum::Prod<
|
|
<Const<R1> as ToTypenum>::Typenum,
|
|
<Const<C1> as ToTypenum>::Typenum,
|
|
>,
|
|
>,
|
|
{
|
|
type Output = ArrayStorage<T, R2, C2>;
|
|
|
|
fn reshape_generic(self, _: Const<R2>, _: Const<C2>) -> Self::Output {
|
|
unsafe {
|
|
let data: [[T; R2]; C2] = std::mem::transmute_copy(&self.0);
|
|
std::mem::forget(self.0);
|
|
ArrayStorage(data)
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
*
|
|
* Serialization.
|
|
*
|
|
*/
|
|
// XXX: open an issue for serde so that it allows the serialization/deserialization of all arrays?
|
|
#[cfg(feature = "serde-serialize")]
|
|
impl<T, const R: usize, const C: usize> Serialize for ArrayStorage<T, R, C>
|
|
where
|
|
T: Scalar + Serialize,
|
|
{
|
|
fn serialize<S>(&self, serializer: S) -> Result<S::Ok, S::Error>
|
|
where
|
|
S: Serializer,
|
|
{
|
|
let mut serializer = serializer.serialize_seq(Some(R * C))?;
|
|
|
|
for e in self.as_slice().iter() {
|
|
serializer.serialize_element(e)?;
|
|
}
|
|
|
|
serializer.end()
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "serde-serialize")]
|
|
impl<'a, T, const R: usize, const C: usize> Deserialize<'a> for ArrayStorage<T, R, C>
|
|
where
|
|
T: Scalar + Deserialize<'a>,
|
|
{
|
|
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
|
|
where
|
|
D: Deserializer<'a>,
|
|
{
|
|
deserializer.deserialize_seq(ArrayStorageVisitor::new())
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "serde-serialize")]
|
|
/// A visitor that produces a matrix array.
|
|
struct ArrayStorageVisitor<T, const R: usize, const C: usize> {
|
|
marker: PhantomData<T>,
|
|
}
|
|
|
|
#[cfg(feature = "serde-serialize")]
|
|
impl<T, const R: usize, const C: usize> ArrayStorageVisitor<T, R, C>
|
|
where
|
|
T: Scalar,
|
|
{
|
|
/// Construct a new sequence visitor.
|
|
pub fn new() -> Self {
|
|
ArrayStorageVisitor {
|
|
marker: PhantomData,
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "serde-serialize")]
|
|
impl<'a, T, const R: usize, const C: usize> Visitor<'a> for ArrayStorageVisitor<T, R, C>
|
|
where
|
|
T: Scalar + Deserialize<'a>,
|
|
{
|
|
type Value = ArrayStorage<T, R, C>;
|
|
|
|
fn expecting(&self, formatter: &mut Formatter) -> fmt::Result {
|
|
formatter.write_str("a matrix array")
|
|
}
|
|
|
|
#[inline]
|
|
fn visit_seq<V>(self, mut visitor: V) -> Result<ArrayStorage<T, R, C>, V::Error>
|
|
where
|
|
V: SeqAccess<'a>,
|
|
{
|
|
let mut out: Self::Value = unsafe { mem::MaybeUninit::uninit().assume_init() };
|
|
let mut curr = 0;
|
|
|
|
while let Some(value) = visitor.next_element()? {
|
|
*out.as_mut_slice()
|
|
.get_mut(curr)
|
|
.ok_or_else(|| V::Error::invalid_length(curr, &self))? = value;
|
|
curr += 1;
|
|
}
|
|
|
|
if curr == R * C {
|
|
Ok(out)
|
|
} else {
|
|
Err(V::Error::invalid_length(curr, &self))
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "bytemuck")]
|
|
unsafe impl<T: Scalar + Copy + bytemuck::Zeroable, const R: usize, const C: usize>
|
|
bytemuck::Zeroable for ArrayStorage<T, R, C>
|
|
{
|
|
}
|
|
|
|
#[cfg(feature = "bytemuck")]
|
|
unsafe impl<T: Scalar + Copy + bytemuck::Pod, const R: usize, const C: usize> bytemuck::Pod
|
|
for ArrayStorage<T, R, C>
|
|
{
|
|
}
|
|
|
|
#[cfg(feature = "abomonation-serialize")]
|
|
impl<T, const R: usize, const C: usize> Abomonation for ArrayStorage<T, R, C>
|
|
where
|
|
T: Scalar + Abomonation,
|
|
{
|
|
unsafe fn entomb<W: Write>(&self, writer: &mut W) -> IOResult<()> {
|
|
for element in self.as_slice() {
|
|
element.entomb(writer)?;
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
unsafe fn exhume<'a, 'b>(&'a mut self, mut bytes: &'b mut [u8]) -> Option<&'b mut [u8]> {
|
|
for element in self.as_mut_slice() {
|
|
let temp = bytes;
|
|
bytes = if let Some(remainder) = element.exhume(temp) {
|
|
remainder
|
|
} else {
|
|
return None;
|
|
}
|
|
}
|
|
Some(bytes)
|
|
}
|
|
|
|
fn extent(&self) -> usize {
|
|
self.as_slice().iter().fold(0, |acc, e| acc + e.extent())
|
|
}
|
|
}
|