#[cfg(feature = "alloc")]
use alloc::vec::Vec;
use core::ops::DerefMut;
#[cfg(feature = "alloc")]
use core::ptr;
use crate::traits::Sealed;
pub trait Option_v1_40<T: DerefMut>: Sealed<Option<T>> {
fn as_deref_mut(&mut self) -> Option<&mut T::Target>;
fn as_deref(&self) -> Option<&T::Target>;
}
impl<T: DerefMut> Option_v1_40<T> for Option<T> {
fn as_deref_mut(&mut self) -> Option<&mut T::Target> {
self.as_mut().map(|t| t.deref_mut())
}
fn as_deref(&self) -> Option<&T::Target> {
self.as_ref().map(|t| t.deref())
}
}
pub trait Option_v1_40_<T>: Sealed<Option<Option<T>>> {
fn flatten(self) -> Option<T>;
}
impl<T> Option_v1_40_<T> for Option<Option<T>> {
fn flatten(self) -> Option<T> {
self.and_then(core::convert::identity)
}
}
pub trait f32_v1_40: Sealed<f32> {
fn to_be_bytes(self) -> [u8; 4];
fn to_le_bytes(self) -> [u8; 4];
fn to_ne_bytes(self) -> [u8; 4];
fn from_be_bytes(bytes: [u8; 4]) -> Self;
fn from_le_bytes(bytes: [u8; 4]) -> Self;
fn from_ne_bytes(bytes: [u8; 4]) -> Self;
}
impl f32_v1_40 for f32 {
#[inline]
fn to_be_bytes(self) -> [u8; 4] {
self.to_bits().to_be_bytes()
}
#[inline]
fn to_le_bytes(self) -> [u8; 4] {
self.to_bits().to_le_bytes()
}
#[inline]
fn to_ne_bytes(self) -> [u8; 4] {
self.to_bits().to_ne_bytes()
}
#[inline]
fn from_be_bytes(bytes: [u8; 4]) -> Self {
Self::from_bits(u32::from_be_bytes(bytes))
}
#[inline]
fn from_le_bytes(bytes: [u8; 4]) -> Self {
Self::from_bits(u32::from_le_bytes(bytes))
}
#[inline]
fn from_ne_bytes(bytes: [u8; 4]) -> Self {
Self::from_bits(u32::from_ne_bytes(bytes))
}
}
pub trait f64_v1_40: Sealed<f64> {
fn to_be_bytes(self) -> [u8; 8];
fn to_le_bytes(self) -> [u8; 8];
fn to_ne_bytes(self) -> [u8; 8];
fn from_be_bytes(bytes: [u8; 8]) -> Self;
fn from_le_bytes(bytes: [u8; 8]) -> Self;
fn from_ne_bytes(bytes: [u8; 8]) -> Self;
}
impl f64_v1_40 for f64 {
#[inline]
fn to_be_bytes(self) -> [u8; 8] {
self.to_bits().to_be_bytes()
}
#[inline]
fn to_le_bytes(self) -> [u8; 8] {
self.to_bits().to_le_bytes()
}
#[inline]
fn to_ne_bytes(self) -> [u8; 8] {
self.to_bits().to_ne_bytes()
}
#[inline]
fn from_be_bytes(bytes: [u8; 8]) -> Self {
Self::from_bits(u64::from_be_bytes(bytes))
}
#[inline]
fn from_le_bytes(bytes: [u8; 8]) -> Self {
Self::from_bits(u64::from_le_bytes(bytes))
}
#[inline]
fn from_ne_bytes(bytes: [u8; 8]) -> Self {
Self::from_bits(u64::from_ne_bytes(bytes))
}
}
pub fn take<T: Default>(dest: &mut T) -> T {
core::mem::replace(dest, T::default())
}
#[cfg(feature = "alloc")]
pub trait slice_v1_40<T>: Sealed<[T]> {
fn repeat(&self, n: usize) -> Vec<T>
where
T: Copy;
}
#[cfg(feature = "alloc")]
impl<T: Copy> slice_v1_40<T> for [T] {
fn repeat(&self, n: usize) -> Vec<T> {
if n == 0 {
return Vec::new();
}
let mut buf = Vec::with_capacity(self.len().checked_mul(n).expect("capacity overflow"));
buf.extend(self);
{
let mut m = n >> 1;
while m > 0 {
unsafe {
ptr::copy_nonoverlapping(
buf.as_ptr(),
(buf.as_mut_ptr() as *mut T).add(buf.len()),
buf.len(),
);
let buf_len = buf.len();
buf.set_len(buf_len * 2);
}
m >>= 1;
}
}
let rem_len = self.len() * n - buf.len();
if rem_len > 0 {
unsafe {
ptr::copy_nonoverlapping(
buf.as_ptr(),
(buf.as_mut_ptr() as *mut T).add(buf.len()),
rem_len,
);
let buf_cap = buf.capacity();
buf.set_len(buf_cap);
}
}
buf
}
}