
Goodies in this version: - if-let chains 🎉 - Better compiler perf for Zed (https://github.com/rust-lang/rust/pull/138522) For more, see: https://releases.rs/docs/1.88.0/ Release Notes: - N/A --------- Co-authored-by: Junkui Zhang <364772080@qq.com>
300 lines
7.5 KiB
Rust
300 lines
7.5 KiB
Rust
use std::{
|
|
alloc::{self, handle_alloc_error},
|
|
cell::Cell,
|
|
ops::{Deref, DerefMut},
|
|
ptr,
|
|
rc::Rc,
|
|
};
|
|
|
|
struct ArenaElement {
|
|
value: *mut u8,
|
|
drop: unsafe fn(*mut u8),
|
|
}
|
|
|
|
impl Drop for ArenaElement {
|
|
#[inline(always)]
|
|
fn drop(&mut self) {
|
|
unsafe {
|
|
(self.drop)(self.value);
|
|
}
|
|
}
|
|
}
|
|
|
|
struct Chunk {
|
|
start: *mut u8,
|
|
end: *mut u8,
|
|
offset: *mut u8,
|
|
}
|
|
|
|
impl Drop for Chunk {
|
|
fn drop(&mut self) {
|
|
unsafe {
|
|
let chunk_size = self.end.offset_from_unsigned(self.start);
|
|
// this never fails as it succeeded during allocation
|
|
let layout = alloc::Layout::from_size_align(chunk_size, 1).unwrap();
|
|
alloc::dealloc(self.start, layout);
|
|
}
|
|
}
|
|
}
|
|
|
|
impl Chunk {
|
|
fn new(chunk_size: usize) -> Self {
|
|
unsafe {
|
|
// this only fails if chunk_size is unreasonably huge
|
|
let layout = alloc::Layout::from_size_align(chunk_size, 1).unwrap();
|
|
let start = alloc::alloc(layout);
|
|
if start.is_null() {
|
|
handle_alloc_error(layout);
|
|
}
|
|
let end = start.add(chunk_size);
|
|
Self {
|
|
start,
|
|
end,
|
|
offset: start,
|
|
}
|
|
}
|
|
}
|
|
|
|
fn allocate(&mut self, layout: alloc::Layout) -> Option<*mut u8> {
|
|
unsafe {
|
|
let aligned = self.offset.add(self.offset.align_offset(layout.align()));
|
|
let next = aligned.add(layout.size());
|
|
|
|
if next <= self.end {
|
|
self.offset = next;
|
|
Some(aligned)
|
|
} else {
|
|
None
|
|
}
|
|
}
|
|
}
|
|
|
|
fn reset(&mut self) {
|
|
self.offset = self.start;
|
|
}
|
|
}
|
|
|
|
pub struct Arena {
|
|
chunks: Vec<Chunk>,
|
|
elements: Vec<ArenaElement>,
|
|
valid: Rc<Cell<bool>>,
|
|
current_chunk_index: usize,
|
|
chunk_size: usize,
|
|
}
|
|
|
|
impl Drop for Arena {
|
|
fn drop(&mut self) {
|
|
self.clear();
|
|
}
|
|
}
|
|
|
|
impl Arena {
|
|
pub fn new(chunk_size: usize) -> Self {
|
|
assert!(chunk_size > 0);
|
|
Self {
|
|
chunks: vec![Chunk::new(chunk_size)],
|
|
elements: Vec::new(),
|
|
valid: Rc::new(Cell::new(true)),
|
|
current_chunk_index: 0,
|
|
chunk_size,
|
|
}
|
|
}
|
|
|
|
pub fn capacity(&self) -> usize {
|
|
self.chunks.len() * self.chunk_size
|
|
}
|
|
|
|
pub fn clear(&mut self) {
|
|
self.valid.set(false);
|
|
self.valid = Rc::new(Cell::new(true));
|
|
self.elements.clear();
|
|
for chunk_index in 0..=self.current_chunk_index {
|
|
self.chunks[chunk_index].reset();
|
|
}
|
|
self.current_chunk_index = 0;
|
|
}
|
|
|
|
#[inline(always)]
|
|
pub fn alloc<T>(&mut self, f: impl FnOnce() -> T) -> ArenaBox<T> {
|
|
#[inline(always)]
|
|
unsafe fn inner_writer<T, F>(ptr: *mut T, f: F)
|
|
where
|
|
F: FnOnce() -> T,
|
|
{
|
|
unsafe {
|
|
ptr::write(ptr, f());
|
|
}
|
|
}
|
|
|
|
unsafe fn drop<T>(ptr: *mut u8) {
|
|
unsafe {
|
|
std::ptr::drop_in_place(ptr.cast::<T>());
|
|
}
|
|
}
|
|
|
|
unsafe {
|
|
let layout = alloc::Layout::new::<T>();
|
|
let mut current_chunk = &mut self.chunks[self.current_chunk_index];
|
|
let ptr = if let Some(ptr) = current_chunk.allocate(layout) {
|
|
ptr
|
|
} else {
|
|
self.current_chunk_index += 1;
|
|
if self.current_chunk_index >= self.chunks.len() {
|
|
self.chunks.push(Chunk::new(self.chunk_size));
|
|
assert_eq!(self.current_chunk_index, self.chunks.len() - 1);
|
|
log::info!(
|
|
"increased element arena capacity to {}kb",
|
|
self.capacity() / 1024,
|
|
);
|
|
}
|
|
current_chunk = &mut self.chunks[self.current_chunk_index];
|
|
if let Some(ptr) = current_chunk.allocate(layout) {
|
|
ptr
|
|
} else {
|
|
panic!(
|
|
"Arena chunk_size of {} is too small to allocate {} bytes",
|
|
self.chunk_size,
|
|
layout.size()
|
|
);
|
|
}
|
|
};
|
|
|
|
inner_writer(ptr.cast(), f);
|
|
self.elements.push(ArenaElement {
|
|
value: ptr,
|
|
drop: drop::<T>,
|
|
});
|
|
|
|
ArenaBox {
|
|
ptr: ptr.cast(),
|
|
valid: self.valid.clone(),
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
pub struct ArenaBox<T: ?Sized> {
|
|
ptr: *mut T,
|
|
valid: Rc<Cell<bool>>,
|
|
}
|
|
|
|
impl<T: ?Sized> ArenaBox<T> {
|
|
#[inline(always)]
|
|
pub fn map<U: ?Sized>(mut self, f: impl FnOnce(&mut T) -> &mut U) -> ArenaBox<U> {
|
|
ArenaBox {
|
|
ptr: f(&mut self),
|
|
valid: self.valid,
|
|
}
|
|
}
|
|
|
|
#[track_caller]
|
|
fn validate(&self) {
|
|
assert!(
|
|
self.valid.get(),
|
|
"attempted to dereference an ArenaRef after its Arena was cleared"
|
|
);
|
|
}
|
|
}
|
|
|
|
impl<T: ?Sized> Deref for ArenaBox<T> {
|
|
type Target = T;
|
|
|
|
#[inline(always)]
|
|
fn deref(&self) -> &Self::Target {
|
|
self.validate();
|
|
unsafe { &*self.ptr }
|
|
}
|
|
}
|
|
|
|
impl<T: ?Sized> DerefMut for ArenaBox<T> {
|
|
#[inline(always)]
|
|
fn deref_mut(&mut self) -> &mut Self::Target {
|
|
self.validate();
|
|
unsafe { &mut *self.ptr }
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use std::{cell::Cell, rc::Rc};
|
|
|
|
use super::*;
|
|
|
|
#[test]
|
|
fn test_arena() {
|
|
let mut arena = Arena::new(1024);
|
|
let a = arena.alloc(|| 1u64);
|
|
let b = arena.alloc(|| 2u32);
|
|
let c = arena.alloc(|| 3u16);
|
|
let d = arena.alloc(|| 4u8);
|
|
assert_eq!(*a, 1);
|
|
assert_eq!(*b, 2);
|
|
assert_eq!(*c, 3);
|
|
assert_eq!(*d, 4);
|
|
|
|
arena.clear();
|
|
let a = arena.alloc(|| 5u64);
|
|
let b = arena.alloc(|| 6u32);
|
|
let c = arena.alloc(|| 7u16);
|
|
let d = arena.alloc(|| 8u8);
|
|
assert_eq!(*a, 5);
|
|
assert_eq!(*b, 6);
|
|
assert_eq!(*c, 7);
|
|
assert_eq!(*d, 8);
|
|
|
|
// Ensure drop gets called.
|
|
let dropped = Rc::new(Cell::new(false));
|
|
struct DropGuard(Rc<Cell<bool>>);
|
|
impl Drop for DropGuard {
|
|
fn drop(&mut self) {
|
|
self.0.set(true);
|
|
}
|
|
}
|
|
arena.alloc(|| DropGuard(dropped.clone()));
|
|
arena.clear();
|
|
assert!(dropped.get());
|
|
}
|
|
|
|
#[test]
|
|
fn test_arena_grow() {
|
|
let mut arena = Arena::new(8);
|
|
arena.alloc(|| 1u64);
|
|
arena.alloc(|| 2u64);
|
|
|
|
assert_eq!(arena.capacity(), 16);
|
|
|
|
arena.alloc(|| 3u32);
|
|
arena.alloc(|| 4u32);
|
|
|
|
assert_eq!(arena.capacity(), 24);
|
|
}
|
|
|
|
#[test]
|
|
fn test_arena_alignment() {
|
|
let mut arena = Arena::new(256);
|
|
let x1 = arena.alloc(|| 1u8);
|
|
let x2 = arena.alloc(|| 2u16);
|
|
let x3 = arena.alloc(|| 3u32);
|
|
let x4 = arena.alloc(|| 4u64);
|
|
let x5 = arena.alloc(|| 5u64);
|
|
|
|
assert_eq!(*x1, 1);
|
|
assert_eq!(*x2, 2);
|
|
assert_eq!(*x3, 3);
|
|
assert_eq!(*x4, 4);
|
|
assert_eq!(*x5, 5);
|
|
|
|
assert_eq!(x1.ptr.align_offset(std::mem::align_of_val(&*x1)), 0);
|
|
assert_eq!(x2.ptr.align_offset(std::mem::align_of_val(&*x2)), 0);
|
|
}
|
|
|
|
#[test]
|
|
#[should_panic(expected = "attempted to dereference an ArenaRef after its Arena was cleared")]
|
|
fn test_arena_use_after_clear() {
|
|
let mut arena = Arena::new(16);
|
|
let value = arena.alloc(|| 1u64);
|
|
|
|
arena.clear();
|
|
let _read_value = *value;
|
|
}
|
|
}
|