mirror of
https://github.com/aljazceru/turso.git
synced 2025-12-25 20:14:21 +01:00
We assumed page was loaded because before inserting we would move there. `NewRowId` unfortunately moves cursor to the rightmost page causing eviction of root page -- this arose the issue with `insert_into_page` not loading the page we were supposed to have loaded so I added `return_if_locked_maybe_load` which is a utility macro to check if the page is locked and if not, load it if needed.
347 lines
11 KiB
Rust
347 lines
11 KiB
Rust
use std::{cell::RefCell, collections::HashMap, ptr::NonNull};
|
|
|
|
use tracing::{debug, trace};
|
|
|
|
use super::pager::PageRef;
|
|
|
|
// In limbo, page cache is shared by default, meaning that multiple frames from WAL can reside in
|
|
// the cache, meaning, we need a way to differentiate between pages cached in different
|
|
// connections. For this we include the max_frame that a connection will read from so that if two
|
|
// connections have different max_frames, they might or not have different frame read from WAL.
|
|
//
|
|
// WAL was introduced after Shared cache in SQLite, so this is why these two features don't work
|
|
// well together because pages with different snapshots may collide.
|
|
#[derive(Debug, Eq, Hash, PartialEq, Clone)]
|
|
pub struct PageCacheKey {
|
|
pgno: usize,
|
|
max_frame: Option<u64>,
|
|
}
|
|
|
|
#[allow(dead_code)]
|
|
struct PageCacheEntry {
|
|
key: PageCacheKey,
|
|
page: PageRef,
|
|
prev: Option<NonNull<PageCacheEntry>>,
|
|
next: Option<NonNull<PageCacheEntry>>,
|
|
}
|
|
|
|
impl PageCacheEntry {
|
|
fn as_non_null(&mut self) -> NonNull<PageCacheEntry> {
|
|
NonNull::new(&mut *self).unwrap()
|
|
}
|
|
}
|
|
|
|
pub struct DumbLruPageCache {
|
|
capacity: usize,
|
|
map: RefCell<HashMap<PageCacheKey, NonNull<PageCacheEntry>>>,
|
|
head: RefCell<Option<NonNull<PageCacheEntry>>>,
|
|
tail: RefCell<Option<NonNull<PageCacheEntry>>>,
|
|
}
|
|
unsafe impl Send for DumbLruPageCache {}
|
|
unsafe impl Sync for DumbLruPageCache {}
|
|
|
|
impl PageCacheKey {
|
|
pub fn new(pgno: usize, max_frame: Option<u64>) -> Self {
|
|
Self { pgno, max_frame }
|
|
}
|
|
}
|
|
impl DumbLruPageCache {
|
|
pub fn new(capacity: usize) -> Self {
|
|
Self {
|
|
capacity,
|
|
map: RefCell::new(HashMap::new()),
|
|
head: RefCell::new(None),
|
|
tail: RefCell::new(None),
|
|
}
|
|
}
|
|
|
|
pub fn contains_key(&mut self, key: &PageCacheKey) -> bool {
|
|
self.map.borrow().contains_key(key)
|
|
}
|
|
|
|
pub fn insert(&mut self, key: PageCacheKey, value: PageRef) {
|
|
self._delete(key.clone(), false);
|
|
trace!("cache_insert(key={:?})", key);
|
|
let entry = Box::new(PageCacheEntry {
|
|
key: key.clone(),
|
|
next: None,
|
|
prev: None,
|
|
page: value,
|
|
});
|
|
let ptr_raw = Box::into_raw(entry);
|
|
let ptr = unsafe { ptr_raw.as_mut().unwrap().as_non_null() };
|
|
self.touch(ptr);
|
|
|
|
self.map.borrow_mut().insert(key, ptr);
|
|
if self.len() > self.capacity {
|
|
self.pop_if_not_dirty();
|
|
}
|
|
}
|
|
|
|
pub fn delete(&mut self, key: PageCacheKey) {
|
|
trace!("cache_delete(key={:?})", key);
|
|
self._delete(key, true)
|
|
}
|
|
|
|
pub fn _delete(&mut self, key: PageCacheKey, clean_page: bool) {
|
|
let ptr = self.map.borrow_mut().remove(&key);
|
|
if ptr.is_none() {
|
|
return;
|
|
}
|
|
let ptr = ptr.unwrap();
|
|
self.detach(ptr, clean_page);
|
|
unsafe { std::ptr::drop_in_place(ptr.as_ptr()) };
|
|
}
|
|
|
|
fn get_ptr(&mut self, key: &PageCacheKey) -> Option<NonNull<PageCacheEntry>> {
|
|
let m = self.map.borrow_mut();
|
|
let ptr = m.get(key);
|
|
ptr.copied()
|
|
}
|
|
|
|
pub fn get(&mut self, key: &PageCacheKey) -> Option<PageRef> {
|
|
self.peek(key, true)
|
|
}
|
|
|
|
/// Get page without promoting entry
|
|
pub fn peek(&mut self, key: &PageCacheKey, touch: bool) -> Option<PageRef> {
|
|
trace!("cache_get(key={:?})", key);
|
|
let mut ptr = self.get_ptr(key)?;
|
|
let page = unsafe { ptr.as_mut().page.clone() };
|
|
if touch {
|
|
self.detach(ptr, false);
|
|
self.touch(ptr);
|
|
}
|
|
Some(page)
|
|
}
|
|
|
|
pub fn resize(&mut self, capacity: usize) {
|
|
let _ = capacity;
|
|
todo!();
|
|
}
|
|
|
|
fn detach(&mut self, mut entry: NonNull<PageCacheEntry>, clean_page: bool) {
|
|
if clean_page {
|
|
// evict buffer
|
|
let page = unsafe { &entry.as_mut().page };
|
|
page.clear_loaded();
|
|
debug!("cleaning up page {}", page.get().id);
|
|
let _ = page.get().contents.take();
|
|
}
|
|
|
|
let (next, prev) = unsafe {
|
|
let c = entry.as_mut();
|
|
let next = c.next;
|
|
let prev = c.prev;
|
|
c.prev = None;
|
|
c.next = None;
|
|
(next, prev)
|
|
};
|
|
|
|
// detach
|
|
match (prev, next) {
|
|
(None, None) => {
|
|
self.head.replace(None);
|
|
self.tail.replace(None);
|
|
}
|
|
(None, Some(mut n)) => {
|
|
unsafe { n.as_mut().prev = None };
|
|
self.head.borrow_mut().replace(n);
|
|
}
|
|
(Some(mut p), None) => {
|
|
unsafe { p.as_mut().next = None };
|
|
self.tail = RefCell::new(Some(p));
|
|
}
|
|
(Some(mut p), Some(mut n)) => unsafe {
|
|
let p_mut = p.as_mut();
|
|
p_mut.next = Some(n);
|
|
let n_mut = n.as_mut();
|
|
n_mut.prev = Some(p);
|
|
},
|
|
};
|
|
}
|
|
|
|
/// inserts into head, assuming we detached first
|
|
fn touch(&mut self, mut entry: NonNull<PageCacheEntry>) {
|
|
if let Some(mut head) = *self.head.borrow_mut() {
|
|
unsafe {
|
|
entry.as_mut().next.replace(head);
|
|
let head = head.as_mut();
|
|
head.prev = Some(entry);
|
|
}
|
|
}
|
|
|
|
if self.tail.borrow().is_none() {
|
|
self.tail.borrow_mut().replace(entry);
|
|
}
|
|
self.head.borrow_mut().replace(entry);
|
|
}
|
|
|
|
fn pop_if_not_dirty(&mut self) {
|
|
let tail = *self.tail.borrow();
|
|
if tail.is_none() {
|
|
return;
|
|
}
|
|
let mut tail = tail.unwrap();
|
|
let tail_entry = unsafe { tail.as_mut() };
|
|
if tail_entry.page.is_dirty() {
|
|
// TODO: drop from another clean entry?
|
|
return;
|
|
}
|
|
tracing::debug!("pop_if_not_dirty(key={:?})", tail_entry.key);
|
|
self.detach(tail, true);
|
|
assert!(self.map.borrow_mut().remove(&tail_entry.key).is_some());
|
|
}
|
|
|
|
pub fn clear(&mut self) {
|
|
let to_remove: Vec<PageCacheKey> = self.map.borrow().iter().map(|v| v.0.clone()).collect();
|
|
for key in to_remove {
|
|
self.delete(key);
|
|
}
|
|
}
|
|
|
|
pub fn print(&mut self) {
|
|
println!("page_cache={}", self.map.borrow().len());
|
|
println!("page_cache={:?}", self.map.borrow())
|
|
}
|
|
|
|
pub fn len(&self) -> usize {
|
|
self.map.borrow().len()
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use std::{num::NonZeroUsize, sync::Arc};
|
|
|
|
use lru::LruCache;
|
|
use rand_chacha::{
|
|
rand_core::{RngCore, SeedableRng},
|
|
ChaCha8Rng,
|
|
};
|
|
|
|
use crate::{storage::page_cache::DumbLruPageCache, Page};
|
|
|
|
use super::PageCacheKey;
|
|
|
|
#[test]
|
|
fn test_page_cache_evict() {
|
|
let mut cache = DumbLruPageCache::new(1);
|
|
let key1 = insert_page(&mut cache, 1);
|
|
let key2 = insert_page(&mut cache, 2);
|
|
assert_eq!(cache.get(&key2).unwrap().get().id, 2);
|
|
assert!(cache.get(&key1).is_none());
|
|
}
|
|
|
|
#[test]
|
|
fn test_page_cache_fuzz() {
|
|
let seed = std::time::SystemTime::now()
|
|
.duration_since(std::time::UNIX_EPOCH)
|
|
.unwrap()
|
|
.as_secs();
|
|
let mut rng = ChaCha8Rng::seed_from_u64(seed);
|
|
tracing::info!("super seed: {}", seed);
|
|
let max_pages = 10;
|
|
let mut cache = DumbLruPageCache::new(10);
|
|
let mut lru = LruCache::new(NonZeroUsize::new(10).unwrap());
|
|
|
|
for _ in 0..10000 {
|
|
match rng.next_u64() % 3 {
|
|
0 => {
|
|
// add
|
|
let id_page = rng.next_u64() % max_pages;
|
|
let id_frame = rng.next_u64() % max_pages;
|
|
let key = PageCacheKey::new(id_page as usize, Some(id_frame));
|
|
#[allow(clippy::arc_with_non_send_sync)]
|
|
let page = Arc::new(Page::new(id_page as usize));
|
|
// println!("inserting page {:?}", key);
|
|
cache.insert(key.clone(), page.clone());
|
|
lru.push(key, page);
|
|
assert!(cache.len() <= 10);
|
|
}
|
|
1 => {
|
|
// remove
|
|
let random = rng.next_u64() % 2 == 0;
|
|
let key = if random || lru.is_empty() {
|
|
let id_page = rng.next_u64() % max_pages;
|
|
let id_frame = rng.next_u64() % max_pages;
|
|
let key = PageCacheKey::new(id_page as usize, Some(id_frame));
|
|
key
|
|
} else {
|
|
let i = rng.next_u64() as usize % lru.len();
|
|
let key = lru.iter().skip(i).next().unwrap().0.clone();
|
|
key
|
|
};
|
|
// println!("removing page {:?}", key);
|
|
lru.pop(&key);
|
|
cache.delete(key);
|
|
}
|
|
2 => {
|
|
// test contents
|
|
for (key, page) in &lru {
|
|
// println!("getting page {:?}", key);
|
|
cache.peek(&key, false).unwrap();
|
|
assert_eq!(page.get().id, key.pgno);
|
|
}
|
|
}
|
|
_ => unreachable!(),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn test_page_cache_insert_and_get() {
|
|
let mut cache = DumbLruPageCache::new(2);
|
|
let key1 = insert_page(&mut cache, 1);
|
|
let key2 = insert_page(&mut cache, 2);
|
|
assert_eq!(cache.get(&key1).unwrap().get().id, 1);
|
|
assert_eq!(cache.get(&key2).unwrap().get().id, 2);
|
|
}
|
|
|
|
#[test]
|
|
fn test_page_cache_over_capacity() {
|
|
let mut cache = DumbLruPageCache::new(2);
|
|
let key1 = insert_page(&mut cache, 1);
|
|
let key2 = insert_page(&mut cache, 2);
|
|
let key3 = insert_page(&mut cache, 3);
|
|
assert!(cache.get(&key1).is_none());
|
|
assert_eq!(cache.get(&key2).unwrap().get().id, 2);
|
|
assert_eq!(cache.get(&key3).unwrap().get().id, 3);
|
|
}
|
|
|
|
#[test]
|
|
fn test_page_cache_delete() {
|
|
let mut cache = DumbLruPageCache::new(2);
|
|
let key1 = insert_page(&mut cache, 1);
|
|
cache.delete(key1.clone());
|
|
assert!(cache.get(&key1).is_none());
|
|
}
|
|
|
|
#[test]
|
|
fn test_page_cache_clear() {
|
|
let mut cache = DumbLruPageCache::new(2);
|
|
let key1 = insert_page(&mut cache, 1);
|
|
let key2 = insert_page(&mut cache, 2);
|
|
cache.clear();
|
|
assert!(cache.get(&key1).is_none());
|
|
assert!(cache.get(&key2).is_none());
|
|
}
|
|
|
|
fn insert_page(cache: &mut DumbLruPageCache, id: usize) -> PageCacheKey {
|
|
let key = PageCacheKey::new(id, None);
|
|
#[allow(clippy::arc_with_non_send_sync)]
|
|
let page = Arc::new(Page::new(id));
|
|
cache.insert(key.clone(), page.clone());
|
|
key
|
|
}
|
|
|
|
#[test]
|
|
fn test_page_cache_insert_sequential() {
|
|
let mut cache = DumbLruPageCache::new(2);
|
|
for i in 0..10000 {
|
|
let key = insert_page(&mut cache, i);
|
|
assert_eq!(cache.peek(&key, false).unwrap().get().id, i);
|
|
}
|
|
}
|
|
}
|