From 3831e0db39280f7bfc3dd755c41ca1e525233a0e Mon Sep 17 00:00:00 2001 From: pedrocarlo Date: Mon, 28 Jul 2025 15:37:32 -0300 Subject: [PATCH] convert must_use compile warnings to `unused_variables` to track locations where we need to refactor in the future --- bindings/javascript/src/lib.rs | 6 +++--- bindings/rust/src/lib.rs | 2 +- bindings/wasm/lib.rs | 6 +++--- core/storage/btree.rs | 13 +++++++------ core/storage/database.rs | 12 ++++++------ core/storage/sqlite3_ondisk.rs | 10 +++++----- core/storage/wal.rs | 6 +++--- core/vdbe/execute.rs | 2 +- core/vdbe/sorter.rs | 4 ++-- simulator/runner/file.rs | 2 +- tests/integration/query_processing/test_btree.rs | 2 +- 11 files changed, 33 insertions(+), 32 deletions(-) diff --git a/bindings/javascript/src/lib.rs b/bindings/javascript/src/lib.rs index ecb5d84ff..0247f8eee 100644 --- a/bindings/javascript/src/lib.rs +++ b/bindings/javascript/src/lib.rs @@ -702,7 +702,7 @@ impl turso_core::DatabaseStorage for DatabaseFile { return Err(turso_core::LimboError::NotADB); } let pos = (page_idx - 1) * size; - self.file.pread(pos, c.into()) + self.file.pread(pos, c) } fn write_page( @@ -713,11 +713,11 @@ impl turso_core::DatabaseStorage for DatabaseFile { ) -> turso_core::Result { let size = buffer.borrow().len(); let pos = (page_idx - 1) * size; - self.file.pwrite(pos, buffer, c.into()) + self.file.pwrite(pos, buffer, c) } fn sync(&self, c: turso_core::Completion) -> turso_core::Result { - self.file.sync(c.into()) + self.file.sync(c) } fn size(&self) -> turso_core::Result { diff --git a/bindings/rust/src/lib.rs b/bindings/rust/src/lib.rs index a4f758308..3abaed46e 100644 --- a/bindings/rust/src/lib.rs +++ b/bindings/rust/src/lib.rs @@ -266,7 +266,7 @@ impl Connection { .inner .lock() .map_err(|e| Error::MutexError(e.to_string()))?; - conn.cacheflush()?; + let res = conn.cacheflush()?; Ok(()) } diff --git a/bindings/wasm/lib.rs b/bindings/wasm/lib.rs index e244f0c6b..bd08d535b 100644 --- a/bindings/wasm/lib.rs +++ b/bindings/wasm/lib.rs @@ -346,7 +346,7 @@ impl turso_core::DatabaseStorage for DatabaseFile { return Err(turso_core::LimboError::NotADB); } let pos = (page_idx - 1) * size; - self.file.pread(pos, c.into()) + self.file.pread(pos, c) } fn write_page( @@ -357,11 +357,11 @@ impl turso_core::DatabaseStorage for DatabaseFile { ) -> Result { let size = buffer.borrow().len(); let pos = (page_idx - 1) * size; - self.file.pwrite(pos, buffer, c.into()) + self.file.pwrite(pos, buffer, c) } fn sync(&self, c: turso_core::Completion) -> Result { - self.file.sync(c.into()) + self.file.sync(c) } fn size(&self) -> Result { diff --git a/core/storage/btree.rs b/core/storage/btree.rs index 32f512324..bf3277f61 100644 --- a/core/storage/btree.rs +++ b/core/storage/btree.rs @@ -2443,7 +2443,7 @@ impl BTreeCursor { } if !self.stack.has_parent() { - self.balance_root()?; + let res = self.balance_root()?; } let write_info = self.state.mut_write_info().unwrap(); @@ -5256,7 +5256,8 @@ impl BTreeCursor { let new_payload = &mut *new_payload; // if it all fits in local space and old_local_size is enough, do an in-place overwrite if new_payload.len() == *old_local_size { - self.overwrite_content(page_ref.clone(), *old_offset, new_payload)?; + let res = + self.overwrite_content(page_ref.clone(), *old_offset, new_payload)?; return Ok(IOResult::Done(())); } @@ -7751,7 +7752,7 @@ mod tests { tracing::info!("seed: {seed}"); for i in 0..inserts { pager.begin_read_tx().unwrap(); - pager.begin_write_tx().unwrap(); + let res = pager.begin_write_tx().unwrap(); let key = { let result; loop { @@ -7921,7 +7922,7 @@ mod tests { for i in 0..operations { let print_progress = i % 100 == 0; pager.begin_read_tx().unwrap(); - pager.begin_write_tx().unwrap(); + let res = pager.begin_write_tx().unwrap(); // Decide whether to insert or delete (80% chance of insert) let is_insert = rng.next_u64() % 100 < (insert_chance * 100.0) as u64; @@ -8302,7 +8303,7 @@ mod tests { let _ = run_until_done(|| pager.allocate_page1(), &pager); for _ in 0..(database_size - 1) { - pager.allocate_page().unwrap(); + let res = pager.allocate_page().unwrap(); } header_accessor::set_page_size(&pager, page_size).unwrap(); @@ -8334,7 +8335,7 @@ mod tests { ))); let c = Completion::new_write(|_| {}); #[allow(clippy::arc_with_non_send_sync)] - pager + let c = pager .db_file .write_page(current_page as usize, buf.clone(), c)?; pager.io.run_once()?; diff --git a/core/storage/database.rs b/core/storage/database.rs index 75278dbdb..8e539a5f3 100644 --- a/core/storage/database.rs +++ b/core/storage/database.rs @@ -41,7 +41,7 @@ impl DatabaseStorage for DatabaseFile { return Err(LimboError::NotADB); } let pos = (page_idx - 1) * size; - self.file.pread(pos, c.into()) + self.file.pread(pos, c) } #[instrument(skip_all, level = Level::DEBUG)] @@ -57,12 +57,12 @@ impl DatabaseStorage for DatabaseFile { assert!(buffer_size <= 65536); assert_eq!(buffer_size & (buffer_size - 1), 0); let pos = (page_idx - 1) * buffer_size; - self.file.pwrite(pos, buffer, c.into()) + self.file.pwrite(pos, buffer, c) } #[instrument(skip_all, level = Level::DEBUG)] fn sync(&self, c: Completion) -> Result { - self.file.sync(c.into()) + self.file.sync(c) } #[instrument(skip_all, level = Level::DEBUG)] @@ -95,7 +95,7 @@ impl DatabaseStorage for FileMemoryStorage { return Err(LimboError::NotADB); } let pos = (page_idx - 1) * size; - self.file.pread(pos, c.into()) + self.file.pread(pos, c) } #[instrument(skip_all, level = Level::DEBUG)] @@ -110,12 +110,12 @@ impl DatabaseStorage for FileMemoryStorage { assert!(buffer_size <= 65536); assert_eq!(buffer_size & (buffer_size - 1), 0); let pos = (page_idx - 1) * buffer_size; - self.file.pwrite(pos, buffer, c.into()) + self.file.pwrite(pos, buffer, c) } #[instrument(skip_all, level = Level::DEBUG)] fn sync(&self, c: Completion) -> Result { - self.file.sync(c.into()) + self.file.sync(c) } #[instrument(skip_all, level = Level::DEBUG)] diff --git a/core/storage/sqlite3_ondisk.rs b/core/storage/sqlite3_ondisk.rs index 5a1ab7305..83f669a0d 100644 --- a/core/storage/sqlite3_ondisk.rs +++ b/core/storage/sqlite3_ondisk.rs @@ -860,7 +860,7 @@ pub fn begin_sync(db_file: Arc, syncing: Rc>) *syncing.borrow_mut() = false; }); #[allow(clippy::arc_with_non_send_sync)] - db_file.sync(completion)?; + let c = db_file.sync(completion)?; Ok(()) } @@ -1565,7 +1565,7 @@ pub fn read_entire_wal_dumb(file: &Arc) -> Result, header: &WalHeader) -> Result< }; #[allow(clippy::arc_with_non_send_sync)] let c = Completion::new_write(write_complete); - io.pwrite(0, buffer.clone(), c.into())?; + let c = io.pwrite(0, buffer.clone(), c)?; Ok(()) } diff --git a/core/storage/wal.rs b/core/storage/wal.rs index 705484e6d..af470eb1b 100644 --- a/core/storage/wal.rs +++ b/core/storage/wal.rs @@ -626,7 +626,7 @@ impl Wal for WalFile { let frame = frame.clone(); finish_read_page(page.get().id, buf, frame).unwrap(); }); - begin_read_wal_frame( + let c = begin_read_wal_frame( &self.get_shared().file, offset + WAL_FRAME_HEADER_SIZE, buffer_pool, @@ -784,7 +784,7 @@ impl Wal for WalFile { *write_counter.borrow_mut() -= 1; } }); - let result = shared.file.pwrite(offset, frame_bytes.clone(), c.into()); + let result = shared.file.pwrite(offset, frame_bytes.clone(), c); if let Err(err) = result { *write_counter.borrow_mut() -= 1; return Err(err); @@ -1001,7 +1001,7 @@ impl Wal for WalFile { syncing.set(false); }); let shared = self.get_shared(); - shared.file.sync(completion.into())?; + let c = shared.file.sync(completion)?; self.sync_state.set(SyncState::Syncing); Ok(IOResult::IO) } diff --git a/core/vdbe/execute.rs b/core/vdbe/execute.rs index af34c58a4..65ed2796f 100644 --- a/core/vdbe/execute.rs +++ b/core/vdbe/execute.rs @@ -6356,7 +6356,7 @@ pub fn op_open_ephemeral( } else { BTreeCursor::new_table(mv_cursor, pager.clone(), root_page as usize, num_columns) }; - cursor.rewind()?; // Will never return io + let res = cursor.rewind()?; // Will never return io let mut cursors: std::cell::RefMut<'_, Vec>> = state.cursors.borrow_mut(); diff --git a/core/vdbe/sorter.rs b/core/vdbe/sorter.rs index dd84fa195..72edca61a 100644 --- a/core/vdbe/sorter.rs +++ b/core/vdbe/sorter.rs @@ -402,7 +402,7 @@ impl SortedChunk { read_buffer_ref, read_complete, ))); - self.file.pread(self.total_bytes_read.get(), c)?; + let c = self.file.pread(self.total_bytes_read.get(), c)?; Ok(()) } @@ -448,7 +448,7 @@ impl SortedChunk { }); let c = Completion::new(CompletionType::Write(WriteCompletion::new(write_complete))); - self.file.pwrite(0, buffer_ref, c)?; + let c = self.file.pwrite(0, buffer_ref, c)?; Ok(()) } } diff --git a/simulator/runner/file.rs b/simulator/runner/file.rs index 3905a3ce7..1be2cb48b 100644 --- a/simulator/runner/file.rs +++ b/simulator/runner/file.rs @@ -121,7 +121,7 @@ impl SimulatorFile { if queued_io[i].time <= now { let io = queued_io.remove(i); // your code here - (io.op)(self)?; + let c = (io.op)(self)?; } else { i += 1; } diff --git a/tests/integration/query_processing/test_btree.rs b/tests/integration/query_processing/test_btree.rs index 4dc19f1c8..a50100b42 100644 --- a/tests/integration/query_processing/test_btree.rs +++ b/tests/integration/query_processing/test_btree.rs @@ -433,7 +433,7 @@ fn write_at(io: &impl IO, file: Arc, offset: usize, data: &[u8]) { let drop_fn = Rc::new(move |_| {}); #[allow(clippy::arc_with_non_send_sync)] let buffer = Arc::new(RefCell::new(Buffer::new(Pin::new(data.to_vec()), drop_fn))); - let result = file.pwrite(offset, buffer, completion.into()).unwrap(); + let result = file.pwrite(offset, buffer, completion).unwrap(); while !result.is_completed() { io.run_once().unwrap(); }