From 6fb7c131ed48d33b8290ce2bf970c37fd9781828 Mon Sep 17 00:00:00 2001 From: Andy Wang Date: Sun, 5 Jun 2022 20:47:01 +0100 Subject: [PATCH] Remove unused lifetimes --- src/concurrency/data_race.rs | 6 +----- src/concurrency/weak_memory.rs | 2 +- 2 files changed, 2 insertions(+), 6 deletions(-) diff --git a/src/concurrency/data_race.rs b/src/concurrency/data_race.rs index f6f0ce528edb..28b09d2f909a 100644 --- a/src/concurrency/data_race.rs +++ b/src/concurrency/data_race.rs @@ -939,11 +939,7 @@ impl VClockAlloc { /// Detect racing atomic read and writes (not data races) /// on every byte of the current access range - pub(super) fn race_free_with_atomic<'tcx>( - &self, - range: AllocRange, - global: &GlobalState, - ) -> bool { + pub(super) fn race_free_with_atomic(&self, range: AllocRange, global: &GlobalState) -> bool { if global.race_detecting() { let (_, clocks) = global.current_thread_state(); let alloc_ranges = self.alloc_ranges.borrow(); diff --git a/src/concurrency/weak_memory.rs b/src/concurrency/weak_memory.rs index dc32a3ddca41..a771a09ed1ec 100644 --- a/src/concurrency/weak_memory.rs +++ b/src/concurrency/weak_memory.rs @@ -138,7 +138,7 @@ impl StoreBufferAlloc { /// before without data race, we can determine that the non-atomic access fully happens /// after all the prior atomic accesses so the location no longer needs to exhibit /// any weak memory behaviours until further atomic accesses. - pub fn memory_accessed<'tcx>(&self, range: AllocRange, global: &GlobalState) { + pub fn memory_accessed(&self, range: AllocRange, global: &GlobalState) { if !global.ongoing_action_data_race_free() { let mut buffers = self.store_buffers.borrow_mut(); let access_type = buffers.access_type(range);