From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id B33E638423A; Fri, 3 Apr 2026 11:07:07 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1775214427; cv=none; b=Qjy6zbxsveQCza4IRibBKF/V3rs+PfuGUw+M5xDwOEUs+hIsiZWxkPUM55aM5vKSTLqbddGKEk0B41jjUmOCGyQnXxBWWYw2iRWe5CcNQqwO7f9JD7XIeZz9hXVgwu4iaiNZ+NHEADdongR4grT8+WCajFm92+1RjlFQwgLUifo= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1775214427; c=relaxed/simple; bh=6SLhoBxw3+UYcVaok8LUtY/DUiO9M8L+BkRyzKi2j+g=; h=From:Date:Subject:MIME-Version:Content-Type:Message-Id:References: In-Reply-To:To:Cc; b=f3a6PGEt8f0+9Fpt/b0t3K6XmdtiPJ2a7ZkgwtCOAt3HfSdWTGiTcneW/Pd98GCNx5Kkwdxb7NN/PvkhMThIsmlMpLHqZz0Ogw2cAXW4oG2uNl8eKTxHAXa8fU8in/c213cBWMG8Fg8KXMYiKL5vvVQZKMe8/x/8d7Qw0wUBaK0= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=qVQTtsNZ; arc=none smtp.client-ip=10.30.226.201 Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="qVQTtsNZ" Received: by smtp.kernel.org (Postfix) with ESMTPS id 7E9ECC19423; Fri, 3 Apr 2026 11:07:07 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1775214427; bh=6SLhoBxw3+UYcVaok8LUtY/DUiO9M8L+BkRyzKi2j+g=; h=From:Date:Subject:References:In-Reply-To:To:Cc:Reply-To:From; b=qVQTtsNZKOFhhY2qO/SsbSWGkhkXVwQ6Q88TPES6dI5ADS0MhXZEytreVUjHvhwAj hZChR6IefANSXNf7uPLAQu9shBxmGJL6DNPphnRBdmPoWYZ7Ga+4qxOcz8wo+hlngm 6Cso6UvP1sgkphOyDC0FnucPw7AKcSlXxAj0HuEPGS/HN0mjVIwXiS1ovfolv3pmMB +Rd5DqWRg2L8fCiZ1tPT5BYJhi2CIy4tu8FFbs6RVHbU6X7Ad0FNqYDtR1b6CLxuz4 2VXF6Pcy3MCjjjoUtH0JItEYbNSbP0fKC3yeoX7+P4SAm5McPNsTVHjgDIBPvBjwUn GlBQD+t+AZxrg== Received: from aws-us-west-2-korg-lkml-1.web.codeaurora.org (localhost.localdomain [127.0.0.1]) by smtp.lore.kernel.org (Postfix) with ESMTP id 7544AE7E373; Fri, 3 Apr 2026 11:07:07 +0000 (UTC) From: Aakash Bollineni via B4 Relay Date: Fri, 03 Apr 2026 16:37:06 +0530 Subject: [PATCH v3 2/3] rust: workqueue: add safe cancellation and status methods Precedence: bulk X-Mailing-List: rust-for-linux@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: 7bit Message-Id: <20260403-workqueue-v3-final-v3-2-6ffc3950d804@multicorewareinc.com> References: <20260403-workqueue-v3-final-v3-0-6ffc3950d804@multicorewareinc.com> In-Reply-To: <20260403-workqueue-v3-final-v3-0-6ffc3950d804@multicorewareinc.com> To: Miguel Ojeda , Boqun Feng , Gary Guo , =?utf-8?q?Bj=C3=B6rn_Roy_Baron?= , Benno Lossin , Andreas Hindborg , Alice Ryhl , Trevor Gross , Danilo Krummrich Cc: rust-for-linux@vger.kernel.org, linux-kernel@vger.kernel.org, Aakash Bollineni , work@onurozkan.dev X-Mailer: b4 0.13.0 X-Developer-Signature: v=1; a=ed25519-sha256; t=1775214425; l=23363; i=aakash.bollineni@multicorewareinc.com; s=20260402; h=from:subject:message-id; bh=TPAN5xuWvuCvSNJpkbWtXOedWR9hrRXmuBf7M6mPSPw=; b=5VUfwWMMy+eO5ZmnsdE1d97G2wsiqypJn+LmPL6/boxw1lQiSotHHMsm/AwQgD0fusWv2KCmP WfueOF8DfQqDwNhkfGlzSJJQtKLgPSMFY6dIg9ALRjyL7KPniUe0+pP X-Developer-Key: i=aakash.bollineni@multicorewareinc.com; a=ed25519; pk=r3Gonl+2k+8RozN9U/XwfICQdnRlAcLeeAfsExmurdE= X-Endpoint-Received: by B4 Relay for aakash.bollineni@multicorewareinc.com/20260402 with auth_id=711 X-Original-From: Aakash Bollineni Reply-To: aakash.bollineni@multicorewareinc.com From: Aakash Bollineni Modernize the Rust workqueue by adding methods for status checking and cancellation of work and delayed work items. Specifically, this patch adds: - is_pending(): Returns true if the work item is currently enqueued. - cancel(): Attempts to cancel the work item before it runs. If successful, it reclaims and returns ownership of the original pointer (Arc/KBox). - cancel_sync(): Synchronously cancels the work item, waiting for it to finish if it's already running. Reclaims ownership if the work was pending. To support safe pointer reclamation, a new "Handle-based" model is introduced via the WorkItemPointer::reclaim method. This ensures that the "leaked" ownership from an enqueue operation is safely recovered without memory leaks. Additionally, we enforce #[repr(transparent)] on core work wrappers to ensure memory layout compatibility with C pointers and fix pointer arithmetic in the work_container_of implementation. Signed-off-by: Aakash Bollineni --- rust/kernel/workqueue.rs | 473 +++++++++++++++++++++++++++++++++++++++++++---- 1 file changed, 438 insertions(+), 35 deletions(-) diff --git a/rust/kernel/workqueue.rs b/rust/kernel/workqueue.rs index 706e833e9702..eec3210bbcc8 100644 --- a/rust/kernel/workqueue.rs +++ b/rust/kernel/workqueue.rs @@ -448,8 +448,26 @@ pub unsafe trait WorkItemPointer: RawWorkItem { /// The provided `work_struct` pointer must originate from a previous call to [`__enqueue`] /// where the `queue_work_on` closure returned true, and the pointer must still be valid. /// + /// The implementation must ensure that the pointer is reclaimed (e.g., via `from_raw`) + /// before calling the `run` method of the underlying [`WorkItem`]. + /// /// [`__enqueue`]: RawWorkItem::__enqueue unsafe extern "C" fn run(ptr: *mut bindings::work_struct); + + /// Reclaims ownership of the pointer from the work item. + /// + /// This is called when a work item is successfully cancelled, allowing the caller + /// to recover the original pointer (e.g., `Arc` or `KBox`) that was "leaked" + /// during enqueuing. + /// + /// # Safety + /// + /// The provided `work_struct` pointer must originate from a previous call to [`__enqueue`] + /// where the `queue_work_on` closure returned true, and the work item must have been + /// successfully cancelled (i.e., `cancel_work` returned true). + /// + /// [`__enqueue`]: RawWorkItem::__enqueue + unsafe fn reclaim(ptr: *mut bindings::work_struct) -> Self; } /// Defines the method that should be called when this work item is executed. @@ -516,6 +534,156 @@ pub fn new(name: &'static CStr, key: Pin<&'static LockClassKey>) -> impl PinInit }) } + /// Returns whether the work item is currently pending. + /// + /// # Warning + /// + /// This method is inherently racy. A work item can be enqueued or start running + /// immediately after this check returns. Do not rely on this for correctness + /// logic (e.g., as a guard for unsafe operations); use it only for debugging or + /// non-critical status reporting. + /// + /// # Examples + /// + /// ``` + /// # use kernel::workqueue::{self, new_work, Work, WorkItem, HasWork}; + /// # use kernel::impl_has_work; + /// # use kernel::sync::Arc; + /// # #[pin_data] + /// # struct MyStruct { #[pin] work: Work } + /// # impl_has_work! { impl HasWork for MyStruct { self.work } } + /// # impl WorkItem for MyStruct { + /// # type Pointer = Arc; + /// # fn run(_this: Arc) {} + /// # } + /// let my_struct = Arc::pin_init(pin_init!(MyStruct { + /// work <- new_work!("MyStruct::work"), + /// }), kernel::alloc::flags::GFP_KERNEL).unwrap(); + /// assert!(!my_struct.work.is_pending()); + /// workqueue::system().enqueue(my_struct.clone()); + /// assert!(my_struct.work.is_pending()); + /// # let _ = my_struct.work.cancel(); + /// ``` + #[inline] + pub fn is_pending(&self) -> bool { + // SAFETY: `self.work` is a valid pointer to a `work_struct`. + unsafe { bindings::work_pending(self.work.get()) } + } + + /// Cancels the work item. + /// + /// If the work item was successfully cancelled (i.e., it was pending and had not yet + /// started running), the original pointer is reclaimed and returned. + /// + /// # Guarantees + /// + /// This method is non-blocking and may return while the work item is still running + /// on another CPU. If it returns `None`, the work item might be about to start, + /// is currently running, or has already finished. + /// + /// # Safety + /// + /// This is safe because ownership is only reclaimed if the kernel confirms (via + /// `cancel_work` returning true) that the work item is no longer in any queue and + /// will not be executed by the workqueue for this specific enqueue event. + /// + /// [`cancel_sync`]: Work::cancel_sync + /// + /// # Examples + /// + /// ``` + /// # use kernel::workqueue::{self, new_work, Work, WorkItem, HasWork}; + /// # use kernel::impl_has_work; + /// # use kernel::sync::Arc; + /// # #[pin_data] + /// # struct MyStruct { #[pin] work: Work } + /// # impl_has_work! { impl HasWork for MyStruct { self.work } } + /// # impl WorkItem for MyStruct { + /// # type Pointer = Arc; + /// # fn run(_this: Arc) {} + /// # } + /// let my_struct = Arc::pin_init(pin_init!(MyStruct { + /// work <- new_work!("MyStruct::work"), + /// }), kernel::alloc::flags::GFP_KERNEL).unwrap(); + /// workqueue::system().enqueue(my_struct.clone()); + /// assert!(my_struct.work.is_pending()); + /// let reclaimed = my_struct.work.cancel(); + /// assert!(reclaimed.is_some()); + /// assert!(!my_struct.work.is_pending()); + /// ``` + pub fn cancel(&self) -> Option + where + T: WorkItem, + { + let work_ptr = self.work.get(); + // SAFETY: `work_ptr` is a valid pointer to a `work_struct`. + if unsafe { bindings::cancel_work(work_ptr) } { + // SAFETY: The work item was successfully cancelled and is guaranteed not to run, + // so we can safely reclaim the ownership leaked during `enqueue`. + Some(unsafe { T::Pointer::reclaim(work_ptr) }) + } else { + None + } + } + + /// Synchronously cancels the work item. + /// + /// This method waits for the work item to finish if it is currently running. + /// If the work item was successfully cancelled from the queue, the pointer is + /// reclaimed and returned. + /// + /// # Guarantees + /// + /// After this method returns, the work item is guaranteed to be: + /// - Not pending in any queue. + /// - Not running on any CPU. + /// + /// This makes it safe to use during teardown (e.g., driver `remove` or object `drop`) + /// to ensure no background tasks are accessing resources that are about to be freed. + /// + /// # Safety + /// + /// Same as [`cancel`], it only reclaims ownership if the kernel confirms the work + /// was still pending and is now removed. + /// + /// [`cancel`]: Work::cancel + /// + /// # Examples + /// + /// ``` + /// # use kernel::workqueue::{self, new_work, Work, WorkItem, HasWork}; + /// # use kernel::impl_has_work; + /// # use kernel::sync::Arc; + /// # #[pin_data] + /// # struct MyStruct { #[pin] work: Work } + /// # impl_has_work! { impl HasWork for MyStruct { self.work } } + /// # impl WorkItem for MyStruct { + /// # type Pointer = Arc; + /// # fn run(_this: Arc) {} + /// # } + /// let my_struct = Arc::pin_init(pin_init!(MyStruct { + /// work <- new_work!("MyStruct::work"), + /// }), kernel::alloc::flags::GFP_KERNEL).unwrap(); + /// workqueue::system().enqueue(my_struct.clone()); + /// let reclaimed = my_struct.work.cancel_sync(); + /// assert!(reclaimed.is_some()); + /// ``` + pub fn cancel_sync(&self) -> Option + where + T: WorkItem, + { + let work_ptr = self.work.get(); + // SAFETY: `work_ptr` is a valid pointer to a `work_struct`. + if unsafe { bindings::cancel_work_sync(work_ptr) } { + // SAFETY: The work item was successfully cancelled/waited for, and is guaranteed + // not to run again unless re-enqueued. We reclaim the ownership leaked during + // `enqueue`. + Some(unsafe { T::Pointer::reclaim(work_ptr) }) + } else { + None + } + } + /// Get a pointer to the inner `work_struct`. /// /// # Safety @@ -674,25 +842,14 @@ pub fn new( pin_init!(Self { dwork <- Opaque::ffi_init(|slot: *mut bindings::delayed_work| { // SAFETY: The `WorkItemPointer` implementation promises that `run` can be used as - // the work item function. + // the work item function. We use the C-helper to ensure the timer function + // and data are initialized correctly according to kernel macros. unsafe { - bindings::init_work_with_key( - core::ptr::addr_of_mut!((*slot).work), + bindings::init_delayed_work( + slot, Some(T::Pointer::run), - false, work_name.as_char_ptr(), work_key.as_ptr(), - ) - } - - // SAFETY: The `delayed_work_timer_fn` function pointer can be used here because - // the timer is embedded in a `struct delayed_work`, and only ever scheduled via - // the core workqueue code, and configured to run in irqsafe context. - unsafe { - bindings::timer_init_key( - core::ptr::addr_of_mut!((*slot).timer), - Some(bindings::delayed_work_timer_fn), - bindings::TIMER_IRQSAFE, timer_name.as_char_ptr(), timer_key.as_ptr(), ) @@ -702,6 +859,89 @@ pub fn new( }) } + /// Returns whether the work item is currently pending. + /// + /// # Warning + /// + /// This method is inherently racy. See [`Work::is_pending`]. + /// + /// # Examples + /// + /// See [`Work::is_pending`]. + /// + /// [`Work::is_pending`]: Work::is_pending + #[inline] + pub fn is_pending(&self) -> bool { + // SAFETY: `self.dwork` is reaching into a valid Opaque. + unsafe { + let ptr: *mut bindings::delayed_work = self.dwork.get(); + bindings::work_pending(core::ptr::addr_of_mut!((*ptr).work)) + } + } + + /// Cancels the delayed work item. + /// + /// If the work item was successfully cancelled (i.e., it was pending and had not yet + /// started running), the original pointer is reclaimed and returned. + /// + /// # Guarantees + /// + /// See [`Work::cancel`]. + /// + /// # Safety + /// + /// Same as [`Work::cancel`]. + /// + /// [`cancel_sync`]: DelayedWork::cancel_sync + /// [`Work::cancel`]: Work::cancel + /// + /// # Examples + /// + /// See [`Work::cancel`]. + pub fn cancel(&self) -> Option + where + T: WorkItem, + { + let dwork_ptr = self.dwork.get(); + // SAFETY: `dwork_ptr` is a valid pointer to a `delayed_work`. + if unsafe { bindings::cancel_delayed_work(dwork_ptr) } { + // SAFETY: The work item was successfully cancelled and is guaranteed not to run, + // so we can safely reclaim the ownership leaked during `enqueue`. + Some(unsafe { T::Pointer::reclaim(core::ptr::addr_of_mut!((*dwork_ptr).work)) }) + } else { + None + } + } + + /// Synchronously cancels the delayed work item. + /// + /// This method waits for the work item to finish if it is currently running. + /// If the work item was successfully cancelled from the queue, the pointer is + /// reclaimed and returned. + /// + /// # Guarantees + /// + /// See [`Work::cancel_sync`]. + /// + /// # Safety + /// + /// Same as [`Work::cancel_sync`]. + pub fn cancel_sync(&self) -> Option + where + T: WorkItem, + { + let dwork_ptr = self.dwork.get(); + // SAFETY: `dwork_ptr` is a valid pointer to a `delayed_work`. + if unsafe { bindings::cancel_delayed_work_sync(dwork_ptr) } { + // SAFETY: The work item was successfully cancelled/waited for, and is guaranteed + // not to run again unless re-enqueued. We reclaim the ownership leaked during + // `enqueue`. + Some(unsafe { T::Pointer::reclaim(core::ptr::addr_of_mut!((*dwork_ptr).work)) }) + } else { + None + } + } + /// Get a pointer to the inner `delayed_work`. /// /// # Safety @@ -781,22 +1021,11 @@ unsafe fn raw_get_work( unsafe fn work_container_of( ptr: *mut $crate::workqueue::Work<$work_type $(, $id)?>, ) -> *mut Self { - // SAFETY: The caller promises that the pointer points at a field of the right type - // in the right kind of struct. - let ptr = unsafe { $crate::workqueue::Work::raw_get(ptr) }; - - // SAFETY: The caller promises that the pointer points at a field of the right type - // in the right kind of struct. - let delayed_work = unsafe { - $crate::container_of!(ptr, $crate::bindings::delayed_work, work) - }; - - let delayed_work: *mut $crate::workqueue::DelayedWork<$work_type $(, $id)?> = - delayed_work.cast(); - - // SAFETY: The caller promises that the pointer points at a field of the right type - // in the right kind of struct. - unsafe { $crate::container_of!(delayed_work, Self, $field) } + // SAFETY: The caller promises that the pointer points at the `work` field + // of a `delayed_work` struct, which is itself the `dwork` field of a + // `DelayedWork` wrapper, which is the `$field` field of a `Self` struct. + let ptr = ptr.cast::<$crate::workqueue::DelayedWork<$work_type $(, $id)?>>(); + unsafe { $crate::container_of!(ptr, Self, $field) } } } )*}; @@ -827,6 +1056,15 @@ unsafe impl WorkItemPointer for Arc T::run(arc) } + + unsafe fn reclaim(ptr: *mut bindings::work_struct) -> Self { + // The `__enqueue` method always uses a `work_struct` stored in a `Work`. + let ptr = ptr.cast::>(); + // SAFETY: This computes the pointer that `__enqueue` got from `Arc::into_raw`. + let ptr = unsafe { T::work_container_of(ptr) }; + // SAFETY: This pointer comes from `Arc::into_raw` and we've been given back ownership. + unsafe { Arc::from_raw(ptr) } + } } // SAFETY: The `work_struct` raw pointer is guaranteed to be valid for the duration of the call to @@ -874,7 +1112,8 @@ unsafe impl RawDelayedWorkItem for Arc { } -// SAFETY: TODO. +// SAFETY: The `WorkItemPointer` implementation for `Pin>` is safe because `KBox::from_raw` +// correctly reconstructs the box that was leaked during `enqueue` (via `KBox::into_raw`). unsafe impl WorkItemPointer for Pin> where T: WorkItem, @@ -883,18 +1122,35 @@ unsafe impl WorkItemPointer for Pin> unsafe extern "C" fn run(ptr: *mut bindings::work_struct) { // The `__enqueue` method always uses a `work_struct` stored in a `Work`. let ptr = ptr.cast::>(); - // SAFETY: This computes the pointer that `__enqueue` got from `Arc::into_raw`. + // SAFETY: This computes the pointer that `__enqueue` got from `KBox::into_raw`. let ptr = unsafe { T::work_container_of(ptr) }; - // SAFETY: This pointer comes from `Arc::into_raw` and we've been given back ownership. + // SAFETY: This pointer comes from `KBox::into_raw` and we've been given back ownership. let boxed = unsafe { KBox::from_raw(ptr) }; // SAFETY: The box was already pinned when it was enqueued. let pinned = unsafe { Pin::new_unchecked(boxed) }; T::run(pinned) } + + unsafe fn reclaim(ptr: *mut bindings::work_struct) -> Self { + // The `__enqueue` method always uses a `work_struct` stored in a `Work`. + let ptr = ptr.cast::>(); + // SAFETY: This computes the pointer that `__enqueue` got from `KBox::into_raw`. + let ptr = unsafe { T::work_container_of(ptr) }; + // SAFETY: This pointer comes from `KBox::into_raw` and we've been given back ownership. + let boxed = unsafe { KBox::from_raw(ptr) }; + // SAFETY: The box was already pinned when it was enqueued. + unsafe { Pin::new_unchecked(boxed) } + } } -// SAFETY: TODO. +// SAFETY: The `work_struct` raw pointer is guaranteed to be valid for the duration of the call to +// the closure because we have exclusive ownership of the `KBox`, and we don't drop it ourselves. +// If `queue_work_on` returns true, it is further guaranteed to be valid until a call to the +// function pointer in `work_struct` because we leak the memory it points to, and only reclaim it +// if the closure returns false (not reachable for KBox as it must succeed) or in +// `WorkItemPointer::run`, which is what the function pointer in the `work_struct` must be +// pointing to. unsafe impl RawWorkItem for Pin> where T: WorkItem, @@ -1022,3 +1278,150 @@ pub fn system_bh_highpri() -> &'static Queue { // SAFETY: `system_bh_highpri_wq` is a C global, always available. unsafe { Queue::from_raw(bindings::system_bh_highpri_wq) } } + +/// Returns the strong count of an [`Arc`] for testing purposes. +/// +/// # Safety +/// +/// This is intended for use in KUnit tests and sample modules ONLY. +#[cfg(CONFIG_KUNIT)] +#[doc(hidden)] +pub fn arc_count(arc: &Arc) -> i32 { + // SAFETY: ArcInner has refcount as its first field. Arc points to data at DATA_OFFSET. + unsafe { + let ptr = Arc::as_ptr(arc); + let inner_ptr = (ptr as *const u8).sub(Arc::::DATA_OFFSET); + let refcount_ptr = inner_ptr as *const i32; + core::ptr::read_volatile(refcount_ptr) + } +} + +#[macros::kunit_tests(rust_kernel_workqueue)] +mod tests { + use super::*; + use crate::sync::Arc; + + #[pin_data] + struct TestWorkItem { + #[pin] + work: Work, + value: i32, + } + + impl_has_work! { + impl HasWork for TestWorkItem { self.work } + } + + impl WorkItem for TestWorkItem { + type Pointer = Arc; + fn run(_this: Arc) {} + } + + #[pin_data] + struct TestDelayedWorkItem { + #[pin] + delayed_work: DelayedWork, + value: i32, + } + + impl_has_delayed_work! { + impl HasDelayedWork for TestDelayedWorkItem { self.delayed_work } + } + + impl WorkItem for TestDelayedWorkItem { + type Pointer = Arc; + fn run(_this: Arc) {} + } + + #[test] + fn test_work_cancel_reclaim() { + let item = Arc::pin_init( + pin_init!(TestWorkItem { + work <- new_work!("TestWorkItem::work"), + value: 42, + }), + GFP_KERNEL, + ) + .expect("Failed to allocate TestWorkItem"); + + let initial_count = arc_count(&item); + + // Enqueue + let _ = system().enqueue(item.clone()); + + // Cancel and Reclaim (if it was pending) + if let Some(reclaimed) = item.work.cancel() { + assert!(arc_count(&item) == initial_count + 1); + drop(reclaimed); + assert!(arc_count(&item) == initial_count); + } + } + + #[test] + fn test_work_cancel_sync_reclaim() { + let item = Arc::pin_init( + pin_init!(TestWorkItem { + work <- new_work!("TestWorkItem::work"), + value: 42, + }), + GFP_KERNEL, + ) + .expect("Failed to allocate TestWorkItem"); + + let initial_count = arc_count(&item); + + // Enqueue + let _ = system().enqueue(item.clone()); + + // Cancel Sync and Reclaim + if let Some(reclaimed) = item.work.cancel_sync() { + assert!(arc_count(&item) == initial_count + 1); + drop(reclaimed); + assert!(arc_count(&item) == initial_count); + } + } + + #[test] + fn test_work_stress_enqueue_cancel() { + let item = Arc::pin_init( + pin_init!(TestWorkItem { + work <- new_work!("TestWorkItem::work"), + value: 42, + }), + GFP_KERNEL, + ) + .expect("Failed to allocate TestWorkItem"); + + for _ in 0..100 { + if let Ok(_) = system().enqueue(item.clone()) { + let _ = item.work.cancel_sync(); + } + } + + assert_eq!(arc_count(&item), 1); + } + + #[test] + fn test_delayed_work_cancel_reclaim() { + let item = Arc::pin_init( + pin_init!(TestDelayedWorkItem { + delayed_work <- new_delayed_work!("TestDelayedWorkItem::delayed_work"), + value: 42, + }), + GFP_KERNEL, + ) + .expect("Failed to allocate TestDelayedWorkItem"); + + let initial_count = arc_count(&item); + + // Enqueue delayed (use a longer delay to ensure it stays pending) + let _ = system().enqueue_delayed(item.clone(), 1000); + + // Cancel and Reclaim + if let Some(reclaimed) = item.delayed_work.cancel() { + assert!(arc_count(&item) == initial_count + 1); + drop(reclaimed); + assert!(arc_count(&item) == initial_count); + } + } +} -- 2.43.0