public inbox for netdev@vger.kernel.org
 help / color / mirror / Atom feed
From: bsdhenrymartin@gmail.com
To: linux-nfs@vger.kernel.org
Cc: Chuck Lever <chuck.lever@oracle.com>,
	Jeff Layton <jlayton@kernel.org>, NeilBrown <neil@brown.name>,
	Olga Kornievskaia <okorniev@redhat.com>,
	Dai Ngo <Dai.Ngo@oracle.com>, Tom Talpey <tom@talpey.com>,
	Trond Myklebust <trondmy@kernel.org>,
	Anna Schumaker <anna@kernel.org>,
	"David S. Miller" <davem@davemloft.net>,
	Eric Dumazet <edumazet@google.com>,
	Jakub Kicinski <kuba@kernel.org>, Paolo Abeni <pabeni@redhat.com>,
	Simon Horman <horms@kernel.org>,
	netdev@vger.kernel.org, linux-kernel@vger.kernel.org,
	Henry Martin <bsdhenrymartin@gmail.com>,
	stable@vger.kernel.org
Subject: [PATCH] sunrpc: fix TLS connect_worker rpc_clnt lifetime UAF
Date: Mon,  9 Mar 2026 19:19:53 +0800	[thread overview]
Message-ID: <20260309112041.1336519-1-bsdhenrymartin@gmail.com> (raw)

From: Henry Martin <bsdhenrymartin@gmail.com>

In xs_connect(), transport->clnt is assigned from task->tk_client
without taking a reference when a TLS connect worker is queued.

If the RPC task finishes before connect_worker runs, tk_client can be
released and its cl_cred can be freed. Later, xs_tcp_tls_setup_socket()
dereferences upper_clnt->cl_cred and passes it to rpc_create(), where
rpc_new_client() calls get_cred() and triggers a slab-use-after-free.

[   93.358371] ==================================================================
[   93.359597] BUG: KASAN: slab-use-after-free in rpc_new_client+0x387/0xdcc
[   93.360748] Write of size 4 at addr ffff88810d67bfa8 by task kworker/u4:4/44
[   93.361919] 
[   93.362225] CPU: 0 UID: 0 PID: 44 Comm: kworker/u4:4 Tainted: G                 N  7.0.0-rc3 #2 PREEMPT(full) 
[   93.362297] Tainted: [N]=TEST
[   93.362313] Hardware name: QEMU Ubuntu 24.04 PC (i440FX + PIIX, 1996), BIOS 1.16.3-debian-1.16.3-2 04/01/2014
[   93.362348] Workqueue: xprtiod xs_tcp_tls_setup_socket
[   93.362433] Call Trace:
[   93.362447]  <TASK>
[   93.362462]  dump_stack_lvl+0xad/0xf9
[   93.362513]  ? rpc_new_client+0x387/0xdcc
[   93.362574]  print_report+0x171/0x4d6
[   93.362653]  ? __virt_addr_valid+0x353/0x364
[   93.362719]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.362784]  ? kmem_cache_debug_flags+0x11/0x26
[   93.362839]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.362913]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.362978]  ? kasan_complete_mode_report_info+0x1c2/0x1d1
[   93.363057]  ? rpc_new_client+0x387/0xdcc
[   93.363122]  kasan_report+0xb3/0xe2
[   93.363202]  ? rpc_new_client+0x387/0xdcc
[   93.363266]  __asan_report_store4_noabort+0x1b/0x21
[   93.363339]  rpc_new_client+0x387/0xdcc
[   93.363399]  ? __sanitizer_cov_trace_pc+0x24/0x5a
[   93.363451]  rpc_create_xprt+0x1ac/0x3b4
[   93.363519]  rpc_create+0x5f9/0x703
[   93.363588]  ? __pfx_rpc_create+0x10/0x10
[   93.363654]  ? __sanitizer_cov_trace_pc+0x24/0x5a
[   93.363706]  ? __pfx_default_wake_function+0x10/0x10
[   93.363808]  ? __dequeue_entity+0x5d2/0x6c3
[   93.363887]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.363952]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.364016]  ? write_comp_data+0x2e/0x8e
[   93.364063]  xs_tcp_tls_setup_socket+0x476/0xff0
[   93.364151]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.364217]  ? __pfx_xs_tcp_tls_setup_socket+0x10/0x10
[   93.364315]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.364386]  ? __kasan_check_write+0x18/0x1e
[   93.364468]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.364540]  ? set_work_data+0x70/0x9c
[   93.364603]  process_scheduled_works+0x66c/0xa15
[   93.364699]  ? __sanitizer_cov_trace_pc+0x24/0x5a
[   93.364763]  worker_thread+0x440/0x547
[   93.364867]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.364937]  ? __pfx_worker_thread+0x10/0x10
[   93.365024]  kthread+0x375/0x38a
[   93.365097]  ? __pfx_kthread+0x10/0x10
[   93.365185]  ret_from_fork+0xa8/0x872
[   93.365247]  ? __pfx_ret_from_fork+0x10/0x10
[   93.365309]  ? __sanitizer_cov_trace_pc+0x24/0x5a
[   93.365364]  ? srso_alias_return_thunk+0x5/0xfbef5
[   93.365428]  ? __switch_to+0xc44/0xc5a
[   93.365509]  ? __pfx_kthread+0x10/0x10
[   93.365593]  ret_from_fork_asm+0x1a/0x30
[   93.365684]  </TASK>
[   93.365701] 
[   93.405276] Allocated by task 392:
[   93.405852]  kasan_save_stack+0x3c/0x5e
[   93.406581]  kasan_save_track+0x18/0x32
[   93.407230]  kasan_save_alloc_info+0x3b/0x49
[   93.407932]  __kasan_slab_alloc+0x52/0x62
[   93.408606]  kmem_cache_alloc_noprof+0x266/0x304
[   93.409359]  prepare_creds+0x32/0x338
[   93.409965]  copy_creds+0x188/0x425
[   93.410545]  copy_process+0x1022/0x5320
[   93.411208]  kernel_clone+0x23d/0x61a
[   93.411870]  __do_sys_clone+0xf8/0x139
[   93.412530]  __x64_sys_clone+0xde/0xed
[   93.413192]  x64_sys_call+0x33f/0x2105
[   93.413883]  do_syscall_64+0x1b3/0x420
[   93.414588]  entry_SYSCALL_64_after_hwframe+0x77/0x7f
[   93.416895] 
[   93.417169] Freed by task 396:
[   93.417673]  kasan_save_stack+0x3c/0x5e
[   93.418321]  kasan_save_track+0x18/0x32
[   93.418972]  kasan_save_free_info+0x43/0x52
[   93.419652]  poison_slab_object+0x33/0x3c
[   93.420315]  __kasan_slab_free+0x25/0x4a
[   93.420973]  kmem_cache_free+0x1e5/0x2e4
[   93.421616]  put_cred_rcu+0x2e7/0x2f4
[   93.422219]  rcu_do_batch+0x5b6/0xa82
[   93.422833]  rcu_core+0x264/0x298
[   93.423475]  rcu_core_si+0x12/0x18
[   93.424086]  handle_softirqs+0x21c/0x488
[   93.424750]  __do_softirq+0x14/0x1a
[   93.425346] 
[   93.425612] Last potentially related work creation:
[   93.426358]  kasan_save_stack+0x3c/0x5e
[   93.427024]  kasan_record_aux_stack+0x92/0x9e
[   93.427739]  call_rcu+0xe4/0xb2b
[   93.428337]  __put_cred+0x13e/0x14c
[   93.428937]  put_cred_many+0x50/0x5e
[   93.429530]  exit_creds+0x95/0xbc
[   93.430099]  __put_task_struct+0x173/0x26a
[   93.430770]  __put_task_struct_rcu_cb+0x22/0x29
[   93.431513]  rcu_do_batch+0x5b6/0xa82
[   93.432144]  rcu_core+0x264/0x298
[   93.432737]  rcu_core_si+0x12/0x18
[   93.433345]  handle_softirqs+0x21c/0x488
[   93.434030]  __do_softirq+0x14/0x1a
[   93.434632] 
[   93.434910] The buggy address belongs to the object at ffff88810d67bf00
[   93.434910]  which belongs to the cache cred of size 184
[   93.436720] The buggy address is located 168 bytes inside of
[   93.436720]  freed 184-byte region [ffff88810d67bf00, ffff88810d67bfb8)
[   93.438582] 
[   93.438868] The buggy address belongs to the physical page:
[   93.439734] page: refcount:0 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x10d67b
[   93.440982] memcg:ffff88810d67b0c9
[   93.441546] flags: 0x200000000000000(node=0|zone=2)
[   93.442327] page_type: f5(slab)
[   93.442878] raw: 0200000000000000 ffff88810088d140 dead000000000122 0000000000000000
[   93.444091] raw: 0000000000000000 0000010000100010 00000000f5000000 ffff88810d67b0c9
[   93.445365] page dumped because: kasan: bad access detected
[   93.446334] 
[   93.446638] Memory state around the buggy address:
[   93.447505]  ffff88810d67be80: 00 00 00 00 00 00 00 fc fc fc fc fc fc fc fc fc
[   93.448748]  ffff88810d67bf00: fa fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
[   93.449973] >ffff88810d67bf80: fb fb fb fb fb fb fb fc fc fc fc fc fc fc fc fc
[   93.451147]                                   ^
[   93.452039]  ffff88810d67c000: fa fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
[   93.453227]  ffff88810d67c080: fb fb fb fb fb fb fb fb fb fb fb fb fb fc fc fc
[   93.454455] ==================================================================
[   93.577640] Disabling lock debugging due to kernel taint
[ 1206.114037] kworker/u4:1 (26) used greatest stack depth: 24168 bytes left

Fix this by taking a client reference when queuing a TLS connect worker
and dropping that reference when the worker exits. Also release any
still-pinned client in xs_destroy() after cancel_delayed_work_sync() to
cover the case where queued work is canceled before execution.

Fixes: 75eb6af7acdf ("SUNRPC: Add a TCP-with-TLS RPC transport class")
Cc: stable@vger.kernel.org # 6.5+
Signed-off-by: Henry Martin <bsdhenrymartin@gmail.com>
---
 net/sunrpc/xprtsock.c | 12 +++++++++++-
 1 file changed, 11 insertions(+), 1 deletion(-)

diff --git a/net/sunrpc/xprtsock.c b/net/sunrpc/xprtsock.c
index 2e1fe6013361..6bf1cf20a86e 100644
--- a/net/sunrpc/xprtsock.c
+++ b/net/sunrpc/xprtsock.c
@@ -1362,6 +1362,10 @@ static void xs_destroy(struct rpc_xprt *xprt)
 	dprintk("RPC:       xs_destroy xprt %p\n", xprt);
 
 	cancel_delayed_work_sync(&transport->connect_worker);
+	if (transport->clnt != NULL) {
+		rpc_release_client(transport->clnt);
+		transport->clnt = NULL;
+	}
 	xs_close(xprt);
 	cancel_work_sync(&transport->recv_worker);
 	cancel_work_sync(&transport->error_worker);
@@ -2758,6 +2762,8 @@ static void xs_tcp_tls_setup_socket(struct work_struct *work)
 out_unlock:
 	current_restore_flags(pflags, PF_MEMALLOC);
 	upper_transport->clnt = NULL;
+	if (upper_clnt != NULL)
+		rpc_release_client(upper_clnt);
 	xprt_unlock_connect(upper_xprt, upper_transport);
 	return;
 
@@ -2805,7 +2811,11 @@ static void xs_connect(struct rpc_xprt *xprt, struct rpc_task *task)
 	} else
 		dprintk("RPC:       xs_connect scheduled xprt %p\n", xprt);
 
-	transport->clnt = task->tk_client;
+	if (transport->connect_worker.work.func == xs_tcp_tls_setup_socket) {
+		WARN_ON_ONCE(transport->clnt != NULL);
+		refcount_inc(&task->tk_client->cl_count);
+		transport->clnt = task->tk_client;
+	}
 	queue_delayed_work(xprtiod_workqueue,
 			&transport->connect_worker,
 			delay);
-- 
2.43.0

             reply	other threads:[~2026-03-09 11:20 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-03-09 11:19 bsdhenrymartin [this message]
2026-03-09 14:45 ` [PATCH] sunrpc: fix TLS connect_worker rpc_clnt lifetime UAF Jeff Layton
2026-03-11 14:18 ` Benjamin Coddington
2026-03-11 14:20   ` Chuck Lever

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20260309112041.1336519-1-bsdhenrymartin@gmail.com \
    --to=bsdhenrymartin@gmail.com \
    --cc=Dai.Ngo@oracle.com \
    --cc=anna@kernel.org \
    --cc=chuck.lever@oracle.com \
    --cc=davem@davemloft.net \
    --cc=edumazet@google.com \
    --cc=horms@kernel.org \
    --cc=jlayton@kernel.org \
    --cc=kuba@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-nfs@vger.kernel.org \
    --cc=neil@brown.name \
    --cc=netdev@vger.kernel.org \
    --cc=okorniev@redhat.com \
    --cc=pabeni@redhat.com \
    --cc=stable@vger.kernel.org \
    --cc=tom@talpey.com \
    --cc=trondmy@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox