* [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock
@ 2025-05-17 7:32 Tetsuo Handa
2025-05-17 15:09 ` Jakub Kicinski
2025-05-20 9:23 ` Jiri Pirko
0 siblings, 2 replies; 5+ messages in thread
From: Tetsuo Handa @ 2025-05-17 7:32 UTC (permalink / raw)
To: Jiri Pirko, Andrew Lunn, David S. Miller, Eric Dumazet,
Jakub Kicinski, Paolo Abeni, Hillf Danton, Network Development
syzbot is reporting locking order problem between wiphy and team [1].
Jiri Pirko commented [2]:
I wonder, since we already rely on rtnl in lots of team code, perhaps we
can remove team->lock completely and convert the rest of the code to be
protected by rtnl lock as well
I asked syzbot to check whether rtnl is already held when team lock is
held, and the reproducer did not find obvious location [3]. Thus, I wrote
this patch for more testing.
[1] https://syzkaller.appspot.com/bug?extid=705c61d60b091ef42c04
[2] https://lkml.kernel.org/r/ZoZ2RH9BcahEB9Sb@nanopsycho.orion
[3] https://lkml.kernel.org/r/68275676.a00a0220.398d88.0213.GAE@google.com
Signed-off-by: Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>
---
drivers/net/team/team_core.c | 52 +++++++----------------
drivers/net/team/team_mode_activebackup.c | 2 +-
drivers/net/team/team_mode_loadbalance.c | 16 ++++---
include/linux/if_team.h | 3 --
4 files changed, 27 insertions(+), 46 deletions(-)
diff --git a/drivers/net/team/team_core.c b/drivers/net/team/team_core.c
index d8fc0c79745d..f32c3db56088 100644
--- a/drivers/net/team/team_core.c
+++ b/drivers/net/team/team_core.c
@@ -933,7 +933,7 @@ static bool team_port_find(const struct team *team,
* Enable/disable port by adding to enabled port hashlist and setting
* port->index (Might be racy so reader could see incorrect ifindex when
* processing a flying packet, but that is not a problem). Write guarded
- * by team->lock.
+ * by RTNL.
*/
static void team_port_enable(struct team *team,
struct team_port *port)
@@ -1660,8 +1660,6 @@ static int team_init(struct net_device *dev)
goto err_options_register;
netif_carrier_off(dev);
- lockdep_register_key(&team->team_lock_key);
- __mutex_init(&team->lock, "team->team_lock_key", &team->team_lock_key);
netdev_lockdep_set_classes(dev);
return 0;
@@ -1682,7 +1680,7 @@ static void team_uninit(struct net_device *dev)
struct team_port *port;
struct team_port *tmp;
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
list_for_each_entry_safe(port, tmp, &team->port_list, list)
team_port_del(team, port->dev);
@@ -1691,9 +1689,7 @@ static void team_uninit(struct net_device *dev)
team_mcast_rejoin_fini(team);
team_notify_peers_fini(team);
team_queue_override_fini(team);
- mutex_unlock(&team->lock);
netdev_change_features(dev);
- lockdep_unregister_key(&team->team_lock_key);
}
static void team_destructor(struct net_device *dev)
@@ -1811,14 +1807,13 @@ static int team_set_mac_address(struct net_device *dev, void *p)
struct team *team = netdev_priv(dev);
struct team_port *port;
+ ASSERT_RTNL();
if (dev->type == ARPHRD_ETHER && !is_valid_ether_addr(addr->sa_data))
return -EADDRNOTAVAIL;
dev_addr_set(dev, addr->sa_data);
- mutex_lock(&team->lock);
list_for_each_entry(port, &team->port_list, list)
if (team->ops.port_change_dev_addr)
team->ops.port_change_dev_addr(team, port);
- mutex_unlock(&team->lock);
return 0;
}
@@ -1829,10 +1824,10 @@ static int team_change_mtu(struct net_device *dev, int new_mtu)
int err;
/*
- * Alhough this is reader, it's guarded by team lock. It's not possible
+ * Alhough this is reader, it's guarded by RTNL. It's not possible
* to traverse list in reverse under rcu_read_lock
*/
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
team->port_mtu_change_allowed = true;
list_for_each_entry(port, &team->port_list, list) {
err = dev_set_mtu(port->dev, new_mtu);
@@ -1843,7 +1838,6 @@ static int team_change_mtu(struct net_device *dev, int new_mtu)
}
}
team->port_mtu_change_allowed = false;
- mutex_unlock(&team->lock);
WRITE_ONCE(dev->mtu, new_mtu);
@@ -1853,7 +1847,6 @@ static int team_change_mtu(struct net_device *dev, int new_mtu)
list_for_each_entry_continue_reverse(port, &team->port_list, list)
dev_set_mtu(port->dev, dev->mtu);
team->port_mtu_change_allowed = false;
- mutex_unlock(&team->lock);
return err;
}
@@ -1904,23 +1897,21 @@ static int team_vlan_rx_add_vid(struct net_device *dev, __be16 proto, u16 vid)
int err;
/*
- * Alhough this is reader, it's guarded by team lock. It's not possible
+ * Alhough this is reader, it's guarded by RTNL. It's not possible
* to traverse list in reverse under rcu_read_lock
*/
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
list_for_each_entry(port, &team->port_list, list) {
err = vlan_vid_add(port->dev, proto, vid);
if (err)
goto unwind;
}
- mutex_unlock(&team->lock);
return 0;
unwind:
list_for_each_entry_continue_reverse(port, &team->port_list, list)
vlan_vid_del(port->dev, proto, vid);
- mutex_unlock(&team->lock);
return err;
}
@@ -1930,10 +1921,9 @@ static int team_vlan_rx_kill_vid(struct net_device *dev, __be16 proto, u16 vid)
struct team *team = netdev_priv(dev);
struct team_port *port;
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
list_for_each_entry(port, &team->port_list, list)
vlan_vid_del(port->dev, proto, vid);
- mutex_unlock(&team->lock);
return 0;
}
@@ -1955,9 +1945,8 @@ static void team_netpoll_cleanup(struct net_device *dev)
{
struct team *team = netdev_priv(dev);
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
__team_netpoll_cleanup(team);
- mutex_unlock(&team->lock);
}
static int team_netpoll_setup(struct net_device *dev)
@@ -1966,7 +1955,7 @@ static int team_netpoll_setup(struct net_device *dev)
struct team_port *port;
int err = 0;
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
list_for_each_entry(port, &team->port_list, list) {
err = __team_port_enable_netpoll(port);
if (err) {
@@ -1974,7 +1963,6 @@ static int team_netpoll_setup(struct net_device *dev)
break;
}
}
- mutex_unlock(&team->lock);
return err;
}
#endif
@@ -1985,9 +1973,8 @@ static int team_add_slave(struct net_device *dev, struct net_device *port_dev,
struct team *team = netdev_priv(dev);
int err;
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
err = team_port_add(team, port_dev, extack);
- mutex_unlock(&team->lock);
if (!err)
netdev_change_features(dev);
@@ -2000,18 +1987,12 @@ static int team_del_slave(struct net_device *dev, struct net_device *port_dev)
struct team *team = netdev_priv(dev);
int err;
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
err = team_port_del(team, port_dev);
- mutex_unlock(&team->lock);
if (err)
return err;
- if (netif_is_team_master(port_dev)) {
- lockdep_unregister_key(&team->team_lock_key);
- lockdep_register_key(&team->team_lock_key);
- lockdep_set_class(&team->lock, &team->team_lock_key);
- }
netdev_change_features(dev);
return err;
@@ -2308,6 +2289,7 @@ static struct team *team_nl_team_get(struct genl_info *info)
struct net_device *dev;
struct team *team;
+ ASSERT_RTNL();
if (!info->attrs[TEAM_ATTR_TEAM_IFINDEX])
return NULL;
@@ -2319,13 +2301,12 @@ static struct team *team_nl_team_get(struct genl_info *info)
}
team = netdev_priv(dev);
- mutex_lock(&team->lock);
return team;
}
static void team_nl_team_put(struct team *team)
{
- mutex_unlock(&team->lock);
+ ASSERT_RTNL();
dev_put(team->dev);
}
@@ -2961,11 +2942,8 @@ static void __team_port_change_port_removed(struct team_port *port)
static void team_port_change_check(struct team_port *port, bool linkup)
{
- struct team *team = port->team;
-
- mutex_lock(&team->lock);
+ ASSERT_RTNL();
__team_port_change_check(port, linkup);
- mutex_unlock(&team->lock);
}
diff --git a/drivers/net/team/team_mode_activebackup.c b/drivers/net/team/team_mode_activebackup.c
index e0f599e2a51d..4e133451f4d6 100644
--- a/drivers/net/team/team_mode_activebackup.c
+++ b/drivers/net/team/team_mode_activebackup.c
@@ -68,7 +68,7 @@ static void ab_active_port_get(struct team *team, struct team_gsetter_ctx *ctx)
struct team_port *active_port;
active_port = rcu_dereference_protected(ab_priv(team)->active_port,
- lockdep_is_held(&team->lock));
+ rtnl_is_locked());
if (active_port)
ctx->data.u32_val = active_port->dev->ifindex;
else
diff --git a/drivers/net/team/team_mode_loadbalance.c b/drivers/net/team/team_mode_loadbalance.c
index 00f8989c29c0..fb4ead02c504 100644
--- a/drivers/net/team/team_mode_loadbalance.c
+++ b/drivers/net/team/team_mode_loadbalance.c
@@ -302,7 +302,7 @@ static int lb_bpf_func_set(struct team *team, struct team_gsetter_ctx *ctx)
/* Clear old filter data */
__fprog_destroy(lb_priv->ex->orig_fprog);
orig_fp = rcu_dereference_protected(lb_priv->fp,
- lockdep_is_held(&team->lock));
+ rtnl_is_locked());
}
rcu_assign_pointer(lb_priv->fp, fp);
@@ -325,7 +325,7 @@ static void lb_bpf_func_free(struct team *team)
__fprog_destroy(lb_priv->ex->orig_fprog);
fp = rcu_dereference_protected(lb_priv->fp,
- lockdep_is_held(&team->lock));
+ rtnl_is_locked());
bpf_prog_destroy(fp);
}
@@ -336,7 +336,7 @@ static void lb_tx_method_get(struct team *team, struct team_gsetter_ctx *ctx)
char *name;
func = rcu_dereference_protected(lb_priv->select_tx_port_func,
- lockdep_is_held(&team->lock));
+ rtnl_is_locked());
name = lb_select_tx_port_get_name(func);
BUG_ON(!name);
ctx->data.str_val = name;
@@ -478,7 +478,13 @@ static void lb_stats_refresh(struct work_struct *work)
team = lb_priv_ex->team;
lb_priv = get_lb_priv(team);
- if (!mutex_trylock(&team->lock)) {
+ /* Since this function is called from WQ context, RTNL can't be held by the caller. */
+ if (!rtnl_trylock()) {
+ /*
+ * Since RTNL is shared by many callers, and rtnl_unlock() is a slower operation
+ * than plain mutex_unlock(), rtnl_trylock() will be more easier to compate than
+ * mutex_trylock(). Therefore, we might want to delay a bit before retrying.
+ */
schedule_delayed_work(&lb_priv_ex->stats.refresh_dw, 0);
return;
}
@@ -515,7 +521,7 @@ static void lb_stats_refresh(struct work_struct *work)
schedule_delayed_work(&lb_priv_ex->stats.refresh_dw,
(lb_priv_ex->stats.refresh_interval * HZ) / 10);
- mutex_unlock(&team->lock);
+ rtnl_unlock();
}
static void lb_stats_refresh_interval_get(struct team *team,
diff --git a/include/linux/if_team.h b/include/linux/if_team.h
index cdc684e04a2f..ce97d891cf72 100644
--- a/include/linux/if_team.h
+++ b/include/linux/if_team.h
@@ -191,8 +191,6 @@ struct team {
const struct header_ops *header_ops_cache;
- struct mutex lock; /* used for overall locking, e.g. port lists write */
-
/*
* List of enabled ports and their count
*/
@@ -223,7 +221,6 @@ struct team {
atomic_t count_pending;
struct delayed_work dw;
} mcast_rejoin;
- struct lock_class_key team_lock_key;
long mode_priv[TEAM_MODE_PRIV_LONGS];
};
--
2.43.5
^ permalink raw reply related [flat|nested] 5+ messages in thread
* Re: [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock
2025-05-17 7:32 [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock Tetsuo Handa
@ 2025-05-17 15:09 ` Jakub Kicinski
2025-05-21 13:54 ` Tetsuo Handa
2025-05-20 9:23 ` Jiri Pirko
1 sibling, 1 reply; 5+ messages in thread
From: Jakub Kicinski @ 2025-05-17 15:09 UTC (permalink / raw)
To: Tetsuo Handa
Cc: Jiri Pirko, Andrew Lunn, David S. Miller, Eric Dumazet,
Paolo Abeni, Hillf Danton, Network Development
On Sat, 17 May 2025 16:32:20 +0900 Tetsuo Handa wrote:
> - if (!mutex_trylock(&team->lock)) {
> + /* Since this function is called from WQ context, RTNL can't be held by the caller. */
> + if (!rtnl_trylock()) {
> + /*
> + * Since RTNL is shared by many callers, and rtnl_unlock() is a slower operation
> + * than plain mutex_unlock(), rtnl_trylock() will be more easier to compate than
> + * mutex_trylock(). Therefore, we might want to delay a bit before retrying.
> + */
I think this was a trylock because there are places we try to cancel
this work while already holding the lock.
FWIW I'm not opposed to the patch. Could you wait a week and repost,
tho? We have a fix queued up in another tree - 6b1d3c5f675cc7
if we apply your patch to net-next there will be a build failure
on merge. Not a showstopper but we'll merge the trees on Thu so it
can be easily avoided if we wait until then.
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock
2025-05-17 7:32 [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock Tetsuo Handa
2025-05-17 15:09 ` Jakub Kicinski
@ 2025-05-20 9:23 ` Jiri Pirko
1 sibling, 0 replies; 5+ messages in thread
From: Jiri Pirko @ 2025-05-20 9:23 UTC (permalink / raw)
To: Tetsuo Handa
Cc: Andrew Lunn, David S. Miller, Eric Dumazet, Jakub Kicinski,
Paolo Abeni, Hillf Danton, Network Development
Sat, May 17, 2025 at 09:32:20AM +0200, penguin-kernel@I-love.SAKURA.ne.jp wrote:
[..]
>@@ -2319,13 +2301,12 @@ static struct team *team_nl_team_get(struct genl_info *info)
> }
>
> team = netdev_priv(dev);
>- mutex_lock(&team->lock);
> return team;
> }
Why do you think this is safe?
Rtnl is held only for set doit.
>
> static void team_nl_team_put(struct team *team)
> {
>- mutex_unlock(&team->lock);
>+ ASSERT_RTNL();
Did you test this? How? Howcome you didn't hit this assertion?
> dev_put(team->dev);
> }
>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock
2025-05-17 15:09 ` Jakub Kicinski
@ 2025-05-21 13:54 ` Tetsuo Handa
2025-05-21 17:59 ` Jakub Kicinski
0 siblings, 1 reply; 5+ messages in thread
From: Tetsuo Handa @ 2025-05-21 13:54 UTC (permalink / raw)
To: Jakub Kicinski
Cc: Jiri Pirko, Andrew Lunn, David S. Miller, Eric Dumazet,
Paolo Abeni, Hillf Danton, Network Development
On 2025/05/18 0:09, Jakub Kicinski wrote:
> I think this was a trylock because there are places we try to cancel
> this work while already holding the lock.
I checked rtnl_unlock(), and it seems to me that rtnl_unlock() calls mutex_unlock() before
doing operations that might sleep. Then, rtnl_unlock() itself won't increase possibility of
rtnl_trylock() failure...
>
> FWIW I'm not opposed to the patch. Could you wait a week and repost,
> tho? We have a fix queued up in another tree - 6b1d3c5f675cc7
> if we apply your patch to net-next there will be a build failure
> on merge. Not a showstopper but we'll merge the trees on Thu so it
> can be easily avoided if we wait until then.
I reposted https://lkml.kernel.org/r/d4047311-5644-4884-84c0-059bbb9e98ee@I-love.SAKURA.ne.jp
also as a fix for a new report ( https://syzkaller.appspot.com/bug?extid=b191b5ccad8d7a986286 )
triggered by 6b1d3c5f675cc7.
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock
2025-05-21 13:54 ` Tetsuo Handa
@ 2025-05-21 17:59 ` Jakub Kicinski
0 siblings, 0 replies; 5+ messages in thread
From: Jakub Kicinski @ 2025-05-21 17:59 UTC (permalink / raw)
To: Tetsuo Handa
Cc: Jiri Pirko, Andrew Lunn, David S. Miller, Eric Dumazet,
Paolo Abeni, Hillf Danton, Network Development
On Wed, 21 May 2025 22:54:51 +0900 Tetsuo Handa wrote:
> On 2025/05/18 0:09, Jakub Kicinski wrote:
> > I think this was a trylock because there are places we try to cancel
> > this work while already holding the lock.
>
> I checked rtnl_unlock(), and it seems to me that rtnl_unlock() calls mutex_unlock() before
> doing operations that might sleep. Then, rtnl_unlock() itself won't increase possibility of
> rtnl_trylock() failure...
I don't think you understood my comment :(
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2025-05-21 17:59 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-05-17 7:32 [PATCH (EXPERIMENTAL)] team: replace term lock with rtnl lock Tetsuo Handa
2025-05-17 15:09 ` Jakub Kicinski
2025-05-21 13:54 ` Tetsuo Handa
2025-05-21 17:59 ` Jakub Kicinski
2025-05-20 9:23 ` Jiri Pirko
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).