mirror of
https://github.com/torvalds/linux.git
synced 2025-04-11 04:53:02 +00:00

In order to exercise and verify notifiers' locking assumptions, register dummy notifiers (via register_netdevice_notifier_dev_net). Share notifier event handler that enforces the assumptions with lock_debug.c (rename and export rtnl_net_debug_event as netdev_debug_event). Add ops lock asserts to netdev_debug_event. Signed-off-by: Stanislav Fomichev <sdf@fomichev.me> Link: https://patch.msgid.link/20250401163452.622454-6-sdf@fomichev.me Signed-off-by: Jakub Kicinski <kuba@kernel.org>
105 lines
2.7 KiB
C
105 lines
2.7 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
|
|
#ifndef _NET_NETDEV_LOCK_H
|
|
#define _NET_NETDEV_LOCK_H
|
|
|
|
#include <linux/lockdep.h>
|
|
#include <linux/netdevice.h>
|
|
#include <linux/rtnetlink.h>
|
|
|
|
static inline bool netdev_trylock(struct net_device *dev)
|
|
{
|
|
return mutex_trylock(&dev->lock);
|
|
}
|
|
|
|
static inline void netdev_assert_locked(const struct net_device *dev)
|
|
{
|
|
lockdep_assert_held(&dev->lock);
|
|
}
|
|
|
|
static inline void
|
|
netdev_assert_locked_or_invisible(const struct net_device *dev)
|
|
{
|
|
if (dev->reg_state == NETREG_REGISTERED ||
|
|
dev->reg_state == NETREG_UNREGISTERING)
|
|
netdev_assert_locked(dev);
|
|
}
|
|
|
|
static inline bool netdev_need_ops_lock(const struct net_device *dev)
|
|
{
|
|
bool ret = dev->request_ops_lock || !!dev->queue_mgmt_ops;
|
|
|
|
#if IS_ENABLED(CONFIG_NET_SHAPER)
|
|
ret |= !!dev->netdev_ops->net_shaper_ops;
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline void netdev_lock_ops(struct net_device *dev)
|
|
{
|
|
if (netdev_need_ops_lock(dev))
|
|
netdev_lock(dev);
|
|
}
|
|
|
|
static inline void netdev_unlock_ops(struct net_device *dev)
|
|
{
|
|
if (netdev_need_ops_lock(dev))
|
|
netdev_unlock(dev);
|
|
}
|
|
|
|
static inline void netdev_ops_assert_locked(const struct net_device *dev)
|
|
{
|
|
if (netdev_need_ops_lock(dev))
|
|
lockdep_assert_held(&dev->lock);
|
|
else
|
|
ASSERT_RTNL();
|
|
}
|
|
|
|
static inline void
|
|
netdev_ops_assert_locked_or_invisible(const struct net_device *dev)
|
|
{
|
|
if (dev->reg_state == NETREG_REGISTERED ||
|
|
dev->reg_state == NETREG_UNREGISTERING)
|
|
netdev_ops_assert_locked(dev);
|
|
}
|
|
|
|
static inline int netdev_lock_cmp_fn(const struct lockdep_map *a,
|
|
const struct lockdep_map *b)
|
|
{
|
|
/* Only lower devices currently grab the instance lock, so no
|
|
* real ordering issues can occur. In the near future, only
|
|
* hardware devices will grab instance lock which also does not
|
|
* involve any ordering. Suppress lockdep ordering warnings
|
|
* until (if) we start grabbing instance lock on pure SW
|
|
* devices (bond/team/veth/etc).
|
|
*/
|
|
if (a == b)
|
|
return 0;
|
|
return -1;
|
|
}
|
|
|
|
#define netdev_lockdep_set_classes(dev) \
|
|
{ \
|
|
static struct lock_class_key qdisc_tx_busylock_key; \
|
|
static struct lock_class_key qdisc_xmit_lock_key; \
|
|
static struct lock_class_key dev_addr_list_lock_key; \
|
|
static struct lock_class_key dev_instance_lock_key; \
|
|
unsigned int i; \
|
|
\
|
|
(dev)->qdisc_tx_busylock = &qdisc_tx_busylock_key; \
|
|
lockdep_set_class(&(dev)->addr_list_lock, \
|
|
&dev_addr_list_lock_key); \
|
|
lockdep_set_class(&(dev)->lock, \
|
|
&dev_instance_lock_key); \
|
|
lock_set_cmp_fn(&dev->lock, netdev_lock_cmp_fn, NULL); \
|
|
for (i = 0; i < (dev)->num_tx_queues; i++) \
|
|
lockdep_set_class(&(dev)->_tx[i]._xmit_lock, \
|
|
&qdisc_xmit_lock_key); \
|
|
}
|
|
|
|
int netdev_debug_event(struct notifier_block *nb, unsigned long event,
|
|
void *ptr);
|
|
|
|
#endif
|