dect
/
linux-2.6
Archived
13
0
Fork 0

netdev: Move _xmit_lock and xmit_lock_owner into netdev_queue.

Accesses are mostly structured such that when there are multiple TX
queues the code transformations will be a little bit simpler.

Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
David S. Miller 2008-07-08 23:13:53 -07:00
parent eb6aafe3f8
commit c773e847ea
10 changed files with 146 additions and 45 deletions

View File

@ -5019,6 +5019,17 @@ static int bond_check_params(struct bond_params *params)
static struct lock_class_key bonding_netdev_xmit_lock_key; static struct lock_class_key bonding_netdev_xmit_lock_key;
static void bond_set_lockdep_class_one(struct netdev_queue *txq)
{
lockdep_set_class(&txq->_xmit_lock,
&bonding_netdev_xmit_lock_key);
}
static void bond_set_lockdep_class(struct net_device *dev)
{
bond_set_lockdep_class_one(&dev->tx_queue);
}
/* Create a new bond based on the specified name and bonding parameters. /* Create a new bond based on the specified name and bonding parameters.
* If name is NULL, obtain a suitable "bond%d" name for us. * If name is NULL, obtain a suitable "bond%d" name for us.
* Caller must NOT hold rtnl_lock; we need to release it here before we * Caller must NOT hold rtnl_lock; we need to release it here before we
@ -5076,7 +5087,7 @@ int bond_create(char *name, struct bond_params *params)
goto out_bond; goto out_bond;
} }
lockdep_set_class(&bond_dev->_xmit_lock, &bonding_netdev_xmit_lock_key); bond_set_lockdep_class(bond_dev);
netif_carrier_off(bond_dev); netif_carrier_off(bond_dev);

View File

@ -124,6 +124,16 @@ static LIST_HEAD(bpq_devices);
*/ */
static struct lock_class_key bpq_netdev_xmit_lock_key; static struct lock_class_key bpq_netdev_xmit_lock_key;
static void bpq_set_lockdep_class_one(struct netdev_queue *txq)
{
lockdep_set_class(&txq->_xmit_lock, &bpq_netdev_xmit_lock_key);
}
static void bpq_set_lockdep_class(struct net_device *dev)
{
bpq_set_lockdep_class_one(&dev->tx_queue);
}
/* ------------------------------------------------------------------------ */ /* ------------------------------------------------------------------------ */
@ -523,7 +533,7 @@ static int bpq_new_device(struct net_device *edev)
err = register_netdevice(ndev); err = register_netdevice(ndev);
if (err) if (err)
goto error; goto error;
lockdep_set_class(&ndev->_xmit_lock, &bpq_netdev_xmit_lock_key); bpq_set_lockdep_class(ndev);
/* List protected by RTNL */ /* List protected by RTNL */
list_add_rcu(&bpq->bpq_list, &bpq_devices); list_add_rcu(&bpq->bpq_list, &bpq_devices);

View File

@ -277,6 +277,17 @@ static struct lock_class_key macvlan_netdev_xmit_lock_key;
#define MACVLAN_STATE_MASK \ #define MACVLAN_STATE_MASK \
((1<<__LINK_STATE_NOCARRIER) | (1<<__LINK_STATE_DORMANT)) ((1<<__LINK_STATE_NOCARRIER) | (1<<__LINK_STATE_DORMANT))
static void macvlan_set_lockdep_class_one(struct netdev_queue *txq)
{
lockdep_set_class(&txq->_xmit_lock,
&macvlan_netdev_xmit_lock_key);
}
static void macvlan_set_lockdep_class(struct net_device *dev)
{
macvlan_set_lockdep_class_one(&dev->tx_queue);
}
static int macvlan_init(struct net_device *dev) static int macvlan_init(struct net_device *dev)
{ {
struct macvlan_dev *vlan = netdev_priv(dev); struct macvlan_dev *vlan = netdev_priv(dev);
@ -287,7 +298,8 @@ static int macvlan_init(struct net_device *dev)
dev->features = lowerdev->features & MACVLAN_FEATURES; dev->features = lowerdev->features & MACVLAN_FEATURES;
dev->iflink = lowerdev->ifindex; dev->iflink = lowerdev->ifindex;
lockdep_set_class(&dev->_xmit_lock, &macvlan_netdev_xmit_lock_key); macvlan_set_lockdep_class(dev);
return 0; return 0;
} }

View File

@ -3102,6 +3102,16 @@ static void prism2_clear_set_tim_queue(local_info_t *local)
*/ */
static struct lock_class_key hostap_netdev_xmit_lock_key; static struct lock_class_key hostap_netdev_xmit_lock_key;
static void prism2_set_lockdep_class_one(struct netdev_queue *txq)
{
lockdep_set_class(&txq->_xmit_lock,
&hostap_netdev_xmit_lock_key);
}
static void prism2_set_lockdep_class(struct net_device *dev)
{
prism2_set_lockdep_class_one(&dev->tx_queue);
}
static struct net_device * static struct net_device *
prism2_init_local_data(struct prism2_helper_functions *funcs, int card_idx, prism2_init_local_data(struct prism2_helper_functions *funcs, int card_idx,
@ -3268,7 +3278,7 @@ while (0)
if (ret >= 0) if (ret >= 0)
ret = register_netdevice(dev); ret = register_netdevice(dev);
lockdep_set_class(&dev->_xmit_lock, &hostap_netdev_xmit_lock_key); prism2_set_lockdep_class(dev);
rtnl_unlock(); rtnl_unlock();
if (ret < 0) { if (ret < 0) {
printk(KERN_WARNING "%s: register netdevice failed!\n", printk(KERN_WARNING "%s: register netdevice failed!\n",

View File

@ -453,6 +453,8 @@ struct netdev_queue {
struct net_device *dev; struct net_device *dev;
struct Qdisc *qdisc; struct Qdisc *qdisc;
struct sk_buff *gso_skb; struct sk_buff *gso_skb;
spinlock_t _xmit_lock;
int xmit_lock_owner;
struct Qdisc *qdisc_sleeping; struct Qdisc *qdisc_sleeping;
struct list_head qdisc_list; struct list_head qdisc_list;
struct netdev_queue *next_sched; struct netdev_queue *next_sched;
@ -639,12 +641,6 @@ struct net_device
/* /*
* One part is mostly used on xmit path (device) * One part is mostly used on xmit path (device)
*/ */
/* hard_start_xmit synchronizer */
spinlock_t _xmit_lock ____cacheline_aligned_in_smp;
/* cpu id of processor entered to hard_start_xmit or -1,
if nobody entered there.
*/
int xmit_lock_owner;
void *priv; /* pointer to private data */ void *priv; /* pointer to private data */
int (*hard_start_xmit) (struct sk_buff *skb, int (*hard_start_xmit) (struct sk_buff *skb,
struct net_device *dev); struct net_device *dev);
@ -1402,52 +1398,72 @@ static inline void netif_rx_complete(struct net_device *dev,
* *
* Get network device transmit lock * Get network device transmit lock
*/ */
static inline void __netif_tx_lock(struct net_device *dev, int cpu) static inline void __netif_tx_lock(struct netdev_queue *txq, int cpu)
{ {
spin_lock(&dev->_xmit_lock); spin_lock(&txq->_xmit_lock);
dev->xmit_lock_owner = cpu; txq->xmit_lock_owner = cpu;
} }
static inline void netif_tx_lock(struct net_device *dev) static inline void netif_tx_lock(struct net_device *dev)
{ {
__netif_tx_lock(dev, smp_processor_id()); __netif_tx_lock(&dev->tx_queue, smp_processor_id());
}
static inline void __netif_tx_lock_bh(struct netdev_queue *txq)
{
spin_lock_bh(&txq->_xmit_lock);
txq->xmit_lock_owner = smp_processor_id();
} }
static inline void netif_tx_lock_bh(struct net_device *dev) static inline void netif_tx_lock_bh(struct net_device *dev)
{ {
spin_lock_bh(&dev->_xmit_lock); __netif_tx_lock_bh(&dev->tx_queue);
dev->xmit_lock_owner = smp_processor_id(); }
static inline int __netif_tx_trylock(struct netdev_queue *txq)
{
int ok = spin_trylock(&txq->_xmit_lock);
if (likely(ok))
txq->xmit_lock_owner = smp_processor_id();
return ok;
} }
static inline int netif_tx_trylock(struct net_device *dev) static inline int netif_tx_trylock(struct net_device *dev)
{ {
int ok = spin_trylock(&dev->_xmit_lock); return __netif_tx_trylock(&dev->tx_queue);
if (likely(ok)) }
dev->xmit_lock_owner = smp_processor_id();
return ok; static inline void __netif_tx_unlock(struct netdev_queue *txq)
{
txq->xmit_lock_owner = -1;
spin_unlock(&txq->_xmit_lock);
} }
static inline void netif_tx_unlock(struct net_device *dev) static inline void netif_tx_unlock(struct net_device *dev)
{ {
dev->xmit_lock_owner = -1; __netif_tx_unlock(&dev->tx_queue);
spin_unlock(&dev->_xmit_lock); }
static inline void __netif_tx_unlock_bh(struct netdev_queue *txq)
{
txq->xmit_lock_owner = -1;
spin_unlock_bh(&txq->_xmit_lock);
} }
static inline void netif_tx_unlock_bh(struct net_device *dev) static inline void netif_tx_unlock_bh(struct net_device *dev)
{ {
dev->xmit_lock_owner = -1; __netif_tx_unlock_bh(&dev->tx_queue);
spin_unlock_bh(&dev->_xmit_lock);
} }
#define HARD_TX_LOCK(dev, cpu) { \ #define HARD_TX_LOCK(dev, txq, cpu) { \
if ((dev->features & NETIF_F_LLTX) == 0) { \ if ((dev->features & NETIF_F_LLTX) == 0) { \
__netif_tx_lock(dev, cpu); \ __netif_tx_lock(txq, cpu); \
} \ } \
} }
#define HARD_TX_UNLOCK(dev) { \ #define HARD_TX_UNLOCK(dev, txq) { \
if ((dev->features & NETIF_F_LLTX) == 0) { \ if ((dev->features & NETIF_F_LLTX) == 0) { \
netif_tx_unlock(dev); \ __netif_tx_unlock(txq); \
} \ } \
} }

View File

@ -627,6 +627,18 @@ static void vlan_dev_set_rx_mode(struct net_device *vlan_dev)
*/ */
static struct lock_class_key vlan_netdev_xmit_lock_key; static struct lock_class_key vlan_netdev_xmit_lock_key;
static void vlan_dev_set_lockdep_one(struct netdev_queue *txq,
int subclass)
{
lockdep_set_class_and_subclass(&txq->_xmit_lock,
&vlan_netdev_xmit_lock_key, subclass);
}
static void vlan_dev_set_lockdep_class(struct net_device *dev, int subclass)
{
vlan_dev_set_lockdep_one(&dev->tx_queue, subclass);
}
static const struct header_ops vlan_header_ops = { static const struct header_ops vlan_header_ops = {
.create = vlan_dev_hard_header, .create = vlan_dev_hard_header,
.rebuild = vlan_dev_rebuild_header, .rebuild = vlan_dev_rebuild_header,
@ -668,8 +680,7 @@ static int vlan_dev_init(struct net_device *dev)
if (is_vlan_dev(real_dev)) if (is_vlan_dev(real_dev))
subclass = 1; subclass = 1;
lockdep_set_class_and_subclass(&dev->_xmit_lock, vlan_dev_set_lockdep_class(dev, subclass);
&vlan_netdev_xmit_lock_key, subclass);
return 0; return 0;
} }

View File

@ -258,7 +258,7 @@ DEFINE_PER_CPU(struct softnet_data, softnet_data);
#ifdef CONFIG_DEBUG_LOCK_ALLOC #ifdef CONFIG_DEBUG_LOCK_ALLOC
/* /*
* register_netdevice() inits dev->_xmit_lock and sets lockdep class * register_netdevice() inits txq->_xmit_lock and sets lockdep class
* according to dev->type * according to dev->type
*/ */
static const unsigned short netdev_lock_type[] = static const unsigned short netdev_lock_type[] =
@ -1758,19 +1758,19 @@ gso:
if (dev->flags & IFF_UP) { if (dev->flags & IFF_UP) {
int cpu = smp_processor_id(); /* ok because BHs are off */ int cpu = smp_processor_id(); /* ok because BHs are off */
if (dev->xmit_lock_owner != cpu) { if (txq->xmit_lock_owner != cpu) {
HARD_TX_LOCK(dev, cpu); HARD_TX_LOCK(dev, txq, cpu);
if (!netif_queue_stopped(dev) && if (!netif_queue_stopped(dev) &&
!netif_subqueue_stopped(dev, skb)) { !netif_subqueue_stopped(dev, skb)) {
rc = 0; rc = 0;
if (!dev_hard_start_xmit(skb, dev)) { if (!dev_hard_start_xmit(skb, dev)) {
HARD_TX_UNLOCK(dev); HARD_TX_UNLOCK(dev, txq);
goto out; goto out;
} }
} }
HARD_TX_UNLOCK(dev); HARD_TX_UNLOCK(dev, txq);
if (net_ratelimit()) if (net_ratelimit())
printk(KERN_CRIT "Virtual device %s asks to " printk(KERN_CRIT "Virtual device %s asks to "
"queue packet!\n", dev->name); "queue packet!\n", dev->name);
@ -3761,6 +3761,20 @@ static void rollback_registered(struct net_device *dev)
dev_put(dev); dev_put(dev);
} }
static void __netdev_init_queue_locks_one(struct netdev_queue *dev_queue,
struct net_device *dev)
{
spin_lock_init(&dev_queue->_xmit_lock);
netdev_set_lockdep_class(&dev_queue->_xmit_lock, dev->type);
dev_queue->xmit_lock_owner = -1;
}
static void netdev_init_queue_locks(struct net_device *dev)
{
__netdev_init_queue_locks_one(&dev->tx_queue, dev);
__netdev_init_queue_locks_one(&dev->rx_queue, dev);
}
/** /**
* register_netdevice - register a network device * register_netdevice - register a network device
* @dev: device to register * @dev: device to register
@ -3795,9 +3809,7 @@ int register_netdevice(struct net_device *dev)
BUG_ON(!dev_net(dev)); BUG_ON(!dev_net(dev));
net = dev_net(dev); net = dev_net(dev);
spin_lock_init(&dev->_xmit_lock); netdev_init_queue_locks(dev);
netdev_set_lockdep_class(&dev->_xmit_lock, dev->type);
dev->xmit_lock_owner = -1;
dev->iflink = -1; dev->iflink = -1;

View File

@ -74,6 +74,16 @@ static const struct proto_ops nr_proto_ops;
*/ */
static struct lock_class_key nr_netdev_xmit_lock_key; static struct lock_class_key nr_netdev_xmit_lock_key;
static void nr_set_lockdep_one(struct netdev_queue *txq)
{
lockdep_set_class(&txq->_xmit_lock, &nr_netdev_xmit_lock_key);
}
static void nr_set_lockdep_key(struct net_device *dev)
{
nr_set_lockdep_one(&dev->tx_queue);
}
/* /*
* Socket removal during an interrupt is now safe. * Socket removal during an interrupt is now safe.
*/ */
@ -1430,7 +1440,7 @@ static int __init nr_proto_init(void)
free_netdev(dev); free_netdev(dev);
goto fail; goto fail;
} }
lockdep_set_class(&dev->_xmit_lock, &nr_netdev_xmit_lock_key); nr_set_lockdep_key(dev);
dev_nr[i] = dev; dev_nr[i] = dev;
} }

View File

@ -75,6 +75,16 @@ ax25_address rose_callsign;
*/ */
static struct lock_class_key rose_netdev_xmit_lock_key; static struct lock_class_key rose_netdev_xmit_lock_key;
static void rose_set_lockdep_one(struct netdev_queue *txq)
{
lockdep_set_class(&txq->_xmit_lock, &rose_netdev_xmit_lock_key);
}
static void rose_set_lockdep_key(struct net_device *dev)
{
rose_set_lockdep_one(&dev->tx_queue);
}
/* /*
* Convert a ROSE address into text. * Convert a ROSE address into text.
*/ */
@ -1576,7 +1586,7 @@ static int __init rose_proto_init(void)
free_netdev(dev); free_netdev(dev);
goto fail; goto fail;
} }
lockdep_set_class(&dev->_xmit_lock, &rose_netdev_xmit_lock_key); rose_set_lockdep_key(dev);
dev_rose[i] = dev; dev_rose[i] = dev;
} }

View File

@ -92,10 +92,9 @@ static inline int handle_dev_cpu_collision(struct sk_buff *skb,
struct netdev_queue *dev_queue, struct netdev_queue *dev_queue,
struct Qdisc *q) struct Qdisc *q)
{ {
struct net_device *dev = dev_queue->dev;
int ret; int ret;
if (unlikely(dev->xmit_lock_owner == smp_processor_id())) { if (unlikely(dev_queue->xmit_lock_owner == smp_processor_id())) {
/* /*
* Same CPU holding the lock. It may be a transient * Same CPU holding the lock. It may be a transient
* configuration error, when hard_start_xmit() recurses. We * configuration error, when hard_start_xmit() recurses. We
@ -105,7 +104,7 @@ static inline int handle_dev_cpu_collision(struct sk_buff *skb,
kfree_skb(skb); kfree_skb(skb);
if (net_ratelimit()) if (net_ratelimit())
printk(KERN_WARNING "Dead loop on netdevice %s, " printk(KERN_WARNING "Dead loop on netdevice %s, "
"fix it urgently!\n", dev->name); "fix it urgently!\n", dev_queue->dev->name);
ret = qdisc_qlen(q); ret = qdisc_qlen(q);
} else { } else {
/* /*
@ -155,10 +154,10 @@ static inline int qdisc_restart(struct netdev_queue *txq)
dev = txq->dev; dev = txq->dev;
HARD_TX_LOCK(dev, smp_processor_id()); HARD_TX_LOCK(dev, txq, smp_processor_id());
if (!netif_subqueue_stopped(dev, skb)) if (!netif_subqueue_stopped(dev, skb))
ret = dev_hard_start_xmit(skb, dev); ret = dev_hard_start_xmit(skb, dev);
HARD_TX_UNLOCK(dev); HARD_TX_UNLOCK(dev, txq);
spin_lock(&txq->lock); spin_lock(&txq->lock);
q = txq->qdisc; q = txq->qdisc;