Commit e6484930 authored by Tom Herbert's avatar Tom Herbert Committed by David S. Miller

net: allocate tx queues in register_netdevice

This patch introduces netif_alloc_netdev_queues which is called from
register_device instead of alloc_netdev_mq.  This makes TX queue
allocation symmetric with RX allocation.  Also, queue locks allocation
is done in netdev_init_one_queue.  Change set_real_num_tx_queues to
fail if requested number < 1 or greater than number of allocated
queues.
Signed-off-by: default avatarTom Herbert <therbert@google.com>
Acked-by: default avatarEric Dumazet <eric.dumazet@gmail.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent bd25fa7b
...@@ -1696,8 +1696,8 @@ static inline int netif_is_multiqueue(const struct net_device *dev) ...@@ -1696,8 +1696,8 @@ static inline int netif_is_multiqueue(const struct net_device *dev)
return dev->num_tx_queues > 1; return dev->num_tx_queues > 1;
} }
extern void netif_set_real_num_tx_queues(struct net_device *dev, extern int netif_set_real_num_tx_queues(struct net_device *dev,
unsigned int txq); unsigned int txq);
#ifdef CONFIG_RPS #ifdef CONFIG_RPS
extern int netif_set_real_num_rx_queues(struct net_device *dev, extern int netif_set_real_num_rx_queues(struct net_device *dev,
......
...@@ -1553,18 +1553,20 @@ static void dev_queue_xmit_nit(struct sk_buff *skb, struct net_device *dev) ...@@ -1553,18 +1553,20 @@ static void dev_queue_xmit_nit(struct sk_buff *skb, struct net_device *dev)
* Routine to help set real_num_tx_queues. To avoid skbs mapped to queues * Routine to help set real_num_tx_queues. To avoid skbs mapped to queues
* greater then real_num_tx_queues stale skbs on the qdisc must be flushed. * greater then real_num_tx_queues stale skbs on the qdisc must be flushed.
*/ */
void netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq) int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq)
{ {
unsigned int real_num = dev->real_num_tx_queues; if (txq < 1 || txq > dev->num_tx_queues)
return -EINVAL;
if (unlikely(txq > dev->num_tx_queues)) if (dev->reg_state == NETREG_REGISTERED) {
; ASSERT_RTNL();
else if (txq > real_num)
dev->real_num_tx_queues = txq; if (txq < dev->real_num_tx_queues)
else if (txq < real_num) { qdisc_reset_all_tx_gt(dev, txq);
dev->real_num_tx_queues = txq;
qdisc_reset_all_tx_gt(dev, txq);
} }
dev->real_num_tx_queues = txq;
return 0;
} }
EXPORT_SYMBOL(netif_set_real_num_tx_queues); EXPORT_SYMBOL(netif_set_real_num_tx_queues);
...@@ -4928,20 +4930,6 @@ static void rollback_registered(struct net_device *dev) ...@@ -4928,20 +4930,6 @@ static void rollback_registered(struct net_device *dev)
rollback_registered_many(&single); rollback_registered_many(&single);
} }
static void __netdev_init_queue_locks_one(struct net_device *dev,
struct netdev_queue *dev_queue,
void *_unused)
{
spin_lock_init(&dev_queue->_xmit_lock);
netdev_set_xmit_lockdep_class(&dev_queue->_xmit_lock, dev->type);
dev_queue->xmit_lock_owner = -1;
}
static void netdev_init_queue_locks(struct net_device *dev)
{
netdev_for_each_tx_queue(dev, __netdev_init_queue_locks_one, NULL);
}
unsigned long netdev_fix_features(unsigned long features, const char *name) unsigned long netdev_fix_features(unsigned long features, const char *name)
{ {
/* Fix illegal SG+CSUM combinations. */ /* Fix illegal SG+CSUM combinations. */
...@@ -5034,6 +5022,41 @@ static int netif_alloc_rx_queues(struct net_device *dev) ...@@ -5034,6 +5022,41 @@ static int netif_alloc_rx_queues(struct net_device *dev)
return 0; return 0;
} }
static int netif_alloc_netdev_queues(struct net_device *dev)
{
unsigned int count = dev->num_tx_queues;
struct netdev_queue *tx;
BUG_ON(count < 1);
tx = kcalloc(count, sizeof(struct netdev_queue), GFP_KERNEL);
if (!tx) {
pr_err("netdev: Unable to allocate %u tx queues.\n",
count);
return -ENOMEM;
}
dev->_tx = tx;
return 0;
}
static void netdev_init_one_queue(struct net_device *dev,
struct netdev_queue *queue,
void *_unused)
{
queue->dev = dev;
/* Initialize queue lock */
spin_lock_init(&queue->_xmit_lock);
netdev_set_xmit_lockdep_class(&queue->_xmit_lock, dev->type);
queue->xmit_lock_owner = -1;
}
static void netdev_init_queues(struct net_device *dev)
{
netdev_for_each_tx_queue(dev, netdev_init_one_queue, NULL);
spin_lock_init(&dev->tx_global_lock);
}
/** /**
* register_netdevice - register a network device * register_netdevice - register a network device
* @dev: device to register * @dev: device to register
...@@ -5067,7 +5090,6 @@ int register_netdevice(struct net_device *dev) ...@@ -5067,7 +5090,6 @@ int register_netdevice(struct net_device *dev)
spin_lock_init(&dev->addr_list_lock); spin_lock_init(&dev->addr_list_lock);
netdev_set_addr_lockdep_class(dev); netdev_set_addr_lockdep_class(dev);
netdev_init_queue_locks(dev);
dev->iflink = -1; dev->iflink = -1;
...@@ -5075,6 +5097,12 @@ int register_netdevice(struct net_device *dev) ...@@ -5075,6 +5097,12 @@ int register_netdevice(struct net_device *dev)
if (ret) if (ret)
goto out; goto out;
ret = netif_alloc_netdev_queues(dev);
if (ret)
goto out;
netdev_init_queues(dev);
/* Init, if this function is available */ /* Init, if this function is available */
if (dev->netdev_ops->ndo_init) { if (dev->netdev_ops->ndo_init) {
ret = dev->netdev_ops->ndo_init(dev); ret = dev->netdev_ops->ndo_init(dev);
...@@ -5456,19 +5484,6 @@ struct rtnl_link_stats64 *dev_get_stats(struct net_device *dev, ...@@ -5456,19 +5484,6 @@ struct rtnl_link_stats64 *dev_get_stats(struct net_device *dev,
} }
EXPORT_SYMBOL(dev_get_stats); EXPORT_SYMBOL(dev_get_stats);
static void netdev_init_one_queue(struct net_device *dev,
struct netdev_queue *queue,
void *_unused)
{
queue->dev = dev;
}
static void netdev_init_queues(struct net_device *dev)
{
netdev_for_each_tx_queue(dev, netdev_init_one_queue, NULL);
spin_lock_init(&dev->tx_global_lock);
}
struct netdev_queue *dev_ingress_queue_create(struct net_device *dev) struct netdev_queue *dev_ingress_queue_create(struct net_device *dev)
{ {
struct netdev_queue *queue = dev_ingress_queue(dev); struct netdev_queue *queue = dev_ingress_queue(dev);
...@@ -5480,7 +5495,6 @@ struct netdev_queue *dev_ingress_queue_create(struct net_device *dev) ...@@ -5480,7 +5495,6 @@ struct netdev_queue *dev_ingress_queue_create(struct net_device *dev)
if (!queue) if (!queue)
return NULL; return NULL;
netdev_init_one_queue(dev, queue, NULL); netdev_init_one_queue(dev, queue, NULL);
__netdev_init_queue_locks_one(dev, queue, NULL);
queue->qdisc = &noop_qdisc; queue->qdisc = &noop_qdisc;
queue->qdisc_sleeping = &noop_qdisc; queue->qdisc_sleeping = &noop_qdisc;
rcu_assign_pointer(dev->ingress_queue, queue); rcu_assign_pointer(dev->ingress_queue, queue);
...@@ -5502,7 +5516,6 @@ struct netdev_queue *dev_ingress_queue_create(struct net_device *dev) ...@@ -5502,7 +5516,6 @@ struct netdev_queue *dev_ingress_queue_create(struct net_device *dev)
struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name, struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name,
void (*setup)(struct net_device *), unsigned int queue_count) void (*setup)(struct net_device *), unsigned int queue_count)
{ {
struct netdev_queue *tx;
struct net_device *dev; struct net_device *dev;
size_t alloc_size; size_t alloc_size;
struct net_device *p; struct net_device *p;
...@@ -5530,20 +5543,12 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name, ...@@ -5530,20 +5543,12 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name,
return NULL; return NULL;
} }
tx = kcalloc(queue_count, sizeof(struct netdev_queue), GFP_KERNEL);
if (!tx) {
printk(KERN_ERR "alloc_netdev: Unable to allocate "
"tx qdiscs.\n");
goto free_p;
}
dev = PTR_ALIGN(p, NETDEV_ALIGN); dev = PTR_ALIGN(p, NETDEV_ALIGN);
dev->padded = (char *)dev - (char *)p; dev->padded = (char *)dev - (char *)p;
dev->pcpu_refcnt = alloc_percpu(int); dev->pcpu_refcnt = alloc_percpu(int);
if (!dev->pcpu_refcnt) if (!dev->pcpu_refcnt)
goto free_tx; goto free_p;
if (dev_addr_init(dev)) if (dev_addr_init(dev))
goto free_pcpu; goto free_pcpu;
...@@ -5553,7 +5558,6 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name, ...@@ -5553,7 +5558,6 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name,
dev_net_set(dev, &init_net); dev_net_set(dev, &init_net);
dev->_tx = tx;
dev->num_tx_queues = queue_count; dev->num_tx_queues = queue_count;
dev->real_num_tx_queues = queue_count; dev->real_num_tx_queues = queue_count;
...@@ -5564,8 +5568,6 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name, ...@@ -5564,8 +5568,6 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name,
dev->gso_max_size = GSO_MAX_SIZE; dev->gso_max_size = GSO_MAX_SIZE;
netdev_init_queues(dev);
INIT_LIST_HEAD(&dev->ethtool_ntuple_list.list); INIT_LIST_HEAD(&dev->ethtool_ntuple_list.list);
dev->ethtool_ntuple_list.count = 0; dev->ethtool_ntuple_list.count = 0;
INIT_LIST_HEAD(&dev->napi_list); INIT_LIST_HEAD(&dev->napi_list);
...@@ -5576,8 +5578,6 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name, ...@@ -5576,8 +5578,6 @@ struct net_device *alloc_netdev_mq(int sizeof_priv, const char *name,
strcpy(dev->name, name); strcpy(dev->name, name);
return dev; return dev;
free_tx:
kfree(tx);
free_pcpu: free_pcpu:
free_percpu(dev->pcpu_refcnt); free_percpu(dev->pcpu_refcnt);
free_p: free_p:
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment