git://git.onelab.eu
/
linux-2.6.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
fedora core 6 1.2949 + vserver 2.2.0
[linux-2.6.git]
/
net
/
sched
/
sch_generic.c
diff --git
a/net/sched/sch_generic.c
b/net/sched/sch_generic.c
index
0326752
..
bc116bd
100644
(file)
--- a/
net/sched/sch_generic.c
+++ b/
net/sched/sch_generic.c
@@
-14,7
+14,6
@@
#include <asm/uaccess.h>
#include <asm/system.h>
#include <linux/bitops.h>
#include <asm/uaccess.h>
#include <asm/system.h>
#include <linux/bitops.h>
-#include <linux/config.h>
#include <linux/module.h>
#include <linux/types.h>
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/types.h>
#include <linux/kernel.h>
@@
-71,9
+70,9
@@
void qdisc_unlock_tree(struct net_device *dev)
dev->queue_lock serializes queue accesses for this device
AND dev->qdisc pointer itself.
dev->queue_lock serializes queue accesses for this device
AND dev->qdisc pointer itself.
-
dev->xmit
_lock serializes accesses to device driver.
+
netif_tx
_lock serializes accesses to device driver.
- dev->queue_lock and
dev->xmit
_lock are mutually exclusive,
+ dev->queue_lock and
netif_tx
_lock are mutually exclusive,
if one is grabbed, another must be free.
*/
if one is grabbed, another must be free.
*/
@@
-89,14
+88,17
@@
void qdisc_unlock_tree(struct net_device *dev)
NOTE: Called under dev->queue_lock with locally disabled BH.
*/
NOTE: Called under dev->queue_lock with locally disabled BH.
*/
-int qdisc_restart(struct net_device *dev)
+
static inline
int qdisc_restart(struct net_device *dev)
{
struct Qdisc *q = dev->qdisc;
struct sk_buff *skb;
/* Dequeue packet */
{
struct Qdisc *q = dev->qdisc;
struct sk_buff *skb;
/* Dequeue packet */
- if ((
skb = q->dequeue(q)) != NULL
) {
+ if ((
(skb = dev->gso_skb)) || ((skb = q->dequeue(q)))
) {
unsigned nolock = (dev->features & NETIF_F_LLTX);
unsigned nolock = (dev->features & NETIF_F_LLTX);
+
+ dev->gso_skb = NULL;
+
/*
* When the driver has LLTX set it does its own locking
* in start_xmit. No need to add additional overhead by
/*
* When the driver has LLTX set it does its own locking
* in start_xmit. No need to add additional overhead by
@@
-107,7
+109,7
@@
int qdisc_restart(struct net_device *dev)
* will be requeued.
*/
if (!nolock) {
* will be requeued.
*/
if (!nolock) {
- if (!
spin_trylock(&dev->xmit_lock
)) {
+ if (!
netif_tx_trylock(dev
)) {
collision:
/* So, someone grabbed the driver. */
collision:
/* So, someone grabbed the driver. */
@@
-125,8
+127,6
@@
int qdisc_restart(struct net_device *dev)
__get_cpu_var(netdev_rx_stat).cpu_collision++;
goto requeue;
}
__get_cpu_var(netdev_rx_stat).cpu_collision++;
goto requeue;
}
- /* Remember that the driver is grabbed by us. */
- dev->xmit_lock_owner = smp_processor_id();
}
{
}
{
@@
-135,14
+135,11
@@
int qdisc_restart(struct net_device *dev)
if (!netif_queue_stopped(dev)) {
int ret;
if (!netif_queue_stopped(dev)) {
int ret;
- if (netdev_nit)
- dev_queue_xmit_nit(skb, dev);
- ret = dev
->
hard_start_xmit(skb, dev);
+ ret = dev
_
hard_start_xmit(skb, dev);
if (ret == NETDEV_TX_OK) {
if (!nolock) {
if (ret == NETDEV_TX_OK) {
if (!nolock) {
- dev->xmit_lock_owner = -1;
- spin_unlock(&dev->xmit_lock);
+ netif_tx_unlock(dev);
}
spin_lock(&dev->queue_lock);
return -1;
}
spin_lock(&dev->queue_lock);
return -1;
@@
-156,8
+153,7
@@
int qdisc_restart(struct net_device *dev)
/* NETDEV_TX_BUSY - we need to requeue */
/* Release the driver */
if (!nolock) {
/* NETDEV_TX_BUSY - we need to requeue */
/* Release the driver */
if (!nolock) {
- dev->xmit_lock_owner = -1;
- spin_unlock(&dev->xmit_lock);
+ netif_tx_unlock(dev);
}
spin_lock(&dev->queue_lock);
q = dev->qdisc;
}
spin_lock(&dev->queue_lock);
q = dev->qdisc;
@@
-174,7
+170,10
@@
int qdisc_restart(struct net_device *dev)
*/
requeue:
*/
requeue:
- q->ops->requeue(skb, q);
+ if (skb->next)
+ dev->gso_skb = skb;
+ else
+ q->ops->requeue(skb, q);
netif_schedule(dev);
return 1;
}
netif_schedule(dev);
return 1;
}
@@
-182,25
+181,39
@@
requeue:
return q->q.qlen;
}
return q->q.qlen;
}
+void __qdisc_run(struct net_device *dev)
+{
+ if (unlikely(dev->qdisc == &noop_qdisc))
+ goto out;
+
+ while (qdisc_restart(dev) < 0 && !netif_queue_stopped(dev))
+ /* NOTHING */;
+
+out:
+ clear_bit(__LINK_STATE_QDISC_RUNNING, &dev->state);
+}
+
static void dev_watchdog(unsigned long arg)
{
struct net_device *dev = (struct net_device *)arg;
static void dev_watchdog(unsigned long arg)
{
struct net_device *dev = (struct net_device *)arg;
-
spin_lock(&dev->xmit_lock
);
+
netif_tx_lock(dev
);
if (dev->qdisc != &noop_qdisc) {
if (netif_device_present(dev) &&
netif_running(dev) &&
netif_carrier_ok(dev)) {
if (netif_queue_stopped(dev) &&
if (dev->qdisc != &noop_qdisc) {
if (netif_device_present(dev) &&
netif_running(dev) &&
netif_carrier_ok(dev)) {
if (netif_queue_stopped(dev) &&
- (jiffies - dev->trans_start) > dev->watchdog_timeo) {
- printk(KERN_INFO "NETDEV WATCHDOG: %s: transmit timed out\n", dev->name);
+ time_after(jiffies, dev->trans_start + dev->watchdog_timeo)) {
+
+ printk(KERN_INFO "NETDEV WATCHDOG: %s: transmit timed out\n",
+ dev->name);
dev->tx_timeout(dev);
}
if (!mod_timer(&dev->watchdog_timer, jiffies + dev->watchdog_timeo))
dev_hold(dev);
}
}
dev->tx_timeout(dev);
}
if (!mod_timer(&dev->watchdog_timer, jiffies + dev->watchdog_timeo))
dev_hold(dev);
}
}
-
spin_unlock(&dev->xmit_lock
);
+
netif_tx_unlock(dev
);
dev_put(dev);
}
dev_put(dev);
}
@@
-224,17
+237,15
@@
void __netdev_watchdog_up(struct net_device *dev)
static void dev_watchdog_up(struct net_device *dev)
{
static void dev_watchdog_up(struct net_device *dev)
{
- spin_lock_bh(&dev->xmit_lock);
__netdev_watchdog_up(dev);
__netdev_watchdog_up(dev);
- spin_unlock_bh(&dev->xmit_lock);
}
static void dev_watchdog_down(struct net_device *dev)
{
}
static void dev_watchdog_down(struct net_device *dev)
{
-
spin_lock_bh(&dev->xmit_lock
);
+
netif_tx_lock_bh(dev
);
if (del_timer(&dev->watchdog_timer))
if (del_timer(&dev->watchdog_timer))
-
__
dev_put(dev);
-
spin_unlock_bh(&dev->xmit_lock
);
+ dev_put(dev);
+
netif_tx_unlock_bh(dev
);
}
void netif_carrier_on(struct net_device *dev)
}
void netif_carrier_on(struct net_device *dev)
@@
-418,10
+429,9
@@
struct Qdisc *qdisc_alloc(struct net_device *dev, struct Qdisc_ops *ops)
size = QDISC_ALIGN(sizeof(*sch));
size += ops->priv_size + (QDISC_ALIGNTO - 1);
size = QDISC_ALIGN(sizeof(*sch));
size += ops->priv_size + (QDISC_ALIGNTO - 1);
- p = k
m
alloc(size, GFP_KERNEL);
+ p = k
z
alloc(size, GFP_KERNEL);
if (!p)
goto errout;
if (!p)
goto errout;
- memset(p, 0, size);
sch = (struct Qdisc *) QDISC_ALIGN((unsigned long) p);
sch->padded = (char *) sch - (char *) p;
sch = (struct Qdisc *) QDISC_ALIGN((unsigned long) p);
sch->padded = (char *) sch - (char *) p;
@@
-440,13
+450,15
@@
errout:
return ERR_PTR(-err);
}
return ERR_PTR(-err);
}
-struct Qdisc * qdisc_create_dflt(struct net_device *dev, struct Qdisc_ops *ops)
+struct Qdisc * qdisc_create_dflt(struct net_device *dev, struct Qdisc_ops *ops,
+ unsigned int parentid)
{
struct Qdisc *sch;
sch = qdisc_alloc(dev, ops);
if (IS_ERR(sch))
goto errout;
{
struct Qdisc *sch;
sch = qdisc_alloc(dev, ops);
if (IS_ERR(sch))
goto errout;
+ sch->parent = parentid;
if (!ops->init || ops->init(sch, NULL) == 0)
return sch;
if (!ops->init || ops->init(sch, NULL) == 0)
return sch;
@@
-510,7
+522,8
@@
void dev_activate(struct net_device *dev)
if (dev->qdisc_sleeping == &noop_qdisc) {
struct Qdisc *qdisc;
if (dev->tx_queue_len) {
if (dev->qdisc_sleeping == &noop_qdisc) {
struct Qdisc *qdisc;
if (dev->tx_queue_len) {
- qdisc = qdisc_create_dflt(dev, &pfifo_fast_ops);
+ qdisc = qdisc_create_dflt(dev, &pfifo_fast_ops,
+ TC_H_ROOT);
if (qdisc == NULL) {
printk(KERN_INFO "%s: activation failed\n", dev->name);
return;
if (qdisc == NULL) {
printk(KERN_INFO "%s: activation failed\n", dev->name);
return;
@@
-553,10
+566,17
@@
void dev_deactivate(struct net_device *dev)
dev_watchdog_down(dev);
dev_watchdog_down(dev);
- while (test_bit(__LINK_STATE_SCHED, &dev->state))
+ /* Wait for outstanding dev_queue_xmit calls. */
+ synchronize_rcu();
+
+ /* Wait for outstanding qdisc_run calls. */
+ while (test_bit(__LINK_STATE_QDISC_RUNNING, &dev->state))
yield();
yield();
- spin_unlock_wait(&dev->xmit_lock);
+ if (dev->gso_skb) {
+ kfree_skb(dev->gso_skb);
+ dev->gso_skb = NULL;
+ }
}
void dev_init_scheduler(struct net_device *dev)
}
void dev_init_scheduler(struct net_device *dev)
@@
-589,15
+609,11
@@
void dev_shutdown(struct net_device *dev)
qdisc_unlock_tree(dev);
}
qdisc_unlock_tree(dev);
}
-EXPORT_SYMBOL(__netdev_watchdog_up);
EXPORT_SYMBOL(netif_carrier_on);
EXPORT_SYMBOL(netif_carrier_off);
EXPORT_SYMBOL(noop_qdisc);
EXPORT_SYMBOL(netif_carrier_on);
EXPORT_SYMBOL(netif_carrier_off);
EXPORT_SYMBOL(noop_qdisc);
-EXPORT_SYMBOL(noop_qdisc_ops);
EXPORT_SYMBOL(qdisc_create_dflt);
EXPORT_SYMBOL(qdisc_create_dflt);
-EXPORT_SYMBOL(qdisc_alloc);
EXPORT_SYMBOL(qdisc_destroy);
EXPORT_SYMBOL(qdisc_reset);
EXPORT_SYMBOL(qdisc_destroy);
EXPORT_SYMBOL(qdisc_reset);
-EXPORT_SYMBOL(qdisc_restart);
EXPORT_SYMBOL(qdisc_lock_tree);
EXPORT_SYMBOL(qdisc_unlock_tree);
EXPORT_SYMBOL(qdisc_lock_tree);
EXPORT_SYMBOL(qdisc_unlock_tree);