Commit 2b3b2908 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'master' of master.kernel.org:/pub/scm/linux/kernel/git/davem/net-2.6

* 'master' of master.kernel.org:/pub/scm/linux/kernel/git/davem/net-2.6:
  [IEEE80211]: avoid integer underflow for runt rx frames
  [TCP]: secure_tcp_sequence_number() should not use a too fast clock
  [SFQ]: Remove artificial limitation for queue limit.
parents e80eaf99 04045f98
...@@ -1550,11 +1550,13 @@ __u32 secure_tcp_sequence_number(__be32 saddr, __be32 daddr, ...@@ -1550,11 +1550,13 @@ __u32 secure_tcp_sequence_number(__be32 saddr, __be32 daddr,
* As close as possible to RFC 793, which * As close as possible to RFC 793, which
* suggests using a 250 kHz clock. * suggests using a 250 kHz clock.
* Further reading shows this assumes 2 Mb/s networks. * Further reading shows this assumes 2 Mb/s networks.
* For 10 Gb/s Ethernet, a 1 GHz clock is appropriate. * For 10 Mb/s Ethernet, a 1 MHz clock is appropriate.
* That's funny, Linux has one built in! Use it! * For 10 Gb/s Ethernet, a 1 GHz clock should be ok, but
* (Networks are faster now - should this be increased?) * we also need to limit the resolution so that the u32 seq
* overlaps less than one time per MSL (2 minutes).
* Choosing a clock of 64 ns period is OK. (period of 274 s)
*/ */
seq += ktime_get_real().tv64; seq += ktime_get_real().tv64 >> 6;
#if 0 #if 0
printk("init_seq(%lx, %lx, %d, %d) = %d\n", printk("init_seq(%lx, %lx, %d, %d) = %d\n",
saddr, daddr, sport, dport, seq); saddr, daddr, sport, dport, seq);
......
...@@ -366,6 +366,12 @@ int ieee80211_rx(struct ieee80211_device *ieee, struct sk_buff *skb, ...@@ -366,6 +366,12 @@ int ieee80211_rx(struct ieee80211_device *ieee, struct sk_buff *skb,
frag = WLAN_GET_SEQ_FRAG(sc); frag = WLAN_GET_SEQ_FRAG(sc);
hdrlen = ieee80211_get_hdrlen(fc); hdrlen = ieee80211_get_hdrlen(fc);
if (skb->len < hdrlen) {
printk(KERN_INFO "%s: invalid SKB length %d\n",
dev->name, skb->len);
goto rx_dropped;
}
/* Put this code here so that we avoid duplicating it in all /* Put this code here so that we avoid duplicating it in all
* Rx paths. - Jean II */ * Rx paths. - Jean II */
#ifdef CONFIG_WIRELESS_EXT #ifdef CONFIG_WIRELESS_EXT
......
...@@ -19,6 +19,7 @@ ...@@ -19,6 +19,7 @@
#include <linux/init.h> #include <linux/init.h>
#include <linux/ipv6.h> #include <linux/ipv6.h>
#include <linux/skbuff.h> #include <linux/skbuff.h>
#include <linux/jhash.h>
#include <net/ip.h> #include <net/ip.h>
#include <net/netlink.h> #include <net/netlink.h>
#include <net/pkt_sched.h> #include <net/pkt_sched.h>
...@@ -95,7 +96,7 @@ struct sfq_sched_data ...@@ -95,7 +96,7 @@ struct sfq_sched_data
/* Variables */ /* Variables */
struct timer_list perturb_timer; struct timer_list perturb_timer;
int perturbation; u32 perturbation;
sfq_index tail; /* Index of current slot in round */ sfq_index tail; /* Index of current slot in round */
sfq_index max_depth; /* Maximal depth */ sfq_index max_depth; /* Maximal depth */
...@@ -109,12 +110,7 @@ struct sfq_sched_data ...@@ -109,12 +110,7 @@ struct sfq_sched_data
static __inline__ unsigned sfq_fold_hash(struct sfq_sched_data *q, u32 h, u32 h1) static __inline__ unsigned sfq_fold_hash(struct sfq_sched_data *q, u32 h, u32 h1)
{ {
int pert = q->perturbation; return jhash_2words(h, h1, q->perturbation) & (SFQ_HASH_DIVISOR - 1);
/* Have we any rotation primitives? If not, WHY? */
h ^= (h1<<pert) ^ (h1>>(0x1F - pert));
h ^= h>>10;
return h & 0x3FF;
} }
static unsigned sfq_hash(struct sfq_sched_data *q, struct sk_buff *skb) static unsigned sfq_hash(struct sfq_sched_data *q, struct sk_buff *skb)
...@@ -256,6 +252,13 @@ sfq_enqueue(struct sk_buff *skb, struct Qdisc* sch) ...@@ -256,6 +252,13 @@ sfq_enqueue(struct sk_buff *skb, struct Qdisc* sch)
q->ht[hash] = x = q->dep[SFQ_DEPTH].next; q->ht[hash] = x = q->dep[SFQ_DEPTH].next;
q->hash[x] = hash; q->hash[x] = hash;
} }
/* If selected queue has length q->limit, this means that
* all another queues are empty and that we do simple tail drop,
* i.e. drop _this_ packet.
*/
if (q->qs[x].qlen >= q->limit)
return qdisc_drop(skb, sch);
sch->qstats.backlog += skb->len; sch->qstats.backlog += skb->len;
__skb_queue_tail(&q->qs[x], skb); __skb_queue_tail(&q->qs[x], skb);
sfq_inc(q, x); sfq_inc(q, x);
...@@ -294,6 +297,19 @@ sfq_requeue(struct sk_buff *skb, struct Qdisc* sch) ...@@ -294,6 +297,19 @@ sfq_requeue(struct sk_buff *skb, struct Qdisc* sch)
} }
sch->qstats.backlog += skb->len; sch->qstats.backlog += skb->len;
__skb_queue_head(&q->qs[x], skb); __skb_queue_head(&q->qs[x], skb);
/* If selected queue has length q->limit+1, this means that
* all another queues are empty and we do simple tail drop.
* This packet is still requeued at head of queue, tail packet
* is dropped.
*/
if (q->qs[x].qlen > q->limit) {
skb = q->qs[x].prev;
__skb_unlink(skb, &q->qs[x]);
sch->qstats.drops++;
sch->qstats.backlog -= skb->len;
kfree_skb(skb);
return NET_XMIT_CN;
}
sfq_inc(q, x); sfq_inc(q, x);
if (q->qs[x].qlen == 1) { /* The flow is new */ if (q->qs[x].qlen == 1) { /* The flow is new */
if (q->tail == SFQ_DEPTH) { /* It is the first flow */ if (q->tail == SFQ_DEPTH) { /* It is the first flow */
...@@ -370,12 +386,10 @@ static void sfq_perturbation(unsigned long arg) ...@@ -370,12 +386,10 @@ static void sfq_perturbation(unsigned long arg)
struct Qdisc *sch = (struct Qdisc*)arg; struct Qdisc *sch = (struct Qdisc*)arg;
struct sfq_sched_data *q = qdisc_priv(sch); struct sfq_sched_data *q = qdisc_priv(sch);
q->perturbation = net_random()&0x1F; get_random_bytes(&q->perturbation, 4);
if (q->perturb_period) { if (q->perturb_period)
q->perturb_timer.expires = jiffies + q->perturb_period; mod_timer(&q->perturb_timer, jiffies + q->perturb_period);
add_timer(&q->perturb_timer);
}
} }
static int sfq_change(struct Qdisc *sch, struct rtattr *opt) static int sfq_change(struct Qdisc *sch, struct rtattr *opt)
...@@ -391,7 +405,7 @@ static int sfq_change(struct Qdisc *sch, struct rtattr *opt) ...@@ -391,7 +405,7 @@ static int sfq_change(struct Qdisc *sch, struct rtattr *opt)
q->quantum = ctl->quantum ? : psched_mtu(sch->dev); q->quantum = ctl->quantum ? : psched_mtu(sch->dev);
q->perturb_period = ctl->perturb_period*HZ; q->perturb_period = ctl->perturb_period*HZ;
if (ctl->limit) if (ctl->limit)
q->limit = min_t(u32, ctl->limit, SFQ_DEPTH - 2); q->limit = min_t(u32, ctl->limit, SFQ_DEPTH - 1);
qlen = sch->q.qlen; qlen = sch->q.qlen;
while (sch->q.qlen > q->limit) while (sch->q.qlen > q->limit)
...@@ -400,8 +414,8 @@ static int sfq_change(struct Qdisc *sch, struct rtattr *opt) ...@@ -400,8 +414,8 @@ static int sfq_change(struct Qdisc *sch, struct rtattr *opt)
del_timer(&q->perturb_timer); del_timer(&q->perturb_timer);
if (q->perturb_period) { if (q->perturb_period) {
q->perturb_timer.expires = jiffies + q->perturb_period; mod_timer(&q->perturb_timer, jiffies + q->perturb_period);
add_timer(&q->perturb_timer); get_random_bytes(&q->perturbation, 4);
} }
sch_tree_unlock(sch); sch_tree_unlock(sch);
return 0; return 0;
...@@ -423,12 +437,13 @@ static int sfq_init(struct Qdisc *sch, struct rtattr *opt) ...@@ -423,12 +437,13 @@ static int sfq_init(struct Qdisc *sch, struct rtattr *opt)
q->dep[i+SFQ_DEPTH].next = i+SFQ_DEPTH; q->dep[i+SFQ_DEPTH].next = i+SFQ_DEPTH;
q->dep[i+SFQ_DEPTH].prev = i+SFQ_DEPTH; q->dep[i+SFQ_DEPTH].prev = i+SFQ_DEPTH;
} }
q->limit = SFQ_DEPTH - 2; q->limit = SFQ_DEPTH - 1;
q->max_depth = 0; q->max_depth = 0;
q->tail = SFQ_DEPTH; q->tail = SFQ_DEPTH;
if (opt == NULL) { if (opt == NULL) {
q->quantum = psched_mtu(sch->dev); q->quantum = psched_mtu(sch->dev);
q->perturb_period = 0; q->perturb_period = 0;
get_random_bytes(&q->perturbation, 4);
} else { } else {
int err = sfq_change(sch, opt); int err = sfq_change(sch, opt);
if (err) if (err)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment