-diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c
---- madwifi.old/ath/if_ath.c 2007-06-01 11:26:04.181689864 +0200
-+++ madwifi.dev/ath/if_ath.c 2007-06-01 11:39:53.078678368 +0200
-@@ -167,7 +167,7 @@
- int, u_int32_t);
+Index: madwifi-trunk-r3314/ath/if_ath.c
+===================================================================
+--- madwifi-trunk-r3314.orig/ath/if_ath.c 2008-03-06 23:49:24.344018025 +0100
++++ madwifi-trunk-r3314/ath/if_ath.c 2008-03-07 00:38:01.010996975 +0100
+@@ -184,7 +184,11 @@
+ struct sk_buff *, int, int, u_int64_t);
static void ath_setdefantenna(struct ath_softc *, u_int);
static struct ath_txq *ath_txq_setup(struct ath_softc *, int, int);
-static void ath_rx_tasklet(TQUEUE_ARG);
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++static int ath_rx_poll(struct napi_struct *napi, int budget);
++#else
+static int ath_rx_poll(struct net_device *dev, int *budget);
++#endif
static int ath_hardstart(struct sk_buff *, struct net_device *);
static int ath_mgtstart(struct ieee80211com *, struct sk_buff *);
#ifdef ATH_SUPERG_COMP
-@@ -443,7 +443,6 @@
- ATH_TXBUF_LOCK_INIT(sc);
- ATH_RXBUF_LOCK_INIT(sc);
+@@ -374,6 +378,9 @@
+ u_int32_t new_clamped_maxtxpower);
+ static u_int32_t ath_get_real_maxtxpower(struct ath_softc *sc);
+
++static void ath_poll_disable(struct net_device *dev);
++static void ath_poll_enable(struct net_device *dev);
++
+ /* calibrate every 30 secs in steady state but check every second at first. */
+ static int ath_calinterval = ATH_SHORT_CALINTERVAL;
+ static int ath_countrycode = CTRY_DEFAULT; /* country code */
+@@ -545,7 +552,6 @@
+
+ atomic_set(&sc->sc_txbuf_counter, 0);
- ATH_INIT_TQUEUE(&sc->sc_rxtq, ath_rx_tasklet, dev);
ATH_INIT_TQUEUE(&sc->sc_txtq, ath_tx_tasklet, dev);
ATH_INIT_TQUEUE(&sc->sc_bmisstq, ath_bmiss_tasklet, dev);
ATH_INIT_TQUEUE(&sc->sc_bstucktq, ath_bstuck_tasklet, dev);
-@@ -700,6 +699,8 @@
+@@ -819,6 +825,12 @@
dev->set_mac_address = ath_set_mac_address;
- dev->change_mtu = ath_change_mtu;
- dev->tx_queue_len = ATH_TXBUF - 1; /* 1 for mgmt frame */
+ dev->change_mtu = ath_change_mtu;
+ dev->tx_queue_len = ATH_TXBUF - ATH_TXBUF_MGT_RESERVED;
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ netif_napi_add(dev, &sc->sc_napi, ath_rx_poll, 64);
++#else
+ dev->poll = ath_rx_poll;
+ dev->weight = 64;
++#endif
#ifdef USE_HEADERLEN_RESV
dev->hard_header_len += sizeof(struct ieee80211_qosframe) +
sizeof(struct llc) +
-@@ -1665,6 +1666,7 @@
- */
- ath_hal_getisr(ah, &status); /* NB: clears ISR too */
- DPRINTF(sc, ATH_DEBUG_INTR, "%s: status 0x%x\n", __func__, status);
+@@ -2213,6 +2225,7 @@
+ (status & HAL_INT_GLOBAL) ? " HAL_INT_GLOBAL" : ""
+ );
+
+ sc->sc_isr = status;
status &= sc->sc_imask; /* discard unasked for bits */
- if (status & HAL_INT_FATAL) {
- sc->sc_stats.ast_hardware++;
-@@ -1700,7 +1702,12 @@
- if (status & HAL_INT_RX) {
- sc->sc_tsf = ath_hal_gettsf64(ah);
- ath_uapsd_processtriggers(sc);
+ /* As soon as we know we have a real interrupt we intend to service,
+ * we will check to see if we need an initial hardware TSF reading.
+@@ -2270,7 +2283,23 @@
+ }
+ if (status & (HAL_INT_RX | HAL_INT_RXPHY)) {
+ ath_uapsd_processtriggers(sc, hw_tsf);
- ATH_SCHEDULE_TQUEUE(&sc->sc_rxtq, &needmark);
+ sc->sc_isr &= ~HAL_INT_RX;
-+ if (netif_rx_schedule_prep(dev)) {
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ if (netif_rx_schedule_prep(dev, &sc->sc_napi))
++#else
++ if (netif_rx_schedule_prep(dev))
++#endif
++ {
++#ifndef ATH_PRECISE_TSF
+ sc->sc_imask &= ~HAL_INT_RX;
+ ath_hal_intrset(ah, sc->sc_imask);
++#endif
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ __netif_rx_schedule(dev, &sc->sc_napi);
++#else
+ __netif_rx_schedule(dev);
++#endif
+ }
}
if (status & HAL_INT_TX) {
#ifdef ATH_SUPERG_DYNTURBO
-@@ -1726,6 +1733,11 @@
+@@ -2296,6 +2325,11 @@
}
}
#endif
ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, &needmark);
}
if (status & HAL_INT_BMISS) {
-@@ -3296,10 +3308,10 @@
- *
- * XXX Using in_softirq is not right since we might
- * be called from other soft irq contexts than
-- * ath_rx_tasklet.
+@@ -2508,6 +2542,7 @@
+ if (sc->sc_tx99 != NULL)
+ sc->sc_tx99->start(sc->sc_tx99);
+ #endif
++ ath_poll_enable(dev);
+
+ done:
+ ATH_UNLOCK(sc);
+@@ -2548,6 +2583,9 @@
+ if (sc->sc_tx99 != NULL)
+ sc->sc_tx99->stop(sc->sc_tx99);
+ #endif
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ ath_poll_disable(dev);
++#endif
+ netif_stop_queue(dev); /* XXX re-enabled by ath_newstate */
+ dev->flags &= ~IFF_RUNNING; /* NB: avoid recursion */
+ ieee80211_stop_running(ic); /* stop all VAPs */
+@@ -4006,6 +4044,39 @@
+ return ath_keyset(sc, k, mac, vap->iv_bss);
+ }
+
++static void ath_poll_disable(struct net_device *dev)
++{
++ struct ath_softc *sc = dev->priv;
++
++ /*
++ * XXX Using in_softirq is not right since we might
++ * be called from other soft irq contexts than
+ * ath_rx_poll
++ */
++ if (!in_softirq()) {
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ napi_disable(&sc->sc_napi);
++#else
++ netif_poll_disable(dev);
++#endif
++ }
++}
++
++static void ath_poll_enable(struct net_device *dev)
++{
++ struct ath_softc *sc = dev->priv;
++
++ /* NB: see above */
++ if (!in_softirq()) {
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ napi_enable(&sc->sc_napi);
++#else
++ netif_poll_enable(dev);
++#endif
++ }
++}
++
++
+ /*
+ * Block/unblock tx+rx processing while a key change is done.
+ * We assume the caller serializes key management operations
+@@ -4023,13 +4094,7 @@
+ * When called from the rx tasklet we cannot use
+ * tasklet_disable because it will block waiting
+ * for us to complete execution.
+- *
+- * XXX Using in_softirq is not right since we might
+- * be called from other soft irq contexts than
+- * ath_rx_tasklet.
*/
- if (!in_softirq())
+- if (!in_softirq())
- tasklet_disable(&sc->sc_rxtq);
-+ netif_poll_disable(dev);
netif_stop_queue(dev);
}
-@@ -3312,7 +3324,7 @@
- DPRINTF(sc, ATH_DEBUG_KEYCACHE, "%s:\n", __func__);
- netif_start_queue(dev);
- if (!in_softirq()) /* NB: see above */
+@@ -4040,9 +4105,9 @@
+ struct ath_softc *sc = dev->priv;
+
+ DPRINTF(sc, ATH_DEBUG_KEYCACHE, "End\n");
+- netif_wake_queue(dev);
+- if (!in_softirq()) /* NB: see above */
- tasklet_enable(&sc->sc_rxtq);
-+ netif_poll_enable(dev);
++
++ if (dev->flags&IFF_RUNNING)
++ netif_wake_queue(dev);
}
/*
-@@ -5573,13 +5585,12 @@
+@@ -6347,15 +6412,25 @@
sc->sc_rxotherant = 0;
}
-static void
-ath_rx_tasklet(TQUEUE_ARG data)
+static int
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ath_rx_poll(struct napi_struct *napi, int budget)
++#else
+ath_rx_poll(struct net_device *dev, int *budget)
++#endif
{
#define PA2DESC(_sc, _pa) \
((struct ath_desc *)((caddr_t)(_sc)->sc_rxdma.dd_desc + \
((_pa) - (_sc)->sc_rxdma.dd_desc_paddr)))
- struct net_device *dev = (struct net_device *)data;
- struct ath_buf *bf;
+- struct ath_buf *bf;
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ struct ath_softc *sc = container_of(napi, struct ath_softc, sc_napi);
++ struct net_device *dev = sc->sc_dev;
++ u_int rx_limit = budget;
++#else
struct ath_softc *sc = dev->priv;
++ u_int rx_limit = min(dev->quota, *budget);
++#endif
++ struct ath_buf *bf;
struct ieee80211com *ic = &sc->sc_ic;
-@@ -5591,12 +5602,15 @@
+ struct ath_hal *ah = sc ? sc->sc_ah : NULL;
+ struct ath_desc *ds;
+@@ -6365,8 +6440,10 @@
unsigned int len;
int type;
u_int phyerr;
+ u_int processed = 0, early_stop = 0;
-+ u_int rx_limit = dev->quota;
- /* Let the 802.11 layer know about the new noise floor */
- sc->sc_channoise = ath_hal_get_channel_noise(ah, &(sc->sc_curchan));
- ic->ic_channoise = sc->sc_channoise;
-
- DPRINTF(sc, ATH_DEBUG_RX_PROC, "%s\n", __func__);
+ DPRINTF(sc, ATH_DEBUG_RX_PROC, "invoked\n");
+process_rx_again:
do {
bf = STAILQ_FIRST(&sc->sc_rxbuf);
if (bf == NULL) { /* XXX ??? can this happen */
-@@ -5620,6 +5634,13 @@
+@@ -6390,6 +6467,15 @@
/* NB: never process the self-linked entry at the end */
break;
}
+
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
+ processed++;
++#endif
+ if (rx_limit-- < 0) {
+ early_stop = 1;
+ break;
+ }
+
skb = bf->bf_skb;
- if (skb == NULL) { /* XXX ??? can this happen */
- printk("%s: no skbuff (%s)\n", DEV_NAME(dev), __func__);
-@@ -5658,6 +5679,7 @@
+ if (skb == NULL) {
+ EPRINTF(sc, "Dropping; buffer contains NULL skbuff.\n");
+@@ -6437,6 +6523,7 @@
sc->sc_stats.ast_rx_phyerr++;
phyerr = rs->rs_phyerr & 0x1f;
sc->sc_stats.ast_rx_phy[phyerr]++;
}
if (rs->rs_status & HAL_RXERR_DECRYPT) {
/*
-@@ -5868,9 +5890,29 @@
+@@ -6632,9 +6719,43 @@
STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
ATH_RXBUF_UNLOCK_IRQ(sc);
} while (ath_rxbuf_init(sc, bf) == 0);
+ if (!early_stop) {
++ unsigned long flags;
+ /* Check if more data is received while we were
+ * processing the descriptor chain.
+ */
-+ ATH_DISABLE_INTR();
++#ifndef ATH_PRECISE_TSF
++ local_irq_save(flags);
+ if (sc->sc_isr & HAL_INT_RX) {
++ u_int64_t hw_tsf = ath_hal_gettsf64(ah);
+ sc->sc_isr &= ~HAL_INT_RX;
-+ ATH_ENABLE_INTR();
-+ ath_uapsd_processtriggers(sc);
++ local_irq_restore(flags);
++ ath_uapsd_processtriggers(sc, hw_tsf);
+ goto process_rx_again;
+ }
-+ netif_rx_complete(dev);
-+
++#endif
++#ifndef ATH_PRECISE_TSF
+ sc->sc_imask |= HAL_INT_RX;
+ ath_hal_intrset(ah, sc->sc_imask);
-+ ATH_ENABLE_INTR();
++ local_irq_restore(flags);
++#endif
+ }
+
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ netif_rx_complete(dev, napi);
++#else
++ netif_rx_complete(dev);
+ *budget -= processed;
++ dev->quota -= processed;
++#endif
/* rx signal state monitoring */
ath_hal_rxmonitor(ah, &sc->sc_halstats, &sc->sc_curchan);
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ return processed;
++#else
+ return early_stop;
++#endif
#undef PA2DESC
}
-@@ -7487,11 +7529,22 @@
+@@ -8285,12 +8406,24 @@
+ {
struct net_device *dev = (struct net_device *)data;
struct ath_softc *sc = dev->priv;
++ unsigned long flags;
+process_tx_again:
if (txqactive(sc->sc_ah, 0))
if (txqactive(sc->sc_ah, sc->sc_cabq->axq_qnum))
ath_tx_processq(sc, sc->sc_cabq);
-+ ATH_DISABLE_INTR();
++ local_irq_save(flags);
+ if (sc->sc_isr & HAL_INT_TX) {
+ sc->sc_isr &= ~HAL_INT_TX;
-+ ATH_ENABLE_INTR();
++ local_irq_restore(flags);
+ goto process_tx_again;
+ }
+ sc->sc_imask |= HAL_INT_TX;
+ ath_hal_intrset(sc->sc_ah, sc->sc_imask);
-+ ATH_ENABLE_INTR();
++ local_irq_restore(flags);
+
netif_wake_queue(dev);
if (sc->sc_softled)
-@@ -7508,6 +7561,7 @@
+@@ -8306,7 +8439,9 @@
+ {
struct net_device *dev = (struct net_device *)data;
struct ath_softc *sc = dev->priv;
++ unsigned long flags;
+process_tx_again:
/*
* Process each active queue.
*/
-@@ -7528,6 +7582,16 @@
+@@ -8327,6 +8462,16 @@
if (sc->sc_uapsdq && txqactive(sc->sc_ah, sc->sc_uapsdq->axq_qnum))
ath_tx_processq(sc, sc->sc_uapsdq);
-+ ATH_DISABLE_INTR();
++ local_irq_save(flags);
+ if (sc->sc_isr & HAL_INT_TX) {
+ sc->sc_isr &= ~HAL_INT_TX;
-+ ATH_ENABLE_INTR();
++ local_irq_restore(flags);
+ goto process_tx_again;
+ }
+ sc->sc_imask |= HAL_INT_TX;
+ ath_hal_intrset(sc->sc_ah, sc->sc_imask);
-+ ATH_ENABLE_INTR();
++ local_irq_restore(flags);
+
netif_wake_queue(dev);
if (sc->sc_softled)
-@@ -7545,6 +7609,7 @@
+@@ -8342,13 +8487,25 @@
+ struct net_device *dev = (struct net_device *)data;
+ struct ath_softc *sc = dev->priv;
unsigned int i;
++ unsigned long flags;
- /* Process each active queue. */
+ /* Process each active queue. This includes sc_cabq, sc_xrtq and
+ * sc_uapsdq */
+process_tx_again:
for (i = 0; i < HAL_NUM_TX_QUEUES; i++)
if (ATH_TXQ_SETUP(sc, i) && txqactive(sc->sc_ah, i))
ath_tx_processq(sc, &sc->sc_txq[i]);
-@@ -7553,6 +7618,16 @@
- ath_tx_processq(sc, sc->sc_xrtxq);
- #endif
-+ ATH_DISABLE_INTR();
++ local_irq_save(flags);
+ if (sc->sc_isr & HAL_INT_TX) {
+ sc->sc_isr &= ~HAL_INT_TX;
-+ ATH_ENABLE_INTR();
++ local_irq_restore(flags);
+ goto process_tx_again;
+ }
+ sc->sc_imask |= HAL_INT_TX;
+ ath_hal_intrset(sc->sc_ah, sc->sc_imask);
-+ ATH_ENABLE_INTR();
++ local_irq_restore(flags);
+
netif_wake_queue(dev);
if (sc->sc_softled)
-@@ -7651,6 +7726,7 @@
+@@ -8423,6 +8580,7 @@
ath_draintxq(struct ath_softc *sc)
{
struct ath_hal *ah = sc->sc_ah;
unsigned int i;
/* XXX return value */
-@@ -9170,9 +9246,9 @@
+@@ -10281,9 +10439,9 @@
dev->mtu = mtu;
if ((dev->flags & IFF_RUNNING) && !sc->sc_invalid) {
/* NB: the rx buffers may need to be reallocated */
- tasklet_disable(&sc->sc_rxtq);
-+ netif_poll_disable(dev);
++ ath_poll_disable(dev);
error = ath_reset(dev);
- tasklet_enable(&sc->sc_rxtq);
-+ netif_poll_enable(dev);
++ ath_poll_enable(dev);
}
ATH_UNLOCK(sc);
-diff -ur madwifi.old/ath/if_athvar.h madwifi.dev/ath/if_athvar.h
---- madwifi.old/ath/if_athvar.h 2007-06-01 11:26:04.158693360 +0200
-+++ madwifi.dev/ath/if_athvar.h 2007-06-01 11:33:26.549439744 +0200
-@@ -48,6 +48,10 @@
- #include "if_athioctl.h"
- #include "net80211/ieee80211.h" /* XXX for WME_NUM_AC */
+Index: madwifi-trunk-r3314/ath/if_athvar.h
+===================================================================
+--- madwifi-trunk-r3314.orig/ath/if_athvar.h 2008-03-06 23:49:24.296015306 +0100
++++ madwifi-trunk-r3314/ath/if_athvar.h 2008-03-07 00:34:32.792958008 +0100
+@@ -53,6 +53,10 @@
+ # include <asm/bitops.h>
+ #endif
+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
+#define irqs_disabled() 0
/*
* Deduce if tasklets are available. If not then
* fall back to using the immediate work queue.
-@@ -621,7 +625,6 @@
+@@ -616,6 +620,9 @@
+ struct ath_softc {
+ struct ieee80211com sc_ic; /* NB: must be first */
+ struct net_device *sc_dev;
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ struct napi_struct sc_napi;
++#endif
+ void __iomem *sc_iobase; /* address of the device */
+ struct semaphore sc_lock; /* dev-level lock */
+ struct net_device_stats sc_devstats; /* device statistics */
+@@ -730,7 +737,6 @@
struct ath_buf *sc_rxbufcur; /* current rx buffer */
u_int32_t *sc_rxlink; /* link ptr in last RX desc */
spinlock_t sc_rxbuflock;
- struct ATH_TQ_STRUCT sc_rxtq; /* rx intr tasklet */
struct ATH_TQ_STRUCT sc_rxorntq; /* rxorn intr tasklet */
u_int8_t sc_defant; /* current default antenna */
- u_int8_t sc_rxotherant; /* rx's on non-default antenna*/
-@@ -634,6 +637,7 @@
+ u_int8_t sc_rxotherant; /* RXs on non-default antenna */
+@@ -745,6 +751,7 @@
u_int sc_txintrperiod; /* tx interrupt batching */
struct ath_txq sc_txq[HAL_NUM_TX_QUEUES];
struct ath_txq *sc_ac2q[WME_NUM_AC]; /* WME AC -> h/w qnum */
struct ATH_TQ_STRUCT sc_txtq; /* tx intr tasklet */
u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN];
struct ath_descdma sc_bdma; /* beacon descriptors */
-@@ -714,6 +718,8 @@
- #define ATH_TXBUF_LOCK_ASSERT(_sc) \
- KASSERT(spin_is_locked(&(_sc)->sc_txbuflock), ("txbuf not locked!"))
+@@ -858,6 +865,8 @@
+ #define ATH_TXBUF_LOCK_CHECK(_sc)
+ #endif
+#define ATH_DISABLE_INTR local_irq_disable
+#define ATH_ENABLE_INTR local_irq_enable
#define ATH_RXBUF_LOCK_INIT(_sc) spin_lock_init(&(_sc)->sc_rxbuflock)
#define ATH_RXBUF_LOCK_DESTROY(_sc)
-diff -ur madwifi.old/net80211/ieee80211_input.c madwifi.dev/net80211/ieee80211_input.c
---- madwifi.old/net80211/ieee80211_input.c 2007-06-01 11:26:04.183689560 +0200
-+++ madwifi.dev/net80211/ieee80211_input.c 2007-06-01 11:31:46.931583960 +0200
-@@ -1144,8 +1144,9 @@
- if (ni->ni_vlan != 0 && vap->iv_vlgrp != NULL) {
+Index: madwifi-trunk-r3314/net80211/ieee80211_input.c
+===================================================================
+--- madwifi-trunk-r3314.orig/net80211/ieee80211_input.c 2008-03-06 23:49:11.027259154 +0100
++++ madwifi-trunk-r3314/net80211/ieee80211_input.c 2008-03-07 00:34:41.589459294 +0100
+@@ -1198,7 +1198,7 @@
/* attach vlan tag */
- vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan);
-- } else
-- netif_rx(skb);
-+ } else {
-+ netif_receive_skb(skb);
-+ }
- dev->last_rx = jiffies;
- }
+ struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni;
+ if (vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan) == NET_RX_DROP) {
+- /* If netif_rx dropped the packet because
++ /* If netif_receive_skb dropped the packet because
+ * device was too busy */
+ if (ni_tmp != NULL) {
+ /* node reference was leaked */
+@@ -1209,8 +1209,8 @@
+ skb = NULL; /* SKB is no longer ours */
+ } else {
+ struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni;
+- if (netif_rx(skb) == NET_RX_DROP) {
+- /* If netif_rx dropped the packet because
++ if (netif_receive_skb(skb) == NET_RX_DROP) {
++ /* If netif_receive_skb dropped the packet because
+ * device was too busy */
+ if (ni_tmp != NULL) {
+ /* node reference was leaked */
+@@ -2322,8 +2322,8 @@
+ skb1->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */
+
+ ni_tmp = SKB_CB(skb1)->ni;
+- if (netif_rx(skb1) == NET_RX_DROP) {
+- /* If netif_rx dropped the packet because
++ if (netif_receive_skb(skb1) == NET_RX_DROP) {
++ /* If netif_receive_skb dropped the packet because
+ * device was too busy */
+ if (ni_tmp != NULL) {
+ /* node reference was leaked */
+Index: madwifi-trunk-r3314/net80211/ieee80211_monitor.c
+===================================================================
+--- madwifi-trunk-r3314.orig/net80211/ieee80211_monitor.c 2008-03-06 23:49:11.031259383 +0100
++++ madwifi-trunk-r3314/net80211/ieee80211_monitor.c 2008-03-07 00:33:59.235045655 +0100
+@@ -584,8 +584,8 @@
+ skb1->protocol =
+ __constant_htons(0x0019); /* ETH_P_80211_RAW */
+
+- if (netif_rx(skb1) == NET_RX_DROP) {
+- /* If netif_rx dropped the packet because
++ if (netif_receive_skb(skb1) == NET_RX_DROP) {
++ /* If netif_receive_skb dropped the packet because
+ * device was too busy, reclaim the ref. in
+ * the skb. */
+ if (SKB_CB(skb1)->ni != NULL)
+Index: madwifi-trunk-r3314/net80211/ieee80211_skb.c
+===================================================================
+--- madwifi-trunk-r3314.orig/net80211/ieee80211_skb.c 2008-03-06 23:49:11.039259840 +0100
++++ madwifi-trunk-r3314/net80211/ieee80211_skb.c 2008-03-06 23:49:24.384020316 +0100
+@@ -73,7 +73,7 @@
+ #undef dev_queue_xmit
+ #undef kfree_skb
+ #undef kfree_skb_fast
+-#undef netif_rx
++#undef netif_receive_skb
+ #undef pskb_copy
+ #undef skb_clone
+ #undef skb_copy
+@@ -638,8 +638,8 @@
+ grp, vlan_tag);
+ }
+
+-int netif_rx_debug(struct sk_buff *skb, const char* func, int line) {
+- return netif_rx(untrack_skb(skb, 0, func, line, __func__, __LINE__));
++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line) {
++ return netif_receive_skb(untrack_skb(skb, 0, func, line, __func__, __LINE__));
+ }
+
+ struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask,
+@@ -760,7 +760,7 @@
}
+
+ EXPORT_SYMBOL(vlan_hwaccel_receive_skb_debug);
+-EXPORT_SYMBOL(netif_rx_debug);
++EXPORT_SYMBOL(netif_receive_skb_debug);
+ EXPORT_SYMBOL(alloc_skb_debug);
+ EXPORT_SYMBOL(dev_alloc_skb_debug);
+ EXPORT_SYMBOL(skb_clone_debug);
+Index: madwifi-trunk-r3314/net80211/ieee80211_skb.h
+===================================================================
+--- madwifi-trunk-r3314.orig/net80211/ieee80211_skb.h 2008-03-06 23:49:11.047260294 +0100
++++ madwifi-trunk-r3314/net80211/ieee80211_skb.h 2008-03-06 23:49:24.384020316 +0100
+@@ -116,7 +116,7 @@
+ int vlan_hwaccel_receive_skb_debug(struct sk_buff *skb,
+ struct vlan_group *grp, unsigned short vlan_tag,
+ const char* func, int line);
+-int netif_rx_debug(struct sk_buff *skb, const char* func, int line);
++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line);
+ struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask,
+ const char *func, int line);
+ struct sk_buff * dev_alloc_skb_debug(unsigned int length,
+@@ -151,7 +151,7 @@
+ #undef dev_queue_xmit
+ #undef kfree_skb
+ #undef kfree_skb_fast
+-#undef netif_rx
++#undef netif_receive_skb
+ #undef pskb_copy
+ #undef skb_clone
+ #undef skb_copy
+@@ -168,8 +168,8 @@
+ skb_copy_expand_debug(_skb, _newheadroom, _newtailroom, _gfp_mask, __func__, __LINE__)
+ #define vlan_hwaccel_receive_skb(_skb, _grp, _tag) \
+ vlan_hwaccel_receive_skb_debug(_skb, _grp, _tag, __func__, __LINE__)
+-#define netif_rx(_skb) \
+- netif_rx_debug(_skb, __func__, __LINE__)
++#define netif_receive_skb(_skb) \
++ netif_receive_skb_debug(_skb, __func__, __LINE__)
+ #define alloc_skb(_length, _gfp_mask) \
+ alloc_skb_debug(_length, _gfp_mask, __func__, __LINE__)
+ #define dev_alloc_skb(_length) \