X-Git-Url: https://git.rohieb.name/openwrt.git/blobdiff_plain/c78b7963259d8ae2d3af609f1df275235666dd56..8d22e2e2411a2de86d6d95dd989d4876029330ab:/package/madwifi/patches/300-napi_polling.patch diff --git a/package/madwifi/patches/300-napi_polling.patch b/package/madwifi/patches/300-napi_polling.patch index f8135710e..bde768494 100644 --- a/package/madwifi/patches/300-napi_polling.patch +++ b/package/madwifi/patches/300-napi_polling.patch @@ -1,55 +1,81 @@ -diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c ---- madwifi.old/ath/if_ath.c 2007-06-01 11:26:04.181689864 +0200 -+++ madwifi.dev/ath/if_ath.c 2007-06-01 11:39:53.078678368 +0200 -@@ -167,7 +167,7 @@ - int, u_int32_t); +--- a/ath/if_ath.c ++++ b/ath/if_ath.c +@@ -184,7 +184,11 @@ static void ath_recv_mgmt(struct ieee802 + struct sk_buff *, int, int, u_int64_t); static void ath_setdefantenna(struct ath_softc *, u_int); static struct ath_txq *ath_txq_setup(struct ath_softc *, int, int); -static void ath_rx_tasklet(TQUEUE_ARG); ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++static int ath_rx_poll(struct napi_struct *napi, int budget); ++#else +static int ath_rx_poll(struct net_device *dev, int *budget); ++#endif static int ath_hardstart(struct sk_buff *, struct net_device *); static int ath_mgtstart(struct ieee80211com *, struct sk_buff *); #ifdef ATH_SUPERG_COMP -@@ -443,7 +443,6 @@ - ATH_TXBUF_LOCK_INIT(sc); - ATH_RXBUF_LOCK_INIT(sc); +@@ -376,6 +380,9 @@ static u_int32_t ath_set_clamped_maxtxpo + u_int32_t new_clamped_maxtxpower); + static u_int32_t ath_get_real_maxtxpower(struct ath_softc *sc); + ++static void ath_poll_disable(struct net_device *dev); ++static void ath_poll_enable(struct net_device *dev); ++ + /* calibrate every 30 secs in steady state but check every second at first. */ + static int ath_calinterval = ATH_SHORT_CALINTERVAL; + static int ath_countrycode = CTRY_DEFAULT; /* country code */ +@@ -547,7 +554,6 @@ ath_attach(u_int16_t devid, struct net_d + + atomic_set(&sc->sc_txbuf_counter, 0); - ATH_INIT_TQUEUE(&sc->sc_rxtq, ath_rx_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_txtq, ath_tx_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_bmisstq, ath_bmiss_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_bstucktq, ath_bstuck_tasklet, dev); -@@ -700,6 +699,8 @@ +@@ -821,6 +827,12 @@ ath_attach(u_int16_t devid, struct net_d dev->set_mac_address = ath_set_mac_address; - dev->change_mtu = ath_change_mtu; - dev->tx_queue_len = ATH_TXBUF - 1; /* 1 for mgmt frame */ + dev->change_mtu = ath_change_mtu; + dev->tx_queue_len = ATH_TXBUF - ATH_TXBUF_MGT_RESERVED; ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ netif_napi_add(dev, &sc->sc_napi, ath_rx_poll, 64); ++#else + dev->poll = ath_rx_poll; + dev->weight = 64; ++#endif #ifdef USE_HEADERLEN_RESV dev->hard_header_len += sizeof(struct ieee80211_qosframe) + sizeof(struct llc) + -@@ -1665,6 +1666,7 @@ - */ - ath_hal_getisr(ah, &status); /* NB: clears ISR too */ - DPRINTF(sc, ATH_DEBUG_INTR, "%s: status 0x%x\n", __func__, status); +@@ -2220,6 +2232,7 @@ ath_intr(int irq, void *dev_id, struct p + (status & HAL_INT_GLOBAL) ? " HAL_INT_GLOBAL" : "" + ); + + sc->sc_isr = status; status &= sc->sc_imask; /* discard unasked for bits */ - if (status & HAL_INT_FATAL) { - sc->sc_stats.ast_hardware++; -@@ -1700,7 +1702,12 @@ - if (status & HAL_INT_RX) { - sc->sc_tsf = ath_hal_gettsf64(ah); - ath_uapsd_processtriggers(sc); + /* As soon as we know we have a real interrupt we intend to service, + * we will check to see if we need an initial hardware TSF reading. +@@ -2277,7 +2290,21 @@ ath_intr(int irq, void *dev_id, struct p + } + if (status & (HAL_INT_RX | HAL_INT_RXPHY)) { + ath_uapsd_processtriggers(sc, hw_tsf); - ATH_SCHEDULE_TQUEUE(&sc->sc_rxtq, &needmark); + sc->sc_isr &= ~HAL_INT_RX; -+ if (netif_rx_schedule_prep(dev)) { ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ if (netif_rx_schedule_prep(dev, &sc->sc_napi)) ++#else ++ if (netif_rx_schedule_prep(dev)) ++#endif ++ { + sc->sc_imask &= ~HAL_INT_RX; + ath_hal_intrset(ah, sc->sc_imask); ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ __netif_rx_schedule(dev, &sc->sc_napi); ++#else + __netif_rx_schedule(dev); ++#endif + } } if (status & HAL_INT_TX) { #ifdef ATH_SUPERG_DYNTURBO -@@ -1726,6 +1733,11 @@ +@@ -2303,6 +2330,11 @@ ath_intr(int irq, void *dev_id, struct p } } #endif @@ -61,75 +87,160 @@ diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, &needmark); } if (status & HAL_INT_BMISS) { -@@ -3296,10 +3308,10 @@ - * - * XXX Using in_softirq is not right since we might - * be called from other soft irq contexts than -- * ath_rx_tasklet. +@@ -2515,6 +2547,7 @@ ath_init(struct net_device *dev) + if (sc->sc_tx99 != NULL) + sc->sc_tx99->start(sc->sc_tx99); + #endif ++ ath_poll_enable(dev); + + done: + ATH_UNLOCK(sc); +@@ -2555,6 +2588,9 @@ ath_stop_locked(struct net_device *dev) + if (sc->sc_tx99 != NULL) + sc->sc_tx99->stop(sc->sc_tx99); + #endif ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ ath_poll_disable(dev); ++#endif + netif_stop_queue(dev); /* XXX re-enabled by ath_newstate */ + dev->flags &= ~IFF_RUNNING; /* NB: avoid recursion */ + ieee80211_stop_running(ic); /* stop all VAPs */ +@@ -4013,12 +4049,47 @@ ath_key_set(struct ieee80211vap *vap, co + return ath_keyset(sc, k, mac, vap->iv_bss); + } + ++static void ath_poll_disable(struct net_device *dev) ++{ ++ struct ath_softc *sc = dev->priv; ++ ++ /* ++ * XXX Using in_softirq is not right since we might ++ * be called from other soft irq contexts than + * ath_rx_poll ++ */ ++ if (!in_softirq()) { ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ napi_disable(&sc->sc_napi); ++#else ++ netif_poll_disable(dev); ++#endif ++ } ++} ++ ++static void ath_poll_enable(struct net_device *dev) ++{ ++ struct ath_softc *sc = dev->priv; ++ ++ /* NB: see above */ ++ if (!in_softirq()) { ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ napi_enable(&sc->sc_napi); ++#else ++ netif_poll_enable(dev); ++#endif ++ } ++} ++ ++ + /* + * Block/unblock tx+rx processing while a key change is done. + * We assume the caller serializes key management operations + * so we only need to worry about synchronization with other + * uses that originate in the driver. + */ ++#define IS_UP(_dev) \ ++ (((_dev)->flags & (IFF_RUNNING|IFF_UP)) == (IFF_RUNNING|IFF_UP)) + static void + ath_key_update_begin(struct ieee80211vap *vap) + { +@@ -4032,14 +4103,9 @@ ath_key_update_begin(struct ieee80211vap + * When called from the rx tasklet we cannot use + * tasklet_disable because it will block waiting + * for us to complete execution. +- * +- * XXX Using in_softirq is not right since we might +- * be called from other soft irq contexts than +- * ath_rx_tasklet. */ - if (!in_softirq()) +- if (!in_softirq()) - tasklet_disable(&sc->sc_rxtq); -+ netif_poll_disable(dev); - netif_stop_queue(dev); +- netif_stop_queue(dev); ++ if (IS_UP(vap->iv_dev)) ++ netif_stop_queue(dev); } -@@ -3312,7 +3324,7 @@ - DPRINTF(sc, ATH_DEBUG_KEYCACHE, "%s:\n", __func__); - netif_start_queue(dev); - if (!in_softirq()) /* NB: see above */ + static void +@@ -4051,9 +4117,9 @@ ath_key_update_end(struct ieee80211vap * + #endif + + DPRINTF(sc, ATH_DEBUG_KEYCACHE, "End\n"); +- netif_wake_queue(dev); +- if (!in_softirq()) /* NB: see above */ - tasklet_enable(&sc->sc_rxtq); -+ netif_poll_enable(dev); ++ ++ if (IS_UP(vap->iv_dev)) ++ netif_wake_queue(dev); } /* -@@ -5573,13 +5585,12 @@ +@@ -6360,15 +6426,25 @@ ath_setdefantenna(struct ath_softc *sc, sc->sc_rxotherant = 0; } -static void -ath_rx_tasklet(TQUEUE_ARG data) +static int ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ath_rx_poll(struct napi_struct *napi, int budget) ++#else +ath_rx_poll(struct net_device *dev, int *budget) ++#endif { #define PA2DESC(_sc, _pa) \ ((struct ath_desc *)((caddr_t)(_sc)->sc_rxdma.dd_desc + \ ((_pa) - (_sc)->sc_rxdma.dd_desc_paddr))) - struct net_device *dev = (struct net_device *)data; - struct ath_buf *bf; +- struct ath_buf *bf; ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ struct ath_softc *sc = container_of(napi, struct ath_softc, sc_napi); ++ struct net_device *dev = sc->sc_dev; ++ u_int rx_limit = budget; ++#else struct ath_softc *sc = dev->priv; ++ u_int rx_limit = min(dev->quota, *budget); ++#endif ++ struct ath_buf *bf; struct ieee80211com *ic = &sc->sc_ic; -@@ -5591,12 +5602,15 @@ + struct ath_hal *ah = sc ? sc->sc_ah : NULL; + struct ath_desc *ds; +@@ -6378,8 +6454,10 @@ ath_rx_tasklet(TQUEUE_ARG data) unsigned int len; int type; u_int phyerr; + u_int processed = 0, early_stop = 0; -+ u_int rx_limit = dev->quota; - /* Let the 802.11 layer know about the new noise floor */ - sc->sc_channoise = ath_hal_get_channel_noise(ah, &(sc->sc_curchan)); - ic->ic_channoise = sc->sc_channoise; - - DPRINTF(sc, ATH_DEBUG_RX_PROC, "%s\n", __func__); + DPRINTF(sc, ATH_DEBUG_RX_PROC, "invoked\n"); +process_rx_again: do { bf = STAILQ_FIRST(&sc->sc_rxbuf); if (bf == NULL) { /* XXX ??? can this happen */ -@@ -5620,6 +5634,13 @@ +@@ -6403,6 +6481,15 @@ ath_rx_tasklet(TQUEUE_ARG data) /* NB: never process the self-linked entry at the end */ break; } + -+ processed++; -+ if (rx_limit-- < 0) { ++ if (rx_limit-- < 2) { + early_stop = 1; + break; + } ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ processed++; ++#endif + skb = bf->bf_skb; - if (skb == NULL) { /* XXX ??? can this happen */ - printk("%s: no skbuff (%s)\n", DEV_NAME(dev), __func__); -@@ -5658,6 +5679,7 @@ + if (skb == NULL) { + EPRINTF(sc, "Dropping; buffer contains NULL skbuff.\n"); +@@ -6450,6 +6537,7 @@ ath_rx_tasklet(TQUEUE_ARG data) sc->sc_stats.ast_rx_phyerr++; phyerr = rs->rs_phyerr & 0x1f; sc->sc_stats.ast_rx_phy[phyerr]++; @@ -137,39 +248,51 @@ diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c } if (rs->rs_status & HAL_RXERR_DECRYPT) { /* -@@ -5868,9 +5890,29 @@ +@@ -6645,9 +6733,39 @@ rx_next: STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list); ATH_RXBUF_UNLOCK_IRQ(sc); } while (ath_rxbuf_init(sc, bf) == 0); + if (!early_stop) { ++ unsigned long flags; + /* Check if more data is received while we were + * processing the descriptor chain. + */ -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_RX) { ++ u_int64_t hw_tsf = ath_hal_gettsf64(ah); + sc->sc_isr &= ~HAL_INT_RX; -+ ATH_ENABLE_INTR(); -+ ath_uapsd_processtriggers(sc); ++ local_irq_restore(flags); ++ ath_uapsd_processtriggers(sc, hw_tsf); + goto process_rx_again; + } -+ netif_rx_complete(dev); -+ -+ sc->sc_imask |= HAL_INT_RX; -+ ath_hal_intrset(ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + } + ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ netif_rx_complete(dev, napi); ++#else ++ netif_rx_complete(dev); + *budget -= processed; ++ dev->quota -= processed; ++#endif ++ sc->sc_imask |= HAL_INT_RX; ++ ath_hal_intrset(ah, sc->sc_imask); /* rx signal state monitoring */ ath_hal_rxmonitor(ah, &sc->sc_halstats, &sc->sc_curchan); ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ return processed; ++#else + return early_stop; ++#endif #undef PA2DESC } -@@ -7487,11 +7529,22 @@ +@@ -8298,12 +8416,24 @@ ath_tx_tasklet_q0(TQUEUE_ARG data) + { struct net_device *dev = (struct net_device *)data; struct ath_softc *sc = dev->priv; ++ unsigned long flags; +process_tx_again: if (txqactive(sc->sc_ah, 0)) @@ -177,95 +300,89 @@ diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c if (txqactive(sc->sc_ah, sc->sc_cabq->axq_qnum)) ath_tx_processq(sc, sc->sc_cabq); -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) -@@ -7508,6 +7561,7 @@ +@@ -8319,7 +8449,9 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data) + { struct net_device *dev = (struct net_device *)data; struct ath_softc *sc = dev->priv; ++ unsigned long flags; +process_tx_again: /* * Process each active queue. */ -@@ -7528,6 +7582,16 @@ +@@ -8340,6 +8472,16 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data) if (sc->sc_uapsdq && txqactive(sc->sc_ah, sc->sc_uapsdq->axq_qnum)) ath_tx_processq(sc, sc->sc_uapsdq); -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) -@@ -7545,6 +7609,7 @@ +@@ -8355,13 +8497,25 @@ ath_tx_tasklet(TQUEUE_ARG data) + struct net_device *dev = (struct net_device *)data; + struct ath_softc *sc = dev->priv; unsigned int i; ++ unsigned long flags; - /* Process each active queue. */ + /* Process each active queue. This includes sc_cabq, sc_xrtq and + * sc_uapsdq */ +process_tx_again: for (i = 0; i < HAL_NUM_TX_QUEUES; i++) if (ATH_TXQ_SETUP(sc, i) && txqactive(sc->sc_ah, i)) ath_tx_processq(sc, &sc->sc_txq[i]); -@@ -7553,6 +7618,16 @@ - ath_tx_processq(sc, sc->sc_xrtxq); - #endif -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) -@@ -7651,6 +7726,7 @@ - ath_draintxq(struct ath_softc *sc) - { - struct ath_hal *ah = sc->sc_ah; -+ int npend = 0; - unsigned int i; - - /* XXX return value */ -@@ -9170,9 +9246,9 @@ +@@ -10296,9 +10450,9 @@ ath_change_mtu(struct net_device *dev, i dev->mtu = mtu; if ((dev->flags & IFF_RUNNING) && !sc->sc_invalid) { /* NB: the rx buffers may need to be reallocated */ - tasklet_disable(&sc->sc_rxtq); -+ netif_poll_disable(dev); ++ ath_poll_disable(dev); error = ath_reset(dev); - tasklet_enable(&sc->sc_rxtq); -+ netif_poll_enable(dev); ++ ath_poll_enable(dev); } ATH_UNLOCK(sc); -diff -ur madwifi.old/ath/if_athvar.h madwifi.dev/ath/if_athvar.h ---- madwifi.old/ath/if_athvar.h 2007-06-01 11:26:04.158693360 +0200 -+++ madwifi.dev/ath/if_athvar.h 2007-06-01 11:33:26.549439744 +0200 -@@ -48,6 +48,10 @@ - #include "if_athioctl.h" - #include "net80211/ieee80211.h" /* XXX for WME_NUM_AC */ +--- a/ath/if_athvar.h ++++ b/ath/if_athvar.h +@@ -53,6 +53,10 @@ + # include + #endif +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) +#define irqs_disabled() 0 @@ -274,15 +391,25 @@ diff -ur madwifi.old/ath/if_athvar.h madwifi.dev/ath/if_athvar.h /* * Deduce if tasklets are available. If not then * fall back to using the immediate work queue. -@@ -621,7 +625,6 @@ +@@ -616,6 +620,9 @@ struct ath_rp { + struct ath_softc { + struct ieee80211com sc_ic; /* NB: must be first */ + struct net_device *sc_dev; ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ struct napi_struct sc_napi; ++#endif + void __iomem *sc_iobase; /* address of the device */ + struct semaphore sc_lock; /* dev-level lock */ + struct net_device_stats sc_devstats; /* device statistics */ +@@ -730,7 +737,6 @@ struct ath_softc { struct ath_buf *sc_rxbufcur; /* current rx buffer */ u_int32_t *sc_rxlink; /* link ptr in last RX desc */ spinlock_t sc_rxbuflock; - struct ATH_TQ_STRUCT sc_rxtq; /* rx intr tasklet */ struct ATH_TQ_STRUCT sc_rxorntq; /* rxorn intr tasklet */ u_int8_t sc_defant; /* current default antenna */ - u_int8_t sc_rxotherant; /* rx's on non-default antenna*/ -@@ -634,6 +637,7 @@ + u_int8_t sc_rxotherant; /* RXs on non-default antenna */ +@@ -745,6 +751,7 @@ struct ath_softc { u_int sc_txintrperiod; /* tx interrupt batching */ struct ath_txq sc_txq[HAL_NUM_TX_QUEUES]; struct ath_txq *sc_ac2q[WME_NUM_AC]; /* WME AC -> h/w qnum */ @@ -290,27 +417,120 @@ diff -ur madwifi.old/ath/if_athvar.h madwifi.dev/ath/if_athvar.h struct ATH_TQ_STRUCT sc_txtq; /* tx intr tasklet */ u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN]; struct ath_descdma sc_bdma; /* beacon descriptors */ -@@ -714,6 +718,8 @@ - #define ATH_TXBUF_LOCK_ASSERT(_sc) \ - KASSERT(spin_is_locked(&(_sc)->sc_txbuflock), ("txbuf not locked!")) +@@ -858,6 +865,8 @@ typedef void (*ath_callback) (struct ath + #define ATH_TXBUF_LOCK_CHECK(_sc) + #endif +#define ATH_DISABLE_INTR local_irq_disable +#define ATH_ENABLE_INTR local_irq_enable #define ATH_RXBUF_LOCK_INIT(_sc) spin_lock_init(&(_sc)->sc_rxbuflock) #define ATH_RXBUF_LOCK_DESTROY(_sc) -diff -ur madwifi.old/net80211/ieee80211_input.c madwifi.dev/net80211/ieee80211_input.c ---- madwifi.old/net80211/ieee80211_input.c 2007-06-01 11:26:04.183689560 +0200 -+++ madwifi.dev/net80211/ieee80211_input.c 2007-06-01 11:31:46.931583960 +0200 -@@ -1144,8 +1144,9 @@ - if (ni->ni_vlan != 0 && vap->iv_vlgrp != NULL) { +--- a/net80211/ieee80211_input.c ++++ b/net80211/ieee80211_input.c +@@ -1198,7 +1198,7 @@ ieee80211_deliver_data(struct ieee80211_ /* attach vlan tag */ - vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan); -- } else -- netif_rx(skb); -+ } else { -+ netif_receive_skb(skb); -+ } - dev->last_rx = jiffies; - } + struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; + if (vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ /* If netif_receive_skb dropped the packet because + * device was too busy */ + if (ni_tmp != NULL) { + /* node reference was leaked */ +@@ -1209,8 +1209,8 @@ ieee80211_deliver_data(struct ieee80211_ + skb = NULL; /* SKB is no longer ours */ + } else { + struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; +- if (netif_rx(skb) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ if (netif_receive_skb(skb) == NET_RX_DROP) { ++ /* If netif_receive_skb dropped the packet because + * device was too busy */ + if (ni_tmp != NULL) { + /* node reference was leaked */ +@@ -2322,8 +2322,8 @@ forward_mgmt_to_app(struct ieee80211vap + skb1->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */ + + ni_tmp = SKB_CB(skb1)->ni; +- if (netif_rx(skb1) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ if (netif_receive_skb(skb1) == NET_RX_DROP) { ++ /* If netif_receive_skb dropped the packet because + * device was too busy */ + if (ni_tmp != NULL) { + /* node reference was leaked */ +--- a/net80211/ieee80211_monitor.c ++++ b/net80211/ieee80211_monitor.c +@@ -584,8 +584,8 @@ ieee80211_input_monitor(struct ieee80211 + skb1->protocol = + __constant_htons(0x0019); /* ETH_P_80211_RAW */ + +- if (netif_rx(skb1) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ if (netif_receive_skb(skb1) == NET_RX_DROP) { ++ /* If netif_receive_skb dropped the packet because + * device was too busy, reclaim the ref. in + * the skb. */ + if (SKB_CB(skb1)->ni != NULL) +--- a/net80211/ieee80211_skb.c ++++ b/net80211/ieee80211_skb.c +@@ -73,7 +73,7 @@ + #undef dev_queue_xmit + #undef kfree_skb + #undef kfree_skb_fast +-#undef netif_rx ++#undef netif_receive_skb + #undef pskb_copy + #undef skb_clone + #undef skb_copy +@@ -638,8 +638,8 @@ int vlan_hwaccel_receive_skb_debug(stru + grp, vlan_tag); + } + +-int netif_rx_debug(struct sk_buff *skb, const char* func, int line) { +- return netif_rx(untrack_skb(skb, 0, func, line, __func__, __LINE__)); ++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line) { ++ return netif_receive_skb(untrack_skb(skb, 0, func, line, __func__, __LINE__)); + } + + struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, +@@ -760,7 +760,7 @@ struct sk_buff * skb_copy_expand_debug(c } + + EXPORT_SYMBOL(vlan_hwaccel_receive_skb_debug); +-EXPORT_SYMBOL(netif_rx_debug); ++EXPORT_SYMBOL(netif_receive_skb_debug); + EXPORT_SYMBOL(alloc_skb_debug); + EXPORT_SYMBOL(dev_alloc_skb_debug); + EXPORT_SYMBOL(skb_clone_debug); +--- a/net80211/ieee80211_skb.h ++++ b/net80211/ieee80211_skb.h +@@ -116,7 +116,7 @@ int ieee80211_skb_references(void); + int vlan_hwaccel_receive_skb_debug(struct sk_buff *skb, + struct vlan_group *grp, unsigned short vlan_tag, + const char* func, int line); +-int netif_rx_debug(struct sk_buff *skb, const char* func, int line); ++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line); + struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, + const char *func, int line); + struct sk_buff * dev_alloc_skb_debug(unsigned int length, +@@ -151,7 +151,7 @@ struct sk_buff * skb_copy_expand_debug(c + #undef dev_queue_xmit + #undef kfree_skb + #undef kfree_skb_fast +-#undef netif_rx ++#undef netif_receive_skb + #undef pskb_copy + #undef skb_clone + #undef skb_copy +@@ -168,8 +168,8 @@ struct sk_buff * skb_copy_expand_debug(c + skb_copy_expand_debug(_skb, _newheadroom, _newtailroom, _gfp_mask, __func__, __LINE__) + #define vlan_hwaccel_receive_skb(_skb, _grp, _tag) \ + vlan_hwaccel_receive_skb_debug(_skb, _grp, _tag, __func__, __LINE__) +-#define netif_rx(_skb) \ +- netif_rx_debug(_skb, __func__, __LINE__) ++#define netif_receive_skb(_skb) \ ++ netif_receive_skb_debug(_skb, __func__, __LINE__) + #define alloc_skb(_length, _gfp_mask) \ + alloc_skb_debug(_length, _gfp_mask, __func__, __LINE__) + #define dev_alloc_skb(_length) \