X-Git-Url: http://git.rohieb.name/openwrt.git/blobdiff_plain/c8ff8610bfef399cbbd968acd395aca5acb6902f..835a68409cb230cccd703fe8e5ffcba3a1096c8e:/package/madwifi/patches/300-napi_polling.patch?ds=inline diff --git a/package/madwifi/patches/300-napi_polling.patch b/package/madwifi/patches/300-napi_polling.patch index 69e2b106b..bc62d7b4a 100644 --- a/package/madwifi/patches/300-napi_polling.patch +++ b/package/madwifi/patches/300-napi_polling.patch @@ -1,57 +1,83 @@ -diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c ---- madwifi.old/ath/if_ath.c 2007-05-31 06:48:28.561525488 +0200 -+++ madwifi.dev/ath/if_ath.c 2007-05-31 06:50:07.143538744 +0200 -@@ -167,7 +167,7 @@ - int, u_int32_t); +--- a/ath/if_ath.c ++++ b/ath/if_ath.c +@@ -184,7 +184,11 @@ static void ath_recv_mgmt(struct ieee802 + struct sk_buff *, int, int, u_int64_t); static void ath_setdefantenna(struct ath_softc *, u_int); static struct ath_txq *ath_txq_setup(struct ath_softc *, int, int); -static void ath_rx_tasklet(TQUEUE_ARG); ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++static int ath_rx_poll(struct napi_struct *napi, int budget); ++#else +static int ath_rx_poll(struct net_device *dev, int *budget); ++#endif static int ath_hardstart(struct sk_buff *, struct net_device *); static int ath_mgtstart(struct ieee80211com *, struct sk_buff *); #ifdef ATH_SUPERG_COMP -@@ -429,7 +429,6 @@ - ATH_TXBUF_LOCK_INIT(sc); - ATH_RXBUF_LOCK_INIT(sc); +@@ -376,6 +380,9 @@ static u_int32_t ath_set_clamped_maxtxpo + u_int32_t new_clamped_maxtxpower); + static u_int32_t ath_get_real_maxtxpower(struct ath_softc *sc); + ++static void ath_poll_disable(struct net_device *dev); ++static void ath_poll_enable(struct net_device *dev); ++ + /* calibrate every 30 secs in steady state but check every second at first. */ + static int ath_calinterval = ATH_SHORT_CALINTERVAL; + static int ath_countrycode = CTRY_DEFAULT; /* country code */ +@@ -547,7 +554,6 @@ ath_attach(u_int16_t devid, struct net_d + + atomic_set(&sc->sc_txbuf_counter, 0); - ATH_INIT_TQUEUE(&sc->sc_rxtq, ath_rx_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_txtq, ath_tx_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_bmisstq, ath_bmiss_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_bstucktq, ath_bstuck_tasklet, dev); -@@ -685,6 +684,8 @@ +@@ -821,6 +827,12 @@ ath_attach(u_int16_t devid, struct net_d dev->set_mac_address = ath_set_mac_address; - dev->change_mtu = ath_change_mtu; - dev->tx_queue_len = ATH_TXBUF - 1; /* 1 for mgmt frame */ + dev->change_mtu = ath_change_mtu; + dev->tx_queue_len = ATH_TXBUF - ATH_TXBUF_MGT_RESERVED; ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ netif_napi_add(dev, &sc->sc_napi, ath_rx_poll, 64); ++#else + dev->poll = ath_rx_poll; + dev->weight = 64; ++#endif #ifdef USE_HEADERLEN_RESV dev->hard_header_len += sizeof(struct ieee80211_qosframe) + sizeof(struct llc) + -@@ -1643,6 +1644,7 @@ - */ - ath_hal_getisr(ah, &status); /* NB: clears ISR too */ - DPRINTF(sc, ATH_DEBUG_INTR, "%s: status 0x%x\n", __func__, status); +@@ -2220,6 +2232,7 @@ ath_intr(int irq, void *dev_id, struct p + (status & HAL_INT_GLOBAL) ? " HAL_INT_GLOBAL" : "" + ); + + sc->sc_isr = status; status &= sc->sc_imask; /* discard unasked for bits */ - if (status & HAL_INT_FATAL) { - sc->sc_stats.ast_hardware++; -@@ -1678,7 +1680,12 @@ - if (status & HAL_INT_RX) { - sc->sc_tsf = ath_hal_gettsf64(ah); - ath_uapsd_processtriggers(sc); + /* As soon as we know we have a real interrupt we intend to service, + * we will check to see if we need an initial hardware TSF reading. +@@ -2277,7 +2290,21 @@ ath_intr(int irq, void *dev_id, struct p + } + if (status & (HAL_INT_RX | HAL_INT_RXPHY)) { + ath_uapsd_processtriggers(sc, hw_tsf); - ATH_SCHEDULE_TQUEUE(&sc->sc_rxtq, &needmark); + sc->sc_isr &= ~HAL_INT_RX; -+ if (netif_rx_schedule_prep(dev)) { ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ if (netif_rx_schedule_prep(dev, &sc->sc_napi)) ++#else ++ if (netif_rx_schedule_prep(dev)) ++#endif ++ { + sc->sc_imask &= ~HAL_INT_RX; + ath_hal_intrset(ah, sc->sc_imask); ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ __netif_rx_schedule(dev, &sc->sc_napi); ++#else + __netif_rx_schedule(dev); ++#endif + } } if (status & HAL_INT_TX) { #ifdef ATH_SUPERG_DYNTURBO -@@ -1704,6 +1711,11 @@ +@@ -2303,6 +2330,11 @@ ath_intr(int irq, void *dev_id, struct p } - } + } #endif + /* disable transmit interrupt */ + sc->sc_isr &= ~HAL_INT_TX; @@ -61,358 +87,160 @@ diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, &needmark); } if (status & HAL_INT_BMISS) { -@@ -2166,12 +2178,13 @@ - * Insert the frame on the outbound list and - * pass it on to the hardware. - */ -- ATH_TXQ_LOCK(txq); -+ ATH_TXQ_LOCK_IRQ(txq); - if (ni && ni->ni_vap && txq == &ATH_VAP(ni->ni_vap)->av_mcastq) { - /* - * The CAB queue is started from the SWBA handler since - * frames only go out on DTIM and to avoid possible races. - */ -+ sc->sc_imask &= ~HAL_INT_SWBA; - ath_hal_intrset(ah, sc->sc_imask & ~HAL_INT_SWBA); - ATH_TXQ_INSERT_TAIL(txq, bf, bf_list); - DPRINTF(sc, ATH_DEBUG_TX_PROC, "%s: txq depth = %d\n", __func__, txq->axq_depth); -@@ -2187,6 +2200,7 @@ - ito64(bf->bf_daddr), bf->bf_desc); - } - txq->axq_link = &lastds->ds_link; -+ sc->sc_imask |= HAL_INT_SWBA; - ath_hal_intrset(ah, sc->sc_imask); - } else { - ATH_TXQ_INSERT_TAIL(txq, bf, bf_list); -@@ -2222,7 +2236,7 @@ - } - } - } -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_IRQ(txq); +@@ -2515,6 +2547,7 @@ ath_init(struct net_device *dev) + if (sc->sc_tx99 != NULL) + sc->sc_tx99->start(sc->sc_tx99); + #endif ++ ath_poll_enable(dev); - sc->sc_devstats.tx_packets++; - sc->sc_devstats.tx_bytes += framelen; -@@ -2373,12 +2387,18 @@ - unsigned int pktlen; - int framecnt; + done: + ATH_UNLOCK(sc); +@@ -2555,6 +2588,9 @@ ath_stop_locked(struct net_device *dev) + if (sc->sc_tx99 != NULL) + sc->sc_tx99->stop(sc->sc_tx99); + #endif ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ ath_poll_disable(dev); ++#endif + netif_stop_queue(dev); /* XXX re-enabled by ath_newstate */ + dev->flags &= ~IFF_RUNNING; /* NB: avoid recursion */ + ieee80211_stop_running(ic); /* stop all VAPs */ +@@ -4013,12 +4049,47 @@ ath_key_set(struct ieee80211vap *vap, co + return ath_keyset(sc, k, mac, vap->iv_bss); + } ++static void ath_poll_disable(struct net_device *dev) ++{ ++ struct ath_softc *sc = dev->priv; ++ + /* -+ * NB: using _BH style locking even though this function may be called -+ * at interrupt time (within tasklet or bh). This should be harmless -+ * and this function calls others (i.e., ath_tx_start()) which do -+ * the same. ++ * XXX Using in_softirq is not right since we might ++ * be called from other soft irq contexts than ++ * ath_rx_poll + */ - for (;;) { -- ATH_TXQ_LOCK(txq); -+ ATH_TXQ_LOCK_BH(txq); - - bf_ff = TAILQ_LAST(&txq->axq_stageq, axq_headtype); - if ((!bf_ff) || ath_ff_flushdonetest(txq, bf_ff)) { -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - break; - } - -@@ -2388,7 +2408,7 @@ - ATH_NODE(ni)->an_tx_ffbuf[bf_ff->bf_skb->priority] = NULL; - TAILQ_REMOVE(&txq->axq_stageq, bf_ff, bf_stagelist); - -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - - /* encap and xmit */ - bf_ff->bf_skb = ieee80211_encap(ni, bf_ff->bf_skb, &framecnt); -@@ -2409,15 +2429,16 @@ - } - bf_ff->bf_node = NULL; - -- ATH_TXBUF_LOCK_IRQ(sc); -+ ATH_TXBUF_LOCK_BH(sc); - STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf_ff, bf_list); -- ATH_TXBUF_UNLOCK_IRQ(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - } -+ ATH_TXQ_UNLOCK_BH(txq); - } - #endif - - #define ATH_HARDSTART_GET_TX_BUF_WITH_LOCK \ -- ATH_TXBUF_LOCK_IRQ(sc); \ -+ ATH_TXBUF_LOCK_BH(sc); \ - bf = STAILQ_FIRST(&sc->sc_txbuf); \ - if (bf != NULL) { \ - STAILQ_REMOVE_HEAD(&sc->sc_txbuf, bf_list); \ -@@ -2432,11 +2453,23 @@ - sc->sc_devstopped = 1; \ - ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, NULL); \ - } \ -- ATH_TXBUF_UNLOCK_IRQ(sc); \ ++ if (!in_softirq()) { ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ napi_disable(&sc->sc_napi); ++#else ++ netif_poll_disable(dev); ++#endif ++ } ++} ++ ++static void ath_poll_enable(struct net_device *dev) ++{ ++ struct ath_softc *sc = dev->priv; + -+#define ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF \ -+ ATH_TXBUF_UNLOCK_BH(sc); \ -+ if (bf == NULL) { /* NB: should not happen */ \ -+ DPRINTF(sc,ATH_DEBUG_XMIT,"%s: discard, no xmit buf\n", __func__); \ -+ sc->sc_stats.ast_tx_nobuf++; \ -+ goto hardstart_fail; \ ++ /* NB: see above */ ++ if (!in_softirq()) { ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ napi_enable(&sc->sc_napi); ++#else ++ netif_poll_enable(dev); ++#endif + } ++} ++ + -+#define ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_ON \ -+ ATH_TXBUF_UNLOCK_BH(sc); \ - if (bf == NULL) { /* NB: should not happen */ \ - DPRINTF(sc,ATH_DEBUG_XMIT, \ - "%s: discard, no xmit buf\n", __func__); \ -+ ATH_TXQ_UNLOCK_BH(txq); \ - sc->sc_stats.ast_tx_nobuf++; \ -+ goto hardstart_fail; \ - } - /* -@@ -2498,6 +2531,7 @@ - if (M_FLAG_GET(skb, M_UAPSD)) { - /* bypass FF handling */ - ATH_HARDSTART_GET_TX_BUF_WITH_LOCK; -+ ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF; - if (bf == NULL) - goto hardstart_fail; - goto ff_bypass; -@@ -2519,7 +2553,7 @@ - /* NB: use this lock to protect an->an_ff_txbuf in athff_can_aggregate() - * call too. - */ -- ATH_TXQ_LOCK(txq); -+ ATH_TXQ_LOCK_BH(txq); - if (athff_can_aggregate(sc, eh, an, skb, vap->iv_fragthreshold, &ff_flush)) { - - if (an->an_tx_ffbuf[skb->priority]) { /* i.e., frame on the staging queue */ -@@ -2529,7 +2563,7 @@ - TAILQ_REMOVE(&txq->axq_stageq, bf, bf_stagelist); - an->an_tx_ffbuf[skb->priority] = NULL; - -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - - /* - * chain skbs and add FF magic -@@ -2556,8 +2590,9 @@ - * to give the buffer back. - */ - ATH_HARDSTART_GET_TX_BUF_WITH_LOCK; -+ ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_ON; - if (bf == NULL) { -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - goto hardstart_fail; - } - DPRINTF(sc, ATH_DEBUG_XMIT | ATH_DEBUG_FF, -@@ -2570,7 +2605,7 @@ - - TAILQ_INSERT_HEAD(&txq->axq_stageq, bf, bf_stagelist); - -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - - return 0; - } -@@ -2581,7 +2616,7 @@ - TAILQ_REMOVE(&txq->axq_stageq, bf_ff, bf_stagelist); - an->an_tx_ffbuf[skb->priority] = NULL; - -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - - /* encap and xmit */ - bf_ff->bf_skb = ieee80211_encap(ni, bf_ff->bf_skb, &framecnt); -@@ -2611,9 +2646,9 @@ - } - bf_ff->bf_node = NULL; - -- ATH_TXBUF_LOCK(sc); -+ ATH_TXBUF_LOCK_BH(sc); - STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf_ff, bf_list); -- ATH_TXBUF_UNLOCK(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - goto ff_flushdone; - } - /* -@@ -2623,14 +2658,13 @@ - else if (an->an_tx_ffbuf[skb->priority]) { - DPRINTF(sc, ATH_DEBUG_XMIT | ATH_DEBUG_FF, - "%s: Out-Of-Order fast-frame\n", __func__); -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - } else -- ATH_TXQ_UNLOCK(txq); -+ ATH_TXQ_UNLOCK_BH(txq); - - ff_flushdone: - ATH_HARDSTART_GET_TX_BUF_WITH_LOCK; -- if (bf == NULL) -- goto hardstart_fail; -+ ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF; - } - - ff_bypass: -@@ -2638,6 +2672,7 @@ - #else /* ATH_SUPERG_FF */ - - ATH_HARDSTART_GET_TX_BUF_WITH_LOCK; -+ ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF; - - #endif /* ATH_SUPERG_FF */ - -@@ -2659,7 +2694,7 @@ - * Allocate 1 ath_buf for each frame given 1 was - * already alloc'd - */ -- ATH_TXBUF_LOCK(sc); -+ ATH_TXBUF_LOCK_BH(sc); - for (bfcnt = 1; bfcnt < framecnt; ++bfcnt) { - if ((tbf = STAILQ_FIRST(&sc->sc_txbuf)) != NULL) { - STAILQ_REMOVE_HEAD(&sc->sc_txbuf, bf_list); -@@ -2680,11 +2715,11 @@ - STAILQ_INSERT_TAIL(&sc->sc_txbuf, tbf, bf_list); - } - } -- ATH_TXBUF_UNLOCK(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - STAILQ_INIT(&bf_head); - goto hardstart_fail; - } -- ATH_TXBUF_UNLOCK(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - - while ((bf = STAILQ_FIRST(&bf_head)) != NULL && skb != NULL) { - unsigned int nextfraglen = 0; -@@ -2720,7 +2755,7 @@ - - hardstart_fail: - if (!STAILQ_EMPTY(&bf_head)) { -- ATH_TXBUF_LOCK(sc); -+ ATH_TXBUF_LOCK_BH(sc); - STAILQ_FOREACH_SAFE(tbf, &bf_head, bf_list, tempbf) { - tbf->bf_skb = NULL; - tbf->bf_node = NULL; -@@ -2730,7 +2765,7 @@ - - STAILQ_INSERT_TAIL(&sc->sc_txbuf, tbf, bf_list); - } -- ATH_TXBUF_UNLOCK(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - } - - /* free sk_buffs */ -@@ -2773,7 +2808,7 @@ - /* - * Grab a TX buffer and associated resources. - */ -- ATH_TXBUF_LOCK_IRQ(sc); -+ ATH_TXBUF_LOCK_BH(sc); - bf = STAILQ_FIRST(&sc->sc_txbuf); - if (bf != NULL) - STAILQ_REMOVE_HEAD(&sc->sc_txbuf, bf_list); -@@ -2784,7 +2819,7 @@ - sc->sc_devstopped=1; - ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, NULL); - } -- ATH_TXBUF_UNLOCK_IRQ(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - if (bf == NULL) { - printk("ath_mgtstart: discard, no xmit buf\n"); - sc->sc_stats.ast_tx_nobufmgt++; -@@ -2813,9 +2848,9 @@ - bf->bf_skb = NULL; - bf->bf_node = NULL; - -- ATH_TXBUF_LOCK_IRQ(sc); -+ ATH_TXBUF_LOCK_BH(sc); - STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf, bf_list); -- ATH_TXBUF_UNLOCK_IRQ(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - } - dev_kfree_skb_any(skb); - skb = NULL; -@@ -3283,10 +3318,10 @@ - * - * XXX Using in_softirq is not right since we might - * be called from other soft irq contexts than + * Block/unblock tx+rx processing while a key change is done. + * We assume the caller serializes key management operations + * so we only need to worry about synchronization with other + * uses that originate in the driver. + */ ++#define IS_UP(_dev) \ ++ (((_dev)->flags & (IFF_RUNNING|IFF_UP)) == (IFF_RUNNING|IFF_UP)) + static void + ath_key_update_begin(struct ieee80211vap *vap) + { +@@ -4032,14 +4103,9 @@ ath_key_update_begin(struct ieee80211vap + * When called from the rx tasklet we cannot use + * tasklet_disable because it will block waiting + * for us to complete execution. +- * +- * XXX Using in_softirq is not right since we might +- * be called from other soft irq contexts than - * ath_rx_tasklet. -+ * ath_rx_poll */ - if (!in_softirq()) +- if (!in_softirq()) - tasklet_disable(&sc->sc_rxtq); -+ netif_poll_disable(dev); - netif_stop_queue(dev); +- netif_stop_queue(dev); ++ if (IS_UP(vap->iv_dev)) ++ netif_stop_queue(dev); } -@@ -3299,7 +3334,7 @@ - DPRINTF(sc, ATH_DEBUG_KEYCACHE, "%s:\n", __func__); - netif_start_queue(dev); - if (!in_softirq()) /* NB: see above */ + static void +@@ -4051,9 +4117,9 @@ ath_key_update_end(struct ieee80211vap * + #endif + + DPRINTF(sc, ATH_DEBUG_KEYCACHE, "End\n"); +- netif_wake_queue(dev); +- if (!in_softirq()) /* NB: see above */ - tasklet_enable(&sc->sc_rxtq); -+ netif_poll_enable(dev); ++ ++ if (IS_UP(vap->iv_dev)) ++ netif_wake_queue(dev); } /* -@@ -4866,9 +4901,9 @@ - bf->bf_node = NULL; - bf->bf_desc->ds_link = 0; - -- ATH_TXBUF_LOCK_IRQ(sc); -+ ATH_TXBUF_LOCK_BH(sc); - STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf, bf_list); -- ATH_TXBUF_UNLOCK_IRQ(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - - an->an_uapsd_overflowqdepth--; - } -@@ -5544,13 +5579,12 @@ +@@ -6360,15 +6426,25 @@ ath_setdefantenna(struct ath_softc *sc, sc->sc_rxotherant = 0; } -static void -ath_rx_tasklet(TQUEUE_ARG data) +static int ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ath_rx_poll(struct napi_struct *napi, int budget) ++#else +ath_rx_poll(struct net_device *dev, int *budget) ++#endif { #define PA2DESC(_sc, _pa) \ ((struct ath_desc *)((caddr_t)(_sc)->sc_rxdma.dd_desc + \ ((_pa) - (_sc)->sc_rxdma.dd_desc_paddr))) - struct net_device *dev = (struct net_device *)data; - struct ath_buf *bf; +- struct ath_buf *bf; ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ struct ath_softc *sc = container_of(napi, struct ath_softc, sc_napi); ++ struct net_device *dev = sc->sc_dev; ++ u_int rx_limit = budget; ++#else struct ath_softc *sc = dev->priv; ++ u_int rx_limit = min(dev->quota, *budget); ++#endif ++ struct ath_buf *bf; struct ieee80211com *ic = &sc->sc_ic; -@@ -5562,12 +5596,15 @@ + struct ath_hal *ah = sc ? sc->sc_ah : NULL; + struct ath_desc *ds; +@@ -6378,8 +6454,10 @@ ath_rx_tasklet(TQUEUE_ARG data) unsigned int len; int type; u_int phyerr; + u_int processed = 0, early_stop = 0; -+ u_int rx_limit = dev->quota; - /* Let the 802.11 layer know about the new noise floor */ - sc->sc_channoise = ath_hal_get_channel_noise(ah, &(sc->sc_curchan)); - ic->ic_channoise = sc->sc_channoise; - - DPRINTF(sc, ATH_DEBUG_RX_PROC, "%s\n", __func__); + DPRINTF(sc, ATH_DEBUG_RX_PROC, "invoked\n"); +process_rx_again: do { bf = STAILQ_FIRST(&sc->sc_rxbuf); if (bf == NULL) { /* XXX ??? can this happen */ -@@ -5591,6 +5628,13 @@ +@@ -6403,6 +6481,15 @@ ath_rx_tasklet(TQUEUE_ARG data) /* NB: never process the self-linked entry at the end */ break; } + -+ processed++; -+ if (rx_limit-- < 0) { ++ if (rx_limit-- < 2) { + early_stop = 1; + break; + } ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ processed++; ++#endif + skb = bf->bf_skb; - if (skb == NULL) { /* XXX ??? can this happen */ - printk("%s: no skbuff (%s)\n", dev->name, __func__); -@@ -5629,6 +5673,7 @@ + if (skb == NULL) { + EPRINTF(sc, "Dropping; buffer contains NULL skbuff.\n"); +@@ -6450,6 +6537,7 @@ ath_rx_tasklet(TQUEUE_ARG data) sc->sc_stats.ast_rx_phyerr++; phyerr = rs->rs_phyerr & 0x1f; sc->sc_stats.ast_rx_phy[phyerr]++; @@ -420,88 +248,51 @@ diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c } if (rs->rs_status & HAL_RXERR_DECRYPT) { /* -@@ -5840,9 +5885,29 @@ +@@ -6645,9 +6733,39 @@ rx_next: STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list); ATH_RXBUF_UNLOCK_IRQ(sc); } while (ath_rxbuf_init(sc, bf) == 0); + if (!early_stop) { ++ unsigned long flags; + /* Check if more data is received while we were + * processing the descriptor chain. + */ -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_RX) { ++ u_int64_t hw_tsf = ath_hal_gettsf64(ah); + sc->sc_isr &= ~HAL_INT_RX; -+ ATH_ENABLE_INTR(); -+ ath_uapsd_processtriggers(sc); ++ local_irq_restore(flags); ++ ath_uapsd_processtriggers(sc, hw_tsf); + goto process_rx_again; + } -+ netif_rx_complete(dev); -+ -+ sc->sc_imask |= HAL_INT_RX; -+ ath_hal_intrset(ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + } + ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ netif_rx_complete(dev, napi); ++#else ++ netif_rx_complete(dev); + *budget -= processed; - ++ dev->quota -= processed; ++#endif ++ sc->sc_imask |= HAL_INT_RX; ++ ath_hal_intrset(ah, sc->sc_imask); + /* rx signal state monitoring */ ath_hal_rxmonitor(ah, &sc->sc_halstats, &sc->sc_curchan); ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ return processed; ++#else + return early_stop; ++#endif #undef PA2DESC } -@@ -6118,22 +6183,22 @@ - } - } - -- ATH_TXBUF_LOCK_IRQ(sc); -+ ATH_TXBUF_LOCK_BH(sc); - bf = STAILQ_FIRST(&sc->sc_grppollbuf); - if (bf != NULL) - STAILQ_REMOVE_HEAD(&sc->sc_grppollbuf, bf_list); - else { - DPRINTF(sc, ATH_DEBUG_XMIT, "%s: No more TxBufs\n", __func__); -- ATH_TXBUF_UNLOCK_IRQ_EARLY(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - return; - } - /* XXX use a counter and leave at least one for mgmt frames */ - if (STAILQ_EMPTY(&sc->sc_grppollbuf)) { - DPRINTF(sc, ATH_DEBUG_XMIT, "%s: No more TxBufs left\n", __func__); -- ATH_TXBUF_UNLOCK_IRQ_EARLY(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - return; - } -- ATH_TXBUF_UNLOCK_IRQ(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - - bf->bf_skbaddr = bus_map_single(sc->sc_bdev, - skb->data, skb->len, BUS_DMA_TODEVICE); -@@ -6599,9 +6664,9 @@ - dev_kfree_skb(lastbuf->bf_skb); - lastbuf->bf_skb = NULL; - ieee80211_unref_node(&lastbuf->bf_node); -- ATH_TXBUF_LOCK_IRQ(sc); -+ ATH_TXBUF_LOCK_BH(sc); - STAILQ_INSERT_TAIL(&sc->sc_txbuf, lastbuf, bf_list); -- ATH_TXBUF_UNLOCK_IRQ(sc); -+ ATH_TXBUF_UNLOCK_BH(sc); - - /* - * move oldest from overflow to delivery -@@ -7426,9 +7491,6 @@ - if (sc->sc_reapcount > ATH_TXBUF_FREE_THRESHOLD) { - if (!sc->sc_dfswait) - netif_start_queue(sc->sc_dev); -- DPRINTF(sc, ATH_DEBUG_TX_PROC, -- "%s: tx tasklet restart the queue\n", -- __func__); - sc->sc_reapcount = 0; - sc->sc_devstopped = 0; - } else -@@ -7463,11 +7525,22 @@ +@@ -8298,12 +8416,24 @@ ath_tx_tasklet_q0(TQUEUE_ARG data) + { struct net_device *dev = (struct net_device *)data; struct ath_softc *sc = dev->priv; ++ unsigned long flags; +process_tx_again: if (txqactive(sc->sc_ah, 0)) @@ -509,96 +300,89 @@ diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c if (txqactive(sc->sc_ah, sc->sc_cabq->axq_qnum)) ath_tx_processq(sc, sc->sc_cabq); -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) -@@ -7484,6 +7557,7 @@ +@@ -8319,7 +8449,9 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data) + { struct net_device *dev = (struct net_device *)data; struct ath_softc *sc = dev->priv; ++ unsigned long flags; +process_tx_again: /* * Process each active queue. */ -@@ -7504,6 +7578,16 @@ +@@ -8340,6 +8472,16 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data) if (sc->sc_uapsdq && txqactive(sc->sc_ah, sc->sc_uapsdq->axq_qnum)) ath_tx_processq(sc, sc->sc_uapsdq); -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) -@@ -7521,6 +7605,7 @@ +@@ -8355,13 +8497,25 @@ ath_tx_tasklet(TQUEUE_ARG data) + struct net_device *dev = (struct net_device *)data; + struct ath_softc *sc = dev->priv; unsigned int i; ++ unsigned long flags; - /* Process each active queue. */ + /* Process each active queue. This includes sc_cabq, sc_xrtq and + * sc_uapsdq */ +process_tx_again: for (i = 0; i < HAL_NUM_TX_QUEUES; i++) if (ATH_TXQ_SETUP(sc, i) && txqactive(sc->sc_ah, i)) ath_tx_processq(sc, &sc->sc_txq[i]); -@@ -7529,6 +7614,16 @@ - ath_tx_processq(sc, sc->sc_xrtxq); - #endif -+ ATH_DISABLE_INTR(); ++ local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ ATH_ENABLE_INTR(); ++ local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) -@@ -7627,6 +7722,7 @@ - ath_draintxq(struct ath_softc *sc) - { - struct ath_hal *ah = sc->sc_ah; -+ int npend = 0; - unsigned int i; - - /* XXX return value */ -@@ -9160,9 +9256,9 @@ +@@ -10296,9 +10450,9 @@ ath_change_mtu(struct net_device *dev, i dev->mtu = mtu; if ((dev->flags & IFF_RUNNING) && !sc->sc_invalid) { /* NB: the rx buffers may need to be reallocated */ - tasklet_disable(&sc->sc_rxtq); -+ netif_poll_disable(dev); ++ ath_poll_disable(dev); error = ath_reset(dev); - tasklet_enable(&sc->sc_rxtq); -+ netif_poll_enable(dev); ++ ath_poll_enable(dev); } ATH_UNLOCK(sc); -Only in madwifi.dev/ath: if_ath.c.orig -diff -ur madwifi.old/ath/if_athvar.h madwifi.dev/ath/if_athvar.h ---- madwifi.old/ath/if_athvar.h 2007-05-31 06:48:28.555526400 +0200 -+++ madwifi.dev/ath/if_athvar.h 2007-05-31 06:48:57.847073408 +0200 -@@ -48,6 +48,10 @@ - #include "if_athioctl.h" - #include "net80211/ieee80211.h" /* XXX for WME_NUM_AC */ +--- a/ath/if_athvar.h ++++ b/ath/if_athvar.h +@@ -53,6 +53,10 @@ + # include + #endif +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) +#define irqs_disabled() 0 @@ -607,424 +391,146 @@ diff -ur madwifi.old/ath/if_athvar.h madwifi.dev/ath/if_athvar.h /* * Deduce if tasklets are available. If not then * fall back to using the immediate work queue. -@@ -478,8 +482,12 @@ - #define ATH_TXQ_LOCK_DESTROY(_tq) - #define ATH_TXQ_LOCK(_tq) spin_lock(&(_tq)->axq_lock) - #define ATH_TXQ_UNLOCK(_tq) spin_unlock(&(_tq)->axq_lock) --#define ATH_TXQ_LOCK_BH(_tq) spin_lock_bh(&(_tq)->axq_lock) --#define ATH_TXQ_UNLOCK_BH(_tq) spin_unlock_bh(&(_tq)->axq_lock) -+#define ATH_TXQ_LOCK_BH(_tq) \ -+ if (!irqs_disabled()) \ -+ spin_lock_bh(&(_tq)->axq_lock) -+#define ATH_TXQ_UNLOCK_BH(_tq) \ -+ if (!irqs_disabled()) \ -+ spin_unlock_bh(&(_tq)->axq_lock) - #define ATH_TXQ_LOCK_IRQ(_tq) do { \ - unsigned long __axq_lockflags; \ - spin_lock_irqsave(&(_tq)->axq_lock, __axq_lockflags); -@@ -624,7 +632,6 @@ +@@ -616,6 +620,9 @@ struct ath_rp { + struct ath_softc { + struct ieee80211com sc_ic; /* NB: must be first */ + struct net_device *sc_dev; ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) ++ struct napi_struct sc_napi; ++#endif + void __iomem *sc_iobase; /* address of the device */ + struct semaphore sc_lock; /* dev-level lock */ + struct net_device_stats sc_devstats; /* device statistics */ +@@ -730,7 +737,6 @@ struct ath_softc { struct ath_buf *sc_rxbufcur; /* current rx buffer */ u_int32_t *sc_rxlink; /* link ptr in last RX desc */ - spinlock_t sc_rxbuflock; + spinlock_t sc_rxbuflock; - struct ATH_TQ_STRUCT sc_rxtq; /* rx intr tasklet */ struct ATH_TQ_STRUCT sc_rxorntq; /* rxorn intr tasklet */ u_int8_t sc_defant; /* current default antenna */ - u_int8_t sc_rxotherant; /* rx's on non-default antenna*/ -@@ -637,6 +644,7 @@ + u_int8_t sc_rxotherant; /* RXs on non-default antenna */ +@@ -745,6 +751,7 @@ struct ath_softc { u_int sc_txintrperiod; /* tx interrupt batching */ struct ath_txq sc_txq[HAL_NUM_TX_QUEUES]; - struct ath_txq *sc_ac2q[WME_NUM_AC]; /* WME AC -> h/w qnum */ + struct ath_txq *sc_ac2q[WME_NUM_AC]; /* WME AC -> h/w qnum */ + HAL_INT sc_isr; /* unmasked ISR state */ struct ATH_TQ_STRUCT sc_txtq; /* tx intr tasklet */ - u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN]; + u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN]; struct ath_descdma sc_bdma; /* beacon descriptors */ -@@ -703,8 +711,12 @@ - #define ATH_TXBUF_LOCK_DESTROY(_sc) - #define ATH_TXBUF_LOCK(_sc) spin_lock(&(_sc)->sc_txbuflock) - #define ATH_TXBUF_UNLOCK(_sc) spin_unlock(&(_sc)->sc_txbuflock) --#define ATH_TXBUF_LOCK_BH(_sc) spin_lock_bh(&(_sc)->sc_txbuflock) --#define ATH_TXBUF_UNLOCK_BH(_sc) spin_unlock_bh(&(_sc)->sc_txbuflock) -+#define ATH_TXBUF_LOCK_BH(_sc) \ -+ if (!irqs_disabled()) \ -+ spin_lock_bh(&(_sc)->sc_txbuflock) -+#define ATH_TXBUF_UNLOCK_BH(_sc) \ -+ if (!irqs_disabled()) \ -+ spin_unlock_bh(&(_sc)->sc_txbuflock) - #define ATH_TXBUF_LOCK_IRQ(_sc) do { \ - unsigned long __txbuflockflags; \ - spin_lock_irqsave(&(_sc)->sc_txbuflock, __txbuflockflags); -@@ -722,8 +734,12 @@ - #define ATH_RXBUF_LOCK_DESTROY(_sc) - #define ATH_RXBUF_LOCK(_sc) spin_lock(&(_sc)->sc_rxbuflock) - #define ATH_RXBUF_UNLOCK(_sc) spin_unlock(&(_sc)->sc_rxbuflock) --#define ATH_RXBUF_LOCK_BH(_sc) spin_lock_bh(&(_sc)->sc_rxbuflock) --#define ATH_RXBUF_UNLOCK_BH(_sc) spin_unlock_bh(&(_sc)->sc_rxbuflock) -+#define ATH_RXBUF_LOCK_BH(_sc) \ -+ if (!irqs_disabled()) \ -+ spin_lock_bh(&(_sc)->sc_rxbuflock) -+#define ATH_RXBUF_UNLOCK_BH(_sc) \ -+ if (!irqs_disabled()) \ -+ spin_unlock_bh(&(_sc)->sc_rxbuflock) - #define ATH_RXBUF_LOCK_IRQ(_sc) do { \ - unsigned long __rxbuflockflags; \ - spin_lock_irqsave(&(_sc)->sc_rxbuflock, __rxbuflockflags); -@@ -733,6 +749,8 @@ - #define ATH_RXBUF_UNLOCK_IRQ_EARLY(_sc) \ - spin_unlock_irqrestore(&(_sc)->sc_rxbuflock, __rxbuflockflags); +@@ -858,6 +865,8 @@ typedef void (*ath_callback) (struct ath + #define ATH_TXBUF_LOCK_CHECK(_sc) + #endif +#define ATH_DISABLE_INTR local_irq_disable +#define ATH_ENABLE_INTR local_irq_enable - /* Protects the device from concurrent accesses */ - #define ATH_LOCK_INIT(_sc) init_MUTEX(&(_sc)->sc_lock) -Only in madwifi.dev/ath: if_athvar.h.orig -diff -ur madwifi.old/net80211/ieee80211_beacon.c madwifi.dev/net80211/ieee80211_beacon.c ---- madwifi.old/net80211/ieee80211_beacon.c 2007-05-31 06:48:28.555526400 +0200 -+++ madwifi.dev/net80211/ieee80211_beacon.c 2007-05-31 06:48:57.847073408 +0200 -@@ -286,7 +286,7 @@ - int len_changed = 0; - u_int16_t capinfo; - -- IEEE80211_LOCK(ic); -+ IEEE80211_BEACON_LOCK(ic); - - if ((ic->ic_flags & IEEE80211_F_DOTH) && - (vap->iv_flags & IEEE80211_F_CHANSWITCH) && -@@ -547,7 +547,7 @@ - vap->iv_flags_ext &= ~IEEE80211_FEXT_APPIE_UPDATE; - } - -- IEEE80211_UNLOCK(ic); -+ IEEE80211_BEACON_UNLOCK(ic); - - return len_changed; - } -diff -ur madwifi.old/net80211/ieee80211_input.c madwifi.dev/net80211/ieee80211_input.c ---- madwifi.old/net80211/ieee80211_input.c 2007-05-31 06:48:28.557526096 +0200 -+++ madwifi.dev/net80211/ieee80211_input.c 2007-05-31 06:48:57.849073104 +0200 -@@ -1169,8 +1169,9 @@ - if (ni->ni_vlan != 0 && vap->iv_vlgrp != NULL) { + #define ATH_RXBUF_LOCK_INIT(_sc) spin_lock_init(&(_sc)->sc_rxbuflock) + #define ATH_RXBUF_LOCK_DESTROY(_sc) +--- a/net80211/ieee80211_input.c ++++ b/net80211/ieee80211_input.c +@@ -1198,7 +1198,7 @@ ieee80211_deliver_data(struct ieee80211_ /* attach vlan tag */ - vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan); -- } else -- netif_rx(skb); -+ } else { -+ netif_receive_skb(skb); -+ } - dev->last_rx = jiffies; - } - } -@@ -3675,9 +3676,9 @@ - } - - /* Okay, take the first queued packet and put it out... */ -- IEEE80211_NODE_SAVEQ_LOCK(ni); -+ IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni); - IEEE80211_NODE_SAVEQ_DEQUEUE(ni, skb, qlen); -- IEEE80211_NODE_SAVEQ_UNLOCK(ni); -+ IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni); - if (skb == NULL) { - IEEE80211_NOTE_MAC(vap, IEEE80211_MSG_POWER, wh->i_addr2, - "%s", "recv ps-poll, but queue empty"); -Only in madwifi.dev/net80211: ieee80211_input.c.orig -diff -ur madwifi.old/net80211/ieee80211_linux.h madwifi.dev/net80211/ieee80211_linux.h ---- madwifi.old/net80211/ieee80211_linux.h 2007-05-24 19:31:37.000000000 +0200 -+++ madwifi.dev/net80211/ieee80211_linux.h 2007-05-31 06:48:57.849073104 +0200 -@@ -31,6 +31,10 @@ - - #include - -+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) -+#define irqs_disabled() 0 -+#endif -+ - /* - * Task deferral - * -@@ -86,8 +90,12 @@ - } while (0) - #define IEEE80211_UNLOCK_IRQ_EARLY(_ic) \ - spin_unlock_irqrestore(&(_ic)->ic_comlock, __ilockflags); --#define IEEE80211_LOCK_BH(_ic) spin_lock_bh(&(_ic)->ic_comlock) --#define IEEE80211_UNLOCK_BH(_ic) spin_unlock_bh(&(_ic)->ic_comlock) -+#define IEEE80211_LOCK_BH(_ic) \ -+ if (!irqs_disabled()) \ -+ spin_lock_bh(&(_ic)->ic_comlock) -+#define IEEE80211_UNLOCK_BH(_ic) \ -+ if (!irqs_disabled()) \ -+ spin_unlock_bh(&(_ic)->ic_comlock) - #define IEEE80211_LOCK(_ic) spin_lock(&(_ic)->ic_comlock) - #define IEEE80211_UNLOCK(_ic) spin_unlock(&(_ic)->ic_comlock) - -@@ -104,15 +112,22 @@ - #define IEEE80211_VAPS_LOCK_DESTROY(_ic) - #define IEEE80211_VAPS_LOCK(_ic) spin_lock(&(_ic)->ic_vapslock); - #define IEEE80211_VAPS_UNLOCK(_ic) spin_unlock(&(_ic)->ic_vapslock); --#define IEEE80211_VAPS_LOCK_BH(_ic) spin_lock_bh(&(_ic)->ic_vapslock); --#define IEEE80211_VAPS_UNLOCK_BH(_ic) spin_unlock_bh(&(_ic)->ic_vapslock); --#define IEEE80211_VAPS_LOCK_IRQ(_ic) do { \ -- int _vaps_lockflags; \ -- spin_lock_irqsave(&(_ic)->ic_vapslock, _vaps_lockflags); --#define IEEE80211_VAPS_UNLOCK_IRQ(_ic) \ -- spin_unlock_irqrestore(&(_ic)->ic_vapslock, _vaps_lockflags); \ --} while (0) --#define IEEE80211_VAPS_UNLOCK_IRQ_EARLY(_ic) spin_unlock_irqrestore(&(_ic)->ic_vapslock, _vaps_lockflags) -+#define IEEE80211_VAPS_LOCK_BH(_ic) \ -+ if (!irqs_disabled()) \ -+ spin_lock_bh(&(_ic)->ic_vapslock); -+#define IEEE80211_VAPS_UNLOCK_BH(_ic) \ -+ if (!irqs_disabled()) \ -+ spin_unlock_bh(&(_ic)->ic_vapslock); -+#define IEEE80211_VAPS_LOCK_IRQ(_ic) do { \ -+ unsigned long __vlockflags=0; \ -+ unsigned int __vlocked=0; \ -+ __vlocked=spin_is_locked(&(_ic)->ic_vapslock); \ -+ if(!__vlocked) spin_lock_irqsave(&(_ic)->ic_vapslock, __vlockflags); -+#define IEEE80211_VAPS_UNLOCK_IRQ(_ic) \ -+ if(!__vlocked) spin_unlock_irqrestore(&(_ic)->ic_vapslock, __vlockflags); \ -+} while (0); -+#define IEEE80211_VAPS_UNLOCK_IRQ_EARLY(_ic) \ -+ if (!__vlocked) spin_unlock_irqrestore(&(_ic)->ic_vapslock, _vaps_lockflags) - - #if (defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK)) && defined(spin_is_locked) - #define IEEE80211_VAPS_LOCK_ASSERT(_ic) \ -@@ -122,6 +137,11 @@ - #define IEEE80211_VAPS_LOCK_ASSERT(_ic) - #endif - -+/* -+ * Beacon locking definitions; piggyback on com lock. -+ */ -+#define IEEE80211_BEACON_LOCK(_ic) IEEE80211_LOCK_IRQ(_ic) -+#define IEEE80211_BEACON_UNLOCK(_ic) IEEE80211_UNLOCK_IRQ(_ic) - - /* - * Node locking definitions. -@@ -191,8 +211,12 @@ - typedef spinlock_t ieee80211_scan_lock_t; - #define IEEE80211_SCAN_LOCK_INIT(_nt, _name) spin_lock_init(&(_nt)->nt_scanlock) - #define IEEE80211_SCAN_LOCK_DESTROY(_nt) --#define IEEE80211_SCAN_LOCK_BH(_nt) spin_lock_bh(&(_nt)->nt_scanlock) --#define IEEE80211_SCAN_UNLOCK_BH(_nt) spin_unlock_bh(&(_nt)->nt_scanlock) -+#define IEEE80211_SCAN_LOCK_BH(_nt) \ -+ if (!irqs_disabled()) \ -+ spin_lock_bh(&(_nt)->nt_scanlock) -+#define IEEE80211_SCAN_UNLOCK_BH(_nt) \ -+ if (!irqs_disabled()) \ -+ spin_unlock_bh(&(_nt)->nt_scanlock) - #define IEEE80211_SCAN_LOCK_IRQ(_nt) do { \ - unsigned long __scan_lockflags; \ - spin_lock_irqsave(&(_nt)->nt_scanlock, __scan_lockflags); -@@ -217,8 +241,12 @@ - #define ACL_LOCK_DESTROY(_as) - #define ACL_LOCK(_as) spin_lock(&(_as)->as_lock) - #define ACL_UNLOCK(_as) spin_unlock(&(_as)->as_lock) --#define ACL_LOCK_BH(_as) spin_lock_bh(&(_as)->as_lock) --#define ACL_UNLOCK_BH(_as) spin_unlock_bh(&(_as)->as_lock) -+#define ACL_LOCK_BH(_as) \ -+ if (!irqs_disabled()) \ -+ spin_lock_bh(&(_as)->as_lock) -+#define ACL_UNLOCK_BH(_as) \ -+ if (!irqs_disabled()) \ -+ spin_unlock_bh(&(_as)->as_lock) - - #if (defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK)) && defined(spin_is_locked) - #define ACL_LOCK_ASSERT(_as) \ -diff -ur madwifi.old/net80211/ieee80211_node.c madwifi.dev/net80211/ieee80211_node.c ---- madwifi.old/net80211/ieee80211_node.c 2007-05-31 06:48:28.558525944 +0200 -+++ madwifi.dev/net80211/ieee80211_node.c 2007-05-31 06:48:57.868070216 +0200 -@@ -1584,7 +1584,7 @@ - struct ieee80211_node *ni; - u_int gen; - -- IEEE80211_SCAN_LOCK_IRQ(nt); -+ IEEE80211_SCAN_LOCK_BH(nt); - gen = ++nt->nt_scangen; - - restart: -@@ -1604,7 +1604,7 @@ - } - IEEE80211_NODE_TABLE_UNLOCK_IRQ(nt); - -- IEEE80211_SCAN_UNLOCK_IRQ(nt); -+ IEEE80211_SCAN_UNLOCK_BH(nt); - } - EXPORT_SYMBOL(ieee80211_iterate_dev_nodes); - -Only in madwifi.dev/net80211: ieee80211_node.c.orig -diff -ur madwifi.old/net80211/ieee80211_power.c madwifi.dev/net80211/ieee80211_power.c ---- madwifi.old/net80211/ieee80211_power.c 2007-05-24 19:31:37.000000000 +0200 -+++ madwifi.dev/net80211/ieee80211_power.c 2007-05-31 06:48:57.868070216 +0200 -@@ -147,7 +147,7 @@ - #endif - struct sk_buff *skb; - -- IEEE80211_NODE_SAVEQ_LOCK(ni); -+ IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni); - while ((skb = skb_peek(&ni->ni_savedq)) != NULL && - M_AGE_GET(skb) < IEEE80211_INACT_WAIT) { - IEEE80211_NOTE(vap, IEEE80211_MSG_POWER, ni, -@@ -159,7 +159,7 @@ - } - if (skb != NULL) - M_AGE_SUB(skb, IEEE80211_INACT_WAIT); -- IEEE80211_NODE_SAVEQ_UNLOCK(ni); -+ IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni); - - IEEE80211_NOTE(vap, IEEE80211_MSG_POWER, ni, - "discard %u frames for age", discard); -@@ -185,7 +185,7 @@ - KASSERT(aid < vap->iv_max_aid, - ("bogus aid %u, max %u", aid, vap->iv_max_aid)); - -- IEEE80211_LOCK(ni->ni_ic); -+ IEEE80211_BEACON_LOCK(ni->ni_ic); - if (set != (isset(vap->iv_tim_bitmap, aid) != 0)) { - if (set) { - setbit(vap->iv_tim_bitmap, aid); -@@ -196,7 +196,7 @@ - } - vap->iv_flags |= IEEE80211_F_TIMUPDATE; - } -- IEEE80211_UNLOCK(ni->ni_ic); -+ IEEE80211_BEACON_UNLOCK(ni->ni_ic); - } - - /* -@@ -295,9 +295,9 @@ - struct sk_buff *skb; - int qlen; - -- IEEE80211_NODE_SAVEQ_LOCK(ni); -+ IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni); - IEEE80211_NODE_SAVEQ_DEQUEUE(ni, skb, qlen); -- IEEE80211_NODE_SAVEQ_UNLOCK(ni); -+ IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni); - if (skb == NULL) - break; - /* -@@ -362,9 +362,9 @@ - for (;;) { - struct sk_buff *skb; - -- IEEE80211_NODE_SAVEQ_LOCK(ni); -+ IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni); - skb = __skb_dequeue(&ni->ni_savedq); -- IEEE80211_NODE_SAVEQ_UNLOCK(ni); -+ IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni); - if (skb == NULL) - break; - ieee80211_parent_queue_xmit(skb); -diff -ur madwifi.old/net80211/ieee80211_proto.c madwifi.dev/net80211/ieee80211_proto.c ---- madwifi.old/net80211/ieee80211_proto.c 2007-05-31 06:48:28.564525032 +0200 -+++ madwifi.dev/net80211/ieee80211_proto.c 2007-05-31 06:48:57.869070064 +0200 -@@ -635,9 +635,9 @@ - { - struct ieee80211com *ic = vap->iv_ic; - -- IEEE80211_LOCK(ic); -+ IEEE80211_BEACON_LOCK(ic); - ieee80211_wme_initparams_locked(vap); -- IEEE80211_UNLOCK(ic); -+ IEEE80211_BEACON_UNLOCK(ic); + struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; + if (vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ /* If netif_receive_skb dropped the packet because + * device was too busy */ + if (ni_tmp != NULL) { + /* node reference was leaked */ +@@ -1209,8 +1209,8 @@ ieee80211_deliver_data(struct ieee80211_ + skb = NULL; /* SKB is no longer ours */ + } else { + struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; +- if (netif_rx(skb) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ if (netif_receive_skb(skb) == NET_RX_DROP) { ++ /* If netif_receive_skb dropped the packet because + * device was too busy */ + if (ni_tmp != NULL) { + /* node reference was leaked */ +@@ -2322,8 +2322,8 @@ forward_mgmt_to_app(struct ieee80211vap + skb1->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */ + + ni_tmp = SKB_CB(skb1)->ni; +- if (netif_rx(skb1) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ if (netif_receive_skb(skb1) == NET_RX_DROP) { ++ /* If netif_receive_skb dropped the packet because + * device was too busy */ + if (ni_tmp != NULL) { + /* node reference was leaked */ +--- a/net80211/ieee80211_monitor.c ++++ b/net80211/ieee80211_monitor.c +@@ -584,8 +584,8 @@ ieee80211_input_monitor(struct ieee80211 + skb1->protocol = + __constant_htons(0x0019); /* ETH_P_80211_RAW */ + +- if (netif_rx(skb1) == NET_RX_DROP) { +- /* If netif_rx dropped the packet because ++ if (netif_receive_skb(skb1) == NET_RX_DROP) { ++ /* If netif_receive_skb dropped the packet because + * device was too busy, reclaim the ref. in + * the skb. */ + if (SKB_CB(skb1)->ni != NULL) +--- a/net80211/ieee80211_skb.c ++++ b/net80211/ieee80211_skb.c +@@ -73,7 +73,7 @@ + #undef dev_queue_xmit + #undef kfree_skb + #undef kfree_skb_fast +-#undef netif_rx ++#undef netif_receive_skb + #undef pskb_copy + #undef skb_clone + #undef skb_copy +@@ -638,8 +638,8 @@ int vlan_hwaccel_receive_skb_debug(stru + grp, vlan_tag); } - void -@@ -920,9 +920,9 @@ - struct ieee80211com *ic = vap->iv_ic; - - if (ic->ic_caps & IEEE80211_C_WME) { -- IEEE80211_LOCK(ic); -+ IEEE80211_BEACON_LOCK(ic); - ieee80211_wme_updateparams_locked(vap); -- IEEE80211_UNLOCK(ic); -+ IEEE80211_BEACON_UNLOCK(ic); - } - } - -diff -ur madwifi.old/net80211/ieee80211_scan_sta.c madwifi.dev/net80211/ieee80211_scan_sta.c ---- madwifi.old/net80211/ieee80211_scan_sta.c 2007-05-21 17:53:39.000000000 +0200 -+++ madwifi.dev/net80211/ieee80211_scan_sta.c 2007-05-31 06:48:57.870069912 +0200 -@@ -163,9 +163,11 @@ - { - struct sta_table *st = ss->ss_priv; - -- spin_lock(&st->st_lock); -+ if (!irqs_disabled()) -+ spin_lock_bh(&st->st_lock); - sta_flush_table(st); -- spin_unlock(&st->st_lock); -+ if (!irqs_disabled()) -+ spin_unlock_bh(&st->st_lock); - ss->ss_last = 0; - return 0; +-int netif_rx_debug(struct sk_buff *skb, const char* func, int line) { +- return netif_rx(untrack_skb(skb, 0, func, line, __func__, __LINE__)); ++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line) { ++ return netif_receive_skb(untrack_skb(skb, 0, func, line, __func__, __LINE__)); } -@@ -215,7 +217,8 @@ - int hash; - - hash = STA_HASH(macaddr); -- spin_lock(&st->st_lock); -+ if (!irqs_disabled()) -+ spin_lock_bh(&st->st_lock); - LIST_FOREACH(se, &st->st_hash[hash], se_hash) - if (IEEE80211_ADDR_EQ(se->base.se_macaddr, macaddr) && - sp->ssid[1] == se->base.se_ssid[1] && -@@ -225,7 +228,7 @@ - MALLOC(se, struct sta_entry *, sizeof(struct sta_entry), - M_80211_SCAN, M_NOWAIT | M_ZERO); - if (se == NULL) { -- spin_unlock(&st->st_lock); -+ spin_unlock_bh(&st->st_lock); - return 0; - } - se->se_scangen = st->st_scangen-1; -@@ -287,7 +290,8 @@ - se->se_seen = 1; - se->se_notseen = 0; - -- spin_unlock(&st->st_lock); -+ if (!irqs_disabled()) -+ spin_unlock_bh(&st->st_lock); - - /* - * If looking for a quick choice and nothing's -@@ -1063,7 +1067,8 @@ - u_int gen; - int res = 0; -- spin_lock(&st->st_scanlock); -+ if (!irqs_disabled()) -+ spin_lock_bh(&st->st_scanlock); - gen = st->st_scangen++; - restart: - spin_lock(&st->st_lock); -@@ -1086,7 +1091,8 @@ - spin_unlock(&st->st_lock); - - done: -- spin_unlock(&st->st_scanlock); -+ if (!irqs_disabled()) -+ spin_unlock_bh(&st->st_scanlock); - - return res; + struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, +@@ -760,7 +760,7 @@ struct sk_buff * skb_copy_expand_debug(c } -@@ -1235,7 +1241,8 @@ - bestchan = NULL; - bestrssi = -1; -- spin_lock(&st->st_lock); -+ if (!irqs_disabled()) -+ spin_lock_bh(&st->st_lock); - for (i = 0; i < ss->ss_last; i++) { - c = ss->ss_chans[i]; - maxrssi = 0; -@@ -1248,7 +1255,8 @@ - if (bestchan == NULL || maxrssi < bestrssi) - bestchan = c; - } -- spin_unlock(&st->st_lock); -+ if (!irqs_disabled()) -+ spin_unlock_bh(&st->st_lock); - - return bestchan; - } + EXPORT_SYMBOL(vlan_hwaccel_receive_skb_debug); +-EXPORT_SYMBOL(netif_rx_debug); ++EXPORT_SYMBOL(netif_receive_skb_debug); + EXPORT_SYMBOL(alloc_skb_debug); + EXPORT_SYMBOL(dev_alloc_skb_debug); + EXPORT_SYMBOL(skb_clone_debug); +--- a/net80211/ieee80211_skb.h ++++ b/net80211/ieee80211_skb.h +@@ -116,7 +116,7 @@ int ieee80211_skb_references(void); + int vlan_hwaccel_receive_skb_debug(struct sk_buff *skb, + struct vlan_group *grp, unsigned short vlan_tag, + const char* func, int line); +-int netif_rx_debug(struct sk_buff *skb, const char* func, int line); ++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line); + struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, + const char *func, int line); + struct sk_buff * dev_alloc_skb_debug(unsigned int length, +@@ -151,7 +151,7 @@ struct sk_buff * skb_copy_expand_debug(c + #undef dev_queue_xmit + #undef kfree_skb + #undef kfree_skb_fast +-#undef netif_rx ++#undef netif_receive_skb + #undef pskb_copy + #undef skb_clone + #undef skb_copy +@@ -168,8 +168,8 @@ struct sk_buff * skb_copy_expand_debug(c + skb_copy_expand_debug(_skb, _newheadroom, _newtailroom, _gfp_mask, __func__, __LINE__) + #define vlan_hwaccel_receive_skb(_skb, _grp, _tag) \ + vlan_hwaccel_receive_skb_debug(_skb, _grp, _tag, __func__, __LINE__) +-#define netif_rx(_skb) \ +- netif_rx_debug(_skb, __func__, __LINE__) ++#define netif_receive_skb(_skb) \ ++ netif_receive_skb_debug(_skb, __func__, __LINE__) + #define alloc_skb(_length, _gfp_mask) \ + alloc_skb_debug(_length, _gfp_mask, __func__, __LINE__) + #define dev_alloc_skb(_length) \