ath9k: add a temporary uci option for setting the channel bandwidth
[openwrt.git] / package / madwifi / patches / 300-napi_polling.patch
index 1cca8f5..bde7684 100644 (file)
@@ -1,58 +1,81 @@
-Index: madwifi-dfs-r3053/ath/if_ath.c
-===================================================================
---- madwifi-dfs-r3053.orig/ath/if_ath.c        2007-12-13 05:25:13.342385808 +0100
-+++ madwifi-dfs-r3053/ath/if_ath.c     2007-12-13 05:25:13.662404044 +0100
-@@ -189,7 +189,7 @@
+--- a/ath/if_ath.c
++++ b/ath/if_ath.c
+@@ -184,7 +184,11 @@ static void ath_recv_mgmt(struct ieee802
        struct sk_buff *, int, int, u_int64_t);
  static void ath_setdefantenna(struct ath_softc *, u_int);
  static struct ath_txq *ath_txq_setup(struct ath_softc *, int, int);
 -static void ath_rx_tasklet(TQUEUE_ARG);
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++static int ath_rx_poll(struct napi_struct *napi, int budget);
++#else
 +static int ath_rx_poll(struct net_device *dev, int *budget);
++#endif
  static int ath_hardstart(struct sk_buff *, struct net_device *);
  static int ath_mgtstart(struct ieee80211com *, struct sk_buff *);
  #ifdef ATH_SUPERG_COMP
-@@ -535,7 +535,6 @@
-       ATH_TXBUF_LOCK_INIT(sc);
-       ATH_RXBUF_LOCK_INIT(sc);
+@@ -376,6 +380,9 @@ static u_int32_t ath_set_clamped_maxtxpo
+               u_int32_t new_clamped_maxtxpower);
+ static u_int32_t ath_get_real_maxtxpower(struct ath_softc *sc);
++static void ath_poll_disable(struct net_device *dev);
++static void ath_poll_enable(struct net_device *dev);
++
+ /* calibrate every 30 secs in steady state but check every second at first. */
+ static int ath_calinterval = ATH_SHORT_CALINTERVAL;
+ static int ath_countrycode = CTRY_DEFAULT;    /* country code */
+@@ -547,7 +554,6 @@ ath_attach(u_int16_t devid, struct net_d
+       atomic_set(&sc->sc_txbuf_counter, 0);
  
 -      ATH_INIT_TQUEUE(&sc->sc_rxtq,     ath_rx_tasklet,       dev);
-       ATH_INIT_TQUEUE(&sc->sc_txtq,     ath_tx_tasklet,       dev);
+       ATH_INIT_TQUEUE(&sc->sc_txtq,     ath_tx_tasklet,       dev);
        ATH_INIT_TQUEUE(&sc->sc_bmisstq,  ath_bmiss_tasklet,    dev);
        ATH_INIT_TQUEUE(&sc->sc_bstucktq, ath_bstuck_tasklet,   dev);
-@@ -808,6 +807,8 @@
+@@ -821,6 +827,12 @@ ath_attach(u_int16_t devid, struct net_d
        dev->set_mac_address = ath_set_mac_address;
        dev->change_mtu = ath_change_mtu;
        dev->tx_queue_len = ATH_TXBUF - ATH_TXBUF_MGT_RESERVED;
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++      netif_napi_add(dev, &sc->sc_napi, ath_rx_poll, 64);
++#else
 +      dev->poll = ath_rx_poll;
 +      dev->weight = 64;
++#endif
  #ifdef USE_HEADERLEN_RESV
        dev->hard_header_len += sizeof(struct ieee80211_qosframe) +
                                sizeof(struct llc) +
-@@ -2150,6 +2151,7 @@
-               (status & HAL_INT_RXPHY)   ? " HAL_INT_RXPHY"   : "",
-               (status & HAL_INT_SWBA)    ? " HAL_INT_SWBA"    : "");
+@@ -2220,6 +2232,7 @@ ath_intr(int irq, void *dev_id, struct p
+               (status & HAL_INT_GLOBAL)       ? " HAL_INT_GLOBAL"     : ""
+               );
  
 +      sc->sc_isr = status;
        status &= sc->sc_imask;                 /* discard unasked for bits */
-       if (status & HAL_INT_FATAL) {
-               sc->sc_stats.ast_hardware++;
-@@ -2201,7 +2203,14 @@
+       /* As soon as we know we have a real interrupt we intend to service, 
+        * we will check to see if we need an initial hardware TSF reading. 
+@@ -2277,7 +2290,21 @@ ath_intr(int irq, void *dev_id, struct p
                }
                if (status & (HAL_INT_RX | HAL_INT_RXPHY)) {
-                       ath_uapsd_processtriggers(sc);
+                       ath_uapsd_processtriggers(sc, hw_tsf);
 -                      ATH_SCHEDULE_TQUEUE(&sc->sc_rxtq, &needmark);
 +                      sc->sc_isr &= ~HAL_INT_RX;
-+                      if (netif_rx_schedule_prep(dev)) {
-+#ifndef ATH_PRECISE_TSF
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++                      if (netif_rx_schedule_prep(dev, &sc->sc_napi))
++#else
++                      if (netif_rx_schedule_prep(dev))
++#endif
++                      {
 +                              sc->sc_imask &= ~HAL_INT_RX;
 +                              ath_hal_intrset(ah, sc->sc_imask);
-+#endif
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++                              __netif_rx_schedule(dev, &sc->sc_napi);
++#else
 +                              __netif_rx_schedule(dev);
++#endif
 +                      }
                }
                if (status & HAL_INT_TX) {
  #ifdef ATH_SUPERG_DYNTURBO
-@@ -2227,6 +2236,11 @@
+@@ -2303,6 +2330,11 @@ ath_intr(int irq, void *dev_id, struct p
                                }
                        }
  #endif
@@ -64,71 +87,160 @@ Index: madwifi-dfs-r3053/ath/if_ath.c
                        ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, &needmark);
                }
                if (status & HAL_INT_BMISS) {
-@@ -3904,10 +3918,10 @@
-        *
-        * XXX Using in_softirq is not right since we might
-        * be called from other soft irq contexts than
--       * ath_rx_tasklet.
+@@ -2515,6 +2547,7 @@ ath_init(struct net_device *dev)
+       if (sc->sc_tx99 != NULL)
+               sc->sc_tx99->start(sc->sc_tx99);
+ #endif
++      ath_poll_enable(dev);
+ done:
+       ATH_UNLOCK(sc);
+@@ -2555,6 +2588,9 @@ ath_stop_locked(struct net_device *dev)
+               if (sc->sc_tx99 != NULL)
+                       sc->sc_tx99->stop(sc->sc_tx99);
+ #endif
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++              ath_poll_disable(dev);
++#endif
+               netif_stop_queue(dev);  /* XXX re-enabled by ath_newstate */
+               dev->flags &= ~IFF_RUNNING;     /* NB: avoid recursion */
+               ieee80211_stop_running(ic);     /* stop all VAPs */
+@@ -4013,12 +4049,47 @@ ath_key_set(struct ieee80211vap *vap, co
+       return ath_keyset(sc, k, mac, vap->iv_bss);
+ }
++static void ath_poll_disable(struct net_device *dev)
++{
++      struct ath_softc *sc = dev->priv;
++
++      /*
++       * XXX Using in_softirq is not right since we might
++       * be called from other soft irq contexts than
 +       * ath_rx_poll
++       */
++      if (!in_softirq()) {
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++              napi_disable(&sc->sc_napi);
++#else
++              netif_poll_disable(dev);
++#endif
++      }
++}
++
++static void ath_poll_enable(struct net_device *dev)
++{
++      struct ath_softc *sc = dev->priv;
++
++      /* NB: see above */
++      if (!in_softirq()) {
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++              napi_enable(&sc->sc_napi);
++#else
++              netif_poll_enable(dev);
++#endif
++      }
++}
++
++
+ /*
+  * Block/unblock tx+rx processing while a key change is done.
+  * We assume the caller serializes key management operations
+  * so we only need to worry about synchronization with other
+  * uses that originate in the driver.
+  */
++#define       IS_UP(_dev) \
++      (((_dev)->flags & (IFF_RUNNING|IFF_UP)) == (IFF_RUNNING|IFF_UP))
+ static void
+ ath_key_update_begin(struct ieee80211vap *vap)
+ {
+@@ -4032,14 +4103,9 @@ ath_key_update_begin(struct ieee80211vap
+        * When called from the rx tasklet we cannot use
+        * tasklet_disable because it will block waiting
+        * for us to complete execution.
+-       *
+-       * XXX Using in_softirq is not right since we might
+-       * be called from other soft irq contexts than
+-       * ath_rx_tasklet.
         */
-       if (!in_softirq())
+-      if (!in_softirq())
 -              tasklet_disable(&sc->sc_rxtq);
-+              netif_poll_disable(dev);
-       netif_stop_queue(dev);
+-      netif_stop_queue(dev);
++      if (IS_UP(vap->iv_dev))
++              netif_stop_queue(dev);
  }
  
-@@ -3920,7 +3934,7 @@
-       DPRINTF(sc, ATH_DEBUG_KEYCACHE, "%s:\n", __func__);
-       netif_start_queue(dev);
-       if (!in_softirq())              /* NB: see above */
+ static void
+@@ -4051,9 +4117,9 @@ ath_key_update_end(struct ieee80211vap *
+ #endif
+       DPRINTF(sc, ATH_DEBUG_KEYCACHE, "End\n");
+-      netif_wake_queue(dev);
+-      if (!in_softirq())              /* NB: see above */
 -              tasklet_enable(&sc->sc_rxtq);
-+              netif_poll_enable(dev);
++
++      if (IS_UP(vap->iv_dev))
++              netif_wake_queue(dev);
  }
  
  /*
-@@ -6287,13 +6301,12 @@
+@@ -6360,15 +6426,25 @@ ath_setdefantenna(struct ath_softc *sc,
        sc->sc_rxotherant = 0;
  }
  
 -static void
 -ath_rx_tasklet(TQUEUE_ARG data)
 +static int
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++ath_rx_poll(struct napi_struct *napi, int budget)
++#else
 +ath_rx_poll(struct net_device *dev, int *budget)
++#endif
  {
  #define       PA2DESC(_sc, _pa) \
        ((struct ath_desc *)((caddr_t)(_sc)->sc_rxdma.dd_desc + \
                ((_pa) - (_sc)->sc_rxdma.dd_desc_paddr)))
 -      struct net_device *dev = (struct net_device *)data;
-       struct ath_buf *bf;
+-      struct ath_buf *bf;
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++      struct ath_softc *sc = container_of(napi, struct ath_softc, sc_napi);
++      struct net_device *dev = sc->sc_dev;
++      u_int rx_limit = budget;
++#else
        struct ath_softc *sc = dev->priv;
++      u_int rx_limit = min(dev->quota, *budget);
++#endif
++      struct ath_buf *bf;
        struct ieee80211com *ic = &sc->sc_ic;
-@@ -6305,8 +6318,11 @@
+       struct ath_hal *ah = sc ? sc->sc_ah : NULL;
+       struct ath_desc *ds;
+@@ -6378,8 +6454,10 @@ ath_rx_tasklet(TQUEUE_ARG data)
        unsigned int len;
        int type;
        u_int phyerr;
 +      u_int processed = 0, early_stop = 0;
-+      u_int rx_limit = dev->quota;
  
-       DPRINTF(sc, ATH_DEBUG_RX_PROC, "%s\n", __func__);
+       DPRINTF(sc, ATH_DEBUG_RX_PROC, "invoked\n");
 +process_rx_again:
        do {
                bf = STAILQ_FIRST(&sc->sc_rxbuf);
                if (bf == NULL) {               /* XXX ??? can this happen */
-@@ -6330,6 +6346,13 @@
+@@ -6403,6 +6481,15 @@ ath_rx_tasklet(TQUEUE_ARG data)
                        /* NB: never process the self-linked entry at the end */
                        break;
                }
 +
-+              processed++;
-+              if (rx_limit-- < 0) {
++              if (rx_limit-- < 2) {
 +                      early_stop = 1;
 +                      break;
 +              }
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++              processed++;
++#endif
 +
                skb = bf->bf_skb;
                if (skb == NULL) {
-                       printk("%s: no skbuff (%s)\n", DEV_NAME(dev), __func__);
-@@ -6374,6 +6397,7 @@
+                       EPRINTF(sc, "Dropping; buffer contains NULL skbuff.\n");
+@@ -6450,6 +6537,7 @@ ath_rx_tasklet(TQUEUE_ARG data)
                                sc->sc_stats.ast_rx_phyerr++;
                                phyerr = rs->rs_phyerr & 0x1f;
                                sc->sc_stats.ast_rx_phy[phyerr]++;
@@ -136,43 +248,51 @@ Index: madwifi-dfs-r3053/ath/if_ath.c
                        }
                        if (rs->rs_status & HAL_RXERR_DECRYPT) {
                                /*
-@@ -6571,9 +6595,33 @@
+@@ -6645,9 +6733,39 @@ rx_next:
                STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
                ATH_RXBUF_UNLOCK_IRQ(sc);
        } while (ath_rxbuf_init(sc, bf) == 0);
 +      if (!early_stop) {
++              unsigned long flags;
 +              /* Check if more data is received while we were
 +               * processing the descriptor chain.
 +               */
-+#ifndef ATH_PRECISE_TSF
-+              ATH_DISABLE_INTR();
++              local_irq_save(flags);
 +              if (sc->sc_isr & HAL_INT_RX) {
++                      u_int64_t hw_tsf = ath_hal_gettsf64(ah);
 +                      sc->sc_isr &= ~HAL_INT_RX;
-+                      ATH_ENABLE_INTR();
-+                      ath_uapsd_processtriggers(sc);
++                      local_irq_restore(flags);
++                      ath_uapsd_processtriggers(sc, hw_tsf);
 +                      goto process_rx_again;
 +              }
-+#endif
-+              netif_rx_complete(dev);
-+
-+#ifndef ATH_PRECISE_TSF
-+              sc->sc_imask |= HAL_INT_RX;
-+              ath_hal_intrset(ah, sc->sc_imask);
-+              ATH_ENABLE_INTR();
-+#endif
++              local_irq_restore(flags);
 +      }
 +
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++      netif_rx_complete(dev, napi);
++#else
++      netif_rx_complete(dev);
 +      *budget -= processed;
++      dev->quota -= processed;
++#endif
++      sc->sc_imask |= HAL_INT_RX;
++      ath_hal_intrset(ah, sc->sc_imask);
  
        /* rx signal state monitoring */
        ath_hal_rxmonitor(ah, &sc->sc_halstats, &sc->sc_curchan);
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++      return processed;
++#else
 +      return early_stop;
++#endif
  #undef PA2DESC
  }
  
-@@ -8192,11 +8240,22 @@
+@@ -8298,12 +8416,24 @@ ath_tx_tasklet_q0(TQUEUE_ARG data)
+ {
        struct net_device *dev = (struct net_device *)data;
        struct ath_softc *sc = dev->priv;
++      unsigned long flags;
  
 +process_tx_again:
        if (txqactive(sc->sc_ah, 0))
@@ -180,96 +300,89 @@ Index: madwifi-dfs-r3053/ath/if_ath.c
        if (txqactive(sc->sc_ah, sc->sc_cabq->axq_qnum))
                ath_tx_processq(sc, sc->sc_cabq);
  
-+      ATH_DISABLE_INTR();
++      local_irq_save(flags);
 +      if (sc->sc_isr & HAL_INT_TX) {
 +              sc->sc_isr &= ~HAL_INT_TX;
-+              ATH_ENABLE_INTR();
++              local_irq_restore(flags);
 +              goto process_tx_again;
 +      }
 +      sc->sc_imask |= HAL_INT_TX;
 +      ath_hal_intrset(sc->sc_ah, sc->sc_imask);
-+      ATH_ENABLE_INTR();
++      local_irq_restore(flags);
 +
        netif_wake_queue(dev);
  
        if (sc->sc_softled)
-@@ -8213,6 +8272,7 @@
+@@ -8319,7 +8449,9 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data)
+ {
        struct net_device *dev = (struct net_device *)data;
        struct ath_softc *sc = dev->priv;
++      unsigned long flags;
  
 +process_tx_again:
        /*
         * Process each active queue.
         */
-@@ -8233,6 +8293,16 @@
+@@ -8340,6 +8472,16 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data)
        if (sc->sc_uapsdq && txqactive(sc->sc_ah, sc->sc_uapsdq->axq_qnum))
                ath_tx_processq(sc, sc->sc_uapsdq);
  
-+      ATH_DISABLE_INTR();
++      local_irq_save(flags);
 +      if (sc->sc_isr & HAL_INT_TX) {
 +              sc->sc_isr &= ~HAL_INT_TX;
-+              ATH_ENABLE_INTR();
++              local_irq_restore(flags);
 +              goto process_tx_again;
 +      }
 +      sc->sc_imask |= HAL_INT_TX;
 +      ath_hal_intrset(sc->sc_ah, sc->sc_imask);
-+      ATH_ENABLE_INTR();
++      local_irq_restore(flags);
 +
        netif_wake_queue(dev);
  
        if (sc->sc_softled)
-@@ -8250,6 +8320,7 @@
+@@ -8355,13 +8497,25 @@ ath_tx_tasklet(TQUEUE_ARG data)
+       struct net_device *dev = (struct net_device *)data;
+       struct ath_softc *sc = dev->priv;
        unsigned int i;
++      unsigned long flags;
  
-       /* Process each active queue. */
+       /* Process each active queue. This includes sc_cabq, sc_xrtq and
+        * sc_uapsdq */
 +process_tx_again:
        for (i = 0; i < HAL_NUM_TX_QUEUES; i++)
                if (ATH_TXQ_SETUP(sc, i) && txqactive(sc->sc_ah, i))
                        ath_tx_processq(sc, &sc->sc_txq[i]);
-@@ -8258,6 +8329,16 @@
-               ath_tx_processq(sc, sc->sc_xrtxq);
- #endif
  
-+      ATH_DISABLE_INTR();
++      local_irq_save(flags);
 +      if (sc->sc_isr & HAL_INT_TX) {
 +              sc->sc_isr &= ~HAL_INT_TX;
-+              ATH_ENABLE_INTR();
++              local_irq_restore(flags);
 +              goto process_tx_again;
 +      }
 +      sc->sc_imask |= HAL_INT_TX;
 +      ath_hal_intrset(sc->sc_ah, sc->sc_imask);
-+      ATH_ENABLE_INTR();
++      local_irq_restore(flags);
 +
        netif_wake_queue(dev);
  
        if (sc->sc_softled)
-@@ -8332,6 +8413,7 @@
- ath_draintxq(struct ath_softc *sc)
- {
-       struct ath_hal *ah = sc->sc_ah;
-+      int npend = 0;
-       unsigned int i;
-       /* XXX return value */
-@@ -10095,9 +10177,9 @@
+@@ -10296,9 +10450,9 @@ ath_change_mtu(struct net_device *dev, i
        dev->mtu = mtu;
        if ((dev->flags & IFF_RUNNING) && !sc->sc_invalid) {
                /* NB: the rx buffers may need to be reallocated */
 -              tasklet_disable(&sc->sc_rxtq);
-+              netif_poll_disable(dev);
++              ath_poll_disable(dev);
                error = ath_reset(dev);
 -              tasklet_enable(&sc->sc_rxtq);
-+              netif_poll_enable(dev);
++              ath_poll_enable(dev);
        }
        ATH_UNLOCK(sc);
  
-Index: madwifi-dfs-r3053/ath/if_athvar.h
-===================================================================
---- madwifi-dfs-r3053.orig/ath/if_athvar.h     2007-12-13 05:25:12.842357313 +0100
-+++ madwifi-dfs-r3053/ath/if_athvar.h  2007-12-13 05:25:13.666404270 +0100
-@@ -50,6 +50,10 @@
- #include <asm/io.h>
- #include <linux/list.h>
+--- a/ath/if_athvar.h
++++ b/ath/if_athvar.h
+@@ -53,6 +53,10 @@
+ # include     <asm/bitops.h>
+ #endif
  
 +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
 +#define irqs_disabled()                       0
@@ -278,7 +391,17 @@ Index: madwifi-dfs-r3053/ath/if_athvar.h
  /*
   * Deduce if tasklets are available.  If not then
   * fall back to using the immediate work queue.
-@@ -728,7 +732,6 @@
+@@ -616,6 +620,9 @@ struct ath_rp {
+ struct ath_softc {
+       struct ieee80211com sc_ic;              /* NB: must be first */
+       struct net_device *sc_dev;
++#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24)
++      struct napi_struct sc_napi;
++#endif
+       void __iomem *sc_iobase;                /* address of the device */
+       struct semaphore sc_lock;               /* dev-level lock */
+       struct net_device_stats sc_devstats;    /* device statistics */
+@@ -730,7 +737,6 @@ struct ath_softc {
        struct ath_buf *sc_rxbufcur;            /* current rx buffer */
        u_int32_t *sc_rxlink;                   /* link ptr in last RX desc */
        spinlock_t sc_rxbuflock;
@@ -286,7 +409,7 @@ Index: madwifi-dfs-r3053/ath/if_athvar.h
        struct ATH_TQ_STRUCT sc_rxorntq;        /* rxorn intr tasklet */
        u_int8_t sc_defant;                     /* current default antenna */
        u_int8_t sc_rxotherant;                 /* RXs on non-default antenna */
-@@ -741,6 +744,7 @@
+@@ -745,6 +751,7 @@ struct ath_softc {
        u_int sc_txintrperiod;                  /* tx interrupt batching */
        struct ath_txq sc_txq[HAL_NUM_TX_QUEUES];
        struct ath_txq *sc_ac2q[WME_NUM_AC];    /* WME AC -> h/w qnum */
@@ -294,7 +417,7 @@ Index: madwifi-dfs-r3053/ath/if_athvar.h
        struct ATH_TQ_STRUCT sc_txtq;           /* tx intr tasklet */
        u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN];
        struct ath_descdma sc_bdma;             /* beacon descriptors */
-@@ -853,6 +857,8 @@
+@@ -858,6 +865,8 @@ typedef void (*ath_callback) (struct ath
  #define       ATH_TXBUF_LOCK_CHECK(_sc)
  #endif
  
@@ -303,51 +426,111 @@ Index: madwifi-dfs-r3053/ath/if_athvar.h
  
  #define       ATH_RXBUF_LOCK_INIT(_sc)        spin_lock_init(&(_sc)->sc_rxbuflock)
  #define       ATH_RXBUF_LOCK_DESTROY(_sc)
-Index: madwifi-dfs-r3053/net80211/ieee80211_input.c
-===================================================================
---- madwifi-dfs-r3053.orig/net80211/ieee80211_input.c  2007-12-13 05:25:10.374216660 +0100
-+++ madwifi-dfs-r3053/net80211/ieee80211_input.c       2007-12-13 05:25:13.674404726 +0100
-@@ -1178,7 +1178,7 @@
+--- a/net80211/ieee80211_input.c
++++ b/net80211/ieee80211_input.c
+@@ -1198,7 +1198,7 @@ ieee80211_deliver_data(struct ieee80211_
                        /* attach vlan tag */
                        struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni;
                        if (vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan) == NET_RX_DROP) {
 -                              /* If netif_rx dropped the packet because 
-+                              /* If netif_receive_skb dropped the packet because 
++                              /* If netif_receive_skb dropped the packet because
                                 * device was too busy */
                                if (ni_tmp != NULL) {
                                        /* node reference was leaked */
-@@ -1189,8 +1189,8 @@
+@@ -1209,8 +1209,8 @@ ieee80211_deliver_data(struct ieee80211_
                        skb = NULL; /* SKB is no longer ours */
                } else {
                        struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni;
 -                      if (netif_rx(skb) == NET_RX_DROP) {
 -                              /* If netif_rx dropped the packet because 
 +                      if (netif_receive_skb(skb) == NET_RX_DROP) {
-+                              /* If netif_receive_skb dropped the packet because 
++                              /* If netif_receive_skb dropped the packet because
                                 * device was too busy */
                                if (ni_tmp != NULL) {
                                        /* node reference was leaked */
-@@ -2299,8 +2299,8 @@
+@@ -2322,8 +2322,8 @@ forward_mgmt_to_app(struct ieee80211vap
                skb1->protocol = __constant_htons(0x0019);  /* ETH_P_80211_RAW */
  
                ni_tmp = SKB_CB(skb1)->ni;
 -              if (netif_rx(skb1) == NET_RX_DROP) {
 -                      /* If netif_rx dropped the packet because 
 +              if (netif_receive_skb(skb1) == NET_RX_DROP) {
-+                      /* If netif_receive_skb dropped the packet because 
++                      /* If netif_receive_skb dropped the packet because
                         * device was too busy */
                        if (ni_tmp != NULL) {
                                /* node reference was leaked */
-Index: madwifi-dfs-r3053/net80211/ieee80211_monitor.c
-===================================================================
---- madwifi-dfs-r3053.orig/net80211/ieee80211_monitor.c        2007-12-13 05:25:10.386217344 +0100
-+++ madwifi-dfs-r3053/net80211/ieee80211_monitor.c     2007-12-13 05:25:13.678404955 +0100
-@@ -577,7 +577,7 @@
-                       skb1->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */
+--- a/net80211/ieee80211_monitor.c
++++ b/net80211/ieee80211_monitor.c
+@@ -584,8 +584,8 @@ ieee80211_input_monitor(struct ieee80211
+                       skb1->protocol = 
+                               __constant_htons(0x0019); /* ETH_P_80211_RAW */
  
-                       ni_tmp = SKB_CB(skb1)->ni;
 -                      if (netif_rx(skb1) == NET_RX_DROP) {
+-                              /* If netif_rx dropped the packet because 
 +                      if (netif_receive_skb(skb1) == NET_RX_DROP) {
-                               /* If netif_rx dropped the packet because 
-                                * device was too busy */
-                               if (ni_tmp != NULL) {
++                              /* If netif_receive_skb dropped the packet because
+                                * device was too busy, reclaim the ref. in 
+                                * the skb. */
+                               if (SKB_CB(skb1)->ni != NULL)
+--- a/net80211/ieee80211_skb.c
++++ b/net80211/ieee80211_skb.c
+@@ -73,7 +73,7 @@
+ #undef dev_queue_xmit
+ #undef kfree_skb
+ #undef kfree_skb_fast
+-#undef netif_rx
++#undef netif_receive_skb
+ #undef pskb_copy
+ #undef skb_clone
+ #undef skb_copy
+@@ -638,8 +638,8 @@ int  vlan_hwaccel_receive_skb_debug(stru
+               grp, vlan_tag);
+ }
+-int netif_rx_debug(struct sk_buff *skb, const char* func, int line) {
+-      return netif_rx(untrack_skb(skb, 0, func, line, __func__, __LINE__));
++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line) {
++      return netif_receive_skb(untrack_skb(skb, 0, func, line, __func__, __LINE__));
+ }
+ struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask,
+@@ -760,7 +760,7 @@ struct sk_buff * skb_copy_expand_debug(c
+ }
+ EXPORT_SYMBOL(vlan_hwaccel_receive_skb_debug);
+-EXPORT_SYMBOL(netif_rx_debug);
++EXPORT_SYMBOL(netif_receive_skb_debug);
+ EXPORT_SYMBOL(alloc_skb_debug);
+ EXPORT_SYMBOL(dev_alloc_skb_debug);
+ EXPORT_SYMBOL(skb_clone_debug);
+--- a/net80211/ieee80211_skb.h
++++ b/net80211/ieee80211_skb.h
+@@ -116,7 +116,7 @@ int ieee80211_skb_references(void);
+ int  vlan_hwaccel_receive_skb_debug(struct sk_buff *skb, 
+                                   struct vlan_group *grp, unsigned short vlan_tag, 
+                                   const char* func, int line);
+-int netif_rx_debug(struct sk_buff *skb, const char* func, int line);
++int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line);
+ struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask,
+                                const char *func, int line);
+ struct sk_buff * dev_alloc_skb_debug(unsigned int length,
+@@ -151,7 +151,7 @@ struct sk_buff * skb_copy_expand_debug(c
+ #undef dev_queue_xmit
+ #undef kfree_skb
+ #undef kfree_skb_fast
+-#undef netif_rx
++#undef netif_receive_skb
+ #undef pskb_copy
+ #undef skb_clone
+ #undef skb_copy
+@@ -168,8 +168,8 @@ struct sk_buff * skb_copy_expand_debug(c
+       skb_copy_expand_debug(_skb, _newheadroom, _newtailroom, _gfp_mask, __func__, __LINE__)
+ #define vlan_hwaccel_receive_skb(_skb, _grp, _tag) \
+       vlan_hwaccel_receive_skb_debug(_skb, _grp, _tag, __func__, __LINE__)
+-#define netif_rx(_skb) \
+-      netif_rx_debug(_skb, __func__, __LINE__)
++#define netif_receive_skb(_skb) \
++      netif_receive_skb_debug(_skb, __func__, __LINE__)
+ #define       alloc_skb(_length, _gfp_mask) \
+       alloc_skb_debug(_length, _gfp_mask, __func__, __LINE__)
+ #define       dev_alloc_skb(_length) \
This page took 0.051894 seconds and 4 git commands to generate.