}
}
+static int ag71xx_rx_reserve(struct ag71xx *ag)
+{
+ int reserve = 0;
+
+ if (ag71xx_get_pdata(ag)->is_ar724x) {
+ if (!ag71xx_has_ar8216(ag))
+ reserve = 2;
+
+ if (ag->phy_dev)
+ reserve += 4 - (ag->phy_dev->pkt_align % 4);
+
+ reserve %= 4;
+ }
+
+ return reserve + AG71XX_RX_PKT_RESERVE;
+}
+
+
static int ag71xx_ring_rx_init(struct ag71xx *ag)
{
struct ag71xx_ring *ring = &ag->rx_ring;
+ unsigned int reserve = ag71xx_rx_reserve(ag);
unsigned int i;
int ret;
struct sk_buff *skb;
dma_addr_t dma_addr;
- skb = dev_alloc_skb(AG71XX_RX_PKT_SIZE + AG71XX_RX_PKT_RESERVE);
+ skb = dev_alloc_skb(AG71XX_RX_PKT_SIZE + reserve);
if (!skb) {
ret = -ENOMEM;
break;
}
skb->dev = ag->dev;
- skb_reserve(skb, AG71XX_RX_PKT_RESERVE);
+ skb_reserve(skb, reserve);
dma_addr = dma_map_single(&ag->dev->dev, skb->data,
AG71XX_RX_PKT_SIZE,
static int ag71xx_ring_rx_refill(struct ag71xx *ag)
{
struct ag71xx_ring *ring = &ag->rx_ring;
+ unsigned int reserve = ag71xx_rx_reserve(ag);
unsigned int count;
count = 0;
dma_addr_t dma_addr;
struct sk_buff *skb;
- skb = dev_alloc_skb(AG71XX_RX_PKT_SIZE +
- AG71XX_RX_PKT_RESERVE);
+ skb = dev_alloc_skb(AG71XX_RX_PKT_SIZE + reserve);
if (skb == NULL)
break;
- skb_reserve(skb, AG71XX_RX_PKT_RESERVE);
+ skb_reserve(skb, reserve);
skb->dev = ag->dev;
dma_addr = dma_map_single(&ag->dev->dev, skb->data,
dev->stats.rx_packets++;
dev->stats.rx_bytes += pktlen;
+ skb_put(skb, pktlen);
if (ag71xx_has_ar8216(ag))
- err = ag71xx_remove_ar8216_header(ag, skb);
+ err = ag71xx_remove_ar8216_header(ag, skb, pktlen);
if (err) {
dev->stats.rx_dropped++;
kfree_skb(skb);
} else {
- skb_put(skb, pktlen);
-
skb->dev = dev;
skb->ip_summed = CHECKSUM_NONE;
- skb->protocol = eth_type_trans(skb, dev);
- netif_receive_skb(skb);
+ if (ag->phy_dev) {
+ ag->phy_dev->netif_receive_skb(skb);
+ } else {
+ skb->protocol = eth_type_trans(skb, dev);
+ netif_receive_skb(skb);
+ }
}
ring->buf[i].skb = NULL;