Domen wrote: > > use your platform's dma mapping functions, rather than virt_to_phys() > > > > it might be the exact same implementation, inside the platform > > internals, but drivers should not be using this directly. > > I've replaced this with dma_map_single(), unmatched with > dma_unmap_single(), since bestcomm doesn't have a way to do that > and it's blank on ppc32 anyway. > > Is this OK? PPC guys?
Even though dma_unmap_single() may be a no-op, calls to dma_map_single() must be matched with calls to dma_unmap_single(). Perhaps with the additions below: > +static void mpc52xx_fec_free_rx_buffers(struct bcom_task *s) > +{ > + struct sk_buff *skb; > + > + while (!bcom_queue_empty(s)) { > + skb = bcom_retrieve_buffer(s, NULL, NULL); dma_unmap_single(&skb->dev->dev, skb-data, FEC_RX_BUFFER_SIZE, DMA_FROM_DEVICE); > + kfree_skb(skb); > + } > +} > + > +static int mpc52xx_fec_alloc_rx_buffers(struct net_device *dev, struct > bcom_task *rxtsk) > +{ > + while (!bcom_queue_full(rxtsk)) { > + struct sk_buff *skb; > + struct bcom_fec_bd *bd; > + > + skb = dev_alloc_skb(FEC_RX_BUFFER_SIZE); > + if (skb == NULL) > + return -EAGAIN; skb->dev = dev; > + > + /* zero out the initial receive buffers to aid debugging */ > + memset(skb->data, 0, FEC_RX_BUFFER_SIZE); > + > + bd = (struct bcom_fec_bd *)bcom_prepare_next_buffer(rxtsk); > + > + bd->status = FEC_RX_BUFFER_SIZE; > + bd->skb_pa = dma_map_single(&dev->dev, skb->data, > + FEC_RX_BUFFER_SIZE, DMA_FROM_DEVICE); > + > + bcom_submit_next_buffer(rxtsk, skb); > + } > + > + return 0; > +} [...] > +static int mpc52xx_fec_hard_start_xmit(struct sk_buff *skb, struct > net_device *dev) > +{ > + struct mpc52xx_fec_priv *priv = netdev_priv(dev); > + struct bcom_fec_bd *bd; > + > + if (bcom_queue_full(priv->tx_dmatsk)) { > + if (net_ratelimit()) > + dev_err(&dev->dev, "transmit queue overrun\n"); > + return 1; > + } > + > + spin_lock_irq(&priv->lock); > + dev->trans_start = jiffies; > + > + bd = (struct bcom_fec_bd *) > + bcom_prepare_next_buffer(priv->tx_dmatsk); > + > + bd->status = skb->len | BCOM_FEC_TX_BD_TFD | BCOM_FEC_TX_BD_TC; > + bd->skb_pa = dma_map_single(&dev->dev, skb->data, skb->len, > DMA_TO_DEVICE); > + > + bcom_submit_next_buffer(priv->tx_dmatsk, skb); > + > + if (bcom_queue_full(priv->tx_dmatsk)) { > + netif_stop_queue(dev); > + } > + > + spin_unlock_irq(&priv->lock); > + > + return 0; > +} > + > +/* This handles BestComm transmit task interrupts > + */ > +static irqreturn_t mpc52xx_fec_tx_interrupt(int irq, void *dev_id) > +{ > + struct net_device *dev = dev_id; > + struct mpc52xx_fec_priv *priv = netdev_priv(dev); > + > + spin_lock(&priv->lock); > + > + while (bcom_buffer_done(priv->tx_dmatsk)) { > + struct sk_buff *skb; > + skb = bcom_retrieve_buffer(priv->tx_dmatsk, NULL, NULL); > + /* Here (and in rx routines) would be a good place for > + * dma_unmap_single(), but bcom doesn't return bcom_bd of the > + * finished transfer, and _unmap is empty on this platfrom. > + */ Replace the above comment with: dma_unmap_single(&dev->dev, skb->data, skb->len, DMA_TO_DEVICE); > + > + dev_kfree_skb_irq(skb); > + } > + > + netif_wake_queue(dev); > + > + spin_unlock(&priv->lock); > + > + return IRQ_HANDLED; > +} > + > +static irqreturn_t mpc52xx_fec_rx_interrupt(int irq, void *dev_id) > +{ > + struct net_device *dev = dev_id; > + struct mpc52xx_fec_priv *priv = netdev_priv(dev); > + > + while (bcom_buffer_done(priv->rx_dmatsk)) { > + struct sk_buff *skb; > + struct sk_buff *rskb; > + struct bcom_fec_bd *bd; > + u32 status; > + > + rskb = bcom_retrieve_buffer(priv->rx_dmatsk, &status, NULL); dma_unmap_single(&dev->dev, rskb->data, FEC_RX_BUFFER_SIZE, DMA_FROM_DEVICE); > + > + /* Test for errors in received frame */ > + if (status & BCOM_FEC_RX_BD_ERRORS) { > + /* Drop packet and reuse the buffer */ > + bd = (struct bcom_fec_bd *) > + bcom_prepare_next_buffer(priv->rx_dmatsk); > + > + bd->status = FEC_RX_BUFFER_SIZE; > + bd->skb_pa = dma_map_single(&dev->dev, rskb->data, > + FEC_RX_BUFFER_SIZE, DMA_FROM_DEVICE); > + > + bcom_submit_next_buffer(priv->rx_dmatsk, rskb); > + > + dev->stats.rx_dropped++; > + > + continue; > + } > + > + /* skbs are allocated on open, so now we allocate a new one, > + * and remove the old (with the packet) */ > + skb = dev_alloc_skb(FEC_RX_BUFFER_SIZE); > + if (skb) { > + /* Process the received skb */ > + int length = status & BCOM_FEC_RX_BD_LEN_MASK; skb->dev = dev; > + > + skb_put(rskb, length - 4); /* length without CRC32 > */ > + > + rskb->dev = dev; Above line is no longer needed since we set rskb->dev on skb allocation. > + rskb->protocol = eth_type_trans(rskb, dev); > + > + netif_rx(rskb); > + dev->last_rx = jiffies; > + } else { > + /* Can't get a new one : reuse the same & drop pkt */ > + dev_notice(&dev->dev, "Memory squeeze, dropping > packet.\n"); > + dev->stats.rx_dropped++; > + > + skb = rskb; > + } > + > + bd = (struct bcom_fec_bd *) > + bcom_prepare_next_buffer(priv->rx_dmatsk); > + > + bd->status = FEC_RX_BUFFER_SIZE; > + bd->skb_pa = dma_map_single(&dev->dev, rskb->data, > + FEC_RX_BUFFER_SIZE, DMA_FROM_DEVICE); > + > + bcom_submit_next_buffer(priv->rx_dmatsk, skb); > + } > + > + return IRQ_HANDLED; > +} -Dale Farnsworth _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@ozlabs.org https://ozlabs.org/mailman/listinfo/linuxppc-dev