mirror of
https://github.com/openwrt/openwrt.git
synced 2024-12-21 06:33:41 +00:00
24d73fbd33
This adds some more bgmac patches from mainline kernel. Signed-off-by: Hauke Mehrtens <hauke@hauke-m.de> SVN-Revision: 38637
373 lines
12 KiB
Diff
373 lines
12 KiB
Diff
patches for bgmac backported from net-next/master
|
|
|
|
--- a/drivers/net/ethernet/broadcom/Kconfig
|
|
+++ b/drivers/net/ethernet/broadcom/Kconfig
|
|
@@ -132,7 +132,8 @@ config BNX2X_SRIOV
|
|
|
|
config BGMAC
|
|
tristate "BCMA bus GBit core support"
|
|
- depends on BCMA_HOST_SOC && HAS_DMA
|
|
+ depends on BCMA_HOST_SOC && HAS_DMA && BCM47XX
|
|
+ select PHYLIB
|
|
---help---
|
|
This driver supports GBit MAC and BCM4706 GBit MAC cores on BCMA bus.
|
|
They can be found on BCM47xx SoCs and provide gigabit ethernet.
|
|
--- a/drivers/net/ethernet/broadcom/bgmac.c
|
|
+++ b/drivers/net/ethernet/broadcom/bgmac.c
|
|
@@ -149,6 +149,8 @@ static netdev_tx_t bgmac_dma_tx_add(stru
|
|
dma_desc->ctl0 = cpu_to_le32(ctl0);
|
|
dma_desc->ctl1 = cpu_to_le32(ctl1);
|
|
|
|
+ netdev_sent_queue(net_dev, skb->len);
|
|
+
|
|
wmb();
|
|
|
|
/* Increase ring->end to point empty slot. We tell hardware the first
|
|
@@ -157,6 +159,7 @@ static netdev_tx_t bgmac_dma_tx_add(stru
|
|
if (++ring->end >= BGMAC_TX_RING_SLOTS)
|
|
ring->end = 0;
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_TX_INDEX,
|
|
+ ring->index_base +
|
|
ring->end * sizeof(struct bgmac_dma_desc));
|
|
|
|
/* Always keep one slot free to allow detecting bugged calls. */
|
|
@@ -177,10 +180,13 @@ static void bgmac_dma_tx_free(struct bgm
|
|
struct device *dma_dev = bgmac->core->dma_dev;
|
|
int empty_slot;
|
|
bool freed = false;
|
|
+ unsigned bytes_compl = 0, pkts_compl = 0;
|
|
|
|
/* The last slot that hardware didn't consume yet */
|
|
empty_slot = bgmac_read(bgmac, ring->mmio_base + BGMAC_DMA_TX_STATUS);
|
|
empty_slot &= BGMAC_DMA_TX_STATDPTR;
|
|
+ empty_slot -= ring->index_base;
|
|
+ empty_slot &= BGMAC_DMA_TX_STATDPTR;
|
|
empty_slot /= sizeof(struct bgmac_dma_desc);
|
|
|
|
while (ring->start != empty_slot) {
|
|
@@ -192,6 +198,9 @@ static void bgmac_dma_tx_free(struct bgm
|
|
slot->skb->len, DMA_TO_DEVICE);
|
|
slot->dma_addr = 0;
|
|
|
|
+ bytes_compl += slot->skb->len;
|
|
+ pkts_compl++;
|
|
+
|
|
/* Free memory! :) */
|
|
dev_kfree_skb(slot->skb);
|
|
slot->skb = NULL;
|
|
@@ -205,6 +214,8 @@ static void bgmac_dma_tx_free(struct bgm
|
|
freed = true;
|
|
}
|
|
|
|
+ netdev_completed_queue(bgmac->net_dev, pkts_compl, bytes_compl);
|
|
+
|
|
if (freed && netif_queue_stopped(bgmac->net_dev))
|
|
netif_wake_queue(bgmac->net_dev);
|
|
}
|
|
@@ -241,31 +252,59 @@ static int bgmac_dma_rx_skb_for_slot(str
|
|
struct bgmac_slot_info *slot)
|
|
{
|
|
struct device *dma_dev = bgmac->core->dma_dev;
|
|
+ struct sk_buff *skb;
|
|
+ dma_addr_t dma_addr;
|
|
struct bgmac_rx_header *rx;
|
|
|
|
/* Alloc skb */
|
|
- slot->skb = netdev_alloc_skb(bgmac->net_dev, BGMAC_RX_BUF_SIZE);
|
|
- if (!slot->skb)
|
|
+ skb = netdev_alloc_skb(bgmac->net_dev, BGMAC_RX_BUF_SIZE);
|
|
+ if (!skb)
|
|
return -ENOMEM;
|
|
|
|
/* Poison - if everything goes fine, hardware will overwrite it */
|
|
- rx = (struct bgmac_rx_header *)slot->skb->data;
|
|
+ rx = (struct bgmac_rx_header *)skb->data;
|
|
rx->len = cpu_to_le16(0xdead);
|
|
rx->flags = cpu_to_le16(0xbeef);
|
|
|
|
/* Map skb for the DMA */
|
|
- slot->dma_addr = dma_map_single(dma_dev, slot->skb->data,
|
|
- BGMAC_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
|
- if (dma_mapping_error(dma_dev, slot->dma_addr)) {
|
|
+ dma_addr = dma_map_single(dma_dev, skb->data,
|
|
+ BGMAC_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
|
+ if (dma_mapping_error(dma_dev, dma_addr)) {
|
|
bgmac_err(bgmac, "DMA mapping error\n");
|
|
+ dev_kfree_skb(skb);
|
|
return -ENOMEM;
|
|
}
|
|
+
|
|
+ /* Update the slot */
|
|
+ slot->skb = skb;
|
|
+ slot->dma_addr = dma_addr;
|
|
+
|
|
if (slot->dma_addr & 0xC0000000)
|
|
bgmac_warn(bgmac, "DMA address using 0xC0000000 bit(s), it may need translation trick\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
+static void bgmac_dma_rx_setup_desc(struct bgmac *bgmac,
|
|
+ struct bgmac_dma_ring *ring, int desc_idx)
|
|
+{
|
|
+ struct bgmac_dma_desc *dma_desc = ring->cpu_base + desc_idx;
|
|
+ u32 ctl0 = 0, ctl1 = 0;
|
|
+
|
|
+ if (desc_idx == ring->num_slots - 1)
|
|
+ ctl0 |= BGMAC_DESC_CTL0_EOT;
|
|
+ ctl1 |= BGMAC_RX_BUF_SIZE & BGMAC_DESC_CTL1_LEN;
|
|
+ /* Is there any BGMAC device that requires extension? */
|
|
+ /* ctl1 |= (addrext << B43_DMA64_DCTL1_ADDREXT_SHIFT) &
|
|
+ * B43_DMA64_DCTL1_ADDREXT_MASK;
|
|
+ */
|
|
+
|
|
+ dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[desc_idx].dma_addr));
|
|
+ dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[desc_idx].dma_addr));
|
|
+ dma_desc->ctl0 = cpu_to_le32(ctl0);
|
|
+ dma_desc->ctl1 = cpu_to_le32(ctl1);
|
|
+}
|
|
+
|
|
static int bgmac_dma_rx_read(struct bgmac *bgmac, struct bgmac_dma_ring *ring,
|
|
int weight)
|
|
{
|
|
@@ -274,6 +313,8 @@ static int bgmac_dma_rx_read(struct bgma
|
|
|
|
end_slot = bgmac_read(bgmac, ring->mmio_base + BGMAC_DMA_RX_STATUS);
|
|
end_slot &= BGMAC_DMA_RX_STATDPTR;
|
|
+ end_slot -= ring->index_base;
|
|
+ end_slot &= BGMAC_DMA_RX_STATDPTR;
|
|
end_slot /= sizeof(struct bgmac_dma_desc);
|
|
|
|
ring->end = end_slot;
|
|
@@ -282,7 +323,6 @@ static int bgmac_dma_rx_read(struct bgma
|
|
struct device *dma_dev = bgmac->core->dma_dev;
|
|
struct bgmac_slot_info *slot = &ring->slots[ring->start];
|
|
struct sk_buff *skb = slot->skb;
|
|
- struct sk_buff *new_skb;
|
|
struct bgmac_rx_header *rx;
|
|
u16 len, flags;
|
|
|
|
@@ -295,38 +335,51 @@ static int bgmac_dma_rx_read(struct bgma
|
|
len = le16_to_cpu(rx->len);
|
|
flags = le16_to_cpu(rx->flags);
|
|
|
|
- /* Check for poison and drop or pass the packet */
|
|
- if (len == 0xdead && flags == 0xbeef) {
|
|
- bgmac_err(bgmac, "Found poisoned packet at slot %d, DMA issue!\n",
|
|
- ring->start);
|
|
- } else {
|
|
+ do {
|
|
+ dma_addr_t old_dma_addr = slot->dma_addr;
|
|
+ int err;
|
|
+
|
|
+ /* Check for poison and drop or pass the packet */
|
|
+ if (len == 0xdead && flags == 0xbeef) {
|
|
+ bgmac_err(bgmac, "Found poisoned packet at slot %d, DMA issue!\n",
|
|
+ ring->start);
|
|
+ dma_sync_single_for_device(dma_dev,
|
|
+ slot->dma_addr,
|
|
+ BGMAC_RX_BUF_SIZE,
|
|
+ DMA_FROM_DEVICE);
|
|
+ break;
|
|
+ }
|
|
+
|
|
/* Omit CRC. */
|
|
len -= ETH_FCS_LEN;
|
|
|
|
- new_skb = netdev_alloc_skb_ip_align(bgmac->net_dev, len);
|
|
- if (new_skb) {
|
|
- skb_put(new_skb, len);
|
|
- skb_copy_from_linear_data_offset(skb, BGMAC_RX_FRAME_OFFSET,
|
|
- new_skb->data,
|
|
- len);
|
|
- skb_checksum_none_assert(skb);
|
|
- new_skb->protocol =
|
|
- eth_type_trans(new_skb, bgmac->net_dev);
|
|
- netif_receive_skb(new_skb);
|
|
- handled++;
|
|
- } else {
|
|
- bgmac->net_dev->stats.rx_dropped++;
|
|
- bgmac_err(bgmac, "Allocation of skb for copying packet failed!\n");
|
|
+ /* Prepare new skb as replacement */
|
|
+ err = bgmac_dma_rx_skb_for_slot(bgmac, slot);
|
|
+ if (err) {
|
|
+ /* Poison the old skb */
|
|
+ rx->len = cpu_to_le16(0xdead);
|
|
+ rx->flags = cpu_to_le16(0xbeef);
|
|
+
|
|
+ dma_sync_single_for_device(dma_dev,
|
|
+ slot->dma_addr,
|
|
+ BGMAC_RX_BUF_SIZE,
|
|
+ DMA_FROM_DEVICE);
|
|
+ break;
|
|
}
|
|
+ bgmac_dma_rx_setup_desc(bgmac, ring, ring->start);
|
|
|
|
- /* Poison the old skb */
|
|
- rx->len = cpu_to_le16(0xdead);
|
|
- rx->flags = cpu_to_le16(0xbeef);
|
|
- }
|
|
-
|
|
- /* Make it back accessible to the hardware */
|
|
- dma_sync_single_for_device(dma_dev, slot->dma_addr,
|
|
- BGMAC_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
|
+ /* Unmap old skb, we'll pass it to the netfif */
|
|
+ dma_unmap_single(dma_dev, old_dma_addr,
|
|
+ BGMAC_RX_BUF_SIZE, DMA_FROM_DEVICE);
|
|
+
|
|
+ skb_put(skb, BGMAC_RX_FRAME_OFFSET + len);
|
|
+ skb_pull(skb, BGMAC_RX_FRAME_OFFSET);
|
|
+
|
|
+ skb_checksum_none_assert(skb);
|
|
+ skb->protocol = eth_type_trans(skb, bgmac->net_dev);
|
|
+ netif_receive_skb(skb);
|
|
+ handled++;
|
|
+ } while (0);
|
|
|
|
if (++ring->start >= BGMAC_RX_RING_SLOTS)
|
|
ring->start = 0;
|
|
@@ -418,9 +471,6 @@ static int bgmac_dma_alloc(struct bgmac
|
|
ring = &bgmac->tx_ring[i];
|
|
ring->num_slots = BGMAC_TX_RING_SLOTS;
|
|
ring->mmio_base = ring_base[i];
|
|
- if (bgmac_dma_unaligned(bgmac, ring, BGMAC_DMA_RING_TX))
|
|
- bgmac_warn(bgmac, "TX on ring 0x%X supports unaligned addressing but this feature is not implemented\n",
|
|
- ring->mmio_base);
|
|
|
|
/* Alloc ring of descriptors */
|
|
size = ring->num_slots * sizeof(struct bgmac_dma_desc);
|
|
@@ -435,6 +485,13 @@ static int bgmac_dma_alloc(struct bgmac
|
|
if (ring->dma_base & 0xC0000000)
|
|
bgmac_warn(bgmac, "DMA address using 0xC0000000 bit(s), it may need translation trick\n");
|
|
|
|
+ ring->unaligned = bgmac_dma_unaligned(bgmac, ring,
|
|
+ BGMAC_DMA_RING_TX);
|
|
+ if (ring->unaligned)
|
|
+ ring->index_base = lower_32_bits(ring->dma_base);
|
|
+ else
|
|
+ ring->index_base = 0;
|
|
+
|
|
/* No need to alloc TX slots yet */
|
|
}
|
|
|
|
@@ -444,9 +501,6 @@ static int bgmac_dma_alloc(struct bgmac
|
|
ring = &bgmac->rx_ring[i];
|
|
ring->num_slots = BGMAC_RX_RING_SLOTS;
|
|
ring->mmio_base = ring_base[i];
|
|
- if (bgmac_dma_unaligned(bgmac, ring, BGMAC_DMA_RING_RX))
|
|
- bgmac_warn(bgmac, "RX on ring 0x%X supports unaligned addressing but this feature is not implemented\n",
|
|
- ring->mmio_base);
|
|
|
|
/* Alloc ring of descriptors */
|
|
size = ring->num_slots * sizeof(struct bgmac_dma_desc);
|
|
@@ -462,6 +516,13 @@ static int bgmac_dma_alloc(struct bgmac
|
|
if (ring->dma_base & 0xC0000000)
|
|
bgmac_warn(bgmac, "DMA address using 0xC0000000 bit(s), it may need translation trick\n");
|
|
|
|
+ ring->unaligned = bgmac_dma_unaligned(bgmac, ring,
|
|
+ BGMAC_DMA_RING_RX);
|
|
+ if (ring->unaligned)
|
|
+ ring->index_base = lower_32_bits(ring->dma_base);
|
|
+ else
|
|
+ ring->index_base = 0;
|
|
+
|
|
/* Alloc RX slots */
|
|
for (j = 0; j < ring->num_slots; j++) {
|
|
err = bgmac_dma_rx_skb_for_slot(bgmac, &ring->slots[j]);
|
|
@@ -482,19 +543,19 @@ err_dma_free:
|
|
static void bgmac_dma_init(struct bgmac *bgmac)
|
|
{
|
|
struct bgmac_dma_ring *ring;
|
|
- struct bgmac_dma_desc *dma_desc;
|
|
- u32 ctl0, ctl1;
|
|
int i;
|
|
|
|
for (i = 0; i < BGMAC_MAX_TX_RINGS; i++) {
|
|
ring = &bgmac->tx_ring[i];
|
|
|
|
- /* We don't implement unaligned addressing, so enable first */
|
|
- bgmac_dma_tx_enable(bgmac, ring);
|
|
+ if (!ring->unaligned)
|
|
+ bgmac_dma_tx_enable(bgmac, ring);
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_TX_RINGLO,
|
|
lower_32_bits(ring->dma_base));
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_TX_RINGHI,
|
|
upper_32_bits(ring->dma_base));
|
|
+ if (ring->unaligned)
|
|
+ bgmac_dma_tx_enable(bgmac, ring);
|
|
|
|
ring->start = 0;
|
|
ring->end = 0; /* Points the slot that should *not* be read */
|
|
@@ -505,32 +566,20 @@ static void bgmac_dma_init(struct bgmac
|
|
|
|
ring = &bgmac->rx_ring[i];
|
|
|
|
- /* We don't implement unaligned addressing, so enable first */
|
|
- bgmac_dma_rx_enable(bgmac, ring);
|
|
+ if (!ring->unaligned)
|
|
+ bgmac_dma_rx_enable(bgmac, ring);
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_RX_RINGLO,
|
|
lower_32_bits(ring->dma_base));
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_RX_RINGHI,
|
|
upper_32_bits(ring->dma_base));
|
|
+ if (ring->unaligned)
|
|
+ bgmac_dma_rx_enable(bgmac, ring);
|
|
|
|
- for (j = 0, dma_desc = ring->cpu_base; j < ring->num_slots;
|
|
- j++, dma_desc++) {
|
|
- ctl0 = ctl1 = 0;
|
|
-
|
|
- if (j == ring->num_slots - 1)
|
|
- ctl0 |= BGMAC_DESC_CTL0_EOT;
|
|
- ctl1 |= BGMAC_RX_BUF_SIZE & BGMAC_DESC_CTL1_LEN;
|
|
- /* Is there any BGMAC device that requires extension? */
|
|
- /* ctl1 |= (addrext << B43_DMA64_DCTL1_ADDREXT_SHIFT) &
|
|
- * B43_DMA64_DCTL1_ADDREXT_MASK;
|
|
- */
|
|
-
|
|
- dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[j].dma_addr));
|
|
- dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[j].dma_addr));
|
|
- dma_desc->ctl0 = cpu_to_le32(ctl0);
|
|
- dma_desc->ctl1 = cpu_to_le32(ctl1);
|
|
- }
|
|
+ for (j = 0; j < ring->num_slots; j++)
|
|
+ bgmac_dma_rx_setup_desc(bgmac, ring, j);
|
|
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_RX_INDEX,
|
|
+ ring->index_base +
|
|
ring->num_slots * sizeof(struct bgmac_dma_desc));
|
|
|
|
ring->start = 0;
|
|
@@ -909,9 +958,9 @@ static void bgmac_chip_reset(struct bgma
|
|
u8 et_swtype = 0;
|
|
u8 sw_type = BGMAC_CHIPCTL_1_SW_TYPE_EPHY |
|
|
BGMAC_CHIPCTL_1_IF_TYPE_MII;
|
|
- char buf[2];
|
|
+ char buf[4];
|
|
|
|
- if (bcm47xx_nvram_getenv("et_swtype", buf, 1) > 0) {
|
|
+ if (bcm47xx_nvram_getenv("et_swtype", buf, sizeof(buf)) > 0) {
|
|
if (kstrtou8(buf, 0, &et_swtype))
|
|
bgmac_err(bgmac, "Failed to parse et_swtype (%s)\n",
|
|
buf);
|
|
@@ -970,6 +1019,8 @@ static void bgmac_chip_reset(struct bgma
|
|
bgmac_miiconfig(bgmac);
|
|
bgmac_phy_init(bgmac);
|
|
|
|
+ netdev_reset_queue(bgmac->net_dev);
|
|
+
|
|
bgmac->int_status = 0;
|
|
}
|
|
|
|
--- a/drivers/net/ethernet/broadcom/bgmac.h
|
|
+++ b/drivers/net/ethernet/broadcom/bgmac.h
|
|
@@ -384,6 +384,8 @@ struct bgmac_dma_ring {
|
|
u16 mmio_base;
|
|
struct bgmac_dma_desc *cpu_base;
|
|
dma_addr_t dma_base;
|
|
+ u32 index_base; /* Used for unaligned rings only, otherwise 0 */
|
|
+ bool unaligned;
|
|
|
|
struct bgmac_slot_info slots[BGMAC_RX_RING_SLOTS];
|
|
};
|