blob: fccc5f5a4403a92fca56b70091e048bb70d7464e [file] [log] [blame]
// SPDX-License-Identifier: GPL-2.0+
/*
* Copyright (c) 2020 Sartura Ltd.
*
* Author: Robert Marko <robert.marko@sartura.hr>
*
* Copyright (c) 2021 Toco Technologies FZE <contact@toco.ae>
* Copyright (c) 2021 Gabor Juhos <j4g8y7@gmail.com>
*
* Qualcomm ESS EDMA ethernet driver
*/
#include <asm/io.h>
#include <clk.h>
#include <cpu_func.h>
#include <dm.h>
#include <dm/device_compat.h>
#include <errno.h>
#include <linux/bitfield.h>
#include <linux/bitops.h>
#include <linux/delay.h>
#include <log.h>
#include <miiphy.h>
#include <net.h>
#include <reset.h>
#include "essedma.h"
#define EDMA_MAX_PKT_SIZE (PKTSIZE_ALIGN + PKTALIGN)
#define EDMA_RXQ_ID 0
#define EDMA_TXQ_ID 0
/* descriptor ring */
struct edma_ring {
u16 count; /* number of descriptors in the ring */
void *hw_desc; /* descriptor ring virtual address */
unsigned int hw_size; /* hw descriptor ring length in bytes */
dma_addr_t dma; /* descriptor ring physical address */
u16 head; /* next Tx descriptor to fill */
u16 tail; /* next Tx descriptor to clean */
};
struct ess_switch {
phys_addr_t base;
struct phy_device *phydev[ESS_PORTS_NUM];
u32 phy_mask;
ofnode ports_node;
phy_interface_t port_wrapper_mode;
int num_phy;
};
struct essedma_priv {
phys_addr_t base;
struct udevice *dev;
struct clk ess_clk;
struct reset_ctl ess_rst;
struct udevice *mdio_dev;
struct ess_switch esw;
phys_addr_t psgmii_base;
struct edma_ring tpd_ring;
struct edma_ring rfd_ring;
};
static void esw_port_loopback_set(struct ess_switch *esw, int port,
bool enable)
{
u32 t;
t = readl(esw->base + ESS_PORT_LOOKUP_CTRL(port));
if (enable)
t |= ESS_PORT_LOOP_BACK_EN;
else
t &= ~ESS_PORT_LOOP_BACK_EN;
writel(t, esw->base + ESS_PORT_LOOKUP_CTRL(port));
}
static void esw_port_loopback_set_all(struct ess_switch *esw, bool enable)
{
int i;
for (i = 1; i < ESS_PORTS_NUM; i++)
esw_port_loopback_set(esw, i, enable);
}
static void ess_reset(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
reset_assert(&priv->ess_rst);
mdelay(10);
reset_deassert(&priv->ess_rst);
mdelay(10);
}
void qca8075_ess_reset(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
struct phy_device *psgmii_phy;
int i, val;
/* Find the PSGMII PHY */
psgmii_phy = priv->esw.phydev[priv->esw.num_phy - 1];
/* Fix phy psgmii RX 20bit */
phy_write(psgmii_phy, MDIO_DEVAD_NONE, MII_BMCR, 0x005b);
/* Reset phy psgmii */
phy_write(psgmii_phy, MDIO_DEVAD_NONE, MII_BMCR, 0x001b);
/* Release reset phy psgmii */
phy_write(psgmii_phy, MDIO_DEVAD_NONE, MII_BMCR, 0x005b);
for (i = 0; i < 100; i++) {
val = phy_read_mmd(psgmii_phy, MDIO_MMD_PMAPMD, 0x28);
if (val & 0x1)
break;
mdelay(1);
}
if (i >= 100)
printf("QCA807x PSGMII PLL_VCO_CALIB Not Ready\n");
/*
* Check qca8075 psgmii calibration done end.
* Freeze phy psgmii RX CDR
*/
phy_write(psgmii_phy, MDIO_DEVAD_NONE, 0x1a, 0x2230);
ess_reset(dev);
/* Check ipq psgmii calibration done start */
for (i = 0; i < 100; i++) {
val = readl(priv->psgmii_base + PSGMIIPHY_VCO_CALIBRATION_CTRL_REGISTER_2);
if (val & 0x1)
break;
mdelay(1);
}
if (i >= 100)
printf("PSGMII PLL_VCO_CALIB Not Ready\n");
/*
* Check ipq psgmii calibration done end.
* Relesae phy psgmii RX CDR
*/
phy_write(psgmii_phy, MDIO_DEVAD_NONE, 0x1a, 0x3230);
/* Release phy psgmii RX 20bit */
phy_write(psgmii_phy, MDIO_DEVAD_NONE, MII_BMCR, 0x005f);
}
#define PSGMII_ST_NUM_RETRIES 20
#define PSGMII_ST_PKT_COUNT (4 * 1024)
#define PSGMII_ST_PKT_SIZE 1504
/*
* Transmitting one byte over a 1000Mbps link requires 8 ns.
* Additionally, use + 1 ns for safety to compensate latencies
* and such.
*/
#define PSGMII_ST_TRAFFIC_TIMEOUT_NS \
(PSGMII_ST_PKT_COUNT * PSGMII_ST_PKT_SIZE * (8 + 1))
#define PSGMII_ST_TRAFFIC_TIMEOUT \
DIV_ROUND_UP(PSGMII_ST_TRAFFIC_TIMEOUT_NS, 1000000)
static bool psgmii_self_test_repeat;
static void psgmii_st_phy_power_down(struct phy_device *phydev)
{
int val;
val = phy_read(phydev, MDIO_DEVAD_NONE, MII_BMCR);
val |= QCA807X_POWER_DOWN;
phy_write(phydev, MDIO_DEVAD_NONE, MII_BMCR, val);
}
static void psgmii_st_phy_prepare(struct phy_device *phydev)
{
int val;
/* check phydev combo port */
val = phy_read(phydev, MDIO_DEVAD_NONE,
QCA807X_CHIP_CONFIGURATION);
if (val) {
/* Select copper page */
val |= QCA807X_MEDIA_PAGE_SELECT;
phy_write(phydev, MDIO_DEVAD_NONE,
QCA807X_CHIP_CONFIGURATION, val);
}
/* Force no link by power down */
psgmii_st_phy_power_down(phydev);
/* Packet number (Non documented) */
phy_write_mmd(phydev, MDIO_MMD_AN, 0x8021, PSGMII_ST_PKT_COUNT);
phy_write_mmd(phydev, MDIO_MMD_AN, 0x8062, PSGMII_ST_PKT_SIZE);
/* Fix MDI status */
val = phy_read(phydev, MDIO_DEVAD_NONE, QCA807X_FUNCTION_CONTROL);
val &= ~QCA807X_MDI_CROSSOVER_MODE_MASK;
val |= FIELD_PREP(QCA807X_MDI_CROSSOVER_MODE_MASK,
QCA807X_MDI_CROSSOVER_MODE_MANUAL_MDI);
val &= ~QCA807X_POLARITY_REVERSAL;
phy_write(phydev, MDIO_DEVAD_NONE, QCA807X_FUNCTION_CONTROL, val);
}
static void psgmii_st_phy_recover(struct phy_device *phydev)
{
int val;
/* Packet number (Non documented) */
phy_write_mmd(phydev, MDIO_MMD_AN, 0x8021, 0x0);
/* Disable CRC checker and packet counter */
val = phy_read_mmd(phydev, MDIO_MMD_AN, QCA807X_MMD7_CRC_PACKET_COUNTER);
val &= ~QCA807X_MMD7_PACKET_COUNTER_SELFCLR;
val &= ~QCA807X_MMD7_CRC_PACKET_COUNTER_EN;
phy_write_mmd(phydev, MDIO_MMD_AN, QCA807X_MMD7_CRC_PACKET_COUNTER, val);
/* Disable traffic (Undocumented) */
phy_write_mmd(phydev, MDIO_MMD_AN, 0x8020, 0x0);
}
static void psgmii_st_phy_start_traffic(struct phy_device *phydev)
{
int val;
/* Enable CRC checker and packet counter */
val = phy_read_mmd(phydev, MDIO_MMD_AN, QCA807X_MMD7_CRC_PACKET_COUNTER);
val |= QCA807X_MMD7_CRC_PACKET_COUNTER_EN;
phy_write_mmd(phydev, MDIO_MMD_AN, QCA807X_MMD7_CRC_PACKET_COUNTER, val);
/* Start traffic (Undocumented) */
phy_write_mmd(phydev, MDIO_MMD_AN, 0x8020, 0xa000);
}
static bool psgmii_st_phy_check_counters(struct phy_device *phydev)
{
u32 tx_ok;
/*
* The number of test packets is limited to 65535 so
* only read the lower 16 bits of the counter.
*/
tx_ok = phy_read_mmd(phydev, MDIO_MMD_AN,
QCA807X_MMD7_VALID_EGRESS_COUNTER_2);
return (tx_ok == PSGMII_ST_PKT_COUNT);
}
static void psgmii_st_phy_reset_loopback(struct phy_device *phydev)
{
/* reset the PHY */
phy_write(phydev, MDIO_DEVAD_NONE, MII_BMCR, 0x9000);
/* enable loopback mode */
phy_write(phydev, MDIO_DEVAD_NONE, MII_BMCR, 0x4140);
}
static inline bool psgmii_st_phy_link_is_up(struct phy_device *phydev)
{
int val;
val = phy_read(phydev, MDIO_DEVAD_NONE, QCA807X_PHY_SPECIFIC);
return !!(val & QCA807X_PHY_SPECIFIC_LINK);
}
static bool psgmii_st_phy_wait(struct ess_switch *esw, u32 mask,
int retries, int delay,
bool (*check)(struct phy_device *))
{
int i;
for (i = 0; i < retries; i++) {
int phy;
for (phy = 0; phy < esw->num_phy - 1; phy++) {
u32 phybit = BIT(phy);
if (!(mask & phybit))
continue;
if (check(esw->phydev[phy]))
mask &= ~phybit;
}
if (!mask)
break;
mdelay(delay);
}
return (!mask);
}
static bool psgmii_st_phy_wait_link(struct ess_switch *esw, u32 mask)
{
return psgmii_st_phy_wait(esw, mask, 100, 10,
psgmii_st_phy_link_is_up);
}
static bool psgmii_st_phy_wait_tx_complete(struct ess_switch *esw, u32 mask)
{
return psgmii_st_phy_wait(esw, mask, PSGMII_ST_TRAFFIC_TIMEOUT, 1,
psgmii_st_phy_check_counters);
}
static bool psgmii_st_run_test_serial(struct ess_switch *esw)
{
bool result = true;
int i;
for (i = 0; i < esw->num_phy - 1; i++) {
struct phy_device *phydev = esw->phydev[i];
psgmii_st_phy_reset_loopback(phydev);
psgmii_st_phy_wait_link(esw, BIT(i));
psgmii_st_phy_start_traffic(phydev);
/* wait for the traffic to complete */
result &= psgmii_st_phy_wait_tx_complete(esw, BIT(i));
/* Power down */
psgmii_st_phy_power_down(phydev);
if (!result)
break;
}
return result;
}
static bool psgmii_st_run_test_parallel(struct ess_switch *esw)
{
bool result;
int i;
/* enable loopback mode on all PHYs */
for (i = 0; i < esw->num_phy - 1; i++)
psgmii_st_phy_reset_loopback(esw->phydev[i]);
psgmii_st_phy_wait_link(esw, esw->phy_mask);
/* start traffic on all PHYs parallely */
for (i = 0; i < esw->num_phy - 1; i++)
psgmii_st_phy_start_traffic(esw->phydev[i]);
/* wait for the traffic to complete on all PHYs */
result = psgmii_st_phy_wait_tx_complete(esw, esw->phy_mask);
/* Power down all PHYs */
for (i = 0; i < esw->num_phy - 1; i++)
psgmii_st_phy_power_down(esw->phydev[i]);
return result;
}
struct psgmii_st_stats {
int succeed;
int failed;
int failed_max;
int failed_cont;
};
static void psgmii_st_update_stats(struct psgmii_st_stats *stats,
bool success)
{
if (success) {
stats->succeed++;
stats->failed_cont = 0;
return;
}
stats->failed++;
stats->failed_cont++;
if (stats->failed_max < stats->failed_cont)
stats->failed_max = stats->failed_cont;
}
static void psgmii_self_test(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
struct ess_switch *esw = &priv->esw;
struct psgmii_st_stats stats;
bool result = false;
unsigned long tm;
int i;
memset(&stats, 0, sizeof(stats));
tm = get_timer(0);
for (i = 0; i < esw->num_phy - 1; i++)
psgmii_st_phy_prepare(esw->phydev[i]);
for (i = 0; i < PSGMII_ST_NUM_RETRIES; i++) {
qca8075_ess_reset(dev);
/* enable loopback mode on the switch's ports */
esw_port_loopback_set_all(esw, true);
/* run test on each PHYs individually after each other */
result = psgmii_st_run_test_serial(esw);
if (result) {
/* run test on each PHYs parallely */
result = psgmii_st_run_test_parallel(esw);
}
psgmii_st_update_stats(&stats, result);
if (psgmii_self_test_repeat)
continue;
if (result)
break;
}
for (i = 0; i < esw->num_phy - 1; i++) {
/* Configuration recover */
psgmii_st_phy_recover(esw->phydev[i]);
/* Disable loopback */
phy_write(esw->phydev[i], MDIO_DEVAD_NONE,
QCA807X_FUNCTION_CONTROL, 0x6860);
phy_write(esw->phydev[i], MDIO_DEVAD_NONE, MII_BMCR, 0x9040);
}
/* disable loopback mode on the switch's ports */
esw_port_loopback_set_all(esw, false);
tm = get_timer(tm);
dev_dbg(priv->dev, "\nPSGMII self-test: succeed %d, failed %d (max %d), duration %lu.%03lu secs\n",
stats.succeed, stats.failed, stats.failed_max,
tm / 1000, tm % 1000);
}
static int ess_switch_disable_lookup(struct ess_switch *esw)
{
int val;
int i;
/* Disable port lookup for all ports*/
for (i = 0; i < ESS_PORTS_NUM; i++) {
int ess_port_vid;
val = readl(esw->base + ESS_PORT_LOOKUP_CTRL(i));
val &= ~ESS_PORT_VID_MEM_MASK;
switch (i) {
case 0:
fallthrough;
case 5:
/* CPU,WAN port -> nothing */
ess_port_vid = 0;
break;
case 1 ... 4:
/* LAN ports -> all other LAN ports */
ess_port_vid = GENMASK(4, 1);
ess_port_vid &= ~BIT(i);
break;
default:
return -EINVAL;
}
val |= FIELD_PREP(ESS_PORT_VID_MEM_MASK, ess_port_vid);
writel(val, esw->base + ESS_PORT_LOOKUP_CTRL(i));
}
/* Set magic value for the global forwarding register 1 */
writel(0x3e3e3e, esw->base + ESS_GLOBAL_FW_CTRL1);
return 0;
}
static int ess_switch_enable_lookup(struct ess_switch *esw)
{
int val;
int i;
/* Enable port lookup for all ports*/
for (i = 0; i < ESS_PORTS_NUM; i++) {
int ess_port_vid;
val = readl(esw->base + ESS_PORT_LOOKUP_CTRL(i));
val &= ~ESS_PORT_VID_MEM_MASK;
switch (i) {
case 0:
/* CPU port -> all other ports */
ess_port_vid = GENMASK(5, 1);
break;
case 1 ... 4:
/* LAN ports -> CPU and all other LAN ports */
ess_port_vid = GENMASK(4, 0);
ess_port_vid &= ~BIT(i);
break;
case 5:
/* WAN port -> CPU port only */
ess_port_vid = BIT(0);
break;
default:
return -EINVAL;
}
val |= FIELD_PREP(ESS_PORT_VID_MEM_MASK, ess_port_vid);
writel(val, esw->base + ESS_PORT_LOOKUP_CTRL(i));
}
/* Set magic value for the global forwarding register 1 */
writel(0x3f3f3f, esw->base + ESS_GLOBAL_FW_CTRL1);
return 0;
}
static void ess_switch_init(struct ess_switch *esw)
{
int val = 0;
int i;
/* Set magic value for the global forwarding register 1 */
writel(0x3e3e3e, esw->base + ESS_GLOBAL_FW_CTRL1);
/* Set 1000M speed, full duplex and RX/TX flow control for the CPU port*/
val &= ~ESS_PORT_SPEED_MASK;
val |= FIELD_PREP(ESS_PORT_SPEED_MASK, ESS_PORT_SPEED_1000);
val |= ESS_PORT_DUPLEX_MODE;
val |= ESS_PORT_TX_FLOW_EN;
val |= ESS_PORT_RX_FLOW_EN;
writel(val, esw->base + ESS_PORT0_STATUS);
/* Disable port lookup for all ports*/
for (i = 0; i < ESS_PORTS_NUM; i++) {
val = readl(esw->base + ESS_PORT_LOOKUP_CTRL(i));
val &= ~ESS_PORT_VID_MEM_MASK;
writel(val, esw->base + ESS_PORT_LOOKUP_CTRL(i));
}
/* Set HOL settings for all ports*/
for (i = 0; i < ESS_PORTS_NUM; i++) {
val = 0;
val |= FIELD_PREP(EG_PORT_QUEUE_NUM_MASK, 30);
if (i == 0 || i == 5) {
val |= FIELD_PREP(EG_PRI5_QUEUE_NUM_MASK, 4);
val |= FIELD_PREP(EG_PRI4_QUEUE_NUM_MASK, 4);
}
val |= FIELD_PREP(EG_PRI3_QUEUE_NUM_MASK, 4);
val |= FIELD_PREP(EG_PRI2_QUEUE_NUM_MASK, 4);
val |= FIELD_PREP(EG_PRI1_QUEUE_NUM_MASK, 4);
val |= FIELD_PREP(EG_PRI0_QUEUE_NUM_MASK, 4);
writel(val, esw->base + ESS_PORT_HOL_CTRL0(i));
val = readl(esw->base + ESS_PORT_HOL_CTRL1(i));
val &= ~ESS_ING_BUF_NUM_0_MASK;
val |= FIELD_PREP(ESS_ING_BUF_NUM_0_MASK, 6);
writel(val, esw->base + ESS_PORT_HOL_CTRL1(i));
}
/* Give switch some time */
mdelay(1);
/* Enable RX and TX MAC-s */
val = readl(esw->base + ESS_PORT0_STATUS);
val |= ESS_PORT_TXMAC_EN;
val |= ESS_PORT_RXMAC_EN;
writel(val, esw->base + ESS_PORT0_STATUS);
/* Set magic value for the global forwarding register 1 */
writel(0x7f7f7f, esw->base + ESS_GLOBAL_FW_CTRL1);
}
static int essedma_of_phy(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
struct ess_switch *esw = &priv->esw;
int num_phy = 0, ret = 0;
ofnode node;
int i;
ofnode_for_each_subnode(node, esw->ports_node) {
struct ofnode_phandle_args phandle_args;
struct phy_device *phydev;
u32 phy_addr;
if (ofnode_is_enabled(node)) {
if (ofnode_parse_phandle_with_args(node, "phy-handle", NULL, 0, 0,
&phandle_args)) {
dev_dbg(priv->dev, "Failed to find phy-handle\n");
return -ENODEV;
}
ret = ofnode_read_u32(phandle_args.node, "reg", &phy_addr);
if (ret) {
dev_dbg(priv->dev, "Missing reg property in PHY node %s\n",
ofnode_get_name(phandle_args.node));
return ret;
}
phydev = dm_mdio_phy_connect(priv->mdio_dev, phy_addr,
dev, priv->esw.port_wrapper_mode);
if (!phydev) {
dev_dbg(priv->dev, "Failed to find phy on addr %d\n", phy_addr);
return -ENODEV;
}
phydev->node = phandle_args.node;
ret = phy_config(phydev);
esw->phydev[num_phy] = phydev;
num_phy++;
}
}
esw->num_phy = num_phy;
for (i = 0; i < esw->num_phy - 1; i++)
esw->phy_mask |= BIT(i);
return ret;
}
static int essedma_of_switch(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
int port_wrapper_mode = -1;
priv->esw.ports_node = ofnode_find_subnode(dev_ofnode(dev), "ports");
if (!ofnode_valid(priv->esw.ports_node)) {
printf("Failed to find ports node\n");
return -EINVAL;
}
port_wrapper_mode = ofnode_read_phy_mode(priv->esw.ports_node);
if (port_wrapper_mode == -1)
return -EINVAL;
priv->esw.port_wrapper_mode = port_wrapper_mode;
return essedma_of_phy(dev);
}
static void ipq40xx_edma_start_rx_tx(struct essedma_priv *priv)
{
volatile u32 data;
/* enable RX queues */
data = readl(priv->base + EDMA_REG_RXQ_CTRL);
data |= EDMA_RXQ_CTRL_EN;
writel(data, priv->base + EDMA_REG_RXQ_CTRL);
/* enable TX queues */
data = readl(priv->base + EDMA_REG_TXQ_CTRL);
data |= EDMA_TXQ_CTRL_TXQ_EN;
writel(data, priv->base + EDMA_REG_TXQ_CTRL);
}
/*
* ipq40xx_edma_init_desc()
* Update descriptor ring size,
* Update buffer and producer/consumer index
*/
static void ipq40xx_edma_init_desc(struct essedma_priv *priv)
{
struct edma_ring *rfd_ring;
struct edma_ring *etdr;
volatile u32 data = 0;
u16 hw_cons_idx = 0;
/* Set the base address of every TPD ring. */
etdr = &priv->tpd_ring;
/* Update TX descriptor ring base address. */
writel((u32)(etdr->dma & 0xffffffff),
priv->base + EDMA_REG_TPD_BASE_ADDR_Q(EDMA_TXQ_ID));
data = readl(priv->base + EDMA_REG_TPD_IDX_Q(EDMA_TXQ_ID));
/* Calculate hardware consumer index for Tx. */
hw_cons_idx = FIELD_GET(EDMA_TPD_CONS_IDX_MASK, data);
etdr->head = hw_cons_idx;
etdr->tail = hw_cons_idx;
data &= ~EDMA_TPD_PROD_IDX_MASK;
data |= hw_cons_idx;
/* Update producer index for Tx. */
writel(data, priv->base + EDMA_REG_TPD_IDX_Q(EDMA_TXQ_ID));
/* Update SW consumer index register for Tx. */
writel(hw_cons_idx,
priv->base + EDMA_REG_TX_SW_CONS_IDX_Q(EDMA_TXQ_ID));
/* Set TPD ring size. */
writel((u32)(etdr->count & EDMA_TPD_RING_SIZE_MASK),
priv->base + EDMA_REG_TPD_RING_SIZE);
/* Configure Rx ring. */
rfd_ring = &priv->rfd_ring;
/* Update Receive Free descriptor ring base address. */
writel((u32)(rfd_ring->dma & 0xffffffff),
priv->base + EDMA_REG_RFD_BASE_ADDR_Q(EDMA_RXQ_ID));
data = readl(priv->base + EDMA_REG_RFD_BASE_ADDR_Q(EDMA_RXQ_ID));
/* Update RFD ring size and RX buffer size. */
data = (rfd_ring->count & EDMA_RFD_RING_SIZE_MASK)
<< EDMA_RFD_RING_SIZE_SHIFT;
data |= (EDMA_MAX_PKT_SIZE & EDMA_RX_BUF_SIZE_MASK)
<< EDMA_RX_BUF_SIZE_SHIFT;
writel(data, priv->base + EDMA_REG_RX_DESC0);
/* Disable TX FIFO low watermark and high watermark */
writel(0, priv->base + EDMA_REG_TXF_WATER_MARK);
/* Load all of base address above */
data = readl(priv->base + EDMA_REG_TX_SRAM_PART);
data |= 1 << EDMA_LOAD_PTR_SHIFT;
writel(data, priv->base + EDMA_REG_TX_SRAM_PART);
}
static void ipq40xx_edma_init_rfd_ring(struct essedma_priv *priv)
{
struct edma_ring *erdr = &priv->rfd_ring;
struct edma_rfd *rfds = erdr->hw_desc;
int i;
for (i = 0; i < erdr->count; i++)
rfds[i].buffer_addr = virt_to_phys(net_rx_packets[i]);
flush_dcache_range(erdr->dma, erdr->dma + erdr->hw_size);
/* setup producer index */
erdr->head = erdr->count - 1;
writel(erdr->head, priv->base + EDMA_REG_RFD_IDX_Q(EDMA_RXQ_ID));
}
static void ipq40xx_edma_configure(struct essedma_priv *priv)
{
u32 tmp;
int i;
/* Set RSS type */
writel(IPQ40XX_EDMA_RSS_TYPE_NONE, priv->base + EDMA_REG_RSS_TYPE);
/* Configure RSS indirection table.
* 128 hash will be configured in the following
* pattern: hash{0,1,2,3} = {Q0,Q2,Q4,Q6} respectively
* and so on
*/
for (i = 0; i < EDMA_NUM_IDT; i++)
writel(EDMA_RSS_IDT_VALUE, priv->base + EDMA_REG_RSS_IDT(i));
/* Set RFD burst number */
tmp = (EDMA_RFD_BURST << EDMA_RXQ_RFD_BURST_NUM_SHIFT);
/* Set RFD prefetch threshold */
tmp |= (EDMA_RFD_THR << EDMA_RXQ_RFD_PF_THRESH_SHIFT);
/* Set RFD in host ring low threshold to generte interrupt */
tmp |= (EDMA_RFD_LTHR << EDMA_RXQ_RFD_LOW_THRESH_SHIFT);
writel(tmp, priv->base + EDMA_REG_RX_DESC1);
/* configure reception control data. */
/* Set Rx FIFO threshold to start to DMA data to host */
tmp = EDMA_FIFO_THRESH_128_BYTE;
/* Set RX remove vlan bit */
tmp |= EDMA_RXQ_CTRL_RMV_VLAN;
writel(tmp, priv->base + EDMA_REG_RXQ_CTRL);
/* Configure transmission control data */
tmp = (EDMA_TPD_BURST << EDMA_TXQ_NUM_TPD_BURST_SHIFT);
tmp |= EDMA_TXQ_CTRL_TPD_BURST_EN;
tmp |= (EDMA_TXF_BURST << EDMA_TXQ_TXF_BURST_NUM_SHIFT);
writel(tmp, priv->base + EDMA_REG_TXQ_CTRL);
}
static void ipq40xx_edma_stop_rx_tx(struct essedma_priv *priv)
{
volatile u32 data;
data = readl(priv->base + EDMA_REG_RXQ_CTRL);
data &= ~EDMA_RXQ_CTRL_EN;
writel(data, priv->base + EDMA_REG_RXQ_CTRL);
data = readl(priv->base + EDMA_REG_TXQ_CTRL);
data &= ~EDMA_TXQ_CTRL_TXQ_EN;
writel(data, priv->base + EDMA_REG_TXQ_CTRL);
}
static int ipq40xx_eth_recv(struct udevice *dev, int flags, uchar **packetp)
{
struct essedma_priv *priv = dev_get_priv(dev);
struct edma_ring *erdr = &priv->rfd_ring;
struct edma_rrd *rrd;
u32 hw_tail;
u8 *rx_pkt;
hw_tail = readl(priv->base + EDMA_REG_RFD_IDX_Q(EDMA_RXQ_ID));
hw_tail = FIELD_GET(EDMA_RFD_CONS_IDX_MASK, hw_tail);
if (hw_tail == erdr->tail)
return -EAGAIN;
rx_pkt = net_rx_packets[erdr->tail];
invalidate_dcache_range((unsigned long)rx_pkt,
(unsigned long)(rx_pkt + EDMA_MAX_PKT_SIZE));
rrd = (struct edma_rrd *)rx_pkt;
/* Check if RRD is valid */
if (!(rrd->rrd7 & EDMA_RRD7_DESC_VALID))
return 0;
*packetp = rx_pkt + EDMA_RRD_SIZE;
/* get the packet size */
return rrd->rrd6;
}
static int ipq40xx_eth_free_pkt(struct udevice *dev, uchar *packet,
int length)
{
struct essedma_priv *priv = dev_get_priv(dev);
struct edma_ring *erdr;
erdr = &priv->rfd_ring;
/* Update the producer index */
writel(erdr->head, priv->base + EDMA_REG_RFD_IDX_Q(EDMA_RXQ_ID));
erdr->head++;
if (erdr->head == erdr->count)
erdr->head = 0;
/* Update the consumer index */
erdr->tail++;
if (erdr->tail == erdr->count)
erdr->tail = 0;
writel(erdr->tail,
priv->base + EDMA_REG_RX_SW_CONS_IDX_Q(EDMA_RXQ_ID));
return 0;
}
static int ipq40xx_eth_start(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
ipq40xx_edma_init_rfd_ring(priv);
ipq40xx_edma_start_rx_tx(priv);
ess_switch_enable_lookup(&priv->esw);
return 0;
}
/*
* One TPD would be enough for sending a packet, however because the
* minimal cache line size is larger than the size of a TPD it is not
* possible to flush only one at once. To overcome this limitation
* multiple TPDs are used for sending a single packet.
*/
#define EDMA_TPDS_PER_PACKET 4
#define EDMA_TPD_MIN_BYTES 4
#define EDMA_MIN_PKT_SIZE (EDMA_TPDS_PER_PACKET * EDMA_TPD_MIN_BYTES)
#define EDMA_TX_COMPLETE_TIMEOUT 1000000
static int ipq40xx_eth_send(struct udevice *dev, void *packet, int length)
{
struct essedma_priv *priv = dev_get_priv(dev);
struct edma_tpd *first_tpd;
struct edma_tpd *tpds;
int i;
if (length < EDMA_MIN_PKT_SIZE)
return 0;
flush_dcache_range((unsigned long)(packet),
(unsigned long)(packet) +
roundup(length, ARCH_DMA_MINALIGN));
tpds = priv->tpd_ring.hw_desc;
for (i = 0; i < EDMA_TPDS_PER_PACKET; i++) {
struct edma_tpd *tpd;
void *frag;
frag = packet + (i * EDMA_TPD_MIN_BYTES);
/* get the next TPD */
tpd = &tpds[priv->tpd_ring.head];
if (i == 0)
first_tpd = tpd;
/* update the software index */
priv->tpd_ring.head++;
if (priv->tpd_ring.head == priv->tpd_ring.count)
priv->tpd_ring.head = 0;
tpd->svlan_tag = 0;
tpd->addr = virt_to_phys(frag);
tpd->word3 = EDMA_PORT_ENABLE_ALL << EDMA_TPD_PORT_BITMAP_SHIFT;
if (i < (EDMA_TPDS_PER_PACKET - 1)) {
tpd->len = EDMA_TPD_MIN_BYTES;
tpd->word1 = 0;
} else {
tpd->len = length;
tpd->word1 = 1 << EDMA_TPD_EOP_SHIFT;
}
length -= EDMA_TPD_MIN_BYTES;
}
/* make sure that memory writing completes */
wmb();
flush_dcache_range((unsigned long)first_tpd,
(unsigned long)first_tpd +
EDMA_TPDS_PER_PACKET * sizeof(struct edma_tpd));
/* update the TX producer index */
writel(priv->tpd_ring.head,
priv->base + EDMA_REG_TPD_IDX_Q(EDMA_TXQ_ID));
/* Wait for TX DMA completion */
for (i = 0; i < EDMA_TX_COMPLETE_TIMEOUT; i++) {
u32 r, prod, cons;
r = readl(priv->base + EDMA_REG_TPD_IDX_Q(EDMA_TXQ_ID));
prod = FIELD_GET(EDMA_TPD_PROD_IDX_MASK, r);
cons = FIELD_GET(EDMA_TPD_CONS_IDX_MASK, r);
if (cons == prod)
break;
udelay(1);
}
if (i == EDMA_TX_COMPLETE_TIMEOUT)
printf("TX timeout: packet not sent!\n");
/* update the software TX consumer index register */
writel(priv->tpd_ring.head,
priv->base + EDMA_REG_TX_SW_CONS_IDX_Q(EDMA_TXQ_ID));
return 0;
}
static void ipq40xx_eth_stop(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
ess_switch_disable_lookup(&priv->esw);
ipq40xx_edma_stop_rx_tx(priv);
}
static void ipq40xx_edma_free_ring(struct edma_ring *ring)
{
free(ring->hw_desc);
}
/*
* Free Tx and Rx rings
*/
static void ipq40xx_edma_free_rings(struct essedma_priv *priv)
{
ipq40xx_edma_free_ring(&priv->tpd_ring);
ipq40xx_edma_free_ring(&priv->rfd_ring);
}
/*
* ipq40xx_edma_alloc_ring()
* allocate edma ring descriptor.
*/
static int ipq40xx_edma_alloc_ring(struct edma_ring *erd,
unsigned int desc_size)
{
erd->head = 0;
erd->tail = 0;
/* Alloc HW descriptors */
erd->hw_size = roundup(desc_size * erd->count,
ARCH_DMA_MINALIGN);
erd->hw_desc = memalign(CONFIG_SYS_CACHELINE_SIZE, erd->hw_size);
if (!erd->hw_desc)
return -ENOMEM;
memset(erd->hw_desc, 0, erd->hw_size);
erd->dma = virt_to_phys(erd->hw_desc);
return 0;
}
/*
* ipq40xx_allocate_tx_rx_rings()
*/
static int ipq40xx_edma_alloc_tx_rx_rings(struct essedma_priv *priv)
{
int ret;
ret = ipq40xx_edma_alloc_ring(&priv->tpd_ring,
sizeof(struct edma_tpd));
if (ret)
return ret;
ret = ipq40xx_edma_alloc_ring(&priv->rfd_ring,
sizeof(struct edma_rfd));
if (ret)
goto err_free_tpd;
return 0;
err_free_tpd:
ipq40xx_edma_free_ring(&priv->tpd_ring);
return ret;
}
static int ipq40xx_eth_write_hwaddr(struct udevice *dev)
{
struct eth_pdata *pdata = dev_get_plat(dev);
struct essedma_priv *priv = dev_get_priv(dev);
unsigned char *mac = pdata->enetaddr;
u32 mac_lo, mac_hi;
mac_hi = ((u32)mac[0]) << 8 | (u32)mac[1];
mac_lo = ((u32)mac[2]) << 24 | ((u32)mac[3]) << 16 |
((u32)mac[4]) << 8 | (u32)mac[5];
writel(mac_lo, priv->base + REG_MAC_CTRL0);
writel(mac_hi, priv->base + REG_MAC_CTRL1);
return 0;
}
static int edma_init(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
int ret;
priv->tpd_ring.count = IPQ40XX_EDMA_TX_RING_SIZE;
priv->rfd_ring.count = PKTBUFSRX;
ret = ipq40xx_edma_alloc_tx_rx_rings(priv);
if (ret)
return -ENOMEM;
ipq40xx_edma_stop_rx_tx(priv);
/* Configure EDMA. */
ipq40xx_edma_configure(priv);
/* Configure descriptor Ring */
ipq40xx_edma_init_desc(priv);
ess_switch_disable_lookup(&priv->esw);
return 0;
}
static int essedma_probe(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
int ret;
priv->dev = dev;
priv->base = dev_read_addr_name(dev, "edma");
if (priv->base == FDT_ADDR_T_NONE)
return -EINVAL;
priv->psgmii_base = dev_read_addr_name(dev, "psgmii_phy");
if (priv->psgmii_base == FDT_ADDR_T_NONE)
return -EINVAL;
priv->esw.base = dev_read_addr_name(dev, "base");
if (priv->esw.base == FDT_ADDR_T_NONE)
return -EINVAL;
ret = clk_get_by_name(dev, "ess", &priv->ess_clk);
if (ret)
return ret;
ret = reset_get_by_name(dev, "ess", &priv->ess_rst);
if (ret)
return ret;
ret = clk_enable(&priv->ess_clk);
if (ret)
return ret;
ess_reset(dev);
ret = uclass_get_device_by_driver(UCLASS_MDIO,
DM_DRIVER_GET(ipq4019_mdio),
&priv->mdio_dev);
if (ret) {
dev_dbg(dev, "Cant find IPQ4019 MDIO: %d\n", ret);
goto err;
}
/* OF switch and PHY parsing and configuration */
ret = essedma_of_switch(dev);
if (ret)
goto err;
switch (priv->esw.port_wrapper_mode) {
case PHY_INTERFACE_MODE_PSGMII:
writel(PSGMIIPHY_PLL_VCO_VAL,
priv->psgmii_base + PSGMIIPHY_PLL_VCO_RELATED_CTRL);
writel(PSGMIIPHY_VCO_VAL, priv->psgmii_base +
PSGMIIPHY_VCO_CALIBRATION_CTRL_REGISTER_1);
/* wait for 10ms */
mdelay(10);
writel(PSGMIIPHY_VCO_RST_VAL, priv->psgmii_base +
PSGMIIPHY_VCO_CALIBRATION_CTRL_REGISTER_1);
break;
case PHY_INTERFACE_MODE_RGMII:
writel(0x1, RGMII_TCSR_ESS_CFG);
writel(0x400, priv->esw.base + ESS_RGMII_CTRL);
break;
default:
printf("Unknown MII interface\n");
}
if (priv->esw.port_wrapper_mode == PHY_INTERFACE_MODE_PSGMII)
psgmii_self_test(dev);
ess_switch_init(&priv->esw);
ret = edma_init(dev);
if (ret)
goto err;
return 0;
err:
reset_assert(&priv->ess_rst);
clk_disable(&priv->ess_clk);
return ret;
}
static int essedma_remove(struct udevice *dev)
{
struct essedma_priv *priv = dev_get_priv(dev);
ipq40xx_edma_free_rings(priv);
clk_disable(&priv->ess_clk);
reset_assert(&priv->ess_rst);
return 0;
}
static const struct eth_ops essedma_eth_ops = {
.start = ipq40xx_eth_start,
.send = ipq40xx_eth_send,
.recv = ipq40xx_eth_recv,
.free_pkt = ipq40xx_eth_free_pkt,
.stop = ipq40xx_eth_stop,
.write_hwaddr = ipq40xx_eth_write_hwaddr,
};
static const struct udevice_id essedma_ids[] = {
{ .compatible = "qcom,ipq4019-ess", },
{ }
};
U_BOOT_DRIVER(essedma) = {
.name = "essedma",
.id = UCLASS_ETH,
.of_match = essedma_ids,
.probe = essedma_probe,
.remove = essedma_remove,
.priv_auto = sizeof(struct essedma_priv),
.plat_auto = sizeof(struct eth_pdata),
.ops = &essedma_eth_ops,
.flags = DM_FLAG_ALLOC_PRIV_DMA,
};