blob: 5ab74add9bd199b8eea5f007b343dcbfdebcdb6d [file] [log] [blame]
/* SPDX-License-Identifier: GPL-2.0
*
* Copyright (c) 2022 MediaTek Inc.
* Author: Henry Yen <henry.yen@mediatek.com>
*/
#include <linux/regmap.h>
#include "mtk_eth_soc.h"
#include "mtk_eth_dbg.h"
#include "mtk_eth_reset.h"
char* mtk_reset_event_name[32] = {
[MTK_EVENT_FORCE] = "Force",
[MTK_EVENT_WARM_CNT] = "Warm",
[MTK_EVENT_COLD_CNT] = "Cold",
[MTK_EVENT_TOTAL_CNT] = "Total",
[MTK_EVENT_FQ_EMPTY] = "FQ Empty",
[MTK_EVENT_TSO_FAIL] = "TSO Fail",
[MTK_EVENT_TSO_ILLEGAL] = "TSO Illegal",
[MTK_EVENT_TSO_ALIGN] = "TSO Align",
[MTK_EVENT_RFIFO_OV] = "RFIFO OV",
[MTK_EVENT_RFIFO_UF] = "RFIFO UF",
};
static int mtk_wifi_num = 0;
static int mtk_rest_cnt = 0;
u32 mtk_reset_flag = MTK_FE_START_RESET;
bool mtk_stop_fail;
typedef u32 (*mtk_monitor_xdma_func) (struct mtk_eth *eth);
void mtk_reset_event_update(struct mtk_eth *eth, u32 id)
{
struct mtk_reset_event *reset_event = &eth->reset_event;
reset_event->count[id]++;
}
int mtk_eth_cold_reset(struct mtk_eth *eth)
{
u32 reset_bits = 0;
if (MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V2) ||
MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V3))
regmap_write(eth->ethsys, ETHSYS_FE_RST_CHK_IDLE_EN, 0);
reset_bits = RSTCTRL_ETH | RSTCTRL_FE | RSTCTRL_PPE0;
if (MTK_HAS_CAPS(eth->soc->caps, MTK_RSTCTRL_PPE1))
reset_bits |= RSTCTRL_PPE1;
#if defined(CONFIG_MEDIATEK_NETSYS_V3)
if (MTK_HAS_CAPS(eth->soc->caps, MTK_RSTCTRL_PPE2))
reset_bits |= RSTCTRL_PPE2;
if (MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V3))
reset_bits |= RSTCTRL_WDMA0 | RSTCTRL_WDMA1 | RSTCTRL_WDMA2;
#endif
ethsys_reset(eth, reset_bits);
if (MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V2))
regmap_write(eth->ethsys, ETHSYS_FE_RST_CHK_IDLE_EN, 0x3ffffff);
if (MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V3))
regmap_write(eth->ethsys, ETHSYS_FE_RST_CHK_IDLE_EN, 0x6F8FF);
return 0;
}
int mtk_eth_warm_reset(struct mtk_eth *eth)
{
u32 reset_bits = 0, i = 0, done = 0;
u32 val1 = 0, val2 = 0, val3 = 0;
mdelay(100);
reset_bits |= RSTCTRL_FE;
regmap_update_bits(eth->ethsys, ETHSYS_RSTCTRL,
reset_bits, reset_bits);
while (i < 1000) {
regmap_read(eth->ethsys, ETHSYS_RSTCTRL, &val1);
if (val1 & RSTCTRL_FE)
break;
i++;
udelay(1);
}
if (i < 1000) {
reset_bits = RSTCTRL_ETH | RSTCTRL_PPE0;
if (MTK_HAS_CAPS(eth->soc->caps, MTK_RSTCTRL_PPE1))
reset_bits |= RSTCTRL_PPE1;
#if defined(CONFIG_MEDIATEK_NETSYS_V3)
if (MTK_HAS_CAPS(eth->soc->caps, MTK_RSTCTRL_PPE2))
reset_bits |= RSTCTRL_PPE2;
if (mtk_reset_flag == MTK_FE_START_RESET)
reset_bits |= RSTCTRL_WDMA0 |
RSTCTRL_WDMA1 | RSTCTRL_WDMA2;
#endif
regmap_update_bits(eth->ethsys, ETHSYS_RSTCTRL,
reset_bits, reset_bits);
udelay(1);
regmap_read(eth->ethsys, ETHSYS_RSTCTRL, &val2);
if (!(val2 & reset_bits))
pr_info("[%s] error val2=0x%x reset_bits=0x%x !\n",
__func__, val2, reset_bits);
reset_bits |= RSTCTRL_FE;
regmap_update_bits(eth->ethsys, ETHSYS_RSTCTRL,
reset_bits, ~reset_bits);
udelay(1);
regmap_read(eth->ethsys, ETHSYS_RSTCTRL, &val3);
if (val3 & reset_bits)
pr_info("[%s] error val3=0x%x reset_bits=0x%x !\n",
__func__, val3, reset_bits);
done = 1;
mtk_reset_event_update(eth, MTK_EVENT_WARM_CNT);
}
pr_info("[%s] reset record val1=0x%x, val2=0x%x, val3=0x%x i:%d done:%d\n",
__func__, val1, val2, val3, i, done);
if (!done)
mtk_eth_cold_reset(eth);
return 0;
}
u32 mtk_check_reset_event(struct mtk_eth *eth, u32 status)
{
u32 ret = 0, val = 0;
if ((status & MTK_FE_INT_FQ_EMPTY) ||
(status & MTK_FE_INT_RFIFO_UF) ||
(status & MTK_FE_INT_RFIFO_OV) ||
(status & MTK_FE_INT_TSO_FAIL) ||
(status & MTK_FE_INT_TSO_ALIGN) ||
(status & MTK_FE_INT_TSO_ILLEGAL)) {
while (status) {
val = ffs((unsigned int)status) - 1;
mtk_reset_event_update(eth, val);
status &= ~(1 << val);
}
ret = 1;
}
if (atomic_read(&force)) {
mtk_reset_event_update(eth, MTK_EVENT_FORCE);
ret = 1;
}
if (ret) {
mtk_reset_event_update(eth, MTK_EVENT_TOTAL_CNT);
if (dbg_show_level)
mtk_dump_netsys_info(eth);
}
return ret;
}
irqreturn_t mtk_handle_fe_irq(int irq, void *_eth)
{
struct mtk_eth *eth = _eth;
u32 status = 0, val = 0;
status = mtk_r32(eth, MTK_FE_INT_STATUS);
pr_info("[%s] Trigger FE Misc ISR: 0x%x\n", __func__, status);
while (status) {
val = ffs((unsigned int)status) - 1;
status &= ~(1 << val);
if ((val == MTK_EVENT_TSO_FAIL) ||
(val == MTK_EVENT_TSO_ILLEGAL) ||
(val == MTK_EVENT_TSO_ALIGN) ||
(val == MTK_EVENT_RFIFO_OV) ||
(val == MTK_EVENT_RFIFO_UF))
pr_info("[%s] Detect reset event: %s !\n", __func__,
mtk_reset_event_name[val]);
}
mtk_w32(eth, 0xFFFFFFFF, MTK_FE_INT_STATUS);
return IRQ_HANDLED;
}
static void mtk_dump_reg(void *_eth, char *name, u32 offset, u32 range)
{
struct mtk_eth *eth = _eth;
u32 cur = offset;
pr_info("\n============ %s ============\n", name);
while(cur < offset + range) {
pr_info("0x%x: %08x %08x %08x %08x\n",
cur, mtk_r32(eth, cur), mtk_r32(eth, cur + 0x4),
mtk_r32(eth, cur + 0x8), mtk_r32(eth, cur + 0xc));
cur += 0x10;
}
}
void mtk_dump_netsys_info(void *_eth)
{
struct mtk_eth *eth = _eth;
u32 id = 0;
mtk_dump_reg(eth, "FE", 0x0, 0x500);
mtk_dump_reg(eth, "ADMA", PDMA_BASE, 0x300);
for (id = 0; id < MTK_QDMA_PAGE_NUM; id++){
mtk_w32(eth, id, MTK_QDMA_PAGE);
pr_info("\nQDMA PAGE:%x ",mtk_r32(eth, MTK_QDMA_PAGE));
mtk_dump_reg(eth, "QDMA", QDMA_BASE, 0x100);
mtk_w32(eth, 0, MTK_QDMA_PAGE);
}
mtk_dump_reg(eth, "QDMA", MTK_QRX_BASE_PTR0, 0x300);
mtk_dump_reg(eth, "WDMA", WDMA_BASE(0), 0x600);
mtk_dump_reg(eth, "PPE", 0x2200, 0x200);
mtk_dump_reg(eth, "GMAC", 0x10000, 0x300);
if (MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V3)) {
mtk_dump_reg(eth, "XGMAC0", 0x12000, 0x300);
mtk_dump_reg(eth, "XGMAC1", 0x13000, 0x300);
mtk_dump_usxgmii(eth->usxgmii->pcs[0].regmap,
"USXGMII0", 0, 0x1000);
mtk_dump_usxgmii(eth->usxgmii->pcs[1].regmap,
"USXGMII1", 0, 0x1000);
}
}
u32 mtk_monitor_wdma_tx(struct mtk_eth *eth)
{
static u32 pre_dtx[MTK_WDMA_CNT];
static u32 err_cnt[MTK_WDMA_CNT];
u32 i = 0, cur_dtx = 0, tx_busy = 0, tx_rdy = 0, err_flag = 0;
u32 dbg_mon = 0;
for (i = 0; i < MTK_WDMA_CNT; i++) {
cur_dtx = mtk_r32(eth, MTK_WDMA_DTX_PTR(i));
tx_busy = mtk_r32(eth, MTK_WDMA_GLO_CFG(i)) & MTK_TX_DMA_BUSY;
dbg_mon = mtk_r32(eth, MTK_WDMA_TX_DBG_MON0(i));
tx_rdy = !(dbg_mon & MTK_CDM_TXFIFO_RDY);
if (cur_dtx == pre_dtx[i] && tx_busy && tx_rdy) {
err_cnt[i]++;
if (err_cnt[i] >= 3) {
pr_info("WDMA %d Info\n", i);
pr_info("err_cnt = %d", err_cnt[i]);
pr_info("prev_dtx = 0x%x | cur_dtx = 0x%x\n",
pre_dtx[i], cur_dtx);
pr_info("WDMA_CTX_PTR = 0x%x\n",
mtk_r32(eth, MTK_WDMA_CTX_PTR(i)));
pr_info("WDMA_DTX_PTR = 0x%x\n",
mtk_r32(eth, MTK_WDMA_DTX_PTR(i)));
pr_info("WDMA_GLO_CFG = 0x%x\n",
mtk_r32(eth, MTK_WDMA_GLO_CFG(i)));
pr_info("WDMA_TX_DBG_MON0 = 0x%x\n",
mtk_r32(eth, MTK_WDMA_TX_DBG_MON0(i)));
pr_info("==============================\n");
err_flag = 1;
}
} else
err_cnt[i] = 0;
pre_dtx[i] = cur_dtx;
}
if (err_flag)
return MTK_FE_START_RESET;
else
return 0;
}
u32 mtk_monitor_wdma_rx(struct mtk_eth *eth)
{
static u32 pre_drx[MTK_WDMA_CNT];
static u32 pre_opq[MTK_WDMA_CNT];
static u32 err_cnt[MTK_WDMA_CNT];
u32 i = 0, cur_drx = 0, rx_busy = 0, err_flag = 0;
u32 cur_opq = 0;
for (i = 0; i < MTK_WDMA_CNT; i++) {
cur_drx = mtk_r32(eth, MTK_WDMA_DRX_PTR(i));
rx_busy = mtk_r32(eth, MTK_WDMA_GLO_CFG(i)) & MTK_RX_DMA_BUSY;
if (i == 0)
cur_opq = (mtk_r32(eth, MTK_PSE_OQ_STA(5)) & 0x1FF);
else if (i == 1)
cur_opq = (mtk_r32(eth, MTK_PSE_OQ_STA(5)) & 0x1FF0000);
else
cur_opq = (mtk_r32(eth, MTK_PSE_OQ_STA(7)) & 0x1FF0000);
if (cur_drx == pre_drx[i] && rx_busy && cur_opq != 0 &&
cur_opq == pre_opq[i]) {
err_cnt[i]++;
if (err_cnt[i] >= 3) {
pr_info("WDMA %d Info\n", i);
pr_info("err_cnt = %d", err_cnt[i]);
pr_info("prev_drx = 0x%x | cur_drx = 0x%x\n",
pre_drx[i], cur_drx);
pr_info("WDMA_CRX_PTR = 0x%x\n",
mtk_r32(eth, MTK_WDMA_CRX_PTR(i)));
pr_info("WDMA_DRX_PTR = 0x%x\n",
mtk_r32(eth, MTK_WDMA_DRX_PTR(i)));
pr_info("WDMA_GLO_CFG = 0x%x\n",
mtk_r32(eth, MTK_WDMA_GLO_CFG(i)));
pr_info("==============================\n");
err_flag = 1;
}
} else
err_cnt[i] = 0;
pre_drx[i] = cur_drx;
pre_opq[i] = cur_opq;
}
if (err_flag)
return MTK_FE_START_RESET;
else
return 0;
}
u32 mtk_monitor_rx_fc(struct mtk_eth *eth)
{
u32 i = 0, mib_base = 0, gdm_fc = 0;
for (i = 0; i < MTK_MAC_COUNT; i++) {
mib_base = MTK_GDM1_TX_GBCNT + MTK_STAT_OFFSET*i;
gdm_fc = mtk_r32(eth, mib_base);
if (gdm_fc < 1)
return 1;
}
return 0;
}
u32 mtk_monitor_qdma_tx(struct mtk_eth *eth)
{
static u32 err_cnt_qtx;
u32 err_flag = 0;
u32 i = 0, is_rx_fc = 0;
u32 is_qfsm_hang = (mtk_r32(eth, MTK_QDMA_FSM) & 0xF00) != 0;
u32 is_qfwd_hang = mtk_r32(eth, MTK_QDMA_FWD_CNT) == 0;
is_rx_fc = mtk_monitor_rx_fc(eth);
if (is_qfsm_hang && is_qfwd_hang && is_rx_fc) {
err_cnt_qtx++;
if (err_cnt_qtx >= 3) {
pr_info("QDMA Tx Info\n");
pr_info("err_cnt = %d", err_cnt_qtx);
pr_info("is_qfsm_hang = %d\n", is_qfsm_hang);
pr_info("is_qfwd_hang = %d\n", is_qfwd_hang);
pr_info("-- -- -- -- -- -- --\n");
pr_info("MTK_QDMA_FSM = 0x%x\n",
mtk_r32(eth, MTK_QDMA_FSM));
pr_info("MTK_QDMA_FWD_CNT = 0x%x\n",
mtk_r32(eth, MTK_QDMA_FWD_CNT));
pr_info("MTK_QDMA_FQ_CNT = 0x%x\n",
mtk_r32(eth, MTK_QDMA_FQ_CNT));
pr_info("==============================\n");
err_flag = 1;
}
} else
err_cnt_qtx = 0;
if (err_flag)
return MTK_FE_STOP_TRAFFIC;
else
return 0;
}
u32 mtk_monitor_qdma_rx(struct mtk_eth *eth)
{
static u32 err_cnt_qrx;
static u32 pre_fq_head, pre_fq_tail;
u32 err_flag = 0;
u32 qrx_fsm = (mtk_r32(eth, MTK_QDMA_FSM) & 0x1F) == 9;
u32 fq_head = mtk_r32(eth, MTK_QDMA_FQ_HEAD);
u32 fq_tail = mtk_r32(eth, MTK_QDMA_FQ_TAIL);
if (qrx_fsm && fq_head == pre_fq_head &&
fq_tail == pre_fq_tail) {
err_cnt_qrx++;
if (err_cnt_qrx >= 3) {
pr_info("QDMA Rx Info\n");
pr_info("err_cnt = %d", err_cnt_qrx);
pr_info("MTK_QDMA_FSM = %d\n",
mtk_r32(eth, MTK_QDMA_FSM));
pr_info("FQ_HEAD = 0x%x\n",
mtk_r32(eth, MTK_QDMA_FQ_HEAD));
pr_info("FQ_TAIL = 0x%x\n",
mtk_r32(eth, MTK_QDMA_FQ_TAIL));
err_flag = 1;
} else
err_cnt_qrx = 0;
}
pre_fq_head = fq_head;
pre_fq_tail = fq_tail;
if (err_flag)
return MTK_FE_STOP_TRAFFIC;
else
return 0;
}
u32 mtk_monitor_adma_rx(struct mtk_eth *eth)
{
static u32 err_cnt_arx;
u32 err_flag = 0;
u32 opq0 = (mtk_r32(eth, MTK_PSE_OQ_STA(0)) & 0x1FF) != 0;
u32 cdm1_fsm = (mtk_r32(eth, MTK_FE_CDM1_FSM) & 0xFFFF0000) != 0;
u32 cur_stat = ((mtk_r32(eth, MTK_ADMA_RX_DBG0) & 0x1F) == 0);
u32 fifo_rdy = ((mtk_r32(eth, MTK_ADMA_RX_DBG0) & 0x40) == 0);
if (opq0 && cdm1_fsm && cur_stat && fifo_rdy) {
err_cnt_arx++;
if (err_cnt_arx >= 3) {
pr_info("ADMA Rx Info\n");
pr_info("err_cnt = %d", err_cnt_arx);
pr_info("CDM1_FSM = %d\n",
mtk_r32(eth, MTK_FE_CDM1_FSM));
pr_info("MTK_PSE_OQ_STA1 = 0x%x\n",
mtk_r32(eth, MTK_PSE_OQ_STA(0)));
pr_info("MTK_ADMA_RX_DBG0 = 0x%x\n",
mtk_r32(eth, MTK_ADMA_RX_DBG0));
pr_info("MTK_ADMA_RX_DBG1 = 0x%x\n",
mtk_r32(eth, MTK_ADMA_RX_DBG1));
pr_info("==============================\n");
err_flag = 1;
}
} else
err_cnt_arx = 0;
if (err_flag)
return MTK_FE_STOP_TRAFFIC;
else
return 0;
}
u32 mtk_monitor_tdma_tx(struct mtk_eth *eth)
{
static u32 err_cnt_ttx;
static u32 pre_fsm;
u32 err_flag = 0;
u32 cur_fsm = 0;
u32 tx_busy = 0;
if (MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V3)) {
cur_fsm = (mtk_r32(eth, MTK_FE_CDM6_FSM) & 0x1FFF) != 0;
tx_busy = ((mtk_r32(eth, MTK_TDMA_GLO_CFG) & 0x2) != 0);
if (cur_fsm == pre_fsm && cur_fsm != 0 && tx_busy) {
err_cnt_ttx++;
if (err_cnt_ttx >= 3) {
pr_info("TDMA Tx Info\n");
pr_info("err_cnt = %d", err_cnt_ttx);
pr_info("CDM6_FSM = %d\n",
mtk_r32(eth, MTK_FE_CDM6_FSM));
pr_info("DMA CFG = 0x%x\n",
mtk_r32(eth, MTK_TDMA_GLO_CFG));
pr_info("==============================\n");
err_flag = 1;
}
} else
err_cnt_ttx = 0;
pre_fsm = cur_fsm;
}
if (err_flag)
return MTK_FE_STOP_TRAFFIC;
else
return 0;
}
u32 mtk_monitor_tdma_rx(struct mtk_eth *eth)
{
static u32 err_cnt_trx;
static u32 pre_fsm;
u32 err_flag = 0;
u32 cur_fsm = 0;
u32 rx_busy = 0;
if (MTK_HAS_CAPS(eth->soc->caps, MTK_NETSYS_V3)) {
cur_fsm = (mtk_r32(eth, MTK_FE_CDM6_FSM) & 0xFFF0000) != 0;
rx_busy = ((mtk_r32(eth, MTK_TDMA_GLO_CFG) & 0x8) != 0);
if (cur_fsm == pre_fsm && cur_fsm != 0 && rx_busy) {
err_cnt_trx++;
if (err_cnt_trx >= 3) {
pr_info("TDMA Rx Info\n");
pr_info("err_cnt = %d", err_cnt_trx);
pr_info("CDM6_FSM = %d\n",
mtk_r32(eth, MTK_FE_CDM6_FSM));
pr_info("DMA CFG = 0x%x\n",
mtk_r32(eth, MTK_TDMA_GLO_CFG));
pr_info("==============================\n");
err_flag = 1;
}
} else
err_cnt_trx = 0;
pre_fsm = cur_fsm;
}
if (err_flag)
return MTK_FE_STOP_TRAFFIC;
else
return 0;
}
static const mtk_monitor_xdma_func mtk_reset_monitor_func[] = {
[0] = mtk_monitor_wdma_tx,
[1] = mtk_monitor_wdma_rx,
[2] = mtk_monitor_qdma_tx,
[3] = mtk_monitor_qdma_rx,
[4] = mtk_monitor_adma_rx,
[5] = mtk_monitor_tdma_tx,
[6] = mtk_monitor_tdma_rx,
};
void mtk_dma_monitor(struct timer_list *t)
{
struct mtk_eth *eth = from_timer(eth, t, mtk_dma_monitor_timer);
u32 i = 0, ret = 0;
for (i = 0; i < 6; i++) {
ret = (*mtk_reset_monitor_func[i]) (eth);
if ((ret == MTK_FE_START_RESET) ||
(ret == MTK_FE_STOP_TRAFFIC)) {
if ((atomic_read(&reset_lock) == 0) &&
(atomic_read(&force) == 1)) {
mtk_reset_flag = ret;
schedule_work(&eth->pending_work);
}
break;
}
}
mod_timer(&eth->mtk_dma_monitor_timer, jiffies + 1 * HZ);
}
void mtk_prepare_reset_fe(struct mtk_eth *eth)
{
u32 i = 0, val = 0, mcr = 0;
/* Disable NETSYS Interrupt */
mtk_w32(eth, 0, MTK_FE_INT_ENABLE);
mtk_w32(eth, 0, MTK_PDMA_INT_MASK);
mtk_w32(eth, 0, MTK_QDMA_INT_MASK);
/* Disable Linux netif Tx path */
for (i = 0; i < MTK_MAC_COUNT; i++) {
if (!eth->netdev[i])
continue;
netif_tx_disable(eth->netdev[i]);
}
/* Disable QDMA Tx */
val = mtk_r32(eth, MTK_QDMA_GLO_CFG);
mtk_w32(eth, val & ~(MTK_TX_DMA_EN), MTK_QDMA_GLO_CFG);
for (i = 0; i < MTK_MAC_COUNT; i++) {
pr_info("[%s] i:%d type:%d id:%d\n",
__func__, i, eth->mac[i]->type, eth->mac[i]->id);
if (eth->mac[i]->type == MTK_XGDM_TYPE &&
eth->mac[i]->id != MTK_GMAC1_ID) {
mcr = mtk_r32(eth, MTK_XMAC_MCR(eth->mac[i]->id));
mcr &= 0xfffffff0;
mcr |= XMAC_MCR_TRX_DISABLE;
pr_info("disable XMAC TX/RX\n");
mtk_w32(eth, mcr, MTK_XMAC_MCR(eth->mac[i]->id));
}
if (eth->mac[i]->type == MTK_GDM_TYPE) {
mcr = mtk_r32(eth, MTK_MAC_MCR(eth->mac[i]->id));
mcr &= ~(MAC_MCR_TX_EN | MAC_MCR_RX_EN);
mtk_w32(eth, mcr, MTK_MAC_MCR(eth->mac[i]->id));
pr_info("disable GMAC TX/RX\n");
}
}
/* Enable GDM drop */
for (i = 0; i < MTK_MAC_COUNT; i++)
mtk_gdm_config(eth, i, MTK_GDMA_DROP_ALL);
/* Disable ADMA Rx */
val = mtk_r32(eth, MTK_PDMA_GLO_CFG);
mtk_w32(eth, val & ~(MTK_RX_DMA_EN), MTK_PDMA_GLO_CFG);
}
void mtk_prepare_reset_ppe(struct mtk_eth *eth, u32 ppe_id)
{
u32 i = 0, poll_time = 5000, val;
/* Disable KA */
mtk_m32(eth, MTK_PPE_KA_CFG_MASK, 0, MTK_PPE_TB_CFG(ppe_id));
mtk_m32(eth, MTK_PPE_NTU_KA_MASK, 0, MTK_PPE_BIND_LMT_1(ppe_id));
mtk_w32(eth, 0, MTK_PPE_KA(ppe_id));
mdelay(10);
/* Set KA timer to maximum */
mtk_m32(eth, MTK_PPE_NTU_KA_MASK, (0xFF << 16), MTK_PPE_BIND_LMT_1(ppe_id));
mtk_w32(eth, 0xFFFFFFFF, MTK_PPE_KA(ppe_id));
/* Set KA tick select */
mtk_m32(eth, MTK_PPE_TICK_SEL_MASK, (0x1 << 24), MTK_PPE_TB_CFG(ppe_id));
mtk_m32(eth, MTK_PPE_KA_CFG_MASK, (0x3 << 12), MTK_PPE_TB_CFG(ppe_id));
mdelay(10);
/* Disable scan mode */
mtk_m32(eth, MTK_PPE_SCAN_MODE_MASK, 0, MTK_PPE_TB_CFG(ppe_id));
mdelay(10);
/* Check PPE idle */
while (i++ < poll_time) {
val = mtk_r32(eth, MTK_PPE_GLO_CFG(ppe_id));
if (!(val & MTK_PPE_BUSY))
break;
mdelay(1);
}
if (i >= poll_time) {
pr_info("[%s] PPE keeps busy !\n", __func__);
mtk_dump_reg(eth, "FE", 0x0, 0x500);
mtk_dump_reg(eth, "PPE", 0x2200, 0x200);
}
}
static int mtk_eth_netdevice_event(struct notifier_block *unused,
unsigned long event, void *ptr)
{
switch (event) {
case MTK_WIFI_RESET_DONE:
case MTK_FE_STOP_TRAFFIC_DONE:
pr_info("%s rcv done event:%x\n", __func__, event);
mtk_rest_cnt--;
if(!mtk_rest_cnt) {
complete(&wait_ser_done);
mtk_rest_cnt = mtk_wifi_num;
}
break;
case MTK_WIFI_CHIP_ONLINE:
mtk_wifi_num++;
mtk_rest_cnt = mtk_wifi_num;
break;
case MTK_WIFI_CHIP_OFFLINE:
mtk_wifi_num--;
mtk_rest_cnt = mtk_wifi_num;
break;
case MTK_FE_STOP_TRAFFIC_DONE_FAIL:
mtk_stop_fail = true;
mtk_reset_flag = MTK_FE_START_RESET;
pr_info("%s rcv done event:%x\n", __func__, event);
complete(&wait_ser_done);
mtk_rest_cnt = mtk_wifi_num;
break;
default:
break;
}
return NOTIFY_DONE;
}
struct notifier_block mtk_eth_netdevice_nb __read_mostly = {
.notifier_call = mtk_eth_netdevice_event,
};