Aaron Williams | 737ce25 | 2022-04-07 09:11:37 +0200 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0 |
| 2 | /* |
| 3 | * Copyright (C) 2018-2022 Marvell International Ltd. |
| 4 | */ |
| 5 | |
| 6 | #include <errno.h> |
| 7 | #include <log.h> |
| 8 | #include <time.h> |
| 9 | #include <linux/delay.h> |
| 10 | |
| 11 | #include <mach/cvmx-regs.h> |
| 12 | #include <mach/cvmx-csr.h> |
| 13 | #include <mach/cvmx-bootmem.h> |
| 14 | #include <mach/octeon-model.h> |
| 15 | #include <mach/cvmx-fuse.h> |
| 16 | #include <mach/octeon-feature.h> |
| 17 | #include <mach/cvmx-qlm.h> |
| 18 | #include <mach/octeon_qlm.h> |
| 19 | #include <mach/cvmx-pcie.h> |
| 20 | #include <mach/cvmx-coremask.h> |
| 21 | |
| 22 | #include <mach/cvmx-agl-defs.h> |
| 23 | #include <mach/cvmx-bgxx-defs.h> |
| 24 | #include <mach/cvmx-ciu-defs.h> |
| 25 | #include <mach/cvmx-gmxx-defs.h> |
| 26 | #include <mach/cvmx-gserx-defs.h> |
| 27 | #include <mach/cvmx-ilk-defs.h> |
| 28 | #include <mach/cvmx-ipd-defs.h> |
| 29 | #include <mach/cvmx-pcsx-defs.h> |
| 30 | #include <mach/cvmx-pcsxx-defs.h> |
| 31 | #include <mach/cvmx-pki-defs.h> |
| 32 | #include <mach/cvmx-pko-defs.h> |
| 33 | #include <mach/cvmx-xcv-defs.h> |
| 34 | |
| 35 | #include <mach/cvmx-hwpko.h> |
| 36 | #include <mach/cvmx-ilk.h> |
| 37 | #include <mach/cvmx-pki.h> |
| 38 | #include <mach/cvmx-pko3.h> |
| 39 | #include <mach/cvmx-pko3-queue.h> |
| 40 | #include <mach/cvmx-pko3-resources.h> |
| 41 | |
| 42 | #include <mach/cvmx-helper.h> |
| 43 | #include <mach/cvmx-helper-board.h> |
| 44 | #include <mach/cvmx-helper-cfg.h> |
| 45 | |
| 46 | #include <mach/cvmx-helper-bgx.h> |
| 47 | #include <mach/cvmx-helper-cfg.h> |
| 48 | #include <mach/cvmx-helper-util.h> |
| 49 | #include <mach/cvmx-helper-pki.h> |
| 50 | |
| 51 | static const int debug; |
| 52 | |
| 53 | #define CVMX_DUMP_REGX(reg) \ |
| 54 | if (debug) \ |
| 55 | debug("%s=%#llx\n", #reg, (long long)csr_rd_node(node, reg)) |
| 56 | |
| 57 | static int cvmx_pko_setup_macs(int node); |
| 58 | |
| 59 | /* |
| 60 | * PKO descriptor queue operation error string |
| 61 | * |
| 62 | * @param dqstatus is the enumeration returned from hardware, |
| 63 | * PKO_QUERY_RTN_S[DQSTATUS]. |
| 64 | * |
| 65 | * @return static constant string error description |
| 66 | */ |
| 67 | const char *pko_dqstatus_error(pko_query_dqstatus_t dqstatus) |
| 68 | { |
| 69 | char *str = "PKO Undefined error"; |
| 70 | |
| 71 | switch (dqstatus) { |
| 72 | case PKO_DQSTATUS_PASS: |
| 73 | str = "No error"; |
| 74 | break; |
| 75 | case PKO_DQSTATUS_BADSTATE: |
| 76 | str = "PKO queue not ready"; |
| 77 | break; |
| 78 | case PKO_DQSTATUS_NOFPABUF: |
| 79 | str = "PKO failed to allocate buffer from FPA"; |
| 80 | break; |
| 81 | case PKO_DQSTATUS_NOPKOBUF: |
| 82 | str = "PKO out of buffers"; |
| 83 | break; |
| 84 | case PKO_DQSTATUS_FAILRTNPTR: |
| 85 | str = "PKO failed to return buffer to FPA"; |
| 86 | break; |
| 87 | case PKO_DQSTATUS_ALREADY: |
| 88 | str = "PKO queue already opened"; |
| 89 | break; |
| 90 | case PKO_DQSTATUS_NOTCREATED: |
| 91 | str = "PKO queue has not been created"; |
| 92 | break; |
| 93 | case PKO_DQSTATUS_NOTEMPTY: |
| 94 | str = "PKO queue is not empty"; |
| 95 | break; |
| 96 | case PKO_DQSTATUS_SENDPKTDROP: |
| 97 | str = "Illegal PKO command construct"; |
| 98 | break; |
| 99 | } |
| 100 | return str; |
| 101 | } |
| 102 | |
| 103 | /* |
| 104 | * PKO global initialization for 78XX. |
| 105 | * |
| 106 | * @param node is the node on which PKO block is initialized. |
| 107 | * @return none. |
| 108 | */ |
| 109 | int cvmx_pko3_hw_init_global(int node, uint16_t aura) |
| 110 | { |
| 111 | cvmx_pko_dpfi_flush_t pko_flush; |
| 112 | cvmx_pko_dpfi_fpa_aura_t pko_aura; |
| 113 | cvmx_pko_dpfi_ena_t dpfi_enable; |
| 114 | cvmx_pko_ptf_iobp_cfg_t ptf_iobp_cfg; |
| 115 | cvmx_pko_pdm_cfg_t pko_pdm_cfg; |
| 116 | cvmx_pko_enable_t pko_enable; |
| 117 | cvmx_pko_dpfi_status_t dpfi_status; |
| 118 | cvmx_pko_status_t pko_status; |
| 119 | cvmx_pko_shaper_cfg_t shaper_cfg; |
| 120 | u64 cycles; |
| 121 | const unsigned int timeout = 100; /* 100 milliseconds */ |
| 122 | |
| 123 | if (node != (aura >> 10)) |
| 124 | cvmx_printf("WARNING: AURA vs PKO node mismatch\n"); |
| 125 | |
| 126 | pko_enable.u64 = csr_rd_node(node, CVMX_PKO_ENABLE); |
| 127 | if (pko_enable.s.enable) { |
| 128 | cvmx_printf("WARNING: %s: PKO already enabled on node %u\n", |
| 129 | __func__, node); |
| 130 | return 0; |
| 131 | } |
| 132 | /* Enable color awareness. */ |
| 133 | shaper_cfg.u64 = csr_rd_node(node, CVMX_PKO_SHAPER_CFG); |
| 134 | shaper_cfg.s.color_aware = 1; |
| 135 | csr_wr_node(node, CVMX_PKO_SHAPER_CFG, shaper_cfg.u64); |
| 136 | |
| 137 | /* Clear FLUSH command to be sure */ |
| 138 | pko_flush.u64 = 0; |
| 139 | pko_flush.s.flush_en = 0; |
| 140 | csr_wr_node(node, CVMX_PKO_DPFI_FLUSH, pko_flush.u64); |
| 141 | |
| 142 | /* set the aura number in pko, use aura node from parameter */ |
| 143 | pko_aura.u64 = 0; |
| 144 | pko_aura.s.node = aura >> 10; |
| 145 | pko_aura.s.laura = aura; |
| 146 | csr_wr_node(node, CVMX_PKO_DPFI_FPA_AURA, pko_aura.u64); |
| 147 | |
| 148 | CVMX_DUMP_REGX(CVMX_PKO_DPFI_FPA_AURA); |
| 149 | |
| 150 | dpfi_enable.u64 = 0; |
| 151 | dpfi_enable.s.enable = 1; |
| 152 | csr_wr_node(node, CVMX_PKO_DPFI_ENA, dpfi_enable.u64); |
| 153 | |
| 154 | /* Prepare timeout */ |
| 155 | cycles = get_timer(0); |
| 156 | |
| 157 | /* Wait until all pointers have been returned */ |
| 158 | do { |
| 159 | pko_status.u64 = csr_rd_node(node, CVMX_PKO_STATUS); |
| 160 | if (get_timer(cycles) > timeout) |
| 161 | break; |
| 162 | } while (!pko_status.s.pko_rdy); |
| 163 | |
| 164 | if (!pko_status.s.pko_rdy) { |
| 165 | dpfi_status.u64 = csr_rd_node(node, CVMX_PKO_DPFI_STATUS); |
| 166 | cvmx_printf("ERROR: %s: PKO DFPI failed, PKO_STATUS=%#llx DPFI_STATUS=%#llx\n", |
| 167 | __func__, (unsigned long long)pko_status.u64, |
| 168 | (unsigned long long)dpfi_status.u64); |
| 169 | return -1; |
| 170 | } |
| 171 | |
| 172 | /* Set max outstanding requests in IOBP for any FIFO.*/ |
| 173 | ptf_iobp_cfg.u64 = csr_rd_node(node, CVMX_PKO_PTF_IOBP_CFG); |
| 174 | if (OCTEON_IS_MODEL(OCTEON_CN78XX)) |
| 175 | ptf_iobp_cfg.s.max_read_size = 0x10; /* Recommended by HRM.*/ |
| 176 | else |
| 177 | /* Reduce the value from recommended 0x10 to avoid |
| 178 | * getting "underflow" condition in the BGX TX FIFO. |
| 179 | */ |
| 180 | ptf_iobp_cfg.s.max_read_size = 3; |
| 181 | csr_wr_node(node, CVMX_PKO_PTF_IOBP_CFG, ptf_iobp_cfg.u64); |
| 182 | |
| 183 | /* Set minimum packet size per Ethernet standard */ |
| 184 | pko_pdm_cfg.u64 = 0; |
| 185 | pko_pdm_cfg.s.pko_pad_minlen = 0x3c; /* 60 bytes before FCS */ |
| 186 | csr_wr_node(node, CVMX_PKO_PDM_CFG, pko_pdm_cfg.u64); |
| 187 | |
| 188 | /* Initialize MACs and FIFOs */ |
| 189 | cvmx_pko_setup_macs(node); |
| 190 | |
| 191 | /* enable PKO, although interfaces and queues are not up yet */ |
| 192 | pko_enable.u64 = 0; |
| 193 | pko_enable.s.enable = 1; |
| 194 | csr_wr_node(node, CVMX_PKO_ENABLE, pko_enable.u64); |
| 195 | |
| 196 | /* PKO_RDY set indicates successful initialization */ |
| 197 | pko_status.u64 = csr_rd_node(node, CVMX_PKO_STATUS); |
| 198 | if (pko_status.s.pko_rdy) |
| 199 | return 0; |
| 200 | |
| 201 | cvmx_printf("ERROR: %s: failed, PKO_STATUS=%#llx\n", __func__, |
| 202 | (unsigned long long)pko_status.u64); |
| 203 | return -1; |
| 204 | } |
| 205 | |
| 206 | /* |
| 207 | * Configure Channel credit level in PKO. |
| 208 | * |
| 209 | * @param node is to specify the node to which this configuration is applied. |
| 210 | * @param level specifies the level at which pko channel queues will be configured, |
| 211 | * @return returns 0 if successful and -1 on failure. |
| 212 | */ |
| 213 | int cvmx_pko3_channel_credit_level(int node, enum cvmx_pko3_level_e level) |
| 214 | { |
| 215 | union cvmx_pko_channel_level channel_level; |
| 216 | |
| 217 | channel_level.u64 = 0; |
| 218 | |
| 219 | if (level == CVMX_PKO_L2_QUEUES) |
| 220 | channel_level.s.cc_level = 0; |
| 221 | else if (level == CVMX_PKO_L3_QUEUES) |
| 222 | channel_level.s.cc_level = 1; |
| 223 | else |
| 224 | return -1; |
| 225 | |
| 226 | csr_wr_node(node, CVMX_PKO_CHANNEL_LEVEL, channel_level.u64); |
| 227 | |
| 228 | return 0; |
| 229 | } |
| 230 | |
| 231 | /** Open configured descriptor queues before queueing packets into them. |
| 232 | * |
| 233 | * @param node is to specify the node to which this configuration is applied. |
| 234 | * @param dq is the descriptor queue number to be opened. |
| 235 | * @return returns 0 on success or -1 on failure. |
| 236 | */ |
| 237 | int cvmx_pko_dq_open(int node, int dq) |
| 238 | { |
| 239 | cvmx_pko_query_rtn_t pko_status; |
| 240 | pko_query_dqstatus_t dqstatus; |
| 241 | cvmx_pko3_dq_params_t *p_param; |
| 242 | |
| 243 | if (debug) |
| 244 | debug("%s: DEBUG: dq %u\n", __func__, dq); |
| 245 | |
| 246 | __cvmx_pko3_dq_param_setup(node); |
| 247 | |
| 248 | pko_status = __cvmx_pko3_do_dma(node, dq, NULL, 0, CVMX_PKO_DQ_OPEN); |
| 249 | |
| 250 | dqstatus = pko_status.s.dqstatus; |
| 251 | |
| 252 | if (dqstatus == PKO_DQSTATUS_ALREADY) |
| 253 | return 0; |
| 254 | if (dqstatus != PKO_DQSTATUS_PASS) { |
| 255 | cvmx_printf("%s: ERROR: Failed to open dq :%u: %s\n", __func__, |
| 256 | dq, pko_dqstatus_error(dqstatus)); |
| 257 | return -1; |
| 258 | } |
| 259 | |
| 260 | /* Setup the descriptor queue software parameters */ |
| 261 | p_param = cvmx_pko3_dq_parameters(node, dq); |
| 262 | if (p_param) { |
| 263 | p_param->depth = pko_status.s.depth; |
| 264 | if (p_param->limit == 0) |
| 265 | p_param->limit = 1024; /* last-resort default */ |
| 266 | } |
| 267 | |
| 268 | return 0; |
| 269 | } |
| 270 | |
| 271 | /* |
| 272 | * PKO initialization of MACs and FIFOs |
| 273 | * |
| 274 | * All MACs are configured and assigned a specific FIFO, |
| 275 | * and each FIFO is configured with size for a best utilization |
| 276 | * of available FIFO resources. |
| 277 | * |
| 278 | * @param node is to specify which node's pko block for this setup. |
| 279 | * @return returns 0 if successful and -1 on failure. |
| 280 | * |
| 281 | * Note: This function contains model-specific code. |
| 282 | */ |
| 283 | static int cvmx_pko_setup_macs(int node) |
| 284 | { |
| 285 | unsigned int interface; |
| 286 | unsigned int port, num_ports; |
| 287 | unsigned int mac_num, fifo, pri, cnt; |
| 288 | cvmx_helper_interface_mode_t mode; |
| 289 | const unsigned int num_interfaces = |
| 290 | cvmx_helper_get_number_of_interfaces(); |
| 291 | u8 fifo_group_cfg[8]; |
| 292 | u8 fifo_group_spd[8]; |
| 293 | unsigned int fifo_count = 0; |
| 294 | unsigned int max_fifos = 0, fifo_groups = 0; |
| 295 | struct { |
| 296 | u8 fifo_cnt; |
| 297 | u8 fifo_id; |
| 298 | u8 pri; |
| 299 | u8 spd; |
| 300 | u8 mac_fifo_cnt; |
| 301 | } cvmx_pko3_mac_table[32]; |
| 302 | |
| 303 | if (OCTEON_IS_MODEL(OCTEON_CN78XX)) { |
| 304 | max_fifos = 28; /* exclusive of NULL FIFO */ |
| 305 | fifo_groups = 8; /* inclusive of NULL PTGF */ |
| 306 | } |
| 307 | if (OCTEON_IS_MODEL(OCTEON_CN73XX) || OCTEON_IS_MODEL(OCTEON_CNF75XX)) { |
| 308 | max_fifos = 16; |
| 309 | fifo_groups = 5; |
| 310 | } |
| 311 | |
| 312 | /* Initialize FIFO allocation table */ |
| 313 | memset(&fifo_group_cfg, 0, sizeof(fifo_group_cfg)); |
| 314 | memset(&fifo_group_spd, 0, sizeof(fifo_group_spd)); |
| 315 | memset(cvmx_pko3_mac_table, 0, sizeof(cvmx_pko3_mac_table)); |
| 316 | |
| 317 | /* Initialize all MACs as disabled */ |
| 318 | for (mac_num = 0; mac_num < __cvmx_pko3_num_macs(); mac_num++) { |
| 319 | cvmx_pko3_mac_table[mac_num].pri = 0; |
| 320 | cvmx_pko3_mac_table[mac_num].fifo_cnt = 0; |
| 321 | cvmx_pko3_mac_table[mac_num].fifo_id = 0x1f; |
| 322 | } |
| 323 | |
| 324 | for (interface = 0; interface < num_interfaces; interface++) { |
| 325 | int xiface = |
| 326 | cvmx_helper_node_interface_to_xiface(node, interface); |
| 327 | /* Interface type for ALL interfaces */ |
| 328 | mode = cvmx_helper_interface_get_mode(xiface); |
| 329 | num_ports = cvmx_helper_interface_enumerate(xiface); |
| 330 | |
| 331 | if (mode == CVMX_HELPER_INTERFACE_MODE_DISABLED) |
| 332 | continue; |
| 333 | /* |
| 334 | * Non-BGX interfaces: |
| 335 | * Each of these interfaces has a single MAC really. |
| 336 | */ |
| 337 | if (mode == CVMX_HELPER_INTERFACE_MODE_ILK || |
| 338 | mode == CVMX_HELPER_INTERFACE_MODE_NPI || |
| 339 | mode == CVMX_HELPER_INTERFACE_MODE_LOOP) |
| 340 | num_ports = 1; |
| 341 | |
| 342 | for (port = 0; port < num_ports; port++) { |
| 343 | int i; |
| 344 | |
| 345 | /* Get the per-port mode for BGX-interfaces */ |
| 346 | if (interface < CVMX_HELPER_MAX_GMX) |
| 347 | mode = cvmx_helper_bgx_get_mode(xiface, port); |
| 348 | /* In MIXED mode, LMACs can run different protocols */ |
| 349 | |
| 350 | /* convert interface/port to mac number */ |
| 351 | i = __cvmx_pko3_get_mac_num(xiface, port); |
| 352 | if (i < 0 || i >= (int)__cvmx_pko3_num_macs()) { |
| 353 | cvmx_printf("%s: ERROR: interface %d:%u port %d has no MAC %d/%d\n", |
| 354 | __func__, node, interface, port, i, |
| 355 | __cvmx_pko3_num_macs()); |
| 356 | continue; |
| 357 | } |
| 358 | |
| 359 | if (mode == CVMX_HELPER_INTERFACE_MODE_RXAUI) { |
| 360 | unsigned int bgx_fifo_size = |
| 361 | __cvmx_helper_bgx_fifo_size(xiface, |
| 362 | port); |
| 363 | |
| 364 | cvmx_pko3_mac_table[i].mac_fifo_cnt = |
| 365 | bgx_fifo_size / |
| 366 | (CVMX_BGX_TX_FIFO_SIZE / 4); |
| 367 | cvmx_pko3_mac_table[i].pri = 2; |
| 368 | cvmx_pko3_mac_table[i].spd = 10; |
| 369 | cvmx_pko3_mac_table[i].fifo_cnt = 2; |
| 370 | } else if (mode == CVMX_HELPER_INTERFACE_MODE_XLAUI) { |
| 371 | unsigned int bgx_fifo_size = |
| 372 | __cvmx_helper_bgx_fifo_size(xiface, |
| 373 | port); |
| 374 | |
| 375 | cvmx_pko3_mac_table[i].mac_fifo_cnt = |
| 376 | bgx_fifo_size / |
| 377 | (CVMX_BGX_TX_FIFO_SIZE / 4); |
| 378 | cvmx_pko3_mac_table[i].pri = 4; |
| 379 | cvmx_pko3_mac_table[i].spd = 40; |
| 380 | cvmx_pko3_mac_table[i].fifo_cnt = 4; |
| 381 | } else if (mode == CVMX_HELPER_INTERFACE_MODE_XAUI) { |
| 382 | unsigned int bgx_fifo_size = |
| 383 | __cvmx_helper_bgx_fifo_size(xiface, |
| 384 | port); |
| 385 | |
| 386 | cvmx_pko3_mac_table[i].mac_fifo_cnt = |
| 387 | bgx_fifo_size / |
| 388 | (CVMX_BGX_TX_FIFO_SIZE / 4); |
| 389 | cvmx_pko3_mac_table[i].pri = 3; |
| 390 | cvmx_pko3_mac_table[i].fifo_cnt = 4; |
| 391 | /* DXAUI at 20G, or XAU at 10G */ |
| 392 | cvmx_pko3_mac_table[i].spd = 20; |
| 393 | } else if (mode == CVMX_HELPER_INTERFACE_MODE_XFI) { |
| 394 | unsigned int bgx_fifo_size = |
| 395 | __cvmx_helper_bgx_fifo_size(xiface, |
| 396 | port); |
| 397 | |
| 398 | cvmx_pko3_mac_table[i].mac_fifo_cnt = |
| 399 | bgx_fifo_size / |
| 400 | (CVMX_BGX_TX_FIFO_SIZE / 4); |
| 401 | cvmx_pko3_mac_table[i].pri = 3; |
| 402 | cvmx_pko3_mac_table[i].fifo_cnt = 4; |
| 403 | cvmx_pko3_mac_table[i].spd = 10; |
| 404 | } else if (mode == CVMX_HELPER_INTERFACE_MODE_LOOP) { |
| 405 | cvmx_pko3_mac_table[i].fifo_cnt = 1; |
| 406 | cvmx_pko3_mac_table[i].pri = 1; |
| 407 | cvmx_pko3_mac_table[i].spd = 1; |
| 408 | cvmx_pko3_mac_table[i].mac_fifo_cnt = 1; |
| 409 | } else if (mode == CVMX_HELPER_INTERFACE_MODE_ILK || |
| 410 | mode == CVMX_HELPER_INTERFACE_MODE_SRIO) { |
| 411 | cvmx_pko3_mac_table[i].fifo_cnt = 4; |
| 412 | cvmx_pko3_mac_table[i].pri = 3; |
| 413 | /* ILK/SRIO: speed depends on lane count */ |
| 414 | cvmx_pko3_mac_table[i].spd = 40; |
| 415 | cvmx_pko3_mac_table[i].mac_fifo_cnt = 4; |
| 416 | } else if (mode == CVMX_HELPER_INTERFACE_MODE_NPI) { |
| 417 | cvmx_pko3_mac_table[i].fifo_cnt = 4; |
| 418 | cvmx_pko3_mac_table[i].pri = 2; |
| 419 | /* Actual speed depends on PCIe lanes/mode */ |
| 420 | cvmx_pko3_mac_table[i].spd = 50; |
| 421 | /* SLI Tx FIFO size to be revisitted */ |
| 422 | cvmx_pko3_mac_table[i].mac_fifo_cnt = 1; |
| 423 | } else { |
| 424 | /* Other BGX interface modes: SGMII/RGMII */ |
| 425 | unsigned int bgx_fifo_size = |
| 426 | __cvmx_helper_bgx_fifo_size(xiface, |
| 427 | port); |
| 428 | |
| 429 | cvmx_pko3_mac_table[i].mac_fifo_cnt = |
| 430 | bgx_fifo_size / |
| 431 | (CVMX_BGX_TX_FIFO_SIZE / 4); |
| 432 | cvmx_pko3_mac_table[i].fifo_cnt = 1; |
| 433 | cvmx_pko3_mac_table[i].pri = 1; |
| 434 | cvmx_pko3_mac_table[i].spd = 1; |
| 435 | } |
| 436 | |
| 437 | if (debug) |
| 438 | debug("%s: intf %d:%u port %u %s mac %02u cnt %u macfifo %uk spd %u\n", |
| 439 | __func__, node, interface, port, |
| 440 | cvmx_helper_interface_mode_to_string(mode), |
| 441 | i, cvmx_pko3_mac_table[i].fifo_cnt, |
| 442 | cvmx_pko3_mac_table[i].mac_fifo_cnt * 8, |
| 443 | cvmx_pko3_mac_table[i].spd); |
| 444 | |
| 445 | } /* for port */ |
| 446 | } /* for interface */ |
| 447 | |
| 448 | /* Count the number of requested FIFOs */ |
| 449 | for (fifo_count = mac_num = 0; mac_num < __cvmx_pko3_num_macs(); |
| 450 | mac_num++) |
| 451 | fifo_count += cvmx_pko3_mac_table[mac_num].fifo_cnt; |
| 452 | |
| 453 | if (debug) |
| 454 | debug("%s: initially requested FIFO count %u\n", __func__, |
| 455 | fifo_count); |
| 456 | |
| 457 | /* Heuristically trim FIFO count to fit in available number */ |
| 458 | pri = 1; |
| 459 | cnt = 4; |
| 460 | while (fifo_count > max_fifos) { |
| 461 | for (mac_num = 0; mac_num < __cvmx_pko3_num_macs(); mac_num++) { |
| 462 | if (cvmx_pko3_mac_table[mac_num].fifo_cnt == cnt && |
| 463 | cvmx_pko3_mac_table[mac_num].pri <= pri) { |
| 464 | cvmx_pko3_mac_table[mac_num].fifo_cnt >>= 1; |
| 465 | fifo_count -= |
| 466 | cvmx_pko3_mac_table[mac_num].fifo_cnt; |
| 467 | } |
| 468 | if (fifo_count <= max_fifos) |
| 469 | break; |
| 470 | } |
| 471 | if (pri >= 4) { |
| 472 | pri = 1; |
| 473 | cnt >>= 1; |
| 474 | } else { |
| 475 | pri++; |
| 476 | } |
| 477 | if (cnt == 0) |
| 478 | break; |
| 479 | } |
| 480 | |
| 481 | if (debug) |
| 482 | debug("%s: adjusted FIFO count %u\n", __func__, fifo_count); |
| 483 | |
| 484 | /* Special case for NULL Virtual FIFO */ |
| 485 | fifo_group_cfg[fifo_groups - 1] = 0; |
| 486 | /* there is no MAC connected to NULL FIFO */ |
| 487 | |
| 488 | /* Configure MAC units, and attach a FIFO to each */ |
| 489 | for (fifo = 0, cnt = 4; cnt > 0; cnt >>= 1) { |
| 490 | unsigned int g; |
| 491 | |
| 492 | for (mac_num = 0; mac_num < __cvmx_pko3_num_macs(); mac_num++) { |
| 493 | if (cvmx_pko3_mac_table[mac_num].fifo_cnt < cnt || |
| 494 | cvmx_pko3_mac_table[mac_num].fifo_id != 0x1f) |
| 495 | continue; |
| 496 | |
| 497 | /* Attach FIFO to MAC */ |
| 498 | cvmx_pko3_mac_table[mac_num].fifo_id = fifo; |
| 499 | g = fifo >> 2; |
| 500 | /* Sum speed for FIFO group */ |
| 501 | fifo_group_spd[g] += cvmx_pko3_mac_table[mac_num].spd; |
| 502 | |
| 503 | if (cnt == 4) |
| 504 | fifo_group_cfg[g] = 4; /* 10k,0,0,0 */ |
| 505 | else if (cnt == 2 && (fifo & 0x3) == 0) |
| 506 | fifo_group_cfg[g] = 3; /* 5k,0,5k,0 */ |
| 507 | else if (cnt == 2 && fifo_group_cfg[g] == 3) |
| 508 | /* no change */; |
| 509 | else if (cnt == 1 && (fifo & 0x2) && |
| 510 | fifo_group_cfg[g] == 3) |
| 511 | fifo_group_cfg[g] = 1; /* 5k,0,2.5k 2.5k*/ |
| 512 | else if (cnt == 1 && (fifo & 0x3) == 0x3) |
| 513 | /* no change */; |
| 514 | else if (cnt == 1) |
| 515 | fifo_group_cfg[g] = 0; /* 2.5k x 4 */ |
| 516 | else |
| 517 | cvmx_printf("ERROR: %s: internal error\n", |
| 518 | __func__); |
| 519 | |
| 520 | fifo += cnt; |
| 521 | } |
| 522 | } |
| 523 | |
| 524 | /* Check if there was no error in FIFO allocation */ |
| 525 | if (fifo > max_fifos) { |
| 526 | cvmx_printf("ERROR: %s: Internal error FIFO %u\n", __func__, |
| 527 | fifo); |
| 528 | return -1; |
| 529 | } |
| 530 | |
| 531 | if (debug) |
| 532 | debug("%s: used %u of FIFOs\n", __func__, fifo); |
| 533 | |
| 534 | /* Now configure all FIFO groups */ |
| 535 | for (fifo = 0; fifo < fifo_groups; fifo++) { |
| 536 | cvmx_pko_ptgfx_cfg_t pko_ptgfx_cfg; |
| 537 | |
| 538 | pko_ptgfx_cfg.u64 = csr_rd_node(node, CVMX_PKO_PTGFX_CFG(fifo)); |
| 539 | if (pko_ptgfx_cfg.s.size != fifo_group_cfg[fifo]) |
| 540 | pko_ptgfx_cfg.s.reset = 1; |
| 541 | pko_ptgfx_cfg.s.size = fifo_group_cfg[fifo]; |
| 542 | if (fifo_group_spd[fifo] >= 40) |
| 543 | if (pko_ptgfx_cfg.s.size >= 3) |
| 544 | pko_ptgfx_cfg.s.rate = 3; /* 50 Gbps */ |
| 545 | else |
| 546 | pko_ptgfx_cfg.s.rate = 2; /* 25 Gbps */ |
| 547 | else if (fifo_group_spd[fifo] >= 20) |
| 548 | pko_ptgfx_cfg.s.rate = 2; /* 25 Gbps */ |
| 549 | else if (fifo_group_spd[fifo] >= 10) |
| 550 | pko_ptgfx_cfg.s.rate = 1; /* 12.5 Gbps */ |
| 551 | else |
| 552 | pko_ptgfx_cfg.s.rate = 0; /* 6.25 Gbps */ |
| 553 | |
| 554 | if (debug) |
| 555 | debug("%s: FIFO %#x-%#x size=%u speed=%d rate=%d\n", |
| 556 | __func__, fifo * 4, fifo * 4 + 3, |
| 557 | pko_ptgfx_cfg.s.size, fifo_group_spd[fifo], |
| 558 | pko_ptgfx_cfg.s.rate); |
| 559 | |
| 560 | csr_wr_node(node, CVMX_PKO_PTGFX_CFG(fifo), pko_ptgfx_cfg.u64); |
| 561 | pko_ptgfx_cfg.s.reset = 0; |
| 562 | csr_wr_node(node, CVMX_PKO_PTGFX_CFG(fifo), pko_ptgfx_cfg.u64); |
| 563 | } |
| 564 | |
| 565 | /* Configure all MACs assigned FIFO number */ |
| 566 | for (mac_num = 0; mac_num < __cvmx_pko3_num_macs(); mac_num++) { |
| 567 | cvmx_pko_macx_cfg_t pko_mac_cfg; |
| 568 | |
| 569 | if (debug) |
| 570 | debug("%s: mac#%02u: fifo=%#x cnt=%u speed=%d\n", |
| 571 | __func__, mac_num, |
| 572 | cvmx_pko3_mac_table[mac_num].fifo_id, |
| 573 | cvmx_pko3_mac_table[mac_num].fifo_cnt, |
| 574 | cvmx_pko3_mac_table[mac_num].spd); |
| 575 | |
| 576 | pko_mac_cfg.u64 = csr_rd_node(node, CVMX_PKO_MACX_CFG(mac_num)); |
| 577 | pko_mac_cfg.s.fifo_num = cvmx_pko3_mac_table[mac_num].fifo_id; |
| 578 | csr_wr_node(node, CVMX_PKO_MACX_CFG(mac_num), pko_mac_cfg.u64); |
| 579 | } |
| 580 | |
| 581 | /* Setup PKO MCI0/MCI1/SKID credits */ |
| 582 | for (mac_num = 0; mac_num < __cvmx_pko3_num_macs(); mac_num++) { |
| 583 | cvmx_pko_mci0_max_credx_t pko_mci0_max_cred; |
| 584 | cvmx_pko_mci1_max_credx_t pko_mci1_max_cred; |
| 585 | cvmx_pko_macx_cfg_t pko_mac_cfg; |
| 586 | unsigned int fifo_credit, mac_credit, skid_credit; |
| 587 | unsigned int pko_fifo_cnt, fifo_size; |
| 588 | unsigned int mac_fifo_cnt; |
| 589 | unsigned int tmp; |
| 590 | int saved_fifo_num; |
| 591 | |
| 592 | pko_fifo_cnt = cvmx_pko3_mac_table[mac_num].fifo_cnt; |
| 593 | mac_fifo_cnt = cvmx_pko3_mac_table[mac_num].mac_fifo_cnt; |
| 594 | |
| 595 | /* Skip unused MACs */ |
| 596 | if (pko_fifo_cnt == 0) |
| 597 | continue; |
| 598 | |
| 599 | /* Check for sanity */ |
| 600 | if (pko_fifo_cnt > 4) |
| 601 | pko_fifo_cnt = 1; |
| 602 | |
| 603 | fifo_size = (2 * 1024) + (1024 / 2); /* 2.5KiB */ |
| 604 | fifo_credit = pko_fifo_cnt * fifo_size; |
| 605 | |
| 606 | if (mac_num == 0) { |
| 607 | /* loopback */ |
| 608 | mac_credit = 4096; /* From HRM Sec 13.0 */ |
| 609 | skid_credit = 0; |
| 610 | } else if (mac_num == 1) { |
| 611 | /* DPI */ |
| 612 | mac_credit = 2 * 1024; |
| 613 | skid_credit = 0; |
| 614 | } else if (octeon_has_feature(OCTEON_FEATURE_ILK) && |
| 615 | (mac_num & 0xfe) == 2) { |
| 616 | /* ILK0, ILK1: MAC 2,3 */ |
| 617 | mac_credit = 4 * 1024; /* 4KB fifo */ |
| 618 | skid_credit = 0; |
| 619 | } else if (octeon_has_feature(OCTEON_FEATURE_SRIO) && |
| 620 | (mac_num >= 6) && (mac_num <= 9)) { |
| 621 | /* SRIO0, SRIO1: MAC 6..9 */ |
| 622 | mac_credit = 1024 / 2; |
| 623 | skid_credit = 0; |
| 624 | } else { |
| 625 | /* BGX */ |
| 626 | mac_credit = mac_fifo_cnt * 8 * 1024; |
| 627 | skid_credit = mac_fifo_cnt * 256; |
| 628 | } |
| 629 | |
| 630 | if (debug) |
| 631 | debug("%s: mac %u pko_fifo_credit=%u mac_credit=%u\n", |
| 632 | __func__, mac_num, fifo_credit, mac_credit); |
| 633 | |
| 634 | tmp = (fifo_credit + mac_credit) / 16; |
| 635 | pko_mci0_max_cred.u64 = 0; |
| 636 | pko_mci0_max_cred.s.max_cred_lim = tmp; |
| 637 | |
| 638 | /* Check for overflow */ |
| 639 | if (pko_mci0_max_cred.s.max_cred_lim != tmp) { |
| 640 | cvmx_printf("WARNING: %s: MCI0 credit overflow\n", |
| 641 | __func__); |
| 642 | pko_mci0_max_cred.s.max_cred_lim = 0xfff; |
| 643 | } |
| 644 | |
| 645 | /* Pass 2 PKO hardware does not use the MCI0 credits */ |
| 646 | if (OCTEON_IS_MODEL(OCTEON_CN78XX_PASS1_X)) |
| 647 | csr_wr_node(node, CVMX_PKO_MCI0_MAX_CREDX(mac_num), |
| 648 | pko_mci0_max_cred.u64); |
| 649 | |
| 650 | /* The original CSR formula is the correct one after all */ |
| 651 | tmp = (mac_credit) / 16; |
| 652 | pko_mci1_max_cred.u64 = 0; |
| 653 | pko_mci1_max_cred.s.max_cred_lim = tmp; |
| 654 | |
| 655 | /* Check for overflow */ |
| 656 | if (pko_mci1_max_cred.s.max_cred_lim != tmp) { |
| 657 | cvmx_printf("WARNING: %s: MCI1 credit overflow\n", |
| 658 | __func__); |
| 659 | pko_mci1_max_cred.s.max_cred_lim = 0xfff; |
| 660 | } |
| 661 | |
| 662 | csr_wr_node(node, CVMX_PKO_MCI1_MAX_CREDX(mac_num), |
| 663 | pko_mci1_max_cred.u64); |
| 664 | |
| 665 | tmp = (skid_credit / 256) >> 1; /* valid 0,1,2 */ |
| 666 | pko_mac_cfg.u64 = csr_rd_node(node, CVMX_PKO_MACX_CFG(mac_num)); |
| 667 | |
| 668 | /* The PKO_MACX_CFG bits cannot be changed unless FIFO_MUM=0x1f (unused fifo) */ |
| 669 | saved_fifo_num = pko_mac_cfg.s.fifo_num; |
| 670 | pko_mac_cfg.s.fifo_num = 0x1f; |
| 671 | pko_mac_cfg.s.skid_max_cnt = tmp; |
| 672 | csr_wr_node(node, CVMX_PKO_MACX_CFG(mac_num), pko_mac_cfg.u64); |
| 673 | |
| 674 | pko_mac_cfg.u64 = csr_rd_node(node, CVMX_PKO_MACX_CFG(mac_num)); |
| 675 | pko_mac_cfg.s.fifo_num = saved_fifo_num; |
| 676 | csr_wr_node(node, CVMX_PKO_MACX_CFG(mac_num), pko_mac_cfg.u64); |
| 677 | |
| 678 | if (debug) { |
| 679 | pko_mci0_max_cred.u64 = |
| 680 | csr_rd_node(node, CVMX_PKO_MCI0_MAX_CREDX(mac_num)); |
| 681 | pko_mci1_max_cred.u64 = |
| 682 | csr_rd_node(node, CVMX_PKO_MCI1_MAX_CREDX(mac_num)); |
| 683 | pko_mac_cfg.u64 = |
| 684 | csr_rd_node(node, CVMX_PKO_MACX_CFG(mac_num)); |
| 685 | debug("%s: mac %u PKO_MCI0_MAX_CREDX=%u PKO_MCI1_MAX_CREDX=%u PKO_MACX_CFG[SKID_MAX_CNT]=%u\n", |
| 686 | __func__, mac_num, |
| 687 | pko_mci0_max_cred.s.max_cred_lim, |
| 688 | pko_mci1_max_cred.s.max_cred_lim, |
| 689 | pko_mac_cfg.s.skid_max_cnt); |
| 690 | } |
| 691 | } /* for mac_num */ |
| 692 | |
| 693 | return 0; |
| 694 | } |
| 695 | |
| 696 | /** Set MAC options |
| 697 | * |
| 698 | * The options supported are the parameters below: |
| 699 | * |
| 700 | * @param xiface The physical interface number |
| 701 | * @param index The physical sub-interface port |
| 702 | * @param fcs_enable Enable FCS generation |
| 703 | * @param pad_enable Enable padding to minimum packet size |
| 704 | * @param fcs_sop_off Number of bytes at start of packet to exclude from FCS |
| 705 | * |
| 706 | * The typical use for `fcs_sop_off` is when the interface is configured |
| 707 | * to use a header such as HighGig to precede every Ethernet packet, |
| 708 | * such a header usually does not partake in the CRC32 computation stream, |
| 709 | * and its size must be set with this parameter. |
| 710 | * |
| 711 | * @return Returns 0 on success, -1 if interface/port is invalid. |
| 712 | */ |
| 713 | int cvmx_pko3_interface_options(int xiface, int index, bool fcs_enable, |
| 714 | bool pad_enable, unsigned int fcs_sop_off) |
| 715 | { |
| 716 | int mac_num; |
| 717 | cvmx_pko_macx_cfg_t pko_mac_cfg; |
| 718 | unsigned int fifo_num; |
| 719 | struct cvmx_xiface xi = cvmx_helper_xiface_to_node_interface(xiface); |
| 720 | |
| 721 | if (debug) |
| 722 | debug("%s: intf %u:%u/%u fcs=%d pad=%d\n", __func__, xi.node, |
| 723 | xi.interface, index, fcs_enable, pad_enable); |
| 724 | |
| 725 | mac_num = __cvmx_pko3_get_mac_num(xiface, index); |
| 726 | if (mac_num < 0) { |
| 727 | cvmx_printf("ERROR: %s: invalid interface %u:%u/%u\n", __func__, |
| 728 | xi.node, xi.interface, index); |
| 729 | return -1; |
| 730 | } |
| 731 | |
| 732 | pko_mac_cfg.u64 = csr_rd_node(xi.node, CVMX_PKO_MACX_CFG(mac_num)); |
| 733 | |
| 734 | /* If MAC is not assigned, return an error */ |
| 735 | if (pko_mac_cfg.s.fifo_num == 0x1f) { |
| 736 | cvmx_printf("ERROR: %s: unused interface %u:%u/%u\n", __func__, |
| 737 | xi.node, xi.interface, index); |
| 738 | return -1; |
| 739 | } |
| 740 | |
| 741 | if (pko_mac_cfg.s.min_pad_ena == pad_enable && |
| 742 | pko_mac_cfg.s.fcs_ena == fcs_enable) { |
| 743 | if (debug) |
| 744 | debug("%s: mac %#x unchanged\n", __func__, mac_num); |
| 745 | return 0; |
| 746 | } |
| 747 | |
| 748 | /* WORKAROUND: Pass1 won't allow change any bits unless FIFO_NUM=0x1f */ |
| 749 | fifo_num = pko_mac_cfg.s.fifo_num; |
| 750 | pko_mac_cfg.s.fifo_num = 0x1f; |
| 751 | |
| 752 | pko_mac_cfg.s.min_pad_ena = pad_enable; |
| 753 | pko_mac_cfg.s.fcs_ena = fcs_enable; |
| 754 | pko_mac_cfg.s.fcs_sop_off = fcs_sop_off; |
| 755 | |
| 756 | csr_wr_node(xi.node, CVMX_PKO_MACX_CFG(mac_num), pko_mac_cfg.u64); |
| 757 | |
| 758 | pko_mac_cfg.s.fifo_num = fifo_num; |
| 759 | csr_wr_node(xi.node, CVMX_PKO_MACX_CFG(mac_num), pko_mac_cfg.u64); |
| 760 | |
| 761 | if (debug) |
| 762 | debug("%s: PKO_MAC[%u]CFG=%#llx\n", __func__, mac_num, |
| 763 | (unsigned long long)csr_rd_node(xi.node, CVMX_PKO_MACX_CFG(mac_num))); |
| 764 | |
| 765 | return 0; |
| 766 | } |
| 767 | |
| 768 | /** Set Descriptor Queue options |
| 769 | * |
| 770 | * The `min_pad` parameter must be in agreement with the interface-level |
| 771 | * padding option for all descriptor queues assigned to that particular |
| 772 | * interface/port. |
| 773 | * |
| 774 | * @param node on which to operate |
| 775 | * @param dq descriptor queue to set |
| 776 | * @param min_pad minimum padding to set for dq |
| 777 | */ |
| 778 | void cvmx_pko3_dq_options(unsigned int node, unsigned int dq, bool min_pad) |
| 779 | { |
| 780 | cvmx_pko_pdm_dqx_minpad_t reg; |
| 781 | |
| 782 | dq &= (1 << 10) - 1; |
| 783 | reg.u64 = csr_rd_node(node, CVMX_PKO_PDM_DQX_MINPAD(dq)); |
| 784 | reg.s.minpad = min_pad; |
| 785 | csr_wr_node(node, CVMX_PKO_PDM_DQX_MINPAD(dq), reg.u64); |
| 786 | } |