blob: 3520488b345356a13b058c268750ebae6d35142a [file] [log] [blame]
Xiaowei Bao13b277f2020-07-09 23:31:33 +08001// SPDX-License-Identifier: GPL-2.0+
2/*
3 * Copyright 2020 NXP
4 * Layerscape PCIe EP driver
5 */
6
Tom Rinidec7ea02024-05-20 13:35:03 -06007#include <config.h>
Wasim Khan8b3d2d22020-09-28 16:26:04 +05308#include <asm/arch/fsl_serdes.h>
Xiaowei Bao13b277f2020-07-09 23:31:33 +08009#include <dm.h>
Simon Glass3ba929a2020-10-30 21:38:53 -060010#include <asm/global_data.h>
Xiaowei Bao13b277f2020-07-09 23:31:33 +080011#include <dm/devres.h>
12#include <errno.h>
13#include <pci_ep.h>
14#include <asm/io.h>
15#include <linux/sizes.h>
16#include <linux/log2.h>
17#include "pcie_layerscape.h"
18
19DECLARE_GLOBAL_DATA_PTR;
20
21static void ls_pcie_ep_enable_cfg(struct ls_pcie_ep *pcie_ep)
22{
23 struct ls_pcie *pcie = pcie_ep->pcie;
24 u32 config;
25
26 config = ctrl_readl(pcie, PCIE_PF_CONFIG);
27 config |= PCIE_CONFIG_READY;
28 ctrl_writel(pcie, config, PCIE_PF_CONFIG);
29}
30
31static int ls_ep_set_bar(struct udevice *dev, uint fn, struct pci_bar *ep_bar)
32{
33 struct ls_pcie_ep *pcie_ep = dev_get_priv(dev);
34 struct ls_pcie *pcie = pcie_ep->pcie;
35 dma_addr_t bar_phys = ep_bar->phys_addr;
36 enum pci_barno bar = ep_bar->barno;
37 u32 reg = PCI_BASE_ADDRESS_0 + (4 * bar);
38 int flags = ep_bar->flags;
39 int type, idx;
40 u64 size;
41
42 idx = bar;
43 /* BAR size is 2^(aperture + 11) */
44 size = max_t(size_t, ep_bar->size, FSL_PCIE_EP_MIN_APERTURE);
45
46 if (!(flags & PCI_BASE_ADDRESS_SPACE))
47 type = PCIE_ATU_TYPE_MEM;
48 else
49 type = PCIE_ATU_TYPE_IO;
50
Xiaowei Baoecb85db2020-07-09 23:31:39 +080051 ls_pcie_atu_inbound_set(pcie, fn, 0, type, idx, bar, bar_phys);
Xiaowei Bao13b277f2020-07-09 23:31:33 +080052
53 dbi_writel(pcie, lower_32_bits(size - 1), reg + PCIE_NO_SRIOV_BAR_BASE);
54 dbi_writel(pcie, flags, reg);
55
56 if (flags & PCI_BASE_ADDRESS_MEM_TYPE_64) {
57 dbi_writel(pcie, upper_32_bits(size - 1),
58 reg + 4 + PCIE_NO_SRIOV_BAR_BASE);
59 dbi_writel(pcie, 0, reg + 4);
60 }
61
62 return 0;
63}
64
65static struct pci_ep_ops ls_pcie_ep_ops = {
66 .set_bar = ls_ep_set_bar,
67};
68
Xiaowei Baob4d63b02020-07-09 23:31:36 +080069static void ls_pcie_ep_setup_atu(struct ls_pcie_ep *pcie_ep, u32 pf)
Xiaowei Bao13b277f2020-07-09 23:31:33 +080070{
71 struct ls_pcie *pcie = pcie_ep->pcie;
Xiaowei Baoecb85db2020-07-09 23:31:39 +080072 u32 vf_flag = 0;
Xiaowei Baob4d63b02020-07-09 23:31:36 +080073 u64 phys = 0;
Xiaowei Bao13b277f2020-07-09 23:31:33 +080074
Tom Rini56af6592022-11-16 13:10:33 -050075 phys = CFG_SYS_PCI_EP_MEMORY_BASE + pf * SZ_64M;
Xiaowei Baob4d63b02020-07-09 23:31:36 +080076
77 phys = ALIGN(phys, PCIE_BAR0_SIZE);
Xiaowei Bao13b277f2020-07-09 23:31:33 +080078 /* ATU 0 : INBOUND : map BAR0 */
Xiaowei Baoecb85db2020-07-09 23:31:39 +080079 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
Xiaowei Baob4d63b02020-07-09 23:31:36 +080080 0 + pf * BAR_NUM, 0, phys);
Xiaowei Bao13b277f2020-07-09 23:31:33 +080081 /* ATU 1 : INBOUND : map BAR1 */
Xiaowei Baob4d63b02020-07-09 23:31:36 +080082 phys = ALIGN(phys + PCIE_BAR0_SIZE, PCIE_BAR1_SIZE);
Xiaowei Baoecb85db2020-07-09 23:31:39 +080083 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
Xiaowei Baob4d63b02020-07-09 23:31:36 +080084 1 + pf * BAR_NUM, 1, phys);
Xiaowei Bao13b277f2020-07-09 23:31:33 +080085 /* ATU 2 : INBOUND : map BAR2 */
Xiaowei Baob4d63b02020-07-09 23:31:36 +080086 phys = ALIGN(phys + PCIE_BAR1_SIZE, PCIE_BAR2_SIZE);
Xiaowei Baoecb85db2020-07-09 23:31:39 +080087 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
Xiaowei Baob4d63b02020-07-09 23:31:36 +080088 2 + pf * BAR_NUM, 2, phys);
89 /* ATU 3 : INBOUND : map BAR2 */
90 phys = ALIGN(phys + PCIE_BAR2_SIZE, PCIE_BAR4_SIZE);
Xiaowei Baoecb85db2020-07-09 23:31:39 +080091 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
Xiaowei Baob4d63b02020-07-09 23:31:36 +080092 3 + pf * BAR_NUM, 4, phys);
Xiaowei Bao13b277f2020-07-09 23:31:33 +080093
Xiaowei Baoecb85db2020-07-09 23:31:39 +080094 if (pcie_ep->sriov_flag) {
95 vf_flag = 1;
96 /* ATU 4 : INBOUND : map BAR0 */
97 phys = ALIGN(phys + PCIE_BAR4_SIZE, PCIE_BAR0_SIZE);
98 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
99 4 + pf * BAR_NUM, 0, phys);
100 /* ATU 5 : INBOUND : map BAR1 */
101 phys = ALIGN(phys + PCIE_BAR0_SIZE * PCIE_VF_NUM,
102 PCIE_BAR1_SIZE);
103 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
104 5 + pf * BAR_NUM, 1, phys);
105 /* ATU 6 : INBOUND : map BAR2 */
106 phys = ALIGN(phys + PCIE_BAR1_SIZE * PCIE_VF_NUM,
107 PCIE_BAR2_SIZE);
108 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
109 6 + pf * BAR_NUM, 2, phys);
110 /* ATU 7 : INBOUND : map BAR4 */
111 phys = ALIGN(phys + PCIE_BAR2_SIZE * PCIE_VF_NUM,
112 PCIE_BAR4_SIZE);
113 ls_pcie_atu_inbound_set(pcie, pf, vf_flag, PCIE_ATU_TYPE_MEM,
114 7 + pf * BAR_NUM, 4, phys);
115 }
116
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800117 /* ATU: OUTBOUND : map MEM */
118 ls_pcie_atu_outbound_set(pcie, pf, PCIE_ATU_TYPE_MEM,
119 (u64)pcie_ep->addr_res.start +
Tom Rini56af6592022-11-16 13:10:33 -0500120 pf * CFG_SYS_PCI_MEMORY_SIZE,
121 0, CFG_SYS_PCI_MEMORY_SIZE);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800122}
123
124/* BAR0 and BAR1 are 32bit BAR2 and BAR4 are 64bit */
125static void ls_pcie_ep_setup_bar(void *bar_base, int bar, u32 size)
126{
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800127 u32 mask;
128
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800129 /* The least inbound window is 4KiB */
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800130 if (size < SZ_4K)
131 mask = 0;
132 else
133 mask = size - 1;
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800134
135 switch (bar) {
136 case 0:
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800137 writel(mask, bar_base + PCI_BASE_ADDRESS_0);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800138 break;
139 case 1:
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800140 writel(mask, bar_base + PCI_BASE_ADDRESS_1);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800141 break;
142 case 2:
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800143 writel(mask, bar_base + PCI_BASE_ADDRESS_2);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800144 writel(0, bar_base + PCI_BASE_ADDRESS_3);
145 break;
146 case 4:
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800147 writel(mask, bar_base + PCI_BASE_ADDRESS_4);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800148 writel(0, bar_base + PCI_BASE_ADDRESS_5);
149 break;
150 default:
151 break;
152 }
153}
154
155static void ls_pcie_ep_setup_bars(void *bar_base)
156{
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800157 /* BAR0 - 32bit - MEM */
158 ls_pcie_ep_setup_bar(bar_base, 0, PCIE_BAR0_SIZE);
159 /* BAR1 - 32bit - MEM*/
160 ls_pcie_ep_setup_bar(bar_base, 1, PCIE_BAR1_SIZE);
161 /* BAR2 - 64bit - MEM */
162 ls_pcie_ep_setup_bar(bar_base, 2, PCIE_BAR2_SIZE);
163 /* BAR4 - 64bit - MEM */
164 ls_pcie_ep_setup_bar(bar_base, 4, PCIE_BAR4_SIZE);
165}
166
167static void ls_pcie_ep_setup_vf_bars(void *bar_base)
168{
169 /* VF BAR0 MASK register at offset 0x19c*/
170 bar_base += PCIE_SRIOV_VFBAR0 - PCI_BASE_ADDRESS_0;
171
172 /* VF-BAR0 - 32bit - MEM */
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800173 ls_pcie_ep_setup_bar(bar_base, 0, PCIE_BAR0_SIZE);
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800174 /* VF-BAR1 - 32bit - MEM*/
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800175 ls_pcie_ep_setup_bar(bar_base, 1, PCIE_BAR1_SIZE);
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800176 /* VF-BAR2 - 64bit - MEM */
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800177 ls_pcie_ep_setup_bar(bar_base, 2, PCIE_BAR2_SIZE);
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800178 /* VF-BAR4 - 64bit - MEM */
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800179 ls_pcie_ep_setup_bar(bar_base, 4, PCIE_BAR4_SIZE);
180}
181
182static void ls_pcie_setup_ep(struct ls_pcie_ep *pcie_ep)
183{
184 u32 sriov;
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800185 u32 pf, vf;
186 void *bar_base = NULL;
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800187 struct ls_pcie *pcie = pcie_ep->pcie;
188
189 sriov = readl(pcie->dbi + PCIE_SRIOV);
190 if (PCI_EXT_CAP_ID(sriov) == PCI_EXT_CAP_ID_SRIOV) {
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800191 pcie_ep->sriov_flag = 1;
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800192 for (pf = 0; pf < PCIE_PF_NUM; pf++) {
Xiaowei Bao06a329a2020-07-09 23:31:37 +0800193 /*
194 * The VF_BARn_REG register's Prefetchable and Type bit
195 * fields are overwritten by a write to VF's BAR Mask
196 * register. Before writing to the VF_BARn_MASK_REG
197 * register, write 0b to the PCIE_MISC_CONTROL_1_OFF
198 * register.
199 */
200 writel(0, pcie->dbi + PCIE_MISC_CONTROL_1_OFF);
201
Xiaowei Baocdecb972020-07-09 23:31:38 +0800202 bar_base = pcie->dbi +
Xiaowei Bao4a602b52020-07-09 23:31:41 +0800203 PCIE_MASK_OFFSET(pcie_ep->cfg2_flag, pf,
204 pcie_ep->pf1_offset);
Xiaowei Baocdecb972020-07-09 23:31:38 +0800205
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800206 if (pcie_ep->cfg2_flag) {
Xiaowei Baocdecb972020-07-09 23:31:38 +0800207 ctrl_writel(pcie,
208 PCIE_LCTRL0_VAL(pf, 0),
209 PCIE_PF_VF_CTRL);
210 ls_pcie_ep_setup_bars(bar_base);
211
212 for (vf = 1; vf <= PCIE_VF_NUM; vf++) {
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800213 ctrl_writel(pcie,
214 PCIE_LCTRL0_VAL(pf, vf),
215 PCIE_PF_VF_CTRL);
Xiaowei Baocdecb972020-07-09 23:31:38 +0800216 ls_pcie_ep_setup_vf_bars(bar_base);
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800217 }
Xiaowei Baocdecb972020-07-09 23:31:38 +0800218 } else {
219 ls_pcie_ep_setup_bars(bar_base);
220 ls_pcie_ep_setup_vf_bars(bar_base);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800221 }
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800222
223 ls_pcie_ep_setup_atu(pcie_ep, pf);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800224 }
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800225
226 if (pcie_ep->cfg2_flag) /* Disable CFG2 */
227 ctrl_writel(pcie, 0, PCIE_PF_VF_CTRL);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800228 } else {
229 ls_pcie_ep_setup_bars(pcie->dbi + PCIE_NO_SRIOV_BAR_BASE);
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800230 ls_pcie_ep_setup_atu(pcie_ep, 0);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800231 }
232
Xiaowei Baobb3f7132020-07-09 23:31:40 +0800233 ls_pcie_dump_atu(pcie, PCIE_ATU_REGION_NUM_SRIOV,
234 PCIE_ATU_REGION_INBOUND);
235
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800236 ls_pcie_ep_enable_cfg(pcie_ep);
237}
238
239static int ls_pcie_ep_probe(struct udevice *dev)
240{
241 struct ls_pcie_ep *pcie_ep = dev_get_priv(dev);
242 struct ls_pcie *pcie;
243 u16 link_sta;
244 int ret;
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800245 u32 svr;
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800246
Hou Zhiqiangbae8c332021-03-11 15:30:51 +0800247 pcie = devm_kzalloc(dev, sizeof(*pcie), GFP_KERNEL);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800248 if (!pcie)
249 return -ENOMEM;
250
251 pcie_ep->pcie = pcie;
252
Johan Jonkerb52189e2023-03-13 01:32:31 +0100253 pcie->dbi = devfdt_get_addr_index_ptr(dev, 0);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800254 if (!pcie->dbi)
Johan Jonkerb52189e2023-03-13 01:32:31 +0100255 return -EINVAL;
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800256
Johan Jonkerb52189e2023-03-13 01:32:31 +0100257 pcie->ctrl = devfdt_get_addr_index_ptr(dev, 1);
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800258 if (!pcie->ctrl)
Johan Jonkerb52189e2023-03-13 01:32:31 +0100259 return -EINVAL;
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800260
261 ret = fdt_get_named_resource(gd->fdt_blob, dev_of_offset(dev),
262 "reg", "reg-names",
263 "addr_space", &pcie_ep->addr_res);
264 if (ret) {
265 printf("%s: resource \"addr_space\" not found\n", dev->name);
266 return ret;
267 }
268
269 pcie->idx = ((unsigned long)pcie->dbi - PCIE_SYS_BASE_ADDR) /
270 PCIE_CCSR_SIZE;
271
Hou Zhiqiang85ffe6c2021-05-13 14:54:32 +0800272 /* This controller is disabled by RCW */
273 if (!is_serdes_configured(PCIE_SRDS_PRTCL(pcie->idx)))
274 return 0;
275
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800276 pcie->big_endian = fdtdec_get_bool(gd->fdt_blob, dev_of_offset(dev),
277 "big-endian");
278
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800279 svr = SVR_SOC_VER(get_svr());
280
Meenakshi Aggarwalccb5d5d2020-10-29 19:16:16 +0530281 if (svr == SVR_LX2160A || svr == SVR_LX2162A ||
282 svr == SVR_LX2120A || svr == SVR_LX2080A ||
283 svr == SVR_LX2122A || svr == SVR_LX2082A)
Xiaowei Bao4a602b52020-07-09 23:31:41 +0800284 pcie_ep->pf1_offset = LX2160_PCIE_PF1_OFFSET;
285 else
286 pcie_ep->pf1_offset = LS_PCIE_PF1_OFFSET;
287
Xiaowei Baob4d63b02020-07-09 23:31:36 +0800288 if (svr == SVR_LS2080A || svr == SVR_LS2085A)
289 pcie_ep->cfg2_flag = 1;
290 else
291 pcie_ep->cfg2_flag = 0;
292
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800293 pcie->mode = readb(pcie->dbi + PCI_HEADER_TYPE) & 0x7f;
294 if (pcie->mode != PCI_HEADER_TYPE_NORMAL)
295 return 0;
296
297 pcie_ep->max_functions = fdtdec_get_int(gd->fdt_blob,
298 dev_of_offset(dev),
299 "max-functions", 1);
300 pcie_ep->num_ib_wins = fdtdec_get_int(gd->fdt_blob, dev_of_offset(dev),
301 "num-ib-windows", 8);
302 pcie_ep->num_ob_wins = fdtdec_get_int(gd->fdt_blob, dev_of_offset(dev),
303 "num-ob-windows", 8);
304
Wasim Khan8b3d2d22020-09-28 16:26:04 +0530305 printf("PCIe%u: %s %s", PCIE_SRDS_PRTCL(pcie->idx), dev->name,
306 "Endpoint");
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800307 ls_pcie_setup_ep(pcie_ep);
308
309 if (!ls_pcie_link_up(pcie)) {
310 /* Let the user know there's no PCIe link */
311 printf(": no link\n");
312 return 0;
313 }
314
315 /* Print the negotiated PCIe link width */
316 link_sta = readw(pcie->dbi + PCIE_LINK_STA);
317 printf(": x%d gen%d\n", (link_sta & PCIE_LINK_WIDTH_MASK) >> 4,
318 link_sta & PCIE_LINK_SPEED_MASK);
319
320 return 0;
321}
322
323static int ls_pcie_ep_remove(struct udevice *dev)
324{
325 return 0;
326}
327
328const struct udevice_id ls_pcie_ep_ids[] = {
329 { .compatible = "fsl,ls-pcie-ep" },
330 { }
331};
332
333U_BOOT_DRIVER(pci_layerscape_ep) = {
334 .name = "pci_layerscape_ep",
335 .id = UCLASS_PCI_EP,
336 .of_match = ls_pcie_ep_ids,
337 .ops = &ls_pcie_ep_ops,
338 .probe = ls_pcie_ep_probe,
339 .remove = ls_pcie_ep_remove,
Simon Glass8a2b47f2020-12-03 16:55:17 -0700340 .priv_auto = sizeof(struct ls_pcie_ep),
Xiaowei Bao13b277f2020-07-09 23:31:33 +0800341};