xref: /linux/drivers/accel/habanalabs/common/pci/pci.c (revision 306ec721d043bbe5e818d59fbb37c28d999b5d8b)
1 // SPDX-License-Identifier: GPL-2.0
2 
3 /*
4  * Copyright 2016-2019 HabanaLabs, Ltd.
5  * All Rights Reserved.
6  */
7 
8 #include "../habanalabs.h"
9 #include "../../include/hw_ip/pci/pci_general.h"
10 
11 #include <linux/pci.h>
12 
13 #include <trace/events/habanalabs.h>
14 
15 #define HL_PLDM_PCI_ELBI_TIMEOUT_MSEC	(HL_PCI_ELBI_TIMEOUT_MSEC * 100)
16 
17 #define IATU_REGION_CTRL_REGION_EN_MASK		BIT(31)
18 #define IATU_REGION_CTRL_MATCH_MODE_MASK	BIT(30)
19 #define IATU_REGION_CTRL_NUM_MATCH_EN_MASK	BIT(19)
20 #define IATU_REGION_CTRL_BAR_NUM_MASK		GENMASK(10, 8)
21 
22 /**
23  * hl_pci_bars_map() - Map PCI BARs.
24  * @hdev: Pointer to hl_device structure.
25  * @name: Array of BAR names.
26  * @is_wc: Array with flag per BAR whether a write-combined mapping is needed.
27  *
28  * Request PCI regions and map them to kernel virtual addresses.
29  *
30  * Return: 0 on success, non-zero for failure.
31  */
32 int hl_pci_bars_map(struct hl_device *hdev, const char * const name[3],
33 			bool is_wc[3])
34 {
35 	struct pci_dev *pdev = hdev->pdev;
36 	int rc, i, bar;
37 
38 	rc = pci_request_regions(pdev, HL_NAME);
39 	if (rc) {
40 		dev_err(hdev->dev, "Cannot obtain PCI resources\n");
41 		return rc;
42 	}
43 
44 	for (i = 0 ; i < 3 ; i++) {
45 		bar = i * 2; /* 64-bit BARs */
46 		hdev->pcie_bar[bar] = is_wc[i] ?
47 				pci_ioremap_wc_bar(pdev, bar) :
48 				pci_ioremap_bar(pdev, bar);
49 		if (!hdev->pcie_bar[bar]) {
50 			dev_err(hdev->dev, "pci_ioremap%s_bar failed for %s\n",
51 					is_wc[i] ? "_wc" : "", name[i]);
52 			rc = -ENODEV;
53 			goto err;
54 		}
55 	}
56 
57 	return 0;
58 
59 err:
60 	for (i = 2 ; i >= 0 ; i--) {
61 		bar = i * 2; /* 64-bit BARs */
62 		if (hdev->pcie_bar[bar])
63 			iounmap(hdev->pcie_bar[bar]);
64 	}
65 
66 	pci_release_regions(pdev);
67 
68 	return rc;
69 }
70 
71 /**
72  * hl_pci_bars_unmap() - Unmap PCI BARS.
73  * @hdev: Pointer to hl_device structure.
74  *
75  * Release all PCI BARs and unmap their virtual addresses.
76  */
77 static void hl_pci_bars_unmap(struct hl_device *hdev)
78 {
79 	struct pci_dev *pdev = hdev->pdev;
80 	int i, bar;
81 
82 	for (i = 2 ; i >= 0 ; i--) {
83 		bar = i * 2; /* 64-bit BARs */
84 		iounmap(hdev->pcie_bar[bar]);
85 	}
86 
87 	pci_release_regions(pdev);
88 }
89 
90 int hl_pci_elbi_read(struct hl_device *hdev, u64 addr, u32 *data)
91 {
92 	struct pci_dev *pdev = hdev->pdev;
93 	ktime_t timeout;
94 	u64 msec;
95 	u32 val;
96 
97 	if (hdev->pldm)
98 		msec = HL_PLDM_PCI_ELBI_TIMEOUT_MSEC;
99 	else
100 		msec = HL_PCI_ELBI_TIMEOUT_MSEC;
101 
102 	/* Clear previous status */
103 	pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, 0);
104 
105 	pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_ADDR, (u32) addr);
106 	pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_CTRL, 0);
107 
108 	timeout = ktime_add_ms(ktime_get(), msec);
109 	for (;;) {
110 		pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, &val);
111 		if (val & PCI_CONFIG_ELBI_STS_MASK)
112 			break;
113 		if (ktime_compare(ktime_get(), timeout) > 0) {
114 			pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS,
115 						&val);
116 			break;
117 		}
118 
119 		usleep_range(300, 500);
120 	}
121 
122 	if ((val & PCI_CONFIG_ELBI_STS_MASK) == PCI_CONFIG_ELBI_STS_DONE) {
123 		pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_DATA, data);
124 
125 		if (unlikely(trace_habanalabs_elbi_read_enabled()))
126 			trace_habanalabs_elbi_read(hdev->dev, (u32) addr, val);
127 
128 		return 0;
129 	}
130 
131 	if (val & PCI_CONFIG_ELBI_STS_ERR) {
132 		dev_err(hdev->dev, "Error reading from ELBI\n");
133 		return -EIO;
134 	}
135 
136 	if (!(val & PCI_CONFIG_ELBI_STS_MASK)) {
137 		dev_err(hdev->dev, "ELBI read didn't finish in time\n");
138 		return -EIO;
139 	}
140 
141 	dev_err(hdev->dev, "ELBI read has undefined bits in status\n");
142 	return -EIO;
143 }
144 
145 /**
146  * hl_pci_elbi_write() - Write through the ELBI interface.
147  * @hdev: Pointer to hl_device structure.
148  * @addr: Address to write to
149  * @data: Data to write
150  *
151  * Return: 0 on success, negative value for failure.
152  */
153 static int hl_pci_elbi_write(struct hl_device *hdev, u64 addr, u32 data)
154 {
155 	struct pci_dev *pdev = hdev->pdev;
156 	ktime_t timeout;
157 	u64 msec;
158 	u32 val;
159 
160 	if (hdev->pldm)
161 		msec = HL_PLDM_PCI_ELBI_TIMEOUT_MSEC;
162 	else
163 		msec = HL_PCI_ELBI_TIMEOUT_MSEC;
164 
165 	/* Clear previous status */
166 	pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, 0);
167 
168 	pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_ADDR, (u32) addr);
169 	pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_DATA, data);
170 	pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_CTRL,
171 				PCI_CONFIG_ELBI_CTRL_WRITE);
172 
173 	timeout = ktime_add_ms(ktime_get(), msec);
174 	for (;;) {
175 		pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, &val);
176 		if (val & PCI_CONFIG_ELBI_STS_MASK)
177 			break;
178 		if (ktime_compare(ktime_get(), timeout) > 0) {
179 			pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS,
180 						&val);
181 			break;
182 		}
183 
184 		usleep_range(300, 500);
185 	}
186 
187 	if ((val & PCI_CONFIG_ELBI_STS_MASK) == PCI_CONFIG_ELBI_STS_DONE) {
188 		if (unlikely(trace_habanalabs_elbi_write_enabled()))
189 			trace_habanalabs_elbi_write(hdev->dev, (u32) addr, val);
190 		return 0;
191 	}
192 
193 	if (val & PCI_CONFIG_ELBI_STS_ERR)
194 		return -EIO;
195 
196 	if (!(val & PCI_CONFIG_ELBI_STS_MASK)) {
197 		dev_err(hdev->dev, "ELBI write didn't finish in time\n");
198 		return -EIO;
199 	}
200 
201 	dev_err(hdev->dev, "ELBI write has undefined bits in status\n");
202 	return -EIO;
203 }
204 
205 /**
206  * hl_pci_iatu_write() - iatu write routine.
207  * @hdev: Pointer to hl_device structure.
208  * @addr: Address to write to
209  * @data: Data to write
210  *
211  * Return: 0 on success, negative value for failure.
212  */
213 int hl_pci_iatu_write(struct hl_device *hdev, u32 addr, u32 data)
214 {
215 	struct asic_fixed_properties *prop = &hdev->asic_prop;
216 	u32 dbi_offset;
217 	int rc;
218 
219 	dbi_offset = addr & 0xFFF;
220 
221 	/* Ignore result of writing to pcie_aux_dbi_reg_addr as it could fail
222 	 * in case the firmware security is enabled
223 	 */
224 	hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0x00300000);
225 
226 	rc = hl_pci_elbi_write(hdev, prop->pcie_dbi_base_address + dbi_offset,
227 				data);
228 
229 	if (rc)
230 		return -EIO;
231 
232 	return 0;
233 }
234 
235 /**
236  * hl_pci_set_inbound_region() - Configure inbound region
237  * @hdev: Pointer to hl_device structure.
238  * @region: Inbound region number.
239  * @pci_region: Inbound region parameters.
240  *
241  * Configure the iATU inbound region.
242  *
243  * Return: 0 on success, negative value for failure.
244  */
245 int hl_pci_set_inbound_region(struct hl_device *hdev, u8 region,
246 		struct hl_inbound_pci_region *pci_region)
247 {
248 	struct asic_fixed_properties *prop = &hdev->asic_prop;
249 	u64 bar_phys_base, region_base, region_end_address;
250 	u32 offset, ctrl_reg_val;
251 	int rc = 0;
252 
253 	/* region offset */
254 	offset = (0x200 * region) + 0x100;
255 
256 	if (pci_region->mode == PCI_ADDRESS_MATCH_MODE) {
257 		bar_phys_base = hdev->pcie_bar_phys[pci_region->bar];
258 		region_base = bar_phys_base + pci_region->offset_in_bar;
259 		region_end_address = region_base + pci_region->size - 1;
260 
261 		rc |= hl_pci_iatu_write(hdev, offset + 0x8,
262 				lower_32_bits(region_base));
263 		rc |= hl_pci_iatu_write(hdev, offset + 0xC,
264 				upper_32_bits(region_base));
265 		rc |= hl_pci_iatu_write(hdev, offset + 0x10,
266 				lower_32_bits(region_end_address));
267 	}
268 
269 	/* Point to the specified address */
270 	rc |= hl_pci_iatu_write(hdev, offset + 0x14, lower_32_bits(pci_region->addr));
271 	rc |= hl_pci_iatu_write(hdev, offset + 0x18, upper_32_bits(pci_region->addr));
272 
273 	/* Set bar type as memory */
274 	rc |= hl_pci_iatu_write(hdev, offset + 0x0, 0);
275 
276 	/* Enable + bar/address match + match enable + bar number */
277 	ctrl_reg_val = FIELD_PREP(IATU_REGION_CTRL_REGION_EN_MASK, 1);
278 	ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_MATCH_MODE_MASK, pci_region->mode);
279 	ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_NUM_MATCH_EN_MASK, 1);
280 
281 	if (pci_region->mode == PCI_BAR_MATCH_MODE)
282 		ctrl_reg_val |= FIELD_PREP(IATU_REGION_CTRL_BAR_NUM_MASK, pci_region->bar);
283 
284 	rc |= hl_pci_iatu_write(hdev, offset + 0x4, ctrl_reg_val);
285 
286 	/* Return the DBI window to the default location
287 	 * Ignore result of writing to pcie_aux_dbi_reg_addr as it could fail
288 	 * in case the firmware security is enabled
289 	 */
290 	hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0);
291 
292 	if (rc)
293 		dev_err(hdev->dev, "failed to map bar %u to 0x%08llx\n",
294 				pci_region->bar, pci_region->addr);
295 
296 	return rc;
297 }
298 
299 /**
300  * hl_pci_set_outbound_region() - Configure outbound region 0
301  * @hdev: Pointer to hl_device structure.
302  * @pci_region: Outbound region parameters.
303  *
304  * Configure the iATU outbound region 0.
305  *
306  * Return: 0 on success, negative value for failure.
307  */
308 int hl_pci_set_outbound_region(struct hl_device *hdev,
309 		struct hl_outbound_pci_region *pci_region)
310 {
311 	struct asic_fixed_properties *prop = &hdev->asic_prop;
312 	u64 outbound_region_end_address;
313 	int rc = 0;
314 
315 	/* Outbound Region 0 */
316 	outbound_region_end_address =
317 			pci_region->addr + pci_region->size - 1;
318 	rc |= hl_pci_iatu_write(hdev, 0x008,
319 				lower_32_bits(pci_region->addr));
320 	rc |= hl_pci_iatu_write(hdev, 0x00C,
321 				upper_32_bits(pci_region->addr));
322 	rc |= hl_pci_iatu_write(hdev, 0x010,
323 				lower_32_bits(outbound_region_end_address));
324 	rc |= hl_pci_iatu_write(hdev, 0x014, 0);
325 
326 	rc |= hl_pci_iatu_write(hdev, 0x018, 0);
327 
328 	rc |= hl_pci_iatu_write(hdev, 0x020,
329 				upper_32_bits(outbound_region_end_address));
330 	/* Increase region size */
331 	rc |= hl_pci_iatu_write(hdev, 0x000, 0x00002000);
332 	/* Enable */
333 	rc |= hl_pci_iatu_write(hdev, 0x004, 0x80000000);
334 
335 	/* Return the DBI window to the default location
336 	 * Ignore result of writing to pcie_aux_dbi_reg_addr as it could fail
337 	 * in case the firmware security is enabled
338 	 */
339 	hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0);
340 
341 	return rc;
342 }
343 
344 /**
345  * hl_get_pci_memory_region() - get PCI region for given address
346  * @hdev: Pointer to hl_device structure.
347  * @addr: device address
348  *
349  * @return region index on success, otherwise PCI_REGION_NUMBER (invalid
350  *         region index)
351  */
352 enum pci_region hl_get_pci_memory_region(struct hl_device *hdev, u64 addr)
353 {
354 	int i;
355 
356 	for  (i = 0 ; i < PCI_REGION_NUMBER ; i++) {
357 		struct pci_mem_region *region = &hdev->pci_mem_region[i];
358 
359 		if (!region->used)
360 			continue;
361 
362 		if ((addr >= region->region_base) &&
363 			(addr < region->region_base + region->region_size))
364 			return i;
365 	}
366 
367 	return PCI_REGION_NUMBER;
368 }
369 
370 /**
371  * hl_pci_init() - PCI initialization code.
372  * @hdev: Pointer to hl_device structure.
373  *
374  * Set DMA masks, initialize the PCI controller and map the PCI BARs.
375  *
376  * Return: 0 on success, non-zero for failure.
377  */
378 int hl_pci_init(struct hl_device *hdev)
379 {
380 	struct asic_fixed_properties *prop = &hdev->asic_prop;
381 	struct pci_dev *pdev = hdev->pdev;
382 	int rc;
383 
384 	rc = pci_enable_device_mem(pdev);
385 	if (rc) {
386 		dev_err(hdev->dev, "can't enable PCI device\n");
387 		return rc;
388 	}
389 
390 	pci_set_master(pdev);
391 
392 	rc = hdev->asic_funcs->pci_bars_map(hdev);
393 	if (rc) {
394 		dev_err(hdev->dev, "Failed to map PCI BAR addresses\n");
395 		goto disable_device;
396 	}
397 
398 	rc = hdev->asic_funcs->init_iatu(hdev);
399 	if (rc) {
400 		dev_err(hdev->dev, "PCI controller was not initialized successfully\n");
401 		goto unmap_pci_bars;
402 	}
403 
404 	/* Driver must sleep in order for FW to finish the iATU configuration */
405 	if (hdev->asic_prop.iatu_done_by_fw)
406 		usleep_range(2000, 3000);
407 
408 	rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(prop->dma_mask));
409 	if (rc) {
410 		dev_err(hdev->dev,
411 			"Failed to set dma mask to %d bits, error %d\n",
412 			prop->dma_mask, rc);
413 		goto unmap_pci_bars;
414 	}
415 
416 	dma_set_max_seg_size(&pdev->dev, U32_MAX);
417 
418 	return 0;
419 
420 unmap_pci_bars:
421 	hl_pci_bars_unmap(hdev);
422 disable_device:
423 	pci_disable_device(pdev);
424 
425 	return rc;
426 }
427 
428 /**
429  * hl_pci_fini() - PCI finalization code.
430  * @hdev: Pointer to hl_device structure
431  *
432  * Unmap PCI bars and disable PCI device.
433  */
434 void hl_pci_fini(struct hl_device *hdev)
435 {
436 	hl_pci_bars_unmap(hdev);
437 
438 	pci_disable_device(hdev->pdev);
439 }
440