1 /* 2 * Copyright 2014 IBM Corp. 3 * 4 * This program is free software; you can redistribute it and/or 5 * modify it under the terms of the GNU General Public License 6 * as published by the Free Software Foundation; either version 7 * 2 of the License, or (at your option) any later version. 8 */ 9 10 #include <linux/pci.h> 11 #include <misc/cxl.h> 12 #include "cxl.h" 13 14 static int cxl_dma_set_mask(struct pci_dev *pdev, u64 dma_mask) 15 { 16 if (dma_mask < DMA_BIT_MASK(64)) { 17 pr_info("%s only 64bit DMA supported on CXL", __func__); 18 return -EIO; 19 } 20 21 *(pdev->dev.dma_mask) = dma_mask; 22 return 0; 23 } 24 25 static int cxl_pci_probe_mode(struct pci_bus *bus) 26 { 27 return PCI_PROBE_NORMAL; 28 } 29 30 static int cxl_setup_msi_irqs(struct pci_dev *pdev, int nvec, int type) 31 { 32 return -ENODEV; 33 } 34 35 static void cxl_teardown_msi_irqs(struct pci_dev *pdev) 36 { 37 /* 38 * MSI should never be set but need still need to provide this call 39 * back. 40 */ 41 } 42 43 static bool cxl_pci_enable_device_hook(struct pci_dev *dev) 44 { 45 struct pci_controller *phb; 46 struct cxl_afu *afu; 47 struct cxl_context *ctx; 48 49 phb = pci_bus_to_host(dev->bus); 50 afu = (struct cxl_afu *)phb->private_data; 51 52 if (!cxl_adapter_link_ok(afu->adapter)) { 53 dev_warn(&dev->dev, "%s: Device link is down, refusing to enable AFU\n", __func__); 54 return false; 55 } 56 57 set_dma_ops(&dev->dev, &dma_direct_ops); 58 set_dma_offset(&dev->dev, PAGE_OFFSET); 59 60 /* 61 * Allocate a context to do cxl things too. If we eventually do real 62 * DMA ops, we'll need a default context to attach them to 63 */ 64 ctx = cxl_dev_context_init(dev); 65 if (!ctx) 66 return false; 67 dev->dev.archdata.cxl_ctx = ctx; 68 69 return (cxl_afu_check_and_enable(afu) == 0); 70 } 71 72 static void cxl_pci_disable_device(struct pci_dev *dev) 73 { 74 struct cxl_context *ctx = cxl_get_context(dev); 75 76 if (ctx) { 77 if (ctx->status == STARTED) { 78 dev_err(&dev->dev, "Default context started\n"); 79 return; 80 } 81 dev->dev.archdata.cxl_ctx = NULL; 82 cxl_release_context(ctx); 83 } 84 } 85 86 static resource_size_t cxl_pci_window_alignment(struct pci_bus *bus, 87 unsigned long type) 88 { 89 return 1; 90 } 91 92 static void cxl_pci_reset_secondary_bus(struct pci_dev *dev) 93 { 94 /* Should we do an AFU reset here ? */ 95 } 96 97 static int cxl_pcie_cfg_record(u8 bus, u8 devfn) 98 { 99 return (bus << 8) + devfn; 100 } 101 102 static unsigned long cxl_pcie_cfg_addr(struct pci_controller* phb, 103 u8 bus, u8 devfn, int offset) 104 { 105 int record = cxl_pcie_cfg_record(bus, devfn); 106 107 return (unsigned long)phb->cfg_addr + ((unsigned long)phb->cfg_data * record) + offset; 108 } 109 110 111 static int cxl_pcie_config_info(struct pci_bus *bus, unsigned int devfn, 112 int offset, int len, 113 volatile void __iomem **ioaddr, 114 u32 *mask, int *shift) 115 { 116 struct pci_controller *phb; 117 struct cxl_afu *afu; 118 unsigned long addr; 119 120 phb = pci_bus_to_host(bus); 121 if (phb == NULL) 122 return PCIBIOS_DEVICE_NOT_FOUND; 123 afu = (struct cxl_afu *)phb->private_data; 124 125 if (cxl_pcie_cfg_record(bus->number, devfn) > afu->crs_num) 126 return PCIBIOS_DEVICE_NOT_FOUND; 127 if (offset >= (unsigned long)phb->cfg_data) 128 return PCIBIOS_BAD_REGISTER_NUMBER; 129 addr = cxl_pcie_cfg_addr(phb, bus->number, devfn, offset); 130 131 *ioaddr = (void *)(addr & ~0x3ULL); 132 *shift = ((addr & 0x3) * 8); 133 switch (len) { 134 case 1: 135 *mask = 0xff; 136 break; 137 case 2: 138 *mask = 0xffff; 139 break; 140 default: 141 *mask = 0xffffffff; 142 break; 143 } 144 return 0; 145 } 146 147 148 static inline bool cxl_config_link_ok(struct pci_bus *bus) 149 { 150 struct pci_controller *phb; 151 struct cxl_afu *afu; 152 153 /* Config space IO is based on phb->cfg_addr, which is based on 154 * afu_desc_mmio. This isn't safe to read/write when the link 155 * goes down, as EEH tears down MMIO space. 156 * 157 * Check if the link is OK before proceeding. 158 */ 159 160 phb = pci_bus_to_host(bus); 161 if (phb == NULL) 162 return false; 163 afu = (struct cxl_afu *)phb->private_data; 164 return cxl_adapter_link_ok(afu->adapter); 165 } 166 167 static int cxl_pcie_read_config(struct pci_bus *bus, unsigned int devfn, 168 int offset, int len, u32 *val) 169 { 170 volatile void __iomem *ioaddr; 171 int shift, rc; 172 u32 mask; 173 174 rc = cxl_pcie_config_info(bus, devfn, offset, len, &ioaddr, 175 &mask, &shift); 176 if (rc) 177 return rc; 178 179 if (!cxl_config_link_ok(bus)) 180 return PCIBIOS_DEVICE_NOT_FOUND; 181 182 /* Can only read 32 bits */ 183 *val = (in_le32(ioaddr) >> shift) & mask; 184 return PCIBIOS_SUCCESSFUL; 185 } 186 187 static int cxl_pcie_write_config(struct pci_bus *bus, unsigned int devfn, 188 int offset, int len, u32 val) 189 { 190 volatile void __iomem *ioaddr; 191 u32 v, mask; 192 int shift, rc; 193 194 rc = cxl_pcie_config_info(bus, devfn, offset, len, &ioaddr, 195 &mask, &shift); 196 if (rc) 197 return rc; 198 199 if (!cxl_config_link_ok(bus)) 200 return PCIBIOS_DEVICE_NOT_FOUND; 201 202 /* Can only write 32 bits so do read-modify-write */ 203 mask <<= shift; 204 val <<= shift; 205 206 v = (in_le32(ioaddr) & ~mask) || (val & mask); 207 208 out_le32(ioaddr, v); 209 return PCIBIOS_SUCCESSFUL; 210 } 211 212 static struct pci_ops cxl_pcie_pci_ops = 213 { 214 .read = cxl_pcie_read_config, 215 .write = cxl_pcie_write_config, 216 }; 217 218 219 static struct pci_controller_ops cxl_pci_controller_ops = 220 { 221 .probe_mode = cxl_pci_probe_mode, 222 .enable_device_hook = cxl_pci_enable_device_hook, 223 .disable_device = cxl_pci_disable_device, 224 .release_device = cxl_pci_disable_device, 225 .window_alignment = cxl_pci_window_alignment, 226 .reset_secondary_bus = cxl_pci_reset_secondary_bus, 227 .setup_msi_irqs = cxl_setup_msi_irqs, 228 .teardown_msi_irqs = cxl_teardown_msi_irqs, 229 .dma_set_mask = cxl_dma_set_mask, 230 }; 231 232 int cxl_pci_vphb_add(struct cxl_afu *afu) 233 { 234 struct pci_dev *phys_dev; 235 struct pci_controller *phb, *phys_phb; 236 237 phys_dev = to_pci_dev(afu->adapter->dev.parent); 238 phys_phb = pci_bus_to_host(phys_dev->bus); 239 240 /* Alloc and setup PHB data structure */ 241 phb = pcibios_alloc_controller(phys_phb->dn); 242 243 if (!phb) 244 return -ENODEV; 245 246 /* Setup parent in sysfs */ 247 phb->parent = &phys_dev->dev; 248 249 /* Setup the PHB using arch provided callback */ 250 phb->ops = &cxl_pcie_pci_ops; 251 phb->cfg_addr = afu->afu_desc_mmio + afu->crs_offset; 252 phb->cfg_data = (void *)(u64)afu->crs_len; 253 phb->private_data = afu; 254 phb->controller_ops = cxl_pci_controller_ops; 255 256 /* Scan the bus */ 257 pcibios_scan_phb(phb); 258 if (phb->bus == NULL) 259 return -ENXIO; 260 261 /* Claim resources. This might need some rework as well depending 262 * whether we are doing probe-only or not, like assigning unassigned 263 * resources etc... 264 */ 265 pcibios_claim_one_bus(phb->bus); 266 267 /* Add probed PCI devices to the device model */ 268 pci_bus_add_devices(phb->bus); 269 270 afu->phb = phb; 271 272 return 0; 273 } 274 275 void cxl_pci_vphb_reconfigure(struct cxl_afu *afu) 276 { 277 /* When we are reconfigured, the AFU's MMIO space is unmapped 278 * and remapped. We need to reflect this in the PHB's view of 279 * the world. 280 */ 281 afu->phb->cfg_addr = afu->afu_desc_mmio + afu->crs_offset; 282 } 283 284 void cxl_pci_vphb_remove(struct cxl_afu *afu) 285 { 286 struct pci_controller *phb; 287 288 /* If there is no configuration record we won't have one of these */ 289 if (!afu || !afu->phb) 290 return; 291 292 phb = afu->phb; 293 294 pci_remove_root_bus(phb->bus); 295 } 296 297 struct cxl_afu *cxl_pci_to_afu(struct pci_dev *dev) 298 { 299 struct pci_controller *phb; 300 301 phb = pci_bus_to_host(dev->bus); 302 303 return (struct cxl_afu *)phb->private_data; 304 } 305 EXPORT_SYMBOL_GPL(cxl_pci_to_afu); 306 307 unsigned int cxl_pci_to_cfg_record(struct pci_dev *dev) 308 { 309 return cxl_pcie_cfg_record(dev->bus->number, dev->devfn); 310 } 311 EXPORT_SYMBOL_GPL(cxl_pci_to_cfg_record); 312