1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * drivers/mfd/mfd-core.c 4 * 5 * core MFD support 6 * Copyright (c) 2006 Ian Molton 7 * Copyright (c) 2007,2008 Dmitry Baryshkov 8 */ 9 10 #include <linux/kernel.h> 11 #include <linux/platform_device.h> 12 #include <linux/acpi.h> 13 #include <linux/list.h> 14 #include <linux/property.h> 15 #include <linux/mfd/core.h> 16 #include <linux/pm_runtime.h> 17 #include <linux/slab.h> 18 #include <linux/module.h> 19 #include <linux/irqdomain.h> 20 #include <linux/of.h> 21 #include <linux/of_address.h> 22 #include <linux/regulator/consumer.h> 23 24 static LIST_HEAD(mfd_of_node_list); 25 26 struct mfd_of_node_entry { 27 struct list_head list; 28 struct device *dev; 29 struct device_node *np; 30 }; 31 32 static struct device_type mfd_dev_type = { 33 .name = "mfd_device", 34 }; 35 36 #if IS_ENABLED(CONFIG_ACPI) 37 struct match_ids_walk_data { 38 struct acpi_device_id *ids; 39 struct acpi_device *adev; 40 }; 41 42 static int match_device_ids(struct acpi_device *adev, void *data) 43 { 44 struct match_ids_walk_data *wd = data; 45 46 if (!acpi_match_device_ids(adev, wd->ids)) { 47 wd->adev = adev; 48 return 1; 49 } 50 51 return 0; 52 } 53 54 static void mfd_acpi_add_device(const struct mfd_cell *cell, 55 struct platform_device *pdev) 56 { 57 const struct mfd_cell_acpi_match *match = cell->acpi_match; 58 struct acpi_device *adev = NULL; 59 struct acpi_device *parent; 60 61 parent = ACPI_COMPANION(pdev->dev.parent); 62 if (!parent) 63 return; 64 65 /* 66 * MFD child device gets its ACPI handle either from the ACPI device 67 * directly under the parent that matches the either _HID or _CID, or 68 * _ADR or it will use the parent handle if is no ID is given. 69 * 70 * Note that use of _ADR is a grey area in the ACPI specification, 71 * though at least Intel Galileo Gen 2 is using it to distinguish 72 * the children devices. 73 */ 74 if (match) { 75 if (match->pnpid) { 76 struct acpi_device_id ids[2] = {}; 77 struct match_ids_walk_data wd = { 78 .adev = NULL, 79 .ids = ids, 80 }; 81 82 strscpy(ids[0].id, match->pnpid, sizeof(ids[0].id)); 83 acpi_dev_for_each_child(parent, match_device_ids, &wd); 84 adev = wd.adev; 85 } else { 86 adev = acpi_find_child_device(parent, match->adr, false); 87 } 88 } 89 90 ACPI_COMPANION_SET(&pdev->dev, adev ?: parent); 91 } 92 #else 93 static inline void mfd_acpi_add_device(const struct mfd_cell *cell, 94 struct platform_device *pdev) 95 { 96 } 97 #endif 98 99 static int mfd_match_of_node_to_dev(struct platform_device *pdev, 100 struct device_node *np, 101 const struct mfd_cell *cell) 102 { 103 #if IS_ENABLED(CONFIG_OF) 104 struct mfd_of_node_entry *of_entry; 105 u64 of_node_addr; 106 107 /* Skip if OF node has previously been allocated to a device */ 108 list_for_each_entry(of_entry, &mfd_of_node_list, list) 109 if (of_entry->np == np) 110 return -EAGAIN; 111 112 if (!cell->use_of_reg) 113 /* No of_reg defined - allocate first free compatible match */ 114 goto allocate_of_node; 115 116 /* We only care about each node's first defined address */ 117 if (of_property_read_reg(np, 0, &of_node_addr, NULL)) 118 /* OF node does not contatin a 'reg' property to match to */ 119 return -EAGAIN; 120 121 if (cell->of_reg != of_node_addr) 122 /* No match */ 123 return -EAGAIN; 124 125 allocate_of_node: 126 of_entry = kzalloc(sizeof(*of_entry), GFP_KERNEL); 127 if (!of_entry) 128 return -ENOMEM; 129 130 of_entry->dev = &pdev->dev; 131 of_entry->np = np; 132 list_add_tail(&of_entry->list, &mfd_of_node_list); 133 134 pdev->dev.of_node = np; 135 pdev->dev.fwnode = &np->fwnode; 136 #endif 137 return 0; 138 } 139 140 static int mfd_add_device(struct device *parent, int id, 141 const struct mfd_cell *cell, 142 struct resource *mem_base, 143 int irq_base, struct irq_domain *domain) 144 { 145 struct resource *res; 146 struct platform_device *pdev; 147 struct device_node *np = NULL; 148 struct mfd_of_node_entry *of_entry, *tmp; 149 int ret = -ENOMEM; 150 int platform_id; 151 int r; 152 153 if (id == PLATFORM_DEVID_AUTO) 154 platform_id = id; 155 else 156 platform_id = id + cell->id; 157 158 pdev = platform_device_alloc(cell->name, platform_id); 159 if (!pdev) 160 goto fail_alloc; 161 162 pdev->mfd_cell = kmemdup(cell, sizeof(*cell), GFP_KERNEL); 163 if (!pdev->mfd_cell) 164 goto fail_device; 165 166 res = kcalloc(cell->num_resources, sizeof(*res), GFP_KERNEL); 167 if (!res) 168 goto fail_device; 169 170 pdev->dev.parent = parent; 171 pdev->dev.type = &mfd_dev_type; 172 pdev->dev.dma_mask = parent->dma_mask; 173 pdev->dev.dma_parms = parent->dma_parms; 174 pdev->dev.coherent_dma_mask = parent->coherent_dma_mask; 175 176 ret = regulator_bulk_register_supply_alias( 177 &pdev->dev, cell->parent_supplies, 178 parent, cell->parent_supplies, 179 cell->num_parent_supplies); 180 if (ret < 0) 181 goto fail_res; 182 183 if (IS_ENABLED(CONFIG_OF) && parent->of_node && cell->of_compatible) { 184 for_each_child_of_node(parent->of_node, np) { 185 if (of_device_is_compatible(np, cell->of_compatible)) { 186 /* Ignore 'disabled' devices error free */ 187 if (!of_device_is_available(np)) { 188 of_node_put(np); 189 ret = 0; 190 goto fail_alias; 191 } 192 193 ret = mfd_match_of_node_to_dev(pdev, np, cell); 194 if (ret == -EAGAIN) 195 continue; 196 of_node_put(np); 197 if (ret) 198 goto fail_alias; 199 200 break; 201 } 202 } 203 204 if (!pdev->dev.of_node) 205 pr_warn("%s: Failed to locate of_node [id: %d]\n", 206 cell->name, platform_id); 207 } 208 209 mfd_acpi_add_device(cell, pdev); 210 211 if (cell->pdata_size) { 212 ret = platform_device_add_data(pdev, 213 cell->platform_data, cell->pdata_size); 214 if (ret) 215 goto fail_of_entry; 216 } 217 218 if (cell->swnode) { 219 ret = device_add_software_node(&pdev->dev, cell->swnode); 220 if (ret) 221 goto fail_of_entry; 222 } 223 224 for (r = 0; r < cell->num_resources; r++) { 225 res[r].name = cell->resources[r].name; 226 res[r].flags = cell->resources[r].flags; 227 228 /* Find out base to use */ 229 if ((cell->resources[r].flags & IORESOURCE_MEM) && mem_base) { 230 res[r].parent = mem_base; 231 res[r].start = mem_base->start + 232 cell->resources[r].start; 233 res[r].end = mem_base->start + 234 cell->resources[r].end; 235 } else if (cell->resources[r].flags & IORESOURCE_IRQ) { 236 if (domain) { 237 /* Unable to create mappings for IRQ ranges. */ 238 WARN_ON(cell->resources[r].start != 239 cell->resources[r].end); 240 res[r].start = res[r].end = irq_create_mapping( 241 domain, cell->resources[r].start); 242 } else { 243 res[r].start = irq_base + 244 cell->resources[r].start; 245 res[r].end = irq_base + 246 cell->resources[r].end; 247 } 248 } else { 249 res[r].parent = cell->resources[r].parent; 250 res[r].start = cell->resources[r].start; 251 res[r].end = cell->resources[r].end; 252 } 253 254 if (!cell->ignore_resource_conflicts) { 255 if (has_acpi_companion(&pdev->dev)) { 256 ret = acpi_check_resource_conflict(&res[r]); 257 if (ret) 258 goto fail_res_conflict; 259 } 260 } 261 } 262 263 ret = platform_device_add_resources(pdev, res, cell->num_resources); 264 if (ret) 265 goto fail_res_conflict; 266 267 ret = platform_device_add(pdev); 268 if (ret) 269 goto fail_res_conflict; 270 271 if (cell->pm_runtime_no_callbacks) 272 pm_runtime_no_callbacks(&pdev->dev); 273 274 kfree(res); 275 276 return 0; 277 278 fail_res_conflict: 279 if (cell->swnode) 280 device_remove_software_node(&pdev->dev); 281 fail_of_entry: 282 list_for_each_entry_safe(of_entry, tmp, &mfd_of_node_list, list) 283 if (of_entry->dev == &pdev->dev) { 284 list_del(&of_entry->list); 285 kfree(of_entry); 286 } 287 fail_alias: 288 regulator_bulk_unregister_supply_alias(&pdev->dev, 289 cell->parent_supplies, 290 cell->num_parent_supplies); 291 fail_res: 292 kfree(res); 293 fail_device: 294 platform_device_put(pdev); 295 fail_alloc: 296 return ret; 297 } 298 299 /** 300 * mfd_add_devices - register child devices 301 * 302 * @parent: Pointer to parent device. 303 * @id: Can be PLATFORM_DEVID_AUTO to let the Platform API take care 304 * of device numbering, or will be added to a device's cell_id. 305 * @cells: Array of (struct mfd_cell)s describing child devices. 306 * @n_devs: Number of child devices to register. 307 * @mem_base: Parent register range resource for child devices. 308 * @irq_base: Base of the range of virtual interrupt numbers allocated for 309 * this MFD device. Unused if @domain is specified. 310 * @domain: Interrupt domain to create mappings for hardware interrupts. 311 */ 312 int mfd_add_devices(struct device *parent, int id, 313 const struct mfd_cell *cells, int n_devs, 314 struct resource *mem_base, 315 int irq_base, struct irq_domain *domain) 316 { 317 int i; 318 int ret; 319 320 for (i = 0; i < n_devs; i++) { 321 ret = mfd_add_device(parent, id, cells + i, mem_base, 322 irq_base, domain); 323 if (ret) 324 goto fail; 325 } 326 327 return 0; 328 329 fail: 330 if (i) 331 mfd_remove_devices(parent); 332 333 return ret; 334 } 335 EXPORT_SYMBOL(mfd_add_devices); 336 337 static int mfd_remove_devices_fn(struct device *dev, void *data) 338 { 339 struct platform_device *pdev; 340 const struct mfd_cell *cell; 341 struct mfd_of_node_entry *of_entry, *tmp; 342 int *level = data; 343 344 if (dev->type != &mfd_dev_type) 345 return 0; 346 347 pdev = to_platform_device(dev); 348 cell = mfd_get_cell(pdev); 349 350 if (level && cell->level > *level) 351 return 0; 352 353 if (cell->swnode) 354 device_remove_software_node(&pdev->dev); 355 356 list_for_each_entry_safe(of_entry, tmp, &mfd_of_node_list, list) 357 if (of_entry->dev == &pdev->dev) { 358 list_del(&of_entry->list); 359 kfree(of_entry); 360 } 361 362 regulator_bulk_unregister_supply_alias(dev, cell->parent_supplies, 363 cell->num_parent_supplies); 364 365 platform_device_unregister(pdev); 366 return 0; 367 } 368 369 void mfd_remove_devices_late(struct device *parent) 370 { 371 int level = MFD_DEP_LEVEL_HIGH; 372 373 device_for_each_child_reverse(parent, &level, mfd_remove_devices_fn); 374 } 375 EXPORT_SYMBOL(mfd_remove_devices_late); 376 377 void mfd_remove_devices(struct device *parent) 378 { 379 int level = MFD_DEP_LEVEL_NORMAL; 380 381 device_for_each_child_reverse(parent, &level, mfd_remove_devices_fn); 382 } 383 EXPORT_SYMBOL(mfd_remove_devices); 384 385 static void devm_mfd_dev_release(struct device *dev, void *res) 386 { 387 mfd_remove_devices(dev); 388 } 389 390 /** 391 * devm_mfd_add_devices - Resource managed version of mfd_add_devices() 392 * 393 * Returns 0 on success or an appropriate negative error number on failure. 394 * All child-devices of the MFD will automatically be removed when it gets 395 * unbinded. 396 * 397 * @dev: Pointer to parent device. 398 * @id: Can be PLATFORM_DEVID_AUTO to let the Platform API take care 399 * of device numbering, or will be added to a device's cell_id. 400 * @cells: Array of (struct mfd_cell)s describing child devices. 401 * @n_devs: Number of child devices to register. 402 * @mem_base: Parent register range resource for child devices. 403 * @irq_base: Base of the range of virtual interrupt numbers allocated for 404 * this MFD device. Unused if @domain is specified. 405 * @domain: Interrupt domain to create mappings for hardware interrupts. 406 */ 407 int devm_mfd_add_devices(struct device *dev, int id, 408 const struct mfd_cell *cells, int n_devs, 409 struct resource *mem_base, 410 int irq_base, struct irq_domain *domain) 411 { 412 struct device **ptr; 413 int ret; 414 415 ptr = devres_alloc(devm_mfd_dev_release, sizeof(*ptr), GFP_KERNEL); 416 if (!ptr) 417 return -ENOMEM; 418 419 ret = mfd_add_devices(dev, id, cells, n_devs, mem_base, 420 irq_base, domain); 421 if (ret < 0) { 422 devres_free(ptr); 423 return ret; 424 } 425 426 *ptr = dev; 427 devres_add(dev, ptr); 428 429 return ret; 430 } 431 EXPORT_SYMBOL(devm_mfd_add_devices); 432 433 MODULE_LICENSE("GPL"); 434 MODULE_AUTHOR("Ian Molton, Dmitry Baryshkov"); 435