mfd-core.c (11067B)
1// SPDX-License-Identifier: GPL-2.0-only 2/* 3 * drivers/mfd/mfd-core.c 4 * 5 * core MFD support 6 * Copyright (c) 2006 Ian Molton 7 * Copyright (c) 2007,2008 Dmitry Baryshkov 8 */ 9 10#include <linux/kernel.h> 11#include <linux/platform_device.h> 12#include <linux/acpi.h> 13#include <linux/list.h> 14#include <linux/property.h> 15#include <linux/mfd/core.h> 16#include <linux/pm_runtime.h> 17#include <linux/slab.h> 18#include <linux/module.h> 19#include <linux/irqdomain.h> 20#include <linux/of.h> 21#include <linux/of_address.h> 22#include <linux/regulator/consumer.h> 23 24static LIST_HEAD(mfd_of_node_list); 25 26struct mfd_of_node_entry { 27 struct list_head list; 28 struct device *dev; 29 struct device_node *np; 30}; 31 32static struct device_type mfd_dev_type = { 33 .name = "mfd_device", 34}; 35 36int mfd_cell_enable(struct platform_device *pdev) 37{ 38 const struct mfd_cell *cell = mfd_get_cell(pdev); 39 40 if (!cell->enable) { 41 dev_dbg(&pdev->dev, "No .enable() call-back registered\n"); 42 return 0; 43 } 44 45 return cell->enable(pdev); 46} 47EXPORT_SYMBOL(mfd_cell_enable); 48 49int mfd_cell_disable(struct platform_device *pdev) 50{ 51 const struct mfd_cell *cell = mfd_get_cell(pdev); 52 53 if (!cell->disable) { 54 dev_dbg(&pdev->dev, "No .disable() call-back registered\n"); 55 return 0; 56 } 57 58 return cell->disable(pdev); 59} 60EXPORT_SYMBOL(mfd_cell_disable); 61 62#if IS_ENABLED(CONFIG_ACPI) 63static void mfd_acpi_add_device(const struct mfd_cell *cell, 64 struct platform_device *pdev) 65{ 66 const struct mfd_cell_acpi_match *match = cell->acpi_match; 67 struct acpi_device *parent, *child; 68 struct acpi_device *adev = NULL; 69 70 parent = ACPI_COMPANION(pdev->dev.parent); 71 if (!parent) 72 return; 73 74 /* 75 * MFD child device gets its ACPI handle either from the ACPI device 76 * directly under the parent that matches the either _HID or _CID, or 77 * _ADR or it will use the parent handle if is no ID is given. 78 * 79 * Note that use of _ADR is a grey area in the ACPI specification, 80 * though at least Intel Galileo Gen 2 is using it to distinguish 81 * the children devices. 82 */ 83 if (match) { 84 if (match->pnpid) { 85 struct acpi_device_id ids[2] = {}; 86 87 strlcpy(ids[0].id, match->pnpid, sizeof(ids[0].id)); 88 list_for_each_entry(child, &parent->children, node) { 89 if (!acpi_match_device_ids(child, ids)) { 90 adev = child; 91 break; 92 } 93 } 94 } else { 95 adev = acpi_find_child_device(parent, match->adr, false); 96 } 97 } 98 99 ACPI_COMPANION_SET(&pdev->dev, adev ?: parent); 100} 101#else 102static inline void mfd_acpi_add_device(const struct mfd_cell *cell, 103 struct platform_device *pdev) 104{ 105} 106#endif 107 108static int mfd_match_of_node_to_dev(struct platform_device *pdev, 109 struct device_node *np, 110 const struct mfd_cell *cell) 111{ 112#if IS_ENABLED(CONFIG_OF) 113 struct mfd_of_node_entry *of_entry; 114 const __be32 *reg; 115 u64 of_node_addr; 116 117 /* Skip if OF node has previously been allocated to a device */ 118 list_for_each_entry(of_entry, &mfd_of_node_list, list) 119 if (of_entry->np == np) 120 return -EAGAIN; 121 122 if (!cell->use_of_reg) 123 /* No of_reg defined - allocate first free compatible match */ 124 goto allocate_of_node; 125 126 /* We only care about each node's first defined address */ 127 reg = of_get_address(np, 0, NULL, NULL); 128 if (!reg) 129 /* OF node does not contatin a 'reg' property to match to */ 130 return -EAGAIN; 131 132 of_node_addr = of_read_number(reg, of_n_addr_cells(np)); 133 134 if (cell->of_reg != of_node_addr) 135 /* No match */ 136 return -EAGAIN; 137 138allocate_of_node: 139 of_entry = kzalloc(sizeof(*of_entry), GFP_KERNEL); 140 if (!of_entry) 141 return -ENOMEM; 142 143 of_entry->dev = &pdev->dev; 144 of_entry->np = np; 145 list_add_tail(&of_entry->list, &mfd_of_node_list); 146 147 pdev->dev.of_node = np; 148 pdev->dev.fwnode = &np->fwnode; 149#endif 150 return 0; 151} 152 153static int mfd_add_device(struct device *parent, int id, 154 const struct mfd_cell *cell, 155 struct resource *mem_base, 156 int irq_base, struct irq_domain *domain) 157{ 158 struct resource *res; 159 struct platform_device *pdev; 160 struct device_node *np = NULL; 161 struct mfd_of_node_entry *of_entry, *tmp; 162 int ret = -ENOMEM; 163 int platform_id; 164 int r; 165 166 if (id == PLATFORM_DEVID_AUTO) 167 platform_id = id; 168 else 169 platform_id = id + cell->id; 170 171 pdev = platform_device_alloc(cell->name, platform_id); 172 if (!pdev) 173 goto fail_alloc; 174 175 pdev->mfd_cell = kmemdup(cell, sizeof(*cell), GFP_KERNEL); 176 if (!pdev->mfd_cell) 177 goto fail_device; 178 179 res = kcalloc(cell->num_resources, sizeof(*res), GFP_KERNEL); 180 if (!res) 181 goto fail_device; 182 183 pdev->dev.parent = parent; 184 pdev->dev.type = &mfd_dev_type; 185 pdev->dev.dma_mask = parent->dma_mask; 186 pdev->dev.dma_parms = parent->dma_parms; 187 pdev->dev.coherent_dma_mask = parent->coherent_dma_mask; 188 189 ret = regulator_bulk_register_supply_alias( 190 &pdev->dev, cell->parent_supplies, 191 parent, cell->parent_supplies, 192 cell->num_parent_supplies); 193 if (ret < 0) 194 goto fail_res; 195 196 if (IS_ENABLED(CONFIG_OF) && parent->of_node && cell->of_compatible) { 197 for_each_child_of_node(parent->of_node, np) { 198 if (of_device_is_compatible(np, cell->of_compatible)) { 199 /* Ignore 'disabled' devices error free */ 200 if (!of_device_is_available(np)) { 201 of_node_put(np); 202 ret = 0; 203 goto fail_alias; 204 } 205 206 ret = mfd_match_of_node_to_dev(pdev, np, cell); 207 if (ret == -EAGAIN) 208 continue; 209 of_node_put(np); 210 if (ret) 211 goto fail_alias; 212 213 break; 214 } 215 } 216 217 if (!pdev->dev.of_node) 218 pr_warn("%s: Failed to locate of_node [id: %d]\n", 219 cell->name, platform_id); 220 } 221 222 mfd_acpi_add_device(cell, pdev); 223 224 if (cell->pdata_size) { 225 ret = platform_device_add_data(pdev, 226 cell->platform_data, cell->pdata_size); 227 if (ret) 228 goto fail_of_entry; 229 } 230 231 if (cell->swnode) { 232 ret = device_add_software_node(&pdev->dev, cell->swnode); 233 if (ret) 234 goto fail_of_entry; 235 } 236 237 for (r = 0; r < cell->num_resources; r++) { 238 res[r].name = cell->resources[r].name; 239 res[r].flags = cell->resources[r].flags; 240 241 /* Find out base to use */ 242 if ((cell->resources[r].flags & IORESOURCE_MEM) && mem_base) { 243 res[r].parent = mem_base; 244 res[r].start = mem_base->start + 245 cell->resources[r].start; 246 res[r].end = mem_base->start + 247 cell->resources[r].end; 248 } else if (cell->resources[r].flags & IORESOURCE_IRQ) { 249 if (domain) { 250 /* Unable to create mappings for IRQ ranges. */ 251 WARN_ON(cell->resources[r].start != 252 cell->resources[r].end); 253 res[r].start = res[r].end = irq_create_mapping( 254 domain, cell->resources[r].start); 255 } else { 256 res[r].start = irq_base + 257 cell->resources[r].start; 258 res[r].end = irq_base + 259 cell->resources[r].end; 260 } 261 } else { 262 res[r].parent = cell->resources[r].parent; 263 res[r].start = cell->resources[r].start; 264 res[r].end = cell->resources[r].end; 265 } 266 267 if (!cell->ignore_resource_conflicts) { 268 if (has_acpi_companion(&pdev->dev)) { 269 ret = acpi_check_resource_conflict(&res[r]); 270 if (ret) 271 goto fail_res_conflict; 272 } 273 } 274 } 275 276 ret = platform_device_add_resources(pdev, res, cell->num_resources); 277 if (ret) 278 goto fail_res_conflict; 279 280 ret = platform_device_add(pdev); 281 if (ret) 282 goto fail_res_conflict; 283 284 if (cell->pm_runtime_no_callbacks) 285 pm_runtime_no_callbacks(&pdev->dev); 286 287 kfree(res); 288 289 return 0; 290 291fail_res_conflict: 292 if (cell->swnode) 293 device_remove_software_node(&pdev->dev); 294fail_of_entry: 295 list_for_each_entry_safe(of_entry, tmp, &mfd_of_node_list, list) 296 if (of_entry->dev == &pdev->dev) { 297 list_del(&of_entry->list); 298 kfree(of_entry); 299 } 300fail_alias: 301 regulator_bulk_unregister_supply_alias(&pdev->dev, 302 cell->parent_supplies, 303 cell->num_parent_supplies); 304fail_res: 305 kfree(res); 306fail_device: 307 platform_device_put(pdev); 308fail_alloc: 309 return ret; 310} 311 312/** 313 * mfd_add_devices - register child devices 314 * 315 * @parent: Pointer to parent device. 316 * @id: Can be PLATFORM_DEVID_AUTO to let the Platform API take care 317 * of device numbering, or will be added to a device's cell_id. 318 * @cells: Array of (struct mfd_cell)s describing child devices. 319 * @n_devs: Number of child devices to register. 320 * @mem_base: Parent register range resource for child devices. 321 * @irq_base: Base of the range of virtual interrupt numbers allocated for 322 * this MFD device. Unused if @domain is specified. 323 * @domain: Interrupt domain to create mappings for hardware interrupts. 324 */ 325int mfd_add_devices(struct device *parent, int id, 326 const struct mfd_cell *cells, int n_devs, 327 struct resource *mem_base, 328 int irq_base, struct irq_domain *domain) 329{ 330 int i; 331 int ret; 332 333 for (i = 0; i < n_devs; i++) { 334 ret = mfd_add_device(parent, id, cells + i, mem_base, 335 irq_base, domain); 336 if (ret) 337 goto fail; 338 } 339 340 return 0; 341 342fail: 343 if (i) 344 mfd_remove_devices(parent); 345 346 return ret; 347} 348EXPORT_SYMBOL(mfd_add_devices); 349 350static int mfd_remove_devices_fn(struct device *dev, void *data) 351{ 352 struct platform_device *pdev; 353 const struct mfd_cell *cell; 354 int *level = data; 355 356 if (dev->type != &mfd_dev_type) 357 return 0; 358 359 pdev = to_platform_device(dev); 360 cell = mfd_get_cell(pdev); 361 362 if (level && cell->level > *level) 363 return 0; 364 365 if (cell->swnode) 366 device_remove_software_node(&pdev->dev); 367 368 regulator_bulk_unregister_supply_alias(dev, cell->parent_supplies, 369 cell->num_parent_supplies); 370 371 platform_device_unregister(pdev); 372 return 0; 373} 374 375void mfd_remove_devices_late(struct device *parent) 376{ 377 int level = MFD_DEP_LEVEL_HIGH; 378 379 device_for_each_child_reverse(parent, &level, mfd_remove_devices_fn); 380} 381EXPORT_SYMBOL(mfd_remove_devices_late); 382 383void mfd_remove_devices(struct device *parent) 384{ 385 int level = MFD_DEP_LEVEL_NORMAL; 386 387 device_for_each_child_reverse(parent, &level, mfd_remove_devices_fn); 388} 389EXPORT_SYMBOL(mfd_remove_devices); 390 391static void devm_mfd_dev_release(struct device *dev, void *res) 392{ 393 mfd_remove_devices(dev); 394} 395 396/** 397 * devm_mfd_add_devices - Resource managed version of mfd_add_devices() 398 * 399 * Returns 0 on success or an appropriate negative error number on failure. 400 * All child-devices of the MFD will automatically be removed when it gets 401 * unbinded. 402 * 403 * @dev: Pointer to parent device. 404 * @id: Can be PLATFORM_DEVID_AUTO to let the Platform API take care 405 * of device numbering, or will be added to a device's cell_id. 406 * @cells: Array of (struct mfd_cell)s describing child devices. 407 * @n_devs: Number of child devices to register. 408 * @mem_base: Parent register range resource for child devices. 409 * @irq_base: Base of the range of virtual interrupt numbers allocated for 410 * this MFD device. Unused if @domain is specified. 411 * @domain: Interrupt domain to create mappings for hardware interrupts. 412 */ 413int devm_mfd_add_devices(struct device *dev, int id, 414 const struct mfd_cell *cells, int n_devs, 415 struct resource *mem_base, 416 int irq_base, struct irq_domain *domain) 417{ 418 struct device **ptr; 419 int ret; 420 421 ptr = devres_alloc(devm_mfd_dev_release, sizeof(*ptr), GFP_KERNEL); 422 if (!ptr) 423 return -ENOMEM; 424 425 ret = mfd_add_devices(dev, id, cells, n_devs, mem_base, 426 irq_base, domain); 427 if (ret < 0) { 428 devres_free(ptr); 429 return ret; 430 } 431 432 *ptr = dev; 433 devres_add(dev, ptr); 434 435 return ret; 436} 437EXPORT_SYMBOL(devm_mfd_add_devices); 438 439MODULE_LICENSE("GPL"); 440MODULE_AUTHOR("Ian Molton, Dmitry Baryshkov");