| /linux/lib/ |
| H A D | devres.c | 36 dev_to_node(dev)); in __devm_ioremap() 286 dev_to_node(dev)); in devm_ioport_map() 337 dev_to_node(dev)); in devm_arch_phys_wc_add() 383 dev_to_node(dev)); in devm_arch_io_reserve_memtype_wc()
|
| /linux/include/linux/ |
| H A D | async.h | 90 return async_schedule_node(func, dev, dev_to_node(dev)); in async_schedule_dev() 114 return async_schedule_node_domain(func, dev, dev_to_node(dev), domain); in async_schedule_dev_domain()
|
| /linux/drivers/usb/host/ |
| H A D | xhci-mem.c | 38 seg = kzalloc_node(sizeof(*seg), flags, dev_to_node(dev)); in xhci_segment_alloc() 50 dev_to_node(dev)); in xhci_segment_alloc() 377 ring = kzalloc_node(sizeof(*ring), flags, dev_to_node(dev)); in xhci_ring_alloc() 461 ctx = kzalloc_node(sizeof(*ctx), flags, dev_to_node(dev)); in xhci_alloc_container_ctx() 493 ctx = kzalloc_node(sizeof(*ctx), flags, dev_to_node(dev)); in xhci_alloc_port_bw_ctx() 631 dev_to_node(dev)); in xhci_alloc_stream_info() 641 dev_to_node(dev)); in xhci_alloc_stream_info() 842 dev_to_node(dev)); in xhci_alloc_tt_info() 1656 dev_to_node(dev)); in scratchpad_alloc() 1667 flags, dev_to_node(dev)); in scratchpad_alloc() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | qat_crypto.c | 60 if ((node == dev_to_node(&GET_DEV(tmp_dev)) || in qat_crypto_get_instance_node() 61 dev_to_node(&GET_DEV(tmp_dev)) < 0) && in qat_crypto_get_instance_node() 153 dev_to_node(&GET_DEV(accel_dev))); in qat_crypto_create_instances()
|
| H A D | adf_transport.c | 407 dev_to_node(&GET_DEV(accel_dev))); in adf_init_bank() 430 dev_to_node(&GET_DEV(accel_dev))); in adf_init_bank() 483 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data() 490 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data()
|
| /linux/net/core/ |
| H A D | devmem.c | 208 dev_to_node(&dev->dev)); in net_devmem_bind_dmabuf() 257 dev_to_node(&dev->dev)); in net_devmem_bind_dmabuf() 271 dev_to_node(&dev->dev)); in net_devmem_bind_dmabuf() 283 dma_addr, len, dev_to_node(&dev->dev), in net_devmem_bind_dmabuf()
|
| /linux/drivers/crypto/intel/qat/qat_dh895xccvf/ |
| H A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_c62xvf/ |
| H A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/drivers/crypto/intel/qat/qat_c3xxxvf/ |
| H A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
| /linux/arch/riscv/include/asm/ |
| H A D | pci.h | 21 return dev_to_node(&bus->dev); in pcibus_to_node()
|
| /linux/drivers/base/test/ |
| H A D | test_async_driver_probe.c | 48 dev_to_node(dev) != numa_node_id()) { in test_probe() 50 dev_to_node(dev), numa_node_id()); in test_probe()
|
| /linux/arch/arm64/kernel/ |
| H A D | pci.c | 38 return dev_to_node(&bus->dev); in pcibus_to_node()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | ptp.c | 355 node = dev_to_node(mlx5_core_dma_dev(mdev)); in mlx5e_ptp_alloc_txqsq() 490 err = mlx5e_ptp_alloc_traffic_db(ptpsq, dev_to_node(mlx5_core_dma_dev(c->mdev))); in mlx5e_ptp_open_txqsq() 572 ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev)); in mlx5e_ptp_open_tx_cqs() 622 ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev)); in mlx5e_ptp_open_rx_cq() 727 int node = dev_to_node(c->mdev->device); in mlx5e_ptp_open_rq() 885 c = kvzalloc_node(sizeof(*c), GFP_KERNEL, dev_to_node(mlx5_core_dma_dev(mdev))); in mlx5e_ptp_open()
|
| /linux/drivers/dma/idxd/ |
| H A D | dma.c | 204 idxd_dma = kzalloc_node(sizeof(*idxd_dma), GFP_KERNEL, dev_to_node(dev)); in idxd_register_dma_device() 257 idxd_chan = kzalloc_node(sizeof(*idxd_chan), GFP_KERNEL, dev_to_node(dev)); in idxd_register_dma_channel()
|
| /linux/drivers/nvdimm/ |
| H A D | of_pmem.c | 63 ndr_desc.numa_node = dev_to_node(&pdev->dev); in of_pmem_region_probe()
|
| /linux/drivers/perf/ |
| H A D | fujitsu_uncore_pmu.c | 417 node = dev_to_node(uncorepmu->dev); in fujitsu_uncore_pmu_online_cpu() 434 node = dev_to_node(uncorepmu->dev); in fujitsu_uncore_pmu_offline_cpu() 466 uncorepmu->cpu = cpumask_local_spread(0, dev_to_node(dev)); in fujitsu_uncore_pmu_probe()
|
| /linux/drivers/crypto/virtio/ |
| H A D | virtio_crypto_skcipher_algs.c | 355 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req() 360 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req() 417 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req()
|
| H A D | virtio_crypto_core.c | 369 if (num_possible_nodes() > 1 && dev_to_node(&vdev->dev) < 0) { in virtcrypto_probe() 380 dev_to_node(&vdev->dev)); in virtcrypto_probe()
|
| H A D | virtio_crypto_akcipher_algs.c | 222 int node = dev_to_node(&vcrypto->vdev->dev); in __virtio_crypto_akcipher_do_req() 285 GFP_KERNEL, dev_to_node(&vcrypto->vdev->dev)); in virtio_crypto_rsa_do_req()
|
| /linux/kernel/dma/ |
| H A D | ops_helpers.c | 71 page = alloc_pages_node(dev_to_node(dev), gfp, get_order(size)); in dma_common_alloc_pages()
|
| /linux/arch/loongarch/kernel/ |
| H A D | numa.c | 265 return dev_to_node(&bus->dev); in pcibus_to_node()
|
| /linux/kernel/ |
| H A D | async.c | 272 __async_schedule_node_domain(func, dev, dev_to_node(dev), in async_schedule_dev_nocall()
|
| /linux/drivers/net/ethernet/fungible/funcore/ |
| H A D | fun_queue.c | 24 int dev_node = dev_to_node(dma_dev); in fun_alloc_ring_mem() 189 int i, node = dev_to_node(dev); in fun_fill_rq()
|
| /linux/arch/x86/pci/ |
| H A D | intel_mid.c | 273 ioapic_set_alloc_attr(&info, dev_to_node(&dev->dev), 1, polarity_low); in intel_mid_pci_irq_enable()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | alloc.c | 65 original_node = dev_to_node(device); in mlx5_dma_zalloc_coherent_node()
|