| /linux/arch/mips/include/asm/ |
| H A D | smp.h | 72 extern const struct plat_smp_ops *mp_ops; /* private */ in arch_smp_send_reschedule() 74 mp_ops->send_ipi_single(cpu, SMP_RESCHEDULE_YOURSELF); in arch_smp_send_reschedule() 80 extern const struct plat_smp_ops *mp_ops; /* private */ in __cpu_disable() 82 return mp_ops->cpu_disable(); in __cpu_disable() 87 extern const struct plat_smp_ops *mp_ops; /* private */ in __cpu_die() 89 mp_ops->cpu_die(cpu); in __cpu_die() 98 extern const struct plat_smp_ops *mp_ops; /* private */ in kexec_nonboot_cpu() 100 return mp_ops->kexec_nonboot_cpu(); in kexec_nonboot_cpu() 105 extern const struct plat_smp_ops *mp_ops; /* private */ in kexec_nonboot_cpu_func() 107 return mp_ops->kexec_nonboot_cpu; in kexec_nonboot_cpu_func() [all …]
|
| H A D | smp-ops.h | 45 extern const struct plat_smp_ops *mp_ops; /* private */ in plat_smp_setup() 47 mp_ops->smp_setup(); in plat_smp_setup()
|
| /linux/arch/sh/include/asm/ |
| H A D | smp.h | 51 extern struct plat_smp_ops *mp_ops; /* private */ in __cpu_die() 53 mp_ops->cpu_die(cpu); in __cpu_die() 59 extern struct plat_smp_ops *mp_ops; /* private */ in hard_smp_processor_id() 61 if (!mp_ops) in hard_smp_processor_id() 64 return mp_ops->smp_processor_id(); in hard_smp_processor_id()
|
| H A D | smp-ops.h | 16 extern struct plat_smp_ops *mp_ops; 23 BUG_ON(!mp_ops); in plat_smp_setup() 24 mp_ops->smp_setup(); in plat_smp_setup() 29 mp_ops->play_dead(); in play_dead()
|
| /linux/arch/sh/kernel/ |
| H A D | smp.c | 36 struct plat_smp_ops *mp_ops = NULL; variable 43 if (mp_ops) in register_smp_ops() 46 mp_ops = ops; in register_smp_ops() 64 mp_ops->prepare_cpus(max_cpus); in smp_prepare_cpus() 129 ret = mp_ops->cpu_disable(cpu); in __cpu_disable() 230 mp_ops->start_cpu(cpu, (unsigned long)_stext); in __cpu_up() 263 mp_ops->send_ipi(cpu, SMP_MSG_RESCHEDULE); in arch_smp_send_reschedule() 276 mp_ops->send_ipi(cpu, SMP_MSG_FUNCTION); in arch_send_call_function_ipi_mask() 281 mp_ops->send_ipi(cpu, SMP_MSG_FUNCTION_SINGLE); in arch_send_call_function_single_ipi() 290 mp_ops->send_ipi(cpu, SMP_MSG_TIMER); in tick_broadcast()
|
| /linux/arch/mips/kernel/ |
| H A D | smp.c | 160 const struct plat_smp_ops *mp_ops; variable 161 EXPORT_SYMBOL(mp_ops); 165 if (mp_ops) in register_smp_ops() 168 mp_ops = ops; in register_smp_ops() 365 mp_ops->init_secondary(); in start_secondary() 411 mp_ops->smp_finish(); in start_secondary() 442 mp_ops->prepare_cpus(max_cpus); in smp_prepare_cpus() 455 if (mp_ops->prepare_boot_cpu) in smp_prepare_boot_cpu() 456 mp_ops->prepare_boot_cpu(); in smp_prepare_boot_cpu() 464 return mp_ops->boot_secondary(cpu, tidle); in arch_cpuhp_kick_ap_alive() [all …]
|
| H A D | smp-cps.c | 960 extern const struct plat_smp_ops *mp_ops; in mips_cps_smp_in_use() 961 return mp_ops == &cps_smp_ops; in mips_cps_smp_in_use()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | smp.h | 22 extern struct smp_ops mp_ops; 101 mp_ops.send_ipi_single(cpu, ACTION_CALL_FUNCTION); in arch_send_call_function_single_ipi() 106 mp_ops.send_ipi_mask(mask, ACTION_CALL_FUNCTION); in arch_send_call_function_ipi_mask()
|
| /linux/arch/loongarch/kernel/ |
| H A D | paravirt.c | 209 native_ops = mp_ops; in pv_ipi_init() 210 mp_ops.init_ipi = pv_init_ipi; in pv_ipi_init() 211 mp_ops.send_ipi_single = pv_send_ipi_single; in pv_ipi_init() 212 mp_ops.send_ipi_mask = pv_send_ipi_mask; in pv_ipi_init()
|
| H A D | irq.c | 114 mp_ops.init_ipi(); in init_IRQ()
|
| H A D | smp.c | 255 mp_ops.send_ipi_single(cpu, ACTION_RESCHEDULE); in arch_smp_send_reschedule() 262 mp_ops.send_ipi_single(smp_processor_id(), ACTION_IRQ_WORK); in arch_irq_work_raise() 310 struct smp_ops mp_ops = { variable
|
| /linux/net/core/ |
| H A D | page_pool.c | 282 pool->mp_ops = rxq->mp_params.mp_ops; in page_pool_init() 285 if (pool->mp_ops) { in page_pool_init() 291 if (WARN_ON(!is_kernel_rodata((unsigned long)pool->mp_ops))) { in page_pool_init() 296 err = pool->mp_ops->init(pool); in page_pool_init() 664 if (static_branch_unlikely(&page_pool_mem_providers) && pool->mp_ops) in page_pool_alloc_netmems() 665 netmem = pool->mp_ops->alloc_netmems(pool, gfp); in page_pool_alloc_netmems() 759 if (static_branch_unlikely(&page_pool_mem_providers) && pool->mp_ops) in page_pool_return_netmem() 760 put = pool->mp_ops->release_netmem(pool, netmem); in page_pool_return_netmem() 1129 if (pool->mp_ops) { in __page_pool_destroy() [all...] |
| H A D | page_pool_user.c | 251 if (pool->mp_ops && pool->mp_ops->nl_fill(pool->mp_priv, rsp, NULL)) in page_pool_nl_fill()
|
| H A D | devmem.c | 138 .mp_ops = &dmabuf_devmem_ops, in net_devmem_unbind_dmabuf() 155 .mp_ops = &dmabuf_devmem_ops, in net_devmem_bind_dmabuf_to_queue()
|
| H A D | dev.c | 12273 if (p->mp_ops && p->mp_ops->uninstall) in dev_memory_provider_uninstall() 12274 p->mp_ops->uninstall(rxq->mp_params.mp_priv, rxq); in dev_memory_provider_uninstall()
|
| /linux/include/net/page_pool/ |
| H A D | types.h | 163 const struct memory_provider_ops *mp_ops; member 226 const struct memory_provider_ops *mp_ops; member
|
| /linux/drivers/irqchip/ |
| H A D | irq-loongarch-avec.c | 96 mp_ops.send_ipi_single(adata->prev_cpu, ACTION_CLEAR_VECTOR); in avecintc_sync() 194 mp_ops.send_ipi_single(cpu, ACTION_CLEAR_VECTOR); in complete_irq_moving()
|
| /linux/io_uring/ |
| H A D | zcrx.c | 510 .mp_ops = &io_uring_pp_zc_ops, in io_close_queue() 807 mp_param.mp_ops = &io_uring_pp_zc_ops; in io_register_zcrx_ifq() 1052 p->mp_ops = NULL; in io_pp_uninstall() 1305 if (!pp || pp->mp_ops != &io_uring_pp_zc_ops || io_pp_to_ifq(pp) != ifq) in io_zcrx_recv_frag()
|