| /linux/tools/testing/selftests/kvm/lib/loongarch/ |
| H A D | processor.c | 190 struct handlers *handlers; in route_exception() local 192 handlers = (struct handlers *)exception_handlers; in route_exception() 194 if (handlers && handlers->exception_handlers[vector]) in route_exception() 195 return handlers->exception_handlers[vector](regs); in route_exception() 208 vm->handlers = __vm_alloc(vm, sizeof(struct handlers), in vm_init_descriptor_tables() 212 addr = addr_gva2hva(vm, vm->handlers); in vm_init_descriptor_tables() 219 struct handlers *handlers = addr_gva2hva(vm, vm->handlers); vm_install_exception_handler() local [all...] |
| /linux/include/net/ |
| H A D | rtnetlink.h | 56 int __rtnl_register_many(const struct rtnl_msg_handler *handlers, int n); 57 void __rtnl_unregister_many(const struct rtnl_msg_handler *handlers, int n); 59 #define rtnl_register_many(handlers) \ argument 60 __rtnl_register_many(handlers, ARRAY_SIZE(handlers)) 61 #define rtnl_unregister_many(handlers) \ argument 62 __rtnl_unregister_many(handlers, ARRAY_SIZE(handlers))
|
| /linux/tools/testing/selftests/kvm/lib/riscv/ |
| H A D | processor.c | 416 struct handlers { in route_exception() 422 struct handlers *handlers = (struct handlers *)exception_handlers; in route_exception() 435 if (handlers && handlers->exception_handlers[vector][ec]) in vm_init_vector_tables() 436 return handlers->exception_handlers[vector][ec](regs); in vm_init_vector_tables() 451 vm->handlers = __vm_alloc(vm, sizeof(struct handlers), vm->page_size, in vm_install_interrupt_handler() 454 *(gva_t *)addr_gva2hva(vm, (gva_t)(&exception_handlers)) = vm->handlers; in vm_install_interrupt_handler() 401 struct handlers { global() struct 402 exception_handlershandlers global() argument 407 struct handlers *handlers = (struct handlers *)exception_handlers; route_exception() local 444 struct handlers *handlers = addr_gva2hva(vm, vm->handlers); vm_install_exception_handler() local 452 struct handlers *handlers = addr_gva2hva(vm, vm->handlers); vm_install_interrupt_handler() local [all...] |
| /linux/tools/testing/selftests/kvm/lib/arm64/ |
| H A D | processor.c | 492 struct handlers { in assert_on_unhandled_exception() 505 struct handlers *handlers = (struct handlers *)exception_handlers; in vcpu_init_descriptor_tables() 529 if (handlers && handlers->exception_handlers[vector][ec]) in route_exception() 530 return handlers->exception_handlers[vector][ec](regs); in route_exception() 538 vm->handlers = __vm_alloc(vm, sizeof(struct handlers), vm->page_size, in route_exception() 541 *(gva_t *)addr_gva2hva(vm, (gva_t)(&exception_handlers)) = vm->handlers; in vm_init_descriptor_tables() 496 struct handlers { global() struct 497 exception_handlershandlers global() argument 509 struct handlers *handlers = (struct handlers *)exception_handlers; route_exception() local 551 struct handlers *handlers = addr_gva2hva(vm, vm->handlers); vm_install_sync_handler() local 562 struct handlers *handlers = addr_gva2hva(vm, vm->handlers); vm_install_exception_handler() local [all...] |
| /linux/arch/powerpc/platforms/powermac/ |
| H A D | pfunc_core.c | 124 #define PMF_PARSE_CALL(name, cmd, handlers, p...) \ argument 128 if (handlers == NULL) \ 130 if (handlers->name) \ 131 return handlers->name(cmd->func, cmd->instdata, \ 542 struct pmf_handlers *handlers; member 580 struct pmf_handlers *handlers, in pmf_parse_one() argument 596 handlers ? "executing" : "parsing"); in pmf_parse_one() 621 rc = pmf_parsers[ccode](&cmd, handlers); in pmf_parse_one() 630 if (handlers == NULL) in pmf_parse_one() 702 struct pmf_handlers *handlers, in pmf_register_driver() argument [all …]
|
| /linux/fs/xfs/ |
| H A D | xfs_fsmap.c | 1160 struct xfs_getfsmap_dev handlers[XFS_GETFSMAP_DEVS]; in xfs_getfsmap() local 1182 memset(handlers, 0, sizeof(handlers)); in xfs_getfsmap() 1183 handlers[0].nr_sectors = XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks); in xfs_getfsmap() 1184 handlers[0].dev = xfs_getfsmap_device(mp, XFS_DEV_DATA); in xfs_getfsmap() 1186 handlers[0].fn = xfs_getfsmap_datadev_rmapbt; in xfs_getfsmap() 1188 handlers[0].fn = xfs_getfsmap_datadev_bnobt; in xfs_getfsmap() 1190 handlers[1].nr_sectors = XFS_FSB_TO_BB(mp, in xfs_getfsmap() 1192 handlers[1].dev = xfs_getfsmap_device(mp, XFS_DEV_LOG); in xfs_getfsmap() 1193 handlers[1].fn = xfs_getfsmap_logdev; in xfs_getfsmap() 1201 handlers[2].nr_sectors = XFS_FSB_TO_BB(mp, mp->m_sb.sb_rblocks); in xfs_getfsmap() [all …]
|
| /linux/fs/btrfs/ |
| H A D | props.c | 48 const struct hlist_head *handlers) in find_prop_handler() argument 52 if (!handlers) { in find_prop_handler() 55 handlers = find_prop_handlers_by_hash(hash); in find_prop_handler() 56 if (!handlers) in find_prop_handler() 60 hlist_for_each_entry(h, handlers, node) in find_prop_handler() 167 const struct hlist_head *handlers; in iterate_object_props() local 187 handlers = find_prop_handlers_by_hash(key.offset); in iterate_object_props() 188 if (!handlers) in iterate_object_props() 223 handler = find_prop_handler(name_buf, handlers); in iterate_object_props()
|
| /linux/fs/ext4/ |
| H A D | fsmap.c | 707 struct ext4_getfsmap_dev handlers[EXT4_GETFSMAP_DEVS]; in ext4_getfsmap() local 721 memset(handlers, 0, sizeof(handlers)); in ext4_getfsmap() 722 handlers[0].gfd_dev = new_encode_dev(sb->s_bdev->bd_dev); in ext4_getfsmap() 723 handlers[0].gfd_fn = ext4_getfsmap_datadev; in ext4_getfsmap() 725 handlers[1].gfd_dev = new_encode_dev( in ext4_getfsmap() 727 handlers[1].gfd_fn = ext4_getfsmap_logdev; in ext4_getfsmap() 730 sort(handlers, EXT4_GETFSMAP_DEVS, sizeof(struct ext4_getfsmap_dev), in ext4_getfsmap() 762 if (!handlers[i].gfd_fn) in ext4_getfsmap() 764 if (head->fmh_keys[0].fmr_device > handlers[i].gfd_dev) in ext4_getfsmap() 766 if (head->fmh_keys[1].fmr_device < handlers[i].gfd_dev) in ext4_getfsmap() [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/dvm/ |
| H A D | rx.c | 947 void (**handlers)(struct iwl_priv *priv, struct iwl_rx_cmd_buffer *rxb); in iwl_setup_rx_handlers() local 949 handlers = priv->rx_handlers; in iwl_setup_rx_handlers() 951 handlers[REPLY_ERROR] = iwlagn_rx_reply_error; in iwl_setup_rx_handlers() 952 handlers[CHANNEL_SWITCH_NOTIFICATION] = iwlagn_rx_csa; in iwl_setup_rx_handlers() 953 handlers[SPECTRUM_MEASURE_NOTIFICATION] = in iwl_setup_rx_handlers() 955 handlers[PM_SLEEP_NOTIFICATION] = iwlagn_rx_pm_sleep_notif; in iwl_setup_rx_handlers() 956 handlers[PM_DEBUG_STATISTIC_NOTIFIC] = in iwl_setup_rx_handlers() 958 handlers[BEACON_NOTIFICATION] = iwlagn_rx_beacon_notif; in iwl_setup_rx_handlers() 959 handlers[REPLY_ADD_STA] = iwl_add_sta_callback; in iwl_setup_rx_handlers() 961 handlers[REPLY_WIPAN_NOA_NOTIFICATION] = iwlagn_rx_noa_notification; in iwl_setup_rx_handlers() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | cxgb3_offload.c | 708 t3c_tid->client->handlers && in do_act_open_rpl() 709 t3c_tid->client->handlers[CPL_ACT_OPEN_RPL]) { in do_act_open_rpl() 710 return t3c_tid->client->handlers[CPL_ACT_OPEN_RPL] (dev, skb, in do_act_open_rpl() 727 if (t3c_tid && t3c_tid->ctx && t3c_tid->client->handlers && in do_stid_rpl() 728 t3c_tid->client->handlers[p->opcode]) { in do_stid_rpl() 729 return t3c_tid->client->handlers[p->opcode] (dev, skb, in do_stid_rpl() 745 if (t3c_tid && t3c_tid->ctx && t3c_tid->client->handlers && in do_hwtid_rpl() 746 t3c_tid->client->handlers[p->opcode]) { in do_hwtid_rpl() 747 return t3c_tid->client->handlers[p->opcode] in do_hwtid_rpl() 772 if (t3c_tid && t3c_tid->ctx && t3c_tid->client->handlers && in do_cr() [all …]
|
| /linux/Documentation/driver-api/acpi/ |
| H A D | scan_handlers.rst | 39 called ACPI scan handlers represented by the following structure:: 49 take care of, list_node is the hook to the global list of ACPI scan handlers 57 available scan handlers. If a matching scan handler is found, its .attach() 72 callbacks from the scan handlers of all device nodes in the given namespace 73 scope (if they have scan handlers). Next, it unregisters all of the device 76 ACPI scan handlers can be added to the list maintained by the ACPI core with the 78 handler as an argument. The order in which scan handlers are added to the list
|
| /linux/drivers/soc/fsl/qbman/ |
| H A D | qman_test_stash.c | 161 struct list_head handlers; member 326 INIT_LIST_HEAD(&hp_cpu->handlers); in create_per_cpu_handlers() 337 list_add_tail(&handler->node, &hp_cpu->handlers); in create_per_cpu_handlers() 350 list_for_each_safe(loop, tmp, &hp_cpu->handlers) { in destroy_per_cpu_handlers() 453 &hp_cpu->handlers, in init_phase2() 477 handler = list_first_entry(&hp_cpu->handlers, struct hp_handler, node); in init_phase2() 496 &hp_cpu->handlers, in init_phase3()
|
| /linux/drivers/char/ipmi/ |
| H A D | ipmi_ipmb.c | 44 struct ipmi_smi_handlers handlers; member 523 iidev->handlers.flags = IPMI_SMI_CAN_HANDLE_IPMB_DIRECT; in ipmi_ipmb_probe() 524 iidev->handlers.start_processing = ipmi_ipmb_start_processing; in ipmi_ipmb_probe() 525 iidev->handlers.shutdown = ipmi_ipmb_shutdown; in ipmi_ipmb_probe() 526 iidev->handlers.sender = ipmi_ipmb_sender; in ipmi_ipmb_probe() 527 iidev->handlers.request_events = ipmi_ipmb_request_events; in ipmi_ipmb_probe() 542 rv = ipmi_register_smi(&iidev->handlers, in ipmi_ipmb_probe()
|
| H A D | ipmi_msghandler.c | 460 const struct ipmi_smi_handlers *handlers; member 1020 if (!intf->handlers->set_need_watch) in smi_add_watch() 1035 intf->handlers->set_need_watch(intf->send_info, in smi_add_watch() 1045 if (!intf->handlers->set_need_watch) in smi_remove_watch() 1068 intf->handlers->set_need_watch(intf->send_info, in smi_remove_watch() 1329 if (!intf->handlers->get_smi_info) in ipmi_get_smi_info() 1332 rv = intf->handlers->get_smi_info(intf->send_info, data); in ipmi_get_smi_info() 1554 if (intf->handlers->set_maintenance_mode) in maintenance_mode_update() 1560 intf->handlers->set_maintenance_mode( in maintenance_mode_update() 1889 const struct ipmi_smi_handlers *handlers, in smi_send() argument [all …]
|
| /linux/drivers/iio/ |
| H A D | TODO | 11 property handlers. 14 to state struct and using property handlers and readers.
|
| /linux/drivers/acpi/ |
| H A D | prmt.c | 72 struct prm_handler_info handlers[] __counted_by(handler_count); 108 module_info_size = struct_size(tm, handlers, module_info->handler_info_count); in acpi_parse_prmt() 150 th = &tm->handlers[cur_handler]; in acpi_parse_prmt() 224 cur_handler = &cur_module->handlers[i]; in find_guid_info()
|
| /linux/Documentation/RCU/ |
| H A D | NMI-RCU.rst | 9 handlers, as well as dynamic irq handlers. This document describes 97 Since NMI handlers disable preemption, synchronize_rcu() is guaranteed 98 not to return until all ongoing NMI handlers exit. It is therefore safe
|
| /linux/Documentation/power/ |
| H A D | suspend-and-interrupts.rst | 21 interrupt handlers for shared IRQs that device drivers implementing them were 51 user of it. Thus, if the IRQ is shared, all of the interrupt handlers installed 90 already suspended. However, after suspend_device_irqs() interrupt handlers are 122 interrupts (interrupt handlers are invoked after suspend_device_irqs()) are 124 handlers are not invoked after suspend_device_irqs()). 127 to individual interrupt handlers, so sharing an IRQ between a system wakeup
|
| /linux/net/wireless/ |
| H A D | wext-core.c | 647 /* IW handlers */ 701 const struct iw_handler_def *handlers = NULL; in get_handler() 709 handlers = dev->ieee80211_ptr->wiphy->wext; in get_handler() 714 handlers = dev->wireless_handlers; in get_handler() 717 if (!handlers) in get_handler() 722 if (index < handlers->num_standard) in get_handler() 723 return handlers->standard[index]; in get_handler() 728 if (index < handlers->num_private) in get_handler() 729 return handlers->private[index]; in get_handler() 702 const struct iw_handler_def *handlers = NULL; get_handler() local
|
| /linux/Documentation/trace/ |
| H A D | events-nmi.rst | 14 NMI handlers are hogging large amounts of CPU time. The kernel 15 will warn if it sees long-running handlers::
|
| /linux/drivers/firewire/ |
| H A D | core-transaction.c | 1027 struct fw_address_handler *handler, **handlers; in handle_fcp_region_request() local 1050 handlers = buffer_on_kernel_stack; in handle_fcp_region_request() 1059 if (handlers == buffer_on_kernel_stack) in handle_fcp_region_request() 1062 buffer_on_kernel_heap = handlers; in handle_fcp_region_request() 1073 if (handlers == buffer_on_kernel_stack) { in handle_fcp_region_request() 1078 handlers = buffer_on_kernel_heap; in handle_fcp_region_request() 1082 handlers[count++] = handler; in handle_fcp_region_request() 1088 handler = handlers[i]; in handle_fcp_region_request() 1095 if (handlers != buffer_on_kernel_stack) in handle_fcp_region_request() 1096 kfree(handlers); in handle_fcp_region_request()
|
| /linux/Documentation/core-api/ |
| H A D | genericirq.rst | 32 Originally, Russell King identified different types of handlers to build 51 This split implementation of high-level IRQ handlers allows us to 77 demultiplexing handlers which can be found in embedded platforms on 104 2. High-level IRQ flow handlers 154 High-level IRQ flow handlers 173 The interrupt flow handlers (either pre-defined or architecture 359 - ``irq_eoi`` - Optional, required for EOI flow handlers
|
| /linux/drivers/media/platform/amphion/ |
| H A D | vpu_msgs.c | 199 static struct vpu_msg_handler handlers[] = { variable 232 for (i = 0; i < ARRAY_SIZE(handlers); i++) { in vpu_session_handle_msg() 233 if (handlers[i].id == msg_id) { in vpu_session_handle_msg() 234 handler = &handlers[i]; in vpu_session_handle_msg()
|
| /linux/drivers/soc/qcom/ |
| H A D | qmi_interface.c | 398 if (!qmi->handlers) in qmi_invoke_handler() 401 for (handler = qmi->handlers; handler->fn; handler++) { in qmi_invoke_handler() 623 const struct qmi_msg_handler *handlers) in qmi_handle_init() argument 638 qmi->handlers = handlers; in qmi_handle_init()
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | v4l2grab-example.rst | 11 libv4l handlers. The advantage is that this grabber can potentially work
|