Home
last modified time | relevance | path

Searched refs:event_list (Results 1 – 25 of 45) sorted by relevance

12

/linux/drivers/usb/usbip/
H A Dusbip_event.c20 static LIST_HEAD(event_list);
47 if (!list_empty(&event_list)) { in get_event()
48 ue = list_first_entry(&event_list, struct usbip_event, node); in get_event()
156 list_for_each_entry_reverse(ue, &event_list, node) { in usbip_event_add()
167 list_add_tail(&ue->node, &event_list); in usbip_event_add()
/linux/tools/perf/scripts/python/
H A Dnetdev-times.py122 event_list = hunk['event_list']
123 for i in range(len(event_list)):
124 event = event_list[i]
129 if i == len(event_list) - 1:
348 event_list = 0
354 event_list = net_rx_dic[cpu]['event_list']
356 if irq_list == [] or event_list == 0:
359 'irq_list':irq_list, 'event_list':event_list}
367 event_list = net_rx_dic[cpu]['event_list']
371 event_list.append(rec_data)
[all …]
/linux/drivers/dma/idxd/
H A Dperfmon.c69 idxd_pmu->event_list[n] = leader; in perfmon_collect_events()
70 idxd_pmu->event_list[n]->hw.idx = n; in perfmon_collect_events()
85 idxd_pmu->event_list[n] = event; in perfmon_collect_events()
86 idxd_pmu->event_list[n]->hw.idx = n; in perfmon_collect_events()
154 event = fake_pmu->event_list[i]; in perfmon_validate_group()
254 event = idxd->idxd_pmu->event_list[i]; in perfmon_counter_overflow()
352 if (event != idxd->idxd_pmu->event_list[i]) in perfmon_pmu_event_stop()
356 idxd->idxd_pmu->event_list[i - 1] = idxd->idxd_pmu->event_list[i]; in perfmon_pmu_event_stop()
/linux/drivers/gpu/drm/
H A Ddrm_file.c156 INIT_LIST_HEAD(&file->event_list); in drm_file_alloc()
215 list_for_each_entry_safe(e, et, &file_priv->event_list, link) { in drm_events_release()
271 WARN_ON(!list_empty(&file->event_list)); in drm_file_free()
555 if (!list_empty(&file_priv->event_list)) { in drm_read()
556 e = list_first_entry(&file_priv->event_list, in drm_read()
574 !list_empty(&file_priv->event_list)); in drm_read()
586 list_add(&e->link, &file_priv->event_list); in drm_read()
631 if (!list_empty(&file_priv->event_list)) in drm_poll()
772 &e->file_priv->event_list); in drm_send_event_helper()
/linux/drivers/infiniband/hw/mlx5/
H A Ddevx.c81 struct list_head event_list; /* headed in ev_file->event_list or in member
100 struct list_head event_list; member
1740 struct list_head event_list; member
1754 INIT_LIST_HEAD(&ev_queue->event_list); in devx_init_event_queue()
1798 INIT_LIST_HEAD(&ev_file->event_list); in UVERBS_HANDLER()
1822 list_add_tail(&async_data->list, &ev_queue->event_list); in devx_query_callback()
2167 list_add_tail(&event_sub->event_list, &sub_list); in UVERBS_HANDLER()
2192 list_for_each_entry_safe(event_sub, tmp_sub, &sub_list, event_list) { in UVERBS_HANDLER()
2196 list_del_init(&event_sub->event_list); in UVERBS_HANDLER()
2225 list_for_each_entry_safe(event_sub, tmp_sub, &sub_list, event_list) { in UVERBS_HANDLER()
[all …]
/linux/drivers/gpu/drm/exynos/
H A Dexynos_drm_g2d.c229 struct list_head event_list; member
376 list_add_tail(&node->event->base.link, &file_priv->event_list); in g2d_add_cmdlist_to_inuse()
919 if (list_empty(&runqueue_node->event_list)) in g2d_finish_event()
922 e = list_first_entry(&runqueue_node->event_list, in g2d_finish_event()
1307 struct list_head *event_list; in exynos_g2d_exec_ioctl() local
1314 event_list = &runqueue_node->event_list; in exynos_g2d_exec_ioctl()
1316 INIT_LIST_HEAD(event_list); in exynos_g2d_exec_ioctl()
1321 list_splice_init(&file_priv->event_list, event_list); in exynos_g2d_exec_ioctl()
1353 INIT_LIST_HEAD(&file_priv->event_list); in g2d_open()
/linux/drivers/infiniband/core/
H A Duverbs_main.c172 list_for_each_entry_safe(evt, tmp, &uobj->event_list, obj_list) { in ib_uverbs_release_uevent()
238 while (list_empty(&ev_queue->event_list)) { in ib_uverbs_event_read()
249 (!list_empty(&ev_queue->event_list) || in ib_uverbs_event_read()
256 event = list_entry(ev_queue->event_list.next, struct ib_uverbs_event, list); in ib_uverbs_event_read()
262 list_del(ev_queue->event_list.next); in ib_uverbs_event_read()
312 if (!list_empty(&ev_queue->event_list)) in ib_uverbs_event_poll()
396 list_add_tail(&entry->list, &ev_queue->event_list); in ib_uverbs_comp_handler()
431 list_add_tail(&entry->list, &async_file->ev_queue.event_list); in ib_uverbs_async_handler()
445 &eobj->event_list, &eobj->events_reported); in uverbs_uobj_event()
484 INIT_LIST_HEAD(&ev_queue->event_list); in ib_uverbs_init_event_queue()
H A Ducma.c82 struct list_head event_list; member
326 list_add_tail(&uevent->list, &ctx->file->event_list); in ucma_connect_event_handler()
361 list_add_tail(&uevent->list, &ctx->file->event_list); in ucma_event_handler()
393 while (list_empty(&file->event_list)) { in ucma_get_event()
400 !list_empty(&file->event_list))) in ucma_get_event()
406 uevent = list_first_entry(&file->event_list, struct ucma_event, list); in ucma_get_event()
517 list_for_each_entry_safe(uevent, tmp, &mc->ctx->file->event_list, list) { in ucma_cleanup_mc_events()
536 list_for_each_entry_safe(uevent, tmp, &ctx->file->event_list, list) { in ucma_cleanup_ctx_events()
1686 LIST_HEAD(event_list); in ucma_migrate_id()
1727 list_for_each_entry_safe(uevent, tmp, &cur_file->event_list, list) in ucma_migrate_id()
[all …]
H A Duverbs_std_types_wq.c69 INIT_LIST_HEAD(&obj->uevent.event_list); in UVERBS_HANDLER()
H A Duverbs_std_types.c148 list_for_each_entry_safe(entry, tmp, &event_queue->event_list, list) { in ib_uverbs_free_event_queue()
H A Duverbs_std_types_srq.c103 INIT_LIST_HEAD(&obj->uevent.event_list); in UVERBS_HANDLER()
H A Duverbs_std_types_cq.c119 INIT_LIST_HEAD(&obj->uevent.event_list); in UVERBS_HANDLER()
/linux/include/media/i2c/
H A Dsaa6588.h22 poll_table *event_list; member
/linux/arch/x86/events/intel/
H A Duncore.c401 box->event_list[n] = leader; in uncore_collect_events()
416 box->event_list[n] = event; in uncore_collect_events()
464 c = uncore_get_event_constraint(box, box->event_list[i]); in uncore_assign_events()
472 hwc = &box->event_list[i]->hw; in uncore_assign_events()
498 uncore_put_event_constraint(box, box->event_list[i]); in uncore_assign_events()
611 event = box->event_list[i]; in uncore_pmu_event_add()
629 event = box->event_list[i]; in uncore_pmu_event_add()
664 if (event == box->event_list[i]) { in uncore_pmu_event_del()
668 box->event_list[i - 1] = box->event_list[i]; in uncore_pmu_event_del()
/linux/drivers/net/wireless/ath/ath10k/
H A Dqmi.h97 struct list_head event_list; member
H A Dqmi.c888 list_add_tail(&event->list, &qmi->event_list); in ath10k_qmi_driver_event_post()
1034 while (!list_empty(&qmi->event_list)) { in ath10k_qmi_driver_event_work()
1035 event = list_first_entry(&qmi->event_list, in ath10k_qmi_driver_event_work()
1104 INIT_LIST_HEAD(&qmi->event_list); in ath10k_qmi_init()
/linux/drivers/scsi/
H A Dvirtio_scsi.c81 struct virtio_scsi_event_node event_list[VIRTIO_SCSI_EVENT_LEN]; member
265 vscsi->event_list[i].vscsi = vscsi; in virtscsi_kick_event_all()
266 vscsi->event_list[i].event = &vscsi->events[i]; in virtscsi_kick_event_all()
267 virtscsi_kick_event(vscsi, &vscsi->event_list[i]); in virtscsi_kick_event_all()
283 cancel_work_sync(&vscsi->event_list[i].work); in virtscsi_cancel_event_work()
H A Dscsi_lib.c2687 LIST_HEAD(event_list); in scsi_evt_thread()
2701 list_splice_init(&sdev->event_list, &event_list); in scsi_evt_thread()
2704 if (list_empty(&event_list)) in scsi_evt_thread()
2707 list_for_each_safe(this, tmp, &event_list) { in scsi_evt_thread()
2738 list_add_tail(&evt->node, &sdev->event_list); in sdev_evt_send()
/linux/drivers/firewire/
H A Dcore-cdev.c60 struct list_head event_list; member
303 INIT_LIST_HEAD(&client->event_list); in fw_device_op_open()
328 list_add_tail(&event->link, &client->event_list); in queue_event()
342 !list_empty(&client->event_list) || in dequeue_event()
347 if (list_empty(&client->event_list) && in dequeue_event()
352 event = list_first_entry(&client->event_list, struct event, link); in dequeue_event()
1879 list_for_each_entry_safe(event, next_event, &client->event_list, link) in fw_device_op_release()
1896 if (!list_empty(&client->event_list)) in fw_device_op_poll()
/linux/net/vmw_vsock/
H A Dvirtio_transport.c78 struct virtio_vsock_event event_list[8]; member
404 for (i = 0; i < ARRAY_SIZE(vsock->event_list); i++) { in virtio_vsock_event_fill()
405 struct virtio_vsock_event *event = &vsock->event_list[i]; in virtio_vsock_event_fill()
/linux/arch/x86/events/
H A Dcore.c1041 c = static_call(x86_pmu_get_event_constraints)(cpuc, i, cpuc->event_list[i]); in x86_schedule_events()
1055 hwc = &cpuc->event_list[i]->hw; in x86_schedule_events()
1126 e = cpuc->event_list[i]; in x86_schedule_events()
1173 cpuc->event_list[n] = event; in collect_event()
1339 event = cpuc->event_list[i]; in x86_pmu_enable()
1367 event = cpuc->event_list[i]; in x86_pmu_enable()
1667 if (event == cpuc->event_list[i]) in x86_pmu_del()
1682 cpuc->event_list[i-1] = cpuc->event_list[i]; in x86_pmu_del()
/linux/drivers/iommu/intel/
H A Dperfmon.c410 iommu_pmu->event_list[idx] = event; in iommu_pmu_assign_event()
466 iommu_pmu->event_list[idx] = NULL; in iommu_pmu_del()
505 event = iommu_pmu->event_list[i]; in iommu_pmu_counter_overflow()
/linux/kernel/events/
H A Dinternal.h36 struct list_head event_list; member
/linux/net/sctp/
H A Dulpqueue.c742 struct sk_buff_head *event_list; in sctp_ulpq_retrieve_ordered() local
751 event_list = (struct sk_buff_head *) sctp_event2skb(event)->prev; in sctp_ulpq_retrieve_ordered()
776 __skb_queue_tail(event_list, pos); in sctp_ulpq_retrieve_ordered()
/linux/net/wireless/
H A Dsme.c1071 list_add_tail(&ev->list, &wdev->event_list); in cfg80211_connect_done()
1276 list_add_tail(&ev->list, &wdev->event_list); in cfg80211_roamed()
1336 list_add_tail(&ev->list, &wdev->event_list); in cfg80211_port_authorized()
1427 list_add_tail(&ev->list, &wdev->event_list); in cfg80211_disconnected()

12