| /linux/drivers/iommu/ |
| H A D | io-pgfault.c | 46 list_for_each_entry_safe(iopf, next, &group->faults, list) { in __iopf_free_group() 99 INIT_LIST_HEAD(&group->faults); in iopf_group_alloc() 101 list_add(&group->last_fault.list, &group->faults); in iopf_group_alloc() 108 list_move(&iopf->list, &group->faults); in iopf_group_alloc() 110 list_add(&group->pending_node, &iopf_param->faults); in iopf_group_alloc() 113 group->fault_count = list_count_nodes(&group->faults); in iopf_group_alloc() 410 INIT_LIST_HEAD(&fault_param->faults); in iopf_queue_add_device() 471 list_for_each_entry_safe(group, temp, &fault_param->faults, pending_node) { in iopf_queue_remove_device()
|
| /linux/Documentation/gpu/rfc/ |
| H A D | i915_vm_bind.rst | 96 newer VM_BIND mode, the VM_BIND mode with GPU page faults and possible future 98 The older execbuf mode and the newer VM_BIND mode without page faults manages 99 residency of backing storage using dma_fence. The VM_BIND mode with page faults 108 In future, when GPU page faults are supported, we can potentially use a 124 When GPU page faults are supported, the execbuf path do not take any of these 180 Where GPU page faults are not available, kernel driver upon buffer invalidation 210 GPU page faults 212 GPU page faults when supported (in future), will only be supported in the 214 binding will require using dma-fence to ensure residency, the GPU page faults 240 faults enabled.
|
| H A D | gpusvm.rst | 94 * Concurrent GPU faults 95 * CPU faults are concurrent so makes sense to have concurrent GPU 96 faults.
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | userfaultfd.rst | 10 memory page faults, something otherwise only the kernel code could do. 19 regions of virtual memory with it. Then, any page faults which occur within the 26 1) ``read/POLLIN`` protocol to notify a userland thread of the faults 58 handle kernel page faults have been a useful tool for exploiting the kernel). 63 - Any user can always create a userfaultfd which traps userspace page faults 67 - In order to also trap kernel page faults for the address space, either the 80 to /dev/userfaultfd can always create userfaultfds that trap kernel page faults; 102 other than page faults are supported. These events are described in more 127 bitmask will specify to the kernel which kind of faults to track for 132 hugetlbfs), or all types of intercepted faults. [all …]
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-class-led-flash | 54 Space separated list of flash faults that may have occurred. 55 Flash faults are re-read after strobing the flash. Possible 56 flash faults:
|
| H A D | sysfs-bus-iio-thermocouple | 16 Open-circuit fault. The detection of open-circuit faults,
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_submitqueue.c | 285 size_t size = min_t(size_t, args->len, sizeof(queue->faults)); in msm_submitqueue_query_faults() 290 args->len = sizeof(queue->faults); in msm_submitqueue_query_faults() 297 ret = copy_to_user(u64_to_user_ptr(args->data), &queue->faults, size); in msm_submitqueue_query_faults()
|
| /linux/Documentation/virt/kvm/devices/ |
| H A D | s390_flic.rst | 18 - enable/disable for the guest transparent async page faults 58 Enables async page faults for the guest. So in case of a major page fault 64 Disables async page faults for the guest and waits until already pending 65 async page faults are done. This is necessary to trigger a completion interrupt
|
| /linux/tools/testing/selftests/mm/ |
| H A D | hmm-tests.c | 45 uint64_t faults; 204 buffer->faults = cmd.faults; in hmm_dmirror_cmd() 346 ASSERT_EQ(buffer->faults, 1); in TEST_F() 456 ASSERT_EQ(buffer->faults, 1); in TEST_F() 500 ASSERT_EQ(buffer->faults, 1); in TEST_F() 522 ASSERT_EQ(buffer->faults, 1); in TEST_F() 626 ASSERT_EQ(buffer->faults, 1); in TEST_F() 708 ASSERT_EQ(buffer->faults, 1); in TEST_F() 764 ASSERT_EQ(buffer->faults, in TEST_F() 44 uint64_t faults; global() member [all...] |
| /linux/kernel/sched/ |
| H A D | fair.c | 1665 unsigned long faults[]; member 1832 return ng->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults() 1833 ng->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults() 1838 return group->faults[task_faults_idx(NUMA_CPU, nid, 0)] + in group_faults_cpu() 1839 group->faults[task_faults_idx(NUMA_CPU, nid, 1)]; in group_faults_cpu() 1844 unsigned long faults = 0; in group_faults_priv() local 1848 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 1)]; in group_faults_priv() 1851 return faults; in group_faults_priv() 1856 unsigned long faults = 0; in group_faults_shared() local 1860 faults += ng->faults[task_faults_idx(NUMA_MEM, node, 0)]; in group_faults_shared() [all …]
|
| /linux/lib/ |
| H A D | test_hmm_uapi.h | 28 __u64 faults; member
|
| /linux/kernel/trace/rv/monitors/pagefault/ |
| H A D | Kconfig | 13 Monitor that real-time tasks do not raise page faults, causing
|
| /linux/Documentation/i2c/ |
| H A D | fault-codes.rst | 11 Not all fault reports imply errors; "page faults" should be a familiar 13 faults. There may be fancier recovery schemes that are appropriate in 86 about probe faults other than ENXIO and ENODEV.)
|
| /linux/tools/perf/tests/shell/ |
| H A D | stat+std_output.sh | 14 event_name=(cpu-clock task-clock context-switches cpu-migrations page-faults stalled-cycles-fronten…
|
| H A D | record_bpf_filter.sh | 123 -e page-faults --filter 'ip < 0xffffffff00000000' \
|
| /linux/Documentation/arch/arm64/ |
| H A D | memory-tagging-extension.rst | 75 thread, asynchronously following one or multiple tag check faults, 87 - ``PR_MTE_TCF_NONE`` - *Ignore* tag check faults 92 If no modes are specified, tag check faults are ignored. If a single 172 - No tag checking modes are selected (tag check faults ignored) 321 * tag check faults (based on per-CPU preference) and allow all
|
| /linux/Documentation/devicetree/bindings/iommu/ |
| H A D | ti,omap-iommu.txt | 22 back a bus error response on MMU faults.
|
| /linux/drivers/ras/ |
| H A D | Kconfig | 14 faults.
|
| /linux/Documentation/trace/rv/ |
| H A D | monitor_rtapp.rst | 15 - Page faults: A real-time thread may access memory that does not have a 32 The `pagefault` monitor reports real-time tasks raising page faults. Its
|
| /linux/Documentation/arch/x86/ |
| H A D | pti.rst | 183 * Double faults: overflowing the kernel stack because of page 184 faults upon page faults. Caused by touching non-pti-mapped
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_vm.c | 617 guard(spinlock)(&vm->faults.lock); in xe_vm_add_fault_entry_pf() 623 if (vm->faults.len >= MAX_FAULTS_SAVED_PER_VM) { in xe_vm_add_fault_entry_pf() 641 list_add_tail(&e->list, &vm->faults.list); in xe_vm_add_fault_entry_pf() 642 vm->faults.len++; in xe_vm_add_fault_entry_pf() 649 guard(spinlock)(&vm->faults.lock); in xe_vm_clear_fault_entries() 650 list_for_each_entry_safe(e, tmp, &vm->faults.list, list) { in xe_vm_clear_fault_entries() 654 vm->faults.len = 0; in xe_vm_clear_fault_entries() 1630 INIT_LIST_HEAD(&vm->faults.list); in xe_vm_create() 1631 spin_lock_init(&vm->faults.lock); in xe_vm_create() 4094 spin_lock(&vm->faults.lock); in fill_faults() [all …]
|
| H A D | xe_vm_types.h | 374 } faults; member
|
| /linux/drivers/soc/samsung/ |
| H A D | Kconfig | 73 faults with better precision.
|
| /linux/Documentation/leds/ |
| H A D | leds-class-flash.rst | 64 bitmask of flash faults that the LED flash class
|
| /linux/arch/arm/nwfpe/ |
| H A D | entry.S | 104 @ ??? For some reason, faults can happen at .Lx2 even with a
|