| /linux/tools/perf/pmu-events/arch/common/common/ |
| H A D | software.json | 18 "EventName": "faults", 19 "BriefDescription": "Number of page faults [This event is an alias of page-faults]", 24 "EventName": "page-faults", 25 "BriefDescription": "Number of page faults [This event is an alias of faults]", 54 "EventName": "minor-faults", 55 "BriefDescription": "Number of minor page faults. Minor faults don't require I/O to handle", 60 "EventName": "major-faults", 61 "BriefDescription": "Number of major page faults. Major faults require I/O to handle", 66 "EventName": "alignment-faults", 67 "BriefDescription": "Number of kernel handled memory alignment faults", [all …]
|
| H A D | metrics.json | 30 "BriefDescription": "Page faults per CPU second", 31 …"MetricExpr": "(software@page\\-faults\\,name\\=page\\-faults@ * 1e9) / (software@cpu\\-clock\\,na… 34 "ScaleUnit": "1faults/sec",
|
| /linux/drivers/iommu/ |
| H A D | io-pgfault.c | 3 * Handle device page faults 46 list_for_each_entry_safe(iopf, next, &group->faults, list) { in __iopf_free_group() 99 INIT_LIST_HEAD(&group->faults); in iopf_group_alloc() 101 list_add(&group->last_fault.list, &group->faults); in iopf_group_alloc() 103 /* See if we have partial faults for this group */ in iopf_group_alloc() 108 list_move(&iopf->list, &group->faults); in iopf_group_alloc() 110 list_add(&group->pending_node, &iopf_param->faults); in iopf_group_alloc() 113 group->fault_count = list_count_nodes(&group->faults); in iopf_group_alloc() 134 * managed PASID table. Therefore page faults for in find_fault_handler() 181 * them before reporting faults [all...] |
| /linux/Documentation/admin-guide/mm/ |
| H A D | userfaultfd.rst | 10 memory page faults, something otherwise only the kernel code could do. 19 regions of virtual memory with it. Then, any page faults which occur within the 26 1) ``read/POLLIN`` protocol to notify a userland thread of the faults 58 handle kernel page faults have been a useful tool for exploiting the kernel). 63 - Any user can always create a userfaultfd which traps userspace page faults 67 - In order to also trap kernel page faults for the address space, either the 80 to /dev/userfaultfd can always create userfaultfds that trap kernel page faults; 102 other than page faults are supported. These events are described in more 127 bitmask will specify to the kernel which kind of faults to track for 132 hugetlbfs), or all types of intercepted faults. [all …]
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | vas-fault.c | 24 * 8MB FIFO can be used if expects more faults for each VAS 57 * It can raise a single interrupt for multiple faults. Expects OS to 58 * process all valid faults and return credit for each fault on user 78 * VAS can interrupt with multiple page faults. So process all in vas_fault_thread_fn() 92 * fifo_in_progress is set. Means these new faults will be in vas_fault_thread_fn() 153 * NX sees faults only with user space windows. in vas_fault_thread_fn() 176 * NX can generate an interrupt for multiple faults. So the in vas_fault_handler() 178 * entry. In case if NX sees continuous faults, it is possible in vas_fault_handler() 197 * FIFO upon page faults.
|
| /linux/Documentation/gpu/rfc/ |
| H A D | i915_vm_bind.rst | 96 newer VM_BIND mode, the VM_BIND mode with GPU page faults and possible future 98 The older execbuf mode and the newer VM_BIND mode without page faults manages 99 residency of backing storage using dma_fence. The VM_BIND mode with page faults 108 In future, when GPU page faults are supported, we can potentially use a 124 When GPU page faults are supported, the execbuf path do not take any of these 180 Where GPU page faults are not available, kernel driver upon buffer invalidation 210 GPU page faults 212 GPU page faults when supported (in future), will only be supported in the 214 binding will require using dma-fence to ensure residency, the GPU page faults 240 faults enabled.
|
| H A D | gpusvm.rst | 94 * Concurrent GPU faults 95 * CPU faults are concurrent so makes sense to have concurrent GPU 96 faults.
|
| /linux/tools/testing/selftests/powerpc/mm/ |
| H A D | stress_code_patching.sh | 20 echo "Testing for spurious faults when mapping kernel memory..." 44 echo "FAILED: Mapping kernel memory causes spurious faults" 1>&2 47 echo "OK: Mapping kernel memory does not cause spurious faults"
|
| /linux/drivers/hwmon/ |
| H A D | ltc4260.c | 98 if (fault) /* Clear reported faults in chip register */ in ltc4260_bool_show() 110 * UV/OV faults are associated with the input voltage, and the POWER BAD and 111 * FET SHORT faults are associated with the output voltage. 156 /* Clear faults */ in ltc4260_probe()
|
| H A D | ltc4222.c | 112 if (fault) /* Clear reported faults in chip register */ in ltc4222_bool_show() 126 * UV/OV faults are associated with the input voltage, and power bad and fet 127 * faults are associated with the output voltage. 192 /* Clear faults */ in ltc4222_probe()
|
| /linux/include/uapi/linux/ |
| H A D | virtio_balloon.h | 66 #define VIRTIO_BALLOON_S_MAJFLT 2 /* Number of major faults */ 67 #define VIRTIO_BALLOON_S_MINFLT 3 /* Number of minor faults */ 85 VIRTIO_BALLOON_S_NAMES_prefix "major-faults", \ 86 VIRTIO_BALLOON_S_NAMES_prefix "minor-faults", \
|
| H A D | userfaultfd.h | 206 * UFFD_FEATURE_MINOR_HUGETLBFS indicates that minor faults 214 * faults would be provided and the offset within the page would not be 318 * NOTE: Write protecting a region (WP=1) is unrelated to page faults, 382 * Create a userfaultfd that can handle page faults only in user mode.
|
| /linux/lib/ |
| H A D | test_hmm_uapi.h | 21 * @faults: (out) number of device page faults seen 28 __u64 faults; member
|
| /linux/arch/x86/mm/ |
| H A D | fault.c | 416 * Note we only handle faults in kernel here. 561 * ones are faults accessing the GDT, or LDT. Perhaps in show_fault_oops() 618 * kernel addresses are always protection faults. in sanitize_error_code() 691 /* Only not-present faults should be handled by KFENCE. */ in page_fault_oops() 892 * 3. T1 : faults... in bad_area_access_error() 967 * Spurious faults may only occur if the TLB contains an entry with 969 * and reserved bit (R = 1) faults are never spurious. 991 * spurious faults. in spurious_kernel_fault() 996 * faults. in spurious_kernel_fault() 1076 * faults just to hit a X86_PF_PK as soon as we fill in a in access_error() [all …]
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-class-led-flash | 54 Space separated list of flash faults that may have occurred. 55 Flash faults are re-read after strobing the flash. Possible 56 flash faults:
|
| /linux/Documentation/virt/kvm/devices/ |
| H A D | s390_flic.rst | 18 - enable/disable for the guest transparent async page faults 58 Enables async page faults for the guest. So in case of a major page fault 64 Disables async page faults for the guest and waits until already pending 65 async page faults are done. This is necessary to trigger a completion interrupt
|
| /linux/tools/perf/pmu-events/arch/x86/amdzen2/ |
| H A D | floating-point.json | 119 "BriefDescription": "Floating Point Dispatch Faults. YMM spill fault.", 125 "BriefDescription": "Floating Point Dispatch Faults. YMM fill fault.", 131 "BriefDescription": "Floating Point Dispatch Faults. XMM fill fault.", 137 "BriefDescription": "Floating Point Dispatch Faults. x87 fill fault.",
|
| /linux/Documentation/arch/arm64/ |
| H A D | memory-tagging-extension.rst | 58 Tag Check Faults 75 thread, asynchronously following one or multiple tag check faults, 87 - ``PR_MTE_TCF_NONE`` - *Ignore* tag check faults 92 If no modes are specified, tag check faults are ignored. If a single 172 - No tag checking modes are selected (tag check faults ignored) 321 * tag check faults (based on per-CPU preference) and allow all
|
| /linux/tools/perf/pmu-events/arch/x86/amdzen3/ |
| H A D | floating-point.json | 118 "BriefDescription": "Floating Point Dispatch Faults. YMM spill fault.", 124 "BriefDescription": "Floating Point Dispatch Faults. YMM fill fault.", 130 "BriefDescription": "Floating Point Dispatch Faults. XMM fill fault.", 136 "BriefDescription": "Floating Point Dispatch Faults. x87 fill fault.",
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_svm.c | 344 * client, with replayable faults enabled). in nouveau_svmm_init() 750 /* Sort parsed faults by instance pointer to prevent unnecessary in nouveau_svm_fault() 752 * type to reduce the amount of work when handling the faults. in nouveau_svm_fault() 771 /* Process list of faults. */ in nouveau_svm_fault() 782 /* Cancel any faults from non-SVM channels. */ in nouveau_svm_fault() 789 /* We try and group handling of faults within a small in nouveau_svm_fault() 800 * permissions based on pending faults. in nouveau_svm_fault() 844 * same SVMM as faults are ordered by access type such in nouveau_svm_fault() 847 * ie. WRITE faults appear first, thus any handling of in nouveau_svm_fault() 848 * pending READ faults wil in nouveau_svm_fault() [all...] |
| /linux/Documentation/i2c/ |
| H A D | fault-codes.rst | 11 Not all fault reports imply errors; "page faults" should be a familiar 13 faults. There may be fancier recovery schemes that are appropriate in 86 about probe faults other than ENXIO and ENODEV.)
|
| /linux/kernel/trace/rv/monitors/pagefault/ |
| H A D | pagefault.c | 68 .description = "Monitor that RT tasks do not raise page faults", 88 MODULE_DESCRIPTION("pagefault: Monitor that RT tasks do not raise page faults");
|
| /linux/arch/x86/include/asm/ |
| H A D | kfence.h | 75 * We need to avoid IPIs, as we may get KFENCE allocations or faults in kfence_protect_page() 78 * lazy fault handling takes care of faults after the page is PRESENT. in kfence_protect_page()
|
| /linux/tools/testing/selftests/mm/ |
| H A D | uffd-common.c | 22 * to avoid non alignment faults on non-x86 archs. 406 /* Write protection page faults */ in wp_range() 469 /* Write protect page faults */ in uffd_handle_page_fault() 477 * Minor page faults in uffd_handle_page_fault() 498 * Missing page faults. in uffd_handle_page_fault() 501 * faults. It's guaranteed because the only threads that in uffd_handle_page_fault() 502 * will trigger uffd faults are the locking threads, and in uffd_handle_page_fault()
|
| /linux/drivers/hv/ |
| H A D | mshv_regions.c | 386 * mshv_region_hmm_fault_and_lock - Handle HMM faults and lock the memory region 393 * 3. Handles HMM faults for the specified range. 426 * mshv_region_range_fault - Handle memory range faults for a given region. 431 * This function resolves memory faults for a specified range of pages 484 /* Map more pages than requested to reduce the number of faults. */ in mshv_region_handle_gfn_fault()
|