/linux/tools/perf/pmu-events/arch/x86/goldmontplus/ |
H A D | other.json | 20 "BriefDescription": "Cycles hardware interrupts are masked", 23 "EventName": "HW_INTERRUPTS.MASKED", 24 …"PublicDescription": "Counts the number of core cycles during which interrupts are masked (disable… 29 "BriefDescription": "Cycles pending interrupts are masked", 33 …core cycles during which there are pending interrupts, but interrupts are masked (EFLAGS.IF = 0).",
|
/linux/tools/perf/pmu-events/arch/x86/goldmont/ |
H A D | other.json | 20 "BriefDescription": "Cycles hardware interrupts are masked", 23 "EventName": "HW_INTERRUPTS.MASKED", 24 …"PublicDescription": "Counts the number of core cycles during which interrupts are masked (disable… 29 "BriefDescription": "Cycles pending interrupts are masked", 33 …core cycles during which there are pending interrupts, but interrupts are masked (EFLAGS.IF = 0).",
|
/linux/arch/x86/math-emu/ |
H A D | errors.c | 322 /* Set summary bits iff exception isn't masked */ in FPU_exception() 387 /* Masked response */ in real_1op_NaN() 398 /* The masked response */ in real_1op_NaN() 435 /* Masked response */ in real_2op_NaN() 503 /* The masked response */ in arith_invalid() 518 /* The masked response */ in FPU_divide_by_zero() 535 partial_status |= flags; /* The masked response */ in set_precision_flag() 547 partial_status |= (SW_Precision | SW_C1); /* The masked response */ in set_precision_flag_up() 555 if (control_word & CW_Precision) { /* The masked response */ in set_precision_flag_down() 564 if (control_word & CW_Denormal) { /* The masked response */ in denormal_operand() [all …]
|
/linux/drivers/net/wireless/ath/ath9k/ |
H A D | ar9002_mac.c | 32 static bool ar9002_hw_get_isr(struct ath_hw *ah, enum ath9k_int *masked, in ar9002_hw_get_isr() argument 53 *masked = 0; in ar9002_hw_get_isr() 58 *masked = 0; in ar9002_hw_get_isr() 91 *masked = 0; in ar9002_hw_get_isr() 95 *masked = isr & ATH9K_INT_COMMON; in ar9002_hw_get_isr() 99 *masked |= ATH9K_INT_RX; in ar9002_hw_get_isr() 106 *masked |= ATH9K_INT_TX; in ar9002_hw_get_isr() 134 *masked |= mask2; in ar9002_hw_get_isr() 153 *masked |= ATH9K_INT_GENTIMER; in ar9002_hw_get_isr() 157 *masked |= ATH9K_INT_TIM_TIMER; in ar9002_hw_get_isr() [all …]
|
H A D | ar9003_mac.c | 183 static bool ar9003_hw_get_isr(struct ath_hw *ah, enum ath9k_int *masked, in ar9003_hw_get_isr() argument 207 *masked = 0; in ar9003_hw_get_isr() 244 *masked = 0; in ar9003_hw_get_isr() 248 *masked = isr & ATH9K_INT_COMMON; in ar9003_hw_get_isr() 252 *masked |= ATH9K_INT_RXLP; in ar9003_hw_get_isr() 256 *masked |= ATH9K_INT_TX; in ar9003_hw_get_isr() 259 *masked |= ATH9K_INT_RXLP; in ar9003_hw_get_isr() 262 *masked |= ATH9K_INT_RXHP; in ar9003_hw_get_isr() 265 *masked |= ATH9K_INT_TX; in ar9003_hw_get_isr() 294 *masked |= ATH9K_INT_GENTIMER; in ar9003_hw_get_isr() [all …]
|
/linux/drivers/vfio/platform/ |
H A D | vfio_platform_irq.c | 24 if (!irq_ctx->masked) { in vfio_platform_mask() 26 irq_ctx->masked = true; in vfio_platform_mask() 84 if (irq_ctx->masked) { in vfio_platform_unmask() 86 irq_ctx->masked = false; in vfio_platform_unmask() 157 if (!irq_ctx->masked) { in vfio_automasked_irq_handler() 162 irq_ctx->masked = true; in vfio_automasked_irq_handler() 204 * irq->masked effectively provides nested disables within the overall in vfio_set_trigger() 208 * irq->masked is initially false. in vfio_set_trigger() 320 vdev->irqs[i].masked = false; in vfio_platform_irq_init()
|
/linux/drivers/vfio/pci/ |
H A D | vfio_pci_intrs.c | 31 bool masked; member 101 /* Returns true if the INTx vfio_pci_irq_ctx.masked value is changed. */ 129 if (!ctx->masked) { in __vfio_pci_intx_mask() 139 ctx->masked = true; in __vfio_pci_intx_mask() 185 if (ctx->masked && !vdev->virq_disabled) { in vfio_pci_intx_unmask_handler() 197 ctx->masked = (ret > 0); in vfio_pci_intx_unmask_handler() 234 ctx->masked = true; in vfio_intx_handler() 236 } else if (!ctx->masked && /* may be shared */ in vfio_intx_handler() 238 ctx->masked = true; in vfio_intx_handler() 280 * Fill the initial masked state based on virq_disabled. After in vfio_intx_enable() [all …]
|
/linux/tools/testing/selftests/net/ |
H A D | fib_rule_tests.sh | 275 "$getnomatch" "sport and dport masked redirect to table" \ 276 "sport and dport masked no redirect to table" 321 "$getnomatch" "dscp masked redirect to table" \ 322 "dscp masked no redirect to table" 328 "$getnomatch" "iif dscp masked redirect to table" \ 329 "iif dscp masked no redirect to table" 352 "$getnomatch" "flowlabel masked redirect to table" \ 353 "flowlabel masked no redirect to table" 359 "$getnomatch" "iif flowlabel masked redirect to table" \ 360 "iif flowlabel masked no redirect to table" [all …]
|
/linux/kernel/irq/ |
H A D | migration.c | 106 bool masked; in __irq_move_irq() local 119 * Be careful vs. already masked interrupts. If this is a in __irq_move_irq() 123 masked = irqd_irq_masked(idata); in __irq_move_irq() 124 if (!masked) in __irq_move_irq() 127 if (!masked) in __irq_move_irq()
|
/linux/arch/powerpc/kvm/ |
H A D | book3s_xive.c | 317 * XXX We could check if the interrupt is masked here and in xive_vm_h_xirr() 320 * if (masked) { in xive_vm_h_xirr() 322 * if (masked) { in xive_vm_h_xirr() 395 /* For each priority that is now masked */ in xive_vm_scan_for_rerouted_irqs() 572 if (state->guest_priority == MASKED) { in xive_vm_h_eoi() 574 if (state->guest_priority != MASKED) { in xive_vm_h_eoi() 915 * interrupt, thus leaving it effectively masked after in kvmppc_xive_attach_escalation() 1094 * Take the lock, set masked, try again if racing in xive_lock_and_mask() 1100 state->guest_priority = MASKED; in xive_lock_and_mask() 1109 if (old_prio == MASKED) in xive_lock_and_mask() [all …]
|
H A D | book3s_xics.c | 154 if ((state->masked_pending || state->resend) && priority != MASKED) { in write_xive() 264 write_xive(xics, ics, state, state->server, MASKED, state->priority); in kvmppc_xics_int_off() 432 * If masked, bail out in icp_deliver_irq() 434 * Note: PAPR doesn't mention anything about masked pending in icp_deliver_irq() 437 * However that would have the effect of losing a masked in icp_deliver_irq() 440 * losing interrupts that occur while masked. in icp_deliver_irq() 446 if (state->priority == MASKED) { in icp_deliver_irq() 447 XICS_DBG("irq %#x masked pending\n", new_irq); in icp_deliver_irq() 1003 …seq_printf(m, "irq 0x%06x: server %#x prio %#x save prio %#x pq_state %d resend %d masked pending … in xics_debug_show() 1048 ics->irq_state[i].priority = MASKED; in kvmppc_xics_create_ics() [all …]
|
/linux/net/openvswitch/ |
H A D | actions.c | 213 /* 'src' is already properly masked. */ 385 const __be32 mask[4], __be32 masked[4]) in mask_ipv6_addr() 387 masked[0] = OVS_MASKED(old[0], addr[0], mask[0]); in mask_ipv6_addr() 388 masked[1] = OVS_MASKED(old[1], addr[1], mask[1]); in mask_ipv6_addr() 389 masked[2] = OVS_MASKED(old[2], addr[2], mask[2]); in mask_ipv6_addr() 390 masked[3] = OVS_MASKED(old[3], addr[3], mask[3]); in mask_ipv6_addr() 525 __be32 masked[4]; in set_ipv6() local 527 mask_ipv6_addr(saddr, key->ipv6_src, mask->ipv6_src, masked); in set_ipv6() 529 if (unlikely(memcmp(saddr, masked, sizeof(masked)))) { in set_ipv6() 530 set_ipv6_addr(skb, flow_key->ip.proto, saddr, masked, in set_ipv6() [all …]
|
/linux/drivers/xen/events/ |
H A D | events_2l.c | 128 * the interrupt edge' if the channel is masked. in evtchn_2l_unmask() 203 * If we masked out all events, wrap to beginning. in evtchn_2l_handle_events() 237 /* If we masked out all events, move on. */ in evtchn_2l_handle_events() 284 printk("%d: masked=%d pending=%d event_sel %0*"PRI_xen_ulong"\n ", i, in xen_debug_interrupt() 337 ? "" : " globally-masked", in xen_debug_interrupt() 339 ? "" : " locally-masked"); in xen_debug_interrupt()
|
/linux/tools/perf/pmu-events/arch/x86/snowridgex/ |
H A D | other.json | 84 …"BriefDescription": "Counts the number of core cycles during which interrupts are masked (disabled… 87 "EventName": "HW_INTERRUPTS.MASKED", 88 …"PublicDescription": "Counts the number of core cycles during which interrupts are masked (disable… 93 … of core cycles during which there are pending interrupts while interrupts are masked (disabled).", 97 …re cycles during which there are pending interrupts while interrupts are masked (disabled). Increm…
|
/linux/tools/perf/pmu-events/arch/x86/elkhartlake/ |
H A D | other.json | 84 …"BriefDescription": "Counts the number of core cycles during which interrupts are masked (disabled… 87 "EventName": "HW_INTERRUPTS.MASKED", 88 …"PublicDescription": "Counts the number of core cycles during which interrupts are masked (disable… 93 … of core cycles during which there are pending interrupts while interrupts are masked (disabled).", 97 …re cycles during which there are pending interrupts while interrupts are masked (disabled). Increm…
|
/linux/arch/mips/include/asm/ |
H A D | msc01_ic.h | 25 #define MSC01_IC_ISAL_OFS 0x00160 /* Masked int_in 31:0 */ 26 #define MSC01_IC_ISAH_OFS 0x00168 /* Masked int_in 63:32 */ 30 #define MSC01_IC_OSA_OFS 0x00190 /* Masked int_out */ 43 #define MSC01_IC_ISA_OFS 0x00860 /* Masked int_in 63:0 */
|
/linux/Documentation/virt/kvm/devices/ |
H A D | xics.rst | 82 * Masked flag, 1 bit 84 This bit is set to 1 if the interrupt is masked (cannot be delivered 86 call, or 0 if it is not masked.
|
/linux/arch/powerpc/include/asm/ |
H A D | hw_irq.h | 42 * Some soft-masked interrupts must be hard masked until they are replayed 43 * (e.g., because the soft-masked handler does not clear the exception). 44 * Interrupt replay itself must remain hard masked too. 342 * is a different soft-masked interrupt pending that requires hard 358 * soft-masked. in should_hard_irq_enable()
|
/linux/include/linux/soc/qcom/ |
H A D | irq.h | 14 * IRQ_DOMAIN_FLAG_QCOM_PDC_WAKEUP: Line must be masked at TLMM and the 27 * requires the interrupt be masked at the child interrupt controller.
|
/linux/drivers/gpio/ |
H A D | gpio-virtio.c | 35 bool masked; member 214 if (WARN_ON(irq_line->queued || irq_line->masked || irq_line->disabled)) in virtio_gpio_irq_prepare() 241 irq_line->masked = false; in virtio_gpio_irq_enable() 256 irq_line->masked = true; in virtio_gpio_irq_disable() 270 irq_line->masked = true; in virtio_gpio_irq_mask() 281 irq_line->masked = false; in virtio_gpio_irq_unmask() 363 if (irq_line->masked || irq_line->disabled) { in ignore_irq() 608 vgpio->irq_lines[i].masked = true; in virtio_gpio_probe()
|
H A D | gpio-adp5585.c | 327 bool masked = test_bit(hwirq, &adp5585_gpio->irq_mask); in adp5585_irq_bus_sync_unlock() local 331 if (masked && !enabled) in adp5585_irq_bus_sync_unlock() 333 if (!masked && enabled) in adp5585_irq_bus_sync_unlock() 339 masked ? 0 : bit); in adp5585_irq_bus_sync_unlock() 340 assign_bit(hwirq, &adp5585_gpio->irq_en, !masked); in adp5585_irq_bus_sync_unlock() 464 /* everything masked by default */ in adp5585_gpio_probe()
|
/linux/arch/arm64/include/asm/ |
H A D | irqflags.h | 16 * Masking debug exceptions causes all other exceptions to be masked too/ 18 * always masked and unmasked together, and have no side effects for other 155 * state if interrupts are already disabled/masked. in __pmr_local_irq_save()
|
/linux/Documentation/admin-guide/perf/ |
H A D | imx-ddr.rst | 40 - 0: corresponding bit is masked. 41 - 1: corresponding bit is not masked, i.e. used to do the matching. 44 When non-masked bits are matching corresponding AXI_ID bits then counter is
|
/linux/drivers/irqchip/ |
H A D | irq-mtk-cirq.c | 207 bool pending, masked; in mtk_cirq_suspend() local 222 * - For each interrupt, inspect its pending and masked status at GIC in mtk_cirq_suspend() 238 &masked); in mtk_cirq_suspend() 241 (pending && !masked)) in mtk_cirq_suspend()
|
/linux/drivers/net/ethernet/ti/ |
H A D | cpts.h | 33 u32 intstat_masked; /* Time sync interrupt status masked */ 66 #define TS_PEND (1<<0) /* masked interrupt read (after enable) */ 67 #define TS_PEND_EN (1<<0) /* masked interrupt enable */
|