| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | deny_namespace.c | 51 __u32 cap_mask = 1ULL << CAP_SYS_ADMIN; in test_userns_create_bpf() local 54 cap_enable_effective(cap_mask, &old_caps); in test_userns_create_bpf() 58 cap_disable_effective(cap_mask, &old_caps); in test_userns_create_bpf() 62 if (cap_mask & old_caps) in test_userns_create_bpf() 63 cap_enable_effective(cap_mask, NULL); in test_userns_create_bpf() 68 __u32 cap_mask = 1ULL << CAP_SYS_ADMIN; in test_unpriv_userns_create_no_bpf() local 71 cap_disable_effective(cap_mask, &old_caps); in test_unpriv_userns_create_no_bpf() 75 if (cap_mask & old_caps) in test_unpriv_userns_create_no_bpf() 76 cap_enable_effective(cap_mask, NULL); in test_unpriv_userns_create_no_bpf()
|
| /linux/drivers/dma/ |
| H A D | dmaengine.c | 330 if (!dma_has_cap(cap, device->cap_mask) || in min_chan() 331 dma_has_cap(DMA_PRIVATE, device->cap_mask)) in min_chan() 376 if (dma_has_cap(DMA_PRIVATE, device->cap_mask)) in dma_channel_rebalance() 399 bitmap_and(has.bits, want->bits, device->cap_mask.bits, in dma_device_satisfies_mask() 478 if (!dma_has_cap(DMA_PRIVATE, chan->device->cap_mask)) in dma_chan_get() 563 if (dma_has_cap(DMA_PRIVATE, device->cap_mask)) in dma_issue_pending_all() 583 if (!(test_bit(DMA_SLAVE, device->cap_mask.bits) || in dma_get_slave_caps() 584 test_bit(DMA_CYCLIC, device->cap_mask.bits))) in dma_get_slave_caps() 634 if (dev->chancnt > 1 && !dma_has_cap(DMA_PRIVATE, dev->cap_mask)) in private_candidate() 671 dma_cap_set(DMA_PRIVATE, device->cap_mask); in find_candidate() [all …]
|
| H A D | mv_xor.c | 1036 int idx, dma_cap_mask_t cap_mask, int irq) in mv_xor_channel_add() argument 1088 dma_dev->cap_mask = cap_mask; in mv_xor_channel_add() 1099 if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) in mv_xor_channel_add() 1101 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in mv_xor_channel_add() 1103 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in mv_xor_channel_add() 1137 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { in mv_xor_channel_add() 1144 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in mv_xor_channel_add() 1153 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "", in mv_xor_channel_add() 1154 dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "", in mv_xor_channel_add() 1155 dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : ""); in mv_xor_channel_add() [all …]
|
| H A D | at_hdmac.c | 518 dma_cap_mask_t cap_mask; member 1952 dma_cap_set(DMA_MEMCPY, at91sam9rl_config.cap_mask); in at_dma_probe() 1953 dma_cap_set(DMA_INTERLEAVE, at91sam9g45_config.cap_mask); in at_dma_probe() 1954 dma_cap_set(DMA_MEMCPY, at91sam9g45_config.cap_mask); in at_dma_probe() 1955 dma_cap_set(DMA_MEMSET, at91sam9g45_config.cap_mask); in at_dma_probe() 1956 dma_cap_set(DMA_MEMSET_SG, at91sam9g45_config.cap_mask); in at_dma_probe() 1957 dma_cap_set(DMA_PRIVATE, at91sam9g45_config.cap_mask); in at_dma_probe() 1958 dma_cap_set(DMA_SLAVE, at91sam9g45_config.cap_mask); in at_dma_probe() 1980 atdma->dma_device.cap_mask = plat_dat->cap_mask; in at_dma_probe() 2048 if (dma_has_cap(DMA_INTERLEAVE, atdma->dma_device.cap_mask)) in at_dma_probe() [all …]
|
| H A D | bcm-sba-raid.c | 1553 dma_cap_zero(dma_dev->cap_mask); in sba_async_register() 1554 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in sba_async_register() 1555 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in sba_async_register() 1556 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in sba_async_register() 1557 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in sba_async_register() 1573 if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) in sba_async_register() 1577 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in sba_async_register() 1581 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in sba_async_register() 1587 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in sba_async_register() 1605 dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "interrupt " : "", in sba_async_register() [all …]
|
| H A D | ep93xx_dma.c | 1449 dma_cap_mask_t mask = edma->dma_dev.cap_mask; in ep93xx_m2p_dma_of_xlate() 1480 dma_cap_mask_t mask = edma->dma_dev.cap_mask; in ep93xx_m2m_dma_of_xlate() 1514 dma_cap_zero(dma_dev->cap_mask); in ep93xx_dma_probe() 1515 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in ep93xx_dma_probe() 1516 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in ep93xx_dma_probe() 1532 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ep93xx_dma_probe() 1540 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ep93xx_dma_probe()
|
| H A D | dmatest.c | 858 dev->cap_mask) && in dmatest_func() 1037 if (dma_has_cap(DMA_COMPLETION_NO_ORDER, dma_dev->cap_mask) && in dmatest_add_channel() 1043 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { in dmatest_add_channel() 1050 if (dma_has_cap(DMA_MEMSET, dma_dev->cap_mask)) { in dmatest_add_channel() 1057 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in dmatest_add_channel() 1061 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in dmatest_add_channel()
|
| H A D | img-mdc-dma.c | 912 dma_cap_zero(mdma->dma_dev.cap_mask); in mdc_dma_probe() 913 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 914 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 915 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); in mdc_dma_probe() 916 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); in mdc_dma_probe()
|
| H A D | moxart-dma.c | 580 dma_cap_zero(mdc->dma_slave.cap_mask); in moxart_probe() 581 dma_cap_set(DMA_SLAVE, mdc->dma_slave.cap_mask); in moxart_probe() 582 dma_cap_set(DMA_PRIVATE, mdc->dma_slave.cap_mask); in moxart_probe()
|
| /linux/drivers/dma/ioat/ |
| H A D | sysfs.c | 22 dma_has_cap(DMA_PQ, dma->cap_mask) ? " pq" : "", in cap_show() 23 dma_has_cap(DMA_PQ_VAL, dma->cap_mask) ? " pq_val" : "", in cap_show() 24 dma_has_cap(DMA_XOR, dma->cap_mask) ? " xor" : "", in cap_show() 25 dma_has_cap(DMA_XOR_VAL, dma->cap_mask) ? " xor_val" : "", in cap_show() 26 dma_has_cap(DMA_INTERRUPT, dma->cap_mask) ? " intr" : ""); in cap_show()
|
| H A D | init.c | 512 dma_cap_set(DMA_MEMCPY, dma->cap_mask); in ioat_probe() 795 if (!dma_has_cap(DMA_XOR, dma->cap_mask)) in ioat_xor_val_self_test() 903 if (!dma_has_cap(DMA_XOR_VAL, dma_chan->device->cap_mask)) in ioat_xor_val_self_test() 1092 dma_cap_set(DMA_INTERRUPT, dma->cap_mask); in ioat3_dma_probe() 1108 dma_cap_set(DMA_XOR, dma->cap_mask); in ioat3_dma_probe() 1111 dma_cap_set(DMA_XOR_VAL, dma->cap_mask); in ioat3_dma_probe() 1119 dma_cap_set(DMA_PQ, dma->cap_mask); in ioat3_dma_probe() 1120 dma_cap_set(DMA_PQ_VAL, dma->cap_mask); in ioat3_dma_probe() 1130 dma_cap_set(DMA_XOR, dma->cap_mask); in ioat3_dma_probe() 1131 dma_cap_set(DMA_XOR_VAL, dma->cap_mask); in ioat3_dma_probe() [all …]
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | transmeta.c | 24 unsigned int cap_mask, uk, max, dummy; in init_transmeta() local 86 rdmsr(0x80860004, cap_mask, uk); in init_transmeta() 89 wrmsr(0x80860004, cap_mask, uk); in init_transmeta()
|
| /linux/arch/arm/plat-orion/ |
| H A D | common.c | 614 dma_cap_set(DMA_MEMCPY, orion_xor0_channels_data[0].cap_mask); in orion_xor0_init() 615 dma_cap_set(DMA_XOR, orion_xor0_channels_data[0].cap_mask); in orion_xor0_init() 617 dma_cap_set(DMA_MEMCPY, orion_xor0_channels_data[1].cap_mask); in orion_xor0_init() 618 dma_cap_set(DMA_XOR, orion_xor0_channels_data[1].cap_mask); in orion_xor0_init() 675 dma_cap_set(DMA_MEMCPY, orion_xor1_channels_data[0].cap_mask); in orion_xor1_init() 676 dma_cap_set(DMA_XOR, orion_xor1_channels_data[0].cap_mask); in orion_xor1_init() 678 dma_cap_set(DMA_MEMCPY, orion_xor1_channels_data[1].cap_mask); in orion_xor1_init() 679 dma_cap_set(DMA_XOR, orion_xor1_channels_data[1].cap_mask); in orion_xor1_init()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_deny_namespace.c | 20 __u64 cap_mask = 1ULL << CAP_SYS_ADMIN; in BPF_PROG() local 26 if (caps.val & cap_mask) in BPF_PROG()
|
| /linux/crypto/async_tx/ |
| H A D | async_tx.c | 50 dma_has_cap(tx_type, depend_tx->chan->device->cap_mask)) in __async_tx_find_channel() 89 if (dma_has_cap(DMA_INTERRUPT, device->cap_mask)) in async_tx_channel_switch() 235 if (device && !dma_has_cap(DMA_INTERRUPT, device->cap_mask)) in async_trigger_callback()
|
| /linux/drivers/dma/idxd/ |
| H A D | dma.c | 212 dma_cap_set(DMA_INTERRUPT, dma->cap_mask); in idxd_register_dma_device() 213 dma_cap_set(DMA_PRIVATE, dma->cap_mask); in idxd_register_dma_device() 214 dma_cap_set(DMA_COMPLETION_NO_ORDER, dma->cap_mask); in idxd_register_dma_device() 219 dma_cap_set(DMA_MEMCPY, dma->cap_mask); in idxd_register_dma_device()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | driver.c | 50 module_param_cb(cap_mask, &cap_ops, &hfi1_cap_mask, S_IWUSR | S_IRUGO); 51 MODULE_PARM_DESC(cap_mask, "Bit mask of enabled/disabled HW features"); 73 cap_mask = *cap_mask_ptr, value, diff, in hfi1_caps_set() local 83 diff = value ^ (cap_mask & ~HFI1_CAP_LOCKED_SMASK); in hfi1_caps_set() 95 cap_mask &= ~diff; in hfi1_caps_set() 97 cap_mask |= (value & diff); in hfi1_caps_set() 99 diff = (cap_mask & (HFI1_CAP_MUST_HAVE_KERN << HFI1_CAP_USER_SHIFT)) ^ in hfi1_caps_set() 100 ((cap_mask & HFI1_CAP_MUST_HAVE_KERN) << HFI1_CAP_USER_SHIFT); in hfi1_caps_set() 101 cap_mask &= ~diff; in hfi1_caps_set() 103 *cap_mask_ptr = cap_mask; in hfi1_caps_set() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qlcnic/ |
| H A D | qlcnic_minidump.c | 295 hdr->drv_cap_mask = hdr->cap_mask; in qlcnic_82xx_cache_tmpl_hdr_values() 296 fw_dump->cap_mask = hdr->cap_mask; in qlcnic_82xx_cache_tmpl_hdr_values() 349 hdr->drv_cap_mask = hdr->cap_mask; in qlcnic_83xx_cache_tmpl_hdr_values() 350 fw_dump->cap_mask = hdr->cap_mask; in qlcnic_83xx_cache_tmpl_hdr_values() 1277 fw_dump->cap_mask); in qlcnic_fw_cmd_get_minidump_temp() 1321 if (i & fw_dump->cap_mask) in qlcnic_dump_fw() 1352 if (!(entry->hdr.mask & fw_dump->cap_mask)) { in qlcnic_dump_fw() 1449 fw_dump->cap_mask = 0x1f; in qlcnic_83xx_get_minidump_template()
|
| /linux/include/linux/platform_data/ |
| H A D | dma-mv_xor.h | 15 dma_cap_mask_t cap_mask; member
|
| H A D | dma-iop32x.h | 102 dma_cap_mask_t cap_mask; member
|
| /linux/drivers/dma/ppc4xx/ |
| H A D | adma.c | 1400 if (dma_has_cap(cap, ref->chan->device->cap_mask)) { in ppc440spe_async_tx_find_best_channel() 3771 dma_cap_set(DMA_MEMCPY, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3772 dma_cap_set(DMA_INTERRUPT, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3773 dma_cap_set(DMA_PQ, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3774 dma_cap_set(DMA_PQ_VAL, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3775 dma_cap_set(DMA_XOR_VAL, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3778 dma_cap_set(DMA_XOR, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3779 dma_cap_set(DMA_PQ, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3780 dma_cap_set(DMA_INTERRUPT, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3781 adev->common.cap_mask = adev->common.cap_mask; in ppc440spe_adma_init_capabilities() [all …]
|
| /linux/sound/soc/qcom/qdsp6/ |
| H A D | q6afe-dai.c | 161 unsigned int cap_mask; in q6tdm_set_tdm_slot() local 174 cap_mask = 0x03; in q6tdm_set_tdm_slot() 177 cap_mask = 0x0F; in q6tdm_set_tdm_slot() 180 cap_mask = 0xFF; in q6tdm_set_tdm_slot() 183 cap_mask = 0xFFFF; in q6tdm_set_tdm_slot() 196 tdm->slot_mask = ((dai->id & 0x1) ? tx_mask : rx_mask) & cap_mask; in q6tdm_set_tdm_slot()
|
| /linux/drivers/dma/xilinx/ |
| H A D | xdma.c | 1255 dma_cap_set(DMA_SLAVE, xdev->dma_dev.cap_mask); in xdma_probe() 1256 dma_cap_set(DMA_PRIVATE, xdev->dma_dev.cap_mask); in xdma_probe() 1257 dma_cap_set(DMA_CYCLIC, xdev->dma_dev.cap_mask); in xdma_probe() 1258 dma_cap_set(DMA_INTERLEAVE, xdev->dma_dev.cap_mask); in xdma_probe() 1259 dma_cap_set(DMA_REPEAT, xdev->dma_dev.cap_mask); in xdma_probe() 1260 dma_cap_set(DMA_LOAD_EOT, xdev->dma_dev.cap_mask); in xdma_probe()
|
| /linux/drivers/platform/x86/intel/speed_select_if/ |
| H A D | isst_tpmi_core.c | 70 u8 cap_mask; member 630 core_power.supported = !!(power_domain_info->sst_header.cap_mask & BIT(0)); in isst_if_core_power_state() 875 perf_level.enabled = !!(power_domain_info->sst_header.cap_mask & BIT(1)); in isst_if_get_perf_level() 1747 if (!(pd_info->sst_header.cap_mask & SST_PP_CAP_CP_ENABLE)) in tpmi_sst_dev_suspend() 1758 if (!(pd_info->sst_header.cap_mask & SST_PP_CAP_PP_ENABLE)) in tpmi_sst_dev_suspend() 1788 if (!(pd_info->sst_header.cap_mask & SST_PP_CAP_CP_ENABLE)) in tpmi_sst_dev_resume() 1799 if (!(pd_info->sst_header.cap_mask & SST_PP_CAP_PP_ENABLE)) in tpmi_sst_dev_resume()
|
| /linux/drivers/dma/dw-edma/ |
| H A D | dw-edma-core.c | 797 dma_cap_zero(dma->cap_mask); in dw_edma_channel_setup() 798 dma_cap_set(DMA_SLAVE, dma->cap_mask); in dw_edma_channel_setup() 799 dma_cap_set(DMA_CYCLIC, dma->cap_mask); in dw_edma_channel_setup() 800 dma_cap_set(DMA_PRIVATE, dma->cap_mask); in dw_edma_channel_setup() 801 dma_cap_set(DMA_INTERLEAVE, dma->cap_mask); in dw_edma_channel_setup()
|