| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | cpuid_test.c | 32 for (i = 0; i < guest_cpuid->nent; i++) { in test_guest_cpuids() 89 TEST_ASSERT(cpuid1->nent == cpuid2->nent, in compare_cpuids() 90 "CPUID nent mismatch: %d vs. %d", cpuid1->nent, cpuid2->nent); in compare_cpuids() 92 for (i = 0; i < cpuid1->nent; i++) { in compare_cpuids() 145 int size = sizeof(*cpuid) + cpuid->nent * sizeof(cpuid->entries[0]); in vcpu_alloc_cpuid() 200 struct kvm_cpuid2 *cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent + 1); in test_get_cpuid2() 204 TEST_ASSERT(cpuid->nent == vcpu->cpuid->nent, in test_get_cpuid2() 206 vcpu->cpuid->nent, cpuid->nent); in test_get_cpuid2() 208 for (i = 0; i < vcpu->cpuid->nent; i++) { in test_get_cpuid2() 209 cpuid->nent = i; in test_get_cpuid2() [all …]
|
| H A D | hyperv_cpuid.c | 38 TEST_ASSERT(hv_cpuid_entries->nent == nent_expected, in test_hv_cpuid() 41 nent_expected, hv_cpuid_entries->nent); in test_hv_cpuid() 43 for (i = 0; i < hv_cpuid_entries->nent; i++) { in test_hv_cpuid() 117 static struct kvm_cpuid2 cpuid = {.nent = 0}; in test_hv_cpuid_e2big()
|
| H A D | hyperv_features.c | 154 prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent); in guest_test_msrs_access() 494 memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent)); in guest_test_msrs_access() 546 prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent); in guest_test_hcalls_access() 663 memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent)); in guest_test_hcalls_access()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| H A D | gk20a.c | 41 int nent; in gk20a_gr_av_to_init_() local 44 nent = (blob->size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_init_() 45 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_av_to_init_() 52 for (i = 0; i < nent; i++) { in gk20a_gr_av_to_init_() 84 int nent; in gk20a_gr_aiv_to_init() local 87 nent = (blob->size / sizeof(struct gk20a_fw_aiv)); in gk20a_gr_aiv_to_init() 88 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_aiv_to_init() 95 for (i = 0; i < nent; i++) { in gk20a_gr_aiv_to_init() 117 int nent; in gk20a_gr_av_to_method() local 120 nent = (blob->size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_method() [all …]
|
| /linux/arch/x86/kvm/ |
| H A D | cpuid.c | 88 struct kvm_cpuid_entry2 *entries, int nent, u32 function, u64 index) in kvm_find_cpuid_entry2() argument 105 for (i = 0; i < nent; i++) { in kvm_find_cpuid_entry2() 177 int nent) in kvm_cpuid_check_equal() argument 191 if (nent != vcpu->arch.cpuid_nent) in kvm_cpuid_check_equal() 194 for (i = 0; i < nent; i++) { in kvm_cpuid_check_equal() 510 int nent) in kvm_set_cpuid() argument 534 swap(vcpu->arch.cpuid_nent, nent); in kvm_set_cpuid() 554 r = kvm_cpuid_check_equal(vcpu, e2, nent); in kvm_set_cpuid() 584 swap(vcpu->arch.cpuid_nent, nent); in kvm_set_cpuid() 597 if (cpuid->nent > KVM_MAX_CPUID_ENTRIES) in kvm_vcpu_ioctl_set_cpuid() [all …]
|
| /linux/drivers/hid/intel-thc-hid/intel-thc/ |
| H A D | intel-thc-dma.c | 225 unsigned int i, nent = PRD_ENTRIES_NUM; in setup_dma_buffers() local 250 config->sgls[i] = sgl_alloc(buf_sz, GFP_KERNEL, &nent); in setup_dma_buffers() 251 if (!config->sgls[i] || nent > PRD_ENTRIES_NUM) { in setup_dma_buffers() 253 i, nent); in setup_dma_buffers() 256 count = dma_map_sg(dev->dev, config->sgls[i], nent, dir); in setup_dma_buffers() 258 config->sgls_nent_pages[i] = nent; in setup_dma_buffers() 450 size_t mes_len, u8 *nent) in calc_prd_entries_num() argument 452 *nent = DIV_ROUND_UP(mes_len, THC_MIN_BYTES_PER_SG_LIST_ENTRY); in calc_prd_entries_num() 453 if (*nent > PRD_ENTRIES_NUM) in calc_prd_entries_num() 459 static size_t calc_message_len(struct thc_prd_table *prd_tbl, u8 *nent) in calc_message_len() argument [all …]
|
| /linux/drivers/tty/vt/ |
| H A D | conmakehash.c | 82 int i, nuni, nent; in main() local 269 nent = 0; in main() 272 while ( nent >= unicount[fp0] ) in main() 275 nent = 0; in main() 277 printf("0x%04x", unitable[fp0][nent++]); in main()
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_eq.c | 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci() 230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe() 466 int nent, in mthca_create_eq() argument 479 eq->nent = roundup_pow_of_two(max(nent, 2)); in mthca_create_eq() 480 npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE; in mthca_create_eq() 510 for (i = 0; i < eq->nent; ++i) in mthca_create_eq() 534 eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24); in mthca_create_eq() 559 eq->eqn, eq->nent); in mthca_create_eq() 592 int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZE + PAGE_SIZE - 1) / in mthca_free_eq()
|
| H A D | mthca_allocator.c | 155 int mthca_array_init(struct mthca_array *array, int nent) in mthca_array_init() argument 157 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init() 172 void mthca_array_cleanup(struct mthca_array *array, int nent) in mthca_array_cleanup() argument 176 for (i = 0; i < (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; ++i) in mthca_array_cleanup()
|
| H A D | mthca_cq.c | 348 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent) in mthca_alloc_cq_buf() argument 353 ret = mthca_buf_alloc(dev, nent * MTHCA_CQ_ENTRY_SIZE, in mthca_alloc_cq_buf() 360 for (i = 0; i < nent; ++i) in mthca_alloc_cq_buf() 768 int mthca_init_cq(struct mthca_dev *dev, int nent, in mthca_init_cq() argument 776 cq->ibcq.cqe = nent - 1; in mthca_init_cq() 814 err = mthca_alloc_cq_buf(dev, &cq->buf, nent); in mthca_init_cq() 828 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq()
|
| H A D | mthca_dev.h | 421 int mthca_array_init(struct mthca_array *array, int nent); 422 void mthca_array_cleanup(struct mthca_array *array, int nent); 487 int mthca_init_cq(struct mthca_dev *dev, int nent, 498 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent);
|
| H A D | mthca_cmd.c | 663 int nent = 0; in mthca_map_cmd() local 693 pages[nent * 2] = cpu_to_be64(virt); in mthca_map_cmd() 697 pages[nent * 2 + 1] = in mthca_map_cmd() 703 if (++nent == MTHCA_MAILBOX_SIZE / 16) { in mthca_map_cmd() 704 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd() 708 nent = 0; in mthca_map_cmd() 713 if (nent) in mthca_map_cmd() 714 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd()
|
| H A D | mthca_provider.c | 585 int nent; in mthca_create_cq() local 621 for (nent = 1; nent <= entries; nent <<= 1) in mthca_create_cq() 624 err = mthca_init_cq(to_mdev(ibdev), nent, context, in mthca_create_cq()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | eq.c | 257 u8 log_eq_size = order_base_2(param->nent + MLX5_NUM_SPARE_EQE); in create_map_eq() 662 .nent = MLX5_NUM_CMD_EQE, in create_async_eqs() 675 .nent = async_eq_depth_devlink_param_get(dev), in create_async_eqs() 691 .nent = /* TODO: sriov max_vf + */ 1, in create_async_eqs() 788 u32 nent = eq_get_size(eq); in mlx5_eq_get_eqe() local 791 eqe = get_eqe(eq, ci & (nent - 1)); in mlx5_eq_get_eqe() 792 eqe = ((eqe->owner & 1) ^ !!(ci & nent)) ? NULL : eqe; in mlx5_eq_get_eqe() 991 int nent; in create_comp_eq() local 1005 nent = comp_eq_depth_devlink_param_get(dev); in create_comp_eq() 1022 .nent = nent, in create_comp_eq()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | eq.c | 110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe() 124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw() 783 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 786 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 808 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 810 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 820 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 823 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 970 static int mlx4_create_eq(struct mlx4_dev *dev, int nent, in mlx4_create_eq() argument 984 eq->nent = roundup_pow_of_two(max(nent, 2)); in mlx4_create_eq() [all …]
|
| H A D | cq.c | 342 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, in mlx4_cq_alloc() argument 382 cpu_to_be32((ilog2(nent) << 24) | in mlx4_cq_alloc() 394 err = mlx4_init_user_cqes(buf_addr, nent, in mlx4_cq_alloc() 399 mlx4_init_kernel_cqes(buf_addr, nent, in mlx4_cq_alloc()
|
| /linux/drivers/tty/serial/ |
| H A D | pch_uart.c | 231 int nent; member 743 for (i = 0; i < priv->nent; i++, sg++) in pch_dma_tx_complete() 749 priv->nent = 0; in pch_dma_tx_complete() 855 int nent; in dma_handle_tx() local 931 nent = dma_map_sg(port->dev, sg, num, DMA_TO_DEVICE); in dma_handle_tx() 932 if (!nent) { in dma_handle_tx() 937 priv->nent = nent; in dma_handle_tx() 939 for (i = 0; i < nent; i++, sg++) { in dma_handle_tx() 943 if (i == (nent - 1)) in dma_handle_tx() 950 priv->sg_tx_p, nent, DMA_MEM_TO_DEV, in dma_handle_tx()
|
| /linux/drivers/iommu/ |
| H A D | omap-iommu.c | 704 int nent = 1; in iopgtable_clear_entry_core() local 718 nent *= 16; in iopgtable_clear_entry_core() 722 bytes *= nent; in iopgtable_clear_entry_core() 723 memset(iopte, 0, nent * sizeof(*iopte)); in iopgtable_clear_entry_core() 725 flush_iopte_range(obj->dev, pt_dma, pt_offset, nent); in iopgtable_clear_entry_core() 736 nent = 1; /* for the next L1 entry */ in iopgtable_clear_entry_core() 740 nent *= 16; in iopgtable_clear_entry_core() 744 bytes *= nent; in iopgtable_clear_entry_core() 746 memset(iopgd, 0, nent * sizeof(*iopgd)); in iopgtable_clear_entry_core() 747 flush_iopte_range(obj->dev, obj->pd_dma, pd_offset, nent); in iopgtable_clear_entry_core()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | free_page_reporting.rst | 29 nent-1中。 当页面被报告函数处理时,分配器将无法访问它们。一旦报告函数完成,这些
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | cq.c | 77 static u8 sw_ownership_bit(int n, int nent) in sw_ownership_bit() argument 79 return (n & nent) ? 1 : 0; in sw_ownership_bit() 673 int nent, in alloc_cq_frag_buf() argument 682 nent * cqe_size, in alloc_cq_frag_buf() 691 buf->nent = nent; in alloc_cq_frag_buf() 880 for (i = 0; i < buf->nent; i++) { in init_cq_frag_buf() 1261 (i + 1) & cq->resize_buf->nent); in copy_resize_cqes() 1263 sw_own = sw_ownership_bit(i + 1, cq->resize_buf->nent); in copy_resize_cqes()
|
| /linux/include/linux/mlx5/ |
| H A D | eq.h | 16 int nent; member
|
| /linux/drivers/vfio/pci/mlx5/ |
| H A D | cmd.c | 1054 struct mlx5_vhca_cq_buf *buf, int nent, in alloc_cq_frag_buf() argument 1062 err = mlx5_frag_buf_alloc_node(mdev, nent * cqe_size, frag_buf, in alloc_cq_frag_buf() 1069 buf->nent = nent; in alloc_cq_frag_buf() 1079 for (i = 0; i < buf->nent; i++) { in init_cq_frag_buf() 1617 u32 nent = size / entry_size; in set_report_output() local 1631 (nent > qp->max_msg_size / entry_size))) in set_report_output() 1641 nent_to_set = min(nent, nent_in_page); in set_report_output() 1651 nent -= nent_to_set; in set_report_output() 1652 } while (nent); in set_report_output()
|
| /linux/lib/ |
| H A D | scatterlist.c | 620 unsigned int nent, nalloc; in sgl_alloc_order() local 623 nent = round_up(length, PAGE_SIZE << order) >> (PAGE_SHIFT + order); in sgl_alloc_order() 625 if (length > (nent << (PAGE_SHIFT + order))) in sgl_alloc_order() 627 nalloc = nent; in sgl_alloc_order() 654 *nent_p = nent; in sgl_alloc_order()
|
| /linux/drivers/mailbox/ |
| H A D | bcm-pdc-mailbox.c | 1198 int nent; in pdc_send_data() local 1207 nent = dma_map_sg(dev, mssg->spu.src, src_nent, DMA_TO_DEVICE); in pdc_send_data() 1208 if (unlikely(nent == 0)) in pdc_send_data() 1214 nent = dma_map_sg(dev, mssg->spu.dst, dst_nent, in pdc_send_data() 1216 if (unlikely(nent == 0)) { in pdc_send_data()
|
| /linux/tools/testing/selftests/kvm/lib/x86/ |
| H A D | processor.c | 913 for (i = 0; i < cpuid->nent; i++) { in __kvm_cpu_has() 1004 if (vcpu->cpuid && vcpu->cpuid->nent < cpuid->nent) { in vcpu_init_cpuid() 1010 vcpu->cpuid = allocate_kvm_cpuid2(cpuid->nent); in vcpu_init_cpuid() 1012 memcpy(vcpu->cpuid, cpuid, kvm_cpuid2_size(cpuid->nent)); in vcpu_init_cpuid() 1330 for (i = 0; i < cpuid->nent; i++) { in get_cpuid_entry()
|