| /linux/arch/riscv/kvm/ |
| H A D | vm.c | 95 struct kvm_irq_routing_entry *ents; in kvm_riscv_setup_default_irq_routing() local 98 ents = kcalloc(lines, sizeof(*ents), GFP_KERNEL); in kvm_riscv_setup_default_irq_routing() 99 if (!ents) in kvm_riscv_setup_default_irq_routing() 103 ents[i].gsi = i; in kvm_riscv_setup_default_irq_routing() 104 ents[i].type = KVM_IRQ_ROUTING_IRQCHIP; in kvm_riscv_setup_default_irq_routing() 105 ents[i].u.irqchip.irqchip = 0; in kvm_riscv_setup_default_irq_routing() 106 ents[i].u.irqchip.pin = i; in kvm_riscv_setup_default_irq_routing() 108 rc = kvm_set_irq_routing(kvm, ents, lines, 0); in kvm_riscv_setup_default_irq_routing() 109 kfree(ents); in kvm_riscv_setup_default_irq_routing()
|
| /linux/net/can/j1939/ |
| H A D | bus.c | 58 ent = &priv->ents[ecu->addr]; in j1939_ecu_map_locked() 85 ent = &priv->ents[ecu->addr]; in j1939_ecu_unmap_locked() 103 for (i = 0; i < ARRAY_SIZE(priv->ents); i++) in j1939_ecu_unmap_all() 104 if (priv->ents[i].ecu) in j1939_ecu_unmap_all() 105 j1939_ecu_unmap_locked(priv->ents[i].ecu); in j1939_ecu_unmap_all() 177 return priv->ents[addr].ecu; in j1939_ecu_find_by_addr_locked() 284 priv->ents[sa].nusers++; in j1939_local_ecu_get() 295 priv->ents[sa].nusers--; in j1939_local_ecu_get() 303 priv->ents[ecu->addr].nusers++; in j1939_local_ecu_get() 318 priv->ents[sa].nusers--; in j1939_local_ecu_put() [all …]
|
| H A D | main.c | 98 priv->ents[skcb->addr.sa].nusers) in j1939_can_recv() 101 priv->ents[skcb->addr.da].nusers) in j1939_can_recv()
|
| H A D | j1939-priv.h | 71 } ents[256]; member
|
| H A D | transport.c | 362 priv->ents[skcb->addr.da].nusers) in j1939_session_skb_queue() 2013 priv->ents[skcb->addr.da].nusers) in j1939_tp_send()
|
| /linux/drivers/gpu/drm/virtio/ |
| H A D | virtgpu_prime.c | 146 int virtgpu_dma_buf_import_sgt(struct virtio_gpu_mem_entry **ents, in virtgpu_dma_buf_import_sgt() argument 167 *ents = kvmalloc_array(sgt->nents, in virtgpu_dma_buf_import_sgt() 170 if (!(*ents)) { in virtgpu_dma_buf_import_sgt() 177 (*ents)[i].addr = cpu_to_le64(sg_dma_address(sl)); in virtgpu_dma_buf_import_sgt() 178 (*ents)[i].length = cpu_to_le32(sg_dma_len(sl)); in virtgpu_dma_buf_import_sgt() 179 (*ents)[i].padding = 0; in virtgpu_dma_buf_import_sgt() 235 struct virtio_gpu_mem_entry *ents = NULL; in virtgpu_dma_buf_init_obj() local 251 ret = virtgpu_dma_buf_import_sgt(&ents, &nents, bo, attach); in virtgpu_dma_buf_init_obj() 261 ents, nents); in virtgpu_dma_buf_init_obj()
|
| /linux/drivers/mtd/ubi/ |
| H A D | cdev.c | 694 if (req->ents[i].vol_id < 0 || in rename_volumes() 695 req->ents[i].vol_id >= ubi->vtbl_slots) in rename_volumes() 697 if (req->ents[i].name_len < 0) in rename_volumes() 699 if (req->ents[i].name_len > UBI_VOL_NAME_MAX) in rename_volumes() 701 req->ents[i].name[req->ents[i].name_len] = '\0'; in rename_volumes() 702 n = strlen(req->ents[i].name); in rename_volumes() 703 if (n != req->ents[i].name_len) in rename_volumes() 710 if (req->ents[i].vol_id == req->ents[n].vol_id) { in rename_volumes() 712 req->ents[i].vol_id); in rename_volumes() 715 if (!strcmp(req->ents[i].name, req->ents[n].name)) { in rename_volumes() [all …]
|
| /linux/kernel/dma/ |
| H A D | mapping.c | 234 int ents; in __dma_map_sg_attrs() local 243 ents = dma_direct_map_sg(dev, sg, nents, dir, attrs); in __dma_map_sg_attrs() 245 ents = iommu_dma_map_sg(dev, sg, nents, dir, attrs); in __dma_map_sg_attrs() 247 ents = ops->map_sg(dev, sg, nents, dir, attrs); in __dma_map_sg_attrs() 249 if (ents > 0) { in __dma_map_sg_attrs() 251 trace_dma_map_sg(dev, sg, nents, ents, dir, attrs); in __dma_map_sg_attrs() 252 debug_dma_map_sg(dev, sg, nents, ents, dir, attrs); in __dma_map_sg_attrs() 253 } else if (WARN_ON_ONCE(ents != -EINVAL && ents != -ENOMEM && in __dma_map_sg_attrs() 254 ents != -EIO && ents != -EREMOTEIO)) { in __dma_map_sg_attrs() 255 trace_dma_map_sg_err(dev, sg, nents, ents, dir, attrs); in __dma_map_sg_attrs() [all …]
|
| /linux/drivers/cxl/core/ |
| H A D | features.c | 106 hdr_size = struct_size(mbox_out, ents, 0); in get_supported_features() 119 alloc_size = struct_size(mbox_out, ents, max_feats); in get_supported_features() 123 alloc_size = struct_size(mbox_out, ents, remain_feats); in get_supported_features() 163 memcpy(entry, mbox_out->ents, retrieved); in get_supported_features() 417 struct_size(feat_out, ents, requested); in cxlctl_get_supported_features() 424 rpc_out->size = struct_size(feat_out, ents, requested); in cxlctl_get_supported_features() 427 for (i = start, pos = &feat_out->ents[0]; in cxlctl_get_supported_features()
|
| /linux/tools/testing/selftests/filesystems/ |
| H A D | file_stressor.c | 155 char ents[1024]; variable 177 nr_read = syscall(SYS_getdents64, self->fd_proc_pid[i], ents, sizeof(ents));
|
| /linux/lib/crypto/mpi/ |
| H A D | mpicoder.c | 335 int x, j, z, lzeros, ents; in mpi_read_raw_from_sgl() local 341 ents = sg_nents_for_len(sgl, nbytes); in mpi_read_raw_from_sgl() 342 if (ents < 0) in mpi_read_raw_from_sgl() 345 sg_miter_start(&miter, sgl, ents, SG_MITER_ATOMIC | SG_MITER_FROM_SG); in mpi_read_raw_from_sgl()
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_dir2_priv.h | 26 struct xfs_dir2_leaf_entry *ents; member 108 struct xfs_dir2_leaf_entry *ents, int *indexp, 126 struct xfs_dir2_leaf_entry *ents, int index, int compact,
|
| H A D | xfs_da_btree.c | 775 size = (int)((char *)&leafhdr.ents[leafhdr.count] - in xfs_da3_root_split() 2472 struct xfs_dir2_leaf_entry *ents; in xfs_da3_swap_lastblock() local 2477 ents = leafhdr.ents; in xfs_da3_swap_lastblock() 2479 dead_hash = be32_to_cpu(ents[leafhdr.count - 1].hashval); in xfs_da3_swap_lastblock()
|
| /linux/include/trace/events/ |
| H A D | dma.h | 280 int ents, enum dma_data_direction dir, unsigned long attrs), 281 TP_ARGS(dev, sgl, nents, ents, dir, attrs), 286 __dynamic_array(u64, dma_addrs, ents) 287 __dynamic_array(unsigned int, lengths, ents) 299 for_each_sg(sgl, sg, ents, i) {
|
| /linux/arch/sparc/mm/ |
| H A D | init_64.c | 115 int ents, ret, i; in read_obp_memory() local 117 ents = prop_size / sizeof(struct linux_prom64_registers); in read_obp_memory() 118 if (ents > MAX_BANKS) { in read_obp_memory() 135 for (i = 0; i < ents; i++) { in read_obp_memory() 156 (ents - i - 1) * sizeof(regs[0])); in read_obp_memory() 158 ents--; in read_obp_memory() 165 *num_ents = ents; in read_obp_memory() 167 sort(regs, ents, sizeof(struct linux_prom64_registers), in read_obp_memory() 600 int n, node, ents, first, last, i; in read_obp_translations() local 622 ents = n; in read_obp_translations() [all …]
|
| /linux/include/uapi/cxl/ |
| H A D | features.h | 105 struct cxl_feat_entry ents[] __counted_by_le(num_entries);
|
| /linux/include/uapi/mtd/ |
| H A D | ubi-user.h | 414 } ents[UBI_MAX_RNVOL]; member
|
| /linux/tools/testing/cxl/test/ |
| H A D | mem.c | 1519 if (count < struct_size(out, ents, 0)) { in mock_get_supported_features() 1526 if (count < struct_size(out, ents, MAX_CXL_TEST_FEATS)) { in mock_get_supported_features() 1532 feat = out->ents; in mock_get_supported_features()
|
| /linux/drivers/gpu/drm/qxl/ |
| H A D | qxl_dev.h | 822 uint32_t ents[]; member
|
| /linux/net/core/ |
| H A D | sysctl_net_core.c | 181 sock_table->ents[i] = RPS_NO_CPU; in rps_sock_flow_sysctl()
|
| H A D | dev.c | 5090 ident = READ_ONCE(sock_flow_table->ents[hash & sock_flow_table->mask]); in get_rps_cpu()
|
| /linux/fs/proc/ |
| H A D | base.c | 2739 const struct pid_entry *ents, unsigned int nents) in proc_pident_readdir() argument 2753 for (p = ents + (ctx->pos - 2); p < ents + nents; p++) { in proc_pident_readdir()
|