| /linux/arch/arm64/kvm/vgic/ |
| H A D | vgic-init.c | 54 struct vgic_dist *dist = &kvm->arch.vgic; in kvm_vgic_early_init() local 56 xa_init_flags(&dist->lpi_xa, XA_FLAGS_LOCK_IRQ); in kvm_vgic_early_init() 197 struct vgic_dist *dist = &kvm->arch.vgic; in kvm_vgic_dist_init() local 201 dist->active_spis = (atomic_t)ATOMIC_INIT(0); in kvm_vgic_dist_init() 202 dist->spis = kcalloc(nr_spis, sizeof(struct vgic_irq), GFP_KERNEL_ACCOUNT); in kvm_vgic_dist_init() 203 if (!dist->spis) in kvm_vgic_dist_init() 215 struct vgic_irq *irq = &dist->spis[i]; in kvm_vgic_dist_init() 223 switch (dist->vgic_model) { in kvm_vgic_dist_init() 233 kfree(dist->spis); in kvm_vgic_dist_init() 234 dist->spis = NULL; in kvm_vgic_dist_init() [all …]
|
| H A D | vgic-v4.c | 192 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_v4_configure_vsgis() local 201 if (dist->nassgireq) in vgic_v4_configure_vsgis() 244 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_v4_init() local 254 if (dist->its_vm.vpes) in vgic_v4_init() 259 dist->its_vm.vpes = kcalloc(nr_vcpus, sizeof(*dist->its_vm.vpes), in vgic_v4_init() 261 if (!dist->its_vm.vpes) in vgic_v4_init() 264 dist->its_vm.nr_vpes = nr_vcpus; in vgic_v4_init() 267 dist->its_vm.vpes[i] = &vcpu->arch.vgic_cpu.vgic_v3.its_vpe; in vgic_v4_init() 269 ret = its_alloc_vcpu_irqs(&dist->its_vm); in vgic_v4_init() 272 kfree(dist->its_vm.vpes); in vgic_v4_init() [all …]
|
| H A D | vgic-debug.c | 37 struct vgic_dist *dist = &kvm->arch.vgic; in iter_next() local 51 xa_find_after(&dist->lpi_xa, &iter->intid, in iter_next() 66 struct vgic_dist *dist = &kvm->arch.vgic; in iter_mark_lpis() local 71 xa_lock_irqsave(&dist->lpi_xa, flags); in iter_mark_lpis() 73 xa_for_each(&dist->lpi_xa, intid, irq) { in iter_mark_lpis() 77 __xa_set_mark(&dist->lpi_xa, intid, LPI_XA_MARK_DEBUG_ITER); in iter_mark_lpis() 81 xa_unlock_irqrestore(&dist->lpi_xa, flags); in iter_mark_lpis() 88 struct vgic_dist *dist = &kvm->arch.vgic; in iter_unmark_lpis() local 92 xa_for_each_marked(&dist->lpi_xa, intid, irq, LPI_XA_MARK_DEBUG_ITER) { in iter_unmark_lpis() 93 xa_lock_irqsave(&dist->lpi_xa, flags); in iter_unmark_lpis() [all …]
|
| H A D | vgic-mmio-v3.c | 43 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_has_its() local 45 if (dist->vgic_model != KVM_DEV_TYPE_ARM_VGIC_V3) in vgic_has_its() 48 return dist->has_its; in vgic_has_its() 125 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_mmio_write_v3_misc() local 133 was_enabled = dist->enabled; in vgic_mmio_write_v3_misc() 134 is_hwsgi = dist->nassgireq; in vgic_mmio_write_v3_misc() 136 dist->enabled = val & GICD_CTLR_ENABLE_SS_G1; in vgic_mmio_write_v3_misc() 143 if (was_enabled && dist->enabled) { in vgic_mmio_write_v3_misc() 149 dist->nassgireq = val & GICD_CTLR_nASSGIreq; in vgic_mmio_write_v3_misc() 150 if (is_hwsgi != dist->nassgireq) in vgic_mmio_write_v3_misc() [all …]
|
| H A D | vgic-v2.c | 422 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_v2_map_resources() local 426 if (IS_VGIC_ADDR_UNDEF(dist->vgic_dist_base) || in vgic_v2_map_resources() 427 IS_VGIC_ADDR_UNDEF(dist->vgic_cpu_base)) { in vgic_v2_map_resources() 432 if (!vgic_v2_check_base(dist->vgic_dist_base, dist->vgic_cpu_base)) { in vgic_v2_map_resources() 447 len = vgic_v2_init_cpuif_iodev(&dist->cpuif_iodev); in vgic_v2_map_resources() 448 dist->cpuif_iodev.base_addr = dist->vgic_cpu_base; in vgic_v2_map_resources() 449 dist->cpuif_iodev.iodev_type = IODEV_CPUIF; in vgic_v2_map_resources() 450 dist->cpuif_iodev.redist_vcpu = NULL; in vgic_v2_map_resources() 452 ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS, dist->vgic_cpu_base, in vgic_v2_map_resources() 453 len, &dist->cpuif_iodev.dev); in vgic_v2_map_resources() [all …]
|
| H A D | vgic.c | 68 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_get_lpi() local 73 irq = xa_load(&dist->lpi_xa, intid); in vgic_get_lpi() 117 static void vgic_release_lpi_locked(struct vgic_dist *dist, struct vgic_irq *irq) in vgic_release_lpi_locked() argument 119 lockdep_assert_held(&dist->lpi_xa.xa_lock); in vgic_release_lpi_locked() 120 __xa_erase(&dist->lpi_xa, irq->intid); in vgic_release_lpi_locked() 143 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_put_irq() local 152 guard(spinlock_irqsave)(&dist->lpi_xa.xa_lock); in vgic_put_irq() 158 xa_lock_irqsave(&dist->lpi_xa, flags); in vgic_put_irq() 159 vgic_release_lpi_locked(dist, irq); in vgic_put_irq() 160 xa_unlock_irqrestore(&dist->lpi_xa, flags); in vgic_put_irq() [all …]
|
| H A D | vgic-v3.c | 567 struct vgic_dist *dist = &kvm->arch.vgic; in unmap_all_vpes() local 570 for (i = 0; i < dist->its_vm.nr_vpes; i++) in unmap_all_vpes() 571 free_irq(dist->its_vm.vpes[i]->irq, kvm_get_vcpu(kvm, i)); in unmap_all_vpes() 576 struct vgic_dist *dist = &kvm->arch.vgic; in map_all_vpes() local 579 for (i = 0; i < dist->its_vm.nr_vpes; i++) in map_all_vpes() 581 dist->its_vm.vpes[i]->irq)); in map_all_vpes() 590 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_v3_save_pending_tables() local 611 xa_for_each(&dist->lpi_xa, index, irq) { in vgic_v3_save_pending_tables() 752 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_v3_map_resources() local 765 if (IS_VGIC_ADDR_UNDEF(dist->vgic_dist_base)) { in vgic_v3_map_resources()
|
| H A D | vgic-mmio-v2.c | 56 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_mmio_write_v2_misc() local 57 bool was_enabled = dist->enabled; in vgic_mmio_write_v2_misc() 61 dist->enabled = val & GICD_ENABLE; in vgic_mmio_write_v2_misc() 62 if (!was_enabled && dist->enabled) in vgic_mmio_write_v2_misc() 76 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in vgic_mmio_uaccess_write_v2_misc() local 99 dist->implementation_rev = reg; in vgic_mmio_uaccess_write_v2_misc()
|
| H A D | vgic-irqfd.c | 139 struct vgic_dist *dist = &kvm->arch.vgic; in kvm_vgic_setup_default_irq_routing() local 140 u32 nr = dist->nr_spis; in kvm_vgic_setup_default_irq_routing()
|
| H A D | vgic-its.c | 79 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_add_lpi() local 92 ret = xa_reserve_irq(&dist->lpi_xa, intid, GFP_KERNEL_ACCOUNT); in vgic_add_lpi() 107 xa_lock_irqsave(&dist->lpi_xa, flags); in vgic_add_lpi() 113 oldirq = xa_load(&dist->lpi_xa, intid); in vgic_add_lpi() 119 ret = xa_err(__xa_store(&dist->lpi_xa, intid, irq, 0)); in vgic_add_lpi() 122 xa_unlock_irqrestore(&dist->lpi_xa, flags); in vgic_add_lpi() 125 xa_release(&dist->lpi_xa, intid); in vgic_add_lpi() 392 struct vgic_dist *dist = &vcpu->kvm->arch.vgic; in its_sync_lpi_pending_table() local 399 xa_for_each(&dist->lpi_xa, intid, irq) { in its_sync_lpi_pending_table() 1297 struct vgic_dist *dist = &kvm->arch.vgic; in vgic_its_invall() local [all …]
|
| /linux/fs/smb/client/compress/ |
| H A D | lz77.c | 78 static __always_inline void *lz77_write_match(void *dst, void **nib, u32 dist, u32 len) in lz77_write_match() argument 81 dist--; in lz77_write_match() 82 dist <<= 3; in lz77_write_match() 85 lz77_write16(dst, dist + len); in lz77_write_match() 90 dist |= 7; in lz77_write_match() 91 lz77_write16(dst, dist); in lz77_write_match() 153 u32 dist, len = 0; in lz77_compress() local 160 dist = srcp - wnd; in lz77_compress() 162 if (dist && dist < LZ77_MATCH_MAX_DIST) in lz77_compress() 192 dstp = lz77_write_match(dstp, &nib, dist, len); in lz77_compress()
|
| /linux/lib/zlib_inflate/ |
| H A D | inffast.c | 92 unsigned dist; /* match distance */ in inflate_fast() local 159 dist = (unsigned)(this.val); in inflate_fast() 169 dist += (unsigned)hold & ((1U << op) - 1); in inflate_fast() 171 if (dist > dmax) { in inflate_fast() 180 if (dist > op) { /* see if copy from window */ in inflate_fast() 181 op = dist - op; /* distance back in window */ in inflate_fast() 195 from = out - dist; /* rest from output */ in inflate_fast() 213 from = out - dist; /* rest from output */ in inflate_fast() 224 from = out - dist; /* rest from output */ in inflate_fast() 243 from = out - dist; /* copy direct from output */ in inflate_fast() [all …]
|
| H A D | inflate.c | 123 unsigned copy, dist; in zlib_updatewindow() local 135 dist = state->wsize - state->write; in zlib_updatewindow() 136 if (dist > copy) dist = copy; in zlib_updatewindow() 137 memcpy(state->window + state->write, strm->next_out - copy, dist); in zlib_updatewindow() 138 copy -= dist; in zlib_updatewindow() 145 state->write += dist; in zlib_updatewindow() 147 if (state->whave < state->wsize) state->whave += dist; in zlib_updatewindow()
|
| /linux/lib/ |
| H A D | cpu_rmap.c | 54 rmap->near[cpu].dist = CPU_RMAP_DIST_INF; in alloc_cpu_rmap() 95 const struct cpumask *mask, u16 dist) in cpu_rmap_copy_neigh() argument 100 if (rmap->near[cpu].dist > dist && in cpu_rmap_copy_neigh() 101 rmap->near[neigh].dist <= dist) { in cpu_rmap_copy_neigh() 103 rmap->near[cpu].dist = dist; in cpu_rmap_copy_neigh() 121 cpu, index, rmap->near[cpu].dist); in debug_print_rmap() 181 rmap->near[cpu].dist = CPU_RMAP_DIST_INF; in cpu_rmap_update() 193 rmap->near[cpu].dist = 0; in cpu_rmap_update()
|
| /linux/lib/zlib_deflate/ |
| H A D | deftree.c | 159 #define d_code(dist) \ argument 160 ((dist) < 256 ? dist_code[dist] : dist_code[256+((dist)>>7)]) 178 int dist; /* distance index */ in tr_static_init() local 200 dist = 0; in tr_static_init() 202 base_dist[code] = dist; in tr_static_init() 204 dist_code[dist++] = (uch)code; in tr_static_init() 207 Assert (dist == 256, "tr_static_init: dist != 256"); in tr_static_init() 208 dist >>= 7; /* from now on, all distances are divided by 128 */ in tr_static_init() 210 base_dist[code] = dist << 7; in tr_static_init() 212 dist_code[256 + dist++] = (uch)code; in tr_static_init() [all …]
|
| /linux/arch/hexagon/kernel/ |
| H A D | module.c | 102 int dist = (int)(value - (uint32_t)location); in apply_relocate_add() local 103 if ((dist < -0x00800000) || in apply_relocate_add() 104 (dist >= 0x00800000)) { in apply_relocate_add() 109 dist, value, (uint32_t)location, in apply_relocate_add() 116 *location |= 0x00003fff & dist; in apply_relocate_add() 117 *location |= 0x01ff0000 & (dist<<2); in apply_relocate_add()
|
| /linux/mm/ |
| H A D | numa_emulation.c | 500 int dist; in numa_emulation() local 502 if (get_option(&emu_cmdline, &dist) == 2) in numa_emulation() 505 dist = physi == physj ? in numa_emulation() 508 dist = phys_dist[physi * numa_dist_cnt + physj]; in numa_emulation() 510 numa_set_distance(i, j, dist); in numa_emulation() 516 u8 dist; in numa_emulation() local 530 dist = phys_dist[physi * numa_dist_cnt + physj]; in numa_emulation() 531 numa_set_distance(i, j, dist); in numa_emulation()
|
| /linux/net/ipv4/ |
| H A D | tcp_bic.c | 103 __u32 dist = (ca->last_max_cwnd - cwnd) in bictcp_update() local 106 if (dist > max_increment) in bictcp_update() 109 else if (dist <= 1U) in bictcp_update() 114 ca->cnt = cwnd / dist; in bictcp_update()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| H A D | ctxgm200.c | 50 u32 dist[TPC_MAX / 4] = {}; in gm200_grctx_generate_smid_config() local 57 dist[sm / 4] |= ((gpc << 4) | tpc) << ((sm % 4) * 8); in gm200_grctx_generate_smid_config() 62 nvkm_wr32(device, 0x405b60 + (i * 4), dist[i]); in gm200_grctx_generate_smid_config()
|
| H A D | ctxgp100.c | 110 u32 dist[TPC_MAX / 4] = {}, gpcs[16] = {}; in gp100_grctx_generate_smid_config() local 116 dist[sm / 4] |= ((gpc << 4) | tpc) << ((sm % 4) * 8); in gp100_grctx_generate_smid_config() 121 nvkm_wr32(device, 0x405b60 + (i * 4), dist[i]); in gp100_grctx_generate_smid_config()
|
| /linux/tools/testing/selftests/kvm/arm64/ |
| H A D | vgic_init.c | 148 struct vgic_region_attr dist; in subtest_dist_rdist() local 152 dist = VGIC_DEV_IS_V3(v->gic_dev_type) ? gic_v3_dist_region in subtest_dist_rdist() 156 kvm_has_device_attr(v->gic_fd, KVM_DEV_ARM_VGIC_GRP_ADDR, dist.attr); in subtest_dist_rdist() 165 addr = dist.alignment / 0x10; in subtest_dist_rdist() 167 dist.attr, &addr); in subtest_dist_rdist() 168 TEST_ASSERT(ret && errno == EINVAL, "GIC dist base not aligned"); in subtest_dist_rdist() 178 dist.attr, &addr); in subtest_dist_rdist() 179 TEST_ASSERT(ret && errno == E2BIG, "dist address beyond IPA limit"); in subtest_dist_rdist() 186 addr = max_phys_size - dist.alignment; in subtest_dist_rdist() 220 dist in subtest_dist_rdist() [all...] |
| /linux/drivers/ata/ |
| H A D | sata_sx4.c | 942 long dist; in pdc20621_get_from_dimm() local 960 dist = min(size, window_size - offset); in pdc20621_get_from_dimm() 961 memcpy_fromio(psource, dimm_mmio + offset / 4, dist); in pdc20621_get_from_dimm() 963 psource += dist; in pdc20621_get_from_dimm() 964 size -= dist; in pdc20621_get_from_dimm() 992 long dist; in pdc20621_put_to_dimm() local 1007 dist = min(size, window_size - offset); in pdc20621_put_to_dimm() 1008 memcpy_toio(dimm_mmio + offset / 4, psource, dist); in pdc20621_put_to_dimm() 1012 psource += dist; in pdc20621_put_to_dimm() 1013 size -= dist; in pdc20621_put_to_dimm()
|
| /linux/lib/xz/ |
| H A D | xz_dec_lzma2.c | 323 static inline uint32_t dict_get(const struct dictionary *dict, uint32_t dist) in dict_get() argument 325 size_t offset = dict->pos - dist - 1; in dict_get() 327 if (dist >= dict->pos) in dict_get() 349 static bool dict_repeat(struct dictionary *dict, uint32_t *len, uint32_t dist) in dict_repeat() argument 354 if (dist >= dict->full || dist >= dict->size) in dict_repeat() 360 back = dict->pos - dist - 1; in dict_repeat() 361 if (dist >= dict->pos) in dict_repeat()
|
| /linux/arch/arm64/kernel/ |
| H A D | hw_breakpoint.c | 756 u64 min_dist = -1, dist; in do_watchpoint() local 789 dist = get_distance_from_watchpoint(addr, val, &ctrl); in do_watchpoint() 790 if (dist < min_dist) { in do_watchpoint() 791 min_dist = dist; in do_watchpoint() 795 if (dist != 0) in do_watchpoint()
|
| /linux/drivers/irqchip/ |
| H A D | irq-gic.c | 1583 struct acpi_madt_generic_distributor *dist; in gic_validate_dist() local 1584 dist = (struct acpi_madt_generic_distributor *)header; in gic_validate_dist() 1586 return (dist->version == ape->driver_data && in gic_validate_dist() 1587 (dist->version != ACPI_MADT_GIC_VERSION_NONE || in gic_validate_dist() 1640 struct acpi_madt_generic_distributor *dist; in gic_v2_acpi_init() local 1658 dist = (struct acpi_madt_generic_distributor *)header; in gic_v2_acpi_init() 1659 gic->raw_dist_base = ioremap(dist->base_address, in gic_v2_acpi_init() 1678 gsi_domain_handle = irq_domain_alloc_fwnode(&dist->base_address); in gic_v2_acpi_init()
|