| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_csr.h | 21 #define HINIC_CSR_DMA_ATTR_ADDR(idx) \ argument 22 (HINIC_DMA_ATTR_BASE + (idx) * HINIC_DMA_ATTR_STRIDE) 26 #define HINIC_CSR_PPF_ELECTION_ADDR(idx) \ argument 27 (HINIC_ELECTION_BASE + (idx) * HINIC_PPF_ELECTION_STRIDE) 34 #define HINIC_CSR_API_CMD_CHAIN_HEAD_HI_ADDR(idx) \ argument 35 (HINIC_CSR_API_CMD_BASE + 0x0 + (idx) * HINIC_CSR_API_CMD_STRIDE) 37 #define HINIC_CSR_API_CMD_CHAIN_HEAD_LO_ADDR(idx) \ argument 38 (HINIC_CSR_API_CMD_BASE + 0x4 + (idx) * HINIC_CSR_API_CMD_STRIDE) 40 #define HINIC_CSR_API_CMD_STATUS_HI_ADDR(idx) \ argument 41 (HINIC_CSR_API_CMD_BASE + 0x8 + (idx) * HINIC_CSR_API_CMD_STRIDE) [all …]
|
| /linux/kernel/events/ |
| H A D | hw_breakpoint_test.c | 31 static struct perf_event *register_test_bp(int cpu, struct task_struct *tsk, int idx) in register_test_bp() argument 35 if (WARN_ON(idx < 0 || idx >= MAX_TEST_BREAKPOINTS)) in register_test_bp() 39 attr.bp_addr = (unsigned long)&break_vars[idx]; in register_test_bp() 124 int idx = 0; in test_one_cpu() local 126 fill_bp_slots(test, &idx, get_test_cpu(0), NULL, 0); in test_one_cpu() 127 TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx)); in test_one_cpu() 128 TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), NULL, idx)); in test_one_cpu() 133 int idx = 0; in test_many_cpus() local 138 bool do_continue = fill_bp_slots(test, &idx, cpu, NULL, 0); in test_many_cpus() 140 TEST_EXPECT_NOSPC(register_test_bp(cpu, NULL, idx)); in test_many_cpus() [all …]
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | uprobe_multi.c | 27 #define NAME(name, idx) PASTE(name, idx) argument 29 #define DEF(name, idx) int __attribute__((weak)) NAME(name, idx)(void) { return 0; } argument 30 #define CALL(name, idx) NAME(name, idx)(); argument 32 #define F(body, name, idx) body(name, idx) argument 34 #define F10(body, name, idx) \ argument 35 F(body, PASTE(name, idx), 40 F100(body,name,idx) global() argument 46 F1000(body,name,idx) global() argument 52 F10000(body,name,idx) global() argument [all...] |
| /linux/drivers/net/ethernet/microchip/lan966x/ |
| H A D | lan966x_ethtool.c | 300 uint idx = i * lan966x->num_stats; in lan966x_stats_update() local 308 lan966x_add_cnt(&lan966x->stats[idx++], in lan966x_stats_update() 362 u32 idx; in lan966x_get_eth_mac_stats() local 366 idx = port->chip_port * lan966x->num_stats; in lan966x_get_eth_mac_stats() 371 lan966x->stats[idx + SYS_COUNT_TX_UC] + in lan966x_get_eth_mac_stats() 372 lan966x->stats[idx + SYS_COUNT_TX_MC] + in lan966x_get_eth_mac_stats() 373 lan966x->stats[idx + SYS_COUNT_TX_BC] + in lan966x_get_eth_mac_stats() 374 lan966x->stats[idx + SYS_COUNT_TX_PMAC_UC] + in lan966x_get_eth_mac_stats() 375 lan966x->stats[idx + SYS_COUNT_TX_PMAC_MC] + in lan966x_get_eth_mac_stats() 376 lan966x->stats[idx + SYS_COUNT_TX_PMAC_BC]; in lan966x_get_eth_mac_stats() [all …]
|
| /linux/drivers/net/can/sja1000/ |
| H A D | sja1000_isa.c | 122 int idx = pdev->id; in sja1000_isa_probe() local 125 dev_dbg(&pdev->dev, "probing idx=%d: port=%#lx, mem=%#lx, irq=%d\n", in sja1000_isa_probe() 126 idx, port[idx], mem[idx], irq[idx]); in sja1000_isa_probe() 128 if (mem[idx]) { in sja1000_isa_probe() 129 if (!request_mem_region(mem[idx], iosize, DRV_NAME)) { in sja1000_isa_probe() 133 base = ioremap(mem[idx], iosize); in sja1000_isa_probe() 139 if (indirect[idx] > 0 || in sja1000_isa_probe() 140 (indirect[idx] == -1 && indirect[0] > 0)) in sja1000_isa_probe() 142 if (!request_region(port[idx], iosize, DRV_NAME)) { in sja1000_isa_probe() 155 dev->irq = irq[idx]; in sja1000_isa_probe() [all …]
|
| /linux/drivers/net/ethernet/microchip/vcap/ |
| H A D | vcap_api_debugfs_kunit.c | 41 int idx; in test_val_keyset() local 46 for (idx = 0; idx < kslist->cnt; idx++) { in test_val_keyset() 47 if (kslist->keysets[idx] == VCAP_KFS_ETAG) in test_val_keyset() 48 return kslist->keysets[idx]; in test_val_keyset() 49 if (kslist->keysets[idx] == in test_val_keyset() 51 return kslist->keysets[idx]; in test_val_keyset() 52 if (kslist->keysets[idx] == in test_val_keyset() 54 return kslist->keysets[idx]; in test_val_keyset() 55 if (kslist->keysets[idx] == in test_val_keyset() 57 return kslist->keysets[idx]; in test_val_keyset() [all …]
|
| /linux/drivers/net/can/cc770/ |
| H A D | cc770_isa.c | 167 int idx = pdev->id; in cc770_isa_probe() local 171 dev_dbg(&pdev->dev, "probing idx=%d: port=%#lx, mem=%#lx, irq=%d\n", in cc770_isa_probe() 172 idx, port[idx], mem[idx], irq[idx]); in cc770_isa_probe() 173 if (mem[idx]) { in cc770_isa_probe() 174 if (!request_mem_region(mem[idx], iosize, KBUILD_MODNAME)) { in cc770_isa_probe() 178 base = ioremap(mem[idx], iosize); in cc770_isa_probe() 184 if (indirect[idx] > 0 || in cc770_isa_probe() 185 (indirect[idx] == -1 && indirect[0] > 0)) in cc770_isa_probe() 187 if (!request_region(port[idx], iosize, KBUILD_MODNAME)) { in cc770_isa_probe() 200 dev->irq = irq[idx]; in cc770_isa_probe() [all …]
|
| /linux/kernel/sched/ |
| H A D | cpudeadline.c | 26 static void cpudl_heapify_down(struct cpudl *cp, int idx) in cpudl_heapify_down() argument 30 int orig_cpu = cp->elements[idx].cpu; in cpudl_heapify_down() 31 u64 orig_dl = cp->elements[idx].dl; in cpudl_heapify_down() 33 if (left_child(idx) >= cp->size) in cpudl_heapify_down() 40 l = left_child(idx); in cpudl_heapify_down() 41 r = right_child(idx); in cpudl_heapify_down() 42 largest = idx; in cpudl_heapify_down() 54 if (largest == idx) in cpudl_heapify_down() 57 /* pull largest child onto idx */ in cpudl_heapify_down() 58 cp->elements[idx].cpu = cp->elements[largest].cpu; in cpudl_heapify_down() [all …]
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | sparsebit.c | 31 * sparsebit_is_set(s, idx) 32 * sparsebit_is_clear(s, idx) 38 * sparsebit_set(s, idx) 39 * sparsebit_clear(s, idx) 40 * sparsebit_set_num(s, idx, num); 41 * sparsebit_clear_num(s, idx, num); 46 * sparsebit_idx_t idx; 48 * idx = sparsebit_first_set(s); 51 * idx = sparsebit_next_set(s, idx); 52 * } while (idx != 0); [all …]
|
| /linux/arch/x86/um/ |
| H A D | tls_32.c | 34 int idx = info->entry_number - host_gdt_entry_tls_min; in do_set_thread_area() local 37 data->arch_data.tls[idx] = *info; in do_set_thread_area() 38 data->arch_data.sync |= BIT(idx); in do_set_thread_area() 63 int idx; in get_free_idx() local 65 for (idx = 0; idx < GDT_ENTRY_TLS_ENTRIES; idx++) in get_free_idx() 66 if (!t->arch.tls_array[idx].present) in get_free_idx() 67 return idx + GDT_ENTRY_TLS_MIN; in get_free_idx() 89 int idx; in load_TLS() local 91 for (idx = GDT_ENTRY_TLS_MIN; idx < GDT_ENTRY_TLS_MAX; idx++) { in load_TLS() 93 &to->thread.arch.tls_array[idx - GDT_ENTRY_TLS_MIN]; in load_TLS() [all …]
|
| /linux/kernel/bpf/ |
| H A D | mprog.c | 80 /* The bpf_mprog_{replace,delete}() operate on exact idx position with the 82 * case of front idx is -1, in case of back idx is bpf_mprog_total(entry). 86 * idx + before: 88 * Insert P4 before P3: idx for old array is 1, idx for new array is 2, 89 * hence we adjust target idx for the new array, so that memmove copies 90 * P1 and P2 to the new entry, and we insert P4 into idx 2. Inserting 91 * before P1 would have old idx -1 and new idx 0. 97 * idx + after: 99 * Insert P4 after P2: idx for old array is 2, idx for new array is 2. 101 * into idx 2. Inserting after P3 would have both old/new idx at 4 aka [all …]
|
| /linux/drivers/net/wireless/mediatek/mt7601u/ |
| H A D | main.c | 49 unsigned int idx = 0; in mt7601u_add_interface() local 50 unsigned int wcid = GROUP_WCID(idx); in mt7601u_add_interface() 55 * - shift vif idx in mt7601u_add_interface() 57 mvif->idx = idx; in mt7601u_add_interface() 65 mvif->group_wcid.idx = wcid; in mt7601u_add_interface() 76 unsigned int wcid = mvif->group_wcid.idx; in mt7601u_remove_interface() 189 int i, idx = 0; in mt76_wcid_alloc() local 192 idx = ffs(~dev->wcid_mask[i]); in mt76_wcid_alloc() 193 if (!idx) in mt76_wcid_alloc() 196 idx--; in mt76_wcid_alloc() [all …]
|
| /linux/mm/ |
| H A D | hugetlb_cgroup.c | 36 __hugetlb_cgroup_counter_from_cgroup(struct hugetlb_cgroup *h_cg, int idx, in __hugetlb_cgroup_counter_from_cgroup() argument 40 return &h_cg->rsvd_hugepage[idx]; in __hugetlb_cgroup_counter_from_cgroup() 41 return &h_cg->hugepage[idx]; in __hugetlb_cgroup_counter_from_cgroup() 45 hugetlb_cgroup_counter_from_cgroup(struct hugetlb_cgroup *h_cg, int idx) in hugetlb_cgroup_counter_from_cgroup() argument 47 return __hugetlb_cgroup_counter_from_cgroup(h_cg, idx, false); in hugetlb_cgroup_counter_from_cgroup() 51 hugetlb_cgroup_counter_from_cgroup_rsvd(struct hugetlb_cgroup *h_cg, int idx) in hugetlb_cgroup_counter_from_cgroup_rsvd() argument 53 return __hugetlb_cgroup_counter_from_cgroup(h_cg, idx, true); in hugetlb_cgroup_counter_from_cgroup_rsvd() 94 int idx; in hugetlb_cgroup_init() local 96 for (idx = 0; idx < HUGE_MAX_HSTATE; idx++) { in hugetlb_cgroup_init() 103 parent_h_cgroup, idx); in hugetlb_cgroup_init() [all …]
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | evergreen_cs.c | 757 unsigned idx) in evergreen_cs_track_validate_texture() argument 765 texdw[0] = radeon_get_ib_value(p, idx + 0); in evergreen_cs_track_validate_texture() 766 texdw[1] = radeon_get_ib_value(p, idx + 1); in evergreen_cs_track_validate_texture() 767 texdw[2] = radeon_get_ib_value(p, idx + 2); in evergreen_cs_track_validate_texture() 768 texdw[3] = radeon_get_ib_value(p, idx + 3); in evergreen_cs_track_validate_texture() 769 texdw[4] = radeon_get_ib_value(p, idx + 4); in evergreen_cs_track_validate_texture() 770 texdw[5] = radeon_get_ib_value(p, idx + 5); in evergreen_cs_track_validate_texture() 771 texdw[6] = radeon_get_ib_value(p, idx + 6); in evergreen_cs_track_validate_texture() 772 texdw[7] = radeon_get_ib_value(p, idx + 7); in evergreen_cs_track_validate_texture() 1051 unsigned idx, unsigned reg) in evergreen_packet0_check() argument [all …]
|
| H A D | r600_cs.c | 842 r = radeon_cs_packet_parse(p, &wait_reg_mem, p->idx); in r600_cs_common_vline_parse() 853 wait_reg_mem_info = radeon_get_ib_value(p, wait_reg_mem.idx + 1); in r600_cs_common_vline_parse() 869 if ((radeon_get_ib_value(p, wait_reg_mem.idx + 2) << 2) != vline_status[0]) { in r600_cs_common_vline_parse() 874 if (radeon_get_ib_value(p, wait_reg_mem.idx + 5) != RADEON_VLINE_STAT) { in r600_cs_common_vline_parse() 880 r = radeon_cs_packet_parse(p, &p3reloc, p->idx + wait_reg_mem.count + 2); in r600_cs_common_vline_parse() 884 h_idx = p->idx - 2; in r600_cs_common_vline_parse() 885 p->idx += wait_reg_mem.count + 2; in r600_cs_common_vline_parse() 886 p->idx += p3reloc.count + 2; in r600_cs_common_vline_parse() 923 unsigned idx, unsigned reg) in r600_packet0_check() argument 932 idx, reg); in r600_packet0_check() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb/ |
| H A D | fpga_defs.h | 215 #define MAC_REG_ADDR(idx, reg) (MAC_REG_BASE + (idx) * 128 + (reg)) argument 217 #define MAC_REG_IDLO(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_LO) argument 218 #define MAC_REG_IDHI(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_HI) argument 219 #define MAC_REG_CSR(idx) MAC_REG_ADDR(idx, A_GMAC_CSR) argument 220 #define MAC_REG_IFS(idx) MAC_REG_ADDR(idx, A_GMAC_IFS) argument 221 #define MAC_REG_LARGEFRAMELENGTH(idx) MAC_REG_ADDR(idx, A_GMAC_JUMBO_FRAME_LEN) argument 222 #define MAC_REG_LINKDLY(idx) MAC_REG_ADDR(idx, A_GMAC_LNK_DLY) argument 223 #define MAC_REG_PAUSETIME(idx) MAC_REG_ADDR(idx, A_GMAC_PAUSETIME) argument 224 #define MAC_REG_CASTLO(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_LO) argument 225 #define MAC_REG_MCASTHI(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_HI) argument [all …]
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | hyperv_features.c | 25 uint32_t idx; member 47 GUEST_ASSERT(msr->idx); in guest_msr() 50 vector = wrmsr_safe(msr->idx, msr->write_val); in guest_msr() 52 if (!vector && (!msr->write || !is_write_only_msr(msr->idx))) in guest_msr() 53 vector = rdmsr_safe(msr->idx, &msr_val); in guest_msr() 58 msr->write ? "WR" : "RD", msr->idx, ex_str(vector)); in guest_msr() 62 msr->write ? "WR" : "RD", msr->idx, ex_str(vector)); in guest_msr() 64 if (vector || is_write_only_msr(msr->idx)) in guest_msr() 70 msr->idx, msr->write_val, msr_val); in guest_msr() 73 if (msr->idx == HV_X64_MSR_TSC_INVARIANT_CONTROL) { in guest_msr() [all …]
|
| /linux/tools/perf/scripts/perl/Perf-Trace-Util/lib/Perf/Trace/ |
| H A D | Core.pm | 39 foreach my $idx (sort {$a <=> $b} keys %trace_flags) { 40 if (!$value && !$idx) { 45 if ($idx && ($value & $idx) == $idx) { 49 $string .= "$trace_flags{$idx}"; 51 $value &= ~$idx; 69 foreach my $idx (sort {$a <=> $b} keys %{$flag_fields{$event_name}{$field_name}{"values"}}) { 70 if (!$value && !$idx) { 71 $string .= "$flag_fields{$event_name}{$field_name}{'values'}{$idx}"; 74 if ($idx && ($value & $idx) == $idx) { 78 $string .= "$flag_fields{$event_name}{$field_name}{'values'}{$idx}"; [all …]
|
| /linux/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/ |
| H A D | Core.py | 35 for idx in sorted(flag_fields[event_name][field_name]['values']): 36 if not value and not idx: 37 string += flag_fields[event_name][field_name]['values'][idx] 39 if idx and (value & idx) == idx: 42 string += flag_fields[event_name][field_name]['values'][idx] 44 value &= ~idx 52 for idx in sorted(symbolic_fields[event_name][field_name]['values']): 53 if not value and not idx: 54 string = symbolic_fields[event_name][field_name]['values'][idx] 56 if (value == idx): [all …]
|
| /linux/include/asm-generic/ |
| H A D | fixmap.h | 26 * 'index to address' translation. If anyone tries to use the idx 30 static __always_inline unsigned long fix_to_virt(const unsigned int idx) in fix_to_virt() argument 32 BUILD_BUG_ON(idx >= __end_of_fixed_addresses); in fix_to_virt() 33 return __fix_to_virt(idx); in fix_to_virt() 64 #define set_fixmap(idx, phys) \ argument 65 __set_fixmap(idx, phys, FIXMAP_PAGE_NORMAL) 69 #define clear_fixmap(idx) \ argument 70 __set_fixmap(idx, 0, FIXMAP_PAGE_CLEAR) 74 #define __set_fixmap_offset(idx, phys, flags) \ argument 77 __set_fixmap(idx, phys, flags); \ [all …]
|
| /linux/include/trace/events/ |
| H A D | cpuhp.h | 14 int idx, 17 TP_ARGS(cpu, target, idx, fun), 22 __field( int, idx ) 29 __entry->idx = idx; 34 __entry->cpu, __entry->target, __entry->idx, __entry->fun) 41 int idx, 45 TP_ARGS(cpu, target, idx, fun, node), 50 __field( int, idx ) 57 __entry->idx = idx; 62 __entry->cpu, __entry->target, __entry->idx, __entry->fun) [all …]
|
| /linux/drivers/net/ethernet/microchip/sparx5/ |
| H A D | sparx5_mirror.c | 24 static u64 sparx5_mirror_port_get(struct sparx5 *sparx5, u32 idx) in sparx5_mirror_port_get() argument 28 val = spx5_rd(sparx5, ANA_AC_PROBE_PORT_CFG(idx)); in sparx5_mirror_port_get() 31 val |= (u64)spx5_rd(sparx5, ANA_AC_PROBE_PORT_CFG1(idx)) << 32; in sparx5_mirror_port_get() 37 static void sparx5_mirror_port_add(struct sparx5 *sparx5, u32 idx, u32 portno) in sparx5_mirror_port_add() argument 45 return spx5_rmw(val, val, sparx5, ANA_AC_PROBE_PORT_CFG(idx)); in sparx5_mirror_port_add() 47 return spx5_rmw(val, val, sparx5, ANA_AC_PROBE_PORT_CFG1(idx)); in sparx5_mirror_port_add() 51 static void sparx5_mirror_port_del(struct sparx5 *sparx5, u32 idx, u32 portno) in sparx5_mirror_port_del() argument 59 return spx5_rmw(0, val, sparx5, ANA_AC_PROBE_PORT_CFG(idx)); in sparx5_mirror_port_del() 61 return spx5_rmw(0, val, sparx5, ANA_AC_PROBE_PORT_CFG1(idx)); in sparx5_mirror_port_del() 65 static bool sparx5_mirror_contains(struct sparx5 *sparx5, u32 idx, u32 portno) in sparx5_mirror_contains() argument [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/clk/ |
| H A D | gk104.c | 145 read_clk(struct gk104_clk *clk, int idx) in read_clk() argument 148 u32 sctl = nvkm_rd32(device, 0x137250 + (idx * 4)); in read_clk() 151 if (idx < 7) { in read_clk() 153 if (ssel & (1 << idx)) { in read_clk() 154 sclk = read_pll(clk, 0x137000 + (idx * 0x20)); in read_clk() 157 sclk = read_div(clk, idx, 0x137160, 0x1371d0); in read_clk() 161 u32 ssrc = nvkm_rd32(device, 0x137160 + (idx * 0x04)); in read_clk() 163 sclk = read_div(clk, idx, 0x137160, 0x1371d0); in read_clk() 172 sclk = read_div(clk, idx, 0x137160, 0x1371d0); in read_clk() 223 calc_div(struct gk104_clk *clk, int idx, u32 ref, u32 freq, u32 *ddiv) in calc_div() argument [all …]
|
| /linux/tools/testing/radix-tree/ |
| H A D | main.c | 17 long idx; in __gang_check() local 22 for (idx = -down; idx < up; idx++) in __gang_check() 23 item_insert(&tree, middle + idx); in __gang_check() 26 for (idx = -down; idx < up; idx++) in __gang_check() 27 item_check_present(&tree, middle + idx); in __gang_check() 126 …ix_tree_root *tree, unsigned long start, unsigned long end, unsigned long *idx, int count, int fro… in check_copied_tags() argument 133 if (idx[i] < start || idx[i] > end) { in check_copied_tags() 134 if (item_tag_get(tree, idx[i], totag)) { in check_copied_tags() 136 end, idx[i], item_tag_get(tree, idx[i], in check_copied_tags() 138 item_tag_get(tree, idx[i], totag)); in check_copied_tags() [all …]
|
| /linux/drivers/media/pci/pt3/ |
| H A D | pt3_dma.c | 16 static u32 get_dma_base(int idx) in get_dma_base() argument 20 i = (idx == 1 || idx == 2) ? 3 - idx : idx; in get_dma_base() 61 static u8 *next_unit(struct pt3_adapter *adap, int *idx, int *ofs) in next_unit() argument 66 (*idx)++; in next_unit() 67 if (*idx == adap->num_bufs) in next_unit() 68 *idx = 0; in next_unit() 70 return &adap->buffer[*idx].data[*ofs]; in next_unit() 75 int idx, ofs; in pt3_proc_dma() local 77 idx = adap->buf_idx; in pt3_proc_dma() 80 if (adap->buffer[idx].data[ofs] == PT3_BUF_CANARY) in pt3_proc_dma() [all …]
|