Home
last modified time | relevance | path

Searched refs:ix (Results 1 – 25 of 82) sorted by relevance

1234

/linux/arch/sh/kernel/cpu/sh2a/
H A Dfpu.c96 unsigned int ix, iy; in denormal_mulf() local
100 ix = hx & 0x7fffffff; in denormal_mulf()
102 if (iy < 0x00800000 || ix == 0) in denormal_mulf()
106 ix &= 0x007fffff; in denormal_mulf()
108 m = (unsigned long long)ix * iy; in denormal_mulf()
116 ix = ((int) (m >> (w - 23)) & 0x007fffff) | (exp << 23); in denormal_mulf()
118 ix = (int) (m >> (w - 22 - exp)) & 0x007fffff; in denormal_mulf()
120 ix = 0; in denormal_mulf()
122 ix |= (hx ^ hy) & 0x80000000; in denormal_mulf()
123 return ix; in denormal_mulf()
[all …]
/linux/io_uring/
H A Dxattr.c27 struct io_xattr *ix = io_kiocb_to_cmd(req, struct io_xattr); in io_xattr_cleanup() local
29 dismiss_delayed_filename(&ix->filename); in io_xattr_cleanup()
30 kfree(ix->ctx.kname); in io_xattr_cleanup()
31 kvfree(ix->ctx.kvalue); in io_xattr_cleanup()
45 struct io_xattr *ix = io_kiocb_to_cmd(req, struct io_xattr); in __io_getxattr_prep() local
49 INIT_DELAYED_FILENAME(&ix->filename); in __io_getxattr_prep()
50 ix->ctx.kvalue = NULL; in __io_getxattr_prep()
52 ix->ctx.value = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_getxattr_prep()
53 ix->ctx.size = READ_ONCE(sqe->len); in __io_getxattr_prep()
54 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_getxattr_prep()
[all …]
/linux/arch/mips/math-emu/
H A Dsp_sqrt.c14 int ix, s, q, m, t, i; in ieee754sp_sqrt() local
56 ix = x.bits; in ieee754sp_sqrt()
59 m = (ix >> 23); in ieee754sp_sqrt()
61 for (i = 0; (ix & 0x00800000) == 0; i++) in ieee754sp_sqrt()
62 ix <<= 1; in ieee754sp_sqrt()
66 ix = (ix & 0x007fffff) | 0x00800000; in ieee754sp_sqrt()
68 ix += ix; in ieee754sp_sqrt()
72 ix += ix; in ieee754sp_sqrt()
79 if (t <= ix) { in ieee754sp_sqrt()
81 ix -= t; in ieee754sp_sqrt()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Drx_res.c356 int ix; in mlx5e_rx_res_channels_init() local
368 for (ix = 0; ix < res->max_nch; ix++) { in mlx5e_rx_res_channels_init()
369 err = mlx5e_rqt_init_direct(&res->channels[ix].direct_rqt, in mlx5e_rx_res_channels_init()
374 err, ix); in mlx5e_rx_res_channels_init()
379 for (ix = 0; ix < res->max_nch; ix++) { in mlx5e_rx_res_channels_init()
381 mlx5e_rqt_get_rqtn(&res->channels[ix].direct_rqt), in mlx5e_rx_res_channels_init()
388 err = mlx5e_tir_init(&res->channels[ix].direct_tir, builder, res->mdev, true); in mlx5e_rx_res_channels_init()
391 err, ix); in mlx5e_rx_res_channels_init()
401 while (--ix >= 0) in mlx5e_rx_res_channels_init()
402 mlx5e_tir_destroy(&res->channels[ix].direct_tir); in mlx5e_rx_res_channels_init()
[all …]
H A Dchannels.c14 static struct mlx5e_channel *mlx5e_channels_get(struct mlx5e_channels *chs, unsigned int ix) in mlx5e_channels_get() argument
16 WARN_ON_ONCE(ix >= mlx5e_channels_get_num(chs)); in mlx5e_channels_get()
17 return chs->c[ix]; in mlx5e_channels_get()
20 bool mlx5e_channels_is_xsk(struct mlx5e_channels *chs, unsigned int ix) in mlx5e_channels_is_xsk() argument
22 struct mlx5e_channel *c = mlx5e_channels_get(chs, ix); in mlx5e_channels_is_xsk()
27 void mlx5e_channels_get_regular_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn, in mlx5e_channels_get_regular_rqn() argument
30 struct mlx5e_channel *c = mlx5e_channels_get(chs, ix); in mlx5e_channels_get_regular_rqn()
37 void mlx5e_channels_get_xsk_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn, in mlx5e_channels_get_xsk_rqn() argument
40 struct mlx5e_channel *c = mlx5e_channels_get(chs, ix); in mlx5e_channels_get_xsk_rqn()
H A Dfs_tt_redirect.c148 int ix = 0; in fs_udp_create_groups() local
178 MLX5_SET_CFG(in, start_flow_index, ix); in fs_udp_create_groups()
179 ix += MLX5E_FS_UDP_GROUP1_SIZE; in fs_udp_create_groups()
180 MLX5_SET_CFG(in, end_flow_index, ix - 1); in fs_udp_create_groups()
188 MLX5_SET_CFG(in, start_flow_index, ix); in fs_udp_create_groups()
189 ix += MLX5E_FS_UDP_GROUP2_SIZE; in fs_udp_create_groups()
190 MLX5_SET_CFG(in, end_flow_index, ix - 1); in fs_udp_create_groups()
430 int ix = 0; in fs_any_create_groups() local
449 MLX5_SET_CFG(in, start_flow_index, ix); in fs_any_create_groups()
450 ix += MLX5E_FS_ANY_GROUP1_SIZE; in fs_any_create_groups()
[all …]
H A Drqt.c118 unsigned int ix = i; in mlx5e_calc_indir_rqns() local
121 ix = mlx5e_bits_invert(ix, ilog2(indir->actual_table_size)); in mlx5e_calc_indir_rqns()
123 ix = indir->table[ix]; in mlx5e_calc_indir_rqns()
125 if (WARN_ON(ix >= num_rqns)) in mlx5e_calc_indir_rqns()
130 rss_rqns[i] = rqns[ix]; in mlx5e_calc_indir_rqns()
132 rss_vhca_ids[i] = vhca_ids[ix]; in mlx5e_calc_indir_rqns()
H A Dchannels.h12 bool mlx5e_channels_is_xsk(struct mlx5e_channels *chs, unsigned int ix);
13 void mlx5e_channels_get_regular_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn,
15 void mlx5e_channels_get_xsk_rqn(struct mlx5e_channels *chs, unsigned int ix, u32 *rqn,
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dpool.c45 static int mlx5e_xsk_add_pool(struct mlx5e_xsk *xsk, struct xsk_buff_pool *pool, u16 ix) in mlx5e_xsk_add_pool() argument
53 xsk->pools[ix] = pool; in mlx5e_xsk_add_pool()
57 static void mlx5e_xsk_remove_pool(struct mlx5e_xsk *xsk, u16 ix) in mlx5e_xsk_remove_pool() argument
59 xsk->pools[ix] = NULL; in mlx5e_xsk_remove_pool()
78 struct xsk_buff_pool *pool, u16 ix) in mlx5e_xsk_enable_locked() argument
85 if (unlikely(mlx5e_xsk_get_pool(&priv->channels.params, &priv->xsk, ix))) in mlx5e_xsk_enable_locked()
91 err = mlx5e_xsk_map_pool(mlx5_sd_ch_ix_get_dev(priv->mdev, ix), pool); in mlx5e_xsk_enable_locked()
95 err = mlx5e_xsk_add_pool(&priv->xsk, pool, ix); in mlx5e_xsk_enable_locked()
122 c = priv->channels.c[ix]; in mlx5e_xsk_enable_locked()
135 mlx5e_rx_res_xsk_update(priv->rx_res, &priv->channels, ix, true); in mlx5e_xsk_enable_locked()
[all …]
H A Dpool.h10 struct mlx5e_xsk *xsk, u16 ix) in mlx5e_xsk_get_pool() argument
15 if (unlikely(ix >= params->num_channels)) in mlx5e_xsk_get_pool()
18 return xsk->pools[ix]; in mlx5e_xsk_get_pool()
H A Drx.h11 int mlx5e_xsk_alloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix);
12 int mlx5e_xsk_alloc_rx_wqes_batched(struct mlx5e_rq *rq, u16 ix, int wqe_bulk);
13 int mlx5e_xsk_alloc_rx_wqes(struct mlx5e_rq *rq, u16 ix, int wqe_bulk);
/linux/drivers/net/ethernet/mellanox/mlx5/core/lib/
H A Dfs_ttc.c467 int ix = *next_ix; in mlx5_create_ttc_table_ipsec_groups() local
474 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_ttc_table_ipsec_groups()
475 ix += groups->group_size[ttc->num_groups]; in mlx5_create_ttc_table_ipsec_groups()
476 MLX5_SET_CFG(in, end_flow_index, ix - 1); in mlx5_create_ttc_table_ipsec_groups()
492 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_ttc_table_ipsec_groups()
493 ix += groups->group_size[ttc->num_groups]; in mlx5_create_ttc_table_ipsec_groups()
494 MLX5_SET_CFG(in, end_flow_index, ix - 1); in mlx5_create_ttc_table_ipsec_groups()
513 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_create_ttc_table_ipsec_groups()
514 ix += groups->group_size[ttc->num_groups]; in mlx5_create_ttc_table_ipsec_groups()
515 MLX5_SET_CFG(in, end_flow_index, ix - 1); in mlx5_create_ttc_table_ipsec_groups()
[all …]
H A Dipsec_fs_roce.c347 int ix = 0; in ipsec_fs_roce_tx_mpv_create_group_rules() local
355 MLX5_SET_CFG(in, start_flow_index, ix); in ipsec_fs_roce_tx_mpv_create_group_rules()
356 ix += MLX5_TX_ROCE_GROUP_SIZE; in ipsec_fs_roce_tx_mpv_create_group_rules()
357 MLX5_SET_CFG(in, end_flow_index, ix - 1); in ipsec_fs_roce_tx_mpv_create_group_rules()
452 int ix = 0; in ipsec_fs_roce_rx_mpv_create() local
510 MLX5_SET_CFG(in, start_flow_index, ix); in ipsec_fs_roce_rx_mpv_create()
511 ix += 1; in ipsec_fs_roce_rx_mpv_create()
512 MLX5_SET_CFG(in, end_flow_index, ix - 1); in ipsec_fs_roce_rx_mpv_create()
609 int ix = 0; in mlx5_ipsec_fs_roce_tx_create() local
639 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5_ipsec_fs_roce_tx_create()
[all …]
H A Dmpfs.h57 int ix = MLX5_L2_ADDR_HASH(mac); \
61 hlist_for_each_entry(ptr, &(hash)[ix], node.hlist) \
72 int ix = MLX5_L2_ADDR_HASH(mac); \
78 hlist_add_head(&ptr->node.hlist, &(hash)[ix]);\
/linux/arch/s390/lib/
H A Dspinlock.c83 int ix; in arch_spin_lock_setup() local
86 for (ix = 0; ix < 4; ix++, node++) { in arch_spin_lock_setup()
89 (ix << _Q_TAIL_IDX_OFFSET); in arch_spin_lock_setup()
138 int ix, cpu; in arch_spin_decode_tail() local
140 ix = (lock & _Q_TAIL_IDX_MASK) >> _Q_TAIL_IDX_OFFSET; in arch_spin_decode_tail()
142 return per_cpu_ptr(&spin_wait[ix], cpu - 1); in arch_spin_decode_tail()
159 int lockval, ix, node_id, tail_id, old, new, owner, count; in arch_spin_lock_queued() local
161 ix = get_lowcore()->spinlock_index++; in arch_spin_lock_queued()
164 node = this_cpu_ptr(&spin_wait[ix]); in arch_spin_lock_queued()
/linux/drivers/media/dvb-frontends/
H A Dmxl692.c196 u32 ix, div_size; in mxl692_checksum() local
203 for (ix = 0; ix < div_size; ix++) in mxl692_checksum()
204 cur_cksum += be32_to_cpu(buf[ix]); in mxl692_checksum()
215 u32 ix, temp; in mxl692_validate_fw_header() local
235 for (ix = 16; ix < buf_len; ix++) in mxl692_validate_fw_header()
236 temp_cksum += buffer[ix]; in mxl692_validate_fw_header()
251 u32 ix = 0, total_len = 0, addr = 0, chunk_len = 0, prevchunk_len = 0; in mxl692_write_fw_block() local
255 ix = *index; in mxl692_write_fw_block()
257 if (buffer[ix] == 0x53) { in mxl692_write_fw_block()
258 total_len = buffer[ix + 1] << 16 | buffer[ix + 2] << 8 | buffer[ix + 3]; in mxl692_write_fw_block()
[all …]
/linux/drivers/input/misc/
H A Dyealink.c284 int ix, len; in yealink_set_ringtone() local
300 ix = 0; in yealink_set_ringtone()
301 while (size != ix) { in yealink_set_ringtone()
302 len = size - ix; in yealink_set_ringtone()
306 p->offset = cpu_to_be16(ix); in yealink_set_ringtone()
307 memcpy(p->data, &buf[ix], len); in yealink_set_ringtone()
309 ix += len; in yealink_set_ringtone()
319 int i, ix, len; in yealink_do_idle_tasks() local
321 ix = yld->stat_ix; in yealink_do_idle_tasks()
329 if (ix >= sizeof(yld->master)) { in yealink_do_idle_tasks()
[all …]
/linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/
H A Dsmu_helper.h155 PHM_GET_FIELD(cgs_read_ind_register(device, port, ix##reg), \
159 PHM_GET_FIELD(cgs_read_ind_register(device, port, ix##reg), \
167 cgs_write_ind_register(device, port, ix##reg, \
168 PHM_SET_FIELD(cgs_read_ind_register(device, port, ix##reg), \
172 cgs_write_ind_register(device, port, ix##reg, \
173 PHM_SET_FIELD(cgs_read_ind_register(device, port, ix##reg), \
181 PHM_WAIT_INDIRECT_REGISTER_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask)
192 PHM_WAIT_INDIRECT_REGISTER_UNEQUAL_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask)
206 PHM_WAIT_VFPF_INDIRECT_REGISTER_UNEQUAL_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask)
220 PHM_WAIT_VFPF_INDIRECT_REGISTER_GIVEN_INDEX(hwmgr, port, ix##reg, value, mask)
/linux/fs/netfs/
H A Drolling_buffer.c122 int nr, ix, to; in rolling_buffer_load_from_ra() local
132 ix = fq->vec.nr; in rolling_buffer_load_from_ra()
133 to = ix + nr; in rolling_buffer_load_from_ra()
135 for (; ix < to; ix++) { in rolling_buffer_load_from_ra()
136 struct folio *folio = folioq_folio(fq, ix); in rolling_buffer_load_from_ra()
139 fq->orders[ix] = order; in rolling_buffer_load_from_ra()
H A Diterator.c113 unsigned int nbv = iter->nr_segs, ix = 0, nsegs = 0; in netfs_limit_bvec() local
122 while (n && ix < nbv && skip) { in netfs_limit_bvec()
123 len = bvecs[ix].bv_len; in netfs_limit_bvec()
128 ix++; in netfs_limit_bvec()
131 while (n && ix < nbv) { in netfs_limit_bvec()
132 len = min3(n, bvecs[ix].bv_len - skip, max_size); in netfs_limit_bvec()
135 ix++; in netfs_limit_bvec()
/linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.c103 unsigned int ix; in mlx5_fpga_conn_post_recv() local
115 ix = conn->qp.rq.pc & (conn->qp.rq.size - 1); in mlx5_fpga_conn_post_recv()
116 data = mlx5_wq_cyc_get_wqe(&conn->qp.wq.rq, ix); in mlx5_fpga_conn_post_recv()
122 conn->qp.rq.bufs[ix] = buf; in mlx5_fpga_conn_post_recv()
146 unsigned int ix, sgi; in mlx5_fpga_conn_post_send() local
149 ix = conn->qp.sq.pc & (conn->qp.sq.size - 1); in mlx5_fpga_conn_post_send()
151 ctrl = mlx5_wq_cyc_get_wqe(&conn->qp.wq.sq, ix); in mlx5_fpga_conn_post_send()
171 conn->qp.sq.bufs[ix] = buf; in mlx5_fpga_conn_post_send()
254 int ix, err; in mlx5_fpga_conn_rq_cqe() local
256 ix = be16_to_cpu(cqe->wqe_counter) & (conn->qp.rq.size - 1); in mlx5_fpga_conn_rq_cqe()
[all …]
/linux/net/rxrpc/
H A Dcall_event.c64 unsigned int ix = req->seq & RXRPC_TXQ_MASK; in rxrpc_retransmit_data() local
65 struct rxrpc_txbuf *txb = tq->bufs[ix]; in rxrpc_retransmit_data()
67 _enter("%x,%x,%x,%x", tq->qbase, req->seq, ix, txb->debug_id); in rxrpc_retransmit_data()
111 unsigned int ix = __ffs(lost); in rxrpc_resend() local
112 struct rxrpc_txbuf *txb = tq->bufs[ix]; in rxrpc_resend()
114 __clear_bit(ix, &lost); in rxrpc_resend()
118 req.seq = tq->qbase + ix; in rxrpc_resend()
231 int ix; in rxrpc_transmit_fresh_data() local
234 ix = seq & RXRPC_TXQ_MASK; in rxrpc_transmit_fresh_data()
235 if (!ix) { in rxrpc_transmit_fresh_data()
[all …]
H A Dinput.c207 int ix) in rxrpc_add_data_rtt_sample() argument
209 ktime_t xmit_ts = ktime_add_us(tq->xmit_ts_base, tq->segment_xmit_ts[ix]); in rxrpc_add_data_rtt_sample()
214 __clear_bit(ix, &tq->rtt_samples); /* Prevent repeat RTT sample */ in rxrpc_add_data_rtt_sample()
251 unsigned int ix = seq - call->tx_qbase; in rxrpc_rotate_tx_window() local
253 _debug("tq=%x seq=%x i=%d f=%x", tq->qbase, seq, ix, tq->bufs[ix]->flags); in rxrpc_rotate_tx_window()
254 if (tq->bufs[ix]->flags & RXRPC_LAST_PACKET) { in rxrpc_rotate_tx_window()
259 if (summary->acked_serial == tq->segment_serial[ix] && in rxrpc_rotate_tx_window()
260 test_bit(ix, &tq->rtt_samples)) in rxrpc_rotate_tx_window()
261 rxrpc_add_data_rtt_sample(call, summary, tq, ix); in rxrpc_rotate_tx_window()
263 if (ix == tq->nr_reported_acks) { in rxrpc_rotate_tx_window()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_fs.c117 int ix = mlx5e_hash_l2(addr); in mlx5e_add_l2_to_hash() local
120 hlist_for_each_entry(hn, &hash[ix], hlist) in mlx5e_add_l2_to_hash()
138 hlist_add_head(&hn->hlist, &hash[ix]); in mlx5e_add_l2_to_hash()
1036 int ix = 0; in mlx5e_create_l2_table_groups() local
1057 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5e_create_l2_table_groups()
1058 ix += MLX5E_L2_GROUP1_SIZE; in mlx5e_create_l2_table_groups()
1059 MLX5_SET_CFG(in, end_flow_index, ix - 1); in mlx5e_create_l2_table_groups()
1068 MLX5_SET_CFG(in, start_flow_index, ix); in mlx5e_create_l2_table_groups()
1069 ix += MLX5E_L2_GROUP2_SIZE; in mlx5e_create_l2_table_groups()
1070 MLX5_SET_CFG(in, end_flow_index, ix - 1); in mlx5e_create_l2_table_groups()
[all …]
/linux/drivers/net/wireless/marvell/mwifiex/
H A Dutil.c801 int ix; in mwifiex_hist_data_reset() local
805 for (ix = 0; ix < MWIFIEX_MAX_AC_RX_RATES; ix++) in mwifiex_hist_data_reset()
806 atomic_set(&phist_data->rx_rate[ix], 0); in mwifiex_hist_data_reset()
807 for (ix = 0; ix < MWIFIEX_MAX_SNR; ix++) in mwifiex_hist_data_reset()
808 atomic_set(&phist_data->snr[ix], 0); in mwifiex_hist_data_reset()
809 for (ix = 0; ix < MWIFIEX_MAX_NOISE_FLR; ix++) in mwifiex_hist_data_reset()
810 atomic_set(&phist_data->noise_flr[ix], 0); in mwifiex_hist_data_reset()
811 for (ix = 0; ix < MWIFIEX_MAX_SIG_STRENGTH; ix++) in mwifiex_hist_data_reset()
812 atomic_set(&phist_data->sig_str[ix], 0); in mwifiex_hist_data_reset()

1234