Home
last modified time | relevance | path

Searched +full:1 +full:q (Results 1 – 25 of 1205) sorted by relevance

12345678910>>...49

/linux/drivers/media/common/videobuf2/
H A Dvideobuf2-core.c36 #define PLANE_INDEX_MASK (BIT_MASK(PLANE_INDEX_BITS) - 1)
38 #define BUFFER_INDEX_MASK (MAX_BUFFER_INDEX - 1)
47 #define dprintk(q, level, fmt, arg...) \ argument
50 pr_info("[%s] %s: " fmt, (q)->name, __func__, \
103 #define log_qop(q, op) \ argument
104 dprintk(q, 2, "call_qop(%s)%s\n", #op, \
105 (q)->ops->op ? "" : " (nop)")
107 #define call_qop(q, op, args...) \ argument
111 log_qop(q, op); \
112 err = (q)->ops->op ? (q)->ops->op(args) : 0; \
[all …]
/linux/net/sched/
H A Dsch_choke.c46 #define CHOKE_MAX_QUEUE (128*1024 - 1)
69 unsigned int tab_mask; /* size - 1 */
75 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument
77 return (q->tail - q->head) & q->tab_mask; in choke_len()
81 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument
83 return q->flags & TC_RED_ECN; in use_ecn()
87 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument
89 return q->flags & TC_RED_HARDDROP; in use_harddrop()
93 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument
96 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes()
[all …]
H A Dsch_dualpi2.c14 * - Supports ECN L4S-identifier (IP.ECN==0b*1)
41 /* 32b enable to support flows with windows up to ~8.6 * 1e9 packets
52 * netlink-provided values can use at most 31b, i.e. be at most (2^23)-1
53 * (~4MHz) as those are given in 1/256th. This enable to tune alpha/beta to
56 #define ALPHA_BETA_MAX ((1U << 31) - 1)
61 * scale_alpha_beta(1) will not round down to 0.
88 bool step_in_packets; /* Step thresh in packets (1) or time (0) */
101 bool drop_early; /* Drop at enqueue (1) instead of dequeue (0) */
102 bool drop_overload; /* Drop (1) on overload, or overflow (0) */
103 bool split_gso; /* Split aggregated skb (1) or leave as is (0) */
[all …]
H A Dsch_netem.c35 Sources: [1] Mark Carson, Darrin Santay, "NIST Net - A Linux-based
60 [1] NetemCLG Home http://netgroup.uniroma2.it/NetemCLG
63 in the Netem module in the Linux kernel", available in [1]
125 TX_IN_GAP_PERIOD = 1,
132 GOOD_STATE = 1,
145 u32 a4; /* p14 for 4-states or 1-k for GE */
200 rho = (u64)state->rho + 1; in get_crandom()
201 answer = (value * ((1ull<<32) - rho) + state->last * rho) >> 32; in get_crandom()
210 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument
212 struct clgstate *clg = &q->clg; in loss_4state()
[all …]
H A Dsch_sfq.c68 - max mtu to 2^18-1;
76 #define SFQ_MAX_FLOWS (0x10000 - SFQ_MAX_DEPTH - 1) /* max number of flows */
80 /* This type should contain at least SFQ_MAX_DEPTH + 1 + SFQ_MAX_FLOWS values */
85 * Small indexes [0 ... SFQ_MAX_FLOWS - 1] are 'pointers' to slots[] array
126 struct sfq_head dep[SFQ_MAX_DEPTH + 1];
129 * dep[1] : list of flows with 1 packet
143 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument
146 return &q->slots[val].dep; in sfq_dep_head()
147 return &q->dep[val - SFQ_MAX_FLOWS]; in sfq_dep_head()
150 static unsigned int sfq_hash(const struct sfq_sched_data *q, in sfq_hash() argument
[all …]
H A Dsch_cake.c101 * @rec_inv_sqrt: reciprocal value of sqrt(count) >> 1
261 CAKE_FLAG_AUTORATE_INGRESS = BIT(1),
300 static u16 quantum_div[CAKE_QUEUES + 1] = {0};
306 1, 1, 1, 1, 1, 1, 1, 1,
316 2, 0, 1, 2, 4, 2, 2, 2,
317 1, 2, 1, 2, 1, 2, 1, 2,
327 0, 1, 0, 0, 2, 0, 0, 0,
328 1, 0, 0, 0, 0, 0, 0, 0,
338 0, 1, 0, 0, 2, 0, 0, 0,
339 1, 0, 0, 0, 0, 0, 0, 0,
[all …]
H A Dsch_skbprio.c40 static u16 calc_new_high_prio(const struct skbprio_sched_data *q) in calc_new_high_prio() argument
44 for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) { in calc_new_high_prio()
45 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_high_prio()
53 static u16 calc_new_low_prio(const struct skbprio_sched_data *q) in calc_new_low_prio() argument
57 for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) { in calc_new_low_prio()
58 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_low_prio()
62 /* SKB queue is empty, return SKBPRIO_MAX_PRIORITY - 1 in calc_new_low_prio()
65 return SKBPRIO_MAX_PRIORITY - 1; in calc_new_low_prio()
71 const unsigned int max_priority = SKBPRIO_MAX_PRIORITY - 1; in skbprio_enqueue()
72 struct skbprio_sched_data *q = qdisc_priv(sch); in skbprio_enqueue() local
[all …]
H A Dsch_fq.c66 * If packets have monotically increasing time_to_send, they are placed in O(1)
75 unsigned long age; /* (jiffies | 1UL) when flow was emptied, for gc */
79 /* Following field is only used for q->internal,
80 * because q->internal is not hashed in fq_root[]
93 struct rb_node rate_node; /* anchor in q->delayed tree */
109 #define FQ_PRIO2BAND_CRUMB_SIZE ((TC_PRIO_MAX + 1) >> 2)
176 f->age = jiffies | 1UL; in fq_flow_set_detached()
181 return !!(f->age & 1UL); in fq_flow_is_detached()
197 static void fq_flow_add_tail(struct fq_sched_data *q, struct fq_flow *flow, in fq_flow_add_tail() argument
200 struct fq_perband_flows *pband = &q->band_flows[flow->band]; in fq_flow_add_tail()
[all …]
H A Dsch_multiq.c32 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_classify() local
35 struct tcf_proto *fl = rcu_dereference_bh(q->filter_list); in multiq_classify()
53 if (band >= q->bands) in multiq_classify()
54 return q->queues[0]; in multiq_classify()
56 return q->queues[band]; in multiq_classify()
79 sch->q.qlen++; in multiq_enqueue()
89 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_dequeue() local
94 for (band = 0; band < q->bands; band++) { in multiq_dequeue()
96 q->curband++; in multiq_dequeue()
97 if (q in multiq_dequeue()
120 struct multiq_sched_data *q = qdisc_priv(sch); multiq_peek() local
151 struct multiq_sched_data *q = qdisc_priv(sch); multiq_reset() local
162 struct multiq_sched_data *q = qdisc_priv(sch); multiq_destroy() local
174 struct multiq_sched_data *q = qdisc_priv(sch); multiq_tune() local
238 struct multiq_sched_data *q = qdisc_priv(sch); multiq_init() local
263 struct multiq_sched_data *q = qdisc_priv(sch); multiq_dump() local
283 struct multiq_sched_data *q = qdisc_priv(sch); multiq_graft() local
296 struct multiq_sched_data *q = qdisc_priv(sch); multiq_leaf() local
304 struct multiq_sched_data *q = qdisc_priv(sch); multiq_find() local
319 multiq_unbind(struct Qdisc * q,unsigned long cl) multiq_unbind() argument
326 struct multiq_sched_data *q = qdisc_priv(sch); multiq_dump_class() local
336 struct multiq_sched_data *q = qdisc_priv(sch); multiq_dump_class_stats() local
349 struct multiq_sched_data *q = qdisc_priv(sch); multiq_walk() local
364 struct multiq_sched_data *q = qdisc_priv(sch); multiq_tcf_block() local
[all...]
H A Dsch_qfq.c25 [1] Paolo Valente,
52 aggregates. See [1] and [2] for a full description.
73 Lmax=1<<MTU_SHIFT, w_min = 1 .
80 inv_w = (1<<FRAC_BITS)/w_i
98 * in the range [1, 2^MAX_WSHIFT], and we try to map each aggregate i to the
108 #define QFQ_MAX_WEIGHT (1<<QFQ_MAX_WSHIFT) /* see qfq_slot_insert */
112 #define ONE_FP (1UL << FRAC_BITS)
116 #define QFQ_MAX_LMAX (1UL << QFQ_MTU_SHIFT)
189 struct qfq_group groups[QFQ_MAX_INDEX + 1]; /* The groups. */
212 struct qfq_sched *q = qdisc_priv(sch); in qfq_find_class() local
[all …]
H A Dsch_ets.c21 * has no traffic then band 1, etc.
54 u8 prio2band[TC_PRIO_MAX + 1];
58 static const struct nla_policy ets_policy[TCA_ETS_MAX + 1] = {
65 static const struct nla_policy ets_priomap_policy[TCA_ETS_MAX + 1] = {
69 static const struct nla_policy ets_quanta_policy[TCA_ETS_MAX + 1] = {
73 static const struct nla_policy ets_class_policy[TCA_ETS_MAX + 1] = {
97 struct ets_sched *q = qdisc_priv(sch); in ets_class_from_arg() local
99 if (arg == 0 || arg > q->nbands) in ets_class_from_arg()
101 return &q->classes[arg - 1]; in ets_class_from_arg()
106 struct ets_sched *q = qdisc_priv(sch); in ets_class_id() local
[all …]
/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_rx.c22 /* Per packet headroom in non-XDP mode. Present only for 1-frag packets. */
33 FUN_XDP_FLUSH_REDIR = 1,
50 static void cache_offer(struct funeth_rxq *q, const struct funeth_rxbuf *buf) in cache_offer() argument
52 struct funeth_rx_cache *c = &q->cache; in cache_offer()
58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_offer()
67 static bool cache_get(struct funeth_rxq *q, struct funeth_rxbuf *rb) in cache_get() argument
69 struct funeth_rx_cache *c = &q->cache; in cache_get()
77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, in cache_get()
88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_get()
98 static int funeth_alloc_page(struct funeth_rxq *q, struct funeth_rxbuf *rb, in funeth_alloc_page() argument
[all …]
H A Dfuneth_tx.c48 dma_unmap_single(dev, addr[-1], data_len, DMA_TO_DEVICE); in fun_map_pkt()
56 static void *txq_end(const struct funeth_txq *q) in txq_end() argument
58 return (void *)q->hw_wb; in txq_end()
64 static unsigned int txq_to_end(const struct funeth_txq *q, void *p) in txq_to_end() argument
66 return txq_end(q) - p; in txq_to_end()
78 static struct fun_dataop_gl *fun_write_gl(const struct funeth_txq *q, in fun_write_gl() argument
90 i < ngle && txq_to_end(q, gle); i++, gle++) in fun_write_gl()
93 if (txq_to_end(q, gle) == 0) { in fun_write_gl()
94 gle = (struct fun_dataop_gl *)q->desc; in fun_write_gl()
107 static struct sk_buff *fun_tls_tx(struct sk_buff *skb, struct funeth_txq *q, in fun_tls_tx() argument
[all …]
/linux/drivers/net/wireless/broadcom/b43/
H A Dpio.c24 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument
37 cookie = (((u16)q->index + 1) << 12); in generate_cookie()
49 struct b43_pio_txqueue *q = NULL; in parse_cookie() local
54 q = pio->tx_queue_AC_BK; in parse_cookie()
57 q = pio->tx_queue_AC_BE; in parse_cookie()
60 q = pio->tx_queue_AC_VI; in parse_cookie()
63 q = pio->tx_queue_AC_VO; in parse_cookie()
66 q = pio->tx_queue_mcast; in parse_cookie()
69 if (B43_WARN_ON(!q)) in parse_cookie()
72 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb3/
H A Dsge.c66 #define FL1_PG_ORDER (PAGE_SIZE > 8192 ? 0 : 1)
94 TXQ_RUNNING = 1 << 0, /* fetch engine is running */
95 TXQ_LAST_PKT_DB = 1 << 1, /* last packet rang the doorbell */
141 dma_addr_t addr[MAX_SKB_FRAGS + 1];
148 * desc = 1 + (flits - 2) / (WR_FLITS - 1).
154 #if SGE_NUM_GENBITS == 1
155 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
160 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
165 # error "SGE_NUM_GENBITS must be 1 or 2"
169 static inline struct sge_qset *rspq_to_qset(const struct sge_rspq *q) in rspq_to_qset() argument
[all …]
/linux/sound/core/seq/oss/
H A Dseq_oss_readq.c20 //#define SNDRV_SEQ_OSS_MAX_TIMEOUT (unsigned long)(-1)
35 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local
37 q = kzalloc(sizeof(*q), GFP_KERNEL); in snd_seq_oss_readq_new()
38 if (!q) in snd_seq_oss_readq_new()
41 q->q = kcalloc(maxlen, sizeof(union evrec), GFP_KERNEL); in snd_seq_oss_readq_new()
42 if (!q->q) { in snd_seq_oss_readq_new()
43 kfree(q); in snd_seq_oss_readq_new()
62 snd_seq_oss_readq_delete(struct seq_oss_readq * q) snd_seq_oss_readq_delete() argument
74 snd_seq_oss_readq_clear(struct seq_oss_readq * q) snd_seq_oss_readq_clear() argument
89 snd_seq_oss_readq_puts(struct seq_oss_readq * q,int dev,unsigned char * data,int len) snd_seq_oss_readq_puts() argument
123 snd_seq_oss_readq_sysex(struct seq_oss_readq * q,int dev,struct snd_seq_event * ev) snd_seq_oss_readq_sysex() argument
141 snd_seq_oss_readq_put_event(struct seq_oss_readq * q,union evrec * ev) snd_seq_oss_readq_put_event() argument
169 snd_seq_oss_readq_pick(struct seq_oss_readq * q,union evrec * rec) snd_seq_oss_readq_pick() argument
181 snd_seq_oss_readq_wait(struct seq_oss_readq * q) snd_seq_oss_readq_wait() argument
193 snd_seq_oss_readq_free(struct seq_oss_readq * q) snd_seq_oss_readq_free() argument
206 snd_seq_oss_readq_poll(struct seq_oss_readq * q,struct file * file,poll_table * wait) snd_seq_oss_readq_poll() argument
216 snd_seq_oss_readq_put_timestamp(struct seq_oss_readq * q,unsigned long curt,int seq_mode) snd_seq_oss_readq_put_timestamp() argument
244 snd_seq_oss_readq_info_read(struct seq_oss_readq * q,struct snd_info_buffer * buf) snd_seq_oss_readq_info_read() argument
[all...]
/linux/drivers/s390/cio/
H A Dqdio_main.c44 " lgr 1,%[schid]\n" in do_siga_sync()
52 : CC_CLOBBER_LIST("0", "1", "2", "3")); in do_siga_sync()
63 " lgr 1,%[schid]\n" in do_siga_input()
69 : CC_CLOBBER_LIST("0", "1", "2")); in do_siga_input()
92 " lgr 1,%[schid]\n" in do_siga_output()
100 : CC_CLOBBER_LIST("0", "1", "2", "3")); in do_siga_output()
107 * @q: queue to manipulate
116 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument
119 int tmp_count = count, tmp_start = start, nr = q->nr; in qdio_do_eqbs()
122 qperf_inc(q, eqbs); in qdio_do_eqbs()
[all …]
/linux/lib/crypto/
H A Dgf128mul.c25 1. distributions of this source code include the above copyright
57 #define gf128mul_dat(q) { \ argument
58 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\
59 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\
60 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\
61 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\
62 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\
63 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\
64 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\
65 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\
[all …]
/linux/drivers/gpu/drm/xe/
H A Dxe_guc_submit.c51 exec_queue_to_guc(struct xe_exec_queue *q) in exec_queue_to_guc() argument
53 return &q->gt->uc.guc; in exec_queue_to_guc()
61 #define EXEC_QUEUE_STATE_REGISTERED (1 << 0)
62 #define EXEC_QUEUE_STATE_ENABLED (1 << 1)
63 #define EXEC_QUEUE_STATE_PENDING_ENABLE (1 << 2)
64 #define EXEC_QUEUE_STATE_PENDING_DISABLE (1 << 3)
65 #define EXEC_QUEUE_STATE_DESTROYED (1 << 4)
66 #define EXEC_QUEUE_STATE_SUSPENDED (1 << 5)
67 #define EXEC_QUEUE_STATE_RESET (1 << 6)
68 #define EXEC_QUEUE_STATE_KILLED (1 << 7)
[all …]
H A Dxe_exec_queue.c33 XE_EXEC_QUEUE_TIMESLICE = 1,
38 static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q,
41 static void __xe_exec_queue_free(struct xe_exec_queue *q) in __xe_exec_queue_free() argument
46 if (q->tlb_inval[i].dep_scheduler) in __xe_exec_queue_free()
47 xe_dep_scheduler_fini(q->tlb_inval[i].dep_scheduler); in __xe_exec_queue_free()
49 if (xe_exec_queue_uses_pxp(q)) in __xe_exec_queue_free()
50 xe_pxp_exec_queue_remove(gt_to_xe(q->gt)->pxp, q); in __xe_exec_queue_free()
51 if (q->vm) in __xe_exec_queue_free()
52 xe_vm_put(q->vm); in __xe_exec_queue_free()
54 if (q->xef) in __xe_exec_queue_free()
[all …]
/linux/sound/core/seq/
H A Dseq_queue.c50 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument
57 queue_list[i] = q; in queue_list_add()
58 q->queue = i; in queue_list_add()
63 return -1; in queue_list_add()
68 struct snd_seq_queue *q; in queue_list_remove() local
71 q = queue_list[id]; in queue_list_remove()
72 if (q) { in queue_list_remove()
73 guard(spinlock)(&q->owner_lock); in queue_list_remove()
74 if (q->owner == client) { in queue_list_remove()
76 q in queue_list_remove()
90 struct snd_seq_queue *q; queue_new() local
121 queue_delete(struct snd_seq_queue * q) queue_delete() argument
162 struct snd_seq_queue *q; snd_seq_queue_alloc() local
181 struct snd_seq_queue *q; snd_seq_queue_delete() local
197 struct snd_seq_queue *q; queueptr() local
212 struct snd_seq_queue *q; snd_seq_queue_find_name() local
230 snd_seq_check_queue(struct snd_seq_queue * q,int atomic,int hop) snd_seq_check_queue() argument
289 struct snd_seq_queue *q; snd_seq_enqueue_event() local
340 check_access(struct snd_seq_queue * q,int client) check_access() argument
348 queue_access_lock(struct snd_seq_queue * q,int client) queue_access_lock() argument
360 queue_access_unlock(struct snd_seq_queue * q) queue_access_unlock() argument
369 struct snd_seq_queue *q = queueptr(queueid); snd_seq_queue_check_access() local
387 struct snd_seq_queue *q = queueptr(queueid); snd_seq_queue_set_owner() local
453 struct snd_seq_queue *q = queueptr(queueid); snd_seq_queue_timer_set_tempo() local
517 struct snd_seq_queue *q; snd_seq_queue_is_used() local
538 struct snd_seq_queue *q; snd_seq_queue_client_leave() local
571 struct snd_seq_queue *q; snd_seq_queue_remove_cells() local
592 queue_broadcast_event(struct snd_seq_queue * q,struct snd_seq_event * ev,int atomic,int hop) queue_broadcast_event() argument
615 snd_seq_queue_process_event(struct snd_seq_queue * q,struct snd_seq_event * ev,int atomic,int hop) snd_seq_queue_process_event() argument
670 struct snd_seq_queue *q; snd_seq_control_queue() local
700 struct snd_seq_queue *q; snd_seq_info_queues_read() local
[all...]
/linux/drivers/infiniband/sw/rxe/
H A Drxe_queue.h18 * - The driver indices are always masked off to q->index_mask
26 * - By passing the type in the parameter list separate from q
83 void rxe_queue_reset(struct rxe_queue *q);
88 int rxe_queue_resize(struct rxe_queue *q, unsigned int *num_elem_p,
95 static inline u32 queue_next_index(struct rxe_queue *q, int index) in queue_next_index() argument
97 return (index + 1) & q->index_mask; in queue_next_index()
100 static inline u32 queue_get_producer(const struct rxe_queue *q, in queue_get_producer() argument
108 prod = smp_load_acquire(&q->buf->producer_index); in queue_get_producer()
112 prod = q->index; in queue_get_producer()
116 prod = q->buf->producer_index; in queue_get_producer()
[all …]
/linux/drivers/spi/
H A Dspi-fsl-qspi.c59 #define QUADSPI_MCR_SWRSTHD_MASK BIT(1)
101 #define QUADSPI_SR_IP_ACC_MASK BIT(1)
125 #define QUADSPI_LCKER_UNLOCK BIT(1)
134 #define LUT_CMD 1
158 #define LUT_PAD(x) (fls(x) - 1)
175 #define QUADSPI_QUIRK_4X_INT_CLK BIT(1)
277 static inline int needs_swap_endian(struct fsl_qspi *q) in needs_swap_endian() argument
279 return q->devtype_data->quirks & QUADSPI_QUIRK_SWAP_ENDIAN; in needs_swap_endian()
282 static inline int needs_4x_clock(struct fsl_qspi *q) in needs_4x_clock() argument
284 return q->devtype_data->quirks & QUADSPI_QUIRK_4X_INT_CLK; in needs_4x_clock()
[all …]
/linux/drivers/net/ethernet/pensando/ionic/
H A Dionic_txrx.c15 static dma_addr_t ionic_tx_map_single(struct ionic_queue *q,
18 static dma_addr_t ionic_tx_map_frag(struct ionic_queue *q,
22 static void ionic_tx_desc_unmap_bufs(struct ionic_queue *q,
25 static void ionic_tx_clean(struct ionic_queue *q,
30 static inline void ionic_txq_post(struct ionic_queue *q, bool ring_dbell) in ionic_txq_post() argument
36 ionic_q_post(q, ring_dbell); in ionic_txq_post()
39 static inline void ionic_rxq_post(struct ionic_queue *q, bool ring_dbell) in ionic_rxq_post() argument
41 ionic_q_post(q, ring_dbell); in ionic_rxq_post()
44 bool ionic_txq_poke_doorbell(struct ionic_queue *q) in ionic_txq_poke_doorbell() argument
50 netdev = q->lif->netdev; in ionic_txq_poke_doorbell()
[all …]
/linux/drivers/net/ethernet/mellanox/mlxsw/
H A Dpci.c101 struct mlxsw_pci_queue *q; member
174 static char *__mlxsw_pci_queue_elem_get(struct mlxsw_pci_queue *q, in __mlxsw_pci_queue_elem_get() argument
177 return q->mem_item.buf + (elem_size * elem_index); in __mlxsw_pci_queue_elem_get()
181 mlxsw_pci_queue_elem_info_get(struct mlxsw_pci_queue *q, int elem_index) in mlxsw_pci_queue_elem_info_get() argument
183 return &q->elem_info[elem_index]; in mlxsw_pci_queue_elem_info_get()
187 mlxsw_pci_queue_elem_info_producer_get(struct mlxsw_pci_queue *q) in mlxsw_pci_queue_elem_info_producer_get() argument
189 int index = q->producer_counter & (q->count - 1); in mlxsw_pci_queue_elem_info_producer_get()
191 if ((u16) (q->producer_counter - q->consumer_counter) == q->count) in mlxsw_pci_queue_elem_info_producer_get()
193 return mlxsw_pci_queue_elem_info_get(q, index); in mlxsw_pci_queue_elem_info_producer_get()
197 mlxsw_pci_queue_elem_info_consumer_get(struct mlxsw_pci_queue *q) in mlxsw_pci_queue_elem_info_consumer_get() argument
[all …]

12345678910>>...49