| /linux/sound/core/seq/oss/ |
| H A D | seq_oss_readq.c | 48 q->qlen = 0; in snd_seq_oss_readq_new() 76 if (q->qlen) { in snd_seq_oss_readq_clear() 77 q->qlen = 0; in snd_seq_oss_readq_clear() 144 if (q->qlen >= q->maxlen - 1) in snd_seq_oss_readq_put_event() 149 q->qlen++; in snd_seq_oss_readq_put_event() 165 if (q->qlen == 0) in snd_seq_oss_readq_pick() 178 (q->qlen > 0 || q->head == q->tail), in snd_seq_oss_readq_wait() 189 if (q->qlen > 0) { in snd_seq_oss_readq_free() 191 q->qlen--; in snd_seq_oss_readq_free() 203 return q->qlen; in snd_seq_oss_readq_poll() [all …]
|
| /linux/net/sched/ |
| H A D | sch_sfq.c | 97 sfq_index qlen; /* number of skbs in skblist */ member 200 int qlen = slot->qlen; in sfq_link() local 202 p = qlen + SFQ_MAX_FLOWS; in sfq_link() 203 n = q->dep[qlen].next; in sfq_link() 208 q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */ in sfq_link() 228 d = q->slots[x].qlen--; in sfq_dec() 241 d = ++q->slots[x].qlen; in sfq_inc() 303 sch->q.qlen--; in sfq_drop() 346 sfq_index x, qlen; in sfq_enqueue() local 420 if (slot->qlen >= q->maxdepth) { in sfq_enqueue() [all …]
|
| H A D | sch_choke.c | 126 --sch->q.qlen; in choke_drop_by_idx() 220 q->vars.qavg = red_calc_qavg(p, &q->vars, sch->q.qlen); in choke_enqueue() 267 if (sch->q.qlen < q->limit) { in choke_enqueue() 270 ++sch->q.qlen; in choke_enqueue() 297 --sch->q.qlen; in choke_dequeue() 380 unsigned int oqlen = sch->q.qlen, tail = 0; in choke_change() 395 --sch->q.qlen; in choke_change() 398 qdisc_tree_reduce_backlog(sch, oqlen - sch->q.qlen, dropped); in choke_change()
|
| H A D | sch_drr.c | 267 __u32 qlen = qdisc_qlen_sum(cl->qdisc); in drr_dump_class_stats() local 272 if (qlen) in drr_dump_class_stats() 277 gnet_stats_copy_queue(d, cl_q->cpu_qstats, &cl_q->qstats, qlen) < 0) in drr_dump_class_stats() 369 sch->q.qlen++; in drr_enqueue() 396 if (cl->qdisc->q.qlen == 0) in drr_dequeue() 402 sch->q.qlen--; in drr_dequeue() 437 if (cl->qdisc->q.qlen) in drr_reset_qdisc()
|
| H A D | sch_skbprio.c | 84 if (sch->q.qlen < READ_ONCE(sch->limit)) { in skbprio_enqueue() 96 sch->q.qlen++; in skbprio_enqueue() 148 sch->q.qlen--; in skbprio_dequeue() 254 q->qstats[cl - 1].qlen) < 0) in skbprio_dump_class_stats()
|
| H A D | sch_fq.c | 86 int qlen; /* number of packets in flow queue */ member 302 * 2) FQ qlen == 0 OR 319 if (sch->q.qlen != 0) { in fq_fastpath_check() 331 if (q->internal.qlen >= 8) in fq_fastpath_check() 500 sch->q.qlen--; in fq_dequeue_skb() 579 if (unlikely(f->qlen >= q->flow_plimit)) { in fq_enqueue() 594 if (f->qlen == 0) in fq_enqueue() 598 f->qlen++; in fq_enqueue() 603 sch->q.qlen++; in fq_enqueue() 662 if (!sch->q.qlen) in fq_dequeue() [all...] |
| H A D | sch_fq_pie.c | 49 u32 qlen; member 188 sch->q.qlen++; in fq_pie_qdisc_enqueue() 194 sel_flow->qlen = 0; in fq_pie_qdisc_enqueue() 197 sel_flow->qlen++; in fq_pie_qdisc_enqueue() 266 sch->q.qlen--; in fq_pie_qdisc_dequeue() 279 flow->qlen--; in fq_pie_qdisc_dequeue() 367 while (sch->q.qlen > sch->limit) { in fq_pie_change()
|
| H A D | sch_fq_codel.c | 181 sch->q.qlen -= i; in fq_codel_drop() 218 if (++sch->q.qlen <= sch->limit && !memory_limited) in fq_codel_enqueue() 222 prev_qlen = sch->q.qlen; in fq_codel_enqueue() 233 prev_qlen -= sch->q.qlen; in fq_codel_enqueue() 268 sch->q.qlen--; in dequeue_func() 443 while (sch->q.qlen > sch->limit || in fq_codel_change() 676 qs.qlen++; in fq_codel_dump_class_stats() 684 if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0) in fq_codel_dump_class_stats()
|
| /linux/sound/core/seq/ |
| H A D | seq_midi_event.c | 50 int qlen; member 141 dev->qlen = 0; in reset_encode() 193 dev->qlen = status_event[dev->type].qlen; in snd_midi_event_encode_byte() 195 if (dev->qlen > 0) { in snd_midi_event_encode_byte() 199 dev->qlen--; in snd_midi_event_encode_byte() 203 dev->qlen = status_event[dev->type].qlen - 1; in snd_midi_event_encode_byte() 207 if (dev->qlen == 0) { in snd_midi_event_encode_byte() 312 int qlen; in snd_midi_event_decode() local 323 qlen = status_event[type].qlen + 1; in snd_midi_event_decode() 328 qlen = status_event[type].qlen; in snd_midi_event_decode() [all …]
|
| /linux/include/net/ |
| H A D | sch_generic.h | 62 __u32 qlen; 193 return !READ_ONCE(qdisc->q.qlen); 539 return q->q.qlen; in qdisc_skb_cb() 544 __u32 qlen = q->qstats.qlen; in qdisc_lock() 549 qlen += per_cpu_ptr(q->cpu_qstats, i)->qlen; in qdisc_root() 551 qlen += q->q.qlen; in qdisc_root() 554 return qlen; in qdisc_root_bh() 69 __u32 qlen; global() member 524 __u32 qlen = q->qstats.qlen; qdisc_qlen_sum() local 961 __u32 qlen = qdisc_qlen_sum(sch); qdisc_qstats_copy() local 966 qdisc_qstats_qlen_backlog(struct Qdisc * sch,__u32 * qlen,__u32 * backlog) qdisc_qstats_qlen_backlog() argument 978 __u32 qlen, backlog; qdisc_purge_queue() local [all...] |
| /linux/drivers/md/ |
| H A D | dm-ps-queue-length.c | 40 atomic_t qlen; /* the number of in-flight I/Os */ member 101 DMEMIT("%d ", atomic_read(&pi->qlen)); in ql_status() 153 atomic_set(&pi->qlen, 0); in ql_add_path() 204 (atomic_read(&pi->qlen) < atomic_read(&best->qlen))) in ql_select_path() 207 if (!atomic_read(&best->qlen)) in ql_select_path() 228 atomic_inc(&pi->qlen); in ql_start_io() 238 atomic_dec(&pi->qlen); in ql_end_io()
|
| /linux/drivers/usb/gadget/legacy/ |
| H A D | gmidi.c | 53 static unsigned int qlen = 32; variable 54 module_param(qlen, uint, S_IRUGO); 55 MODULE_PARM_DESC(qlen, "USB read and write request queue length"); 157 midi_opts->qlen = qlen; in midi_bind()
|
| H A D | printer.c | 51 static unsigned qlen = 10; variable 52 module_param(qlen, uint, S_IRUGO|S_IWUSR); 53 MODULE_PARM_DESC(qlen, "The number of 8k buffers to use per endpoint"); 55 #define QLEN qlen
|
| H A D | zero.c | 66 .qlen = GZERO_QLEN, 263 module_param_named(qlen, gzero_options.qlen, uint, S_IRUGO|S_IWUSR); 264 MODULE_PARM_DESC(qlen, "depth of loopback queue"); 322 lb_opts->qlen = gzero_options.qlen; in zero_bind()
|
| /linux/net/core/ |
| H A D | gen_stats.c | 348 qstats->qlen += qcpu->qlen; in gnet_stats_add_queue_cpu() 363 qstats->qlen += q->qlen; in gnet_stats_add_queue() 389 struct gnet_stats_queue *q, __u32 qlen) in gnet_stats_copy_queue() argument 394 qstats.qlen = qlen; in gnet_stats_copy_queue() 398 d->tc_stats.qlen = qstats.qlen; in gnet_stats_copy_queue()
|
| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cptvf_main.c | 99 pqinfo->qlen = 0; in free_pending_queues() 103 static int alloc_pending_queues(struct pending_qinfo *pqinfo, u32 qlen, in alloc_pending_queues() argument 111 pqinfo->qlen = qlen; in alloc_pending_queues() 114 queue->head = kzalloc_objs(*queue->head, qlen); in alloc_pending_queues() 136 static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues) in init_pending_queues() argument 144 ret = alloc_pending_queues(&cptvf->pqinfo, qlen, nr_queues); in init_pending_queues() 203 u32 qlen) in alloc_command_queues() argument 213 cptvf->qsize = min(qlen, cqinfo->qchunksize) * in alloc_command_queues() 216 q_size = qlen * cqinfo->cmd_size; in alloc_command_queues() 277 static int init_command_queues(struct cpt_vf *cptvf, u32 qlen) in init_command_queues() argument [all …]
|
| /linux/drivers/usb/gadget/function/ |
| H A D | f_loopback.c | 34 unsigned qlen; member 321 for (i = 0; i < loop->qlen && result == 0; i++) { in alloc_requests() 439 loop->qlen = lb_opts->qlen; in loopback_alloc() 440 if (!loop->qlen) in loopback_alloc() 441 loop->qlen = 32; in loopback_alloc() 477 result = sprintf(page, "%d\n", opts->qlen); in f_lb_opts_qlen_show() 500 opts->qlen = num; in f_lb_opts_qlen_store() 507 CONFIGFS_ATTR(f_lb_opts_, qlen); 577 lb_opts->qlen = GZERO_QLEN; in loopback_alloc_instance()
|
| /linux/drivers/nfc/st-nci/ |
| H A D | ndlc.c | 94 if (ndlc->send_q.qlen) in llt_ndlc_send_queue() 96 ndlc->send_q.qlen, ndlc->ack_pending_q.qlen); in llt_ndlc_send_queue() 98 while (ndlc->send_q.qlen) { in llt_ndlc_send_queue() 153 if (ndlc->rcv_q.qlen) in llt_ndlc_rcv_queue() 154 pr_debug("rcvQlen=%d\n", ndlc->rcv_q.qlen); in llt_ndlc_rcv_queue()
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf_main.c | 88 queue->qlen = 0; in free_pending_queues() 93 static int alloc_pending_queues(struct otx_cpt_pending_qinfo *pqinfo, u32 qlen, in alloc_pending_queues() argument 103 queue->head = kzalloc_objs(*queue->head, qlen); in alloc_pending_queues() 112 queue->qlen = qlen; in alloc_pending_queues() 125 static int init_pending_queues(struct otx_cptvf *cptvf, u32 qlen, in init_pending_queues() argument 134 ret = alloc_pending_queues(&cptvf->pqinfo, qlen, num_queues); in init_pending_queues() 188 u32 qlen) in alloc_command_queues() argument 199 cptvf->qsize = min(qlen, cqinfo->qchunksize) * in alloc_command_queues() 202 q_size = qlen * OTX_CPT_INST_SIZE; in alloc_command_queues() 263 static int init_command_queues(struct otx_cptvf *cptvf, u32 qlen) in init_command_queues() argument [all …]
|
| H A D | otx_cptvf_reqmgr.c | 59 int qlen) in get_free_pending_entry() argument 68 if (unlikely(q->rear == qlen)) in get_free_pending_entry() 325 pentry = get_free_pending_entry(pqueue, pqueue->qlen); in process_request() 331 pentry = get_free_pending_entry(pqueue, pqueue->qlen); in process_request() 346 pqueue->pending_count > (pqueue->qlen - CPT_IQ_STOP_MARGIN)) { in process_request() 567 resume_index = modulo_inc(pqueue->front, pqueue->qlen, in process_pending_queue() 593 pqueue->front = modulo_inc(pqueue->front, pqueue->qlen, 1); in process_pending_queue()
|
| /linux/tools/testing/selftests/net/tcp_ao/lib/ |
| H A D | repair.c | 42 static void test_sock_checkpoint_queue(int sk, int queue, int qlen, in test_sock_checkpoint_queue() argument 56 if (!qlen) { in test_sock_checkpoint_queue() 61 q->buf = malloc(qlen); in test_sock_checkpoint_queue() 64 ret = recv(sk, q->buf, qlen, MSG_PEEK | MSG_DONTWAIT); in test_sock_checkpoint_queue() 65 if (ret != qlen) in test_sock_checkpoint_queue() 66 test_error("recv(%d): %d", qlen, ret); in test_sock_checkpoint_queue()
|
| /linux/net/kcm/ |
| H A D | kcmproc.c | 119 kcm->sk.sk_receive_queue.qlen, in kcm_format_sock() 121 kcm->sk.sk_write_queue.qlen, in kcm_format_sock() 149 psock->sk->sk_receive_queue.qlen, in kcm_format_psock() 151 psock->sk->sk_write_queue.qlen, in kcm_format_psock() 167 if (psock->sk->sk_receive_queue.qlen) { in kcm_format_psock()
|
| /linux/net/nfc/hci/ |
| H A D | llc_shdlc.c | 322 if (shdlc->send_q.qlen == 0) { in llc_shdlc_rcv_s_frame() 462 if (shdlc->rcv_q.qlen) in llc_shdlc_handle_rcv_queue() 463 pr_debug("rcvQlen=%d\n", shdlc->rcv_q.qlen); in llc_shdlc_handle_rcv_queue() 518 if (shdlc->send_q.qlen) in llc_shdlc_handle_send_queue() 520 shdlc->send_q.qlen, shdlc->ns, shdlc->dnr, in llc_shdlc_handle_send_queue() 523 shdlc->ack_pending_q.qlen); in llc_shdlc_handle_send_queue() 525 while (shdlc->send_q.qlen && shdlc->ack_pending_q.qlen < shdlc->w && in llc_shdlc_handle_send_queue()
|
| /linux/tools/testing/selftests/net/ |
| H A D | tfo.c | 48 unsigned long qlen = 32; in run_server() local 66 if (setsockopt(fd, SOL_TCP, TCP_FASTOPEN, &qlen, sizeof(qlen)) < 0) in run_server()
|
| /linux/drivers/firmware/samsung/ |
| H A D | exynos-acpm.c | 83 u32 qlen; member 149 unsigned int qlen; member 279 i = (i + 1) % achan->qlen; in acpm_get_rx() 441 idx = (tx_front + 1) % achan->qlen; in acpm_do_xfer() 482 achan->qlen = readl(&chan_shmem->qlen); in acpm_chan_shmem_get_params() 493 achan->id, achan->poll_completion, achan->mlen, achan->qlen); in acpm_chan_shmem_get_params()
|