| /linux/fs/gfs2/ |
| H A D | quota.c | 108 struct gfs2_quota_data *qd = container_of(rcu, struct gfs2_quota_data, qd_rcu); in gfs2_qd_dealloc() local 109 struct gfs2_sbd *sdp = qd->qd_sbd; in gfs2_qd_dealloc() 111 kmem_cache_free(gfs2_quotad_cachep, qd); in gfs2_qd_dealloc() 116 static void gfs2_qd_dispose(struct gfs2_quota_data *qd) in gfs2_qd_dispose() argument 118 struct gfs2_sbd *sdp = qd->qd_sbd; in gfs2_qd_dispose() 121 list_del(&qd->qd_list); in gfs2_qd_dispose() 124 spin_lock_bucket(qd->qd_hash); in gfs2_qd_dispose() 125 hlist_bl_del_rcu(&qd->qd_hlist); in gfs2_qd_dispose() 126 spin_unlock_bucket(qd->qd_hash); in gfs2_qd_dispose() 129 gfs2_assert_warn(sdp, !qd->qd_change); in gfs2_qd_dispose() [all …]
|
| /linux/fs/erofs/ |
| H A D | namei.c | 17 const struct erofs_qstr *qd, in erofs_dirnamecmp() argument 27 DBG_BUGON(qd->name > qd->end); in erofs_dirnamecmp() 31 while (qd->name + i < qd->end && qd->name[i] != '\0') { in erofs_dirnamecmp() 32 if (qn->name[i] != qd->name[i]) { in erofs_dirnamecmp() 34 return qn->name[i] > qd->name[i] ? 1 : -1; in erofs_dirnamecmp()
|
| /linux/net/sched/ |
| H A D | sch_cake.c | 1374 static u32 cake_calc_overhead(struct cake_sched_data *qd, u32 len, u32 off) in cake_calc_overhead() argument 1376 struct cake_sched_config *q = qd->config; in cake_calc_overhead() 1381 if (qd->max_netlen < len) in cake_calc_overhead() 1382 qd->max_netlen = len; in cake_calc_overhead() 1383 if (qd->min_netlen > len) in cake_calc_overhead() 1384 qd->min_netlen = len; in cake_calc_overhead() 1403 if (qd->max_adjlen < len) in cake_calc_overhead() 1404 qd->max_adjlen = len; in cake_calc_overhead() 1405 if (qd->min_adjlen > len) in cake_calc_overhead() 1406 qd->min_adjlen = len; in cake_calc_overhead() [all …]
|
| /linux/drivers/s390/block/ |
| H A D | scm_blk.c | 282 const struct blk_mq_queue_data *qd) in scm_blk_request() argument 287 struct request *req = qd->rq; in scm_blk_request() 322 if (qd->last || scmrq->aob->request.msb_count == nr_requests_per_io) { in scm_blk_request() 333 struct scm_queue *qd = kzalloc_obj(*qd); in scm_blk_init_hctx() local 335 if (!qd) in scm_blk_init_hctx() 338 spin_lock_init(&qd->lock); in scm_blk_init_hctx() 339 hctx->driver_data = qd; in scm_blk_init_hctx() 346 struct scm_queue *qd = hctx->driver_data; in scm_blk_exit_hctx() local 348 WARN_ON(qd->scmrq); in scm_blk_exit_hctx()
|
| H A D | dasd.c | 3019 const struct blk_mq_queue_data *qd) in do_dasd_request() argument 3023 struct request *req = qd->rq; in do_dasd_request()
|
| /linux/block/ |
| H A D | blk-iolatency.c | 315 static inline unsigned long scale_amount(unsigned long qd, bool up) in scale_amount() argument 317 return max(up ? qd >> SCALE_UP_FACTOR : qd >> SCALE_DOWN_FACTOR, 1UL); in scale_amount() 333 unsigned long qd = blkiolat->rqos.disk->queue->nr_requests; in scale_cookie_change() local 334 unsigned long scale = scale_amount(qd, up); in scale_cookie_change() 336 unsigned long max_scale = qd << 1; in scale_cookie_change() 346 else if (diff > qd) in scale_cookie_change() 357 if (diff > qd) { in scale_cookie_change() 375 unsigned long qd = iolat->blkiolat->rqos.disk->queue->nr_requests; in scale_change() local 376 unsigned long scale = scale_amount(qd, up); in scale_change() 379 if (old > qd) in scale_change() [all …]
|
| /linux/drivers/scsi/aacraid/ |
| H A D | aachba.c | 3259 struct aac_query_disk qd; in query_disk() local 3265 if (copy_from_user(&qd, arg, sizeof (struct aac_query_disk))) in query_disk() 3267 if (qd.cnum == -1) { in query_disk() 3268 if (qd.id < 0 || qd.id >= dev->maximum_num_containers) in query_disk() 3270 qd.cnum = qd.id; in query_disk() 3271 } else if ((qd.bus == -1) && (qd.id == -1) && (qd.lun == -1)) { in query_disk() 3272 if (qd.cnum < 0 || qd.cnum >= dev->maximum_num_containers) in query_disk() 3274 qd.instance = dev->scsi_host_ptr->host_no; in query_disk() 3275 qd.bus = 0; in query_disk() 3276 qd.id = CONTAINER_TO_ID(qd.cnum); in query_disk() [all …]
|
| /linux/arch/arm/boot/dts/nxp/imx/ |
| H A D | imx6q-dfi-fs700-m60.dts | 16 compatible = "dfi,fs700-m60-6qd", "dfi,fs700e-m60", "fsl,imx6q";
|
| /linux/drivers/mailbox/ |
| H A D | ti-msgmgr.c | 641 const struct ti_msgmgr_valid_queue_desc *qd, in ti_msgmgr_queue_setup() argument 647 qinst->proxy_id = qd->proxy_id; in ti_msgmgr_queue_setup() 648 qinst->queue_id = qd->queue_id; in ti_msgmgr_queue_setup() 682 qinst->is_tx = qd->is_tx; in ti_msgmgr_queue_setup()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_device_queue_manager.c | 626 const struct kfd_criu_queue_priv_data *qd, in create_queue_nocpsch() argument 667 retval = allocate_sdma_queue(dqm, q, qd ? &qd->sdma_id : NULL); in create_queue_nocpsch() 673 retval = allocate_doorbell(qpd, q, qd ? &qd->doorbell_id : NULL); in create_queue_nocpsch() 687 if (qd) in create_queue_nocpsch() 690 qd->ctl_stack_size); in create_queue_nocpsch() 1970 const struct kfd_criu_queue_priv_data *qd, in create_queue_cpsch() argument 1987 retval = allocate_sdma_queue(dqm, q, qd ? &qd->sdma_id : NULL); in create_queue_cpsch() 1993 retval = allocate_doorbell(qpd, q, qd ? &qd->doorbell_id : NULL); in create_queue_cpsch() 2021 if (qd) in create_queue_cpsch() 2024 qd->ctl_stack_size); in create_queue_cpsch()
|
| /linux/drivers/clk/at91/ |
| H A D | clk-audio-pll.c | 47 #define AUDIO_PLL_QDPAD(qd, div) ((AT91_PMC_AUDIO_PLL_QDPAD_EXTDIV(qd) & \ argument
|
| /linux/drivers/s390/crypto/ |
| H A D | ap_card.c | 46 return sysfs_emit(buf, "%d\n", ac->hwinfo.qd + 1); in depth_show()
|
| H A D | ap_queue.c | 288 if (aq->queue_count < aq->card->hwinfo.qd + 1) { in ap_sm_write()
|
| /linux/drivers/block/ |
| H A D | xen-blkfront.c | 894 const struct blk_mq_queue_data *qd) in blkif_queue_rq() argument 902 blk_mq_start_request(qd->rq); in blkif_queue_rq() 915 if (unlikely(req_op(qd->rq) == REQ_OP_FLUSH && !info->feature_flush)) in blkif_queue_rq() 920 if (blkif_queue_request(qd->rq, rinfo)) in blkif_queue_rq() 933 blk_mq_end_request(qd->rq, BLK_STS_OK); in blkif_queue_rq()
|
| H A D | loop.c | 1828 int qd, ret; in loop_set_hw_queue_depth() local 1830 ret = kstrtoint(s, 0, &qd); in loop_set_hw_queue_depth() 1833 if (qd < 1) in loop_set_hw_queue_depth() 1835 hw_queue_depth = qd; in loop_set_hw_queue_depth()
|
| /linux/tools/perf/util/ |
| H A D | auxtrace.c | 1175 struct queue_data *qd = data; in auxtrace_queue_data_cb() local 1179 if (qd->events && event->header.type == PERF_RECORD_AUXTRACE) { in auxtrace_queue_data_cb() 1187 if (!qd->samples || event->header.type != PERF_RECORD_SAMPLE) in auxtrace_queue_data_cb() 1207 struct queue_data qd = { in auxtrace_queue_data() local 1223 auxtrace_queue_data_cb, &qd); in auxtrace_queue_data()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | chip.c | 9662 struct qsfp_data *qd; in qsfp_event() local 9666 qd = container_of(work, struct qsfp_data, qsfp_work); in qsfp_event() 9667 ppd = qd->ppd; in qsfp_event() 9687 if (qd->cache_refresh_required) { in qsfp_event() 9701 if (qd->check_interrupt_flags) { in qsfp_event()
|
| /linux/drivers/ufs/core/ |
| H A D | ufshcd.c | 10725 const struct blk_mq_queue_data *qd) in ufshcd_queue_tmf() argument
|