/linux/fs/xfs/ |
H A D | xfs_quotaops.c | 26 struct xfs_def_quota *defq; in xfs_qm_fill_state() local 35 defq = xfs_get_defquota(mp->m_quotainfo, type); in xfs_qm_fill_state() 41 tstate->spc_timelimit = (u32)defq->blk.time; in xfs_qm_fill_state() 42 tstate->ino_timelimit = (u32)defq->ino.time; in xfs_qm_fill_state() 43 tstate->rt_spc_timelimit = (u32)defq->rtb.time; in xfs_qm_fill_state()
|
H A D | xfs_qm.c | 554 struct xfs_def_quota *defq; in xfs_qm_set_defquota() local 561 defq = xfs_get_defquota(qinf, xfs_dquot_type(dqp)); in xfs_qm_set_defquota() 567 defq->blk.hard = dqp->q_blk.hardlimit; in xfs_qm_set_defquota() 568 defq->blk.soft = dqp->q_blk.softlimit; in xfs_qm_set_defquota() 569 defq->ino.hard = dqp->q_ino.hardlimit; in xfs_qm_set_defquota() 570 defq->ino.soft = dqp->q_ino.softlimit; in xfs_qm_set_defquota() 571 defq->rtb.hard = dqp->q_rtb.hardlimit; in xfs_qm_set_defquota() 572 defq->rtb.soft = dqp->q_rtb.softlimit; in xfs_qm_set_defquota() 583 struct xfs_def_quota *defq; in xfs_qm_init_timelimits() local 587 defq = xfs_get_defquota(qinf, type); in xfs_qm_init_timelimits() [all …]
|
H A D | xfs_qm_syscalls.c | 281 struct xfs_def_quota *defq; in xfs_qm_scall_setqlim() local 305 defq = xfs_get_defquota(q, xfs_dquot_type(dqp)); in xfs_qm_scall_setqlim() 340 qlim = id == 0 ? &defq->blk : NULL; in xfs_qm_scall_setqlim() 355 qlim = id == 0 ? &defq->rtb : NULL; in xfs_qm_scall_setqlim() 369 qlim = id == 0 ? &defq->ino : NULL; in xfs_qm_scall_setqlim()
|
/linux/net/tipc/ |
H A D | bcast.c | 784 void tipc_mcast_filter_msg(struct net *net, struct sk_buff_head *defq, in tipc_mcast_filter_msg() argument 798 if (likely(!msg_is_syn(hdr) && skb_queue_empty(defq))) in tipc_mcast_filter_msg() 808 skb_queue_walk(defq, _skb) { in tipc_mcast_filter_msg() 822 __skb_queue_tail(defq, skb); in tipc_mcast_filter_msg() 831 __skb_queue_tail(defq, skb); in tipc_mcast_filter_msg() 838 __skb_queue_tail(defq, skb); in tipc_mcast_filter_msg() 843 __skb_unlink(_skb, defq); in tipc_mcast_filter_msg() 853 skb_queue_walk_safe(defq, _skb, tmp) { in tipc_mcast_filter_msg() 861 __skb_unlink(_skb, defq); in tipc_mcast_filter_msg()
|
H A D | group.c | 472 static void tipc_group_sort_msg(struct sk_buff *skb, struct sk_buff_head *defq) in tipc_group_sort_msg() argument 481 skb_queue_walk_safe(defq, _skb, tmp) { in tipc_group_sort_msg() 485 __skb_queue_before(defq, _skb, skb); in tipc_group_sort_msg() 491 __skb_queue_tail(defq, skb); in tipc_group_sort_msg() 501 struct sk_buff_head *defq; in tipc_group_filter_msg() local 525 defq = &m->deferredq; in tipc_group_filter_msg() 526 tipc_group_sort_msg(skb, defq); in tipc_group_filter_msg() 528 while ((skb = skb_peek(defq))) { in tipc_group_filter_msg() 564 __skb_dequeue(defq); in tipc_group_filter_msg() 574 __skb_queue_purge(defq); in tipc_group_filter_msg()
|
H A D | bcast.h | 109 void tipc_mcast_filter_msg(struct net *net, struct sk_buff_head *defq,
|
H A D | link.c | 1767 struct sk_buff_head *defq = &l->deferdq; in tipc_link_rcv() local 1804 if (!__tipc_skb_queue_sorted(defq, seqno, skb)) in tipc_link_rcv() 1822 } while ((skb = __tipc_skb_dequeue(defq, l->rcv_nxt))); in tipc_link_rcv()
|
/linux/fs/xfs/scrub/ |
H A D | quota_repair.c | 304 struct xfs_def_quota *defq = xfs_get_defquota(qi, dqtype); in xrep_quota_block() local 358 defq->blk.time); in xrep_quota_block() 362 defq->ino.time); in xrep_quota_block() 366 defq->rtb.time); in xrep_quota_block()
|
/linux/drivers/net/ethernet/freescale/dpaa/ |
H A D | dpaa_eth.c | 1225 struct dpaa_fq *defq, in dpaa_eth_init_tx_port() argument 1242 params.specific_params.non_rx_params.dflt_fqid = defq->fqid; in dpaa_eth_init_tx_port() 1266 struct dpaa_fq *defq, struct dpaa_fq *pcdq, in dpaa_eth_init_rx_port() argument 1285 rx_p->dflt_fqid = defq->fqid; in dpaa_eth_init_rx_port()
|
/linux/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cxgb4.h | 1900 unsigned int flags, unsigned int defq);
|
H A D | t4_hw.c | 5234 unsigned int flags, unsigned int defq) in t4_config_vi_rss() argument 5244 FW_RSS_VI_CONFIG_CMD_DEFAULTQ_V(defq)); in t4_config_vi_rss()
|