Home
last modified time | relevance | path

Searched refs:rdev (Results 1 – 25 of 66) sorted by relevance

123

/freebsd/sys/dev/bnxt/bnxt_re/
H A Dmain.c109 static int bnxt_re_query_hwrm_intf_version(struct bnxt_re_dev *rdev);
110 static int bnxt_re_hwrm_qcfg(struct bnxt_re_dev *rdev, u32 *db_len,
112 static int bnxt_re_ib_init(struct bnxt_re_dev *rdev);
113 static void bnxt_re_ib_init_2(struct bnxt_re_dev *rdev);
117 static int bnxt_re_hwrm_dbr_pacing_qcfg(struct bnxt_re_dev *rdev);
133 void bnxt_re_set_dma_device(struct ib_device *ibdev, struct bnxt_re_dev *rdev) in bnxt_re_set_dma_device() argument
135 ibdev->dma_device = &rdev->en_dev->pdev->dev; in bnxt_re_set_dma_device()
138 void bnxt_re_init_resolve_wq(struct bnxt_re_dev *rdev) in bnxt_re_init_resolve_wq() argument
140 rdev->resolve_wq = create_singlethread_workqueue("bnxt_re_resolve_wq"); in bnxt_re_init_resolve_wq()
141 INIT_LIST_HEAD(&rdev->mac_wq_list); in bnxt_re_init_resolve_wq()
[all …]
H A Dbnxt_re.h119 int ib_register_device_compat(struct bnxt_re_dev *rdev);
164 #define BNXT_RE_HWRM_CMD_TIMEOUT(rdev) \ argument
165 ((rdev)->chip_ctx->hwrm_cmd_max_timeout * 1000)
205 struct bnxt_re_dev *rdev; member
313 struct bnxt_re_dev *rdev; member
338 struct bnxt_re_dev *rdev; member
439 struct bnxt_re_dev *rdev; member
445 struct bnxt_re_dev *rdev; member
615 struct bnxt_re_dev *rdev; member
632 static inline bool __bnxt_re_is_rdev_valid(struct bnxt_re_dev *rdev) in __bnxt_re_is_rdev_valid() argument
[all …]
H A Dib_verbs.c55 struct bnxt_re_dev *rdev; in bnxt_re_resolve_dmac_task() local
60 rdev = dmac_work->rdev; in bnxt_re_resolve_dmac_task()
62 rc = ib_resolve_eth_dmac(&rdev->ibdev, ah_attr); in bnxt_re_resolve_dmac_task()
64 dev_err(rdev_to_dev(dmac_work->rdev), in bnxt_re_resolve_dmac_task()
111 static int bnxt_re_copy_to_udata(struct bnxt_re_dev *rdev, void *data, in bnxt_re_copy_to_udata() argument
118 dev_err(rdev_to_dev(rdev), in bnxt_re_copy_to_udata()
128 struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); in bnxt_re_get_netdev() local
133 if (!rdev || !rdev->netdev) in bnxt_re_get_netdev()
136 netdev = rdev->netdev; in bnxt_re_get_netdev()
151 struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); in bnxt_re_query_device() local
[all …]
H A Dstats.c34 int bnxt_re_get_flow_stats_from_service_pf(struct bnxt_re_dev *rdev, in bnxt_re_get_flow_stats_from_service_pf() argument
40 struct bnxt_en_dev *en_dev = rdev->en_dev; in bnxt_re_get_flow_stats_from_service_pf()
51 bnxt_re_init_hwrm_hdr(rdev, (void *)&req, HWRM_CFA_FLOW_STATS, -1, target_id); in bnxt_re_get_flow_stats_from_service_pf()
63 sizeof(resp), BNXT_RE_HWRM_CMD_TIMEOUT(rdev)); in bnxt_re_get_flow_stats_from_service_pf()
66 dev_err(rdev_to_dev(rdev), in bnxt_re_get_flow_stats_from_service_pf()
88 int bnxt_re_get_qos_stats(struct bnxt_re_dev *rdev) in bnxt_re_get_qos_stats() argument
94 struct bnxt_en_dev *en_dev = rdev->en_dev; in bnxt_re_get_qos_stats()
103 bnxt_re_init_hwrm_hdr(rdev, (void *)&req, HWRM_CFA_FLOW_STATS, -1, -1); in bnxt_re_get_qos_stats()
115 sizeof(resp), BNXT_RE_HWRM_CMD_TIMEOUT(rdev)); in bnxt_re_get_qos_stats()
118 dev_err(rdev_to_dev(rdev), in bnxt_re_get_qos_stats()
[all …]
H A Dib_verbs.h127 struct bnxt_re_dev *rdev);
130 struct bnxt_re_dev *rdev);
133 bnxt_re_check_if_vlan_valid(struct bnxt_re_dev *rdev, u16 vlan_id);
159 struct bnxt_re_dev *rdev; member
166 struct bnxt_re_dev *rdev; member
172 struct bnxt_re_dev *rdev; member
197 struct bnxt_re_dev *rdev; member
216 struct bnxt_re_dev *rdev; member
236 struct bnxt_re_dev *rdev; member
247 struct bnxt_re_dev *rdev; member
[all …]
H A Dstats.h348 int bnxt_re_get_device_stats(struct bnxt_re_dev *rdev);
349 int bnxt_re_get_flow_stats_from_service_pf(struct bnxt_re_dev *rdev,
352 int bnxt_re_get_qos_stats(struct bnxt_re_dev *rdev);
/freebsd/sys/dev/cxgbe/iw_cxgbe/
H A Dresource.c42 static int c4iw_init_qid_table(struct c4iw_rdev *rdev) in c4iw_init_qid_table() argument
46 if (c4iw_id_table_alloc(&rdev->resource.qid_table, in c4iw_init_qid_table()
47 rdev->adap->vres.qp.start, in c4iw_init_qid_table()
48 rdev->adap->vres.qp.size, in c4iw_init_qid_table()
49 rdev->adap->vres.qp.size, 0)) { in c4iw_init_qid_table()
54 for (i = rdev->adap->vres.qp.start; in c4iw_init_qid_table()
55 i < rdev->adap->vres.qp.start + rdev->adap->vres.qp.size; i++) in c4iw_init_qid_table()
56 if (!(i & rdev->qpmask)) in c4iw_init_qid_table()
57 c4iw_id_free(&rdev in c4iw_init_qid_table()
62 c4iw_init_resource(struct c4iw_rdev * rdev,u32 nr_tpt,u32 nr_pdid) c4iw_init_resource() argument
104 c4iw_get_cqid(struct c4iw_rdev * rdev,struct c4iw_dev_ucontext * uctx) c4iw_get_cqid() argument
159 c4iw_put_cqid(struct c4iw_rdev * rdev,u32 qid,struct c4iw_dev_ucontext * uctx) c4iw_put_cqid() argument
174 c4iw_get_qpid(struct c4iw_rdev * rdev,struct c4iw_dev_ucontext * uctx) c4iw_get_qpid() argument
229 c4iw_put_qpid(struct c4iw_rdev * rdev,u32 qid,struct c4iw_dev_ucontext * uctx) c4iw_put_qpid() argument
255 c4iw_pblpool_alloc(struct c4iw_rdev * rdev,int size) c4iw_pblpool_alloc() argument
274 c4iw_pblpool_free(struct c4iw_rdev * rdev,u32 addr,int size) c4iw_pblpool_free() argument
283 c4iw_pblpool_create(struct c4iw_rdev * rdev) c4iw_pblpool_create() argument
295 c4iw_pblpool_destroy(struct c4iw_rdev * rdev) c4iw_pblpool_destroy() argument
304 c4iw_rqtpool_alloc(struct c4iw_rdev * rdev,int size) c4iw_rqtpool_alloc() argument
328 c4iw_rqtpool_free(struct c4iw_rdev * rdev,u32 addr,int size) c4iw_rqtpool_free() argument
338 c4iw_rqtpool_create(struct c4iw_rdev * rdev) c4iw_rqtpool_create() argument
350 c4iw_rqtpool_destroy(struct c4iw_rdev * rdev) c4iw_rqtpool_destroy() argument
[all...]
H A Ddevice.c49 c4iw_release_dev_ucontext(struct c4iw_rdev *rdev, in c4iw_release_dev_ucontext() argument
59 if (!(entry->qid & rdev->qpmask)) { in c4iw_release_dev_ucontext()
60 c4iw_put_resource(&rdev->resource.qid_table, in c4iw_release_dev_ucontext()
62 mutex_lock(&rdev->stats.lock); in c4iw_release_dev_ucontext()
63 rdev->stats.qid.cur -= rdev->qpmask + 1; in c4iw_release_dev_ucontext()
64 mutex_unlock(&rdev->stats.lock); in c4iw_release_dev_ucontext()
78 c4iw_init_dev_ucontext(struct c4iw_rdev *rdev, struct c4iw_dev_ucontext *uctx) in c4iw_init_dev_ucontext() argument
87 c4iw_rdev_open(struct c4iw_rdev *rdev) in c4iw_rdev_open() argument
89 struct adapter *sc = rdev in c4iw_rdev_open()
186 c4iw_rdev_close(struct c4iw_rdev * rdev) c4iw_rdev_close() argument
[all...]
H A Dmem.c54 return (is_t5(dev->rdev.adap) && length >= 8*1024*1024*1024ULL); in mr_exceeds_hw_limits()
58 _c4iw_write_mem_dma_aligned(struct c4iw_rdev *rdev, u32 addr, u32 len, in _c4iw_write_mem_dma_aligned() argument
61 struct adapter *sc = rdev->adap; in _c4iw_write_mem_dma_aligned()
79 ret = c4iw_wait_for_reply(rdev, &wr_wait, 0, 0, NULL, __func__); in _c4iw_write_mem_dma_aligned()
85 _c4iw_write_mem_inline(struct c4iw_rdev *rdev, u32 addr, u32 len, void *data) in _c4iw_write_mem_dma_aligned()
87 struct adapter *sc = rdev->adap; in _c4iw_write_mem_dma_aligned()
114 ret = c4iw_wait_for_reply(rdev, &wr_wait, 0, 0, NULL, __func__); in _c4iw_write_mem_inline()
119 _c4iw_write_mem_dma(struct c4iw_rdev *rdev, u32 addr, u32 len, void *data) in _c4iw_write_mem_inline()
121 struct c4iw_dev *rhp = rdev_to_c4iw_dev(rdev); in _c4iw_write_mem_inline()
142 ret = _c4iw_write_mem_dma_aligned(rdev, add in _c4iw_write_mem_inline()
108 _c4iw_write_mem_inline(struct c4iw_rdev * rdev,u32 addr,u32 len,void * data) _c4iw_write_mem_inline() argument
180 _c4iw_write_mem_dma(struct c4iw_rdev * rdev,u32 addr,u32 len,void * data) _c4iw_write_mem_dma() argument
223 write_adapter_mem(struct c4iw_rdev * rdev,u32 addr,u32 len,void * data) write_adapter_mem() argument
248 write_tpt_entry(struct c4iw_rdev * rdev,u32 reset_tpt_entry,u32 * stag,u8 stag_state,u32 pdid,enum fw_ri_stag_type type,enum fw_ri_mem_perms perm,int bind_enabled,u32 zbva,u64 to,u64 len,u8 page_size,u32 pbl_size,u32 pbl_addr) write_tpt_entry() argument
320 write_pbl(struct c4iw_rdev * rdev,__be64 * pbl,u32 pbl_addr,u32 pbl_size) write_pbl() argument
332 dereg_mem(struct c4iw_rdev * rdev,u32 stag,u32 pbl_size,u32 pbl_addr) dereg_mem() argument
339 allocate_window(struct c4iw_rdev * rdev,u32 * stag,u32 pdid) allocate_window() argument
346 deallocate_window(struct c4iw_rdev * rdev,u32 stag) deallocate_window() argument
352 allocate_stag(struct c4iw_rdev * rdev,u32 * stag,u32 pdid,u32 pbl_size,u32 pbl_addr) allocate_stag() argument
[all...]
H A Dprovider.c106 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in c4iw_dealloc_ucontext()
121 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext()
130 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext()
150 mm->addr = vtophys(rhp->rdev.status_page); in c4iw_alloc_ucontext()
165 struct c4iw_rdev *rdev; in c4iw_mmap() local
183 rdev = &(to_c4iw_dev(context->device)->rdev); in c4iw_mmap()
198 if (rdev->adap->iwt.wc_en && addr >= rdev->bar2_pa && in c4iw_mmap()
199 addr < rdev->bar2_pa + rdev->bar2_len) in c4iw_mmap()
217 c4iw_put_resource(&rhp->rdev.resource.pdid_table, php->pdid); in c4iw_deallocate_pd()
218 mutex_lock(&rhp->rdev.stats.lock); in c4iw_deallocate_pd()
[all …]
H A Diw_cxgbe.h169 static inline int c4iw_stopped(struct c4iw_rdev *rdev)
171 return rdev->flags & T4_IW_STOPPED; in c4iw_stopped()
174 static inline int c4iw_num_stags(struct c4iw_rdev *rdev)
176 return (int)(rdev->adap->vres.stag.size >> 5); in c4iw_num_stags()
179 static inline int t4_max_fr_depth(struct c4iw_rdev *rdev, bool use_dsgl)
181 if (rdev->adap->params.ulptx_memwrite_dsgl && use_dsgl) in t4_max_fr_depth()
182 return rdev->adap->params.dev_512sgl_mr ? T4_MAX_FR_FW_DSGL_DEPTH : T4_MAX_FR_DSGL_DEPTH; in t4_max_fr_depth()
207 c4iw_wait_for_reply(struct c4iw_rdev *rdev, struct c4iw_wr_wait *wr_waitp,
210 struct adapter *sc = rdev->adap; in c4iw_wait_for_reply()
216 if (c4iw_stopped(rdev)) { in c4iw_wait_for_reply()
170 c4iw_stopped(struct c4iw_rdev * rdev) c4iw_stopped() argument
175 c4iw_num_stags(struct c4iw_rdev * rdev) c4iw_num_stags() argument
180 t4_max_fr_depth(struct c4iw_rdev * rdev,bool use_dsgl) t4_max_fr_depth() argument
208 c4iw_wait_for_reply(struct c4iw_rdev * rdev,struct c4iw_wr_wait * wr_waitp,u32 hwtid,u32 qpid,struct socket * so,const char * func) c4iw_wait_for_reply() argument
269 struct c4iw_rdev rdev; global() member
284 rdev_to_c4iw_dev(struct c4iw_rdev * rdev) rdev_to_c4iw_dev() argument
[all...]
H A Dcq.c52 static int destroy_cq(struct c4iw_rdev *rdev, struct t4_cq *cq, in destroy_cq() argument
55 struct adapter *sc = rdev->adap; in destroy_cq()
56 struct c4iw_dev *rhp = rdev_to_c4iw_dev(rdev); in destroy_cq()
84 c4iw_wait_for_reply(rdev, &wr_wait, 0, 0, NULL, __func__); in destroy_cq()
90 c4iw_put_cqid(rdev, cq->cqid, uctx); in destroy_cq()
95 create_cq(struct c4iw_rdev *rdev, struct t4_cq *cq, in create_cq() argument
98 struct adapter *sc = rdev->adap; in create_cq()
99 struct c4iw_dev *rhp = rdev_to_c4iw_dev(rdev); in create_cq()
103 int user = (uctx != &rdev->uctx); in create_cq()
109 if (__predict_false(c4iw_stopped(rdev))) in create_cq()
[all …]
H A Dqp.c83 device_get_nameunit(dev->rdev.adap->dev)); in alloc_ird()
103 static int destroy_qp(struct c4iw_rdev *rdev, struct t4_wq *wq, in destroy_qp() argument
106 struct c4iw_dev *rhp = rdev_to_c4iw_dev(rdev); in destroy_qp()
117 c4iw_rqtpool_free(rdev, wq->rq.rqt_hwaddr, wq->rq.rqt_size); in destroy_qp()
120 c4iw_put_qpid(rdev, wq->rq.qid, uctx); in destroy_qp()
121 c4iw_put_qpid(rdev, wq->sq.qid, uctx); in destroy_qp()
125 static int create_qp(struct c4iw_rdev *rdev, struct t4_wq *wq, in create_qp() argument
129 struct adapter *sc = rdev->adap; in create_qp()
130 struct c4iw_dev *rhp = rdev_to_c4iw_dev(rdev); in create_qp()
131 int user = (uctx != &rdev->uctx); in create_qp()
[all …]
/freebsd/tests/sys/fs/fusefs/
H A Dmknod.cc81 in.body.mknod.rdev == (uint32_t)dev && in expect_mknod()
92 out.body.entry.attr.rdev = dev; in expect_mknod()
124 in.body.mknod.rdev == (uint32_t)dev && in expect_mknod()
134 out.body.entry.attr.rdev = dev; in expect_mknod()
150 dev_t rdev = 0xfe00; /* /dev/vda's device number on Linux */ in TEST_F() local
155 expect_mknod(FUSE_ROOT_ID, RELPATH, ino, mode, rdev); in TEST_F()
157 EXPECT_EQ(0, mknod(FULLPATH, mode, rdev)) << strerror(errno); in TEST_F()
165 dev_t rdev = 54; /* /dev/fuse's device number */ in TEST_F() local
170 expect_mknod(FUSE_ROOT_ID, RELPATH, ino, mode, rdev); in TEST_F()
172 EXPECT_EQ(0, mknod(FULLPATH, mode, rdev)) << strerror(errno); in TEST_F()
[all …]
/freebsd/contrib/ofed/libbnxtre/
H A Dmain.c291 struct bnxt_re_dev *rdev; in bnxt_re_init_context() local
294 rdev = to_bnxt_re_dev(&vdev->device); in bnxt_re_init_context()
309 ret = _bnxt_re_init_context(rdev, cntx, &resp, cmd_fd); in bnxt_re_init_context()
313 cntx->rdev = rdev; in bnxt_re_init_context()
314 ret = bnxt_re_query_device_compat(&cntx->ibvctx, &rdev->devattr); in bnxt_re_init_context()
323 struct bnxt_re_dev *rdev; in bnxt_re_uninit_context() local
326 rdev = cntx->rdev; in bnxt_re_uninit_context()
327 _bnxt_re_uninit_context(rdev, cntx); in bnxt_re_uninit_context()
H A Dverbs.c276 if (depth > cntx->rdev->max_cq_depth + 1) in bnxt_re_alloc_cqslab()
277 depth = cntx->rdev->max_cq_depth + 1; in bnxt_re_alloc_cqslab()
280 sz = get_aligned((depth * cntx->rdev->cqe_size), cntx->rdev->pg_size); in bnxt_re_alloc_cqslab()
281 mem = bnxt_re_alloc_mem(sz, cntx->rdev->pg_size); in bnxt_re_alloc_cqslab()
484 cq->cqq->stride = cntx->rdev->cqe_size; in bnxt_re_resize_cq_complete()
1178 struct bnxt_re_dev *rdev; in bnxt_re_check_qp_limits() local
1180 rdev = cntx->rdev; in bnxt_re_check_qp_limits()
1181 devattr = &rdev->devattr; in bnxt_re_check_qp_limits()
1202 static int bnxt_re_get_rq_slots(struct bnxt_re_dev *rdev, uint8_t qpmode, in bnxt_re_get_rq_slots() argument
1211 max_wqesz = bnxt_re_calc_wqe_sz(rdev->devattr.max_sge); in bnxt_re_get_rq_slots()
[all …]
/freebsd/libexec/flua/lfs/
H A Dlfs.c245 PUSH_INTEGER(rdev, rdev) in PUSH_INTEGER()
303 PUSH_ENTRY(rdev),
/freebsd/sys/dev/qlnx/qlnxr/
H A Dqlnxr_def.h868 #define QLNX_IS_IWARP(rdev) IS_IWARP(ECORE_LEADING_HWFN(rdev->cdev)) argument
869 #define QLNX_IS_ROCE(rdev) IS_ROCE(ECORE_LEADING_HWFN(rdev->cdev)) argument
/freebsd/sys/fs/tarfs/
H A Dtarfs.h99 dev_t rdev; member
231 unsigned int flags, const char *linkname, dev_t rdev,
H A Dtarfs_vfsops.c446 dev_t rdev; in tarfs_alloc_one() local
552 rdev = NODEV; in tarfs_alloc_one()
782 rdev = makedev(major, minor); in tarfs_alloc_one()
784 0, 0, mtime, uid, gid, mode, flags, NULL, rdev, in tarfs_alloc_one()
804 rdev = makedev(major, minor); in tarfs_alloc_one()
806 0, 0, mtime, uid, gid, mode, flags, NULL, rdev, in tarfs_alloc_one()
/freebsd/sys/dev/cxgb/
H A Dbin2h.pl57 my ($dev,$ino,$mode,$nlink,$uid,$gid,$rdev,$size,
/freebsd/contrib/ntp/ntpd/
H A Dntp_ppsdev.c156 dev_t rdev) in findDevByDevId() argument
172 if (sb.st_rdev == rdev) { in findDevByDevId()
/freebsd/crypto/openssl/providers/implementations/rands/seeding/
H A Drand_unix.c415 dev_t rdev; member
523 && rd->rdev == st.st_rdev; in check_random_device()
547 rd->rdev = st.st_rdev; in get_random_device()
/freebsd/sbin/restore/
H A Drestore.h110 int rdev; /* device number of file */ member
/freebsd/sys/contrib/openzfs/module/zfs/
H A Dzfs_sa.c314 uint64_t uid, gid, mode, rdev, xattr, parent, tmp_gen; in zfs_sa_upgrade() local
355 SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_RDEV(zfsvfs), NULL, &rdev, 8); in zfs_sa_upgrade()
404 &rdev, 8); in zfs_sa_upgrade()

123