1 /* SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause) */ 2 /* 3 * Copyright (C) 2014-2016 Freescale Semiconductor, Inc. 4 * Copyright 2016-2019 NXP 5 * 6 */ 7 #ifndef __FSL_QBMAN_PORTAL_H 8 #define __FSL_QBMAN_PORTAL_H 9 10 #include <soc/fsl/dpaa2-fd.h> 11 12 #define QMAN_REV_4000 0x04000000 13 #define QMAN_REV_4100 0x04010000 14 #define QMAN_REV_4101 0x04010001 15 #define QMAN_REV_5000 0x05000000 16 17 #define QMAN_REV_MASK 0xffff0000 18 19 struct dpaa2_dq; 20 struct qbman_swp; 21 22 /* qbman software portal descriptor structure */ 23 struct qbman_swp_desc { 24 void *cena_bar; /* Cache-enabled portal base address */ 25 void __iomem *cinh_bar; /* Cache-inhibited portal base address */ 26 u32 qman_version; 27 u32 qman_clk; 28 u32 qman_256_cycles_per_ns; 29 }; 30 31 #define QBMAN_SWP_INTERRUPT_EQRI 0x01 32 #define QBMAN_SWP_INTERRUPT_EQDI 0x02 33 #define QBMAN_SWP_INTERRUPT_DQRI 0x04 34 #define QBMAN_SWP_INTERRUPT_RCRI 0x08 35 #define QBMAN_SWP_INTERRUPT_RCDI 0x10 36 #define QBMAN_SWP_INTERRUPT_VDCI 0x20 37 38 /* the structure for pull dequeue descriptor */ 39 struct qbman_pull_desc { 40 u8 verb; 41 u8 numf; 42 u8 tok; 43 u8 reserved; 44 __le32 dq_src; 45 __le64 rsp_addr; 46 u64 rsp_addr_virt; 47 u8 padding[40]; 48 }; 49 50 enum qbman_pull_type_e { 51 /* dequeue with priority precedence, respect intra-class scheduling */ 52 qbman_pull_type_prio = 1, 53 /* dequeue with active FQ precedence, respect ICS */ 54 qbman_pull_type_active, 55 /* dequeue with active FQ precedence, no ICS */ 56 qbman_pull_type_active_noics 57 }; 58 59 /* Definitions for parsing dequeue entries */ 60 #define QBMAN_RESULT_MASK 0x7f 61 #define QBMAN_RESULT_DQ 0x60 62 #define QBMAN_RESULT_FQRN 0x21 63 #define QBMAN_RESULT_FQRNI 0x22 64 #define QBMAN_RESULT_FQPN 0x24 65 #define QBMAN_RESULT_FQDAN 0x25 66 #define QBMAN_RESULT_CDAN 0x26 67 #define QBMAN_RESULT_CSCN_MEM 0x27 68 #define QBMAN_RESULT_CGCU 0x28 69 #define QBMAN_RESULT_BPSCN 0x29 70 #define QBMAN_RESULT_CSCN_WQ 0x2a 71 72 /* QBMan FQ management command codes */ 73 #define QBMAN_FQ_SCHEDULE 0x48 74 #define QBMAN_FQ_FORCE 0x49 75 #define QBMAN_FQ_XON 0x4d 76 #define QBMAN_FQ_XOFF 0x4e 77 78 /* structure of enqueue descriptor */ 79 struct qbman_eq_desc { 80 u8 verb; 81 u8 dca; 82 __le16 seqnum; 83 __le16 orpid; 84 __le16 reserved1; 85 __le32 tgtid; 86 __le32 tag; 87 __le16 qdbin; 88 u8 qpri; 89 u8 reserved[3]; 90 u8 wae; 91 u8 rspid; 92 __le64 rsp_addr; 93 }; 94 95 struct qbman_eq_desc_with_fd { 96 struct qbman_eq_desc desc; 97 u8 fd[32]; 98 }; 99 100 /* buffer release descriptor */ 101 struct qbman_release_desc { 102 u8 verb; 103 u8 reserved; 104 __le16 bpid; 105 __le32 reserved2; 106 __le64 buf[7]; 107 }; 108 109 /* Management command result codes */ 110 #define QBMAN_MC_RSLT_OK 0xf0 111 112 #define CODE_CDAN_WE_EN 0x1 113 #define CODE_CDAN_WE_CTX 0x4 114 115 /* portal data structure */ 116 struct qbman_swp { 117 const struct qbman_swp_desc *desc; 118 void *addr_cena; 119 void __iomem *addr_cinh; 120 121 /* Management commands */ 122 struct { 123 u32 valid_bit; /* 0x00 or 0x80 */ 124 } mc; 125 126 /* Management response */ 127 struct { 128 u32 valid_bit; /* 0x00 or 0x80 */ 129 } mr; 130 131 /* Push dequeues */ 132 u32 sdq; 133 134 /* Volatile dequeues */ 135 struct { 136 atomic_t available; /* indicates if a command can be sent */ 137 u32 valid_bit; /* 0x00 or 0x80 */ 138 struct dpaa2_dq *storage; /* NULL if DQRR */ 139 } vdq; 140 141 /* DQRR */ 142 struct { 143 u32 next_idx; 144 u32 valid_bit; 145 u8 dqrr_size; 146 int reset_bug; /* indicates dqrr reset workaround is needed */ 147 } dqrr; 148 149 struct { 150 u32 pi; 151 u32 pi_vb; 152 u32 pi_ring_size; 153 u32 pi_ci_mask; 154 u32 ci; 155 int available; 156 u32 pend; 157 u32 no_pfdr; 158 } eqcr; 159 160 spinlock_t access_spinlock; 161 162 /* Interrupt coalescing */ 163 u32 irq_threshold; 164 u32 irq_holdoff; 165 int use_adaptive_rx_coalesce; 166 }; 167 168 /* Function pointers */ 169 extern 170 int (*qbman_swp_enqueue_ptr)(struct qbman_swp *s, 171 const struct qbman_eq_desc *d, 172 const struct dpaa2_fd *fd); 173 extern 174 int (*qbman_swp_enqueue_multiple_ptr)(struct qbman_swp *s, 175 const struct qbman_eq_desc *d, 176 const struct dpaa2_fd *fd, 177 uint32_t *flags, 178 int num_frames); 179 extern 180 int (*qbman_swp_enqueue_multiple_desc_ptr)(struct qbman_swp *s, 181 const struct qbman_eq_desc *d, 182 const struct dpaa2_fd *fd, 183 int num_frames); 184 extern 185 int (*qbman_swp_pull_ptr)(struct qbman_swp *s, struct qbman_pull_desc *d); 186 extern 187 const struct dpaa2_dq *(*qbman_swp_dqrr_next_ptr)(struct qbman_swp *s); 188 extern 189 int (*qbman_swp_release_ptr)(struct qbman_swp *s, 190 const struct qbman_release_desc *d, 191 const u64 *buffers, 192 unsigned int num_buffers); 193 194 /* Functions */ 195 struct qbman_swp *qbman_swp_init(const struct qbman_swp_desc *d); 196 void qbman_swp_finish(struct qbman_swp *p); 197 u32 qbman_swp_interrupt_read_status(struct qbman_swp *p); 198 void qbman_swp_interrupt_clear_status(struct qbman_swp *p, u32 mask); 199 u32 qbman_swp_interrupt_get_trigger(struct qbman_swp *p); 200 void qbman_swp_interrupt_set_trigger(struct qbman_swp *p, u32 mask); 201 int qbman_swp_interrupt_get_inhibit(struct qbman_swp *p); 202 void qbman_swp_interrupt_set_inhibit(struct qbman_swp *p, int inhibit); 203 204 void qbman_swp_push_get(struct qbman_swp *p, u8 channel_idx, int *enabled); 205 void qbman_swp_push_set(struct qbman_swp *p, u8 channel_idx, int enable); 206 207 void qbman_pull_desc_clear(struct qbman_pull_desc *d); 208 void qbman_pull_desc_set_storage(struct qbman_pull_desc *d, 209 struct dpaa2_dq *storage, 210 dma_addr_t storage_phys, 211 int stash); 212 void qbman_pull_desc_set_numframes(struct qbman_pull_desc *d, u8 numframes); 213 void qbman_pull_desc_set_fq(struct qbman_pull_desc *d, u32 fqid); 214 void qbman_pull_desc_set_wq(struct qbman_pull_desc *d, u32 wqid, 215 enum qbman_pull_type_e dct); 216 void qbman_pull_desc_set_channel(struct qbman_pull_desc *d, u32 chid, 217 enum qbman_pull_type_e dct); 218 219 void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq); 220 221 int qbman_result_has_new_result(struct qbman_swp *p, const struct dpaa2_dq *dq); 222 223 void qbman_eq_desc_clear(struct qbman_eq_desc *d); 224 void qbman_eq_desc_set_no_orp(struct qbman_eq_desc *d, int respond_success); 225 void qbman_eq_desc_set_token(struct qbman_eq_desc *d, u8 token); 226 void qbman_eq_desc_set_fq(struct qbman_eq_desc *d, u32 fqid); 227 void qbman_eq_desc_set_qd(struct qbman_eq_desc *d, u32 qdid, 228 u32 qd_bin, u32 qd_prio); 229 230 231 void qbman_release_desc_clear(struct qbman_release_desc *d); 232 void qbman_release_desc_set_bpid(struct qbman_release_desc *d, u16 bpid); 233 void qbman_release_desc_set_rcdi(struct qbman_release_desc *d, int enable); 234 235 int qbman_swp_acquire(struct qbman_swp *s, u16 bpid, u64 *buffers, 236 unsigned int num_buffers); 237 int qbman_swp_alt_fq_state(struct qbman_swp *s, u32 fqid, 238 u8 alt_fq_verb); 239 int qbman_swp_CDAN_set(struct qbman_swp *s, u16 channelid, 240 u8 we_mask, u8 cdan_en, 241 u64 ctx); 242 243 void *qbman_swp_mc_start(struct qbman_swp *p); 244 void qbman_swp_mc_submit(struct qbman_swp *p, void *cmd, u8 cmd_verb); 245 void *qbman_swp_mc_result(struct qbman_swp *p); 246 247 /** 248 * qbman_swp_enqueue() - Issue an enqueue command 249 * @s: the software portal used for enqueue 250 * @d: the enqueue descriptor 251 * @fd: the frame descriptor to be enqueued 252 * 253 * Return 0 for successful enqueue, -EBUSY if the EQCR is not ready. 254 */ 255 static inline int 256 qbman_swp_enqueue(struct qbman_swp *s, const struct qbman_eq_desc *d, 257 const struct dpaa2_fd *fd) 258 { 259 return qbman_swp_enqueue_ptr(s, d, fd); 260 } 261 262 /** 263 * qbman_swp_enqueue_multiple() - Issue a multi enqueue command 264 * using one enqueue descriptor 265 * @s: the software portal used for enqueue 266 * @d: the enqueue descriptor 267 * @fd: table pointer of frame descriptor table to be enqueued 268 * @flags: table pointer of QBMAN_ENQUEUE_FLAG_DCA flags, not used if NULL 269 * @num_frames: number of fd to be enqueued 270 * 271 * Return the number of fd enqueued, or a negative error number. 272 */ 273 static inline int 274 qbman_swp_enqueue_multiple(struct qbman_swp *s, 275 const struct qbman_eq_desc *d, 276 const struct dpaa2_fd *fd, 277 uint32_t *flags, 278 int num_frames) 279 { 280 return qbman_swp_enqueue_multiple_ptr(s, d, fd, flags, num_frames); 281 } 282 283 /** 284 * qbman_swp_enqueue_multiple_desc() - Issue a multi enqueue command 285 * using multiple enqueue descriptor 286 * @s: the software portal used for enqueue 287 * @d: table of minimal enqueue descriptor 288 * @fd: table pointer of frame descriptor table to be enqueued 289 * @num_frames: number of fd to be enqueued 290 * 291 * Return the number of fd enqueued, or a negative error number. 292 */ 293 static inline int 294 qbman_swp_enqueue_multiple_desc(struct qbman_swp *s, 295 const struct qbman_eq_desc *d, 296 const struct dpaa2_fd *fd, 297 int num_frames) 298 { 299 return qbman_swp_enqueue_multiple_desc_ptr(s, d, fd, num_frames); 300 } 301 302 /** 303 * qbman_result_is_DQ() - check if the dequeue result is a dequeue response 304 * @dq: the dequeue result to be checked 305 * 306 * DQRR entries may contain non-dequeue results, ie. notifications 307 */ 308 static inline int qbman_result_is_DQ(const struct dpaa2_dq *dq) 309 { 310 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_DQ); 311 } 312 313 /** 314 * qbman_result_is_SCN() - Check the dequeue result is notification or not 315 * @dq: the dequeue result to be checked 316 * 317 */ 318 static inline int qbman_result_is_SCN(const struct dpaa2_dq *dq) 319 { 320 return !qbman_result_is_DQ(dq); 321 } 322 323 /* FQ Data Availability */ 324 static inline int qbman_result_is_FQDAN(const struct dpaa2_dq *dq) 325 { 326 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQDAN); 327 } 328 329 /* Channel Data Availability */ 330 static inline int qbman_result_is_CDAN(const struct dpaa2_dq *dq) 331 { 332 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CDAN); 333 } 334 335 /* Congestion State Change */ 336 static inline int qbman_result_is_CSCN(const struct dpaa2_dq *dq) 337 { 338 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CSCN_WQ); 339 } 340 341 /* Buffer Pool State Change */ 342 static inline int qbman_result_is_BPSCN(const struct dpaa2_dq *dq) 343 { 344 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_BPSCN); 345 } 346 347 /* Congestion Group Count Update */ 348 static inline int qbman_result_is_CGCU(const struct dpaa2_dq *dq) 349 { 350 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CGCU); 351 } 352 353 /* Retirement */ 354 static inline int qbman_result_is_FQRN(const struct dpaa2_dq *dq) 355 { 356 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQRN); 357 } 358 359 /* Retirement Immediate */ 360 static inline int qbman_result_is_FQRNI(const struct dpaa2_dq *dq) 361 { 362 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQRNI); 363 } 364 365 /* Park */ 366 static inline int qbman_result_is_FQPN(const struct dpaa2_dq *dq) 367 { 368 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQPN); 369 } 370 371 /** 372 * qbman_result_SCN_state() - Get the state field in State-change notification 373 */ 374 static inline u8 qbman_result_SCN_state(const struct dpaa2_dq *scn) 375 { 376 return scn->scn.state; 377 } 378 379 #define SCN_RID_MASK 0x00FFFFFF 380 381 /** 382 * qbman_result_SCN_rid() - Get the resource id in State-change notification 383 */ 384 static inline u32 qbman_result_SCN_rid(const struct dpaa2_dq *scn) 385 { 386 return le32_to_cpu(scn->scn.rid_tok) & SCN_RID_MASK; 387 } 388 389 /** 390 * qbman_result_SCN_ctx() - Get the context data in State-change notification 391 */ 392 static inline u64 qbman_result_SCN_ctx(const struct dpaa2_dq *scn) 393 { 394 return le64_to_cpu(scn->scn.ctx); 395 } 396 397 /** 398 * qbman_swp_fq_schedule() - Move the fq to the scheduled state 399 * @s: the software portal object 400 * @fqid: the index of frame queue to be scheduled 401 * 402 * There are a couple of different ways that a FQ can end up parked state, 403 * This schedules it. 404 * 405 * Return 0 for success, or negative error code for failure. 406 */ 407 static inline int qbman_swp_fq_schedule(struct qbman_swp *s, u32 fqid) 408 { 409 return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_SCHEDULE); 410 } 411 412 /** 413 * qbman_swp_fq_force() - Force the FQ to fully scheduled state 414 * @s: the software portal object 415 * @fqid: the index of frame queue to be forced 416 * 417 * Force eligible will force a tentatively-scheduled FQ to be fully-scheduled 418 * and thus be available for selection by any channel-dequeuing behaviour (push 419 * or pull). If the FQ is subsequently "dequeued" from the channel and is still 420 * empty at the time this happens, the resulting dq_entry will have no FD. 421 * (qbman_result_DQ_fd() will return NULL.) 422 * 423 * Return 0 for success, or negative error code for failure. 424 */ 425 static inline int qbman_swp_fq_force(struct qbman_swp *s, u32 fqid) 426 { 427 return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_FORCE); 428 } 429 430 /** 431 * qbman_swp_fq_xon() - sets FQ flow-control to XON 432 * @s: the software portal object 433 * @fqid: the index of frame queue 434 * 435 * This setting doesn't affect enqueues to the FQ, just dequeues. 436 * 437 * Return 0 for success, or negative error code for failure. 438 */ 439 static inline int qbman_swp_fq_xon(struct qbman_swp *s, u32 fqid) 440 { 441 return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_XON); 442 } 443 444 /** 445 * qbman_swp_fq_xoff() - sets FQ flow-control to XOFF 446 * @s: the software portal object 447 * @fqid: the index of frame queue 448 * 449 * This setting doesn't affect enqueues to the FQ, just dequeues. 450 * XOFF FQs will remain in the tenatively-scheduled state, even when 451 * non-empty, meaning they won't be selected for scheduled dequeuing. 452 * If a FQ is changed to XOFF after it had already become truly-scheduled 453 * to a channel, and a pull dequeue of that channel occurs that selects 454 * that FQ for dequeuing, then the resulting dq_entry will have no FD. 455 * (qbman_result_DQ_fd() will return NULL.) 456 * 457 * Return 0 for success, or negative error code for failure. 458 */ 459 static inline int qbman_swp_fq_xoff(struct qbman_swp *s, u32 fqid) 460 { 461 return qbman_swp_alt_fq_state(s, fqid, QBMAN_FQ_XOFF); 462 } 463 464 /* If the user has been allocated a channel object that is going to generate 465 * CDANs to another channel, then the qbman_swp_CDAN* functions will be 466 * necessary. 467 * 468 * CDAN-enabled channels only generate a single CDAN notification, after which 469 * they need to be reenabled before they'll generate another. The idea is 470 * that pull dequeuing will occur in reaction to the CDAN, followed by a 471 * reenable step. Each function generates a distinct command to hardware, so a 472 * combination function is provided if the user wishes to modify the "context" 473 * (which shows up in each CDAN message) each time they reenable, as a single 474 * command to hardware. 475 */ 476 477 /** 478 * qbman_swp_CDAN_set_context() - Set CDAN context 479 * @s: the software portal object 480 * @channelid: the channel index 481 * @ctx: the context to be set in CDAN 482 * 483 * Return 0 for success, or negative error code for failure. 484 */ 485 static inline int qbman_swp_CDAN_set_context(struct qbman_swp *s, u16 channelid, 486 u64 ctx) 487 { 488 return qbman_swp_CDAN_set(s, channelid, 489 CODE_CDAN_WE_CTX, 490 0, ctx); 491 } 492 493 /** 494 * qbman_swp_CDAN_enable() - Enable CDAN for the channel 495 * @s: the software portal object 496 * @channelid: the index of the channel to generate CDAN 497 * 498 * Return 0 for success, or negative error code for failure. 499 */ 500 static inline int qbman_swp_CDAN_enable(struct qbman_swp *s, u16 channelid) 501 { 502 return qbman_swp_CDAN_set(s, channelid, 503 CODE_CDAN_WE_EN, 504 1, 0); 505 } 506 507 /** 508 * qbman_swp_CDAN_disable() - disable CDAN for the channel 509 * @s: the software portal object 510 * @channelid: the index of the channel to generate CDAN 511 * 512 * Return 0 for success, or negative error code for failure. 513 */ 514 static inline int qbman_swp_CDAN_disable(struct qbman_swp *s, u16 channelid) 515 { 516 return qbman_swp_CDAN_set(s, channelid, 517 CODE_CDAN_WE_EN, 518 0, 0); 519 } 520 521 /** 522 * qbman_swp_CDAN_set_context_enable() - Set CDAN contest and enable CDAN 523 * @s: the software portal object 524 * @channelid: the index of the channel to generate CDAN 525 * @ctx:i the context set in CDAN 526 * 527 * Return 0 for success, or negative error code for failure. 528 */ 529 static inline int qbman_swp_CDAN_set_context_enable(struct qbman_swp *s, 530 u16 channelid, 531 u64 ctx) 532 { 533 return qbman_swp_CDAN_set(s, channelid, 534 CODE_CDAN_WE_EN | CODE_CDAN_WE_CTX, 535 1, ctx); 536 } 537 538 /* Wraps up submit + poll-for-result */ 539 static inline void *qbman_swp_mc_complete(struct qbman_swp *swp, void *cmd, 540 u8 cmd_verb) 541 { 542 int loopvar = 2000; 543 544 qbman_swp_mc_submit(swp, cmd, cmd_verb); 545 546 do { 547 cmd = qbman_swp_mc_result(swp); 548 } while (!cmd && loopvar--); 549 550 WARN_ON(!loopvar); 551 552 return cmd; 553 } 554 555 /* Query APIs */ 556 struct qbman_fq_query_np_rslt { 557 u8 verb; 558 u8 rslt; 559 u8 st1; 560 u8 st2; 561 u8 reserved[2]; 562 __le16 od1_sfdr; 563 __le16 od2_sfdr; 564 __le16 od3_sfdr; 565 __le16 ra1_sfdr; 566 __le16 ra2_sfdr; 567 __le32 pfdr_hptr; 568 __le32 pfdr_tptr; 569 __le32 frm_cnt; 570 __le32 byte_cnt; 571 __le16 ics_surp; 572 u8 is; 573 u8 reserved2[29]; 574 }; 575 576 int qbman_fq_query_state(struct qbman_swp *s, u32 fqid, 577 struct qbman_fq_query_np_rslt *r); 578 u32 qbman_fq_state_frame_count(const struct qbman_fq_query_np_rslt *r); 579 u32 qbman_fq_state_byte_count(const struct qbman_fq_query_np_rslt *r); 580 581 struct qbman_bp_query_rslt { 582 u8 verb; 583 u8 rslt; 584 u8 reserved[4]; 585 u8 bdi; 586 u8 state; 587 __le32 fill; 588 __le32 hdotr; 589 __le16 swdet; 590 __le16 swdxt; 591 __le16 hwdet; 592 __le16 hwdxt; 593 __le16 swset; 594 __le16 swsxt; 595 __le16 vbpid; 596 __le16 icid; 597 __le64 bpscn_addr; 598 __le64 bpscn_ctx; 599 __le16 hw_targ; 600 u8 dbe; 601 u8 reserved2; 602 u8 sdcnt; 603 u8 hdcnt; 604 u8 sscnt; 605 u8 reserved3[9]; 606 }; 607 608 int qbman_bp_query(struct qbman_swp *s, u16 bpid, 609 struct qbman_bp_query_rslt *r); 610 611 u32 qbman_bp_info_num_free_bufs(struct qbman_bp_query_rslt *a); 612 613 /** 614 * qbman_swp_release() - Issue a buffer release command 615 * @s: the software portal object 616 * @d: the release descriptor 617 * @buffers: a pointer pointing to the buffer address to be released 618 * @num_buffers: number of buffers to be released, must be less than 8 619 * 620 * Return 0 for success, -EBUSY if the release command ring is not ready. 621 */ 622 static inline int qbman_swp_release(struct qbman_swp *s, 623 const struct qbman_release_desc *d, 624 const u64 *buffers, 625 unsigned int num_buffers) 626 { 627 return qbman_swp_release_ptr(s, d, buffers, num_buffers); 628 } 629 630 /** 631 * qbman_swp_pull() - Issue the pull dequeue command 632 * @s: the software portal object 633 * @d: the software portal descriptor which has been configured with 634 * the set of qbman_pull_desc_set_*() calls 635 * 636 * Return 0 for success, and -EBUSY if the software portal is not ready 637 * to do pull dequeue. 638 */ 639 static inline int qbman_swp_pull(struct qbman_swp *s, 640 struct qbman_pull_desc *d) 641 { 642 return qbman_swp_pull_ptr(s, d); 643 } 644 645 /** 646 * qbman_swp_dqrr_next() - Get an valid DQRR entry 647 * @s: the software portal object 648 * 649 * Return NULL if there are no unconsumed DQRR entries. Return a DQRR entry 650 * only once, so repeated calls can return a sequence of DQRR entries, without 651 * requiring they be consumed immediately or in any particular order. 652 */ 653 static inline const struct dpaa2_dq *qbman_swp_dqrr_next(struct qbman_swp *s) 654 { 655 return qbman_swp_dqrr_next_ptr(s); 656 } 657 658 int qbman_swp_set_irq_coalescing(struct qbman_swp *p, u32 irq_threshold, 659 u32 irq_holdoff); 660 661 void qbman_swp_get_irq_coalescing(struct qbman_swp *p, u32 *irq_threshold, 662 u32 *irq_holdoff); 663 664 #endif /* __FSL_QBMAN_PORTAL_H */ 665