| /linux/drivers/rtc/ |
| H A D | rtc-gamecube.c | 92 void __iomem *iob; member 99 void __iomem *iob = d->iob; in exi_read() local 106 iowrite32be(RTC_EXICSR, iob + EXICSR); in exi_read() 107 iowrite32be(reg << 8, iob + EXIDATA); in exi_read() 108 iowrite32be(RTC_EXICR_W, iob + EXICR); in exi_read() 109 while (!(ioread32be(iob + EXICSR) & EXICSR_INTSET)) in exi_read() 113 iowrite32be(RTC_EXICSR, iob + EXICSR); in exi_read() 114 iowrite32be(RTC_EXICR_R, iob + EXICR); in exi_read() 115 while (!(ioread32be(iob + EXICSR) & EXICSR_INTSET)) in exi_read() 117 *data = ioread32be(iob + EXIDATA); in exi_read() [all …]
|
| /linux/arch/powerpc/platforms/pasemi/ |
| H A D | iommu.c | 69 static void __iomem *iob; variable 97 out_le32(iob+IOB_AT_INVAL_TLB_REG, bus_addr >> 14); in iobmap_build() 121 out_le32(iob+IOB_AT_INVAL_TLB_REG, bus_addr >> 14); in iobmap_free() 220 iob = ioremap(IOB_BASE, IOB_SIZE); in iob_init() 221 if (!iob) in iob_init() 228 out_le32(iob+IOB_XLT_L1_REGBASE+i*4, regword); in iob_init() 232 regword = in_le32(iob+IOB_AD_REG); in iob_init() 235 out_le32(iob+IOB_AD_REG, regword); in iob_init() 238 regword = in_le32(iob+IOBCOM_REG); in iob_init() 240 out_le32(iob+IOBCOM_REG, regword); in iob_init()
|
| /linux/arch/mips/sgi-ip22/ |
| H A D | ip22-mc.c | 191 iob(); in prom_cleanup() 194 iob(); in prom_cleanup() 197 iob(); in prom_cleanup() 200 iob(); in prom_cleanup()
|
| /linux/drivers/s390/net/ |
| H A D | qeth_core.h | 624 void (*finalize)(struct qeth_card *card, struct qeth_cmd_buffer *iob); 625 bool (*match)(struct qeth_cmd_buffer *iob, 627 void (*callback)(struct qeth_card *card, struct qeth_cmd_buffer *iob, 632 static inline void qeth_get_cmd(struct qeth_cmd_buffer *iob) in qeth_get_cmd() argument 634 refcount_inc(&iob->ref_count); in qeth_get_cmd() 637 static inline struct qeth_ipa_cmd *__ipa_reply(struct qeth_cmd_buffer *iob) in __ipa_reply() argument 639 if (!IS_IPA(iob->data)) in __ipa_reply() 642 return (struct qeth_ipa_cmd *) PDU_ENCAPSULATION(iob->data); in __ipa_reply() 645 static inline struct qeth_ipa_cmd *__ipa_cmd(struct qeth_cmd_buffer *iob) in __ipa_cmd() argument 647 return (struct qeth_ipa_cmd *)(iob->data + IPA_PDU_HEADER_SIZE); in __ipa_cmd() [all …]
|
| /linux/arch/mips/txx9/rbtx4927/ |
| H A D | setup.c | 85 iob(); in tx4927_pci_setup() 101 iob(); in tx4927_pci_setup() 132 iob(); in tx4937_pci_setup() 148 iob(); in tx4937_pci_setup()
|
| /linux/arch/mips/dec/ |
| H A D | kn02-irq.c | 49 iob(); in ack_kn02_irq() 69 iob(); in init_kn02_irqs()
|
| H A D | ecc-berr.c | 37 iob(); in dec_ecc_be_ack() 147 iob(); in dec_ecc_be_backend() 240 iob(); in dec_kn02_be_init()
|
| H A D | kn01-berr.c | 55 iob(); in dec_kn01_be_ack() 190 iob(); in dec_kn01_be_init()
|
| H A D | kn02xa-berr.c | 34 iob(); in dec_kn02xa_be_ack()
|
| /linux/arch/mips/include/asm/ |
| H A D | barrier.h | 72 #define iob() wbflush() macro 77 #define iob() fast_iob() macro
|
| /linux/include/linux/ |
| H A D | blk-mq.h | 749 int blk_rq_poll(struct request *rq, struct io_comp_batch *iob, 894 struct io_comp_batch *iob, bool is_error, in blk_mq_add_to_batch() argument 904 if (!iob) in blk_mq_add_to_batch() 915 if (!iob->complete) in blk_mq_add_to_batch() 916 iob->complete = complete; in blk_mq_add_to_batch() 917 else if (iob->complete != complete) in blk_mq_add_to_batch() 919 iob->need_ts |= blk_mq_need_time_stamp(req); in blk_mq_add_to_batch() 920 rq_list_add_tail(&iob->req_list, req); in blk_mq_add_to_batch()
|
| /linux/io_uring/ |
| H A D | rw.c | 1244 static int io_uring_classic_poll(struct io_kiocb *req, struct io_comp_batch *iob, in io_uring_classic_poll() argument 1253 return file->f_op->uring_cmd_iopoll(ioucmd, iob, poll_flags); in io_uring_classic_poll() 1257 return file->f_op->iopoll(&rw->kiocb, iob, poll_flags); in io_uring_classic_poll() 1296 struct io_comp_batch *iob, unsigned int poll_flags) in io_uring_hybrid_poll() argument 1303 ret = io_uring_classic_poll(req, iob, poll_flags); in io_uring_hybrid_poll() 1320 DEFINE_IO_COMP_BATCH(iob); in io_do_iopoll() 1343 ret = io_uring_hybrid_poll(req, &iob, poll_flags); in io_do_iopoll() 1345 ret = io_uring_classic_poll(req, &iob, poll_flags); in io_do_iopoll() 1353 if (!rq_list_empty(&iob.req_list) || in io_do_iopoll() 1358 if (!rq_list_empty(&iob.req_list)) in io_do_iopoll() [all …]
|
| /linux/arch/mips/mti-malta/ |
| H A D | malta-int.c | 71 iob(); /* sync */ in mips_pcibios_iack() 74 iob(); /* sync */ in mips_pcibios_iack()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | barrier.h | 47 #define iob() c_sync() macro
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-pko-defs.h | 1296 uint64_t iob:1; member 1322 uint64_t iob:1; 1331 uint64_t iob:1; member 1359 uint64_t iob:1; 1369 uint64_t iob:1; member 1403 uint64_t iob:1; 1413 uint64_t iob:1; member 1447 uint64_t iob:1;
|
| H A D | cvmx-npi-defs.h | 2218 uint64_t iob:1; member 2276 uint64_t iob:1; 2285 uint64_t iob:1; member 2347 uint64_t iob:1; 2356 uint64_t iob:1; member 2418 uint64_t iob:1; 2426 uint64_t iob:1; member 2482 uint64_t iob:1;
|
| /linux/arch/mips/pci/ |
| H A D | ops-bcm63xx.c | 111 iob(); in bcm63xx_do_cfg_read() 132 iob(); in bcm63xx_do_cfg_write()
|
| H A D | pci-bcm63xx.c | 105 iob(); in bcm63xx_int_cfg_readl()
|
| /linux/arch/mips/mm/ |
| H A D | c-r3k.c | 278 iob(); in r3k_dma_cache_wback_inv()
|
| /linux/drivers/nvme/host/ |
| H A D | pci.c | 1346 static void nvme_pci_complete_batch(struct io_comp_batch *iob) in nvme_pci_complete_batch() argument 1348 nvme_complete_batch(iob, nvme_pci_unmap_rq); in nvme_pci_complete_batch() 1376 struct io_comp_batch *iob, u16 idx) in nvme_handle_cqe() argument 1404 !blk_mq_add_to_batch(req, iob, in nvme_handle_cqe() 1423 struct io_comp_batch *iob) in nvme_poll_cq() argument 1434 nvme_handle_cqe(nvmeq, iob, nvmeq->cq_head); in nvme_poll_cq() 1446 DEFINE_IO_COMP_BATCH(iob); in nvme_irq() 1448 if (nvme_poll_cq(nvmeq, &iob)) { in nvme_irq() 1449 if (!rq_list_empty(&iob.req_list)) in nvme_irq() 1450 nvme_pci_complete_batch(&iob); in nvme_irq() [all …]
|
| H A D | ioctl.c | 674 struct io_comp_batch *iob, in nvme_ns_chr_uring_cmd_iopoll() argument 681 return blk_rq_poll(req, iob, poll_flags); in nvme_ns_chr_uring_cmd_iopoll()
|
| /linux/Documentation/admin-guide/perf/ |
| H A D | xgene-pmu.rst | 39 / # perf list | grep -e l3c -e iob -e mcb -e mc
|
| /linux/block/ |
| H A D | blk-mq.c | 55 struct io_comp_batch *iob, unsigned int flags); 1188 void blk_mq_end_request_batch(struct io_comp_batch *iob) in blk_mq_end_request_batch() argument 1195 if (iob->need_ts) in blk_mq_end_request_batch() 1198 while ((rq = rq_list_pop(&iob->req_list)) != NULL) { in blk_mq_end_request_batch() 1203 if (iob->need_ts) in blk_mq_end_request_batch() 5193 struct io_comp_batch *iob, unsigned int flags) in blk_hctx_poll() argument 5198 ret = q->mq_ops->poll(hctx, iob); in blk_hctx_poll() 5212 struct io_comp_batch *iob, unsigned int flags) in blk_mq_poll() argument 5216 return blk_hctx_poll(q, q->queue_hw_ctx[cookie], iob, flags); in blk_mq_poll() 5219 int blk_rq_poll(struct request *rq, struct io_comp_batch *iob, in blk_rq_poll() argument [all...] |
| H A D | blk-mq.h | 46 int blk_mq_poll(struct request_queue *q, blk_qc_t cookie, struct io_comp_batch *iob,
|
| /linux/drivers/net/ethernet/amd/ |
| H A D | declance.c | 299 iob(); in writereg() 384 iob(); in cp_to_buf() 524 iob(); in lance_init_ring()
|