Home
last modified time | relevance | path

Searched refs:iop (Results 1 – 25 of 43) sorted by relevance

12

/Linux-v5.4/lib/
Dirq_poll.c27 void irq_poll_sched(struct irq_poll *iop) in irq_poll_sched() argument
31 if (test_bit(IRQ_POLL_F_DISABLE, &iop->state)) in irq_poll_sched()
33 if (test_and_set_bit(IRQ_POLL_F_SCHED, &iop->state)) in irq_poll_sched()
37 list_add_tail(&iop->list, this_cpu_ptr(&blk_cpu_iopoll)); in irq_poll_sched()
51 static void __irq_poll_complete(struct irq_poll *iop) in __irq_poll_complete() argument
53 list_del(&iop->list); in __irq_poll_complete()
55 clear_bit_unlock(IRQ_POLL_F_SCHED, &iop->state); in __irq_poll_complete()
68 void irq_poll_complete(struct irq_poll *iop) in irq_poll_complete() argument
73 __irq_poll_complete(iop); in irq_poll_complete()
87 struct irq_poll *iop; in irq_poll_softirq() local
[all …]
/Linux-v5.4/arch/m68k/mac/
Diop.c164 static __inline__ void iop_loadaddr(volatile struct mac_iop *iop, __u16 addr) in iop_loadaddr() argument
166 iop->ram_addr_lo = addr; in iop_loadaddr()
167 iop->ram_addr_hi = addr >> 8; in iop_loadaddr()
170 static __inline__ __u8 iop_readb(volatile struct mac_iop *iop, __u16 addr) in iop_readb() argument
172 iop->ram_addr_lo = addr; in iop_readb()
173 iop->ram_addr_hi = addr >> 8; in iop_readb()
174 return iop->ram_data; in iop_readb()
177 static __inline__ void iop_writeb(volatile struct mac_iop *iop, __u16 addr, __u8 data) in iop_writeb() argument
179 iop->ram_addr_lo = addr; in iop_writeb()
180 iop->ram_addr_hi = addr >> 8; in iop_writeb()
[all …]
DMakefile6 obj-y := config.o macints.o iop.o via.o oss.o psc.o \
/Linux-v5.4/drivers/iommu/
Dio-pgtable.c33 struct io_pgtable *iop; in alloc_io_pgtable_ops() local
43 iop = fns->alloc(cfg, cookie); in alloc_io_pgtable_ops()
44 if (!iop) in alloc_io_pgtable_ops()
47 iop->fmt = fmt; in alloc_io_pgtable_ops()
48 iop->cookie = cookie; in alloc_io_pgtable_ops()
49 iop->cfg = *cfg; in alloc_io_pgtable_ops()
51 return &iop->ops; in alloc_io_pgtable_ops()
61 struct io_pgtable *iop; in free_io_pgtable_ops() local
66 iop = container_of(ops, struct io_pgtable, ops); in free_io_pgtable_ops()
67 io_pgtable_tlb_flush_all(iop); in free_io_pgtable_ops()
[all …]
Dio-pgtable-arm.c29 container_of((x), struct arm_lpae_io_pgtable, iop)
181 struct io_pgtable iop; member
305 if (data->iop.cfg.quirks & IO_PGTABLE_QUIRK_ARM_NS) in __arm_lpae_init_pte()
308 if (data->iop.fmt != ARM_MALI_LPAE && lvl == ARM_LPAE_MAX_LEVELS - 1) in __arm_lpae_init_pte()
313 if (data->iop.fmt != ARM_MALI_LPAE) in __arm_lpae_init_pte()
318 __arm_lpae_set_pte(ptep, pte, &data->iop.cfg); in __arm_lpae_init_pte()
328 if (iopte_leaf(pte, lvl, data->iop.fmt)) { in arm_lpae_init_pte()
389 struct io_pgtable_cfg *cfg = &data->iop.cfg; in __arm_lpae_map()
416 if (pte && !iopte_leaf(pte, lvl, data->iop.fmt)) { in __arm_lpae_map()
433 if (data->iop.fmt == ARM_64_LPAE_S1 || in arm_lpae_prot_to_pte()
[all …]
Dio-pgtable-arm-v7s.c40 container_of((x), struct arm_v7s_io_pgtable, iop)
167 struct io_pgtable iop; member
229 return phys_to_virt(iopte_to_paddr(pte, lvl, &data->iop.cfg)); in iopte_deref()
235 struct io_pgtable_cfg *cfg = &data->iop.cfg; in __arm_v7s_alloc_table()
283 struct io_pgtable_cfg *cfg = &data->iop.cfg; in __arm_v7s_free_table()
416 struct io_pgtable_cfg *cfg = &data->iop.cfg; in arm_v7s_init_pte()
477 struct io_pgtable_cfg *cfg = &data->iop.cfg; in __arm_v7s_map()
524 struct io_pgtable *iop = &data->iop; in arm_v7s_map() local
531 if (WARN_ON(iova >= (1ULL << data->iop.cfg.ias) || in arm_v7s_map()
532 paddr >= (1ULL << data->iop.cfg.oas))) in arm_v7s_map()
[all …]
Dmtk_iommu.c107 struct io_pgtable_ops *iop; member
333 dom->iop = alloc_io_pgtable_ops(ARM_V7S, &dom->cfg, data); in mtk_iommu_domain_finalise()
334 if (!dom->iop) { in mtk_iommu_domain_finalise()
378 free_io_pgtable_ops(dom->iop); in mtk_iommu_domain_free()
427 ret = dom->iop->map(dom->iop, iova, paddr, size, prot); in mtk_iommu_map()
442 unmapsz = dom->iop->unmap(dom->iop, iova, size, gather); in mtk_iommu_unmap()
468 pa = dom->iop->iova_to_phys(dom->iop, iova); in mtk_iommu_iova_to_phys()
Dipmmu-vmsa.c76 struct io_pgtable_ops *iop; member
504 domain->iop = alloc_io_pgtable_ops(ARM_32_LPAE_S1, &domain->cfg, in ipmmu_domain_init_context()
506 if (!domain->iop) { in ipmmu_domain_init_context()
658 free_io_pgtable_ops(domain->iop); in ipmmu_domain_free()
734 return domain->iop->map(domain->iop, iova, paddr, size, prot); in ipmmu_map()
742 return domain->iop->unmap(domain->iop, iova, size, gather); in ipmmu_unmap()
766 return domain->iop->iova_to_phys(domain->iop, iova); in ipmmu_iova_to_phys()
Dmsm_iommu.c45 struct io_pgtable_ops *iop; member
361 priv->iop = alloc_io_pgtable_ops(ARM_V7S, &priv->cfg, priv); in msm_iommu_domain_config()
362 if (!priv->iop) { in msm_iommu_domain_config()
488 free_io_pgtable_ops(priv->iop); in msm_iommu_detach_dev()
514 ret = priv->iop->map(priv->iop, iova, pa, len, prot); in msm_iommu_map()
527 len = priv->iop->unmap(priv->iop, iova, len, gather); in msm_iommu_unmap()
/Linux-v5.4/drivers/md/bcache/
Drequest.c486 struct data_insert_op iop; member
503 s->iop.status = bio->bi_status; in bch_cache_read_endio()
505 ptr_stale(s->iop.c, &b->key, 0)) { in bch_cache_read_endio()
506 atomic_long_inc(&s->iop.c->cache_read_races); in bch_cache_read_endio()
507 s->iop.status = BLK_STS_IOERR; in bch_cache_read_endio()
510 bch_bbio_endio(s->iop.c, bio, bio->bi_status, "reading from cache"); in bch_cache_read_endio()
524 if (bkey_cmp(k, &KEY(s->iop.inode, bio->bi_iter.bi_sector, 0)) <= 0) in cache_lookup_fn()
527 if (KEY_INODE(k) != s->iop.inode || in cache_lookup_fn()
530 unsigned int sectors = KEY_INODE(k) == s->iop.inode in cache_lookup_fn()
561 bch_cut_front(&KEY(s->iop.inode, n->bi_iter.bi_sector, 0), bio_key); in cache_lookup_fn()
[all …]
/Linux-v5.4/include/linux/
Dio-pgtable.h192 static inline void io_pgtable_tlb_flush_all(struct io_pgtable *iop) in io_pgtable_tlb_flush_all() argument
194 iop->cfg.tlb->tlb_flush_all(iop->cookie); in io_pgtable_tlb_flush_all()
198 io_pgtable_tlb_flush_walk(struct io_pgtable *iop, unsigned long iova, in io_pgtable_tlb_flush_walk() argument
201 iop->cfg.tlb->tlb_flush_walk(iova, size, granule, iop->cookie); in io_pgtable_tlb_flush_walk()
205 io_pgtable_tlb_flush_leaf(struct io_pgtable *iop, unsigned long iova, in io_pgtable_tlb_flush_leaf() argument
208 iop->cfg.tlb->tlb_flush_leaf(iova, size, granule, iop->cookie); in io_pgtable_tlb_flush_leaf()
212 io_pgtable_tlb_add_page(struct io_pgtable *iop, in io_pgtable_tlb_add_page() argument
216 if (iop->cfg.tlb->tlb_add_page) in io_pgtable_tlb_add_page()
217 iop->cfg.tlb->tlb_add_page(gather, iova, granule, iop->cookie); in io_pgtable_tlb_add_page()
229 void (*free)(struct io_pgtable *iop);
/Linux-v5.4/arch/powerpc/platforms/8xx/
Dcpm1.c311 struct cpm_ioport32e __iomem *iop; in cpm1_set_pin32() local
315 iop = (struct cpm_ioport32e __iomem *) in cpm1_set_pin32()
318 iop = (struct cpm_ioport32e __iomem *) in cpm1_set_pin32()
322 setbits32(&iop->dir, pin); in cpm1_set_pin32()
324 clrbits32(&iop->dir, pin); in cpm1_set_pin32()
327 setbits32(&iop->par, pin); in cpm1_set_pin32()
329 clrbits32(&iop->par, pin); in cpm1_set_pin32()
340 setbits32(&iop->sor, pin); in cpm1_set_pin32()
342 clrbits32(&iop->sor, pin); in cpm1_set_pin32()
353 struct cpm_ioport16 __iomem *iop = in cpm1_set_pin16() local
[all …]
/Linux-v5.4/fs/iomap/
Dbuffered-io.c25 struct iomap_page *iop = to_iomap_page(page); in iomap_page_create() local
27 if (iop || i_blocksize(inode) == PAGE_SIZE) in iomap_page_create()
28 return iop; in iomap_page_create()
30 iop = kmalloc(sizeof(*iop), GFP_NOFS | __GFP_NOFAIL); in iomap_page_create()
31 atomic_set(&iop->read_count, 0); in iomap_page_create()
32 atomic_set(&iop->write_count, 0); in iomap_page_create()
33 bitmap_zero(iop->uptodate, PAGE_SIZE / SECTOR_SIZE); in iomap_page_create()
40 set_page_private(page, (unsigned long)iop); in iomap_page_create()
42 return iop; in iomap_page_create()
48 struct iomap_page *iop = to_iomap_page(page); in iomap_page_release() local
[all …]
/Linux-v5.4/arch/powerpc/sysdev/
Dcpm_common.c111 struct cpm2_ioports __iomem *iop = mm_gc->regs; in cpm2_gpio32_save_regs() local
113 cpm2_gc->cpdata = in_be32(&iop->dat); in cpm2_gpio32_save_regs()
119 struct cpm2_ioports __iomem *iop = mm_gc->regs; in cpm2_gpio32_get() local
124 return !!(in_be32(&iop->dat) & pin_mask); in cpm2_gpio32_get()
131 struct cpm2_ioports __iomem *iop = mm_gc->regs; in __cpm2_gpio32_set() local
138 out_be32(&iop->dat, cpm2_gc->cpdata); in __cpm2_gpio32_set()
159 struct cpm2_ioports __iomem *iop = mm_gc->regs; in cpm2_gpio32_dir_out() local
165 setbits32(&iop->dir, pin_mask); in cpm2_gpio32_dir_out()
177 struct cpm2_ioports __iomem *iop = mm_gc->regs; in cpm2_gpio32_dir_in() local
183 clrbits32(&iop->dir, pin_mask); in cpm2_gpio32_dir_in()
Dcpm2.c332 struct cpm2_ioports __iomem *iop = in cpm2_set_pin() local
338 setbits32(&iop[port].dir, pin); in cpm2_set_pin()
340 clrbits32(&iop[port].dir, pin); in cpm2_set_pin()
343 setbits32(&iop[port].par, pin); in cpm2_set_pin()
345 clrbits32(&iop[port].par, pin); in cpm2_set_pin()
348 setbits32(&iop[port].sor, pin); in cpm2_set_pin()
350 clrbits32(&iop[port].sor, pin); in cpm2_set_pin()
353 setbits32(&iop[port].odr, pin); in cpm2_set_pin()
355 clrbits32(&iop[port].odr, pin); in cpm2_set_pin()
/Linux-v5.4/drivers/net/fddi/skfp/h/
Dskfbi.h702 #define ADDR(a) (char far *) smc->hw.iop+(a)
703 #define ADDRS(smc,a) (char far *) (smc)->hw.iop+(a)
705 #define ADDR(a) (((a)>>7) ? (outp(smc->hw.iop+B0_RAP,(a)>>7), \
706 (smc->hw.iop+(((a)&0x7F)|((a)>>7 ? 0x80:0)))) : \
707 (smc->hw.iop+(((a)&0x7F)|((a)>>7 ? 0x80:0))))
708 #define ADDRS(smc,a) (((a)>>7) ? (outp((smc)->hw.iop+B0_RAP,(a)>>7), \
709 ((smc)->hw.iop+(((a)&0x7F)|((a)>>7 ? 0x80:0)))) : \
710 ((smc)->hw.iop+(((a)&0x7F)|((a)>>7 ? 0x80:0))))
747 #define GET_ISR_SMP(iop) inpd((iop)+B0_ISRC) argument
749 #define CHECK_ISR_SMP(iop) (inpd((iop)+B0_ISRC) & inpd((iop)+B0_IMSK)) argument
[all …]
Dtargetos.h51 #define ADDR(a) (smc->hw.iop+(a))
53 …ADDR(a) (((a)>>7) ? (outp(smc->hw.iop+B0_RAP,(a)>>7), (smc->hw.iop+( ((a)&0x7F) | ((a)>>7 ? 0x80:0…
Dtargethw.h60 HW_PTR iop ; /* IO base address */ member
/Linux-v5.4/drivers/infiniband/core/
Dcq.c125 static int ib_poll_handler(struct irq_poll *iop, int budget) in ib_poll_handler() argument
127 struct ib_cq *cq = container_of(iop, struct ib_cq, iop); in ib_poll_handler()
133 irq_poll_complete(&cq->iop); in ib_poll_handler()
135 irq_poll_sched(&cq->iop); in ib_poll_handler()
146 irq_poll_sched(&cq->iop); in ib_cq_completion_softirq()
226 irq_poll_init(&cq->iop, IB_POLL_BUDGET_IRQ, ib_poll_handler); in __ib_alloc_cq_user()
297 irq_poll_disable(&cq->iop); in ib_free_cq_user()
/Linux-v5.4/include/uapi/linux/
Di2o-dev.h50 unsigned int iop; /* IOP unit number */ member
55 unsigned int iop; /* IOP unit number */ member
60 unsigned int iop; /* IOP unit number */ member
66 unsigned int iop; /* IOP unit number */ member
75 unsigned int iop; /* IOP unit number */ member
86 unsigned int iop; /* IOP unit number */ member
98 unsigned int iop; member
/Linux-v5.4/drivers/scsi/
Dhptiop.c52 req = readl(&hba->u.itl.iop->inbound_queue); in iop_wait_ready_itl()
59 writel(req, &hba->u.itl.iop->outbound_queue); in iop_wait_ready_itl()
60 readl(&hba->u.itl.iop->outbound_intstatus); in iop_wait_ready_itl()
90 while ((req = readl(&hba->u.itl.iop->outbound_queue)) != in hptiop_drain_outbound_queue_itl()
99 ((char __iomem *)hba->u.itl.iop + req); in hptiop_drain_outbound_queue_itl()
115 struct hpt_iopmu_itl __iomem *iop = hba->u.itl.iop; in iop_intr_itl() local
123 status = readl(&iop->outbound_intstatus); in iop_intr_itl()
126 u32 msg = readl(&iop->outbound_msgaddr0); in iop_intr_itl()
129 writel(IOPMU_OUTBOUND_INT_MSG0, &iop->outbound_intstatus); in iop_intr_itl()
306 writel((unsigned long)req - (unsigned long)hba->u.itl.iop, in iop_send_sync_request_itl()
[all …]
/Linux-v5.4/fs/xfs/
Dxfs_aops.c64 struct iomap_page *iop = to_iomap_page(bvec->bv_page); in xfs_finish_page_writeback() local
71 ASSERT(iop || i_blocksize(inode) == PAGE_SIZE); in xfs_finish_page_writeback()
72 ASSERT(!iop || atomic_read(&iop->write_count) > 0); in xfs_finish_page_writeback()
74 if (!iop || atomic_dec_and_test(&iop->write_count)) in xfs_finish_page_writeback()
759 struct iomap_page *iop, in xfs_add_to_ioend() argument
789 if (iop && !same_page) in xfs_add_to_ioend()
790 atomic_inc(&iop->write_count); in xfs_add_to_ioend()
874 struct iomap_page *iop = to_iomap_page(page); in xfs_writepage_map() local
880 ASSERT(iop || i_blocksize(inode) == PAGE_SIZE); in xfs_writepage_map()
881 ASSERT(!iop || atomic_read(&iop->write_count) == 0); in xfs_writepage_map()
[all …]
/Linux-v5.4/arch/alpha/kernel/
Dcore_wildfire.c187 wildfire_iop *iop; in wildfire_hardware_probe() local
286 iop = WILDFIRE_iop(soft_qbb); in wildfire_hardware_probe()
290 if ((iop->iop_hose[i].init.csr & 1) == 1 && in wildfire_hardware_probe()
579 wildfire_iop *iop = WILDFIRE_iop(qbbno); in wildfire_dump_iop_regs() local
582 printk(KERN_ERR "IOP registers for QBB %d (%p)\n", qbbno, iop); in wildfire_dump_iop_regs()
584 printk(KERN_ERR " IOA_CONFIG: 0x%16lx\n", iop->ioa_config.csr); in wildfire_dump_iop_regs()
585 printk(KERN_ERR " IOD_CONFIG: 0x%16lx\n", iop->iod_config.csr); in wildfire_dump_iop_regs()
587 iop->iop_switch_credits.csr); in wildfire_dump_iop_regs()
589 iop->iop_hose_credits.csr); in wildfire_dump_iop_regs()
593 i, iop->iop_hose[i].init.csr); in wildfire_dump_iop_regs()
[all …]
/Linux-v5.4/fs/fuse/
Dcontrol.c224 const struct inode_operations *iop, in fuse_ctl_add_dentry() argument
247 if (iop) in fuse_ctl_add_dentry()
248 inode->i_op = iop; in fuse_ctl_add_dentry()
/Linux-v5.4/drivers/video/fbdev/
Dcyber2000fb.c1662 unsigned char __iomem *iop; in cyberpro_pci_enable_mmio() local
1664 iop = ioremap(0x3000000, 0x5000); in cyberpro_pci_enable_mmio()
1665 if (iop == NULL) { in cyberpro_pci_enable_mmio()
1670 writeb(0x18, iop + 0x46e8); in cyberpro_pci_enable_mmio()
1671 writeb(0x01, iop + 0x102); in cyberpro_pci_enable_mmio()
1672 writeb(0x08, iop + 0x46e8); in cyberpro_pci_enable_mmio()
1673 writeb(EXT_BIU_MISC, iop + 0x3ce); in cyberpro_pci_enable_mmio()
1674 writeb(EXT_BIU_MISC_LIN_ENABLE, iop + 0x3cf); in cyberpro_pci_enable_mmio()
1676 iounmap(iop); in cyberpro_pci_enable_mmio()

12