Lines Matching refs:mt76
359 if (is_mt7986(&dev->mt76)) in mt7915_reg_map_l1()
362 l1_remap = is_mt7915(&dev->mt76) ? in mt7915_reg_map_l1()
365 dev->bus_ops->rmw(&dev->mt76, l1_remap, in mt7915_reg_map_l1()
369 dev->bus_ops->rr(&dev->mt76, l1_remap); in mt7915_reg_map_l1()
378 if (is_mt7915(&dev->mt76)) { in mt7915_reg_map_l2()
382 dev->bus_ops->rmw(&dev->mt76, MT_HIF_REMAP_L2, in mt7915_reg_map_l2()
387 dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L2); in mt7915_reg_map_l2()
389 u32 ofs = is_mt7986(&dev->mt76) ? 0x400000 : 0; in mt7915_reg_map_l2()
394 dev->bus_ops->rmw(&dev->mt76, MT_HIF_REMAP_L2_MT7916 + ofs, in mt7915_reg_map_l2()
399 dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L2_MT7916 + ofs); in mt7915_reg_map_l2()
415 dev_err(dev->mt76.dev, "err: reg_map is null\n"); in __mt7915_reg_addr()
437 if (dev_is_pci(dev->mt76.dev) && in __mt7915_reg_addr()
453 struct mt7915_dev *dev = container_of(mdev, struct mt7915_dev, mt76); in mt7915_rr()
461 struct mt7915_dev *dev = container_of(mdev, struct mt7915_dev, mt76); in mt7915_wr()
469 struct mt7915_dev *dev = container_of(mdev, struct mt7915_dev, mt76); in mt7915_rmw()
482 dev = container_of(mdev, struct mt7915_dev, mt76); in mt7915_mmio_init()
483 mt76_mmio_init(&dev->mt76, mem_base); in mt7915_mmio_init()
508 dev->bus_ops = dev->mt76.bus; in mt7915_mmio_init()
509 bus_ops = devm_kmemdup(dev->mt76.dev, dev->bus_ops, sizeof(*bus_ops), in mt7915_mmio_init()
517 dev->mt76.bus = bus_ops; in mt7915_mmio_init()
530 struct mt76_dev *mdev = &dev->mt76; in mt7915_dual_hif_set_irq_mask()
549 struct mt7915_dev *dev = container_of(mdev, struct mt7915_dev, mt76); in mt7915_rx_poll_complete()
558 struct mtk_wed_device *wed = &dev->mt76.mmio.wed; in mt7915_irq_tasklet()
563 intr = mtk_wed_device_irq_get(wed, dev->mt76.mmio.irqmask); in mt7915_irq_tasklet()
570 intr &= dev->mt76.mmio.irqmask; in mt7915_irq_tasklet()
576 intr1 &= dev->mt76.mmio.irqmask; in mt7915_irq_tasklet()
582 trace_dev_irq(&dev->mt76, intr, dev->mt76.mmio.irqmask); in mt7915_irq_tasklet()
591 napi_schedule(&dev->mt76.tx_napi); in mt7915_irq_tasklet()
594 napi_schedule(&dev->mt76.napi[MT_RXQ_MAIN]); in mt7915_irq_tasklet()
597 napi_schedule(&dev->mt76.napi[MT_RXQ_BAND1]); in mt7915_irq_tasklet()
600 napi_schedule(&dev->mt76.napi[MT_RXQ_MCU]); in mt7915_irq_tasklet()
603 napi_schedule(&dev->mt76.napi[MT_RXQ_MCU_WA]); in mt7915_irq_tasklet()
605 if (!is_mt7915(&dev->mt76) && in mt7915_irq_tasklet()
607 napi_schedule(&dev->mt76.napi[MT_RXQ_MAIN_WA]); in mt7915_irq_tasklet()
610 napi_schedule(&dev->mt76.napi[MT_RXQ_BAND1_WA]); in mt7915_irq_tasklet()
618 queue_work(dev->mt76.wq, &dev->reset_work); in mt7915_irq_tasklet()
627 struct mtk_wed_device *wed = &dev->mt76.mmio.wed; in mt7915_irq_handler()
674 dev = container_of(mdev, struct mt7915_dev, mt76); in mt7915_mmio_probe()
685 mt76_free_device(&dev->mt76); in mt7915_mmio_probe()