Lines Matching refs:oct
225 int octeon_setup_interrupt(struct octeon_device *oct, u32 num_ioqs);
229 int lio_wait_for_clean_oq(struct octeon_device *oct);
238 int lio_setup_glists(struct octeon_device *oct, struct lio *lio, int num_qs);
260 *recv_buffer_alloc(struct octeon_device *oct, in recv_buffer_alloc() argument
286 pg_info->dma = dma_map_page(&oct->pci_dev->dev, page, 0, in recv_buffer_alloc()
290 if (dma_mapping_error(&oct->pci_dev->dev, pg_info->dma)) { in recv_buffer_alloc()
331 recv_buffer_recycle(struct octeon_device *oct, void *buf) in recv_buffer_recycle() argument
336 dev_err(&oct->pci_dev->dev, "%s: pg_info->page NULL\n", in recv_buffer_recycle()
343 dma_unmap_page(&oct->pci_dev->dev, in recv_buffer_recycle()
363 *recv_buffer_reuse(struct octeon_device *oct, void *buf) in recv_buffer_reuse() argument
370 dma_unmap_page(&oct->pci_dev->dev, in recv_buffer_reuse()
432 #define lio_dma_alloc(oct, size, dma_addr) \ argument
433 dma_alloc_coherent(&(oct)->pci_dev->dev, size, dma_addr, GFP_KERNEL)
434 #define lio_dma_free(oct, size, virt_addr, dma_addr) \ argument
435 dma_free_coherent(&(oct)->pci_dev->dev, size, virt_addr, dma_addr)
535 static inline int wait_for_pending_requests(struct octeon_device *oct) in wait_for_pending_requests() argument
541 &oct->response_list[OCTEON_ORDERED_SC_LIST] in wait_for_pending_requests()
602 static inline int skb_iq(struct octeon_device *oct, struct sk_buff *skb) in skb_iq() argument
604 return skb->queue_mapping % oct->num_iqs; in skb_iq()