/Linux-v5.4/drivers/staging/qlge/ |
D | qlge_dbg.c | 1836 void ql_dump_tx_desc(struct tx_buf_desc *tbd) in ql_dump_tx_desc() argument 1839 le64_to_cpu((u64) tbd->addr)); in ql_dump_tx_desc() 1841 le32_to_cpu(tbd->len & TX_DESC_LEN_MASK)); in ql_dump_tx_desc() 1843 tbd->len & TX_DESC_C ? "C" : ".", in ql_dump_tx_desc() 1844 tbd->len & TX_DESC_E ? "E" : "."); in ql_dump_tx_desc() 1845 tbd++; in ql_dump_tx_desc() 1847 le64_to_cpu((u64) tbd->addr)); in ql_dump_tx_desc() 1849 le32_to_cpu(tbd->len & TX_DESC_LEN_MASK)); in ql_dump_tx_desc() 1851 tbd->len & TX_DESC_C ? "C" : ".", in ql_dump_tx_desc() 1852 tbd->len & TX_DESC_E ? "E" : "."); in ql_dump_tx_desc() [all …]
|
D | qlge_main.c | 1316 struct tx_buf_desc *tbd = mac_iocb_ptr->tbd; in ql_map_send() local 1336 tbd->len = cpu_to_le32(len); in ql_map_send() 1337 tbd->addr = cpu_to_le64(map); in ql_map_send() 1351 tbd++; in ql_map_send() 1383 tbd->addr = cpu_to_le64(map); in ql_map_send() 1389 tbd->len = in ql_map_send() 1396 tbd = (struct tx_buf_desc *)&tx_ring_desc->oal; in ql_map_send() 1411 tbd->addr = cpu_to_le64(map); in ql_map_send() 1412 tbd->len = cpu_to_le32(skb_frag_size(frag)); in ql_map_send() 1421 tbd->len = cpu_to_le32(le32_to_cpu(tbd->len) | TX_DESC_E); in ql_map_send()
|
D | qlge.h | 1097 struct tx_buf_desc tbd[TX_DESC_PER_IOCB]; member 1142 struct tx_buf_desc tbd[TX_DESC_PER_IOCB]; member 2329 void ql_dump_tx_desc(struct tx_buf_desc *tbd);
|
/Linux-v5.4/drivers/net/ethernet/i825xx/ |
D | lib82596.c | 202 u32 tbd; member 973 struct i596_tbd *tbd; in i596_start_xmit() local 989 tbd = lp->dma->tbds + lp->next_tx_cmd; in i596_start_xmit() 1001 tx_cmd->tbd = SWAP32(virt_to_dma(lp, tbd)); in i596_start_xmit() 1002 tbd->next = I596_NULL; in i596_start_xmit() 1009 tbd->pad = 0; in i596_start_xmit() 1010 tbd->size = SWAP16(EOF | length); in i596_start_xmit() 1014 tbd->data = SWAP32(tx_cmd->dma_addr); in i596_start_xmit() 1018 DMA_WBACK_INV(dev, tbd, sizeof(struct i596_tbd)); in i596_start_xmit()
|
D | 82596.c | 235 struct i596_tbd *tbd; member 1054 struct i596_tbd *tbd; in i596_start_xmit() local 1068 tbd = lp->tbds + lp->next_tx_cmd; in i596_start_xmit() 1079 tx_cmd->tbd = WSWAPtbd(virt_to_bus(tbd)); in i596_start_xmit() 1080 tbd->next = I596_NULL; in i596_start_xmit() 1087 tbd->pad = 0; in i596_start_xmit() 1088 tbd->size = EOF | length; in i596_start_xmit() 1090 tbd->data = WSWAPchar(virt_to_bus(skb->data)); in i596_start_xmit()
|
D | ether1.c | 674 tbd_t tbd; in ether1_sendpacket() local 705 tbd.tbd_opts = TBD_EOL | skb->len; in ether1_sendpacket() 706 tbd.tbd_link = I82586_NULL; in ether1_sendpacket() 707 tbd.tbd_bufl = dataddr; in ether1_sendpacket() 708 tbd.tbd_bufh = 0; in ether1_sendpacket() 715 ether1_writebuffer (dev, &tbd, tbdaddr, TBD_SIZE); in ether1_sendpacket()
|
/Linux-v5.4/arch/m68k/include/asm/ |
D | quicc_simple.h | 26 typedef int (handle_tx_error_routine)(int scc_num, int channel_num, QUICC_BD *tbd);
|
/Linux-v5.4/drivers/atm/ |
D | idt77252.c | 707 struct scqe *tbd; in push_on_scq() local 742 tbd = &IDT77252_PRV_TBD(skb); in push_on_scq() 745 scq->next->word_1 = cpu_to_le32(tbd->word_1 | in push_on_scq() 747 scq->next->word_2 = cpu_to_le32(tbd->word_2); in push_on_scq() 748 scq->next->word_3 = cpu_to_le32(tbd->word_3); in push_on_scq() 749 scq->next->word_4 = cpu_to_le32(tbd->word_4); in push_on_scq() 834 struct scqe *tbd; in queue_skb() local 847 tbd = &IDT77252_PRV_TBD(skb); in queue_skb() 859 tbd->word_1 = SAR_TBD_OAM | ATM_CELL_PAYLOAD | SAR_TBD_EPDU; in queue_skb() 860 tbd->word_2 = IDT77252_PRV_PADDR(skb) + 4; in queue_skb() [all …]
|
D | idt77252.h | 789 struct scqe tbd; /* Transmit Buffer Descriptor */ member 795 (((struct idt77252_skb_prv *)(ATM_SKB(skb)+1))->tbd)
|
D | nicstar.c | 133 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd, 1717 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd, in push_scqe() argument 1748 *scq->next = *tbd; in push_scqe() 1754 card->index, le32_to_cpu(tbd->word_1), le32_to_cpu(tbd->word_2), in push_scqe() 1755 le32_to_cpu(tbd->word_3), le32_to_cpu(tbd->word_4), in push_scqe()
|
/Linux-v5.4/drivers/net/wireless/intel/ipw2x00/ |
D | ipw2100.c | 2809 struct ipw2100_bd *tbd; in __ipw2100_tx_process() local 2822 tbd = &txq->drv[packet->index]; in __ipw2100_tx_process() 2834 descriptors_used = tbd->num_fragments; in __ipw2100_tx_process() 2835 frag_num = tbd->num_fragments - 1; in __ipw2100_tx_process() 2919 tbd = &txq->drv[(packet->index + 1 + i) % txq->entries]; in __ipw2100_tx_process() 2923 tbd->host_addr, tbd->buf_length); in __ipw2100_tx_process() 2926 tbd->host_addr, in __ipw2100_tx_process() 2927 tbd->buf_length, PCI_DMA_TODEVICE); in __ipw2100_tx_process() 3001 struct ipw2100_bd *tbd; in ipw2100_tx_send_commands() local 3029 tbd = &txq->drv[txq->next]; in ipw2100_tx_send_commands() [all …]
|
/Linux-v5.4/drivers/net/ethernet/intel/ |
D | e100.c | 495 } tbd; member 1766 cb->u.tcb.tbd_array = cb->dma_addr + offsetof(struct cb, u.tcb.tbd); in e100_xmit_prepare() 1770 cb->u.tcb.tbd.buf_addr = cpu_to_le32(dma_addr); in e100_xmit_prepare() 1771 cb->u.tcb.tbd.size = cpu_to_le16(skb->len); in e100_xmit_prepare() 1835 le32_to_cpu(cb->u.tcb.tbd.buf_addr), in e100_tx_clean() 1836 le16_to_cpu(cb->u.tcb.tbd.size), in e100_tx_clean() 1862 le32_to_cpu(cb->u.tcb.tbd.buf_addr), in e100_clean_cbs() 1863 le16_to_cpu(cb->u.tcb.tbd.size), in e100_clean_cbs()
|
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 2926 unsigned int tbd[SGE_TXQ_PER_SET] = {0, 0}; in sge_timer_tx() local 2930 tbd[TXQ_ETH] = reclaim_completed_tx(adap, &qs->txq[TXQ_ETH], in sge_timer_tx() 2936 tbd[TXQ_OFLD] = reclaim_completed_tx(adap, &qs->txq[TXQ_OFLD], in sge_timer_tx() 2942 (max(tbd[TXQ_ETH], tbd[TXQ_OFLD]) / in sge_timer_tx()
|
/Linux-v5.4/Documentation/input/devices/ |
D | atarikbd.rst | 443 be used sparingly, and the output should not be shut of for more than <tbd>
|