Lines Matching full:peer
24 * the peer memory windows.
91 * inbound memory windows for each peer (where N is the number of peers).
101 int peer, peer_widx; in ntb_msi_setup_mws() local
114 for (peer = 0; peer < ntb_peer_port_count(ntb); peer++) { in ntb_msi_setup_mws()
115 peer_widx = ntb_peer_highest_mw_idx(ntb, peer); in ntb_msi_setup_mws()
119 ret = ntb_mw_get_align(ntb, peer, peer_widx, &addr_align, in ntb_msi_setup_mws()
127 for (peer = 0; peer < ntb_peer_port_count(ntb); peer++) { in ntb_msi_setup_mws()
128 peer_widx = ntb_peer_highest_mw_idx(ntb, peer); in ntb_msi_setup_mws()
134 ret = ntb_mw_get_align(ntb, peer, peer_widx, NULL, in ntb_msi_setup_mws()
144 ret = ntb_mw_set_trans(ntb, peer, peer_widx, in ntb_msi_setup_mws()
156 for (i = 0; i < peer; i++) { in ntb_msi_setup_mws()
157 peer_widx = ntb_peer_highest_mw_idx(ntb, peer); in ntb_msi_setup_mws()
176 int peer; in ntb_msi_clear_mws() local
179 for (peer = 0; peer < ntb_peer_port_count(ntb); peer++) { in ntb_msi_clear_mws()
180 peer_widx = ntb_peer_highest_mw_idx(ntb, peer); in ntb_msi_clear_mws()
184 ntb_mw_clear_trans(ntb, peer, peer_widx); in ntb_msi_clear_mws()
268 * it. The descriptor can then be sent to a peer to trigger
352 * ntb_msi_peer_trigger() - Trigger an interrupt handler on a peer
354 * @peer: Peer index
357 * This function triggers an interrupt on a peer. It requires
358 * the descriptor structure to have been passed from that peer
363 int ntb_msi_peer_trigger(struct ntb_dev *ntb, int peer, in ntb_msi_peer_trigger() argument
371 idx = desc->addr_offset / sizeof(*ntb->msi->peer_mws[peer]); in ntb_msi_peer_trigger()
373 iowrite32(desc->data, &ntb->msi->peer_mws[peer][idx]); in ntb_msi_peer_trigger()
380 * ntb_msi_peer_addr() - Get the DMA address to trigger a peer's MSI interrupt
382 * @peer: Peer index
393 int ntb_msi_peer_addr(struct ntb_dev *ntb, int peer, in ntb_msi_peer_addr() argument
397 int peer_widx = ntb_peer_mw_count(ntb) - 1 - peer; in ntb_msi_peer_addr()