Lines Matching refs:nets
83 static struct rionet_net nets[RIONET_MAX_NETS]; variable
185 add_num = nets[rnet->mport->id].nact; in rionet_start_xmit()
200 if (nets[rnet->mport->id].active[i]) { in rionet_start_xmit()
202 nets[rnet->mport->id].active[i]); in rionet_start_xmit()
209 if (nets[rnet->mport->id].active[destid]) in rionet_start_xmit()
211 nets[rnet->mport->id].active[destid]); in rionet_start_xmit()
242 if (!nets[netid].active[sid]) { in rionet_dbell_event()
243 spin_lock(&nets[netid].lock); in rionet_dbell_event()
244 list_for_each_entry(peer, &nets[netid].peers, node) { in rionet_dbell_event()
246 nets[netid].active[sid] = peer->rdev; in rionet_dbell_event()
247 nets[netid].nact++; in rionet_dbell_event()
250 spin_unlock(&nets[netid].lock); in rionet_dbell_event()
256 spin_lock(&nets[netid].lock); in rionet_dbell_event()
257 if (nets[netid].active[sid]) { in rionet_dbell_event()
258 nets[netid].active[sid] = NULL; in rionet_dbell_event()
259 nets[netid].nact--; in rionet_dbell_event()
261 spin_unlock(&nets[netid].lock); in rionet_dbell_event()
357 spin_lock_irqsave(&nets[netid].lock, flags); in rionet_open()
358 list_for_each_entry(peer, &nets[netid].peers, node) { in rionet_open()
362 spin_unlock_irqrestore(&nets[netid].lock, flags); in rionet_open()
387 spin_lock_irqsave(&nets[netid].lock, flags); in rionet_close()
388 list_for_each_entry(peer, &nets[netid].peers, node) { in rionet_close()
389 if (nets[netid].active[peer->rdev->destid]) { in rionet_close()
391 nets[netid].active[peer->rdev->destid] = NULL; in rionet_close()
396 spin_unlock_irqrestore(&nets[netid].lock, flags); in rionet_close()
417 spin_lock_irqsave(&nets[netid].lock, flags); in rionet_remove_dev()
418 list_for_each_entry(peer, &nets[netid].peers, node) { in rionet_remove_dev()
421 if (nets[netid].active[rdev->destid]) { in rionet_remove_dev()
428 nets[netid].active[rdev->destid] = NULL; in rionet_remove_dev()
429 nets[netid].nact--; in rionet_remove_dev()
435 spin_unlock_irqrestore(&nets[netid].lock, flags); in rionet_remove_dev()
492 nets[mport->id].active = (struct rio_dev **)__get_free_pages(GFP_KERNEL, in rionet_setup_netdev()
494 if (!nets[mport->id].active) { in rionet_setup_netdev()
498 memset((void *)nets[mport->id].active, 0, rionet_active_bytes); in rionet_setup_netdev()
530 free_pages((unsigned long)nets[mport->id].active, in rionet_setup_netdev()
564 if (!nets[netid].ndev) { in rionet_add_dev()
591 INIT_LIST_HEAD(&nets[netid].peers); in rionet_add_dev()
592 spin_lock_init(&nets[netid].lock); in rionet_add_dev()
593 nets[netid].nact = 0; in rionet_add_dev()
594 nets[netid].ndev = ndev; in rionet_add_dev()
605 rnet = netdev_priv(nets[netid].ndev); in rionet_add_dev()
623 spin_lock_irqsave(&nets[netid].lock, flags); in rionet_add_dev()
624 list_add_tail(&peer->node, &nets[netid].peers); in rionet_add_dev()
625 spin_unlock_irqrestore(&nets[netid].lock, flags); in rionet_add_dev()
649 if (!nets[i].ndev) in rionet_shutdown()
652 spin_lock_irqsave(&nets[i].lock, flags); in rionet_shutdown()
653 list_for_each_entry(peer, &nets[i].peers, node) { in rionet_shutdown()
654 if (nets[i].active[peer->rdev->destid]) { in rionet_shutdown()
657 nets[i].active[peer->rdev->destid] = NULL; in rionet_shutdown()
660 spin_unlock_irqrestore(&nets[i].lock, flags); in rionet_shutdown()
675 WARN(nets[id].nact, "%s called when connected to %d peers\n", in rionet_remove_mport()
676 __func__, nets[id].nact); in rionet_remove_mport()
677 WARN(!nets[id].ndev, "%s called for mport without NDEV\n", in rionet_remove_mport()
680 if (nets[id].ndev) { in rionet_remove_mport()
681 ndev = nets[id].ndev; in rionet_remove_mport()
685 free_pages((unsigned long)nets[id].active, in rionet_remove_mport()
688 nets[id].active = NULL; in rionet_remove_mport()
690 nets[id].ndev = NULL; in rionet_remove_mport()