Lines Matching refs:perf

144 	struct perf_ctx	*perf;  member
166 struct perf_ctx *perf; member
202 int (*cmd_recv)(struct perf_ctx *perf, int *pidx, enum perf_cmd *cmd,
244 static void perf_terminate_test(struct perf_ctx *perf);
250 link = ntb_link_is_up(peer->perf->ntb, NULL, NULL); in perf_link_is_up()
257 struct perf_ctx *perf = peer->perf; in perf_spad_cmd_send() local
261 dev_dbg(&perf->ntb->dev, "CMD send: %d 0x%llx\n", cmd, data); in perf_spad_cmd_send()
274 sts = ntb_peer_spad_read(perf->ntb, peer->pidx, in perf_spad_cmd_send()
275 PERF_SPAD_CMD(perf->gidx)); in perf_spad_cmd_send()
281 ntb_peer_spad_write(perf->ntb, peer->pidx, in perf_spad_cmd_send()
282 PERF_SPAD_LDATA(perf->gidx), in perf_spad_cmd_send()
284 ntb_peer_spad_write(perf->ntb, peer->pidx, in perf_spad_cmd_send()
285 PERF_SPAD_HDATA(perf->gidx), in perf_spad_cmd_send()
288 ntb_peer_spad_write(perf->ntb, peer->pidx, in perf_spad_cmd_send()
289 PERF_SPAD_CMD(perf->gidx), in perf_spad_cmd_send()
292 ntb_peer_db_set(perf->ntb, PERF_SPAD_NOTIFY(peer->gidx)); in perf_spad_cmd_send()
294 dev_dbg(&perf->ntb->dev, "DB ring peer %#llx\n", in perf_spad_cmd_send()
303 static int perf_spad_cmd_recv(struct perf_ctx *perf, int *pidx, in perf_spad_cmd_recv() argument
309 ntb_db_clear(perf->ntb, PERF_SPAD_NOTIFY(perf->gidx)); in perf_spad_cmd_recv()
317 for (*pidx = 0; *pidx < perf->pcnt; (*pidx)++) { in perf_spad_cmd_recv()
318 peer = &perf->peers[*pidx]; in perf_spad_cmd_recv()
323 val = ntb_spad_read(perf->ntb, PERF_SPAD_CMD(peer->gidx)); in perf_spad_cmd_recv()
329 val = ntb_spad_read(perf->ntb, PERF_SPAD_LDATA(peer->gidx)); in perf_spad_cmd_recv()
332 val = ntb_spad_read(perf->ntb, PERF_SPAD_HDATA(peer->gidx)); in perf_spad_cmd_recv()
336 ntb_spad_write(perf->ntb, PERF_SPAD_CMD(peer->gidx), in perf_spad_cmd_recv()
339 dev_dbg(&perf->ntb->dev, "CMD recv: %d 0x%llx\n", *cmd, *data); in perf_spad_cmd_recv()
350 struct perf_ctx *perf = peer->perf; in perf_msg_cmd_send() local
354 dev_dbg(&perf->ntb->dev, "CMD send: %d 0x%llx\n", cmd, data); in perf_msg_cmd_send()
363 outbits = ntb_msg_outbits(perf->ntb); in perf_msg_cmd_send()
368 ret = ntb_msg_clear_sts(perf->ntb, outbits); in perf_msg_cmd_send()
372 ntb_peer_msg_write(perf->ntb, peer->pidx, PERF_MSG_LDATA, in perf_msg_cmd_send()
375 if (ntb_msg_read_sts(perf->ntb) & outbits) { in perf_msg_cmd_send()
380 ntb_peer_msg_write(perf->ntb, peer->pidx, PERF_MSG_HDATA, in perf_msg_cmd_send()
385 ntb_peer_msg_write(perf->ntb, peer->pidx, PERF_MSG_CMD, cmd); in perf_msg_cmd_send()
393 static int perf_msg_cmd_recv(struct perf_ctx *perf, int *pidx, in perf_msg_cmd_recv() argument
399 inbits = ntb_msg_inbits(perf->ntb); in perf_msg_cmd_recv()
401 if (hweight64(ntb_msg_read_sts(perf->ntb) & inbits) < 3) in perf_msg_cmd_recv()
404 val = ntb_msg_read(perf->ntb, pidx, PERF_MSG_CMD); in perf_msg_cmd_recv()
407 val = ntb_msg_read(perf->ntb, pidx, PERF_MSG_LDATA); in perf_msg_cmd_recv()
410 val = ntb_msg_read(perf->ntb, pidx, PERF_MSG_HDATA); in perf_msg_cmd_recv()
414 ntb_msg_clear_sts(perf->ntb, inbits); in perf_msg_cmd_recv()
416 dev_dbg(&perf->ntb->dev, "CMD recv: %d 0x%llx\n", *cmd, *data); in perf_msg_cmd_recv()
423 struct perf_ctx *perf = peer->perf; in perf_cmd_send() local
426 return perf->cmd_send(peer, cmd, data); in perf_cmd_send()
428 dev_err(&perf->ntb->dev, "Send invalid command\n"); in perf_cmd_send()
442 dev_err(&peer->perf->ntb->dev, "Exec invalid command\n"); in perf_cmd_exec()
449 dev_dbg(&peer->perf->ntb->dev, "CMD exec: %d\n", cmd); in perf_cmd_exec()
456 static int perf_cmd_recv(struct perf_ctx *perf) in perf_cmd_recv() argument
462 while (!(ret = perf->cmd_recv(perf, &pidx, &cmd, &data))) { in perf_cmd_recv()
463 peer = &perf->peers[pidx]; in perf_cmd_recv()
473 dev_err(&perf->ntb->dev, "Recv invalid command\n"); in perf_cmd_recv()
484 struct perf_ctx *perf = ctx; in perf_link_event() local
489 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_link_event()
490 peer = &perf->peers[pidx]; in perf_link_event()
506 struct perf_ctx *perf = ctx; in perf_db_event() local
508 dev_dbg(&perf->ntb->dev, "DB vec %d mask %#llx bits %#llx\n", vec, in perf_db_event()
509 ntb_db_vector_mask(perf->ntb, vec), ntb_db_read(perf->ntb)); in perf_db_event()
512 (void)perf_cmd_recv(perf); in perf_db_event()
517 struct perf_ctx *perf = ctx; in perf_msg_event() local
519 dev_dbg(&perf->ntb->dev, "Msg status bits %#llx\n", in perf_msg_event()
520 ntb_msg_read_sts(perf->ntb)); in perf_msg_event()
523 (void)perf_cmd_recv(perf); in perf_msg_event()
534 (void)ntb_peer_mw_clear_trans(peer->perf->ntb, peer->pidx, peer->gidx); in perf_free_outbuf()
539 struct perf_ctx *perf = peer->perf; in perf_setup_outbuf() local
543 ret = ntb_peer_mw_set_trans(perf->ntb, peer->pidx, peer->gidx, in perf_setup_outbuf()
546 dev_err(&perf->ntb->dev, "Failed to set outbuf translation\n"); in perf_setup_outbuf()
561 (void)ntb_mw_clear_trans(peer->perf->ntb, peer->pidx, peer->gidx); in perf_free_inbuf()
562 dma_free_coherent(&peer->perf->ntb->dev, peer->inbuf_size, in perf_free_inbuf()
570 struct perf_ctx *perf = peer->perf; in perf_setup_inbuf() local
574 ret = ntb_mw_get_align(perf->ntb, peer->pidx, perf->gidx, in perf_setup_inbuf()
577 dev_err(&perf->ntb->dev, "Couldn't get inbuf restrictions\n"); in perf_setup_inbuf()
582 dev_err(&perf->ntb->dev, "Too big inbuf size %pa > %pa\n", in perf_setup_inbuf()
591 peer->inbuf = dma_alloc_coherent(&perf->ntb->dev, peer->inbuf_size, in perf_setup_inbuf()
594 dev_err(&perf->ntb->dev, "Failed to alloc inbuf of %pa\n", in perf_setup_inbuf()
599 dev_err(&perf->ntb->dev, "Unaligned inbuf allocated\n"); in perf_setup_inbuf()
603 ret = ntb_mw_set_trans(perf->ntb, peer->pidx, peer->gidx, in perf_setup_inbuf()
606 dev_err(&perf->ntb->dev, "Failed to set inbuf translation\n"); in perf_setup_inbuf()
643 if (test_bit(0, &peer->perf->busy_flag) && in perf_service_work()
644 peer == peer->perf->test_peer) { in perf_service_work()
645 dev_warn(&peer->perf->ntb->dev, in perf_service_work()
647 perf_terminate_test(peer->perf); in perf_service_work()
654 static int perf_init_service(struct perf_ctx *perf) in perf_init_service() argument
658 if (ntb_peer_mw_count(perf->ntb) < perf->pcnt + 1) { in perf_init_service()
659 dev_err(&perf->ntb->dev, "Not enough memory windows\n"); in perf_init_service()
663 if (ntb_msg_count(perf->ntb) >= PERF_MSG_CNT) { in perf_init_service()
664 perf->cmd_send = perf_msg_cmd_send; in perf_init_service()
665 perf->cmd_recv = perf_msg_cmd_recv; in perf_init_service()
667 dev_dbg(&perf->ntb->dev, "Message service initialized\n"); in perf_init_service()
672 dev_dbg(&perf->ntb->dev, "Message service unsupported\n"); in perf_init_service()
674 mask = GENMASK_ULL(perf->pcnt, 0); in perf_init_service()
675 if (ntb_spad_count(perf->ntb) >= PERF_SPAD_CNT(perf->pcnt) && in perf_init_service()
676 (ntb_db_valid_mask(perf->ntb) & mask) == mask) { in perf_init_service()
677 perf->cmd_send = perf_spad_cmd_send; in perf_init_service()
678 perf->cmd_recv = perf_spad_cmd_recv; in perf_init_service()
680 dev_dbg(&perf->ntb->dev, "Scratchpad service initialized\n"); in perf_init_service()
685 dev_dbg(&perf->ntb->dev, "Scratchpad service unsupported\n"); in perf_init_service()
687 dev_err(&perf->ntb->dev, "Command services unsupported\n"); in perf_init_service()
692 static int perf_enable_service(struct perf_ctx *perf) in perf_enable_service() argument
697 mask = ntb_db_valid_mask(perf->ntb); in perf_enable_service()
698 (void)ntb_db_set_mask(perf->ntb, mask); in perf_enable_service()
700 ret = ntb_set_ctx(perf->ntb, perf, &perf_ops); in perf_enable_service()
704 if (perf->cmd_send == perf_msg_cmd_send) { in perf_enable_service()
707 inbits = ntb_msg_inbits(perf->ntb); in perf_enable_service()
708 outbits = ntb_msg_outbits(perf->ntb); in perf_enable_service()
709 (void)ntb_msg_set_mask(perf->ntb, inbits | outbits); in perf_enable_service()
712 ret = ntb_msg_clear_mask(perf->ntb, incmd_bit); in perf_enable_service()
714 dev_dbg(&perf->ntb->dev, "MSG sts unmasked %#llx\n", incmd_bit); in perf_enable_service()
716 scnt = ntb_spad_count(perf->ntb); in perf_enable_service()
718 ntb_spad_write(perf->ntb, sidx, PERF_CMD_INVAL); in perf_enable_service()
719 incmd_bit = PERF_SPAD_NOTIFY(perf->gidx); in perf_enable_service()
720 ret = ntb_db_clear_mask(perf->ntb, incmd_bit); in perf_enable_service()
722 dev_dbg(&perf->ntb->dev, "DB bits unmasked %#llx\n", incmd_bit); in perf_enable_service()
725 ntb_clear_ctx(perf->ntb); in perf_enable_service()
729 ntb_link_enable(perf->ntb, NTB_SPEED_AUTO, NTB_WIDTH_AUTO); in perf_enable_service()
731 ntb_link_event(perf->ntb); in perf_enable_service()
736 static void perf_disable_service(struct perf_ctx *perf) in perf_disable_service() argument
740 ntb_link_disable(perf->ntb); in perf_disable_service()
742 if (perf->cmd_send == perf_msg_cmd_send) { in perf_disable_service()
745 inbits = ntb_msg_inbits(perf->ntb); in perf_disable_service()
746 (void)ntb_msg_set_mask(perf->ntb, inbits); in perf_disable_service()
748 (void)ntb_db_set_mask(perf->ntb, PERF_SPAD_NOTIFY(perf->gidx)); in perf_disable_service()
751 ntb_clear_ctx(perf->ntb); in perf_disable_service()
753 for (pidx = 0; pidx < perf->pcnt; pidx++) in perf_disable_service()
754 perf_cmd_exec(&perf->peers[pidx], PERF_CMD_CLEAR); in perf_disable_service()
756 for (pidx = 0; pidx < perf->pcnt; pidx++) in perf_disable_service()
757 flush_work(&perf->peers[pidx].service); in perf_disable_service()
841 return likely(atomic_read(&pthr->perf->tsync) > 0) ? 0 : -EINTR; in perf_copy_chunk()
851 struct perf_ctx *perf = data; in perf_dma_filter() local
854 node = dev_to_node(&perf->ntb->dev); in perf_dma_filter()
861 struct perf_ctx *perf = pthr->perf; in perf_init_test() local
864 pthr->src = kmalloc_node(perf->test_peer->outbuf_size, GFP_KERNEL, in perf_init_test()
865 dev_to_node(&perf->ntb->dev)); in perf_init_test()
869 get_random_bytes(pthr->src, perf->test_peer->outbuf_size); in perf_init_test()
876 pthr->dma_chan = dma_request_channel(dma_mask, perf_dma_filter, perf); in perf_init_test()
878 dev_err(&perf->ntb->dev, "%d: Failed to get DMA channel\n", in perf_init_test()
880 atomic_dec(&perf->tsync); in perf_init_test()
881 wake_up(&perf->twait); in perf_init_test()
893 struct perf_peer *peer = pthr->perf->test_peer; in perf_run_test()
894 struct perf_ctx *perf = pthr->perf; in perf_run_test() local
914 dev_err(&perf->ntb->dev, "%d: Got error %d on test\n", in perf_run_test()
937 struct perf_ctx *perf = pthr->perf; in perf_sync_test() local
944 atomic_read(&perf->tsync) < 0)); in perf_sync_test()
946 if (atomic_read(&perf->tsync) < 0) in perf_sync_test()
952 dev_dbg(&perf->ntb->dev, "%d: copied %llu bytes\n", in perf_sync_test()
955 dev_dbg(&perf->ntb->dev, "%d: lasted %llu usecs\n", in perf_sync_test()
958 dev_dbg(&perf->ntb->dev, "%d: %llu MBytes/s\n", pthr->tidx, in perf_sync_test()
966 struct perf_ctx *perf = pthr->perf; in perf_clear_test() local
980 atomic_dec(&perf->tsync); in perf_clear_test()
981 wake_up(&perf->twait); in perf_clear_test()
1015 static int perf_set_tcnt(struct perf_ctx *perf, u8 tcnt) in perf_set_tcnt() argument
1020 if (test_and_set_bit_lock(0, &perf->busy_flag)) in perf_set_tcnt()
1023 perf->tcnt = tcnt; in perf_set_tcnt()
1025 clear_bit_unlock(0, &perf->busy_flag); in perf_set_tcnt()
1030 static void perf_terminate_test(struct perf_ctx *perf) in perf_terminate_test() argument
1034 atomic_set(&perf->tsync, -1); in perf_terminate_test()
1035 wake_up(&perf->twait); in perf_terminate_test()
1038 wake_up(&perf->threads[tidx].dma_wait); in perf_terminate_test()
1039 cancel_work_sync(&perf->threads[tidx].work); in perf_terminate_test()
1045 struct perf_ctx *perf = peer->perf; in perf_submit_test() local
1052 if (test_and_set_bit_lock(0, &perf->busy_flag)) in perf_submit_test()
1055 perf->test_peer = peer; in perf_submit_test()
1056 atomic_set(&perf->tsync, perf->tcnt); in perf_submit_test()
1059 pthr = &perf->threads[tidx]; in perf_submit_test()
1064 if (tidx < perf->tcnt) in perf_submit_test()
1068 ret = wait_event_interruptible(perf->twait, in perf_submit_test()
1069 atomic_read(&perf->tsync) <= 0); in perf_submit_test()
1071 perf_terminate_test(perf); in perf_submit_test()
1075 clear_bit_unlock(0, &perf->busy_flag); in perf_submit_test()
1080 static int perf_read_stats(struct perf_ctx *perf, char *buf, in perf_read_stats() argument
1086 if (test_and_set_bit_lock(0, &perf->busy_flag)) in perf_read_stats()
1090 " Peer %d test statistics:\n", perf->test_peer->pidx); in perf_read_stats()
1093 pthr = &perf->threads[tidx]; in perf_read_stats()
1110 clear_bit_unlock(0, &perf->busy_flag); in perf_read_stats()
1115 static void perf_init_threads(struct perf_ctx *perf) in perf_init_threads() argument
1120 perf->tcnt = DEF_THREADS_CNT; in perf_init_threads()
1121 perf->test_peer = &perf->peers[0]; in perf_init_threads()
1122 init_waitqueue_head(&perf->twait); in perf_init_threads()
1125 pthr = &perf->threads[tidx]; in perf_init_threads()
1127 pthr->perf = perf; in perf_init_threads()
1135 static void perf_clear_threads(struct perf_ctx *perf) in perf_clear_threads() argument
1137 perf_terminate_test(perf); in perf_clear_threads()
1148 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_read_info() local
1165 "Local port %d, Global index %d\n", ntb_port_number(perf->ntb), in perf_dbgfs_read_info()
1166 perf->gidx); in perf_dbgfs_read_info()
1168 if (test_bit(0, &perf->busy_flag)) { in perf_dbgfs_read_info()
1171 ntb_peer_port_number(perf->ntb, perf->test_peer->pidx), in perf_dbgfs_read_info()
1172 perf->test_peer->pidx); in perf_dbgfs_read_info()
1177 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_dbgfs_read_info()
1178 peer = &perf->peers[pidx]; in perf_dbgfs_read_info()
1182 ntb_peer_port_number(perf->ntb, peer->pidx), peer->pidx, in perf_dbgfs_read_info()
1228 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_read_run() local
1236 ret = perf_read_stats(perf, buf, PERF_BUF_LEN, &pos); in perf_dbgfs_read_run()
1250 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_write_run() local
1258 if (pidx < 0 || pidx >= perf->pcnt) in perf_dbgfs_write_run()
1261 peer = &perf->peers[pidx]; in perf_dbgfs_write_run()
1279 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_read_tcnt() local
1283 pos = scnprintf(buf, sizeof(buf), "%hhu\n", perf->tcnt); in perf_dbgfs_read_tcnt()
1292 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_write_tcnt() local
1300 ret = perf_set_tcnt(perf, val); in perf_dbgfs_write_tcnt()
1313 static void perf_setup_dbgfs(struct perf_ctx *perf) in perf_setup_dbgfs() argument
1315 struct pci_dev *pdev = perf->ntb->pdev; in perf_setup_dbgfs()
1317 perf->dbgfs_dir = debugfs_create_dir(pci_name(pdev), perf_dbgfs_topdir); in perf_setup_dbgfs()
1318 if (!perf->dbgfs_dir) { in perf_setup_dbgfs()
1319 dev_warn(&perf->ntb->dev, "DebugFS unsupported\n"); in perf_setup_dbgfs()
1323 debugfs_create_file("info", 0600, perf->dbgfs_dir, perf, in perf_setup_dbgfs()
1326 debugfs_create_file("run", 0600, perf->dbgfs_dir, perf, in perf_setup_dbgfs()
1329 debugfs_create_file("threads_count", 0600, perf->dbgfs_dir, perf, in perf_setup_dbgfs()
1333 debugfs_create_u8("chunk_order", 0500, perf->dbgfs_dir, &chunk_order); in perf_setup_dbgfs()
1335 debugfs_create_u8("total_order", 0500, perf->dbgfs_dir, &total_order); in perf_setup_dbgfs()
1337 debugfs_create_bool("use_dma", 0500, perf->dbgfs_dir, &use_dma); in perf_setup_dbgfs()
1340 static void perf_clear_dbgfs(struct perf_ctx *perf) in perf_clear_dbgfs() argument
1342 debugfs_remove_recursive(perf->dbgfs_dir); in perf_clear_dbgfs()
1352 struct perf_ctx *perf; in perf_create_data() local
1354 perf = devm_kzalloc(&ntb->dev, sizeof(*perf), GFP_KERNEL); in perf_create_data()
1355 if (!perf) in perf_create_data()
1358 perf->pcnt = ntb_peer_port_count(ntb); in perf_create_data()
1359 perf->peers = devm_kcalloc(&ntb->dev, perf->pcnt, sizeof(*perf->peers), in perf_create_data()
1361 if (!perf->peers) in perf_create_data()
1364 perf->ntb = ntb; in perf_create_data()
1366 return perf; in perf_create_data()
1371 struct perf_ctx *perf = peer->perf; in perf_setup_peer_mw() local
1376 ret = ntb_peer_mw_get_addr(perf->ntb, peer->gidx, &phys_addr, in perf_setup_peer_mw()
1381 peer->outbuf = devm_ioremap_wc(&perf->ntb->dev, phys_addr, in perf_setup_peer_mw()
1388 dev_warn(&peer->perf->ntb->dev, in perf_setup_peer_mw()
1396 static int perf_init_peers(struct perf_ctx *perf) in perf_init_peers() argument
1401 lport = ntb_port_number(perf->ntb); in perf_init_peers()
1402 perf->gidx = -1; in perf_init_peers()
1403 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_init_peers()
1404 peer = &perf->peers[pidx]; in perf_init_peers()
1406 peer->perf = perf; in perf_init_peers()
1408 if (lport < ntb_peer_port_number(perf->ntb, pidx)) { in perf_init_peers()
1409 if (perf->gidx == -1) in perf_init_peers()
1410 perf->gidx = pidx; in perf_init_peers()
1417 if (perf->gidx == -1) in perf_init_peers()
1418 perf->gidx = pidx; in perf_init_peers()
1420 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_init_peers()
1421 ret = perf_setup_peer_mw(&perf->peers[pidx]); in perf_init_peers()
1426 dev_dbg(&perf->ntb->dev, "Global port index %d\n", perf->gidx); in perf_init_peers()
1433 struct perf_ctx *perf; in perf_probe() local
1436 perf = perf_create_data(ntb); in perf_probe()
1437 if (IS_ERR(perf)) in perf_probe()
1438 return PTR_ERR(perf); in perf_probe()
1440 ret = perf_init_peers(perf); in perf_probe()
1444 perf_init_threads(perf); in perf_probe()
1446 ret = perf_init_service(perf); in perf_probe()
1450 ret = perf_enable_service(perf); in perf_probe()
1454 perf_setup_dbgfs(perf); in perf_probe()
1461 struct perf_ctx *perf = ntb->ctx; in perf_remove() local
1463 perf_clear_dbgfs(perf); in perf_remove()
1465 perf_disable_service(perf); in perf_remove()
1467 perf_clear_threads(perf); in perf_remove()