Lines Matching refs:perf
144 struct perf_ctx *perf; member
166 struct perf_ctx *perf; member
202 int (*cmd_recv)(struct perf_ctx *perf, int *pidx, enum perf_cmd *cmd,
244 static void perf_terminate_test(struct perf_ctx *perf);
250 link = ntb_link_is_up(peer->perf->ntb, NULL, NULL); in perf_link_is_up()
257 struct perf_ctx *perf = peer->perf; in perf_spad_cmd_send() local
261 dev_dbg(&perf->ntb->dev, "CMD send: %d 0x%llx\n", cmd, data); in perf_spad_cmd_send()
274 sts = ntb_peer_spad_read(perf->ntb, peer->pidx, in perf_spad_cmd_send()
275 PERF_SPAD_CMD(perf->gidx)); in perf_spad_cmd_send()
281 ntb_peer_spad_write(perf->ntb, peer->pidx, in perf_spad_cmd_send()
282 PERF_SPAD_LDATA(perf->gidx), in perf_spad_cmd_send()
284 ntb_peer_spad_write(perf->ntb, peer->pidx, in perf_spad_cmd_send()
285 PERF_SPAD_HDATA(perf->gidx), in perf_spad_cmd_send()
287 ntb_peer_spad_write(perf->ntb, peer->pidx, in perf_spad_cmd_send()
288 PERF_SPAD_CMD(perf->gidx), in perf_spad_cmd_send()
290 ntb_peer_db_set(perf->ntb, PERF_SPAD_NOTIFY(peer->gidx)); in perf_spad_cmd_send()
292 dev_dbg(&perf->ntb->dev, "DB ring peer %#llx\n", in perf_spad_cmd_send()
301 static int perf_spad_cmd_recv(struct perf_ctx *perf, int *pidx, in perf_spad_cmd_recv() argument
307 ntb_db_clear(perf->ntb, PERF_SPAD_NOTIFY(perf->gidx)); in perf_spad_cmd_recv()
315 for (*pidx = 0; *pidx < perf->pcnt; (*pidx)++) { in perf_spad_cmd_recv()
316 peer = &perf->peers[*pidx]; in perf_spad_cmd_recv()
321 val = ntb_spad_read(perf->ntb, PERF_SPAD_CMD(peer->gidx)); in perf_spad_cmd_recv()
327 val = ntb_spad_read(perf->ntb, PERF_SPAD_LDATA(peer->gidx)); in perf_spad_cmd_recv()
330 val = ntb_spad_read(perf->ntb, PERF_SPAD_HDATA(peer->gidx)); in perf_spad_cmd_recv()
334 ntb_spad_write(perf->ntb, PERF_SPAD_CMD(peer->gidx), in perf_spad_cmd_recv()
337 dev_dbg(&perf->ntb->dev, "CMD recv: %d 0x%llx\n", *cmd, *data); in perf_spad_cmd_recv()
348 struct perf_ctx *perf = peer->perf; in perf_msg_cmd_send() local
352 dev_dbg(&perf->ntb->dev, "CMD send: %d 0x%llx\n", cmd, data); in perf_msg_cmd_send()
361 outbits = ntb_msg_outbits(perf->ntb); in perf_msg_cmd_send()
366 ret = ntb_msg_clear_sts(perf->ntb, outbits); in perf_msg_cmd_send()
370 ntb_peer_msg_write(perf->ntb, peer->pidx, PERF_MSG_LDATA, in perf_msg_cmd_send()
373 if (ntb_msg_read_sts(perf->ntb) & outbits) { in perf_msg_cmd_send()
378 ntb_peer_msg_write(perf->ntb, peer->pidx, PERF_MSG_HDATA, in perf_msg_cmd_send()
382 ntb_peer_msg_write(perf->ntb, peer->pidx, PERF_MSG_CMD, cmd); in perf_msg_cmd_send()
390 static int perf_msg_cmd_recv(struct perf_ctx *perf, int *pidx, in perf_msg_cmd_recv() argument
396 inbits = ntb_msg_inbits(perf->ntb); in perf_msg_cmd_recv()
398 if (hweight64(ntb_msg_read_sts(perf->ntb) & inbits) < 3) in perf_msg_cmd_recv()
401 val = ntb_msg_read(perf->ntb, pidx, PERF_MSG_CMD); in perf_msg_cmd_recv()
404 val = ntb_msg_read(perf->ntb, pidx, PERF_MSG_LDATA); in perf_msg_cmd_recv()
407 val = ntb_msg_read(perf->ntb, pidx, PERF_MSG_HDATA); in perf_msg_cmd_recv()
411 ntb_msg_clear_sts(perf->ntb, inbits); in perf_msg_cmd_recv()
413 dev_dbg(&perf->ntb->dev, "CMD recv: %d 0x%llx\n", *cmd, *data); in perf_msg_cmd_recv()
420 struct perf_ctx *perf = peer->perf; in perf_cmd_send() local
423 return perf->cmd_send(peer, cmd, data); in perf_cmd_send()
425 dev_err(&perf->ntb->dev, "Send invalid command\n"); in perf_cmd_send()
439 dev_err(&peer->perf->ntb->dev, "Exec invalid command\n"); in perf_cmd_exec()
446 dev_dbg(&peer->perf->ntb->dev, "CMD exec: %d\n", cmd); in perf_cmd_exec()
453 static int perf_cmd_recv(struct perf_ctx *perf) in perf_cmd_recv() argument
459 while (!(ret = perf->cmd_recv(perf, &pidx, &cmd, &data))) { in perf_cmd_recv()
460 peer = &perf->peers[pidx]; in perf_cmd_recv()
470 dev_err(&perf->ntb->dev, "Recv invalid command\n"); in perf_cmd_recv()
481 struct perf_ctx *perf = ctx; in perf_link_event() local
486 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_link_event()
487 peer = &perf->peers[pidx]; in perf_link_event()
503 struct perf_ctx *perf = ctx; in perf_db_event() local
505 dev_dbg(&perf->ntb->dev, "DB vec %d mask %#llx bits %#llx\n", vec, in perf_db_event()
506 ntb_db_vector_mask(perf->ntb, vec), ntb_db_read(perf->ntb)); in perf_db_event()
509 (void)perf_cmd_recv(perf); in perf_db_event()
514 struct perf_ctx *perf = ctx; in perf_msg_event() local
516 dev_dbg(&perf->ntb->dev, "Msg status bits %#llx\n", in perf_msg_event()
517 ntb_msg_read_sts(perf->ntb)); in perf_msg_event()
520 (void)perf_cmd_recv(perf); in perf_msg_event()
531 (void)ntb_peer_mw_clear_trans(peer->perf->ntb, peer->pidx, peer->gidx); in perf_free_outbuf()
536 struct perf_ctx *perf = peer->perf; in perf_setup_outbuf() local
540 ret = ntb_peer_mw_set_trans(perf->ntb, peer->pidx, peer->gidx, in perf_setup_outbuf()
543 dev_err(&perf->ntb->dev, "Failed to set outbuf translation\n"); in perf_setup_outbuf()
558 (void)ntb_mw_clear_trans(peer->perf->ntb, peer->pidx, peer->gidx); in perf_free_inbuf()
559 dma_free_coherent(&peer->perf->ntb->dev, peer->inbuf_size, in perf_free_inbuf()
567 struct perf_ctx *perf = peer->perf; in perf_setup_inbuf() local
571 ret = ntb_mw_get_align(perf->ntb, peer->pidx, perf->gidx, in perf_setup_inbuf()
574 dev_err(&perf->ntb->dev, "Couldn't get inbuf restrictions\n"); in perf_setup_inbuf()
579 dev_err(&perf->ntb->dev, "Too big inbuf size %pa > %pa\n", in perf_setup_inbuf()
588 peer->inbuf = dma_alloc_coherent(&perf->ntb->dev, peer->inbuf_size, in perf_setup_inbuf()
591 dev_err(&perf->ntb->dev, "Failed to alloc inbuf of %pa\n", in perf_setup_inbuf()
596 dev_err(&perf->ntb->dev, "Unaligned inbuf allocated\n"); in perf_setup_inbuf()
600 ret = ntb_mw_set_trans(perf->ntb, peer->pidx, peer->gidx, in perf_setup_inbuf()
603 dev_err(&perf->ntb->dev, "Failed to set inbuf translation\n"); in perf_setup_inbuf()
640 if (test_bit(0, &peer->perf->busy_flag) && in perf_service_work()
641 peer == peer->perf->test_peer) { in perf_service_work()
642 dev_warn(&peer->perf->ntb->dev, in perf_service_work()
644 perf_terminate_test(peer->perf); in perf_service_work()
651 static int perf_init_service(struct perf_ctx *perf) in perf_init_service() argument
655 if (ntb_peer_mw_count(perf->ntb) < perf->pcnt + 1) { in perf_init_service()
656 dev_err(&perf->ntb->dev, "Not enough memory windows\n"); in perf_init_service()
660 if (ntb_msg_count(perf->ntb) >= PERF_MSG_CNT) { in perf_init_service()
661 perf->cmd_send = perf_msg_cmd_send; in perf_init_service()
662 perf->cmd_recv = perf_msg_cmd_recv; in perf_init_service()
664 dev_dbg(&perf->ntb->dev, "Message service initialized\n"); in perf_init_service()
669 dev_dbg(&perf->ntb->dev, "Message service unsupported\n"); in perf_init_service()
671 mask = GENMASK_ULL(perf->pcnt, 0); in perf_init_service()
672 if (ntb_spad_count(perf->ntb) >= PERF_SPAD_CNT(perf->pcnt) && in perf_init_service()
673 (ntb_db_valid_mask(perf->ntb) & mask) == mask) { in perf_init_service()
674 perf->cmd_send = perf_spad_cmd_send; in perf_init_service()
675 perf->cmd_recv = perf_spad_cmd_recv; in perf_init_service()
677 dev_dbg(&perf->ntb->dev, "Scratchpad service initialized\n"); in perf_init_service()
682 dev_dbg(&perf->ntb->dev, "Scratchpad service unsupported\n"); in perf_init_service()
684 dev_err(&perf->ntb->dev, "Command services unsupported\n"); in perf_init_service()
689 static int perf_enable_service(struct perf_ctx *perf) in perf_enable_service() argument
694 mask = ntb_db_valid_mask(perf->ntb); in perf_enable_service()
695 (void)ntb_db_set_mask(perf->ntb, mask); in perf_enable_service()
697 ret = ntb_set_ctx(perf->ntb, perf, &perf_ops); in perf_enable_service()
701 if (perf->cmd_send == perf_msg_cmd_send) { in perf_enable_service()
704 inbits = ntb_msg_inbits(perf->ntb); in perf_enable_service()
705 outbits = ntb_msg_outbits(perf->ntb); in perf_enable_service()
706 (void)ntb_msg_set_mask(perf->ntb, inbits | outbits); in perf_enable_service()
709 ret = ntb_msg_clear_mask(perf->ntb, incmd_bit); in perf_enable_service()
711 dev_dbg(&perf->ntb->dev, "MSG sts unmasked %#llx\n", incmd_bit); in perf_enable_service()
713 scnt = ntb_spad_count(perf->ntb); in perf_enable_service()
715 ntb_spad_write(perf->ntb, sidx, PERF_CMD_INVAL); in perf_enable_service()
716 incmd_bit = PERF_SPAD_NOTIFY(perf->gidx); in perf_enable_service()
717 ret = ntb_db_clear_mask(perf->ntb, incmd_bit); in perf_enable_service()
719 dev_dbg(&perf->ntb->dev, "DB bits unmasked %#llx\n", incmd_bit); in perf_enable_service()
722 ntb_clear_ctx(perf->ntb); in perf_enable_service()
726 ntb_link_enable(perf->ntb, NTB_SPEED_AUTO, NTB_WIDTH_AUTO); in perf_enable_service()
728 ntb_link_event(perf->ntb); in perf_enable_service()
733 static void perf_disable_service(struct perf_ctx *perf) in perf_disable_service() argument
737 if (perf->cmd_send == perf_msg_cmd_send) { in perf_disable_service()
740 inbits = ntb_msg_inbits(perf->ntb); in perf_disable_service()
741 (void)ntb_msg_set_mask(perf->ntb, inbits); in perf_disable_service()
743 (void)ntb_db_set_mask(perf->ntb, PERF_SPAD_NOTIFY(perf->gidx)); in perf_disable_service()
746 ntb_clear_ctx(perf->ntb); in perf_disable_service()
748 for (pidx = 0; pidx < perf->pcnt; pidx++) in perf_disable_service()
749 perf_cmd_exec(&perf->peers[pidx], PERF_CMD_CLEAR); in perf_disable_service()
751 for (pidx = 0; pidx < perf->pcnt; pidx++) in perf_disable_service()
752 flush_work(&perf->peers[pidx].service); in perf_disable_service()
754 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_disable_service()
755 struct perf_peer *peer = &perf->peers[pidx]; in perf_disable_service()
757 ntb_spad_write(perf->ntb, PERF_SPAD_CMD(peer->gidx), 0); in perf_disable_service()
760 ntb_db_clear(perf->ntb, PERF_SPAD_NOTIFY(perf->gidx)); in perf_disable_service()
762 ntb_link_disable(perf->ntb); in perf_disable_service()
846 return likely(atomic_read(&pthr->perf->tsync) > 0) ? 0 : -EINTR; in perf_copy_chunk()
856 struct perf_ctx *perf = data; in perf_dma_filter() local
859 node = dev_to_node(&perf->ntb->dev); in perf_dma_filter()
866 struct perf_ctx *perf = pthr->perf; in perf_init_test() local
869 pthr->src = kmalloc_node(perf->test_peer->outbuf_size, GFP_KERNEL, in perf_init_test()
870 dev_to_node(&perf->ntb->dev)); in perf_init_test()
874 get_random_bytes(pthr->src, perf->test_peer->outbuf_size); in perf_init_test()
881 pthr->dma_chan = dma_request_channel(dma_mask, perf_dma_filter, perf); in perf_init_test()
883 dev_err(&perf->ntb->dev, "%d: Failed to get DMA channel\n", in perf_init_test()
885 atomic_dec(&perf->tsync); in perf_init_test()
886 wake_up(&perf->twait); in perf_init_test()
898 struct perf_peer *peer = pthr->perf->test_peer; in perf_run_test()
899 struct perf_ctx *perf = pthr->perf; in perf_run_test() local
919 dev_err(&perf->ntb->dev, "%d: Got error %d on test\n", in perf_run_test()
942 struct perf_ctx *perf = pthr->perf; in perf_sync_test() local
949 atomic_read(&perf->tsync) < 0)); in perf_sync_test()
951 if (atomic_read(&perf->tsync) < 0) in perf_sync_test()
957 dev_dbg(&perf->ntb->dev, "%d: copied %llu bytes\n", in perf_sync_test()
960 dev_dbg(&perf->ntb->dev, "%d: lasted %llu usecs\n", in perf_sync_test()
963 dev_dbg(&perf->ntb->dev, "%d: %llu MBytes/s\n", pthr->tidx, in perf_sync_test()
971 struct perf_ctx *perf = pthr->perf; in perf_clear_test() local
985 atomic_dec(&perf->tsync); in perf_clear_test()
986 wake_up(&perf->twait); in perf_clear_test()
1020 static int perf_set_tcnt(struct perf_ctx *perf, u8 tcnt) in perf_set_tcnt() argument
1025 if (test_and_set_bit_lock(0, &perf->busy_flag)) in perf_set_tcnt()
1028 perf->tcnt = tcnt; in perf_set_tcnt()
1030 clear_bit_unlock(0, &perf->busy_flag); in perf_set_tcnt()
1035 static void perf_terminate_test(struct perf_ctx *perf) in perf_terminate_test() argument
1039 atomic_set(&perf->tsync, -1); in perf_terminate_test()
1040 wake_up(&perf->twait); in perf_terminate_test()
1043 wake_up(&perf->threads[tidx].dma_wait); in perf_terminate_test()
1044 cancel_work_sync(&perf->threads[tidx].work); in perf_terminate_test()
1050 struct perf_ctx *perf = peer->perf; in perf_submit_test() local
1057 if (test_and_set_bit_lock(0, &perf->busy_flag)) in perf_submit_test()
1060 perf->test_peer = peer; in perf_submit_test()
1061 atomic_set(&perf->tsync, perf->tcnt); in perf_submit_test()
1064 pthr = &perf->threads[tidx]; in perf_submit_test()
1069 if (tidx < perf->tcnt) in perf_submit_test()
1073 ret = wait_event_interruptible(perf->twait, in perf_submit_test()
1074 atomic_read(&perf->tsync) <= 0); in perf_submit_test()
1076 perf_terminate_test(perf); in perf_submit_test()
1080 clear_bit_unlock(0, &perf->busy_flag); in perf_submit_test()
1085 static int perf_read_stats(struct perf_ctx *perf, char *buf, in perf_read_stats() argument
1091 if (test_and_set_bit_lock(0, &perf->busy_flag)) in perf_read_stats()
1095 " Peer %d test statistics:\n", perf->test_peer->pidx); in perf_read_stats()
1098 pthr = &perf->threads[tidx]; in perf_read_stats()
1115 clear_bit_unlock(0, &perf->busy_flag); in perf_read_stats()
1120 static void perf_init_threads(struct perf_ctx *perf) in perf_init_threads() argument
1125 perf->tcnt = DEF_THREADS_CNT; in perf_init_threads()
1126 perf->test_peer = &perf->peers[0]; in perf_init_threads()
1127 init_waitqueue_head(&perf->twait); in perf_init_threads()
1130 pthr = &perf->threads[tidx]; in perf_init_threads()
1132 pthr->perf = perf; in perf_init_threads()
1140 static void perf_clear_threads(struct perf_ctx *perf) in perf_clear_threads() argument
1142 perf_terminate_test(perf); in perf_clear_threads()
1153 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_read_info() local
1170 "Local port %d, Global index %d\n", ntb_port_number(perf->ntb), in perf_dbgfs_read_info()
1171 perf->gidx); in perf_dbgfs_read_info()
1173 if (test_bit(0, &perf->busy_flag)) { in perf_dbgfs_read_info()
1176 ntb_peer_port_number(perf->ntb, perf->test_peer->pidx), in perf_dbgfs_read_info()
1177 perf->test_peer->pidx); in perf_dbgfs_read_info()
1182 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_dbgfs_read_info()
1183 peer = &perf->peers[pidx]; in perf_dbgfs_read_info()
1187 ntb_peer_port_number(perf->ntb, peer->pidx), peer->pidx, in perf_dbgfs_read_info()
1233 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_read_run() local
1241 ret = perf_read_stats(perf, buf, PERF_BUF_LEN, &pos); in perf_dbgfs_read_run()
1255 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_write_run() local
1263 if (pidx < 0 || pidx >= perf->pcnt) in perf_dbgfs_write_run()
1266 peer = &perf->peers[pidx]; in perf_dbgfs_write_run()
1284 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_read_tcnt() local
1288 pos = scnprintf(buf, sizeof(buf), "%hhu\n", perf->tcnt); in perf_dbgfs_read_tcnt()
1297 struct perf_ctx *perf = filep->private_data; in perf_dbgfs_write_tcnt() local
1305 ret = perf_set_tcnt(perf, val); in perf_dbgfs_write_tcnt()
1318 static void perf_setup_dbgfs(struct perf_ctx *perf) in perf_setup_dbgfs() argument
1320 struct pci_dev *pdev = perf->ntb->pdev; in perf_setup_dbgfs()
1322 perf->dbgfs_dir = debugfs_create_dir(pci_name(pdev), perf_dbgfs_topdir); in perf_setup_dbgfs()
1323 if (!perf->dbgfs_dir) { in perf_setup_dbgfs()
1324 dev_warn(&perf->ntb->dev, "DebugFS unsupported\n"); in perf_setup_dbgfs()
1328 debugfs_create_file("info", 0600, perf->dbgfs_dir, perf, in perf_setup_dbgfs()
1331 debugfs_create_file("run", 0600, perf->dbgfs_dir, perf, in perf_setup_dbgfs()
1334 debugfs_create_file("threads_count", 0600, perf->dbgfs_dir, perf, in perf_setup_dbgfs()
1338 debugfs_create_u8("chunk_order", 0500, perf->dbgfs_dir, &chunk_order); in perf_setup_dbgfs()
1340 debugfs_create_u8("total_order", 0500, perf->dbgfs_dir, &total_order); in perf_setup_dbgfs()
1342 debugfs_create_bool("use_dma", 0500, perf->dbgfs_dir, &use_dma); in perf_setup_dbgfs()
1345 static void perf_clear_dbgfs(struct perf_ctx *perf) in perf_clear_dbgfs() argument
1347 debugfs_remove_recursive(perf->dbgfs_dir); in perf_clear_dbgfs()
1357 struct perf_ctx *perf; in perf_create_data() local
1359 perf = devm_kzalloc(&ntb->dev, sizeof(*perf), GFP_KERNEL); in perf_create_data()
1360 if (!perf) in perf_create_data()
1363 perf->pcnt = ntb_peer_port_count(ntb); in perf_create_data()
1364 perf->peers = devm_kcalloc(&ntb->dev, perf->pcnt, sizeof(*perf->peers), in perf_create_data()
1366 if (!perf->peers) in perf_create_data()
1369 perf->ntb = ntb; in perf_create_data()
1371 return perf; in perf_create_data()
1376 struct perf_ctx *perf = peer->perf; in perf_setup_peer_mw() local
1381 ret = ntb_peer_mw_get_addr(perf->ntb, perf->gidx, &phys_addr, in perf_setup_peer_mw()
1386 peer->outbuf = devm_ioremap_wc(&perf->ntb->dev, phys_addr, in perf_setup_peer_mw()
1393 dev_warn(&peer->perf->ntb->dev, in perf_setup_peer_mw()
1401 static int perf_init_peers(struct perf_ctx *perf) in perf_init_peers() argument
1406 lport = ntb_port_number(perf->ntb); in perf_init_peers()
1407 perf->gidx = -1; in perf_init_peers()
1408 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_init_peers()
1409 peer = &perf->peers[pidx]; in perf_init_peers()
1411 peer->perf = perf; in perf_init_peers()
1413 if (lport < ntb_peer_port_number(perf->ntb, pidx)) { in perf_init_peers()
1414 if (perf->gidx == -1) in perf_init_peers()
1415 perf->gidx = pidx; in perf_init_peers()
1422 if (perf->gidx == -1) in perf_init_peers()
1423 perf->gidx = pidx; in perf_init_peers()
1425 for (pidx = 0; pidx < perf->pcnt; pidx++) { in perf_init_peers()
1426 ret = perf_setup_peer_mw(&perf->peers[pidx]); in perf_init_peers()
1431 dev_dbg(&perf->ntb->dev, "Global port index %d\n", perf->gidx); in perf_init_peers()
1438 struct perf_ctx *perf; in perf_probe() local
1441 perf = perf_create_data(ntb); in perf_probe()
1442 if (IS_ERR(perf)) in perf_probe()
1443 return PTR_ERR(perf); in perf_probe()
1445 ret = perf_init_peers(perf); in perf_probe()
1449 perf_init_threads(perf); in perf_probe()
1451 ret = perf_init_service(perf); in perf_probe()
1455 ret = perf_enable_service(perf); in perf_probe()
1459 perf_setup_dbgfs(perf); in perf_probe()
1466 struct perf_ctx *perf = ntb->ctx; in perf_remove() local
1468 perf_clear_dbgfs(perf); in perf_remove()
1470 perf_disable_service(perf); in perf_remove()
1472 perf_clear_threads(perf); in perf_remove()