/Linux-v6.1/tools/perf/util/ |
D | thread_map.c | 35 struct perf_thread_map *threads; in thread_map__new_by_pid() local 46 threads = thread_map__alloc(items); in thread_map__new_by_pid() 47 if (threads != NULL) { in thread_map__new_by_pid() 49 perf_thread_map__set_pid(threads, i, atoi(namelist[i]->d_name)); in thread_map__new_by_pid() 50 threads->nr = items; in thread_map__new_by_pid() 51 refcount_set(&threads->refcnt, 1); in thread_map__new_by_pid() 58 return threads; in thread_map__new_by_pid() 63 struct perf_thread_map *threads = thread_map__alloc(1); in thread_map__new_by_tid() local 65 if (threads != NULL) { in thread_map__new_by_tid() 66 perf_thread_map__set_pid(threads, 0, tid); in thread_map__new_by_tid() [all …]
|
/Linux-v6.1/tools/perf/tests/ |
D | thread-map.c | 68 struct perf_thread_map *threads; in process_event() local 74 threads = thread_map__new_event(&event->thread_map); in process_event() 75 TEST_ASSERT_VAL("failed to alloc map", threads); in process_event() 77 TEST_ASSERT_VAL("wrong nr", threads->nr == 1); in process_event() 79 perf_thread_map__pid(threads, 0) == getpid()); in process_event() 81 perf_thread_map__comm(threads, 0) && in process_event() 82 !strcmp(perf_thread_map__comm(threads, 0), NAME)); in process_event() 84 refcount_read(&threads->refcnt) == 1); in process_event() 85 perf_thread_map__put(threads); in process_event() 91 struct perf_thread_map *threads; in test__thread_map_synthesize() local [all …]
|
D | mmap-thread-lookup.c | 21 #define THREADS 4 macro 32 static struct thread_data threads[THREADS]; variable 81 struct thread_data *td = &threads[i]; in thread_create() 101 struct thread_data *td0 = &threads[0]; in threads_create() 110 for (i = 1; !err && i < THREADS; i++) in threads_create() 118 struct thread_data *td0 = &threads[0]; in threads_destroy() 126 for (i = 1; !err && i < THREADS; i++) in threads_destroy() 127 err = pthread_join(threads[i].pt, NULL); in threads_destroy() 162 * The threads_create will not return before all threads in mmap_events() 168 TEST_ASSERT_VAL("failed to create threads", !threads_create()); in mmap_events() [all …]
|
/Linux-v6.1/tools/lib/perf/tests/ |
D | test-evsel.c | 56 struct perf_thread_map *threads; in test_stat_thread() local 64 threads = perf_thread_map__new_dummy(); in test_stat_thread() 65 __T("failed to create threads", threads); in test_stat_thread() 67 perf_thread_map__set_pid(threads, 0, 0); in test_stat_thread() 72 err = perf_evsel__open(evsel, NULL, threads); in test_stat_thread() 81 perf_thread_map__put(threads); in test_stat_thread() 88 struct perf_thread_map *threads; in test_stat_thread_enable() local 97 threads = perf_thread_map__new_dummy(); in test_stat_thread_enable() 98 __T("failed to create threads", threads); in test_stat_thread_enable() 100 perf_thread_map__set_pid(threads, 0, 0); in test_stat_thread_enable() [all …]
|
D | test-threadmap.c | 16 struct perf_thread_map *threads; in test_threadmap_array() local 19 threads = perf_thread_map__new_array(nr, array); in test_threadmap_array() 20 __T("Failed to allocate new thread map", threads); in test_threadmap_array() 22 __T("Unexpected number of threads", perf_thread_map__nr(threads) == nr); in test_threadmap_array() 26 perf_thread_map__pid(threads, i) == (array ? array[i] : -1)); in test_threadmap_array() 30 perf_thread_map__set_pid(threads, i, i * 100); in test_threadmap_array() 33 perf_thread_map__pid(threads, 0) == (array ? array[0] : -1)); in test_threadmap_array() 37 perf_thread_map__pid(threads, i) == i * 100); in test_threadmap_array() 40 perf_thread_map__put(threads); in test_threadmap_array() 48 struct perf_thread_map *threads; in test_threadmap() local [all …]
|
D | test-evlist.c | 95 struct perf_thread_map *threads; in test_stat_thread() local 108 threads = perf_thread_map__new_dummy(); in test_stat_thread() 109 __T("failed to create threads", threads); in test_stat_thread() 111 perf_thread_map__set_pid(threads, 0, 0); in test_stat_thread() 130 perf_evlist__set_maps(evlist, NULL, threads); in test_stat_thread() 143 perf_thread_map__put(threads); in test_stat_thread() 150 struct perf_thread_map *threads; in test_stat_thread_enable() local 165 threads = perf_thread_map__new_dummy(); in test_stat_thread_enable() 166 __T("failed to create threads", threads); in test_stat_thread_enable() 168 perf_thread_map__set_pid(threads, 0, 0); in test_stat_thread_enable() [all …]
|
/Linux-v6.1/tools/lib/perf/ |
D | threadmap.c | 47 struct perf_thread_map *threads = thread_map__alloc(nr_threads); in perf_thread_map__new_array() local 50 if (!threads) in perf_thread_map__new_array() 54 perf_thread_map__set_pid(threads, i, array ? array[i] : -1); in perf_thread_map__new_array() 56 threads->nr = nr_threads; in perf_thread_map__new_array() 57 refcount_set(&threads->refcnt, 1); in perf_thread_map__new_array() 59 return threads; in perf_thread_map__new_array() 67 static void perf_thread_map__delete(struct perf_thread_map *threads) in perf_thread_map__delete() argument 69 if (threads) { in perf_thread_map__delete() 72 WARN_ONCE(refcount_read(&threads->refcnt) != 0, in perf_thread_map__delete() 74 for (i = 0; i < threads->nr; i++) in perf_thread_map__delete() [all …]
|
/Linux-v6.1/tools/perf/bench/ |
D | breakpoint.c | 33 OPT_UINTEGER('t', "threads", &thread_params.nthreads, "Specify amount of threads"), 83 pthread_t *threads; in breakpoint_thread() local 85 threads = calloc(thread_params.nthreads, sizeof(threads[0])); in breakpoint_thread() 86 if (!threads) in breakpoint_thread() 92 if (pthread_create(&threads[i], NULL, passive_thread, &done)) in breakpoint_thread() 98 pthread_join(threads[i], NULL); in breakpoint_thread() 100 free(threads); in breakpoint_thread() 105 // then starts nparallel threads which create and join bench_repeat batches of nthreads threads. 143 printf("# Created/joined %d threads with %d breakpoints and %d parallelism\n", in bench_breakpoint_thread() 173 OPT_UINTEGER('p', "passive", &enable_params.npassive, "Specify amount of passive threads"), [all …]
|
D | synthesize.c | 34 OPT_UINTEGER('m', "min-threads", &min_threads, 35 "Minimum number of threads in multithreaded bench"), 36 OPT_UINTEGER('M', "max-threads", &max_threads, 37 "Maximum number of threads in multithreaded bench"), 62 struct perf_thread_map *threads, in do_run_single_threaded() argument 81 target, threads, in do_run_single_threaded() 116 struct perf_thread_map *threads; in run_single_threaded() local 125 threads = thread_map__new_by_pid(getpid()); in run_single_threaded() 126 if (!threads) { in run_single_threaded() 136 err = do_run_single_threaded(session, threads, &target, false); in run_single_threaded() [all …]
|
D | futex-wake.c | 5 * futex-wake: Block a bunch of threads on a futex and wake'em up, N at a time. 33 /* all threads will block on the same futex */ 53 OPT_UINTEGER('t', "threads", ¶ms.nthreads, "Specify amount of threads"), 54 OPT_UINTEGER('w', "nwakes", ¶ms.nwakes, "Specify amount of threads to wake at once"), 91 printf("Wokeup %d of %d threads in %.4f ms (+-%.2f%%)\n", in print_summary() 111 /* create and block all threads */ in block_threads() 174 printf("Run summary [PID %d]: blocking on %d threads (at [%s] futex %p), " in bench_futex_wake() 190 /* create, launch & block all threads */ in bench_futex_wake() 193 /* make sure all threads are already blocked */ in bench_futex_wake() 202 /* Ok, all threads are patiently blocked, start waking folks up */ in bench_futex_wake() [all …]
|
D | futex-wake-parallel.c | 5 * Block a bunch of threads and let parallel waker threads wakeup an 48 /* all threads will block on the same futex -- hash bucket chaos ;) */ 63 OPT_UINTEGER('t', "threads", ¶ms.nthreads, "Specify amount of threads"), 64 OPT_UINTEGER('w', "nwakers", ¶ms.nwakes, "Specify amount of waking threads"), 106 /* create and block all threads */ in wakeup_threads() 159 /* create and block all threads */ in block_threads() 195 printf("[Run %d]: Avg per-thread latency (waking %d/%d threads) " in print_run() 210 printf("Avg per-thread latency (waking %d/%d threads) in %.4f ms (+-%.2f%%)\n", in print_summary() 289 printf("Run summary [PID %d]: blocking on %d threads (at [%s] " in bench_futex_wake_parallel() 290 "futex %p), %d threads waking up %d at a time.\n\n", in bench_futex_wake_parallel() [all …]
|
/Linux-v6.1/Documentation/driver-api/dmaengine/ |
D | dmatest.rst | 16 test multiple channels at the same time, and it can start multiple threads 73 (shared) parameters used for all threads will use the new values. 74 After the channels are specified, each thread is set as pending. All threads 82 Once started a message like " dmatest: Added 1 threads using dma0chan0" is 171 dmatest: Added 1 threads using dma0chan2 179 dmatest: Added 1 threads using dma0chan1 181 dmatest: Added 1 threads using dma0chan2 191 dmatest: Added 1 threads using dma0chan0 192 dmatest: Added 1 threads using dma0chan3 193 dmatest: Added 1 threads using dma0chan4 [all …]
|
/Linux-v6.1/Documentation/x86/ |
D | topology.rst | 24 threads, cores, packages, etc. 37 - threads 95 A core consists of 1 or more threads. It does not matter whether the threads 96 are SMT- or CMT-type threads. 105 The number of threads in a core. The number of threads in a package can be 111 Threads chapter 116 AMDs nomenclature for CMT threads is "Compute Unit Core". The kernel always 123 The cpumask contains all online threads in the package to which a thread 126 The number of online threads is also printed in /proc/cpuinfo "siblings." 130 The cpumask contains all online threads in the core to which a thread [all …]
|
/Linux-v6.1/Documentation/power/ |
D | freezing-of-tasks.rst | 11 kernel threads are controlled during hibernation or system-wide suspend (on some 19 PF_NOFREEZE unset (all user space processes and some kernel threads) are 29 fake signal to all user space processes, and wakes up all the kernel threads. 37 frozen before kernel threads. 44 signal-handling code, but the freezable kernel threads need to call it 64 threads must call try_to_freeze() somewhere or use one of the 80 - freezes all tasks (including kernel threads) because we can't freeze 81 kernel threads without freezing userspace tasks 84 - thaws only kernel threads; this is particularly useful if we need to do 85 anything special in between thawing of kernel threads and thawing of [all …]
|
/Linux-v6.1/include/uapi/linux/ |
D | membarrier.h | 34 * @MEMBARRIER_CMD_GLOBAL: Execute a memory barrier on all running threads. 36 * is ensured that all running threads have passed 40 * (non-running threads are de facto in such a 41 * state). This covers threads from all processes 44 * Execute a memory barrier on all running threads 48 * is ensured that all running threads have passed 52 * (non-running threads are de facto in such a 53 * state). This only covers threads from processes 70 * threads siblings have passed through a state 74 * (non-running threads are de facto in such a [all …]
|
/Linux-v6.1/tools/testing/selftests/powerpc/dscr/ |
D | dscr_default_test.c | 6 * it's sysfs interface and then verifies that all threads 16 static unsigned long result[THREADS]; 62 pthread_t threads[THREADS]; in dscr_default() local 63 unsigned long i, *status[THREADS]; in dscr_default() 74 /* Spawn all testing threads */ in dscr_default() 75 for (i = 0; i < THREADS; i++) { in dscr_default() 76 if (pthread_create(&threads[i], NULL, do_test, (void *)i)) { in dscr_default() 104 for (i = 0; i < THREADS; i++) { in dscr_default() 105 if (pthread_join(threads[i], (void **)&(status[i]))) { in dscr_default()
|
/Linux-v6.1/tools/testing/selftests/vm/ |
D | migration.c | 24 pthread_t *threads; in FIXTURE() local 50 self->threads = malloc(self->nthreads * sizeof(*self->threads)); in FIXTURE_SETUP() 51 ASSERT_NE(self->threads, NULL); in FIXTURE_SETUP() 58 free(self->threads); in FIXTURE_TEARDOWN() 111 * between nodes whilst other threads try and access them triggering the 120 SKIP(return, "Not enough threads or NUMA nodes available"); 128 if (pthread_create(&self->threads[i], NULL, access_mem, ptr)) 133 ASSERT_EQ(pthread_cancel(self->threads[i]), 0); 146 SKIP(return, "Not enough threads or NUMA nodes available"); 175 SKIP(return, "Not enough threads or NUMA nodes available"); [all …]
|
/Linux-v6.1/arch/powerpc/kvm/ |
D | book3s_hv_ras.c | 257 * - On TB error, HMI interrupt is reported on all the threads of the core 266 * All threads need to co-ordinate before making opal hmi handler. 267 * All threads will use sibling_subcore_state->in_guest[] (shared by all 268 * threads in the core) in paca which holds information about whether 271 * subcore status. Only primary threads from each subcore is responsible 278 * primary threads to decide who takes up the responsibility. 285 * - All other threads which are in host will call 290 * - Once all primary threads clear in_guest[0-3], all of them will invoke 292 * - Now all threads will wait for TB resync to complete by invoking 296 * - All other threads will now come out of resync wait loop and proceed [all …]
|
/Linux-v6.1/samples/pktgen/ |
D | pktgen_sample06_numa_awared_queue_irq_affinity.sh | 3 # Multiqueue: Using pktgen threads for sending on multiple CPUs 4 # * adding devices to kernel threads which are in the same NUMA node 5 # * bound devices queue's irq affinity to the threads, 1:1 mapping 29 [ $THREADS -gt ${#irq_array[*]} -o $THREADS -gt ${#cpu_array[*]} ] && \ 30 err 1 "Thread number $THREADS exceeds: min (${#irq_array[*]},${#cpu_array[*]})" 49 # Threads are specified with parameter -t value in $THREADS 50 for ((i = 0; i < $THREADS; i++)); do 106 for ((i = 0; i < $THREADS; i++)); do
|
/Linux-v6.1/tools/testing/selftests/dma/ |
D | dma_map_benchmark.c | 29 int threads = 1, seconds = 20, node = -1; in main() local 41 threads = atoi(optarg); in main() 66 if (threads <= 0 || threads > DMA_MAP_MAX_THREADS) { in main() 67 fprintf(stderr, "invalid number of threads, must be in 1-%d\n", in main() 109 map.threads = threads; in main() 121 printf("dma mapping benchmark: threads:%d seconds:%d node:%d dir:%s granule: %d\n", in main() 122 threads, seconds, node, dir[directions], granule); in main()
|
/Linux-v6.1/tools/testing/selftests/powerpc/math/ |
D | fpu_signal.c | 26 * worker threads 77 int i, j, rc, threads; in test_signal_fpu() local 81 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_signal_fpu() 82 tids = malloc(threads * sizeof(pthread_t)); in test_signal_fpu() 86 threads_starting = threads; in test_signal_fpu() 87 for (i = 0; i < threads; i++) { in test_signal_fpu() 98 printf("\tSending signals to all threads %d times...", ITERATIONS); in test_signal_fpu() 100 for (j = 0; j < threads; j++) { in test_signal_fpu() 109 for (i = 0; i < threads; i++) { in test_signal_fpu()
|
D | fpu_preempt.c | 8 * no way to be sure preemption happened so this test just uses many threads 28 * worker threads 57 int i, rc, threads; in test_preempt_fpu() local 60 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_preempt_fpu() 61 tids = malloc((threads) * sizeof(pthread_t)); in test_preempt_fpu() 65 threads_starting = threads; in test_preempt_fpu() 66 for (i = 0; i < threads; i++) { in test_preempt_fpu() 88 for (i = 0; i < threads; i++) { in test_preempt_fpu()
|
D | vmx_signal.c | 28 * worker threads 95 int i, j, rc, threads; in test_signal_vmx() local 102 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_signal_vmx() 103 tids = malloc(threads * sizeof(pthread_t)); in test_signal_vmx() 107 threads_starting = threads; in test_signal_vmx() 108 for (i = 0; i < threads; i++) { in test_signal_vmx() 114 printf("\tWaiting for %d workers to start... %d", threads, threads_starting); in test_signal_vmx() 122 printf("\tSending signals to all threads %d times...", ITERATIONS); in test_signal_vmx() 124 for (j = 0; j < threads; j++) { in test_signal_vmx() 133 for (i = 0; i < threads; i++) { in test_signal_vmx()
|
/Linux-v6.1/tools/lib/perf/Documentation/examples/ |
D | counting.c | 23 struct perf_thread_map *threads; in main() local 40 threads = perf_thread_map__new_dummy(); in main() 41 if (!threads) { in main() 42 fprintf(stderr, "failed to create threads\n"); in main() 45 perf_thread_map__set_pid(threads, 0, 0); in main() 63 perf_evlist__set_maps(evlist, NULL, threads); in main() 81 perf_thread_map__put(threads); in main()
|
/Linux-v6.1/tools/perf/python/ |
D | twatch.py | 13 threads = perf.thread_map(thread) 21 """What we want are just the PERF_RECORD_ lifetime events for threads, 25 threads comes and goes... So use (perf.TYPE_SOFTWARE, perf_COUNT_SW_DUMMY, 28 evsel.open(cpus = cpus, threads = threads); 29 evlist = perf.evlist(cpus, threads) 56 to figure out if this is a context switch in or out of the monitored threads.
|