| /Linux-v5.15/fs/xfs/ | 
| D | xfs_stats.h | 165 	per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++;	\166 	per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++;	\
 171 	per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--;	\
 172 	per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--;	\
 177 	per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc);	\
 178 	per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc);	\
 183 	per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++;	\
 184 	per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++;	\
 189 	per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off];	\
 190 	per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off];	\
 [all …]
 
 | 
| D | xfs_stats.c | 15 		val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx));  in counter_val()71 		xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes;  in xfs_stats_format()
 72 		xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes;  in xfs_stats_format()
 73 		xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes;  in xfs_stats_format()
 74 		defer_relog += per_cpu_ptr(stats, i)->s.defer_relog;  in xfs_stats_format()
 100 		vn_active = per_cpu_ptr(stats, c)->s.vn_active;  in xfs_stats_clearall()
 101 		memset(per_cpu_ptr(stats, c), 0, sizeof(*stats));  in xfs_stats_clearall()
 102 		per_cpu_ptr(stats, c)->s.vn_active = vn_active;  in xfs_stats_clearall()
 
 | 
| /Linux-v5.15/arch/x86/events/amd/ | 
| D | uncore.c | 75 		return *per_cpu_ptr(amd_uncore_nb, event->cpu);  in event_to_amd_uncore()77 		return *per_cpu_ptr(amd_uncore_llc, event->cpu);  in event_to_amd_uncore()
 388 		*per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb;  in amd_uncore_cpu_up_prepare()
 402 		*per_cpu_ptr(amd_uncore_llc, cpu) = uncore_llc;  in amd_uncore_cpu_up_prepare()
 409 		*per_cpu_ptr(amd_uncore_nb, cpu) = NULL;  in amd_uncore_cpu_up_prepare()
 422 		that = *per_cpu_ptr(uncores, cpu);  in amd_uncore_find_online_sibling()
 447 		uncore = *per_cpu_ptr(amd_uncore_nb, cpu);  in amd_uncore_cpu_starting()
 452 		*per_cpu_ptr(amd_uncore_nb, cpu) = uncore;  in amd_uncore_cpu_starting()
 456 		uncore = *per_cpu_ptr(amd_uncore_llc, cpu);  in amd_uncore_cpu_starting()
 460 		*per_cpu_ptr(amd_uncore_llc, cpu) = uncore;  in amd_uncore_cpu_starting()
 [all …]
 
 | 
| /Linux-v5.15/drivers/infiniband/ulp/rtrs/ | 
| D | rtrs-clt-stats.c | 27 		s = per_cpu_ptr(stats->pcpu_stats, con->cpu);  in rtrs_clt_update_wc_stats()54 		s = per_cpu_ptr(stats->pcpu_stats, cpu);  in rtrs_clt_stats_migration_cnt_to_str()
 61 		s = per_cpu_ptr(stats->pcpu_stats, cpu);  in rtrs_clt_stats_migration_cnt_to_str()
 88 		r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma;  in rtrs_clt_stats_rdma_to_str()
 118 		s = per_cpu_ptr(stats->pcpu_stats, cpu);  in rtrs_clt_reset_rdma_stats()
 134 		s = per_cpu_ptr(stats->pcpu_stats, cpu);  in rtrs_clt_reset_cpu_migr_stats()
 
 | 
| /Linux-v5.15/kernel/bpf/ | 
| D | percpu_freelist.c | 15 		struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu);  in pcpu_freelist_init()65 		head = per_cpu_ptr(s->freelist, cpu);  in ___pcpu_freelist_push_nmi()
 112 		head = per_cpu_ptr(s->freelist, cpu);  in pcpu_freelist_populate()
 132 		head = per_cpu_ptr(s->freelist, cpu);  in ___pcpu_freelist_pop()
 166 		head = per_cpu_ptr(s->freelist, cpu);  in ___pcpu_freelist_pop_nmi()
 
 | 
| D | bpf_lru_list.c | 408 	l = per_cpu_ptr(lru->percpu_lru, cpu);  in bpf_percpu_lru_pop_free()441 	loc_l = per_cpu_ptr(clru->local_list, cpu);  in bpf_common_lru_pop_free()
 470 		steal_loc_l = per_cpu_ptr(clru->local_list, steal);  in bpf_common_lru_pop_free()
 515 		loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu);  in bpf_common_lru_push_free()
 542 	l = per_cpu_ptr(lru->percpu_lru, node->cpu);  in bpf_percpu_lru_push_free()
 592 		l = per_cpu_ptr(lru->percpu_lru, cpu);  in bpf_percpu_lru_populate()
 659 			l = per_cpu_ptr(lru->percpu_lru, cpu);  in bpf_lru_init()
 673 			loc_l = per_cpu_ptr(clru->local_list, cpu);  in bpf_lru_init()
 
 | 
| /Linux-v5.15/kernel/irq/ | 
| D | matrix.c | 138 		cm = per_cpu_ptr(m->maps, cpu);  in matrix_find_best_cpu()159 		cm = per_cpu_ptr(m->maps, cpu);  in matrix_find_best_cpu_managed()
 215 		struct cpumap *cm = per_cpu_ptr(m->maps, cpu);  in irq_matrix_reserve_managed()
 256 		struct cpumap *cm = per_cpu_ptr(m->maps, cpu);  in irq_matrix_remove_managed()
 299 	cm = per_cpu_ptr(m->maps, cpu);  in irq_matrix_alloc_managed()
 394 	cm = per_cpu_ptr(m->maps, cpu);  in irq_matrix_alloc()
 421 	struct cpumap *cm = per_cpu_ptr(m->maps, cpu);  in irq_matrix_free()
 504 		struct cpumap *cm = per_cpu_ptr(m->maps, cpu);  in irq_matrix_debug_show()
 
 | 
| /Linux-v5.15/kernel/sched/ | 
| D | topology.c | 880 		sibling = *per_cpu_ptr(sdd->sd, i);  in build_balance_mask()939 	sg->sgc = *per_cpu_ptr(sdd->sgc, cpu);  in init_overlap_sched_group()
 999 		sibling = *per_cpu_ptr(sdd->sd, i);  in build_overlap_sched_groups()
 1149 	struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu);  in get_group()
 1157 	sg = *per_cpu_ptr(sdd->sg, cpu);  in get_group()
 1158 	sg->sgc = *per_cpu_ptr(sdd->sgc, cpu);  in get_group()
 1462 	WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd);  in claim_allocations()
 1463 	*per_cpu_ptr(sdd->sd, cpu) = NULL;  in claim_allocations()
 1465 	if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref))  in claim_allocations()
 1466 		*per_cpu_ptr(sdd->sds, cpu) = NULL;  in claim_allocations()
 [all …]
 
 | 
| D | cpuacct.c | 102 	struct cpuacct_usage *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);  in cpuacct_cpuusage_read()137 	struct cpuacct_usage *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);  in cpuacct_cpuusage_write()
 246 		struct cpuacct_usage *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);  in cpuacct_all_seq_show()
 279 		u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat;  in cpuacct_stats_show()
 
 | 
| /Linux-v5.15/kernel/ | 
| D | smpboot.c | 172 	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);  in __smpboot_create_thread()197 	*per_cpu_ptr(ht->store, cpu) = tsk;  in __smpboot_create_thread()
 230 	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);  in smpboot_unpark_thread()
 249 	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);  in smpboot_park_thread()
 272 		struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);  in smpboot_destroy_threads()
 277 			*per_cpu_ptr(ht->store, cpu) = NULL;  in smpboot_destroy_threads()
 
 | 
| D | relay.c | 207 	*per_cpu_ptr(chan->buf, buf->cpu) = NULL;  in relay_destroy_buf()330 	if (chan->is_global && (buf = *per_cpu_ptr(chan->buf, 0))) {  in relay_reset()
 337 		if ((buf = *per_cpu_ptr(chan->buf, i)))  in relay_reset()
 385 		return *per_cpu_ptr(chan->buf, 0);  in relay_open_buf()
 409 		*per_cpu_ptr(chan->buf, 0) = buf;  in relay_open_buf()
 443 		if ((buf = *per_cpu_ptr(chan->buf, cpu)))  in relay_prepare_cpu()
 451 		*per_cpu_ptr(chan->buf, cpu) = buf;  in relay_prepare_cpu()
 523 		*per_cpu_ptr(chan->buf, i) = buf;  in relay_open()
 532 		if ((buf = *per_cpu_ptr(chan->buf, i)))  in relay_open()
 597 		buf = *per_cpu_ptr(chan->buf, 0);  in relay_late_setup_files()
 [all …]
 
 | 
| D | cpu.c | 169 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in cpuhp_invoke_callback()559 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in bringup_wait_for_ap()
 712 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in cpuhp_create()
 810 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in cpuhp_invoke_ap_callback()
 859 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in cpuhp_kick_ap_work()
 1021 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in takedown_cpu()
 1116 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in _cpu_down()
 1278 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in notify_cpu_starting()
 1318 	struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in _cpu_up()
 1930 		struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);  in cpuhp_rollback_install()
 [all …]
 
 | 
| /Linux-v5.15/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ | 
| D | percpu.h | 28 #define per_cpu_ptr(ptr, cpu) \  macro69 		THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), thread_cpu_id), \
 76 		THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), this_cpu_add_impl_cpu), \
 
 | 
| /Linux-v5.15/fs/squashfs/ | 
| D | decompressor_multi_percpu.c | 40 		stream = per_cpu_ptr(percpu, cpu);  in squashfs_decompressor_create()54 		stream = per_cpu_ptr(percpu, cpu);  in squashfs_decompressor_create()
 71 			stream = per_cpu_ptr(percpu, cpu);  in squashfs_decompressor_destroy()
 
 | 
| /Linux-v5.15/include/linux/ | 
| D | part_stat.h | 28 	(per_cpu_ptr((part)->bd_stats, (cpu))->field)38 		res += per_cpu_ptr((part)->bd_stats, _cpu)->field; \
 47 		memset(per_cpu_ptr(part->bd_stats, i), value,  in part_stat_set_all()
 
 | 
| /Linux-v5.15/drivers/powercap/ | 
| D | idle_inject.c | 92 		iit = per_cpu_ptr(&idle_inject_thread, cpu);  in idle_inject_wakeup()137 	iit = per_cpu_ptr(&idle_inject_thread, cpu);  in idle_inject_fn()
 254 		iit = per_cpu_ptr(&idle_inject_thread, cpu);  in idle_inject_stop()
 284 		per_cpu_ptr(&idle_inject_thread, cpu);  in idle_inject_should_run()
 
 | 
| /Linux-v5.15/kernel/trace/ | 
| D | trace_functions_graph.c | 173 	data = per_cpu_ptr(tr->array_buffer.data, cpu);  in trace_graph_entry()253 	data = per_cpu_ptr(tr->array_buffer.data, cpu);  in trace_graph_return()
 391 	last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);  in verif_pid()
 644 		cpu_data = per_cpu_ptr(data->cpu_data, cpu);  in print_graph_entry_leaf()
 688 		cpu_data = per_cpu_ptr(data->cpu_data, cpu);  in print_graph_entry_nested()
 788 	depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);  in check_irq_entry()
 834 	depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);  in check_irq_return()
 918 		cpu_data = per_cpu_ptr(data->cpu_data, cpu);  in print_graph_return()
 980 		depth = per_cpu_ptr(data->cpu_data, iter->cpu)->depth;  in print_graph_comment()
 1046 	if (data && per_cpu_ptr(data->cpu_data, cpu)->ignore) {  in print_graph_function_flags()
 [all …]
 
 | 
| D | trace_sched_wakeup.c | 85 	*data = per_cpu_ptr(tr->array_buffer.data, cpu);  in func_prolog_preempt_disable()457 	disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);  in probe_wakeup_sched_switch()
 471 	data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu);  in probe_wakeup_sched_switch()
 494 	atomic_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);  in probe_wakeup_sched_switch()
 549 	disabled = atomic_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);  in probe_wakeup()
 581 	data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu);  in probe_wakeup()
 596 	atomic_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);  in probe_wakeup()
 
 | 
| /Linux-v5.15/drivers/clocksource/ | 
| D | timer-mp-csky.c | 78 	struct timer_of *to = per_cpu_ptr(&csky_to, cpu);  in csky_mptimer_starting_cpu()145 		to = per_cpu_ptr(&csky_to, cpu);  in csky_mptimer_init()
 168 		to = per_cpu_ptr(&csky_to, cpu_rollback);  in csky_mptimer_init()
 
 | 
| D | timer-tegra.c | 134 	struct timer_of *to = per_cpu_ptr(&tegra_to, cpu);  in tegra_timer_setup()159 	struct timer_of *to = per_cpu_ptr(&tegra_to, cpu);  in tegra_timer_stop()
 302 		struct timer_of *cpu_to = per_cpu_ptr(&tegra_to, cpu);  in tegra_init_timer()
 359 		cpu_to = per_cpu_ptr(&tegra_to, cpu);  in tegra_init_timer()
 
 | 
| /Linux-v5.15/tools/testing/radix-tree/linux/ | 
| D | percpu.h | 10 #define per_cpu_ptr(ptr, cpu)   ({ (void)(cpu); (ptr); })  macro11 #define per_cpu(var, cpu)	(*per_cpu_ptr(&(var), cpu))
 
 | 
| /Linux-v5.15/arch/x86/kernel/ | 
| D | irq_64.c | 37 	char *stack = (char *)per_cpu_ptr(&irq_stack_backing_store, cpu);  in map_irq_stack()63 	void *va = per_cpu_ptr(&irq_stack_backing_store, cpu);  in map_irq_stack()
 
 | 
| D | kgdb.c | 208 		bp = *per_cpu_ptr(breakinfo[breakno].pev, cpu);  in kgdb_correct_hw_break()237 		pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);  in hw_break_reserve_slot()
 249 		pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);  in hw_break_reserve_slot()
 264 		pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);  in hw_break_release_slot()
 304 		bp = *per_cpu_ptr(breakinfo[i].pev, cpu);  in kgdb_remove_all_hw_break()
 397 		bp = *per_cpu_ptr(breakinfo[i].pev, cpu);  in kgdb_disable_hw_debug()
 666 			pevent = per_cpu_ptr(breakinfo[i].pev, cpu);  in kgdb_arch_late()
 
 | 
| /Linux-v5.15/drivers/hv/ | 
| D | hv.c | 122 		hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu);  in hv_synic_alloc()134 		hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu);  in hv_synic_alloc()
 175 			= per_cpu_ptr(hv_context.cpu_context, cpu);  in hv_synic_free()
 195 		= per_cpu_ptr(hv_context.cpu_context, cpu);  in hv_synic_enable_regs()
 
 | 
| /Linux-v5.15/kernel/rcu/ | 
| D | tree_exp.h | 244 		rdp = per_cpu_ptr(&rcu_data, cpu);  in rcu_report_exp_cpu_mult()282 	struct rcu_data *rdp = per_cpu_ptr(&rcu_data, raw_smp_processor_id());  in exp_funnel_lock()
 353 		struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);  in sync_rcu_exp_select_node_cpus()
 381 		struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);  in sync_rcu_exp_select_node_cpus()
 505 				rdp = per_cpu_ptr(&rcu_data, cpu);  in synchronize_rcu_expedited_wait()
 537 				rdp = per_cpu_ptr(&rcu_data, cpu);  in synchronize_rcu_expedited_wait()
 758 	rdp = per_cpu_ptr(&rcu_data, cpu);  in sync_sched_exp_online_cleanup()
 
 |