Home
last modified time | relevance | path

Searched refs:per_cpu_ptr (Results 1 – 25 of 284) sorted by relevance

12345678910>>...12

/Linux-v5.4/fs/xfs/
Dxfs_stats.h164 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \
165 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \
170 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \
171 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \
176 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \
177 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \
182 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \
183 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \
188 per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \
189 per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]; \
[all …]
Dxfs_stats.c15 val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx)); in counter_val()
70 xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes; in xfs_stats_format()
71 xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes; in xfs_stats_format()
72 xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes; in xfs_stats_format()
96 vn_active = per_cpu_ptr(stats, c)->s.vn_active; in xfs_stats_clearall()
97 memset(per_cpu_ptr(stats, c), 0, sizeof(*stats)); in xfs_stats_clearall()
98 per_cpu_ptr(stats, c)->s.vn_active = vn_active; in xfs_stats_clearall()
/Linux-v5.4/arch/x86/events/amd/
Duncore.c75 return *per_cpu_ptr(amd_uncore_nb, event->cpu); in event_to_amd_uncore()
77 return *per_cpu_ptr(amd_uncore_llc, event->cpu); in event_to_amd_uncore()
344 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb; in amd_uncore_cpu_up_prepare()
358 *per_cpu_ptr(amd_uncore_llc, cpu) = uncore_llc; in amd_uncore_cpu_up_prepare()
365 *per_cpu_ptr(amd_uncore_nb, cpu) = NULL; in amd_uncore_cpu_up_prepare()
378 that = *per_cpu_ptr(uncores, cpu); in amd_uncore_find_online_sibling()
403 uncore = *per_cpu_ptr(amd_uncore_nb, cpu); in amd_uncore_cpu_starting()
408 *per_cpu_ptr(amd_uncore_nb, cpu) = uncore; in amd_uncore_cpu_starting()
412 uncore = *per_cpu_ptr(amd_uncore_llc, cpu); in amd_uncore_cpu_starting()
416 *per_cpu_ptr(amd_uncore_llc, cpu) = uncore; in amd_uncore_cpu_starting()
[all …]
/Linux-v5.4/kernel/sched/
Dtopology.c850 sibling = *per_cpu_ptr(sdd->sd, i); in build_balance_mask()
909 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in init_overlap_sched_group()
944 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups()
1059 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1067 sg = *per_cpu_ptr(sdd->sg, cpu); in get_group()
1068 sg->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group()
1265 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations()
1266 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations()
1268 if (atomic_read(&(*per_cpu_ptr(sdd->sds, cpu))->ref)) in claim_allocations()
1269 *per_cpu_ptr(sdd->sds, cpu) = NULL; in claim_allocations()
[all …]
/Linux-v5.4/kernel/irq/
Dmatrix.c138 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu()
159 cm = per_cpu_ptr(m->maps, cpu); in matrix_find_best_cpu_managed()
215 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_reserve_managed()
256 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_remove_managed()
298 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc_managed()
387 cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_alloc()
414 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_free()
495 struct cpumap *cm = per_cpu_ptr(m->maps, cpu); in irq_matrix_debug_show()
/Linux-v5.4/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/
Dpercpu.h28 #define per_cpu_ptr(ptr, cpu) \ macro
69 THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), thread_cpu_id), \
76 THIS_CPU_ADD_HELPER(per_cpu_ptr(&(pcp), this_cpu_add_impl_cpu), \
/Linux-v5.4/fs/squashfs/
Ddecompressor_multi_percpu.c38 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create()
51 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create()
68 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_destroy()
/Linux-v5.4/kernel/
Dsmpboot.c173 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in __smpboot_create_thread()
197 *per_cpu_ptr(ht->store, cpu) = tsk; in __smpboot_create_thread()
230 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_unpark_thread()
249 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_park_thread()
272 struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); in smpboot_destroy_threads()
277 *per_cpu_ptr(ht->store, cpu) = NULL; in smpboot_destroy_threads()
Drelay.c218 *per_cpu_ptr(chan->buf, buf->cpu) = NULL; in relay_destroy_buf()
395 if (chan->is_global && (buf = *per_cpu_ptr(chan->buf, 0))) { in relay_reset()
402 if ((buf = *per_cpu_ptr(chan->buf, i))) in relay_reset()
450 return *per_cpu_ptr(chan->buf, 0); in relay_open_buf()
474 *per_cpu_ptr(chan->buf, 0) = buf; in relay_open_buf()
529 if ((buf = *per_cpu_ptr(chan->buf, cpu))) in relay_prepare_cpu()
537 *per_cpu_ptr(chan->buf, cpu) = buf; in relay_prepare_cpu()
602 *per_cpu_ptr(chan->buf, i) = buf; in relay_open()
611 if ((buf = *per_cpu_ptr(chan->buf, i))) in relay_open()
676 buf = *per_cpu_ptr(chan->buf, 0); in relay_late_setup_files()
[all …]
Dcpu.c151 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_invoke_callback()
521 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in bringup_wait_for_ap()
617 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_create()
721 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_invoke_ap_callback()
770 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in cpuhp_kick_ap_work()
883 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in takedown_cpu()
887 kthread_park(per_cpu_ptr(&cpuhp_state, cpu)->thread); in takedown_cpu()
903 kthread_unpark(per_cpu_ptr(&cpuhp_state, cpu)->thread); in takedown_cpu()
980 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in _cpu_down()
1074 struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu); in notify_cpu_starting()
[all …]
/Linux-v5.4/drivers/block/zram/
Dzcomp.c163 if (WARN_ON(*per_cpu_ptr(comp->stream, cpu))) in zcomp_cpu_up_prepare()
171 *per_cpu_ptr(comp->stream, cpu) = zstrm; in zcomp_cpu_up_prepare()
180 zstrm = *per_cpu_ptr(comp->stream, cpu); in zcomp_cpu_dead()
183 *per_cpu_ptr(comp->stream, cpu) = NULL; in zcomp_cpu_dead()
/Linux-v5.4/drivers/powercap/
Didle_inject.c89 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_wakeup()
134 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_fn()
240 iit = per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_stop()
272 per_cpu_ptr(&idle_inject_thread, cpu); in idle_inject_should_run()
/Linux-v5.4/net/xfrm/
Dxfrm_ipcomp.c45 u8 *scratch = *per_cpu_ptr(ipcomp_scratches, cpu); in ipcomp_decompress()
46 struct crypto_comp *tfm = *per_cpu_ptr(ipcd->tfms, cpu); in ipcomp_decompress()
212 vfree(*per_cpu_ptr(scratches, i)); in ipcomp_free_scratches()
237 *per_cpu_ptr(scratches, i) = scratch; in ipcomp_alloc_scratches()
265 struct crypto_comp *tfm = *per_cpu_ptr(tfms, cpu); in ipcomp_free_tfms()
307 *per_cpu_ptr(tfms, cpu) = tfm; in ipcomp_alloc_tfms()
/Linux-v5.4/kernel/trace/
Dtrace_functions_graph.c174 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in trace_graph_entry()
255 data = per_cpu_ptr(tr->trace_buffer.data, cpu); in trace_graph_return()
393 last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid); in verif_pid()
646 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_leaf()
690 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_nested()
790 depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq); in check_irq_entry()
836 depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq); in check_irq_return()
920 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_return()
982 depth = per_cpu_ptr(data->cpu_data, iter->cpu)->depth; in print_graph_comment()
1048 if (data && per_cpu_ptr(data->cpu_data, cpu)->ignore) { in print_graph_function_flags()
[all …]
/Linux-v5.4/kernel/bpf/
Dpercpu_freelist.c15 struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_init()
72 head = per_cpu_ptr(s->freelist, cpu); in pcpu_freelist_populate()
92 head = per_cpu_ptr(s->freelist, cpu); in __pcpu_freelist_pop()
Dbpf_lru_list.c408 l = per_cpu_ptr(lru->percpu_lru, cpu); in bpf_percpu_lru_pop_free()
441 loc_l = per_cpu_ptr(clru->local_list, cpu); in bpf_common_lru_pop_free()
470 steal_loc_l = per_cpu_ptr(clru->local_list, steal); in bpf_common_lru_pop_free()
514 loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu); in bpf_common_lru_push_free()
541 l = per_cpu_ptr(lru->percpu_lru, node->cpu); in bpf_percpu_lru_push_free()
591 l = per_cpu_ptr(lru->percpu_lru, cpu); in bpf_percpu_lru_populate()
658 l = per_cpu_ptr(lru->percpu_lru, cpu); in bpf_lru_init()
672 loc_l = per_cpu_ptr(clru->local_list, cpu); in bpf_lru_init()
/Linux-v5.4/drivers/clocksource/
Dtimer-mp-csky.c78 struct timer_of *to = per_cpu_ptr(&csky_to, cpu); in csky_mptimer_starting_cpu()
145 to = per_cpu_ptr(&csky_to, cpu); in csky_mptimer_init()
168 to = per_cpu_ptr(&csky_to, cpu_rollback); in csky_mptimer_init()
Dhyperv_timer.c117 ce = per_cpu_ptr(hv_clock_event, cpu); in hv_stimer_init()
142 ce = per_cpu_ptr(hv_clock_event, cpu); in hv_stimer_cleanup()
197 ce = per_cpu_ptr(hv_clock_event, cpu); in hv_stimer_global_cleanup()
Dtimer-tegra.c134 struct timer_of *to = per_cpu_ptr(&tegra_to, cpu); in tegra_timer_setup()
159 struct timer_of *to = per_cpu_ptr(&tegra_to, cpu); in tegra_timer_stop()
302 struct timer_of *cpu_to = per_cpu_ptr(&tegra_to, cpu); in tegra_init_timer()
359 cpu_to = per_cpu_ptr(&tegra_to, cpu); in tegra_init_timer()
/Linux-v5.4/tools/testing/radix-tree/linux/
Dpercpu.h10 #define per_cpu_ptr(ptr, cpu) ({ (void)(cpu); (ptr); }) macro
11 #define per_cpu(var, cpu) (*per_cpu_ptr(&(var), cpu))
/Linux-v5.4/drivers/hv/
Dhv.c85 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()
97 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()
140 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_free()
160 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_enable_regs()
/Linux-v5.4/arch/x86/kernel/
Dirq_64.c35 char *stack = (char *)per_cpu_ptr(&irq_stack_backing_store, cpu); in map_irq_stack()
60 void *va = per_cpu_ptr(&irq_stack_backing_store, cpu); in map_irq_stack()
Dkgdb.c208 bp = *per_cpu_ptr(breakinfo[breakno].pev, cpu); in kgdb_correct_hw_break()
237 pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); in hw_break_reserve_slot()
249 pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); in hw_break_reserve_slot()
264 pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); in hw_break_release_slot()
304 bp = *per_cpu_ptr(breakinfo[i].pev, cpu); in kgdb_remove_all_hw_break()
397 bp = *per_cpu_ptr(breakinfo[i].pev, cpu); in kgdb_disable_hw_debug()
665 pevent = per_cpu_ptr(breakinfo[i].pev, cpu); in kgdb_arch_late()
/Linux-v5.4/drivers/cpuidle/governors/
Dteo.c119 struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); in teo_update()
234 struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); in teo_select()
387 struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); in teo_reflect()
412 struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); in teo_enable_device()
/Linux-v5.4/drivers/cpuidle/
Dcpuidle-haltpoll.c57 dev = per_cpu_ptr(haltpoll_cpuidle_devices, cpu); in haltpoll_cpu_online()
74 dev = per_cpu_ptr(haltpoll_cpuidle_devices, cpu); in haltpoll_cpu_offline()

12345678910>>...12