/Linux-v5.4/kernel/ |
D | irq_work.c | 65 if (llist_add(&work->llnode, this_cpu_ptr(&lazy_list)) && in __irq_work_queue_local() 69 if (llist_add(&work->llnode, this_cpu_ptr(&raised_list))) in __irq_work_queue_local() 129 raised = this_cpu_ptr(&raised_list); in irq_work_needs_cpu() 130 lazy = this_cpu_ptr(&lazy_list); in irq_work_needs_cpu() 180 irq_work_run_list(this_cpu_ptr(&raised_list)); in irq_work_run() 181 irq_work_run_list(this_cpu_ptr(&lazy_list)); in irq_work_run() 187 struct llist_head *raised = this_cpu_ptr(&raised_list); in irq_work_tick() 191 irq_work_run_list(this_cpu_ptr(&lazy_list)); in irq_work_tick()
|
D | watchdog.c | 356 complete(this_cpu_ptr(&softlockup_completion)); in softlockup_fn() 376 if (completion_done(this_cpu_ptr(&softlockup_completion))) { in watchdog_timer_fn() 377 reinit_completion(this_cpu_ptr(&softlockup_completion)); in watchdog_timer_fn() 380 this_cpu_ptr(&softlockup_stop_work)); in watchdog_timer_fn() 481 struct hrtimer *hrtimer = this_cpu_ptr(&watchdog_hrtimer); in watchdog_enable() 482 struct completion *done = this_cpu_ptr(&softlockup_completion); in watchdog_enable() 507 struct hrtimer *hrtimer = this_cpu_ptr(&watchdog_hrtimer); in watchdog_disable() 518 wait_for_completion(this_cpu_ptr(&softlockup_completion)); in watchdog_disable()
|
D | user-return-notifier.c | 18 hlist_add_head(&urn->link, this_cpu_ptr(&return_notifier_list)); in user_return_notifier_register() 29 if (hlist_empty(this_cpu_ptr(&return_notifier_list))) in user_return_notifier_unregister()
|
/Linux-v5.4/arch/x86/kernel/cpu/mce/ |
D | intel.c | 136 if (machine_check_poll(0, this_cpu_ptr(&mce_banks_owned))) in mce_intel_cmci_poll() 159 owned = this_cpu_ptr(mce_banks_owned); in cmci_toggle_interrupt_mode() 256 machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); in intel_threshold_interrupt() 266 unsigned long *owned = (void *)this_cpu_ptr(&mce_banks_owned); in cmci_discover() 288 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 312 __clear_bit(i, this_cpu_ptr(mce_poll_banks)); in cmci_discover() 323 WARN_ON(!test_bit(i, this_cpu_ptr(mce_poll_banks))); in cmci_discover() 348 machine_check_poll(0, this_cpu_ptr(&mce_banks_owned)); in cmci_recheck() 357 if (!test_bit(bank, this_cpu_ptr(mce_banks_owned))) in __cmci_disable_bank() 362 __clear_bit(bank, this_cpu_ptr(mce_banks_owned)); in __cmci_disable_bank()
|
D | core.c | 401 return *(u64 *)((char *)this_cpu_ptr(&injectm) + offset); in mce_rdmsrl() 423 *(u64 *)((char *)this_cpu_ptr(&injectm) + offset) = v; in mce_wrmsrl() 694 struct mce_bank *mce_banks = this_cpu_ptr(mce_banks_array); in machine_check_poll() 1142 struct mce_bank *mce_banks = this_cpu_ptr(mce_banks_array); in __mc_scan_banks() 1262 final = this_cpu_ptr(&mces_seen); in do_machine_check() 1409 struct timer_list *cpu_t = this_cpu_ptr(&mce_timer); in mce_timer_fn() 1416 if (mce_available(this_cpu_ptr(&cpu_info))) { in mce_timer_fn() 1417 machine_check_poll(0, this_cpu_ptr(&mce_poll_banks)); in mce_timer_fn() 1444 struct timer_list *t = this_cpu_ptr(&mce_timer); in mce_timer_kick() 1486 struct mce_bank *mce_banks = this_cpu_ptr(mce_banks_array); in __mcheck_cpu_mce_banks_init() [all …]
|
/Linux-v5.4/net/core/ |
D | dst_cache.c | 68 return dst_cache_per_cpu_get(dst_cache, this_cpu_ptr(dst_cache->cache)); in dst_cache_get() 80 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip4() 98 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip4() 113 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_set_ip6() 114 dst_cache_per_cpu_dst_set(this_cpu_ptr(dst_cache->cache), dst, in dst_cache_set_ip6() 129 idst = this_cpu_ptr(dst_cache->cache); in dst_cache_get_ip6()
|
/Linux-v5.4/kernel/time/ |
D | tick-sched.c | 246 irq_work_queue(this_cpu_ptr(&nohz_full_kick_work)); in tick_nohz_full_kick() 382 ts = this_cpu_ptr(&tick_cpu_sched); in __tick_nohz_task_switch() 476 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_tick_stopped() 973 __tick_nohz_idle_stop_tick(this_cpu_ptr(&tick_cpu_sched)); in tick_nohz_idle_stop_tick() 978 tick_nohz_retain_tick(this_cpu_ptr(&tick_cpu_sched)); in tick_nohz_idle_retain_tick() 999 ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_enter() 1019 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_irq_exit() 1032 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_idle_got_tick() 1062 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_get_sleep_length() 1112 struct tick_sched *ts = this_cpu_ptr(&tick_cpu_sched); in tick_nohz_get_idle_calls() [all …]
|
/Linux-v5.4/arch/xtensa/kernel/ |
D | hw_breakpoint.c | 177 i = alloc_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_install_hw_breakpoint() 184 i = alloc_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_install_hw_breakpoint() 215 i = free_slot(this_cpu_ptr(bp_on_reg), XCHAL_NUM_IBREAK, bp); in arch_uninstall_hw_breakpoint() 223 i = free_slot(this_cpu_ptr(wp_on_reg), XCHAL_NUM_DBREAK, bp); in arch_uninstall_hw_breakpoint() 268 struct perf_event *bp = this_cpu_ptr(wp_on_reg)[i]; in restore_dbreak() 280 struct perf_event **bp = this_cpu_ptr(bp_on_reg); in check_hw_breakpoint() 289 struct perf_event **bp = this_cpu_ptr(wp_on_reg); in check_hw_breakpoint()
|
/Linux-v5.4/arch/x86/events/intel/ |
D | bts.c | 222 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in __bts_event_start() 250 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_start() 251 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_start() 281 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in __bts_event_stop() 295 struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events); in bts_event_stop() 296 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in bts_event_stop() 329 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_enable_local() 349 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_disable_local() 440 struct debug_store *ds = this_cpu_ptr(&cpu_hw_events)->ds; in intel_bts_interrupt() 441 struct bts_ctx *bts = this_cpu_ptr(&bts_ctx); in intel_bts_interrupt() [all …]
|
/Linux-v5.4/block/ |
D | blk-softirq.c | 28 cpu_list = this_cpu_ptr(&blk_cpu_done); in blk_done_softirq() 49 list = this_cpu_ptr(&blk_cpu_done); in trigger_softirq() 91 this_cpu_ptr(&blk_cpu_done)); in blk_softirq_cpu_dead() 130 list = this_cpu_ptr(&blk_cpu_done); in __blk_complete_request()
|
/Linux-v5.4/arch/powerpc/include/asm/ |
D | xics.h | 110 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_push_cppr() 123 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_pop_cppr() 133 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_set_base_cppr() 145 struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr); in xics_cppr_top()
|
/Linux-v5.4/arch/s390/kernel/ |
D | perf_cpum_cf.c | 123 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_enable() 146 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_disable() 365 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_start() 396 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_stop() 417 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_add() 441 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_del() 469 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_start_txn() 489 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_cancel_txn() 510 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_pmu_commit_txn()
|
D | perf_cpum_cf_diag.c | 109 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_trailer() 132 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_enable() 158 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_disable() 371 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_getctrset() 523 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_start() 524 struct cf_diag_csd *csd = this_cpu_ptr(&cf_diag_csd); in cf_diag_start() 546 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_stop() 547 struct cf_diag_csd *csd = this_cpu_ptr(&cf_diag_csd); in cf_diag_stop() 566 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_add() 590 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cf_diag_del()
|
D | perf_cpum_cf_common.c | 48 cpuhw = this_cpu_ptr(&cpu_cf_events); in cpumf_measurement_alert() 76 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in cpum_cf_setup_cpu() 130 struct cpu_cf_events *cpuhw = this_cpu_ptr(&cpu_cf_events); in kernel_cpumcf_alert()
|
/Linux-v5.4/arch/arm/kernel/ |
D | perf_event_xscale.c | 149 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_handle_irq() 209 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_enable_event() 245 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_disable_event() 304 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_start() 316 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale1pmu_stop() 501 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_handle_irq() 555 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_enable_event() 601 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_disable_event() 667 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_start() 679 struct pmu_hw_events *events = this_cpu_ptr(cpu_pmu->hw_events); in xscale2pmu_stop()
|
/Linux-v5.4/arch/arm64/kernel/ |
D | hw_breakpoint.c | 236 slots = this_cpu_ptr(bp_on_reg); in hw_breakpoint_control() 243 slots = this_cpu_ptr(wp_on_reg); in hw_breakpoint_control() 588 slots = this_cpu_ptr(bp_on_reg); in toggle_bp_registers() 592 slots = this_cpu_ptr(wp_on_reg); in toggle_bp_registers() 630 slots = this_cpu_ptr(bp_on_reg); in breakpoint_handler() 680 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in breakpoint_handler() 745 slots = this_cpu_ptr(wp_on_reg); in watchpoint_handler() 823 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in watchpoint_handler() 848 kernel_step = this_cpu_ptr(&stepping_kernel_bp); in reinstall_suspended_bps() 947 for (slots = this_cpu_ptr(bp_on_reg), i = 0; i < core_num_brps; ++i) { in hw_breakpoint_reset() [all …]
|
D | fpsimd.c | 289 this_cpu_ptr(&fpsimd_last_state); in fpsimd_save() 1093 this_cpu_ptr(&fpsimd_last_state); in fpsimd_bind_task_to_cpu() 1115 this_cpu_ptr(&fpsimd_last_state); in fpsimd_bind_state_to_cpu() 1305 char *sve_state = this_cpu_ptr(efi_sve_state); in __efi_fpsimd_begin() 1310 &this_cpu_ptr(&efi_fpsimd_state)->fpsr); in __efi_fpsimd_begin() 1312 fpsimd_save_state(this_cpu_ptr(&efi_fpsimd_state)); in __efi_fpsimd_begin() 1332 char const *sve_state = this_cpu_ptr(efi_sve_state); in __efi_fpsimd_end() 1335 &this_cpu_ptr(&efi_fpsimd_state)->fpsr, in __efi_fpsimd_end() 1340 fpsimd_load_state(this_cpu_ptr(&efi_fpsimd_state)); in __efi_fpsimd_end()
|
/Linux-v5.4/net/netfilter/ |
D | nft_counter.c | 36 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_do_eval() 37 myseq = this_cpu_ptr(&nft_counter_seq); in nft_counter_do_eval() 68 this_cpu = this_cpu_ptr(cpu_stats); in nft_counter_do_init() 110 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_reset() 242 this_cpu = this_cpu_ptr(cpu_stats); in nft_counter_clone()
|
/Linux-v5.4/arch/x86/xen/ |
D | multicalls.c | 58 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_flush() 140 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in __xen_mc_entry() 171 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_extend_args() 201 struct mc_buffer *b = this_cpu_ptr(&mc_buffer); in xen_mc_callback()
|
/Linux-v5.4/arch/powerpc/platforms/powernv/ |
D | opal-tracepoints.c | 55 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_entry() 76 depth = this_cpu_ptr(&opal_trace_depth); in __trace_opal_exit()
|
/Linux-v5.4/lib/ |
D | irq_poll.c | 37 list_add_tail(&iop->list, this_cpu_ptr(&blk_cpu_iopoll)); in irq_poll_sched() 80 struct list_head *list = this_cpu_ptr(&blk_cpu_iopoll); in irq_poll_softirq() 196 this_cpu_ptr(&blk_cpu_iopoll)); in irq_poll_cpu_dead()
|
/Linux-v5.4/arch/arm64/include/asm/ |
D | hardirq.h | 42 struct nmi_ctx *nmi_ctx = this_cpu_ptr(&nmi_contexts); \ 54 struct nmi_ctx *nmi_ctx = this_cpu_ptr(&nmi_contexts); \
|
/Linux-v5.4/net/sched/ |
D | act_gact.c | 164 bstats_cpu_update(this_cpu_ptr(gact->common.cpu_bstats), skb); in tcf_gact_act() 166 qstats_drop_inc(this_cpu_ptr(gact->common.cpu_qstats)); in tcf_gact_act() 180 _bstats_cpu_update(this_cpu_ptr(gact->common.cpu_bstats), bytes, in tcf_gact_stats_update() 183 this_cpu_ptr(gact->common.cpu_qstats)->drops += packets; in tcf_gact_stats_update() 186 _bstats_cpu_update(this_cpu_ptr(gact->common.cpu_bstats_hw), in tcf_gact_stats_update()
|
/Linux-v5.4/arch/sh/kernel/ |
D | kprobes.c | 102 struct kprobe *saved = this_cpu_ptr(&saved_next_opcode); in arch_remove_kprobe() 111 saved = this_cpu_ptr(&saved_next_opcode2); in arch_remove_kprobe() 153 op1 = this_cpu_ptr(&saved_next_opcode); in prepare_singlestep() 154 op2 = this_cpu_ptr(&saved_next_opcode2); in prepare_singlestep() 378 p = this_cpu_ptr(&saved_next_opcode); in post_kprobe_handler() 390 p = this_cpu_ptr(&saved_next_opcode2); in post_kprobe_handler()
|
/Linux-v5.4/kernel/irq/ |
D | matrix.c | 80 struct cpumap *cm = this_cpu_ptr(m->maps); in irq_matrix_online() 101 struct cpumap *cm = this_cpu_ptr(m->maps); in irq_matrix_offline() 184 struct cpumap *cm = this_cpu_ptr(m->maps); in irq_matrix_assign_system() 323 struct cpumap *cm = this_cpu_ptr(m->maps); in irq_matrix_assign() 443 struct cpumap *cm = this_cpu_ptr(m->maps); in irq_matrix_available() 467 struct cpumap *cm = this_cpu_ptr(m->maps); in irq_matrix_allocated()
|