/Linux-v5.4/mm/ |
D | percpu-vm.c | 85 unsigned int cpu, tcpu; in pcpu_alloc_pages() local 105 for_each_possible_cpu(tcpu) { in pcpu_alloc_pages() 106 if (tcpu == cpu) in pcpu_alloc_pages() 109 __free_page(pages[pcpu_page_idx(tcpu, i)]); in pcpu_alloc_pages() 216 unsigned int cpu, tcpu; in pcpu_map_pages() local 232 for_each_possible_cpu(tcpu) { in pcpu_map_pages() 233 if (tcpu == cpu) in pcpu_map_pages() 235 __pcpu_unmap_pages(pcpu_chunk_addr(chunk, tcpu, page_start), in pcpu_map_pages()
|
D | percpu.c | 2541 unsigned int cpu, tcpu; in pcpu_build_alloc_info() local 2573 for_each_possible_cpu(tcpu) { in pcpu_build_alloc_info() 2574 if (cpu == tcpu) in pcpu_build_alloc_info() 2576 if (group_map[tcpu] == group && cpu_distance_fn && in pcpu_build_alloc_info() 2577 (cpu_distance_fn(cpu, tcpu) > LOCAL_DISTANCE || in pcpu_build_alloc_info() 2578 cpu_distance_fn(tcpu, cpu) > LOCAL_DISTANCE)) { in pcpu_build_alloc_info()
|
/Linux-v5.4/drivers/media/pci/b2c2/ |
D | flexcop-dma.c | 12 u8 *tcpu; in flexcop_dma_allocate() local 20 tcpu = pci_alloc_consistent(pdev, size, &tdma); in flexcop_dma_allocate() 21 if (tcpu != NULL) { in flexcop_dma_allocate() 23 dma->cpu_addr0 = tcpu; in flexcop_dma_allocate() 25 dma->cpu_addr1 = tcpu + size/2; in flexcop_dma_allocate()
|
/Linux-v5.4/arch/s390/kvm/ |
D | diag.c | 156 struct kvm_vcpu *tcpu; in __diag_time_slice_end_directed() local 166 tcpu = kvm_get_vcpu_by_id(vcpu->kvm, tid); in __diag_time_slice_end_directed() 167 if (tcpu) in __diag_time_slice_end_directed() 168 kvm_vcpu_yield_to(tcpu); in __diag_time_slice_end_directed()
|
/Linux-v5.4/arch/x86/platform/uv/ |
D | uv_nmi.c | 754 int tcpu; in uv_nmi_dump_state() local 764 for_each_online_cpu(tcpu) { in uv_nmi_dump_state() 765 if (cpumask_test_cpu(tcpu, uv_nmi_cpu_mask)) in uv_nmi_dump_state() 767 else if (tcpu == cpu) in uv_nmi_dump_state() 768 uv_nmi_dump_state_cpu(tcpu, regs); in uv_nmi_dump_state() 770 uv_nmi_trigger_dump(tcpu); in uv_nmi_dump_state()
|
D | tlb_uv.c | 779 int tcpu; in disable_for_period() local 789 for_each_present_cpu(tcpu) { in disable_for_period() 790 tbcp = &per_cpu(bau_control, tcpu); in disable_for_period() 990 int tcpu; in check_enable() local 998 for_each_present_cpu(tcpu) { in check_enable() 999 tbcp = &per_cpu(bau_control, tcpu); in check_enable()
|
/Linux-v5.4/drivers/xen/events/ |
D | events_base.c | 1297 static int xen_rebind_evtchn_to_cpu(int evtchn, unsigned int tcpu) in xen_rebind_evtchn_to_cpu() argument 1310 bind_vcpu.vcpu = xen_vcpu_nr(tcpu); in xen_rebind_evtchn_to_cpu() 1324 bind_evtchn_to_cpu(evtchn, tcpu); in xen_rebind_evtchn_to_cpu() 1335 unsigned tcpu = cpumask_first_and(dest, cpu_online_mask); in set_affinity_irq() local 1336 int ret = xen_rebind_evtchn_to_cpu(evtchn_from_irq(data->irq), tcpu); in set_affinity_irq() 1339 irq_data_update_effective_affinity(data, cpumask_of(tcpu)); in set_affinity_irq() 1345 int xen_set_affinity_evtchn(struct irq_desc *desc, unsigned int tcpu) in xen_set_affinity_evtchn() argument 1349 return set_affinity_irq(d, cpumask_of(tcpu), false); in xen_set_affinity_evtchn()
|
/Linux-v5.4/include/xen/ |
D | events.h | 63 int xen_set_affinity_evtchn(struct irq_desc *desc, unsigned int tcpu);
|
/Linux-v5.4/net/core/ |
D | dev.c | 3971 u32 tcpu; in get_rps_cpu() local 4016 tcpu = rflow->cpu; in get_rps_cpu() 4029 if (unlikely(tcpu != next_cpu) && in get_rps_cpu() 4030 (tcpu >= nr_cpu_ids || !cpu_online(tcpu) || in get_rps_cpu() 4031 ((int)(per_cpu(softnet_data, tcpu).input_queue_head - in get_rps_cpu() 4033 tcpu = next_cpu; in get_rps_cpu() 4037 if (tcpu < nr_cpu_ids && cpu_online(tcpu)) { in get_rps_cpu() 4039 cpu = tcpu; in get_rps_cpu() 4047 tcpu = map->cpus[reciprocal_scale(hash, map->len)]; in get_rps_cpu() 4048 if (cpu_online(tcpu)) { in get_rps_cpu() [all …]
|
/Linux-v5.4/Documentation/ia64/ |
D | err_inject.rst | 964 …printf("\t\tcpu,loop,interval,err_type_info,err_struct_info[,err_data_buffer[0],err_data_buffer[1]…
|