Searched refs:hv_context (Results 1 – 4 of 4) sorted by relevance
24 struct hv_context hv_context; variable33 hv_context.cpu_context = alloc_percpu(struct hv_per_cpu_context); in hv_init()34 if (!hv_context.cpu_context) in hv_init()55 hv_cpu = get_cpu_ptr(hv_context.cpu_context); in hv_post_message()85 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()89 hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask), in hv_synic_alloc()91 if (hv_context.hv_numa_map == NULL) { in hv_synic_alloc()97 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()140 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_free()147 kfree(hv_context.hv_numa_map); in hv_synic_free()[all …]
143 struct hv_context { struct158 extern struct hv_context hv_context; argument
347 = this_cpu_ptr(hv_context.cpu_context); in percpu_channel_enq()698 alloced_mask = &hv_context.hv_numa_map[primary->numa_node]; in init_vp_index()789 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_wait_for_unload()818 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_wait_for_unload()
1198 = this_cpu_ptr(hv_context.cpu_context); in vmbus_isr()2410 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_exit()