Searched refs:hv_context (Results 1 – 4 of 4) sorted by relevance
36 struct hv_context hv_context = { variable62 hv_context.cpu_context = alloc_percpu(struct hv_per_cpu_context); in hv_init()63 if (!hv_context.cpu_context) in hv_init()87 hv_cpu = get_cpu_ptr(hv_context.cpu_context); in hv_post_message()116 hv_cpu = this_cpu_ptr(hv_context.cpu_context); in hv_stimer0_isr()193 hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask), in hv_synic_alloc()195 if (hv_context.hv_numa_map == NULL) { in hv_synic_alloc()202 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()259 = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_free()267 kfree(hv_context.hv_numa_map); in hv_synic_free()[all …]
225 struct hv_context { struct244 extern struct hv_context hv_context; argument
361 = this_cpu_ptr(hv_context.cpu_context); in percpu_channel_enq()645 alloced_mask = &hv_context.hv_numa_map[primary->numa_node]; in init_vp_index()732 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_wait_for_unload()761 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_wait_for_unload()
990 = this_cpu_ptr(hv_context.cpu_context); in vmbus_isr()1897 = per_cpu_ptr(hv_context.cpu_context, cpu); in vmbus_exit()