Lines Matching refs:hv_cpu

86 	struct hv_per_cpu_context *hv_cpu;  in hv_post_message()  local
92 hv_cpu = get_cpu_ptr(hv_context.cpu_context); in hv_post_message()
93 aligned_msg = hv_cpu->post_msg_page; in hv_post_message()
106 put_cpu_ptr(hv_cpu); in hv_post_message()
114 struct hv_per_cpu_context *hv_cpu; in hv_synic_alloc() local
122 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()
123 memset(hv_cpu, 0, sizeof(*hv_cpu)); in hv_synic_alloc()
134 hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu); in hv_synic_alloc()
136 tasklet_init(&hv_cpu->msg_dpc, in hv_synic_alloc()
137 vmbus_on_msg_dpc, (unsigned long) hv_cpu); in hv_synic_alloc()
139 hv_cpu->synic_message_page = in hv_synic_alloc()
141 if (hv_cpu->synic_message_page == NULL) { in hv_synic_alloc()
146 hv_cpu->synic_event_page = (void *)get_zeroed_page(GFP_ATOMIC); in hv_synic_alloc()
147 if (hv_cpu->synic_event_page == NULL) { in hv_synic_alloc()
152 hv_cpu->post_msg_page = (void *)get_zeroed_page(GFP_ATOMIC); in hv_synic_alloc()
153 if (hv_cpu->post_msg_page == NULL) { in hv_synic_alloc()
174 struct hv_per_cpu_context *hv_cpu in hv_synic_free() local
177 free_page((unsigned long)hv_cpu->synic_event_page); in hv_synic_free()
178 free_page((unsigned long)hv_cpu->synic_message_page); in hv_synic_free()
179 free_page((unsigned long)hv_cpu->post_msg_page); in hv_synic_free()
194 struct hv_per_cpu_context *hv_cpu in hv_synic_enable_regs() local
204 simp.base_simp_gpa = virt_to_phys(hv_cpu->synic_message_page) in hv_synic_enable_regs()
212 siefp.base_siefp_gpa = virt_to_phys(hv_cpu->synic_event_page) in hv_synic_enable_regs()
308 struct hv_per_cpu_context *hv_cpu = this_cpu_ptr(hv_context.cpu_context); in hv_synic_event_pending() local
310 (union hv_synic_event_flags *)hv_cpu->synic_event_page + VMBUS_MESSAGE_SINT; in hv_synic_event_pending()