Lines Matching defs:vmx

191 	struct vcpu_vmx *vmx = to_vmx(vcpu);  in nested_vmx_fail()  local
221 static void vmx_disable_shadow_vmcs(struct vcpu_vmx *vmx) in vmx_disable_shadow_vmcs()
232 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_evmcs() local
249 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_evmcs_handle_vmclear() local
273 static void vmx_sync_vmcs_host_state(struct vcpu_vmx *vmx, in vmx_sync_vmcs_host_state()
294 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_switch_vmcs() local
319 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_put_vmcs12_pages() local
333 struct vcpu_vmx *vmx = to_vmx(vcpu); in free_nested() local
412 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_ept_inject_page_fault() local
465 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_ept_new_eptp() local
623 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_prepare_msr_bitmap() local
727 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_cache_shadow_vmcs12() local
746 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_flush_cached_shadow_vmcs12() local
945 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_max_atomic_switch_msrs() local
1001 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_get_vmexit_msr_value() local
1099 struct vcpu_vmx *vmx = to_vmx(vcpu); in prepare_vmx_msr_autostore_list() local
1194 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_transition_tlb_flush() local
1255 static int vmx_restore_vmx_basic(struct vcpu_vmx *vmx, u64 data) in vmx_restore_vmx_basic()
1326 vmx_restore_control_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) in vmx_restore_control_msr()
1349 static int vmx_restore_vmx_misc(struct vcpu_vmx *vmx, u64 data) in vmx_restore_vmx_misc()
1397 static int vmx_restore_vmx_ept_vpid_cap(struct vcpu_vmx *vmx, u64 data) in vmx_restore_vmx_ept_vpid_cap()
1423 static int vmx_restore_fixed0_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data) in vmx_restore_fixed0_msr()
1445 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_vmx_msr() local
1593 static void copy_shadow_to_vmcs12(struct vcpu_vmx *vmx) in copy_shadow_to_vmcs12()
1620 static void copy_vmcs12_to_shadow(struct vcpu_vmx *vmx) in copy_vmcs12_to_shadow()
1654 static void copy_enlightened_to_vmcs12(struct vcpu_vmx *vmx, u32 hv_clean_fields) in copy_enlightened_to_vmcs12()
1902 static void copy_vmcs12_to_enlightened(struct vcpu_vmx *vmx) in copy_vmcs12_to_enlightened()
2088 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_handle_enlightened_vmptrld() local
2176 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_sync_vmcs12_to_shadow() local
2188 struct vcpu_vmx *vmx = in vmx_preemption_timer_fn() local
2200 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_calc_preemption_timer_value() local
2217 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_start_preemption_timer() local
2239 static u64 nested_vmx_calc_efer(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) in nested_vmx_calc_efer()
2250 static void prepare_vmcs02_constant_state(struct vcpu_vmx *vmx) in prepare_vmcs02_constant_state()
2314 static void prepare_vmcs02_early_rare(struct vcpu_vmx *vmx, in prepare_vmcs02_early_rare()
2340 static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, in prepare_vmcs02_early()
2507 static void prepare_vmcs02_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12) in prepare_vmcs02_rare()
2641 struct vcpu_vmx *vmx = to_vmx(vcpu); in prepare_vmcs02() local
2792 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_check_eptp() local
2841 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_execution_controls() local
2900 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_exit_controls() local
2917 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_entry_controls() local
3096 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_check_vmcs_link_ptr() local
3204 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_check_vmentry_hw() local
3283 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_evmcs_page() local
3313 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vmcs12_pages() local
3428 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_write_pml_buffer() local
3500 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_enter_non_root_mode() local
3662 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_run() local
3919 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_complete_nested_posted_interrupt() local
4069 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_has_nested_events() local
4195 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_check_nested_events() local
4467 struct vcpu_vmx *vmx = to_vmx(vcpu); in sync_vmcs02_to_vmcs12_rare() local
4512 struct vcpu_vmx *vmx = to_vmx(vcpu); in copy_vmcs02_to_vmcs12_rare() local
4540 struct vcpu_vmx *vmx = to_vmx(vcpu); in sync_vmcs02_to_vmcs12() local
4800 static inline u64 nested_vmx_get_vmcs01_guest_efer(struct vcpu_vmx *vmx) in nested_vmx_get_vmcs01_guest_efer()
4826 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_restore_host_state() local
4936 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_vmexit() local
5288 struct vcpu_vmx *vmx = to_vmx(vcpu); in alloc_shadow_vmcs() local
5311 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_vmx_operation() local
5365 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmxon() local
5453 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_vmcs12() local
5495 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmclear() local
5553 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmread() local
5659 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmwrite() local
5760 static void set_current_vmptr(struct vcpu_vmx *vmx, gpa_t vmptr) in set_current_vmptr()
5776 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmptrld() local
5874 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invept() local
5954 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invvpid() local
6086 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmfunc() local
6573 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_reflect_vmexit() local
6627 struct vcpu_vmx *vmx; in vmx_get_nested_state() local
6759 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_nested_state() local