@@ -5122,7 +5122,7 @@ void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
51225122static int kvm_vcpu_ioctl_get_lapic (struct kvm_vcpu * vcpu ,
51235123 struct kvm_lapic_state * s )
51245124{
5125- static_call_cond (kvm_x86_sync_pir_to_irr )(vcpu );
5125+ static_call (kvm_x86_sync_pir_to_irr )(vcpu );
51265126
51275127 return kvm_apic_get_state (vcpu , s );
51285128}
@@ -9336,7 +9336,7 @@ int x86_emulate_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
93369336 kvm_rip_write (vcpu , ctxt -> eip );
93379337 if (r && (ctxt -> tf || (vcpu -> guest_debug & KVM_GUESTDBG_SINGLESTEP )))
93389338 r = kvm_vcpu_do_singlestep (vcpu );
9339- static_call_cond (kvm_x86_update_emulated_instruction )(vcpu );
9339+ static_call (kvm_x86_update_emulated_instruction )(vcpu );
93409340 __kvm_set_rflags (vcpu , ctxt -> eflags );
93419341 }
93429342
@@ -10759,7 +10759,7 @@ static void vcpu_scan_ioapic(struct kvm_vcpu *vcpu)
1075910759
1076010760 bitmap_zero (vcpu -> arch .ioapic_handled_vectors , 256 );
1076110761
10762- static_call_cond (kvm_x86_sync_pir_to_irr )(vcpu );
10762+ static_call (kvm_x86_sync_pir_to_irr )(vcpu );
1076310763
1076410764 if (irqchip_split (vcpu -> kvm ))
1076510765 kvm_scan_ioapic_routes (vcpu , vcpu -> arch .ioapic_handled_vectors );
@@ -10784,25 +10784,25 @@ static void vcpu_load_eoi_exitmap(struct kvm_vcpu *vcpu)
1078410784 bitmap_or ((ulong * )eoi_exit_bitmap ,
1078510785 vcpu -> arch .ioapic_handled_vectors ,
1078610786 to_hv_synic (vcpu )-> vec_bitmap , 256 );
10787- static_call_cond (kvm_x86_load_eoi_exitmap )(vcpu , eoi_exit_bitmap );
10787+ static_call (kvm_x86_load_eoi_exitmap )(vcpu , eoi_exit_bitmap );
1078810788 return ;
1078910789 }
1079010790#endif
10791- static_call_cond (kvm_x86_load_eoi_exitmap )(
10791+ static_call (kvm_x86_load_eoi_exitmap )(
1079210792 vcpu , (u64 * )vcpu -> arch .ioapic_handled_vectors );
1079310793}
1079410794
1079510795void kvm_arch_guest_memory_reclaimed (struct kvm * kvm )
1079610796{
10797- static_call_cond (kvm_x86_guest_memory_reclaimed )(kvm );
10797+ static_call (kvm_x86_guest_memory_reclaimed )(kvm );
1079810798}
1079910799
1080010800static void kvm_vcpu_reload_apic_access_page (struct kvm_vcpu * vcpu )
1080110801{
1080210802 if (!lapic_in_kernel (vcpu ))
1080310803 return ;
1080410804
10805- static_call_cond (kvm_x86_set_apic_access_page_addr )(vcpu );
10805+ static_call (kvm_x86_set_apic_access_page_addr )(vcpu );
1080610806}
1080710807
1080810808/*
@@ -11050,7 +11050,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
1105011050 * i.e. they can post interrupts even if APICv is temporarily disabled.
1105111051 */
1105211052 if (kvm_lapic_enabled (vcpu ))
11053- static_call_cond (kvm_x86_sync_pir_to_irr )(vcpu );
11053+ static_call (kvm_x86_sync_pir_to_irr )(vcpu );
1105411054
1105511055 if (kvm_vcpu_exit_request (vcpu )) {
1105611056 vcpu -> mode = OUTSIDE_GUEST_MODE ;
@@ -11099,7 +11099,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
1109911099 break ;
1110011100
1110111101 if (kvm_lapic_enabled (vcpu ))
11102- static_call_cond (kvm_x86_sync_pir_to_irr )(vcpu );
11102+ static_call (kvm_x86_sync_pir_to_irr )(vcpu );
1110311103
1110411104 if (unlikely (kvm_vcpu_exit_request (vcpu ))) {
1110511105 exit_fastpath = EXIT_FASTPATH_EXIT_HANDLED ;
@@ -11873,7 +11873,7 @@ static int __set_sregs_common(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs,
1187311873 * mmu_reset_needed |= kvm_read_cr3 (vcpu ) != sregs -> cr3 ;
1187411874 vcpu -> arch .cr3 = sregs -> cr3 ;
1187511875 kvm_register_mark_dirty (vcpu , VCPU_EXREG_CR3 );
11876- static_call_cond (kvm_x86_post_set_cr3 )(vcpu , sregs -> cr3 );
11876+ static_call (kvm_x86_post_set_cr3 )(vcpu , sregs -> cr3 );
1187711877
1187811878 kvm_set_cr8 (vcpu , sregs -> cr8 );
1187911879
@@ -12822,7 +12822,7 @@ void kvm_arch_destroy_vm(struct kvm *kvm)
1282212822 mutex_unlock (& kvm -> slots_lock );
1282312823 }
1282412824 kvm_unload_vcpu_mmus (kvm );
12825- static_call_cond (kvm_x86_vm_destroy )(kvm );
12825+ static_call (kvm_x86_vm_destroy )(kvm );
1282612826 kvm_free_msr_filter (srcu_dereference_check (kvm -> arch .msr_filter , & kvm -> srcu , 1 ));
1282712827 kvm_pic_destroy (kvm );
1282812828 kvm_ioapic_destroy (kvm );
@@ -13513,7 +13513,7 @@ bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu)
1351313513void kvm_arch_start_assignment (struct kvm * kvm )
1351413514{
1351513515 if (atomic_inc_return (& kvm -> arch .assigned_device_count ) == 1 )
13516- static_call_cond (kvm_x86_pi_start_assignment )(kvm );
13516+ static_call (kvm_x86_pi_start_assignment )(kvm );
1351713517}
1351813518EXPORT_SYMBOL_GPL (kvm_arch_start_assignment );
1351913519
@@ -13650,7 +13650,7 @@ int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_ord
1365013650#ifdef CONFIG_HAVE_KVM_GMEM_INVALIDATE
1365113651void kvm_arch_gmem_invalidate (kvm_pfn_t start , kvm_pfn_t end )
1365213652{
13653- static_call_cond (kvm_x86_gmem_invalidate )(start , end );
13653+ static_call (kvm_x86_gmem_invalidate )(start , end );
1365413654}
1365513655#endif
1365613656
0 commit comments