@@ -272,7 +272,7 @@ static void kvm_xen_update_runstate_guest(struct kvm_vcpu *v, bool atomic)
272
272
* gfn_to_pfn caches that cover the region.
273
273
*/
274
274
read_lock_irqsave (& gpc1 -> lock , flags );
275
- while (!kvm_gpc_check (v -> kvm , gpc1 , gpc1 -> gpa , user_len1 )) {
275
+ while (!kvm_gpc_check (gpc1 , gpc1 -> gpa , user_len1 )) {
276
276
read_unlock_irqrestore (& gpc1 -> lock , flags );
277
277
278
278
/* When invoked from kvm_sched_out() we cannot sleep */
@@ -308,7 +308,7 @@ static void kvm_xen_update_runstate_guest(struct kvm_vcpu *v, bool atomic)
308
308
*/
309
309
read_lock (& gpc2 -> lock );
310
310
311
- if (!kvm_gpc_check (v -> kvm , gpc2 , gpc2 -> gpa , user_len2 )) {
311
+ if (!kvm_gpc_check (gpc2 , gpc2 -> gpa , user_len2 )) {
312
312
read_unlock (& gpc2 -> lock );
313
313
read_unlock_irqrestore (& gpc1 -> lock , flags );
314
314
@@ -488,8 +488,7 @@ void kvm_xen_inject_pending_events(struct kvm_vcpu *v)
488
488
* little more honest about it.
489
489
*/
490
490
read_lock_irqsave (& gpc -> lock , flags );
491
- while (!kvm_gpc_check (v -> kvm , gpc , gpc -> gpa ,
492
- sizeof (struct vcpu_info ))) {
491
+ while (!kvm_gpc_check (gpc , gpc -> gpa , sizeof (struct vcpu_info ))) {
493
492
read_unlock_irqrestore (& gpc -> lock , flags );
494
493
495
494
if (kvm_gpc_refresh (v -> kvm , gpc , gpc -> gpa ,
@@ -553,8 +552,7 @@ int __kvm_xen_has_interrupt(struct kvm_vcpu *v)
553
552
sizeof_field (struct compat_vcpu_info , evtchn_upcall_pending ));
554
553
555
554
read_lock_irqsave (& gpc -> lock , flags );
556
- while (!kvm_gpc_check (v -> kvm , gpc , gpc -> gpa ,
557
- sizeof (struct vcpu_info ))) {
555
+ while (!kvm_gpc_check (gpc , gpc -> gpa , sizeof (struct vcpu_info ))) {
558
556
read_unlock_irqrestore (& gpc -> lock , flags );
559
557
560
558
/*
@@ -1158,7 +1156,7 @@ static bool wait_pending_event(struct kvm_vcpu *vcpu, int nr_ports,
1158
1156
1159
1157
read_lock_irqsave (& gpc -> lock , flags );
1160
1158
idx = srcu_read_lock (& kvm -> srcu );
1161
- if (!kvm_gpc_check (kvm , gpc , gpc -> gpa , PAGE_SIZE ))
1159
+ if (!kvm_gpc_check (gpc , gpc -> gpa , PAGE_SIZE ))
1162
1160
goto out_rcu ;
1163
1161
1164
1162
ret = false;
@@ -1580,7 +1578,7 @@ int kvm_xen_set_evtchn_fast(struct kvm_xen_evtchn *xe, struct kvm *kvm)
1580
1578
idx = srcu_read_lock (& kvm -> srcu );
1581
1579
1582
1580
read_lock_irqsave (& gpc -> lock , flags );
1583
- if (!kvm_gpc_check (kvm , gpc , gpc -> gpa , PAGE_SIZE ))
1581
+ if (!kvm_gpc_check (gpc , gpc -> gpa , PAGE_SIZE ))
1584
1582
goto out_rcu ;
1585
1583
1586
1584
if (IS_ENABLED (CONFIG_64BIT ) && kvm -> arch .xen .long_mode ) {
@@ -1614,7 +1612,7 @@ int kvm_xen_set_evtchn_fast(struct kvm_xen_evtchn *xe, struct kvm *kvm)
1614
1612
gpc = & vcpu -> arch .xen .vcpu_info_cache ;
1615
1613
1616
1614
read_lock_irqsave (& gpc -> lock , flags );
1617
- if (!kvm_gpc_check (kvm , gpc , gpc -> gpa , sizeof (struct vcpu_info ))) {
1615
+ if (!kvm_gpc_check (gpc , gpc -> gpa , sizeof (struct vcpu_info ))) {
1618
1616
/*
1619
1617
* Could not access the vcpu_info. Set the bit in-kernel
1620
1618
* and prod the vCPU to deliver it for itself.
0 commit comments