Linux Audio

Check our new training course

Loading...
v6.2
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Copyright © 2019 Oracle and/or its affiliates. All rights reserved.
  4 * Copyright © 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved.
  5 *
  6 * KVM Xen emulation
  7 */
  8
  9#ifndef __ARCH_X86_KVM_XEN_H__
 10#define __ARCH_X86_KVM_XEN_H__
 11
 
 
 12#ifdef CONFIG_KVM_XEN
 13#include <linux/jump_label_ratelimit.h>
 14
 15extern struct static_key_false_deferred kvm_xen_enabled;
 16
 17int __kvm_xen_has_interrupt(struct kvm_vcpu *vcpu);
 18void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu);
 19int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
 20int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
 21int kvm_xen_hvm_set_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
 22int kvm_xen_hvm_get_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
 23int kvm_xen_hvm_evtchn_send(struct kvm *kvm, struct kvm_irq_routing_xen_evtchn *evt);
 24int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data);
 25int kvm_xen_hvm_config(struct kvm *kvm, struct kvm_xen_hvm_config *xhc);
 26void kvm_xen_init_vm(struct kvm *kvm);
 27void kvm_xen_destroy_vm(struct kvm *kvm);
 28void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu);
 29void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu);
 30int kvm_xen_set_evtchn_fast(struct kvm_xen_evtchn *xe,
 31			    struct kvm *kvm);
 32int kvm_xen_setup_evtchn(struct kvm *kvm,
 33			 struct kvm_kernel_irq_routing_entry *e,
 34			 const struct kvm_irq_routing_entry *ue);
 
 35
 36static inline bool kvm_xen_msr_enabled(struct kvm *kvm)
 37{
 38	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 39		kvm->arch.xen_hvm_config.msr;
 40}
 41
 42static inline bool kvm_xen_hypercall_enabled(struct kvm *kvm)
 43{
 44	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 45		(kvm->arch.xen_hvm_config.flags &
 46		 KVM_XEN_HVM_CONFIG_INTERCEPT_HCALL);
 47}
 48
 49static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
 50{
 51	if (static_branch_unlikely(&kvm_xen_enabled.key) &&
 52	    vcpu->arch.xen.vcpu_info_cache.active &&
 53	    vcpu->kvm->arch.xen.upcall_vector)
 54		return __kvm_xen_has_interrupt(vcpu);
 55
 56	return 0;
 57}
 58
 59static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
 60{
 61	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 62		vcpu->arch.xen.evtchn_pending_sel;
 63}
 64
 65static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
 66{
 67	return !!vcpu->arch.xen.timer_virq;
 68}
 69
 70static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
 71{
 72	if (kvm_xen_hypercall_enabled(vcpu->kvm) && kvm_xen_timer_enabled(vcpu))
 73		return atomic_read(&vcpu->arch.xen.timer_pending);
 74
 75	return 0;
 76}
 77
 78void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu);
 79#else
 80static inline int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data)
 81{
 82	return 1;
 83}
 84
 85static inline void kvm_xen_init_vm(struct kvm *kvm)
 86{
 87}
 88
 89static inline void kvm_xen_destroy_vm(struct kvm *kvm)
 90{
 91}
 92
 93static inline void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu)
 94{
 95}
 96
 97static inline void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu)
 98{
 99}
100
101static inline bool kvm_xen_msr_enabled(struct kvm *kvm)
102{
103	return false;
104}
105
106static inline bool kvm_xen_hypercall_enabled(struct kvm *kvm)
107{
108	return false;
109}
110
111static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
112{
113	return 0;
114}
115
116static inline void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu)
117{
118}
119
120static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
121{
122	return false;
123}
124
125static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
126{
127	return 0;
128}
129
130static inline void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu)
131{
132}
133
134static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
135{
136	return false;
 
 
 
 
137}
138#endif
139
140int kvm_xen_hypercall(struct kvm_vcpu *vcpu);
141
142#include <asm/pvclock-abi.h>
143#include <asm/xen/interface.h>
144#include <xen/interface/vcpu.h>
145
146void kvm_xen_update_runstate(struct kvm_vcpu *vcpu, int state);
147
148static inline void kvm_xen_runstate_set_running(struct kvm_vcpu *vcpu)
149{
150	kvm_xen_update_runstate(vcpu, RUNSTATE_running);
151}
152
153static inline void kvm_xen_runstate_set_preempted(struct kvm_vcpu *vcpu)
154{
155	/*
156	 * If the vCPU wasn't preempted but took a normal exit for
157	 * some reason (hypercalls, I/O, etc.), that is accounted as
158	 * still RUNSTATE_running, as the VMM is still operating on
159	 * behalf of the vCPU. Only if the VMM does actually block
160	 * does it need to enter RUNSTATE_blocked.
161	 */
162	if (WARN_ON_ONCE(!vcpu->preempted))
163		return;
164
165	kvm_xen_update_runstate(vcpu, RUNSTATE_runnable);
166}
167
168/* 32-bit compatibility definitions, also used natively in 32-bit build */
169struct compat_arch_vcpu_info {
170	unsigned int cr2;
171	unsigned int pad[5];
172};
173
174struct compat_vcpu_info {
175	uint8_t evtchn_upcall_pending;
176	uint8_t evtchn_upcall_mask;
177	uint16_t pad;
178	uint32_t evtchn_pending_sel;
179	struct compat_arch_vcpu_info arch;
180	struct pvclock_vcpu_time_info time;
181}; /* 64 bytes (x86) */
182
183struct compat_arch_shared_info {
184	unsigned int max_pfn;
185	unsigned int pfn_to_mfn_frame_list_list;
186	unsigned int nmi_reason;
187	unsigned int p2m_cr3;
188	unsigned int p2m_vaddr;
189	unsigned int p2m_generation;
190	uint32_t wc_sec_hi;
191};
192
193struct compat_shared_info {
194	struct compat_vcpu_info vcpu_info[MAX_VIRT_CPUS];
195	uint32_t evtchn_pending[32];
196	uint32_t evtchn_mask[32];
197	struct pvclock_wall_clock wc;
198	struct compat_arch_shared_info arch;
199};
200
201#define COMPAT_EVTCHN_2L_NR_CHANNELS (8 *				\
202				      sizeof_field(struct compat_shared_info, \
203						   evtchn_pending))
204struct compat_vcpu_runstate_info {
205    int state;
206    uint64_t state_entry_time;
207    uint64_t time[4];
208} __attribute__((packed));
209
210struct compat_sched_poll {
211	/* This is actually a guest virtual address which points to ports. */
212	uint32_t ports;
213	unsigned int nr_ports;
214	uint64_t timeout;
215};
216
217#endif /* __ARCH_X86_KVM_XEN_H__ */
v6.8
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Copyright © 2019 Oracle and/or its affiliates. All rights reserved.
  4 * Copyright © 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved.
  5 *
  6 * KVM Xen emulation
  7 */
  8
  9#ifndef __ARCH_X86_KVM_XEN_H__
 10#define __ARCH_X86_KVM_XEN_H__
 11
 12#include <asm/xen/hypervisor.h>
 13
 14#ifdef CONFIG_KVM_XEN
 15#include <linux/jump_label_ratelimit.h>
 16
 17extern struct static_key_false_deferred kvm_xen_enabled;
 18
 19int __kvm_xen_has_interrupt(struct kvm_vcpu *vcpu);
 20void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu);
 21int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
 22int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
 23int kvm_xen_hvm_set_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
 24int kvm_xen_hvm_get_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
 25int kvm_xen_hvm_evtchn_send(struct kvm *kvm, struct kvm_irq_routing_xen_evtchn *evt);
 26int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data);
 27int kvm_xen_hvm_config(struct kvm *kvm, struct kvm_xen_hvm_config *xhc);
 28void kvm_xen_init_vm(struct kvm *kvm);
 29void kvm_xen_destroy_vm(struct kvm *kvm);
 30void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu);
 31void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu);
 32int kvm_xen_set_evtchn_fast(struct kvm_xen_evtchn *xe,
 33			    struct kvm *kvm);
 34int kvm_xen_setup_evtchn(struct kvm *kvm,
 35			 struct kvm_kernel_irq_routing_entry *e,
 36			 const struct kvm_irq_routing_entry *ue);
 37void kvm_xen_update_tsc_info(struct kvm_vcpu *vcpu);
 38
 39static inline bool kvm_xen_msr_enabled(struct kvm *kvm)
 40{
 41	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 42		kvm->arch.xen_hvm_config.msr;
 43}
 44
 45static inline bool kvm_xen_hypercall_enabled(struct kvm *kvm)
 46{
 47	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 48		(kvm->arch.xen_hvm_config.flags &
 49		 KVM_XEN_HVM_CONFIG_INTERCEPT_HCALL);
 50}
 51
 52static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
 53{
 54	if (static_branch_unlikely(&kvm_xen_enabled.key) &&
 55	    vcpu->arch.xen.vcpu_info_cache.active &&
 56	    vcpu->kvm->arch.xen.upcall_vector)
 57		return __kvm_xen_has_interrupt(vcpu);
 58
 59	return 0;
 60}
 61
 62static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
 63{
 64	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 65		vcpu->arch.xen.evtchn_pending_sel;
 66}
 67
 68static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
 69{
 70	return !!vcpu->arch.xen.timer_virq;
 71}
 72
 73static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
 74{
 75	if (kvm_xen_hypercall_enabled(vcpu->kvm) && kvm_xen_timer_enabled(vcpu))
 76		return atomic_read(&vcpu->arch.xen.timer_pending);
 77
 78	return 0;
 79}
 80
 81void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu);
 82#else
 83static inline int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data)
 84{
 85	return 1;
 86}
 87
 88static inline void kvm_xen_init_vm(struct kvm *kvm)
 89{
 90}
 91
 92static inline void kvm_xen_destroy_vm(struct kvm *kvm)
 93{
 94}
 95
 96static inline void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu)
 97{
 98}
 99
100static inline void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu)
101{
102}
103
104static inline bool kvm_xen_msr_enabled(struct kvm *kvm)
105{
106	return false;
107}
108
109static inline bool kvm_xen_hypercall_enabled(struct kvm *kvm)
110{
111	return false;
112}
113
114static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
115{
116	return 0;
117}
118
119static inline void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu)
120{
121}
122
123static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
124{
125	return false;
126}
127
128static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
129{
130	return 0;
131}
132
133static inline void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu)
134{
135}
136
137static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
138{
139	return false;
140}
141
142static inline void kvm_xen_update_tsc_info(struct kvm_vcpu *vcpu)
143{
144}
145#endif
146
147int kvm_xen_hypercall(struct kvm_vcpu *vcpu);
148
149#include <asm/pvclock-abi.h>
150#include <asm/xen/interface.h>
151#include <xen/interface/vcpu.h>
152
153void kvm_xen_update_runstate(struct kvm_vcpu *vcpu, int state);
154
155static inline void kvm_xen_runstate_set_running(struct kvm_vcpu *vcpu)
156{
157	kvm_xen_update_runstate(vcpu, RUNSTATE_running);
158}
159
160static inline void kvm_xen_runstate_set_preempted(struct kvm_vcpu *vcpu)
161{
162	/*
163	 * If the vCPU wasn't preempted but took a normal exit for
164	 * some reason (hypercalls, I/O, etc.), that is accounted as
165	 * still RUNSTATE_running, as the VMM is still operating on
166	 * behalf of the vCPU. Only if the VMM does actually block
167	 * does it need to enter RUNSTATE_blocked.
168	 */
169	if (WARN_ON_ONCE(!vcpu->preempted))
170		return;
171
172	kvm_xen_update_runstate(vcpu, RUNSTATE_runnable);
173}
174
175/* 32-bit compatibility definitions, also used natively in 32-bit build */
176struct compat_arch_vcpu_info {
177	unsigned int cr2;
178	unsigned int pad[5];
179};
180
181struct compat_vcpu_info {
182	uint8_t evtchn_upcall_pending;
183	uint8_t evtchn_upcall_mask;
184	uint16_t pad;
185	uint32_t evtchn_pending_sel;
186	struct compat_arch_vcpu_info arch;
187	struct pvclock_vcpu_time_info time;
188}; /* 64 bytes (x86) */
189
190struct compat_arch_shared_info {
191	unsigned int max_pfn;
192	unsigned int pfn_to_mfn_frame_list_list;
193	unsigned int nmi_reason;
194	unsigned int p2m_cr3;
195	unsigned int p2m_vaddr;
196	unsigned int p2m_generation;
197	uint32_t wc_sec_hi;
198};
199
200struct compat_shared_info {
201	struct compat_vcpu_info vcpu_info[MAX_VIRT_CPUS];
202	uint32_t evtchn_pending[32];
203	uint32_t evtchn_mask[32];
204	struct pvclock_wall_clock wc;
205	struct compat_arch_shared_info arch;
206};
207
208#define COMPAT_EVTCHN_2L_NR_CHANNELS (8 *				\
209				      sizeof_field(struct compat_shared_info, \
210						   evtchn_pending))
211struct compat_vcpu_runstate_info {
212    int state;
213    uint64_t state_entry_time;
214    uint64_t time[4];
215} __attribute__((packed));
216
217struct compat_sched_poll {
218	/* This is actually a guest virtual address which points to ports. */
219	uint32_t ports;
220	unsigned int nr_ports;
221	uint64_t timeout;
222};
223
224#endif /* __ARCH_X86_KVM_XEN_H__ */