Linux Audio

Check our new training course

Loading...
Note: File does not exist in v5.4.
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Copyright © 2019 Oracle and/or its affiliates. All rights reserved.
  4 * Copyright © 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved.
  5 *
  6 * KVM Xen emulation
  7 */
  8
  9#ifndef __ARCH_X86_KVM_XEN_H__
 10#define __ARCH_X86_KVM_XEN_H__
 11
 12#include <asm/xen/hypervisor.h>
 13
 14#ifdef CONFIG_KVM_XEN
 15#include <linux/jump_label_ratelimit.h>
 16
 17extern struct static_key_false_deferred kvm_xen_enabled;
 18
 19int __kvm_xen_has_interrupt(struct kvm_vcpu *vcpu);
 20void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu);
 21void kvm_xen_inject_vcpu_vector(struct kvm_vcpu *vcpu);
 22int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
 23int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
 24int kvm_xen_hvm_set_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
 25int kvm_xen_hvm_get_attr(struct kvm *kvm, struct kvm_xen_hvm_attr *data);
 26int kvm_xen_hvm_evtchn_send(struct kvm *kvm, struct kvm_irq_routing_xen_evtchn *evt);
 27int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data);
 28int kvm_xen_hvm_config(struct kvm *kvm, struct kvm_xen_hvm_config *xhc);
 29void kvm_xen_init_vm(struct kvm *kvm);
 30void kvm_xen_destroy_vm(struct kvm *kvm);
 31void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu);
 32void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu);
 33int kvm_xen_set_evtchn_fast(struct kvm_xen_evtchn *xe,
 34			    struct kvm *kvm);
 35int kvm_xen_setup_evtchn(struct kvm *kvm,
 36			 struct kvm_kernel_irq_routing_entry *e,
 37			 const struct kvm_irq_routing_entry *ue);
 38void kvm_xen_update_tsc_info(struct kvm_vcpu *vcpu);
 39
 40static inline void kvm_xen_sw_enable_lapic(struct kvm_vcpu *vcpu)
 41{
 42	/*
 43	 * The local APIC is being enabled. If the per-vCPU upcall vector is
 44	 * set and the vCPU's evtchn_upcall_pending flag is set, inject the
 45	 * interrupt.
 46	 */
 47	if (static_branch_unlikely(&kvm_xen_enabled.key) &&
 48	    vcpu->arch.xen.vcpu_info_cache.active &&
 49	    vcpu->arch.xen.upcall_vector && __kvm_xen_has_interrupt(vcpu))
 50		kvm_xen_inject_vcpu_vector(vcpu);
 51}
 52
 53static inline bool kvm_xen_msr_enabled(struct kvm *kvm)
 54{
 55	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 56		kvm->arch.xen_hvm_config.msr;
 57}
 58
 59static inline bool kvm_xen_hypercall_enabled(struct kvm *kvm)
 60{
 61	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 62		(kvm->arch.xen_hvm_config.flags &
 63		 KVM_XEN_HVM_CONFIG_INTERCEPT_HCALL);
 64}
 65
 66static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
 67{
 68	if (static_branch_unlikely(&kvm_xen_enabled.key) &&
 69	    vcpu->arch.xen.vcpu_info_cache.active &&
 70	    vcpu->kvm->arch.xen.upcall_vector)
 71		return __kvm_xen_has_interrupt(vcpu);
 72
 73	return 0;
 74}
 75
 76static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
 77{
 78	return static_branch_unlikely(&kvm_xen_enabled.key) &&
 79		vcpu->arch.xen.evtchn_pending_sel;
 80}
 81
 82static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
 83{
 84	return !!vcpu->arch.xen.timer_virq;
 85}
 86
 87static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
 88{
 89	if (kvm_xen_hypercall_enabled(vcpu->kvm) && kvm_xen_timer_enabled(vcpu))
 90		return atomic_read(&vcpu->arch.xen.timer_pending);
 91
 92	return 0;
 93}
 94
 95void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu);
 96#else
 97static inline int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data)
 98{
 99	return 1;
100}
101
102static inline void kvm_xen_init_vm(struct kvm *kvm)
103{
104}
105
106static inline void kvm_xen_destroy_vm(struct kvm *kvm)
107{
108}
109
110static inline void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu)
111{
112}
113
114static inline void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu)
115{
116}
117
118static inline void kvm_xen_sw_enable_lapic(struct kvm_vcpu *vcpu)
119{
120}
121
122static inline bool kvm_xen_msr_enabled(struct kvm *kvm)
123{
124	return false;
125}
126
127static inline bool kvm_xen_hypercall_enabled(struct kvm *kvm)
128{
129	return false;
130}
131
132static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
133{
134	return 0;
135}
136
137static inline void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu)
138{
139}
140
141static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
142{
143	return false;
144}
145
146static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
147{
148	return 0;
149}
150
151static inline void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu)
152{
153}
154
155static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
156{
157	return false;
158}
159
160static inline void kvm_xen_update_tsc_info(struct kvm_vcpu *vcpu)
161{
162}
163#endif
164
165int kvm_xen_hypercall(struct kvm_vcpu *vcpu);
166
167#include <asm/pvclock-abi.h>
168#include <asm/xen/interface.h>
169#include <xen/interface/vcpu.h>
170
171void kvm_xen_update_runstate(struct kvm_vcpu *vcpu, int state);
172
173static inline void kvm_xen_runstate_set_running(struct kvm_vcpu *vcpu)
174{
175	kvm_xen_update_runstate(vcpu, RUNSTATE_running);
176}
177
178static inline void kvm_xen_runstate_set_preempted(struct kvm_vcpu *vcpu)
179{
180	/*
181	 * If the vCPU wasn't preempted but took a normal exit for
182	 * some reason (hypercalls, I/O, etc.), that is accounted as
183	 * still RUNSTATE_running, as the VMM is still operating on
184	 * behalf of the vCPU. Only if the VMM does actually block
185	 * does it need to enter RUNSTATE_blocked.
186	 */
187	if (WARN_ON_ONCE(!vcpu->preempted))
188		return;
189
190	kvm_xen_update_runstate(vcpu, RUNSTATE_runnable);
191}
192
193/* 32-bit compatibility definitions, also used natively in 32-bit build */
194struct compat_arch_vcpu_info {
195	unsigned int cr2;
196	unsigned int pad[5];
197};
198
199struct compat_vcpu_info {
200	uint8_t evtchn_upcall_pending;
201	uint8_t evtchn_upcall_mask;
202	uint16_t pad;
203	uint32_t evtchn_pending_sel;
204	struct compat_arch_vcpu_info arch;
205	struct pvclock_vcpu_time_info time;
206}; /* 64 bytes (x86) */
207
208struct compat_arch_shared_info {
209	unsigned int max_pfn;
210	unsigned int pfn_to_mfn_frame_list_list;
211	unsigned int nmi_reason;
212	unsigned int p2m_cr3;
213	unsigned int p2m_vaddr;
214	unsigned int p2m_generation;
215	uint32_t wc_sec_hi;
216};
217
218struct compat_shared_info {
219	struct compat_vcpu_info vcpu_info[MAX_VIRT_CPUS];
220	uint32_t evtchn_pending[32];
221	uint32_t evtchn_mask[32];
222	struct pvclock_wall_clock wc;
223	struct compat_arch_shared_info arch;
224};
225
226#define COMPAT_EVTCHN_2L_NR_CHANNELS (8 *				\
227				      sizeof_field(struct compat_shared_info, \
228						   evtchn_pending))
229struct compat_vcpu_runstate_info {
230    int state;
231    uint64_t state_entry_time;
232    uint64_t time[4];
233} __attribute__((packed));
234
235struct compat_sched_poll {
236	/* This is actually a guest virtual address which points to ports. */
237	uint32_t ports;
238	unsigned int nr_ports;
239	uint64_t timeout;
240};
241
242#endif /* __ARCH_X86_KVM_XEN_H__ */