Linux Audio

Check our new training course

Loading...
  1/* SPDX-License-Identifier: GPL-2.0-only */
  2/*
  3 * A stand-alone ticket spinlock implementation for use by the non-VHE
  4 * KVM hypervisor code running at EL2.
  5 *
  6 * Copyright (C) 2020 Google LLC
  7 * Author: Will Deacon <will@kernel.org>
  8 *
  9 * Heavily based on the implementation removed by c11090474d70 which was:
 10 * Copyright (C) 2012 ARM Ltd.
 11 */
 12
 13#ifndef __ARM64_KVM_NVHE_SPINLOCK_H__
 14#define __ARM64_KVM_NVHE_SPINLOCK_H__
 15
 16#include <asm/alternative.h>
 17#include <asm/lse.h>
 18#include <asm/rwonce.h>
 19
 20typedef union hyp_spinlock {
 21	u32	__val;
 22	struct {
 23#ifdef __AARCH64EB__
 24		u16 next, owner;
 25#else
 26		u16 owner, next;
 27#endif
 28	};
 29} hyp_spinlock_t;
 30
 31#define __HYP_SPIN_LOCK_INITIALIZER \
 32	{ .__val = 0 }
 33
 34#define __HYP_SPIN_LOCK_UNLOCKED \
 35	((hyp_spinlock_t) __HYP_SPIN_LOCK_INITIALIZER)
 36
 37#define DEFINE_HYP_SPINLOCK(x)	hyp_spinlock_t x = __HYP_SPIN_LOCK_UNLOCKED
 38
 39#define hyp_spin_lock_init(l)						\
 40do {									\
 41	*(l) = __HYP_SPIN_LOCK_UNLOCKED;				\
 42} while (0)
 43
 44static inline void hyp_spin_lock(hyp_spinlock_t *lock)
 45{
 46	u32 tmp;
 47	hyp_spinlock_t lockval, newval;
 48
 49	asm volatile(
 50	/* Atomically increment the next ticket. */
 51	ARM64_LSE_ATOMIC_INSN(
 52	/* LL/SC */
 53"	prfm	pstl1strm, %3\n"
 54"1:	ldaxr	%w0, %3\n"
 55"	add	%w1, %w0, #(1 << 16)\n"
 56"	stxr	%w2, %w1, %3\n"
 57"	cbnz	%w2, 1b\n",
 58	/* LSE atomics */
 59"	mov	%w2, #(1 << 16)\n"
 60"	ldadda	%w2, %w0, %3\n"
 61	__nops(3))
 62
 63	/* Did we get the lock? */
 64"	eor	%w1, %w0, %w0, ror #16\n"
 65"	cbz	%w1, 3f\n"
 66	/*
 67	 * No: spin on the owner. Send a local event to avoid missing an
 68	 * unlock before the exclusive load.
 69	 */
 70"	sevl\n"
 71"2:	wfe\n"
 72"	ldaxrh	%w2, %4\n"
 73"	eor	%w1, %w2, %w0, lsr #16\n"
 74"	cbnz	%w1, 2b\n"
 75	/* We got the lock. Critical section starts here. */
 76"3:"
 77	: "=&r" (lockval), "=&r" (newval), "=&r" (tmp), "+Q" (*lock)
 78	: "Q" (lock->owner)
 79	: "memory");
 80}
 81
 82static inline void hyp_spin_unlock(hyp_spinlock_t *lock)
 83{
 84	u64 tmp;
 85
 86	asm volatile(
 87	ARM64_LSE_ATOMIC_INSN(
 88	/* LL/SC */
 89	"	ldrh	%w1, %0\n"
 90	"	add	%w1, %w1, #1\n"
 91	"	stlrh	%w1, %0",
 92	/* LSE atomics */
 93	"	mov	%w1, #1\n"
 94	"	staddlh	%w1, %0\n"
 95	__nops(1))
 96	: "=Q" (lock->owner), "=&r" (tmp)
 97	:
 98	: "memory");
 99}
100
101static inline bool hyp_spin_is_locked(hyp_spinlock_t *lock)
102{
103	hyp_spinlock_t lockval = READ_ONCE(*lock);
104
105	return lockval.owner != lockval.next;
106}
107
108#ifdef CONFIG_NVHE_EL2_DEBUG
109static inline void hyp_assert_lock_held(hyp_spinlock_t *lock)
110{
111	/*
112	 * The __pkvm_init() path accesses protected data-structures without
113	 * holding locks as the other CPUs are guaranteed to not enter EL2
114	 * concurrently at this point in time. The point by which EL2 is
115	 * initialized on all CPUs is reflected in the pkvm static key, so
116	 * wait until it is set before checking the lock state.
117	 */
118	if (static_branch_likely(&kvm_protected_mode_initialized))
119		BUG_ON(!hyp_spin_is_locked(lock));
120}
121#else
122static inline void hyp_assert_lock_held(hyp_spinlock_t *lock) { }
123#endif
124
125#endif /* __ARM64_KVM_NVHE_SPINLOCK_H__ */