Linux Audio

Check our new training course

Loading...
Note: File does not exist in v4.17.
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * PREEMPT_RT substitution for spin/rw_locks
  4 *
  5 * spinlocks and rwlocks on RT are based on rtmutexes, with a few twists to
  6 * resemble the non RT semantics:
  7 *
  8 * - Contrary to plain rtmutexes, spinlocks and rwlocks are state
  9 *   preserving. The task state is saved before blocking on the underlying
 10 *   rtmutex, and restored when the lock has been acquired. Regular wakeups
 11 *   during that time are redirected to the saved state so no wake up is
 12 *   missed.
 13 *
 14 * - Non RT spin/rwlocks disable preemption and eventually interrupts.
 15 *   Disabling preemption has the side effect of disabling migration and
 16 *   preventing RCU grace periods.
 17 *
 18 *   The RT substitutions explicitly disable migration and take
 19 *   rcu_read_lock() across the lock held section.
 20 */
 21#include <linux/spinlock.h>
 22#include <linux/export.h>
 23
 24#define RT_MUTEX_BUILD_SPINLOCKS
 25#include "rtmutex.c"
 26
 27/*
 28 * __might_resched() skips the state check as rtlocks are state
 29 * preserving. Take RCU nesting into account as spin/read/write_lock() can
 30 * legitimately nest into an RCU read side critical section.
 31 */
 32#define RTLOCK_RESCHED_OFFSETS						\
 33	(rcu_preempt_depth() << MIGHT_RESCHED_RCU_SHIFT)
 34
 35#define rtlock_might_resched()						\
 36	__might_resched(__FILE__, __LINE__, RTLOCK_RESCHED_OFFSETS)
 37
 38static __always_inline void rtlock_lock(struct rt_mutex_base *rtm)
 39{
 40	if (unlikely(!rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
 41		rtlock_slowlock(rtm);
 42}
 43
 44static __always_inline void __rt_spin_lock(spinlock_t *lock)
 45{
 46	rtlock_might_resched();
 47	rtlock_lock(&lock->lock);
 48	rcu_read_lock();
 49	migrate_disable();
 50}
 51
 52void __sched rt_spin_lock(spinlock_t *lock)
 53{
 54	spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
 55	__rt_spin_lock(lock);
 56}
 57EXPORT_SYMBOL(rt_spin_lock);
 58
 59#ifdef CONFIG_DEBUG_LOCK_ALLOC
 60void __sched rt_spin_lock_nested(spinlock_t *lock, int subclass)
 61{
 62	spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_);
 63	__rt_spin_lock(lock);
 64}
 65EXPORT_SYMBOL(rt_spin_lock_nested);
 66
 67void __sched rt_spin_lock_nest_lock(spinlock_t *lock,
 68				    struct lockdep_map *nest_lock)
 69{
 70	spin_acquire_nest(&lock->dep_map, 0, 0, nest_lock, _RET_IP_);
 71	__rt_spin_lock(lock);
 72}
 73EXPORT_SYMBOL(rt_spin_lock_nest_lock);
 74#endif
 75
 76void __sched rt_spin_unlock(spinlock_t *lock)
 77{
 78	spin_release(&lock->dep_map, _RET_IP_);
 79	migrate_enable();
 80	rcu_read_unlock();
 81
 82	if (unlikely(!rt_mutex_cmpxchg_release(&lock->lock, current, NULL)))
 83		rt_mutex_slowunlock(&lock->lock);
 84}
 85EXPORT_SYMBOL(rt_spin_unlock);
 86
 87/*
 88 * Wait for the lock to get unlocked: instead of polling for an unlock
 89 * (like raw spinlocks do), lock and unlock, to force the kernel to
 90 * schedule if there's contention:
 91 */
 92void __sched rt_spin_lock_unlock(spinlock_t *lock)
 93{
 94	spin_lock(lock);
 95	spin_unlock(lock);
 96}
 97EXPORT_SYMBOL(rt_spin_lock_unlock);
 98
 99static __always_inline int __rt_spin_trylock(spinlock_t *lock)
100{
101	int ret = 1;
102
103	if (unlikely(!rt_mutex_cmpxchg_acquire(&lock->lock, NULL, current)))
104		ret = rt_mutex_slowtrylock(&lock->lock);
105
106	if (ret) {
107		spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
108		rcu_read_lock();
109		migrate_disable();
110	}
111	return ret;
112}
113
114int __sched rt_spin_trylock(spinlock_t *lock)
115{
116	return __rt_spin_trylock(lock);
117}
118EXPORT_SYMBOL(rt_spin_trylock);
119
120int __sched rt_spin_trylock_bh(spinlock_t *lock)
121{
122	int ret;
123
124	local_bh_disable();
125	ret = __rt_spin_trylock(lock);
126	if (!ret)
127		local_bh_enable();
128	return ret;
129}
130EXPORT_SYMBOL(rt_spin_trylock_bh);
131
132#ifdef CONFIG_DEBUG_LOCK_ALLOC
133void __rt_spin_lock_init(spinlock_t *lock, const char *name,
134			 struct lock_class_key *key, bool percpu)
135{
136	u8 type = percpu ? LD_LOCK_PERCPU : LD_LOCK_NORMAL;
137
138	debug_check_no_locks_freed((void *)lock, sizeof(*lock));
139	lockdep_init_map_type(&lock->dep_map, name, key, 0, LD_WAIT_CONFIG,
140			      LD_WAIT_INV, type);
141}
142EXPORT_SYMBOL(__rt_spin_lock_init);
143#endif
144
145/*
146 * RT-specific reader/writer locks
147 */
148#define rwbase_set_and_save_current_state(state)	\
149	current_save_and_set_rtlock_wait_state()
150
151#define rwbase_restore_current_state()			\
152	current_restore_rtlock_saved_state()
153
154static __always_inline int
155rwbase_rtmutex_lock_state(struct rt_mutex_base *rtm, unsigned int state)
156{
157	if (unlikely(!rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
158		rtlock_slowlock(rtm);
159	return 0;
160}
161
162static __always_inline int
163rwbase_rtmutex_slowlock_locked(struct rt_mutex_base *rtm, unsigned int state)
164{
165	rtlock_slowlock_locked(rtm);
166	return 0;
167}
168
169static __always_inline void rwbase_rtmutex_unlock(struct rt_mutex_base *rtm)
170{
171	if (likely(rt_mutex_cmpxchg_acquire(rtm, current, NULL)))
172		return;
173
174	rt_mutex_slowunlock(rtm);
175}
176
177static __always_inline int  rwbase_rtmutex_trylock(struct rt_mutex_base *rtm)
178{
179	if (likely(rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
180		return 1;
181
182	return rt_mutex_slowtrylock(rtm);
183}
184
185#define rwbase_signal_pending_state(state, current)	(0)
186
187#define rwbase_schedule()				\
188	schedule_rtlock()
189
190#include "rwbase_rt.c"
191/*
192 * The common functions which get wrapped into the rwlock API.
193 */
194int __sched rt_read_trylock(rwlock_t *rwlock)
195{
196	int ret;
197
198	ret = rwbase_read_trylock(&rwlock->rwbase);
199	if (ret) {
200		rwlock_acquire_read(&rwlock->dep_map, 0, 1, _RET_IP_);
201		rcu_read_lock();
202		migrate_disable();
203	}
204	return ret;
205}
206EXPORT_SYMBOL(rt_read_trylock);
207
208int __sched rt_write_trylock(rwlock_t *rwlock)
209{
210	int ret;
211
212	ret = rwbase_write_trylock(&rwlock->rwbase);
213	if (ret) {
214		rwlock_acquire(&rwlock->dep_map, 0, 1, _RET_IP_);
215		rcu_read_lock();
216		migrate_disable();
217	}
218	return ret;
219}
220EXPORT_SYMBOL(rt_write_trylock);
221
222void __sched rt_read_lock(rwlock_t *rwlock)
223{
224	rtlock_might_resched();
225	rwlock_acquire_read(&rwlock->dep_map, 0, 0, _RET_IP_);
226	rwbase_read_lock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
227	rcu_read_lock();
228	migrate_disable();
229}
230EXPORT_SYMBOL(rt_read_lock);
231
232void __sched rt_write_lock(rwlock_t *rwlock)
233{
234	rtlock_might_resched();
235	rwlock_acquire(&rwlock->dep_map, 0, 0, _RET_IP_);
236	rwbase_write_lock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
237	rcu_read_lock();
238	migrate_disable();
239}
240EXPORT_SYMBOL(rt_write_lock);
241
242#ifdef CONFIG_DEBUG_LOCK_ALLOC
243void __sched rt_write_lock_nested(rwlock_t *rwlock, int subclass)
244{
245	rtlock_might_resched();
246	rwlock_acquire(&rwlock->dep_map, subclass, 0, _RET_IP_);
247	rwbase_write_lock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
248	rcu_read_lock();
249	migrate_disable();
250}
251EXPORT_SYMBOL(rt_write_lock_nested);
252#endif
253
254void __sched rt_read_unlock(rwlock_t *rwlock)
255{
256	rwlock_release(&rwlock->dep_map, _RET_IP_);
257	migrate_enable();
258	rcu_read_unlock();
259	rwbase_read_unlock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
260}
261EXPORT_SYMBOL(rt_read_unlock);
262
263void __sched rt_write_unlock(rwlock_t *rwlock)
264{
265	rwlock_release(&rwlock->dep_map, _RET_IP_);
266	rcu_read_unlock();
267	migrate_enable();
268	rwbase_write_unlock(&rwlock->rwbase);
269}
270EXPORT_SYMBOL(rt_write_unlock);
271
272#ifdef CONFIG_DEBUG_LOCK_ALLOC
273void __rt_rwlock_init(rwlock_t *rwlock, const char *name,
274		      struct lock_class_key *key)
275{
276	debug_check_no_locks_freed((void *)rwlock, sizeof(*rwlock));
277	lockdep_init_map_wait(&rwlock->dep_map, name, key, 0, LD_WAIT_CONFIG);
278}
279EXPORT_SYMBOL(__rt_rwlock_init);
280#endif