Linux Audio

Check our new training course

Loading...
v5.14.15
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Out-of-line refcount functions.
  4 */
  5
  6#include <linux/mutex.h>
  7#include <linux/refcount.h>
  8#include <linux/spinlock.h>
  9#include <linux/bug.h>
 10
 11#define REFCOUNT_WARN(str)	WARN_ONCE(1, "refcount_t: " str ".\n")
 12
 13void refcount_warn_saturate(refcount_t *r, enum refcount_saturation_type t)
 14{
 15	refcount_set(r, REFCOUNT_SATURATED);
 16
 17	switch (t) {
 18	case REFCOUNT_ADD_NOT_ZERO_OVF:
 19		REFCOUNT_WARN("saturated; leaking memory");
 20		break;
 21	case REFCOUNT_ADD_OVF:
 22		REFCOUNT_WARN("saturated; leaking memory");
 23		break;
 24	case REFCOUNT_ADD_UAF:
 25		REFCOUNT_WARN("addition on 0; use-after-free");
 26		break;
 27	case REFCOUNT_SUB_UAF:
 28		REFCOUNT_WARN("underflow; use-after-free");
 29		break;
 30	case REFCOUNT_DEC_LEAK:
 31		REFCOUNT_WARN("decrement hit 0; leaking memory");
 32		break;
 33	default:
 34		REFCOUNT_WARN("unknown saturation event!?");
 35	}
 36}
 37EXPORT_SYMBOL(refcount_warn_saturate);
 38
 39/**
 40 * refcount_dec_if_one - decrement a refcount if it is 1
 41 * @r: the refcount
 42 *
 43 * No atomic_t counterpart, it attempts a 1 -> 0 transition and returns the
 44 * success thereof.
 45 *
 46 * Like all decrement operations, it provides release memory order and provides
 47 * a control dependency.
 48 *
 49 * It can be used like a try-delete operator; this explicit case is provided
 50 * and not cmpxchg in generic, because that would allow implementing unsafe
 51 * operations.
 52 *
 53 * Return: true if the resulting refcount is 0, false otherwise
 54 */
 55bool refcount_dec_if_one(refcount_t *r)
 56{
 57	int val = 1;
 58
 59	return atomic_try_cmpxchg_release(&r->refs, &val, 0);
 60}
 61EXPORT_SYMBOL(refcount_dec_if_one);
 62
 63/**
 64 * refcount_dec_not_one - decrement a refcount if it is not 1
 65 * @r: the refcount
 66 *
 67 * No atomic_t counterpart, it decrements unless the value is 1, in which case
 68 * it will return false.
 69 *
 70 * Was often done like: atomic_add_unless(&var, -1, 1)
 71 *
 72 * Return: true if the decrement operation was successful, false otherwise
 73 */
 74bool refcount_dec_not_one(refcount_t *r)
 75{
 76	unsigned int new, val = atomic_read(&r->refs);
 77
 78	do {
 79		if (unlikely(val == REFCOUNT_SATURATED))
 80			return true;
 81
 82		if (val == 1)
 83			return false;
 84
 85		new = val - 1;
 86		if (new > val) {
 87			WARN_ONCE(new > val, "refcount_t: underflow; use-after-free.\n");
 88			return true;
 89		}
 90
 91	} while (!atomic_try_cmpxchg_release(&r->refs, &val, new));
 92
 93	return true;
 94}
 95EXPORT_SYMBOL(refcount_dec_not_one);
 96
 97/**
 98 * refcount_dec_and_mutex_lock - return holding mutex if able to decrement
 99 *                               refcount to 0
100 * @r: the refcount
101 * @lock: the mutex to be locked
102 *
103 * Similar to atomic_dec_and_mutex_lock(), it will WARN on underflow and fail
104 * to decrement when saturated at REFCOUNT_SATURATED.
105 *
106 * Provides release memory ordering, such that prior loads and stores are done
107 * before, and provides a control dependency such that free() must come after.
108 * See the comment on top.
109 *
110 * Return: true and hold mutex if able to decrement refcount to 0, false
111 *         otherwise
112 */
113bool refcount_dec_and_mutex_lock(refcount_t *r, struct mutex *lock)
114{
115	if (refcount_dec_not_one(r))
116		return false;
117
118	mutex_lock(lock);
119	if (!refcount_dec_and_test(r)) {
120		mutex_unlock(lock);
121		return false;
122	}
123
124	return true;
125}
126EXPORT_SYMBOL(refcount_dec_and_mutex_lock);
127
128/**
129 * refcount_dec_and_lock - return holding spinlock if able to decrement
130 *                         refcount to 0
131 * @r: the refcount
132 * @lock: the spinlock to be locked
133 *
134 * Similar to atomic_dec_and_lock(), it will WARN on underflow and fail to
135 * decrement when saturated at REFCOUNT_SATURATED.
136 *
137 * Provides release memory ordering, such that prior loads and stores are done
138 * before, and provides a control dependency such that free() must come after.
139 * See the comment on top.
140 *
141 * Return: true and hold spinlock if able to decrement refcount to 0, false
142 *         otherwise
143 */
144bool refcount_dec_and_lock(refcount_t *r, spinlock_t *lock)
145{
146	if (refcount_dec_not_one(r))
147		return false;
148
149	spin_lock(lock);
150	if (!refcount_dec_and_test(r)) {
151		spin_unlock(lock);
152		return false;
153	}
154
155	return true;
156}
157EXPORT_SYMBOL(refcount_dec_and_lock);
158
159/**
160 * refcount_dec_and_lock_irqsave - return holding spinlock with disabled
161 *                                 interrupts if able to decrement refcount to 0
162 * @r: the refcount
163 * @lock: the spinlock to be locked
164 * @flags: saved IRQ-flags if the is acquired
165 *
166 * Same as refcount_dec_and_lock() above except that the spinlock is acquired
167 * with disabled interrupts.
168 *
169 * Return: true and hold spinlock if able to decrement refcount to 0, false
170 *         otherwise
171 */
172bool refcount_dec_and_lock_irqsave(refcount_t *r, spinlock_t *lock,
173				   unsigned long *flags)
174{
175	if (refcount_dec_not_one(r))
176		return false;
177
178	spin_lock_irqsave(lock, *flags);
179	if (!refcount_dec_and_test(r)) {
180		spin_unlock_irqrestore(lock, *flags);
181		return false;
182	}
183
184	return true;
185}
186EXPORT_SYMBOL(refcount_dec_and_lock_irqsave);
v6.13.7
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Out-of-line refcount functions.
  4 */
  5
  6#include <linux/mutex.h>
  7#include <linux/refcount.h>
  8#include <linux/spinlock.h>
  9#include <linux/bug.h>
 10
 11#define REFCOUNT_WARN(str)	WARN_ONCE(1, "refcount_t: " str ".\n")
 12
 13void refcount_warn_saturate(refcount_t *r, enum refcount_saturation_type t)
 14{
 15	refcount_set(r, REFCOUNT_SATURATED);
 16
 17	switch (t) {
 18	case REFCOUNT_ADD_NOT_ZERO_OVF:
 19		REFCOUNT_WARN("saturated; leaking memory");
 20		break;
 21	case REFCOUNT_ADD_OVF:
 22		REFCOUNT_WARN("saturated; leaking memory");
 23		break;
 24	case REFCOUNT_ADD_UAF:
 25		REFCOUNT_WARN("addition on 0; use-after-free");
 26		break;
 27	case REFCOUNT_SUB_UAF:
 28		REFCOUNT_WARN("underflow; use-after-free");
 29		break;
 30	case REFCOUNT_DEC_LEAK:
 31		REFCOUNT_WARN("decrement hit 0; leaking memory");
 32		break;
 33	default:
 34		REFCOUNT_WARN("unknown saturation event!?");
 35	}
 36}
 37EXPORT_SYMBOL(refcount_warn_saturate);
 38
 39/**
 40 * refcount_dec_if_one - decrement a refcount if it is 1
 41 * @r: the refcount
 42 *
 43 * No atomic_t counterpart, it attempts a 1 -> 0 transition and returns the
 44 * success thereof.
 45 *
 46 * Like all decrement operations, it provides release memory order and provides
 47 * a control dependency.
 48 *
 49 * It can be used like a try-delete operator; this explicit case is provided
 50 * and not cmpxchg in generic, because that would allow implementing unsafe
 51 * operations.
 52 *
 53 * Return: true if the resulting refcount is 0, false otherwise
 54 */
 55bool refcount_dec_if_one(refcount_t *r)
 56{
 57	int val = 1;
 58
 59	return atomic_try_cmpxchg_release(&r->refs, &val, 0);
 60}
 61EXPORT_SYMBOL(refcount_dec_if_one);
 62
 63/**
 64 * refcount_dec_not_one - decrement a refcount if it is not 1
 65 * @r: the refcount
 66 *
 67 * No atomic_t counterpart, it decrements unless the value is 1, in which case
 68 * it will return false.
 69 *
 70 * Was often done like: atomic_add_unless(&var, -1, 1)
 71 *
 72 * Return: true if the decrement operation was successful, false otherwise
 73 */
 74bool refcount_dec_not_one(refcount_t *r)
 75{
 76	unsigned int new, val = atomic_read(&r->refs);
 77
 78	do {
 79		if (unlikely(val == REFCOUNT_SATURATED))
 80			return true;
 81
 82		if (val == 1)
 83			return false;
 84
 85		new = val - 1;
 86		if (new > val) {
 87			WARN_ONCE(new > val, "refcount_t: underflow; use-after-free.\n");
 88			return true;
 89		}
 90
 91	} while (!atomic_try_cmpxchg_release(&r->refs, &val, new));
 92
 93	return true;
 94}
 95EXPORT_SYMBOL(refcount_dec_not_one);
 96
 97/**
 98 * refcount_dec_and_mutex_lock - return holding mutex if able to decrement
 99 *                               refcount to 0
100 * @r: the refcount
101 * @lock: the mutex to be locked
102 *
103 * Similar to atomic_dec_and_mutex_lock(), it will WARN on underflow and fail
104 * to decrement when saturated at REFCOUNT_SATURATED.
105 *
106 * Provides release memory ordering, such that prior loads and stores are done
107 * before, and provides a control dependency such that free() must come after.
108 * See the comment on top.
109 *
110 * Return: true and hold mutex if able to decrement refcount to 0, false
111 *         otherwise
112 */
113bool refcount_dec_and_mutex_lock(refcount_t *r, struct mutex *lock)
114{
115	if (refcount_dec_not_one(r))
116		return false;
117
118	mutex_lock(lock);
119	if (!refcount_dec_and_test(r)) {
120		mutex_unlock(lock);
121		return false;
122	}
123
124	return true;
125}
126EXPORT_SYMBOL(refcount_dec_and_mutex_lock);
127
128/**
129 * refcount_dec_and_lock - return holding spinlock if able to decrement
130 *                         refcount to 0
131 * @r: the refcount
132 * @lock: the spinlock to be locked
133 *
134 * Similar to atomic_dec_and_lock(), it will WARN on underflow and fail to
135 * decrement when saturated at REFCOUNT_SATURATED.
136 *
137 * Provides release memory ordering, such that prior loads and stores are done
138 * before, and provides a control dependency such that free() must come after.
139 * See the comment on top.
140 *
141 * Return: true and hold spinlock if able to decrement refcount to 0, false
142 *         otherwise
143 */
144bool refcount_dec_and_lock(refcount_t *r, spinlock_t *lock)
145{
146	if (refcount_dec_not_one(r))
147		return false;
148
149	spin_lock(lock);
150	if (!refcount_dec_and_test(r)) {
151		spin_unlock(lock);
152		return false;
153	}
154
155	return true;
156}
157EXPORT_SYMBOL(refcount_dec_and_lock);
158
159/**
160 * refcount_dec_and_lock_irqsave - return holding spinlock with disabled
161 *                                 interrupts if able to decrement refcount to 0
162 * @r: the refcount
163 * @lock: the spinlock to be locked
164 * @flags: saved IRQ-flags if the is acquired
165 *
166 * Same as refcount_dec_and_lock() above except that the spinlock is acquired
167 * with disabled interrupts.
168 *
169 * Return: true and hold spinlock if able to decrement refcount to 0, false
170 *         otherwise
171 */
172bool refcount_dec_and_lock_irqsave(refcount_t *r, spinlock_t *lock,
173				   unsigned long *flags)
174{
175	if (refcount_dec_not_one(r))
176		return false;
177
178	spin_lock_irqsave(lock, *flags);
179	if (!refcount_dec_and_test(r)) {
180		spin_unlock_irqrestore(lock, *flags);
181		return false;
182	}
183
184	return true;
185}
186EXPORT_SYMBOL(refcount_dec_and_lock_irqsave);