Linux Audio

Check our new training course

Linux debugging, profiling, tracing and performance analysis training

Apr 14-17, 2025
Register
Loading...
v6.8
  1// SPDX-License-Identifier: GPL-2.0-or-later
  2/* Cache data I/O routines
  3 *
  4 * Copyright (C) 2021 Red Hat, Inc. All Rights Reserved.
  5 * Written by David Howells (dhowells@redhat.com)
  6 */
  7#define FSCACHE_DEBUG_LEVEL OPERATION
  8#include <linux/fscache-cache.h>
  9#include <linux/uio.h>
 10#include <linux/bvec.h>
 11#include <linux/slab.h>
 12#include <linux/uio.h>
 13#include "internal.h"
 14
 15/**
 16 * fscache_wait_for_operation - Wait for an object become accessible
 17 * @cres: The cache resources for the operation being performed
 18 * @want_state: The minimum state the object must be at
 19 *
 20 * See if the target cache object is at the specified minimum state of
 21 * accessibility yet, and if not, wait for it.
 22 */
 23bool fscache_wait_for_operation(struct netfs_cache_resources *cres,
 24				enum fscache_want_state want_state)
 25{
 26	struct fscache_cookie *cookie = fscache_cres_cookie(cres);
 27	enum fscache_cookie_state state;
 28
 29again:
 30	if (!fscache_cache_is_live(cookie->volume->cache)) {
 31		_leave(" [broken]");
 32		return false;
 33	}
 34
 35	state = fscache_cookie_state(cookie);
 36	_enter("c=%08x{%u},%x", cookie->debug_id, state, want_state);
 37
 38	switch (state) {
 39	case FSCACHE_COOKIE_STATE_CREATING:
 40	case FSCACHE_COOKIE_STATE_INVALIDATING:
 41		if (want_state == FSCACHE_WANT_PARAMS)
 42			goto ready; /* There can be no content */
 43		fallthrough;
 44	case FSCACHE_COOKIE_STATE_LOOKING_UP:
 45	case FSCACHE_COOKIE_STATE_LRU_DISCARDING:
 46		wait_var_event(&cookie->state,
 47			       fscache_cookie_state(cookie) != state);
 48		goto again;
 49
 50	case FSCACHE_COOKIE_STATE_ACTIVE:
 51		goto ready;
 52	case FSCACHE_COOKIE_STATE_DROPPED:
 53	case FSCACHE_COOKIE_STATE_RELINQUISHING:
 54	default:
 55		_leave(" [not live]");
 56		return false;
 57	}
 58
 59ready:
 60	if (!cres->cache_priv2)
 61		return cookie->volume->cache->ops->begin_operation(cres, want_state);
 62	return true;
 63}
 64EXPORT_SYMBOL(fscache_wait_for_operation);
 65
 66/*
 67 * Begin an I/O operation on the cache, waiting till we reach the right state.
 68 *
 69 * Attaches the resources required to the operation resources record.
 70 */
 71static int fscache_begin_operation(struct netfs_cache_resources *cres,
 72				   struct fscache_cookie *cookie,
 73				   enum fscache_want_state want_state,
 74				   enum fscache_access_trace why)
 75{
 76	enum fscache_cookie_state state;
 77	long timeo;
 78	bool once_only = false;
 79
 80	cres->ops		= NULL;
 81	cres->cache_priv	= cookie;
 82	cres->cache_priv2	= NULL;
 83	cres->debug_id		= cookie->debug_id;
 84	cres->inval_counter	= cookie->inval_counter;
 85
 86	if (!fscache_begin_cookie_access(cookie, why))
 
 87		return -ENOBUFS;
 
 88
 89again:
 90	spin_lock(&cookie->lock);
 91
 92	state = fscache_cookie_state(cookie);
 93	_enter("c=%08x{%u},%x", cookie->debug_id, state, want_state);
 94
 95	switch (state) {
 96	case FSCACHE_COOKIE_STATE_LOOKING_UP:
 97	case FSCACHE_COOKIE_STATE_LRU_DISCARDING:
 98	case FSCACHE_COOKIE_STATE_INVALIDATING:
 99		goto wait_for_file_wrangling;
100	case FSCACHE_COOKIE_STATE_CREATING:
101		if (want_state == FSCACHE_WANT_PARAMS)
102			goto ready; /* There can be no content */
103		goto wait_for_file_wrangling;
104	case FSCACHE_COOKIE_STATE_ACTIVE:
105		goto ready;
106	case FSCACHE_COOKIE_STATE_DROPPED:
107	case FSCACHE_COOKIE_STATE_RELINQUISHING:
108		WARN(1, "Can't use cookie in state %u\n", cookie->state);
109		goto not_live;
110	default:
111		goto not_live;
112	}
113
114ready:
115	spin_unlock(&cookie->lock);
116	if (!cookie->volume->cache->ops->begin_operation(cres, want_state))
117		goto failed;
118	return 0;
119
120wait_for_file_wrangling:
121	spin_unlock(&cookie->lock);
122	trace_fscache_access(cookie->debug_id, refcount_read(&cookie->ref),
123			     atomic_read(&cookie->n_accesses),
124			     fscache_access_io_wait);
125	timeo = wait_var_event_timeout(&cookie->state,
126				       fscache_cookie_state(cookie) != state, 20 * HZ);
127	if (timeo <= 1 && !once_only) {
128		pr_warn("%s: cookie state change wait timed out: cookie->state=%u state=%u",
129			__func__, fscache_cookie_state(cookie), state);
130		fscache_print_cookie(cookie, 'O');
131		once_only = true;
132	}
133	goto again;
134
135not_live:
136	spin_unlock(&cookie->lock);
137failed:
138	cres->cache_priv = NULL;
139	cres->ops = NULL;
140	fscache_end_cookie_access(cookie, fscache_access_io_not_live);
141	_leave(" = -ENOBUFS");
142	return -ENOBUFS;
143}
144
145int __fscache_begin_read_operation(struct netfs_cache_resources *cres,
146				   struct fscache_cookie *cookie)
147{
148	return fscache_begin_operation(cres, cookie, FSCACHE_WANT_PARAMS,
149				       fscache_access_io_read);
150}
151EXPORT_SYMBOL(__fscache_begin_read_operation);
152
153int __fscache_begin_write_operation(struct netfs_cache_resources *cres,
154				    struct fscache_cookie *cookie)
155{
156	return fscache_begin_operation(cres, cookie, FSCACHE_WANT_PARAMS,
157				       fscache_access_io_write);
158}
159EXPORT_SYMBOL(__fscache_begin_write_operation);
160
161struct fscache_write_request {
162	struct netfs_cache_resources cache_resources;
163	struct address_space	*mapping;
164	loff_t			start;
165	size_t			len;
166	bool			set_bits;
167	netfs_io_terminated_t	term_func;
168	void			*term_func_priv;
169};
170
171void __fscache_clear_page_bits(struct address_space *mapping,
172			       loff_t start, size_t len)
173{
174	pgoff_t first = start / PAGE_SIZE;
175	pgoff_t last = (start + len - 1) / PAGE_SIZE;
176	struct page *page;
177
178	if (len) {
179		XA_STATE(xas, &mapping->i_pages, first);
180
181		rcu_read_lock();
182		xas_for_each(&xas, page, last) {
183			end_page_fscache(page);
184		}
185		rcu_read_unlock();
186	}
187}
188EXPORT_SYMBOL(__fscache_clear_page_bits);
189
190/*
191 * Deal with the completion of writing the data to the cache.
192 */
193static void fscache_wreq_done(void *priv, ssize_t transferred_or_error,
194			      bool was_async)
195{
196	struct fscache_write_request *wreq = priv;
197
198	fscache_clear_page_bits(wreq->mapping, wreq->start, wreq->len,
199				wreq->set_bits);
200
201	if (wreq->term_func)
202		wreq->term_func(wreq->term_func_priv, transferred_or_error,
203				was_async);
204	fscache_end_operation(&wreq->cache_resources);
205	kfree(wreq);
206}
207
208void __fscache_write_to_cache(struct fscache_cookie *cookie,
209			      struct address_space *mapping,
210			      loff_t start, size_t len, loff_t i_size,
211			      netfs_io_terminated_t term_func,
212			      void *term_func_priv,
213			      bool cond)
214{
215	struct fscache_write_request *wreq;
216	struct netfs_cache_resources *cres;
217	struct iov_iter iter;
218	int ret = -ENOBUFS;
219
220	if (len == 0)
221		goto abandon;
222
223	_enter("%llx,%zx", start, len);
224
225	wreq = kzalloc(sizeof(struct fscache_write_request), GFP_NOFS);
226	if (!wreq)
227		goto abandon;
228	wreq->mapping		= mapping;
229	wreq->start		= start;
230	wreq->len		= len;
231	wreq->set_bits		= cond;
232	wreq->term_func		= term_func;
233	wreq->term_func_priv	= term_func_priv;
234
235	cres = &wreq->cache_resources;
236	if (fscache_begin_operation(cres, cookie, FSCACHE_WANT_WRITE,
237				    fscache_access_io_write) < 0)
238		goto abandon_free;
239
240	ret = cres->ops->prepare_write(cres, &start, &len, len, i_size, false);
241	if (ret < 0)
242		goto abandon_end;
243
244	/* TODO: Consider clearing page bits now for space the write isn't
245	 * covering.  This is more complicated than it appears when THPs are
246	 * taken into account.
247	 */
248
249	iov_iter_xarray(&iter, ITER_SOURCE, &mapping->i_pages, start, len);
250	fscache_write(cres, start, &iter, fscache_wreq_done, wreq);
251	return;
252
253abandon_end:
254	return fscache_wreq_done(wreq, ret, false);
255abandon_free:
256	kfree(wreq);
257abandon:
258	fscache_clear_page_bits(mapping, start, len, cond);
259	if (term_func)
260		term_func(term_func_priv, ret, false);
261}
262EXPORT_SYMBOL(__fscache_write_to_cache);
263
264/*
265 * Change the size of a backing object.
266 */
267void __fscache_resize_cookie(struct fscache_cookie *cookie, loff_t new_size)
268{
269	struct netfs_cache_resources cres;
270
271	trace_fscache_resize(cookie, new_size);
272	if (fscache_begin_operation(&cres, cookie, FSCACHE_WANT_WRITE,
273				    fscache_access_io_resize) == 0) {
274		fscache_stat(&fscache_n_resizes);
275		set_bit(FSCACHE_COOKIE_NEEDS_UPDATE, &cookie->flags);
276
277		/* We cannot defer a resize as we need to do it inside the
278		 * netfs's inode lock so that we're serialised with respect to
279		 * writes.
280		 */
281		cookie->volume->cache->ops->resize_cookie(&cres, new_size);
282		fscache_end_operation(&cres);
283	} else {
284		fscache_stat(&fscache_n_resizes_null);
285	}
286}
287EXPORT_SYMBOL(__fscache_resize_cookie);
v6.9.4
  1// SPDX-License-Identifier: GPL-2.0-or-later
  2/* Cache data I/O routines
  3 *
  4 * Copyright (C) 2021 Red Hat, Inc. All Rights Reserved.
  5 * Written by David Howells (dhowells@redhat.com)
  6 */
  7#define FSCACHE_DEBUG_LEVEL OPERATION
  8#include <linux/fscache-cache.h>
  9#include <linux/uio.h>
 10#include <linux/bvec.h>
 11#include <linux/slab.h>
 12#include <linux/uio.h>
 13#include "internal.h"
 14
 15/**
 16 * fscache_wait_for_operation - Wait for an object become accessible
 17 * @cres: The cache resources for the operation being performed
 18 * @want_state: The minimum state the object must be at
 19 *
 20 * See if the target cache object is at the specified minimum state of
 21 * accessibility yet, and if not, wait for it.
 22 */
 23bool fscache_wait_for_operation(struct netfs_cache_resources *cres,
 24				enum fscache_want_state want_state)
 25{
 26	struct fscache_cookie *cookie = fscache_cres_cookie(cres);
 27	enum fscache_cookie_state state;
 28
 29again:
 30	if (!fscache_cache_is_live(cookie->volume->cache)) {
 31		_leave(" [broken]");
 32		return false;
 33	}
 34
 35	state = fscache_cookie_state(cookie);
 36	_enter("c=%08x{%u},%x", cookie->debug_id, state, want_state);
 37
 38	switch (state) {
 39	case FSCACHE_COOKIE_STATE_CREATING:
 40	case FSCACHE_COOKIE_STATE_INVALIDATING:
 41		if (want_state == FSCACHE_WANT_PARAMS)
 42			goto ready; /* There can be no content */
 43		fallthrough;
 44	case FSCACHE_COOKIE_STATE_LOOKING_UP:
 45	case FSCACHE_COOKIE_STATE_LRU_DISCARDING:
 46		wait_var_event(&cookie->state,
 47			       fscache_cookie_state(cookie) != state);
 48		goto again;
 49
 50	case FSCACHE_COOKIE_STATE_ACTIVE:
 51		goto ready;
 52	case FSCACHE_COOKIE_STATE_DROPPED:
 53	case FSCACHE_COOKIE_STATE_RELINQUISHING:
 54	default:
 55		_leave(" [not live]");
 56		return false;
 57	}
 58
 59ready:
 60	if (!cres->cache_priv2)
 61		return cookie->volume->cache->ops->begin_operation(cres, want_state);
 62	return true;
 63}
 64EXPORT_SYMBOL(fscache_wait_for_operation);
 65
 66/*
 67 * Begin an I/O operation on the cache, waiting till we reach the right state.
 68 *
 69 * Attaches the resources required to the operation resources record.
 70 */
 71static int fscache_begin_operation(struct netfs_cache_resources *cres,
 72				   struct fscache_cookie *cookie,
 73				   enum fscache_want_state want_state,
 74				   enum fscache_access_trace why)
 75{
 76	enum fscache_cookie_state state;
 77	long timeo;
 78	bool once_only = false;
 79
 80	cres->ops		= NULL;
 81	cres->cache_priv	= cookie;
 82	cres->cache_priv2	= NULL;
 83	cres->debug_id		= cookie->debug_id;
 84	cres->inval_counter	= cookie->inval_counter;
 85
 86	if (!fscache_begin_cookie_access(cookie, why)) {
 87		cres->cache_priv = NULL;
 88		return -ENOBUFS;
 89	}
 90
 91again:
 92	spin_lock(&cookie->lock);
 93
 94	state = fscache_cookie_state(cookie);
 95	_enter("c=%08x{%u},%x", cookie->debug_id, state, want_state);
 96
 97	switch (state) {
 98	case FSCACHE_COOKIE_STATE_LOOKING_UP:
 99	case FSCACHE_COOKIE_STATE_LRU_DISCARDING:
100	case FSCACHE_COOKIE_STATE_INVALIDATING:
101		goto wait_for_file_wrangling;
102	case FSCACHE_COOKIE_STATE_CREATING:
103		if (want_state == FSCACHE_WANT_PARAMS)
104			goto ready; /* There can be no content */
105		goto wait_for_file_wrangling;
106	case FSCACHE_COOKIE_STATE_ACTIVE:
107		goto ready;
108	case FSCACHE_COOKIE_STATE_DROPPED:
109	case FSCACHE_COOKIE_STATE_RELINQUISHING:
110		WARN(1, "Can't use cookie in state %u\n", cookie->state);
111		goto not_live;
112	default:
113		goto not_live;
114	}
115
116ready:
117	spin_unlock(&cookie->lock);
118	if (!cookie->volume->cache->ops->begin_operation(cres, want_state))
119		goto failed;
120	return 0;
121
122wait_for_file_wrangling:
123	spin_unlock(&cookie->lock);
124	trace_fscache_access(cookie->debug_id, refcount_read(&cookie->ref),
125			     atomic_read(&cookie->n_accesses),
126			     fscache_access_io_wait);
127	timeo = wait_var_event_timeout(&cookie->state,
128				       fscache_cookie_state(cookie) != state, 20 * HZ);
129	if (timeo <= 1 && !once_only) {
130		pr_warn("%s: cookie state change wait timed out: cookie->state=%u state=%u",
131			__func__, fscache_cookie_state(cookie), state);
132		fscache_print_cookie(cookie, 'O');
133		once_only = true;
134	}
135	goto again;
136
137not_live:
138	spin_unlock(&cookie->lock);
139failed:
140	cres->cache_priv = NULL;
141	cres->ops = NULL;
142	fscache_end_cookie_access(cookie, fscache_access_io_not_live);
143	_leave(" = -ENOBUFS");
144	return -ENOBUFS;
145}
146
147int __fscache_begin_read_operation(struct netfs_cache_resources *cres,
148				   struct fscache_cookie *cookie)
149{
150	return fscache_begin_operation(cres, cookie, FSCACHE_WANT_PARAMS,
151				       fscache_access_io_read);
152}
153EXPORT_SYMBOL(__fscache_begin_read_operation);
154
155int __fscache_begin_write_operation(struct netfs_cache_resources *cres,
156				    struct fscache_cookie *cookie)
157{
158	return fscache_begin_operation(cres, cookie, FSCACHE_WANT_PARAMS,
159				       fscache_access_io_write);
160}
161EXPORT_SYMBOL(__fscache_begin_write_operation);
162
163struct fscache_write_request {
164	struct netfs_cache_resources cache_resources;
165	struct address_space	*mapping;
166	loff_t			start;
167	size_t			len;
168	bool			set_bits;
169	netfs_io_terminated_t	term_func;
170	void			*term_func_priv;
171};
172
173void __fscache_clear_page_bits(struct address_space *mapping,
174			       loff_t start, size_t len)
175{
176	pgoff_t first = start / PAGE_SIZE;
177	pgoff_t last = (start + len - 1) / PAGE_SIZE;
178	struct page *page;
179
180	if (len) {
181		XA_STATE(xas, &mapping->i_pages, first);
182
183		rcu_read_lock();
184		xas_for_each(&xas, page, last) {
185			end_page_fscache(page);
186		}
187		rcu_read_unlock();
188	}
189}
190EXPORT_SYMBOL(__fscache_clear_page_bits);
191
192/*
193 * Deal with the completion of writing the data to the cache.
194 */
195static void fscache_wreq_done(void *priv, ssize_t transferred_or_error,
196			      bool was_async)
197{
198	struct fscache_write_request *wreq = priv;
199
200	fscache_clear_page_bits(wreq->mapping, wreq->start, wreq->len,
201				wreq->set_bits);
202
203	if (wreq->term_func)
204		wreq->term_func(wreq->term_func_priv, transferred_or_error,
205				was_async);
206	fscache_end_operation(&wreq->cache_resources);
207	kfree(wreq);
208}
209
210void __fscache_write_to_cache(struct fscache_cookie *cookie,
211			      struct address_space *mapping,
212			      loff_t start, size_t len, loff_t i_size,
213			      netfs_io_terminated_t term_func,
214			      void *term_func_priv,
215			      bool cond)
216{
217	struct fscache_write_request *wreq;
218	struct netfs_cache_resources *cres;
219	struct iov_iter iter;
220	int ret = -ENOBUFS;
221
222	if (len == 0)
223		goto abandon;
224
225	_enter("%llx,%zx", start, len);
226
227	wreq = kzalloc(sizeof(struct fscache_write_request), GFP_NOFS);
228	if (!wreq)
229		goto abandon;
230	wreq->mapping		= mapping;
231	wreq->start		= start;
232	wreq->len		= len;
233	wreq->set_bits		= cond;
234	wreq->term_func		= term_func;
235	wreq->term_func_priv	= term_func_priv;
236
237	cres = &wreq->cache_resources;
238	if (fscache_begin_operation(cres, cookie, FSCACHE_WANT_WRITE,
239				    fscache_access_io_write) < 0)
240		goto abandon_free;
241
242	ret = cres->ops->prepare_write(cres, &start, &len, len, i_size, false);
243	if (ret < 0)
244		goto abandon_end;
245
246	/* TODO: Consider clearing page bits now for space the write isn't
247	 * covering.  This is more complicated than it appears when THPs are
248	 * taken into account.
249	 */
250
251	iov_iter_xarray(&iter, ITER_SOURCE, &mapping->i_pages, start, len);
252	fscache_write(cres, start, &iter, fscache_wreq_done, wreq);
253	return;
254
255abandon_end:
256	return fscache_wreq_done(wreq, ret, false);
257abandon_free:
258	kfree(wreq);
259abandon:
260	fscache_clear_page_bits(mapping, start, len, cond);
261	if (term_func)
262		term_func(term_func_priv, ret, false);
263}
264EXPORT_SYMBOL(__fscache_write_to_cache);
265
266/*
267 * Change the size of a backing object.
268 */
269void __fscache_resize_cookie(struct fscache_cookie *cookie, loff_t new_size)
270{
271	struct netfs_cache_resources cres;
272
273	trace_fscache_resize(cookie, new_size);
274	if (fscache_begin_operation(&cres, cookie, FSCACHE_WANT_WRITE,
275				    fscache_access_io_resize) == 0) {
276		fscache_stat(&fscache_n_resizes);
277		set_bit(FSCACHE_COOKIE_NEEDS_UPDATE, &cookie->flags);
278
279		/* We cannot defer a resize as we need to do it inside the
280		 * netfs's inode lock so that we're serialised with respect to
281		 * writes.
282		 */
283		cookie->volume->cache->ops->resize_cookie(&cres, new_size);
284		fscache_end_operation(&cres);
285	} else {
286		fscache_stat(&fscache_n_resizes_null);
287	}
288}
289EXPORT_SYMBOL(__fscache_resize_cookie);