Linux Audio

Check our new training course

Loading...
v6.8
  1// SPDX-License-Identifier: GPL-2.0-or-later
  2/* CacheFiles path walking and related routines
  3 *
  4 * Copyright (C) 2021 Red Hat, Inc. All Rights Reserved.
  5 * Written by David Howells (dhowells@redhat.com)
 
 
 
 
 
  6 */
  7
 
 
 
  8#include <linux/fs.h>
 
 
 
 
  9#include <linux/namei.h>
 
 
 10#include "internal.h"
 11
 12/*
 13 * Mark the backing file as being a cache file if it's not already in use.  The
 14 * mark tells the culling request command that it's not allowed to cull the
 15 * file or directory.  The caller must hold the inode lock.
 16 */
 17static bool __cachefiles_mark_inode_in_use(struct cachefiles_object *object,
 18					   struct inode *inode)
 19{
 20	bool can_use = false;
 21
 22	if (!(inode->i_flags & S_KERNEL_FILE)) {
 23		inode->i_flags |= S_KERNEL_FILE;
 24		trace_cachefiles_mark_active(object, inode);
 25		can_use = true;
 26	} else {
 27		trace_cachefiles_mark_failed(object, inode);
 28	}
 29
 30	return can_use;
 31}
 32
 33static bool cachefiles_mark_inode_in_use(struct cachefiles_object *object,
 34					 struct inode *inode)
 35{
 36	bool can_use;
 37
 38	inode_lock(inode);
 39	can_use = __cachefiles_mark_inode_in_use(object, inode);
 40	inode_unlock(inode);
 41	return can_use;
 42}
 43
 44/*
 45 * Unmark a backing inode.  The caller must hold the inode lock.
 46 */
 47static void __cachefiles_unmark_inode_in_use(struct cachefiles_object *object,
 48					     struct inode *inode)
 49{
 50	inode->i_flags &= ~S_KERNEL_FILE;
 51	trace_cachefiles_mark_inactive(object, inode);
 52}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 53
 54static void cachefiles_do_unmark_inode_in_use(struct cachefiles_object *object,
 55					      struct inode *inode)
 56{
 57	inode_lock(inode);
 58	__cachefiles_unmark_inode_in_use(object, inode);
 59	inode_unlock(inode);
 60}
 61
 62/*
 63 * Unmark a backing inode and tell cachefilesd that there's something that can
 64 * be culled.
 65 */
 66void cachefiles_unmark_inode_in_use(struct cachefiles_object *object,
 67				    struct file *file)
 68{
 69	struct cachefiles_cache *cache = object->volume->cache;
 70	struct inode *inode = file_inode(file);
 71
 72	cachefiles_do_unmark_inode_in_use(object, inode);
 73
 74	if (!test_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags)) {
 75		atomic_long_add(inode->i_blocks, &cache->b_released);
 76		if (atomic_inc_return(&cache->f_released))
 77			cachefiles_state_changed(cache);
 78	}
 79}
 80
 81/*
 82 * get a subdirectory
 
 
 
 83 */
 84struct dentry *cachefiles_get_directory(struct cachefiles_cache *cache,
 85					struct dentry *dir,
 86					const char *dirname,
 87					bool *_is_new)
 88{
 89	struct dentry *subdir;
 90	struct path path;
 91	int ret;
 92
 93	_enter(",,%s", dirname);
 
 94
 95	/* search the current directory for the element name */
 96	inode_lock_nested(d_inode(dir), I_MUTEX_PARENT);
 97
 98retry:
 99	ret = cachefiles_inject_read_error();
100	if (ret == 0)
101		subdir = lookup_one_len(dirname, dir, strlen(dirname));
102	else
103		subdir = ERR_PTR(ret);
104	trace_cachefiles_lookup(NULL, dir, subdir);
105	if (IS_ERR(subdir)) {
106		trace_cachefiles_vfs_error(NULL, d_backing_inode(dir),
107					   PTR_ERR(subdir),
108					   cachefiles_trace_lookup_error);
109		if (PTR_ERR(subdir) == -ENOMEM)
110			goto nomem_d_alloc;
111		goto lookup_error;
112	}
113
114	_debug("subdir -> %pd %s",
115	       subdir, d_backing_inode(subdir) ? "positive" : "negative");
 
116
117	/* we need to create the subdir if it doesn't exist yet */
118	if (d_is_negative(subdir)) {
119		ret = cachefiles_has_space(cache, 1, 0,
120					   cachefiles_has_space_for_create);
121		if (ret < 0)
122			goto mkdir_error;
123
124		_debug("attempt mkdir");
125
126		path.mnt = cache->mnt;
127		path.dentry = dir;
128		ret = security_path_mkdir(&path, subdir, 0700);
129		if (ret < 0)
130			goto mkdir_error;
131		ret = cachefiles_inject_write_error();
132		if (ret == 0)
133			ret = vfs_mkdir(&nop_mnt_idmap, d_inode(dir), subdir, 0700);
134		if (ret < 0) {
135			trace_cachefiles_vfs_error(NULL, d_inode(dir), ret,
136						   cachefiles_trace_mkdir_error);
137			goto mkdir_error;
138		}
139		trace_cachefiles_mkdir(dir, subdir);
140
141		if (unlikely(d_unhashed(subdir))) {
142			cachefiles_put_directory(subdir);
143			goto retry;
144		}
145		ASSERT(d_backing_inode(subdir));
146
147		_debug("mkdir -> %pd{ino=%lu}",
148		       subdir, d_backing_inode(subdir)->i_ino);
149		if (_is_new)
150			*_is_new = true;
151	}
152
153	/* Tell rmdir() it's not allowed to delete the subdir */
154	inode_lock(d_inode(subdir));
155	inode_unlock(d_inode(dir));
156
157	if (!__cachefiles_mark_inode_in_use(NULL, d_inode(subdir))) {
158		pr_notice("cachefiles: Inode already in use: %pd (B=%lx)\n",
159			  subdir, d_inode(subdir)->i_ino);
160		goto mark_error;
161	}
 
 
 
 
162
163	inode_unlock(d_inode(subdir));
164
165	/* we need to make sure the subdir is a directory */
166	ASSERT(d_backing_inode(subdir));
167
168	if (!d_can_lookup(subdir)) {
169		pr_err("%s is not a directory\n", dirname);
170		ret = -EIO;
171		goto check_error;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
172	}
173
174	ret = -EPERM;
175	if (!(d_backing_inode(subdir)->i_opflags & IOP_XATTR) ||
176	    !d_backing_inode(subdir)->i_op->lookup ||
177	    !d_backing_inode(subdir)->i_op->mkdir ||
178	    !d_backing_inode(subdir)->i_op->rename ||
179	    !d_backing_inode(subdir)->i_op->rmdir ||
180	    !d_backing_inode(subdir)->i_op->unlink)
181		goto check_error;
182
183	_leave(" = [%lu]", d_backing_inode(subdir)->i_ino);
184	return subdir;
185
186check_error:
187	cachefiles_put_directory(subdir);
188	_leave(" = %d [check]", ret);
189	return ERR_PTR(ret);
190
191mark_error:
192	inode_unlock(d_inode(subdir));
193	dput(subdir);
194	return ERR_PTR(-EBUSY);
195
196mkdir_error:
197	inode_unlock(d_inode(dir));
198	dput(subdir);
199	pr_err("mkdir %s failed with error %d\n", dirname, ret);
200	return ERR_PTR(ret);
201
202lookup_error:
203	inode_unlock(d_inode(dir));
204	ret = PTR_ERR(subdir);
205	pr_err("Lookup %s failed with error %d\n", dirname, ret);
206	return ERR_PTR(ret);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
207
208nomem_d_alloc:
209	inode_unlock(d_inode(dir));
210	_leave(" = -ENOMEM");
211	return ERR_PTR(-ENOMEM);
212}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
213
214/*
215 * Put a subdirectory.
216 */
217void cachefiles_put_directory(struct dentry *dir)
218{
219	if (dir) {
220		cachefiles_do_unmark_inode_in_use(NULL, d_inode(dir));
221		dput(dir);
222	}
223}
224
225/*
226 * Remove a regular file from the cache.
227 */
228static int cachefiles_unlink(struct cachefiles_cache *cache,
229			     struct cachefiles_object *object,
230			     struct dentry *dir, struct dentry *dentry,
231			     enum fscache_why_object_killed why)
232{
233	struct path path = {
234		.mnt	= cache->mnt,
235		.dentry	= dir,
236	};
237	int ret;
238
239	trace_cachefiles_unlink(object, d_inode(dentry)->i_ino, why);
240	ret = security_path_unlink(&path, dentry);
241	if (ret < 0) {
242		cachefiles_io_error(cache, "Unlink security error");
243		return ret;
244	}
245
246	ret = cachefiles_inject_remove_error();
247	if (ret == 0) {
248		ret = vfs_unlink(&nop_mnt_idmap, d_backing_inode(dir), dentry, NULL);
249		if (ret == -EIO)
250			cachefiles_io_error(cache, "Unlink failed");
251	}
252	if (ret != 0)
253		trace_cachefiles_vfs_error(object, d_backing_inode(dir), ret,
254					   cachefiles_trace_unlink_error);
255	return ret;
256}
257
258/*
259 * Delete an object representation from the cache
260 * - File backed objects are unlinked
261 * - Directory backed objects are stuffed into the graveyard for userspace to
262 *   delete
 
263 */
264int cachefiles_bury_object(struct cachefiles_cache *cache,
265			   struct cachefiles_object *object,
266			   struct dentry *dir,
267			   struct dentry *rep,
268			   enum fscache_why_object_killed why)
269{
270	struct dentry *grave, *trap;
271	struct path path, path_to_graveyard;
272	char nbuffer[8 + 8 + 1];
273	int ret;
274
275	_enter(",'%pd','%pd'", dir, rep);
 
 
276
277	if (rep->d_parent != dir) {
278		inode_unlock(d_inode(dir));
279		_leave(" = -ESTALE");
280		return -ESTALE;
281	}
282
283	/* non-directories can just be unlinked */
284	if (!d_is_dir(rep)) {
285		dget(rep); /* Stop the dentry being negated if it's only pinned
286			    * by a file struct.
287			    */
288		ret = cachefiles_unlink(cache, object, dir, rep, why);
289		dput(rep);
 
 
 
 
 
 
 
 
 
 
 
 
 
290
291		inode_unlock(d_inode(dir));
292		_leave(" = %d", ret);
293		return ret;
294	}
295
296	/* directories have to be moved to the graveyard */
297	_debug("move stale object to graveyard");
298	inode_unlock(d_inode(dir));
299
300try_again:
301	/* first step is to make up a grave dentry in the graveyard */
302	sprintf(nbuffer, "%08x%08x",
303		(uint32_t) ktime_get_real_seconds(),
304		(uint32_t) atomic_inc_return(&cache->gravecounter));
305
306	/* do the multiway lock magic */
307	trap = lock_rename(cache->graveyard, dir);
308	if (IS_ERR(trap))
309		return PTR_ERR(trap);
310
311	/* do some checks before getting the grave dentry */
312	if (rep->d_parent != dir || IS_DEADDIR(d_inode(rep))) {
313		/* the entry was probably culled when we dropped the parent dir
314		 * lock */
315		unlock_rename(cache->graveyard, dir);
316		_leave(" = 0 [culled?]");
317		return 0;
318	}
319
320	if (!d_can_lookup(cache->graveyard)) {
321		unlock_rename(cache->graveyard, dir);
322		cachefiles_io_error(cache, "Graveyard no longer a directory");
323		return -EIO;
324	}
325
326	if (trap == rep) {
327		unlock_rename(cache->graveyard, dir);
328		cachefiles_io_error(cache, "May not make directory loop");
329		return -EIO;
330	}
331
332	if (d_mountpoint(rep)) {
333		unlock_rename(cache->graveyard, dir);
334		cachefiles_io_error(cache, "Mountpoint in cache");
335		return -EIO;
336	}
337
338	grave = lookup_one_len(nbuffer, cache->graveyard, strlen(nbuffer));
339	if (IS_ERR(grave)) {
340		unlock_rename(cache->graveyard, dir);
341		trace_cachefiles_vfs_error(object, d_inode(cache->graveyard),
342					   PTR_ERR(grave),
343					   cachefiles_trace_lookup_error);
344
345		if (PTR_ERR(grave) == -ENOMEM) {
346			_leave(" = -ENOMEM");
347			return -ENOMEM;
348		}
349
350		cachefiles_io_error(cache, "Lookup error %ld", PTR_ERR(grave));
 
351		return -EIO;
352	}
353
354	if (d_is_positive(grave)) {
355		unlock_rename(cache->graveyard, dir);
356		dput(grave);
357		grave = NULL;
358		cond_resched();
359		goto try_again;
360	}
361
362	if (d_mountpoint(grave)) {
363		unlock_rename(cache->graveyard, dir);
364		dput(grave);
365		cachefiles_io_error(cache, "Mountpoint in graveyard");
366		return -EIO;
367	}
368
369	/* target should not be an ancestor of source */
370	if (trap == grave) {
371		unlock_rename(cache->graveyard, dir);
372		dput(grave);
373		cachefiles_io_error(cache, "May not make directory loop");
374		return -EIO;
375	}
376
377	/* attempt the rename */
378	path.mnt = cache->mnt;
379	path.dentry = dir;
380	path_to_graveyard.mnt = cache->mnt;
381	path_to_graveyard.dentry = cache->graveyard;
382	ret = security_path_rename(&path, rep, &path_to_graveyard, grave, 0);
383	if (ret < 0) {
384		cachefiles_io_error(cache, "Rename security error %d", ret);
385	} else {
386		struct renamedata rd = {
387			.old_mnt_idmap	= &nop_mnt_idmap,
388			.old_dir	= d_inode(dir),
389			.old_dentry	= rep,
390			.new_mnt_idmap	= &nop_mnt_idmap,
391			.new_dir	= d_inode(cache->graveyard),
392			.new_dentry	= grave,
393		};
394		trace_cachefiles_rename(object, d_inode(rep)->i_ino, why);
395		ret = cachefiles_inject_read_error();
396		if (ret == 0)
397			ret = vfs_rename(&rd);
398		if (ret != 0)
399			trace_cachefiles_vfs_error(object, d_inode(dir), ret,
400						   cachefiles_trace_rename_error);
401		if (ret != 0 && ret != -ENOMEM)
402			cachefiles_io_error(cache,
403					    "Rename failed with error %d", ret);
 
 
 
404	}
405
406	__cachefiles_unmark_inode_in_use(object, d_inode(rep));
407	unlock_rename(cache->graveyard, dir);
408	dput(grave);
409	_leave(" = 0");
410	return 0;
411}
412
413/*
414 * Delete a cache file.
415 */
416int cachefiles_delete_object(struct cachefiles_object *object,
417			     enum fscache_why_object_killed why)
418{
419	struct cachefiles_volume *volume = object->volume;
420	struct dentry *dentry = object->file->f_path.dentry;
421	struct dentry *fan = volume->fanout[(u8)object->cookie->key_hash];
422	int ret;
423
424	_enter(",OBJ%x{%pD}", object->debug_id, object->file);
425
426	/* Stop the dentry being negated if it's only pinned by a file struct. */
427	dget(dentry);
 
428
429	inode_lock_nested(d_backing_inode(fan), I_MUTEX_PARENT);
430	ret = cachefiles_unlink(volume->cache, object, fan, dentry, why);
431	inode_unlock(d_backing_inode(fan));
432	dput(dentry);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
433	return ret;
434}
435
436/*
437 * Create a temporary file and leave it unattached and un-xattr'd until the
438 * time comes to discard the object from memory.
439 */
440struct file *cachefiles_create_tmpfile(struct cachefiles_object *object)
 
 
 
441{
442	struct cachefiles_volume *volume = object->volume;
443	struct cachefiles_cache *cache = volume->cache;
444	const struct cred *saved_cred;
445	struct dentry *fan = volume->fanout[(u8)object->cookie->key_hash];
446	struct file *file;
447	const struct path parentpath = { .mnt = cache->mnt, .dentry = fan };
448	uint64_t ni_size;
449	long ret;
450
451
452	cachefiles_begin_secure(cache, &saved_cred);
453
454	ret = cachefiles_inject_write_error();
455	if (ret == 0) {
456		file = kernel_tmpfile_open(&nop_mnt_idmap, &parentpath,
457					   S_IFREG | 0600,
458					   O_RDWR | O_LARGEFILE | O_DIRECT,
459					   cache->cache_cred);
460		ret = PTR_ERR_OR_ZERO(file);
461	}
462	if (ret) {
463		trace_cachefiles_vfs_error(object, d_inode(fan), ret,
464					   cachefiles_trace_tmpfile_error);
465		if (ret == -EIO)
466			cachefiles_io_error_obj(object, "Failed to create tmpfile");
467		goto err;
468	}
469
470	trace_cachefiles_tmpfile(object, file_inode(file));
 
 
471
472	/* This is a newly created file with no other possible user */
473	if (!cachefiles_mark_inode_in_use(object, file_inode(file)))
474		WARN_ON(1);
475
476	ret = cachefiles_ondemand_init_object(object);
477	if (ret < 0)
478		goto err_unuse;
 
 
479
480	ni_size = object->cookie->object_size;
481	ni_size = round_up(ni_size, CACHEFILES_DIO_BLOCK_SIZE);
482
483	if (ni_size > 0) {
484		trace_cachefiles_trunc(object, file_inode(file), 0, ni_size,
485				       cachefiles_trunc_expand_tmpfile);
486		ret = cachefiles_inject_write_error();
487		if (ret == 0)
488			ret = vfs_truncate(&file->f_path, ni_size);
489		if (ret < 0) {
490			trace_cachefiles_vfs_error(
491				object, file_inode(file), ret,
492				cachefiles_trace_trunc_error);
493			goto err_unuse;
494		}
495	}
496
497	ret = -EINVAL;
498	if (unlikely(!file->f_op->read_iter) ||
499	    unlikely(!file->f_op->write_iter)) {
500		fput(file);
501		pr_notice("Cache does not support read_iter and write_iter\n");
502		goto err_unuse;
503	}
504out:
505	cachefiles_end_secure(cache, saved_cred);
506	return file;
507
508err_unuse:
509	cachefiles_do_unmark_inode_in_use(object, file_inode(file));
510	fput(file);
511err:
512	file = ERR_PTR(ret);
513	goto out;
514}
515
516/*
517 * Create a new file.
518 */
519static bool cachefiles_create_file(struct cachefiles_object *object)
520{
521	struct file *file;
522	int ret;
523
524	ret = cachefiles_has_space(object->volume->cache, 1, 0,
525				   cachefiles_has_space_for_create);
526	if (ret < 0)
527		return false;
528
529	file = cachefiles_create_tmpfile(object);
530	if (IS_ERR(file))
531		return false;
532
533	set_bit(FSCACHE_COOKIE_NEEDS_UPDATE, &object->cookie->flags);
534	set_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags);
535	_debug("create -> %pD{ino=%lu}", file, file_inode(file)->i_ino);
536	object->file = file;
537	return true;
538}
539
540/*
541 * Open an existing file, checking its attributes and replacing it if it is
542 * stale.
543 */
544static bool cachefiles_open_file(struct cachefiles_object *object,
545				 struct dentry *dentry)
546{
547	struct cachefiles_cache *cache = object->volume->cache;
548	struct file *file;
549	struct path path;
550	int ret;
551
552	_enter("%pd", dentry);
 
553
554	if (!cachefiles_mark_inode_in_use(object, d_inode(dentry))) {
555		pr_notice("cachefiles: Inode already in use: %pd (B=%lx)\n",
556			  dentry, d_inode(dentry)->i_ino);
557		return false;
558	}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
559
560	/* We need to open a file interface onto a data file now as we can't do
561	 * it on demand because writeback called from do_exit() sees
562	 * current->fs == NULL - which breaks d_path() called from ext4 open.
563	 */
564	path.mnt = cache->mnt;
565	path.dentry = dentry;
566	file = kernel_file_open(&path, O_RDWR | O_LARGEFILE | O_DIRECT,
567				d_backing_inode(dentry), cache->cache_cred);
568	if (IS_ERR(file)) {
569		trace_cachefiles_vfs_error(object, d_backing_inode(dentry),
570					   PTR_ERR(file),
571					   cachefiles_trace_open_error);
572		goto error;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
573	}
574
575	if (unlikely(!file->f_op->read_iter) ||
576	    unlikely(!file->f_op->write_iter)) {
577		pr_notice("Cache does not support read_iter and write_iter\n");
578		goto error_fput;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
579	}
580	_debug("file -> %pd positive", dentry);
581
582	ret = cachefiles_ondemand_init_object(object);
583	if (ret < 0)
584		goto error_fput;
585
586	ret = cachefiles_check_auxdata(object, file);
587	if (ret < 0)
588		goto check_failed;
589
590	clear_bit(FSCACHE_COOKIE_NO_DATA_TO_READ, &object->cookie->flags);
 
591
592	object->file = file;
593
594	/* Always update the atime on an object we've just looked up (this is
595	 * used to keep track of culling, and atimes are only updated by read,
596	 * write and readdir but not lookup or open).
597	 */
598	touch_atime(&file->f_path);
599	dput(dentry);
600	return true;
601
602check_failed:
603	fscache_cookie_lookup_negative(object->cookie);
604	cachefiles_unmark_inode_in_use(object, file);
605	fput(file);
606	dput(dentry);
607	if (ret == -ESTALE)
608		return cachefiles_create_file(object);
609	return false;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
610
611error_fput:
612	fput(file);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
613error:
614	cachefiles_do_unmark_inode_in_use(object, d_inode(dentry));
615	dput(dentry);
616	return false;
 
 
 
 
617}
618
619/*
620 * walk from the parent object to the child object through the backing
621 * filesystem, creating directories as we go
622 */
623bool cachefiles_look_up_object(struct cachefiles_object *object)
 
 
624{
625	struct cachefiles_volume *volume = object->volume;
626	struct dentry *dentry, *fan = volume->fanout[(u8)object->cookie->key_hash];
 
627	int ret;
628
629	_enter("OBJ%x,%s,", object->debug_id, object->d_name);
630
631	/* Look up path "cache/vol/fanout/file". */
632	ret = cachefiles_inject_read_error();
633	if (ret == 0)
634		dentry = lookup_positive_unlocked(object->d_name, fan,
635						  object->d_name_len);
636	else
637		dentry = ERR_PTR(ret);
638	trace_cachefiles_lookup(object, fan, dentry);
639	if (IS_ERR(dentry)) {
640		if (dentry == ERR_PTR(-ENOENT))
641			goto new_file;
642		if (dentry == ERR_PTR(-EIO))
643			cachefiles_io_error_obj(object, "Lookup failed");
644		return false;
645	}
646
647	if (!d_is_reg(dentry)) {
648		pr_err("%pd is not a file\n", dentry);
649		inode_lock_nested(d_inode(fan), I_MUTEX_PARENT);
650		ret = cachefiles_bury_object(volume->cache, object, fan, dentry,
651					     FSCACHE_OBJECT_IS_WEIRD);
652		dput(dentry);
653		if (ret < 0)
654			return false;
655		goto new_file;
656	}
657
658	if (!cachefiles_open_file(object, dentry))
659		return false;
660
661	_leave(" = t [%lu]", file_inode(object->file)->i_ino);
662	return true;
 
 
 
663
664new_file:
665	fscache_cookie_lookup_negative(object->cookie);
666	return cachefiles_create_file(object);
667}
668
669/*
670 * Attempt to link a temporary file into its rightful place in the cache.
671 */
672bool cachefiles_commit_tmpfile(struct cachefiles_cache *cache,
673			       struct cachefiles_object *object)
674{
675	struct cachefiles_volume *volume = object->volume;
676	struct dentry *dentry, *fan = volume->fanout[(u8)object->cookie->key_hash];
677	bool success = false;
678	int ret;
679
680	_enter(",%pD", object->file);
681
682	inode_lock_nested(d_inode(fan), I_MUTEX_PARENT);
683	ret = cachefiles_inject_read_error();
684	if (ret == 0)
685		dentry = lookup_one_len(object->d_name, fan, object->d_name_len);
686	else
687		dentry = ERR_PTR(ret);
688	if (IS_ERR(dentry)) {
689		trace_cachefiles_vfs_error(object, d_inode(fan), PTR_ERR(dentry),
690					   cachefiles_trace_lookup_error);
691		_debug("lookup fail %ld", PTR_ERR(dentry));
692		goto out_unlock;
693	}
694
695	if (!d_is_negative(dentry)) {
696		if (d_backing_inode(dentry) == file_inode(object->file)) {
697			success = true;
698			goto out_dput;
699		}
700
701		ret = cachefiles_unlink(volume->cache, object, fan, dentry,
702					FSCACHE_OBJECT_IS_STALE);
703		if (ret < 0)
704			goto out_dput;
705
706		dput(dentry);
707		ret = cachefiles_inject_read_error();
708		if (ret == 0)
709			dentry = lookup_one_len(object->d_name, fan, object->d_name_len);
710		else
711			dentry = ERR_PTR(ret);
712		if (IS_ERR(dentry)) {
713			trace_cachefiles_vfs_error(object, d_inode(fan), PTR_ERR(dentry),
714						   cachefiles_trace_lookup_error);
715			_debug("lookup fail %ld", PTR_ERR(dentry));
716			goto out_unlock;
717		}
718	}
719
720	ret = cachefiles_inject_read_error();
721	if (ret == 0)
722		ret = vfs_link(object->file->f_path.dentry, &nop_mnt_idmap,
723			       d_inode(fan), dentry, NULL);
724	if (ret < 0) {
725		trace_cachefiles_vfs_error(object, d_inode(fan), ret,
726					   cachefiles_trace_link_error);
727		_debug("link fail %d", ret);
728	} else {
729		trace_cachefiles_link(object, file_inode(object->file));
730		spin_lock(&object->lock);
731		/* TODO: Do we want to switch the file pointer to the new dentry? */
732		clear_bit(CACHEFILES_OBJECT_USING_TMPFILE, &object->flags);
733		spin_unlock(&object->lock);
734		success = true;
735	}
736
737out_dput:
738	dput(dentry);
739out_unlock:
740	inode_unlock(d_inode(fan));
741	_leave(" = %u", success);
742	return success;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
743}
744
745/*
746 * Look up an inode to be checked or culled.  Return -EBUSY if the inode is
747 * marked in use.
748 */
749static struct dentry *cachefiles_lookup_for_cull(struct cachefiles_cache *cache,
750						 struct dentry *dir,
751						 char *filename)
 
 
752{
 
 
753	struct dentry *victim;
754	int ret = -ENOENT;
 
 
 
 
755
756	inode_lock_nested(d_inode(dir), I_MUTEX_PARENT);
 
757
 
758	victim = lookup_one_len(filename, dir, strlen(filename));
 
759	if (IS_ERR(victim))
760		goto lookup_error;
761	if (d_is_negative(victim))
762		goto lookup_put;
763	if (d_inode(victim)->i_flags & S_KERNEL_FILE)
764		goto lookup_busy;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
765	return victim;
766
767lookup_busy:
768	ret = -EBUSY;
769lookup_put:
770	inode_unlock(d_inode(dir));
771	dput(victim);
772	return ERR_PTR(ret);
 
773
774lookup_error:
775	inode_unlock(d_inode(dir));
776	ret = PTR_ERR(victim);
777	if (ret == -ENOENT)
778		return ERR_PTR(-ESTALE); /* Probably got retired by the netfs */
 
 
 
779
780	if (ret == -EIO) {
781		cachefiles_io_error(cache, "Lookup failed");
782	} else if (ret != -ENOMEM) {
783		pr_err("Internal error: %d\n", ret);
784		ret = -EIO;
785	}
786
 
787	return ERR_PTR(ret);
788}
789
790/*
791 * Cull an object if it's not in use
792 * - called only by cache manager daemon
793 */
794int cachefiles_cull(struct cachefiles_cache *cache, struct dentry *dir,
795		    char *filename)
796{
797	struct dentry *victim;
798	struct inode *inode;
799	int ret;
800
801	_enter(",%pd/,%s", dir, filename);
 
802
803	victim = cachefiles_lookup_for_cull(cache, dir, filename);
804	if (IS_ERR(victim))
805		return PTR_ERR(victim);
806
807	/* check to see if someone is using this object */
808	inode = d_inode(victim);
809	inode_lock(inode);
810	if (inode->i_flags & S_KERNEL_FILE) {
811		ret = -EBUSY;
812	} else {
813		/* Stop the cache from picking it back up */
814		inode->i_flags |= S_KERNEL_FILE;
815		ret = 0;
816	}
817	inode_unlock(inode);
818	if (ret < 0)
819		goto error_unlock;
820
821	ret = cachefiles_bury_object(cache, NULL, dir, victim,
822				     FSCACHE_OBJECT_WAS_CULLED);
 
 
823	if (ret < 0)
824		goto error;
825
826	fscache_count_culled();
827	dput(victim);
828	_leave(" = 0");
829	return 0;
830
831error_unlock:
832	inode_unlock(d_inode(dir));
833error:
834	dput(victim);
835	if (ret == -ENOENT)
836		return -ESTALE; /* Probably got retired by the netfs */
 
 
 
837
838	if (ret != -ENOMEM) {
839		pr_err("Internal error: %d\n", ret);
840		ret = -EIO;
841	}
842
843	_leave(" = %d", ret);
844	return ret;
845}
846
847/*
848 * Find out if an object is in use or not
849 * - called only by cache manager daemon
850 * - returns -EBUSY or 0 to indicate whether an object is in use or not
851 */
852int cachefiles_check_in_use(struct cachefiles_cache *cache, struct dentry *dir,
853			    char *filename)
854{
855	struct dentry *victim;
856	int ret = 0;
857
858	victim = cachefiles_lookup_for_cull(cache, dir, filename);
 
 
 
859	if (IS_ERR(victim))
860		return PTR_ERR(victim);
861
862	inode_unlock(d_inode(dir));
863	dput(victim);
864	return ret;
 
865}
v3.1
 
  1/* CacheFiles path walking and related routines
  2 *
  3 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
  4 * Written by David Howells (dhowells@redhat.com)
  5 *
  6 * This program is free software; you can redistribute it and/or
  7 * modify it under the terms of the GNU General Public Licence
  8 * as published by the Free Software Foundation; either version
  9 * 2 of the Licence, or (at your option) any later version.
 10 */
 11
 12#include <linux/module.h>
 13#include <linux/sched.h>
 14#include <linux/file.h>
 15#include <linux/fs.h>
 16#include <linux/fsnotify.h>
 17#include <linux/quotaops.h>
 18#include <linux/xattr.h>
 19#include <linux/mount.h>
 20#include <linux/namei.h>
 21#include <linux/security.h>
 22#include <linux/slab.h>
 23#include "internal.h"
 24
 25#define CACHEFILES_KEYBUF_SIZE 512
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 26
 27/*
 28 * dump debugging info about an object
 29 */
 30static noinline
 31void __cachefiles_printk_object(struct cachefiles_object *object,
 32				const char *prefix,
 33				u8 *keybuf)
 34{
 35	struct fscache_cookie *cookie;
 36	unsigned keylen, loop;
 37
 38	printk(KERN_ERR "%sobject: OBJ%x\n",
 39	       prefix, object->fscache.debug_id);
 40	printk(KERN_ERR "%sobjstate=%s fl=%lx wbusy=%x ev=%lx[%lx]\n",
 41	       prefix, fscache_object_states[object->fscache.state],
 42	       object->fscache.flags, work_busy(&object->fscache.work),
 43	       object->fscache.events,
 44	       object->fscache.event_mask & FSCACHE_OBJECT_EVENTS_MASK);
 45	printk(KERN_ERR "%sops=%u inp=%u exc=%u\n",
 46	       prefix, object->fscache.n_ops, object->fscache.n_in_progress,
 47	       object->fscache.n_exclusive);
 48	printk(KERN_ERR "%sparent=%p\n",
 49	       prefix, object->fscache.parent);
 50
 51	spin_lock(&object->fscache.lock);
 52	cookie = object->fscache.cookie;
 53	if (cookie) {
 54		printk(KERN_ERR "%scookie=%p [pr=%p nd=%p fl=%lx]\n",
 55		       prefix,
 56		       object->fscache.cookie,
 57		       object->fscache.cookie->parent,
 58		       object->fscache.cookie->netfs_data,
 59		       object->fscache.cookie->flags);
 60		if (keybuf)
 61			keylen = cookie->def->get_key(cookie->netfs_data, keybuf,
 62						      CACHEFILES_KEYBUF_SIZE);
 63		else
 64			keylen = 0;
 65	} else {
 66		printk(KERN_ERR "%scookie=NULL\n", prefix);
 67		keylen = 0;
 68	}
 69	spin_unlock(&object->fscache.lock);
 70
 71	if (keylen) {
 72		printk(KERN_ERR "%skey=[%u] '", prefix, keylen);
 73		for (loop = 0; loop < keylen; loop++)
 74			printk("%02x", keybuf[loop]);
 75		printk("'\n");
 76	}
 77}
 78
 79/*
 80 * dump debugging info about a pair of objects
 
 81 */
 82static noinline void cachefiles_printk_object(struct cachefiles_object *object,
 83					      struct cachefiles_object *xobject)
 84{
 85	u8 *keybuf;
 
 86
 87	keybuf = kmalloc(CACHEFILES_KEYBUF_SIZE, GFP_NOIO);
 88	if (object)
 89		__cachefiles_printk_object(object, "", keybuf);
 90	if (xobject)
 91		__cachefiles_printk_object(xobject, "x", keybuf);
 92	kfree(keybuf);
 
 93}
 94
 95/*
 96 * mark the owner of a dentry, if there is one, to indicate that that dentry
 97 * has been preemptively deleted
 98 * - the caller must hold the i_mutex on the dentry's parent as required to
 99 *   call vfs_unlink(), vfs_rmdir() or vfs_rename()
100 */
101static void cachefiles_mark_object_buried(struct cachefiles_cache *cache,
102					  struct dentry *dentry)
 
 
103{
104	struct cachefiles_object *object;
105	struct rb_node *p;
 
106
107	_enter(",'%*.*s'",
108	       dentry->d_name.len, dentry->d_name.len, dentry->d_name.name);
109
110	write_lock(&cache->active_lock);
 
111
112	p = cache->active_nodes.rb_node;
113	while (p) {
114		object = rb_entry(p, struct cachefiles_object, active_node);
115		if (object->dentry > dentry)
116			p = p->rb_left;
117		else if (object->dentry < dentry)
118			p = p->rb_right;
119		else
120			goto found_dentry;
 
 
 
 
 
121	}
122
123	write_unlock(&cache->active_lock);
124	_leave(" [no owner]");
125	return;
126
127	/* found the dentry for  */
128found_dentry:
129	kdebug("preemptive burial: OBJ%x [%s] %p",
130	       object->fscache.debug_id,
131	       fscache_object_states[object->fscache.state],
132	       dentry);
 
 
133
134	if (object->fscache.state < FSCACHE_OBJECT_DYING) {
135		printk(KERN_ERR "\n");
136		printk(KERN_ERR "CacheFiles: Error:"
137		       " Can't preemptively bury live object\n");
138		cachefiles_printk_object(object, NULL);
139	} else if (test_and_set_bit(CACHEFILES_OBJECT_BURIED, &object->flags)) {
140		printk(KERN_ERR "CacheFiles: Error:"
141		       " Object already preemptively buried\n");
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142	}
143
144	write_unlock(&cache->active_lock);
145	_leave(" [owner marked]");
146}
147
148/*
149 * record the fact that an object is now active
150 */
151static int cachefiles_mark_object_active(struct cachefiles_cache *cache,
152					 struct cachefiles_object *object)
153{
154	struct cachefiles_object *xobject;
155	struct rb_node **_p, *_parent = NULL;
156	struct dentry *dentry;
157
158	_enter(",%p", object);
159
160try_again:
161	write_lock(&cache->active_lock);
162
163	if (test_and_set_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags)) {
164		printk(KERN_ERR "CacheFiles: Error: Object already active\n");
165		cachefiles_printk_object(object, NULL);
166		BUG();
167	}
168
169	dentry = object->dentry;
170	_p = &cache->active_nodes.rb_node;
171	while (*_p) {
172		_parent = *_p;
173		xobject = rb_entry(_parent,
174				   struct cachefiles_object, active_node);
175
176		ASSERT(xobject != object);
177
178		if (xobject->dentry > dentry)
179			_p = &(*_p)->rb_left;
180		else if (xobject->dentry < dentry)
181			_p = &(*_p)->rb_right;
182		else
183			goto wait_for_old_object;
184	}
185
186	rb_link_node(&object->active_node, _parent, _p);
187	rb_insert_color(&object->active_node, &cache->active_nodes);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
188
189	write_unlock(&cache->active_lock);
190	_leave(" = 0");
191	return 0;
 
 
192
193	/* an old object from a previous incarnation is hogging the slot - we
194	 * need to wait for it to be destroyed */
195wait_for_old_object:
196	if (xobject->fscache.state < FSCACHE_OBJECT_DYING) {
197		printk(KERN_ERR "\n");
198		printk(KERN_ERR "CacheFiles: Error:"
199		       " Unexpected object collision\n");
200		cachefiles_printk_object(object, xobject);
201		BUG();
202	}
203	atomic_inc(&xobject->usage);
204	write_unlock(&cache->active_lock);
205
206	if (test_bit(CACHEFILES_OBJECT_ACTIVE, &xobject->flags)) {
207		wait_queue_head_t *wq;
208
209		signed long timeout = 60 * HZ;
210		wait_queue_t wait;
211		bool requeue;
212
213		/* if the object we're waiting for is queued for processing,
214		 * then just put ourselves on the queue behind it */
215		if (work_pending(&xobject->fscache.work)) {
216			_debug("queue OBJ%x behind OBJ%x immediately",
217			       object->fscache.debug_id,
218			       xobject->fscache.debug_id);
219			goto requeue;
220		}
221
222		/* otherwise we sleep until either the object we're waiting for
223		 * is done, or the fscache_object is congested */
224		wq = bit_waitqueue(&xobject->flags, CACHEFILES_OBJECT_ACTIVE);
225		init_wait(&wait);
226		requeue = false;
227		do {
228			prepare_to_wait(wq, &wait, TASK_UNINTERRUPTIBLE);
229			if (!test_bit(CACHEFILES_OBJECT_ACTIVE, &xobject->flags))
230				break;
231
232			requeue = fscache_object_sleep_till_congested(&timeout);
233		} while (timeout > 0 && !requeue);
234		finish_wait(wq, &wait);
235
236		if (requeue &&
237		    test_bit(CACHEFILES_OBJECT_ACTIVE, &xobject->flags)) {
238			_debug("queue OBJ%x behind OBJ%x after wait",
239			       object->fscache.debug_id,
240			       xobject->fscache.debug_id);
241			goto requeue;
242		}
243
244		if (timeout <= 0) {
245			printk(KERN_ERR "\n");
246			printk(KERN_ERR "CacheFiles: Error: Overlong"
247			       " wait for old active object to go away\n");
248			cachefiles_printk_object(object, xobject);
249			goto requeue;
250		}
 
251	}
 
252
253	ASSERT(!test_bit(CACHEFILES_OBJECT_ACTIVE, &xobject->flags));
 
 
 
 
 
 
 
 
 
 
 
 
254
255	cache->cache.ops->put_object(&xobject->fscache);
256	goto try_again;
 
 
 
 
257
258requeue:
259	clear_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags);
260	cache->cache.ops->put_object(&xobject->fscache);
261	_leave(" = -ETIMEDOUT");
262	return -ETIMEDOUT;
 
 
 
 
 
263}
264
265/*
266 * delete an object representation from the cache
267 * - file backed objects are unlinked
268 * - directory backed objects are stuffed into the graveyard for userspace to
269 *   delete
270 * - unlocks the directory mutex
271 */
272static int cachefiles_bury_object(struct cachefiles_cache *cache,
273				  struct dentry *dir,
274				  struct dentry *rep,
275				  bool preemptive)
 
276{
277	struct dentry *grave, *trap;
278	struct path path, path_to_graveyard;
279	char nbuffer[8 + 8 + 1];
280	int ret;
281
282	_enter(",'%*.*s','%*.*s'",
283	       dir->d_name.len, dir->d_name.len, dir->d_name.name,
284	       rep->d_name.len, rep->d_name.len, rep->d_name.name);
285
286	_debug("remove %p from %p", rep, dir);
 
 
 
 
287
288	/* non-directories can just be unlinked */
289	if (!S_ISDIR(rep->d_inode->i_mode)) {
290		_debug("unlink stale object");
291
292		path.mnt = cache->mnt;
293		path.dentry = dir;
294		ret = security_path_unlink(&path, rep);
295		if (ret < 0) {
296			cachefiles_io_error(cache, "Unlink security error");
297		} else {
298			ret = vfs_unlink(dir->d_inode, rep);
299
300			if (preemptive)
301				cachefiles_mark_object_buried(cache, rep);
302		}
303
304		mutex_unlock(&dir->d_inode->i_mutex);
305
306		if (ret == -EIO)
307			cachefiles_io_error(cache, "Unlink failed");
308
 
309		_leave(" = %d", ret);
310		return ret;
311	}
312
313	/* directories have to be moved to the graveyard */
314	_debug("move stale object to graveyard");
315	mutex_unlock(&dir->d_inode->i_mutex);
316
317try_again:
318	/* first step is to make up a grave dentry in the graveyard */
319	sprintf(nbuffer, "%08x%08x",
320		(uint32_t) get_seconds(),
321		(uint32_t) atomic_inc_return(&cache->gravecounter));
322
323	/* do the multiway lock magic */
324	trap = lock_rename(cache->graveyard, dir);
 
 
325
326	/* do some checks before getting the grave dentry */
327	if (rep->d_parent != dir) {
328		/* the entry was probably culled when we dropped the parent dir
329		 * lock */
330		unlock_rename(cache->graveyard, dir);
331		_leave(" = 0 [culled?]");
332		return 0;
333	}
334
335	if (!S_ISDIR(cache->graveyard->d_inode->i_mode)) {
336		unlock_rename(cache->graveyard, dir);
337		cachefiles_io_error(cache, "Graveyard no longer a directory");
338		return -EIO;
339	}
340
341	if (trap == rep) {
342		unlock_rename(cache->graveyard, dir);
343		cachefiles_io_error(cache, "May not make directory loop");
344		return -EIO;
345	}
346
347	if (d_mountpoint(rep)) {
348		unlock_rename(cache->graveyard, dir);
349		cachefiles_io_error(cache, "Mountpoint in cache");
350		return -EIO;
351	}
352
353	grave = lookup_one_len(nbuffer, cache->graveyard, strlen(nbuffer));
354	if (IS_ERR(grave)) {
355		unlock_rename(cache->graveyard, dir);
 
 
 
356
357		if (PTR_ERR(grave) == -ENOMEM) {
358			_leave(" = -ENOMEM");
359			return -ENOMEM;
360		}
361
362		cachefiles_io_error(cache, "Lookup error %ld",
363				    PTR_ERR(grave));
364		return -EIO;
365	}
366
367	if (grave->d_inode) {
368		unlock_rename(cache->graveyard, dir);
369		dput(grave);
370		grave = NULL;
371		cond_resched();
372		goto try_again;
373	}
374
375	if (d_mountpoint(grave)) {
376		unlock_rename(cache->graveyard, dir);
377		dput(grave);
378		cachefiles_io_error(cache, "Mountpoint in graveyard");
379		return -EIO;
380	}
381
382	/* target should not be an ancestor of source */
383	if (trap == grave) {
384		unlock_rename(cache->graveyard, dir);
385		dput(grave);
386		cachefiles_io_error(cache, "May not make directory loop");
387		return -EIO;
388	}
389
390	/* attempt the rename */
391	path.mnt = cache->mnt;
392	path.dentry = dir;
393	path_to_graveyard.mnt = cache->mnt;
394	path_to_graveyard.dentry = cache->graveyard;
395	ret = security_path_rename(&path, rep, &path_to_graveyard, grave);
396	if (ret < 0) {
397		cachefiles_io_error(cache, "Rename security error %d", ret);
398	} else {
399		ret = vfs_rename(dir->d_inode, rep,
400				 cache->graveyard->d_inode, grave);
 
 
 
 
 
 
 
 
 
 
 
 
 
401		if (ret != 0 && ret != -ENOMEM)
402			cachefiles_io_error(cache,
403					    "Rename failed with error %d", ret);
404
405		if (preemptive)
406			cachefiles_mark_object_buried(cache, rep);
407	}
408
 
409	unlock_rename(cache->graveyard, dir);
410	dput(grave);
411	_leave(" = 0");
412	return 0;
413}
414
415/*
416 * delete an object representation from the cache
417 */
418int cachefiles_delete_object(struct cachefiles_cache *cache,
419			     struct cachefiles_object *object)
420{
421	struct dentry *dir;
 
 
422	int ret;
423
424	_enter(",OBJ%x{%p}", object->fscache.debug_id, object->dentry);
425
426	ASSERT(object->dentry);
427	ASSERT(object->dentry->d_inode);
428	ASSERT(object->dentry->d_parent);
429
430	dir = dget_parent(object->dentry);
431
432	mutex_lock_nested(&dir->d_inode->i_mutex, I_MUTEX_PARENT);
433
434	if (test_bit(CACHEFILES_OBJECT_BURIED, &object->flags)) {
435		/* object allocation for the same key preemptively deleted this
436		 * object's file so that it could create its own file */
437		_debug("object preemptively buried");
438		mutex_unlock(&dir->d_inode->i_mutex);
439		ret = 0;
440	} else {
441		/* we need to check that our parent is _still_ our parent - it
442		 * may have been renamed */
443		if (dir == object->dentry->d_parent) {
444			ret = cachefiles_bury_object(cache, dir,
445						     object->dentry, false);
446		} else {
447			/* it got moved, presumably by cachefilesd culling it,
448			 * so it's no longer in the key path and we can ignore
449			 * it */
450			mutex_unlock(&dir->d_inode->i_mutex);
451			ret = 0;
452		}
453	}
454
455	dput(dir);
456	_leave(" = %d", ret);
457	return ret;
458}
459
460/*
461 * walk from the parent object to the child object through the backing
462 * filesystem, creating directories as we go
463 */
464int cachefiles_walk_to_object(struct cachefiles_object *parent,
465			      struct cachefiles_object *object,
466			      const char *key,
467			      struct cachefiles_xattr *auxdata)
468{
469	struct cachefiles_cache *cache;
470	struct dentry *dir, *next = NULL;
471	struct path path;
472	unsigned long start;
473	const char *name;
474	int ret, nlen;
475
476	_enter("OBJ%x{%p},OBJ%x,%s,",
477	       parent->fscache.debug_id, parent->dentry,
478	       object->fscache.debug_id, key);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
479
480	cache = container_of(parent->fscache.cache,
481			     struct cachefiles_cache, cache);
482	path.mnt = cache->mnt;
483
484	ASSERT(parent->dentry);
485	ASSERT(parent->dentry->d_inode);
 
486
487	if (!(S_ISDIR(parent->dentry->d_inode->i_mode))) {
488		// TODO: convert file to dir
489		_leave("looking up in none directory");
490		return -ENOBUFS;
491	}
492
493	dir = dget(parent->dentry);
 
494
495advance:
496	/* attempt to transit the first directory component */
497	name = key;
498	nlen = strlen(key);
 
 
 
 
 
 
 
 
 
499
500	/* key ends in a double NUL */
501	key = key + nlen + 1;
502	if (!*key)
503		key = NULL;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
504
505lookup_again:
506	/* search the current directory for the element name */
507	_debug("lookup '%s'", name);
 
 
 
 
508
509	mutex_lock_nested(&dir->d_inode->i_mutex, I_MUTEX_PARENT);
 
 
 
510
511	start = jiffies;
512	next = lookup_one_len(name, dir, nlen);
513	cachefiles_hist(cachefiles_lookup_histogram, start);
514	if (IS_ERR(next))
515		goto lookup_error;
 
 
 
 
 
516
517	_debug("next -> %p %s", next, next->d_inode ? "positive" : "negative");
 
 
 
 
 
 
 
 
 
 
518
519	if (!key)
520		object->new = !next->d_inode;
521
522	/* if this element of the path doesn't exist, then the lookup phase
523	 * failed, and we can release any readers in the certain knowledge that
524	 * there's nothing for them to actually read */
525	if (!next->d_inode)
526		fscache_object_lookup_negative(&object->fscache);
527
528	/* we need to create the object if it's negative */
529	if (key || object->type == FSCACHE_COOKIE_TYPE_INDEX) {
530		/* index objects and intervening tree levels must be subdirs */
531		if (!next->d_inode) {
532			ret = cachefiles_has_space(cache, 1, 0);
533			if (ret < 0)
534				goto create_error;
535
536			path.dentry = dir;
537			ret = security_path_mkdir(&path, next, 0);
538			if (ret < 0)
539				goto create_error;
540			start = jiffies;
541			ret = vfs_mkdir(dir->d_inode, next, 0);
542			cachefiles_hist(cachefiles_mkdir_histogram, start);
543			if (ret < 0)
544				goto create_error;
545
546			ASSERT(next->d_inode);
547
548			_debug("mkdir -> %p{%p{ino=%lu}}",
549			       next, next->d_inode, next->d_inode->i_ino);
550
551		} else if (!S_ISDIR(next->d_inode->i_mode)) {
552			kerror("inode %lu is not a directory",
553			       next->d_inode->i_ino);
554			ret = -ENOBUFS;
555			goto error;
556		}
557
558	} else {
559		/* non-index objects start out life as files */
560		if (!next->d_inode) {
561			ret = cachefiles_has_space(cache, 1, 0);
562			if (ret < 0)
563				goto create_error;
564
565			path.dentry = dir;
566			ret = security_path_mknod(&path, next, S_IFREG, 0);
567			if (ret < 0)
568				goto create_error;
569			start = jiffies;
570			ret = vfs_create(dir->d_inode, next, S_IFREG, NULL);
571			cachefiles_hist(cachefiles_create_histogram, start);
572			if (ret < 0)
573				goto create_error;
574
575			ASSERT(next->d_inode);
576
577			_debug("create -> %p{%p{ino=%lu}}",
578			       next, next->d_inode, next->d_inode->i_ino);
579
580		} else if (!S_ISDIR(next->d_inode->i_mode) &&
581			   !S_ISREG(next->d_inode->i_mode)
582			   ) {
583			kerror("inode %lu is not a file or directory",
584			       next->d_inode->i_ino);
585			ret = -ENOBUFS;
586			goto error;
587		}
588	}
589
590	/* process the next component */
591	if (key) {
592		_debug("advance");
593		mutex_unlock(&dir->d_inode->i_mutex);
594		dput(dir);
595		dir = next;
596		next = NULL;
597		goto advance;
598	}
599
600	/* we've found the object we were looking for */
601	object->dentry = next;
602
603	/* if we've found that the terminal object exists, then we need to
604	 * check its attributes and delete it if it's out of date */
605	if (!object->new) {
606		_debug("validate '%*.*s'",
607		       next->d_name.len, next->d_name.len, next->d_name.name);
608
609		ret = cachefiles_check_object_xattr(object, auxdata);
610		if (ret == -ESTALE) {
611			/* delete the object (the deleter drops the directory
612			 * mutex) */
613			object->dentry = NULL;
614
615			ret = cachefiles_bury_object(cache, dir, next, true);
616			dput(next);
617			next = NULL;
618
619			if (ret < 0)
620				goto delete_error;
621
622			_debug("redo lookup");
623			goto lookup_again;
624		}
625	}
 
626
627	/* note that we're now using this object */
628	ret = cachefiles_mark_object_active(cache, object);
 
629
630	mutex_unlock(&dir->d_inode->i_mutex);
631	dput(dir);
632	dir = NULL;
633
634	if (ret == -ETIMEDOUT)
635		goto mark_active_timed_out;
636
637	_debug("=== OBTAINED_OBJECT ===");
638
639	if (object->new) {
640		/* attach data to a newly constructed terminal object */
641		ret = cachefiles_set_object_xattr(object, auxdata);
642		if (ret < 0)
643			goto check_error;
644	} else {
645		/* always update the atime on an object we've just looked up
646		 * (this is used to keep track of culling, and atimes are only
647		 * updated by read, write and readdir but not lookup or
648		 * open) */
649		touch_atime(cache->mnt, next);
650	}
651
652	/* open a file interface onto a data file */
653	if (object->type != FSCACHE_COOKIE_TYPE_INDEX) {
654		if (S_ISREG(object->dentry->d_inode->i_mode)) {
655			const struct address_space_operations *aops;
656
657			ret = -EPERM;
658			aops = object->dentry->d_inode->i_mapping->a_ops;
659			if (!aops->bmap)
660				goto check_error;
661
662			object->backer = object->dentry;
663		} else {
664			BUG(); // TODO: open file in data-class subdir
665		}
666	}
667
668	object->new = 0;
669	fscache_obtained_object(&object->fscache);
670
671	_leave(" = 0 [%lu]", object->dentry->d_inode->i_ino);
672	return 0;
673
674create_error:
675	_debug("create error %d", ret);
676	if (ret == -EIO)
677		cachefiles_io_error(cache, "Create/mkdir failed");
678	goto error;
679
680mark_active_timed_out:
681	_debug("mark active timed out");
682	goto release_dentry;
683
684check_error:
685	_debug("check error %d", ret);
686	write_lock(&cache->active_lock);
687	rb_erase(&object->active_node, &cache->active_nodes);
688	clear_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags);
689	wake_up_bit(&object->flags, CACHEFILES_OBJECT_ACTIVE);
690	write_unlock(&cache->active_lock);
691release_dentry:
692	dput(object->dentry);
693	object->dentry = NULL;
694	goto error_out;
695
696delete_error:
697	_debug("delete error %d", ret);
698	goto error_out2;
699
700lookup_error:
701	_debug("lookup error %ld", PTR_ERR(next));
702	ret = PTR_ERR(next);
703	if (ret == -EIO)
704		cachefiles_io_error(cache, "Lookup failed");
705	next = NULL;
706error:
707	mutex_unlock(&dir->d_inode->i_mutex);
708	dput(next);
709error_out2:
710	dput(dir);
711error_out:
712	_leave(" = error %d", -ret);
713	return ret;
714}
715
716/*
717 * get a subdirectory
 
718 */
719struct dentry *cachefiles_get_directory(struct cachefiles_cache *cache,
720					struct dentry *dir,
721					const char *dirname)
722{
723	struct dentry *subdir;
724	unsigned long start;
725	struct path path;
726	int ret;
727
728	_enter(",,%s", dirname);
729
730	/* search the current directory for the element name */
731	mutex_lock(&dir->d_inode->i_mutex);
732
733	start = jiffies;
734	subdir = lookup_one_len(dirname, dir, strlen(dirname));
735	cachefiles_hist(cachefiles_lookup_histogram, start);
736	if (IS_ERR(subdir)) {
737		if (PTR_ERR(subdir) == -ENOMEM)
738			goto nomem_d_alloc;
739		goto lookup_error;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
740	}
741
742	_debug("subdir -> %p %s",
743	       subdir, subdir->d_inode ? "positive" : "negative");
744
745	/* we need to create the subdir if it doesn't exist yet */
746	if (!subdir->d_inode) {
747		ret = cachefiles_has_space(cache, 1, 0);
748		if (ret < 0)
749			goto mkdir_error;
750
751		_debug("attempt mkdir");
 
 
 
752
753		path.mnt = cache->mnt;
754		path.dentry = dir;
755		ret = security_path_mkdir(&path, subdir, 0700);
756		if (ret < 0)
757			goto mkdir_error;
758		ret = vfs_mkdir(dir->d_inode, subdir, 0700);
759		if (ret < 0)
760			goto mkdir_error;
 
 
761
762		ASSERT(subdir->d_inode);
763
764		_debug("mkdir -> %p{%p{ino=%lu}}",
765		       subdir,
766		       subdir->d_inode,
767		       subdir->d_inode->i_ino);
768	}
 
 
 
 
 
 
 
 
 
 
 
 
 
769
770	mutex_unlock(&dir->d_inode->i_mutex);
 
 
 
771
772	/* we need to make sure the subdir is a directory */
773	ASSERT(subdir->d_inode);
 
 
 
 
 
 
 
 
 
 
 
774
775	if (!S_ISDIR(subdir->d_inode->i_mode)) {
776		kerror("%s is not a directory", dirname);
777		ret = -EIO;
778		goto check_error;
 
 
 
 
 
 
 
 
 
 
 
779	}
780
781	ret = -EPERM;
782	if (!subdir->d_inode->i_op ||
783	    !subdir->d_inode->i_op->setxattr ||
784	    !subdir->d_inode->i_op->getxattr ||
785	    !subdir->d_inode->i_op->lookup ||
786	    !subdir->d_inode->i_op->mkdir ||
787	    !subdir->d_inode->i_op->create ||
788	    !subdir->d_inode->i_op->rename ||
789	    !subdir->d_inode->i_op->rmdir ||
790	    !subdir->d_inode->i_op->unlink)
791		goto check_error;
792
793	_leave(" = [%lu]", subdir->d_inode->i_ino);
794	return subdir;
795
796check_error:
797	dput(subdir);
798	_leave(" = %d [check]", ret);
799	return ERR_PTR(ret);
800
801mkdir_error:
802	mutex_unlock(&dir->d_inode->i_mutex);
803	dput(subdir);
804	kerror("mkdir %s failed with error %d", dirname, ret);
805	return ERR_PTR(ret);
806
807lookup_error:
808	mutex_unlock(&dir->d_inode->i_mutex);
809	ret = PTR_ERR(subdir);
810	kerror("Lookup %s failed with error %d", dirname, ret);
811	return ERR_PTR(ret);
812
813nomem_d_alloc:
814	mutex_unlock(&dir->d_inode->i_mutex);
815	_leave(" = -ENOMEM");
816	return ERR_PTR(-ENOMEM);
817}
818
819/*
820 * find out if an object is in use or not
821 * - if finds object and it's not in use:
822 *   - returns a pointer to the object and a reference on it
823 *   - returns with the directory locked
824 */
825static struct dentry *cachefiles_check_active(struct cachefiles_cache *cache,
826					      struct dentry *dir,
827					      char *filename)
828{
829	struct cachefiles_object *object;
830	struct rb_node *_n;
831	struct dentry *victim;
832	unsigned long start;
833	int ret;
834
835	//_enter(",%*.*s/,%s",
836	//       dir->d_name.len, dir->d_name.len, dir->d_name.name, filename);
837
838	/* look up the victim */
839	mutex_lock_nested(&dir->d_inode->i_mutex, 1);
840
841	start = jiffies;
842	victim = lookup_one_len(filename, dir, strlen(filename));
843	cachefiles_hist(cachefiles_lookup_histogram, start);
844	if (IS_ERR(victim))
845		goto lookup_error;
846
847	//_debug("victim -> %p %s",
848	//       victim, victim->d_inode ? "positive" : "negative");
849
850	/* if the object is no longer there then we probably retired the object
851	 * at the netfs's request whilst the cull was in progress
852	 */
853	if (!victim->d_inode) {
854		mutex_unlock(&dir->d_inode->i_mutex);
855		dput(victim);
856		_leave(" = -ENOENT [absent]");
857		return ERR_PTR(-ENOENT);
858	}
859
860	/* check to see if we're using this object */
861	read_lock(&cache->active_lock);
862
863	_n = cache->active_nodes.rb_node;
864
865	while (_n) {
866		object = rb_entry(_n, struct cachefiles_object, active_node);
867
868		if (object->dentry > victim)
869			_n = _n->rb_left;
870		else if (object->dentry < victim)
871			_n = _n->rb_right;
872		else
873			goto object_in_use;
874	}
875
876	read_unlock(&cache->active_lock);
877
878	//_leave(" = %p", victim);
879	return victim;
880
881object_in_use:
882	read_unlock(&cache->active_lock);
883	mutex_unlock(&dir->d_inode->i_mutex);
 
884	dput(victim);
885	//_leave(" = -EBUSY [in use]");
886	return ERR_PTR(-EBUSY);
887
888lookup_error:
889	mutex_unlock(&dir->d_inode->i_mutex);
890	ret = PTR_ERR(victim);
891	if (ret == -ENOENT) {
892		/* file or dir now absent - probably retired by netfs */
893		_leave(" = -ESTALE [absent]");
894		return ERR_PTR(-ESTALE);
895	}
896
897	if (ret == -EIO) {
898		cachefiles_io_error(cache, "Lookup failed");
899	} else if (ret != -ENOMEM) {
900		kerror("Internal error: %d", ret);
901		ret = -EIO;
902	}
903
904	_leave(" = %d", ret);
905	return ERR_PTR(ret);
906}
907
908/*
909 * cull an object if it's not in use
910 * - called only by cache manager daemon
911 */
912int cachefiles_cull(struct cachefiles_cache *cache, struct dentry *dir,
913		    char *filename)
914{
915	struct dentry *victim;
 
916	int ret;
917
918	_enter(",%*.*s/,%s",
919	       dir->d_name.len, dir->d_name.len, dir->d_name.name, filename);
920
921	victim = cachefiles_check_active(cache, dir, filename);
922	if (IS_ERR(victim))
923		return PTR_ERR(victim);
924
925	_debug("victim -> %p %s",
926	       victim, victim->d_inode ? "positive" : "negative");
927
928	/* okay... the victim is not being used so we can cull it
929	 * - start by marking it as stale
930	 */
931	_debug("victim is cullable");
932
933	ret = cachefiles_remove_object_xattr(cache, victim);
 
 
934	if (ret < 0)
935		goto error_unlock;
936
937	/*  actually remove the victim (drops the dir mutex) */
938	_debug("bury");
939
940	ret = cachefiles_bury_object(cache, dir, victim, false);
941	if (ret < 0)
942		goto error;
943
 
944	dput(victim);
945	_leave(" = 0");
946	return 0;
947
948error_unlock:
949	mutex_unlock(&dir->d_inode->i_mutex);
950error:
951	dput(victim);
952	if (ret == -ENOENT) {
953		/* file or dir now absent - probably retired by netfs */
954		_leave(" = -ESTALE [absent]");
955		return -ESTALE;
956	}
957
958	if (ret != -ENOMEM) {
959		kerror("Internal error: %d", ret);
960		ret = -EIO;
961	}
962
963	_leave(" = %d", ret);
964	return ret;
965}
966
967/*
968 * find out if an object is in use or not
969 * - called only by cache manager daemon
970 * - returns -EBUSY or 0 to indicate whether an object is in use or not
971 */
972int cachefiles_check_in_use(struct cachefiles_cache *cache, struct dentry *dir,
973			    char *filename)
974{
975	struct dentry *victim;
 
976
977	//_enter(",%*.*s/,%s",
978	//       dir->d_name.len, dir->d_name.len, dir->d_name.name, filename);
979
980	victim = cachefiles_check_active(cache, dir, filename);
981	if (IS_ERR(victim))
982		return PTR_ERR(victim);
983
984	mutex_unlock(&dir->d_inode->i_mutex);
985	dput(victim);
986	//_leave(" = 0");
987	return 0;
988}