Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 | /* SPDX-License-Identifier: GPL-2.0-or-later */ /* NFS filesystem cache interface definitions * * Copyright (C) 2008 Red Hat, Inc. All Rights Reserved. * Written by David Howells (dhowells@redhat.com) */ #ifndef _NFS_FSCACHE_H #define _NFS_FSCACHE_H #include <linux/swap.h> #include <linux/nfs_fs.h> #include <linux/nfs_mount.h> #include <linux/nfs4_mount.h> #include <linux/fscache.h> #include <linux/iversion.h> #ifdef CONFIG_NFS_FSCACHE /* * Definition of the auxiliary data attached to NFS inode storage objects * within the cache. * * The contents of this struct are recorded in the on-disk local cache in the * auxiliary data attached to the data storage object backing an inode. This * permits coherency to be managed when a new inode binds to an already extant * cache object. */ struct nfs_fscache_inode_auxdata { s64 mtime_sec; s64 mtime_nsec; s64 ctime_sec; s64 ctime_nsec; u64 change_attr; }; struct nfs_netfs_io_data { /* * NFS may split a netfs_io_subrequest into multiple RPCs, each * with their own read completion. In netfs, we can only call * netfs_subreq_terminated() once for each subrequest. Use the * refcount here to double as a marker of the last RPC completion, * and only call netfs via netfs_subreq_terminated() once. */ refcount_t refcount; struct netfs_io_subrequest *sreq; /* * Final disposition of the netfs_io_subrequest, sent in * netfs_subreq_terminated() */ atomic64_t transferred; int error; }; static inline void nfs_netfs_get(struct nfs_netfs_io_data *netfs) { refcount_inc(&netfs->refcount); } static inline void nfs_netfs_put(struct nfs_netfs_io_data *netfs) { ssize_t final_len; /* Only the last RPC completion should call netfs_subreq_terminated() */ if (!refcount_dec_and_test(&netfs->refcount)) return; /* * The NFS pageio interface may read a complete page, even when netfs * only asked for a partial page. Specifically, this may be seen when * one thread is truncating a file while another one is reading the last * page of the file. * Correct the final length here to be no larger than the netfs subrequest * length, and thus avoid netfs's "Subreq overread" warning message. */ final_len = min_t(s64, netfs->sreq->len, atomic64_read(&netfs->transferred)); netfs_subreq_terminated(netfs->sreq, netfs->error ?: final_len, false); kfree(netfs); } static inline void nfs_netfs_inode_init(struct nfs_inode *nfsi) { netfs_inode_init(&nfsi->netfs, &nfs_netfs_ops, false); } extern void nfs_netfs_initiate_read(struct nfs_pgio_header *hdr); extern void nfs_netfs_read_completion(struct nfs_pgio_header *hdr); extern int nfs_netfs_folio_unlock(struct folio *folio); /* * fscache.c */ extern int nfs_fscache_get_super_cookie(struct super_block *, const char *, int); extern void nfs_fscache_release_super_cookie(struct super_block *); extern void nfs_fscache_init_inode(struct inode *); extern void nfs_fscache_clear_inode(struct inode *); extern void nfs_fscache_open_file(struct inode *, struct file *); extern void nfs_fscache_release_file(struct inode *, struct file *); extern int nfs_netfs_readahead(struct readahead_control *ractl); extern int nfs_netfs_read_folio(struct file *file, struct folio *folio); static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) { if (folio_test_fscache(folio)) { if (current_is_kswapd() || !(gfp & __GFP_FS)) return false; folio_wait_fscache(folio); } fscache_note_page_release(netfs_i_cookie(netfs_inode(folio->mapping->host))); return true; } static inline void nfs_fscache_update_auxdata(struct nfs_fscache_inode_auxdata *auxdata, struct inode *inode) { memset(auxdata, 0, sizeof(*auxdata)); auxdata->mtime_sec = inode_get_mtime(inode).tv_sec; auxdata->mtime_nsec = inode_get_mtime(inode).tv_nsec; auxdata->ctime_sec = inode_get_ctime(inode).tv_sec; auxdata->ctime_nsec = inode_get_ctime(inode).tv_nsec; if (NFS_SERVER(inode)->nfs_client->rpc_ops->version == 4) auxdata->change_attr = inode_peek_iversion_raw(inode); } /* * Invalidate the contents of fscache for this inode. This will not sleep. */ static inline void nfs_fscache_invalidate(struct inode *inode, int flags) { struct nfs_fscache_inode_auxdata auxdata; struct fscache_cookie *cookie = netfs_i_cookie(&NFS_I(inode)->netfs); nfs_fscache_update_auxdata(&auxdata, inode); fscache_invalidate(cookie, &auxdata, i_size_read(inode), flags); } /* * indicate the client caching state as readable text */ static inline const char *nfs_server_fscache_state(struct nfs_server *server) { if (server->fscache) return "yes"; return "no "; } static inline void nfs_netfs_set_pgio_header(struct nfs_pgio_header *hdr, struct nfs_pageio_descriptor *desc) { hdr->netfs = desc->pg_netfs; } static inline void nfs_netfs_set_pageio_descriptor(struct nfs_pageio_descriptor *desc, struct nfs_pgio_header *hdr) { desc->pg_netfs = hdr->netfs; } static inline void nfs_netfs_reset_pageio_descriptor(struct nfs_pageio_descriptor *desc) { desc->pg_netfs = NULL; } #else /* CONFIG_NFS_FSCACHE */ static inline void nfs_netfs_inode_init(struct nfs_inode *nfsi) {} static inline void nfs_netfs_initiate_read(struct nfs_pgio_header *hdr) {} static inline void nfs_netfs_read_completion(struct nfs_pgio_header *hdr) {} static inline int nfs_netfs_folio_unlock(struct folio *folio) { return 1; } static inline void nfs_fscache_release_super_cookie(struct super_block *sb) {} static inline void nfs_fscache_init_inode(struct inode *inode) {} static inline void nfs_fscache_clear_inode(struct inode *inode) {} static inline void nfs_fscache_open_file(struct inode *inode, struct file *filp) {} static inline void nfs_fscache_release_file(struct inode *inode, struct file *file) {} static inline int nfs_netfs_readahead(struct readahead_control *ractl) { return -ENOBUFS; } static inline int nfs_netfs_read_folio(struct file *file, struct folio *folio) { return -ENOBUFS; } static inline bool nfs_fscache_release_folio(struct folio *folio, gfp_t gfp) { return true; /* may release folio */ } static inline void nfs_fscache_invalidate(struct inode *inode, int flags) {} static inline const char *nfs_server_fscache_state(struct nfs_server *server) { return "no "; } static inline void nfs_netfs_set_pgio_header(struct nfs_pgio_header *hdr, struct nfs_pageio_descriptor *desc) {} static inline void nfs_netfs_set_pageio_descriptor(struct nfs_pageio_descriptor *desc, struct nfs_pgio_header *hdr) {} static inline void nfs_netfs_reset_pageio_descriptor(struct nfs_pageio_descriptor *desc) {} #endif /* CONFIG_NFS_FSCACHE */ #endif /* _NFS_FSCACHE_H */ |