Merge tag 'sched_ext-for-6.12' of git://git.kernel.org/pub/scm/linux/kernel/git/tj...
[sfrench/cifs-2.6.git] / fs / cachefiles / cache.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /* Manage high-level VFS aspects of a cache.
3  *
4  * Copyright (C) 2007, 2021 Red Hat, Inc. All Rights Reserved.
5  * Written by David Howells (dhowells@redhat.com)
6  */
7
8 #include <linux/slab.h>
9 #include <linux/statfs.h>
10 #include <linux/namei.h>
11 #include <trace/events/fscache.h>
12 #include "internal.h"
13
14 /*
15  * Bring a cache online.
16  */
17 int cachefiles_add_cache(struct cachefiles_cache *cache)
18 {
19         struct fscache_cache *cache_cookie;
20         struct path path;
21         struct kstatfs stats;
22         struct dentry *graveyard, *cachedir, *root;
23         const struct cred *saved_cred;
24         int ret;
25
26         _enter("");
27
28         cache_cookie = fscache_acquire_cache(cache->tag);
29         if (IS_ERR(cache_cookie))
30                 return PTR_ERR(cache_cookie);
31
32         /* we want to work under the module's security ID */
33         ret = cachefiles_get_security_ID(cache);
34         if (ret < 0)
35                 goto error_getsec;
36
37         cachefiles_begin_secure(cache, &saved_cred);
38
39         /* look up the directory at the root of the cache */
40         ret = kern_path(cache->rootdirname, LOOKUP_DIRECTORY, &path);
41         if (ret < 0)
42                 goto error_open_root;
43
44         cache->mnt = path.mnt;
45         root = path.dentry;
46
47         ret = -EINVAL;
48         if (is_idmapped_mnt(path.mnt)) {
49                 pr_warn("File cache on idmapped mounts not supported");
50                 goto error_unsupported;
51         }
52
53         /* Check features of the backing filesystem:
54          * - Directories must support looking up and directory creation
55          * - We create tmpfiles to handle invalidation
56          * - We use xattrs to store metadata
57          * - We need to be able to query the amount of space available
58          * - We want to be able to sync the filesystem when stopping the cache
59          * - We use DIO to/from pages, so the blocksize mustn't be too big.
60          */
61         ret = -EOPNOTSUPP;
62         if (d_is_negative(root) ||
63             !d_backing_inode(root)->i_op->lookup ||
64             !d_backing_inode(root)->i_op->mkdir ||
65             !d_backing_inode(root)->i_op->tmpfile ||
66             !(d_backing_inode(root)->i_opflags & IOP_XATTR) ||
67             !root->d_sb->s_op->statfs ||
68             !root->d_sb->s_op->sync_fs ||
69             root->d_sb->s_blocksize > PAGE_SIZE)
70                 goto error_unsupported;
71
72         ret = -EROFS;
73         if (sb_rdonly(root->d_sb))
74                 goto error_unsupported;
75
76         /* determine the security of the on-disk cache as this governs
77          * security ID of files we create */
78         ret = cachefiles_determine_cache_security(cache, root, &saved_cred);
79         if (ret < 0)
80                 goto error_unsupported;
81
82         /* get the cache size and blocksize */
83         ret = vfs_statfs(&path, &stats);
84         if (ret < 0)
85                 goto error_unsupported;
86
87         ret = -ERANGE;
88         if (stats.f_bsize <= 0)
89                 goto error_unsupported;
90
91         ret = -EOPNOTSUPP;
92         if (stats.f_bsize > PAGE_SIZE)
93                 goto error_unsupported;
94
95         cache->bsize = stats.f_bsize;
96         cache->bshift = ilog2(stats.f_bsize);
97
98         _debug("blksize %u (shift %u)",
99                cache->bsize, cache->bshift);
100
101         _debug("size %llu, avail %llu",
102                (unsigned long long) stats.f_blocks,
103                (unsigned long long) stats.f_bavail);
104
105         /* set up caching limits */
106         do_div(stats.f_files, 100);
107         cache->fstop = stats.f_files * cache->fstop_percent;
108         cache->fcull = stats.f_files * cache->fcull_percent;
109         cache->frun  = stats.f_files * cache->frun_percent;
110
111         _debug("limits {%llu,%llu,%llu} files",
112                (unsigned long long) cache->frun,
113                (unsigned long long) cache->fcull,
114                (unsigned long long) cache->fstop);
115
116         do_div(stats.f_blocks, 100);
117         cache->bstop = stats.f_blocks * cache->bstop_percent;
118         cache->bcull = stats.f_blocks * cache->bcull_percent;
119         cache->brun  = stats.f_blocks * cache->brun_percent;
120
121         _debug("limits {%llu,%llu,%llu} blocks",
122                (unsigned long long) cache->brun,
123                (unsigned long long) cache->bcull,
124                (unsigned long long) cache->bstop);
125
126         /* get the cache directory and check its type */
127         cachedir = cachefiles_get_directory(cache, root, "cache", NULL);
128         if (IS_ERR(cachedir)) {
129                 ret = PTR_ERR(cachedir);
130                 goto error_unsupported;
131         }
132
133         cache->store = cachedir;
134
135         /* get the graveyard directory */
136         graveyard = cachefiles_get_directory(cache, root, "graveyard", NULL);
137         if (IS_ERR(graveyard)) {
138                 ret = PTR_ERR(graveyard);
139                 goto error_unsupported;
140         }
141
142         cache->graveyard = graveyard;
143         cache->cache = cache_cookie;
144
145         ret = fscache_add_cache(cache_cookie, &cachefiles_cache_ops, cache);
146         if (ret < 0)
147                 goto error_add_cache;
148
149         /* done */
150         set_bit(CACHEFILES_READY, &cache->flags);
151         dput(root);
152
153         pr_info("File cache on %s registered\n", cache_cookie->name);
154
155         /* check how much space the cache has */
156         cachefiles_has_space(cache, 0, 0, cachefiles_has_space_check);
157         cachefiles_end_secure(cache, saved_cred);
158         _leave(" = 0 [%px]", cache->cache);
159         return 0;
160
161 error_add_cache:
162         cachefiles_put_directory(cache->graveyard);
163         cache->graveyard = NULL;
164 error_unsupported:
165         cachefiles_put_directory(cache->store);
166         cache->store = NULL;
167         mntput(cache->mnt);
168         cache->mnt = NULL;
169         dput(root);
170 error_open_root:
171         cachefiles_end_secure(cache, saved_cred);
172         put_cred(cache->cache_cred);
173         cache->cache_cred = NULL;
174 error_getsec:
175         fscache_relinquish_cache(cache_cookie);
176         cache->cache = NULL;
177         pr_err("Failed to register: %d\n", ret);
178         return ret;
179 }
180
181 /*
182  * See if we have space for a number of pages and/or a number of files in the
183  * cache
184  */
185 int cachefiles_has_space(struct cachefiles_cache *cache,
186                          unsigned fnr, unsigned bnr,
187                          enum cachefiles_has_space_for reason)
188 {
189         struct kstatfs stats;
190         u64 b_avail, b_writing;
191         int ret;
192
193         struct path path = {
194                 .mnt    = cache->mnt,
195                 .dentry = cache->mnt->mnt_root,
196         };
197
198         //_enter("{%llu,%llu,%llu,%llu,%llu,%llu},%u,%u",
199         //       (unsigned long long) cache->frun,
200         //       (unsigned long long) cache->fcull,
201         //       (unsigned long long) cache->fstop,
202         //       (unsigned long long) cache->brun,
203         //       (unsigned long long) cache->bcull,
204         //       (unsigned long long) cache->bstop,
205         //       fnr, bnr);
206
207         /* find out how many pages of blockdev are available */
208         memset(&stats, 0, sizeof(stats));
209
210         ret = vfs_statfs(&path, &stats);
211         if (ret < 0) {
212                 trace_cachefiles_vfs_error(NULL, d_inode(path.dentry), ret,
213                                            cachefiles_trace_statfs_error);
214                 if (ret == -EIO)
215                         cachefiles_io_error(cache, "statfs failed");
216                 _leave(" = %d", ret);
217                 return ret;
218         }
219
220         b_avail = stats.f_bavail;
221         b_writing = atomic_long_read(&cache->b_writing);
222         if (b_avail > b_writing)
223                 b_avail -= b_writing;
224         else
225                 b_avail = 0;
226
227         //_debug("avail %llu,%llu",
228         //       (unsigned long long)stats.f_ffree,
229         //       (unsigned long long)b_avail);
230
231         /* see if there is sufficient space */
232         if (stats.f_ffree > fnr)
233                 stats.f_ffree -= fnr;
234         else
235                 stats.f_ffree = 0;
236
237         if (b_avail > bnr)
238                 b_avail -= bnr;
239         else
240                 b_avail = 0;
241
242         ret = -ENOBUFS;
243         if (stats.f_ffree < cache->fstop ||
244             b_avail < cache->bstop)
245                 goto stop_and_begin_cull;
246
247         ret = 0;
248         if (stats.f_ffree < cache->fcull ||
249             b_avail < cache->bcull)
250                 goto begin_cull;
251
252         if (test_bit(CACHEFILES_CULLING, &cache->flags) &&
253             stats.f_ffree >= cache->frun &&
254             b_avail >= cache->brun &&
255             test_and_clear_bit(CACHEFILES_CULLING, &cache->flags)
256             ) {
257                 _debug("cease culling");
258                 cachefiles_state_changed(cache);
259         }
260
261         //_leave(" = 0");
262         return 0;
263
264 stop_and_begin_cull:
265         switch (reason) {
266         case cachefiles_has_space_for_write:
267                 fscache_count_no_write_space();
268                 break;
269         case cachefiles_has_space_for_create:
270                 fscache_count_no_create_space();
271                 break;
272         default:
273                 break;
274         }
275 begin_cull:
276         if (!test_and_set_bit(CACHEFILES_CULLING, &cache->flags)) {
277                 _debug("### CULL CACHE ###");
278                 cachefiles_state_changed(cache);
279         }
280
281         _leave(" = %d", ret);
282         return ret;
283 }
284
285 /*
286  * Mark all the objects as being out of service and queue them all for cleanup.
287  */
288 static void cachefiles_withdraw_objects(struct cachefiles_cache *cache)
289 {
290         struct cachefiles_object *object;
291         unsigned int count = 0;
292
293         _enter("");
294
295         spin_lock(&cache->object_list_lock);
296
297         while (!list_empty(&cache->object_list)) {
298                 object = list_first_entry(&cache->object_list,
299                                           struct cachefiles_object, cache_link);
300                 cachefiles_see_object(object, cachefiles_obj_see_withdrawal);
301                 list_del_init(&object->cache_link);
302                 fscache_withdraw_cookie(object->cookie);
303                 count++;
304                 if ((count & 63) == 0) {
305                         spin_unlock(&cache->object_list_lock);
306                         cond_resched();
307                         spin_lock(&cache->object_list_lock);
308                 }
309         }
310
311         spin_unlock(&cache->object_list_lock);
312         _leave(" [%u objs]", count);
313 }
314
315 /*
316  * Withdraw fscache volumes.
317  */
318 static void cachefiles_withdraw_fscache_volumes(struct cachefiles_cache *cache)
319 {
320         struct list_head *cur;
321         struct cachefiles_volume *volume;
322         struct fscache_volume *vcookie;
323
324         _enter("");
325 retry:
326         spin_lock(&cache->object_list_lock);
327         list_for_each(cur, &cache->volumes) {
328                 volume = list_entry(cur, struct cachefiles_volume, cache_link);
329
330                 if (atomic_read(&volume->vcookie->n_accesses) == 0)
331                         continue;
332
333                 vcookie = fscache_try_get_volume(volume->vcookie,
334                                                  fscache_volume_get_withdraw);
335                 if (vcookie) {
336                         spin_unlock(&cache->object_list_lock);
337                         fscache_withdraw_volume(vcookie);
338                         fscache_put_volume(vcookie, fscache_volume_put_withdraw);
339                         goto retry;
340                 }
341         }
342         spin_unlock(&cache->object_list_lock);
343
344         _leave("");
345 }
346
347 /*
348  * Withdraw cachefiles volumes.
349  */
350 static void cachefiles_withdraw_volumes(struct cachefiles_cache *cache)
351 {
352         _enter("");
353
354         for (;;) {
355                 struct fscache_volume *vcookie = NULL;
356                 struct cachefiles_volume *volume = NULL;
357
358                 spin_lock(&cache->object_list_lock);
359                 if (!list_empty(&cache->volumes)) {
360                         volume = list_first_entry(&cache->volumes,
361                                                   struct cachefiles_volume, cache_link);
362                         vcookie = fscache_try_get_volume(volume->vcookie,
363                                                          fscache_volume_get_withdraw);
364                         if (!vcookie) {
365                                 spin_unlock(&cache->object_list_lock);
366                                 cpu_relax();
367                                 continue;
368                         }
369                         list_del_init(&volume->cache_link);
370                 }
371                 spin_unlock(&cache->object_list_lock);
372                 if (!volume)
373                         break;
374
375                 cachefiles_withdraw_volume(volume);
376                 fscache_put_volume(vcookie, fscache_volume_put_withdraw);
377         }
378
379         _leave("");
380 }
381
382 /*
383  * Sync a cache to backing disk.
384  */
385 static void cachefiles_sync_cache(struct cachefiles_cache *cache)
386 {
387         const struct cred *saved_cred;
388         int ret;
389
390         _enter("%s", cache->cache->name);
391
392         /* make sure all pages pinned by operations on behalf of the netfs are
393          * written to disc */
394         cachefiles_begin_secure(cache, &saved_cred);
395         down_read(&cache->mnt->mnt_sb->s_umount);
396         ret = sync_filesystem(cache->mnt->mnt_sb);
397         up_read(&cache->mnt->mnt_sb->s_umount);
398         cachefiles_end_secure(cache, saved_cred);
399
400         if (ret == -EIO)
401                 cachefiles_io_error(cache,
402                                     "Attempt to sync backing fs superblock returned error %d",
403                                     ret);
404 }
405
406 /*
407  * Withdraw cache objects.
408  */
409 void cachefiles_withdraw_cache(struct cachefiles_cache *cache)
410 {
411         struct fscache_cache *fscache = cache->cache;
412
413         pr_info("File cache on %s unregistering\n", fscache->name);
414
415         fscache_withdraw_cache(fscache);
416         cachefiles_withdraw_fscache_volumes(cache);
417
418         /* we now have to destroy all the active objects pertaining to this
419          * cache - which we do by passing them off to thread pool to be
420          * disposed of */
421         cachefiles_withdraw_objects(cache);
422         fscache_wait_for_objects(fscache);
423
424         cachefiles_withdraw_volumes(cache);
425         cachefiles_sync_cache(cache);
426         cache->cache = NULL;
427         fscache_relinquish_cache(fscache);
428 }