1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * linux/drivers/staging/erofs/super.c
4 *
5 * Copyright (C) 2017-2018 HUAWEI, Inc.
6 * http://www.huawei.com/
7 * Created by Gao Xiang <gaoxiang25@huawei.com>
8 *
9 * This file is subject to the terms and conditions of the GNU General Public
10 * License. See the file COPYING in the main directory of the Linux
11 * distribution for more details.
12 */
13 #include <linux/module.h>
14 #include <linux/buffer_head.h>
15 #include <linux/statfs.h>
16 #include <linux/parser.h>
17 #include <linux/seq_file.h>
18 #include "internal.h"
19
20 #define CREATE_TRACE_POINTS
21 #include <trace/events/erofs.h>
22
23 static struct kmem_cache *erofs_inode_cachep __read_mostly;
24
init_once(void * ptr)25 static void init_once(void *ptr)
26 {
27 struct erofs_vnode *vi = ptr;
28
29 inode_init_once(&vi->vfs_inode);
30 }
31
erofs_init_inode_cache(void)32 static int erofs_init_inode_cache(void)
33 {
34 erofs_inode_cachep = kmem_cache_create("erofs_inode",
35 sizeof(struct erofs_vnode), 0,
36 SLAB_RECLAIM_ACCOUNT, init_once);
37
38 return erofs_inode_cachep != NULL ? 0 : -ENOMEM;
39 }
40
erofs_exit_inode_cache(void)41 static void erofs_exit_inode_cache(void)
42 {
43 BUG_ON(erofs_inode_cachep == NULL);
44 kmem_cache_destroy(erofs_inode_cachep);
45 }
46
alloc_inode(struct super_block * sb)47 static struct inode *alloc_inode(struct super_block *sb)
48 {
49 struct erofs_vnode *vi =
50 kmem_cache_alloc(erofs_inode_cachep, GFP_KERNEL);
51
52 if (vi == NULL)
53 return NULL;
54
55 /* zero out everything except vfs_inode */
56 memset(vi, 0, offsetof(struct erofs_vnode, vfs_inode));
57 return &vi->vfs_inode;
58 }
59
i_callback(struct rcu_head * head)60 static void i_callback(struct rcu_head *head)
61 {
62 struct inode *inode = container_of(head, struct inode, i_rcu);
63 struct erofs_vnode *vi = EROFS_V(inode);
64
65 /* be careful RCU symlink path (see ext4_inode_info->i_data)! */
66 if (is_inode_fast_symlink(inode))
67 kfree(inode->i_link);
68
69 kfree(vi->xattr_shared_xattrs);
70
71 kmem_cache_free(erofs_inode_cachep, vi);
72 }
73
destroy_inode(struct inode * inode)74 static void destroy_inode(struct inode *inode)
75 {
76 call_rcu(&inode->i_rcu, i_callback);
77 }
78
superblock_read(struct super_block * sb)79 static int superblock_read(struct super_block *sb)
80 {
81 struct erofs_sb_info *sbi;
82 struct buffer_head *bh;
83 struct erofs_super_block *layout;
84 unsigned blkszbits;
85 int ret;
86
87 bh = sb_bread(sb, 0);
88
89 if (bh == NULL) {
90 errln("cannot read erofs superblock");
91 return -EIO;
92 }
93
94 sbi = EROFS_SB(sb);
95 layout = (struct erofs_super_block *)((u8 *)bh->b_data
96 + EROFS_SUPER_OFFSET);
97
98 ret = -EINVAL;
99 if (le32_to_cpu(layout->magic) != EROFS_SUPER_MAGIC_V1) {
100 errln("cannot find valid erofs superblock");
101 goto out;
102 }
103
104 blkszbits = layout->blkszbits;
105 /* 9(512 bytes) + LOG_SECTORS_PER_BLOCK == LOG_BLOCK_SIZE */
106 if (unlikely(blkszbits != LOG_BLOCK_SIZE)) {
107 errln("blksize %u isn't supported on this platform",
108 1 << blkszbits);
109 goto out;
110 }
111
112 sbi->blocks = le32_to_cpu(layout->blocks);
113 sbi->meta_blkaddr = le32_to_cpu(layout->meta_blkaddr);
114 #ifdef CONFIG_EROFS_FS_XATTR
115 sbi->xattr_blkaddr = le32_to_cpu(layout->xattr_blkaddr);
116 #endif
117 sbi->islotbits = ffs(sizeof(struct erofs_inode_v1)) - 1;
118 #ifdef CONFIG_EROFS_FS_ZIP
119 sbi->clusterbits = 12;
120
121 if (1 << (sbi->clusterbits - 12) > Z_EROFS_CLUSTER_MAX_PAGES)
122 errln("clusterbits %u is not supported on this kernel",
123 sbi->clusterbits);
124 #endif
125
126 sbi->root_nid = le16_to_cpu(layout->root_nid);
127 sbi->inos = le64_to_cpu(layout->inos);
128
129 sbi->build_time = le64_to_cpu(layout->build_time);
130 sbi->build_time_nsec = le32_to_cpu(layout->build_time_nsec);
131
132 memcpy(&sb->s_uuid, layout->uuid, sizeof(layout->uuid));
133 memcpy(sbi->volume_name, layout->volume_name,
134 sizeof(layout->volume_name));
135
136 ret = 0;
137 out:
138 brelse(bh);
139 return ret;
140 }
141
142 #ifdef CONFIG_EROFS_FAULT_INJECTION
143 char *erofs_fault_name[FAULT_MAX] = {
144 [FAULT_KMALLOC] = "kmalloc",
145 };
146
erofs_build_fault_attr(struct erofs_sb_info * sbi,unsigned int rate)147 static void erofs_build_fault_attr(struct erofs_sb_info *sbi,
148 unsigned int rate)
149 {
150 struct erofs_fault_info *ffi = &sbi->fault_info;
151
152 if (rate) {
153 atomic_set(&ffi->inject_ops, 0);
154 ffi->inject_rate = rate;
155 ffi->inject_type = (1 << FAULT_MAX) - 1;
156 } else {
157 memset(ffi, 0, sizeof(struct erofs_fault_info));
158 }
159 }
160 #endif
161
default_options(struct erofs_sb_info * sbi)162 static void default_options(struct erofs_sb_info *sbi)
163 {
164 #ifdef CONFIG_EROFS_FS_XATTR
165 set_opt(sbi, XATTR_USER);
166 #endif
167
168 #ifdef CONFIG_EROFS_FS_POSIX_ACL
169 set_opt(sbi, POSIX_ACL);
170 #endif
171 }
172
173 enum {
174 Opt_user_xattr,
175 Opt_nouser_xattr,
176 Opt_acl,
177 Opt_noacl,
178 Opt_fault_injection,
179 Opt_err
180 };
181
182 static match_table_t erofs_tokens = {
183 {Opt_user_xattr, "user_xattr"},
184 {Opt_nouser_xattr, "nouser_xattr"},
185 {Opt_acl, "acl"},
186 {Opt_noacl, "noacl"},
187 {Opt_fault_injection, "fault_injection=%u"},
188 {Opt_err, NULL}
189 };
190
parse_options(struct super_block * sb,char * options)191 static int parse_options(struct super_block *sb, char *options)
192 {
193 substring_t args[MAX_OPT_ARGS];
194 char *p;
195 int arg = 0;
196
197 if (!options)
198 return 0;
199
200 while ((p = strsep(&options, ",")) != NULL) {
201 int token;
202
203 if (!*p)
204 continue;
205
206 args[0].to = args[0].from = NULL;
207 token = match_token(p, erofs_tokens, args);
208
209 switch (token) {
210 #ifdef CONFIG_EROFS_FS_XATTR
211 case Opt_user_xattr:
212 set_opt(EROFS_SB(sb), XATTR_USER);
213 break;
214 case Opt_nouser_xattr:
215 clear_opt(EROFS_SB(sb), XATTR_USER);
216 break;
217 #else
218 case Opt_user_xattr:
219 infoln("user_xattr options not supported");
220 break;
221 case Opt_nouser_xattr:
222 infoln("nouser_xattr options not supported");
223 break;
224 #endif
225 #ifdef CONFIG_EROFS_FS_POSIX_ACL
226 case Opt_acl:
227 set_opt(EROFS_SB(sb), POSIX_ACL);
228 break;
229 case Opt_noacl:
230 clear_opt(EROFS_SB(sb), POSIX_ACL);
231 break;
232 #else
233 case Opt_acl:
234 infoln("acl options not supported");
235 break;
236 case Opt_noacl:
237 infoln("noacl options not supported");
238 break;
239 #endif
240 case Opt_fault_injection:
241 if (args->from && match_int(args, &arg))
242 return -EINVAL;
243 #ifdef CONFIG_EROFS_FAULT_INJECTION
244 erofs_build_fault_attr(EROFS_SB(sb), arg);
245 set_opt(EROFS_SB(sb), FAULT_INJECTION);
246 #else
247 infoln("FAULT_INJECTION was not selected");
248 #endif
249 break;
250 default:
251 errln("Unrecognized mount option \"%s\" "
252 "or missing value", p);
253 return -EINVAL;
254 }
255 }
256 return 0;
257 }
258
259 #ifdef EROFS_FS_HAS_MANAGED_CACHE
260
261 static const struct address_space_operations managed_cache_aops;
262
managed_cache_releasepage(struct page * page,gfp_t gfp_mask)263 static int managed_cache_releasepage(struct page *page, gfp_t gfp_mask)
264 {
265 int ret = 1; /* 0 - busy */
266 struct address_space *const mapping = page->mapping;
267
268 BUG_ON(!PageLocked(page));
269 BUG_ON(mapping->a_ops != &managed_cache_aops);
270
271 if (PagePrivate(page))
272 ret = erofs_try_to_free_cached_page(mapping, page);
273
274 return ret;
275 }
276
managed_cache_invalidatepage(struct page * page,unsigned int offset,unsigned int length)277 static void managed_cache_invalidatepage(struct page *page,
278 unsigned int offset, unsigned int length)
279 {
280 const unsigned int stop = length + offset;
281
282 BUG_ON(!PageLocked(page));
283
284 /* Check for overflow */
285 BUG_ON(stop > PAGE_SIZE || stop < length);
286
287 if (offset == 0 && stop == PAGE_SIZE)
288 while (!managed_cache_releasepage(page, GFP_NOFS))
289 cond_resched();
290 }
291
292 static const struct address_space_operations managed_cache_aops = {
293 .releasepage = managed_cache_releasepage,
294 .invalidatepage = managed_cache_invalidatepage,
295 };
296
erofs_init_managed_cache(struct super_block * sb)297 static struct inode *erofs_init_managed_cache(struct super_block *sb)
298 {
299 struct inode *inode = new_inode(sb);
300
301 if (unlikely(inode == NULL))
302 return ERR_PTR(-ENOMEM);
303
304 set_nlink(inode, 1);
305 inode->i_size = OFFSET_MAX;
306
307 inode->i_mapping->a_ops = &managed_cache_aops;
308 mapping_set_gfp_mask(inode->i_mapping,
309 GFP_NOFS | __GFP_HIGHMEM |
310 __GFP_MOVABLE | __GFP_NOFAIL);
311 return inode;
312 }
313
314 #endif
315
erofs_read_super(struct super_block * sb,const char * dev_name,void * data,int silent)316 static int erofs_read_super(struct super_block *sb,
317 const char *dev_name, void *data, int silent)
318 {
319 struct inode *inode;
320 struct erofs_sb_info *sbi;
321 int err = -EINVAL;
322
323 infoln("read_super, device -> %s", dev_name);
324 infoln("options -> %s", (char *)data);
325
326 if (unlikely(!sb_set_blocksize(sb, EROFS_BLKSIZ))) {
327 errln("failed to set erofs blksize");
328 goto err;
329 }
330
331 sbi = kzalloc(sizeof(struct erofs_sb_info), GFP_KERNEL);
332 if (unlikely(sbi == NULL)) {
333 err = -ENOMEM;
334 goto err;
335 }
336 sb->s_fs_info = sbi;
337
338 err = superblock_read(sb);
339 if (err)
340 goto err_sbread;
341
342 sb->s_magic = EROFS_SUPER_MAGIC;
343 sb->s_flags |= SB_RDONLY | SB_NOATIME;
344 sb->s_maxbytes = MAX_LFS_FILESIZE;
345 sb->s_time_gran = 1;
346
347 sb->s_op = &erofs_sops;
348
349 #ifdef CONFIG_EROFS_FS_XATTR
350 sb->s_xattr = erofs_xattr_handlers;
351 #endif
352
353 /* set erofs default mount options */
354 default_options(sbi);
355
356 err = parse_options(sb, data);
357 if (err)
358 goto err_parseopt;
359
360 if (!silent)
361 infoln("root inode @ nid %llu", ROOT_NID(sbi));
362
363 #ifdef CONFIG_EROFS_FS_ZIP
364 INIT_RADIX_TREE(&sbi->workstn_tree, GFP_ATOMIC);
365 #endif
366
367 #ifdef EROFS_FS_HAS_MANAGED_CACHE
368 sbi->managed_cache = erofs_init_managed_cache(sb);
369 if (IS_ERR(sbi->managed_cache)) {
370 err = PTR_ERR(sbi->managed_cache);
371 goto err_init_managed_cache;
372 }
373 #endif
374
375 /* get the root inode */
376 inode = erofs_iget(sb, ROOT_NID(sbi), true);
377 if (IS_ERR(inode)) {
378 err = PTR_ERR(inode);
379 goto err_iget;
380 }
381
382 if (!S_ISDIR(inode->i_mode)) {
383 errln("rootino(nid %llu) is not a directory(i_mode %o)",
384 ROOT_NID(sbi), inode->i_mode);
385 err = -EINVAL;
386 goto err_isdir;
387 }
388
389 sb->s_root = d_make_root(inode);
390 if (sb->s_root == NULL) {
391 err = -ENOMEM;
392 goto err_makeroot;
393 }
394
395 /* save the device name to sbi */
396 sbi->dev_name = __getname();
397 if (sbi->dev_name == NULL) {
398 err = -ENOMEM;
399 goto err_devname;
400 }
401
402 snprintf(sbi->dev_name, PATH_MAX, "%s", dev_name);
403 sbi->dev_name[PATH_MAX - 1] = '\0';
404
405 erofs_register_super(sb);
406
407 /*
408 * We already have a positive dentry, which was instantiated
409 * by d_make_root. Just need to d_rehash it.
410 */
411 d_rehash(sb->s_root);
412
413 if (!silent)
414 infoln("mounted on %s with opts: %s.", dev_name,
415 (char *)data);
416 return 0;
417 /*
418 * please add a label for each exit point and use
419 * the following name convention, thus new features
420 * can be integrated easily without renaming labels.
421 */
422 err_devname:
423 dput(sb->s_root);
424 err_makeroot:
425 err_isdir:
426 if (sb->s_root == NULL)
427 iput(inode);
428 err_iget:
429 #ifdef EROFS_FS_HAS_MANAGED_CACHE
430 iput(sbi->managed_cache);
431 err_init_managed_cache:
432 #endif
433 err_parseopt:
434 err_sbread:
435 sb->s_fs_info = NULL;
436 kfree(sbi);
437 err:
438 return err;
439 }
440
441 /*
442 * could be triggered after deactivate_locked_super()
443 * is called, thus including umount and failed to initialize.
444 */
erofs_put_super(struct super_block * sb)445 static void erofs_put_super(struct super_block *sb)
446 {
447 struct erofs_sb_info *sbi = EROFS_SB(sb);
448
449 /* for cases which are failed in "read_super" */
450 if (sbi == NULL)
451 return;
452
453 WARN_ON(sb->s_magic != EROFS_SUPER_MAGIC);
454
455 infoln("unmounted for %s", sbi->dev_name);
456 __putname(sbi->dev_name);
457
458 #ifdef EROFS_FS_HAS_MANAGED_CACHE
459 iput(sbi->managed_cache);
460 #endif
461
462 mutex_lock(&sbi->umount_mutex);
463
464 #ifdef CONFIG_EROFS_FS_ZIP
465 erofs_workstation_cleanup_all(sb);
466 #endif
467
468 erofs_unregister_super(sb);
469 mutex_unlock(&sbi->umount_mutex);
470
471 kfree(sbi);
472 sb->s_fs_info = NULL;
473 }
474
475
476 struct erofs_mount_private {
477 const char *dev_name;
478 char *options;
479 };
480
481 /* support mount_bdev() with options */
erofs_fill_super(struct super_block * sb,void * _priv,int silent)482 static int erofs_fill_super(struct super_block *sb,
483 void *_priv, int silent)
484 {
485 struct erofs_mount_private *priv = _priv;
486
487 return erofs_read_super(sb, priv->dev_name,
488 priv->options, silent);
489 }
490
erofs_mount(struct file_system_type * fs_type,int flags,const char * dev_name,void * data)491 static struct dentry *erofs_mount(
492 struct file_system_type *fs_type, int flags,
493 const char *dev_name, void *data)
494 {
495 struct erofs_mount_private priv = {
496 .dev_name = dev_name,
497 .options = data
498 };
499
500 return mount_bdev(fs_type, flags, dev_name,
501 &priv, erofs_fill_super);
502 }
503
erofs_kill_sb(struct super_block * sb)504 static void erofs_kill_sb(struct super_block *sb)
505 {
506 kill_block_super(sb);
507 }
508
509 static struct shrinker erofs_shrinker_info = {
510 .scan_objects = erofs_shrink_scan,
511 .count_objects = erofs_shrink_count,
512 .seeks = DEFAULT_SEEKS,
513 };
514
515 static struct file_system_type erofs_fs_type = {
516 .owner = THIS_MODULE,
517 .name = "erofs",
518 .mount = erofs_mount,
519 .kill_sb = erofs_kill_sb,
520 .fs_flags = FS_REQUIRES_DEV,
521 };
522 MODULE_ALIAS_FS("erofs");
523
524 #ifdef CONFIG_EROFS_FS_ZIP
525 extern int z_erofs_init_zip_subsystem(void);
526 extern void z_erofs_exit_zip_subsystem(void);
527 #endif
528
erofs_module_init(void)529 static int __init erofs_module_init(void)
530 {
531 int err;
532
533 erofs_check_ondisk_layout_definitions();
534 infoln("initializing erofs " EROFS_VERSION);
535
536 err = erofs_init_inode_cache();
537 if (err)
538 goto icache_err;
539
540 err = register_shrinker(&erofs_shrinker_info);
541 if (err)
542 goto shrinker_err;
543
544 #ifdef CONFIG_EROFS_FS_ZIP
545 err = z_erofs_init_zip_subsystem();
546 if (err)
547 goto zip_err;
548 #endif
549
550 err = register_filesystem(&erofs_fs_type);
551 if (err)
552 goto fs_err;
553
554 infoln("successfully to initialize erofs");
555 return 0;
556
557 fs_err:
558 #ifdef CONFIG_EROFS_FS_ZIP
559 z_erofs_exit_zip_subsystem();
560 zip_err:
561 #endif
562 unregister_shrinker(&erofs_shrinker_info);
563 shrinker_err:
564 erofs_exit_inode_cache();
565 icache_err:
566 return err;
567 }
568
erofs_module_exit(void)569 static void __exit erofs_module_exit(void)
570 {
571 unregister_filesystem(&erofs_fs_type);
572 #ifdef CONFIG_EROFS_FS_ZIP
573 z_erofs_exit_zip_subsystem();
574 #endif
575 unregister_shrinker(&erofs_shrinker_info);
576 erofs_exit_inode_cache();
577 infoln("successfully finalize erofs");
578 }
579
580 /* get filesystem statistics */
erofs_statfs(struct dentry * dentry,struct kstatfs * buf)581 static int erofs_statfs(struct dentry *dentry, struct kstatfs *buf)
582 {
583 struct super_block *sb = dentry->d_sb;
584 struct erofs_sb_info *sbi = EROFS_SB(sb);
585 u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
586
587 buf->f_type = sb->s_magic;
588 buf->f_bsize = EROFS_BLKSIZ;
589 buf->f_blocks = sbi->blocks;
590 buf->f_bfree = buf->f_bavail = 0;
591
592 buf->f_files = ULLONG_MAX;
593 buf->f_ffree = ULLONG_MAX - sbi->inos;
594
595 buf->f_namelen = EROFS_NAME_LEN;
596
597 buf->f_fsid.val[0] = (u32)id;
598 buf->f_fsid.val[1] = (u32)(id >> 32);
599 return 0;
600 }
601
erofs_show_options(struct seq_file * seq,struct dentry * root)602 static int erofs_show_options(struct seq_file *seq, struct dentry *root)
603 {
604 struct erofs_sb_info *sbi __maybe_unused = EROFS_SB(root->d_sb);
605
606 #ifdef CONFIG_EROFS_FS_XATTR
607 if (test_opt(sbi, XATTR_USER))
608 seq_puts(seq, ",user_xattr");
609 else
610 seq_puts(seq, ",nouser_xattr");
611 #endif
612 #ifdef CONFIG_EROFS_FS_POSIX_ACL
613 if (test_opt(sbi, POSIX_ACL))
614 seq_puts(seq, ",acl");
615 else
616 seq_puts(seq, ",noacl");
617 #endif
618 #ifdef CONFIG_EROFS_FAULT_INJECTION
619 if (test_opt(sbi, FAULT_INJECTION))
620 seq_printf(seq, ",fault_injection=%u",
621 sbi->fault_info.inject_rate);
622 #endif
623 return 0;
624 }
625
erofs_remount(struct super_block * sb,int * flags,char * data)626 static int erofs_remount(struct super_block *sb, int *flags, char *data)
627 {
628 BUG_ON(!sb_rdonly(sb));
629
630 *flags |= SB_RDONLY;
631 return 0;
632 }
633
634 const struct super_operations erofs_sops = {
635 .put_super = erofs_put_super,
636 .alloc_inode = alloc_inode,
637 .destroy_inode = destroy_inode,
638 .statfs = erofs_statfs,
639 .show_options = erofs_show_options,
640 .remount_fs = erofs_remount,
641 };
642
643 module_init(erofs_module_init);
644 module_exit(erofs_module_exit);
645
646 MODULE_DESCRIPTION("Enhanced ROM File System");
647 MODULE_AUTHOR("Gao Xiang, Yu Chao, Miao Xie, CONSUMER BG, HUAWEI Inc.");
648 MODULE_LICENSE("GPL");
649
650