2 * linux/fs/hfsplus/super.c
5 * Brad Boyer (flar@allandria.com)
6 * (C) 2003 Ardis Technologies <roman@ardistech.com>
10 #include <linux/module.h>
11 #include <linux/init.h>
12 #include <linux/pagemap.h>
13 #include <linux/blkdev.h>
15 #include <linux/slab.h>
16 #include <linux/vfs.h>
17 #include <linux/nls.h>
19 static struct inode *hfsplus_alloc_inode(struct super_block *sb);
20 static void hfsplus_destroy_inode(struct inode *inode);
22 #include "hfsplus_fs.h"
24 static int hfsplus_system_read_inode(struct inode *inode)
26 struct hfsplus_vh *vhdr = HFSPLUS_SB(inode->i_sb)->s_vhdr;
28 switch (inode->i_ino) {
29 case HFSPLUS_EXT_CNID:
30 hfsplus_inode_read_fork(inode, &vhdr->ext_file);
31 inode->i_mapping->a_ops = &hfsplus_btree_aops;
33 case HFSPLUS_CAT_CNID:
34 hfsplus_inode_read_fork(inode, &vhdr->cat_file);
35 inode->i_mapping->a_ops = &hfsplus_btree_aops;
37 case HFSPLUS_ALLOC_CNID:
38 hfsplus_inode_read_fork(inode, &vhdr->alloc_file);
39 inode->i_mapping->a_ops = &hfsplus_aops;
41 case HFSPLUS_START_CNID:
42 hfsplus_inode_read_fork(inode, &vhdr->start_file);
44 case HFSPLUS_ATTR_CNID:
45 hfsplus_inode_read_fork(inode, &vhdr->attr_file);
46 inode->i_mapping->a_ops = &hfsplus_btree_aops;
55 struct inode *hfsplus_iget(struct super_block *sb, unsigned long ino)
57 struct hfs_find_data fd;
61 inode = iget_locked(sb, ino);
63 return ERR_PTR(-ENOMEM);
64 if (!(inode->i_state & I_NEW))
67 INIT_LIST_HEAD(&HFSPLUS_I(inode)->open_dir_list);
68 mutex_init(&HFSPLUS_I(inode)->extents_lock);
69 HFSPLUS_I(inode)->flags = 0;
70 HFSPLUS_I(inode)->extent_state = 0;
71 HFSPLUS_I(inode)->rsrc_inode = NULL;
72 atomic_set(&HFSPLUS_I(inode)->opencnt, 0);
74 if (inode->i_ino >= HFSPLUS_FIRSTUSER_CNID ||
75 inode->i_ino == HFSPLUS_ROOT_CNID) {
76 err = hfs_find_init(HFSPLUS_SB(inode->i_sb)->cat_tree, &fd);
78 err = hfsplus_find_cat(inode->i_sb, inode->i_ino, &fd);
80 err = hfsplus_cat_read_inode(inode, &fd);
84 err = hfsplus_system_read_inode(inode);
92 unlock_new_inode(inode);
96 static int hfsplus_system_write_inode(struct inode *inode)
98 struct hfsplus_sb_info *sbi = HFSPLUS_SB(inode->i_sb);
99 struct hfsplus_vh *vhdr = sbi->s_vhdr;
100 struct hfsplus_fork_raw *fork;
101 struct hfs_btree *tree = NULL;
103 switch (inode->i_ino) {
104 case HFSPLUS_EXT_CNID:
105 fork = &vhdr->ext_file;
106 tree = sbi->ext_tree;
108 case HFSPLUS_CAT_CNID:
109 fork = &vhdr->cat_file;
110 tree = sbi->cat_tree;
112 case HFSPLUS_ALLOC_CNID:
113 fork = &vhdr->alloc_file;
115 case HFSPLUS_START_CNID:
116 fork = &vhdr->start_file;
118 case HFSPLUS_ATTR_CNID:
119 fork = &vhdr->attr_file;
120 tree = sbi->attr_tree;
126 if (fork->total_size != cpu_to_be64(inode->i_size)) {
127 set_bit(HFSPLUS_SB_WRITEBACKUP, &sbi->flags);
128 hfsplus_mark_mdb_dirty(inode->i_sb);
130 hfsplus_inode_write_fork(inode, fork);
132 hfs_btree_write(tree);
136 static int hfsplus_write_inode(struct inode *inode,
137 struct writeback_control *wbc)
141 dprint(DBG_INODE, "hfsplus_write_inode: %lu\n", inode->i_ino);
143 err = hfsplus_ext_write_extent(inode);
147 if (inode->i_ino >= HFSPLUS_FIRSTUSER_CNID ||
148 inode->i_ino == HFSPLUS_ROOT_CNID)
149 return hfsplus_cat_write_inode(inode);
151 return hfsplus_system_write_inode(inode);
154 static void hfsplus_evict_inode(struct inode *inode)
156 dprint(DBG_INODE, "hfsplus_evict_inode: %lu\n", inode->i_ino);
157 truncate_inode_pages(&inode->i_data, 0);
159 if (HFSPLUS_IS_RSRC(inode)) {
160 HFSPLUS_I(HFSPLUS_I(inode)->rsrc_inode)->rsrc_inode = NULL;
161 iput(HFSPLUS_I(inode)->rsrc_inode);
165 static int hfsplus_sync_fs(struct super_block *sb, int wait)
167 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
168 struct hfsplus_vh *vhdr = sbi->s_vhdr;
169 int write_backup = 0;
175 dprint(DBG_SUPER, "hfsplus_sync_fs\n");
178 * Explicitly write out the special metadata inodes.
180 * While these special inodes are marked as hashed and written
181 * out peridocically by the flusher threads we redirty them
182 * during writeout of normal inodes, and thus the life lock
183 * prevents us from getting the latest state to disk.
185 error = filemap_write_and_wait(sbi->cat_tree->inode->i_mapping);
186 error2 = filemap_write_and_wait(sbi->ext_tree->inode->i_mapping);
189 if (sbi->attr_tree) {
191 filemap_write_and_wait(sbi->attr_tree->inode->i_mapping);
195 error2 = filemap_write_and_wait(sbi->alloc_file->i_mapping);
199 mutex_lock(&sbi->vh_mutex);
200 mutex_lock(&sbi->alloc_mutex);
201 vhdr->free_blocks = cpu_to_be32(sbi->free_blocks);
202 vhdr->next_cnid = cpu_to_be32(sbi->next_cnid);
203 vhdr->folder_count = cpu_to_be32(sbi->folder_count);
204 vhdr->file_count = cpu_to_be32(sbi->file_count);
206 if (test_and_clear_bit(HFSPLUS_SB_WRITEBACKUP, &sbi->flags)) {
207 memcpy(sbi->s_backup_vhdr, sbi->s_vhdr, sizeof(*sbi->s_vhdr));
211 error2 = hfsplus_submit_bio(sb,
212 sbi->part_start + HFSPLUS_VOLHEAD_SECTOR,
213 sbi->s_vhdr_buf, NULL, WRITE_SYNC);
219 error2 = hfsplus_submit_bio(sb,
220 sbi->part_start + sbi->sect_count - 2,
221 sbi->s_backup_vhdr_buf, NULL, WRITE_SYNC);
225 mutex_unlock(&sbi->alloc_mutex);
226 mutex_unlock(&sbi->vh_mutex);
228 if (!test_bit(HFSPLUS_SB_NOBARRIER, &sbi->flags))
229 blkdev_issue_flush(sb->s_bdev, GFP_KERNEL, NULL);
234 static void delayed_sync_fs(struct work_struct *work)
236 struct hfsplus_sb_info *sbi;
238 sbi = container_of(work, struct hfsplus_sb_info, sync_work.work);
240 spin_lock(&sbi->work_lock);
241 sbi->work_queued = 0;
242 spin_unlock(&sbi->work_lock);
244 hfsplus_sync_fs(sbi->alloc_file->i_sb, 1);
247 void hfsplus_mark_mdb_dirty(struct super_block *sb)
249 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
252 if (sb->s_flags & MS_RDONLY)
255 spin_lock(&sbi->work_lock);
256 if (!sbi->work_queued) {
257 delay = msecs_to_jiffies(dirty_writeback_interval * 10);
258 queue_delayed_work(system_long_wq, &sbi->sync_work, delay);
259 sbi->work_queued = 1;
261 spin_unlock(&sbi->work_lock);
264 static void hfsplus_put_super(struct super_block *sb)
266 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
268 dprint(DBG_SUPER, "hfsplus_put_super\n");
270 cancel_delayed_work_sync(&sbi->sync_work);
272 if (!(sb->s_flags & MS_RDONLY) && sbi->s_vhdr) {
273 struct hfsplus_vh *vhdr = sbi->s_vhdr;
275 vhdr->modify_date = hfsp_now2mt();
276 vhdr->attributes |= cpu_to_be32(HFSPLUS_VOL_UNMNT);
277 vhdr->attributes &= cpu_to_be32(~HFSPLUS_VOL_INCNSTNT);
279 hfsplus_sync_fs(sb, 1);
282 hfs_btree_close(sbi->attr_tree);
283 hfs_btree_close(sbi->cat_tree);
284 hfs_btree_close(sbi->ext_tree);
285 iput(sbi->alloc_file);
286 iput(sbi->hidden_dir);
287 kfree(sbi->s_vhdr_buf);
288 kfree(sbi->s_backup_vhdr_buf);
289 unload_nls(sbi->nls);
290 kfree(sb->s_fs_info);
291 sb->s_fs_info = NULL;
294 static int hfsplus_statfs(struct dentry *dentry, struct kstatfs *buf)
296 struct super_block *sb = dentry->d_sb;
297 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
298 u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
300 buf->f_type = HFSPLUS_SUPER_MAGIC;
301 buf->f_bsize = sb->s_blocksize;
302 buf->f_blocks = sbi->total_blocks << sbi->fs_shift;
303 buf->f_bfree = sbi->free_blocks << sbi->fs_shift;
304 buf->f_bavail = buf->f_bfree;
305 buf->f_files = 0xFFFFFFFF;
306 buf->f_ffree = 0xFFFFFFFF - sbi->next_cnid;
307 buf->f_fsid.val[0] = (u32)id;
308 buf->f_fsid.val[1] = (u32)(id >> 32);
309 buf->f_namelen = HFSPLUS_MAX_STRLEN;
314 static int hfsplus_remount(struct super_block *sb, int *flags, char *data)
316 if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY))
318 if (!(*flags & MS_RDONLY)) {
319 struct hfsplus_vh *vhdr = HFSPLUS_SB(sb)->s_vhdr;
322 if (!hfsplus_parse_options_remount(data, &force))
325 if (!(vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_UNMNT))) {
326 printk(KERN_WARNING "hfs: filesystem was "
327 "not cleanly unmounted, "
328 "running fsck.hfsplus is recommended. "
329 "leaving read-only.\n");
330 sb->s_flags |= MS_RDONLY;
334 } else if (vhdr->attributes &
335 cpu_to_be32(HFSPLUS_VOL_SOFTLOCK)) {
336 printk(KERN_WARNING "hfs: filesystem is marked locked, "
337 "leaving read-only.\n");
338 sb->s_flags |= MS_RDONLY;
340 } else if (vhdr->attributes &
341 cpu_to_be32(HFSPLUS_VOL_JOURNALED)) {
342 printk(KERN_WARNING "hfs: filesystem is "
344 "leaving read-only.\n");
345 sb->s_flags |= MS_RDONLY;
352 static const struct super_operations hfsplus_sops = {
353 .alloc_inode = hfsplus_alloc_inode,
354 .destroy_inode = hfsplus_destroy_inode,
355 .write_inode = hfsplus_write_inode,
356 .evict_inode = hfsplus_evict_inode,
357 .put_super = hfsplus_put_super,
358 .sync_fs = hfsplus_sync_fs,
359 .statfs = hfsplus_statfs,
360 .remount_fs = hfsplus_remount,
361 .show_options = hfsplus_show_options,
364 static int hfsplus_fill_super(struct super_block *sb, void *data, int silent)
366 struct hfsplus_vh *vhdr;
367 struct hfsplus_sb_info *sbi;
368 hfsplus_cat_entry entry;
369 struct hfs_find_data fd;
370 struct inode *root, *inode;
372 struct nls_table *nls = NULL;
373 u64 last_fs_block, last_fs_page;
377 sbi = kzalloc(sizeof(*sbi), GFP_KERNEL);
382 mutex_init(&sbi->alloc_mutex);
383 mutex_init(&sbi->vh_mutex);
384 spin_lock_init(&sbi->work_lock);
385 INIT_DELAYED_WORK(&sbi->sync_work, delayed_sync_fs);
386 hfsplus_fill_defaults(sbi);
389 if (!hfsplus_parse_options(data, sbi)) {
390 printk(KERN_ERR "hfs: unable to parse mount options\n");
394 /* temporarily use utf8 to correctly find the hidden dir below */
396 sbi->nls = load_nls("utf8");
398 printk(KERN_ERR "hfs: unable to load nls for utf8\n");
402 /* Grab the volume header */
403 if (hfsplus_read_wrapper(sb)) {
405 printk(KERN_WARNING "hfs: unable to find HFS+ superblock\n");
410 /* Copy parts of the volume header into the superblock */
411 sb->s_magic = HFSPLUS_VOLHEAD_SIG;
412 if (be16_to_cpu(vhdr->version) < HFSPLUS_MIN_VERSION ||
413 be16_to_cpu(vhdr->version) > HFSPLUS_CURRENT_VERSION) {
414 printk(KERN_ERR "hfs: wrong filesystem version\n");
417 sbi->total_blocks = be32_to_cpu(vhdr->total_blocks);
418 sbi->free_blocks = be32_to_cpu(vhdr->free_blocks);
419 sbi->next_cnid = be32_to_cpu(vhdr->next_cnid);
420 sbi->file_count = be32_to_cpu(vhdr->file_count);
421 sbi->folder_count = be32_to_cpu(vhdr->folder_count);
422 sbi->data_clump_blocks =
423 be32_to_cpu(vhdr->data_clump_sz) >> sbi->alloc_blksz_shift;
424 if (!sbi->data_clump_blocks)
425 sbi->data_clump_blocks = 1;
426 sbi->rsrc_clump_blocks =
427 be32_to_cpu(vhdr->rsrc_clump_sz) >> sbi->alloc_blksz_shift;
428 if (!sbi->rsrc_clump_blocks)
429 sbi->rsrc_clump_blocks = 1;
432 last_fs_block = sbi->total_blocks - 1;
433 last_fs_page = (last_fs_block << sbi->alloc_blksz_shift) >>
436 if ((last_fs_block > (sector_t)(~0ULL) >> (sbi->alloc_blksz_shift - 9)) ||
437 (last_fs_page > (pgoff_t)(~0ULL))) {
438 printk(KERN_ERR "hfs: filesystem size too large.\n");
442 /* Set up operations so we can load metadata */
443 sb->s_op = &hfsplus_sops;
444 sb->s_maxbytes = MAX_LFS_FILESIZE;
446 if (!(vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_UNMNT))) {
447 printk(KERN_WARNING "hfs: Filesystem was "
448 "not cleanly unmounted, "
449 "running fsck.hfsplus is recommended. "
450 "mounting read-only.\n");
451 sb->s_flags |= MS_RDONLY;
452 } else if (test_and_clear_bit(HFSPLUS_SB_FORCE, &sbi->flags)) {
454 } else if (vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_SOFTLOCK)) {
455 printk(KERN_WARNING "hfs: Filesystem is marked locked, mounting read-only.\n");
456 sb->s_flags |= MS_RDONLY;
457 } else if ((vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_JOURNALED)) &&
458 !(sb->s_flags & MS_RDONLY)) {
459 printk(KERN_WARNING "hfs: write access to "
460 "a journaled filesystem is not supported, "
461 "use the force option at your own risk, "
462 "mounting read-only.\n");
463 sb->s_flags |= MS_RDONLY;
468 /* Load metadata objects (B*Trees) */
469 sbi->ext_tree = hfs_btree_open(sb, HFSPLUS_EXT_CNID);
470 if (!sbi->ext_tree) {
471 printk(KERN_ERR "hfs: failed to load extents file\n");
474 sbi->cat_tree = hfs_btree_open(sb, HFSPLUS_CAT_CNID);
475 if (!sbi->cat_tree) {
476 printk(KERN_ERR "hfs: failed to load catalog file\n");
477 goto out_close_ext_tree;
479 if (vhdr->attr_file.total_blocks != 0) {
480 sbi->attr_tree = hfs_btree_open(sb, HFSPLUS_ATTR_CNID);
481 if (!sbi->attr_tree) {
482 printk(KERN_ERR "hfs: failed to load attributes file\n");
483 goto out_close_cat_tree;
487 inode = hfsplus_iget(sb, HFSPLUS_ALLOC_CNID);
489 printk(KERN_ERR "hfs: failed to load allocation file\n");
490 err = PTR_ERR(inode);
491 goto out_close_attr_tree;
493 sbi->alloc_file = inode;
495 /* Load the root directory */
496 root = hfsplus_iget(sb, HFSPLUS_ROOT_CNID);
498 printk(KERN_ERR "hfs: failed to load root directory\n");
500 goto out_put_alloc_file;
503 sb->s_d_op = &hfsplus_dentry_operations;
504 sb->s_root = d_make_root(root);
507 goto out_put_alloc_file;
510 str.len = sizeof(HFSP_HIDDENDIR_NAME) - 1;
511 str.name = HFSP_HIDDENDIR_NAME;
512 err = hfs_find_init(sbi->cat_tree, &fd);
515 hfsplus_cat_build_key(sb, fd.search_key, HFSPLUS_ROOT_CNID, &str);
516 if (!hfs_brec_read(&fd, &entry, sizeof(entry))) {
518 if (entry.type != cpu_to_be16(HFSPLUS_FOLDER))
520 inode = hfsplus_iget(sb, be32_to_cpu(entry.folder.id));
522 err = PTR_ERR(inode);
525 sbi->hidden_dir = inode;
529 if (!(sb->s_flags & MS_RDONLY)) {
531 * H+LX == hfsplusutils, H+Lx == this driver, H+lx is unused
532 * all three are registered with Apple for our use
534 vhdr->last_mount_vers = cpu_to_be32(HFSP_MOUNT_VERSION);
535 vhdr->modify_date = hfsp_now2mt();
536 be32_add_cpu(&vhdr->write_count, 1);
537 vhdr->attributes &= cpu_to_be32(~HFSPLUS_VOL_UNMNT);
538 vhdr->attributes |= cpu_to_be32(HFSPLUS_VOL_INCNSTNT);
539 hfsplus_sync_fs(sb, 1);
541 if (!sbi->hidden_dir) {
542 mutex_lock(&sbi->vh_mutex);
543 sbi->hidden_dir = hfsplus_new_inode(sb, S_IFDIR);
544 if (!sbi->hidden_dir) {
545 mutex_unlock(&sbi->vh_mutex);
549 err = hfsplus_create_cat(sbi->hidden_dir->i_ino, root,
550 &str, sbi->hidden_dir);
551 mutex_unlock(&sbi->vh_mutex);
553 goto out_put_hidden_dir;
555 hfsplus_mark_inode_dirty(sbi->hidden_dir,
556 HFSPLUS_I_CAT_DIRTY);
560 unload_nls(sbi->nls);
565 iput(sbi->hidden_dir);
570 iput(sbi->alloc_file);
572 hfs_btree_close(sbi->attr_tree);
574 hfs_btree_close(sbi->cat_tree);
576 hfs_btree_close(sbi->ext_tree);
578 kfree(sbi->s_vhdr_buf);
579 kfree(sbi->s_backup_vhdr_buf);
581 unload_nls(sbi->nls);
588 MODULE_AUTHOR("Brad Boyer");
589 MODULE_DESCRIPTION("Extended Macintosh Filesystem");
590 MODULE_LICENSE("GPL");
592 static struct kmem_cache *hfsplus_inode_cachep;
594 static struct inode *hfsplus_alloc_inode(struct super_block *sb)
596 struct hfsplus_inode_info *i;
598 i = kmem_cache_alloc(hfsplus_inode_cachep, GFP_KERNEL);
599 return i ? &i->vfs_inode : NULL;
602 static void hfsplus_i_callback(struct rcu_head *head)
604 struct inode *inode = container_of(head, struct inode, i_rcu);
606 kmem_cache_free(hfsplus_inode_cachep, HFSPLUS_I(inode));
609 static void hfsplus_destroy_inode(struct inode *inode)
611 call_rcu(&inode->i_rcu, hfsplus_i_callback);
614 #define HFSPLUS_INODE_SIZE sizeof(struct hfsplus_inode_info)
616 static struct dentry *hfsplus_mount(struct file_system_type *fs_type,
617 int flags, const char *dev_name, void *data)
619 return mount_bdev(fs_type, flags, dev_name, data, hfsplus_fill_super);
622 static struct file_system_type hfsplus_fs_type = {
623 .owner = THIS_MODULE,
625 .mount = hfsplus_mount,
626 .kill_sb = kill_block_super,
627 .fs_flags = FS_REQUIRES_DEV,
630 static void hfsplus_init_once(void *p)
632 struct hfsplus_inode_info *i = p;
634 inode_init_once(&i->vfs_inode);
637 static int __init init_hfsplus_fs(void)
641 hfsplus_inode_cachep = kmem_cache_create("hfsplus_icache",
642 HFSPLUS_INODE_SIZE, 0, SLAB_HWCACHE_ALIGN,
644 if (!hfsplus_inode_cachep)
646 err = hfsplus_create_attr_tree_cache();
648 goto destroy_inode_cache;
649 err = register_filesystem(&hfsplus_fs_type);
651 goto destroy_attr_tree_cache;
654 destroy_attr_tree_cache:
655 hfsplus_destroy_attr_tree_cache();
658 kmem_cache_destroy(hfsplus_inode_cachep);
663 static void __exit exit_hfsplus_fs(void)
665 unregister_filesystem(&hfsplus_fs_type);
666 hfsplus_destroy_attr_tree_cache();
669 * Make sure all delayed rcu free inodes are flushed before we
673 kmem_cache_destroy(hfsplus_inode_cachep);
676 module_init(init_hfsplus_fs)
677 module_exit(exit_hfsplus_fs)