2 * JFFS2 -- Journalling Flash File System, Version 2.
4 * Copyright © 2001-2007 Red Hat, Inc.
5 * Copyright © 2004-2010 David Woodhouse <dwmw2@infradead.org>
7 * Created by David Woodhouse <dwmw2@infradead.org>
9 * For licensing information, see the file 'LICENCE' in this directory.
13 #include <linux/capability.h>
14 #include <linux/kernel.h>
15 #include <linux/sched.h>
17 #include <linux/list.h>
18 #include <linux/mtd/mtd.h>
19 #include <linux/pagemap.h>
20 #include <linux/slab.h>
21 #include <linux/vmalloc.h>
22 #include <linux/vfs.h>
23 #include <linux/crc32.h>
26 static int jffs2_flash_setup(struct jffs2_sb_info *c);
28 int jffs2_do_setattr (struct inode *inode, struct iattr *iattr)
30 struct jffs2_full_dnode *old_metadata, *new_metadata;
31 struct jffs2_inode_info *f = JFFS2_INODE_INFO(inode);
32 struct jffs2_sb_info *c = JFFS2_SB_INFO(inode->i_sb);
33 struct jffs2_raw_inode *ri;
34 union jffs2_device_node dev;
35 unsigned char *mdata = NULL;
40 int alloc_type = ALLOC_NORMAL;
42 jffs2_dbg(1, "%s(): ino #%lu\n", __func__, inode->i_ino);
44 /* Special cases - we don't want more than one data node
45 for these types on the medium at any time. So setattr
46 must read the original data associated with the node
47 (i.e. the device numbers or the target name) and write
48 it out again with the appropriate data attached */
49 if (S_ISBLK(inode->i_mode) || S_ISCHR(inode->i_mode)) {
50 /* For these, we don't actually need to read the old node */
51 mdatalen = jffs2_encode_dev(&dev, inode->i_rdev);
53 jffs2_dbg(1, "%s(): Writing %d bytes of kdev_t\n",
55 } else if (S_ISLNK(inode->i_mode)) {
57 mdatalen = f->metadata->size;
58 mdata = kmalloc(f->metadata->size, GFP_USER);
60 mutex_unlock(&f->sem);
63 ret = jffs2_read_dnode(c, f, f->metadata, mdata, 0, mdatalen);
65 mutex_unlock(&f->sem);
69 mutex_unlock(&f->sem);
70 jffs2_dbg(1, "%s(): Writing %d bytes of symlink target\n",
74 ri = jffs2_alloc_raw_inode();
76 if (S_ISLNK(inode->i_mode))
81 ret = jffs2_reserve_space(c, sizeof(*ri) + mdatalen, &alloclen,
82 ALLOC_NORMAL, JFFS2_SUMMARY_INODE_SIZE);
84 jffs2_free_raw_inode(ri);
85 if (S_ISLNK(inode->i_mode))
90 ivalid = iattr->ia_valid;
92 ri->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
93 ri->nodetype = cpu_to_je16(JFFS2_NODETYPE_INODE);
94 ri->totlen = cpu_to_je32(sizeof(*ri) + mdatalen);
95 ri->hdr_crc = cpu_to_je32(crc32(0, ri, sizeof(struct jffs2_unknown_node)-4));
97 ri->ino = cpu_to_je32(inode->i_ino);
98 ri->version = cpu_to_je32(++f->highest_version);
100 ri->uid = cpu_to_je16((ivalid & ATTR_UID)?iattr->ia_uid:inode->i_uid);
101 ri->gid = cpu_to_je16((ivalid & ATTR_GID)?iattr->ia_gid:inode->i_gid);
103 if (ivalid & ATTR_MODE)
104 ri->mode = cpu_to_jemode(iattr->ia_mode);
106 ri->mode = cpu_to_jemode(inode->i_mode);
109 ri->isize = cpu_to_je32((ivalid & ATTR_SIZE)?iattr->ia_size:inode->i_size);
110 ri->atime = cpu_to_je32(I_SEC((ivalid & ATTR_ATIME)?iattr->ia_atime:inode->i_atime));
111 ri->mtime = cpu_to_je32(I_SEC((ivalid & ATTR_MTIME)?iattr->ia_mtime:inode->i_mtime));
112 ri->ctime = cpu_to_je32(I_SEC((ivalid & ATTR_CTIME)?iattr->ia_ctime:inode->i_ctime));
114 ri->offset = cpu_to_je32(0);
115 ri->csize = ri->dsize = cpu_to_je32(mdatalen);
116 ri->compr = JFFS2_COMPR_NONE;
117 if (ivalid & ATTR_SIZE && inode->i_size < iattr->ia_size) {
118 /* It's an extension. Make it a hole node */
119 ri->compr = JFFS2_COMPR_ZERO;
120 ri->dsize = cpu_to_je32(iattr->ia_size - inode->i_size);
121 ri->offset = cpu_to_je32(inode->i_size);
122 } else if (ivalid & ATTR_SIZE && !iattr->ia_size) {
123 /* For truncate-to-zero, treat it as deletion because
124 it'll always be obsoleting all previous nodes */
125 alloc_type = ALLOC_DELETION;
127 ri->node_crc = cpu_to_je32(crc32(0, ri, sizeof(*ri)-8));
129 ri->data_crc = cpu_to_je32(crc32(0, mdata, mdatalen));
131 ri->data_crc = cpu_to_je32(0);
133 new_metadata = jffs2_write_dnode(c, f, ri, mdata, mdatalen, alloc_type);
134 if (S_ISLNK(inode->i_mode))
137 if (IS_ERR(new_metadata)) {
138 jffs2_complete_reservation(c);
139 jffs2_free_raw_inode(ri);
140 mutex_unlock(&f->sem);
141 return PTR_ERR(new_metadata);
143 /* It worked. Update the inode */
144 inode->i_atime = ITIME(je32_to_cpu(ri->atime));
145 inode->i_ctime = ITIME(je32_to_cpu(ri->ctime));
146 inode->i_mtime = ITIME(je32_to_cpu(ri->mtime));
147 inode->i_mode = jemode_to_cpu(ri->mode);
148 inode->i_uid = je16_to_cpu(ri->uid);
149 inode->i_gid = je16_to_cpu(ri->gid);
152 old_metadata = f->metadata;
154 if (ivalid & ATTR_SIZE && inode->i_size > iattr->ia_size)
155 jffs2_truncate_fragtree (c, &f->fragtree, iattr->ia_size);
157 if (ivalid & ATTR_SIZE && inode->i_size < iattr->ia_size) {
158 jffs2_add_full_dnode_to_inode(c, f, new_metadata);
159 inode->i_size = iattr->ia_size;
160 inode->i_blocks = (inode->i_size + 511) >> 9;
163 f->metadata = new_metadata;
166 jffs2_mark_node_obsolete(c, old_metadata->raw);
167 jffs2_free_full_dnode(old_metadata);
169 jffs2_free_raw_inode(ri);
171 mutex_unlock(&f->sem);
172 jffs2_complete_reservation(c);
174 /* We have to do the truncate_setsize() without f->sem held, since
175 some pages may be locked and waiting for it in readpage().
176 We are protected from a simultaneous write() extending i_size
177 back past iattr->ia_size, because do_truncate() holds the
178 generic inode semaphore. */
179 if (ivalid & ATTR_SIZE && inode->i_size > iattr->ia_size) {
180 truncate_setsize(inode, iattr->ia_size);
181 inode->i_blocks = (inode->i_size + 511) >> 9;
187 int jffs2_setattr(struct dentry *dentry, struct iattr *iattr)
191 rc = inode_change_ok(dentry->d_inode, iattr);
195 rc = jffs2_do_setattr(dentry->d_inode, iattr);
196 if (!rc && (iattr->ia_valid & ATTR_MODE))
197 rc = jffs2_acl_chmod(dentry->d_inode);
202 int jffs2_statfs(struct dentry *dentry, struct kstatfs *buf)
204 struct jffs2_sb_info *c = JFFS2_SB_INFO(dentry->d_sb);
207 buf->f_type = JFFS2_SUPER_MAGIC;
208 buf->f_bsize = 1 << PAGE_SHIFT;
209 buf->f_blocks = c->flash_size >> PAGE_SHIFT;
212 buf->f_namelen = JFFS2_MAX_NAME_LEN;
213 buf->f_fsid.val[0] = JFFS2_SUPER_MAGIC;
214 buf->f_fsid.val[1] = c->mtd->index;
216 spin_lock(&c->erase_completion_lock);
217 avail = c->dirty_size + c->free_size;
218 if (avail > c->sector_size * c->resv_blocks_write)
219 avail -= c->sector_size * c->resv_blocks_write;
222 spin_unlock(&c->erase_completion_lock);
224 buf->f_bavail = buf->f_bfree = avail >> PAGE_SHIFT;
230 void jffs2_evict_inode (struct inode *inode)
232 /* We can forget about this inode for now - drop all
233 * the nodelists associated with it, etc.
235 struct jffs2_sb_info *c = JFFS2_SB_INFO(inode->i_sb);
236 struct jffs2_inode_info *f = JFFS2_INODE_INFO(inode);
238 jffs2_dbg(1, "%s(): ino #%lu mode %o\n",
239 __func__, inode->i_ino, inode->i_mode);
240 truncate_inode_pages(&inode->i_data, 0);
241 end_writeback(inode);
242 jffs2_do_clear_inode(c, f);
245 struct inode *jffs2_iget(struct super_block *sb, unsigned long ino)
247 struct jffs2_inode_info *f;
248 struct jffs2_sb_info *c;
249 struct jffs2_raw_inode latest_node;
250 union jffs2_device_node jdev;
255 jffs2_dbg(1, "%s(): ino == %lu\n", __func__, ino);
257 inode = iget_locked(sb, ino);
259 return ERR_PTR(-ENOMEM);
260 if (!(inode->i_state & I_NEW))
263 f = JFFS2_INODE_INFO(inode);
264 c = JFFS2_SB_INFO(inode->i_sb);
266 jffs2_init_inode_info(f);
269 ret = jffs2_do_read_inode(c, f, inode->i_ino, &latest_node);
272 mutex_unlock(&f->sem);
276 inode->i_mode = jemode_to_cpu(latest_node.mode);
277 inode->i_uid = je16_to_cpu(latest_node.uid);
278 inode->i_gid = je16_to_cpu(latest_node.gid);
279 inode->i_size = je32_to_cpu(latest_node.isize);
280 inode->i_atime = ITIME(je32_to_cpu(latest_node.atime));
281 inode->i_mtime = ITIME(je32_to_cpu(latest_node.mtime));
282 inode->i_ctime = ITIME(je32_to_cpu(latest_node.ctime));
284 set_nlink(inode, f->inocache->pino_nlink);
286 inode->i_blocks = (inode->i_size + 511) >> 9;
288 switch (inode->i_mode & S_IFMT) {
291 inode->i_op = &jffs2_symlink_inode_operations;
296 struct jffs2_full_dirent *fd;
297 set_nlink(inode, 2); /* parent and '.' */
299 for (fd=f->dents; fd; fd = fd->next) {
300 if (fd->type == DT_DIR && fd->ino)
303 /* Root dir gets i_nlink 3 for some reason */
304 if (inode->i_ino == 1)
307 inode->i_op = &jffs2_dir_inode_operations;
308 inode->i_fop = &jffs2_dir_operations;
312 inode->i_op = &jffs2_file_inode_operations;
313 inode->i_fop = &jffs2_file_operations;
314 inode->i_mapping->a_ops = &jffs2_file_address_operations;
315 inode->i_mapping->nrpages = 0;
320 /* Read the device numbers from the media */
321 if (f->metadata->size != sizeof(jdev.old_id) &&
322 f->metadata->size != sizeof(jdev.new_id)) {
323 printk(KERN_NOTICE "Device node has strange size %d\n", f->metadata->size);
326 jffs2_dbg(1, "Reading device numbers from flash\n");
327 ret = jffs2_read_dnode(c, f, f->metadata, (char *)&jdev, 0, f->metadata->size);
330 printk(KERN_NOTICE "Read device numbers for inode %lu failed\n", (unsigned long)inode->i_ino);
333 if (f->metadata->size == sizeof(jdev.old_id))
334 rdev = old_decode_dev(je16_to_cpu(jdev.old_id));
336 rdev = new_decode_dev(je32_to_cpu(jdev.new_id));
340 inode->i_op = &jffs2_file_inode_operations;
341 init_special_inode(inode, inode->i_mode, rdev);
345 printk(KERN_WARNING "jffs2_read_inode(): Bogus imode %o for ino %lu\n", inode->i_mode, (unsigned long)inode->i_ino);
348 mutex_unlock(&f->sem);
350 jffs2_dbg(1, "jffs2_read_inode() returning\n");
351 unlock_new_inode(inode);
357 mutex_unlock(&f->sem);
358 jffs2_do_clear_inode(c, f);
363 void jffs2_dirty_inode(struct inode *inode, int flags)
367 if (!(inode->i_state & I_DIRTY_DATASYNC)) {
368 jffs2_dbg(2, "%s(): not calling setattr() for ino #%lu\n",
369 __func__, inode->i_ino);
373 jffs2_dbg(1, "%s(): calling setattr() for ino #%lu\n",
374 __func__, inode->i_ino);
376 iattr.ia_valid = ATTR_MODE|ATTR_UID|ATTR_GID|ATTR_ATIME|ATTR_MTIME|ATTR_CTIME;
377 iattr.ia_mode = inode->i_mode;
378 iattr.ia_uid = inode->i_uid;
379 iattr.ia_gid = inode->i_gid;
380 iattr.ia_atime = inode->i_atime;
381 iattr.ia_mtime = inode->i_mtime;
382 iattr.ia_ctime = inode->i_ctime;
384 jffs2_do_setattr(inode, &iattr);
387 int jffs2_do_remount_fs(struct super_block *sb, int *flags, char *data)
389 struct jffs2_sb_info *c = JFFS2_SB_INFO(sb);
391 if (c->flags & JFFS2_SB_FLAG_RO && !(sb->s_flags & MS_RDONLY))
394 /* We stop if it was running, then restart if it needs to.
395 This also catches the case where it was stopped and this
396 is just a remount to restart it.
397 Flush the writebuffer, if neccecary, else we loose it */
398 if (!(sb->s_flags & MS_RDONLY)) {
399 jffs2_stop_garbage_collect_thread(c);
400 mutex_lock(&c->alloc_sem);
401 jffs2_flush_wbuf_pad(c);
402 mutex_unlock(&c->alloc_sem);
405 if (!(*flags & MS_RDONLY))
406 jffs2_start_garbage_collect_thread(c);
408 *flags |= MS_NOATIME;
412 /* jffs2_new_inode: allocate a new inode and inocache, add it to the hash,
413 fill in the raw_inode while you're at it. */
414 struct inode *jffs2_new_inode (struct inode *dir_i, umode_t mode, struct jffs2_raw_inode *ri)
417 struct super_block *sb = dir_i->i_sb;
418 struct jffs2_sb_info *c;
419 struct jffs2_inode_info *f;
422 jffs2_dbg(1, "%s(): dir_i %ld, mode 0x%x\n",
423 __func__, dir_i->i_ino, mode);
425 c = JFFS2_SB_INFO(sb);
427 inode = new_inode(sb);
430 return ERR_PTR(-ENOMEM);
432 f = JFFS2_INODE_INFO(inode);
433 jffs2_init_inode_info(f);
436 memset(ri, 0, sizeof(*ri));
437 /* Set OS-specific defaults for new inodes */
438 ri->uid = cpu_to_je16(current_fsuid());
440 if (dir_i->i_mode & S_ISGID) {
441 ri->gid = cpu_to_je16(dir_i->i_gid);
445 ri->gid = cpu_to_je16(current_fsgid());
448 /* POSIX ACLs have to be processed now, at least partly.
449 The umask is only applied if there's no default ACL */
450 ret = jffs2_init_acl_pre(dir_i, inode, &mode);
452 make_bad_inode(inode);
456 ret = jffs2_do_new_inode (c, f, mode, ri);
458 make_bad_inode(inode);
463 inode->i_ino = je32_to_cpu(ri->ino);
464 inode->i_mode = jemode_to_cpu(ri->mode);
465 inode->i_gid = je16_to_cpu(ri->gid);
466 inode->i_uid = je16_to_cpu(ri->uid);
467 inode->i_atime = inode->i_ctime = inode->i_mtime = CURRENT_TIME_SEC;
468 ri->atime = ri->mtime = ri->ctime = cpu_to_je32(I_SEC(inode->i_mtime));
473 if (insert_inode_locked(inode) < 0) {
474 make_bad_inode(inode);
476 return ERR_PTR(-EINVAL);
482 static int calculate_inocache_hashsize(uint32_t flash_size)
485 * Pick a inocache hash size based on the size of the medium.
486 * Count how many megabytes we're dealing with, apply a hashsize twice
487 * that size, but rounding down to the usual big powers of 2. And keep
488 * to sensible bounds.
491 int size_mb = flash_size / 1024 / 1024;
492 int hashsize = (size_mb * 2) & ~0x3f;
494 if (hashsize < INOCACHE_HASHSIZE_MIN)
495 return INOCACHE_HASHSIZE_MIN;
496 if (hashsize > INOCACHE_HASHSIZE_MAX)
497 return INOCACHE_HASHSIZE_MAX;
502 int jffs2_do_fill_super(struct super_block *sb, void *data, int silent)
504 struct jffs2_sb_info *c;
505 struct inode *root_i;
509 c = JFFS2_SB_INFO(sb);
511 #ifndef CONFIG_JFFS2_FS_WRITEBUFFER
512 if (c->mtd->type == MTD_NANDFLASH) {
513 printk(KERN_ERR "jffs2: Cannot operate on NAND flash unless jffs2 NAND support is compiled in.\n");
516 if (c->mtd->type == MTD_DATAFLASH) {
517 printk(KERN_ERR "jffs2: Cannot operate on DataFlash unless jffs2 DataFlash support is compiled in.\n");
522 c->flash_size = c->mtd->size;
523 c->sector_size = c->mtd->erasesize;
524 blocks = c->flash_size / c->sector_size;
527 * Size alignment check
529 if ((c->sector_size * blocks) != c->flash_size) {
530 c->flash_size = c->sector_size * blocks;
531 printk(KERN_INFO "jffs2: Flash size not aligned to erasesize, reducing to %dKiB\n",
532 c->flash_size / 1024);
535 if (c->flash_size < 5*c->sector_size) {
536 printk(KERN_ERR "jffs2: Too few erase blocks (%d)\n", c->flash_size / c->sector_size);
540 c->cleanmarker_size = sizeof(struct jffs2_unknown_node);
542 /* NAND (or other bizarre) flash... do setup accordingly */
543 ret = jffs2_flash_setup(c);
547 c->inocache_hashsize = calculate_inocache_hashsize(c->flash_size);
548 c->inocache_list = kcalloc(c->inocache_hashsize, sizeof(struct jffs2_inode_cache *), GFP_KERNEL);
549 if (!c->inocache_list) {
554 jffs2_init_xattr_subsystem(c);
556 if ((ret = jffs2_do_mount_fs(c)))
559 jffs2_dbg(1, "%s(): Getting root inode\n", __func__);
560 root_i = jffs2_iget(sb, 1);
561 if (IS_ERR(root_i)) {
562 jffs2_dbg(1, "get root inode failed\n");
563 ret = PTR_ERR(root_i);
569 jffs2_dbg(1, "%s(): d_alloc_root()\n", __func__);
570 sb->s_root = d_alloc_root(root_i);
574 sb->s_maxbytes = 0xFFFFFFFF;
575 sb->s_blocksize = PAGE_CACHE_SIZE;
576 sb->s_blocksize_bits = PAGE_CACHE_SHIFT;
577 sb->s_magic = JFFS2_SUPER_MAGIC;
578 if (!(sb->s_flags & MS_RDONLY))
579 jffs2_start_garbage_collect_thread(c);
585 jffs2_free_ino_caches(c);
586 jffs2_free_raw_node_refs(c);
587 if (jffs2_blocks_use_vmalloc(c))
592 jffs2_clear_xattr_subsystem(c);
593 kfree(c->inocache_list);
595 jffs2_flash_cleanup(c);
600 void jffs2_gc_release_inode(struct jffs2_sb_info *c,
601 struct jffs2_inode_info *f)
603 iput(OFNI_EDONI_2SFFJ(f));
606 struct jffs2_inode_info *jffs2_gc_fetch_inode(struct jffs2_sb_info *c,
607 int inum, int unlinked)
610 struct jffs2_inode_cache *ic;
613 /* The inode has zero nlink but its nodes weren't yet marked
614 obsolete. This has to be because we're still waiting for
615 the final (close() and) iput() to happen.
617 There's a possibility that the final iput() could have
618 happened while we were contemplating. In order to ensure
619 that we don't cause a new read_inode() (which would fail)
620 for the inode in question, we use ilookup() in this case
623 The nlink can't _become_ zero at this point because we're
624 holding the alloc_sem, and jffs2_do_unlink() would also
625 need that while decrementing nlink on any inode.
627 inode = ilookup(OFNI_BS_2SFFJ(c), inum);
629 jffs2_dbg(1, "ilookup() failed for ino #%u; inode is probably deleted.\n",
632 spin_lock(&c->inocache_lock);
633 ic = jffs2_get_ino_cache(c, inum);
635 jffs2_dbg(1, "Inode cache for ino #%u is gone\n",
637 spin_unlock(&c->inocache_lock);
640 if (ic->state != INO_STATE_CHECKEDABSENT) {
641 /* Wait for progress. Don't just loop */
642 jffs2_dbg(1, "Waiting for ino #%u in state %d\n",
644 sleep_on_spinunlock(&c->inocache_wq, &c->inocache_lock);
646 spin_unlock(&c->inocache_lock);
652 /* Inode has links to it still; they're not going away because
653 jffs2_do_unlink() would need the alloc_sem and we have it.
654 Just iget() it, and if read_inode() is necessary that's OK.
656 inode = jffs2_iget(OFNI_BS_2SFFJ(c), inum);
658 return ERR_CAST(inode);
660 if (is_bad_inode(inode)) {
661 printk(KERN_NOTICE "Eep. read_inode() failed for ino #%u. unlinked %d\n",
663 /* NB. This will happen again. We need to do something appropriate here. */
665 return ERR_PTR(-EIO);
668 return JFFS2_INODE_INFO(inode);
671 unsigned char *jffs2_gc_fetch_page(struct jffs2_sb_info *c,
672 struct jffs2_inode_info *f,
673 unsigned long offset,
676 struct inode *inode = OFNI_EDONI_2SFFJ(f);
679 pg = read_cache_page_async(inode->i_mapping, offset >> PAGE_CACHE_SHIFT,
680 (void *)jffs2_do_readpage_unlock, inode);
684 *priv = (unsigned long)pg;
688 void jffs2_gc_release_page(struct jffs2_sb_info *c,
692 struct page *pg = (void *)*priv;
695 page_cache_release(pg);
698 static int jffs2_flash_setup(struct jffs2_sb_info *c) {
701 if (jffs2_cleanmarker_oob(c)) {
702 /* NAND flash... do setup accordingly */
703 ret = jffs2_nand_flash_setup(c);
709 if (jffs2_dataflash(c)) {
710 ret = jffs2_dataflash_setup(c);
715 /* and Intel "Sibley" flash */
716 if (jffs2_nor_wbuf_flash(c)) {
717 ret = jffs2_nor_wbuf_flash_setup(c);
722 /* and an UBI volume */
723 if (jffs2_ubivol(c)) {
724 ret = jffs2_ubivol_setup(c);
732 void jffs2_flash_cleanup(struct jffs2_sb_info *c) {
734 if (jffs2_cleanmarker_oob(c)) {
735 jffs2_nand_flash_cleanup(c);
739 if (jffs2_dataflash(c)) {
740 jffs2_dataflash_cleanup(c);
743 /* and Intel "Sibley" flash */
744 if (jffs2_nor_wbuf_flash(c)) {
745 jffs2_nor_wbuf_flash_cleanup(c);
748 /* and an UBI volume */
749 if (jffs2_ubivol(c)) {
750 jffs2_ubivol_cleanup(c);