2 * Copyright (C) 2007 Oracle. All rights reserved.
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public
6 * License v2 as published by the Free Software Foundation.
8 * This program is distributed in the hope that it will be useful,
9 * but WITHOUT ANY WARRANTY; without even the implied warranty of
10 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
11 * General Public License for more details.
13 * You should have received a copy of the GNU General Public
14 * License along with this program; if not, write to the
15 * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
16 * Boston, MA 021110-1307, USA.
19 #include <linux/bio.h>
20 #include <linux/slab.h>
21 #include <linux/pagemap.h>
22 #include <linux/highmem.h>
25 #include "transaction.h"
27 #include "print-tree.h"
28 #include "compression.h"
30 #define __MAX_CSUM_ITEMS(r, size) \
31 ((unsigned long)(((BTRFS_MAX_ITEM_SIZE(r) * 2) / size) - 1))
33 #define MAX_CSUM_ITEMS(r, size) (min_t(u32, __MAX_CSUM_ITEMS(r, size), \
36 #define MAX_ORDERED_SUM_BYTES(r) ((PAGE_SIZE - \
37 sizeof(struct btrfs_ordered_sum)) / \
38 sizeof(u32) * (r)->sectorsize)
40 int btrfs_insert_file_extent(struct btrfs_trans_handle *trans,
41 struct btrfs_root *root,
42 u64 objectid, u64 pos,
43 u64 disk_offset, u64 disk_num_bytes,
44 u64 num_bytes, u64 offset, u64 ram_bytes,
45 u8 compression, u8 encryption, u16 other_encoding)
48 struct btrfs_file_extent_item *item;
49 struct btrfs_key file_key;
50 struct btrfs_path *path;
51 struct extent_buffer *leaf;
53 path = btrfs_alloc_path();
56 file_key.objectid = objectid;
57 file_key.offset = pos;
58 file_key.type = BTRFS_EXTENT_DATA_KEY;
60 path->leave_spinning = 1;
61 ret = btrfs_insert_empty_item(trans, root, path, &file_key,
65 BUG_ON(ret); /* Can't happen */
66 leaf = path->nodes[0];
67 item = btrfs_item_ptr(leaf, path->slots[0],
68 struct btrfs_file_extent_item);
69 btrfs_set_file_extent_disk_bytenr(leaf, item, disk_offset);
70 btrfs_set_file_extent_disk_num_bytes(leaf, item, disk_num_bytes);
71 btrfs_set_file_extent_offset(leaf, item, offset);
72 btrfs_set_file_extent_num_bytes(leaf, item, num_bytes);
73 btrfs_set_file_extent_ram_bytes(leaf, item, ram_bytes);
74 btrfs_set_file_extent_generation(leaf, item, trans->transid);
75 btrfs_set_file_extent_type(leaf, item, BTRFS_FILE_EXTENT_REG);
76 btrfs_set_file_extent_compression(leaf, item, compression);
77 btrfs_set_file_extent_encryption(leaf, item, encryption);
78 btrfs_set_file_extent_other_encoding(leaf, item, other_encoding);
80 btrfs_mark_buffer_dirty(leaf);
82 btrfs_free_path(path);
86 static struct btrfs_csum_item *
87 btrfs_lookup_csum(struct btrfs_trans_handle *trans,
88 struct btrfs_root *root,
89 struct btrfs_path *path,
93 struct btrfs_key file_key;
94 struct btrfs_key found_key;
95 struct btrfs_csum_item *item;
96 struct extent_buffer *leaf;
98 u16 csum_size = btrfs_super_csum_size(root->fs_info->super_copy);
101 file_key.objectid = BTRFS_EXTENT_CSUM_OBJECTID;
102 file_key.offset = bytenr;
103 file_key.type = BTRFS_EXTENT_CSUM_KEY;
104 ret = btrfs_search_slot(trans, root, &file_key, path, 0, cow);
107 leaf = path->nodes[0];
110 if (path->slots[0] == 0)
113 btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]);
114 if (found_key.type != BTRFS_EXTENT_CSUM_KEY)
117 csum_offset = (bytenr - found_key.offset) >>
118 root->fs_info->sb->s_blocksize_bits;
119 csums_in_item = btrfs_item_size_nr(leaf, path->slots[0]);
120 csums_in_item /= csum_size;
122 if (csum_offset == csums_in_item) {
125 } else if (csum_offset > csums_in_item) {
129 item = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_csum_item);
130 item = (struct btrfs_csum_item *)((unsigned char *)item +
131 csum_offset * csum_size);
139 int btrfs_lookup_file_extent(struct btrfs_trans_handle *trans,
140 struct btrfs_root *root,
141 struct btrfs_path *path, u64 objectid,
145 struct btrfs_key file_key;
146 int ins_len = mod < 0 ? -1 : 0;
149 file_key.objectid = objectid;
150 file_key.offset = offset;
151 file_key.type = BTRFS_EXTENT_DATA_KEY;
152 ret = btrfs_search_slot(trans, root, &file_key, path, ins_len, cow);
156 static void btrfs_io_bio_endio_readpage(struct btrfs_io_bio *bio, int err)
158 kfree(bio->csum_allocated);
161 static int __btrfs_lookup_bio_sums(struct btrfs_root *root,
162 struct inode *inode, struct bio *bio,
163 u64 logical_offset, u32 *dst, int dio)
165 struct bio_vec *bvec = bio->bi_io_vec;
166 struct btrfs_io_bio *btrfs_bio = btrfs_io_bio(bio);
167 struct btrfs_csum_item *item = NULL;
168 struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree;
169 struct btrfs_path *path;
172 u64 item_start_offset = 0;
173 u64 item_last_offset = 0;
180 u16 csum_size = btrfs_super_csum_size(root->fs_info->super_copy);
182 path = btrfs_alloc_path();
186 nblocks = bio->bi_iter.bi_size >> inode->i_sb->s_blocksize_bits;
188 if (nblocks * csum_size > BTRFS_BIO_INLINE_CSUM_SIZE) {
189 btrfs_bio->csum_allocated = kmalloc_array(nblocks,
190 csum_size, GFP_NOFS);
191 if (!btrfs_bio->csum_allocated) {
192 btrfs_free_path(path);
195 btrfs_bio->csum = btrfs_bio->csum_allocated;
196 btrfs_bio->end_io = btrfs_io_bio_endio_readpage;
198 btrfs_bio->csum = btrfs_bio->csum_inline;
200 csum = btrfs_bio->csum;
205 if (bio->bi_iter.bi_size > PAGE_SIZE * 8)
206 path->reada = READA_FORWARD;
208 WARN_ON(bio->bi_vcnt <= 0);
211 * the free space stuff is only read when it hasn't been
212 * updated in the current transaction. So, we can safely
213 * read from the commit root and sidestep a nasty deadlock
214 * between reading the free space cache and updating the csum tree.
216 if (btrfs_is_free_space_inode(inode)) {
217 path->search_commit_root = 1;
218 path->skip_locking = 1;
221 disk_bytenr = (u64)bio->bi_iter.bi_sector << 9;
223 offset = logical_offset;
225 page_bytes_left = bvec->bv_len;
226 while (bio_index < bio->bi_vcnt) {
228 offset = page_offset(bvec->bv_page) + bvec->bv_offset;
229 count = btrfs_find_ordered_sum(inode, offset, disk_bytenr,
230 (u32 *)csum, nblocks);
234 if (!item || disk_bytenr < item_start_offset ||
235 disk_bytenr >= item_last_offset) {
236 struct btrfs_key found_key;
240 btrfs_release_path(path);
241 item = btrfs_lookup_csum(NULL, root->fs_info->csum_root,
242 path, disk_bytenr, 0);
245 memset(csum, 0, csum_size);
246 if (BTRFS_I(inode)->root->root_key.objectid ==
247 BTRFS_DATA_RELOC_TREE_OBJECTID) {
248 set_extent_bits(io_tree, offset,
249 offset + root->sectorsize - 1,
252 btrfs_info_rl(BTRFS_I(inode)->root->fs_info,
253 "no csum found for inode %llu start %llu",
254 btrfs_ino(inode), offset);
257 btrfs_release_path(path);
260 btrfs_item_key_to_cpu(path->nodes[0], &found_key,
263 item_start_offset = found_key.offset;
264 item_size = btrfs_item_size_nr(path->nodes[0],
266 item_last_offset = item_start_offset +
267 (item_size / csum_size) *
269 item = btrfs_item_ptr(path->nodes[0], path->slots[0],
270 struct btrfs_csum_item);
273 * this byte range must be able to fit inside
274 * a single leaf so it will also fit inside a u32
276 diff = disk_bytenr - item_start_offset;
277 diff = diff / root->sectorsize;
278 diff = diff * csum_size;
279 count = min_t(int, nblocks, (item_last_offset - disk_bytenr) >>
280 inode->i_sb->s_blocksize_bits);
281 read_extent_buffer(path->nodes[0], csum,
282 ((unsigned long)item) + diff,
285 csum += count * csum_size;
289 disk_bytenr += root->sectorsize;
290 offset += root->sectorsize;
291 page_bytes_left -= root->sectorsize;
292 if (!page_bytes_left) {
295 * make sure we're still inside the
296 * bio before we update page_bytes_left
298 if (bio_index >= bio->bi_vcnt) {
303 page_bytes_left = bvec->bv_len;
310 btrfs_free_path(path);
314 int btrfs_lookup_bio_sums(struct btrfs_root *root, struct inode *inode,
315 struct bio *bio, u32 *dst)
317 return __btrfs_lookup_bio_sums(root, inode, bio, 0, dst, 0);
320 int btrfs_lookup_bio_sums_dio(struct btrfs_root *root, struct inode *inode,
321 struct bio *bio, u64 offset)
323 return __btrfs_lookup_bio_sums(root, inode, bio, offset, NULL, 1);
326 int btrfs_lookup_csums_range(struct btrfs_root *root, u64 start, u64 end,
327 struct list_head *list, int search_commit)
329 struct btrfs_key key;
330 struct btrfs_path *path;
331 struct extent_buffer *leaf;
332 struct btrfs_ordered_sum *sums;
333 struct btrfs_csum_item *item;
335 unsigned long offset;
339 u16 csum_size = btrfs_super_csum_size(root->fs_info->super_copy);
341 ASSERT(IS_ALIGNED(start, root->sectorsize) &&
342 IS_ALIGNED(end + 1, root->sectorsize));
344 path = btrfs_alloc_path();
349 path->skip_locking = 1;
350 path->reada = READA_FORWARD;
351 path->search_commit_root = 1;
354 key.objectid = BTRFS_EXTENT_CSUM_OBJECTID;
356 key.type = BTRFS_EXTENT_CSUM_KEY;
358 ret = btrfs_search_slot(NULL, root, &key, path, 0, 0);
361 if (ret > 0 && path->slots[0] > 0) {
362 leaf = path->nodes[0];
363 btrfs_item_key_to_cpu(leaf, &key, path->slots[0] - 1);
364 if (key.objectid == BTRFS_EXTENT_CSUM_OBJECTID &&
365 key.type == BTRFS_EXTENT_CSUM_KEY) {
366 offset = (start - key.offset) >>
367 root->fs_info->sb->s_blocksize_bits;
368 if (offset * csum_size <
369 btrfs_item_size_nr(leaf, path->slots[0] - 1))
374 while (start <= end) {
375 leaf = path->nodes[0];
376 if (path->slots[0] >= btrfs_header_nritems(leaf)) {
377 ret = btrfs_next_leaf(root, path);
382 leaf = path->nodes[0];
385 btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
386 if (key.objectid != BTRFS_EXTENT_CSUM_OBJECTID ||
387 key.type != BTRFS_EXTENT_CSUM_KEY ||
391 if (key.offset > start)
394 size = btrfs_item_size_nr(leaf, path->slots[0]);
395 csum_end = key.offset + (size / csum_size) * root->sectorsize;
396 if (csum_end <= start) {
401 csum_end = min(csum_end, end + 1);
402 item = btrfs_item_ptr(path->nodes[0], path->slots[0],
403 struct btrfs_csum_item);
404 while (start < csum_end) {
405 size = min_t(size_t, csum_end - start,
406 MAX_ORDERED_SUM_BYTES(root));
407 sums = kzalloc(btrfs_ordered_sum_size(root, size),
414 sums->bytenr = start;
415 sums->len = (int)size;
417 offset = (start - key.offset) >>
418 root->fs_info->sb->s_blocksize_bits;
420 size >>= root->fs_info->sb->s_blocksize_bits;
422 read_extent_buffer(path->nodes[0],
424 ((unsigned long)item) + offset,
427 start += root->sectorsize * size;
428 list_add_tail(&sums->list, &tmplist);
434 while (ret < 0 && !list_empty(&tmplist)) {
435 sums = list_entry(tmplist.next, struct btrfs_ordered_sum, list);
436 list_del(&sums->list);
439 list_splice_tail(&tmplist, list);
441 btrfs_free_path(path);
445 int btrfs_csum_one_bio(struct btrfs_root *root, struct inode *inode,
446 struct bio *bio, u64 file_start, int contig)
448 struct btrfs_ordered_sum *sums;
449 struct btrfs_ordered_extent *ordered;
451 struct bio_vec *bvec = bio->bi_io_vec;
456 unsigned long total_bytes = 0;
457 unsigned long this_sum_bytes = 0;
460 WARN_ON(bio->bi_vcnt <= 0);
461 sums = kzalloc(btrfs_ordered_sum_size(root, bio->bi_iter.bi_size),
466 sums->len = bio->bi_iter.bi_size;
467 INIT_LIST_HEAD(&sums->list);
472 offset = page_offset(bvec->bv_page) + bvec->bv_offset;
474 ordered = btrfs_lookup_ordered_extent(inode, offset);
475 BUG_ON(!ordered); /* Logic error */
476 sums->bytenr = (u64)bio->bi_iter.bi_sector << 9;
479 while (bio_index < bio->bi_vcnt) {
481 offset = page_offset(bvec->bv_page) + bvec->bv_offset;
483 data = kmap_atomic(bvec->bv_page);
485 nr_sectors = BTRFS_BYTES_TO_BLKS(root->fs_info,
486 bvec->bv_len + root->sectorsize
489 for (i = 0; i < nr_sectors; i++) {
490 if (offset >= ordered->file_offset + ordered->len ||
491 offset < ordered->file_offset) {
492 unsigned long bytes_left;
495 sums->len = this_sum_bytes;
497 btrfs_add_ordered_sum(inode, ordered, sums);
498 btrfs_put_ordered_extent(ordered);
500 bytes_left = bio->bi_iter.bi_size - total_bytes;
502 sums = kzalloc(btrfs_ordered_sum_size(root, bytes_left),
504 BUG_ON(!sums); /* -ENOMEM */
505 sums->len = bytes_left;
506 ordered = btrfs_lookup_ordered_extent(inode,
508 ASSERT(ordered); /* Logic error */
509 sums->bytenr = ((u64)bio->bi_iter.bi_sector << 9)
513 data = kmap_atomic(bvec->bv_page);
516 sums->sums[index] = ~(u32)0;
518 = btrfs_csum_data(data + bvec->bv_offset
519 + (i * root->sectorsize),
522 btrfs_csum_final(sums->sums[index],
523 (char *)(sums->sums + index));
525 offset += root->sectorsize;
526 this_sum_bytes += root->sectorsize;
527 total_bytes += root->sectorsize;
536 btrfs_add_ordered_sum(inode, ordered, sums);
537 btrfs_put_ordered_extent(ordered);
542 * helper function for csum removal, this expects the
543 * key to describe the csum pointed to by the path, and it expects
544 * the csum to overlap the range [bytenr, len]
546 * The csum should not be entirely contained in the range and the
547 * range should not be entirely contained in the csum.
549 * This calls btrfs_truncate_item with the correct args based on the
550 * overlap, and fixes up the key as required.
552 static noinline void truncate_one_csum(struct btrfs_root *root,
553 struct btrfs_path *path,
554 struct btrfs_key *key,
557 struct extent_buffer *leaf;
558 u16 csum_size = btrfs_super_csum_size(root->fs_info->super_copy);
560 u64 end_byte = bytenr + len;
561 u32 blocksize_bits = root->fs_info->sb->s_blocksize_bits;
563 leaf = path->nodes[0];
564 csum_end = btrfs_item_size_nr(leaf, path->slots[0]) / csum_size;
565 csum_end <<= root->fs_info->sb->s_blocksize_bits;
566 csum_end += key->offset;
568 if (key->offset < bytenr && csum_end <= end_byte) {
573 * A simple truncate off the end of the item
575 u32 new_size = (bytenr - key->offset) >> blocksize_bits;
576 new_size *= csum_size;
577 btrfs_truncate_item(root, path, new_size, 1);
578 } else if (key->offset >= bytenr && csum_end > end_byte &&
579 end_byte > key->offset) {
584 * we need to truncate from the beginning of the csum
586 u32 new_size = (csum_end - end_byte) >> blocksize_bits;
587 new_size *= csum_size;
589 btrfs_truncate_item(root, path, new_size, 0);
591 key->offset = end_byte;
592 btrfs_set_item_key_safe(root->fs_info, path, key);
599 * deletes the csum items from the csum tree for a given
602 int btrfs_del_csums(struct btrfs_trans_handle *trans,
603 struct btrfs_root *root, u64 bytenr, u64 len)
605 struct btrfs_path *path;
606 struct btrfs_key key;
607 u64 end_byte = bytenr + len;
609 struct extent_buffer *leaf;
611 u16 csum_size = btrfs_super_csum_size(root->fs_info->super_copy);
612 int blocksize_bits = root->fs_info->sb->s_blocksize_bits;
614 root = root->fs_info->csum_root;
616 path = btrfs_alloc_path();
621 key.objectid = BTRFS_EXTENT_CSUM_OBJECTID;
622 key.offset = end_byte - 1;
623 key.type = BTRFS_EXTENT_CSUM_KEY;
625 path->leave_spinning = 1;
626 ret = btrfs_search_slot(trans, root, &key, path, -1, 1);
628 if (path->slots[0] == 0)
631 } else if (ret < 0) {
635 leaf = path->nodes[0];
636 btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
638 if (key.objectid != BTRFS_EXTENT_CSUM_OBJECTID ||
639 key.type != BTRFS_EXTENT_CSUM_KEY) {
643 if (key.offset >= end_byte)
646 csum_end = btrfs_item_size_nr(leaf, path->slots[0]) / csum_size;
647 csum_end <<= blocksize_bits;
648 csum_end += key.offset;
650 /* this csum ends before we start, we're done */
651 if (csum_end <= bytenr)
654 /* delete the entire item, it is inside our range */
655 if (key.offset >= bytenr && csum_end <= end_byte) {
656 ret = btrfs_del_item(trans, root, path);
659 if (key.offset == bytenr)
661 } else if (key.offset < bytenr && csum_end > end_byte) {
662 unsigned long offset;
663 unsigned long shift_len;
664 unsigned long item_offset;
669 * Our bytes are in the middle of the csum,
670 * we need to split this item and insert a new one.
672 * But we can't drop the path because the
673 * csum could change, get removed, extended etc.
675 * The trick here is the max size of a csum item leaves
676 * enough room in the tree block for a single
677 * item header. So, we split the item in place,
678 * adding a new header pointing to the existing
679 * bytes. Then we loop around again and we have
680 * a nicely formed csum item that we can neatly
683 offset = (bytenr - key.offset) >> blocksize_bits;
686 shift_len = (len >> blocksize_bits) * csum_size;
688 item_offset = btrfs_item_ptr_offset(leaf,
691 memset_extent_buffer(leaf, 0, item_offset + offset,
696 * btrfs_split_item returns -EAGAIN when the
697 * item changed size or key
699 ret = btrfs_split_item(trans, root, path, &key, offset);
700 if (ret && ret != -EAGAIN) {
701 btrfs_abort_transaction(trans, root, ret);
705 key.offset = end_byte - 1;
707 truncate_one_csum(root, path, &key, bytenr, len);
708 if (key.offset < bytenr)
711 btrfs_release_path(path);
715 btrfs_free_path(path);
719 int btrfs_csum_file_blocks(struct btrfs_trans_handle *trans,
720 struct btrfs_root *root,
721 struct btrfs_ordered_sum *sums)
723 struct btrfs_key file_key;
724 struct btrfs_key found_key;
725 struct btrfs_path *path;
726 struct btrfs_csum_item *item;
727 struct btrfs_csum_item *item_end;
728 struct extent_buffer *leaf = NULL;
738 u16 csum_size = btrfs_super_csum_size(root->fs_info->super_copy);
740 path = btrfs_alloc_path();
744 next_offset = (u64)-1;
746 bytenr = sums->bytenr + total_bytes;
747 file_key.objectid = BTRFS_EXTENT_CSUM_OBJECTID;
748 file_key.offset = bytenr;
749 file_key.type = BTRFS_EXTENT_CSUM_KEY;
751 item = btrfs_lookup_csum(trans, root, path, bytenr, 1);
754 leaf = path->nodes[0];
755 item_end = btrfs_item_ptr(leaf, path->slots[0],
756 struct btrfs_csum_item);
757 item_end = (struct btrfs_csum_item *)((char *)item_end +
758 btrfs_item_size_nr(leaf, path->slots[0]));
762 if (ret != -EFBIG && ret != -ENOENT)
767 /* we found one, but it isn't big enough yet */
768 leaf = path->nodes[0];
769 item_size = btrfs_item_size_nr(leaf, path->slots[0]);
770 if ((item_size / csum_size) >=
771 MAX_CSUM_ITEMS(root, csum_size)) {
772 /* already at max size, make a new one */
776 int slot = path->slots[0] + 1;
777 /* we didn't find a csum item, insert one */
778 nritems = btrfs_header_nritems(path->nodes[0]);
779 if (!nritems || (path->slots[0] >= nritems - 1)) {
780 ret = btrfs_next_leaf(root, path);
785 slot = path->slots[0];
787 btrfs_item_key_to_cpu(path->nodes[0], &found_key, slot);
788 if (found_key.objectid != BTRFS_EXTENT_CSUM_OBJECTID ||
789 found_key.type != BTRFS_EXTENT_CSUM_KEY) {
793 next_offset = found_key.offset;
799 * at this point, we know the tree has an item, but it isn't big
800 * enough yet to put our csum in. Grow it
802 btrfs_release_path(path);
803 ret = btrfs_search_slot(trans, root, &file_key, path,
809 if (path->slots[0] == 0)
814 leaf = path->nodes[0];
815 btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]);
816 csum_offset = (bytenr - found_key.offset) >>
817 root->fs_info->sb->s_blocksize_bits;
819 if (found_key.type != BTRFS_EXTENT_CSUM_KEY ||
820 found_key.objectid != BTRFS_EXTENT_CSUM_OBJECTID ||
821 csum_offset >= MAX_CSUM_ITEMS(root, csum_size)) {
825 if (csum_offset == btrfs_item_size_nr(leaf, path->slots[0]) /
832 if (btrfs_leaf_free_space(root, leaf) <
833 sizeof(struct btrfs_item) + csum_size * 2)
836 free_space = btrfs_leaf_free_space(root, leaf) -
837 sizeof(struct btrfs_item) - csum_size;
838 tmp = sums->len - total_bytes;
839 tmp >>= root->fs_info->sb->s_blocksize_bits;
842 extend_nr = max_t(int, 1, (int)tmp);
843 diff = (csum_offset + extend_nr) * csum_size;
844 diff = min(diff, MAX_CSUM_ITEMS(root, csum_size) * csum_size);
846 diff = diff - btrfs_item_size_nr(leaf, path->slots[0]);
847 diff = min(free_space, diff);
851 btrfs_extend_item(root, path, diff);
857 btrfs_release_path(path);
862 tmp = sums->len - total_bytes;
863 tmp >>= root->fs_info->sb->s_blocksize_bits;
864 tmp = min(tmp, (next_offset - file_key.offset) >>
865 root->fs_info->sb->s_blocksize_bits);
867 tmp = max((u64)1, tmp);
868 tmp = min(tmp, (u64)MAX_CSUM_ITEMS(root, csum_size));
869 ins_size = csum_size * tmp;
871 ins_size = csum_size;
873 path->leave_spinning = 1;
874 ret = btrfs_insert_empty_item(trans, root, path, &file_key,
876 path->leave_spinning = 0;
879 if (WARN_ON(ret != 0))
881 leaf = path->nodes[0];
883 item = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_csum_item);
884 item_end = (struct btrfs_csum_item *)((unsigned char *)item +
885 btrfs_item_size_nr(leaf, path->slots[0]));
886 item = (struct btrfs_csum_item *)((unsigned char *)item +
887 csum_offset * csum_size);
889 ins_size = (u32)(sums->len - total_bytes) >>
890 root->fs_info->sb->s_blocksize_bits;
891 ins_size *= csum_size;
892 ins_size = min_t(u32, (unsigned long)item_end - (unsigned long)item,
894 write_extent_buffer(leaf, sums->sums + index, (unsigned long)item,
897 ins_size /= csum_size;
898 total_bytes += ins_size * root->sectorsize;
901 btrfs_mark_buffer_dirty(path->nodes[0]);
902 if (total_bytes < sums->len) {
903 btrfs_release_path(path);
908 btrfs_free_path(path);
915 void btrfs_extent_item_to_extent_map(struct inode *inode,
916 const struct btrfs_path *path,
917 struct btrfs_file_extent_item *fi,
918 const bool new_inline,
919 struct extent_map *em)
921 struct btrfs_root *root = BTRFS_I(inode)->root;
922 struct extent_buffer *leaf = path->nodes[0];
923 const int slot = path->slots[0];
924 struct btrfs_key key;
925 u64 extent_start, extent_end;
927 u8 type = btrfs_file_extent_type(leaf, fi);
928 int compress_type = btrfs_file_extent_compression(leaf, fi);
930 em->bdev = root->fs_info->fs_devices->latest_bdev;
931 btrfs_item_key_to_cpu(leaf, &key, slot);
932 extent_start = key.offset;
934 if (type == BTRFS_FILE_EXTENT_REG ||
935 type == BTRFS_FILE_EXTENT_PREALLOC) {
936 extent_end = extent_start +
937 btrfs_file_extent_num_bytes(leaf, fi);
938 } else if (type == BTRFS_FILE_EXTENT_INLINE) {
940 size = btrfs_file_extent_inline_len(leaf, slot, fi);
941 extent_end = ALIGN(extent_start + size, root->sectorsize);
944 em->ram_bytes = btrfs_file_extent_ram_bytes(leaf, fi);
945 if (type == BTRFS_FILE_EXTENT_REG ||
946 type == BTRFS_FILE_EXTENT_PREALLOC) {
947 em->start = extent_start;
948 em->len = extent_end - extent_start;
949 em->orig_start = extent_start -
950 btrfs_file_extent_offset(leaf, fi);
951 em->orig_block_len = btrfs_file_extent_disk_num_bytes(leaf, fi);
952 bytenr = btrfs_file_extent_disk_bytenr(leaf, fi);
954 em->block_start = EXTENT_MAP_HOLE;
957 if (compress_type != BTRFS_COMPRESS_NONE) {
958 set_bit(EXTENT_FLAG_COMPRESSED, &em->flags);
959 em->compress_type = compress_type;
960 em->block_start = bytenr;
961 em->block_len = em->orig_block_len;
963 bytenr += btrfs_file_extent_offset(leaf, fi);
964 em->block_start = bytenr;
965 em->block_len = em->len;
966 if (type == BTRFS_FILE_EXTENT_PREALLOC)
967 set_bit(EXTENT_FLAG_PREALLOC, &em->flags);
969 } else if (type == BTRFS_FILE_EXTENT_INLINE) {
970 em->block_start = EXTENT_MAP_INLINE;
971 em->start = extent_start;
972 em->len = extent_end - extent_start;
974 * Initialize orig_start and block_len with the same values
975 * as in inode.c:btrfs_get_extent().
977 em->orig_start = EXTENT_MAP_HOLE;
978 em->block_len = (u64)-1;
979 if (!new_inline && compress_type != BTRFS_COMPRESS_NONE) {
980 set_bit(EXTENT_FLAG_COMPRESSED, &em->flags);
981 em->compress_type = compress_type;
984 btrfs_err(root->fs_info,
985 "unknown file extent item type %d, inode %llu, offset %llu, root %llu",
986 type, btrfs_ino(inode), extent_start,
987 root->root_key.objectid);