2 * btree.c - NILFS B-tree.
4 * Copyright (C) 2005-2008 Nippon Telegraph and Telephone Corporation.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 * Written by Koji Sato <koji@osrg.net>.
23 #include <linux/slab.h>
24 #include <linux/string.h>
25 #include <linux/errno.h>
26 #include <linux/pagevec.h>
34 static void __nilfs_btree_init(struct nilfs_bmap *bmap);
36 static struct nilfs_btree_path *nilfs_btree_alloc_path(void)
38 struct nilfs_btree_path *path;
39 int level = NILFS_BTREE_LEVEL_DATA;
41 path = kmem_cache_alloc(nilfs_btree_path_cache, GFP_NOFS);
45 for (; level < NILFS_BTREE_LEVEL_MAX; level++) {
46 path[level].bp_bh = NULL;
47 path[level].bp_sib_bh = NULL;
48 path[level].bp_index = 0;
49 path[level].bp_oldreq.bpr_ptr = NILFS_BMAP_INVALID_PTR;
50 path[level].bp_newreq.bpr_ptr = NILFS_BMAP_INVALID_PTR;
51 path[level].bp_op = NULL;
58 static void nilfs_btree_free_path(struct nilfs_btree_path *path)
60 int level = NILFS_BTREE_LEVEL_DATA;
62 for (; level < NILFS_BTREE_LEVEL_MAX; level++)
63 brelse(path[level].bp_bh);
65 kmem_cache_free(nilfs_btree_path_cache, path);
69 * B-tree node operations
71 static int nilfs_btree_get_new_block(const struct nilfs_bmap *btree,
72 __u64 ptr, struct buffer_head **bhp)
74 struct address_space *btnc = &NILFS_BMAP_I(btree)->i_btnode_cache;
75 struct buffer_head *bh;
77 bh = nilfs_btnode_create_block(btnc, ptr);
81 set_buffer_nilfs_volatile(bh);
86 static int nilfs_btree_node_get_flags(const struct nilfs_btree_node *node)
88 return node->bn_flags;
92 nilfs_btree_node_set_flags(struct nilfs_btree_node *node, int flags)
94 node->bn_flags = flags;
97 static int nilfs_btree_node_root(const struct nilfs_btree_node *node)
99 return nilfs_btree_node_get_flags(node) & NILFS_BTREE_NODE_ROOT;
102 static int nilfs_btree_node_get_level(const struct nilfs_btree_node *node)
104 return node->bn_level;
108 nilfs_btree_node_set_level(struct nilfs_btree_node *node, int level)
110 node->bn_level = level;
113 static int nilfs_btree_node_get_nchildren(const struct nilfs_btree_node *node)
115 return le16_to_cpu(node->bn_nchildren);
119 nilfs_btree_node_set_nchildren(struct nilfs_btree_node *node, int nchildren)
121 node->bn_nchildren = cpu_to_le16(nchildren);
124 static int nilfs_btree_node_size(const struct nilfs_bmap *btree)
126 return 1 << btree->b_inode->i_blkbits;
129 static int nilfs_btree_nchildren_per_block(const struct nilfs_bmap *btree)
131 return btree->b_nchildren_per_block;
135 nilfs_btree_node_dkeys(const struct nilfs_btree_node *node)
137 return (__le64 *)((char *)(node + 1) +
138 (nilfs_btree_node_root(node) ?
139 0 : NILFS_BTREE_NODE_EXTRA_PAD_SIZE));
143 nilfs_btree_node_dptrs(const struct nilfs_btree_node *node, int ncmax)
145 return (__le64 *)(nilfs_btree_node_dkeys(node) + ncmax);
149 nilfs_btree_node_get_key(const struct nilfs_btree_node *node, int index)
151 return le64_to_cpu(*(nilfs_btree_node_dkeys(node) + index));
155 nilfs_btree_node_set_key(struct nilfs_btree_node *node, int index, __u64 key)
157 *(nilfs_btree_node_dkeys(node) + index) = cpu_to_le64(key);
161 nilfs_btree_node_get_ptr(const struct nilfs_btree_node *node, int index,
164 return le64_to_cpu(*(nilfs_btree_node_dptrs(node, ncmax) + index));
168 nilfs_btree_node_set_ptr(struct nilfs_btree_node *node, int index, __u64 ptr,
171 *(nilfs_btree_node_dptrs(node, ncmax) + index) = cpu_to_le64(ptr);
174 static void nilfs_btree_node_init(struct nilfs_btree_node *node, int flags,
175 int level, int nchildren, int ncmax,
176 const __u64 *keys, const __u64 *ptrs)
182 nilfs_btree_node_set_flags(node, flags);
183 nilfs_btree_node_set_level(node, level);
184 nilfs_btree_node_set_nchildren(node, nchildren);
186 dkeys = nilfs_btree_node_dkeys(node);
187 dptrs = nilfs_btree_node_dptrs(node, ncmax);
188 for (i = 0; i < nchildren; i++) {
189 dkeys[i] = cpu_to_le64(keys[i]);
190 dptrs[i] = cpu_to_le64(ptrs[i]);
194 /* Assume the buffer heads corresponding to left and right are locked. */
195 static void nilfs_btree_node_move_left(struct nilfs_btree_node *left,
196 struct nilfs_btree_node *right,
197 int n, int lncmax, int rncmax)
199 __le64 *ldkeys, *rdkeys;
200 __le64 *ldptrs, *rdptrs;
201 int lnchildren, rnchildren;
203 ldkeys = nilfs_btree_node_dkeys(left);
204 ldptrs = nilfs_btree_node_dptrs(left, lncmax);
205 lnchildren = nilfs_btree_node_get_nchildren(left);
207 rdkeys = nilfs_btree_node_dkeys(right);
208 rdptrs = nilfs_btree_node_dptrs(right, rncmax);
209 rnchildren = nilfs_btree_node_get_nchildren(right);
211 memcpy(ldkeys + lnchildren, rdkeys, n * sizeof(*rdkeys));
212 memcpy(ldptrs + lnchildren, rdptrs, n * sizeof(*rdptrs));
213 memmove(rdkeys, rdkeys + n, (rnchildren - n) * sizeof(*rdkeys));
214 memmove(rdptrs, rdptrs + n, (rnchildren - n) * sizeof(*rdptrs));
218 nilfs_btree_node_set_nchildren(left, lnchildren);
219 nilfs_btree_node_set_nchildren(right, rnchildren);
222 /* Assume that the buffer heads corresponding to left and right are locked. */
223 static void nilfs_btree_node_move_right(struct nilfs_btree_node *left,
224 struct nilfs_btree_node *right,
225 int n, int lncmax, int rncmax)
227 __le64 *ldkeys, *rdkeys;
228 __le64 *ldptrs, *rdptrs;
229 int lnchildren, rnchildren;
231 ldkeys = nilfs_btree_node_dkeys(left);
232 ldptrs = nilfs_btree_node_dptrs(left, lncmax);
233 lnchildren = nilfs_btree_node_get_nchildren(left);
235 rdkeys = nilfs_btree_node_dkeys(right);
236 rdptrs = nilfs_btree_node_dptrs(right, rncmax);
237 rnchildren = nilfs_btree_node_get_nchildren(right);
239 memmove(rdkeys + n, rdkeys, rnchildren * sizeof(*rdkeys));
240 memmove(rdptrs + n, rdptrs, rnchildren * sizeof(*rdptrs));
241 memcpy(rdkeys, ldkeys + lnchildren - n, n * sizeof(*rdkeys));
242 memcpy(rdptrs, ldptrs + lnchildren - n, n * sizeof(*rdptrs));
246 nilfs_btree_node_set_nchildren(left, lnchildren);
247 nilfs_btree_node_set_nchildren(right, rnchildren);
250 /* Assume that the buffer head corresponding to node is locked. */
251 static void nilfs_btree_node_insert(struct nilfs_btree_node *node, int index,
252 __u64 key, __u64 ptr, int ncmax)
258 dkeys = nilfs_btree_node_dkeys(node);
259 dptrs = nilfs_btree_node_dptrs(node, ncmax);
260 nchildren = nilfs_btree_node_get_nchildren(node);
261 if (index < nchildren) {
262 memmove(dkeys + index + 1, dkeys + index,
263 (nchildren - index) * sizeof(*dkeys));
264 memmove(dptrs + index + 1, dptrs + index,
265 (nchildren - index) * sizeof(*dptrs));
267 dkeys[index] = cpu_to_le64(key);
268 dptrs[index] = cpu_to_le64(ptr);
270 nilfs_btree_node_set_nchildren(node, nchildren);
273 /* Assume that the buffer head corresponding to node is locked. */
274 static void nilfs_btree_node_delete(struct nilfs_btree_node *node, int index,
275 __u64 *keyp, __u64 *ptrp, int ncmax)
283 dkeys = nilfs_btree_node_dkeys(node);
284 dptrs = nilfs_btree_node_dptrs(node, ncmax);
285 key = le64_to_cpu(dkeys[index]);
286 ptr = le64_to_cpu(dptrs[index]);
287 nchildren = nilfs_btree_node_get_nchildren(node);
293 if (index < nchildren - 1) {
294 memmove(dkeys + index, dkeys + index + 1,
295 (nchildren - index - 1) * sizeof(*dkeys));
296 memmove(dptrs + index, dptrs + index + 1,
297 (nchildren - index - 1) * sizeof(*dptrs));
300 nilfs_btree_node_set_nchildren(node, nchildren);
303 static int nilfs_btree_node_lookup(const struct nilfs_btree_node *node,
304 __u64 key, int *indexp)
307 int index, low, high, s;
311 high = nilfs_btree_node_get_nchildren(node) - 1;
314 while (low <= high) {
315 index = (low + high) / 2;
316 nkey = nilfs_btree_node_get_key(node, index);
320 } else if (nkey < key) {
330 if (nilfs_btree_node_get_level(node) > NILFS_BTREE_LEVEL_NODE_MIN) {
331 if (s > 0 && index > 0)
343 * nilfs_btree_node_broken - verify consistency of btree node
344 * @node: btree node block to be examined
345 * @size: node size (in bytes)
346 * @blocknr: block number
348 * Return Value: If node is broken, 1 is returned. Otherwise, 0 is returned.
350 static int nilfs_btree_node_broken(const struct nilfs_btree_node *node,
351 size_t size, sector_t blocknr)
353 int level, flags, nchildren;
356 level = nilfs_btree_node_get_level(node);
357 flags = nilfs_btree_node_get_flags(node);
358 nchildren = nilfs_btree_node_get_nchildren(node);
360 if (unlikely(level < NILFS_BTREE_LEVEL_NODE_MIN ||
361 level >= NILFS_BTREE_LEVEL_MAX ||
362 (flags & NILFS_BTREE_NODE_ROOT) ||
364 nchildren > NILFS_BTREE_NODE_NCHILDREN_MAX(size))) {
365 printk(KERN_CRIT "NILFS: bad btree node (blocknr=%llu): "
366 "level = %d, flags = 0x%x, nchildren = %d\n",
367 (unsigned long long)blocknr, level, flags, nchildren);
374 * nilfs_btree_root_broken - verify consistency of btree root node
375 * @node: btree root node to be examined
378 * Return Value: If node is broken, 1 is returned. Otherwise, 0 is returned.
380 static int nilfs_btree_root_broken(const struct nilfs_btree_node *node,
383 int level, flags, nchildren;
386 level = nilfs_btree_node_get_level(node);
387 flags = nilfs_btree_node_get_flags(node);
388 nchildren = nilfs_btree_node_get_nchildren(node);
390 if (unlikely(level < NILFS_BTREE_LEVEL_NODE_MIN ||
391 level > NILFS_BTREE_LEVEL_MAX ||
393 nchildren > NILFS_BTREE_ROOT_NCHILDREN_MAX)) {
394 pr_crit("NILFS: bad btree root (inode number=%lu): level = %d, flags = 0x%x, nchildren = %d\n",
395 ino, level, flags, nchildren);
401 int nilfs_btree_broken_node_block(struct buffer_head *bh)
405 if (buffer_nilfs_checked(bh))
408 ret = nilfs_btree_node_broken((struct nilfs_btree_node *)bh->b_data,
409 bh->b_size, bh->b_blocknr);
411 set_buffer_nilfs_checked(bh);
415 static struct nilfs_btree_node *
416 nilfs_btree_get_root(const struct nilfs_bmap *btree)
418 return (struct nilfs_btree_node *)btree->b_u.u_data;
421 static struct nilfs_btree_node *
422 nilfs_btree_get_nonroot_node(const struct nilfs_btree_path *path, int level)
424 return (struct nilfs_btree_node *)path[level].bp_bh->b_data;
427 static struct nilfs_btree_node *
428 nilfs_btree_get_sib_node(const struct nilfs_btree_path *path, int level)
430 return (struct nilfs_btree_node *)path[level].bp_sib_bh->b_data;
433 static int nilfs_btree_height(const struct nilfs_bmap *btree)
435 return nilfs_btree_node_get_level(nilfs_btree_get_root(btree)) + 1;
438 static struct nilfs_btree_node *
439 nilfs_btree_get_node(const struct nilfs_bmap *btree,
440 const struct nilfs_btree_path *path,
441 int level, int *ncmaxp)
443 struct nilfs_btree_node *node;
445 if (level == nilfs_btree_height(btree) - 1) {
446 node = nilfs_btree_get_root(btree);
447 *ncmaxp = NILFS_BTREE_ROOT_NCHILDREN_MAX;
449 node = nilfs_btree_get_nonroot_node(path, level);
450 *ncmaxp = nilfs_btree_nchildren_per_block(btree);
456 nilfs_btree_bad_node(struct nilfs_btree_node *node, int level)
458 if (unlikely(nilfs_btree_node_get_level(node) != level)) {
460 printk(KERN_CRIT "NILFS: btree level mismatch: %d != %d\n",
461 nilfs_btree_node_get_level(node), level);
467 struct nilfs_btree_readahead_info {
468 struct nilfs_btree_node *node; /* parent node */
469 int max_ra_blocks; /* max nof blocks to read ahead */
470 int index; /* current index on the parent node */
471 int ncmax; /* nof children in the parent node */
474 static int __nilfs_btree_get_block(const struct nilfs_bmap *btree, __u64 ptr,
475 struct buffer_head **bhp,
476 const struct nilfs_btree_readahead_info *ra)
478 struct address_space *btnc = &NILFS_BMAP_I(btree)->i_btnode_cache;
479 struct buffer_head *bh, *ra_bh;
480 sector_t submit_ptr = 0;
483 ret = nilfs_btnode_submit_block(btnc, ptr, 0, READ, &bh, &submit_ptr);
494 /* read ahead sibling nodes */
495 for (n = ra->max_ra_blocks, i = ra->index + 1;
496 n > 0 && i < ra->ncmax; n--, i++) {
497 ptr2 = nilfs_btree_node_get_ptr(ra->node, i, ra->ncmax);
499 ret = nilfs_btnode_submit_block(btnc, ptr2, 0, READA,
500 &ra_bh, &submit_ptr);
501 if (likely(!ret || ret == -EEXIST))
503 else if (ret != -EBUSY)
505 if (!buffer_locked(bh))
513 if (!buffer_uptodate(bh)) {
519 if (nilfs_btree_broken_node_block(bh)) {
520 clear_buffer_uptodate(bh);
529 static int nilfs_btree_get_block(const struct nilfs_bmap *btree, __u64 ptr,
530 struct buffer_head **bhp)
532 return __nilfs_btree_get_block(btree, ptr, bhp, NULL);
535 static int nilfs_btree_do_lookup(const struct nilfs_bmap *btree,
536 struct nilfs_btree_path *path,
537 __u64 key, __u64 *ptrp, int minlevel,
540 struct nilfs_btree_node *node;
541 struct nilfs_btree_readahead_info p, *ra;
543 int level, index, found, ncmax, ret;
545 node = nilfs_btree_get_root(btree);
546 level = nilfs_btree_node_get_level(node);
547 if (level < minlevel || nilfs_btree_node_get_nchildren(node) <= 0)
550 found = nilfs_btree_node_lookup(node, key, &index);
551 ptr = nilfs_btree_node_get_ptr(node, index,
552 NILFS_BTREE_ROOT_NCHILDREN_MAX);
553 path[level].bp_bh = NULL;
554 path[level].bp_index = index;
556 ncmax = nilfs_btree_nchildren_per_block(btree);
558 while (--level >= minlevel) {
560 if (level == NILFS_BTREE_LEVEL_NODE_MIN && readahead) {
561 p.node = nilfs_btree_get_node(btree, path, level + 1,
567 ret = __nilfs_btree_get_block(btree, ptr, &path[level].bp_bh,
572 node = nilfs_btree_get_nonroot_node(path, level);
573 if (nilfs_btree_bad_node(node, level))
576 found = nilfs_btree_node_lookup(node, key, &index);
580 ptr = nilfs_btree_node_get_ptr(node, index, ncmax);
582 WARN_ON(found || level != NILFS_BTREE_LEVEL_NODE_MIN);
584 ptr = NILFS_BMAP_INVALID_PTR;
586 path[level].bp_index = index;
597 static int nilfs_btree_do_lookup_last(const struct nilfs_bmap *btree,
598 struct nilfs_btree_path *path,
599 __u64 *keyp, __u64 *ptrp)
601 struct nilfs_btree_node *node;
603 int index, level, ncmax, ret;
605 node = nilfs_btree_get_root(btree);
606 index = nilfs_btree_node_get_nchildren(node) - 1;
609 level = nilfs_btree_node_get_level(node);
610 ptr = nilfs_btree_node_get_ptr(node, index,
611 NILFS_BTREE_ROOT_NCHILDREN_MAX);
612 path[level].bp_bh = NULL;
613 path[level].bp_index = index;
614 ncmax = nilfs_btree_nchildren_per_block(btree);
616 for (level--; level > 0; level--) {
617 ret = nilfs_btree_get_block(btree, ptr, &path[level].bp_bh);
620 node = nilfs_btree_get_nonroot_node(path, level);
621 if (nilfs_btree_bad_node(node, level))
623 index = nilfs_btree_node_get_nchildren(node) - 1;
624 ptr = nilfs_btree_node_get_ptr(node, index, ncmax);
625 path[level].bp_index = index;
629 *keyp = nilfs_btree_node_get_key(node, index);
636 static int nilfs_btree_lookup(const struct nilfs_bmap *btree,
637 __u64 key, int level, __u64 *ptrp)
639 struct nilfs_btree_path *path;
642 path = nilfs_btree_alloc_path();
646 ret = nilfs_btree_do_lookup(btree, path, key, ptrp, level, 0);
648 nilfs_btree_free_path(path);
653 static int nilfs_btree_lookup_contig(const struct nilfs_bmap *btree,
654 __u64 key, __u64 *ptrp, unsigned maxblocks)
656 struct nilfs_btree_path *path;
657 struct nilfs_btree_node *node;
658 struct inode *dat = NULL;
661 int level = NILFS_BTREE_LEVEL_NODE_MIN;
662 int ret, cnt, index, maxlevel, ncmax;
663 struct nilfs_btree_readahead_info p;
665 path = nilfs_btree_alloc_path();
669 ret = nilfs_btree_do_lookup(btree, path, key, &ptr, level, 1);
673 if (NILFS_BMAP_USE_VBN(btree)) {
674 dat = nilfs_bmap_get_dat(btree);
675 ret = nilfs_dat_translate(dat, ptr, &blocknr);
681 if (cnt == maxblocks)
684 maxlevel = nilfs_btree_height(btree) - 1;
685 node = nilfs_btree_get_node(btree, path, level, &ncmax);
686 index = path[level].bp_index + 1;
688 while (index < nilfs_btree_node_get_nchildren(node)) {
689 if (nilfs_btree_node_get_key(node, index) !=
692 ptr2 = nilfs_btree_node_get_ptr(node, index, ncmax);
694 ret = nilfs_dat_translate(dat, ptr2, &blocknr);
699 if (ptr2 != ptr + cnt || ++cnt == maxblocks)
704 if (level == maxlevel)
707 /* look-up right sibling node */
708 p.node = nilfs_btree_get_node(btree, path, level + 1, &p.ncmax);
709 p.index = path[level + 1].bp_index + 1;
711 if (p.index >= nilfs_btree_node_get_nchildren(p.node) ||
712 nilfs_btree_node_get_key(p.node, p.index) != key + cnt)
714 ptr2 = nilfs_btree_node_get_ptr(p.node, p.index, p.ncmax);
715 path[level + 1].bp_index = p.index;
717 brelse(path[level].bp_bh);
718 path[level].bp_bh = NULL;
720 ret = __nilfs_btree_get_block(btree, ptr2, &path[level].bp_bh,
724 node = nilfs_btree_get_nonroot_node(path, level);
725 ncmax = nilfs_btree_nchildren_per_block(btree);
727 path[level].bp_index = index;
733 nilfs_btree_free_path(path);
737 static void nilfs_btree_promote_key(struct nilfs_bmap *btree,
738 struct nilfs_btree_path *path,
739 int level, __u64 key)
741 if (level < nilfs_btree_height(btree) - 1) {
743 nilfs_btree_node_set_key(
744 nilfs_btree_get_nonroot_node(path, level),
745 path[level].bp_index, key);
746 if (!buffer_dirty(path[level].bp_bh))
747 mark_buffer_dirty(path[level].bp_bh);
748 } while ((path[level].bp_index == 0) &&
749 (++level < nilfs_btree_height(btree) - 1));
753 if (level == nilfs_btree_height(btree) - 1) {
754 nilfs_btree_node_set_key(nilfs_btree_get_root(btree),
755 path[level].bp_index, key);
759 static void nilfs_btree_do_insert(struct nilfs_bmap *btree,
760 struct nilfs_btree_path *path,
761 int level, __u64 *keyp, __u64 *ptrp)
763 struct nilfs_btree_node *node;
766 if (level < nilfs_btree_height(btree) - 1) {
767 node = nilfs_btree_get_nonroot_node(path, level);
768 ncblk = nilfs_btree_nchildren_per_block(btree);
769 nilfs_btree_node_insert(node, path[level].bp_index,
770 *keyp, *ptrp, ncblk);
771 if (!buffer_dirty(path[level].bp_bh))
772 mark_buffer_dirty(path[level].bp_bh);
774 if (path[level].bp_index == 0)
775 nilfs_btree_promote_key(btree, path, level + 1,
776 nilfs_btree_node_get_key(node,
779 node = nilfs_btree_get_root(btree);
780 nilfs_btree_node_insert(node, path[level].bp_index,
782 NILFS_BTREE_ROOT_NCHILDREN_MAX);
786 static void nilfs_btree_carry_left(struct nilfs_bmap *btree,
787 struct nilfs_btree_path *path,
788 int level, __u64 *keyp, __u64 *ptrp)
790 struct nilfs_btree_node *node, *left;
791 int nchildren, lnchildren, n, move, ncblk;
793 node = nilfs_btree_get_nonroot_node(path, level);
794 left = nilfs_btree_get_sib_node(path, level);
795 nchildren = nilfs_btree_node_get_nchildren(node);
796 lnchildren = nilfs_btree_node_get_nchildren(left);
797 ncblk = nilfs_btree_nchildren_per_block(btree);
800 n = (nchildren + lnchildren + 1) / 2 - lnchildren;
801 if (n > path[level].bp_index) {
802 /* move insert point */
807 nilfs_btree_node_move_left(left, node, n, ncblk, ncblk);
809 if (!buffer_dirty(path[level].bp_bh))
810 mark_buffer_dirty(path[level].bp_bh);
811 if (!buffer_dirty(path[level].bp_sib_bh))
812 mark_buffer_dirty(path[level].bp_sib_bh);
814 nilfs_btree_promote_key(btree, path, level + 1,
815 nilfs_btree_node_get_key(node, 0));
818 brelse(path[level].bp_bh);
819 path[level].bp_bh = path[level].bp_sib_bh;
820 path[level].bp_sib_bh = NULL;
821 path[level].bp_index += lnchildren;
822 path[level + 1].bp_index--;
824 brelse(path[level].bp_sib_bh);
825 path[level].bp_sib_bh = NULL;
826 path[level].bp_index -= n;
829 nilfs_btree_do_insert(btree, path, level, keyp, ptrp);
832 static void nilfs_btree_carry_right(struct nilfs_bmap *btree,
833 struct nilfs_btree_path *path,
834 int level, __u64 *keyp, __u64 *ptrp)
836 struct nilfs_btree_node *node, *right;
837 int nchildren, rnchildren, n, move, ncblk;
839 node = nilfs_btree_get_nonroot_node(path, level);
840 right = nilfs_btree_get_sib_node(path, level);
841 nchildren = nilfs_btree_node_get_nchildren(node);
842 rnchildren = nilfs_btree_node_get_nchildren(right);
843 ncblk = nilfs_btree_nchildren_per_block(btree);
846 n = (nchildren + rnchildren + 1) / 2 - rnchildren;
847 if (n > nchildren - path[level].bp_index) {
848 /* move insert point */
853 nilfs_btree_node_move_right(node, right, n, ncblk, ncblk);
855 if (!buffer_dirty(path[level].bp_bh))
856 mark_buffer_dirty(path[level].bp_bh);
857 if (!buffer_dirty(path[level].bp_sib_bh))
858 mark_buffer_dirty(path[level].bp_sib_bh);
860 path[level + 1].bp_index++;
861 nilfs_btree_promote_key(btree, path, level + 1,
862 nilfs_btree_node_get_key(right, 0));
863 path[level + 1].bp_index--;
866 brelse(path[level].bp_bh);
867 path[level].bp_bh = path[level].bp_sib_bh;
868 path[level].bp_sib_bh = NULL;
869 path[level].bp_index -= nilfs_btree_node_get_nchildren(node);
870 path[level + 1].bp_index++;
872 brelse(path[level].bp_sib_bh);
873 path[level].bp_sib_bh = NULL;
876 nilfs_btree_do_insert(btree, path, level, keyp, ptrp);
879 static void nilfs_btree_split(struct nilfs_bmap *btree,
880 struct nilfs_btree_path *path,
881 int level, __u64 *keyp, __u64 *ptrp)
883 struct nilfs_btree_node *node, *right;
886 int nchildren, n, move, ncblk;
888 node = nilfs_btree_get_nonroot_node(path, level);
889 right = nilfs_btree_get_sib_node(path, level);
890 nchildren = nilfs_btree_node_get_nchildren(node);
891 ncblk = nilfs_btree_nchildren_per_block(btree);
894 n = (nchildren + 1) / 2;
895 if (n > nchildren - path[level].bp_index) {
900 nilfs_btree_node_move_right(node, right, n, ncblk, ncblk);
902 if (!buffer_dirty(path[level].bp_bh))
903 mark_buffer_dirty(path[level].bp_bh);
904 if (!buffer_dirty(path[level].bp_sib_bh))
905 mark_buffer_dirty(path[level].bp_sib_bh);
907 newkey = nilfs_btree_node_get_key(right, 0);
908 newptr = path[level].bp_newreq.bpr_ptr;
911 path[level].bp_index -= nilfs_btree_node_get_nchildren(node);
912 nilfs_btree_node_insert(right, path[level].bp_index,
913 *keyp, *ptrp, ncblk);
915 *keyp = nilfs_btree_node_get_key(right, 0);
916 *ptrp = path[level].bp_newreq.bpr_ptr;
918 brelse(path[level].bp_bh);
919 path[level].bp_bh = path[level].bp_sib_bh;
920 path[level].bp_sib_bh = NULL;
922 nilfs_btree_do_insert(btree, path, level, keyp, ptrp);
924 *keyp = nilfs_btree_node_get_key(right, 0);
925 *ptrp = path[level].bp_newreq.bpr_ptr;
927 brelse(path[level].bp_sib_bh);
928 path[level].bp_sib_bh = NULL;
931 path[level + 1].bp_index++;
934 static void nilfs_btree_grow(struct nilfs_bmap *btree,
935 struct nilfs_btree_path *path,
936 int level, __u64 *keyp, __u64 *ptrp)
938 struct nilfs_btree_node *root, *child;
941 root = nilfs_btree_get_root(btree);
942 child = nilfs_btree_get_sib_node(path, level);
943 ncblk = nilfs_btree_nchildren_per_block(btree);
945 n = nilfs_btree_node_get_nchildren(root);
947 nilfs_btree_node_move_right(root, child, n,
948 NILFS_BTREE_ROOT_NCHILDREN_MAX, ncblk);
949 nilfs_btree_node_set_level(root, level + 1);
951 if (!buffer_dirty(path[level].bp_sib_bh))
952 mark_buffer_dirty(path[level].bp_sib_bh);
954 path[level].bp_bh = path[level].bp_sib_bh;
955 path[level].bp_sib_bh = NULL;
957 nilfs_btree_do_insert(btree, path, level, keyp, ptrp);
959 *keyp = nilfs_btree_node_get_key(child, 0);
960 *ptrp = path[level].bp_newreq.bpr_ptr;
963 static __u64 nilfs_btree_find_near(const struct nilfs_bmap *btree,
964 const struct nilfs_btree_path *path)
966 struct nilfs_btree_node *node;
970 return NILFS_BMAP_INVALID_PTR;
973 level = NILFS_BTREE_LEVEL_NODE_MIN;
974 if (path[level].bp_index > 0) {
975 node = nilfs_btree_get_node(btree, path, level, &ncmax);
976 return nilfs_btree_node_get_ptr(node,
977 path[level].bp_index - 1,
982 level = NILFS_BTREE_LEVEL_NODE_MIN + 1;
983 if (level <= nilfs_btree_height(btree) - 1) {
984 node = nilfs_btree_get_node(btree, path, level, &ncmax);
985 return nilfs_btree_node_get_ptr(node, path[level].bp_index,
989 return NILFS_BMAP_INVALID_PTR;
992 static __u64 nilfs_btree_find_target_v(const struct nilfs_bmap *btree,
993 const struct nilfs_btree_path *path,
998 ptr = nilfs_bmap_find_target_seq(btree, key);
999 if (ptr != NILFS_BMAP_INVALID_PTR)
1000 /* sequential access */
1003 ptr = nilfs_btree_find_near(btree, path);
1004 if (ptr != NILFS_BMAP_INVALID_PTR)
1009 return nilfs_bmap_find_target_in_group(btree);
1012 static int nilfs_btree_prepare_insert(struct nilfs_bmap *btree,
1013 struct nilfs_btree_path *path,
1014 int *levelp, __u64 key, __u64 ptr,
1015 struct nilfs_bmap_stats *stats)
1017 struct buffer_head *bh;
1018 struct nilfs_btree_node *node, *parent, *sib;
1020 int pindex, level, ncmax, ncblk, ret;
1021 struct inode *dat = NULL;
1023 stats->bs_nblocks = 0;
1024 level = NILFS_BTREE_LEVEL_DATA;
1026 /* allocate a new ptr for data block */
1027 if (NILFS_BMAP_USE_VBN(btree)) {
1028 path[level].bp_newreq.bpr_ptr =
1029 nilfs_btree_find_target_v(btree, path, key);
1030 dat = nilfs_bmap_get_dat(btree);
1033 ret = nilfs_bmap_prepare_alloc_ptr(btree, &path[level].bp_newreq, dat);
1037 ncblk = nilfs_btree_nchildren_per_block(btree);
1039 for (level = NILFS_BTREE_LEVEL_NODE_MIN;
1040 level < nilfs_btree_height(btree) - 1;
1042 node = nilfs_btree_get_nonroot_node(path, level);
1043 if (nilfs_btree_node_get_nchildren(node) < ncblk) {
1044 path[level].bp_op = nilfs_btree_do_insert;
1045 stats->bs_nblocks++;
1049 parent = nilfs_btree_get_node(btree, path, level + 1, &ncmax);
1050 pindex = path[level + 1].bp_index;
1054 sibptr = nilfs_btree_node_get_ptr(parent, pindex - 1,
1056 ret = nilfs_btree_get_block(btree, sibptr, &bh);
1058 goto err_out_child_node;
1059 sib = (struct nilfs_btree_node *)bh->b_data;
1060 if (nilfs_btree_node_get_nchildren(sib) < ncblk) {
1061 path[level].bp_sib_bh = bh;
1062 path[level].bp_op = nilfs_btree_carry_left;
1063 stats->bs_nblocks++;
1071 if (pindex < nilfs_btree_node_get_nchildren(parent) - 1) {
1072 sibptr = nilfs_btree_node_get_ptr(parent, pindex + 1,
1074 ret = nilfs_btree_get_block(btree, sibptr, &bh);
1076 goto err_out_child_node;
1077 sib = (struct nilfs_btree_node *)bh->b_data;
1078 if (nilfs_btree_node_get_nchildren(sib) < ncblk) {
1079 path[level].bp_sib_bh = bh;
1080 path[level].bp_op = nilfs_btree_carry_right;
1081 stats->bs_nblocks++;
1089 path[level].bp_newreq.bpr_ptr =
1090 path[level - 1].bp_newreq.bpr_ptr + 1;
1091 ret = nilfs_bmap_prepare_alloc_ptr(btree,
1092 &path[level].bp_newreq, dat);
1094 goto err_out_child_node;
1095 ret = nilfs_btree_get_new_block(btree,
1096 path[level].bp_newreq.bpr_ptr,
1099 goto err_out_curr_node;
1101 stats->bs_nblocks++;
1103 sib = (struct nilfs_btree_node *)bh->b_data;
1104 nilfs_btree_node_init(sib, 0, level, 0, ncblk, NULL, NULL);
1105 path[level].bp_sib_bh = bh;
1106 path[level].bp_op = nilfs_btree_split;
1110 node = nilfs_btree_get_root(btree);
1111 if (nilfs_btree_node_get_nchildren(node) <
1112 NILFS_BTREE_ROOT_NCHILDREN_MAX) {
1113 path[level].bp_op = nilfs_btree_do_insert;
1114 stats->bs_nblocks++;
1119 path[level].bp_newreq.bpr_ptr = path[level - 1].bp_newreq.bpr_ptr + 1;
1120 ret = nilfs_bmap_prepare_alloc_ptr(btree, &path[level].bp_newreq, dat);
1122 goto err_out_child_node;
1123 ret = nilfs_btree_get_new_block(btree, path[level].bp_newreq.bpr_ptr,
1126 goto err_out_curr_node;
1128 nilfs_btree_node_init((struct nilfs_btree_node *)bh->b_data,
1129 0, level, 0, ncblk, NULL, NULL);
1130 path[level].bp_sib_bh = bh;
1131 path[level].bp_op = nilfs_btree_grow;
1134 path[level].bp_op = nilfs_btree_do_insert;
1136 /* a newly-created node block and a data block are added */
1137 stats->bs_nblocks += 2;
1146 nilfs_bmap_abort_alloc_ptr(btree, &path[level].bp_newreq, dat);
1148 for (level--; level > NILFS_BTREE_LEVEL_DATA; level--) {
1149 nilfs_btnode_delete(path[level].bp_sib_bh);
1150 nilfs_bmap_abort_alloc_ptr(btree, &path[level].bp_newreq, dat);
1154 nilfs_bmap_abort_alloc_ptr(btree, &path[level].bp_newreq, dat);
1157 stats->bs_nblocks = 0;
1161 static void nilfs_btree_commit_insert(struct nilfs_bmap *btree,
1162 struct nilfs_btree_path *path,
1163 int maxlevel, __u64 key, __u64 ptr)
1165 struct inode *dat = NULL;
1168 set_buffer_nilfs_volatile((struct buffer_head *)((unsigned long)ptr));
1169 ptr = path[NILFS_BTREE_LEVEL_DATA].bp_newreq.bpr_ptr;
1170 if (NILFS_BMAP_USE_VBN(btree)) {
1171 nilfs_bmap_set_target_v(btree, key, ptr);
1172 dat = nilfs_bmap_get_dat(btree);
1175 for (level = NILFS_BTREE_LEVEL_NODE_MIN; level <= maxlevel; level++) {
1176 nilfs_bmap_commit_alloc_ptr(btree,
1177 &path[level - 1].bp_newreq, dat);
1178 path[level].bp_op(btree, path, level, &key, &ptr);
1181 if (!nilfs_bmap_dirty(btree))
1182 nilfs_bmap_set_dirty(btree);
1185 static int nilfs_btree_insert(struct nilfs_bmap *btree, __u64 key, __u64 ptr)
1187 struct nilfs_btree_path *path;
1188 struct nilfs_bmap_stats stats;
1191 path = nilfs_btree_alloc_path();
1195 ret = nilfs_btree_do_lookup(btree, path, key, NULL,
1196 NILFS_BTREE_LEVEL_NODE_MIN, 0);
1197 if (ret != -ENOENT) {
1203 ret = nilfs_btree_prepare_insert(btree, path, &level, key, ptr, &stats);
1206 nilfs_btree_commit_insert(btree, path, level, key, ptr);
1207 nilfs_inode_add_blocks(btree->b_inode, stats.bs_nblocks);
1210 nilfs_btree_free_path(path);
1214 static void nilfs_btree_do_delete(struct nilfs_bmap *btree,
1215 struct nilfs_btree_path *path,
1216 int level, __u64 *keyp, __u64 *ptrp)
1218 struct nilfs_btree_node *node;
1221 if (level < nilfs_btree_height(btree) - 1) {
1222 node = nilfs_btree_get_nonroot_node(path, level);
1223 ncblk = nilfs_btree_nchildren_per_block(btree);
1224 nilfs_btree_node_delete(node, path[level].bp_index,
1226 if (!buffer_dirty(path[level].bp_bh))
1227 mark_buffer_dirty(path[level].bp_bh);
1228 if (path[level].bp_index == 0)
1229 nilfs_btree_promote_key(btree, path, level + 1,
1230 nilfs_btree_node_get_key(node, 0));
1232 node = nilfs_btree_get_root(btree);
1233 nilfs_btree_node_delete(node, path[level].bp_index,
1235 NILFS_BTREE_ROOT_NCHILDREN_MAX);
1239 static void nilfs_btree_borrow_left(struct nilfs_bmap *btree,
1240 struct nilfs_btree_path *path,
1241 int level, __u64 *keyp, __u64 *ptrp)
1243 struct nilfs_btree_node *node, *left;
1244 int nchildren, lnchildren, n, ncblk;
1246 nilfs_btree_do_delete(btree, path, level, keyp, ptrp);
1248 node = nilfs_btree_get_nonroot_node(path, level);
1249 left = nilfs_btree_get_sib_node(path, level);
1250 nchildren = nilfs_btree_node_get_nchildren(node);
1251 lnchildren = nilfs_btree_node_get_nchildren(left);
1252 ncblk = nilfs_btree_nchildren_per_block(btree);
1254 n = (nchildren + lnchildren) / 2 - nchildren;
1256 nilfs_btree_node_move_right(left, node, n, ncblk, ncblk);
1258 if (!buffer_dirty(path[level].bp_bh))
1259 mark_buffer_dirty(path[level].bp_bh);
1260 if (!buffer_dirty(path[level].bp_sib_bh))
1261 mark_buffer_dirty(path[level].bp_sib_bh);
1263 nilfs_btree_promote_key(btree, path, level + 1,
1264 nilfs_btree_node_get_key(node, 0));
1266 brelse(path[level].bp_sib_bh);
1267 path[level].bp_sib_bh = NULL;
1268 path[level].bp_index += n;
1271 static void nilfs_btree_borrow_right(struct nilfs_bmap *btree,
1272 struct nilfs_btree_path *path,
1273 int level, __u64 *keyp, __u64 *ptrp)
1275 struct nilfs_btree_node *node, *right;
1276 int nchildren, rnchildren, n, ncblk;
1278 nilfs_btree_do_delete(btree, path, level, keyp, ptrp);
1280 node = nilfs_btree_get_nonroot_node(path, level);
1281 right = nilfs_btree_get_sib_node(path, level);
1282 nchildren = nilfs_btree_node_get_nchildren(node);
1283 rnchildren = nilfs_btree_node_get_nchildren(right);
1284 ncblk = nilfs_btree_nchildren_per_block(btree);
1286 n = (nchildren + rnchildren) / 2 - nchildren;
1288 nilfs_btree_node_move_left(node, right, n, ncblk, ncblk);
1290 if (!buffer_dirty(path[level].bp_bh))
1291 mark_buffer_dirty(path[level].bp_bh);
1292 if (!buffer_dirty(path[level].bp_sib_bh))
1293 mark_buffer_dirty(path[level].bp_sib_bh);
1295 path[level + 1].bp_index++;
1296 nilfs_btree_promote_key(btree, path, level + 1,
1297 nilfs_btree_node_get_key(right, 0));
1298 path[level + 1].bp_index--;
1300 brelse(path[level].bp_sib_bh);
1301 path[level].bp_sib_bh = NULL;
1304 static void nilfs_btree_concat_left(struct nilfs_bmap *btree,
1305 struct nilfs_btree_path *path,
1306 int level, __u64 *keyp, __u64 *ptrp)
1308 struct nilfs_btree_node *node, *left;
1311 nilfs_btree_do_delete(btree, path, level, keyp, ptrp);
1313 node = nilfs_btree_get_nonroot_node(path, level);
1314 left = nilfs_btree_get_sib_node(path, level);
1315 ncblk = nilfs_btree_nchildren_per_block(btree);
1317 n = nilfs_btree_node_get_nchildren(node);
1319 nilfs_btree_node_move_left(left, node, n, ncblk, ncblk);
1321 if (!buffer_dirty(path[level].bp_sib_bh))
1322 mark_buffer_dirty(path[level].bp_sib_bh);
1324 nilfs_btnode_delete(path[level].bp_bh);
1325 path[level].bp_bh = path[level].bp_sib_bh;
1326 path[level].bp_sib_bh = NULL;
1327 path[level].bp_index += nilfs_btree_node_get_nchildren(left);
1330 static void nilfs_btree_concat_right(struct nilfs_bmap *btree,
1331 struct nilfs_btree_path *path,
1332 int level, __u64 *keyp, __u64 *ptrp)
1334 struct nilfs_btree_node *node, *right;
1337 nilfs_btree_do_delete(btree, path, level, keyp, ptrp);
1339 node = nilfs_btree_get_nonroot_node(path, level);
1340 right = nilfs_btree_get_sib_node(path, level);
1341 ncblk = nilfs_btree_nchildren_per_block(btree);
1343 n = nilfs_btree_node_get_nchildren(right);
1345 nilfs_btree_node_move_left(node, right, n, ncblk, ncblk);
1347 if (!buffer_dirty(path[level].bp_bh))
1348 mark_buffer_dirty(path[level].bp_bh);
1350 nilfs_btnode_delete(path[level].bp_sib_bh);
1351 path[level].bp_sib_bh = NULL;
1352 path[level + 1].bp_index++;
1355 static void nilfs_btree_shrink(struct nilfs_bmap *btree,
1356 struct nilfs_btree_path *path,
1357 int level, __u64 *keyp, __u64 *ptrp)
1359 struct nilfs_btree_node *root, *child;
1362 nilfs_btree_do_delete(btree, path, level, keyp, ptrp);
1364 root = nilfs_btree_get_root(btree);
1365 child = nilfs_btree_get_nonroot_node(path, level);
1366 ncblk = nilfs_btree_nchildren_per_block(btree);
1368 nilfs_btree_node_delete(root, 0, NULL, NULL,
1369 NILFS_BTREE_ROOT_NCHILDREN_MAX);
1370 nilfs_btree_node_set_level(root, level);
1371 n = nilfs_btree_node_get_nchildren(child);
1372 nilfs_btree_node_move_left(root, child, n,
1373 NILFS_BTREE_ROOT_NCHILDREN_MAX, ncblk);
1375 nilfs_btnode_delete(path[level].bp_bh);
1376 path[level].bp_bh = NULL;
1379 static void nilfs_btree_nop(struct nilfs_bmap *btree,
1380 struct nilfs_btree_path *path,
1381 int level, __u64 *keyp, __u64 *ptrp)
1385 static int nilfs_btree_prepare_delete(struct nilfs_bmap *btree,
1386 struct nilfs_btree_path *path,
1388 struct nilfs_bmap_stats *stats,
1391 struct buffer_head *bh;
1392 struct nilfs_btree_node *node, *parent, *sib;
1394 int pindex, dindex, level, ncmin, ncmax, ncblk, ret;
1397 stats->bs_nblocks = 0;
1398 ncmin = NILFS_BTREE_NODE_NCHILDREN_MIN(nilfs_btree_node_size(btree));
1399 ncblk = nilfs_btree_nchildren_per_block(btree);
1401 for (level = NILFS_BTREE_LEVEL_NODE_MIN, dindex = path[level].bp_index;
1402 level < nilfs_btree_height(btree) - 1;
1404 node = nilfs_btree_get_nonroot_node(path, level);
1405 path[level].bp_oldreq.bpr_ptr =
1406 nilfs_btree_node_get_ptr(node, dindex, ncblk);
1407 ret = nilfs_bmap_prepare_end_ptr(btree,
1408 &path[level].bp_oldreq, dat);
1410 goto err_out_child_node;
1412 if (nilfs_btree_node_get_nchildren(node) > ncmin) {
1413 path[level].bp_op = nilfs_btree_do_delete;
1414 stats->bs_nblocks++;
1418 parent = nilfs_btree_get_node(btree, path, level + 1, &ncmax);
1419 pindex = path[level + 1].bp_index;
1424 sibptr = nilfs_btree_node_get_ptr(parent, pindex - 1,
1426 ret = nilfs_btree_get_block(btree, sibptr, &bh);
1428 goto err_out_curr_node;
1429 sib = (struct nilfs_btree_node *)bh->b_data;
1430 if (nilfs_btree_node_get_nchildren(sib) > ncmin) {
1431 path[level].bp_sib_bh = bh;
1432 path[level].bp_op = nilfs_btree_borrow_left;
1433 stats->bs_nblocks++;
1436 path[level].bp_sib_bh = bh;
1437 path[level].bp_op = nilfs_btree_concat_left;
1438 stats->bs_nblocks++;
1442 nilfs_btree_node_get_nchildren(parent) - 1) {
1444 sibptr = nilfs_btree_node_get_ptr(parent, pindex + 1,
1446 ret = nilfs_btree_get_block(btree, sibptr, &bh);
1448 goto err_out_curr_node;
1449 sib = (struct nilfs_btree_node *)bh->b_data;
1450 if (nilfs_btree_node_get_nchildren(sib) > ncmin) {
1451 path[level].bp_sib_bh = bh;
1452 path[level].bp_op = nilfs_btree_borrow_right;
1453 stats->bs_nblocks++;
1456 path[level].bp_sib_bh = bh;
1457 path[level].bp_op = nilfs_btree_concat_right;
1458 stats->bs_nblocks++;
1460 * When merging right sibling node
1461 * into the current node, pointer to
1462 * the right sibling node must be
1463 * terminated instead. The adjustment
1464 * below is required for that.
1466 dindex = pindex + 1;
1471 /* the only child of the root node */
1472 WARN_ON(level != nilfs_btree_height(btree) - 2);
1473 if (nilfs_btree_node_get_nchildren(node) - 1 <=
1474 NILFS_BTREE_ROOT_NCHILDREN_MAX) {
1475 path[level].bp_op = nilfs_btree_shrink;
1476 stats->bs_nblocks += 2;
1478 path[level].bp_op = nilfs_btree_nop;
1479 goto shrink_root_child;
1481 path[level].bp_op = nilfs_btree_do_delete;
1482 stats->bs_nblocks++;
1488 /* child of the root node is deleted */
1489 path[level].bp_op = nilfs_btree_do_delete;
1490 stats->bs_nblocks++;
1493 node = nilfs_btree_get_root(btree);
1494 path[level].bp_oldreq.bpr_ptr =
1495 nilfs_btree_node_get_ptr(node, dindex,
1496 NILFS_BTREE_ROOT_NCHILDREN_MAX);
1498 ret = nilfs_bmap_prepare_end_ptr(btree, &path[level].bp_oldreq, dat);
1500 goto err_out_child_node;
1509 nilfs_bmap_abort_end_ptr(btree, &path[level].bp_oldreq, dat);
1511 for (level--; level >= NILFS_BTREE_LEVEL_NODE_MIN; level--) {
1512 brelse(path[level].bp_sib_bh);
1513 nilfs_bmap_abort_end_ptr(btree, &path[level].bp_oldreq, dat);
1516 stats->bs_nblocks = 0;
1520 static void nilfs_btree_commit_delete(struct nilfs_bmap *btree,
1521 struct nilfs_btree_path *path,
1522 int maxlevel, struct inode *dat)
1526 for (level = NILFS_BTREE_LEVEL_NODE_MIN; level <= maxlevel; level++) {
1527 nilfs_bmap_commit_end_ptr(btree, &path[level].bp_oldreq, dat);
1528 path[level].bp_op(btree, path, level, NULL, NULL);
1531 if (!nilfs_bmap_dirty(btree))
1532 nilfs_bmap_set_dirty(btree);
1535 static int nilfs_btree_delete(struct nilfs_bmap *btree, __u64 key)
1538 struct nilfs_btree_path *path;
1539 struct nilfs_bmap_stats stats;
1543 path = nilfs_btree_alloc_path();
1547 ret = nilfs_btree_do_lookup(btree, path, key, NULL,
1548 NILFS_BTREE_LEVEL_NODE_MIN, 0);
1553 dat = NILFS_BMAP_USE_VBN(btree) ? nilfs_bmap_get_dat(btree) : NULL;
1555 ret = nilfs_btree_prepare_delete(btree, path, &level, &stats, dat);
1558 nilfs_btree_commit_delete(btree, path, level, dat);
1559 nilfs_inode_sub_blocks(btree->b_inode, stats.bs_nblocks);
1562 nilfs_btree_free_path(path);
1566 static int nilfs_btree_last_key(const struct nilfs_bmap *btree, __u64 *keyp)
1568 struct nilfs_btree_path *path;
1571 path = nilfs_btree_alloc_path();
1575 ret = nilfs_btree_do_lookup_last(btree, path, keyp, NULL);
1577 nilfs_btree_free_path(path);
1582 static int nilfs_btree_check_delete(struct nilfs_bmap *btree, __u64 key)
1584 struct buffer_head *bh;
1585 struct nilfs_btree_node *root, *node;
1586 __u64 maxkey, nextmaxkey;
1590 root = nilfs_btree_get_root(btree);
1591 switch (nilfs_btree_height(btree)) {
1597 nchildren = nilfs_btree_node_get_nchildren(root);
1600 ptr = nilfs_btree_node_get_ptr(root, nchildren - 1,
1601 NILFS_BTREE_ROOT_NCHILDREN_MAX);
1602 ret = nilfs_btree_get_block(btree, ptr, &bh);
1605 node = (struct nilfs_btree_node *)bh->b_data;
1611 nchildren = nilfs_btree_node_get_nchildren(node);
1612 maxkey = nilfs_btree_node_get_key(node, nchildren - 1);
1613 nextmaxkey = (nchildren > 1) ?
1614 nilfs_btree_node_get_key(node, nchildren - 2) : 0;
1618 return (maxkey == key) && (nextmaxkey < NILFS_BMAP_LARGE_LOW);
1621 static int nilfs_btree_gather_data(struct nilfs_bmap *btree,
1622 __u64 *keys, __u64 *ptrs, int nitems)
1624 struct buffer_head *bh;
1625 struct nilfs_btree_node *node, *root;
1629 int nchildren, ncmax, i, ret;
1631 root = nilfs_btree_get_root(btree);
1632 switch (nilfs_btree_height(btree)) {
1636 ncmax = NILFS_BTREE_ROOT_NCHILDREN_MAX;
1639 nchildren = nilfs_btree_node_get_nchildren(root);
1640 WARN_ON(nchildren > 1);
1641 ptr = nilfs_btree_node_get_ptr(root, nchildren - 1,
1642 NILFS_BTREE_ROOT_NCHILDREN_MAX);
1643 ret = nilfs_btree_get_block(btree, ptr, &bh);
1646 node = (struct nilfs_btree_node *)bh->b_data;
1647 ncmax = nilfs_btree_nchildren_per_block(btree);
1654 nchildren = nilfs_btree_node_get_nchildren(node);
1655 if (nchildren < nitems)
1657 dkeys = nilfs_btree_node_dkeys(node);
1658 dptrs = nilfs_btree_node_dptrs(node, ncmax);
1659 for (i = 0; i < nitems; i++) {
1660 keys[i] = le64_to_cpu(dkeys[i]);
1661 ptrs[i] = le64_to_cpu(dptrs[i]);
1671 nilfs_btree_prepare_convert_and_insert(struct nilfs_bmap *btree, __u64 key,
1672 union nilfs_bmap_ptr_req *dreq,
1673 union nilfs_bmap_ptr_req *nreq,
1674 struct buffer_head **bhp,
1675 struct nilfs_bmap_stats *stats)
1677 struct buffer_head *bh;
1678 struct inode *dat = NULL;
1681 stats->bs_nblocks = 0;
1684 /* cannot find near ptr */
1685 if (NILFS_BMAP_USE_VBN(btree)) {
1686 dreq->bpr_ptr = nilfs_btree_find_target_v(btree, NULL, key);
1687 dat = nilfs_bmap_get_dat(btree);
1690 ret = nilfs_bmap_prepare_alloc_ptr(btree, dreq, dat);
1695 stats->bs_nblocks++;
1697 nreq->bpr_ptr = dreq->bpr_ptr + 1;
1698 ret = nilfs_bmap_prepare_alloc_ptr(btree, nreq, dat);
1702 ret = nilfs_btree_get_new_block(btree, nreq->bpr_ptr, &bh);
1707 stats->bs_nblocks++;
1715 nilfs_bmap_abort_alloc_ptr(btree, nreq, dat);
1717 nilfs_bmap_abort_alloc_ptr(btree, dreq, dat);
1718 stats->bs_nblocks = 0;
1724 nilfs_btree_commit_convert_and_insert(struct nilfs_bmap *btree,
1725 __u64 key, __u64 ptr,
1726 const __u64 *keys, const __u64 *ptrs,
1728 union nilfs_bmap_ptr_req *dreq,
1729 union nilfs_bmap_ptr_req *nreq,
1730 struct buffer_head *bh)
1732 struct nilfs_btree_node *node;
1737 /* free resources */
1738 if (btree->b_ops->bop_clear != NULL)
1739 btree->b_ops->bop_clear(btree);
1741 /* ptr must be a pointer to a buffer head. */
1742 set_buffer_nilfs_volatile((struct buffer_head *)((unsigned long)ptr));
1744 /* convert and insert */
1745 dat = NILFS_BMAP_USE_VBN(btree) ? nilfs_bmap_get_dat(btree) : NULL;
1746 __nilfs_btree_init(btree);
1748 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat);
1749 nilfs_bmap_commit_alloc_ptr(btree, nreq, dat);
1751 /* create child node at level 1 */
1752 node = (struct nilfs_btree_node *)bh->b_data;
1753 ncblk = nilfs_btree_nchildren_per_block(btree);
1754 nilfs_btree_node_init(node, 0, 1, n, ncblk, keys, ptrs);
1755 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr, ncblk);
1756 if (!buffer_dirty(bh))
1757 mark_buffer_dirty(bh);
1758 if (!nilfs_bmap_dirty(btree))
1759 nilfs_bmap_set_dirty(btree);
1763 /* create root node at level 2 */
1764 node = nilfs_btree_get_root(btree);
1765 tmpptr = nreq->bpr_ptr;
1766 nilfs_btree_node_init(node, NILFS_BTREE_NODE_ROOT, 2, 1,
1767 NILFS_BTREE_ROOT_NCHILDREN_MAX,
1770 nilfs_bmap_commit_alloc_ptr(btree, dreq, dat);
1772 /* create root node at level 1 */
1773 node = nilfs_btree_get_root(btree);
1774 nilfs_btree_node_init(node, NILFS_BTREE_NODE_ROOT, 1, n,
1775 NILFS_BTREE_ROOT_NCHILDREN_MAX,
1777 nilfs_btree_node_insert(node, n, key, dreq->bpr_ptr,
1778 NILFS_BTREE_ROOT_NCHILDREN_MAX);
1779 if (!nilfs_bmap_dirty(btree))
1780 nilfs_bmap_set_dirty(btree);
1783 if (NILFS_BMAP_USE_VBN(btree))
1784 nilfs_bmap_set_target_v(btree, key, dreq->bpr_ptr);
1788 * nilfs_btree_convert_and_insert -
1796 int nilfs_btree_convert_and_insert(struct nilfs_bmap *btree,
1797 __u64 key, __u64 ptr,
1798 const __u64 *keys, const __u64 *ptrs, int n)
1800 struct buffer_head *bh;
1801 union nilfs_bmap_ptr_req dreq, nreq, *di, *ni;
1802 struct nilfs_bmap_stats stats;
1805 if (n + 1 <= NILFS_BTREE_ROOT_NCHILDREN_MAX) {
1808 } else if ((n + 1) <= NILFS_BTREE_NODE_NCHILDREN_MAX(
1809 1 << btree->b_inode->i_blkbits)) {
1818 ret = nilfs_btree_prepare_convert_and_insert(btree, key, di, ni, &bh,
1822 nilfs_btree_commit_convert_and_insert(btree, key, ptr, keys, ptrs, n,
1824 nilfs_inode_add_blocks(btree->b_inode, stats.bs_nblocks);
1828 static int nilfs_btree_propagate_p(struct nilfs_bmap *btree,
1829 struct nilfs_btree_path *path,
1831 struct buffer_head *bh)
1833 while ((++level < nilfs_btree_height(btree) - 1) &&
1834 !buffer_dirty(path[level].bp_bh))
1835 mark_buffer_dirty(path[level].bp_bh);
1840 static int nilfs_btree_prepare_update_v(struct nilfs_bmap *btree,
1841 struct nilfs_btree_path *path,
1842 int level, struct inode *dat)
1844 struct nilfs_btree_node *parent;
1847 parent = nilfs_btree_get_node(btree, path, level + 1, &ncmax);
1848 path[level].bp_oldreq.bpr_ptr =
1849 nilfs_btree_node_get_ptr(parent, path[level + 1].bp_index,
1851 path[level].bp_newreq.bpr_ptr = path[level].bp_oldreq.bpr_ptr + 1;
1852 ret = nilfs_dat_prepare_update(dat, &path[level].bp_oldreq.bpr_req,
1853 &path[level].bp_newreq.bpr_req);
1857 if (buffer_nilfs_node(path[level].bp_bh)) {
1858 path[level].bp_ctxt.oldkey = path[level].bp_oldreq.bpr_ptr;
1859 path[level].bp_ctxt.newkey = path[level].bp_newreq.bpr_ptr;
1860 path[level].bp_ctxt.bh = path[level].bp_bh;
1861 ret = nilfs_btnode_prepare_change_key(
1862 &NILFS_BMAP_I(btree)->i_btnode_cache,
1863 &path[level].bp_ctxt);
1865 nilfs_dat_abort_update(dat,
1866 &path[level].bp_oldreq.bpr_req,
1867 &path[level].bp_newreq.bpr_req);
1875 static void nilfs_btree_commit_update_v(struct nilfs_bmap *btree,
1876 struct nilfs_btree_path *path,
1877 int level, struct inode *dat)
1879 struct nilfs_btree_node *parent;
1882 nilfs_dat_commit_update(dat, &path[level].bp_oldreq.bpr_req,
1883 &path[level].bp_newreq.bpr_req,
1884 btree->b_ptr_type == NILFS_BMAP_PTR_VS);
1886 if (buffer_nilfs_node(path[level].bp_bh)) {
1887 nilfs_btnode_commit_change_key(
1888 &NILFS_BMAP_I(btree)->i_btnode_cache,
1889 &path[level].bp_ctxt);
1890 path[level].bp_bh = path[level].bp_ctxt.bh;
1892 set_buffer_nilfs_volatile(path[level].bp_bh);
1894 parent = nilfs_btree_get_node(btree, path, level + 1, &ncmax);
1895 nilfs_btree_node_set_ptr(parent, path[level + 1].bp_index,
1896 path[level].bp_newreq.bpr_ptr, ncmax);
1899 static void nilfs_btree_abort_update_v(struct nilfs_bmap *btree,
1900 struct nilfs_btree_path *path,
1901 int level, struct inode *dat)
1903 nilfs_dat_abort_update(dat, &path[level].bp_oldreq.bpr_req,
1904 &path[level].bp_newreq.bpr_req);
1905 if (buffer_nilfs_node(path[level].bp_bh))
1906 nilfs_btnode_abort_change_key(
1907 &NILFS_BMAP_I(btree)->i_btnode_cache,
1908 &path[level].bp_ctxt);
1911 static int nilfs_btree_prepare_propagate_v(struct nilfs_bmap *btree,
1912 struct nilfs_btree_path *path,
1913 int minlevel, int *maxlevelp,
1919 if (!buffer_nilfs_volatile(path[level].bp_bh)) {
1920 ret = nilfs_btree_prepare_update_v(btree, path, level, dat);
1924 while ((++level < nilfs_btree_height(btree) - 1) &&
1925 !buffer_dirty(path[level].bp_bh)) {
1927 WARN_ON(buffer_nilfs_volatile(path[level].bp_bh));
1928 ret = nilfs_btree_prepare_update_v(btree, path, level, dat);
1934 *maxlevelp = level - 1;
1939 while (--level > minlevel)
1940 nilfs_btree_abort_update_v(btree, path, level, dat);
1941 if (!buffer_nilfs_volatile(path[level].bp_bh))
1942 nilfs_btree_abort_update_v(btree, path, level, dat);
1946 static void nilfs_btree_commit_propagate_v(struct nilfs_bmap *btree,
1947 struct nilfs_btree_path *path,
1948 int minlevel, int maxlevel,
1949 struct buffer_head *bh,
1954 if (!buffer_nilfs_volatile(path[minlevel].bp_bh))
1955 nilfs_btree_commit_update_v(btree, path, minlevel, dat);
1957 for (level = minlevel + 1; level <= maxlevel; level++)
1958 nilfs_btree_commit_update_v(btree, path, level, dat);
1961 static int nilfs_btree_propagate_v(struct nilfs_bmap *btree,
1962 struct nilfs_btree_path *path,
1963 int level, struct buffer_head *bh)
1965 int maxlevel = 0, ret;
1966 struct nilfs_btree_node *parent;
1967 struct inode *dat = nilfs_bmap_get_dat(btree);
1972 path[level].bp_bh = bh;
1973 ret = nilfs_btree_prepare_propagate_v(btree, path, level, &maxlevel,
1978 if (buffer_nilfs_volatile(path[level].bp_bh)) {
1979 parent = nilfs_btree_get_node(btree, path, level + 1, &ncmax);
1980 ptr = nilfs_btree_node_get_ptr(parent,
1981 path[level + 1].bp_index,
1983 ret = nilfs_dat_mark_dirty(dat, ptr);
1988 nilfs_btree_commit_propagate_v(btree, path, level, maxlevel, bh, dat);
1991 brelse(path[level].bp_bh);
1992 path[level].bp_bh = NULL;
1996 static int nilfs_btree_propagate(struct nilfs_bmap *btree,
1997 struct buffer_head *bh)
1999 struct nilfs_btree_path *path;
2000 struct nilfs_btree_node *node;
2004 WARN_ON(!buffer_dirty(bh));
2006 path = nilfs_btree_alloc_path();
2010 if (buffer_nilfs_node(bh)) {
2011 node = (struct nilfs_btree_node *)bh->b_data;
2012 key = nilfs_btree_node_get_key(node, 0);
2013 level = nilfs_btree_node_get_level(node);
2015 key = nilfs_bmap_data_get_key(btree, bh);
2016 level = NILFS_BTREE_LEVEL_DATA;
2019 ret = nilfs_btree_do_lookup(btree, path, key, NULL, level + 1, 0);
2021 if (unlikely(ret == -ENOENT))
2022 printk(KERN_CRIT "%s: key = %llu, level == %d\n",
2023 __func__, (unsigned long long)key, level);
2027 ret = NILFS_BMAP_USE_VBN(btree) ?
2028 nilfs_btree_propagate_v(btree, path, level, bh) :
2029 nilfs_btree_propagate_p(btree, path, level, bh);
2032 nilfs_btree_free_path(path);
2037 static int nilfs_btree_propagate_gc(struct nilfs_bmap *btree,
2038 struct buffer_head *bh)
2040 return nilfs_dat_mark_dirty(nilfs_bmap_get_dat(btree), bh->b_blocknr);
2043 static void nilfs_btree_add_dirty_buffer(struct nilfs_bmap *btree,
2044 struct list_head *lists,
2045 struct buffer_head *bh)
2047 struct list_head *head;
2048 struct buffer_head *cbh;
2049 struct nilfs_btree_node *node, *cnode;
2054 node = (struct nilfs_btree_node *)bh->b_data;
2055 key = nilfs_btree_node_get_key(node, 0);
2056 level = nilfs_btree_node_get_level(node);
2057 if (level < NILFS_BTREE_LEVEL_NODE_MIN ||
2058 level >= NILFS_BTREE_LEVEL_MAX) {
2061 "%s: invalid btree level: %d (key=%llu, ino=%lu, "
2063 __func__, level, (unsigned long long)key,
2064 NILFS_BMAP_I(btree)->vfs_inode.i_ino,
2065 (unsigned long long)bh->b_blocknr);
2069 list_for_each(head, &lists[level]) {
2070 cbh = list_entry(head, struct buffer_head, b_assoc_buffers);
2071 cnode = (struct nilfs_btree_node *)cbh->b_data;
2072 ckey = nilfs_btree_node_get_key(cnode, 0);
2076 list_add_tail(&bh->b_assoc_buffers, head);
2079 static void nilfs_btree_lookup_dirty_buffers(struct nilfs_bmap *btree,
2080 struct list_head *listp)
2082 struct address_space *btcache = &NILFS_BMAP_I(btree)->i_btnode_cache;
2083 struct list_head lists[NILFS_BTREE_LEVEL_MAX];
2084 struct pagevec pvec;
2085 struct buffer_head *bh, *head;
2089 for (level = NILFS_BTREE_LEVEL_NODE_MIN;
2090 level < NILFS_BTREE_LEVEL_MAX;
2092 INIT_LIST_HEAD(&lists[level]);
2094 pagevec_init(&pvec, 0);
2096 while (pagevec_lookup_tag(&pvec, btcache, &index, PAGECACHE_TAG_DIRTY,
2098 for (i = 0; i < pagevec_count(&pvec); i++) {
2099 bh = head = page_buffers(pvec.pages[i]);
2101 if (buffer_dirty(bh))
2102 nilfs_btree_add_dirty_buffer(btree,
2104 } while ((bh = bh->b_this_page) != head);
2106 pagevec_release(&pvec);
2110 for (level = NILFS_BTREE_LEVEL_NODE_MIN;
2111 level < NILFS_BTREE_LEVEL_MAX;
2113 list_splice_tail(&lists[level], listp);
2116 static int nilfs_btree_assign_p(struct nilfs_bmap *btree,
2117 struct nilfs_btree_path *path,
2119 struct buffer_head **bh,
2121 union nilfs_binfo *binfo)
2123 struct nilfs_btree_node *parent;
2128 parent = nilfs_btree_get_node(btree, path, level + 1, &ncmax);
2129 ptr = nilfs_btree_node_get_ptr(parent, path[level + 1].bp_index,
2131 if (buffer_nilfs_node(*bh)) {
2132 path[level].bp_ctxt.oldkey = ptr;
2133 path[level].bp_ctxt.newkey = blocknr;
2134 path[level].bp_ctxt.bh = *bh;
2135 ret = nilfs_btnode_prepare_change_key(
2136 &NILFS_BMAP_I(btree)->i_btnode_cache,
2137 &path[level].bp_ctxt);
2140 nilfs_btnode_commit_change_key(
2141 &NILFS_BMAP_I(btree)->i_btnode_cache,
2142 &path[level].bp_ctxt);
2143 *bh = path[level].bp_ctxt.bh;
2146 nilfs_btree_node_set_ptr(parent, path[level + 1].bp_index, blocknr,
2149 key = nilfs_btree_node_get_key(parent, path[level + 1].bp_index);
2150 /* on-disk format */
2151 binfo->bi_dat.bi_blkoff = cpu_to_le64(key);
2152 binfo->bi_dat.bi_level = level;
2157 static int nilfs_btree_assign_v(struct nilfs_bmap *btree,
2158 struct nilfs_btree_path *path,
2160 struct buffer_head **bh,
2162 union nilfs_binfo *binfo)
2164 struct nilfs_btree_node *parent;
2165 struct inode *dat = nilfs_bmap_get_dat(btree);
2168 union nilfs_bmap_ptr_req req;
2171 parent = nilfs_btree_get_node(btree, path, level + 1, &ncmax);
2172 ptr = nilfs_btree_node_get_ptr(parent, path[level + 1].bp_index,
2175 ret = nilfs_dat_prepare_start(dat, &req.bpr_req);
2178 nilfs_dat_commit_start(dat, &req.bpr_req, blocknr);
2180 key = nilfs_btree_node_get_key(parent, path[level + 1].bp_index);
2181 /* on-disk format */
2182 binfo->bi_v.bi_vblocknr = cpu_to_le64(ptr);
2183 binfo->bi_v.bi_blkoff = cpu_to_le64(key);
2188 static int nilfs_btree_assign(struct nilfs_bmap *btree,
2189 struct buffer_head **bh,
2191 union nilfs_binfo *binfo)
2193 struct nilfs_btree_path *path;
2194 struct nilfs_btree_node *node;
2198 path = nilfs_btree_alloc_path();
2202 if (buffer_nilfs_node(*bh)) {
2203 node = (struct nilfs_btree_node *)(*bh)->b_data;
2204 key = nilfs_btree_node_get_key(node, 0);
2205 level = nilfs_btree_node_get_level(node);
2207 key = nilfs_bmap_data_get_key(btree, *bh);
2208 level = NILFS_BTREE_LEVEL_DATA;
2211 ret = nilfs_btree_do_lookup(btree, path, key, NULL, level + 1, 0);
2213 WARN_ON(ret == -ENOENT);
2217 ret = NILFS_BMAP_USE_VBN(btree) ?
2218 nilfs_btree_assign_v(btree, path, level, bh, blocknr, binfo) :
2219 nilfs_btree_assign_p(btree, path, level, bh, blocknr, binfo);
2222 nilfs_btree_free_path(path);
2227 static int nilfs_btree_assign_gc(struct nilfs_bmap *btree,
2228 struct buffer_head **bh,
2230 union nilfs_binfo *binfo)
2232 struct nilfs_btree_node *node;
2236 ret = nilfs_dat_move(nilfs_bmap_get_dat(btree), (*bh)->b_blocknr,
2241 if (buffer_nilfs_node(*bh)) {
2242 node = (struct nilfs_btree_node *)(*bh)->b_data;
2243 key = nilfs_btree_node_get_key(node, 0);
2245 key = nilfs_bmap_data_get_key(btree, *bh);
2247 /* on-disk format */
2248 binfo->bi_v.bi_vblocknr = cpu_to_le64((*bh)->b_blocknr);
2249 binfo->bi_v.bi_blkoff = cpu_to_le64(key);
2254 static int nilfs_btree_mark(struct nilfs_bmap *btree, __u64 key, int level)
2256 struct buffer_head *bh;
2257 struct nilfs_btree_path *path;
2261 path = nilfs_btree_alloc_path();
2265 ret = nilfs_btree_do_lookup(btree, path, key, &ptr, level + 1, 0);
2267 WARN_ON(ret == -ENOENT);
2270 ret = nilfs_btree_get_block(btree, ptr, &bh);
2272 WARN_ON(ret == -ENOENT);
2276 if (!buffer_dirty(bh))
2277 mark_buffer_dirty(bh);
2279 if (!nilfs_bmap_dirty(btree))
2280 nilfs_bmap_set_dirty(btree);
2283 nilfs_btree_free_path(path);
2287 static const struct nilfs_bmap_operations nilfs_btree_ops = {
2288 .bop_lookup = nilfs_btree_lookup,
2289 .bop_lookup_contig = nilfs_btree_lookup_contig,
2290 .bop_insert = nilfs_btree_insert,
2291 .bop_delete = nilfs_btree_delete,
2294 .bop_propagate = nilfs_btree_propagate,
2296 .bop_lookup_dirty_buffers = nilfs_btree_lookup_dirty_buffers,
2298 .bop_assign = nilfs_btree_assign,
2299 .bop_mark = nilfs_btree_mark,
2301 .bop_last_key = nilfs_btree_last_key,
2302 .bop_check_insert = NULL,
2303 .bop_check_delete = nilfs_btree_check_delete,
2304 .bop_gather_data = nilfs_btree_gather_data,
2307 static const struct nilfs_bmap_operations nilfs_btree_ops_gc = {
2309 .bop_lookup_contig = NULL,
2314 .bop_propagate = nilfs_btree_propagate_gc,
2316 .bop_lookup_dirty_buffers = nilfs_btree_lookup_dirty_buffers,
2318 .bop_assign = nilfs_btree_assign_gc,
2321 .bop_last_key = NULL,
2322 .bop_check_insert = NULL,
2323 .bop_check_delete = NULL,
2324 .bop_gather_data = NULL,
2327 static void __nilfs_btree_init(struct nilfs_bmap *bmap)
2329 bmap->b_ops = &nilfs_btree_ops;
2330 bmap->b_nchildren_per_block =
2331 NILFS_BTREE_NODE_NCHILDREN_MAX(nilfs_btree_node_size(bmap));
2334 int nilfs_btree_init(struct nilfs_bmap *bmap)
2338 __nilfs_btree_init(bmap);
2340 if (nilfs_btree_root_broken(nilfs_btree_get_root(bmap),
2341 bmap->b_inode->i_ino))
2346 void nilfs_btree_init_gc(struct nilfs_bmap *bmap)
2348 bmap->b_ops = &nilfs_btree_ops_gc;
2349 bmap->b_nchildren_per_block =
2350 NILFS_BTREE_NODE_NCHILDREN_MAX(nilfs_btree_node_size(bmap));