xref: /linux/fs/gfs2/dir.c (revision 8f8d5745bb520c76b81abef4a2cb3023d0313bfd)
1 /*
2  * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
3  * Copyright (C) 2004-2006 Red Hat, Inc.  All rights reserved.
4  *
5  * This copyrighted material is made available to anyone wishing to use,
6  * modify, copy, or redistribute it subject to the terms and conditions
7  * of the GNU General Public License version 2.
8  */
9 
10 /*
11  * Implements Extendible Hashing as described in:
12  *   "Extendible Hashing" by Fagin, et al in
13  *     __ACM Trans. on Database Systems__, Sept 1979.
14  *
15  *
16  * Here's the layout of dirents which is essentially the same as that of ext2
17  * within a single block. The field de_name_len is the number of bytes
18  * actually required for the name (no null terminator). The field de_rec_len
19  * is the number of bytes allocated to the dirent. The offset of the next
20  * dirent in the block is (dirent + dirent->de_rec_len). When a dirent is
21  * deleted, the preceding dirent inherits its allocated space, ie
22  * prev->de_rec_len += deleted->de_rec_len. Since the next dirent is obtained
23  * by adding de_rec_len to the current dirent, this essentially causes the
24  * deleted dirent to get jumped over when iterating through all the dirents.
25  *
26  * When deleting the first dirent in a block, there is no previous dirent so
27  * the field de_ino is set to zero to designate it as deleted. When allocating
28  * a dirent, gfs2_dirent_alloc iterates through the dirents in a block. If the
29  * first dirent has (de_ino == 0) and de_rec_len is large enough, this first
30  * dirent is allocated. Otherwise it must go through all the 'used' dirents
31  * searching for one in which the amount of total space minus the amount of
32  * used space will provide enough space for the new dirent.
33  *
34  * There are two types of blocks in which dirents reside. In a stuffed dinode,
35  * the dirents begin at offset sizeof(struct gfs2_dinode) from the beginning of
36  * the block.  In leaves, they begin at offset sizeof(struct gfs2_leaf) from the
37  * beginning of the leaf block. The dirents reside in leaves when
38  *
39  * dip->i_diskflags & GFS2_DIF_EXHASH is true
40  *
41  * Otherwise, the dirents are "linear", within a single stuffed dinode block.
42  *
43  * When the dirents are in leaves, the actual contents of the directory file are
44  * used as an array of 64-bit block pointers pointing to the leaf blocks. The
45  * dirents are NOT in the directory file itself. There can be more than one
46  * block pointer in the array that points to the same leaf. In fact, when a
47  * directory is first converted from linear to exhash, all of the pointers
48  * point to the same leaf.
49  *
50  * When a leaf is completely full, the size of the hash table can be
51  * doubled unless it is already at the maximum size which is hard coded into
52  * GFS2_DIR_MAX_DEPTH. After that, leaves are chained together in a linked list,
53  * but never before the maximum hash table size has been reached.
54  */
55 
56 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
57 
58 #include <linux/slab.h>
59 #include <linux/spinlock.h>
60 #include <linux/buffer_head.h>
61 #include <linux/sort.h>
62 #include <linux/gfs2_ondisk.h>
63 #include <linux/crc32.h>
64 #include <linux/vmalloc.h>
65 #include <linux/bio.h>
66 
67 #include "gfs2.h"
68 #include "incore.h"
69 #include "dir.h"
70 #include "glock.h"
71 #include "inode.h"
72 #include "meta_io.h"
73 #include "quota.h"
74 #include "rgrp.h"
75 #include "trans.h"
76 #include "bmap.h"
77 #include "util.h"
78 
79 #define IS_LEAF     1 /* Hashed (leaf) directory */
80 #define IS_DINODE   2 /* Linear (stuffed dinode block) directory */
81 
82 #define MAX_RA_BLOCKS 32 /* max read-ahead blocks */
83 
84 #define gfs2_disk_hash2offset(h) (((u64)(h)) >> 1)
85 #define gfs2_dir_offset2hash(p) ((u32)(((u64)(p)) << 1))
86 #define GFS2_HASH_INDEX_MASK 0xffffc000
87 #define GFS2_USE_HASH_FLAG 0x2000
88 
89 struct qstr gfs2_qdot __read_mostly;
90 struct qstr gfs2_qdotdot __read_mostly;
91 
92 typedef int (*gfs2_dscan_t)(const struct gfs2_dirent *dent,
93 			    const struct qstr *name, void *opaque);
94 
95 int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block,
96 			    struct buffer_head **bhp)
97 {
98 	struct buffer_head *bh;
99 
100 	bh = gfs2_meta_new(ip->i_gl, block);
101 	gfs2_trans_add_meta(ip->i_gl, bh);
102 	gfs2_metatype_set(bh, GFS2_METATYPE_JD, GFS2_FORMAT_JD);
103 	gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header));
104 	*bhp = bh;
105 	return 0;
106 }
107 
108 static int gfs2_dir_get_existing_buffer(struct gfs2_inode *ip, u64 block,
109 					struct buffer_head **bhp)
110 {
111 	struct buffer_head *bh;
112 	int error;
113 
114 	error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh);
115 	if (error)
116 		return error;
117 	if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) {
118 		brelse(bh);
119 		return -EIO;
120 	}
121 	*bhp = bh;
122 	return 0;
123 }
124 
125 static int gfs2_dir_write_stuffed(struct gfs2_inode *ip, const char *buf,
126 				  unsigned int offset, unsigned int size)
127 {
128 	struct buffer_head *dibh;
129 	int error;
130 
131 	error = gfs2_meta_inode_buffer(ip, &dibh);
132 	if (error)
133 		return error;
134 
135 	gfs2_trans_add_meta(ip->i_gl, dibh);
136 	memcpy(dibh->b_data + offset + sizeof(struct gfs2_dinode), buf, size);
137 	if (ip->i_inode.i_size < offset + size)
138 		i_size_write(&ip->i_inode, offset + size);
139 	ip->i_inode.i_mtime = ip->i_inode.i_ctime = current_time(&ip->i_inode);
140 	gfs2_dinode_out(ip, dibh->b_data);
141 
142 	brelse(dibh);
143 
144 	return size;
145 }
146 
147 
148 
149 /**
150  * gfs2_dir_write_data - Write directory information to the inode
151  * @ip: The GFS2 inode
152  * @buf: The buffer containing information to be written
153  * @offset: The file offset to start writing at
154  * @size: The amount of data to write
155  *
156  * Returns: The number of bytes correctly written or error code
157  */
158 static int gfs2_dir_write_data(struct gfs2_inode *ip, const char *buf,
159 			       u64 offset, unsigned int size)
160 {
161 	struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
162 	struct buffer_head *dibh;
163 	u64 lblock, dblock;
164 	u32 extlen = 0;
165 	unsigned int o;
166 	int copied = 0;
167 	int error = 0;
168 	int new = 0;
169 
170 	if (!size)
171 		return 0;
172 
173 	if (gfs2_is_stuffed(ip) && offset + size <= gfs2_max_stuffed_size(ip))
174 		return gfs2_dir_write_stuffed(ip, buf, (unsigned int)offset,
175 					      size);
176 
177 	if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
178 		return -EINVAL;
179 
180 	if (gfs2_is_stuffed(ip)) {
181 		error = gfs2_unstuff_dinode(ip, NULL);
182 		if (error)
183 			return error;
184 	}
185 
186 	lblock = offset;
187 	o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header);
188 
189 	while (copied < size) {
190 		unsigned int amount;
191 		struct buffer_head *bh;
192 
193 		amount = size - copied;
194 		if (amount > sdp->sd_sb.sb_bsize - o)
195 			amount = sdp->sd_sb.sb_bsize - o;
196 
197 		if (!extlen) {
198 			new = 1;
199 			error = gfs2_extent_map(&ip->i_inode, lblock, &new,
200 						&dblock, &extlen);
201 			if (error)
202 				goto fail;
203 			error = -EIO;
204 			if (gfs2_assert_withdraw(sdp, dblock))
205 				goto fail;
206 		}
207 
208 		if (amount == sdp->sd_jbsize || new)
209 			error = gfs2_dir_get_new_buffer(ip, dblock, &bh);
210 		else
211 			error = gfs2_dir_get_existing_buffer(ip, dblock, &bh);
212 
213 		if (error)
214 			goto fail;
215 
216 		gfs2_trans_add_meta(ip->i_gl, bh);
217 		memcpy(bh->b_data + o, buf, amount);
218 		brelse(bh);
219 
220 		buf += amount;
221 		copied += amount;
222 		lblock++;
223 		dblock++;
224 		extlen--;
225 
226 		o = sizeof(struct gfs2_meta_header);
227 	}
228 
229 out:
230 	error = gfs2_meta_inode_buffer(ip, &dibh);
231 	if (error)
232 		return error;
233 
234 	if (ip->i_inode.i_size < offset + copied)
235 		i_size_write(&ip->i_inode, offset + copied);
236 	ip->i_inode.i_mtime = ip->i_inode.i_ctime = current_time(&ip->i_inode);
237 
238 	gfs2_trans_add_meta(ip->i_gl, dibh);
239 	gfs2_dinode_out(ip, dibh->b_data);
240 	brelse(dibh);
241 
242 	return copied;
243 fail:
244 	if (copied)
245 		goto out;
246 	return error;
247 }
248 
249 static int gfs2_dir_read_stuffed(struct gfs2_inode *ip, __be64 *buf,
250 				 unsigned int size)
251 {
252 	struct buffer_head *dibh;
253 	int error;
254 
255 	error = gfs2_meta_inode_buffer(ip, &dibh);
256 	if (!error) {
257 		memcpy(buf, dibh->b_data + sizeof(struct gfs2_dinode), size);
258 		brelse(dibh);
259 	}
260 
261 	return (error) ? error : size;
262 }
263 
264 
265 /**
266  * gfs2_dir_read_data - Read a data from a directory inode
267  * @ip: The GFS2 Inode
268  * @buf: The buffer to place result into
269  * @size: Amount of data to transfer
270  *
271  * Returns: The amount of data actually copied or the error
272  */
273 static int gfs2_dir_read_data(struct gfs2_inode *ip, __be64 *buf,
274 			      unsigned int size)
275 {
276 	struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
277 	u64 lblock, dblock;
278 	u32 extlen = 0;
279 	unsigned int o;
280 	int copied = 0;
281 	int error = 0;
282 
283 	if (gfs2_is_stuffed(ip))
284 		return gfs2_dir_read_stuffed(ip, buf, size);
285 
286 	if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
287 		return -EINVAL;
288 
289 	lblock = 0;
290 	o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header);
291 
292 	while (copied < size) {
293 		unsigned int amount;
294 		struct buffer_head *bh;
295 		int new;
296 
297 		amount = size - copied;
298 		if (amount > sdp->sd_sb.sb_bsize - o)
299 			amount = sdp->sd_sb.sb_bsize - o;
300 
301 		if (!extlen) {
302 			new = 0;
303 			error = gfs2_extent_map(&ip->i_inode, lblock, &new,
304 						&dblock, &extlen);
305 			if (error || !dblock)
306 				goto fail;
307 			BUG_ON(extlen < 1);
308 			bh = gfs2_meta_ra(ip->i_gl, dblock, extlen);
309 		} else {
310 			error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, 0, &bh);
311 			if (error)
312 				goto fail;
313 		}
314 		error = gfs2_metatype_check(sdp, bh, GFS2_METATYPE_JD);
315 		if (error) {
316 			brelse(bh);
317 			goto fail;
318 		}
319 		dblock++;
320 		extlen--;
321 		memcpy(buf, bh->b_data + o, amount);
322 		brelse(bh);
323 		buf += (amount/sizeof(__be64));
324 		copied += amount;
325 		lblock++;
326 		o = sizeof(struct gfs2_meta_header);
327 	}
328 
329 	return copied;
330 fail:
331 	return (copied) ? copied : error;
332 }
333 
334 /**
335  * gfs2_dir_get_hash_table - Get pointer to the dir hash table
336  * @ip: The inode in question
337  *
338  * Returns: The hash table or an error
339  */
340 
341 static __be64 *gfs2_dir_get_hash_table(struct gfs2_inode *ip)
342 {
343 	struct inode *inode = &ip->i_inode;
344 	int ret;
345 	u32 hsize;
346 	__be64 *hc;
347 
348 	BUG_ON(!(ip->i_diskflags & GFS2_DIF_EXHASH));
349 
350 	hc = ip->i_hash_cache;
351 	if (hc)
352 		return hc;
353 
354 	hsize = BIT(ip->i_depth);
355 	hsize *= sizeof(__be64);
356 	if (hsize != i_size_read(&ip->i_inode)) {
357 		gfs2_consist_inode(ip);
358 		return ERR_PTR(-EIO);
359 	}
360 
361 	hc = kmalloc(hsize, GFP_NOFS | __GFP_NOWARN);
362 	if (hc == NULL)
363 		hc = __vmalloc(hsize, GFP_NOFS, PAGE_KERNEL);
364 
365 	if (hc == NULL)
366 		return ERR_PTR(-ENOMEM);
367 
368 	ret = gfs2_dir_read_data(ip, hc, hsize);
369 	if (ret < 0) {
370 		kvfree(hc);
371 		return ERR_PTR(ret);
372 	}
373 
374 	spin_lock(&inode->i_lock);
375 	if (likely(!ip->i_hash_cache)) {
376 		ip->i_hash_cache = hc;
377 		hc = NULL;
378 	}
379 	spin_unlock(&inode->i_lock);
380 	kvfree(hc);
381 
382 	return ip->i_hash_cache;
383 }
384 
385 /**
386  * gfs2_dir_hash_inval - Invalidate dir hash
387  * @ip: The directory inode
388  *
389  * Must be called with an exclusive glock, or during glock invalidation.
390  */
391 void gfs2_dir_hash_inval(struct gfs2_inode *ip)
392 {
393 	__be64 *hc;
394 
395 	spin_lock(&ip->i_inode.i_lock);
396 	hc = ip->i_hash_cache;
397 	ip->i_hash_cache = NULL;
398 	spin_unlock(&ip->i_inode.i_lock);
399 
400 	kvfree(hc);
401 }
402 
403 static inline int gfs2_dirent_sentinel(const struct gfs2_dirent *dent)
404 {
405 	return dent->de_inum.no_addr == 0 || dent->de_inum.no_formal_ino == 0;
406 }
407 
408 static inline int __gfs2_dirent_find(const struct gfs2_dirent *dent,
409 				     const struct qstr *name, int ret)
410 {
411 	if (!gfs2_dirent_sentinel(dent) &&
412 	    be32_to_cpu(dent->de_hash) == name->hash &&
413 	    be16_to_cpu(dent->de_name_len) == name->len &&
414 	    memcmp(dent+1, name->name, name->len) == 0)
415 		return ret;
416 	return 0;
417 }
418 
419 static int gfs2_dirent_find(const struct gfs2_dirent *dent,
420 			    const struct qstr *name,
421 			    void *opaque)
422 {
423 	return __gfs2_dirent_find(dent, name, 1);
424 }
425 
426 static int gfs2_dirent_prev(const struct gfs2_dirent *dent,
427 			    const struct qstr *name,
428 			    void *opaque)
429 {
430 	return __gfs2_dirent_find(dent, name, 2);
431 }
432 
433 /*
434  * name->name holds ptr to start of block.
435  * name->len holds size of block.
436  */
437 static int gfs2_dirent_last(const struct gfs2_dirent *dent,
438 			    const struct qstr *name,
439 			    void *opaque)
440 {
441 	const char *start = name->name;
442 	const char *end = (const char *)dent + be16_to_cpu(dent->de_rec_len);
443 	if (name->len == (end - start))
444 		return 1;
445 	return 0;
446 }
447 
448 /* Look for the dirent that contains the offset specified in data. Once we
449  * find that dirent, there must be space available there for the new dirent */
450 static int gfs2_dirent_find_offset(const struct gfs2_dirent *dent,
451 				  const struct qstr *name,
452 				  void *ptr)
453 {
454 	unsigned required = GFS2_DIRENT_SIZE(name->len);
455 	unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
456 	unsigned totlen = be16_to_cpu(dent->de_rec_len);
457 
458 	if (ptr < (void *)dent || ptr >= (void *)dent + totlen)
459 		return 0;
460 	if (gfs2_dirent_sentinel(dent))
461 		actual = 0;
462 	if (ptr < (void *)dent + actual)
463 		return -1;
464 	if ((void *)dent + totlen >= ptr + required)
465 		return 1;
466 	return -1;
467 }
468 
469 static int gfs2_dirent_find_space(const struct gfs2_dirent *dent,
470 				  const struct qstr *name,
471 				  void *opaque)
472 {
473 	unsigned required = GFS2_DIRENT_SIZE(name->len);
474 	unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
475 	unsigned totlen = be16_to_cpu(dent->de_rec_len);
476 
477 	if (gfs2_dirent_sentinel(dent))
478 		actual = 0;
479 	if (totlen - actual >= required)
480 		return 1;
481 	return 0;
482 }
483 
484 struct dirent_gather {
485 	const struct gfs2_dirent **pdent;
486 	unsigned offset;
487 };
488 
489 static int gfs2_dirent_gather(const struct gfs2_dirent *dent,
490 			      const struct qstr *name,
491 			      void *opaque)
492 {
493 	struct dirent_gather *g = opaque;
494 	if (!gfs2_dirent_sentinel(dent)) {
495 		g->pdent[g->offset++] = dent;
496 	}
497 	return 0;
498 }
499 
500 /*
501  * Other possible things to check:
502  * - Inode located within filesystem size (and on valid block)
503  * - Valid directory entry type
504  * Not sure how heavy-weight we want to make this... could also check
505  * hash is correct for example, but that would take a lot of extra time.
506  * For now the most important thing is to check that the various sizes
507  * are correct.
508  */
509 static int gfs2_check_dirent(struct gfs2_sbd *sdp,
510 			     struct gfs2_dirent *dent, unsigned int offset,
511 			     unsigned int size, unsigned int len, int first)
512 {
513 	const char *msg = "gfs2_dirent too small";
514 	if (unlikely(size < sizeof(struct gfs2_dirent)))
515 		goto error;
516 	msg = "gfs2_dirent misaligned";
517 	if (unlikely(offset & 0x7))
518 		goto error;
519 	msg = "gfs2_dirent points beyond end of block";
520 	if (unlikely(offset + size > len))
521 		goto error;
522 	msg = "zero inode number";
523 	if (unlikely(!first && gfs2_dirent_sentinel(dent)))
524 		goto error;
525 	msg = "name length is greater than space in dirent";
526 	if (!gfs2_dirent_sentinel(dent) &&
527 	    unlikely(sizeof(struct gfs2_dirent)+be16_to_cpu(dent->de_name_len) >
528 		     size))
529 		goto error;
530 	return 0;
531 error:
532 	fs_warn(sdp, "%s: %s (%s)\n",
533 		__func__, msg, first ? "first in block" : "not first in block");
534 	return -EIO;
535 }
536 
537 static int gfs2_dirent_offset(struct gfs2_sbd *sdp, const void *buf)
538 {
539 	const struct gfs2_meta_header *h = buf;
540 	int offset;
541 
542 	BUG_ON(buf == NULL);
543 
544 	switch(be32_to_cpu(h->mh_type)) {
545 	case GFS2_METATYPE_LF:
546 		offset = sizeof(struct gfs2_leaf);
547 		break;
548 	case GFS2_METATYPE_DI:
549 		offset = sizeof(struct gfs2_dinode);
550 		break;
551 	default:
552 		goto wrong_type;
553 	}
554 	return offset;
555 wrong_type:
556 	fs_warn(sdp, "%s: wrong block type %u\n", __func__,
557 		be32_to_cpu(h->mh_type));
558 	return -1;
559 }
560 
561 static struct gfs2_dirent *gfs2_dirent_scan(struct inode *inode, void *buf,
562 					    unsigned int len, gfs2_dscan_t scan,
563 					    const struct qstr *name,
564 					    void *opaque)
565 {
566 	struct gfs2_dirent *dent, *prev;
567 	unsigned offset;
568 	unsigned size;
569 	int ret = 0;
570 
571 	ret = gfs2_dirent_offset(GFS2_SB(inode), buf);
572 	if (ret < 0)
573 		goto consist_inode;
574 
575 	offset = ret;
576 	prev = NULL;
577 	dent = buf + offset;
578 	size = be16_to_cpu(dent->de_rec_len);
579 	if (gfs2_check_dirent(GFS2_SB(inode), dent, offset, size, len, 1))
580 		goto consist_inode;
581 	do {
582 		ret = scan(dent, name, opaque);
583 		if (ret)
584 			break;
585 		offset += size;
586 		if (offset == len)
587 			break;
588 		prev = dent;
589 		dent = buf + offset;
590 		size = be16_to_cpu(dent->de_rec_len);
591 		if (gfs2_check_dirent(GFS2_SB(inode), dent, offset, size,
592 				      len, 0))
593 			goto consist_inode;
594 	} while(1);
595 
596 	switch(ret) {
597 	case 0:
598 		return NULL;
599 	case 1:
600 		return dent;
601 	case 2:
602 		return prev ? prev : dent;
603 	default:
604 		BUG_ON(ret > 0);
605 		return ERR_PTR(ret);
606 	}
607 
608 consist_inode:
609 	gfs2_consist_inode(GFS2_I(inode));
610 	return ERR_PTR(-EIO);
611 }
612 
613 static int dirent_check_reclen(struct gfs2_inode *dip,
614 			       const struct gfs2_dirent *d, const void *end_p)
615 {
616 	const void *ptr = d;
617 	u16 rec_len = be16_to_cpu(d->de_rec_len);
618 
619 	if (unlikely(rec_len < sizeof(struct gfs2_dirent)))
620 		goto broken;
621 	ptr += rec_len;
622 	if (ptr < end_p)
623 		return rec_len;
624 	if (ptr == end_p)
625 		return -ENOENT;
626 broken:
627 	gfs2_consist_inode(dip);
628 	return -EIO;
629 }
630 
631 /**
632  * dirent_next - Next dirent
633  * @dip: the directory
634  * @bh: The buffer
635  * @dent: Pointer to list of dirents
636  *
637  * Returns: 0 on success, error code otherwise
638  */
639 
640 static int dirent_next(struct gfs2_inode *dip, struct buffer_head *bh,
641 		       struct gfs2_dirent **dent)
642 {
643 	struct gfs2_dirent *cur = *dent, *tmp;
644 	char *bh_end = bh->b_data + bh->b_size;
645 	int ret;
646 
647 	ret = dirent_check_reclen(dip, cur, bh_end);
648 	if (ret < 0)
649 		return ret;
650 
651 	tmp = (void *)cur + ret;
652 	ret = dirent_check_reclen(dip, tmp, bh_end);
653 	if (ret == -EIO)
654 		return ret;
655 
656         /* Only the first dent could ever have de_inum.no_addr == 0 */
657 	if (gfs2_dirent_sentinel(tmp)) {
658 		gfs2_consist_inode(dip);
659 		return -EIO;
660 	}
661 
662 	*dent = tmp;
663 	return 0;
664 }
665 
666 /**
667  * dirent_del - Delete a dirent
668  * @dip: The GFS2 inode
669  * @bh: The buffer
670  * @prev: The previous dirent
671  * @cur: The current dirent
672  *
673  */
674 
675 static void dirent_del(struct gfs2_inode *dip, struct buffer_head *bh,
676 		       struct gfs2_dirent *prev, struct gfs2_dirent *cur)
677 {
678 	u16 cur_rec_len, prev_rec_len;
679 
680 	if (gfs2_dirent_sentinel(cur)) {
681 		gfs2_consist_inode(dip);
682 		return;
683 	}
684 
685 	gfs2_trans_add_meta(dip->i_gl, bh);
686 
687 	/* If there is no prev entry, this is the first entry in the block.
688 	   The de_rec_len is already as big as it needs to be.  Just zero
689 	   out the inode number and return.  */
690 
691 	if (!prev) {
692 		cur->de_inum.no_addr = 0;
693 		cur->de_inum.no_formal_ino = 0;
694 		return;
695 	}
696 
697 	/*  Combine this dentry with the previous one.  */
698 
699 	prev_rec_len = be16_to_cpu(prev->de_rec_len);
700 	cur_rec_len = be16_to_cpu(cur->de_rec_len);
701 
702 	if ((char *)prev + prev_rec_len != (char *)cur)
703 		gfs2_consist_inode(dip);
704 	if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size)
705 		gfs2_consist_inode(dip);
706 
707 	prev_rec_len += cur_rec_len;
708 	prev->de_rec_len = cpu_to_be16(prev_rec_len);
709 }
710 
711 
712 static struct gfs2_dirent *do_init_dirent(struct inode *inode,
713 					  struct gfs2_dirent *dent,
714 					  const struct qstr *name,
715 					  struct buffer_head *bh,
716 					  unsigned offset)
717 {
718 	struct gfs2_inode *ip = GFS2_I(inode);
719 	struct gfs2_dirent *ndent;
720 	unsigned totlen;
721 
722 	totlen = be16_to_cpu(dent->de_rec_len);
723 	BUG_ON(offset + name->len > totlen);
724 	gfs2_trans_add_meta(ip->i_gl, bh);
725 	ndent = (struct gfs2_dirent *)((char *)dent + offset);
726 	dent->de_rec_len = cpu_to_be16(offset);
727 	gfs2_qstr2dirent(name, totlen - offset, ndent);
728 	return ndent;
729 }
730 
731 
732 /*
733  * Takes a dent from which to grab space as an argument. Returns the
734  * newly created dent.
735  */
736 static struct gfs2_dirent *gfs2_init_dirent(struct inode *inode,
737 					    struct gfs2_dirent *dent,
738 					    const struct qstr *name,
739 					    struct buffer_head *bh)
740 {
741 	unsigned offset = 0;
742 
743 	if (!gfs2_dirent_sentinel(dent))
744 		offset = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
745 	return do_init_dirent(inode, dent, name, bh, offset);
746 }
747 
748 static struct gfs2_dirent *gfs2_dirent_split_alloc(struct inode *inode,
749 						   struct buffer_head *bh,
750 						   const struct qstr *name,
751 						   void *ptr)
752 {
753 	struct gfs2_dirent *dent;
754 	dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
755 				gfs2_dirent_find_offset, name, ptr);
756 	if (!dent || IS_ERR(dent))
757 		return dent;
758 	return do_init_dirent(inode, dent, name, bh,
759 			      (unsigned)(ptr - (void *)dent));
760 }
761 
762 static int get_leaf(struct gfs2_inode *dip, u64 leaf_no,
763 		    struct buffer_head **bhp)
764 {
765 	int error;
766 
767 	error = gfs2_meta_read(dip->i_gl, leaf_no, DIO_WAIT, 0, bhp);
768 	if (!error && gfs2_metatype_check(GFS2_SB(&dip->i_inode), *bhp, GFS2_METATYPE_LF)) {
769 		/* pr_info("block num=%llu\n", leaf_no); */
770 		error = -EIO;
771 	}
772 
773 	return error;
774 }
775 
776 /**
777  * get_leaf_nr - Get a leaf number associated with the index
778  * @dip: The GFS2 inode
779  * @index:
780  * @leaf_out:
781  *
782  * Returns: 0 on success, error code otherwise
783  */
784 
785 static int get_leaf_nr(struct gfs2_inode *dip, u32 index,
786 		       u64 *leaf_out)
787 {
788 	__be64 *hash;
789 	int error;
790 
791 	hash = gfs2_dir_get_hash_table(dip);
792 	error = PTR_ERR_OR_ZERO(hash);
793 
794 	if (!error)
795 		*leaf_out = be64_to_cpu(*(hash + index));
796 
797 	return error;
798 }
799 
800 static int get_first_leaf(struct gfs2_inode *dip, u32 index,
801 			  struct buffer_head **bh_out)
802 {
803 	u64 leaf_no;
804 	int error;
805 
806 	error = get_leaf_nr(dip, index, &leaf_no);
807 	if (!error)
808 		error = get_leaf(dip, leaf_no, bh_out);
809 
810 	return error;
811 }
812 
813 static struct gfs2_dirent *gfs2_dirent_search(struct inode *inode,
814 					      const struct qstr *name,
815 					      gfs2_dscan_t scan,
816 					      struct buffer_head **pbh)
817 {
818 	struct buffer_head *bh;
819 	struct gfs2_dirent *dent;
820 	struct gfs2_inode *ip = GFS2_I(inode);
821 	int error;
822 
823 	if (ip->i_diskflags & GFS2_DIF_EXHASH) {
824 		struct gfs2_leaf *leaf;
825 		unsigned int hsize = BIT(ip->i_depth);
826 		unsigned int index;
827 		u64 ln;
828 		if (hsize * sizeof(u64) != i_size_read(inode)) {
829 			gfs2_consist_inode(ip);
830 			return ERR_PTR(-EIO);
831 		}
832 
833 		index = name->hash >> (32 - ip->i_depth);
834 		error = get_first_leaf(ip, index, &bh);
835 		if (error)
836 			return ERR_PTR(error);
837 		do {
838 			dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
839 						scan, name, NULL);
840 			if (dent)
841 				goto got_dent;
842 			leaf = (struct gfs2_leaf *)bh->b_data;
843 			ln = be64_to_cpu(leaf->lf_next);
844 			brelse(bh);
845 			if (!ln)
846 				break;
847 
848 			error = get_leaf(ip, ln, &bh);
849 		} while(!error);
850 
851 		return error ? ERR_PTR(error) : NULL;
852 	}
853 
854 
855 	error = gfs2_meta_inode_buffer(ip, &bh);
856 	if (error)
857 		return ERR_PTR(error);
858 	dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, scan, name, NULL);
859 got_dent:
860 	if (unlikely(dent == NULL || IS_ERR(dent))) {
861 		brelse(bh);
862 		bh = NULL;
863 	}
864 	*pbh = bh;
865 	return dent;
866 }
867 
868 static struct gfs2_leaf *new_leaf(struct inode *inode, struct buffer_head **pbh, u16 depth)
869 {
870 	struct gfs2_inode *ip = GFS2_I(inode);
871 	unsigned int n = 1;
872 	u64 bn;
873 	int error;
874 	struct buffer_head *bh;
875 	struct gfs2_leaf *leaf;
876 	struct gfs2_dirent *dent;
877 	struct timespec64 tv = current_time(inode);
878 
879 	error = gfs2_alloc_blocks(ip, &bn, &n, 0, NULL);
880 	if (error)
881 		return NULL;
882 	bh = gfs2_meta_new(ip->i_gl, bn);
883 	if (!bh)
884 		return NULL;
885 
886 	gfs2_trans_add_unrevoke(GFS2_SB(inode), bn, 1);
887 	gfs2_trans_add_meta(ip->i_gl, bh);
888 	gfs2_metatype_set(bh, GFS2_METATYPE_LF, GFS2_FORMAT_LF);
889 	leaf = (struct gfs2_leaf *)bh->b_data;
890 	leaf->lf_depth = cpu_to_be16(depth);
891 	leaf->lf_entries = 0;
892 	leaf->lf_dirent_format = cpu_to_be32(GFS2_FORMAT_DE);
893 	leaf->lf_next = 0;
894 	leaf->lf_inode = cpu_to_be64(ip->i_no_addr);
895 	leaf->lf_dist = cpu_to_be32(1);
896 	leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
897 	leaf->lf_sec = cpu_to_be64(tv.tv_sec);
898 	memset(leaf->lf_reserved2, 0, sizeof(leaf->lf_reserved2));
899 	dent = (struct gfs2_dirent *)(leaf+1);
900 	gfs2_qstr2dirent(&empty_name, bh->b_size - sizeof(struct gfs2_leaf), dent);
901 	*pbh = bh;
902 	return leaf;
903 }
904 
905 /**
906  * dir_make_exhash - Convert a stuffed directory into an ExHash directory
907  * @dip: The GFS2 inode
908  *
909  * Returns: 0 on success, error code otherwise
910  */
911 
912 static int dir_make_exhash(struct inode *inode)
913 {
914 	struct gfs2_inode *dip = GFS2_I(inode);
915 	struct gfs2_sbd *sdp = GFS2_SB(inode);
916 	struct gfs2_dirent *dent;
917 	struct qstr args;
918 	struct buffer_head *bh, *dibh;
919 	struct gfs2_leaf *leaf;
920 	int y;
921 	u32 x;
922 	__be64 *lp;
923 	u64 bn;
924 	int error;
925 
926 	error = gfs2_meta_inode_buffer(dip, &dibh);
927 	if (error)
928 		return error;
929 
930 	/*  Turn over a new leaf  */
931 
932 	leaf = new_leaf(inode, &bh, 0);
933 	if (!leaf)
934 		return -ENOSPC;
935 	bn = bh->b_blocknr;
936 
937 	gfs2_assert(sdp, dip->i_entries < BIT(16));
938 	leaf->lf_entries = cpu_to_be16(dip->i_entries);
939 
940 	/*  Copy dirents  */
941 
942 	gfs2_buffer_copy_tail(bh, sizeof(struct gfs2_leaf), dibh,
943 			     sizeof(struct gfs2_dinode));
944 
945 	/*  Find last entry  */
946 
947 	x = 0;
948 	args.len = bh->b_size - sizeof(struct gfs2_dinode) +
949 		   sizeof(struct gfs2_leaf);
950 	args.name = bh->b_data;
951 	dent = gfs2_dirent_scan(&dip->i_inode, bh->b_data, bh->b_size,
952 				gfs2_dirent_last, &args, NULL);
953 	if (!dent) {
954 		brelse(bh);
955 		brelse(dibh);
956 		return -EIO;
957 	}
958 	if (IS_ERR(dent)) {
959 		brelse(bh);
960 		brelse(dibh);
961 		return PTR_ERR(dent);
962 	}
963 
964 	/*  Adjust the last dirent's record length
965 	   (Remember that dent still points to the last entry.)  */
966 
967 	dent->de_rec_len = cpu_to_be16(be16_to_cpu(dent->de_rec_len) +
968 		sizeof(struct gfs2_dinode) -
969 		sizeof(struct gfs2_leaf));
970 
971 	brelse(bh);
972 
973 	/*  We're done with the new leaf block, now setup the new
974 	    hash table.  */
975 
976 	gfs2_trans_add_meta(dip->i_gl, dibh);
977 	gfs2_buffer_clear_tail(dibh, sizeof(struct gfs2_dinode));
978 
979 	lp = (__be64 *)(dibh->b_data + sizeof(struct gfs2_dinode));
980 
981 	for (x = sdp->sd_hash_ptrs; x--; lp++)
982 		*lp = cpu_to_be64(bn);
983 
984 	i_size_write(inode, sdp->sd_sb.sb_bsize / 2);
985 	gfs2_add_inode_blocks(&dip->i_inode, 1);
986 	dip->i_diskflags |= GFS2_DIF_EXHASH;
987 
988 	for (x = sdp->sd_hash_ptrs, y = -1; x; x >>= 1, y++) ;
989 	dip->i_depth = y;
990 
991 	gfs2_dinode_out(dip, dibh->b_data);
992 
993 	brelse(dibh);
994 
995 	return 0;
996 }
997 
998 /**
999  * dir_split_leaf - Split a leaf block into two
1000  * @dip: The GFS2 inode
1001  * @index:
1002  * @leaf_no:
1003  *
1004  * Returns: 0 on success, error code on failure
1005  */
1006 
1007 static int dir_split_leaf(struct inode *inode, const struct qstr *name)
1008 {
1009 	struct gfs2_inode *dip = GFS2_I(inode);
1010 	struct buffer_head *nbh, *obh, *dibh;
1011 	struct gfs2_leaf *nleaf, *oleaf;
1012 	struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new;
1013 	u32 start, len, half_len, divider;
1014 	u64 bn, leaf_no;
1015 	__be64 *lp;
1016 	u32 index;
1017 	int x;
1018 	int error;
1019 
1020 	index = name->hash >> (32 - dip->i_depth);
1021 	error = get_leaf_nr(dip, index, &leaf_no);
1022 	if (error)
1023 		return error;
1024 
1025 	/*  Get the old leaf block  */
1026 	error = get_leaf(dip, leaf_no, &obh);
1027 	if (error)
1028 		return error;
1029 
1030 	oleaf = (struct gfs2_leaf *)obh->b_data;
1031 	if (dip->i_depth == be16_to_cpu(oleaf->lf_depth)) {
1032 		brelse(obh);
1033 		return 1; /* can't split */
1034 	}
1035 
1036 	gfs2_trans_add_meta(dip->i_gl, obh);
1037 
1038 	nleaf = new_leaf(inode, &nbh, be16_to_cpu(oleaf->lf_depth) + 1);
1039 	if (!nleaf) {
1040 		brelse(obh);
1041 		return -ENOSPC;
1042 	}
1043 	bn = nbh->b_blocknr;
1044 
1045 	/*  Compute the start and len of leaf pointers in the hash table.  */
1046 	len = BIT(dip->i_depth - be16_to_cpu(oleaf->lf_depth));
1047 	half_len = len >> 1;
1048 	if (!half_len) {
1049 		fs_warn(GFS2_SB(inode), "i_depth %u lf_depth %u index %u\n",
1050 			dip->i_depth, be16_to_cpu(oleaf->lf_depth), index);
1051 		gfs2_consist_inode(dip);
1052 		error = -EIO;
1053 		goto fail_brelse;
1054 	}
1055 
1056 	start = (index & ~(len - 1));
1057 
1058 	/* Change the pointers.
1059 	   Don't bother distinguishing stuffed from non-stuffed.
1060 	   This code is complicated enough already. */
1061 	lp = kmalloc_array(half_len, sizeof(__be64), GFP_NOFS);
1062 	if (!lp) {
1063 		error = -ENOMEM;
1064 		goto fail_brelse;
1065 	}
1066 
1067 	/*  Change the pointers  */
1068 	for (x = 0; x < half_len; x++)
1069 		lp[x] = cpu_to_be64(bn);
1070 
1071 	gfs2_dir_hash_inval(dip);
1072 
1073 	error = gfs2_dir_write_data(dip, (char *)lp, start * sizeof(u64),
1074 				    half_len * sizeof(u64));
1075 	if (error != half_len * sizeof(u64)) {
1076 		if (error >= 0)
1077 			error = -EIO;
1078 		goto fail_lpfree;
1079 	}
1080 
1081 	kfree(lp);
1082 
1083 	/*  Compute the divider  */
1084 	divider = (start + half_len) << (32 - dip->i_depth);
1085 
1086 	/*  Copy the entries  */
1087 	dent = (struct gfs2_dirent *)(obh->b_data + sizeof(struct gfs2_leaf));
1088 
1089 	do {
1090 		next = dent;
1091 		if (dirent_next(dip, obh, &next))
1092 			next = NULL;
1093 
1094 		if (!gfs2_dirent_sentinel(dent) &&
1095 		    be32_to_cpu(dent->de_hash) < divider) {
1096 			struct qstr str;
1097 			void *ptr = ((char *)dent - obh->b_data) + nbh->b_data;
1098 			str.name = (char*)(dent+1);
1099 			str.len = be16_to_cpu(dent->de_name_len);
1100 			str.hash = be32_to_cpu(dent->de_hash);
1101 			new = gfs2_dirent_split_alloc(inode, nbh, &str, ptr);
1102 			if (IS_ERR(new)) {
1103 				error = PTR_ERR(new);
1104 				break;
1105 			}
1106 
1107 			new->de_inum = dent->de_inum; /* No endian worries */
1108 			new->de_type = dent->de_type; /* No endian worries */
1109 			be16_add_cpu(&nleaf->lf_entries, 1);
1110 
1111 			dirent_del(dip, obh, prev, dent);
1112 
1113 			if (!oleaf->lf_entries)
1114 				gfs2_consist_inode(dip);
1115 			be16_add_cpu(&oleaf->lf_entries, -1);
1116 
1117 			if (!prev)
1118 				prev = dent;
1119 		} else {
1120 			prev = dent;
1121 		}
1122 		dent = next;
1123 	} while (dent);
1124 
1125 	oleaf->lf_depth = nleaf->lf_depth;
1126 
1127 	error = gfs2_meta_inode_buffer(dip, &dibh);
1128 	if (!gfs2_assert_withdraw(GFS2_SB(&dip->i_inode), !error)) {
1129 		gfs2_trans_add_meta(dip->i_gl, dibh);
1130 		gfs2_add_inode_blocks(&dip->i_inode, 1);
1131 		gfs2_dinode_out(dip, dibh->b_data);
1132 		brelse(dibh);
1133 	}
1134 
1135 	brelse(obh);
1136 	brelse(nbh);
1137 
1138 	return error;
1139 
1140 fail_lpfree:
1141 	kfree(lp);
1142 
1143 fail_brelse:
1144 	brelse(obh);
1145 	brelse(nbh);
1146 	return error;
1147 }
1148 
1149 /**
1150  * dir_double_exhash - Double size of ExHash table
1151  * @dip: The GFS2 dinode
1152  *
1153  * Returns: 0 on success, error code on failure
1154  */
1155 
1156 static int dir_double_exhash(struct gfs2_inode *dip)
1157 {
1158 	struct buffer_head *dibh;
1159 	u32 hsize;
1160 	u32 hsize_bytes;
1161 	__be64 *hc;
1162 	__be64 *hc2, *h;
1163 	int x;
1164 	int error = 0;
1165 
1166 	hsize = BIT(dip->i_depth);
1167 	hsize_bytes = hsize * sizeof(__be64);
1168 
1169 	hc = gfs2_dir_get_hash_table(dip);
1170 	if (IS_ERR(hc))
1171 		return PTR_ERR(hc);
1172 
1173 	hc2 = kmalloc_array(hsize_bytes, 2, GFP_NOFS | __GFP_NOWARN);
1174 	if (hc2 == NULL)
1175 		hc2 = __vmalloc(hsize_bytes * 2, GFP_NOFS, PAGE_KERNEL);
1176 
1177 	if (!hc2)
1178 		return -ENOMEM;
1179 
1180 	h = hc2;
1181 	error = gfs2_meta_inode_buffer(dip, &dibh);
1182 	if (error)
1183 		goto out_kfree;
1184 
1185 	for (x = 0; x < hsize; x++) {
1186 		*h++ = *hc;
1187 		*h++ = *hc;
1188 		hc++;
1189 	}
1190 
1191 	error = gfs2_dir_write_data(dip, (char *)hc2, 0, hsize_bytes * 2);
1192 	if (error != (hsize_bytes * 2))
1193 		goto fail;
1194 
1195 	gfs2_dir_hash_inval(dip);
1196 	dip->i_hash_cache = hc2;
1197 	dip->i_depth++;
1198 	gfs2_dinode_out(dip, dibh->b_data);
1199 	brelse(dibh);
1200 	return 0;
1201 
1202 fail:
1203 	/* Replace original hash table & size */
1204 	gfs2_dir_write_data(dip, (char *)hc, 0, hsize_bytes);
1205 	i_size_write(&dip->i_inode, hsize_bytes);
1206 	gfs2_dinode_out(dip, dibh->b_data);
1207 	brelse(dibh);
1208 out_kfree:
1209 	kvfree(hc2);
1210 	return error;
1211 }
1212 
1213 /**
1214  * compare_dents - compare directory entries by hash value
1215  * @a: first dent
1216  * @b: second dent
1217  *
1218  * When comparing the hash entries of @a to @b:
1219  *   gt: returns 1
1220  *   lt: returns -1
1221  *   eq: returns 0
1222  */
1223 
1224 static int compare_dents(const void *a, const void *b)
1225 {
1226 	const struct gfs2_dirent *dent_a, *dent_b;
1227 	u32 hash_a, hash_b;
1228 	int ret = 0;
1229 
1230 	dent_a = *(const struct gfs2_dirent **)a;
1231 	hash_a = dent_a->de_cookie;
1232 
1233 	dent_b = *(const struct gfs2_dirent **)b;
1234 	hash_b = dent_b->de_cookie;
1235 
1236 	if (hash_a > hash_b)
1237 		ret = 1;
1238 	else if (hash_a < hash_b)
1239 		ret = -1;
1240 	else {
1241 		unsigned int len_a = be16_to_cpu(dent_a->de_name_len);
1242 		unsigned int len_b = be16_to_cpu(dent_b->de_name_len);
1243 
1244 		if (len_a > len_b)
1245 			ret = 1;
1246 		else if (len_a < len_b)
1247 			ret = -1;
1248 		else
1249 			ret = memcmp(dent_a + 1, dent_b + 1, len_a);
1250 	}
1251 
1252 	return ret;
1253 }
1254 
1255 /**
1256  * do_filldir_main - read out directory entries
1257  * @dip: The GFS2 inode
1258  * @ctx: what to feed the entries to
1259  * @darr: an array of struct gfs2_dirent pointers to read
1260  * @entries: the number of entries in darr
1261  * @copied: pointer to int that's non-zero if a entry has been copied out
1262  *
1263  * Jump through some hoops to make sure that if there are hash collsions,
1264  * they are read out at the beginning of a buffer.  We want to minimize
1265  * the possibility that they will fall into different readdir buffers or
1266  * that someone will want to seek to that location.
1267  *
1268  * Returns: errno, >0 if the actor tells you to stop
1269  */
1270 
1271 static int do_filldir_main(struct gfs2_inode *dip, struct dir_context *ctx,
1272 			   struct gfs2_dirent **darr, u32 entries,
1273 			   u32 sort_start, int *copied)
1274 {
1275 	const struct gfs2_dirent *dent, *dent_next;
1276 	u64 off, off_next;
1277 	unsigned int x, y;
1278 	int run = 0;
1279 
1280 	if (sort_start < entries)
1281 		sort(&darr[sort_start], entries - sort_start,
1282 		     sizeof(struct gfs2_dirent *), compare_dents, NULL);
1283 
1284 	dent_next = darr[0];
1285 	off_next = dent_next->de_cookie;
1286 
1287 	for (x = 0, y = 1; x < entries; x++, y++) {
1288 		dent = dent_next;
1289 		off = off_next;
1290 
1291 		if (y < entries) {
1292 			dent_next = darr[y];
1293 			off_next = dent_next->de_cookie;
1294 
1295 			if (off < ctx->pos)
1296 				continue;
1297 			ctx->pos = off;
1298 
1299 			if (off_next == off) {
1300 				if (*copied && !run)
1301 					return 1;
1302 				run = 1;
1303 			} else
1304 				run = 0;
1305 		} else {
1306 			if (off < ctx->pos)
1307 				continue;
1308 			ctx->pos = off;
1309 		}
1310 
1311 		if (!dir_emit(ctx, (const char *)(dent + 1),
1312 				be16_to_cpu(dent->de_name_len),
1313 				be64_to_cpu(dent->de_inum.no_addr),
1314 				be16_to_cpu(dent->de_type)))
1315 			return 1;
1316 
1317 		*copied = 1;
1318 	}
1319 
1320 	/* Increment the ctx->pos by one, so the next time we come into the
1321 	   do_filldir fxn, we get the next entry instead of the last one in the
1322 	   current leaf */
1323 
1324 	ctx->pos++;
1325 
1326 	return 0;
1327 }
1328 
1329 static void *gfs2_alloc_sort_buffer(unsigned size)
1330 {
1331 	void *ptr = NULL;
1332 
1333 	if (size < KMALLOC_MAX_SIZE)
1334 		ptr = kmalloc(size, GFP_NOFS | __GFP_NOWARN);
1335 	if (!ptr)
1336 		ptr = __vmalloc(size, GFP_NOFS, PAGE_KERNEL);
1337 	return ptr;
1338 }
1339 
1340 
1341 static int gfs2_set_cookies(struct gfs2_sbd *sdp, struct buffer_head *bh,
1342 			    unsigned leaf_nr, struct gfs2_dirent **darr,
1343 			    unsigned entries)
1344 {
1345 	int sort_id = -1;
1346 	int i;
1347 
1348 	for (i = 0; i < entries; i++) {
1349 		unsigned offset;
1350 
1351 		darr[i]->de_cookie = be32_to_cpu(darr[i]->de_hash);
1352 		darr[i]->de_cookie = gfs2_disk_hash2offset(darr[i]->de_cookie);
1353 
1354 		if (!sdp->sd_args.ar_loccookie)
1355 			continue;
1356 		offset = (char *)(darr[i]) -
1357 			(bh->b_data + gfs2_dirent_offset(sdp, bh->b_data));
1358 		offset /= GFS2_MIN_DIRENT_SIZE;
1359 		offset += leaf_nr * sdp->sd_max_dents_per_leaf;
1360 		if (offset >= GFS2_USE_HASH_FLAG ||
1361 		    leaf_nr >= GFS2_USE_HASH_FLAG) {
1362 			darr[i]->de_cookie |= GFS2_USE_HASH_FLAG;
1363 			if (sort_id < 0)
1364 				sort_id = i;
1365 			continue;
1366 		}
1367 		darr[i]->de_cookie &= GFS2_HASH_INDEX_MASK;
1368 		darr[i]->de_cookie |= offset;
1369 	}
1370 	return sort_id;
1371 }
1372 
1373 
1374 static int gfs2_dir_read_leaf(struct inode *inode, struct dir_context *ctx,
1375 			      int *copied, unsigned *depth,
1376 			      u64 leaf_no)
1377 {
1378 	struct gfs2_inode *ip = GFS2_I(inode);
1379 	struct gfs2_sbd *sdp = GFS2_SB(inode);
1380 	struct buffer_head *bh;
1381 	struct gfs2_leaf *lf;
1382 	unsigned entries = 0, entries2 = 0;
1383 	unsigned leaves = 0, leaf = 0, offset, sort_offset;
1384 	struct gfs2_dirent **darr, *dent;
1385 	struct dirent_gather g;
1386 	struct buffer_head **larr;
1387 	int error, i, need_sort = 0, sort_id;
1388 	u64 lfn = leaf_no;
1389 
1390 	do {
1391 		error = get_leaf(ip, lfn, &bh);
1392 		if (error)
1393 			goto out;
1394 		lf = (struct gfs2_leaf *)bh->b_data;
1395 		if (leaves == 0)
1396 			*depth = be16_to_cpu(lf->lf_depth);
1397 		entries += be16_to_cpu(lf->lf_entries);
1398 		leaves++;
1399 		lfn = be64_to_cpu(lf->lf_next);
1400 		brelse(bh);
1401 	} while(lfn);
1402 
1403 	if (*depth < GFS2_DIR_MAX_DEPTH || !sdp->sd_args.ar_loccookie) {
1404 		need_sort = 1;
1405 		sort_offset = 0;
1406 	}
1407 
1408 	if (!entries)
1409 		return 0;
1410 
1411 	error = -ENOMEM;
1412 	/*
1413 	 * The extra 99 entries are not normally used, but are a buffer
1414 	 * zone in case the number of entries in the leaf is corrupt.
1415 	 * 99 is the maximum number of entries that can fit in a single
1416 	 * leaf block.
1417 	 */
1418 	larr = gfs2_alloc_sort_buffer((leaves + entries + 99) * sizeof(void *));
1419 	if (!larr)
1420 		goto out;
1421 	darr = (struct gfs2_dirent **)(larr + leaves);
1422 	g.pdent = (const struct gfs2_dirent **)darr;
1423 	g.offset = 0;
1424 	lfn = leaf_no;
1425 
1426 	do {
1427 		error = get_leaf(ip, lfn, &bh);
1428 		if (error)
1429 			goto out_free;
1430 		lf = (struct gfs2_leaf *)bh->b_data;
1431 		lfn = be64_to_cpu(lf->lf_next);
1432 		if (lf->lf_entries) {
1433 			offset = g.offset;
1434 			entries2 += be16_to_cpu(lf->lf_entries);
1435 			dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
1436 						gfs2_dirent_gather, NULL, &g);
1437 			error = PTR_ERR(dent);
1438 			if (IS_ERR(dent))
1439 				goto out_free;
1440 			if (entries2 != g.offset) {
1441 				fs_warn(sdp, "Number of entries corrupt in dir "
1442 						"leaf %llu, entries2 (%u) != "
1443 						"g.offset (%u)\n",
1444 					(unsigned long long)bh->b_blocknr,
1445 					entries2, g.offset);
1446 				gfs2_consist_inode(ip);
1447 				error = -EIO;
1448 				goto out_free;
1449 			}
1450 			error = 0;
1451 			sort_id = gfs2_set_cookies(sdp, bh, leaf, &darr[offset],
1452 						   be16_to_cpu(lf->lf_entries));
1453 			if (!need_sort && sort_id >= 0) {
1454 				need_sort = 1;
1455 				sort_offset = offset + sort_id;
1456 			}
1457 			larr[leaf++] = bh;
1458 		} else {
1459 			larr[leaf++] = NULL;
1460 			brelse(bh);
1461 		}
1462 	} while(lfn);
1463 
1464 	BUG_ON(entries2 != entries);
1465 	error = do_filldir_main(ip, ctx, darr, entries, need_sort ?
1466 				sort_offset : entries, copied);
1467 out_free:
1468 	for(i = 0; i < leaf; i++)
1469 		if (larr[i])
1470 			brelse(larr[i]);
1471 	kvfree(larr);
1472 out:
1473 	return error;
1474 }
1475 
1476 /**
1477  * gfs2_dir_readahead - Issue read-ahead requests for leaf blocks.
1478  *
1479  * Note: we can't calculate each index like dir_e_read can because we don't
1480  * have the leaf, and therefore we don't have the depth, and therefore we
1481  * don't have the length. So we have to just read enough ahead to make up
1482  * for the loss of information.
1483  */
1484 static void gfs2_dir_readahead(struct inode *inode, unsigned hsize, u32 index,
1485 			       struct file_ra_state *f_ra)
1486 {
1487 	struct gfs2_inode *ip = GFS2_I(inode);
1488 	struct gfs2_glock *gl = ip->i_gl;
1489 	struct buffer_head *bh;
1490 	u64 blocknr = 0, last;
1491 	unsigned count;
1492 
1493 	/* First check if we've already read-ahead for the whole range. */
1494 	if (index + MAX_RA_BLOCKS < f_ra->start)
1495 		return;
1496 
1497 	f_ra->start = max((pgoff_t)index, f_ra->start);
1498 	for (count = 0; count < MAX_RA_BLOCKS; count++) {
1499 		if (f_ra->start >= hsize) /* if exceeded the hash table */
1500 			break;
1501 
1502 		last = blocknr;
1503 		blocknr = be64_to_cpu(ip->i_hash_cache[f_ra->start]);
1504 		f_ra->start++;
1505 		if (blocknr == last)
1506 			continue;
1507 
1508 		bh = gfs2_getbuf(gl, blocknr, 1);
1509 		if (trylock_buffer(bh)) {
1510 			if (buffer_uptodate(bh)) {
1511 				unlock_buffer(bh);
1512 				brelse(bh);
1513 				continue;
1514 			}
1515 			bh->b_end_io = end_buffer_read_sync;
1516 			submit_bh(REQ_OP_READ,
1517 				  REQ_RAHEAD | REQ_META | REQ_PRIO,
1518 				  bh);
1519 			continue;
1520 		}
1521 		brelse(bh);
1522 	}
1523 }
1524 
1525 /**
1526  * dir_e_read - Reads the entries from a directory into a filldir buffer
1527  * @dip: dinode pointer
1528  * @ctx: actor to feed the entries to
1529  *
1530  * Returns: errno
1531  */
1532 
1533 static int dir_e_read(struct inode *inode, struct dir_context *ctx,
1534 		      struct file_ra_state *f_ra)
1535 {
1536 	struct gfs2_inode *dip = GFS2_I(inode);
1537 	u32 hsize, len = 0;
1538 	u32 hash, index;
1539 	__be64 *lp;
1540 	int copied = 0;
1541 	int error = 0;
1542 	unsigned depth = 0;
1543 
1544 	hsize = BIT(dip->i_depth);
1545 	hash = gfs2_dir_offset2hash(ctx->pos);
1546 	index = hash >> (32 - dip->i_depth);
1547 
1548 	if (dip->i_hash_cache == NULL)
1549 		f_ra->start = 0;
1550 	lp = gfs2_dir_get_hash_table(dip);
1551 	if (IS_ERR(lp))
1552 		return PTR_ERR(lp);
1553 
1554 	gfs2_dir_readahead(inode, hsize, index, f_ra);
1555 
1556 	while (index < hsize) {
1557 		error = gfs2_dir_read_leaf(inode, ctx,
1558 					   &copied, &depth,
1559 					   be64_to_cpu(lp[index]));
1560 		if (error)
1561 			break;
1562 
1563 		len = BIT(dip->i_depth - depth);
1564 		index = (index & ~(len - 1)) + len;
1565 	}
1566 
1567 	if (error > 0)
1568 		error = 0;
1569 	return error;
1570 }
1571 
1572 int gfs2_dir_read(struct inode *inode, struct dir_context *ctx,
1573 		  struct file_ra_state *f_ra)
1574 {
1575 	struct gfs2_inode *dip = GFS2_I(inode);
1576 	struct gfs2_sbd *sdp = GFS2_SB(inode);
1577 	struct dirent_gather g;
1578 	struct gfs2_dirent **darr, *dent;
1579 	struct buffer_head *dibh;
1580 	int copied = 0;
1581 	int error;
1582 
1583 	if (!dip->i_entries)
1584 		return 0;
1585 
1586 	if (dip->i_diskflags & GFS2_DIF_EXHASH)
1587 		return dir_e_read(inode, ctx, f_ra);
1588 
1589 	if (!gfs2_is_stuffed(dip)) {
1590 		gfs2_consist_inode(dip);
1591 		return -EIO;
1592 	}
1593 
1594 	error = gfs2_meta_inode_buffer(dip, &dibh);
1595 	if (error)
1596 		return error;
1597 
1598 	error = -ENOMEM;
1599 	/* 96 is max number of dirents which can be stuffed into an inode */
1600 	darr = kmalloc_array(96, sizeof(struct gfs2_dirent *), GFP_NOFS);
1601 	if (darr) {
1602 		g.pdent = (const struct gfs2_dirent **)darr;
1603 		g.offset = 0;
1604 		dent = gfs2_dirent_scan(inode, dibh->b_data, dibh->b_size,
1605 					gfs2_dirent_gather, NULL, &g);
1606 		if (IS_ERR(dent)) {
1607 			error = PTR_ERR(dent);
1608 			goto out;
1609 		}
1610 		if (dip->i_entries != g.offset) {
1611 			fs_warn(sdp, "Number of entries corrupt in dir %llu, "
1612 				"ip->i_entries (%u) != g.offset (%u)\n",
1613 				(unsigned long long)dip->i_no_addr,
1614 				dip->i_entries,
1615 				g.offset);
1616 			gfs2_consist_inode(dip);
1617 			error = -EIO;
1618 			goto out;
1619 		}
1620 		gfs2_set_cookies(sdp, dibh, 0, darr, dip->i_entries);
1621 		error = do_filldir_main(dip, ctx, darr,
1622 					dip->i_entries, 0, &copied);
1623 out:
1624 		kfree(darr);
1625 	}
1626 
1627 	if (error > 0)
1628 		error = 0;
1629 
1630 	brelse(dibh);
1631 
1632 	return error;
1633 }
1634 
1635 /**
1636  * gfs2_dir_search - Search a directory
1637  * @dip: The GFS2 dir inode
1638  * @name: The name we are looking up
1639  * @fail_on_exist: Fail if the name exists rather than looking it up
1640  *
1641  * This routine searches a directory for a file or another directory.
1642  * Assumes a glock is held on dip.
1643  *
1644  * Returns: errno
1645  */
1646 
1647 struct inode *gfs2_dir_search(struct inode *dir, const struct qstr *name,
1648 			      bool fail_on_exist)
1649 {
1650 	struct buffer_head *bh;
1651 	struct gfs2_dirent *dent;
1652 	u64 addr, formal_ino;
1653 	u16 dtype;
1654 
1655 	dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh);
1656 	if (dent) {
1657 		struct inode *inode;
1658 		u16 rahead;
1659 
1660 		if (IS_ERR(dent))
1661 			return ERR_CAST(dent);
1662 		dtype = be16_to_cpu(dent->de_type);
1663 		rahead = be16_to_cpu(dent->de_rahead);
1664 		addr = be64_to_cpu(dent->de_inum.no_addr);
1665 		formal_ino = be64_to_cpu(dent->de_inum.no_formal_ino);
1666 		brelse(bh);
1667 		if (fail_on_exist)
1668 			return ERR_PTR(-EEXIST);
1669 		inode = gfs2_inode_lookup(dir->i_sb, dtype, addr, formal_ino,
1670 					  GFS2_BLKST_FREE /* ignore */);
1671 		if (!IS_ERR(inode))
1672 			GFS2_I(inode)->i_rahead = rahead;
1673 		return inode;
1674 	}
1675 	return ERR_PTR(-ENOENT);
1676 }
1677 
1678 int gfs2_dir_check(struct inode *dir, const struct qstr *name,
1679 		   const struct gfs2_inode *ip)
1680 {
1681 	struct buffer_head *bh;
1682 	struct gfs2_dirent *dent;
1683 	int ret = -ENOENT;
1684 
1685 	dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh);
1686 	if (dent) {
1687 		if (IS_ERR(dent))
1688 			return PTR_ERR(dent);
1689 		if (ip) {
1690 			if (be64_to_cpu(dent->de_inum.no_addr) != ip->i_no_addr)
1691 				goto out;
1692 			if (be64_to_cpu(dent->de_inum.no_formal_ino) !=
1693 			    ip->i_no_formal_ino)
1694 				goto out;
1695 			if (unlikely(IF2DT(ip->i_inode.i_mode) !=
1696 			    be16_to_cpu(dent->de_type))) {
1697 				gfs2_consist_inode(GFS2_I(dir));
1698 				ret = -EIO;
1699 				goto out;
1700 			}
1701 		}
1702 		ret = 0;
1703 out:
1704 		brelse(bh);
1705 	}
1706 	return ret;
1707 }
1708 
1709 /**
1710  * dir_new_leaf - Add a new leaf onto hash chain
1711  * @inode: The directory
1712  * @name: The name we are adding
1713  *
1714  * This adds a new dir leaf onto an existing leaf when there is not
1715  * enough space to add a new dir entry. This is a last resort after
1716  * we've expanded the hash table to max size and also split existing
1717  * leaf blocks, so it will only occur for very large directories.
1718  *
1719  * The dist parameter is set to 1 for leaf blocks directly attached
1720  * to the hash table, 2 for one layer of indirection, 3 for two layers
1721  * etc. We are thus able to tell the difference between an old leaf
1722  * with dist set to zero (i.e. "don't know") and a new one where we
1723  * set this information for debug/fsck purposes.
1724  *
1725  * Returns: 0 on success, or -ve on error
1726  */
1727 
1728 static int dir_new_leaf(struct inode *inode, const struct qstr *name)
1729 {
1730 	struct buffer_head *bh, *obh;
1731 	struct gfs2_inode *ip = GFS2_I(inode);
1732 	struct gfs2_leaf *leaf, *oleaf;
1733 	u32 dist = 1;
1734 	int error;
1735 	u32 index;
1736 	u64 bn;
1737 
1738 	index = name->hash >> (32 - ip->i_depth);
1739 	error = get_first_leaf(ip, index, &obh);
1740 	if (error)
1741 		return error;
1742 	do {
1743 		dist++;
1744 		oleaf = (struct gfs2_leaf *)obh->b_data;
1745 		bn = be64_to_cpu(oleaf->lf_next);
1746 		if (!bn)
1747 			break;
1748 		brelse(obh);
1749 		error = get_leaf(ip, bn, &obh);
1750 		if (error)
1751 			return error;
1752 	} while(1);
1753 
1754 	gfs2_trans_add_meta(ip->i_gl, obh);
1755 
1756 	leaf = new_leaf(inode, &bh, be16_to_cpu(oleaf->lf_depth));
1757 	if (!leaf) {
1758 		brelse(obh);
1759 		return -ENOSPC;
1760 	}
1761 	leaf->lf_dist = cpu_to_be32(dist);
1762 	oleaf->lf_next = cpu_to_be64(bh->b_blocknr);
1763 	brelse(bh);
1764 	brelse(obh);
1765 
1766 	error = gfs2_meta_inode_buffer(ip, &bh);
1767 	if (error)
1768 		return error;
1769 	gfs2_trans_add_meta(ip->i_gl, bh);
1770 	gfs2_add_inode_blocks(&ip->i_inode, 1);
1771 	gfs2_dinode_out(ip, bh->b_data);
1772 	brelse(bh);
1773 	return 0;
1774 }
1775 
1776 static u16 gfs2_inode_ra_len(const struct gfs2_inode *ip)
1777 {
1778 	u64 where = ip->i_no_addr + 1;
1779 	if (ip->i_eattr == where)
1780 		return 1;
1781 	return 0;
1782 }
1783 
1784 /**
1785  * gfs2_dir_add - Add new filename into directory
1786  * @inode: The directory inode
1787  * @name: The new name
1788  * @nip: The GFS2 inode to be linked in to the directory
1789  * @da: The directory addition info
1790  *
1791  * If the call to gfs2_diradd_alloc_required resulted in there being
1792  * no need to allocate any new directory blocks, then it will contain
1793  * a pointer to the directory entry and the bh in which it resides. We
1794  * can use that without having to repeat the search. If there was no
1795  * free space, then we must now create more space.
1796  *
1797  * Returns: 0 on success, error code on failure
1798  */
1799 
1800 int gfs2_dir_add(struct inode *inode, const struct qstr *name,
1801 		 const struct gfs2_inode *nip, struct gfs2_diradd *da)
1802 {
1803 	struct gfs2_inode *ip = GFS2_I(inode);
1804 	struct buffer_head *bh = da->bh;
1805 	struct gfs2_dirent *dent = da->dent;
1806 	struct timespec64 tv;
1807 	struct gfs2_leaf *leaf;
1808 	int error;
1809 
1810 	while(1) {
1811 		if (da->bh == NULL) {
1812 			dent = gfs2_dirent_search(inode, name,
1813 						  gfs2_dirent_find_space, &bh);
1814 		}
1815 		if (dent) {
1816 			if (IS_ERR(dent))
1817 				return PTR_ERR(dent);
1818 			dent = gfs2_init_dirent(inode, dent, name, bh);
1819 			gfs2_inum_out(nip, dent);
1820 			dent->de_type = cpu_to_be16(IF2DT(nip->i_inode.i_mode));
1821 			dent->de_rahead = cpu_to_be16(gfs2_inode_ra_len(nip));
1822 			tv = current_time(&ip->i_inode);
1823 			if (ip->i_diskflags & GFS2_DIF_EXHASH) {
1824 				leaf = (struct gfs2_leaf *)bh->b_data;
1825 				be16_add_cpu(&leaf->lf_entries, 1);
1826 				leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
1827 				leaf->lf_sec = cpu_to_be64(tv.tv_sec);
1828 			}
1829 			da->dent = NULL;
1830 			da->bh = NULL;
1831 			brelse(bh);
1832 			ip->i_entries++;
1833 			ip->i_inode.i_mtime = ip->i_inode.i_ctime = tv;
1834 			if (S_ISDIR(nip->i_inode.i_mode))
1835 				inc_nlink(&ip->i_inode);
1836 			mark_inode_dirty(inode);
1837 			error = 0;
1838 			break;
1839 		}
1840 		if (!(ip->i_diskflags & GFS2_DIF_EXHASH)) {
1841 			error = dir_make_exhash(inode);
1842 			if (error)
1843 				break;
1844 			continue;
1845 		}
1846 		error = dir_split_leaf(inode, name);
1847 		if (error == 0)
1848 			continue;
1849 		if (error < 0)
1850 			break;
1851 		if (ip->i_depth < GFS2_DIR_MAX_DEPTH) {
1852 			error = dir_double_exhash(ip);
1853 			if (error)
1854 				break;
1855 			error = dir_split_leaf(inode, name);
1856 			if (error < 0)
1857 				break;
1858 			if (error == 0)
1859 				continue;
1860 		}
1861 		error = dir_new_leaf(inode, name);
1862 		if (!error)
1863 			continue;
1864 		error = -ENOSPC;
1865 		break;
1866 	}
1867 	return error;
1868 }
1869 
1870 
1871 /**
1872  * gfs2_dir_del - Delete a directory entry
1873  * @dip: The GFS2 inode
1874  * @filename: The filename
1875  *
1876  * Returns: 0 on success, error code on failure
1877  */
1878 
1879 int gfs2_dir_del(struct gfs2_inode *dip, const struct dentry *dentry)
1880 {
1881 	const struct qstr *name = &dentry->d_name;
1882 	struct gfs2_dirent *dent, *prev = NULL;
1883 	struct buffer_head *bh;
1884 	struct timespec64 tv = current_time(&dip->i_inode);
1885 
1886 	/* Returns _either_ the entry (if its first in block) or the
1887 	   previous entry otherwise */
1888 	dent = gfs2_dirent_search(&dip->i_inode, name, gfs2_dirent_prev, &bh);
1889 	if (!dent) {
1890 		gfs2_consist_inode(dip);
1891 		return -EIO;
1892 	}
1893 	if (IS_ERR(dent)) {
1894 		gfs2_consist_inode(dip);
1895 		return PTR_ERR(dent);
1896 	}
1897 	/* If not first in block, adjust pointers accordingly */
1898 	if (gfs2_dirent_find(dent, name, NULL) == 0) {
1899 		prev = dent;
1900 		dent = (struct gfs2_dirent *)((char *)dent + be16_to_cpu(prev->de_rec_len));
1901 	}
1902 
1903 	dirent_del(dip, bh, prev, dent);
1904 	if (dip->i_diskflags & GFS2_DIF_EXHASH) {
1905 		struct gfs2_leaf *leaf = (struct gfs2_leaf *)bh->b_data;
1906 		u16 entries = be16_to_cpu(leaf->lf_entries);
1907 		if (!entries)
1908 			gfs2_consist_inode(dip);
1909 		leaf->lf_entries = cpu_to_be16(--entries);
1910 		leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
1911 		leaf->lf_sec = cpu_to_be64(tv.tv_sec);
1912 	}
1913 	brelse(bh);
1914 
1915 	if (!dip->i_entries)
1916 		gfs2_consist_inode(dip);
1917 	dip->i_entries--;
1918 	dip->i_inode.i_mtime = dip->i_inode.i_ctime = tv;
1919 	if (d_is_dir(dentry))
1920 		drop_nlink(&dip->i_inode);
1921 	mark_inode_dirty(&dip->i_inode);
1922 
1923 	return 0;
1924 }
1925 
1926 /**
1927  * gfs2_dir_mvino - Change inode number of directory entry
1928  * @dip: The GFS2 inode
1929  * @filename:
1930  * @new_inode:
1931  *
1932  * This routine changes the inode number of a directory entry.  It's used
1933  * by rename to change ".." when a directory is moved.
1934  * Assumes a glock is held on dvp.
1935  *
1936  * Returns: errno
1937  */
1938 
1939 int gfs2_dir_mvino(struct gfs2_inode *dip, const struct qstr *filename,
1940 		   const struct gfs2_inode *nip, unsigned int new_type)
1941 {
1942 	struct buffer_head *bh;
1943 	struct gfs2_dirent *dent;
1944 
1945 	dent = gfs2_dirent_search(&dip->i_inode, filename, gfs2_dirent_find, &bh);
1946 	if (!dent) {
1947 		gfs2_consist_inode(dip);
1948 		return -EIO;
1949 	}
1950 	if (IS_ERR(dent))
1951 		return PTR_ERR(dent);
1952 
1953 	gfs2_trans_add_meta(dip->i_gl, bh);
1954 	gfs2_inum_out(nip, dent);
1955 	dent->de_type = cpu_to_be16(new_type);
1956 	brelse(bh);
1957 
1958 	dip->i_inode.i_mtime = dip->i_inode.i_ctime = current_time(&dip->i_inode);
1959 	mark_inode_dirty_sync(&dip->i_inode);
1960 	return 0;
1961 }
1962 
1963 /**
1964  * leaf_dealloc - Deallocate a directory leaf
1965  * @dip: the directory
1966  * @index: the hash table offset in the directory
1967  * @len: the number of pointers to this leaf
1968  * @leaf_no: the leaf number
1969  * @leaf_bh: buffer_head for the starting leaf
1970  * last_dealloc: 1 if this is the final dealloc for the leaf, else 0
1971  *
1972  * Returns: errno
1973  */
1974 
1975 static int leaf_dealloc(struct gfs2_inode *dip, u32 index, u32 len,
1976 			u64 leaf_no, struct buffer_head *leaf_bh,
1977 			int last_dealloc)
1978 {
1979 	struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode);
1980 	struct gfs2_leaf *tmp_leaf;
1981 	struct gfs2_rgrp_list rlist;
1982 	struct buffer_head *bh, *dibh;
1983 	u64 blk, nblk;
1984 	unsigned int rg_blocks = 0, l_blocks = 0;
1985 	char *ht;
1986 	unsigned int x, size = len * sizeof(u64);
1987 	int error;
1988 
1989 	error = gfs2_rindex_update(sdp);
1990 	if (error)
1991 		return error;
1992 
1993 	memset(&rlist, 0, sizeof(struct gfs2_rgrp_list));
1994 
1995 	ht = kzalloc(size, GFP_NOFS | __GFP_NOWARN);
1996 	if (ht == NULL)
1997 		ht = __vmalloc(size, GFP_NOFS | __GFP_NOWARN | __GFP_ZERO,
1998 			       PAGE_KERNEL);
1999 	if (!ht)
2000 		return -ENOMEM;
2001 
2002 	error = gfs2_quota_hold(dip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
2003 	if (error)
2004 		goto out;
2005 
2006 	/*  Count the number of leaves  */
2007 	bh = leaf_bh;
2008 
2009 	for (blk = leaf_no; blk; blk = nblk) {
2010 		if (blk != leaf_no) {
2011 			error = get_leaf(dip, blk, &bh);
2012 			if (error)
2013 				goto out_rlist;
2014 		}
2015 		tmp_leaf = (struct gfs2_leaf *)bh->b_data;
2016 		nblk = be64_to_cpu(tmp_leaf->lf_next);
2017 		if (blk != leaf_no)
2018 			brelse(bh);
2019 
2020 		gfs2_rlist_add(dip, &rlist, blk);
2021 		l_blocks++;
2022 	}
2023 
2024 	gfs2_rlist_alloc(&rlist);
2025 
2026 	for (x = 0; x < rlist.rl_rgrps; x++) {
2027 		struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(rlist.rl_ghs[x].gh_gl);
2028 
2029 		rg_blocks += rgd->rd_length;
2030 	}
2031 
2032 	error = gfs2_glock_nq_m(rlist.rl_rgrps, rlist.rl_ghs);
2033 	if (error)
2034 		goto out_rlist;
2035 
2036 	error = gfs2_trans_begin(sdp,
2037 			rg_blocks + (DIV_ROUND_UP(size, sdp->sd_jbsize) + 1) +
2038 			RES_DINODE + RES_STATFS + RES_QUOTA, l_blocks);
2039 	if (error)
2040 		goto out_rg_gunlock;
2041 
2042 	bh = leaf_bh;
2043 
2044 	for (blk = leaf_no; blk; blk = nblk) {
2045 		struct gfs2_rgrpd *rgd;
2046 
2047 		if (blk != leaf_no) {
2048 			error = get_leaf(dip, blk, &bh);
2049 			if (error)
2050 				goto out_end_trans;
2051 		}
2052 		tmp_leaf = (struct gfs2_leaf *)bh->b_data;
2053 		nblk = be64_to_cpu(tmp_leaf->lf_next);
2054 		if (blk != leaf_no)
2055 			brelse(bh);
2056 
2057 		rgd = gfs2_blk2rgrpd(sdp, blk, true);
2058 		gfs2_free_meta(dip, rgd, blk, 1);
2059 		gfs2_add_inode_blocks(&dip->i_inode, -1);
2060 	}
2061 
2062 	error = gfs2_dir_write_data(dip, ht, index * sizeof(u64), size);
2063 	if (error != size) {
2064 		if (error >= 0)
2065 			error = -EIO;
2066 		goto out_end_trans;
2067 	}
2068 
2069 	error = gfs2_meta_inode_buffer(dip, &dibh);
2070 	if (error)
2071 		goto out_end_trans;
2072 
2073 	gfs2_trans_add_meta(dip->i_gl, dibh);
2074 	/* On the last dealloc, make this a regular file in case we crash.
2075 	   (We don't want to free these blocks a second time.)  */
2076 	if (last_dealloc)
2077 		dip->i_inode.i_mode = S_IFREG;
2078 	gfs2_dinode_out(dip, dibh->b_data);
2079 	brelse(dibh);
2080 
2081 out_end_trans:
2082 	gfs2_trans_end(sdp);
2083 out_rg_gunlock:
2084 	gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs);
2085 out_rlist:
2086 	gfs2_rlist_free(&rlist);
2087 	gfs2_quota_unhold(dip);
2088 out:
2089 	kvfree(ht);
2090 	return error;
2091 }
2092 
2093 /**
2094  * gfs2_dir_exhash_dealloc - free all the leaf blocks in a directory
2095  * @dip: the directory
2096  *
2097  * Dealloc all on-disk directory leaves to FREEMETA state
2098  * Change on-disk inode type to "regular file"
2099  *
2100  * Returns: errno
2101  */
2102 
2103 int gfs2_dir_exhash_dealloc(struct gfs2_inode *dip)
2104 {
2105 	struct buffer_head *bh;
2106 	struct gfs2_leaf *leaf;
2107 	u32 hsize, len;
2108 	u32 index = 0, next_index;
2109 	__be64 *lp;
2110 	u64 leaf_no;
2111 	int error = 0, last;
2112 
2113 	hsize = BIT(dip->i_depth);
2114 
2115 	lp = gfs2_dir_get_hash_table(dip);
2116 	if (IS_ERR(lp))
2117 		return PTR_ERR(lp);
2118 
2119 	while (index < hsize) {
2120 		leaf_no = be64_to_cpu(lp[index]);
2121 		if (leaf_no) {
2122 			error = get_leaf(dip, leaf_no, &bh);
2123 			if (error)
2124 				goto out;
2125 			leaf = (struct gfs2_leaf *)bh->b_data;
2126 			len = BIT(dip->i_depth - be16_to_cpu(leaf->lf_depth));
2127 
2128 			next_index = (index & ~(len - 1)) + len;
2129 			last = ((next_index >= hsize) ? 1 : 0);
2130 			error = leaf_dealloc(dip, index, len, leaf_no, bh,
2131 					     last);
2132 			brelse(bh);
2133 			if (error)
2134 				goto out;
2135 			index = next_index;
2136 		} else
2137 			index++;
2138 	}
2139 
2140 	if (index != hsize) {
2141 		gfs2_consist_inode(dip);
2142 		error = -EIO;
2143 	}
2144 
2145 out:
2146 
2147 	return error;
2148 }
2149 
2150 /**
2151  * gfs2_diradd_alloc_required - find if adding entry will require an allocation
2152  * @ip: the file being written to
2153  * @filname: the filename that's going to be added
2154  * @da: The structure to return dir alloc info
2155  *
2156  * Returns: 0 if ok, -ve on error
2157  */
2158 
2159 int gfs2_diradd_alloc_required(struct inode *inode, const struct qstr *name,
2160 			       struct gfs2_diradd *da)
2161 {
2162 	struct gfs2_inode *ip = GFS2_I(inode);
2163 	struct gfs2_sbd *sdp = GFS2_SB(inode);
2164 	const unsigned int extra = sizeof(struct gfs2_dinode) - sizeof(struct gfs2_leaf);
2165 	struct gfs2_dirent *dent;
2166 	struct buffer_head *bh;
2167 
2168 	da->nr_blocks = 0;
2169 	da->bh = NULL;
2170 	da->dent = NULL;
2171 
2172 	dent = gfs2_dirent_search(inode, name, gfs2_dirent_find_space, &bh);
2173 	if (!dent) {
2174 		da->nr_blocks = sdp->sd_max_dirres;
2175 		if (!(ip->i_diskflags & GFS2_DIF_EXHASH) &&
2176 		    (GFS2_DIRENT_SIZE(name->len) < extra))
2177 			da->nr_blocks = 1;
2178 		return 0;
2179 	}
2180 	if (IS_ERR(dent))
2181 		return PTR_ERR(dent);
2182 
2183 	if (da->save_loc) {
2184 		da->bh = bh;
2185 		da->dent = dent;
2186 	} else {
2187 		brelse(bh);
2188 	}
2189 	return 0;
2190 }
2191 
2192