]> Pileus Git - ~andy/linux/blob - fs/gfs2/super.c
[GFS2] Journal extent mapping
[~andy/linux] / fs / gfs2 / super.c
1 /*
2  * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
3  * Copyright (C) 2004-2007 Red Hat, Inc.  All rights reserved.
4  *
5  * This copyrighted material is made available to anyone wishing to use,
6  * modify, copy, or redistribute it subject to the terms and conditions
7  * of the GNU General Public License version 2.
8  */
9
10 #include <linux/sched.h>
11 #include <linux/slab.h>
12 #include <linux/spinlock.h>
13 #include <linux/completion.h>
14 #include <linux/buffer_head.h>
15 #include <linux/crc32.h>
16 #include <linux/gfs2_ondisk.h>
17 #include <linux/bio.h>
18 #include <linux/lm_interface.h>
19
20 #include "gfs2.h"
21 #include "incore.h"
22 #include "bmap.h"
23 #include "dir.h"
24 #include "glock.h"
25 #include "glops.h"
26 #include "inode.h"
27 #include "log.h"
28 #include "meta_io.h"
29 #include "quota.h"
30 #include "recovery.h"
31 #include "rgrp.h"
32 #include "super.h"
33 #include "trans.h"
34 #include "util.h"
35
36 static const u32 gfs2_old_fs_formats[] = {
37         0
38 };
39
40 static const u32 gfs2_old_multihost_formats[] = {
41         0
42 };
43
44 /**
45  * gfs2_tune_init - Fill a gfs2_tune structure with default values
46  * @gt: tune
47  *
48  */
49
50 void gfs2_tune_init(struct gfs2_tune *gt)
51 {
52         spin_lock_init(&gt->gt_spin);
53
54         gt->gt_demote_secs = 300;
55         gt->gt_incore_log_blocks = 1024;
56         gt->gt_log_flush_secs = 60;
57         gt->gt_recoverd_secs = 60;
58         gt->gt_logd_secs = 1;
59         gt->gt_quotad_secs = 5;
60         gt->gt_quota_simul_sync = 64;
61         gt->gt_quota_warn_period = 10;
62         gt->gt_quota_scale_num = 1;
63         gt->gt_quota_scale_den = 1;
64         gt->gt_quota_cache_secs = 300;
65         gt->gt_quota_quantum = 60;
66         gt->gt_atime_quantum = 3600;
67         gt->gt_new_files_jdata = 0;
68         gt->gt_new_files_directio = 0;
69         gt->gt_max_readahead = 1 << 18;
70         gt->gt_stall_secs = 600;
71         gt->gt_complain_secs = 10;
72         gt->gt_statfs_quantum = 30;
73         gt->gt_statfs_slow = 0;
74 }
75
76 /**
77  * gfs2_check_sb - Check superblock
78  * @sdp: the filesystem
79  * @sb: The superblock
80  * @silent: Don't print a message if the check fails
81  *
82  * Checks the version code of the FS is one that we understand how to
83  * read and that the sizes of the various on-disk structures have not
84  * changed.
85  */
86
87 int gfs2_check_sb(struct gfs2_sbd *sdp, struct gfs2_sb_host *sb, int silent)
88 {
89         unsigned int x;
90
91         if (sb->sb_magic != GFS2_MAGIC ||
92             sb->sb_type != GFS2_METATYPE_SB) {
93                 if (!silent)
94                         printk(KERN_WARNING "GFS2: not a GFS2 filesystem\n");
95                 return -EINVAL;
96         }
97
98         /*  If format numbers match exactly, we're done.  */
99
100         if (sb->sb_fs_format == GFS2_FORMAT_FS &&
101             sb->sb_multihost_format == GFS2_FORMAT_MULTI)
102                 return 0;
103
104         if (sb->sb_fs_format != GFS2_FORMAT_FS) {
105                 for (x = 0; gfs2_old_fs_formats[x]; x++)
106                         if (gfs2_old_fs_formats[x] == sb->sb_fs_format)
107                                 break;
108
109                 if (!gfs2_old_fs_formats[x]) {
110                         printk(KERN_WARNING
111                                "GFS2: code version (%u, %u) is incompatible "
112                                "with ondisk format (%u, %u)\n",
113                                GFS2_FORMAT_FS, GFS2_FORMAT_MULTI,
114                                sb->sb_fs_format, sb->sb_multihost_format);
115                         printk(KERN_WARNING
116                                "GFS2: I don't know how to upgrade this FS\n");
117                         return -EINVAL;
118                 }
119         }
120
121         if (sb->sb_multihost_format != GFS2_FORMAT_MULTI) {
122                 for (x = 0; gfs2_old_multihost_formats[x]; x++)
123                         if (gfs2_old_multihost_formats[x] ==
124                             sb->sb_multihost_format)
125                                 break;
126
127                 if (!gfs2_old_multihost_formats[x]) {
128                         printk(KERN_WARNING
129                                "GFS2: code version (%u, %u) is incompatible "
130                                "with ondisk format (%u, %u)\n",
131                                GFS2_FORMAT_FS, GFS2_FORMAT_MULTI,
132                                sb->sb_fs_format, sb->sb_multihost_format);
133                         printk(KERN_WARNING
134                                "GFS2: I don't know how to upgrade this FS\n");
135                         return -EINVAL;
136                 }
137         }
138
139         if (!sdp->sd_args.ar_upgrade) {
140                 printk(KERN_WARNING
141                        "GFS2: code version (%u, %u) is incompatible "
142                        "with ondisk format (%u, %u)\n",
143                        GFS2_FORMAT_FS, GFS2_FORMAT_MULTI,
144                        sb->sb_fs_format, sb->sb_multihost_format);
145                 printk(KERN_INFO
146                        "GFS2: Use the \"upgrade\" mount option to upgrade "
147                        "the FS\n");
148                 printk(KERN_INFO "GFS2: See the manual for more details\n");
149                 return -EINVAL;
150         }
151
152         return 0;
153 }
154
155
156 static void end_bio_io_page(struct bio *bio, int error)
157 {
158         struct page *page = bio->bi_private;
159
160         if (!error)
161                 SetPageUptodate(page);
162         else
163                 printk(KERN_WARNING "gfs2: error %d reading superblock\n", error);
164         unlock_page(page);
165 }
166
167 static void gfs2_sb_in(struct gfs2_sb_host *sb, const void *buf)
168 {
169         const struct gfs2_sb *str = buf;
170
171         sb->sb_magic = be32_to_cpu(str->sb_header.mh_magic);
172         sb->sb_type = be32_to_cpu(str->sb_header.mh_type);
173         sb->sb_format = be32_to_cpu(str->sb_header.mh_format);
174         sb->sb_fs_format = be32_to_cpu(str->sb_fs_format);
175         sb->sb_multihost_format = be32_to_cpu(str->sb_multihost_format);
176         sb->sb_bsize = be32_to_cpu(str->sb_bsize);
177         sb->sb_bsize_shift = be32_to_cpu(str->sb_bsize_shift);
178         sb->sb_master_dir.no_addr = be64_to_cpu(str->sb_master_dir.no_addr);
179         sb->sb_master_dir.no_formal_ino = be64_to_cpu(str->sb_master_dir.no_formal_ino);
180         sb->sb_root_dir.no_addr = be64_to_cpu(str->sb_root_dir.no_addr);
181         sb->sb_root_dir.no_formal_ino = be64_to_cpu(str->sb_root_dir.no_formal_ino);
182
183         memcpy(sb->sb_lockproto, str->sb_lockproto, GFS2_LOCKNAME_LEN);
184         memcpy(sb->sb_locktable, str->sb_locktable, GFS2_LOCKNAME_LEN);
185 }
186
187 /**
188  * gfs2_read_super - Read the gfs2 super block from disk
189  * @sdp: The GFS2 super block
190  * @sector: The location of the super block
191  * @error: The error code to return
192  *
193  * This uses the bio functions to read the super block from disk
194  * because we want to be 100% sure that we never read cached data.
195  * A super block is read twice only during each GFS2 mount and is
196  * never written to by the filesystem. The first time its read no
197  * locks are held, and the only details which are looked at are those
198  * relating to the locking protocol. Once locking is up and working,
199  * the sb is read again under the lock to establish the location of
200  * the master directory (contains pointers to journals etc) and the
201  * root directory.
202  *
203  * Returns: 0 on success or error
204  */
205
206 int gfs2_read_super(struct gfs2_sbd *sdp, sector_t sector)
207 {
208         struct super_block *sb = sdp->sd_vfs;
209         struct gfs2_sb *p;
210         struct page *page;
211         struct bio *bio;
212
213         page = alloc_page(GFP_KERNEL);
214         if (unlikely(!page))
215                 return -ENOBUFS;
216
217         ClearPageUptodate(page);
218         ClearPageDirty(page);
219         lock_page(page);
220
221         bio = bio_alloc(GFP_KERNEL, 1);
222         if (unlikely(!bio)) {
223                 __free_page(page);
224                 return -ENOBUFS;
225         }
226
227         bio->bi_sector = sector * (sb->s_blocksize >> 9);
228         bio->bi_bdev = sb->s_bdev;
229         bio_add_page(bio, page, PAGE_SIZE, 0);
230
231         bio->bi_end_io = end_bio_io_page;
232         bio->bi_private = page;
233         submit_bio(READ_SYNC | (1 << BIO_RW_META), bio);
234         wait_on_page_locked(page);
235         bio_put(bio);
236         if (!PageUptodate(page)) {
237                 __free_page(page);
238                 return -EIO;
239         }
240         p = kmap(page);
241         gfs2_sb_in(&sdp->sd_sb, p);
242         kunmap(page);
243         __free_page(page);
244         return 0;
245 }
246
247 /**
248  * gfs2_read_sb - Read super block
249  * @sdp: The GFS2 superblock
250  * @gl: the glock for the superblock (assumed to be held)
251  * @silent: Don't print message if mount fails
252  *
253  */
254
255 int gfs2_read_sb(struct gfs2_sbd *sdp, struct gfs2_glock *gl, int silent)
256 {
257         u32 hash_blocks, ind_blocks, leaf_blocks;
258         u32 tmp_blocks;
259         unsigned int x;
260         int error;
261
262         error = gfs2_read_super(sdp, GFS2_SB_ADDR >> sdp->sd_fsb2bb_shift);
263         if (error) {
264                 if (!silent)
265                         fs_err(sdp, "can't read superblock\n");
266                 return error;
267         }
268
269         error = gfs2_check_sb(sdp, &sdp->sd_sb, silent);
270         if (error)
271                 return error;
272
273         sdp->sd_fsb2bb_shift = sdp->sd_sb.sb_bsize_shift -
274                                GFS2_BASIC_BLOCK_SHIFT;
275         sdp->sd_fsb2bb = 1 << sdp->sd_fsb2bb_shift;
276         sdp->sd_diptrs = (sdp->sd_sb.sb_bsize -
277                           sizeof(struct gfs2_dinode)) / sizeof(u64);
278         sdp->sd_inptrs = (sdp->sd_sb.sb_bsize -
279                           sizeof(struct gfs2_meta_header)) / sizeof(u64);
280         sdp->sd_jbsize = sdp->sd_sb.sb_bsize - sizeof(struct gfs2_meta_header);
281         sdp->sd_hash_bsize = sdp->sd_sb.sb_bsize / 2;
282         sdp->sd_hash_bsize_shift = sdp->sd_sb.sb_bsize_shift - 1;
283         sdp->sd_hash_ptrs = sdp->sd_hash_bsize / sizeof(u64);
284         sdp->sd_qc_per_block = (sdp->sd_sb.sb_bsize -
285                                 sizeof(struct gfs2_meta_header)) /
286                                 sizeof(struct gfs2_quota_change);
287
288         /* Compute maximum reservation required to add a entry to a directory */
289
290         hash_blocks = DIV_ROUND_UP(sizeof(u64) * (1 << GFS2_DIR_MAX_DEPTH),
291                              sdp->sd_jbsize);
292
293         ind_blocks = 0;
294         for (tmp_blocks = hash_blocks; tmp_blocks > sdp->sd_diptrs;) {
295                 tmp_blocks = DIV_ROUND_UP(tmp_blocks, sdp->sd_inptrs);
296                 ind_blocks += tmp_blocks;
297         }
298
299         leaf_blocks = 2 + GFS2_DIR_MAX_DEPTH;
300
301         sdp->sd_max_dirres = hash_blocks + ind_blocks + leaf_blocks;
302
303         sdp->sd_heightsize[0] = sdp->sd_sb.sb_bsize -
304                                 sizeof(struct gfs2_dinode);
305         sdp->sd_heightsize[1] = sdp->sd_sb.sb_bsize * sdp->sd_diptrs;
306         for (x = 2;; x++) {
307                 u64 space, d;
308                 u32 m;
309
310                 space = sdp->sd_heightsize[x - 1] * sdp->sd_inptrs;
311                 d = space;
312                 m = do_div(d, sdp->sd_inptrs);
313
314                 if (d != sdp->sd_heightsize[x - 1] || m)
315                         break;
316                 sdp->sd_heightsize[x] = space;
317         }
318         sdp->sd_max_height = x;
319         gfs2_assert(sdp, sdp->sd_max_height <= GFS2_MAX_META_HEIGHT);
320
321         sdp->sd_jheightsize[0] = sdp->sd_sb.sb_bsize -
322                                  sizeof(struct gfs2_dinode);
323         sdp->sd_jheightsize[1] = sdp->sd_jbsize * sdp->sd_diptrs;
324         for (x = 2;; x++) {
325                 u64 space, d;
326                 u32 m;
327
328                 space = sdp->sd_jheightsize[x - 1] * sdp->sd_inptrs;
329                 d = space;
330                 m = do_div(d, sdp->sd_inptrs);
331
332                 if (d != sdp->sd_jheightsize[x - 1] || m)
333                         break;
334                 sdp->sd_jheightsize[x] = space;
335         }
336         sdp->sd_max_jheight = x;
337         gfs2_assert(sdp, sdp->sd_max_jheight <= GFS2_MAX_META_HEIGHT);
338
339         return 0;
340 }
341
342 /**
343  * gfs2_jindex_hold - Grab a lock on the jindex
344  * @sdp: The GFS2 superblock
345  * @ji_gh: the holder for the jindex glock
346  *
347  * This is very similar to the gfs2_rindex_hold() function, except that
348  * in general we hold the jindex lock for longer periods of time and
349  * we grab it far less frequently (in general) then the rgrp lock.
350  *
351  * Returns: errno
352  */
353
354 int gfs2_jindex_hold(struct gfs2_sbd *sdp, struct gfs2_holder *ji_gh)
355 {
356         struct gfs2_inode *dip = GFS2_I(sdp->sd_jindex);
357         struct qstr name;
358         char buf[20];
359         struct gfs2_jdesc *jd;
360         int error;
361
362         name.name = buf;
363
364         mutex_lock(&sdp->sd_jindex_mutex);
365
366         for (;;) {
367                 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, ji_gh);
368                 if (error)
369                         break;
370
371                 name.len = sprintf(buf, "journal%u", sdp->sd_journals);
372                 name.hash = gfs2_disk_hash(name.name, name.len);
373
374                 error = gfs2_dir_check(sdp->sd_jindex, &name, NULL);
375                 if (error == -ENOENT) {
376                         error = 0;
377                         break;
378                 }
379
380                 gfs2_glock_dq_uninit(ji_gh);
381
382                 if (error)
383                         break;
384
385                 error = -ENOMEM;
386                 jd = kzalloc(sizeof(struct gfs2_jdesc), GFP_KERNEL);
387                 if (!jd)
388                         break;
389
390                 jd->jd_inode = gfs2_lookupi(sdp->sd_jindex, &name, 1, NULL);
391                 if (!jd->jd_inode || IS_ERR(jd->jd_inode)) {
392                         if (!jd->jd_inode)
393                                 error = -ENOENT;
394                         else
395                                 error = PTR_ERR(jd->jd_inode);
396                         kfree(jd);
397                         break;
398                 }
399
400                 spin_lock(&sdp->sd_jindex_spin);
401                 jd->jd_jid = sdp->sd_journals++;
402                 list_add_tail(&jd->jd_list, &sdp->sd_jindex_list);
403                 spin_unlock(&sdp->sd_jindex_spin);
404         }
405
406         mutex_unlock(&sdp->sd_jindex_mutex);
407
408         return error;
409 }
410
411 /**
412  * gfs2_jindex_free - Clear all the journal index information
413  * @sdp: The GFS2 superblock
414  *
415  */
416
417 void gfs2_jindex_free(struct gfs2_sbd *sdp)
418 {
419         struct list_head list, *head;
420         struct gfs2_jdesc *jd;
421         struct gfs2_journal_extent *jext;
422
423         spin_lock(&sdp->sd_jindex_spin);
424         list_add(&list, &sdp->sd_jindex_list);
425         list_del_init(&sdp->sd_jindex_list);
426         sdp->sd_journals = 0;
427         spin_unlock(&sdp->sd_jindex_spin);
428
429         while (!list_empty(&list)) {
430                 jd = list_entry(list.next, struct gfs2_jdesc, jd_list);
431                 head = &jd->extent_list;
432                 while (!list_empty(head)) {
433                         jext = list_entry(head->next,
434                                           struct gfs2_journal_extent,
435                                           extent_list);
436                         list_del(&jext->extent_list);
437                         kfree(jext);
438                 }
439                 list_del(&jd->jd_list);
440                 iput(jd->jd_inode);
441                 kfree(jd);
442         }
443 }
444
445 static struct gfs2_jdesc *jdesc_find_i(struct list_head *head, unsigned int jid)
446 {
447         struct gfs2_jdesc *jd;
448         int found = 0;
449
450         list_for_each_entry(jd, head, jd_list) {
451                 if (jd->jd_jid == jid) {
452                         found = 1;
453                         break;
454                 }
455         }
456
457         if (!found)
458                 jd = NULL;
459
460         return jd;
461 }
462
463 struct gfs2_jdesc *gfs2_jdesc_find(struct gfs2_sbd *sdp, unsigned int jid)
464 {
465         struct gfs2_jdesc *jd;
466
467         spin_lock(&sdp->sd_jindex_spin);
468         jd = jdesc_find_i(&sdp->sd_jindex_list, jid);
469         spin_unlock(&sdp->sd_jindex_spin);
470
471         return jd;
472 }
473
474 void gfs2_jdesc_make_dirty(struct gfs2_sbd *sdp, unsigned int jid)
475 {
476         struct gfs2_jdesc *jd;
477
478         spin_lock(&sdp->sd_jindex_spin);
479         jd = jdesc_find_i(&sdp->sd_jindex_list, jid);
480         if (jd)
481                 jd->jd_dirty = 1;
482         spin_unlock(&sdp->sd_jindex_spin);
483 }
484
485 struct gfs2_jdesc *gfs2_jdesc_find_dirty(struct gfs2_sbd *sdp)
486 {
487         struct gfs2_jdesc *jd;
488         int found = 0;
489
490         spin_lock(&sdp->sd_jindex_spin);
491
492         list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
493                 if (jd->jd_dirty) {
494                         jd->jd_dirty = 0;
495                         found = 1;
496                         break;
497                 }
498         }
499         spin_unlock(&sdp->sd_jindex_spin);
500
501         if (!found)
502                 jd = NULL;
503
504         return jd;
505 }
506
507 int gfs2_jdesc_check(struct gfs2_jdesc *jd)
508 {
509         struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
510         struct gfs2_sbd *sdp = GFS2_SB(jd->jd_inode);
511         int ar;
512         int error;
513
514         if (ip->i_di.di_size < (8 << 20) || ip->i_di.di_size > (1 << 30) ||
515             (ip->i_di.di_size & (sdp->sd_sb.sb_bsize - 1))) {
516                 gfs2_consist_inode(ip);
517                 return -EIO;
518         }
519         jd->jd_blocks = ip->i_di.di_size >> sdp->sd_sb.sb_bsize_shift;
520
521         error = gfs2_write_alloc_required(ip, 0, ip->i_di.di_size, &ar);
522         if (!error && ar) {
523                 gfs2_consist_inode(ip);
524                 error = -EIO;
525         }
526
527         return error;
528 }
529
530 /**
531  * gfs2_make_fs_rw - Turn a Read-Only FS into a Read-Write one
532  * @sdp: the filesystem
533  *
534  * Returns: errno
535  */
536
537 int gfs2_make_fs_rw(struct gfs2_sbd *sdp)
538 {
539         struct gfs2_inode *ip = GFS2_I(sdp->sd_jdesc->jd_inode);
540         struct gfs2_glock *j_gl = ip->i_gl;
541         struct gfs2_holder t_gh;
542         struct gfs2_log_header_host head;
543         int error;
544
545         error = gfs2_glock_nq_init(sdp->sd_trans_gl, LM_ST_SHARED, 0, &t_gh);
546         if (error)
547                 return error;
548
549         j_gl->gl_ops->go_inval(j_gl, DIO_METADATA);
550
551         error = gfs2_find_jhead(sdp->sd_jdesc, &head);
552         if (error)
553                 goto fail;
554
555         if (!(head.lh_flags & GFS2_LOG_HEAD_UNMOUNT)) {
556                 gfs2_consist(sdp);
557                 error = -EIO;
558                 goto fail;
559         }
560
561         /*  Initialize some head of the log stuff  */
562         sdp->sd_log_sequence = head.lh_sequence + 1;
563         gfs2_log_pointers_init(sdp, head.lh_blkno);
564
565         error = gfs2_quota_init(sdp);
566         if (error)
567                 goto fail;
568
569         set_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags);
570
571         gfs2_glock_dq_uninit(&t_gh);
572
573         return 0;
574
575 fail:
576         t_gh.gh_flags |= GL_NOCACHE;
577         gfs2_glock_dq_uninit(&t_gh);
578
579         return error;
580 }
581
582 /**
583  * gfs2_make_fs_ro - Turn a Read-Write FS into a Read-Only one
584  * @sdp: the filesystem
585  *
586  * Returns: errno
587  */
588
589 int gfs2_make_fs_ro(struct gfs2_sbd *sdp)
590 {
591         struct gfs2_holder t_gh;
592         int error;
593
594         gfs2_quota_sync(sdp);
595         gfs2_statfs_sync(sdp);
596
597         error = gfs2_glock_nq_init(sdp->sd_trans_gl, LM_ST_SHARED, GL_NOCACHE,
598                                    &t_gh);
599         if (error && !test_bit(SDF_SHUTDOWN, &sdp->sd_flags))
600                 return error;
601
602         gfs2_meta_syncfs(sdp);
603         gfs2_log_shutdown(sdp);
604
605         clear_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags);
606
607         if (t_gh.gh_gl)
608                 gfs2_glock_dq_uninit(&t_gh);
609
610         gfs2_quota_cleanup(sdp);
611
612         return error;
613 }
614
615 static void gfs2_statfs_change_in(struct gfs2_statfs_change_host *sc, const void *buf)
616 {
617         const struct gfs2_statfs_change *str = buf;
618
619         sc->sc_total = be64_to_cpu(str->sc_total);
620         sc->sc_free = be64_to_cpu(str->sc_free);
621         sc->sc_dinodes = be64_to_cpu(str->sc_dinodes);
622 }
623
624 static void gfs2_statfs_change_out(const struct gfs2_statfs_change_host *sc, void *buf)
625 {
626         struct gfs2_statfs_change *str = buf;
627
628         str->sc_total = cpu_to_be64(sc->sc_total);
629         str->sc_free = cpu_to_be64(sc->sc_free);
630         str->sc_dinodes = cpu_to_be64(sc->sc_dinodes);
631 }
632
633 int gfs2_statfs_init(struct gfs2_sbd *sdp)
634 {
635         struct gfs2_inode *m_ip = GFS2_I(sdp->sd_statfs_inode);
636         struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
637         struct gfs2_inode *l_ip = GFS2_I(sdp->sd_sc_inode);
638         struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
639         struct buffer_head *m_bh, *l_bh;
640         struct gfs2_holder gh;
641         int error;
642
643         error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE,
644                                    &gh);
645         if (error)
646                 return error;
647
648         error = gfs2_meta_inode_buffer(m_ip, &m_bh);
649         if (error)
650                 goto out;
651
652         if (sdp->sd_args.ar_spectator) {
653                 spin_lock(&sdp->sd_statfs_spin);
654                 gfs2_statfs_change_in(m_sc, m_bh->b_data +
655                                       sizeof(struct gfs2_dinode));
656                 spin_unlock(&sdp->sd_statfs_spin);
657         } else {
658                 error = gfs2_meta_inode_buffer(l_ip, &l_bh);
659                 if (error)
660                         goto out_m_bh;
661
662                 spin_lock(&sdp->sd_statfs_spin);
663                 gfs2_statfs_change_in(m_sc, m_bh->b_data +
664                                       sizeof(struct gfs2_dinode));
665                 gfs2_statfs_change_in(l_sc, l_bh->b_data +
666                                       sizeof(struct gfs2_dinode));
667                 spin_unlock(&sdp->sd_statfs_spin);
668
669                 brelse(l_bh);
670         }
671
672 out_m_bh:
673         brelse(m_bh);
674 out:
675         gfs2_glock_dq_uninit(&gh);
676         return 0;
677 }
678
679 void gfs2_statfs_change(struct gfs2_sbd *sdp, s64 total, s64 free,
680                         s64 dinodes)
681 {
682         struct gfs2_inode *l_ip = GFS2_I(sdp->sd_sc_inode);
683         struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
684         struct buffer_head *l_bh;
685         int error;
686
687         error = gfs2_meta_inode_buffer(l_ip, &l_bh);
688         if (error)
689                 return;
690
691         mutex_lock(&sdp->sd_statfs_mutex);
692         gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1);
693         mutex_unlock(&sdp->sd_statfs_mutex);
694
695         spin_lock(&sdp->sd_statfs_spin);
696         l_sc->sc_total += total;
697         l_sc->sc_free += free;
698         l_sc->sc_dinodes += dinodes;
699         gfs2_statfs_change_out(l_sc, l_bh->b_data + sizeof(struct gfs2_dinode));
700         spin_unlock(&sdp->sd_statfs_spin);
701
702         brelse(l_bh);
703 }
704
705 int gfs2_statfs_sync(struct gfs2_sbd *sdp)
706 {
707         struct gfs2_inode *m_ip = GFS2_I(sdp->sd_statfs_inode);
708         struct gfs2_inode *l_ip = GFS2_I(sdp->sd_sc_inode);
709         struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
710         struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
711         struct gfs2_holder gh;
712         struct buffer_head *m_bh, *l_bh;
713         int error;
714
715         error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE,
716                                    &gh);
717         if (error)
718                 return error;
719
720         error = gfs2_meta_inode_buffer(m_ip, &m_bh);
721         if (error)
722                 goto out;
723
724         spin_lock(&sdp->sd_statfs_spin);
725         gfs2_statfs_change_in(m_sc, m_bh->b_data +
726                               sizeof(struct gfs2_dinode));
727         if (!l_sc->sc_total && !l_sc->sc_free && !l_sc->sc_dinodes) {
728                 spin_unlock(&sdp->sd_statfs_spin);
729                 goto out_bh;
730         }
731         spin_unlock(&sdp->sd_statfs_spin);
732
733         error = gfs2_meta_inode_buffer(l_ip, &l_bh);
734         if (error)
735                 goto out_bh;
736
737         error = gfs2_trans_begin(sdp, 2 * RES_DINODE, 0);
738         if (error)
739                 goto out_bh2;
740
741         mutex_lock(&sdp->sd_statfs_mutex);
742         gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1);
743         mutex_unlock(&sdp->sd_statfs_mutex);
744
745         spin_lock(&sdp->sd_statfs_spin);
746         m_sc->sc_total += l_sc->sc_total;
747         m_sc->sc_free += l_sc->sc_free;
748         m_sc->sc_dinodes += l_sc->sc_dinodes;
749         memset(l_sc, 0, sizeof(struct gfs2_statfs_change));
750         memset(l_bh->b_data + sizeof(struct gfs2_dinode),
751                0, sizeof(struct gfs2_statfs_change));
752         spin_unlock(&sdp->sd_statfs_spin);
753
754         gfs2_trans_add_bh(m_ip->i_gl, m_bh, 1);
755         gfs2_statfs_change_out(m_sc, m_bh->b_data + sizeof(struct gfs2_dinode));
756
757         gfs2_trans_end(sdp);
758
759 out_bh2:
760         brelse(l_bh);
761 out_bh:
762         brelse(m_bh);
763 out:
764         gfs2_glock_dq_uninit(&gh);
765         return error;
766 }
767
768 /**
769  * gfs2_statfs_i - Do a statfs
770  * @sdp: the filesystem
771  * @sg: the sg structure
772  *
773  * Returns: errno
774  */
775
776 int gfs2_statfs_i(struct gfs2_sbd *sdp, struct gfs2_statfs_change_host *sc)
777 {
778         struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
779         struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
780
781         spin_lock(&sdp->sd_statfs_spin);
782
783         *sc = *m_sc;
784         sc->sc_total += l_sc->sc_total;
785         sc->sc_free += l_sc->sc_free;
786         sc->sc_dinodes += l_sc->sc_dinodes;
787
788         spin_unlock(&sdp->sd_statfs_spin);
789
790         if (sc->sc_free < 0)
791                 sc->sc_free = 0;
792         if (sc->sc_free > sc->sc_total)
793                 sc->sc_free = sc->sc_total;
794         if (sc->sc_dinodes < 0)
795                 sc->sc_dinodes = 0;
796
797         return 0;
798 }
799
800 /**
801  * statfs_fill - fill in the sg for a given RG
802  * @rgd: the RG
803  * @sc: the sc structure
804  *
805  * Returns: 0 on success, -ESTALE if the LVB is invalid
806  */
807
808 static int statfs_slow_fill(struct gfs2_rgrpd *rgd,
809                             struct gfs2_statfs_change_host *sc)
810 {
811         gfs2_rgrp_verify(rgd);
812         sc->sc_total += rgd->rd_data;
813         sc->sc_free += rgd->rd_rg.rg_free;
814         sc->sc_dinodes += rgd->rd_rg.rg_dinodes;
815         return 0;
816 }
817
818 /**
819  * gfs2_statfs_slow - Stat a filesystem using asynchronous locking
820  * @sdp: the filesystem
821  * @sc: the sc info that will be returned
822  *
823  * Any error (other than a signal) will cause this routine to fall back
824  * to the synchronous version.
825  *
826  * FIXME: This really shouldn't busy wait like this.
827  *
828  * Returns: errno
829  */
830
831 int gfs2_statfs_slow(struct gfs2_sbd *sdp, struct gfs2_statfs_change_host *sc)
832 {
833         struct gfs2_holder ri_gh;
834         struct gfs2_rgrpd *rgd_next;
835         struct gfs2_holder *gha, *gh;
836         unsigned int slots = 64;
837         unsigned int x;
838         int done;
839         int error = 0, err;
840
841         memset(sc, 0, sizeof(struct gfs2_statfs_change_host));
842         gha = kcalloc(slots, sizeof(struct gfs2_holder), GFP_KERNEL);
843         if (!gha)
844                 return -ENOMEM;
845
846         error = gfs2_rindex_hold(sdp, &ri_gh);
847         if (error)
848                 goto out;
849
850         rgd_next = gfs2_rgrpd_get_first(sdp);
851
852         for (;;) {
853                 done = 1;
854
855                 for (x = 0; x < slots; x++) {
856                         gh = gha + x;
857
858                         if (gh->gh_gl && gfs2_glock_poll(gh)) {
859                                 err = gfs2_glock_wait(gh);
860                                 if (err) {
861                                         gfs2_holder_uninit(gh);
862                                         error = err;
863                                 } else {
864                                         if (!error)
865                                                 error = statfs_slow_fill(
866                                                         gh->gh_gl->gl_object, sc);
867                                         gfs2_glock_dq_uninit(gh);
868                                 }
869                         }
870
871                         if (gh->gh_gl)
872                                 done = 0;
873                         else if (rgd_next && !error) {
874                                 error = gfs2_glock_nq_init(rgd_next->rd_gl,
875                                                            LM_ST_SHARED,
876                                                            GL_ASYNC,
877                                                            gh);
878                                 rgd_next = gfs2_rgrpd_get_next(rgd_next);
879                                 done = 0;
880                         }
881
882                         if (signal_pending(current))
883                                 error = -ERESTARTSYS;
884                 }
885
886                 if (done)
887                         break;
888
889                 yield();
890         }
891
892         gfs2_glock_dq_uninit(&ri_gh);
893
894 out:
895         kfree(gha);
896         return error;
897 }
898
899 struct lfcc {
900         struct list_head list;
901         struct gfs2_holder gh;
902 };
903
904 /**
905  * gfs2_lock_fs_check_clean - Stop all writes to the FS and check that all
906  *                            journals are clean
907  * @sdp: the file system
908  * @state: the state to put the transaction lock into
909  * @t_gh: the hold on the transaction lock
910  *
911  * Returns: errno
912  */
913
914 static int gfs2_lock_fs_check_clean(struct gfs2_sbd *sdp,
915                                     struct gfs2_holder *t_gh)
916 {
917         struct gfs2_inode *ip;
918         struct gfs2_holder ji_gh;
919         struct gfs2_jdesc *jd;
920         struct lfcc *lfcc;
921         LIST_HEAD(list);
922         struct gfs2_log_header_host lh;
923         int error;
924
925         error = gfs2_jindex_hold(sdp, &ji_gh);
926         if (error)
927                 return error;
928
929         list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
930                 lfcc = kmalloc(sizeof(struct lfcc), GFP_KERNEL);
931                 if (!lfcc) {
932                         error = -ENOMEM;
933                         goto out;
934                 }
935                 ip = GFS2_I(jd->jd_inode);
936                 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &lfcc->gh);
937                 if (error) {
938                         kfree(lfcc);
939                         goto out;
940                 }
941                 list_add(&lfcc->list, &list);
942         }
943
944         error = gfs2_glock_nq_init(sdp->sd_trans_gl, LM_ST_DEFERRED,
945                                LM_FLAG_PRIORITY | GL_NOCACHE,
946                                t_gh);
947
948         list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
949                 error = gfs2_jdesc_check(jd);
950                 if (error)
951                         break;
952                 error = gfs2_find_jhead(jd, &lh);
953                 if (error)
954                         break;
955                 if (!(lh.lh_flags & GFS2_LOG_HEAD_UNMOUNT)) {
956                         error = -EBUSY;
957                         break;
958                 }
959         }
960
961         if (error)
962                 gfs2_glock_dq_uninit(t_gh);
963
964 out:
965         while (!list_empty(&list)) {
966                 lfcc = list_entry(list.next, struct lfcc, list);
967                 list_del(&lfcc->list);
968                 gfs2_glock_dq_uninit(&lfcc->gh);
969                 kfree(lfcc);
970         }
971         gfs2_glock_dq_uninit(&ji_gh);
972         return error;
973 }
974
975 /**
976  * gfs2_freeze_fs - freezes the file system
977  * @sdp: the file system
978  *
979  * This function flushes data and meta data for all machines by
980  * aquiring the transaction log exclusively.  All journals are
981  * ensured to be in a clean state as well.
982  *
983  * Returns: errno
984  */
985
986 int gfs2_freeze_fs(struct gfs2_sbd *sdp)
987 {
988         int error = 0;
989
990         mutex_lock(&sdp->sd_freeze_lock);
991
992         if (!sdp->sd_freeze_count++) {
993                 error = gfs2_lock_fs_check_clean(sdp, &sdp->sd_freeze_gh);
994                 if (error)
995                         sdp->sd_freeze_count--;
996         }
997
998         mutex_unlock(&sdp->sd_freeze_lock);
999
1000         return error;
1001 }
1002
1003 /**
1004  * gfs2_unfreeze_fs - unfreezes the file system
1005  * @sdp: the file system
1006  *
1007  * This function allows the file system to proceed by unlocking
1008  * the exclusively held transaction lock.  Other GFS2 nodes are
1009  * now free to acquire the lock shared and go on with their lives.
1010  *
1011  */
1012
1013 void gfs2_unfreeze_fs(struct gfs2_sbd *sdp)
1014 {
1015         mutex_lock(&sdp->sd_freeze_lock);
1016
1017         if (sdp->sd_freeze_count && !--sdp->sd_freeze_count)
1018                 gfs2_glock_dq_uninit(&sdp->sd_freeze_gh);
1019
1020         mutex_unlock(&sdp->sd_freeze_lock);
1021 }
1022