glops.c 20.5 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0-only
D
David Teigland 已提交
2 3
/*
 * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
4
 * Copyright (C) 2004-2008 Red Hat, Inc.  All rights reserved.
D
David Teigland 已提交
5 6 7 8 9
 */

#include <linux/spinlock.h>
#include <linux/completion.h>
#include <linux/buffer_head.h>
10
#include <linux/gfs2_ondisk.h>
11
#include <linux/bio.h>
12
#include <linux/posix_acl.h>
13
#include <linux/security.h>
D
David Teigland 已提交
14 15

#include "gfs2.h"
16
#include "incore.h"
D
David Teigland 已提交
17 18 19 20 21 22 23 24
#include "bmap.h"
#include "glock.h"
#include "glops.h"
#include "inode.h"
#include "log.h"
#include "meta_io.h"
#include "recovery.h"
#include "rgrp.h"
25
#include "util.h"
26
#include "trans.h"
27
#include "dir.h"
A
Abhi Das 已提交
28
#include "lops.h"
D
David Teigland 已提交
29

30 31
struct workqueue_struct *gfs2_freeze_wq;

32 33
extern struct workqueue_struct *gfs2_control_wq;

34 35
static void gfs2_ail_error(struct gfs2_glock *gl, const struct buffer_head *bh)
{
36 37 38
	fs_err(gl->gl_name.ln_sbd,
	       "AIL buffer %p: blocknr %llu state 0x%08lx mapping %p page "
	       "state 0x%lx\n",
39 40
	       bh, (unsigned long long)bh->b_blocknr, bh->b_state,
	       bh->b_page->mapping, bh->b_page->flags);
41
	fs_err(gl->gl_name.ln_sbd, "AIL glock %u:%llu mapping %p\n",
42 43
	       gl->gl_name.ln_type, gl->gl_name.ln_number,
	       gfs2_glock2aspace(gl));
44 45
	gfs2_lm(gl->gl_name.ln_sbd, "AIL error\n");
	gfs2_withdraw(gl->gl_name.ln_sbd);
46 47
}

48
/**
S
Steven Whitehouse 已提交
49
 * __gfs2_ail_flush - remove all buffers for a given lock from the AIL
50
 * @gl: the glock
51
 * @fsync: set when called from fsync (not all buffers will be clean)
52 53 54 55
 *
 * None of the buffers should be dirty, locked, or pinned.
 */

56 57
static void __gfs2_ail_flush(struct gfs2_glock *gl, bool fsync,
			     unsigned int nr_revokes)
58
{
59
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
60
	struct list_head *head = &gl->gl_ail_list;
61
	struct gfs2_bufdata *bd, *tmp;
62
	struct buffer_head *bh;
63
	const unsigned long b_state = (1UL << BH_Dirty)|(1UL << BH_Pinned)|(1UL << BH_Lock);
64

65
	gfs2_log_lock(sdp);
D
Dave Chinner 已提交
66
	spin_lock(&sdp->sd_ail_lock);
67 68 69
	list_for_each_entry_safe_reverse(bd, tmp, head, bd_ail_gl_list) {
		if (nr_revokes == 0)
			break;
70
		bh = bd->bd_bh;
71 72 73
		if (bh->b_state & b_state) {
			if (fsync)
				continue;
74
			gfs2_ail_error(gl, bh);
75
		}
76
		gfs2_trans_add_revoke(sdp, bd);
77
		nr_revokes--;
78
	}
79
	GLOCK_BUG_ON(gl, !fsync && atomic_read(&gl->gl_ail_count));
D
Dave Chinner 已提交
80
	spin_unlock(&sdp->sd_ail_lock);
81
	gfs2_log_unlock(sdp);
S
Steven Whitehouse 已提交
82 83 84
}


85
static int gfs2_ail_empty_gl(struct gfs2_glock *gl)
S
Steven Whitehouse 已提交
86
{
87
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
S
Steven Whitehouse 已提交
88
	struct gfs2_trans tr;
89
	int ret;
S
Steven Whitehouse 已提交
90 91

	memset(&tr, 0, sizeof(tr));
92 93
	INIT_LIST_HEAD(&tr.tr_buf);
	INIT_LIST_HEAD(&tr.tr_databuf);
94 95
	INIT_LIST_HEAD(&tr.tr_ail1_list);
	INIT_LIST_HEAD(&tr.tr_ail2_list);
S
Steven Whitehouse 已提交
96 97
	tr.tr_revokes = atomic_read(&gl->gl_ail_count);

98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121
	if (!tr.tr_revokes) {
		bool have_revokes;
		bool log_in_flight;

		/*
		 * We have nothing on the ail, but there could be revokes on
		 * the sdp revoke queue, in which case, we still want to flush
		 * the log and wait for it to finish.
		 *
		 * If the sdp revoke list is empty too, we might still have an
		 * io outstanding for writing revokes, so we should wait for
		 * it before returning.
		 *
		 * If none of these conditions are true, our revokes are all
		 * flushed and we can return.
		 */
		gfs2_log_lock(sdp);
		have_revokes = !list_empty(&sdp->sd_log_revokes);
		log_in_flight = atomic_read(&sdp->sd_log_in_flight);
		gfs2_log_unlock(sdp);
		if (have_revokes)
			goto flush;
		if (log_in_flight)
			log_flush_wait(sdp);
122
		return 0;
123
	}
S
Steven Whitehouse 已提交
124

125 126 127
	/* A shortened, inline version of gfs2_trans_begin()
         * tr->alloced is not set since the transaction structure is
         * on the stack */
128
	tr.tr_reserved = 1 + gfs2_struct2blk(sdp, tr.tr_revokes);
129
	tr.tr_ip = _RET_IP_;
130 131 132
	ret = gfs2_log_reserve(sdp, tr.tr_reserved);
	if (ret < 0)
		return ret;
133
	WARN_ON_ONCE(current->journal_info);
S
Steven Whitehouse 已提交
134 135
	current->journal_info = &tr;

136
	__gfs2_ail_flush(gl, 0, tr.tr_revokes);
S
Steven Whitehouse 已提交
137 138

	gfs2_trans_end(sdp);
139
flush:
140 141
	gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL |
		       GFS2_LFC_AIL_EMPTY_GL);
142
	return 0;
S
Steven Whitehouse 已提交
143
}
144

145
void gfs2_ail_flush(struct gfs2_glock *gl, bool fsync)
S
Steven Whitehouse 已提交
146
{
147
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
S
Steven Whitehouse 已提交
148
	unsigned int revokes = atomic_read(&gl->gl_ail_count);
149
	unsigned int max_revokes = (sdp->sd_sb.sb_bsize - sizeof(struct gfs2_log_descriptor)) / sizeof(u64);
S
Steven Whitehouse 已提交
150 151 152 153 154
	int ret;

	if (!revokes)
		return;

155 156 157 158
	while (revokes > max_revokes)
		max_revokes += (sdp->sd_sb.sb_bsize - sizeof(struct gfs2_meta_header)) / sizeof(u64);

	ret = gfs2_trans_begin(sdp, 0, max_revokes);
S
Steven Whitehouse 已提交
159 160
	if (ret)
		return;
161
	__gfs2_ail_flush(gl, fsync, max_revokes);
162
	gfs2_trans_end(sdp);
163 164
	gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL |
		       GFS2_LFC_AIL_FLUSH);
165
}
S
Steven Whitehouse 已提交
166 167

/**
S
Steven Whitehouse 已提交
168
 * rgrp_go_sync - sync out the metadata for this glock
D
David Teigland 已提交
169 170 171 172
 * @gl: the glock
 *
 * Called when demoting or unlocking an EX glock.  We must flush
 * to disk all dirty buffers/pages relating to this glock, and must not
173
 * return to caller to demote/unlock the glock until I/O is complete.
D
David Teigland 已提交
174 175
 */

176
static int rgrp_go_sync(struct gfs2_glock *gl)
D
David Teigland 已提交
177
{
178
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
179
	struct address_space *mapping = &sdp->sd_aspace;
180
	struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl);
B
Bob Peterson 已提交
181 182 183
	const unsigned bsize = sdp->sd_sb.sb_bsize;
	loff_t start = (rgd->rd_addr * bsize) & PAGE_MASK;
	loff_t end = PAGE_ALIGN((rgd->rd_addr + rgd->rd_length) * bsize) - 1;
S
Steven Whitehouse 已提交
184 185 186
	int error;

	if (!test_and_clear_bit(GLF_DIRTY, &gl->gl_flags))
187
		return 0;
188
	GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE);
S
Steven Whitehouse 已提交
189

190 191
	gfs2_log_flush(sdp, gl, GFS2_LOG_HEAD_FLUSH_NORMAL |
		       GFS2_LFC_RGRP_GO_SYNC);
B
Bob Peterson 已提交
192 193
	filemap_fdatawrite_range(mapping, start, end);
	error = filemap_fdatawait_range(mapping, start, end);
194
	WARN_ON_ONCE(error && !gfs2_withdrawn(sdp));
195
	mapping_set_error(mapping, error);
196 197
	if (!error)
		error = gfs2_ail_empty_gl(gl);
B
Bob Peterson 已提交
198
	gfs2_free_clones(rgd);
199
	return error;
D
David Teigland 已提交
200 201 202
}

/**
S
Steven Whitehouse 已提交
203
 * rgrp_go_inval - invalidate the metadata for this glock
D
David Teigland 已提交
204 205 206
 * @gl: the glock
 * @flags:
 *
S
Steven Whitehouse 已提交
207 208 209
 * We never used LM_ST_DEFERRED with resource groups, so that we
 * should always see the metadata flag set here.
 *
D
David Teigland 已提交
210 211
 */

S
Steven Whitehouse 已提交
212
static void rgrp_go_inval(struct gfs2_glock *gl, int flags)
D
David Teigland 已提交
213
{
214
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
215
	struct address_space *mapping = &sdp->sd_aspace;
216
	struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl);
B
Bob Peterson 已提交
217 218 219
	const unsigned bsize = sdp->sd_sb.sb_bsize;
	loff_t start = (rgd->rd_addr * bsize) & PAGE_MASK;
	loff_t end = PAGE_ALIGN((rgd->rd_addr + rgd->rd_length) * bsize) - 1;
220

B
Bob Peterson 已提交
221
	gfs2_rgrp_brelse(rgd);
222
	WARN_ON_ONCE(!(flags & DIO_METADATA));
B
Bob Peterson 已提交
223 224
	truncate_inode_pages_range(mapping, start, end);
	rgd->rd_flags &= ~GFS2_RDF_UPTODATE;
D
David Teigland 已提交
225 226
}

227 228 229 230 231 232 233 234 235
static void gfs2_rgrp_go_dump(struct seq_file *seq, struct gfs2_glock *gl,
			      const char *fs_id_buf)
{
	struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl);

	if (rgd)
		gfs2_rgrp_dump(seq, rgd, fs_id_buf);
}

236 237 238 239 240 241 242 243 244 245 246 247
static struct gfs2_inode *gfs2_glock2inode(struct gfs2_glock *gl)
{
	struct gfs2_inode *ip;

	spin_lock(&gl->gl_lockref.lock);
	ip = gl->gl_object;
	if (ip)
		set_bit(GIF_GLOP_PENDING, &ip->i_flags);
	spin_unlock(&gl->gl_lockref.lock);
	return ip;
}

248 249 250 251 252 253 254 255 256 257 258
struct gfs2_rgrpd *gfs2_glock2rgrp(struct gfs2_glock *gl)
{
	struct gfs2_rgrpd *rgd;

	spin_lock(&gl->gl_lockref.lock);
	rgd = gl->gl_object;
	spin_unlock(&gl->gl_lockref.lock);

	return rgd;
}

259 260 261 262 263 264 265 266 267
static void gfs2_clear_glop_pending(struct gfs2_inode *ip)
{
	if (!ip)
		return;

	clear_bit_unlock(GIF_GLOP_PENDING, &ip->i_flags);
	wake_up_bit(&ip->i_flags, GIF_GLOP_PENDING);
}

S
Steven Whitehouse 已提交
268 269 270 271 272 273
/**
 * inode_go_sync - Sync the dirty data and/or metadata for an inode glock
 * @gl: the glock protecting the inode
 *
 */

274
static int inode_go_sync(struct gfs2_glock *gl)
S
Steven Whitehouse 已提交
275
{
276 277
	struct gfs2_inode *ip = gfs2_glock2inode(gl);
	int isreg = ip && S_ISREG(ip->i_inode.i_mode);
278
	struct address_space *metamapping = gfs2_glock2aspace(gl);
279
	int error = 0, ret;
280

281
	if (isreg) {
282 283 284 285
		if (test_and_clear_bit(GIF_SW_PAGED, &ip->i_flags))
			unmap_shared_mapping_range(ip->i_inode.i_mapping, 0, 0);
		inode_dio_wait(&ip->i_inode);
	}
S
Steven Whitehouse 已提交
286
	if (!test_and_clear_bit(GLF_DIRTY, &gl->gl_flags))
287
		goto out;
S
Steven Whitehouse 已提交
288

289
	GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE);
S
Steven Whitehouse 已提交
290

291 292
	gfs2_log_flush(gl->gl_name.ln_sbd, gl, GFS2_LOG_HEAD_FLUSH_NORMAL |
		       GFS2_LFC_INODE_GO_SYNC);
S
Steven Whitehouse 已提交
293
	filemap_fdatawrite(metamapping);
294
	if (isreg) {
S
Steven Whitehouse 已提交
295 296 297 298
		struct address_space *mapping = ip->i_inode.i_mapping;
		filemap_fdatawrite(mapping);
		error = filemap_fdatawait(mapping);
		mapping_set_error(mapping, error);
S
Steven Whitehouse 已提交
299
	}
300 301 302 303
	ret = filemap_fdatawait(metamapping);
	mapping_set_error(metamapping, ret);
	if (!error)
		error = ret;
S
Steven Whitehouse 已提交
304
	gfs2_ail_empty_gl(gl);
305 306 307 308
	/*
	 * Writeback of the data mapping may cause the dirty flag to be set
	 * so we have to clear it again here.
	 */
309
	smp_mb__before_atomic();
310
	clear_bit(GLF_DIRTY, &gl->gl_flags);
311 312 313

out:
	gfs2_clear_glop_pending(ip);
314
	return error;
S
Steven Whitehouse 已提交
315 316
}

D
David Teigland 已提交
317 318 319 320
/**
 * inode_go_inval - prepare a inode glock to be released
 * @gl: the glock
 * @flags:
321 322
 *
 * Normally we invalidate everything, but if we are moving into
S
Steven Whitehouse 已提交
323 324
 * LM_ST_DEFERRED from LM_ST_SHARED or LM_ST_EXCLUSIVE then we
 * can keep hold of the metadata, since it won't have changed.
D
David Teigland 已提交
325 326 327 328 329
 *
 */

static void inode_go_inval(struct gfs2_glock *gl, int flags)
{
330
	struct gfs2_inode *ip = gfs2_glock2inode(gl);
D
David Teigland 已提交
331

S
Steven Whitehouse 已提交
332
	if (flags & DIO_METADATA) {
333
		struct address_space *mapping = gfs2_glock2aspace(gl);
S
Steven Whitehouse 已提交
334
		truncate_inode_pages(mapping, 0);
335
		if (ip) {
336
			set_bit(GIF_INVALID, &ip->i_flags);
337
			forget_all_cached_acls(&ip->i_inode);
338
			security_inode_invalidate_secctx(&ip->i_inode);
339
			gfs2_dir_hash_inval(ip);
340
		}
341 342
	}

343
	if (ip == GFS2_I(gl->gl_name.ln_sbd->sd_rindex)) {
344
		gfs2_log_flush(gl->gl_name.ln_sbd, NULL,
345 346
			       GFS2_LOG_HEAD_FLUSH_NORMAL |
			       GFS2_LFC_INODE_GO_INVAL);
347
		gl->gl_name.ln_sbd->sd_rindex_uptodate = 0;
348
	}
349
	if (ip && S_ISREG(ip->i_inode.i_mode))
350
		truncate_inode_pages(ip->i_inode.i_mapping, 0);
351 352

	gfs2_clear_glop_pending(ip);
D
David Teigland 已提交
353 354 355 356 357 358 359 360 361
}

/**
 * inode_go_demote_ok - Check to see if it's ok to unlock an inode glock
 * @gl: the glock
 *
 * Returns: 1 if it's ok
 */

362
static int inode_go_demote_ok(const struct gfs2_glock *gl)
D
David Teigland 已提交
363
{
364
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
365

366 367
	if (sdp->sd_jindex == gl->gl_object || sdp->sd_rindex == gl->gl_object)
		return 0;
368

369
	return 1;
D
David Teigland 已提交
370 371
}

372 373 374
static int gfs2_dinode_in(struct gfs2_inode *ip, const void *buf)
{
	const struct gfs2_dinode *str = buf;
375
	struct timespec64 atime;
376 377 378 379 380 381 382 383 384 385 386 387 388
	u16 height, depth;

	if (unlikely(ip->i_no_addr != be64_to_cpu(str->di_num.no_addr)))
		goto corrupt;
	ip->i_no_formal_ino = be64_to_cpu(str->di_num.no_formal_ino);
	ip->i_inode.i_mode = be32_to_cpu(str->di_mode);
	ip->i_inode.i_rdev = 0;
	switch (ip->i_inode.i_mode & S_IFMT) {
	case S_IFBLK:
	case S_IFCHR:
		ip->i_inode.i_rdev = MKDEV(be32_to_cpu(str->di_major),
					   be32_to_cpu(str->di_minor));
		break;
389
	}
390

391 392
	i_uid_write(&ip->i_inode, be32_to_cpu(str->di_uid));
	i_gid_write(&ip->i_inode, be32_to_cpu(str->di_gid));
393
	set_nlink(&ip->i_inode, be32_to_cpu(str->di_nlink));
394 395 396 397
	i_size_write(&ip->i_inode, be64_to_cpu(str->di_size));
	gfs2_set_inode_blocks(&ip->i_inode, be64_to_cpu(str->di_blocks));
	atime.tv_sec = be64_to_cpu(str->di_atime);
	atime.tv_nsec = be32_to_cpu(str->di_atime_nsec);
398
	if (timespec64_compare(&ip->i_inode.i_atime, &atime) < 0)
399 400 401 402 403 404 405 406 407 408
		ip->i_inode.i_atime = atime;
	ip->i_inode.i_mtime.tv_sec = be64_to_cpu(str->di_mtime);
	ip->i_inode.i_mtime.tv_nsec = be32_to_cpu(str->di_mtime_nsec);
	ip->i_inode.i_ctime.tv_sec = be64_to_cpu(str->di_ctime);
	ip->i_inode.i_ctime.tv_nsec = be32_to_cpu(str->di_ctime_nsec);

	ip->i_goal = be64_to_cpu(str->di_goal_meta);
	ip->i_generation = be64_to_cpu(str->di_generation);

	ip->i_diskflags = be32_to_cpu(str->di_flags);
S
Steven Whitehouse 已提交
409 410
	ip->i_eattr = be64_to_cpu(str->di_eattr);
	/* i_diskflags and i_eattr must be set before gfs2_set_inode_flags() */
411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454
	gfs2_set_inode_flags(&ip->i_inode);
	height = be16_to_cpu(str->di_height);
	if (unlikely(height > GFS2_MAX_META_HEIGHT))
		goto corrupt;
	ip->i_height = (u8)height;

	depth = be16_to_cpu(str->di_depth);
	if (unlikely(depth > GFS2_DIR_MAX_DEPTH))
		goto corrupt;
	ip->i_depth = (u8)depth;
	ip->i_entries = be32_to_cpu(str->di_entries);

	if (S_ISREG(ip->i_inode.i_mode))
		gfs2_set_aops(&ip->i_inode);

	return 0;
corrupt:
	gfs2_consist_inode(ip);
	return -EIO;
}

/**
 * gfs2_inode_refresh - Refresh the incore copy of the dinode
 * @ip: The GFS2 inode
 *
 * Returns: errno
 */

int gfs2_inode_refresh(struct gfs2_inode *ip)
{
	struct buffer_head *dibh;
	int error;

	error = gfs2_meta_inode_buffer(ip, &dibh);
	if (error)
		return error;

	error = gfs2_dinode_in(ip, dibh->b_data);
	brelse(dibh);
	clear_bit(GIF_INVALID, &ip->i_flags);

	return error;
}

D
David Teigland 已提交
455 456 457 458 459 460 461 462 463 464 465
/**
 * inode_go_lock - operation done after an inode lock is locked by a process
 * @gl: the glock
 * @flags:
 *
 * Returns: errno
 */

static int inode_go_lock(struct gfs2_holder *gh)
{
	struct gfs2_glock *gl = gh->gh_gl;
466
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
467
	struct gfs2_inode *ip = gl->gl_object;
D
David Teigland 已提交
468 469
	int error = 0;

470
	if (!ip || (gh->gh_flags & GL_SKIP))
D
David Teigland 已提交
471 472
		return 0;

473
	if (test_bit(GIF_INVALID, &ip->i_flags)) {
D
David Teigland 已提交
474 475 476 477 478
		error = gfs2_inode_refresh(ip);
		if (error)
			return error;
	}

479 480 481
	if (gh->gh_state != LM_ST_DEFERRED)
		inode_dio_wait(&ip->i_inode);

482
	if ((ip->i_diskflags & GFS2_DIF_TRUNC_IN_PROG) &&
D
David Teigland 已提交
483
	    (gl->gl_state == LM_ST_EXCLUSIVE) &&
484 485 486
	    (gh->gh_state == LM_ST_EXCLUSIVE)) {
		spin_lock(&sdp->sd_trunc_lock);
		if (list_empty(&ip->i_trunc_list))
487
			list_add(&ip->i_trunc_list, &sdp->sd_trunc_list);
488 489 490 491
		spin_unlock(&sdp->sd_trunc_lock);
		wake_up(&sdp->sd_quota_wait);
		return 1;
	}
D
David Teigland 已提交
492 493 494 495

	return error;
}

496 497 498 499
/**
 * inode_go_dump - print information about an inode
 * @seq: The iterator
 * @ip: the inode
500
 * @fs_id_buf: file system id (may be empty)
501 502 503
 *
 */

504 505
static void inode_go_dump(struct seq_file *seq, struct gfs2_glock *gl,
			  const char *fs_id_buf)
506
{
507 508 509 510
	struct gfs2_inode *ip = gl->gl_object;
	struct inode *inode = &ip->i_inode;
	unsigned long nrpages;

511
	if (ip == NULL)
512
		return;
513 514 515 516 517

	xa_lock_irq(&inode->i_data.i_pages);
	nrpages = inode->i_data.nrpages;
	xa_unlock_irq(&inode->i_data.i_pages);

518 519
	gfs2_print_dbg(seq, "%s I: n:%llu/%llu t:%u f:0x%02lx d:0x%08x s:%llu "
		       "p:%lu\n", fs_id_buf,
520 521
		  (unsigned long long)ip->i_no_formal_ino,
		  (unsigned long long)ip->i_no_addr,
522 523
		  IF2DT(ip->i_inode.i_mode), ip->i_flags,
		  (unsigned int)ip->i_diskflags,
524
		  (unsigned long long)i_size_read(inode), nrpages);
525 526
}

D
David Teigland 已提交
527
/**
528
 * freeze_go_sync - promote/demote the freeze glock
D
David Teigland 已提交
529 530 531 532 533 534
 * @gl: the glock
 * @state: the requested state
 * @flags:
 *
 */

535
static int freeze_go_sync(struct gfs2_glock *gl)
D
David Teigland 已提交
536
{
537
	int error = 0;
538
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
D
David Teigland 已提交
539

540
	if (gl->gl_req == LM_ST_EXCLUSIVE && !gfs2_withdrawn(sdp)) {
541 542 543
		atomic_set(&sdp->sd_freeze_state, SFS_STARTING_FREEZE);
		error = freeze_super(sdp->sd_vfs);
		if (error) {
544 545
			fs_info(sdp, "GFS2: couldn't freeze filesystem: %d\n",
				error);
546 547
			if (gfs2_withdrawn(sdp)) {
				atomic_set(&sdp->sd_freeze_state, SFS_UNFROZEN);
548
				return 0;
549
			}
550 551 552
			gfs2_assert_withdraw(sdp, 0);
		}
		queue_work(gfs2_freeze_wq, &sdp->sd_freeze_work);
553 554 555 556 557
		if (test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags))
			gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_FREEZE |
				       GFS2_LFC_FREEZE_GO_SYNC);
		else /* read-only mounts */
			atomic_set(&sdp->sd_freeze_state, SFS_FROZEN);
D
David Teigland 已提交
558
	}
559
	return 0;
D
David Teigland 已提交
560 561 562
}

/**
563
 * freeze_go_xmote_bh - After promoting/demoting the freeze glock
D
David Teigland 已提交
564 565 566 567
 * @gl: the glock
 *
 */

568
static int freeze_go_xmote_bh(struct gfs2_glock *gl, struct gfs2_holder *gh)
D
David Teigland 已提交
569
{
570
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
571
	struct gfs2_inode *ip = GFS2_I(sdp->sd_jdesc->jd_inode);
572
	struct gfs2_glock *j_gl = ip->i_gl;
A
Al Viro 已提交
573
	struct gfs2_log_header_host head;
D
David Teigland 已提交
574 575
	int error;

576
	if (test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) {
577
		j_gl->gl_ops->go_inval(j_gl, DIO_METADATA);
D
David Teigland 已提交
578

A
Abhi Das 已提交
579
		error = gfs2_find_jhead(sdp->sd_jdesc, &head, false);
D
David Teigland 已提交
580 581 582 583 584 585
		if (error)
			gfs2_consist(sdp);
		if (!(head.lh_flags & GFS2_LOG_HEAD_UNMOUNT))
			gfs2_consist(sdp);

		/*  Initialize some head of the log stuff  */
586
		if (!gfs2_withdrawn(sdp)) {
D
David Teigland 已提交
587 588 589 590
			sdp->sd_log_sequence = head.lh_sequence + 1;
			gfs2_log_pointers_init(sdp, head.lh_blkno);
		}
	}
591
	return 0;
D
David Teigland 已提交
592 593
}

594 595 596 597 598 599 600
/**
 * trans_go_demote_ok
 * @gl: the glock
 *
 * Always returns 0
 */

601
static int freeze_go_demote_ok(const struct gfs2_glock *gl)
602 603 604 605
{
	return 0;
}

606 607 608 609
/**
 * iopen_go_callback - schedule the dcache entry for the inode to be deleted
 * @gl: the glock
 *
A
Andreas Gruenbacher 已提交
610
 * gl_lockref.lock lock is held while calling this
611
 */
612
static void iopen_go_callback(struct gfs2_glock *gl, bool remote)
613
{
614
	struct gfs2_inode *ip = gl->gl_object;
615
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
616

617
	if (!remote || sb_rdonly(sdp->sd_vfs))
618
		return;
619 620

	if (gl->gl_demote_state == LM_ST_UNLOCKED &&
621
	    gl->gl_state == LM_ST_SHARED && ip) {
S
Steven Whitehouse 已提交
622
		gl->gl_lockref.count++;
623 624
		if (!queue_delayed_work(gfs2_delete_workqueue,
					&gl->gl_delete, 0))
S
Steven Whitehouse 已提交
625
			gl->gl_lockref.count--;
626 627 628
	}
}

629 630 631 632 633
static int iopen_go_demote_ok(const struct gfs2_glock *gl)
{
       return !gfs2_delete_work_queued(gl);
}

634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700
/**
 * inode_go_free - wake up anyone waiting for dlm's unlock ast to free it
 * @gl: glock being freed
 *
 * For now, this is only used for the journal inode glock. In withdraw
 * situations, we need to wait for the glock to be freed so that we know
 * other nodes may proceed with recovery / journal replay.
 */
static void inode_go_free(struct gfs2_glock *gl)
{
	/* Note that we cannot reference gl_object because it's already set
	 * to NULL by this point in its lifecycle. */
	if (!test_bit(GLF_FREEING, &gl->gl_flags))
		return;
	clear_bit_unlock(GLF_FREEING, &gl->gl_flags);
	wake_up_bit(&gl->gl_flags, GLF_FREEING);
}

/**
 * nondisk_go_callback - used to signal when a node did a withdraw
 * @gl: the nondisk glock
 * @remote: true if this came from a different cluster node
 *
 */
static void nondisk_go_callback(struct gfs2_glock *gl, bool remote)
{
	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;

	/* Ignore the callback unless it's from another node, and it's the
	   live lock. */
	if (!remote || gl->gl_name.ln_number != GFS2_LIVE_LOCK)
		return;

	/* First order of business is to cancel the demote request. We don't
	 * really want to demote a nondisk glock. At best it's just to inform
	 * us of another node's withdraw. We'll keep it in SH mode. */
	clear_bit(GLF_DEMOTE, &gl->gl_flags);
	clear_bit(GLF_PENDING_DEMOTE, &gl->gl_flags);

	/* Ignore the unlock if we're withdrawn, unmounting, or in recovery. */
	if (test_bit(SDF_NORECOVERY, &sdp->sd_flags) ||
	    test_bit(SDF_WITHDRAWN, &sdp->sd_flags) ||
	    test_bit(SDF_REMOTE_WITHDRAW, &sdp->sd_flags))
		return;

	/* We only care when a node wants us to unlock, because that means
	 * they want a journal recovered. */
	if (gl->gl_demote_state != LM_ST_UNLOCKED)
		return;

	if (sdp->sd_args.ar_spectator) {
		fs_warn(sdp, "Spectator node cannot recover journals.\n");
		return;
	}

	fs_warn(sdp, "Some node has withdrawn; checking for recovery.\n");
	set_bit(SDF_REMOTE_WITHDRAW, &sdp->sd_flags);
	/*
	 * We can't call remote_withdraw directly here or gfs2_recover_journal
	 * because this is called from the glock unlock function and the
	 * remote_withdraw needs to enqueue and dequeue the same "live" glock
	 * we were called from. So we queue it to the control work queue in
	 * lock_dlm.
	 */
	queue_delayed_work(gfs2_control_wq, &sdp->sd_control_work, 0);
}

701
const struct gfs2_glock_operations gfs2_meta_glops = {
702
	.go_type = LM_TYPE_META,
703
	.go_flags = GLOF_NONDISK,
D
David Teigland 已提交
704 705
};

706
const struct gfs2_glock_operations gfs2_inode_glops = {
707
	.go_sync = inode_go_sync,
D
David Teigland 已提交
708 709 710
	.go_inval = inode_go_inval,
	.go_demote_ok = inode_go_demote_ok,
	.go_lock = inode_go_lock,
711
	.go_dump = inode_go_dump,
712
	.go_type = LM_TYPE_INODE,
713
	.go_flags = GLOF_ASPACE | GLOF_LRU | GLOF_LVB,
714
	.go_free = inode_go_free,
D
David Teigland 已提交
715 716
};

717
const struct gfs2_glock_operations gfs2_rgrp_glops = {
718
	.go_sync = rgrp_go_sync,
S
Steven Whitehouse 已提交
719
	.go_inval = rgrp_go_inval,
720
	.go_lock = gfs2_rgrp_go_lock,
721
	.go_dump = gfs2_rgrp_go_dump,
722
	.go_type = LM_TYPE_RGRP,
723
	.go_flags = GLOF_LVB,
D
David Teigland 已提交
724 725
};

726 727 728 729
const struct gfs2_glock_operations gfs2_freeze_glops = {
	.go_sync = freeze_go_sync,
	.go_xmote_bh = freeze_go_xmote_bh,
	.go_demote_ok = freeze_go_demote_ok,
730
	.go_type = LM_TYPE_NONDISK,
731
	.go_flags = GLOF_NONDISK,
D
David Teigland 已提交
732 733
};

734
const struct gfs2_glock_operations gfs2_iopen_glops = {
735
	.go_type = LM_TYPE_IOPEN,
736
	.go_callback = iopen_go_callback,
737
	.go_demote_ok = iopen_go_demote_ok,
738
	.go_flags = GLOF_LRU | GLOF_NONDISK,
D
David Teigland 已提交
739 740
};

741
const struct gfs2_glock_operations gfs2_flock_glops = {
742
	.go_type = LM_TYPE_FLOCK,
743
	.go_flags = GLOF_LRU | GLOF_NONDISK,
D
David Teigland 已提交
744 745
};

746
const struct gfs2_glock_operations gfs2_nondisk_glops = {
747
	.go_type = LM_TYPE_NONDISK,
748
	.go_flags = GLOF_NONDISK,
749
	.go_callback = nondisk_go_callback,
D
David Teigland 已提交
750 751
};

752
const struct gfs2_glock_operations gfs2_quota_glops = {
753
	.go_type = LM_TYPE_QUOTA,
754
	.go_flags = GLOF_LVB | GLOF_LRU | GLOF_NONDISK,
D
David Teigland 已提交
755 756
};

757
const struct gfs2_glock_operations gfs2_journal_glops = {
758
	.go_type = LM_TYPE_JOURNAL,
759
	.go_flags = GLOF_NONDISK,
D
David Teigland 已提交
760 761
};

762 763 764 765 766 767 768 769 770 771 772
const struct gfs2_glock_operations *gfs2_glops_list[] = {
	[LM_TYPE_META] = &gfs2_meta_glops,
	[LM_TYPE_INODE] = &gfs2_inode_glops,
	[LM_TYPE_RGRP] = &gfs2_rgrp_glops,
	[LM_TYPE_IOPEN] = &gfs2_iopen_glops,
	[LM_TYPE_FLOCK] = &gfs2_flock_glops,
	[LM_TYPE_NONDISK] = &gfs2_nondisk_glops,
	[LM_TYPE_QUOTA] = &gfs2_quota_glops,
	[LM_TYPE_JOURNAL] = &gfs2_journal_glops,
};