1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
4 * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved.
7 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
9 #include <linux/sched.h>
10 #include <linux/slab.h>
11 #include <linux/spinlock.h>
12 #include <linux/completion.h>
13 #include <linux/buffer_head.h>
14 #include <linux/kallsyms.h>
15 #include <linux/gfs2_ondisk.h>
26 #include "trace_gfs2.h"
28 int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks,
31 struct gfs2_trans *tr;
34 BUG_ON(current->journal_info);
35 BUG_ON(blocks == 0 && revokes == 0);
37 if (!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags))
40 tr = kmem_cache_zalloc(gfs2_trans_cachep, GFP_NOFS);
45 tr->tr_blocks = blocks;
46 tr->tr_revokes = revokes;
48 set_bit(TR_ALLOCED, &tr->tr_flags);
50 tr->tr_reserved += 6 + blocks;
52 tr->tr_reserved += gfs2_struct2blk(sdp, revokes);
53 INIT_LIST_HEAD(&tr->tr_databuf);
54 INIT_LIST_HEAD(&tr->tr_buf);
55 INIT_LIST_HEAD(&tr->tr_ail1_list);
56 INIT_LIST_HEAD(&tr->tr_ail2_list);
58 sb_start_intwrite(sdp->sd_vfs);
60 error = gfs2_log_reserve(sdp, tr->tr_reserved);
64 current->journal_info = tr;
69 sb_end_intwrite(sdp->sd_vfs);
70 kmem_cache_free(gfs2_trans_cachep, tr);
75 static void gfs2_print_trans(struct gfs2_sbd *sdp, const struct gfs2_trans *tr)
77 fs_warn(sdp, "Transaction created at: %pSR\n", (void *)tr->tr_ip);
78 fs_warn(sdp, "blocks=%u revokes=%u reserved=%u touched=%u\n",
79 tr->tr_blocks, tr->tr_revokes, tr->tr_reserved,
80 test_bit(TR_TOUCHED, &tr->tr_flags));
81 fs_warn(sdp, "Buf %u/%u Databuf %u/%u Revoke %u/%u\n",
82 tr->tr_num_buf_new, tr->tr_num_buf_rm,
83 tr->tr_num_databuf_new, tr->tr_num_databuf_rm,
84 tr->tr_num_revoke, tr->tr_num_revoke_rm);
87 void gfs2_trans_end(struct gfs2_sbd *sdp)
89 struct gfs2_trans *tr = current->journal_info;
91 int alloced = test_bit(TR_ALLOCED, &tr->tr_flags);
93 current->journal_info = NULL;
95 if (!test_bit(TR_TOUCHED, &tr->tr_flags)) {
96 gfs2_log_release(sdp, tr->tr_reserved);
98 gfs2_trans_free(sdp, tr);
99 sb_end_intwrite(sdp->sd_vfs);
104 nbuf = tr->tr_num_buf_new + tr->tr_num_databuf_new;
105 nbuf -= tr->tr_num_buf_rm;
106 nbuf -= tr->tr_num_databuf_rm;
108 if (gfs2_assert_withdraw(sdp, (nbuf <= tr->tr_blocks) &&
109 (tr->tr_num_revoke <= tr->tr_revokes)))
110 gfs2_print_trans(sdp, tr);
112 gfs2_log_commit(sdp, tr);
113 if (alloced && !test_bit(TR_ATTACHED, &tr->tr_flags))
114 gfs2_trans_free(sdp, tr);
115 up_read(&sdp->sd_log_flush_lock);
117 if (sdp->sd_vfs->s_flags & SB_SYNCHRONOUS)
118 gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL |
121 sb_end_intwrite(sdp->sd_vfs);
124 static struct gfs2_bufdata *gfs2_alloc_bufdata(struct gfs2_glock *gl,
125 struct buffer_head *bh)
127 struct gfs2_bufdata *bd;
129 bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL);
132 INIT_LIST_HEAD(&bd->bd_list);
138 * gfs2_trans_add_data - Add a databuf to the transaction.
139 * @gl: The inode glock associated with the buffer
140 * @bh: The buffer to add
142 * This is used in journaled data mode.
143 * We need to journal the data block in the same way as metadata in
144 * the functions above. The difference is that here we have a tag
145 * which is two __be64's being the block number (as per meta data)
146 * and a flag which says whether the data block needs escaping or
147 * not. This means we need a new log entry for each 251 or so data
148 * blocks, which isn't an enormous overhead but twice as much as
149 * for normal metadata blocks.
151 void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh)
153 struct gfs2_trans *tr = current->journal_info;
154 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
155 struct gfs2_bufdata *bd;
158 if (buffer_pinned(bh)) {
159 set_bit(TR_TOUCHED, &tr->tr_flags);
165 gfs2_log_unlock(sdp);
167 if (bh->b_private == NULL)
168 bd = gfs2_alloc_bufdata(gl, bh);
174 gfs2_assert(sdp, bd->bd_gl == gl);
175 set_bit(TR_TOUCHED, &tr->tr_flags);
176 if (list_empty(&bd->bd_list)) {
177 set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags);
178 set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags);
179 gfs2_pin(sdp, bd->bd_bh);
180 tr->tr_num_databuf_new++;
181 list_add_tail(&bd->bd_list, &tr->tr_databuf);
183 gfs2_log_unlock(sdp);
188 void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh)
191 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
192 struct gfs2_bufdata *bd;
193 struct gfs2_meta_header *mh;
194 struct gfs2_trans *tr = current->journal_info;
195 enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state);
198 if (buffer_pinned(bh)) {
199 set_bit(TR_TOUCHED, &tr->tr_flags);
205 gfs2_log_unlock(sdp);
207 lock_page(bh->b_page);
208 if (bh->b_private == NULL)
209 bd = gfs2_alloc_bufdata(gl, bh);
212 unlock_page(bh->b_page);
216 gfs2_assert(sdp, bd->bd_gl == gl);
217 set_bit(TR_TOUCHED, &tr->tr_flags);
218 if (!list_empty(&bd->bd_list))
220 set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags);
221 set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags);
222 mh = (struct gfs2_meta_header *)bd->bd_bh->b_data;
223 if (unlikely(mh->mh_magic != cpu_to_be32(GFS2_MAGIC))) {
224 fs_err(sdp, "Attempting to add uninitialised block to "
225 "journal (inplace block=%lld)\n",
226 (unsigned long long)bd->bd_bh->b_blocknr);
229 if (unlikely(state == SFS_FROZEN)) {
230 fs_info(sdp, "GFS2:adding buf while frozen\n");
231 gfs2_assert_withdraw(sdp, 0);
233 if (unlikely(gfs2_withdrawn(sdp))) {
234 fs_info(sdp, "GFS2:adding buf while withdrawn! 0x%llx\n",
235 (unsigned long long)bd->bd_bh->b_blocknr);
237 gfs2_pin(sdp, bd->bd_bh);
238 mh->__pad0 = cpu_to_be64(0);
239 mh->mh_jid = cpu_to_be32(sdp->sd_jdesc->jd_jid);
240 list_add(&bd->bd_list, &tr->tr_buf);
241 tr->tr_num_buf_new++;
243 gfs2_log_unlock(sdp);
248 void gfs2_trans_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd)
250 struct gfs2_trans *tr = current->journal_info;
252 BUG_ON(!list_empty(&bd->bd_list));
253 gfs2_add_revoke(sdp, bd);
254 set_bit(TR_TOUCHED, &tr->tr_flags);
258 void gfs2_trans_remove_revoke(struct gfs2_sbd *sdp, u64 blkno, unsigned int len)
260 struct gfs2_bufdata *bd, *tmp;
261 struct gfs2_trans *tr = current->journal_info;
262 unsigned int n = len;
265 list_for_each_entry_safe(bd, tmp, &sdp->sd_log_revokes, bd_list) {
266 if ((bd->bd_blkno >= blkno) && (bd->bd_blkno < (blkno + len))) {
267 list_del_init(&bd->bd_list);
268 gfs2_assert_withdraw(sdp, sdp->sd_log_num_revoke);
269 sdp->sd_log_num_revoke--;
271 gfs2_glock_remove_revoke(bd->bd_gl);
272 kmem_cache_free(gfs2_bufdata_cachep, bd);
273 tr->tr_num_revoke_rm++;
278 gfs2_log_unlock(sdp);
281 void gfs2_trans_free(struct gfs2_sbd *sdp, struct gfs2_trans *tr)
286 gfs2_assert_warn(sdp, list_empty(&tr->tr_ail1_list));
287 gfs2_assert_warn(sdp, list_empty(&tr->tr_ail2_list));
288 gfs2_assert_warn(sdp, list_empty(&tr->tr_databuf));
289 gfs2_assert_warn(sdp, list_empty(&tr->tr_buf));
290 kmem_cache_free(gfs2_trans_cachep, tr);