cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

trans.c (9386B)


      1// SPDX-License-Identifier: GPL-2.0-only
      2/*
      3 * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
      4 * Copyright (C) 2004-2006 Red Hat, Inc.  All rights reserved.
      5 */
      6
      7#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
      8
      9#include <linux/sched.h>
     10#include <linux/slab.h>
     11#include <linux/spinlock.h>
     12#include <linux/completion.h>
     13#include <linux/buffer_head.h>
     14#include <linux/kallsyms.h>
     15#include <linux/gfs2_ondisk.h>
     16
     17#include "gfs2.h"
     18#include "incore.h"
     19#include "glock.h"
     20#include "inode.h"
     21#include "log.h"
     22#include "lops.h"
     23#include "meta_io.h"
     24#include "trans.h"
     25#include "util.h"
     26#include "trace_gfs2.h"
     27
     28static void gfs2_print_trans(struct gfs2_sbd *sdp, const struct gfs2_trans *tr)
     29{
     30	fs_warn(sdp, "Transaction created at: %pSR\n", (void *)tr->tr_ip);
     31	fs_warn(sdp, "blocks=%u revokes=%u reserved=%u touched=%u\n",
     32		tr->tr_blocks, tr->tr_revokes, tr->tr_reserved,
     33		test_bit(TR_TOUCHED, &tr->tr_flags));
     34	fs_warn(sdp, "Buf %u/%u Databuf %u/%u Revoke %u\n",
     35		tr->tr_num_buf_new, tr->tr_num_buf_rm,
     36		tr->tr_num_databuf_new, tr->tr_num_databuf_rm,
     37		tr->tr_num_revoke);
     38}
     39
     40int __gfs2_trans_begin(struct gfs2_trans *tr, struct gfs2_sbd *sdp,
     41		       unsigned int blocks, unsigned int revokes,
     42		       unsigned long ip)
     43{
     44	unsigned int extra_revokes;
     45
     46	if (current->journal_info) {
     47		gfs2_print_trans(sdp, current->journal_info);
     48		BUG();
     49	}
     50	BUG_ON(blocks == 0 && revokes == 0);
     51
     52	if (!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags))
     53		return -EROFS;
     54
     55	tr->tr_ip = ip;
     56	tr->tr_blocks = blocks;
     57	tr->tr_revokes = revokes;
     58	tr->tr_reserved = GFS2_LOG_FLUSH_MIN_BLOCKS;
     59	if (blocks) {
     60		/*
     61		 * The reserved blocks are either used for data or metadata.
     62		 * We can have mixed data and metadata, each with its own log
     63		 * descriptor block; see calc_reserved().
     64		 */
     65		tr->tr_reserved += blocks + 1 + DIV_ROUND_UP(blocks - 1, databuf_limit(sdp));
     66	}
     67	INIT_LIST_HEAD(&tr->tr_databuf);
     68	INIT_LIST_HEAD(&tr->tr_buf);
     69	INIT_LIST_HEAD(&tr->tr_list);
     70	INIT_LIST_HEAD(&tr->tr_ail1_list);
     71	INIT_LIST_HEAD(&tr->tr_ail2_list);
     72
     73	if (gfs2_assert_warn(sdp, tr->tr_reserved <= sdp->sd_jdesc->jd_blocks))
     74		return -EINVAL;
     75
     76	sb_start_intwrite(sdp->sd_vfs);
     77
     78	/*
     79	 * Try the reservations under sd_log_flush_lock to prevent log flushes
     80	 * from creating inconsistencies between the number of allocated and
     81	 * reserved revokes.  If that fails, do a full-block allocation outside
     82	 * of the lock to avoid stalling log flushes.  Then, allot the
     83	 * appropriate number of blocks to revokes, use as many revokes locally
     84	 * as needed, and "release" the surplus into the revokes pool.
     85	 */
     86
     87	down_read(&sdp->sd_log_flush_lock);
     88	if (gfs2_log_try_reserve(sdp, tr, &extra_revokes))
     89		goto reserved;
     90	up_read(&sdp->sd_log_flush_lock);
     91	gfs2_log_reserve(sdp, tr, &extra_revokes);
     92	down_read(&sdp->sd_log_flush_lock);
     93
     94reserved:
     95	gfs2_log_release_revokes(sdp, extra_revokes);
     96	if (unlikely(!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags))) {
     97		gfs2_log_release_revokes(sdp, tr->tr_revokes);
     98		up_read(&sdp->sd_log_flush_lock);
     99		gfs2_log_release(sdp, tr->tr_reserved);
    100		sb_end_intwrite(sdp->sd_vfs);
    101		return -EROFS;
    102	}
    103
    104	current->journal_info = tr;
    105
    106	return 0;
    107}
    108
    109int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks,
    110		     unsigned int revokes)
    111{
    112	struct gfs2_trans *tr;
    113	int error;
    114
    115	tr = kmem_cache_zalloc(gfs2_trans_cachep, GFP_NOFS);
    116	if (!tr)
    117		return -ENOMEM;
    118	error = __gfs2_trans_begin(tr, sdp, blocks, revokes, _RET_IP_);
    119	if (error)
    120		kmem_cache_free(gfs2_trans_cachep, tr);
    121	return error;
    122}
    123
    124void gfs2_trans_end(struct gfs2_sbd *sdp)
    125{
    126	struct gfs2_trans *tr = current->journal_info;
    127	s64 nbuf;
    128
    129	current->journal_info = NULL;
    130
    131	if (!test_bit(TR_TOUCHED, &tr->tr_flags)) {
    132		gfs2_log_release_revokes(sdp, tr->tr_revokes);
    133		up_read(&sdp->sd_log_flush_lock);
    134		gfs2_log_release(sdp, tr->tr_reserved);
    135		if (!test_bit(TR_ONSTACK, &tr->tr_flags))
    136			gfs2_trans_free(sdp, tr);
    137		sb_end_intwrite(sdp->sd_vfs);
    138		return;
    139	}
    140
    141	gfs2_log_release_revokes(sdp, tr->tr_revokes - tr->tr_num_revoke);
    142
    143	nbuf = tr->tr_num_buf_new + tr->tr_num_databuf_new;
    144	nbuf -= tr->tr_num_buf_rm;
    145	nbuf -= tr->tr_num_databuf_rm;
    146
    147	if (gfs2_assert_withdraw(sdp, nbuf <= tr->tr_blocks) ||
    148	    gfs2_assert_withdraw(sdp, tr->tr_num_revoke <= tr->tr_revokes))
    149		gfs2_print_trans(sdp, tr);
    150
    151	gfs2_log_commit(sdp, tr);
    152	if (!test_bit(TR_ONSTACK, &tr->tr_flags) &&
    153	    !test_bit(TR_ATTACHED, &tr->tr_flags))
    154		gfs2_trans_free(sdp, tr);
    155	up_read(&sdp->sd_log_flush_lock);
    156
    157	if (sdp->sd_vfs->s_flags & SB_SYNCHRONOUS)
    158		gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL |
    159			       GFS2_LFC_TRANS_END);
    160	sb_end_intwrite(sdp->sd_vfs);
    161}
    162
    163static struct gfs2_bufdata *gfs2_alloc_bufdata(struct gfs2_glock *gl,
    164					       struct buffer_head *bh)
    165{
    166	struct gfs2_bufdata *bd;
    167
    168	bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL);
    169	bd->bd_bh = bh;
    170	bd->bd_gl = gl;
    171	INIT_LIST_HEAD(&bd->bd_list);
    172	INIT_LIST_HEAD(&bd->bd_ail_st_list);
    173	INIT_LIST_HEAD(&bd->bd_ail_gl_list);
    174	bh->b_private = bd;
    175	return bd;
    176}
    177
    178/**
    179 * gfs2_trans_add_data - Add a databuf to the transaction.
    180 * @gl: The inode glock associated with the buffer
    181 * @bh: The buffer to add
    182 *
    183 * This is used in journaled data mode.
    184 * We need to journal the data block in the same way as metadata in
    185 * the functions above. The difference is that here we have a tag
    186 * which is two __be64's being the block number (as per meta data)
    187 * and a flag which says whether the data block needs escaping or
    188 * not. This means we need a new log entry for each 251 or so data
    189 * blocks, which isn't an enormous overhead but twice as much as
    190 * for normal metadata blocks.
    191 */
    192void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh)
    193{
    194	struct gfs2_trans *tr = current->journal_info;
    195	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
    196	struct gfs2_bufdata *bd;
    197
    198	lock_buffer(bh);
    199	if (buffer_pinned(bh)) {
    200		set_bit(TR_TOUCHED, &tr->tr_flags);
    201		goto out;
    202	}
    203	gfs2_log_lock(sdp);
    204	bd = bh->b_private;
    205	if (bd == NULL) {
    206		gfs2_log_unlock(sdp);
    207		unlock_buffer(bh);
    208		if (bh->b_private == NULL)
    209			bd = gfs2_alloc_bufdata(gl, bh);
    210		else
    211			bd = bh->b_private;
    212		lock_buffer(bh);
    213		gfs2_log_lock(sdp);
    214	}
    215	gfs2_assert(sdp, bd->bd_gl == gl);
    216	set_bit(TR_TOUCHED, &tr->tr_flags);
    217	if (list_empty(&bd->bd_list)) {
    218		set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags);
    219		set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags);
    220		gfs2_pin(sdp, bd->bd_bh);
    221		tr->tr_num_databuf_new++;
    222		list_add_tail(&bd->bd_list, &tr->tr_databuf);
    223	}
    224	gfs2_log_unlock(sdp);
    225out:
    226	unlock_buffer(bh);
    227}
    228
    229void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh)
    230{
    231
    232	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
    233	struct gfs2_bufdata *bd;
    234	struct gfs2_meta_header *mh;
    235	struct gfs2_trans *tr = current->journal_info;
    236	enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state);
    237
    238	lock_buffer(bh);
    239	if (buffer_pinned(bh)) {
    240		set_bit(TR_TOUCHED, &tr->tr_flags);
    241		goto out;
    242	}
    243	gfs2_log_lock(sdp);
    244	bd = bh->b_private;
    245	if (bd == NULL) {
    246		gfs2_log_unlock(sdp);
    247		unlock_buffer(bh);
    248		lock_page(bh->b_page);
    249		if (bh->b_private == NULL)
    250			bd = gfs2_alloc_bufdata(gl, bh);
    251		else
    252			bd = bh->b_private;
    253		unlock_page(bh->b_page);
    254		lock_buffer(bh);
    255		gfs2_log_lock(sdp);
    256	}
    257	gfs2_assert(sdp, bd->bd_gl == gl);
    258	set_bit(TR_TOUCHED, &tr->tr_flags);
    259	if (!list_empty(&bd->bd_list))
    260		goto out_unlock;
    261	set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags);
    262	set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags);
    263	mh = (struct gfs2_meta_header *)bd->bd_bh->b_data;
    264	if (unlikely(mh->mh_magic != cpu_to_be32(GFS2_MAGIC))) {
    265		fs_err(sdp, "Attempting to add uninitialised block to "
    266		       "journal (inplace block=%lld)\n",
    267		       (unsigned long long)bd->bd_bh->b_blocknr);
    268		BUG();
    269	}
    270	if (unlikely(state == SFS_FROZEN)) {
    271		fs_info(sdp, "GFS2:adding buf while frozen\n");
    272		gfs2_assert_withdraw(sdp, 0);
    273	}
    274	if (unlikely(gfs2_withdrawn(sdp))) {
    275		fs_info(sdp, "GFS2:adding buf while withdrawn! 0x%llx\n",
    276			(unsigned long long)bd->bd_bh->b_blocknr);
    277	}
    278	gfs2_pin(sdp, bd->bd_bh);
    279	mh->__pad0 = cpu_to_be64(0);
    280	mh->mh_jid = cpu_to_be32(sdp->sd_jdesc->jd_jid);
    281	list_add(&bd->bd_list, &tr->tr_buf);
    282	tr->tr_num_buf_new++;
    283out_unlock:
    284	gfs2_log_unlock(sdp);
    285out:
    286	unlock_buffer(bh);
    287}
    288
    289void gfs2_trans_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd)
    290{
    291	struct gfs2_trans *tr = current->journal_info;
    292
    293	BUG_ON(!list_empty(&bd->bd_list));
    294	gfs2_add_revoke(sdp, bd);
    295	set_bit(TR_TOUCHED, &tr->tr_flags);
    296	tr->tr_num_revoke++;
    297}
    298
    299void gfs2_trans_remove_revoke(struct gfs2_sbd *sdp, u64 blkno, unsigned int len)
    300{
    301	struct gfs2_bufdata *bd, *tmp;
    302	unsigned int n = len;
    303
    304	gfs2_log_lock(sdp);
    305	list_for_each_entry_safe(bd, tmp, &sdp->sd_log_revokes, bd_list) {
    306		if ((bd->bd_blkno >= blkno) && (bd->bd_blkno < (blkno + len))) {
    307			list_del_init(&bd->bd_list);
    308			gfs2_assert_withdraw(sdp, sdp->sd_log_num_revoke);
    309			sdp->sd_log_num_revoke--;
    310			if (bd->bd_gl)
    311				gfs2_glock_remove_revoke(bd->bd_gl);
    312			kmem_cache_free(gfs2_bufdata_cachep, bd);
    313			gfs2_log_release_revokes(sdp, 1);
    314			if (--n == 0)
    315				break;
    316		}
    317	}
    318	gfs2_log_unlock(sdp);
    319}
    320
    321void gfs2_trans_free(struct gfs2_sbd *sdp, struct gfs2_trans *tr)
    322{
    323	if (tr == NULL)
    324		return;
    325
    326	gfs2_assert_warn(sdp, list_empty(&tr->tr_ail1_list));
    327	gfs2_assert_warn(sdp, list_empty(&tr->tr_ail2_list));
    328	gfs2_assert_warn(sdp, list_empty(&tr->tr_databuf));
    329	gfs2_assert_warn(sdp, list_empty(&tr->tr_buf));
    330	kmem_cache_free(gfs2_trans_cachep, tr);
    331}