mirror of
https://github.com/torvalds/linux.git
synced 2024-12-28 13:51:44 +00:00
38552ff676
When gfs2_create_inode() fails after creating a new inode, it uses the GIF_FREE_VFS_INODE and GIF_ALLOC_FAILED inode flags to communicate to gfs2_evict_inode() which parts of the inode need to be deallocated and destroyed. In some error cases, the inode ends up being allocated on disk and then accidentally left behind. In others, the inode is partially constructed and then not properly destroyed. Clean this up by completely handling the inode deallocation and destruction in gfs2_evict_inode(). This means that gfs2_evict_inode() may now be faced with partially constructed inodes, so add the necessary checks to cope with that. In particular, make sure that for incompletely constructed inodes, we're not accessing the buffers backing the on-disk blocks; the contents may be undefined. Signed-off-by: Andreas Gruenbacher <agruenba@redhat.com>
555 lines
13 KiB
C
555 lines
13 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
|
|
* Copyright (C) 2004-2008 Red Hat, Inc. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/buffer_head.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/writeback.h>
|
|
#include <linux/swap.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/bio.h>
|
|
#include <linux/gfs2_ondisk.h>
|
|
|
|
#include "gfs2.h"
|
|
#include "incore.h"
|
|
#include "glock.h"
|
|
#include "glops.h"
|
|
#include "inode.h"
|
|
#include "log.h"
|
|
#include "lops.h"
|
|
#include "meta_io.h"
|
|
#include "rgrp.h"
|
|
#include "trans.h"
|
|
#include "util.h"
|
|
#include "trace_gfs2.h"
|
|
|
|
static int gfs2_aspace_writepage(struct page *page, struct writeback_control *wbc)
|
|
{
|
|
struct buffer_head *bh, *head;
|
|
int nr_underway = 0;
|
|
blk_opf_t write_flags = REQ_META | REQ_PRIO | wbc_to_write_flags(wbc);
|
|
|
|
BUG_ON(!PageLocked(page));
|
|
BUG_ON(!page_has_buffers(page));
|
|
|
|
head = page_buffers(page);
|
|
bh = head;
|
|
|
|
do {
|
|
if (!buffer_mapped(bh))
|
|
continue;
|
|
/*
|
|
* If it's a fully non-blocking write attempt and we cannot
|
|
* lock the buffer then redirty the page. Note that this can
|
|
* potentially cause a busy-wait loop from flusher thread and kswapd
|
|
* activity, but those code paths have their own higher-level
|
|
* throttling.
|
|
*/
|
|
if (wbc->sync_mode != WB_SYNC_NONE) {
|
|
lock_buffer(bh);
|
|
} else if (!trylock_buffer(bh)) {
|
|
redirty_page_for_writepage(wbc, page);
|
|
continue;
|
|
}
|
|
if (test_clear_buffer_dirty(bh)) {
|
|
mark_buffer_async_write(bh);
|
|
} else {
|
|
unlock_buffer(bh);
|
|
}
|
|
} while ((bh = bh->b_this_page) != head);
|
|
|
|
/*
|
|
* The page and its buffers are protected by PageWriteback(), so we can
|
|
* drop the bh refcounts early.
|
|
*/
|
|
BUG_ON(PageWriteback(page));
|
|
set_page_writeback(page);
|
|
|
|
do {
|
|
struct buffer_head *next = bh->b_this_page;
|
|
if (buffer_async_write(bh)) {
|
|
submit_bh(REQ_OP_WRITE | write_flags, bh);
|
|
nr_underway++;
|
|
}
|
|
bh = next;
|
|
} while (bh != head);
|
|
unlock_page(page);
|
|
|
|
if (nr_underway == 0)
|
|
end_page_writeback(page);
|
|
|
|
return 0;
|
|
}
|
|
|
|
const struct address_space_operations gfs2_meta_aops = {
|
|
.dirty_folio = block_dirty_folio,
|
|
.invalidate_folio = block_invalidate_folio,
|
|
.writepage = gfs2_aspace_writepage,
|
|
.release_folio = gfs2_release_folio,
|
|
};
|
|
|
|
const struct address_space_operations gfs2_rgrp_aops = {
|
|
.dirty_folio = block_dirty_folio,
|
|
.invalidate_folio = block_invalidate_folio,
|
|
.writepage = gfs2_aspace_writepage,
|
|
.release_folio = gfs2_release_folio,
|
|
};
|
|
|
|
/**
|
|
* gfs2_getbuf - Get a buffer with a given address space
|
|
* @gl: the glock
|
|
* @blkno: the block number (filesystem scope)
|
|
* @create: 1 if the buffer should be created
|
|
*
|
|
* Returns: the buffer
|
|
*/
|
|
|
|
struct buffer_head *gfs2_getbuf(struct gfs2_glock *gl, u64 blkno, int create)
|
|
{
|
|
struct address_space *mapping = gfs2_glock2aspace(gl);
|
|
struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
|
|
struct page *page;
|
|
struct buffer_head *bh;
|
|
unsigned int shift;
|
|
unsigned long index;
|
|
unsigned int bufnum;
|
|
|
|
if (mapping == NULL)
|
|
mapping = &sdp->sd_aspace;
|
|
|
|
shift = PAGE_SHIFT - sdp->sd_sb.sb_bsize_shift;
|
|
index = blkno >> shift; /* convert block to page */
|
|
bufnum = blkno - (index << shift); /* block buf index within page */
|
|
|
|
if (create) {
|
|
for (;;) {
|
|
page = grab_cache_page(mapping, index);
|
|
if (page)
|
|
break;
|
|
yield();
|
|
}
|
|
if (!page_has_buffers(page))
|
|
create_empty_buffers(page, sdp->sd_sb.sb_bsize, 0);
|
|
} else {
|
|
page = find_get_page_flags(mapping, index,
|
|
FGP_LOCK|FGP_ACCESSED);
|
|
if (!page)
|
|
return NULL;
|
|
if (!page_has_buffers(page)) {
|
|
bh = NULL;
|
|
goto out_unlock;
|
|
}
|
|
}
|
|
|
|
/* Locate header for our buffer within our page */
|
|
for (bh = page_buffers(page); bufnum--; bh = bh->b_this_page)
|
|
/* Do nothing */;
|
|
get_bh(bh);
|
|
|
|
if (!buffer_mapped(bh))
|
|
map_bh(bh, sdp->sd_vfs, blkno);
|
|
|
|
out_unlock:
|
|
unlock_page(page);
|
|
put_page(page);
|
|
|
|
return bh;
|
|
}
|
|
|
|
static void meta_prep_new(struct buffer_head *bh)
|
|
{
|
|
struct gfs2_meta_header *mh = (struct gfs2_meta_header *)bh->b_data;
|
|
|
|
lock_buffer(bh);
|
|
clear_buffer_dirty(bh);
|
|
set_buffer_uptodate(bh);
|
|
unlock_buffer(bh);
|
|
|
|
mh->mh_magic = cpu_to_be32(GFS2_MAGIC);
|
|
}
|
|
|
|
/**
|
|
* gfs2_meta_new - Get a block
|
|
* @gl: The glock associated with this block
|
|
* @blkno: The block number
|
|
*
|
|
* Returns: The buffer
|
|
*/
|
|
|
|
struct buffer_head *gfs2_meta_new(struct gfs2_glock *gl, u64 blkno)
|
|
{
|
|
struct buffer_head *bh;
|
|
bh = gfs2_getbuf(gl, blkno, CREATE);
|
|
meta_prep_new(bh);
|
|
return bh;
|
|
}
|
|
|
|
static void gfs2_meta_read_endio(struct bio *bio)
|
|
{
|
|
struct bio_vec *bvec;
|
|
struct bvec_iter_all iter_all;
|
|
|
|
bio_for_each_segment_all(bvec, bio, iter_all) {
|
|
struct page *page = bvec->bv_page;
|
|
struct buffer_head *bh = page_buffers(page);
|
|
unsigned int len = bvec->bv_len;
|
|
|
|
while (bh_offset(bh) < bvec->bv_offset)
|
|
bh = bh->b_this_page;
|
|
do {
|
|
struct buffer_head *next = bh->b_this_page;
|
|
len -= bh->b_size;
|
|
bh->b_end_io(bh, !bio->bi_status);
|
|
bh = next;
|
|
} while (bh && len);
|
|
}
|
|
bio_put(bio);
|
|
}
|
|
|
|
/*
|
|
* Submit several consecutive buffer head I/O requests as a single bio I/O
|
|
* request. (See submit_bh_wbc.)
|
|
*/
|
|
static void gfs2_submit_bhs(blk_opf_t opf, struct buffer_head *bhs[], int num)
|
|
{
|
|
while (num > 0) {
|
|
struct buffer_head *bh = *bhs;
|
|
struct bio *bio;
|
|
|
|
bio = bio_alloc(bh->b_bdev, num, opf, GFP_NOIO);
|
|
bio->bi_iter.bi_sector = bh->b_blocknr * (bh->b_size >> 9);
|
|
while (num > 0) {
|
|
bh = *bhs;
|
|
if (!bio_add_page(bio, bh->b_page, bh->b_size, bh_offset(bh))) {
|
|
BUG_ON(bio->bi_iter.bi_size == 0);
|
|
break;
|
|
}
|
|
bhs++;
|
|
num--;
|
|
}
|
|
bio->bi_end_io = gfs2_meta_read_endio;
|
|
submit_bio(bio);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* gfs2_meta_read - Read a block from disk
|
|
* @gl: The glock covering the block
|
|
* @blkno: The block number
|
|
* @flags: flags
|
|
* @rahead: Do read-ahead
|
|
* @bhp: the place where the buffer is returned (NULL on failure)
|
|
*
|
|
* Returns: errno
|
|
*/
|
|
|
|
int gfs2_meta_read(struct gfs2_glock *gl, u64 blkno, int flags,
|
|
int rahead, struct buffer_head **bhp)
|
|
{
|
|
struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
|
|
struct buffer_head *bh, *bhs[2];
|
|
int num = 0;
|
|
|
|
if (unlikely(gfs2_withdrawn(sdp)) && !gfs2_withdraw_in_prog(sdp)) {
|
|
*bhp = NULL;
|
|
return -EIO;
|
|
}
|
|
|
|
*bhp = bh = gfs2_getbuf(gl, blkno, CREATE);
|
|
|
|
lock_buffer(bh);
|
|
if (buffer_uptodate(bh)) {
|
|
unlock_buffer(bh);
|
|
flags &= ~DIO_WAIT;
|
|
} else {
|
|
bh->b_end_io = end_buffer_read_sync;
|
|
get_bh(bh);
|
|
bhs[num++] = bh;
|
|
}
|
|
|
|
if (rahead) {
|
|
bh = gfs2_getbuf(gl, blkno + 1, CREATE);
|
|
|
|
lock_buffer(bh);
|
|
if (buffer_uptodate(bh)) {
|
|
unlock_buffer(bh);
|
|
brelse(bh);
|
|
} else {
|
|
bh->b_end_io = end_buffer_read_sync;
|
|
bhs[num++] = bh;
|
|
}
|
|
}
|
|
|
|
gfs2_submit_bhs(REQ_OP_READ | REQ_META | REQ_PRIO, bhs, num);
|
|
if (!(flags & DIO_WAIT))
|
|
return 0;
|
|
|
|
bh = *bhp;
|
|
wait_on_buffer(bh);
|
|
if (unlikely(!buffer_uptodate(bh))) {
|
|
struct gfs2_trans *tr = current->journal_info;
|
|
if (tr && test_bit(TR_TOUCHED, &tr->tr_flags))
|
|
gfs2_io_error_bh_wd(sdp, bh);
|
|
brelse(bh);
|
|
*bhp = NULL;
|
|
return -EIO;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* gfs2_meta_wait - Reread a block from disk
|
|
* @sdp: the filesystem
|
|
* @bh: The block to wait for
|
|
*
|
|
* Returns: errno
|
|
*/
|
|
|
|
int gfs2_meta_wait(struct gfs2_sbd *sdp, struct buffer_head *bh)
|
|
{
|
|
if (unlikely(gfs2_withdrawn(sdp)) && !gfs2_withdraw_in_prog(sdp))
|
|
return -EIO;
|
|
|
|
wait_on_buffer(bh);
|
|
|
|
if (!buffer_uptodate(bh)) {
|
|
struct gfs2_trans *tr = current->journal_info;
|
|
if (tr && test_bit(TR_TOUCHED, &tr->tr_flags))
|
|
gfs2_io_error_bh_wd(sdp, bh);
|
|
return -EIO;
|
|
}
|
|
if (unlikely(gfs2_withdrawn(sdp)) && !gfs2_withdraw_in_prog(sdp))
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void gfs2_remove_from_journal(struct buffer_head *bh, int meta)
|
|
{
|
|
struct address_space *mapping = bh->b_page->mapping;
|
|
struct gfs2_sbd *sdp = gfs2_mapping2sbd(mapping);
|
|
struct gfs2_bufdata *bd = bh->b_private;
|
|
struct gfs2_trans *tr = current->journal_info;
|
|
int was_pinned = 0;
|
|
|
|
if (test_clear_buffer_pinned(bh)) {
|
|
trace_gfs2_pin(bd, 0);
|
|
atomic_dec(&sdp->sd_log_pinned);
|
|
list_del_init(&bd->bd_list);
|
|
if (meta == REMOVE_META)
|
|
tr->tr_num_buf_rm++;
|
|
else
|
|
tr->tr_num_databuf_rm++;
|
|
set_bit(TR_TOUCHED, &tr->tr_flags);
|
|
was_pinned = 1;
|
|
brelse(bh);
|
|
}
|
|
if (bd) {
|
|
if (bd->bd_tr) {
|
|
gfs2_trans_add_revoke(sdp, bd);
|
|
} else if (was_pinned) {
|
|
bh->b_private = NULL;
|
|
kmem_cache_free(gfs2_bufdata_cachep, bd);
|
|
} else if (!list_empty(&bd->bd_ail_st_list) &&
|
|
!list_empty(&bd->bd_ail_gl_list)) {
|
|
gfs2_remove_from_ail(bd);
|
|
}
|
|
}
|
|
clear_buffer_dirty(bh);
|
|
clear_buffer_uptodate(bh);
|
|
}
|
|
|
|
/**
|
|
* gfs2_ail1_wipe - remove deleted/freed buffers from the ail1 list
|
|
* @sdp: superblock
|
|
* @bstart: starting block address of buffers to remove
|
|
* @blen: length of buffers to be removed
|
|
*
|
|
* This function is called from gfs2_journal wipe, whose job is to remove
|
|
* buffers, corresponding to deleted blocks, from the journal. If we find any
|
|
* bufdata elements on the system ail1 list, they haven't been written to
|
|
* the journal yet. So we remove them.
|
|
*/
|
|
static void gfs2_ail1_wipe(struct gfs2_sbd *sdp, u64 bstart, u32 blen)
|
|
{
|
|
struct gfs2_trans *tr, *s;
|
|
struct gfs2_bufdata *bd, *bs;
|
|
struct buffer_head *bh;
|
|
u64 end = bstart + blen;
|
|
|
|
gfs2_log_lock(sdp);
|
|
spin_lock(&sdp->sd_ail_lock);
|
|
list_for_each_entry_safe(tr, s, &sdp->sd_ail1_list, tr_list) {
|
|
list_for_each_entry_safe(bd, bs, &tr->tr_ail1_list,
|
|
bd_ail_st_list) {
|
|
bh = bd->bd_bh;
|
|
if (bh->b_blocknr < bstart || bh->b_blocknr >= end)
|
|
continue;
|
|
|
|
gfs2_remove_from_journal(bh, REMOVE_JDATA);
|
|
}
|
|
}
|
|
spin_unlock(&sdp->sd_ail_lock);
|
|
gfs2_log_unlock(sdp);
|
|
}
|
|
|
|
static struct buffer_head *gfs2_getjdatabuf(struct gfs2_inode *ip, u64 blkno)
|
|
{
|
|
struct address_space *mapping = ip->i_inode.i_mapping;
|
|
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
|
|
struct page *page;
|
|
struct buffer_head *bh;
|
|
unsigned int shift = PAGE_SHIFT - sdp->sd_sb.sb_bsize_shift;
|
|
unsigned long index = blkno >> shift; /* convert block to page */
|
|
unsigned int bufnum = blkno - (index << shift);
|
|
|
|
page = find_get_page_flags(mapping, index, FGP_LOCK|FGP_ACCESSED);
|
|
if (!page)
|
|
return NULL;
|
|
if (!page_has_buffers(page)) {
|
|
unlock_page(page);
|
|
put_page(page);
|
|
return NULL;
|
|
}
|
|
/* Locate header for our buffer within our page */
|
|
for (bh = page_buffers(page); bufnum--; bh = bh->b_this_page)
|
|
/* Do nothing */;
|
|
get_bh(bh);
|
|
unlock_page(page);
|
|
put_page(page);
|
|
return bh;
|
|
}
|
|
|
|
/**
|
|
* gfs2_journal_wipe - make inode's buffers so they aren't dirty/pinned anymore
|
|
* @ip: the inode who owns the buffers
|
|
* @bstart: the first buffer in the run
|
|
* @blen: the number of buffers in the run
|
|
*
|
|
*/
|
|
|
|
void gfs2_journal_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen)
|
|
{
|
|
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
|
|
struct buffer_head *bh;
|
|
int ty;
|
|
|
|
if (!ip->i_gl) {
|
|
/* This can only happen during incomplete inode creation. */
|
|
BUG_ON(!test_bit(GIF_ALLOC_FAILED, &ip->i_flags));
|
|
return;
|
|
}
|
|
|
|
gfs2_ail1_wipe(sdp, bstart, blen);
|
|
while (blen) {
|
|
ty = REMOVE_META;
|
|
bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE);
|
|
if (!bh && gfs2_is_jdata(ip)) {
|
|
bh = gfs2_getjdatabuf(ip, bstart);
|
|
ty = REMOVE_JDATA;
|
|
}
|
|
if (bh) {
|
|
lock_buffer(bh);
|
|
gfs2_log_lock(sdp);
|
|
spin_lock(&sdp->sd_ail_lock);
|
|
gfs2_remove_from_journal(bh, ty);
|
|
spin_unlock(&sdp->sd_ail_lock);
|
|
gfs2_log_unlock(sdp);
|
|
unlock_buffer(bh);
|
|
brelse(bh);
|
|
}
|
|
|
|
bstart++;
|
|
blen--;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* gfs2_meta_buffer - Get a metadata buffer
|
|
* @ip: The GFS2 inode
|
|
* @mtype: The block type (GFS2_METATYPE_*)
|
|
* @num: The block number (device relative) of the buffer
|
|
* @bhp: the buffer is returned here
|
|
*
|
|
* Returns: errno
|
|
*/
|
|
|
|
int gfs2_meta_buffer(struct gfs2_inode *ip, u32 mtype, u64 num,
|
|
struct buffer_head **bhp)
|
|
{
|
|
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
|
|
struct gfs2_glock *gl = ip->i_gl;
|
|
struct buffer_head *bh;
|
|
int ret = 0;
|
|
int rahead = 0;
|
|
|
|
if (num == ip->i_no_addr)
|
|
rahead = ip->i_rahead;
|
|
|
|
ret = gfs2_meta_read(gl, num, DIO_WAIT, rahead, &bh);
|
|
if (ret == 0 && gfs2_metatype_check(sdp, bh, mtype)) {
|
|
brelse(bh);
|
|
ret = -EIO;
|
|
} else {
|
|
*bhp = bh;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* gfs2_meta_ra - start readahead on an extent of a file
|
|
* @gl: the glock the blocks belong to
|
|
* @dblock: the starting disk block
|
|
* @extlen: the number of blocks in the extent
|
|
*
|
|
* returns: the first buffer in the extent
|
|
*/
|
|
|
|
struct buffer_head *gfs2_meta_ra(struct gfs2_glock *gl, u64 dblock, u32 extlen)
|
|
{
|
|
struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
|
|
struct buffer_head *first_bh, *bh;
|
|
u32 max_ra = gfs2_tune_get(sdp, gt_max_readahead) >>
|
|
sdp->sd_sb.sb_bsize_shift;
|
|
|
|
BUG_ON(!extlen);
|
|
|
|
if (max_ra < 1)
|
|
max_ra = 1;
|
|
if (extlen > max_ra)
|
|
extlen = max_ra;
|
|
|
|
first_bh = gfs2_getbuf(gl, dblock, CREATE);
|
|
|
|
if (buffer_uptodate(first_bh))
|
|
goto out;
|
|
bh_read_nowait(first_bh, REQ_META | REQ_PRIO);
|
|
|
|
dblock++;
|
|
extlen--;
|
|
|
|
while (extlen) {
|
|
bh = gfs2_getbuf(gl, dblock, CREATE);
|
|
|
|
bh_readahead(bh, REQ_RAHEAD | REQ_META | REQ_PRIO);
|
|
brelse(bh);
|
|
dblock++;
|
|
extlen--;
|
|
if (!buffer_locked(first_bh) && buffer_uptodate(first_bh))
|
|
goto out;
|
|
}
|
|
|
|
wait_on_buffer(first_bh);
|
|
out:
|
|
return first_bh;
|
|
}
|
|
|