mirror of
https://github.com/torvalds/linux.git
synced 2024-12-27 13:22:23 +00:00
btrfs: add a shrinker for extent maps
Extent maps are used either to represent existing file extent items, or to
represent new extents that are going to be written and the respective file
extent items are created when the ordered extent completes.
We currently don't have any limit for how many extent maps we can have,
neither per inode nor globally. Most of the time this not too noticeable
because extent maps are removed in the following situations:
1) When evicting an inode;
2) When releasing folios (pages) through the btrfs_release_folio() address
space operation callback.
However we won't release extent maps in the folio range if the folio is
either dirty or under writeback or if the inode's i_size is less than
or equals to 16M (see try_release_extent_mapping(). This 16M i_size
constraint was added back in 2008 with commit 70dec8079d
("Btrfs:
extent_io and extent_state optimizations"), but there's no explanation
about why we have it or why the 16M value.
This means that for buffered IO we can reach an OOM situation due to too
many extent maps if either of the following happens:
1) There's a set of tasks constantly doing IO on many files with a size
not larger than 16M, specially if they keep the files open for very
long periods, therefore preventing inode eviction.
This requires a really high number of such files, and having many non
mergeable extent maps (due to random 4K writes for example) and a
machine with very little memory;
2) There's a set tasks constantly doing random write IO (therefore
creating many non mergeable extent maps) on files and keeping them
open for long periods of time, so inode eviction doesn't happen and
there's always a lot of dirty pages or pages under writeback,
preventing btrfs_release_folio() from releasing the respective extent
maps.
This second case was actually reported in the thread pointed by the Link
tag below, and it requires a very large file under heavy IO and a machine
with very little amount of RAM, which is probably hard to happen in
practice in a real world use case.
However when using direct IO this is not so hard to happen, because the
page cache is not used, and therefore btrfs_release_folio() is never
called. Which means extent maps are dropped only when evicting the inode,
and that means that if we have tasks that keep a file descriptor open and
keep doing IO on a very large file (or files), we can exhaust memory due
to an unbounded amount of extent maps. This is especially easy to happen
if we have a huge file with millions of small extents and their extent
maps are not mergeable (non contiguous offsets and disk locations).
This was reported in that thread with the following fio test:
$ cat test.sh
#!/bin/bash
DEV=/dev/sdj
MNT=/mnt/sdj
MOUNT_OPTIONS="-o ssd"
MKFS_OPTIONS=""
cat <<EOF > /tmp/fio-job.ini
[global]
name=fio-rand-write
filename=$MNT/fio-rand-write
rw=randwrite
bs=4K
direct=1
numjobs=16
fallocate=none
time_based
runtime=90000
[file1]
size=300G
ioengine=libaio
iodepth=16
EOF
umount $MNT &> /dev/null
mkfs.btrfs -f $MKFS_OPTIONS $DEV
mount $MOUNT_OPTIONS $DEV $MNT
fio /tmp/fio-job.ini
umount $MNT
Monitoring the btrfs_extent_map slab while running the test with:
$ watch -d -n 1 'cat /sys/kernel/slab/btrfs_extent_map/objects \
/sys/kernel/slab/btrfs_extent_map/total_objects'
Shows the number of active and total extent maps skyrocketing to tens of
millions, and on systems with a short amount of memory it's easy and quick
to get into an OOM situation, as reported in that thread.
So to avoid this issue add a shrinker that will remove extents maps, as
long as they are not pinned, and takes proper care with any concurrent
fsync to avoid missing extents (setting the full sync flag while in the
middle of a fast fsync). This shrinker is triggered through the callbacks
nr_cached_objects and free_cached_objects of struct super_operations.
The shrinker will iterate over all roots and over all inodes of each
root, and keeps track of the last scanned root and inode, so that the
next time it runs, it starts from that root and from the next inode.
This is similar to what xfs does for its inode reclaim (implements those
callbacks, and cycles through inodes by starting from where it ended
last time).
Reviewed-by: Josef Bacik <josef@toxicpanda.com>
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Reviewed-by: David Sterba <dsterba@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
This commit is contained in:
parent
f1d97e7691
commit
956a17d9d0
@ -8,6 +8,7 @@
|
||||
#include "extent_map.h"
|
||||
#include "compression.h"
|
||||
#include "btrfs_inode.h"
|
||||
#include "disk-io.h"
|
||||
|
||||
|
||||
static struct kmem_cache *extent_map_cache;
|
||||
@ -1026,3 +1027,162 @@ out_free_pre:
|
||||
free_extent_map(split_pre);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static long btrfs_scan_inode(struct btrfs_inode *inode, long *scanned, long nr_to_scan)
|
||||
{
|
||||
const u64 cur_fs_gen = btrfs_get_fs_generation(inode->root->fs_info);
|
||||
struct extent_map_tree *tree = &inode->extent_tree;
|
||||
long nr_dropped = 0;
|
||||
struct rb_node *node;
|
||||
|
||||
/*
|
||||
* Take the mmap lock so that we serialize with the inode logging phase
|
||||
* of fsync because we may need to set the full sync flag on the inode,
|
||||
* in case we have to remove extent maps in the tree's list of modified
|
||||
* extents. If we set the full sync flag in the inode while an fsync is
|
||||
* in progress, we may risk missing new extents because before the flag
|
||||
* is set, fsync decides to only wait for writeback to complete and then
|
||||
* during inode logging it sees the flag set and uses the subvolume tree
|
||||
* to find new extents, which may not be there yet because ordered
|
||||
* extents haven't completed yet.
|
||||
*
|
||||
* We also do a try lock because otherwise we could deadlock. This is
|
||||
* because the shrinker for this filesystem may be invoked while we are
|
||||
* in a path that is holding the mmap lock in write mode. For example in
|
||||
* a reflink operation while COWing an extent buffer, when allocating
|
||||
* pages for a new extent buffer and under memory pressure, the shrinker
|
||||
* may be invoked, and therefore we would deadlock by attempting to read
|
||||
* lock the mmap lock while we are holding already a write lock on it.
|
||||
*/
|
||||
if (!down_read_trylock(&inode->i_mmap_lock))
|
||||
return 0;
|
||||
|
||||
write_lock(&tree->lock);
|
||||
node = rb_first_cached(&tree->map);
|
||||
while (node) {
|
||||
struct extent_map *em;
|
||||
|
||||
em = rb_entry(node, struct extent_map, rb_node);
|
||||
node = rb_next(node);
|
||||
(*scanned)++;
|
||||
|
||||
if (em->flags & EXTENT_FLAG_PINNED)
|
||||
goto next;
|
||||
|
||||
/*
|
||||
* If the inode is in the list of modified extents (new) and its
|
||||
* generation is the same (or is greater than) the current fs
|
||||
* generation, it means it was not yet persisted so we have to
|
||||
* set the full sync flag so that the next fsync will not miss
|
||||
* it.
|
||||
*/
|
||||
if (!list_empty(&em->list) && em->generation >= cur_fs_gen)
|
||||
btrfs_set_inode_full_sync(inode);
|
||||
|
||||
remove_extent_mapping(inode, em);
|
||||
/* Drop the reference for the tree. */
|
||||
free_extent_map(em);
|
||||
nr_dropped++;
|
||||
next:
|
||||
if (*scanned >= nr_to_scan)
|
||||
break;
|
||||
|
||||
/*
|
||||
* Restart if we had to reschedule, and any extent maps that were
|
||||
* pinned before may have become unpinned after we released the
|
||||
* lock and took it again.
|
||||
*/
|
||||
if (cond_resched_rwlock_write(&tree->lock))
|
||||
node = rb_first_cached(&tree->map);
|
||||
}
|
||||
write_unlock(&tree->lock);
|
||||
up_read(&inode->i_mmap_lock);
|
||||
|
||||
return nr_dropped;
|
||||
}
|
||||
|
||||
static long btrfs_scan_root(struct btrfs_root *root, long *scanned, long nr_to_scan)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = root->fs_info;
|
||||
struct btrfs_inode *inode;
|
||||
long nr_dropped = 0;
|
||||
u64 min_ino = fs_info->extent_map_shrinker_last_ino + 1;
|
||||
|
||||
inode = btrfs_find_first_inode(root, min_ino);
|
||||
while (inode) {
|
||||
nr_dropped += btrfs_scan_inode(inode, scanned, nr_to_scan);
|
||||
|
||||
min_ino = btrfs_ino(inode) + 1;
|
||||
fs_info->extent_map_shrinker_last_ino = btrfs_ino(inode);
|
||||
iput(&inode->vfs_inode);
|
||||
|
||||
if (*scanned >= nr_to_scan)
|
||||
break;
|
||||
|
||||
cond_resched();
|
||||
inode = btrfs_find_first_inode(root, min_ino);
|
||||
}
|
||||
|
||||
if (inode) {
|
||||
/*
|
||||
* There are still inodes in this root or we happened to process
|
||||
* the last one and reached the scan limit. In either case set
|
||||
* the current root to this one, so we'll resume from the next
|
||||
* inode if there is one or we will find out this was the last
|
||||
* one and move to the next root.
|
||||
*/
|
||||
fs_info->extent_map_shrinker_last_root = btrfs_root_id(root);
|
||||
} else {
|
||||
/*
|
||||
* No more inodes in this root, set extent_map_shrinker_last_ino to 0 so
|
||||
* that when processing the next root we start from its first inode.
|
||||
*/
|
||||
fs_info->extent_map_shrinker_last_ino = 0;
|
||||
fs_info->extent_map_shrinker_last_root = btrfs_root_id(root) + 1;
|
||||
}
|
||||
|
||||
return nr_dropped;
|
||||
}
|
||||
|
||||
long btrfs_free_extent_maps(struct btrfs_fs_info *fs_info, long nr_to_scan)
|
||||
{
|
||||
const u64 start_root_id = fs_info->extent_map_shrinker_last_root;
|
||||
u64 next_root_id = start_root_id;
|
||||
bool cycled = false;
|
||||
long nr_dropped = 0;
|
||||
long scanned = 0;
|
||||
|
||||
while (scanned < nr_to_scan) {
|
||||
struct btrfs_root *root;
|
||||
unsigned long count;
|
||||
|
||||
spin_lock(&fs_info->fs_roots_radix_lock);
|
||||
count = radix_tree_gang_lookup(&fs_info->fs_roots_radix,
|
||||
(void **)&root,
|
||||
(unsigned long)next_root_id, 1);
|
||||
if (count == 0) {
|
||||
spin_unlock(&fs_info->fs_roots_radix_lock);
|
||||
if (start_root_id > 0 && !cycled) {
|
||||
next_root_id = 0;
|
||||
fs_info->extent_map_shrinker_last_root = 0;
|
||||
fs_info->extent_map_shrinker_last_ino = 0;
|
||||
cycled = true;
|
||||
continue;
|
||||
}
|
||||
break;
|
||||
}
|
||||
next_root_id = btrfs_root_id(root) + 1;
|
||||
root = btrfs_grab_root(root);
|
||||
spin_unlock(&fs_info->fs_roots_radix_lock);
|
||||
|
||||
if (!root)
|
||||
continue;
|
||||
|
||||
if (is_fstree(btrfs_root_id(root)))
|
||||
nr_dropped += btrfs_scan_root(root, &scanned, nr_to_scan);
|
||||
|
||||
btrfs_put_root(root);
|
||||
}
|
||||
|
||||
return nr_dropped;
|
||||
}
|
||||
|
@ -140,5 +140,6 @@ void btrfs_drop_extent_map_range(struct btrfs_inode *inode,
|
||||
int btrfs_replace_extent_map_range(struct btrfs_inode *inode,
|
||||
struct extent_map *new_em,
|
||||
bool modified);
|
||||
long btrfs_free_extent_maps(struct btrfs_fs_info *fs_info, long nr_to_scan);
|
||||
|
||||
#endif
|
||||
|
@ -631,6 +631,8 @@ struct btrfs_fs_info {
|
||||
s32 delalloc_batch;
|
||||
|
||||
struct percpu_counter evictable_extent_maps;
|
||||
u64 extent_map_shrinker_last_root;
|
||||
u64 extent_map_shrinker_last_ino;
|
||||
|
||||
/* Protected by 'trans_lock'. */
|
||||
struct list_head dirty_cowonly_roots;
|
||||
|
@ -2371,6 +2371,21 @@ static int btrfs_show_devname(struct seq_file *m, struct dentry *root)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static long btrfs_nr_cached_objects(struct super_block *sb, struct shrink_control *sc)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
|
||||
|
||||
return percpu_counter_sum_positive(&fs_info->evictable_extent_maps);
|
||||
}
|
||||
|
||||
static long btrfs_free_cached_objects(struct super_block *sb, struct shrink_control *sc)
|
||||
{
|
||||
const long nr_to_scan = min_t(unsigned long, LONG_MAX, sc->nr_to_scan);
|
||||
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
|
||||
|
||||
return btrfs_free_extent_maps(fs_info, nr_to_scan);
|
||||
}
|
||||
|
||||
static const struct super_operations btrfs_super_ops = {
|
||||
.drop_inode = btrfs_drop_inode,
|
||||
.evict_inode = btrfs_evict_inode,
|
||||
@ -2384,6 +2399,8 @@ static const struct super_operations btrfs_super_ops = {
|
||||
.statfs = btrfs_statfs,
|
||||
.freeze_fs = btrfs_freeze,
|
||||
.unfreeze_fs = btrfs_unfreeze,
|
||||
.nr_cached_objects = btrfs_nr_cached_objects,
|
||||
.free_cached_objects = btrfs_free_cached_objects,
|
||||
};
|
||||
|
||||
static const struct file_operations btrfs_ctl_fops = {
|
||||
|
Loading…
Reference in New Issue
Block a user