radix-tree: delete radix_tree_range_tag_if_tagged()
This is an exceptionally complicated function with just one caller (tag_pages_for_writeback). We devote a large portion of the runtime of the test suite to testing this one function which has one caller. By introducing the new function radix_tree_iter_tag_set(), we can eliminate all of the complexity while keeping the performance. The caller can now use a fairly standard radix_tree_for_each() loop, and it doesn't need to worry about tricksy things like 'start' wrapping. The test suite continues to spend a large amount of time investigating this function, but now it's testing the underlying primitives such as radix_tree_iter_resume() and the radix_tree_for_each_tagged() iterator which are also used by other parts of the kernel. Link: http://lkml.kernel.org/r/1480369871-5271-57-git-send-email-mawilcox@linuxonhyperv.com Signed-off-by: Matthew Wilcox <willy@infradead.org> Tested-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Konstantin Khlebnikov <koct9i@gmail.com> Cc: Ross Zwisler <ross.zwisler@linux.intel.com> Cc: Matthew Wilcox <mawilcox@microsoft.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
478922e2b0
commit
268f42de71
@ -120,6 +120,41 @@ static inline bool radix_tree_empty(struct radix_tree_root *root)
|
|||||||
return root->rnode == NULL;
|
return root->rnode == NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* struct radix_tree_iter - radix tree iterator state
|
||||||
|
*
|
||||||
|
* @index: index of current slot
|
||||||
|
* @next_index: one beyond the last index for this chunk
|
||||||
|
* @tags: bit-mask for tag-iterating
|
||||||
|
* @node: node that contains current slot
|
||||||
|
* @shift: shift for the node that holds our slots
|
||||||
|
*
|
||||||
|
* This radix tree iterator works in terms of "chunks" of slots. A chunk is a
|
||||||
|
* subinterval of slots contained within one radix tree leaf node. It is
|
||||||
|
* described by a pointer to its first slot and a struct radix_tree_iter
|
||||||
|
* which holds the chunk's position in the tree and its size. For tagged
|
||||||
|
* iteration radix_tree_iter also holds the slots' bit-mask for one chosen
|
||||||
|
* radix tree tag.
|
||||||
|
*/
|
||||||
|
struct radix_tree_iter {
|
||||||
|
unsigned long index;
|
||||||
|
unsigned long next_index;
|
||||||
|
unsigned long tags;
|
||||||
|
struct radix_tree_node *node;
|
||||||
|
#ifdef CONFIG_RADIX_TREE_MULTIORDER
|
||||||
|
unsigned int shift;
|
||||||
|
#endif
|
||||||
|
};
|
||||||
|
|
||||||
|
static inline unsigned int iter_shift(const struct radix_tree_iter *iter)
|
||||||
|
{
|
||||||
|
#ifdef CONFIG_RADIX_TREE_MULTIORDER
|
||||||
|
return iter->shift;
|
||||||
|
#else
|
||||||
|
return 0;
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Radix-tree synchronization
|
* Radix-tree synchronization
|
||||||
*
|
*
|
||||||
@ -283,6 +318,8 @@ void *radix_tree_tag_clear(struct radix_tree_root *root,
|
|||||||
unsigned long index, unsigned int tag);
|
unsigned long index, unsigned int tag);
|
||||||
int radix_tree_tag_get(struct radix_tree_root *root,
|
int radix_tree_tag_get(struct radix_tree_root *root,
|
||||||
unsigned long index, unsigned int tag);
|
unsigned long index, unsigned int tag);
|
||||||
|
void radix_tree_iter_tag_set(struct radix_tree_root *root,
|
||||||
|
const struct radix_tree_iter *iter, unsigned int tag);
|
||||||
unsigned int
|
unsigned int
|
||||||
radix_tree_gang_lookup_tag(struct radix_tree_root *root, void **results,
|
radix_tree_gang_lookup_tag(struct radix_tree_root *root, void **results,
|
||||||
unsigned long first_index, unsigned int max_items,
|
unsigned long first_index, unsigned int max_items,
|
||||||
@ -291,10 +328,6 @@ unsigned int
|
|||||||
radix_tree_gang_lookup_tag_slot(struct radix_tree_root *root, void ***results,
|
radix_tree_gang_lookup_tag_slot(struct radix_tree_root *root, void ***results,
|
||||||
unsigned long first_index, unsigned int max_items,
|
unsigned long first_index, unsigned int max_items,
|
||||||
unsigned int tag);
|
unsigned int tag);
|
||||||
unsigned long radix_tree_range_tag_if_tagged(struct radix_tree_root *root,
|
|
||||||
unsigned long *first_indexp, unsigned long last_index,
|
|
||||||
unsigned long nr_to_tag,
|
|
||||||
unsigned int fromtag, unsigned int totag);
|
|
||||||
int radix_tree_tagged(struct radix_tree_root *root, unsigned int tag);
|
int radix_tree_tagged(struct radix_tree_root *root, unsigned int tag);
|
||||||
|
|
||||||
static inline void radix_tree_preload_end(void)
|
static inline void radix_tree_preload_end(void)
|
||||||
@ -302,39 +335,6 @@ static inline void radix_tree_preload_end(void)
|
|||||||
preempt_enable();
|
preempt_enable();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* struct radix_tree_iter - radix tree iterator state
|
|
||||||
*
|
|
||||||
* @index: index of current slot
|
|
||||||
* @next_index: one beyond the last index for this chunk
|
|
||||||
* @tags: bit-mask for tag-iterating
|
|
||||||
* @shift: shift for the node that holds our slots
|
|
||||||
*
|
|
||||||
* This radix tree iterator works in terms of "chunks" of slots. A chunk is a
|
|
||||||
* subinterval of slots contained within one radix tree leaf node. It is
|
|
||||||
* described by a pointer to its first slot and a struct radix_tree_iter
|
|
||||||
* which holds the chunk's position in the tree and its size. For tagged
|
|
||||||
* iteration radix_tree_iter also holds the slots' bit-mask for one chosen
|
|
||||||
* radix tree tag.
|
|
||||||
*/
|
|
||||||
struct radix_tree_iter {
|
|
||||||
unsigned long index;
|
|
||||||
unsigned long next_index;
|
|
||||||
unsigned long tags;
|
|
||||||
#ifdef CONFIG_RADIX_TREE_MULTIORDER
|
|
||||||
unsigned int shift;
|
|
||||||
#endif
|
|
||||||
};
|
|
||||||
|
|
||||||
static inline unsigned int iter_shift(struct radix_tree_iter *iter)
|
|
||||||
{
|
|
||||||
#ifdef CONFIG_RADIX_TREE_MULTIORDER
|
|
||||||
return iter->shift;
|
|
||||||
#else
|
|
||||||
return 0;
|
|
||||||
#endif
|
|
||||||
}
|
|
||||||
|
|
||||||
#define RADIX_TREE_ITER_TAG_MASK 0x00FF /* tag index in lower byte */
|
#define RADIX_TREE_ITER_TAG_MASK 0x00FF /* tag index in lower byte */
|
||||||
#define RADIX_TREE_ITER_TAGGED 0x0100 /* lookup tagged slots */
|
#define RADIX_TREE_ITER_TAGGED 0x0100 /* lookup tagged slots */
|
||||||
#define RADIX_TREE_ITER_CONTIG 0x0200 /* stop at first hole */
|
#define RADIX_TREE_ITER_CONTIG 0x0200 /* stop at first hole */
|
||||||
|
117
lib/radix-tree.c
117
lib/radix-tree.c
@ -219,6 +219,11 @@ radix_tree_find_next_bit(struct radix_tree_node *node, unsigned int tag,
|
|||||||
return RADIX_TREE_MAP_SIZE;
|
return RADIX_TREE_MAP_SIZE;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static unsigned int iter_offset(const struct radix_tree_iter *iter)
|
||||||
|
{
|
||||||
|
return (iter->index >> iter_shift(iter)) & RADIX_TREE_MAP_MASK;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The maximum index which can be stored in a radix tree
|
* The maximum index which can be stored in a radix tree
|
||||||
*/
|
*/
|
||||||
@ -1014,6 +1019,18 @@ static void node_tag_set(struct radix_tree_root *root,
|
|||||||
root_tag_set(root, tag);
|
root_tag_set(root, tag);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* radix_tree_iter_tag_set - set a tag on the current iterator entry
|
||||||
|
* @root: radix tree root
|
||||||
|
* @iter: iterator state
|
||||||
|
* @tag: tag to set
|
||||||
|
*/
|
||||||
|
void radix_tree_iter_tag_set(struct radix_tree_root *root,
|
||||||
|
const struct radix_tree_iter *iter, unsigned int tag)
|
||||||
|
{
|
||||||
|
node_tag_set(root, iter->node, tag, iter_offset(iter));
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* radix_tree_tag_clear - clear a tag on a radix tree node
|
* radix_tree_tag_clear - clear a tag on a radix tree node
|
||||||
* @root: radix tree root
|
* @root: radix tree root
|
||||||
@ -1164,6 +1181,7 @@ void ** __radix_tree_next_slot(void **slot, struct radix_tree_iter *iter,
|
|||||||
if (node == RADIX_TREE_RETRY)
|
if (node == RADIX_TREE_RETRY)
|
||||||
return slot;
|
return slot;
|
||||||
node = entry_to_node(node);
|
node = entry_to_node(node);
|
||||||
|
iter->node = node;
|
||||||
iter->shift = node->shift;
|
iter->shift = node->shift;
|
||||||
|
|
||||||
if (flags & RADIX_TREE_ITER_TAGGED) {
|
if (flags & RADIX_TREE_ITER_TAGGED) {
|
||||||
@ -1266,6 +1284,7 @@ void **radix_tree_next_chunk(struct radix_tree_root *root,
|
|||||||
iter->index = index;
|
iter->index = index;
|
||||||
iter->next_index = maxindex + 1;
|
iter->next_index = maxindex + 1;
|
||||||
iter->tags = 1;
|
iter->tags = 1;
|
||||||
|
iter->node = NULL;
|
||||||
__set_iter_shift(iter, 0);
|
__set_iter_shift(iter, 0);
|
||||||
return (void **)&root->rnode;
|
return (void **)&root->rnode;
|
||||||
}
|
}
|
||||||
@ -1308,6 +1327,7 @@ void **radix_tree_next_chunk(struct radix_tree_root *root,
|
|||||||
/* Update the iterator state */
|
/* Update the iterator state */
|
||||||
iter->index = (index &~ node_maxindex(node)) | (offset << node->shift);
|
iter->index = (index &~ node_maxindex(node)) | (offset << node->shift);
|
||||||
iter->next_index = (index | node_maxindex(node)) + 1;
|
iter->next_index = (index | node_maxindex(node)) + 1;
|
||||||
|
iter->node = node;
|
||||||
__set_iter_shift(iter, node->shift);
|
__set_iter_shift(iter, node->shift);
|
||||||
|
|
||||||
if (flags & RADIX_TREE_ITER_TAGGED)
|
if (flags & RADIX_TREE_ITER_TAGGED)
|
||||||
@ -1317,103 +1337,6 @@ void **radix_tree_next_chunk(struct radix_tree_root *root,
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL(radix_tree_next_chunk);
|
EXPORT_SYMBOL(radix_tree_next_chunk);
|
||||||
|
|
||||||
/**
|
|
||||||
* radix_tree_range_tag_if_tagged - for each item in given range set given
|
|
||||||
* tag if item has another tag set
|
|
||||||
* @root: radix tree root
|
|
||||||
* @first_indexp: pointer to a starting index of a range to scan
|
|
||||||
* @last_index: last index of a range to scan
|
|
||||||
* @nr_to_tag: maximum number items to tag
|
|
||||||
* @iftag: tag index to test
|
|
||||||
* @settag: tag index to set if tested tag is set
|
|
||||||
*
|
|
||||||
* This function scans range of radix tree from first_index to last_index
|
|
||||||
* (inclusive). For each item in the range if iftag is set, the function sets
|
|
||||||
* also settag. The function stops either after tagging nr_to_tag items or
|
|
||||||
* after reaching last_index.
|
|
||||||
*
|
|
||||||
* The tags must be set from the leaf level only and propagated back up the
|
|
||||||
* path to the root. We must do this so that we resolve the full path before
|
|
||||||
* setting any tags on intermediate nodes. If we set tags as we descend, then
|
|
||||||
* we can get to the leaf node and find that the index that has the iftag
|
|
||||||
* set is outside the range we are scanning. This reults in dangling tags and
|
|
||||||
* can lead to problems with later tag operations (e.g. livelocks on lookups).
|
|
||||||
*
|
|
||||||
* The function returns the number of leaves where the tag was set and sets
|
|
||||||
* *first_indexp to the first unscanned index.
|
|
||||||
* WARNING! *first_indexp can wrap if last_index is ULONG_MAX. Caller must
|
|
||||||
* be prepared to handle that.
|
|
||||||
*/
|
|
||||||
unsigned long radix_tree_range_tag_if_tagged(struct radix_tree_root *root,
|
|
||||||
unsigned long *first_indexp, unsigned long last_index,
|
|
||||||
unsigned long nr_to_tag,
|
|
||||||
unsigned int iftag, unsigned int settag)
|
|
||||||
{
|
|
||||||
struct radix_tree_node *node, *child;
|
|
||||||
unsigned long maxindex;
|
|
||||||
unsigned long tagged = 0;
|
|
||||||
unsigned long index = *first_indexp;
|
|
||||||
|
|
||||||
radix_tree_load_root(root, &child, &maxindex);
|
|
||||||
last_index = min(last_index, maxindex);
|
|
||||||
if (index > last_index)
|
|
||||||
return 0;
|
|
||||||
if (!nr_to_tag)
|
|
||||||
return 0;
|
|
||||||
if (!root_tag_get(root, iftag)) {
|
|
||||||
*first_indexp = last_index + 1;
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
if (!radix_tree_is_internal_node(child)) {
|
|
||||||
*first_indexp = last_index + 1;
|
|
||||||
root_tag_set(root, settag);
|
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
node = entry_to_node(child);
|
|
||||||
|
|
||||||
for (;;) {
|
|
||||||
unsigned offset = radix_tree_descend(node, &child, index);
|
|
||||||
if (!child)
|
|
||||||
goto next;
|
|
||||||
if (!tag_get(node, iftag, offset))
|
|
||||||
goto next;
|
|
||||||
/* Sibling slots never have tags set on them */
|
|
||||||
if (radix_tree_is_internal_node(child)) {
|
|
||||||
node = entry_to_node(child);
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
tagged++;
|
|
||||||
node_tag_set(root, node, settag, offset);
|
|
||||||
next:
|
|
||||||
/* Go to next entry in node */
|
|
||||||
index = ((index >> node->shift) + 1) << node->shift;
|
|
||||||
/* Overflow can happen when last_index is ~0UL... */
|
|
||||||
if (index > last_index || !index)
|
|
||||||
break;
|
|
||||||
offset = (index >> node->shift) & RADIX_TREE_MAP_MASK;
|
|
||||||
while (offset == 0) {
|
|
||||||
/*
|
|
||||||
* We've fully scanned this node. Go up. Because
|
|
||||||
* last_index is guaranteed to be in the tree, what
|
|
||||||
* we do below cannot wander astray.
|
|
||||||
*/
|
|
||||||
node = node->parent;
|
|
||||||
offset = (index >> node->shift) & RADIX_TREE_MAP_MASK;
|
|
||||||
}
|
|
||||||
if (is_sibling_entry(node, node->slots[offset]))
|
|
||||||
goto next;
|
|
||||||
if (tagged >= nr_to_tag)
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
*first_indexp = index;
|
|
||||||
|
|
||||||
return tagged;
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(radix_tree_range_tag_if_tagged);
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* radix_tree_gang_lookup - perform multiple lookup on a radix tree
|
* radix_tree_gang_lookup - perform multiple lookup on a radix tree
|
||||||
* @root: radix tree root
|
* @root: radix tree root
|
||||||
|
@ -2106,18 +2106,26 @@ void tag_pages_for_writeback(struct address_space *mapping,
|
|||||||
pgoff_t start, pgoff_t end)
|
pgoff_t start, pgoff_t end)
|
||||||
{
|
{
|
||||||
#define WRITEBACK_TAG_BATCH 4096
|
#define WRITEBACK_TAG_BATCH 4096
|
||||||
unsigned long tagged;
|
unsigned long tagged = 0;
|
||||||
|
struct radix_tree_iter iter;
|
||||||
|
void **slot;
|
||||||
|
|
||||||
do {
|
spin_lock_irq(&mapping->tree_lock);
|
||||||
spin_lock_irq(&mapping->tree_lock);
|
radix_tree_for_each_tagged(slot, &mapping->page_tree, &iter, start,
|
||||||
tagged = radix_tree_range_tag_if_tagged(&mapping->page_tree,
|
PAGECACHE_TAG_DIRTY) {
|
||||||
&start, end, WRITEBACK_TAG_BATCH,
|
if (iter.index > end)
|
||||||
PAGECACHE_TAG_DIRTY, PAGECACHE_TAG_TOWRITE);
|
break;
|
||||||
|
radix_tree_iter_tag_set(&mapping->page_tree, &iter,
|
||||||
|
PAGECACHE_TAG_TOWRITE);
|
||||||
|
tagged++;
|
||||||
|
if ((tagged % WRITEBACK_TAG_BATCH) != 0)
|
||||||
|
continue;
|
||||||
|
slot = radix_tree_iter_resume(slot, &iter);
|
||||||
spin_unlock_irq(&mapping->tree_lock);
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
WARN_ON_ONCE(tagged > WRITEBACK_TAG_BATCH);
|
|
||||||
cond_resched();
|
cond_resched();
|
||||||
/* We check 'start' to handle wrapping when end == ~0UL */
|
spin_lock_irq(&mapping->tree_lock);
|
||||||
} while (tagged >= WRITEBACK_TAG_BATCH && start);
|
}
|
||||||
|
spin_unlock_irq(&mapping->tree_lock);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(tag_pages_for_writeback);
|
EXPORT_SYMBOL(tag_pages_for_writeback);
|
||||||
|
|
||||||
|
@ -205,8 +205,7 @@ void copy_tag_check(void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
// printf("\ncopying tags...\n");
|
// printf("\ncopying tags...\n");
|
||||||
cur = start;
|
tagged = tag_tagged_items(&tree, NULL, start, end, ITEMS, 0, 1);
|
||||||
tagged = radix_tree_range_tag_if_tagged(&tree, &cur, end, ITEMS, 0, 1);
|
|
||||||
|
|
||||||
// printf("checking copied tags\n");
|
// printf("checking copied tags\n");
|
||||||
assert(tagged == count);
|
assert(tagged == count);
|
||||||
@ -214,16 +213,13 @@ void copy_tag_check(void)
|
|||||||
|
|
||||||
/* Copy tags in several rounds */
|
/* Copy tags in several rounds */
|
||||||
// printf("\ncopying tags...\n");
|
// printf("\ncopying tags...\n");
|
||||||
cur = start;
|
tmp = rand() % (count / 10 + 2);
|
||||||
do {
|
tagged = tag_tagged_items(&tree, NULL, start, end, tmp, 0, 2);
|
||||||
tmp = rand() % (count/10+2);
|
assert(tagged == count);
|
||||||
tagged = radix_tree_range_tag_if_tagged(&tree, &cur, end, tmp, 0, 2);
|
|
||||||
} while (tmp == tagged);
|
|
||||||
|
|
||||||
// printf("%lu %lu %lu\n", tagged, tmp, count);
|
// printf("%lu %lu %lu\n", tagged, tmp, count);
|
||||||
// printf("checking copied tags\n");
|
// printf("checking copied tags\n");
|
||||||
check_copied_tags(&tree, start, end, idx, ITEMS, 0, 2);
|
check_copied_tags(&tree, start, end, idx, ITEMS, 0, 2);
|
||||||
assert(tagged < tmp);
|
|
||||||
verify_tag_consistency(&tree, 0);
|
verify_tag_consistency(&tree, 0);
|
||||||
verify_tag_consistency(&tree, 1);
|
verify_tag_consistency(&tree, 1);
|
||||||
verify_tag_consistency(&tree, 2);
|
verify_tag_consistency(&tree, 2);
|
||||||
|
@ -26,7 +26,6 @@ static void __multiorder_tag_test(int index, int order)
|
|||||||
{
|
{
|
||||||
RADIX_TREE(tree, GFP_KERNEL);
|
RADIX_TREE(tree, GFP_KERNEL);
|
||||||
int base, err, i;
|
int base, err, i;
|
||||||
unsigned long first = 0;
|
|
||||||
|
|
||||||
/* our canonical entry */
|
/* our canonical entry */
|
||||||
base = index & ~((1 << order) - 1);
|
base = index & ~((1 << order) - 1);
|
||||||
@ -60,7 +59,7 @@ static void __multiorder_tag_test(int index, int order)
|
|||||||
assert(!radix_tree_tag_get(&tree, i, 1));
|
assert(!radix_tree_tag_get(&tree, i, 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
assert(radix_tree_range_tag_if_tagged(&tree, &first, ~0UL, 10, 0, 1) == 1);
|
assert(tag_tagged_items(&tree, NULL, 0, ~0UL, 10, 0, 1) == 1);
|
||||||
assert(radix_tree_tag_clear(&tree, index, 0));
|
assert(radix_tree_tag_clear(&tree, index, 0));
|
||||||
|
|
||||||
for_each_index(i, base, order) {
|
for_each_index(i, base, order) {
|
||||||
@ -251,7 +250,6 @@ void multiorder_tagged_iteration(void)
|
|||||||
RADIX_TREE(tree, GFP_KERNEL);
|
RADIX_TREE(tree, GFP_KERNEL);
|
||||||
struct radix_tree_iter iter;
|
struct radix_tree_iter iter;
|
||||||
void **slot;
|
void **slot;
|
||||||
unsigned long first = 0;
|
|
||||||
int i, j;
|
int i, j;
|
||||||
|
|
||||||
printf("Multiorder tagged iteration test\n");
|
printf("Multiorder tagged iteration test\n");
|
||||||
@ -296,8 +294,8 @@ void multiorder_tagged_iteration(void)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
radix_tree_range_tag_if_tagged(&tree, &first, ~0UL,
|
assert(tag_tagged_items(&tree, NULL, 0, ~0UL, TAG_ENTRIES, 1, 2) ==
|
||||||
MT_NUM_ENTRIES, 1, 2);
|
TAG_ENTRIES);
|
||||||
|
|
||||||
for (j = 0; j < 256; j++) {
|
for (j = 0; j < 256; j++) {
|
||||||
int mask, k;
|
int mask, k;
|
||||||
@ -323,9 +321,8 @@ void multiorder_tagged_iteration(void)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
first = 1;
|
assert(tag_tagged_items(&tree, NULL, 1, ~0UL, MT_NUM_ENTRIES * 2, 1, 0)
|
||||||
radix_tree_range_tag_if_tagged(&tree, &first, ~0UL,
|
== TAG_ENTRIES);
|
||||||
MT_NUM_ENTRIES, 1, 0);
|
|
||||||
i = 0;
|
i = 0;
|
||||||
radix_tree_for_each_tagged(slot, &tree, &iter, 0, 0) {
|
radix_tree_for_each_tagged(slot, &tree, &iter, 0, 0) {
|
||||||
assert(iter.index == tag_index[i]);
|
assert(iter.index == tag_index[i]);
|
||||||
|
@ -50,6 +50,7 @@
|
|||||||
#include <stdio.h>
|
#include <stdio.h>
|
||||||
|
|
||||||
#include "regression.h"
|
#include "regression.h"
|
||||||
|
#include "test.h"
|
||||||
|
|
||||||
#define PAGECACHE_TAG_DIRTY 0
|
#define PAGECACHE_TAG_DIRTY 0
|
||||||
#define PAGECACHE_TAG_WRITEBACK 1
|
#define PAGECACHE_TAG_WRITEBACK 1
|
||||||
@ -90,7 +91,7 @@ void regression2_test(void)
|
|||||||
/* 1. */
|
/* 1. */
|
||||||
start = 0;
|
start = 0;
|
||||||
end = max_slots - 2;
|
end = max_slots - 2;
|
||||||
radix_tree_range_tag_if_tagged(&mt_tree, &start, end, 1,
|
tag_tagged_items(&mt_tree, NULL, start, end, 1,
|
||||||
PAGECACHE_TAG_DIRTY, PAGECACHE_TAG_TOWRITE);
|
PAGECACHE_TAG_DIRTY, PAGECACHE_TAG_TOWRITE);
|
||||||
|
|
||||||
/* 2. */
|
/* 2. */
|
||||||
|
@ -23,7 +23,7 @@ __simple_checks(struct radix_tree_root *tree, unsigned long index, int tag)
|
|||||||
item_tag_set(tree, index, tag);
|
item_tag_set(tree, index, tag);
|
||||||
ret = item_tag_get(tree, index, tag);
|
ret = item_tag_get(tree, index, tag);
|
||||||
assert(ret != 0);
|
assert(ret != 0);
|
||||||
ret = radix_tree_range_tag_if_tagged(tree, &first, ~0UL, 10, tag, !tag);
|
ret = tag_tagged_items(tree, NULL, first, ~0UL, 10, tag, !tag);
|
||||||
assert(ret == 1);
|
assert(ret == 1);
|
||||||
ret = item_tag_get(tree, index, !tag);
|
ret = item_tag_get(tree, index, !tag);
|
||||||
assert(ret != 0);
|
assert(ret != 0);
|
||||||
@ -320,7 +320,7 @@ static void single_check(void)
|
|||||||
assert(ret == 0);
|
assert(ret == 0);
|
||||||
verify_tag_consistency(&tree, 0);
|
verify_tag_consistency(&tree, 0);
|
||||||
verify_tag_consistency(&tree, 1);
|
verify_tag_consistency(&tree, 1);
|
||||||
ret = radix_tree_range_tag_if_tagged(&tree, &first, 10, 10, 0, 1);
|
ret = tag_tagged_items(&tree, NULL, first, 10, 10, 0, 1);
|
||||||
assert(ret == 1);
|
assert(ret == 1);
|
||||||
ret = radix_tree_gang_lookup_tag(&tree, (void **)items, 0, BATCH, 1);
|
ret = radix_tree_gang_lookup_tag(&tree, (void **)items, 0, BATCH, 1);
|
||||||
assert(ret == 1);
|
assert(ret == 1);
|
||||||
|
@ -151,6 +151,40 @@ void item_full_scan(struct radix_tree_root *root, unsigned long start,
|
|||||||
assert(nfound == 0);
|
assert(nfound == 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Use the same pattern as tag_pages_for_writeback() in mm/page-writeback.c */
|
||||||
|
int tag_tagged_items(struct radix_tree_root *root, pthread_mutex_t *lock,
|
||||||
|
unsigned long start, unsigned long end, unsigned batch,
|
||||||
|
unsigned iftag, unsigned thentag)
|
||||||
|
{
|
||||||
|
unsigned long tagged = 0;
|
||||||
|
struct radix_tree_iter iter;
|
||||||
|
void **slot;
|
||||||
|
|
||||||
|
if (batch == 0)
|
||||||
|
batch = 1;
|
||||||
|
|
||||||
|
if (lock)
|
||||||
|
pthread_mutex_lock(lock);
|
||||||
|
radix_tree_for_each_tagged(slot, root, &iter, start, iftag) {
|
||||||
|
if (iter.index > end)
|
||||||
|
break;
|
||||||
|
radix_tree_iter_tag_set(root, &iter, thentag);
|
||||||
|
tagged++;
|
||||||
|
if ((tagged % batch) != 0)
|
||||||
|
continue;
|
||||||
|
slot = radix_tree_iter_resume(slot, &iter);
|
||||||
|
if (lock) {
|
||||||
|
pthread_mutex_unlock(lock);
|
||||||
|
rcu_barrier();
|
||||||
|
pthread_mutex_lock(lock);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (lock)
|
||||||
|
pthread_mutex_unlock(lock);
|
||||||
|
|
||||||
|
return tagged;
|
||||||
|
}
|
||||||
|
|
||||||
/* Use the same pattern as find_swap_entry() in mm/shmem.c */
|
/* Use the same pattern as find_swap_entry() in mm/shmem.c */
|
||||||
unsigned long find_item(struct radix_tree_root *root, void *item)
|
unsigned long find_item(struct radix_tree_root *root, void *item)
|
||||||
{
|
{
|
||||||
|
@ -25,6 +25,9 @@ void item_full_scan(struct radix_tree_root *root, unsigned long start,
|
|||||||
unsigned long nr, int chunk);
|
unsigned long nr, int chunk);
|
||||||
void item_kill_tree(struct radix_tree_root *root);
|
void item_kill_tree(struct radix_tree_root *root);
|
||||||
|
|
||||||
|
int tag_tagged_items(struct radix_tree_root *, pthread_mutex_t *,
|
||||||
|
unsigned long start, unsigned long end, unsigned batch,
|
||||||
|
unsigned iftag, unsigned thentag);
|
||||||
unsigned long find_item(struct radix_tree_root *, void *item);
|
unsigned long find_item(struct radix_tree_root *, void *item);
|
||||||
|
|
||||||
void tag_check(void);
|
void tag_check(void);
|
||||||
|
Loading…
Reference in New Issue
Block a user