The hardware tag-based KASAN for compatibility with the other modes stores the tag associated to a page in page->flags. Due to this the kernel faults on access when it allocates a page with an initial tag and the user changes the tags. Reset the tag associated by the kernel to a page in all the meaningful places to prevent kernel faults on access. Note: An alternative to this approach could be to modify page_to_virt(). This though could end up being racy, in fact if a CPU checks the PG_mte_tagged bit and decides that the page is not tagged but another CPU maps the same with PROT_MTE and becomes tagged the subsequent kernel access would fail. Link: https://lkml.kernel.org/r/9073d4e973747a6f78d5bdd7ebe17f290d087096.1606161801.git.andreyknvl@google.com Signed-off-by: Vincenzo Frascino <vincenzo.frascino@arm.com> Signed-off-by: Andrey Konovalov <andreyknvl@google.com> Reviewed-by: Catalin Marinas <catalin.marinas@arm.com> Tested-by: Vincenzo Frascino <vincenzo.frascino@arm.com> Cc: Alexander Potapenko <glider@google.com> Cc: Andrey Ryabinin <aryabinin@virtuozzo.com> Cc: Branislav Rankov <Branislav.Rankov@arm.com> Cc: Dmitry Vyukov <dvyukov@google.com> Cc: Evgenii Stepanov <eugenis@google.com> Cc: Kevin Brodsky <kevin.brodsky@arm.com> Cc: Marco Elver <elver@google.com> Cc: Vasily Gorbik <gor@linux.ibm.com> Cc: Will Deacon <will.deacon@arm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
		
			
				
	
	
		
			93 lines
		
	
	
		
			2.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			93 lines
		
	
	
		
			2.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-only
 | |
| 
 | |
| #include <linux/pagemap.h>
 | |
| #include <linux/xarray.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/swap.h>
 | |
| #include <linux/swapops.h>
 | |
| #include <asm/mte.h>
 | |
| 
 | |
| static DEFINE_XARRAY(mte_pages);
 | |
| 
 | |
| void *mte_allocate_tag_storage(void)
 | |
| {
 | |
| 	/* tags granule is 16 bytes, 2 tags stored per byte */
 | |
| 	return kmalloc(PAGE_SIZE / 16 / 2, GFP_KERNEL);
 | |
| }
 | |
| 
 | |
| void mte_free_tag_storage(char *storage)
 | |
| {
 | |
| 	kfree(storage);
 | |
| }
 | |
| 
 | |
| int mte_save_tags(struct page *page)
 | |
| {
 | |
| 	void *tag_storage, *ret;
 | |
| 
 | |
| 	if (!test_bit(PG_mte_tagged, &page->flags))
 | |
| 		return 0;
 | |
| 
 | |
| 	tag_storage = mte_allocate_tag_storage();
 | |
| 	if (!tag_storage)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	mte_save_page_tags(page_address(page), tag_storage);
 | |
| 
 | |
| 	/* page_private contains the swap entry.val set in do_swap_page */
 | |
| 	ret = xa_store(&mte_pages, page_private(page), tag_storage, GFP_KERNEL);
 | |
| 	if (WARN(xa_is_err(ret), "Failed to store MTE tags")) {
 | |
| 		mte_free_tag_storage(tag_storage);
 | |
| 		return xa_err(ret);
 | |
| 	} else if (ret) {
 | |
| 		/* Entry is being replaced, free the old entry */
 | |
| 		mte_free_tag_storage(ret);
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| bool mte_restore_tags(swp_entry_t entry, struct page *page)
 | |
| {
 | |
| 	void *tags = xa_load(&mte_pages, entry.val);
 | |
| 
 | |
| 	if (!tags)
 | |
| 		return false;
 | |
| 
 | |
| 	page_kasan_tag_reset(page);
 | |
| 	/*
 | |
| 	 * We need smp_wmb() in between setting the flags and clearing the
 | |
| 	 * tags because if another thread reads page->flags and builds a
 | |
| 	 * tagged address out of it, there is an actual dependency to the
 | |
| 	 * memory access, but on the current thread we do not guarantee that
 | |
| 	 * the new page->flags are visible before the tags were updated.
 | |
| 	 */
 | |
| 	smp_wmb();
 | |
| 	mte_restore_page_tags(page_address(page), tags);
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| void mte_invalidate_tags(int type, pgoff_t offset)
 | |
| {
 | |
| 	swp_entry_t entry = swp_entry(type, offset);
 | |
| 	void *tags = xa_erase(&mte_pages, entry.val);
 | |
| 
 | |
| 	mte_free_tag_storage(tags);
 | |
| }
 | |
| 
 | |
| void mte_invalidate_tags_area(int type)
 | |
| {
 | |
| 	swp_entry_t entry = swp_entry(type, 0);
 | |
| 	swp_entry_t last_entry = swp_entry(type + 1, 0);
 | |
| 	void *tags;
 | |
| 
 | |
| 	XA_STATE(xa_state, &mte_pages, entry.val);
 | |
| 
 | |
| 	xa_lock(&mte_pages);
 | |
| 	xas_for_each(&xa_state, tags, last_entry.val - 1) {
 | |
| 		__xa_erase(&mte_pages, xa_state.xa_index);
 | |
| 		mte_free_tag_storage(tags);
 | |
| 	}
 | |
| 	xa_unlock(&mte_pages);
 | |
| }
 |