iommu/amd: use new invalidate_range mmu-notifier
Make use of the new invalidate_range mmu_notifier call-back and remove the old logic of assigning an empty page-table between invalidate_range_start and invalidate_range_end. Signed-off-by: Joerg Roedel <jroedel@suse.de> Tested-by: Oded Gabbay <oded.gabbay@amd.com> Reviewed-by: Andrea Arcangeli <aarcange@redhat.com> Reviewed-by: Jérôme Glisse <jglisse@redhat.com> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Rik van Riel <riel@redhat.com> Cc: Hugh Dickins <hughd@google.com> Cc: Mel Gorman <mgorman@suse.de> Cc: Johannes Weiner <jweiner@redhat.com> Cc: Jay Cornwall <Jay.Cornwall@amd.com> Cc: Suravee Suthikulpanit <Suravee.Suthikulpanit@amd.com> Cc: Jesse Barnes <jbarnes@virtuousgeek.org> Cc: David Woodhouse <dwmw2@infradead.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Oded Gabbay <oded.gabbay@amd.com>
This commit is contained in:
parent
0f0a327fa1
commit
e7cc3dd48c
@ -92,13 +92,6 @@ static spinlock_t state_lock;
|
|||||||
|
|
||||||
static struct workqueue_struct *iommu_wq;
|
static struct workqueue_struct *iommu_wq;
|
||||||
|
|
||||||
/*
|
|
||||||
* Empty page table - Used between
|
|
||||||
* mmu_notifier_invalidate_range_start and
|
|
||||||
* mmu_notifier_invalidate_range_end
|
|
||||||
*/
|
|
||||||
static u64 *empty_page_table;
|
|
||||||
|
|
||||||
static void free_pasid_states(struct device_state *dev_state);
|
static void free_pasid_states(struct device_state *dev_state);
|
||||||
|
|
||||||
static u16 device_id(struct pci_dev *pdev)
|
static u16 device_id(struct pci_dev *pdev)
|
||||||
@ -418,46 +411,21 @@ static void mn_invalidate_page(struct mmu_notifier *mn,
|
|||||||
__mn_flush_page(mn, address);
|
__mn_flush_page(mn, address);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void mn_invalidate_range_start(struct mmu_notifier *mn,
|
static void mn_invalidate_range(struct mmu_notifier *mn,
|
||||||
struct mm_struct *mm,
|
struct mm_struct *mm,
|
||||||
unsigned long start, unsigned long end)
|
unsigned long start, unsigned long end)
|
||||||
{
|
{
|
||||||
struct pasid_state *pasid_state;
|
struct pasid_state *pasid_state;
|
||||||
struct device_state *dev_state;
|
struct device_state *dev_state;
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
pasid_state = mn_to_state(mn);
|
pasid_state = mn_to_state(mn);
|
||||||
dev_state = pasid_state->device_state;
|
dev_state = pasid_state->device_state;
|
||||||
|
|
||||||
spin_lock_irqsave(&pasid_state->lock, flags);
|
if ((start ^ (end - 1)) < PAGE_SIZE)
|
||||||
if (pasid_state->mmu_notifier_count == 0) {
|
amd_iommu_flush_page(dev_state->domain, pasid_state->pasid,
|
||||||
amd_iommu_domain_set_gcr3(dev_state->domain,
|
start);
|
||||||
pasid_state->pasid,
|
else
|
||||||
__pa(empty_page_table));
|
amd_iommu_flush_tlb(dev_state->domain, pasid_state->pasid);
|
||||||
}
|
|
||||||
pasid_state->mmu_notifier_count += 1;
|
|
||||||
spin_unlock_irqrestore(&pasid_state->lock, flags);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void mn_invalidate_range_end(struct mmu_notifier *mn,
|
|
||||||
struct mm_struct *mm,
|
|
||||||
unsigned long start, unsigned long end)
|
|
||||||
{
|
|
||||||
struct pasid_state *pasid_state;
|
|
||||||
struct device_state *dev_state;
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
pasid_state = mn_to_state(mn);
|
|
||||||
dev_state = pasid_state->device_state;
|
|
||||||
|
|
||||||
spin_lock_irqsave(&pasid_state->lock, flags);
|
|
||||||
pasid_state->mmu_notifier_count -= 1;
|
|
||||||
if (pasid_state->mmu_notifier_count == 0) {
|
|
||||||
amd_iommu_domain_set_gcr3(dev_state->domain,
|
|
||||||
pasid_state->pasid,
|
|
||||||
__pa(pasid_state->mm->pgd));
|
|
||||||
}
|
|
||||||
spin_unlock_irqrestore(&pasid_state->lock, flags);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void mn_release(struct mmu_notifier *mn, struct mm_struct *mm)
|
static void mn_release(struct mmu_notifier *mn, struct mm_struct *mm)
|
||||||
@ -482,8 +450,7 @@ static struct mmu_notifier_ops iommu_mn = {
|
|||||||
.release = mn_release,
|
.release = mn_release,
|
||||||
.clear_flush_young = mn_clear_flush_young,
|
.clear_flush_young = mn_clear_flush_young,
|
||||||
.invalidate_page = mn_invalidate_page,
|
.invalidate_page = mn_invalidate_page,
|
||||||
.invalidate_range_start = mn_invalidate_range_start,
|
.invalidate_range = mn_invalidate_range,
|
||||||
.invalidate_range_end = mn_invalidate_range_end,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static void set_pri_tag_status(struct pasid_state *pasid_state,
|
static void set_pri_tag_status(struct pasid_state *pasid_state,
|
||||||
@ -954,18 +921,10 @@ static int __init amd_iommu_v2_init(void)
|
|||||||
if (iommu_wq == NULL)
|
if (iommu_wq == NULL)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
ret = -ENOMEM;
|
|
||||||
empty_page_table = (u64 *)get_zeroed_page(GFP_KERNEL);
|
|
||||||
if (empty_page_table == NULL)
|
|
||||||
goto out_destroy_wq;
|
|
||||||
|
|
||||||
amd_iommu_register_ppr_notifier(&ppr_nb);
|
amd_iommu_register_ppr_notifier(&ppr_nb);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
out_destroy_wq:
|
|
||||||
destroy_workqueue(iommu_wq);
|
|
||||||
|
|
||||||
out:
|
out:
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
@ -999,8 +958,6 @@ static void __exit amd_iommu_v2_exit(void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
destroy_workqueue(iommu_wq);
|
destroy_workqueue(iommu_wq);
|
||||||
|
|
||||||
free_page((unsigned long)empty_page_table);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
module_init(amd_iommu_v2_init);
|
module_init(amd_iommu_v2_init);
|
||||||
|
Loading…
Reference in New Issue
Block a user