iommu/arm-smmu: clean up use of `flags' in page table handling code
Commit 972157cac5
("arm/smmu: Use irqsafe spinlock for domain lock")
fixed our page table locks to be the irq{save,restore} variants, since
the DMA mapping API can be invoked from interrupt context.
This patch cleans up our use of the flags variable so we can distinguish
between IRQ flags (now `flags') and pte protection bits (now `prot').
Signed-off-by: Will Deacon <will.deacon@arm.com>
This commit is contained in:
parent
0c9d70d724
commit
b410aed932
@ -1254,7 +1254,7 @@ static bool arm_smmu_pte_is_contiguous_range(unsigned long addr,
|
|||||||
|
|
||||||
static int arm_smmu_alloc_init_pte(struct arm_smmu_device *smmu, pmd_t *pmd,
|
static int arm_smmu_alloc_init_pte(struct arm_smmu_device *smmu, pmd_t *pmd,
|
||||||
unsigned long addr, unsigned long end,
|
unsigned long addr, unsigned long end,
|
||||||
unsigned long pfn, int flags, int stage)
|
unsigned long pfn, int prot, int stage)
|
||||||
{
|
{
|
||||||
pte_t *pte, *start;
|
pte_t *pte, *start;
|
||||||
pteval_t pteval = ARM_SMMU_PTE_PAGE | ARM_SMMU_PTE_AF | ARM_SMMU_PTE_XN;
|
pteval_t pteval = ARM_SMMU_PTE_PAGE | ARM_SMMU_PTE_AF | ARM_SMMU_PTE_XN;
|
||||||
@ -1276,28 +1276,28 @@ static int arm_smmu_alloc_init_pte(struct arm_smmu_device *smmu, pmd_t *pmd,
|
|||||||
|
|
||||||
if (stage == 1) {
|
if (stage == 1) {
|
||||||
pteval |= ARM_SMMU_PTE_AP_UNPRIV | ARM_SMMU_PTE_nG;
|
pteval |= ARM_SMMU_PTE_AP_UNPRIV | ARM_SMMU_PTE_nG;
|
||||||
if (!(flags & IOMMU_WRITE) && (flags & IOMMU_READ))
|
if (!(prot & IOMMU_WRITE) && (prot & IOMMU_READ))
|
||||||
pteval |= ARM_SMMU_PTE_AP_RDONLY;
|
pteval |= ARM_SMMU_PTE_AP_RDONLY;
|
||||||
|
|
||||||
if (flags & IOMMU_CACHE)
|
if (prot & IOMMU_CACHE)
|
||||||
pteval |= (MAIR_ATTR_IDX_CACHE <<
|
pteval |= (MAIR_ATTR_IDX_CACHE <<
|
||||||
ARM_SMMU_PTE_ATTRINDX_SHIFT);
|
ARM_SMMU_PTE_ATTRINDX_SHIFT);
|
||||||
} else {
|
} else {
|
||||||
pteval |= ARM_SMMU_PTE_HAP_FAULT;
|
pteval |= ARM_SMMU_PTE_HAP_FAULT;
|
||||||
if (flags & IOMMU_READ)
|
if (prot & IOMMU_READ)
|
||||||
pteval |= ARM_SMMU_PTE_HAP_READ;
|
pteval |= ARM_SMMU_PTE_HAP_READ;
|
||||||
if (flags & IOMMU_WRITE)
|
if (prot & IOMMU_WRITE)
|
||||||
pteval |= ARM_SMMU_PTE_HAP_WRITE;
|
pteval |= ARM_SMMU_PTE_HAP_WRITE;
|
||||||
if (flags & IOMMU_CACHE)
|
if (prot & IOMMU_CACHE)
|
||||||
pteval |= ARM_SMMU_PTE_MEMATTR_OIWB;
|
pteval |= ARM_SMMU_PTE_MEMATTR_OIWB;
|
||||||
else
|
else
|
||||||
pteval |= ARM_SMMU_PTE_MEMATTR_NC;
|
pteval |= ARM_SMMU_PTE_MEMATTR_NC;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* If no access, create a faulting entry to avoid TLB fills */
|
/* If no access, create a faulting entry to avoid TLB fills */
|
||||||
if (flags & IOMMU_EXEC)
|
if (prot & IOMMU_EXEC)
|
||||||
pteval &= ~ARM_SMMU_PTE_XN;
|
pteval &= ~ARM_SMMU_PTE_XN;
|
||||||
else if (!(flags & (IOMMU_READ | IOMMU_WRITE)))
|
else if (!(prot & (IOMMU_READ | IOMMU_WRITE)))
|
||||||
pteval &= ~ARM_SMMU_PTE_PAGE;
|
pteval &= ~ARM_SMMU_PTE_PAGE;
|
||||||
|
|
||||||
pteval |= ARM_SMMU_PTE_SH_IS;
|
pteval |= ARM_SMMU_PTE_SH_IS;
|
||||||
@ -1359,7 +1359,7 @@ static int arm_smmu_alloc_init_pte(struct arm_smmu_device *smmu, pmd_t *pmd,
|
|||||||
|
|
||||||
static int arm_smmu_alloc_init_pmd(struct arm_smmu_device *smmu, pud_t *pud,
|
static int arm_smmu_alloc_init_pmd(struct arm_smmu_device *smmu, pud_t *pud,
|
||||||
unsigned long addr, unsigned long end,
|
unsigned long addr, unsigned long end,
|
||||||
phys_addr_t phys, int flags, int stage)
|
phys_addr_t phys, int prot, int stage)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
pmd_t *pmd;
|
pmd_t *pmd;
|
||||||
@ -1383,7 +1383,7 @@ static int arm_smmu_alloc_init_pmd(struct arm_smmu_device *smmu, pud_t *pud,
|
|||||||
do {
|
do {
|
||||||
next = pmd_addr_end(addr, end);
|
next = pmd_addr_end(addr, end);
|
||||||
ret = arm_smmu_alloc_init_pte(smmu, pmd, addr, end, pfn,
|
ret = arm_smmu_alloc_init_pte(smmu, pmd, addr, end, pfn,
|
||||||
flags, stage);
|
prot, stage);
|
||||||
phys += next - addr;
|
phys += next - addr;
|
||||||
} while (pmd++, addr = next, addr < end);
|
} while (pmd++, addr = next, addr < end);
|
||||||
|
|
||||||
@ -1392,7 +1392,7 @@ static int arm_smmu_alloc_init_pmd(struct arm_smmu_device *smmu, pud_t *pud,
|
|||||||
|
|
||||||
static int arm_smmu_alloc_init_pud(struct arm_smmu_device *smmu, pgd_t *pgd,
|
static int arm_smmu_alloc_init_pud(struct arm_smmu_device *smmu, pgd_t *pgd,
|
||||||
unsigned long addr, unsigned long end,
|
unsigned long addr, unsigned long end,
|
||||||
phys_addr_t phys, int flags, int stage)
|
phys_addr_t phys, int prot, int stage)
|
||||||
{
|
{
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
pud_t *pud;
|
pud_t *pud;
|
||||||
@ -1416,7 +1416,7 @@ static int arm_smmu_alloc_init_pud(struct arm_smmu_device *smmu, pgd_t *pgd,
|
|||||||
do {
|
do {
|
||||||
next = pud_addr_end(addr, end);
|
next = pud_addr_end(addr, end);
|
||||||
ret = arm_smmu_alloc_init_pmd(smmu, pud, addr, next, phys,
|
ret = arm_smmu_alloc_init_pmd(smmu, pud, addr, next, phys,
|
||||||
flags, stage);
|
prot, stage);
|
||||||
phys += next - addr;
|
phys += next - addr;
|
||||||
} while (pud++, addr = next, addr < end);
|
} while (pud++, addr = next, addr < end);
|
||||||
|
|
||||||
@ -1425,7 +1425,7 @@ static int arm_smmu_alloc_init_pud(struct arm_smmu_device *smmu, pgd_t *pgd,
|
|||||||
|
|
||||||
static int arm_smmu_handle_mapping(struct arm_smmu_domain *smmu_domain,
|
static int arm_smmu_handle_mapping(struct arm_smmu_domain *smmu_domain,
|
||||||
unsigned long iova, phys_addr_t paddr,
|
unsigned long iova, phys_addr_t paddr,
|
||||||
size_t size, int flags)
|
size_t size, int prot)
|
||||||
{
|
{
|
||||||
int ret, stage;
|
int ret, stage;
|
||||||
unsigned long end;
|
unsigned long end;
|
||||||
@ -1433,7 +1433,7 @@ static int arm_smmu_handle_mapping(struct arm_smmu_domain *smmu_domain,
|
|||||||
struct arm_smmu_cfg *root_cfg = &smmu_domain->root_cfg;
|
struct arm_smmu_cfg *root_cfg = &smmu_domain->root_cfg;
|
||||||
pgd_t *pgd = root_cfg->pgd;
|
pgd_t *pgd = root_cfg->pgd;
|
||||||
struct arm_smmu_device *smmu = root_cfg->smmu;
|
struct arm_smmu_device *smmu = root_cfg->smmu;
|
||||||
unsigned long irqflags;
|
unsigned long flags;
|
||||||
|
|
||||||
if (root_cfg->cbar == CBAR_TYPE_S2_TRANS) {
|
if (root_cfg->cbar == CBAR_TYPE_S2_TRANS) {
|
||||||
stage = 2;
|
stage = 2;
|
||||||
@ -1456,14 +1456,14 @@ static int arm_smmu_handle_mapping(struct arm_smmu_domain *smmu_domain,
|
|||||||
if (paddr & ~output_mask)
|
if (paddr & ~output_mask)
|
||||||
return -ERANGE;
|
return -ERANGE;
|
||||||
|
|
||||||
spin_lock_irqsave(&smmu_domain->lock, irqflags);
|
spin_lock_irqsave(&smmu_domain->lock, flags);
|
||||||
pgd += pgd_index(iova);
|
pgd += pgd_index(iova);
|
||||||
end = iova + size;
|
end = iova + size;
|
||||||
do {
|
do {
|
||||||
unsigned long next = pgd_addr_end(iova, end);
|
unsigned long next = pgd_addr_end(iova, end);
|
||||||
|
|
||||||
ret = arm_smmu_alloc_init_pud(smmu, pgd, iova, next, paddr,
|
ret = arm_smmu_alloc_init_pud(smmu, pgd, iova, next, paddr,
|
||||||
flags, stage);
|
prot, stage);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
|
|
||||||
@ -1472,13 +1472,13 @@ static int arm_smmu_handle_mapping(struct arm_smmu_domain *smmu_domain,
|
|||||||
} while (pgd++, iova != end);
|
} while (pgd++, iova != end);
|
||||||
|
|
||||||
out_unlock:
|
out_unlock:
|
||||||
spin_unlock_irqrestore(&smmu_domain->lock, irqflags);
|
spin_unlock_irqrestore(&smmu_domain->lock, flags);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int arm_smmu_map(struct iommu_domain *domain, unsigned long iova,
|
static int arm_smmu_map(struct iommu_domain *domain, unsigned long iova,
|
||||||
phys_addr_t paddr, size_t size, int flags)
|
phys_addr_t paddr, size_t size, int prot)
|
||||||
{
|
{
|
||||||
struct arm_smmu_domain *smmu_domain = domain->priv;
|
struct arm_smmu_domain *smmu_domain = domain->priv;
|
||||||
|
|
||||||
@ -1489,7 +1489,7 @@ static int arm_smmu_map(struct iommu_domain *domain, unsigned long iova,
|
|||||||
if ((phys_addr_t)iova & ~smmu_domain->output_mask)
|
if ((phys_addr_t)iova & ~smmu_domain->output_mask)
|
||||||
return -ERANGE;
|
return -ERANGE;
|
||||||
|
|
||||||
return arm_smmu_handle_mapping(smmu_domain, iova, paddr, size, flags);
|
return arm_smmu_handle_mapping(smmu_domain, iova, paddr, size, prot);
|
||||||
}
|
}
|
||||||
|
|
||||||
static size_t arm_smmu_unmap(struct iommu_domain *domain, unsigned long iova,
|
static size_t arm_smmu_unmap(struct iommu_domain *domain, unsigned long iova,
|
||||||
|
Loading…
Reference in New Issue
Block a user