mirror of
https://github.com/torvalds/linux.git
synced 2024-11-14 08:02:07 +00:00
ebf7cda0f9
Use runtime PM functionality interfaced with hwmod enable/idle functions, to replace direct clock operations and sysconfig handling. Due to reset sequence, pm_runtime_[get|put]_sync must be used, to avoid possible operations with the module under reset. Because of this and given that the driver uses spin_locks to protect their critical sections, we must use pm_runtime_irq_safe in order for the runtime ops to be happy, otherwise might_sleep_if checks in runtime framework will complain. The remaining pm_runtime out of iommu_enable and iommu_disable corresponds to paths that can be accessed through debugfs, some of them doesn't work if the module is not enabled first, but in future if the mmu is idled withouth freeing, these are needed to debug. Signed-off-by: Omar Ramirez Luna <omar.luna@linaro.org> Tested-by: Ohad Ben-Cohen <ohad@wizery.com> Acked-by: Tony Lindgren <tony@atomide.com> Signed-off-by: Joerg Roedel <joro@8bytes.org>
226 lines
5.6 KiB
C
226 lines
5.6 KiB
C
/*
|
|
* omap iommu: main structures
|
|
*
|
|
* Copyright (C) 2008-2009 Nokia Corporation
|
|
*
|
|
* Written by Hiroshi DOYU <Hiroshi.DOYU@nokia.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
|
|
#if defined(CONFIG_ARCH_OMAP1)
|
|
#error "iommu for this processor not implemented yet"
|
|
#endif
|
|
|
|
struct iotlb_entry {
|
|
u32 da;
|
|
u32 pa;
|
|
u32 pgsz, prsvd, valid;
|
|
union {
|
|
u16 ap;
|
|
struct {
|
|
u32 endian, elsz, mixed;
|
|
};
|
|
};
|
|
};
|
|
|
|
struct omap_iommu {
|
|
const char *name;
|
|
struct module *owner;
|
|
void __iomem *regbase;
|
|
struct device *dev;
|
|
void *isr_priv;
|
|
struct iommu_domain *domain;
|
|
|
|
unsigned int refcount;
|
|
spinlock_t iommu_lock; /* global for this whole object */
|
|
|
|
/*
|
|
* We don't change iopgd for a situation like pgd for a task,
|
|
* but share it globally for each iommu.
|
|
*/
|
|
u32 *iopgd;
|
|
spinlock_t page_table_lock; /* protect iopgd */
|
|
|
|
int nr_tlb_entries;
|
|
|
|
struct list_head mmap;
|
|
struct mutex mmap_lock; /* protect mmap */
|
|
|
|
void *ctx; /* iommu context: registres saved area */
|
|
u32 da_start;
|
|
u32 da_end;
|
|
};
|
|
|
|
struct cr_regs {
|
|
union {
|
|
struct {
|
|
u16 cam_l;
|
|
u16 cam_h;
|
|
};
|
|
u32 cam;
|
|
};
|
|
union {
|
|
struct {
|
|
u16 ram_l;
|
|
u16 ram_h;
|
|
};
|
|
u32 ram;
|
|
};
|
|
};
|
|
|
|
/* architecture specific functions */
|
|
struct iommu_functions {
|
|
unsigned long version;
|
|
|
|
int (*enable)(struct omap_iommu *obj);
|
|
void (*disable)(struct omap_iommu *obj);
|
|
void (*set_twl)(struct omap_iommu *obj, bool on);
|
|
u32 (*fault_isr)(struct omap_iommu *obj, u32 *ra);
|
|
|
|
void (*tlb_read_cr)(struct omap_iommu *obj, struct cr_regs *cr);
|
|
void (*tlb_load_cr)(struct omap_iommu *obj, struct cr_regs *cr);
|
|
|
|
struct cr_regs *(*alloc_cr)(struct omap_iommu *obj,
|
|
struct iotlb_entry *e);
|
|
int (*cr_valid)(struct cr_regs *cr);
|
|
u32 (*cr_to_virt)(struct cr_regs *cr);
|
|
void (*cr_to_e)(struct cr_regs *cr, struct iotlb_entry *e);
|
|
ssize_t (*dump_cr)(struct omap_iommu *obj, struct cr_regs *cr,
|
|
char *buf);
|
|
|
|
u32 (*get_pte_attr)(struct iotlb_entry *e);
|
|
|
|
void (*save_ctx)(struct omap_iommu *obj);
|
|
void (*restore_ctx)(struct omap_iommu *obj);
|
|
ssize_t (*dump_ctx)(struct omap_iommu *obj, char *buf, ssize_t len);
|
|
};
|
|
|
|
#ifdef CONFIG_IOMMU_API
|
|
/**
|
|
* dev_to_omap_iommu() - retrieves an omap iommu object from a user device
|
|
* @dev: iommu client device
|
|
*/
|
|
static inline struct omap_iommu *dev_to_omap_iommu(struct device *dev)
|
|
{
|
|
struct omap_iommu_arch_data *arch_data = dev->archdata.iommu;
|
|
|
|
return arch_data->iommu_dev;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* MMU Register offsets
|
|
*/
|
|
#define MMU_REVISION 0x00
|
|
#define MMU_IRQSTATUS 0x18
|
|
#define MMU_IRQENABLE 0x1c
|
|
#define MMU_WALKING_ST 0x40
|
|
#define MMU_CNTL 0x44
|
|
#define MMU_FAULT_AD 0x48
|
|
#define MMU_TTB 0x4c
|
|
#define MMU_LOCK 0x50
|
|
#define MMU_LD_TLB 0x54
|
|
#define MMU_CAM 0x58
|
|
#define MMU_RAM 0x5c
|
|
#define MMU_GFLUSH 0x60
|
|
#define MMU_FLUSH_ENTRY 0x64
|
|
#define MMU_READ_CAM 0x68
|
|
#define MMU_READ_RAM 0x6c
|
|
#define MMU_EMU_FAULT_AD 0x70
|
|
|
|
#define MMU_REG_SIZE 256
|
|
|
|
/*
|
|
* MMU Register bit definitions
|
|
*/
|
|
#define MMU_CAM_VATAG_SHIFT 12
|
|
#define MMU_CAM_VATAG_MASK \
|
|
((~0UL >> MMU_CAM_VATAG_SHIFT) << MMU_CAM_VATAG_SHIFT)
|
|
#define MMU_CAM_P (1 << 3)
|
|
#define MMU_CAM_V (1 << 2)
|
|
#define MMU_CAM_PGSZ_MASK 3
|
|
#define MMU_CAM_PGSZ_1M (0 << 0)
|
|
#define MMU_CAM_PGSZ_64K (1 << 0)
|
|
#define MMU_CAM_PGSZ_4K (2 << 0)
|
|
#define MMU_CAM_PGSZ_16M (3 << 0)
|
|
|
|
#define MMU_RAM_PADDR_SHIFT 12
|
|
#define MMU_RAM_PADDR_MASK \
|
|
((~0UL >> MMU_RAM_PADDR_SHIFT) << MMU_RAM_PADDR_SHIFT)
|
|
|
|
#define MMU_RAM_ENDIAN_MASK (1 << MMU_RAM_ENDIAN_SHIFT)
|
|
#define MMU_RAM_ENDIAN_BIG (1 << MMU_RAM_ENDIAN_SHIFT)
|
|
|
|
#define MMU_RAM_ELSZ_MASK (3 << MMU_RAM_ELSZ_SHIFT)
|
|
#define MMU_RAM_ELSZ_8 (0 << MMU_RAM_ELSZ_SHIFT)
|
|
#define MMU_RAM_ELSZ_16 (1 << MMU_RAM_ELSZ_SHIFT)
|
|
#define MMU_RAM_ELSZ_32 (2 << MMU_RAM_ELSZ_SHIFT)
|
|
#define MMU_RAM_ELSZ_NONE (3 << MMU_RAM_ELSZ_SHIFT)
|
|
#define MMU_RAM_MIXED_SHIFT 6
|
|
#define MMU_RAM_MIXED_MASK (1 << MMU_RAM_MIXED_SHIFT)
|
|
#define MMU_RAM_MIXED MMU_RAM_MIXED_MASK
|
|
|
|
/*
|
|
* utilities for super page(16MB, 1MB, 64KB and 4KB)
|
|
*/
|
|
|
|
#define iopgsz_max(bytes) \
|
|
(((bytes) >= SZ_16M) ? SZ_16M : \
|
|
((bytes) >= SZ_1M) ? SZ_1M : \
|
|
((bytes) >= SZ_64K) ? SZ_64K : \
|
|
((bytes) >= SZ_4K) ? SZ_4K : 0)
|
|
|
|
#define bytes_to_iopgsz(bytes) \
|
|
(((bytes) == SZ_16M) ? MMU_CAM_PGSZ_16M : \
|
|
((bytes) == SZ_1M) ? MMU_CAM_PGSZ_1M : \
|
|
((bytes) == SZ_64K) ? MMU_CAM_PGSZ_64K : \
|
|
((bytes) == SZ_4K) ? MMU_CAM_PGSZ_4K : -1)
|
|
|
|
#define iopgsz_to_bytes(iopgsz) \
|
|
(((iopgsz) == MMU_CAM_PGSZ_16M) ? SZ_16M : \
|
|
((iopgsz) == MMU_CAM_PGSZ_1M) ? SZ_1M : \
|
|
((iopgsz) == MMU_CAM_PGSZ_64K) ? SZ_64K : \
|
|
((iopgsz) == MMU_CAM_PGSZ_4K) ? SZ_4K : 0)
|
|
|
|
#define iopgsz_ok(bytes) (bytes_to_iopgsz(bytes) >= 0)
|
|
|
|
/*
|
|
* global functions
|
|
*/
|
|
extern u32 omap_iommu_arch_version(void);
|
|
|
|
extern void omap_iotlb_cr_to_e(struct cr_regs *cr, struct iotlb_entry *e);
|
|
|
|
extern int
|
|
omap_iopgtable_store_entry(struct omap_iommu *obj, struct iotlb_entry *e);
|
|
|
|
extern void omap_iommu_save_ctx(struct device *dev);
|
|
extern void omap_iommu_restore_ctx(struct device *dev);
|
|
|
|
extern int omap_foreach_iommu_device(void *data,
|
|
int (*fn)(struct device *, void *));
|
|
|
|
extern int omap_install_iommu_arch(const struct iommu_functions *ops);
|
|
extern void omap_uninstall_iommu_arch(const struct iommu_functions *ops);
|
|
|
|
extern ssize_t
|
|
omap_iommu_dump_ctx(struct omap_iommu *obj, char *buf, ssize_t len);
|
|
extern size_t
|
|
omap_dump_tlb_entries(struct omap_iommu *obj, char *buf, ssize_t len);
|
|
|
|
/*
|
|
* register accessors
|
|
*/
|
|
static inline u32 iommu_read_reg(struct omap_iommu *obj, size_t offs)
|
|
{
|
|
return __raw_readl(obj->regbase + offs);
|
|
}
|
|
|
|
static inline void iommu_write_reg(struct omap_iommu *obj, u32 val, size_t offs)
|
|
{
|
|
__raw_writel(val, obj->regbase + offs);
|
|
}
|