forked from Minki/linux
dma, mm/pat: Rename dma_*_writecombine() to dma_*_wc()
Rename dma_*_writecombine() to dma_*_wc(), so that the naming is coherent across the various write-combining APIs. Keep the old names for compatibility for a while, these can be removed at a later time. A guard is left to enable backporting of the rename, and later remove of the old mapping defines seemlessly. Build tested successfully with allmodconfig. The following Coccinelle SmPL patch was used for this simple transformation: @ rename_dma_alloc_writecombine @ expression dev, size, dma_addr, gfp; @@ -dma_alloc_writecombine(dev, size, dma_addr, gfp) +dma_alloc_wc(dev, size, dma_addr, gfp) @ rename_dma_free_writecombine @ expression dev, size, cpu_addr, dma_addr; @@ -dma_free_writecombine(dev, size, cpu_addr, dma_addr) +dma_free_wc(dev, size, cpu_addr, dma_addr) @ rename_dma_mmap_writecombine @ expression dev, vma, cpu_addr, dma_addr, size; @@ -dma_mmap_writecombine(dev, vma, cpu_addr, dma_addr, size) +dma_mmap_wc(dev, vma, cpu_addr, dma_addr, size) We also keep the old names as compatibility helpers, and guard against their definition to make backporting easier. Generated-by: Coccinelle SmPL Suggested-by: Ingo Molnar <mingo@kernel.org> Signed-off-by: Luis R. Rodriguez <mcgrof@suse.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: airlied@linux.ie Cc: akpm@linux-foundation.org Cc: benh@kernel.crashing.org Cc: bhelgaas@google.com Cc: bp@suse.de Cc: dan.j.williams@intel.com Cc: daniel.vetter@ffwll.ch Cc: dhowells@redhat.com Cc: julia.lawall@lip6.fr Cc: konrad.wilk@oracle.com Cc: linux-fbdev@vger.kernel.org Cc: linux-pci@vger.kernel.org Cc: luto@amacapital.net Cc: mst@redhat.com Cc: tomi.valkeinen@ti.com Cc: toshi.kani@hp.com Cc: vinod.koul@intel.com Cc: xen-devel@lists.xensource.com Link: http://lkml.kernel.org/r/1453516462-4844-1-git-send-email-mcgrof@do-not-panic.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
parent
e2857b8f11
commit
f6e45661f9
@ -86,8 +86,8 @@ static int lpc32xx_clcd_setup(struct clcd_fb *fb)
|
||||
{
|
||||
dma_addr_t dma;
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev,
|
||||
PANEL_SIZE, &dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, PANEL_SIZE, &dma,
|
||||
GFP_KERNEL);
|
||||
if (!fb->fb.screen_base) {
|
||||
printk(KERN_ERR "CLCD: unable to map framebuffer\n");
|
||||
return -ENOMEM;
|
||||
@ -116,15 +116,14 @@ static int lpc32xx_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
static int lpc32xx_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
static void lpc32xx_clcd_remove(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -42,8 +42,8 @@ int netx_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
fb->panel = netx_panel;
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev, 1024*1024,
|
||||
&dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, 1024 * 1024, &dma,
|
||||
GFP_KERNEL);
|
||||
if (!fb->fb.screen_base) {
|
||||
printk(KERN_ERR "CLCD: unable to map framebuffer\n");
|
||||
return -ENOMEM;
|
||||
@ -57,16 +57,14 @@ int netx_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
int netx_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
void netx_clcd_remove(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
||||
static AMBA_AHB_DEVICE(fb, "fb", 0, 0x00104000, { NETX_IRQ_LCD }, NULL);
|
||||
|
@ -90,8 +90,8 @@ int nspire_clcd_setup(struct clcd_fb *fb)
|
||||
panel_size = ((panel->mode.xres * panel->mode.yres) * panel->bpp) / 8;
|
||||
panel_size = ALIGN(panel_size, PAGE_SIZE);
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev,
|
||||
panel_size, &dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, panel_size, &dma,
|
||||
GFP_KERNEL);
|
||||
|
||||
if (!fb->fb.screen_base) {
|
||||
pr_err("CLCD: unable to map framebuffer\n");
|
||||
@ -107,13 +107,12 @@ int nspire_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
int nspire_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
void nspire_clcd_remove(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
@ -1300,10 +1300,10 @@ static int iop_adma_probe(struct platform_device *pdev)
|
||||
* note: writecombine gives slightly better performance, but
|
||||
* requires that we explicitly flush the writes
|
||||
*/
|
||||
adev->dma_desc_pool_virt = dma_alloc_writecombine(&pdev->dev,
|
||||
plat_data->pool_size,
|
||||
&adev->dma_desc_pool,
|
||||
GFP_KERNEL);
|
||||
adev->dma_desc_pool_virt = dma_alloc_wc(&pdev->dev,
|
||||
plat_data->pool_size,
|
||||
&adev->dma_desc_pool,
|
||||
GFP_KERNEL);
|
||||
if (!adev->dma_desc_pool_virt) {
|
||||
ret = -ENOMEM;
|
||||
goto err_free_adev;
|
||||
|
@ -964,8 +964,8 @@ mv_xor_channel_add(struct mv_xor_device *xordev,
|
||||
* requires that we explicitly flush the writes
|
||||
*/
|
||||
mv_chan->dma_desc_pool_virt =
|
||||
dma_alloc_writecombine(&pdev->dev, MV_XOR_POOL_SIZE,
|
||||
&mv_chan->dma_desc_pool, GFP_KERNEL);
|
||||
dma_alloc_wc(&pdev->dev, MV_XOR_POOL_SIZE, &mv_chan->dma_desc_pool,
|
||||
GFP_KERNEL);
|
||||
if (!mv_chan->dma_desc_pool_virt)
|
||||
return ERR_PTR(-ENOMEM);
|
||||
|
||||
|
@ -502,8 +502,8 @@ static int bam_alloc_chan(struct dma_chan *chan)
|
||||
return 0;
|
||||
|
||||
/* allocate FIFO descriptor space, but only if necessary */
|
||||
bchan->fifo_virt = dma_alloc_writecombine(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
&bchan->fifo_phys, GFP_KERNEL);
|
||||
bchan->fifo_virt = dma_alloc_wc(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
&bchan->fifo_phys, GFP_KERNEL);
|
||||
|
||||
if (!bchan->fifo_virt) {
|
||||
dev_err(bdev->dev, "Failed to allocate desc fifo\n");
|
||||
@ -538,8 +538,8 @@ static void bam_free_chan(struct dma_chan *chan)
|
||||
bam_reset_channel(bchan);
|
||||
spin_unlock_irqrestore(&bchan->vc.lock, flags);
|
||||
|
||||
dma_free_writecombine(bdev->dev, BAM_DESC_FIFO_SIZE, bchan->fifo_virt,
|
||||
bchan->fifo_phys);
|
||||
dma_free_wc(bdev->dev, BAM_DESC_FIFO_SIZE, bchan->fifo_virt,
|
||||
bchan->fifo_phys);
|
||||
bchan->fifo_virt = NULL;
|
||||
|
||||
/* mask irq for pipe/channel */
|
||||
@ -1231,9 +1231,9 @@ static int bam_dma_remove(struct platform_device *pdev)
|
||||
bam_dma_terminate_all(&bdev->channels[i].vc.chan);
|
||||
tasklet_kill(&bdev->channels[i].vc.task);
|
||||
|
||||
dma_free_writecombine(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
bdev->channels[i].fifo_virt,
|
||||
bdev->channels[i].fifo_phys);
|
||||
dma_free_wc(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
bdev->channels[i].fifo_virt,
|
||||
bdev->channels[i].fifo_phys);
|
||||
}
|
||||
|
||||
tasklet_kill(&bdev->task);
|
||||
|
@ -109,8 +109,8 @@ struct drm_gem_cma_object *drm_gem_cma_create(struct drm_device *drm,
|
||||
if (IS_ERR(cma_obj))
|
||||
return cma_obj;
|
||||
|
||||
cma_obj->vaddr = dma_alloc_writecombine(drm->dev, size,
|
||||
&cma_obj->paddr, GFP_KERNEL | __GFP_NOWARN);
|
||||
cma_obj->vaddr = dma_alloc_wc(drm->dev, size, &cma_obj->paddr,
|
||||
GFP_KERNEL | __GFP_NOWARN);
|
||||
if (!cma_obj->vaddr) {
|
||||
dev_err(drm->dev, "failed to allocate buffer with size %zu\n",
|
||||
size);
|
||||
@ -192,8 +192,8 @@ void drm_gem_cma_free_object(struct drm_gem_object *gem_obj)
|
||||
cma_obj = to_drm_gem_cma_obj(gem_obj);
|
||||
|
||||
if (cma_obj->vaddr) {
|
||||
dma_free_writecombine(gem_obj->dev->dev, cma_obj->base.size,
|
||||
cma_obj->vaddr, cma_obj->paddr);
|
||||
dma_free_wc(gem_obj->dev->dev, cma_obj->base.size,
|
||||
cma_obj->vaddr, cma_obj->paddr);
|
||||
} else if (gem_obj->import_attach) {
|
||||
drm_prime_gem_destroy(gem_obj, cma_obj->sgt);
|
||||
}
|
||||
@ -324,9 +324,8 @@ static int drm_gem_cma_mmap_obj(struct drm_gem_cma_object *cma_obj,
|
||||
vma->vm_flags &= ~VM_PFNMAP;
|
||||
vma->vm_pgoff = 0;
|
||||
|
||||
ret = dma_mmap_writecombine(cma_obj->base.dev->dev, vma,
|
||||
cma_obj->vaddr, cma_obj->paddr,
|
||||
vma->vm_end - vma->vm_start);
|
||||
ret = dma_mmap_wc(cma_obj->base.dev->dev, vma, cma_obj->vaddr,
|
||||
cma_obj->paddr, vma->vm_end - vma->vm_start);
|
||||
if (ret)
|
||||
drm_gem_vm_close(vma);
|
||||
|
||||
|
@ -1113,8 +1113,8 @@ struct etnaviv_cmdbuf *etnaviv_gpu_cmdbuf_new(struct etnaviv_gpu *gpu, u32 size,
|
||||
if (!cmdbuf)
|
||||
return NULL;
|
||||
|
||||
cmdbuf->vaddr = dma_alloc_writecombine(gpu->dev, size, &cmdbuf->paddr,
|
||||
GFP_KERNEL);
|
||||
cmdbuf->vaddr = dma_alloc_wc(gpu->dev, size, &cmdbuf->paddr,
|
||||
GFP_KERNEL);
|
||||
if (!cmdbuf->vaddr) {
|
||||
kfree(cmdbuf);
|
||||
return NULL;
|
||||
@ -1128,8 +1128,8 @@ struct etnaviv_cmdbuf *etnaviv_gpu_cmdbuf_new(struct etnaviv_gpu *gpu, u32 size,
|
||||
|
||||
void etnaviv_gpu_cmdbuf_free(struct etnaviv_cmdbuf *cmdbuf)
|
||||
{
|
||||
dma_free_writecombine(cmdbuf->gpu->dev, cmdbuf->size,
|
||||
cmdbuf->vaddr, cmdbuf->paddr);
|
||||
dma_free_wc(cmdbuf->gpu->dev, cmdbuf->size, cmdbuf->vaddr,
|
||||
cmdbuf->paddr);
|
||||
kfree(cmdbuf);
|
||||
}
|
||||
|
||||
|
@ -573,10 +573,9 @@ static int omap_dmm_remove(struct platform_device *dev)
|
||||
|
||||
kfree(omap_dmm->engines);
|
||||
if (omap_dmm->refill_va)
|
||||
dma_free_writecombine(omap_dmm->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
omap_dmm->refill_va,
|
||||
omap_dmm->refill_pa);
|
||||
dma_free_wc(omap_dmm->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
omap_dmm->refill_va, omap_dmm->refill_pa);
|
||||
if (omap_dmm->dummy_page)
|
||||
__free_page(omap_dmm->dummy_page);
|
||||
|
||||
@ -701,9 +700,9 @@ static int omap_dmm_probe(struct platform_device *dev)
|
||||
omap_dmm->dummy_pa = page_to_phys(omap_dmm->dummy_page);
|
||||
|
||||
/* alloc refill memory */
|
||||
omap_dmm->refill_va = dma_alloc_writecombine(&dev->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
&omap_dmm->refill_pa, GFP_KERNEL);
|
||||
omap_dmm->refill_va = dma_alloc_wc(&dev->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
&omap_dmm->refill_pa, GFP_KERNEL);
|
||||
if (!omap_dmm->refill_va) {
|
||||
dev_err(&dev->dev, "could not allocate refill memory\n");
|
||||
goto fail;
|
||||
|
@ -1330,8 +1330,8 @@ void omap_gem_free_object(struct drm_gem_object *obj)
|
||||
omap_gem_detach_pages(obj);
|
||||
|
||||
if (!is_shmem(obj)) {
|
||||
dma_free_writecombine(dev->dev, obj->size,
|
||||
omap_obj->vaddr, omap_obj->paddr);
|
||||
dma_free_wc(dev->dev, obj->size, omap_obj->vaddr,
|
||||
omap_obj->paddr);
|
||||
} else if (omap_obj->vaddr) {
|
||||
vunmap(omap_obj->vaddr);
|
||||
}
|
||||
@ -1395,8 +1395,8 @@ struct drm_gem_object *omap_gem_new(struct drm_device *dev,
|
||||
/* attempt to allocate contiguous memory if we don't
|
||||
* have DMM for remappign discontiguous buffers
|
||||
*/
|
||||
omap_obj->vaddr = dma_alloc_writecombine(dev->dev, size,
|
||||
&omap_obj->paddr, GFP_KERNEL);
|
||||
omap_obj->vaddr = dma_alloc_wc(dev->dev, size,
|
||||
&omap_obj->paddr, GFP_KERNEL);
|
||||
if (!omap_obj->vaddr) {
|
||||
kfree(omap_obj);
|
||||
|
||||
|
@ -157,17 +157,15 @@ static void sti_cursor_atomic_update(struct drm_plane *drm_plane,
|
||||
cursor->height = src_h;
|
||||
|
||||
if (cursor->pixmap.base)
|
||||
dma_free_writecombine(cursor->dev,
|
||||
cursor->pixmap.size,
|
||||
cursor->pixmap.base,
|
||||
cursor->pixmap.paddr);
|
||||
dma_free_wc(cursor->dev, cursor->pixmap.size,
|
||||
cursor->pixmap.base, cursor->pixmap.paddr);
|
||||
|
||||
cursor->pixmap.size = cursor->width * cursor->height;
|
||||
|
||||
cursor->pixmap.base = dma_alloc_writecombine(cursor->dev,
|
||||
cursor->pixmap.size,
|
||||
&cursor->pixmap.paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
cursor->pixmap.base = dma_alloc_wc(cursor->dev,
|
||||
cursor->pixmap.size,
|
||||
&cursor->pixmap.paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
if (!cursor->pixmap.base) {
|
||||
DRM_ERROR("Failed to allocate memory for pixmap\n");
|
||||
return;
|
||||
@ -252,8 +250,8 @@ struct drm_plane *sti_cursor_create(struct drm_device *drm_dev,
|
||||
|
||||
/* Allocate clut buffer */
|
||||
size = 0x100 * sizeof(unsigned short);
|
||||
cursor->clut = dma_alloc_writecombine(dev, size, &cursor->clut_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
cursor->clut = dma_alloc_wc(dev, size, &cursor->clut_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
|
||||
if (!cursor->clut) {
|
||||
DRM_ERROR("Failed to allocate memory for cursor clut\n");
|
||||
@ -286,7 +284,7 @@ struct drm_plane *sti_cursor_create(struct drm_device *drm_dev,
|
||||
return &cursor->plane.drm_plane;
|
||||
|
||||
err_plane:
|
||||
dma_free_writecombine(dev, size, cursor->clut, cursor->clut_paddr);
|
||||
dma_free_wc(dev, size, cursor->clut, cursor->clut_paddr);
|
||||
err_clut:
|
||||
devm_kfree(dev, cursor);
|
||||
return NULL;
|
||||
|
@ -312,8 +312,7 @@ static void sti_gdp_init(struct sti_gdp *gdp)
|
||||
/* Allocate all the nodes within a single memory page */
|
||||
size = sizeof(struct sti_gdp_node) *
|
||||
GDP_NODE_PER_FIELD * GDP_NODE_NB_BANK;
|
||||
base = dma_alloc_writecombine(gdp->dev,
|
||||
size, &dma_addr, GFP_KERNEL | GFP_DMA);
|
||||
base = dma_alloc_wc(gdp->dev, size, &dma_addr, GFP_KERNEL | GFP_DMA);
|
||||
|
||||
if (!base) {
|
||||
DRM_ERROR("Failed to allocate memory for GDP node\n");
|
||||
|
@ -617,9 +617,9 @@ static void sti_hqvdp_init(struct sti_hqvdp *hqvdp)
|
||||
|
||||
/* Allocate memory for the VDP commands */
|
||||
size = NB_VDP_CMD * sizeof(struct sti_hqvdp_cmd);
|
||||
hqvdp->hqvdp_cmd = dma_alloc_writecombine(hqvdp->dev, size,
|
||||
&hqvdp->hqvdp_cmd_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
hqvdp->hqvdp_cmd = dma_alloc_wc(hqvdp->dev, size,
|
||||
&hqvdp->hqvdp_cmd_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
if (!hqvdp->hqvdp_cmd) {
|
||||
DRM_ERROR("Failed to allocate memory for VDP cmd\n");
|
||||
return;
|
||||
|
@ -175,8 +175,7 @@ static void tegra_bo_free(struct drm_device *drm, struct tegra_bo *bo)
|
||||
sg_free_table(bo->sgt);
|
||||
kfree(bo->sgt);
|
||||
} else if (bo->vaddr) {
|
||||
dma_free_writecombine(drm->dev, bo->gem.size, bo->vaddr,
|
||||
bo->paddr);
|
||||
dma_free_wc(drm->dev, bo->gem.size, bo->vaddr, bo->paddr);
|
||||
}
|
||||
}
|
||||
|
||||
@ -233,8 +232,8 @@ static int tegra_bo_alloc(struct drm_device *drm, struct tegra_bo *bo)
|
||||
} else {
|
||||
size_t size = bo->gem.size;
|
||||
|
||||
bo->vaddr = dma_alloc_writecombine(drm->dev, size, &bo->paddr,
|
||||
GFP_KERNEL | __GFP_NOWARN);
|
||||
bo->vaddr = dma_alloc_wc(drm->dev, size, &bo->paddr,
|
||||
GFP_KERNEL | __GFP_NOWARN);
|
||||
if (!bo->vaddr) {
|
||||
dev_err(drm->dev,
|
||||
"failed to allocate buffer of size %zu\n",
|
||||
@ -472,8 +471,8 @@ int tegra_drm_mmap(struct file *file, struct vm_area_struct *vma)
|
||||
vma->vm_flags &= ~VM_PFNMAP;
|
||||
vma->vm_pgoff = 0;
|
||||
|
||||
ret = dma_mmap_writecombine(gem->dev->dev, vma, bo->vaddr,
|
||||
bo->paddr, gem->size);
|
||||
ret = dma_mmap_wc(gem->dev->dev, vma, bo->vaddr, bo->paddr,
|
||||
gem->size);
|
||||
if (ret) {
|
||||
drm_gem_vm_close(vma);
|
||||
return ret;
|
||||
|
@ -398,9 +398,8 @@ int vc4_mmap(struct file *filp, struct vm_area_struct *vma)
|
||||
vma->vm_flags &= ~VM_PFNMAP;
|
||||
vma->vm_pgoff = 0;
|
||||
|
||||
ret = dma_mmap_writecombine(bo->base.base.dev->dev, vma,
|
||||
bo->base.vaddr, bo->base.paddr,
|
||||
vma->vm_end - vma->vm_start);
|
||||
ret = dma_mmap_wc(bo->base.base.dev->dev, vma, bo->base.vaddr,
|
||||
bo->base.paddr, vma->vm_end - vma->vm_start);
|
||||
if (ret)
|
||||
drm_gem_vm_close(vma);
|
||||
|
||||
|
@ -52,8 +52,8 @@ static void host1x_pushbuffer_destroy(struct push_buffer *pb)
|
||||
struct host1x *host1x = cdma_to_host1x(cdma);
|
||||
|
||||
if (pb->phys != 0)
|
||||
dma_free_writecombine(host1x->dev, pb->size_bytes + 4,
|
||||
pb->mapped, pb->phys);
|
||||
dma_free_wc(host1x->dev, pb->size_bytes + 4, pb->mapped,
|
||||
pb->phys);
|
||||
|
||||
pb->mapped = NULL;
|
||||
pb->phys = 0;
|
||||
@ -76,8 +76,8 @@ static int host1x_pushbuffer_init(struct push_buffer *pb)
|
||||
pb->pos = 0;
|
||||
|
||||
/* allocate and map pushbuffer memory */
|
||||
pb->mapped = dma_alloc_writecombine(host1x->dev, pb->size_bytes + 4,
|
||||
&pb->phys, GFP_KERNEL);
|
||||
pb->mapped = dma_alloc_wc(host1x->dev, pb->size_bytes + 4, &pb->phys,
|
||||
GFP_KERNEL);
|
||||
if (!pb->mapped)
|
||||
goto fail;
|
||||
|
||||
|
@ -467,9 +467,8 @@ static inline int copy_gathers(struct host1x_job *job, struct device *dev)
|
||||
size += g->words * sizeof(u32);
|
||||
}
|
||||
|
||||
job->gather_copy_mapped = dma_alloc_writecombine(dev, size,
|
||||
&job->gather_copy,
|
||||
GFP_KERNEL);
|
||||
job->gather_copy_mapped = dma_alloc_wc(dev, size, &job->gather_copy,
|
||||
GFP_KERNEL);
|
||||
if (!job->gather_copy_mapped) {
|
||||
job->gather_copy_mapped = NULL;
|
||||
return -ENOMEM;
|
||||
@ -578,9 +577,8 @@ void host1x_job_unpin(struct host1x_job *job)
|
||||
job->num_unpins = 0;
|
||||
|
||||
if (job->gather_copy_size)
|
||||
dma_free_writecombine(job->channel->dev, job->gather_copy_size,
|
||||
job->gather_copy_mapped,
|
||||
job->gather_copy);
|
||||
dma_free_wc(job->channel->dev, job->gather_copy_size,
|
||||
job->gather_copy_mapped, job->gather_copy);
|
||||
}
|
||||
EXPORT_SYMBOL(host1x_job_unpin);
|
||||
|
||||
|
@ -1455,9 +1455,9 @@ static int coda_alloc_bitstream_buffer(struct coda_ctx *ctx,
|
||||
return 0;
|
||||
|
||||
ctx->bitstream.size = roundup_pow_of_two(q_data->sizeimage * 2);
|
||||
ctx->bitstream.vaddr = dma_alloc_writecombine(
|
||||
&ctx->dev->plat_dev->dev, ctx->bitstream.size,
|
||||
&ctx->bitstream.paddr, GFP_KERNEL);
|
||||
ctx->bitstream.vaddr = dma_alloc_wc(&ctx->dev->plat_dev->dev,
|
||||
ctx->bitstream.size,
|
||||
&ctx->bitstream.paddr, GFP_KERNEL);
|
||||
if (!ctx->bitstream.vaddr) {
|
||||
v4l2_err(&ctx->dev->v4l2_dev,
|
||||
"failed to allocate bitstream ringbuffer");
|
||||
@ -1474,8 +1474,8 @@ static void coda_free_bitstream_buffer(struct coda_ctx *ctx)
|
||||
if (ctx->bitstream.vaddr == NULL)
|
||||
return;
|
||||
|
||||
dma_free_writecombine(&ctx->dev->plat_dev->dev, ctx->bitstream.size,
|
||||
ctx->bitstream.vaddr, ctx->bitstream.paddr);
|
||||
dma_free_wc(&ctx->dev->plat_dev->dev, ctx->bitstream.size,
|
||||
ctx->bitstream.vaddr, ctx->bitstream.paddr);
|
||||
ctx->bitstream.vaddr = NULL;
|
||||
kfifo_init(&ctx->bitstream_fifo, NULL, 0);
|
||||
}
|
||||
|
@ -1040,8 +1040,8 @@ static int acornfb_probe(struct platform_device *dev)
|
||||
* for the framebuffer if we are not using
|
||||
* VRAM.
|
||||
*/
|
||||
base = dma_alloc_writecombine(current_par.dev, size, &handle,
|
||||
GFP_KERNEL);
|
||||
base = dma_alloc_wc(current_par.dev, size, &handle,
|
||||
GFP_KERNEL);
|
||||
if (base == NULL) {
|
||||
printk(KERN_ERR "acornfb: unable to allocate screen "
|
||||
"memory\n");
|
||||
|
@ -154,8 +154,8 @@ int versatile_clcd_setup_dma(struct clcd_fb *fb, unsigned long framesize)
|
||||
{
|
||||
dma_addr_t dma;
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev, framesize,
|
||||
&dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, framesize, &dma,
|
||||
GFP_KERNEL);
|
||||
if (!fb->fb.screen_base) {
|
||||
pr_err("CLCD: unable to map framebuffer\n");
|
||||
return -ENOMEM;
|
||||
@ -169,14 +169,12 @@ int versatile_clcd_setup_dma(struct clcd_fb *fb, unsigned long framesize)
|
||||
|
||||
int versatile_clcd_mmap_dma(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
void versatile_clcd_remove_dma(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
@ -774,8 +774,8 @@ static int clcdfb_of_dma_setup(struct clcd_fb *fb)
|
||||
|
||||
static int clcdfb_of_dma_mmap(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
static void clcdfb_of_dma_remove(struct clcd_fb *fb)
|
||||
|
@ -414,8 +414,8 @@ static inline void atmel_lcdfb_free_video_memory(struct atmel_lcdfb_info *sinfo)
|
||||
{
|
||||
struct fb_info *info = sinfo->info;
|
||||
|
||||
dma_free_writecombine(info->device, info->fix.smem_len,
|
||||
info->screen_base, info->fix.smem_start);
|
||||
dma_free_wc(info->device, info->fix.smem_len, info->screen_base,
|
||||
info->fix.smem_start);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -435,8 +435,9 @@ static int atmel_lcdfb_alloc_video_memory(struct atmel_lcdfb_info *sinfo)
|
||||
* ((var->bits_per_pixel + 7) / 8));
|
||||
info->fix.smem_len = max(smem_len, sinfo->smem_len);
|
||||
|
||||
info->screen_base = dma_alloc_writecombine(info->device, info->fix.smem_len,
|
||||
(dma_addr_t *)&info->fix.smem_start, GFP_KERNEL);
|
||||
info->screen_base = dma_alloc_wc(info->device, info->fix.smem_len,
|
||||
(dma_addr_t *)&info->fix.smem_start,
|
||||
GFP_KERNEL);
|
||||
|
||||
if (!info->screen_base) {
|
||||
return -ENOMEM;
|
||||
|
@ -316,9 +316,8 @@ static int ep93xxfb_mmap(struct fb_info *info, struct vm_area_struct *vma)
|
||||
unsigned int offset = vma->vm_pgoff << PAGE_SHIFT;
|
||||
|
||||
if (offset < info->fix.smem_len) {
|
||||
return dma_mmap_writecombine(info->dev, vma, info->screen_base,
|
||||
info->fix.smem_start,
|
||||
info->fix.smem_len);
|
||||
return dma_mmap_wc(info->dev, vma, info->screen_base,
|
||||
info->fix.smem_start, info->fix.smem_len);
|
||||
}
|
||||
|
||||
return -EINVAL;
|
||||
@ -428,8 +427,7 @@ static int ep93xxfb_alloc_videomem(struct fb_info *info)
|
||||
/* Maximum 16bpp -> used memory is maximum x*y*2 bytes */
|
||||
fb_size = EP93XXFB_MAX_XRES * EP93XXFB_MAX_YRES * 2;
|
||||
|
||||
virt_addr = dma_alloc_writecombine(info->dev, fb_size,
|
||||
&phys_addr, GFP_KERNEL);
|
||||
virt_addr = dma_alloc_wc(info->dev, fb_size, &phys_addr, GFP_KERNEL);
|
||||
if (!virt_addr)
|
||||
return -ENOMEM;
|
||||
|
||||
|
@ -1185,8 +1185,8 @@ static int gbefb_probe(struct platform_device *p_dev)
|
||||
} else {
|
||||
/* try to allocate memory with the classical allocator
|
||||
* this has high chance to fail on low memory machines */
|
||||
gbe_mem = dma_alloc_writecombine(NULL, gbe_mem_size,
|
||||
&gbe_dma_addr, GFP_KERNEL);
|
||||
gbe_mem = dma_alloc_wc(NULL, gbe_mem_size, &gbe_dma_addr,
|
||||
GFP_KERNEL);
|
||||
if (!gbe_mem) {
|
||||
printk(KERN_ERR "gbefb: couldn't allocate framebuffer memory\n");
|
||||
ret = -ENOMEM;
|
||||
@ -1238,7 +1238,7 @@ static int gbefb_probe(struct platform_device *p_dev)
|
||||
out_gbe_unmap:
|
||||
arch_phys_wc_del(par->wc_cookie);
|
||||
if (gbe_dma_addr)
|
||||
dma_free_writecombine(NULL, gbe_mem_size, gbe_mem, gbe_mem_phys);
|
||||
dma_free_wc(NULL, gbe_mem_size, gbe_mem, gbe_mem_phys);
|
||||
out_tiles_free:
|
||||
dma_free_coherent(NULL, GBE_TLB_SIZE * sizeof(uint16_t),
|
||||
(void *)gbe_tiles.cpu, gbe_tiles.dma);
|
||||
@ -1259,7 +1259,7 @@ static int gbefb_remove(struct platform_device* p_dev)
|
||||
gbe_turn_off();
|
||||
arch_phys_wc_del(par->wc_cookie);
|
||||
if (gbe_dma_addr)
|
||||
dma_free_writecombine(NULL, gbe_mem_size, gbe_mem, gbe_mem_phys);
|
||||
dma_free_wc(NULL, gbe_mem_size, gbe_mem, gbe_mem_phys);
|
||||
dma_free_coherent(NULL, GBE_TLB_SIZE * sizeof(uint16_t),
|
||||
(void *)gbe_tiles.cpu, gbe_tiles.dma);
|
||||
release_mem_region(GBE_BASE, sizeof(struct sgi_gbe));
|
||||
|
@ -937,8 +937,8 @@ static int imxfb_probe(struct platform_device *pdev)
|
||||
}
|
||||
|
||||
fbi->map_size = PAGE_ALIGN(info->fix.smem_len);
|
||||
info->screen_base = dma_alloc_writecombine(&pdev->dev, fbi->map_size,
|
||||
&fbi->map_dma, GFP_KERNEL);
|
||||
info->screen_base = dma_alloc_wc(&pdev->dev, fbi->map_size,
|
||||
&fbi->map_dma, GFP_KERNEL);
|
||||
|
||||
if (!info->screen_base) {
|
||||
dev_err(&pdev->dev, "Failed to allocate video RAM: %d\n", ret);
|
||||
@ -1005,8 +1005,8 @@ failed_cmap:
|
||||
if (pdata && pdata->exit)
|
||||
pdata->exit(fbi->pdev);
|
||||
failed_platform_init:
|
||||
dma_free_writecombine(&pdev->dev, fbi->map_size, info->screen_base,
|
||||
fbi->map_dma);
|
||||
dma_free_wc(&pdev->dev, fbi->map_size, info->screen_base,
|
||||
fbi->map_dma);
|
||||
failed_map:
|
||||
iounmap(fbi->regs);
|
||||
failed_ioremap:
|
||||
@ -1041,8 +1041,8 @@ static int imxfb_remove(struct platform_device *pdev)
|
||||
kfree(info->pseudo_palette);
|
||||
framebuffer_release(info);
|
||||
|
||||
dma_free_writecombine(&pdev->dev, fbi->map_size, info->screen_base,
|
||||
fbi->map_dma);
|
||||
dma_free_wc(&pdev->dev, fbi->map_size, info->screen_base,
|
||||
fbi->map_dma);
|
||||
|
||||
iounmap(fbi->regs);
|
||||
release_mem_region(res->start, resource_size(res));
|
||||
|
@ -1336,9 +1336,8 @@ static int mx3fb_map_video_memory(struct fb_info *fbi, unsigned int mem_len,
|
||||
int retval = 0;
|
||||
dma_addr_t addr;
|
||||
|
||||
fbi->screen_base = dma_alloc_writecombine(fbi->device,
|
||||
mem_len,
|
||||
&addr, GFP_DMA | GFP_KERNEL);
|
||||
fbi->screen_base = dma_alloc_wc(fbi->device, mem_len, &addr,
|
||||
GFP_DMA | GFP_KERNEL);
|
||||
|
||||
if (!fbi->screen_base) {
|
||||
dev_err(fbi->device, "Cannot allocate %u bytes framebuffer memory\n",
|
||||
@ -1378,8 +1377,8 @@ err0:
|
||||
*/
|
||||
static int mx3fb_unmap_video_memory(struct fb_info *fbi)
|
||||
{
|
||||
dma_free_writecombine(fbi->device, fbi->fix.smem_len,
|
||||
fbi->screen_base, fbi->fix.smem_start);
|
||||
dma_free_wc(fbi->device, fbi->fix.smem_len, fbi->screen_base,
|
||||
fbi->fix.smem_start);
|
||||
|
||||
fbi->screen_base = NULL;
|
||||
mutex_lock(&fbi->mm_lock);
|
||||
|
@ -396,8 +396,8 @@ static int nuc900fb_map_video_memory(struct fb_info *info)
|
||||
dev_dbg(fbi->dev, "nuc900fb_map_video_memory(fbi=%p) map_size %lu\n",
|
||||
fbi, map_size);
|
||||
|
||||
info->screen_base = dma_alloc_writecombine(fbi->dev, map_size,
|
||||
&map_dma, GFP_KERNEL);
|
||||
info->screen_base = dma_alloc_wc(fbi->dev, map_size, &map_dma,
|
||||
GFP_KERNEL);
|
||||
|
||||
if (!info->screen_base)
|
||||
return -ENOMEM;
|
||||
@ -411,8 +411,8 @@ static int nuc900fb_map_video_memory(struct fb_info *info)
|
||||
static inline void nuc900fb_unmap_video_memory(struct fb_info *info)
|
||||
{
|
||||
struct nuc900fb_info *fbi = info->par;
|
||||
dma_free_writecombine(fbi->dev, PAGE_ALIGN(info->fix.smem_len),
|
||||
info->screen_base, info->fix.smem_start);
|
||||
dma_free_wc(fbi->dev, PAGE_ALIGN(info->fix.smem_len),
|
||||
info->screen_base, info->fix.smem_start);
|
||||
}
|
||||
|
||||
static irqreturn_t nuc900fb_irqhandler(int irq, void *dev_id)
|
||||
|
@ -612,8 +612,8 @@ static void lcdc_dma_handler(u16 status, void *data)
|
||||
|
||||
static int alloc_palette_ram(void)
|
||||
{
|
||||
lcdc.palette_virt = dma_alloc_writecombine(lcdc.fbdev->dev,
|
||||
MAX_PALETTE_SIZE, &lcdc.palette_phys, GFP_KERNEL);
|
||||
lcdc.palette_virt = dma_alloc_wc(lcdc.fbdev->dev, MAX_PALETTE_SIZE,
|
||||
&lcdc.palette_phys, GFP_KERNEL);
|
||||
if (lcdc.palette_virt == NULL) {
|
||||
dev_err(lcdc.fbdev->dev, "failed to alloc palette memory\n");
|
||||
return -ENOMEM;
|
||||
@ -625,8 +625,8 @@ static int alloc_palette_ram(void)
|
||||
|
||||
static void free_palette_ram(void)
|
||||
{
|
||||
dma_free_writecombine(lcdc.fbdev->dev, MAX_PALETTE_SIZE,
|
||||
lcdc.palette_virt, lcdc.palette_phys);
|
||||
dma_free_wc(lcdc.fbdev->dev, MAX_PALETTE_SIZE, lcdc.palette_virt,
|
||||
lcdc.palette_phys);
|
||||
}
|
||||
|
||||
static int alloc_fbmem(struct omapfb_mem_region *region)
|
||||
@ -642,8 +642,8 @@ static int alloc_fbmem(struct omapfb_mem_region *region)
|
||||
if (region->size > frame_size)
|
||||
frame_size = region->size;
|
||||
lcdc.vram_size = frame_size;
|
||||
lcdc.vram_virt = dma_alloc_writecombine(lcdc.fbdev->dev,
|
||||
lcdc.vram_size, &lcdc.vram_phys, GFP_KERNEL);
|
||||
lcdc.vram_virt = dma_alloc_wc(lcdc.fbdev->dev, lcdc.vram_size,
|
||||
&lcdc.vram_phys, GFP_KERNEL);
|
||||
if (lcdc.vram_virt == NULL) {
|
||||
dev_err(lcdc.fbdev->dev, "unable to allocate FB DMA memory\n");
|
||||
return -ENOMEM;
|
||||
@ -660,8 +660,8 @@ static int alloc_fbmem(struct omapfb_mem_region *region)
|
||||
|
||||
static void free_fbmem(void)
|
||||
{
|
||||
dma_free_writecombine(lcdc.fbdev->dev, lcdc.vram_size,
|
||||
lcdc.vram_virt, lcdc.vram_phys);
|
||||
dma_free_wc(lcdc.fbdev->dev, lcdc.vram_size, lcdc.vram_virt,
|
||||
lcdc.vram_phys);
|
||||
}
|
||||
|
||||
static int setup_fbmem(struct omapfb_mem_desc *req_md)
|
||||
|
@ -680,8 +680,8 @@ static int pxa168fb_probe(struct platform_device *pdev)
|
||||
*/
|
||||
info->fix.smem_len = PAGE_ALIGN(DEFAULT_FB_SIZE);
|
||||
|
||||
info->screen_base = dma_alloc_writecombine(fbi->dev, info->fix.smem_len,
|
||||
&fbi->fb_start_dma, GFP_KERNEL);
|
||||
info->screen_base = dma_alloc_wc(fbi->dev, info->fix.smem_len,
|
||||
&fbi->fb_start_dma, GFP_KERNEL);
|
||||
if (info->screen_base == NULL) {
|
||||
ret = -ENOMEM;
|
||||
goto failed_free_info;
|
||||
@ -804,8 +804,8 @@ static int pxa168fb_remove(struct platform_device *pdev)
|
||||
|
||||
irq = platform_get_irq(pdev, 0);
|
||||
|
||||
dma_free_writecombine(fbi->dev, PAGE_ALIGN(info->fix.smem_len),
|
||||
info->screen_base, info->fix.smem_start);
|
||||
dma_free_wc(fbi->dev, PAGE_ALIGN(info->fix.smem_len),
|
||||
info->screen_base, info->fix.smem_start);
|
||||
|
||||
clk_disable(fbi->clk);
|
||||
|
||||
|
@ -2446,8 +2446,8 @@ static int pxafb_remove(struct platform_device *dev)
|
||||
|
||||
free_pages_exact(fbi->video_mem, fbi->video_mem_size);
|
||||
|
||||
dma_free_writecombine(&dev->dev, fbi->dma_buff_size,
|
||||
fbi->dma_buff, fbi->dma_buff_phys);
|
||||
dma_free_wc(&dev->dev, fbi->dma_buff_size, fbi->dma_buff,
|
||||
fbi->dma_buff_phys);
|
||||
|
||||
iounmap(fbi->mmio_base);
|
||||
|
||||
|
@ -1105,8 +1105,7 @@ static int s3c_fb_alloc_memory(struct s3c_fb *sfb, struct s3c_fb_win *win)
|
||||
|
||||
dev_dbg(sfb->dev, "want %u bytes for window\n", size);
|
||||
|
||||
fbi->screen_base = dma_alloc_writecombine(sfb->dev, size,
|
||||
&map_dma, GFP_KERNEL);
|
||||
fbi->screen_base = dma_alloc_wc(sfb->dev, size, &map_dma, GFP_KERNEL);
|
||||
if (!fbi->screen_base)
|
||||
return -ENOMEM;
|
||||
|
||||
@ -1131,8 +1130,8 @@ static void s3c_fb_free_memory(struct s3c_fb *sfb, struct s3c_fb_win *win)
|
||||
struct fb_info *fbi = win->fbinfo;
|
||||
|
||||
if (fbi->screen_base)
|
||||
dma_free_writecombine(sfb->dev, PAGE_ALIGN(fbi->fix.smem_len),
|
||||
fbi->screen_base, fbi->fix.smem_start);
|
||||
dma_free_wc(sfb->dev, PAGE_ALIGN(fbi->fix.smem_len),
|
||||
fbi->screen_base, fbi->fix.smem_start);
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -645,8 +645,8 @@ static int s3c2410fb_map_video_memory(struct fb_info *info)
|
||||
|
||||
dprintk("map_video_memory(fbi=%p) map_size %u\n", fbi, map_size);
|
||||
|
||||
info->screen_base = dma_alloc_writecombine(fbi->dev, map_size,
|
||||
&map_dma, GFP_KERNEL);
|
||||
info->screen_base = dma_alloc_wc(fbi->dev, map_size, &map_dma,
|
||||
GFP_KERNEL);
|
||||
|
||||
if (info->screen_base) {
|
||||
/* prevent initial garbage on screen */
|
||||
@ -667,8 +667,8 @@ static inline void s3c2410fb_unmap_video_memory(struct fb_info *info)
|
||||
{
|
||||
struct s3c2410fb_info *fbi = info->par;
|
||||
|
||||
dma_free_writecombine(fbi->dev, PAGE_ALIGN(info->fix.smem_len),
|
||||
info->screen_base, info->fix.smem_start);
|
||||
dma_free_wc(fbi->dev, PAGE_ALIGN(info->fix.smem_len),
|
||||
info->screen_base, info->fix.smem_start);
|
||||
}
|
||||
|
||||
static inline void modify_gpio(void __iomem *reg,
|
||||
|
@ -567,8 +567,8 @@ static int sa1100fb_mmap(struct fb_info *info,
|
||||
|
||||
if (off < info->fix.smem_len) {
|
||||
vma->vm_pgoff += 1; /* skip over the palette */
|
||||
return dma_mmap_writecombine(fbi->dev, vma, fbi->map_cpu,
|
||||
fbi->map_dma, fbi->map_size);
|
||||
return dma_mmap_wc(fbi->dev, vma, fbi->map_cpu, fbi->map_dma,
|
||||
fbi->map_size);
|
||||
}
|
||||
|
||||
vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
|
||||
@ -1099,8 +1099,8 @@ static int sa1100fb_map_video_memory(struct sa1100fb_info *fbi)
|
||||
* of the framebuffer.
|
||||
*/
|
||||
fbi->map_size = PAGE_ALIGN(fbi->fb.fix.smem_len + PAGE_SIZE);
|
||||
fbi->map_cpu = dma_alloc_writecombine(fbi->dev, fbi->map_size,
|
||||
&fbi->map_dma, GFP_KERNEL);
|
||||
fbi->map_cpu = dma_alloc_wc(fbi->dev, fbi->map_size, &fbi->map_dma,
|
||||
GFP_KERNEL);
|
||||
|
||||
if (fbi->map_cpu) {
|
||||
fbi->fb.screen_base = fbi->map_cpu + PAGE_SIZE;
|
||||
|
@ -641,31 +641,40 @@ static inline void dmam_release_declared_memory(struct device *dev)
|
||||
}
|
||||
#endif /* CONFIG_HAVE_GENERIC_DMA_COHERENT */
|
||||
|
||||
static inline void *dma_alloc_writecombine(struct device *dev, size_t size,
|
||||
dma_addr_t *dma_addr, gfp_t gfp)
|
||||
static inline void *dma_alloc_wc(struct device *dev, size_t size,
|
||||
dma_addr_t *dma_addr, gfp_t gfp)
|
||||
{
|
||||
DEFINE_DMA_ATTRS(attrs);
|
||||
dma_set_attr(DMA_ATTR_WRITE_COMBINE, &attrs);
|
||||
return dma_alloc_attrs(dev, size, dma_addr, gfp, &attrs);
|
||||
}
|
||||
#ifndef dma_alloc_writecombine
|
||||
#define dma_alloc_writecombine dma_alloc_wc
|
||||
#endif
|
||||
|
||||
static inline void dma_free_writecombine(struct device *dev, size_t size,
|
||||
void *cpu_addr, dma_addr_t dma_addr)
|
||||
static inline void dma_free_wc(struct device *dev, size_t size,
|
||||
void *cpu_addr, dma_addr_t dma_addr)
|
||||
{
|
||||
DEFINE_DMA_ATTRS(attrs);
|
||||
dma_set_attr(DMA_ATTR_WRITE_COMBINE, &attrs);
|
||||
return dma_free_attrs(dev, size, cpu_addr, dma_addr, &attrs);
|
||||
}
|
||||
#ifndef dma_free_writecombine
|
||||
#define dma_free_writecombine dma_free_wc
|
||||
#endif
|
||||
|
||||
static inline int dma_mmap_writecombine(struct device *dev,
|
||||
struct vm_area_struct *vma,
|
||||
void *cpu_addr, dma_addr_t dma_addr,
|
||||
size_t size)
|
||||
static inline int dma_mmap_wc(struct device *dev,
|
||||
struct vm_area_struct *vma,
|
||||
void *cpu_addr, dma_addr_t dma_addr,
|
||||
size_t size)
|
||||
{
|
||||
DEFINE_DMA_ATTRS(attrs);
|
||||
dma_set_attr(DMA_ATTR_WRITE_COMBINE, &attrs);
|
||||
return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, &attrs);
|
||||
}
|
||||
#ifndef dma_mmap_writecombine
|
||||
#define dma_mmap_writecombine dma_mmap_wc
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_NEED_DMA_MAP_STATE
|
||||
#define DEFINE_DMA_UNMAP_ADDR(ADDR_NAME) dma_addr_t ADDR_NAME
|
||||
|
@ -141,10 +141,8 @@ int pxa2xx_pcm_mmap(struct snd_pcm_substream *substream,
|
||||
struct vm_area_struct *vma)
|
||||
{
|
||||
struct snd_pcm_runtime *runtime = substream->runtime;
|
||||
return dma_mmap_writecombine(substream->pcm->card->dev, vma,
|
||||
runtime->dma_area,
|
||||
runtime->dma_addr,
|
||||
runtime->dma_bytes);
|
||||
return dma_mmap_wc(substream->pcm->card->dev, vma, runtime->dma_area,
|
||||
runtime->dma_addr, runtime->dma_bytes);
|
||||
}
|
||||
EXPORT_SYMBOL(pxa2xx_pcm_mmap);
|
||||
|
||||
@ -156,8 +154,7 @@ int pxa2xx_pcm_preallocate_dma_buffer(struct snd_pcm *pcm, int stream)
|
||||
buf->dev.type = SNDRV_DMA_TYPE_DEV;
|
||||
buf->dev.dev = pcm->card->dev;
|
||||
buf->private_data = NULL;
|
||||
buf->area = dma_alloc_writecombine(pcm->card->dev, size,
|
||||
&buf->addr, GFP_KERNEL);
|
||||
buf->area = dma_alloc_wc(pcm->card->dev, size, &buf->addr, GFP_KERNEL);
|
||||
if (!buf->area)
|
||||
return -ENOMEM;
|
||||
buf->bytes = size;
|
||||
@ -178,8 +175,7 @@ void pxa2xx_pcm_free_dma_buffers(struct snd_pcm *pcm)
|
||||
buf = &substream->dma_buffer;
|
||||
if (!buf->area)
|
||||
continue;
|
||||
dma_free_writecombine(pcm->card->dev, buf->bytes,
|
||||
buf->area, buf->addr);
|
||||
dma_free_wc(pcm->card->dev, buf->bytes, buf->area, buf->addr);
|
||||
buf->area = NULL;
|
||||
}
|
||||
}
|
||||
|
@ -217,8 +217,8 @@ static int snd_imx_pcm_mmap(struct snd_pcm_substream *substream,
|
||||
struct snd_pcm_runtime *runtime = substream->runtime;
|
||||
int ret;
|
||||
|
||||
ret = dma_mmap_writecombine(substream->pcm->card->dev, vma,
|
||||
runtime->dma_area, runtime->dma_addr, runtime->dma_bytes);
|
||||
ret = dma_mmap_wc(substream->pcm->card->dev, vma, runtime->dma_area,
|
||||
runtime->dma_addr, runtime->dma_bytes);
|
||||
|
||||
pr_debug("%s: ret: %d %p %pad 0x%08x\n", __func__, ret,
|
||||
runtime->dma_area,
|
||||
@ -247,8 +247,7 @@ static int imx_pcm_preallocate_dma_buffer(struct snd_pcm *pcm, int stream)
|
||||
buf->dev.type = SNDRV_DMA_TYPE_DEV;
|
||||
buf->dev.dev = pcm->card->dev;
|
||||
buf->private_data = NULL;
|
||||
buf->area = dma_alloc_writecombine(pcm->card->dev, size,
|
||||
&buf->addr, GFP_KERNEL);
|
||||
buf->area = dma_alloc_wc(pcm->card->dev, size, &buf->addr, GFP_KERNEL);
|
||||
if (!buf->area)
|
||||
return -ENOMEM;
|
||||
buf->bytes = size;
|
||||
@ -330,8 +329,7 @@ static void imx_pcm_free(struct snd_pcm *pcm)
|
||||
if (!buf->area)
|
||||
continue;
|
||||
|
||||
dma_free_writecombine(pcm->card->dev, buf->bytes,
|
||||
buf->area, buf->addr);
|
||||
dma_free_wc(pcm->card->dev, buf->bytes, buf->area, buf->addr);
|
||||
buf->area = NULL;
|
||||
}
|
||||
}
|
||||
|
@ -267,10 +267,8 @@ static int nuc900_dma_mmap(struct snd_pcm_substream *substream,
|
||||
{
|
||||
struct snd_pcm_runtime *runtime = substream->runtime;
|
||||
|
||||
return dma_mmap_writecombine(substream->pcm->card->dev, vma,
|
||||
runtime->dma_area,
|
||||
runtime->dma_addr,
|
||||
runtime->dma_bytes);
|
||||
return dma_mmap_wc(substream->pcm->card->dev, vma, runtime->dma_area,
|
||||
runtime->dma_addr, runtime->dma_bytes);
|
||||
}
|
||||
|
||||
static struct snd_pcm_ops nuc900_dma_ops = {
|
||||
|
@ -156,10 +156,8 @@ static int omap_pcm_mmap(struct snd_pcm_substream *substream,
|
||||
{
|
||||
struct snd_pcm_runtime *runtime = substream->runtime;
|
||||
|
||||
return dma_mmap_writecombine(substream->pcm->card->dev, vma,
|
||||
runtime->dma_area,
|
||||
runtime->dma_addr,
|
||||
runtime->dma_bytes);
|
||||
return dma_mmap_wc(substream->pcm->card->dev, vma, runtime->dma_area,
|
||||
runtime->dma_addr, runtime->dma_bytes);
|
||||
}
|
||||
|
||||
static struct snd_pcm_ops omap_pcm_ops = {
|
||||
@ -183,8 +181,7 @@ static int omap_pcm_preallocate_dma_buffer(struct snd_pcm *pcm,
|
||||
buf->dev.type = SNDRV_DMA_TYPE_DEV;
|
||||
buf->dev.dev = pcm->card->dev;
|
||||
buf->private_data = NULL;
|
||||
buf->area = dma_alloc_writecombine(pcm->card->dev, size,
|
||||
&buf->addr, GFP_KERNEL);
|
||||
buf->area = dma_alloc_wc(pcm->card->dev, size, &buf->addr, GFP_KERNEL);
|
||||
if (!buf->area)
|
||||
return -ENOMEM;
|
||||
|
||||
@ -207,8 +204,7 @@ static void omap_pcm_free_dma_buffers(struct snd_pcm *pcm)
|
||||
if (!buf->area)
|
||||
continue;
|
||||
|
||||
dma_free_writecombine(pcm->card->dev, buf->bytes,
|
||||
buf->area, buf->addr);
|
||||
dma_free_wc(pcm->card->dev, buf->bytes, buf->area, buf->addr);
|
||||
buf->area = NULL;
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user