mirror of
https://github.com/torvalds/linux.git
synced 2024-11-15 08:31:55 +00:00
5a0e3ad6af
percpu.h is included by sched.h and module.h and thus ends up being included when building most .c files. percpu.h includes slab.h which in turn includes gfp.h making everything defined by the two files universally available and complicating inclusion dependencies. percpu.h -> slab.h dependency is about to be removed. Prepare for this change by updating users of gfp and slab facilities include those headers directly instead of assuming availability. As this conversion needs to touch large number of source files, the following script is used as the basis of conversion. http://userweb.kernel.org/~tj/misc/slabh-sweep.py The script does the followings. * Scan files for gfp and slab usages and update includes such that only the necessary includes are there. ie. if only gfp is used, gfp.h, if slab is used, slab.h. * When the script inserts a new include, it looks at the include blocks and try to put the new include such that its order conforms to its surrounding. It's put in the include block which contains core kernel includes, in the same order that the rest are ordered - alphabetical, Christmas tree, rev-Xmas-tree or at the end if there doesn't seem to be any matching order. * If the script can't find a place to put a new include (mostly because the file doesn't have fitting include block), it prints out an error message indicating which .h file needs to be added to the file. The conversion was done in the following steps. 1. The initial automatic conversion of all .c files updated slightly over 4000 files, deleting around 700 includes and adding ~480 gfp.h and ~3000 slab.h inclusions. The script emitted errors for ~400 files. 2. Each error was manually checked. Some didn't need the inclusion, some needed manual addition while adding it to implementation .h or embedding .c file was more appropriate for others. This step added inclusions to around 150 files. 3. The script was run again and the output was compared to the edits from #2 to make sure no file was left behind. 4. Several build tests were done and a couple of problems were fixed. e.g. lib/decompress_*.c used malloc/free() wrappers around slab APIs requiring slab.h to be added manually. 5. The script was run on all .h files but without automatically editing them as sprinkling gfp.h and slab.h inclusions around .h files could easily lead to inclusion dependency hell. Most gfp.h inclusion directives were ignored as stuff from gfp.h was usually wildly available and often used in preprocessor macros. Each slab.h inclusion directive was examined and added manually as necessary. 6. percpu.h was updated not to include slab.h. 7. Build test were done on the following configurations and failures were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my distributed build env didn't work with gcov compiles) and a few more options had to be turned off depending on archs to make things build (like ipr on powerpc/64 which failed due to missing writeq). * x86 and x86_64 UP and SMP allmodconfig and a custom test config. * powerpc and powerpc64 SMP allmodconfig * sparc and sparc64 SMP allmodconfig * ia64 SMP allmodconfig * s390 SMP allmodconfig * alpha SMP allmodconfig * um on x86_64 SMP allmodconfig 8. percpu.h modifications were reverted so that it could be applied as a separate patch and serve as bisection point. Given the fact that I had only a couple of failures from tests on step 6, I'm fairly confident about the coverage of this conversion patch. If there is a breakage, it's likely to be something in one of the arch headers which should be easily discoverable easily on most builds of the specific arch. Signed-off-by: Tejun Heo <tj@kernel.org> Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org> Cc: Ingo Molnar <mingo@redhat.com> Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
854 lines
20 KiB
C
854 lines
20 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (c) 2005 Silicon Graphics, Inc. All rights reserved.
|
|
*/
|
|
|
|
/*
|
|
* MOATB Core Services driver.
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/module.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/types.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/notifier.h>
|
|
#include <linux/reboot.h>
|
|
#include <linux/init.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/device.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/uio.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/smp_lock.h>
|
|
#include <linux/slab.h>
|
|
#include <asm/io.h>
|
|
#include <asm/uaccess.h>
|
|
#include <asm/system.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/sn/addrs.h>
|
|
#include <asm/sn/intr.h>
|
|
#include <asm/sn/tiocx.h>
|
|
#include "mbcs.h"
|
|
|
|
#define MBCS_DEBUG 0
|
|
#if MBCS_DEBUG
|
|
#define DBG(fmt...) printk(KERN_ALERT fmt)
|
|
#else
|
|
#define DBG(fmt...)
|
|
#endif
|
|
static int mbcs_major;
|
|
|
|
static LIST_HEAD(soft_list);
|
|
|
|
/*
|
|
* file operations
|
|
*/
|
|
static const struct file_operations mbcs_ops = {
|
|
.open = mbcs_open,
|
|
.llseek = mbcs_sram_llseek,
|
|
.read = mbcs_sram_read,
|
|
.write = mbcs_sram_write,
|
|
.mmap = mbcs_gscr_mmap,
|
|
};
|
|
|
|
struct mbcs_callback_arg {
|
|
int minor;
|
|
struct cx_dev *cx_dev;
|
|
};
|
|
|
|
static inline void mbcs_getdma_init(struct getdma *gdma)
|
|
{
|
|
memset(gdma, 0, sizeof(struct getdma));
|
|
gdma->DoneIntEnable = 1;
|
|
}
|
|
|
|
static inline void mbcs_putdma_init(struct putdma *pdma)
|
|
{
|
|
memset(pdma, 0, sizeof(struct putdma));
|
|
pdma->DoneIntEnable = 1;
|
|
}
|
|
|
|
static inline void mbcs_algo_init(struct algoblock *algo_soft)
|
|
{
|
|
memset(algo_soft, 0, sizeof(struct algoblock));
|
|
}
|
|
|
|
static inline void mbcs_getdma_set(void *mmr,
|
|
uint64_t hostAddr,
|
|
uint64_t localAddr,
|
|
uint64_t localRamSel,
|
|
uint64_t numPkts,
|
|
uint64_t amoEnable,
|
|
uint64_t intrEnable,
|
|
uint64_t peerIO,
|
|
uint64_t amoHostDest,
|
|
uint64_t amoModType, uint64_t intrHostDest,
|
|
uint64_t intrVector)
|
|
{
|
|
union dma_control rdma_control;
|
|
union dma_amo_dest amo_dest;
|
|
union intr_dest intr_dest;
|
|
union dma_localaddr local_addr;
|
|
union dma_hostaddr host_addr;
|
|
|
|
rdma_control.dma_control_reg = 0;
|
|
amo_dest.dma_amo_dest_reg = 0;
|
|
intr_dest.intr_dest_reg = 0;
|
|
local_addr.dma_localaddr_reg = 0;
|
|
host_addr.dma_hostaddr_reg = 0;
|
|
|
|
host_addr.dma_sys_addr = hostAddr;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_SYS_ADDR, host_addr.dma_hostaddr_reg);
|
|
|
|
local_addr.dma_ram_addr = localAddr;
|
|
local_addr.dma_ram_sel = localRamSel;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_LOC_ADDR, local_addr.dma_localaddr_reg);
|
|
|
|
rdma_control.dma_op_length = numPkts;
|
|
rdma_control.done_amo_en = amoEnable;
|
|
rdma_control.done_int_en = intrEnable;
|
|
rdma_control.pio_mem_n = peerIO;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_CTRL, rdma_control.dma_control_reg);
|
|
|
|
amo_dest.dma_amo_sys_addr = amoHostDest;
|
|
amo_dest.dma_amo_mod_type = amoModType;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_AMO_DEST, amo_dest.dma_amo_dest_reg);
|
|
|
|
intr_dest.address = intrHostDest;
|
|
intr_dest.int_vector = intrVector;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_INT_DEST, intr_dest.intr_dest_reg);
|
|
|
|
}
|
|
|
|
static inline void mbcs_putdma_set(void *mmr,
|
|
uint64_t hostAddr,
|
|
uint64_t localAddr,
|
|
uint64_t localRamSel,
|
|
uint64_t numPkts,
|
|
uint64_t amoEnable,
|
|
uint64_t intrEnable,
|
|
uint64_t peerIO,
|
|
uint64_t amoHostDest,
|
|
uint64_t amoModType,
|
|
uint64_t intrHostDest, uint64_t intrVector)
|
|
{
|
|
union dma_control wdma_control;
|
|
union dma_amo_dest amo_dest;
|
|
union intr_dest intr_dest;
|
|
union dma_localaddr local_addr;
|
|
union dma_hostaddr host_addr;
|
|
|
|
wdma_control.dma_control_reg = 0;
|
|
amo_dest.dma_amo_dest_reg = 0;
|
|
intr_dest.intr_dest_reg = 0;
|
|
local_addr.dma_localaddr_reg = 0;
|
|
host_addr.dma_hostaddr_reg = 0;
|
|
|
|
host_addr.dma_sys_addr = hostAddr;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_SYS_ADDR, host_addr.dma_hostaddr_reg);
|
|
|
|
local_addr.dma_ram_addr = localAddr;
|
|
local_addr.dma_ram_sel = localRamSel;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_LOC_ADDR, local_addr.dma_localaddr_reg);
|
|
|
|
wdma_control.dma_op_length = numPkts;
|
|
wdma_control.done_amo_en = amoEnable;
|
|
wdma_control.done_int_en = intrEnable;
|
|
wdma_control.pio_mem_n = peerIO;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_CTRL, wdma_control.dma_control_reg);
|
|
|
|
amo_dest.dma_amo_sys_addr = amoHostDest;
|
|
amo_dest.dma_amo_mod_type = amoModType;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_AMO_DEST, amo_dest.dma_amo_dest_reg);
|
|
|
|
intr_dest.address = intrHostDest;
|
|
intr_dest.int_vector = intrVector;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_INT_DEST, intr_dest.intr_dest_reg);
|
|
|
|
}
|
|
|
|
static inline void mbcs_algo_set(void *mmr,
|
|
uint64_t amoHostDest,
|
|
uint64_t amoModType,
|
|
uint64_t intrHostDest,
|
|
uint64_t intrVector, uint64_t algoStepCount)
|
|
{
|
|
union dma_amo_dest amo_dest;
|
|
union intr_dest intr_dest;
|
|
union algo_step step;
|
|
|
|
step.algo_step_reg = 0;
|
|
intr_dest.intr_dest_reg = 0;
|
|
amo_dest.dma_amo_dest_reg = 0;
|
|
|
|
amo_dest.dma_amo_sys_addr = amoHostDest;
|
|
amo_dest.dma_amo_mod_type = amoModType;
|
|
MBCS_MMR_SET(mmr, MBCS_ALG_AMO_DEST, amo_dest.dma_amo_dest_reg);
|
|
|
|
intr_dest.address = intrHostDest;
|
|
intr_dest.int_vector = intrVector;
|
|
MBCS_MMR_SET(mmr, MBCS_ALG_INT_DEST, intr_dest.intr_dest_reg);
|
|
|
|
step.alg_step_cnt = algoStepCount;
|
|
MBCS_MMR_SET(mmr, MBCS_ALG_STEP, step.algo_step_reg);
|
|
}
|
|
|
|
static inline int mbcs_getdma_start(struct mbcs_soft *soft)
|
|
{
|
|
void *mmr_base;
|
|
struct getdma *gdma;
|
|
uint64_t numPkts;
|
|
union cm_control cm_control;
|
|
|
|
mmr_base = soft->mmr_base;
|
|
gdma = &soft->getdma;
|
|
|
|
/* check that host address got setup */
|
|
if (!gdma->hostAddr)
|
|
return -1;
|
|
|
|
numPkts =
|
|
(gdma->bytes + (MBCS_CACHELINE_SIZE - 1)) / MBCS_CACHELINE_SIZE;
|
|
|
|
/* program engine */
|
|
mbcs_getdma_set(mmr_base, tiocx_dma_addr(gdma->hostAddr),
|
|
gdma->localAddr,
|
|
(gdma->localAddr < MB2) ? 0 :
|
|
(gdma->localAddr < MB4) ? 1 :
|
|
(gdma->localAddr < MB6) ? 2 : 3,
|
|
numPkts,
|
|
gdma->DoneAmoEnable,
|
|
gdma->DoneIntEnable,
|
|
gdma->peerIO,
|
|
gdma->amoHostDest,
|
|
gdma->amoModType,
|
|
gdma->intrHostDest, gdma->intrVector);
|
|
|
|
/* start engine */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.rd_dma_go = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
static inline int mbcs_putdma_start(struct mbcs_soft *soft)
|
|
{
|
|
void *mmr_base;
|
|
struct putdma *pdma;
|
|
uint64_t numPkts;
|
|
union cm_control cm_control;
|
|
|
|
mmr_base = soft->mmr_base;
|
|
pdma = &soft->putdma;
|
|
|
|
/* check that host address got setup */
|
|
if (!pdma->hostAddr)
|
|
return -1;
|
|
|
|
numPkts =
|
|
(pdma->bytes + (MBCS_CACHELINE_SIZE - 1)) / MBCS_CACHELINE_SIZE;
|
|
|
|
/* program engine */
|
|
mbcs_putdma_set(mmr_base, tiocx_dma_addr(pdma->hostAddr),
|
|
pdma->localAddr,
|
|
(pdma->localAddr < MB2) ? 0 :
|
|
(pdma->localAddr < MB4) ? 1 :
|
|
(pdma->localAddr < MB6) ? 2 : 3,
|
|
numPkts,
|
|
pdma->DoneAmoEnable,
|
|
pdma->DoneIntEnable,
|
|
pdma->peerIO,
|
|
pdma->amoHostDest,
|
|
pdma->amoModType,
|
|
pdma->intrHostDest, pdma->intrVector);
|
|
|
|
/* start engine */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.wr_dma_go = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
static inline int mbcs_algo_start(struct mbcs_soft *soft)
|
|
{
|
|
struct algoblock *algo_soft = &soft->algo;
|
|
void *mmr_base = soft->mmr_base;
|
|
union cm_control cm_control;
|
|
|
|
if (mutex_lock_interruptible(&soft->algolock))
|
|
return -ERESTARTSYS;
|
|
|
|
atomic_set(&soft->algo_done, 0);
|
|
|
|
mbcs_algo_set(mmr_base,
|
|
algo_soft->amoHostDest,
|
|
algo_soft->amoModType,
|
|
algo_soft->intrHostDest,
|
|
algo_soft->intrVector, algo_soft->algoStepCount);
|
|
|
|
/* start algorithm */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.alg_done_int_en = 1;
|
|
cm_control.alg_go = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
mutex_unlock(&soft->algolock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline ssize_t
|
|
do_mbcs_sram_dmawrite(struct mbcs_soft *soft, uint64_t hostAddr,
|
|
size_t len, loff_t * off)
|
|
{
|
|
int rv = 0;
|
|
|
|
if (mutex_lock_interruptible(&soft->dmawritelock))
|
|
return -ERESTARTSYS;
|
|
|
|
atomic_set(&soft->dmawrite_done, 0);
|
|
|
|
soft->putdma.hostAddr = hostAddr;
|
|
soft->putdma.localAddr = *off;
|
|
soft->putdma.bytes = len;
|
|
|
|
if (mbcs_putdma_start(soft) < 0) {
|
|
DBG(KERN_ALERT "do_mbcs_sram_dmawrite: "
|
|
"mbcs_putdma_start failed\n");
|
|
rv = -EAGAIN;
|
|
goto dmawrite_exit;
|
|
}
|
|
|
|
if (wait_event_interruptible(soft->dmawrite_queue,
|
|
atomic_read(&soft->dmawrite_done))) {
|
|
rv = -ERESTARTSYS;
|
|
goto dmawrite_exit;
|
|
}
|
|
|
|
rv = len;
|
|
*off += len;
|
|
|
|
dmawrite_exit:
|
|
mutex_unlock(&soft->dmawritelock);
|
|
|
|
return rv;
|
|
}
|
|
|
|
static inline ssize_t
|
|
do_mbcs_sram_dmaread(struct mbcs_soft *soft, uint64_t hostAddr,
|
|
size_t len, loff_t * off)
|
|
{
|
|
int rv = 0;
|
|
|
|
if (mutex_lock_interruptible(&soft->dmareadlock))
|
|
return -ERESTARTSYS;
|
|
|
|
atomic_set(&soft->dmawrite_done, 0);
|
|
|
|
soft->getdma.hostAddr = hostAddr;
|
|
soft->getdma.localAddr = *off;
|
|
soft->getdma.bytes = len;
|
|
|
|
if (mbcs_getdma_start(soft) < 0) {
|
|
DBG(KERN_ALERT "mbcs_strategy: mbcs_getdma_start failed\n");
|
|
rv = -EAGAIN;
|
|
goto dmaread_exit;
|
|
}
|
|
|
|
if (wait_event_interruptible(soft->dmaread_queue,
|
|
atomic_read(&soft->dmaread_done))) {
|
|
rv = -ERESTARTSYS;
|
|
goto dmaread_exit;
|
|
}
|
|
|
|
rv = len;
|
|
*off += len;
|
|
|
|
dmaread_exit:
|
|
mutex_unlock(&soft->dmareadlock);
|
|
|
|
return rv;
|
|
}
|
|
|
|
static int mbcs_open(struct inode *ip, struct file *fp)
|
|
{
|
|
struct mbcs_soft *soft;
|
|
int minor;
|
|
|
|
lock_kernel();
|
|
minor = iminor(ip);
|
|
|
|
/* Nothing protects access to this list... */
|
|
list_for_each_entry(soft, &soft_list, list) {
|
|
if (soft->nasid == minor) {
|
|
fp->private_data = soft->cxdev;
|
|
unlock_kernel();
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
unlock_kernel();
|
|
return -ENODEV;
|
|
}
|
|
|
|
static ssize_t mbcs_sram_read(struct file * fp, char __user *buf, size_t len, loff_t * off)
|
|
{
|
|
struct cx_dev *cx_dev = fp->private_data;
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
uint64_t hostAddr;
|
|
int rv = 0;
|
|
|
|
hostAddr = __get_dma_pages(GFP_KERNEL, get_order(len));
|
|
if (hostAddr == 0)
|
|
return -ENOMEM;
|
|
|
|
rv = do_mbcs_sram_dmawrite(soft, hostAddr, len, off);
|
|
if (rv < 0)
|
|
goto exit;
|
|
|
|
if (copy_to_user(buf, (void *)hostAddr, len))
|
|
rv = -EFAULT;
|
|
|
|
exit:
|
|
free_pages(hostAddr, get_order(len));
|
|
|
|
return rv;
|
|
}
|
|
|
|
static ssize_t
|
|
mbcs_sram_write(struct file * fp, const char __user *buf, size_t len, loff_t * off)
|
|
{
|
|
struct cx_dev *cx_dev = fp->private_data;
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
uint64_t hostAddr;
|
|
int rv = 0;
|
|
|
|
hostAddr = __get_dma_pages(GFP_KERNEL, get_order(len));
|
|
if (hostAddr == 0)
|
|
return -ENOMEM;
|
|
|
|
if (copy_from_user((void *)hostAddr, buf, len)) {
|
|
rv = -EFAULT;
|
|
goto exit;
|
|
}
|
|
|
|
rv = do_mbcs_sram_dmaread(soft, hostAddr, len, off);
|
|
|
|
exit:
|
|
free_pages(hostAddr, get_order(len));
|
|
|
|
return rv;
|
|
}
|
|
|
|
static loff_t mbcs_sram_llseek(struct file * filp, loff_t off, int whence)
|
|
{
|
|
loff_t newpos;
|
|
|
|
switch (whence) {
|
|
case SEEK_SET:
|
|
newpos = off;
|
|
break;
|
|
|
|
case SEEK_CUR:
|
|
newpos = filp->f_pos + off;
|
|
break;
|
|
|
|
case SEEK_END:
|
|
newpos = MBCS_SRAM_SIZE + off;
|
|
break;
|
|
|
|
default: /* can't happen */
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (newpos < 0)
|
|
return -EINVAL;
|
|
|
|
filp->f_pos = newpos;
|
|
|
|
return newpos;
|
|
}
|
|
|
|
static uint64_t mbcs_pioaddr(struct mbcs_soft *soft, uint64_t offset)
|
|
{
|
|
uint64_t mmr_base;
|
|
|
|
mmr_base = (uint64_t) (soft->mmr_base + offset);
|
|
|
|
return mmr_base;
|
|
}
|
|
|
|
static void mbcs_debug_pioaddr_set(struct mbcs_soft *soft)
|
|
{
|
|
soft->debug_addr = mbcs_pioaddr(soft, MBCS_DEBUG_START);
|
|
}
|
|
|
|
static void mbcs_gscr_pioaddr_set(struct mbcs_soft *soft)
|
|
{
|
|
soft->gscr_addr = mbcs_pioaddr(soft, MBCS_GSCR_START);
|
|
}
|
|
|
|
static int mbcs_gscr_mmap(struct file *fp, struct vm_area_struct *vma)
|
|
{
|
|
struct cx_dev *cx_dev = fp->private_data;
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
|
|
if (vma->vm_pgoff != 0)
|
|
return -EINVAL;
|
|
|
|
vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
|
|
|
|
/* Remap-pfn-range will mark the range VM_IO and VM_RESERVED */
|
|
if (remap_pfn_range(vma,
|
|
vma->vm_start,
|
|
__pa(soft->gscr_addr) >> PAGE_SHIFT,
|
|
PAGE_SIZE,
|
|
vma->vm_page_prot))
|
|
return -EAGAIN;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mbcs_completion_intr_handler - Primary completion handler.
|
|
* @irq: irq
|
|
* @arg: soft struct for device
|
|
*
|
|
*/
|
|
static irqreturn_t
|
|
mbcs_completion_intr_handler(int irq, void *arg)
|
|
{
|
|
struct mbcs_soft *soft = (struct mbcs_soft *)arg;
|
|
void *mmr_base;
|
|
union cm_status cm_status;
|
|
union cm_control cm_control;
|
|
|
|
mmr_base = soft->mmr_base;
|
|
cm_status.cm_status_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_STATUS);
|
|
|
|
if (cm_status.rd_dma_done) {
|
|
/* stop dma-read engine, clear status */
|
|
cm_control.cm_control_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.rd_dma_clr = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL,
|
|
cm_control.cm_control_reg);
|
|
atomic_set(&soft->dmaread_done, 1);
|
|
wake_up(&soft->dmaread_queue);
|
|
}
|
|
if (cm_status.wr_dma_done) {
|
|
/* stop dma-write engine, clear status */
|
|
cm_control.cm_control_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.wr_dma_clr = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL,
|
|
cm_control.cm_control_reg);
|
|
atomic_set(&soft->dmawrite_done, 1);
|
|
wake_up(&soft->dmawrite_queue);
|
|
}
|
|
if (cm_status.alg_done) {
|
|
/* clear status */
|
|
cm_control.cm_control_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.alg_done_clr = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL,
|
|
cm_control.cm_control_reg);
|
|
atomic_set(&soft->algo_done, 1);
|
|
wake_up(&soft->algo_queue);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/**
|
|
* mbcs_intr_alloc - Allocate interrupts.
|
|
* @dev: device pointer
|
|
*
|
|
*/
|
|
static int mbcs_intr_alloc(struct cx_dev *dev)
|
|
{
|
|
struct sn_irq_info *sn_irq;
|
|
struct mbcs_soft *soft;
|
|
struct getdma *getdma;
|
|
struct putdma *putdma;
|
|
struct algoblock *algo;
|
|
|
|
soft = dev->soft;
|
|
getdma = &soft->getdma;
|
|
putdma = &soft->putdma;
|
|
algo = &soft->algo;
|
|
|
|
soft->get_sn_irq = NULL;
|
|
soft->put_sn_irq = NULL;
|
|
soft->algo_sn_irq = NULL;
|
|
|
|
sn_irq = tiocx_irq_alloc(dev->cx_id.nasid, TIOCX_CORELET, -1, -1, -1);
|
|
if (sn_irq == NULL)
|
|
return -EAGAIN;
|
|
soft->get_sn_irq = sn_irq;
|
|
getdma->intrHostDest = sn_irq->irq_xtalkaddr;
|
|
getdma->intrVector = sn_irq->irq_irq;
|
|
if (request_irq(sn_irq->irq_irq,
|
|
(void *)mbcs_completion_intr_handler, IRQF_SHARED,
|
|
"MBCS get intr", (void *)soft)) {
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
sn_irq = tiocx_irq_alloc(dev->cx_id.nasid, TIOCX_CORELET, -1, -1, -1);
|
|
if (sn_irq == NULL) {
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
soft->put_sn_irq = sn_irq;
|
|
putdma->intrHostDest = sn_irq->irq_xtalkaddr;
|
|
putdma->intrVector = sn_irq->irq_irq;
|
|
if (request_irq(sn_irq->irq_irq,
|
|
(void *)mbcs_completion_intr_handler, IRQF_SHARED,
|
|
"MBCS put intr", (void *)soft)) {
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
sn_irq = tiocx_irq_alloc(dev->cx_id.nasid, TIOCX_CORELET, -1, -1, -1);
|
|
if (sn_irq == NULL) {
|
|
free_irq(soft->put_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
soft->algo_sn_irq = sn_irq;
|
|
algo->intrHostDest = sn_irq->irq_xtalkaddr;
|
|
algo->intrVector = sn_irq->irq_irq;
|
|
if (request_irq(sn_irq->irq_irq,
|
|
(void *)mbcs_completion_intr_handler, IRQF_SHARED,
|
|
"MBCS algo intr", (void *)soft)) {
|
|
tiocx_irq_free(soft->algo_sn_irq);
|
|
free_irq(soft->put_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mbcs_intr_dealloc - Remove interrupts.
|
|
* @dev: device pointer
|
|
*
|
|
*/
|
|
static void mbcs_intr_dealloc(struct cx_dev *dev)
|
|
{
|
|
struct mbcs_soft *soft;
|
|
|
|
soft = dev->soft;
|
|
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
free_irq(soft->put_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->algo_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->algo_sn_irq);
|
|
}
|
|
|
|
static inline int mbcs_hw_init(struct mbcs_soft *soft)
|
|
{
|
|
void *mmr_base = soft->mmr_base;
|
|
union cm_control cm_control;
|
|
union cm_req_timeout cm_req_timeout;
|
|
uint64_t err_stat;
|
|
|
|
cm_req_timeout.cm_req_timeout_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_REQ_TOUT);
|
|
|
|
cm_req_timeout.time_out = MBCS_CM_CONTROL_REQ_TOUT_MASK;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_REQ_TOUT,
|
|
cm_req_timeout.cm_req_timeout_reg);
|
|
|
|
mbcs_gscr_pioaddr_set(soft);
|
|
mbcs_debug_pioaddr_set(soft);
|
|
|
|
/* clear errors */
|
|
err_stat = MBCS_MMR_GET(mmr_base, MBCS_CM_ERR_STAT);
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CLR_ERR_STAT, err_stat);
|
|
MBCS_MMR_ZERO(mmr_base, MBCS_CM_ERROR_DETAIL1);
|
|
|
|
/* enable interrupts */
|
|
/* turn off 2^23 (INT_EN_PIO_REQ_ADDR_INV) */
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_ERR_INT_EN, 0x3ffffff7e00ffUL);
|
|
|
|
/* arm status regs and clear engines */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.rearm_stat_regs = 1;
|
|
cm_control.alg_clr = 1;
|
|
cm_control.wr_dma_clr = 1;
|
|
cm_control.rd_dma_clr = 1;
|
|
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static ssize_t show_algo(struct device *dev, struct device_attribute *attr, char *buf)
|
|
{
|
|
struct cx_dev *cx_dev = to_cx_dev(dev);
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
uint64_t debug0;
|
|
|
|
/*
|
|
* By convention, the first debug register contains the
|
|
* algorithm number and revision.
|
|
*/
|
|
debug0 = *(uint64_t *) soft->debug_addr;
|
|
|
|
return sprintf(buf, "0x%x 0x%x\n",
|
|
upper_32_bits(debug0), lower_32_bits(debug0));
|
|
}
|
|
|
|
static ssize_t store_algo(struct device *dev, struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
int n;
|
|
struct cx_dev *cx_dev = to_cx_dev(dev);
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
|
|
if (count <= 0)
|
|
return 0;
|
|
|
|
n = simple_strtoul(buf, NULL, 0);
|
|
|
|
if (n == 1) {
|
|
mbcs_algo_start(soft);
|
|
if (wait_event_interruptible(soft->algo_queue,
|
|
atomic_read(&soft->algo_done)))
|
|
return -ERESTARTSYS;
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
DEVICE_ATTR(algo, 0644, show_algo, store_algo);
|
|
|
|
/**
|
|
* mbcs_probe - Initialize for device
|
|
* @dev: device pointer
|
|
* @device_id: id table pointer
|
|
*
|
|
*/
|
|
static int mbcs_probe(struct cx_dev *dev, const struct cx_device_id *id)
|
|
{
|
|
struct mbcs_soft *soft;
|
|
|
|
dev->soft = NULL;
|
|
|
|
soft = kzalloc(sizeof(struct mbcs_soft), GFP_KERNEL);
|
|
if (soft == NULL)
|
|
return -ENOMEM;
|
|
|
|
soft->nasid = dev->cx_id.nasid;
|
|
list_add(&soft->list, &soft_list);
|
|
soft->mmr_base = (void *)tiocx_swin_base(dev->cx_id.nasid);
|
|
dev->soft = soft;
|
|
soft->cxdev = dev;
|
|
|
|
init_waitqueue_head(&soft->dmawrite_queue);
|
|
init_waitqueue_head(&soft->dmaread_queue);
|
|
init_waitqueue_head(&soft->algo_queue);
|
|
|
|
mutex_init(&soft->dmawritelock);
|
|
mutex_init(&soft->dmareadlock);
|
|
mutex_init(&soft->algolock);
|
|
|
|
mbcs_getdma_init(&soft->getdma);
|
|
mbcs_putdma_init(&soft->putdma);
|
|
mbcs_algo_init(&soft->algo);
|
|
|
|
mbcs_hw_init(soft);
|
|
|
|
/* Allocate interrupts */
|
|
mbcs_intr_alloc(dev);
|
|
|
|
device_create_file(&dev->dev, &dev_attr_algo);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mbcs_remove(struct cx_dev *dev)
|
|
{
|
|
if (dev->soft) {
|
|
mbcs_intr_dealloc(dev);
|
|
kfree(dev->soft);
|
|
}
|
|
|
|
device_remove_file(&dev->dev, &dev_attr_algo);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct cx_device_id __devinitdata mbcs_id_table[] = {
|
|
{
|
|
.part_num = MBCS_PART_NUM,
|
|
.mfg_num = MBCS_MFG_NUM,
|
|
},
|
|
{
|
|
.part_num = MBCS_PART_NUM_ALG0,
|
|
.mfg_num = MBCS_MFG_NUM,
|
|
},
|
|
{0, 0}
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(cx, mbcs_id_table);
|
|
|
|
static struct cx_drv mbcs_driver = {
|
|
.name = DEVICE_NAME,
|
|
.id_table = mbcs_id_table,
|
|
.probe = mbcs_probe,
|
|
.remove = mbcs_remove,
|
|
};
|
|
|
|
static void __exit mbcs_exit(void)
|
|
{
|
|
unregister_chrdev(mbcs_major, DEVICE_NAME);
|
|
cx_driver_unregister(&mbcs_driver);
|
|
}
|
|
|
|
static int __init mbcs_init(void)
|
|
{
|
|
int rv;
|
|
|
|
if (!ia64_platform_is("sn2"))
|
|
return -ENODEV;
|
|
|
|
// Put driver into chrdevs[]. Get major number.
|
|
rv = register_chrdev(mbcs_major, DEVICE_NAME, &mbcs_ops);
|
|
if (rv < 0) {
|
|
DBG(KERN_ALERT "mbcs_init: can't get major number. %d\n", rv);
|
|
return rv;
|
|
}
|
|
mbcs_major = rv;
|
|
|
|
return cx_driver_register(&mbcs_driver);
|
|
}
|
|
|
|
module_init(mbcs_init);
|
|
module_exit(mbcs_exit);
|
|
|
|
MODULE_AUTHOR("Bruce Losure <blosure@sgi.com>");
|
|
MODULE_DESCRIPTION("Driver for MOATB Core Services");
|
|
MODULE_LICENSE("GPL");
|