mirror of
https://github.com/torvalds/linux.git
synced 2024-11-11 22:51:42 +00:00
scsi: elx: efct: RQ buffer, memory pool allocation and deallocation APIs
Add driver definitions for: - RQ data buffer allocation and deallocate. - Memory pool allocation and deallocation APIs. - Mailbox command submission and completion routines. Link: https://lore.kernel.org/r/20210601235512.20104-21-jsmart2021@gmail.com Reviewed-by: Daniel Wagner <dwagner@suse.de> Reviewed-by: Hannes Reinecke <hare@suse.de> Co-developed-by: Ram Vegesna <ram.vegesna@broadcom.com> Signed-off-by: Ram Vegesna <ram.vegesna@broadcom.com> Signed-off-by: James Smart <jsmart2021@gmail.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
This commit is contained in:
parent
e10fc23359
commit
580c0255e4
@ -1151,3 +1151,409 @@ efct_get_wwpn(struct efct_hw *hw)
|
|||||||
memcpy(p, sli->wwpn, sizeof(p));
|
memcpy(p, sli->wwpn, sizeof(p));
|
||||||
return get_unaligned_be64(p);
|
return get_unaligned_be64(p);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static struct efc_hw_rq_buffer *
|
||||||
|
efct_hw_rx_buffer_alloc(struct efct_hw *hw, u32 rqindex, u32 count,
|
||||||
|
u32 size)
|
||||||
|
{
|
||||||
|
struct efct *efct = hw->os;
|
||||||
|
struct efc_hw_rq_buffer *rq_buf = NULL;
|
||||||
|
struct efc_hw_rq_buffer *prq;
|
||||||
|
u32 i;
|
||||||
|
|
||||||
|
if (!count)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
rq_buf = kmalloc_array(count, sizeof(*rq_buf), GFP_KERNEL);
|
||||||
|
if (!rq_buf)
|
||||||
|
return NULL;
|
||||||
|
memset(rq_buf, 0, sizeof(*rq_buf) * count);
|
||||||
|
|
||||||
|
for (i = 0, prq = rq_buf; i < count; i ++, prq++) {
|
||||||
|
prq->rqindex = rqindex;
|
||||||
|
prq->dma.size = size;
|
||||||
|
prq->dma.virt = dma_alloc_coherent(&efct->pci->dev,
|
||||||
|
prq->dma.size,
|
||||||
|
&prq->dma.phys,
|
||||||
|
GFP_DMA);
|
||||||
|
if (!prq->dma.virt) {
|
||||||
|
efc_log_err(hw->os, "DMA allocation failed\n");
|
||||||
|
kfree(rq_buf);
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return rq_buf;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
efct_hw_rx_buffer_free(struct efct_hw *hw,
|
||||||
|
struct efc_hw_rq_buffer *rq_buf,
|
||||||
|
u32 count)
|
||||||
|
{
|
||||||
|
struct efct *efct = hw->os;
|
||||||
|
u32 i;
|
||||||
|
struct efc_hw_rq_buffer *prq;
|
||||||
|
|
||||||
|
if (rq_buf) {
|
||||||
|
for (i = 0, prq = rq_buf; i < count; i++, prq++) {
|
||||||
|
dma_free_coherent(&efct->pci->dev,
|
||||||
|
prq->dma.size, prq->dma.virt,
|
||||||
|
prq->dma.phys);
|
||||||
|
memset(&prq->dma, 0, sizeof(struct efc_dma));
|
||||||
|
}
|
||||||
|
|
||||||
|
kfree(rq_buf);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
int
|
||||||
|
efct_hw_rx_allocate(struct efct_hw *hw)
|
||||||
|
{
|
||||||
|
struct efct *efct = hw->os;
|
||||||
|
u32 i;
|
||||||
|
int rc = 0;
|
||||||
|
u32 rqindex = 0;
|
||||||
|
u32 hdr_size = EFCT_HW_RQ_SIZE_HDR;
|
||||||
|
u32 payload_size = hw->config.rq_default_buffer_size;
|
||||||
|
|
||||||
|
rqindex = 0;
|
||||||
|
|
||||||
|
for (i = 0; i < hw->hw_rq_count; i++) {
|
||||||
|
struct hw_rq *rq = hw->hw_rq[i];
|
||||||
|
|
||||||
|
/* Allocate header buffers */
|
||||||
|
rq->hdr_buf = efct_hw_rx_buffer_alloc(hw, rqindex,
|
||||||
|
rq->entry_count,
|
||||||
|
hdr_size);
|
||||||
|
if (!rq->hdr_buf) {
|
||||||
|
efc_log_err(efct, "rx_buffer_alloc hdr_buf failed\n");
|
||||||
|
rc = -EIO;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
efc_log_debug(hw->os,
|
||||||
|
"rq[%2d] rq_id %02d header %4d by %4d bytes\n",
|
||||||
|
i, rq->hdr->id, rq->entry_count, hdr_size);
|
||||||
|
|
||||||
|
rqindex++;
|
||||||
|
|
||||||
|
/* Allocate payload buffers */
|
||||||
|
rq->payload_buf = efct_hw_rx_buffer_alloc(hw, rqindex,
|
||||||
|
rq->entry_count,
|
||||||
|
payload_size);
|
||||||
|
if (!rq->payload_buf) {
|
||||||
|
efc_log_err(efct, "rx_buffer_alloc fb_buf failed\n");
|
||||||
|
rc = -EIO;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
efc_log_debug(hw->os,
|
||||||
|
"rq[%2d] rq_id %02d default %4d by %4d bytes\n",
|
||||||
|
i, rq->data->id, rq->entry_count, payload_size);
|
||||||
|
rqindex++;
|
||||||
|
}
|
||||||
|
|
||||||
|
return rc ? -EIO : 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
int
|
||||||
|
efct_hw_rx_post(struct efct_hw *hw)
|
||||||
|
{
|
||||||
|
u32 i;
|
||||||
|
u32 idx;
|
||||||
|
u32 rq_idx;
|
||||||
|
int rc = 0;
|
||||||
|
|
||||||
|
if (!hw->seq_pool) {
|
||||||
|
u32 count = 0;
|
||||||
|
|
||||||
|
for (i = 0; i < hw->hw_rq_count; i++)
|
||||||
|
count += hw->hw_rq[i]->entry_count;
|
||||||
|
|
||||||
|
hw->seq_pool = kmalloc_array(count,
|
||||||
|
sizeof(struct efc_hw_sequence), GFP_KERNEL);
|
||||||
|
if (!hw->seq_pool)
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* In RQ pair mode, we MUST post the header and payload buffer at the
|
||||||
|
* same time.
|
||||||
|
*/
|
||||||
|
for (rq_idx = 0, idx = 0; rq_idx < hw->hw_rq_count; rq_idx++) {
|
||||||
|
struct hw_rq *rq = hw->hw_rq[rq_idx];
|
||||||
|
|
||||||
|
for (i = 0; i < rq->entry_count - 1; i++) {
|
||||||
|
struct efc_hw_sequence *seq;
|
||||||
|
|
||||||
|
seq = hw->seq_pool + idx;
|
||||||
|
idx++;
|
||||||
|
seq->header = &rq->hdr_buf[i];
|
||||||
|
seq->payload = &rq->payload_buf[i];
|
||||||
|
rc = efct_hw_sequence_free(hw, seq);
|
||||||
|
if (rc)
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
if (rc)
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (rc && hw->seq_pool)
|
||||||
|
kfree(hw->seq_pool);
|
||||||
|
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
void
|
||||||
|
efct_hw_rx_free(struct efct_hw *hw)
|
||||||
|
{
|
||||||
|
u32 i;
|
||||||
|
|
||||||
|
/* Free hw_rq buffers */
|
||||||
|
for (i = 0; i < hw->hw_rq_count; i++) {
|
||||||
|
struct hw_rq *rq = hw->hw_rq[i];
|
||||||
|
|
||||||
|
if (rq) {
|
||||||
|
efct_hw_rx_buffer_free(hw, rq->hdr_buf,
|
||||||
|
rq->entry_count);
|
||||||
|
rq->hdr_buf = NULL;
|
||||||
|
efct_hw_rx_buffer_free(hw, rq->payload_buf,
|
||||||
|
rq->entry_count);
|
||||||
|
rq->payload_buf = NULL;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
efct_hw_cmd_submit_pending(struct efct_hw *hw)
|
||||||
|
{
|
||||||
|
int rc = 0;
|
||||||
|
|
||||||
|
/* Assumes lock held */
|
||||||
|
|
||||||
|
/* Only submit MQE if there's room */
|
||||||
|
while (hw->cmd_head_count < (EFCT_HW_MQ_DEPTH - 1) &&
|
||||||
|
!list_empty(&hw->cmd_pending)) {
|
||||||
|
struct efct_command_ctx *ctx;
|
||||||
|
|
||||||
|
ctx = list_first_entry(&hw->cmd_pending,
|
||||||
|
struct efct_command_ctx, list_entry);
|
||||||
|
if (!ctx)
|
||||||
|
break;
|
||||||
|
|
||||||
|
list_del_init(&ctx->list_entry);
|
||||||
|
|
||||||
|
list_add_tail(&ctx->list_entry, &hw->cmd_head);
|
||||||
|
hw->cmd_head_count++;
|
||||||
|
if (sli_mq_write(&hw->sli, hw->mq, ctx->buf) < 0) {
|
||||||
|
efc_log_debug(hw->os,
|
||||||
|
"sli_queue_write failed: %d\n", rc);
|
||||||
|
rc = -EIO;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
int
|
||||||
|
efct_hw_command(struct efct_hw *hw, u8 *cmd, u32 opts, void *cb, void *arg)
|
||||||
|
{
|
||||||
|
int rc = -EIO;
|
||||||
|
unsigned long flags = 0;
|
||||||
|
void *bmbx = NULL;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If the chip is in an error state (UE'd) then reject this mailbox
|
||||||
|
* command.
|
||||||
|
*/
|
||||||
|
if (sli_fw_error_status(&hw->sli) > 0) {
|
||||||
|
efc_log_crit(hw->os, "Chip in an error state - reset needed\n");
|
||||||
|
efc_log_crit(hw->os, "status=%#x error1=%#x error2=%#x\n",
|
||||||
|
sli_reg_read_status(&hw->sli),
|
||||||
|
sli_reg_read_err1(&hw->sli),
|
||||||
|
sli_reg_read_err2(&hw->sli));
|
||||||
|
|
||||||
|
return -EIO;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Send a mailbox command to the hardware, and either wait for
|
||||||
|
* a completion (EFCT_CMD_POLL) or get an optional asynchronous
|
||||||
|
* completion (EFCT_CMD_NOWAIT).
|
||||||
|
*/
|
||||||
|
|
||||||
|
if (opts == EFCT_CMD_POLL) {
|
||||||
|
mutex_lock(&hw->bmbx_lock);
|
||||||
|
bmbx = hw->sli.bmbx.virt;
|
||||||
|
|
||||||
|
memset(bmbx, 0, SLI4_BMBX_SIZE);
|
||||||
|
memcpy(bmbx, cmd, SLI4_BMBX_SIZE);
|
||||||
|
|
||||||
|
if (sli_bmbx_command(&hw->sli) == 0) {
|
||||||
|
rc = 0;
|
||||||
|
memcpy(cmd, bmbx, SLI4_BMBX_SIZE);
|
||||||
|
}
|
||||||
|
mutex_unlock(&hw->bmbx_lock);
|
||||||
|
} else if (opts == EFCT_CMD_NOWAIT) {
|
||||||
|
struct efct_command_ctx *ctx = NULL;
|
||||||
|
|
||||||
|
if (hw->state != EFCT_HW_STATE_ACTIVE) {
|
||||||
|
efc_log_err(hw->os, "Can't send command, HW state=%d\n",
|
||||||
|
hw->state);
|
||||||
|
return -EIO;
|
||||||
|
}
|
||||||
|
|
||||||
|
ctx = mempool_alloc(hw->cmd_ctx_pool, GFP_ATOMIC);
|
||||||
|
if (!ctx)
|
||||||
|
return -ENOSPC;
|
||||||
|
|
||||||
|
memset(ctx, 0, sizeof(struct efct_command_ctx));
|
||||||
|
|
||||||
|
if (cb) {
|
||||||
|
ctx->cb = cb;
|
||||||
|
ctx->arg = arg;
|
||||||
|
}
|
||||||
|
|
||||||
|
memcpy(ctx->buf, cmd, SLI4_BMBX_SIZE);
|
||||||
|
ctx->ctx = hw;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&hw->cmd_lock, flags);
|
||||||
|
|
||||||
|
/* Add to pending list */
|
||||||
|
INIT_LIST_HEAD(&ctx->list_entry);
|
||||||
|
list_add_tail(&ctx->list_entry, &hw->cmd_pending);
|
||||||
|
|
||||||
|
/* Submit as much of the pending list as we can */
|
||||||
|
rc = efct_hw_cmd_submit_pending(hw);
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(&hw->cmd_lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
efct_hw_command_process(struct efct_hw *hw, int status, u8 *mqe,
|
||||||
|
size_t size)
|
||||||
|
{
|
||||||
|
struct efct_command_ctx *ctx = NULL;
|
||||||
|
unsigned long flags = 0;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&hw->cmd_lock, flags);
|
||||||
|
if (!list_empty(&hw->cmd_head)) {
|
||||||
|
ctx = list_first_entry(&hw->cmd_head,
|
||||||
|
struct efct_command_ctx, list_entry);
|
||||||
|
list_del_init(&ctx->list_entry);
|
||||||
|
}
|
||||||
|
if (!ctx) {
|
||||||
|
efc_log_err(hw->os, "no command context\n");
|
||||||
|
spin_unlock_irqrestore(&hw->cmd_lock, flags);
|
||||||
|
return -EIO;
|
||||||
|
}
|
||||||
|
|
||||||
|
hw->cmd_head_count--;
|
||||||
|
|
||||||
|
/* Post any pending requests */
|
||||||
|
efct_hw_cmd_submit_pending(hw);
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(&hw->cmd_lock, flags);
|
||||||
|
|
||||||
|
if (ctx->cb) {
|
||||||
|
memcpy(ctx->buf, mqe, size);
|
||||||
|
ctx->cb(hw, status, ctx->buf, ctx->arg);
|
||||||
|
}
|
||||||
|
|
||||||
|
mempool_free(ctx, hw->cmd_ctx_pool);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
efct_hw_mq_process(struct efct_hw *hw,
|
||||||
|
int status, struct sli4_queue *mq)
|
||||||
|
{
|
||||||
|
u8 mqe[SLI4_BMBX_SIZE];
|
||||||
|
int rc;
|
||||||
|
|
||||||
|
rc = sli_mq_read(&hw->sli, mq, mqe);
|
||||||
|
if (!rc)
|
||||||
|
rc = efct_hw_command_process(hw, status, mqe, mq->size);
|
||||||
|
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int
|
||||||
|
efct_hw_command_cancel(struct efct_hw *hw)
|
||||||
|
{
|
||||||
|
unsigned long flags = 0;
|
||||||
|
int rc = 0;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&hw->cmd_lock, flags);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Manually clean up remaining commands. Note: since this calls
|
||||||
|
* efct_hw_command_process(), we'll also process the cmd_pending
|
||||||
|
* list, so no need to manually clean that out.
|
||||||
|
*/
|
||||||
|
while (!list_empty(&hw->cmd_head)) {
|
||||||
|
u8 mqe[SLI4_BMBX_SIZE] = { 0 };
|
||||||
|
struct efct_command_ctx *ctx;
|
||||||
|
|
||||||
|
ctx = list_first_entry(&hw->cmd_head,
|
||||||
|
struct efct_command_ctx, list_entry);
|
||||||
|
|
||||||
|
efc_log_debug(hw->os, "hung command %08x\n",
|
||||||
|
!ctx ? U32_MAX :
|
||||||
|
(!ctx->buf ? U32_MAX : *((u32 *)ctx->buf)));
|
||||||
|
spin_unlock_irqrestore(&hw->cmd_lock, flags);
|
||||||
|
rc = efct_hw_command_process(hw, -1, mqe, SLI4_BMBX_SIZE);
|
||||||
|
spin_lock_irqsave(&hw->cmd_lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(&hw->cmd_lock, flags);
|
||||||
|
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void
|
||||||
|
efct_mbox_rsp_cb(struct efct_hw *hw, int status, u8 *mqe, void *arg)
|
||||||
|
{
|
||||||
|
struct efct_mbox_rqst_ctx *ctx = arg;
|
||||||
|
|
||||||
|
if (ctx) {
|
||||||
|
if (ctx->callback)
|
||||||
|
(*ctx->callback)(hw->os->efcport, status, mqe,
|
||||||
|
ctx->arg);
|
||||||
|
|
||||||
|
mempool_free(ctx, hw->mbox_rqst_pool);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
int
|
||||||
|
efct_issue_mbox_rqst(void *base, void *cmd, void *cb, void *arg)
|
||||||
|
{
|
||||||
|
struct efct_mbox_rqst_ctx *ctx;
|
||||||
|
struct efct *efct = base;
|
||||||
|
struct efct_hw *hw = &efct->hw;
|
||||||
|
int rc;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Allocate a callback context (which includes the mbox cmd buffer),
|
||||||
|
* we need this to be persistent as the mbox cmd submission may be
|
||||||
|
* queued and executed later execution.
|
||||||
|
*/
|
||||||
|
ctx = mempool_alloc(hw->mbox_rqst_pool, GFP_ATOMIC);
|
||||||
|
if (!ctx)
|
||||||
|
return -EIO;
|
||||||
|
|
||||||
|
ctx->callback = cb;
|
||||||
|
ctx->arg = arg;
|
||||||
|
|
||||||
|
rc = efct_hw_command(hw, cmd, EFCT_CMD_NOWAIT, efct_mbox_rsp_cb, ctx);
|
||||||
|
if (rc) {
|
||||||
|
efc_log_err(efct, "issue mbox rqst failure rc:%d\n", rc);
|
||||||
|
mempool_free(ctx, hw->mbox_rqst_pool);
|
||||||
|
return -EIO;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
@ -602,4 +602,13 @@ efct_get_wwnn(struct efct_hw *hw);
|
|||||||
uint64_t
|
uint64_t
|
||||||
efct_get_wwpn(struct efct_hw *hw);
|
efct_get_wwpn(struct efct_hw *hw);
|
||||||
|
|
||||||
|
int efct_hw_rx_allocate(struct efct_hw *hw);
|
||||||
|
int efct_hw_rx_post(struct efct_hw *hw);
|
||||||
|
void efct_hw_rx_free(struct efct_hw *hw);
|
||||||
|
int
|
||||||
|
efct_hw_command(struct efct_hw *hw, u8 *cmd, u32 opts, void *cb,
|
||||||
|
void *arg);
|
||||||
|
int
|
||||||
|
efct_issue_mbox_rqst(void *base, void *cmd, void *cb, void *arg);
|
||||||
|
|
||||||
#endif /* __EFCT_H__ */
|
#endif /* __EFCT_H__ */
|
||||||
|
Loading…
Reference in New Issue
Block a user