2015-04-29 12:32:35 +00:00
|
|
|
/*
|
|
|
|
* Intel MIC Platform Software Stack (MPSS)
|
|
|
|
*
|
|
|
|
* Copyright(c) 2014 Intel Corporation.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License, version 2, as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful, but
|
|
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* General Public License for more details.
|
|
|
|
*
|
|
|
|
* Intel SCIF driver.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
#include <linux/scif.h>
|
|
|
|
#include "scif_main.h"
|
|
|
|
#include "scif_map.h"
|
|
|
|
|
|
|
|
static const char * const scif_ep_states[] = {
|
|
|
|
"Unbound",
|
|
|
|
"Bound",
|
|
|
|
"Listening",
|
|
|
|
"Connected",
|
|
|
|
"Connecting",
|
|
|
|
"Mapping",
|
|
|
|
"Closing",
|
|
|
|
"Close Listening",
|
|
|
|
"Disconnected",
|
|
|
|
"Zombie"};
|
|
|
|
|
|
|
|
enum conn_async_state {
|
|
|
|
ASYNC_CONN_IDLE = 1, /* ep setup for async connect */
|
|
|
|
ASYNC_CONN_INPROGRESS, /* async connect in progress */
|
|
|
|
ASYNC_CONN_FLUSH_WORK /* async work flush in progress */
|
|
|
|
};
|
|
|
|
|
|
|
|
scif_epd_t scif_open(void)
|
|
|
|
{
|
|
|
|
struct scif_endpt *ep;
|
|
|
|
|
|
|
|
might_sleep();
|
|
|
|
ep = kzalloc(sizeof(*ep), GFP_KERNEL);
|
|
|
|
if (!ep)
|
|
|
|
goto err_ep_alloc;
|
|
|
|
|
|
|
|
ep->qp_info.qp = kzalloc(sizeof(*ep->qp_info.qp), GFP_KERNEL);
|
|
|
|
if (!ep->qp_info.qp)
|
|
|
|
goto err_qp_alloc;
|
|
|
|
|
|
|
|
spin_lock_init(&ep->lock);
|
|
|
|
mutex_init(&ep->sendlock);
|
|
|
|
mutex_init(&ep->recvlock);
|
|
|
|
|
|
|
|
ep->state = SCIFEP_UNBOUND;
|
|
|
|
dev_dbg(scif_info.mdev.this_device,
|
|
|
|
"SCIFAPI open: ep %p success\n", ep);
|
|
|
|
return ep;
|
|
|
|
|
|
|
|
err_qp_alloc:
|
|
|
|
kfree(ep);
|
|
|
|
err_ep_alloc:
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(scif_open);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* scif_disconnect_ep - Disconnects the endpoint if found
|
|
|
|
* @epd: The end point returned from scif_open()
|
|
|
|
*/
|
|
|
|
static struct scif_endpt *scif_disconnect_ep(struct scif_endpt *ep)
|
|
|
|
{
|
|
|
|
struct scifmsg msg;
|
|
|
|
struct scif_endpt *fep = NULL;
|
|
|
|
struct scif_endpt *tmpep;
|
|
|
|
struct list_head *pos, *tmpq;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wake up any threads blocked in send()/recv() before closing
|
|
|
|
* out the connection. Grabbing and releasing the send/recv lock
|
|
|
|
* will ensure that any blocked senders/receivers have exited for
|
|
|
|
* Ring 0 endpoints. It is a Ring 0 bug to call send/recv after
|
|
|
|
* close. Ring 3 endpoints are not affected since close will not
|
|
|
|
* be called while there are IOCTLs executing.
|
|
|
|
*/
|
|
|
|
wake_up_interruptible(&ep->sendwq);
|
|
|
|
wake_up_interruptible(&ep->recvwq);
|
|
|
|
mutex_lock(&ep->sendlock);
|
|
|
|
mutex_unlock(&ep->sendlock);
|
|
|
|
mutex_lock(&ep->recvlock);
|
|
|
|
mutex_unlock(&ep->recvlock);
|
|
|
|
|
|
|
|
/* Remove from the connected list */
|
|
|
|
mutex_lock(&scif_info.connlock);
|
|
|
|
list_for_each_safe(pos, tmpq, &scif_info.connected) {
|
|
|
|
tmpep = list_entry(pos, struct scif_endpt, list);
|
|
|
|
if (tmpep == ep) {
|
|
|
|
list_del(pos);
|
|
|
|
fep = tmpep;
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!fep) {
|
|
|
|
/*
|
|
|
|
* The other side has completed the disconnect before
|
|
|
|
* the end point can be removed from the list. Therefore
|
|
|
|
* the ep lock is not locked, traverse the disconnected
|
|
|
|
* list to find the endpoint and release the conn lock.
|
|
|
|
*/
|
|
|
|
list_for_each_safe(pos, tmpq, &scif_info.disconnected) {
|
|
|
|
tmpep = list_entry(pos, struct scif_endpt, list);
|
|
|
|
if (tmpep == ep) {
|
|
|
|
list_del(pos);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mutex_unlock(&scif_info.connlock);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
init_completion(&ep->discon);
|
|
|
|
msg.uop = SCIF_DISCNCT;
|
|
|
|
msg.src = ep->port;
|
|
|
|
msg.dst = ep->peer;
|
|
|
|
msg.payload[0] = (u64)ep;
|
|
|
|
msg.payload[1] = ep->remote_ep;
|
|
|
|
|
|
|
|
err = scif_nodeqp_send(ep->remote_dev, &msg);
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
mutex_unlock(&scif_info.connlock);
|
|
|
|
|
|
|
|
if (!err)
|
|
|
|
/* Wait for the remote node to respond with SCIF_DISCNT_ACK */
|
|
|
|
wait_for_completion_timeout(&ep->discon,
|
|
|
|
SCIF_NODE_ALIVE_TIMEOUT);
|
|
|
|
return ep;
|
|
|
|
}
|
|
|
|
|
|
|
|
int scif_close(scif_epd_t epd)
|
|
|
|
{
|
|
|
|
struct scif_endpt *ep = (struct scif_endpt *)epd;
|
|
|
|
struct scif_endpt *tmpep;
|
|
|
|
struct list_head *pos, *tmpq;
|
|
|
|
enum scif_epd_state oldstate;
|
|
|
|
bool flush_conn;
|
|
|
|
|
|
|
|
dev_dbg(scif_info.mdev.this_device, "SCIFAPI close: ep %p %s\n",
|
|
|
|
ep, scif_ep_states[ep->state]);
|
|
|
|
might_sleep();
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
flush_conn = (ep->conn_async_state == ASYNC_CONN_INPROGRESS);
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
|
|
|
|
if (flush_conn)
|
|
|
|
flush_work(&scif_info.conn_work);
|
|
|
|
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
oldstate = ep->state;
|
|
|
|
|
|
|
|
ep->state = SCIFEP_CLOSING;
|
|
|
|
|
|
|
|
switch (oldstate) {
|
|
|
|
case SCIFEP_ZOMBIE:
|
|
|
|
case SCIFEP_DISCONNECTED:
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
/* Remove from the disconnected list */
|
|
|
|
mutex_lock(&scif_info.connlock);
|
|
|
|
list_for_each_safe(pos, tmpq, &scif_info.disconnected) {
|
|
|
|
tmpep = list_entry(pos, struct scif_endpt, list);
|
|
|
|
if (tmpep == ep) {
|
|
|
|
list_del(pos);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mutex_unlock(&scif_info.connlock);
|
|
|
|
break;
|
|
|
|
case SCIFEP_UNBOUND:
|
|
|
|
case SCIFEP_BOUND:
|
|
|
|
case SCIFEP_CONNECTING:
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
break;
|
|
|
|
case SCIFEP_MAPPING:
|
|
|
|
case SCIFEP_CONNECTED:
|
|
|
|
case SCIFEP_CLOSING:
|
|
|
|
{
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
scif_disconnect_ep(ep);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case SCIFEP_LISTENING:
|
|
|
|
case SCIFEP_CLLISTEN:
|
|
|
|
{
|
|
|
|
struct scif_conreq *conreq;
|
|
|
|
struct scifmsg msg;
|
|
|
|
struct scif_endpt *aep;
|
|
|
|
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
spin_lock(&scif_info.eplock);
|
|
|
|
|
|
|
|
/* remove from listen list */
|
|
|
|
list_for_each_safe(pos, tmpq, &scif_info.listen) {
|
|
|
|
tmpep = list_entry(pos, struct scif_endpt, list);
|
|
|
|
if (tmpep == ep)
|
|
|
|
list_del(pos);
|
|
|
|
}
|
|
|
|
/* Remove any dangling accepts */
|
|
|
|
while (ep->acceptcnt) {
|
|
|
|
aep = list_first_entry(&ep->li_accept,
|
|
|
|
struct scif_endpt, liacceptlist);
|
|
|
|
list_del(&aep->liacceptlist);
|
|
|
|
scif_put_port(aep->port.port);
|
|
|
|
list_for_each_safe(pos, tmpq, &scif_info.uaccept) {
|
|
|
|
tmpep = list_entry(pos, struct scif_endpt,
|
|
|
|
miacceptlist);
|
|
|
|
if (tmpep == aep) {
|
|
|
|
list_del(pos);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
spin_unlock(&scif_info.eplock);
|
|
|
|
mutex_lock(&scif_info.connlock);
|
|
|
|
list_for_each_safe(pos, tmpq, &scif_info.connected) {
|
|
|
|
tmpep = list_entry(pos,
|
|
|
|
struct scif_endpt, list);
|
|
|
|
if (tmpep == aep) {
|
|
|
|
list_del(pos);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
list_for_each_safe(pos, tmpq, &scif_info.disconnected) {
|
|
|
|
tmpep = list_entry(pos,
|
|
|
|
struct scif_endpt, list);
|
|
|
|
if (tmpep == aep) {
|
|
|
|
list_del(pos);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mutex_unlock(&scif_info.connlock);
|
|
|
|
scif_teardown_ep(aep);
|
|
|
|
spin_lock(&scif_info.eplock);
|
|
|
|
scif_add_epd_to_zombie_list(aep, SCIF_EPLOCK_HELD);
|
|
|
|
ep->acceptcnt--;
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
spin_unlock(&scif_info.eplock);
|
|
|
|
|
|
|
|
/* Remove and reject any pending connection requests. */
|
|
|
|
while (ep->conreqcnt) {
|
|
|
|
conreq = list_first_entry(&ep->conlist,
|
|
|
|
struct scif_conreq, list);
|
|
|
|
list_del(&conreq->list);
|
|
|
|
|
|
|
|
msg.uop = SCIF_CNCT_REJ;
|
|
|
|
msg.dst.node = conreq->msg.src.node;
|
|
|
|
msg.dst.port = conreq->msg.src.port;
|
|
|
|
msg.payload[0] = conreq->msg.payload[0];
|
|
|
|
msg.payload[1] = conreq->msg.payload[1];
|
|
|
|
/*
|
|
|
|
* No Error Handling on purpose for scif_nodeqp_send().
|
|
|
|
* If the remote node is lost we still want free the
|
|
|
|
* connection requests on the self node.
|
|
|
|
*/
|
|
|
|
scif_nodeqp_send(&scif_dev[conreq->msg.src.node],
|
|
|
|
&msg);
|
|
|
|
ep->conreqcnt--;
|
|
|
|
kfree(conreq);
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
/* If a kSCIF accept is waiting wake it up */
|
|
|
|
wake_up_interruptible(&ep->conwq);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
scif_put_port(ep->port.port);
|
|
|
|
scif_teardown_ep(ep);
|
|
|
|
scif_add_epd_to_zombie_list(ep, !SCIF_EPLOCK_HELD);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(scif_close);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* scif_flush() - Wakes up any blocking accepts. The endpoint will no longer
|
|
|
|
* accept new connections.
|
|
|
|
* @epd: The end point returned from scif_open()
|
|
|
|
*/
|
|
|
|
int __scif_flush(scif_epd_t epd)
|
|
|
|
{
|
|
|
|
struct scif_endpt *ep = (struct scif_endpt *)epd;
|
|
|
|
|
|
|
|
switch (ep->state) {
|
|
|
|
case SCIFEP_LISTENING:
|
|
|
|
{
|
|
|
|
ep->state = SCIFEP_CLLISTEN;
|
|
|
|
|
|
|
|
/* If an accept is waiting wake it up */
|
|
|
|
wake_up_interruptible(&ep->conwq);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int scif_bind(scif_epd_t epd, u16 pn)
|
|
|
|
{
|
|
|
|
struct scif_endpt *ep = (struct scif_endpt *)epd;
|
|
|
|
int ret = 0;
|
|
|
|
int tmp;
|
|
|
|
|
|
|
|
dev_dbg(scif_info.mdev.this_device,
|
|
|
|
"SCIFAPI bind: ep %p %s requested port number %d\n",
|
|
|
|
ep, scif_ep_states[ep->state], pn);
|
|
|
|
if (pn) {
|
|
|
|
/*
|
|
|
|
* Similar to IETF RFC 1700, SCIF ports below
|
|
|
|
* SCIF_ADMIN_PORT_END can only be bound by system (or root)
|
|
|
|
* processes or by processes executed by privileged users.
|
|
|
|
*/
|
|
|
|
if (pn < SCIF_ADMIN_PORT_END && !capable(CAP_SYS_ADMIN)) {
|
|
|
|
ret = -EACCES;
|
|
|
|
goto scif_bind_admin_exit;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
if (ep->state == SCIFEP_BOUND) {
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto scif_bind_exit;
|
|
|
|
} else if (ep->state != SCIFEP_UNBOUND) {
|
|
|
|
ret = -EISCONN;
|
|
|
|
goto scif_bind_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (pn) {
|
|
|
|
tmp = scif_rsrv_port(pn);
|
|
|
|
if (tmp != pn) {
|
|
|
|
ret = -EINVAL;
|
|
|
|
goto scif_bind_exit;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
pn = scif_get_new_port();
|
|
|
|
if (!pn) {
|
|
|
|
ret = -ENOSPC;
|
|
|
|
goto scif_bind_exit;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
ep->state = SCIFEP_BOUND;
|
|
|
|
ep->port.node = scif_info.nodeid;
|
|
|
|
ep->port.port = pn;
|
|
|
|
ep->conn_async_state = ASYNC_CONN_IDLE;
|
|
|
|
ret = pn;
|
|
|
|
dev_dbg(scif_info.mdev.this_device,
|
|
|
|
"SCIFAPI bind: bound to port number %d\n", pn);
|
|
|
|
scif_bind_exit:
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
scif_bind_admin_exit:
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(scif_bind);
|
|
|
|
|
|
|
|
int scif_listen(scif_epd_t epd, int backlog)
|
|
|
|
{
|
|
|
|
struct scif_endpt *ep = (struct scif_endpt *)epd;
|
|
|
|
|
|
|
|
dev_dbg(scif_info.mdev.this_device,
|
|
|
|
"SCIFAPI listen: ep %p %s\n", ep, scif_ep_states[ep->state]);
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
switch (ep->state) {
|
|
|
|
case SCIFEP_ZOMBIE:
|
|
|
|
case SCIFEP_CLOSING:
|
|
|
|
case SCIFEP_CLLISTEN:
|
|
|
|
case SCIFEP_UNBOUND:
|
|
|
|
case SCIFEP_DISCONNECTED:
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
return -EINVAL;
|
|
|
|
case SCIFEP_LISTENING:
|
|
|
|
case SCIFEP_CONNECTED:
|
|
|
|
case SCIFEP_CONNECTING:
|
|
|
|
case SCIFEP_MAPPING:
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
return -EISCONN;
|
|
|
|
case SCIFEP_BOUND:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
ep->state = SCIFEP_LISTENING;
|
|
|
|
ep->backlog = backlog;
|
|
|
|
|
|
|
|
ep->conreqcnt = 0;
|
|
|
|
ep->acceptcnt = 0;
|
|
|
|
INIT_LIST_HEAD(&ep->conlist);
|
|
|
|
init_waitqueue_head(&ep->conwq);
|
|
|
|
INIT_LIST_HEAD(&ep->li_accept);
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Listen status is complete so delete the qp information not needed
|
|
|
|
* on a listen before placing on the list of listening ep's
|
|
|
|
*/
|
|
|
|
scif_teardown_ep(ep);
|
|
|
|
ep->qp_info.qp = NULL;
|
|
|
|
|
|
|
|
spin_lock(&scif_info.eplock);
|
|
|
|
list_add_tail(&ep->list, &scif_info.listen);
|
|
|
|
spin_unlock(&scif_info.eplock);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(scif_listen);
|
2015-04-29 12:32:36 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
************************************************************************
|
|
|
|
* SCIF connection flow:
|
|
|
|
*
|
|
|
|
* 1) A SCIF listening endpoint can call scif_accept(..) to wait for SCIF
|
|
|
|
* connections via a SCIF_CNCT_REQ message
|
|
|
|
* 2) A SCIF endpoint can initiate a SCIF connection by calling
|
|
|
|
* scif_connect(..) which calls scif_setup_qp_connect(..) which
|
|
|
|
* allocates the local qp for the endpoint ring buffer and then sends
|
|
|
|
* a SCIF_CNCT_REQ to the remote node and waits for a SCIF_CNCT_GNT or
|
|
|
|
* a SCIF_CNCT_REJ message
|
|
|
|
* 3) The peer node handles a SCIF_CNCT_REQ via scif_cnctreq_resp(..) which
|
|
|
|
* wakes up any threads blocked in step 1 or sends a SCIF_CNCT_REJ
|
|
|
|
* message otherwise
|
|
|
|
* 4) A thread blocked waiting for incoming connections allocates its local
|
|
|
|
* endpoint QP and ring buffer following which it sends a SCIF_CNCT_GNT
|
|
|
|
* and waits for a SCIF_CNCT_GNT(N)ACK. If the allocation fails then
|
|
|
|
* the node sends a SCIF_CNCT_REJ message
|
|
|
|
* 5) Upon receipt of a SCIF_CNCT_GNT or a SCIF_CNCT_REJ message the
|
|
|
|
* connecting endpoint is woken up as part of handling
|
|
|
|
* scif_cnctgnt_resp(..) following which it maps the remote endpoints'
|
|
|
|
* QP, updates its outbound QP and sends a SCIF_CNCT_GNTACK message on
|
|
|
|
* success or a SCIF_CNCT_GNTNACK message on failure and completes
|
|
|
|
* the scif_connect(..) API
|
|
|
|
* 6) Upon receipt of a SCIF_CNCT_GNT(N)ACK the accepting endpoint blocked
|
|
|
|
* in step 4 is woken up and completes the scif_accept(..) API
|
|
|
|
* 7) The SCIF connection is now established between the two SCIF endpoints.
|
|
|
|
*/
|
|
|
|
static int scif_conn_func(struct scif_endpt *ep)
|
|
|
|
{
|
|
|
|
int err = 0;
|
|
|
|
struct scifmsg msg;
|
|
|
|
struct device *spdev;
|
|
|
|
|
|
|
|
/* Initiate the first part of the endpoint QP setup */
|
|
|
|
err = scif_setup_qp_connect(ep->qp_info.qp, &ep->qp_info.qp_offset,
|
|
|
|
SCIF_ENDPT_QP_SIZE, ep->remote_dev);
|
|
|
|
if (err) {
|
|
|
|
dev_err(&ep->remote_dev->sdev->dev,
|
|
|
|
"%s err %d qp_offset 0x%llx\n",
|
|
|
|
__func__, err, ep->qp_info.qp_offset);
|
|
|
|
ep->state = SCIFEP_BOUND;
|
|
|
|
goto connect_error_simple;
|
|
|
|
}
|
|
|
|
|
|
|
|
spdev = scif_get_peer_dev(ep->remote_dev);
|
|
|
|
if (IS_ERR(spdev)) {
|
|
|
|
err = PTR_ERR(spdev);
|
|
|
|
goto cleanup_qp;
|
|
|
|
}
|
|
|
|
/* Format connect message and send it */
|
|
|
|
msg.src = ep->port;
|
|
|
|
msg.dst = ep->conn_port;
|
|
|
|
msg.uop = SCIF_CNCT_REQ;
|
|
|
|
msg.payload[0] = (u64)ep;
|
|
|
|
msg.payload[1] = ep->qp_info.qp_offset;
|
|
|
|
err = _scif_nodeqp_send(ep->remote_dev, &msg);
|
|
|
|
if (err)
|
|
|
|
goto connect_error_dec;
|
|
|
|
scif_put_peer_dev(spdev);
|
|
|
|
/*
|
|
|
|
* Wait for the remote node to respond with SCIF_CNCT_GNT or
|
|
|
|
* SCIF_CNCT_REJ message.
|
|
|
|
*/
|
|
|
|
err = wait_event_timeout(ep->conwq, ep->state != SCIFEP_CONNECTING,
|
|
|
|
SCIF_NODE_ALIVE_TIMEOUT);
|
|
|
|
if (!err) {
|
|
|
|
dev_err(&ep->remote_dev->sdev->dev,
|
|
|
|
"%s %d timeout\n", __func__, __LINE__);
|
|
|
|
ep->state = SCIFEP_BOUND;
|
|
|
|
}
|
|
|
|
spdev = scif_get_peer_dev(ep->remote_dev);
|
|
|
|
if (IS_ERR(spdev)) {
|
|
|
|
err = PTR_ERR(spdev);
|
|
|
|
goto cleanup_qp;
|
|
|
|
}
|
|
|
|
if (ep->state == SCIFEP_MAPPING) {
|
|
|
|
err = scif_setup_qp_connect_response(ep->remote_dev,
|
|
|
|
ep->qp_info.qp,
|
|
|
|
ep->qp_info.gnt_pld);
|
|
|
|
/*
|
|
|
|
* If the resource to map the queue are not available then
|
|
|
|
* we need to tell the other side to terminate the accept
|
|
|
|
*/
|
|
|
|
if (err) {
|
|
|
|
dev_err(&ep->remote_dev->sdev->dev,
|
|
|
|
"%s %d err %d\n", __func__, __LINE__, err);
|
|
|
|
msg.uop = SCIF_CNCT_GNTNACK;
|
|
|
|
msg.payload[0] = ep->remote_ep;
|
|
|
|
_scif_nodeqp_send(ep->remote_dev, &msg);
|
|
|
|
ep->state = SCIFEP_BOUND;
|
|
|
|
goto connect_error_dec;
|
|
|
|
}
|
|
|
|
|
|
|
|
msg.uop = SCIF_CNCT_GNTACK;
|
|
|
|
msg.payload[0] = ep->remote_ep;
|
|
|
|
err = _scif_nodeqp_send(ep->remote_dev, &msg);
|
|
|
|
if (err) {
|
|
|
|
ep->state = SCIFEP_BOUND;
|
|
|
|
goto connect_error_dec;
|
|
|
|
}
|
|
|
|
ep->state = SCIFEP_CONNECTED;
|
|
|
|
mutex_lock(&scif_info.connlock);
|
|
|
|
list_add_tail(&ep->list, &scif_info.connected);
|
|
|
|
mutex_unlock(&scif_info.connlock);
|
|
|
|
dev_dbg(&ep->remote_dev->sdev->dev,
|
|
|
|
"SCIFAPI connect: ep %p connected\n", ep);
|
|
|
|
} else if (ep->state == SCIFEP_BOUND) {
|
|
|
|
dev_dbg(&ep->remote_dev->sdev->dev,
|
|
|
|
"SCIFAPI connect: ep %p connection refused\n", ep);
|
|
|
|
err = -ECONNREFUSED;
|
|
|
|
goto connect_error_dec;
|
|
|
|
}
|
|
|
|
scif_put_peer_dev(spdev);
|
|
|
|
return err;
|
|
|
|
connect_error_dec:
|
|
|
|
scif_put_peer_dev(spdev);
|
|
|
|
cleanup_qp:
|
|
|
|
scif_cleanup_ep_qp(ep);
|
|
|
|
connect_error_simple:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* scif_conn_handler:
|
|
|
|
*
|
|
|
|
* Workqueue handler for servicing non-blocking SCIF connect
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
void scif_conn_handler(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct scif_endpt *ep;
|
|
|
|
|
|
|
|
do {
|
|
|
|
ep = NULL;
|
|
|
|
spin_lock(&scif_info.nb_connect_lock);
|
|
|
|
if (!list_empty(&scif_info.nb_connect_list)) {
|
|
|
|
ep = list_first_entry(&scif_info.nb_connect_list,
|
|
|
|
struct scif_endpt, conn_list);
|
|
|
|
list_del(&ep->conn_list);
|
|
|
|
}
|
|
|
|
spin_unlock(&scif_info.nb_connect_lock);
|
|
|
|
if (ep)
|
|
|
|
ep->conn_err = scif_conn_func(ep);
|
|
|
|
} while (ep);
|
|
|
|
}
|
|
|
|
|
|
|
|
int __scif_connect(scif_epd_t epd, struct scif_port_id *dst, bool non_block)
|
|
|
|
{
|
|
|
|
struct scif_endpt *ep = (struct scif_endpt *)epd;
|
|
|
|
int err = 0;
|
|
|
|
struct scif_dev *remote_dev;
|
|
|
|
struct device *spdev;
|
|
|
|
|
|
|
|
dev_dbg(scif_info.mdev.this_device, "SCIFAPI connect: ep %p %s\n", ep,
|
|
|
|
scif_ep_states[ep->state]);
|
|
|
|
|
|
|
|
if (!scif_dev || dst->node > scif_info.maxid)
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
might_sleep();
|
|
|
|
|
|
|
|
remote_dev = &scif_dev[dst->node];
|
|
|
|
spdev = scif_get_peer_dev(remote_dev);
|
|
|
|
if (IS_ERR(spdev)) {
|
|
|
|
err = PTR_ERR(spdev);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
switch (ep->state) {
|
|
|
|
case SCIFEP_ZOMBIE:
|
|
|
|
case SCIFEP_CLOSING:
|
|
|
|
err = -EINVAL;
|
|
|
|
break;
|
|
|
|
case SCIFEP_DISCONNECTED:
|
|
|
|
if (ep->conn_async_state == ASYNC_CONN_INPROGRESS)
|
|
|
|
ep->conn_async_state = ASYNC_CONN_FLUSH_WORK;
|
|
|
|
else
|
|
|
|
err = -EINVAL;
|
|
|
|
break;
|
|
|
|
case SCIFEP_LISTENING:
|
|
|
|
case SCIFEP_CLLISTEN:
|
|
|
|
err = -EOPNOTSUPP;
|
|
|
|
break;
|
|
|
|
case SCIFEP_CONNECTING:
|
|
|
|
case SCIFEP_MAPPING:
|
|
|
|
if (ep->conn_async_state == ASYNC_CONN_INPROGRESS)
|
|
|
|
err = -EINPROGRESS;
|
|
|
|
else
|
|
|
|
err = -EISCONN;
|
|
|
|
break;
|
|
|
|
case SCIFEP_CONNECTED:
|
|
|
|
if (ep->conn_async_state == ASYNC_CONN_INPROGRESS)
|
|
|
|
ep->conn_async_state = ASYNC_CONN_FLUSH_WORK;
|
|
|
|
else
|
|
|
|
err = -EISCONN;
|
|
|
|
break;
|
|
|
|
case SCIFEP_UNBOUND:
|
|
|
|
ep->port.port = scif_get_new_port();
|
|
|
|
if (!ep->port.port) {
|
|
|
|
err = -ENOSPC;
|
|
|
|
} else {
|
|
|
|
ep->port.node = scif_info.nodeid;
|
|
|
|
ep->conn_async_state = ASYNC_CONN_IDLE;
|
|
|
|
}
|
|
|
|
/* Fall through */
|
|
|
|
case SCIFEP_BOUND:
|
|
|
|
/*
|
|
|
|
* If a non-blocking connect has been already initiated
|
|
|
|
* (conn_async_state is either ASYNC_CONN_INPROGRESS or
|
|
|
|
* ASYNC_CONN_FLUSH_WORK), the end point could end up in
|
|
|
|
* SCIF_BOUND due an error in the connection process
|
|
|
|
* (e.g., connection refused) If conn_async_state is
|
|
|
|
* ASYNC_CONN_INPROGRESS - transition to ASYNC_CONN_FLUSH_WORK
|
|
|
|
* so that the error status can be collected. If the state is
|
|
|
|
* already ASYNC_CONN_FLUSH_WORK - then set the error to
|
|
|
|
* EINPROGRESS since some other thread is waiting to collect
|
|
|
|
* error status.
|
|
|
|
*/
|
|
|
|
if (ep->conn_async_state == ASYNC_CONN_INPROGRESS) {
|
|
|
|
ep->conn_async_state = ASYNC_CONN_FLUSH_WORK;
|
|
|
|
} else if (ep->conn_async_state == ASYNC_CONN_FLUSH_WORK) {
|
|
|
|
err = -EINPROGRESS;
|
|
|
|
} else {
|
|
|
|
ep->conn_port = *dst;
|
|
|
|
init_waitqueue_head(&ep->sendwq);
|
|
|
|
init_waitqueue_head(&ep->recvwq);
|
|
|
|
init_waitqueue_head(&ep->conwq);
|
|
|
|
ep->conn_async_state = 0;
|
|
|
|
|
|
|
|
if (unlikely(non_block))
|
|
|
|
ep->conn_async_state = ASYNC_CONN_INPROGRESS;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (err || ep->conn_async_state == ASYNC_CONN_FLUSH_WORK)
|
|
|
|
goto connect_simple_unlock1;
|
|
|
|
|
|
|
|
ep->state = SCIFEP_CONNECTING;
|
|
|
|
ep->remote_dev = &scif_dev[dst->node];
|
|
|
|
ep->qp_info.qp->magic = SCIFEP_MAGIC;
|
|
|
|
if (ep->conn_async_state == ASYNC_CONN_INPROGRESS) {
|
|
|
|
spin_lock(&scif_info.nb_connect_lock);
|
|
|
|
list_add_tail(&ep->conn_list, &scif_info.nb_connect_list);
|
|
|
|
spin_unlock(&scif_info.nb_connect_lock);
|
|
|
|
err = -EINPROGRESS;
|
|
|
|
schedule_work(&scif_info.conn_work);
|
|
|
|
}
|
|
|
|
connect_simple_unlock1:
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
scif_put_peer_dev(spdev);
|
|
|
|
if (err) {
|
|
|
|
return err;
|
|
|
|
} else if (ep->conn_async_state == ASYNC_CONN_FLUSH_WORK) {
|
|
|
|
flush_work(&scif_info.conn_work);
|
|
|
|
err = ep->conn_err;
|
|
|
|
spin_lock(&ep->lock);
|
|
|
|
ep->conn_async_state = ASYNC_CONN_IDLE;
|
|
|
|
spin_unlock(&ep->lock);
|
|
|
|
} else {
|
|
|
|
err = scif_conn_func(ep);
|
|
|
|
}
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
int scif_connect(scif_epd_t epd, struct scif_port_id *dst)
|
|
|
|
{
|
|
|
|
return __scif_connect(epd, dst, false);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(scif_connect);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* scif_accept() - Accept a connection request from the remote node
|
|
|
|
*
|
|
|
|
* The function accepts a connection request from the remote node. Successful
|
|
|
|
* complete is indicate by a new end point being created and passed back
|
|
|
|
* to the caller for future reference.
|
|
|
|
*
|
|
|
|
* Upon successful complete a zero will be returned and the peer information
|
|
|
|
* will be filled in.
|
|
|
|
*
|
|
|
|
* If the end point is not in the listening state -EINVAL will be returned.
|
|
|
|
*
|
|
|
|
* If during the connection sequence resource allocation fails the -ENOMEM
|
|
|
|
* will be returned.
|
|
|
|
*
|
|
|
|
* If the function is called with the ASYNC flag set and no connection requests
|
|
|
|
* are pending it will return -EAGAIN.
|
|
|
|
*
|
|
|
|
* If the remote side is not sending any connection requests the caller may
|
|
|
|
* terminate this function with a signal. If so a -EINTR will be returned.
|
|
|
|
*/
|
|
|
|
int scif_accept(scif_epd_t epd, struct scif_port_id *peer,
|
|
|
|
scif_epd_t *newepd, int flags)
|
|
|
|
{
|
|
|
|
struct scif_endpt *lep = (struct scif_endpt *)epd;
|
|
|
|
struct scif_endpt *cep;
|
|
|
|
struct scif_conreq *conreq;
|
|
|
|
struct scifmsg msg;
|
|
|
|
int err;
|
|
|
|
struct device *spdev;
|
|
|
|
|
|
|
|
dev_dbg(scif_info.mdev.this_device,
|
|
|
|
"SCIFAPI accept: ep %p %s\n", lep, scif_ep_states[lep->state]);
|
|
|
|
|
|
|
|
if (flags & ~SCIF_ACCEPT_SYNC)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (!peer || !newepd)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
might_sleep();
|
|
|
|
spin_lock(&lep->lock);
|
|
|
|
if (lep->state != SCIFEP_LISTENING) {
|
|
|
|
spin_unlock(&lep->lock);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!lep->conreqcnt && !(flags & SCIF_ACCEPT_SYNC)) {
|
|
|
|
/* No connection request present and we do not want to wait */
|
|
|
|
spin_unlock(&lep->lock);
|
|
|
|
return -EAGAIN;
|
|
|
|
}
|
|
|
|
|
|
|
|
lep->files = current->files;
|
|
|
|
retry_connection:
|
|
|
|
spin_unlock(&lep->lock);
|
|
|
|
/* Wait for the remote node to send us a SCIF_CNCT_REQ */
|
|
|
|
err = wait_event_interruptible(lep->conwq,
|
|
|
|
(lep->conreqcnt ||
|
|
|
|
(lep->state != SCIFEP_LISTENING)));
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
|
|
|
|
if (lep->state != SCIFEP_LISTENING)
|
|
|
|
return -EINTR;
|
|
|
|
|
|
|
|
spin_lock(&lep->lock);
|
|
|
|
|
|
|
|
if (!lep->conreqcnt)
|
|
|
|
goto retry_connection;
|
|
|
|
|
|
|
|
/* Get the first connect request off the list */
|
|
|
|
conreq = list_first_entry(&lep->conlist, struct scif_conreq, list);
|
|
|
|
list_del(&conreq->list);
|
|
|
|
lep->conreqcnt--;
|
|
|
|
spin_unlock(&lep->lock);
|
|
|
|
|
|
|
|
/* Fill in the peer information */
|
|
|
|
peer->node = conreq->msg.src.node;
|
|
|
|
peer->port = conreq->msg.src.port;
|
|
|
|
|
|
|
|
cep = kzalloc(sizeof(*cep), GFP_KERNEL);
|
|
|
|
if (!cep) {
|
|
|
|
err = -ENOMEM;
|
|
|
|
goto scif_accept_error_epalloc;
|
|
|
|
}
|
|
|
|
spin_lock_init(&cep->lock);
|
|
|
|
mutex_init(&cep->sendlock);
|
|
|
|
mutex_init(&cep->recvlock);
|
|
|
|
cep->state = SCIFEP_CONNECTING;
|
|
|
|
cep->remote_dev = &scif_dev[peer->node];
|
|
|
|
cep->remote_ep = conreq->msg.payload[0];
|
|
|
|
|
|
|
|
cep->qp_info.qp = kzalloc(sizeof(*cep->qp_info.qp), GFP_KERNEL);
|
|
|
|
if (!cep->qp_info.qp) {
|
|
|
|
err = -ENOMEM;
|
|
|
|
goto scif_accept_error_qpalloc;
|
|
|
|
}
|
|
|
|
|
|
|
|
cep->qp_info.qp->magic = SCIFEP_MAGIC;
|
|
|
|
spdev = scif_get_peer_dev(cep->remote_dev);
|
|
|
|
if (IS_ERR(spdev)) {
|
|
|
|
err = PTR_ERR(spdev);
|
|
|
|
goto scif_accept_error_map;
|
|
|
|
}
|
|
|
|
err = scif_setup_qp_accept(cep->qp_info.qp, &cep->qp_info.qp_offset,
|
|
|
|
conreq->msg.payload[1], SCIF_ENDPT_QP_SIZE,
|
|
|
|
cep->remote_dev);
|
|
|
|
if (err) {
|
|
|
|
dev_dbg(&cep->remote_dev->sdev->dev,
|
|
|
|
"SCIFAPI accept: ep %p new %p scif_setup_qp_accept %d qp_offset 0x%llx\n",
|
|
|
|
lep, cep, err, cep->qp_info.qp_offset);
|
|
|
|
scif_put_peer_dev(spdev);
|
|
|
|
goto scif_accept_error_map;
|
|
|
|
}
|
|
|
|
|
|
|
|
cep->port.node = lep->port.node;
|
|
|
|
cep->port.port = lep->port.port;
|
|
|
|
cep->peer.node = peer->node;
|
|
|
|
cep->peer.port = peer->port;
|
|
|
|
init_waitqueue_head(&cep->sendwq);
|
|
|
|
init_waitqueue_head(&cep->recvwq);
|
|
|
|
init_waitqueue_head(&cep->conwq);
|
|
|
|
|
|
|
|
msg.uop = SCIF_CNCT_GNT;
|
|
|
|
msg.src = cep->port;
|
|
|
|
msg.payload[0] = cep->remote_ep;
|
|
|
|
msg.payload[1] = cep->qp_info.qp_offset;
|
|
|
|
msg.payload[2] = (u64)cep;
|
|
|
|
|
|
|
|
err = _scif_nodeqp_send(cep->remote_dev, &msg);
|
|
|
|
scif_put_peer_dev(spdev);
|
|
|
|
if (err)
|
|
|
|
goto scif_accept_error_map;
|
|
|
|
retry:
|
|
|
|
/* Wait for the remote node to respond with SCIF_CNCT_GNT(N)ACK */
|
|
|
|
err = wait_event_timeout(cep->conwq, cep->state != SCIFEP_CONNECTING,
|
|
|
|
SCIF_NODE_ACCEPT_TIMEOUT);
|
|
|
|
if (!err && scifdev_alive(cep))
|
|
|
|
goto retry;
|
|
|
|
err = !err ? -ENODEV : 0;
|
|
|
|
if (err)
|
|
|
|
goto scif_accept_error_map;
|
|
|
|
kfree(conreq);
|
|
|
|
|
|
|
|
spin_lock(&cep->lock);
|
|
|
|
|
|
|
|
if (cep->state == SCIFEP_CLOSING) {
|
|
|
|
/*
|
|
|
|
* Remote failed to allocate resources and NAKed the grant.
|
|
|
|
* There is at this point nothing referencing the new end point.
|
|
|
|
*/
|
|
|
|
spin_unlock(&cep->lock);
|
|
|
|
scif_teardown_ep(cep);
|
|
|
|
kfree(cep);
|
|
|
|
|
|
|
|
/* If call with sync flag then go back and wait. */
|
|
|
|
if (flags & SCIF_ACCEPT_SYNC) {
|
|
|
|
spin_lock(&lep->lock);
|
|
|
|
goto retry_connection;
|
|
|
|
}
|
|
|
|
return -EAGAIN;
|
|
|
|
}
|
|
|
|
|
|
|
|
scif_get_port(cep->port.port);
|
|
|
|
*newepd = (scif_epd_t)cep;
|
|
|
|
spin_unlock(&cep->lock);
|
|
|
|
return 0;
|
|
|
|
scif_accept_error_map:
|
|
|
|
scif_teardown_ep(cep);
|
|
|
|
scif_accept_error_qpalloc:
|
|
|
|
kfree(cep);
|
|
|
|
scif_accept_error_epalloc:
|
|
|
|
msg.uop = SCIF_CNCT_REJ;
|
|
|
|
msg.dst.node = conreq->msg.src.node;
|
|
|
|
msg.dst.port = conreq->msg.src.port;
|
|
|
|
msg.payload[0] = conreq->msg.payload[0];
|
|
|
|
msg.payload[1] = conreq->msg.payload[1];
|
|
|
|
scif_nodeqp_send(&scif_dev[conreq->msg.src.node], &msg);
|
|
|
|
kfree(conreq);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(scif_accept);
|