mirror of
https://github.com/torvalds/linux.git
synced 2024-11-13 23:51:39 +00:00
57af281e53
Tidy up the abort generation infrastructure in the following ways: (1) Create an enum and string mapping table to list the reasons an abort might be generated in tracing. (2) Replace the 3-char string with the values from (1) in the places that use that to log the abort source. This gets rid of a memcpy() in the tracepoint. (3) Subsume the rxrpc_rx_eproto tracepoint with the rxrpc_abort tracepoint and use values from (1) to indicate the trace reason. (4) Always make a call to an abort function at the point of the abort rather than stashing the values into variables and using goto to get to a place where it reported. The C optimiser will collapse the calls together as appropriate. The abort functions return a value that can be returned directly if appropriate. Note that this extends into afs also at the points where that generates an abort. To aid with this, the afs sources need to #define RXRPC_TRACE_ONLY_DEFINE_ENUMS before including the rxrpc tracing header because they don't have access to the rxrpc internal structures that some of the tracepoints make use of. Signed-off-by: David Howells <dhowells@redhat.com> cc: Marc Dionne <marc.dionne@auristor.com> cc: linux-afs@lists.infradead.org
627 lines
16 KiB
C
627 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/* In-kernel rxperf server for testing purposes.
|
|
*
|
|
* Copyright (C) 2022 Red Hat, Inc. All Rights Reserved.
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
*/
|
|
|
|
#define pr_fmt(fmt) "rxperf: " fmt
|
|
#include <linux/module.h>
|
|
#include <linux/slab.h>
|
|
#include <net/sock.h>
|
|
#include <net/af_rxrpc.h>
|
|
#define RXRPC_TRACE_ONLY_DEFINE_ENUMS
|
|
#include <trace/events/rxrpc.h>
|
|
|
|
MODULE_DESCRIPTION("rxperf test server (afs)");
|
|
MODULE_AUTHOR("Red Hat, Inc.");
|
|
MODULE_LICENSE("GPL");
|
|
|
|
#define RXPERF_PORT 7009
|
|
#define RX_PERF_SERVICE 147
|
|
#define RX_PERF_VERSION 3
|
|
#define RX_PERF_SEND 0
|
|
#define RX_PERF_RECV 1
|
|
#define RX_PERF_RPC 3
|
|
#define RX_PERF_FILE 4
|
|
#define RX_PERF_MAGIC_COOKIE 0x4711
|
|
|
|
struct rxperf_proto_params {
|
|
__be32 version;
|
|
__be32 type;
|
|
__be32 rsize;
|
|
__be32 wsize;
|
|
} __packed;
|
|
|
|
static const u8 rxperf_magic_cookie[] = { 0x00, 0x00, 0x47, 0x11 };
|
|
static const u8 secret[8] = { 0xa7, 0x83, 0x8a, 0xcb, 0xc7, 0x83, 0xec, 0x94 };
|
|
|
|
enum rxperf_call_state {
|
|
RXPERF_CALL_SV_AWAIT_PARAMS, /* Server: Awaiting parameter block */
|
|
RXPERF_CALL_SV_AWAIT_REQUEST, /* Server: Awaiting request data */
|
|
RXPERF_CALL_SV_REPLYING, /* Server: Replying */
|
|
RXPERF_CALL_SV_AWAIT_ACK, /* Server: Awaiting final ACK */
|
|
RXPERF_CALL_COMPLETE, /* Completed or failed */
|
|
};
|
|
|
|
struct rxperf_call {
|
|
struct rxrpc_call *rxcall;
|
|
struct iov_iter iter;
|
|
struct kvec kvec[1];
|
|
struct work_struct work;
|
|
const char *type;
|
|
size_t iov_len;
|
|
size_t req_len; /* Size of request blob */
|
|
size_t reply_len; /* Size of reply blob */
|
|
unsigned int debug_id;
|
|
unsigned int operation_id;
|
|
struct rxperf_proto_params params;
|
|
__be32 tmp[2];
|
|
s32 abort_code;
|
|
enum rxperf_call_state state;
|
|
short error;
|
|
unsigned short unmarshal;
|
|
u16 service_id;
|
|
int (*deliver)(struct rxperf_call *call);
|
|
void (*processor)(struct work_struct *work);
|
|
};
|
|
|
|
static struct socket *rxperf_socket;
|
|
static struct key *rxperf_sec_keyring; /* Ring of security/crypto keys */
|
|
static struct workqueue_struct *rxperf_workqueue;
|
|
|
|
static void rxperf_deliver_to_call(struct work_struct *work);
|
|
static int rxperf_deliver_param_block(struct rxperf_call *call);
|
|
static int rxperf_deliver_request(struct rxperf_call *call);
|
|
static int rxperf_process_call(struct rxperf_call *call);
|
|
static void rxperf_charge_preallocation(struct work_struct *work);
|
|
|
|
static DECLARE_WORK(rxperf_charge_preallocation_work,
|
|
rxperf_charge_preallocation);
|
|
|
|
static inline void rxperf_set_call_state(struct rxperf_call *call,
|
|
enum rxperf_call_state to)
|
|
{
|
|
call->state = to;
|
|
}
|
|
|
|
static inline void rxperf_set_call_complete(struct rxperf_call *call,
|
|
int error, s32 remote_abort)
|
|
{
|
|
if (call->state != RXPERF_CALL_COMPLETE) {
|
|
call->abort_code = remote_abort;
|
|
call->error = error;
|
|
call->state = RXPERF_CALL_COMPLETE;
|
|
}
|
|
}
|
|
|
|
static void rxperf_rx_discard_new_call(struct rxrpc_call *rxcall,
|
|
unsigned long user_call_ID)
|
|
{
|
|
kfree((struct rxperf_call *)user_call_ID);
|
|
}
|
|
|
|
static void rxperf_rx_new_call(struct sock *sk, struct rxrpc_call *rxcall,
|
|
unsigned long user_call_ID)
|
|
{
|
|
queue_work(rxperf_workqueue, &rxperf_charge_preallocation_work);
|
|
}
|
|
|
|
static void rxperf_queue_call_work(struct rxperf_call *call)
|
|
{
|
|
queue_work(rxperf_workqueue, &call->work);
|
|
}
|
|
|
|
static void rxperf_notify_rx(struct sock *sk, struct rxrpc_call *rxcall,
|
|
unsigned long call_user_ID)
|
|
{
|
|
struct rxperf_call *call = (struct rxperf_call *)call_user_ID;
|
|
|
|
if (call->state != RXPERF_CALL_COMPLETE)
|
|
rxperf_queue_call_work(call);
|
|
}
|
|
|
|
static void rxperf_rx_attach(struct rxrpc_call *rxcall, unsigned long user_call_ID)
|
|
{
|
|
struct rxperf_call *call = (struct rxperf_call *)user_call_ID;
|
|
|
|
call->rxcall = rxcall;
|
|
}
|
|
|
|
static void rxperf_notify_end_reply_tx(struct sock *sock,
|
|
struct rxrpc_call *rxcall,
|
|
unsigned long call_user_ID)
|
|
{
|
|
rxperf_set_call_state((struct rxperf_call *)call_user_ID,
|
|
RXPERF_CALL_SV_AWAIT_ACK);
|
|
}
|
|
|
|
/*
|
|
* Charge the incoming call preallocation.
|
|
*/
|
|
static void rxperf_charge_preallocation(struct work_struct *work)
|
|
{
|
|
struct rxperf_call *call;
|
|
|
|
for (;;) {
|
|
call = kzalloc(sizeof(*call), GFP_KERNEL);
|
|
if (!call)
|
|
break;
|
|
|
|
call->type = "unset";
|
|
call->debug_id = atomic_inc_return(&rxrpc_debug_id);
|
|
call->deliver = rxperf_deliver_param_block;
|
|
call->state = RXPERF_CALL_SV_AWAIT_PARAMS;
|
|
call->service_id = RX_PERF_SERVICE;
|
|
call->iov_len = sizeof(call->params);
|
|
call->kvec[0].iov_len = sizeof(call->params);
|
|
call->kvec[0].iov_base = &call->params;
|
|
iov_iter_kvec(&call->iter, READ, call->kvec, 1, call->iov_len);
|
|
INIT_WORK(&call->work, rxperf_deliver_to_call);
|
|
|
|
if (rxrpc_kernel_charge_accept(rxperf_socket,
|
|
rxperf_notify_rx,
|
|
rxperf_rx_attach,
|
|
(unsigned long)call,
|
|
GFP_KERNEL,
|
|
call->debug_id) < 0)
|
|
break;
|
|
call = NULL;
|
|
}
|
|
|
|
kfree(call);
|
|
}
|
|
|
|
/*
|
|
* Open an rxrpc socket and bind it to be a server for callback notifications
|
|
* - the socket is left in blocking mode and non-blocking ops use MSG_DONTWAIT
|
|
*/
|
|
static int rxperf_open_socket(void)
|
|
{
|
|
struct sockaddr_rxrpc srx;
|
|
struct socket *socket;
|
|
int ret;
|
|
|
|
ret = sock_create_kern(&init_net, AF_RXRPC, SOCK_DGRAM, PF_INET6,
|
|
&socket);
|
|
if (ret < 0)
|
|
goto error_1;
|
|
|
|
socket->sk->sk_allocation = GFP_NOFS;
|
|
|
|
/* bind the callback manager's address to make this a server socket */
|
|
memset(&srx, 0, sizeof(srx));
|
|
srx.srx_family = AF_RXRPC;
|
|
srx.srx_service = RX_PERF_SERVICE;
|
|
srx.transport_type = SOCK_DGRAM;
|
|
srx.transport_len = sizeof(srx.transport.sin6);
|
|
srx.transport.sin6.sin6_family = AF_INET6;
|
|
srx.transport.sin6.sin6_port = htons(RXPERF_PORT);
|
|
|
|
ret = rxrpc_sock_set_min_security_level(socket->sk,
|
|
RXRPC_SECURITY_ENCRYPT);
|
|
if (ret < 0)
|
|
goto error_2;
|
|
|
|
ret = rxrpc_sock_set_security_keyring(socket->sk, rxperf_sec_keyring);
|
|
|
|
ret = kernel_bind(socket, (struct sockaddr *)&srx, sizeof(srx));
|
|
if (ret < 0)
|
|
goto error_2;
|
|
|
|
rxrpc_kernel_new_call_notification(socket, rxperf_rx_new_call,
|
|
rxperf_rx_discard_new_call);
|
|
|
|
ret = kernel_listen(socket, INT_MAX);
|
|
if (ret < 0)
|
|
goto error_2;
|
|
|
|
rxperf_socket = socket;
|
|
rxperf_charge_preallocation(&rxperf_charge_preallocation_work);
|
|
return 0;
|
|
|
|
error_2:
|
|
sock_release(socket);
|
|
error_1:
|
|
pr_err("Can't set up rxperf socket: %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* close the rxrpc socket rxperf was using
|
|
*/
|
|
static void rxperf_close_socket(void)
|
|
{
|
|
kernel_listen(rxperf_socket, 0);
|
|
kernel_sock_shutdown(rxperf_socket, SHUT_RDWR);
|
|
flush_workqueue(rxperf_workqueue);
|
|
sock_release(rxperf_socket);
|
|
}
|
|
|
|
/*
|
|
* Log remote abort codes that indicate that we have a protocol disagreement
|
|
* with the server.
|
|
*/
|
|
static void rxperf_log_error(struct rxperf_call *call, s32 remote_abort)
|
|
{
|
|
static int max = 0;
|
|
const char *msg;
|
|
int m;
|
|
|
|
switch (remote_abort) {
|
|
case RX_EOF: msg = "unexpected EOF"; break;
|
|
case RXGEN_CC_MARSHAL: msg = "client marshalling"; break;
|
|
case RXGEN_CC_UNMARSHAL: msg = "client unmarshalling"; break;
|
|
case RXGEN_SS_MARSHAL: msg = "server marshalling"; break;
|
|
case RXGEN_SS_UNMARSHAL: msg = "server unmarshalling"; break;
|
|
case RXGEN_DECODE: msg = "opcode decode"; break;
|
|
case RXGEN_SS_XDRFREE: msg = "server XDR cleanup"; break;
|
|
case RXGEN_CC_XDRFREE: msg = "client XDR cleanup"; break;
|
|
case -32: msg = "insufficient data"; break;
|
|
default:
|
|
return;
|
|
}
|
|
|
|
m = max;
|
|
if (m < 3) {
|
|
max = m + 1;
|
|
pr_info("Peer reported %s failure on %s\n", msg, call->type);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* deliver messages to a call
|
|
*/
|
|
static void rxperf_deliver_to_call(struct work_struct *work)
|
|
{
|
|
struct rxperf_call *call = container_of(work, struct rxperf_call, work);
|
|
enum rxperf_call_state state;
|
|
u32 abort_code, remote_abort = 0;
|
|
int ret = 0;
|
|
|
|
if (call->state == RXPERF_CALL_COMPLETE)
|
|
return;
|
|
|
|
while (state = call->state,
|
|
state == RXPERF_CALL_SV_AWAIT_PARAMS ||
|
|
state == RXPERF_CALL_SV_AWAIT_REQUEST ||
|
|
state == RXPERF_CALL_SV_AWAIT_ACK
|
|
) {
|
|
if (state == RXPERF_CALL_SV_AWAIT_ACK) {
|
|
if (!rxrpc_kernel_check_life(rxperf_socket, call->rxcall))
|
|
goto call_complete;
|
|
return;
|
|
}
|
|
|
|
ret = call->deliver(call);
|
|
if (ret == 0)
|
|
ret = rxperf_process_call(call);
|
|
|
|
switch (ret) {
|
|
case 0:
|
|
continue;
|
|
case -EINPROGRESS:
|
|
case -EAGAIN:
|
|
return;
|
|
case -ECONNABORTED:
|
|
rxperf_log_error(call, call->abort_code);
|
|
goto call_complete;
|
|
case -EOPNOTSUPP:
|
|
abort_code = RXGEN_OPCODE;
|
|
rxrpc_kernel_abort_call(rxperf_socket, call->rxcall,
|
|
abort_code, ret,
|
|
rxperf_abort_op_not_supported);
|
|
goto call_complete;
|
|
case -ENOTSUPP:
|
|
abort_code = RX_USER_ABORT;
|
|
rxrpc_kernel_abort_call(rxperf_socket, call->rxcall,
|
|
abort_code, ret,
|
|
rxperf_abort_op_not_supported);
|
|
goto call_complete;
|
|
case -EIO:
|
|
pr_err("Call %u in bad state %u\n",
|
|
call->debug_id, call->state);
|
|
fallthrough;
|
|
case -ENODATA:
|
|
case -EBADMSG:
|
|
case -EMSGSIZE:
|
|
case -ENOMEM:
|
|
case -EFAULT:
|
|
rxrpc_kernel_abort_call(rxperf_socket, call->rxcall,
|
|
RXGEN_SS_UNMARSHAL, ret,
|
|
rxperf_abort_unmarshal_error);
|
|
goto call_complete;
|
|
default:
|
|
rxrpc_kernel_abort_call(rxperf_socket, call->rxcall,
|
|
RX_CALL_DEAD, ret,
|
|
rxperf_abort_general_error);
|
|
goto call_complete;
|
|
}
|
|
}
|
|
|
|
call_complete:
|
|
rxperf_set_call_complete(call, ret, remote_abort);
|
|
/* The call may have been requeued */
|
|
rxrpc_kernel_end_call(rxperf_socket, call->rxcall);
|
|
cancel_work(&call->work);
|
|
kfree(call);
|
|
}
|
|
|
|
/*
|
|
* Extract a piece of data from the received data socket buffers.
|
|
*/
|
|
static int rxperf_extract_data(struct rxperf_call *call, bool want_more)
|
|
{
|
|
u32 remote_abort = 0;
|
|
int ret;
|
|
|
|
ret = rxrpc_kernel_recv_data(rxperf_socket, call->rxcall, &call->iter,
|
|
&call->iov_len, want_more, &remote_abort,
|
|
&call->service_id);
|
|
pr_debug("Extract i=%zu l=%zu m=%u ret=%d\n",
|
|
iov_iter_count(&call->iter), call->iov_len, want_more, ret);
|
|
if (ret == 0 || ret == -EAGAIN)
|
|
return ret;
|
|
|
|
if (ret == 1) {
|
|
switch (call->state) {
|
|
case RXPERF_CALL_SV_AWAIT_REQUEST:
|
|
rxperf_set_call_state(call, RXPERF_CALL_SV_REPLYING);
|
|
break;
|
|
case RXPERF_CALL_COMPLETE:
|
|
pr_debug("premature completion %d", call->error);
|
|
return call->error;
|
|
default:
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
rxperf_set_call_complete(call, ret, remote_abort);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Grab the operation ID from an incoming manager call.
|
|
*/
|
|
static int rxperf_deliver_param_block(struct rxperf_call *call)
|
|
{
|
|
u32 version;
|
|
int ret;
|
|
|
|
/* Extract the parameter block */
|
|
ret = rxperf_extract_data(call, true);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
version = ntohl(call->params.version);
|
|
call->operation_id = ntohl(call->params.type);
|
|
call->deliver = rxperf_deliver_request;
|
|
|
|
if (version != RX_PERF_VERSION) {
|
|
pr_info("Version mismatch %x\n", version);
|
|
return -ENOTSUPP;
|
|
}
|
|
|
|
switch (call->operation_id) {
|
|
case RX_PERF_SEND:
|
|
call->type = "send";
|
|
call->reply_len = 0;
|
|
call->iov_len = 4; /* Expect req size */
|
|
break;
|
|
case RX_PERF_RECV:
|
|
call->type = "recv";
|
|
call->req_len = 0;
|
|
call->iov_len = 4; /* Expect reply size */
|
|
break;
|
|
case RX_PERF_RPC:
|
|
call->type = "rpc";
|
|
call->iov_len = 8; /* Expect req size and reply size */
|
|
break;
|
|
case RX_PERF_FILE:
|
|
call->type = "file";
|
|
fallthrough;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
rxperf_set_call_state(call, RXPERF_CALL_SV_AWAIT_REQUEST);
|
|
return call->deliver(call);
|
|
}
|
|
|
|
/*
|
|
* Deliver the request data.
|
|
*/
|
|
static int rxperf_deliver_request(struct rxperf_call *call)
|
|
{
|
|
int ret;
|
|
|
|
switch (call->unmarshal) {
|
|
case 0:
|
|
call->kvec[0].iov_len = call->iov_len;
|
|
call->kvec[0].iov_base = call->tmp;
|
|
iov_iter_kvec(&call->iter, READ, call->kvec, 1, call->iov_len);
|
|
call->unmarshal++;
|
|
fallthrough;
|
|
case 1:
|
|
ret = rxperf_extract_data(call, true);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
switch (call->operation_id) {
|
|
case RX_PERF_SEND:
|
|
call->type = "send";
|
|
call->req_len = ntohl(call->tmp[0]);
|
|
call->reply_len = 0;
|
|
break;
|
|
case RX_PERF_RECV:
|
|
call->type = "recv";
|
|
call->req_len = 0;
|
|
call->reply_len = ntohl(call->tmp[0]);
|
|
break;
|
|
case RX_PERF_RPC:
|
|
call->type = "rpc";
|
|
call->req_len = ntohl(call->tmp[0]);
|
|
call->reply_len = ntohl(call->tmp[1]);
|
|
break;
|
|
default:
|
|
pr_info("Can't parse extra params\n");
|
|
return -EIO;
|
|
}
|
|
|
|
pr_debug("CALL op=%s rq=%zx rp=%zx\n",
|
|
call->type, call->req_len, call->reply_len);
|
|
|
|
call->iov_len = call->req_len;
|
|
iov_iter_discard(&call->iter, READ, call->req_len);
|
|
call->unmarshal++;
|
|
fallthrough;
|
|
case 2:
|
|
ret = rxperf_extract_data(call, false);
|
|
if (ret < 0)
|
|
return ret;
|
|
call->unmarshal++;
|
|
fallthrough;
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Process a call for which we've received the request.
|
|
*/
|
|
static int rxperf_process_call(struct rxperf_call *call)
|
|
{
|
|
struct msghdr msg = {};
|
|
struct bio_vec bv[1];
|
|
struct kvec iov[1];
|
|
ssize_t n;
|
|
size_t reply_len = call->reply_len, len;
|
|
|
|
rxrpc_kernel_set_tx_length(rxperf_socket, call->rxcall,
|
|
reply_len + sizeof(rxperf_magic_cookie));
|
|
|
|
while (reply_len > 0) {
|
|
len = min_t(size_t, reply_len, PAGE_SIZE);
|
|
bv[0].bv_page = ZERO_PAGE(0);
|
|
bv[0].bv_offset = 0;
|
|
bv[0].bv_len = len;
|
|
iov_iter_bvec(&msg.msg_iter, WRITE, bv, 1, len);
|
|
msg.msg_flags = MSG_MORE;
|
|
n = rxrpc_kernel_send_data(rxperf_socket, call->rxcall, &msg,
|
|
len, rxperf_notify_end_reply_tx);
|
|
if (n < 0)
|
|
return n;
|
|
if (n == 0)
|
|
return -EIO;
|
|
reply_len -= n;
|
|
}
|
|
|
|
len = sizeof(rxperf_magic_cookie);
|
|
iov[0].iov_base = (void *)rxperf_magic_cookie;
|
|
iov[0].iov_len = len;
|
|
iov_iter_kvec(&msg.msg_iter, WRITE, iov, 1, len);
|
|
msg.msg_flags = 0;
|
|
n = rxrpc_kernel_send_data(rxperf_socket, call->rxcall, &msg, len,
|
|
rxperf_notify_end_reply_tx);
|
|
if (n >= 0)
|
|
return 0; /* Success */
|
|
|
|
if (n == -ENOMEM)
|
|
rxrpc_kernel_abort_call(rxperf_socket, call->rxcall,
|
|
RXGEN_SS_MARSHAL, -ENOMEM,
|
|
rxperf_abort_oom);
|
|
return n;
|
|
}
|
|
|
|
/*
|
|
* Add a key to the security keyring.
|
|
*/
|
|
static int rxperf_add_key(struct key *keyring)
|
|
{
|
|
key_ref_t kref;
|
|
int ret;
|
|
|
|
kref = key_create_or_update(make_key_ref(keyring, true),
|
|
"rxrpc_s",
|
|
__stringify(RX_PERF_SERVICE) ":2",
|
|
secret,
|
|
sizeof(secret),
|
|
KEY_POS_VIEW | KEY_POS_READ | KEY_POS_SEARCH
|
|
| KEY_USR_VIEW,
|
|
KEY_ALLOC_NOT_IN_QUOTA);
|
|
|
|
if (IS_ERR(kref)) {
|
|
pr_err("Can't allocate rxperf server key: %ld\n", PTR_ERR(kref));
|
|
return PTR_ERR(kref);
|
|
}
|
|
|
|
ret = key_link(keyring, key_ref_to_ptr(kref));
|
|
if (ret < 0)
|
|
pr_err("Can't link rxperf server key: %d\n", ret);
|
|
key_ref_put(kref);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Initialise the rxperf server.
|
|
*/
|
|
static int __init rxperf_init(void)
|
|
{
|
|
struct key *keyring;
|
|
int ret = -ENOMEM;
|
|
|
|
pr_info("Server registering\n");
|
|
|
|
rxperf_workqueue = alloc_workqueue("rxperf", 0, 0);
|
|
if (!rxperf_workqueue)
|
|
goto error_workqueue;
|
|
|
|
keyring = keyring_alloc("rxperf_server",
|
|
GLOBAL_ROOT_UID, GLOBAL_ROOT_GID, current_cred(),
|
|
KEY_POS_VIEW | KEY_POS_READ | KEY_POS_SEARCH |
|
|
KEY_POS_WRITE |
|
|
KEY_USR_VIEW | KEY_USR_READ | KEY_USR_SEARCH |
|
|
KEY_USR_WRITE |
|
|
KEY_OTH_VIEW | KEY_OTH_READ | KEY_OTH_SEARCH,
|
|
KEY_ALLOC_NOT_IN_QUOTA,
|
|
NULL, NULL);
|
|
if (IS_ERR(keyring)) {
|
|
pr_err("Can't allocate rxperf server keyring: %ld\n",
|
|
PTR_ERR(keyring));
|
|
goto error_keyring;
|
|
}
|
|
rxperf_sec_keyring = keyring;
|
|
ret = rxperf_add_key(keyring);
|
|
if (ret < 0)
|
|
goto error_key;
|
|
|
|
ret = rxperf_open_socket();
|
|
if (ret < 0)
|
|
goto error_socket;
|
|
return 0;
|
|
|
|
error_socket:
|
|
error_key:
|
|
key_put(rxperf_sec_keyring);
|
|
error_keyring:
|
|
destroy_workqueue(rxperf_workqueue);
|
|
rcu_barrier();
|
|
error_workqueue:
|
|
pr_err("Failed to register: %d\n", ret);
|
|
return ret;
|
|
}
|
|
late_initcall(rxperf_init); /* Must be called after net/ to create socket */
|
|
|
|
static void __exit rxperf_exit(void)
|
|
{
|
|
pr_info("Server unregistering.\n");
|
|
|
|
rxperf_close_socket();
|
|
key_put(rxperf_sec_keyring);
|
|
destroy_workqueue(rxperf_workqueue);
|
|
rcu_barrier();
|
|
}
|
|
module_exit(rxperf_exit);
|
|
|