mirror of
https://github.com/torvalds/linux.git
synced 2024-12-27 13:22:23 +00:00
a0c313d08d
The .remove() callback for a platform driver returns an int which makes many driver authors wrongly assume it's possible to do error handling by returning an error code. However the value returned is ignored (apart from emitting a warning) and this typically results in resource leaks. To improve here there is a quest to make the remove callback return void. In the first step of this quest all drivers are converted to .remove_new(), which already returns void. Eventually after all drivers are converted, .remove_new() will be renamed to .remove(). Trivially convert this driver from always returning zero in the remove callback to the void returning variant. Signed-off-by: Uwe Kleine-König <u.kleine-koenig@pengutronix.de> Signed-off-by: Jassi Brar <jaswinder.singh@linaro.org>
1083 lines
27 KiB
C
1083 lines
27 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (c) 2018 Pengutronix, Oleksij Rempel <o.rempel@pengutronix.de>
|
|
* Copyright 2022 NXP, Peng Fan <peng.fan@nxp.com>
|
|
*/
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/firmware/imx/ipc.h>
|
|
#include <linux/firmware/imx/s4.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/io.h>
|
|
#include <linux/iopoll.h>
|
|
#include <linux/jiffies.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mailbox_controller.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/suspend.h>
|
|
#include <linux/slab.h>
|
|
|
|
#include "mailbox.h"
|
|
|
|
#define IMX_MU_CHANS 24
|
|
/* TX0/RX0/RXDB[0-3] */
|
|
#define IMX_MU_SCU_CHANS 6
|
|
/* TX0/RX0 */
|
|
#define IMX_MU_S4_CHANS 2
|
|
#define IMX_MU_CHAN_NAME_SIZE 20
|
|
|
|
#define IMX_MU_NUM_RR 4
|
|
|
|
#define IMX_MU_SECO_TX_TOUT (msecs_to_jiffies(3000))
|
|
#define IMX_MU_SECO_RX_TOUT (msecs_to_jiffies(3000))
|
|
|
|
/* Please not change TX & RX */
|
|
enum imx_mu_chan_type {
|
|
IMX_MU_TYPE_TX = 0, /* Tx */
|
|
IMX_MU_TYPE_RX = 1, /* Rx */
|
|
IMX_MU_TYPE_TXDB = 2, /* Tx doorbell */
|
|
IMX_MU_TYPE_RXDB = 3, /* Rx doorbell */
|
|
IMX_MU_TYPE_RST = 4, /* Reset */
|
|
IMX_MU_TYPE_TXDB_V2 = 5, /* Tx doorbell with S/W ACK */
|
|
};
|
|
|
|
enum imx_mu_xcr {
|
|
IMX_MU_CR,
|
|
IMX_MU_GIER,
|
|
IMX_MU_GCR,
|
|
IMX_MU_TCR,
|
|
IMX_MU_RCR,
|
|
IMX_MU_xCR_MAX,
|
|
};
|
|
|
|
enum imx_mu_xsr {
|
|
IMX_MU_SR,
|
|
IMX_MU_GSR,
|
|
IMX_MU_TSR,
|
|
IMX_MU_RSR,
|
|
IMX_MU_xSR_MAX,
|
|
};
|
|
|
|
struct imx_sc_rpc_msg_max {
|
|
struct imx_sc_rpc_msg hdr;
|
|
u32 data[30];
|
|
};
|
|
|
|
struct imx_s4_rpc_msg_max {
|
|
struct imx_s4_rpc_msg hdr;
|
|
u32 data[254];
|
|
};
|
|
|
|
struct imx_mu_con_priv {
|
|
unsigned int idx;
|
|
char irq_desc[IMX_MU_CHAN_NAME_SIZE];
|
|
enum imx_mu_chan_type type;
|
|
struct mbox_chan *chan;
|
|
struct tasklet_struct txdb_tasklet;
|
|
};
|
|
|
|
struct imx_mu_priv {
|
|
struct device *dev;
|
|
void __iomem *base;
|
|
void *msg;
|
|
spinlock_t xcr_lock; /* control register lock */
|
|
|
|
struct mbox_controller mbox;
|
|
struct mbox_chan mbox_chans[IMX_MU_CHANS];
|
|
|
|
struct imx_mu_con_priv con_priv[IMX_MU_CHANS];
|
|
const struct imx_mu_dcfg *dcfg;
|
|
struct clk *clk;
|
|
int irq[IMX_MU_CHANS];
|
|
bool suspend;
|
|
|
|
u32 xcr[IMX_MU_xCR_MAX];
|
|
|
|
bool side_b;
|
|
};
|
|
|
|
enum imx_mu_type {
|
|
IMX_MU_V1,
|
|
IMX_MU_V2 = BIT(1),
|
|
IMX_MU_V2_S4 = BIT(15),
|
|
IMX_MU_V2_IRQ = BIT(16),
|
|
};
|
|
|
|
struct imx_mu_dcfg {
|
|
int (*tx)(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp, void *data);
|
|
int (*rx)(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp);
|
|
int (*rxdb)(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp);
|
|
void (*init)(struct imx_mu_priv *priv);
|
|
enum imx_mu_type type;
|
|
u32 xTR; /* Transmit Register0 */
|
|
u32 xRR; /* Receive Register0 */
|
|
u32 xSR[IMX_MU_xSR_MAX]; /* Status Registers */
|
|
u32 xCR[IMX_MU_xCR_MAX]; /* Control Registers */
|
|
};
|
|
|
|
#define IMX_MU_xSR_GIPn(type, x) (type & IMX_MU_V2 ? BIT(x) : BIT(28 + (3 - (x))))
|
|
#define IMX_MU_xSR_RFn(type, x) (type & IMX_MU_V2 ? BIT(x) : BIT(24 + (3 - (x))))
|
|
#define IMX_MU_xSR_TEn(type, x) (type & IMX_MU_V2 ? BIT(x) : BIT(20 + (3 - (x))))
|
|
|
|
/* General Purpose Interrupt Enable */
|
|
#define IMX_MU_xCR_GIEn(type, x) (type & IMX_MU_V2 ? BIT(x) : BIT(28 + (3 - (x))))
|
|
/* Receive Interrupt Enable */
|
|
#define IMX_MU_xCR_RIEn(type, x) (type & IMX_MU_V2 ? BIT(x) : BIT(24 + (3 - (x))))
|
|
/* Transmit Interrupt Enable */
|
|
#define IMX_MU_xCR_TIEn(type, x) (type & IMX_MU_V2 ? BIT(x) : BIT(20 + (3 - (x))))
|
|
/* General Purpose Interrupt Request */
|
|
#define IMX_MU_xCR_GIRn(type, x) (type & IMX_MU_V2 ? BIT(x) : BIT(16 + (3 - (x))))
|
|
/* MU reset */
|
|
#define IMX_MU_xCR_RST(type) (type & IMX_MU_V2 ? BIT(0) : BIT(5))
|
|
#define IMX_MU_xSR_RST(type) (type & IMX_MU_V2 ? BIT(0) : BIT(7))
|
|
|
|
|
|
static struct imx_mu_priv *to_imx_mu_priv(struct mbox_controller *mbox)
|
|
{
|
|
return container_of(mbox, struct imx_mu_priv, mbox);
|
|
}
|
|
|
|
static void imx_mu_write(struct imx_mu_priv *priv, u32 val, u32 offs)
|
|
{
|
|
iowrite32(val, priv->base + offs);
|
|
}
|
|
|
|
static u32 imx_mu_read(struct imx_mu_priv *priv, u32 offs)
|
|
{
|
|
return ioread32(priv->base + offs);
|
|
}
|
|
|
|
static int imx_mu_tx_waiting_write(struct imx_mu_priv *priv, u32 val, u32 idx)
|
|
{
|
|
u64 timeout_time = get_jiffies_64() + IMX_MU_SECO_TX_TOUT;
|
|
u32 status;
|
|
u32 can_write;
|
|
|
|
dev_dbg(priv->dev, "Trying to write %.8x to idx %d\n", val, idx);
|
|
|
|
do {
|
|
status = imx_mu_read(priv, priv->dcfg->xSR[IMX_MU_TSR]);
|
|
can_write = status & IMX_MU_xSR_TEn(priv->dcfg->type, idx % 4);
|
|
} while (!can_write && time_is_after_jiffies64(timeout_time));
|
|
|
|
if (!can_write) {
|
|
dev_err(priv->dev, "timeout trying to write %.8x at %d(%.8x)\n",
|
|
val, idx, status);
|
|
return -ETIME;
|
|
}
|
|
|
|
imx_mu_write(priv, val, priv->dcfg->xTR + (idx % 4) * 4);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_mu_rx_waiting_read(struct imx_mu_priv *priv, u32 *val, u32 idx)
|
|
{
|
|
u64 timeout_time = get_jiffies_64() + IMX_MU_SECO_RX_TOUT;
|
|
u32 status;
|
|
u32 can_read;
|
|
|
|
dev_dbg(priv->dev, "Trying to read from idx %d\n", idx);
|
|
|
|
do {
|
|
status = imx_mu_read(priv, priv->dcfg->xSR[IMX_MU_RSR]);
|
|
can_read = status & IMX_MU_xSR_RFn(priv->dcfg->type, idx % 4);
|
|
} while (!can_read && time_is_after_jiffies64(timeout_time));
|
|
|
|
if (!can_read) {
|
|
dev_err(priv->dev, "timeout trying to read idx %d (%.8x)\n",
|
|
idx, status);
|
|
return -ETIME;
|
|
}
|
|
|
|
*val = imx_mu_read(priv, priv->dcfg->xRR + (idx % 4) * 4);
|
|
dev_dbg(priv->dev, "Read %.8x\n", *val);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static u32 imx_mu_xcr_rmw(struct imx_mu_priv *priv, enum imx_mu_xcr type, u32 set, u32 clr)
|
|
{
|
|
unsigned long flags;
|
|
u32 val;
|
|
|
|
spin_lock_irqsave(&priv->xcr_lock, flags);
|
|
val = imx_mu_read(priv, priv->dcfg->xCR[type]);
|
|
val &= ~clr;
|
|
val |= set;
|
|
imx_mu_write(priv, val, priv->dcfg->xCR[type]);
|
|
spin_unlock_irqrestore(&priv->xcr_lock, flags);
|
|
|
|
return val;
|
|
}
|
|
|
|
static int imx_mu_generic_tx(struct imx_mu_priv *priv,
|
|
struct imx_mu_con_priv *cp,
|
|
void *data)
|
|
{
|
|
u32 *arg = data;
|
|
|
|
switch (cp->type) {
|
|
case IMX_MU_TYPE_TX:
|
|
imx_mu_write(priv, *arg, priv->dcfg->xTR + cp->idx * 4);
|
|
imx_mu_xcr_rmw(priv, IMX_MU_TCR, IMX_MU_xCR_TIEn(priv->dcfg->type, cp->idx), 0);
|
|
break;
|
|
case IMX_MU_TYPE_TXDB:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_GCR, IMX_MU_xCR_GIRn(priv->dcfg->type, cp->idx), 0);
|
|
tasklet_schedule(&cp->txdb_tasklet);
|
|
break;
|
|
case IMX_MU_TYPE_TXDB_V2:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_GCR, IMX_MU_xCR_GIRn(priv->dcfg->type, cp->idx), 0);
|
|
break;
|
|
default:
|
|
dev_warn_ratelimited(priv->dev, "Send data on wrong channel type: %d\n", cp->type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_mu_generic_rx(struct imx_mu_priv *priv,
|
|
struct imx_mu_con_priv *cp)
|
|
{
|
|
u32 dat;
|
|
|
|
dat = imx_mu_read(priv, priv->dcfg->xRR + (cp->idx) * 4);
|
|
mbox_chan_received_data(cp->chan, (void *)&dat);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_mu_generic_rxdb(struct imx_mu_priv *priv,
|
|
struct imx_mu_con_priv *cp)
|
|
{
|
|
imx_mu_write(priv, IMX_MU_xSR_GIPn(priv->dcfg->type, cp->idx),
|
|
priv->dcfg->xSR[IMX_MU_GSR]);
|
|
mbox_chan_received_data(cp->chan, NULL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_mu_specific_tx(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp, void *data)
|
|
{
|
|
u32 *arg = data;
|
|
int i, ret;
|
|
u32 xsr;
|
|
u32 size, max_size, num_tr;
|
|
|
|
if (priv->dcfg->type & IMX_MU_V2_S4) {
|
|
size = ((struct imx_s4_rpc_msg_max *)data)->hdr.size;
|
|
max_size = sizeof(struct imx_s4_rpc_msg_max);
|
|
num_tr = 8;
|
|
} else {
|
|
size = ((struct imx_sc_rpc_msg_max *)data)->hdr.size;
|
|
max_size = sizeof(struct imx_sc_rpc_msg_max);
|
|
num_tr = 4;
|
|
}
|
|
|
|
switch (cp->type) {
|
|
case IMX_MU_TYPE_TX:
|
|
/*
|
|
* msg->hdr.size specifies the number of u32 words while
|
|
* sizeof yields bytes.
|
|
*/
|
|
|
|
if (size > max_size / 4) {
|
|
/*
|
|
* The real message size can be different to
|
|
* struct imx_sc_rpc_msg_max/imx_s4_rpc_msg_max size
|
|
*/
|
|
dev_err(priv->dev, "Maximal message size (%u bytes) exceeded on TX; got: %i bytes\n", max_size, size << 2);
|
|
return -EINVAL;
|
|
}
|
|
|
|
for (i = 0; i < num_tr && i < size; i++)
|
|
imx_mu_write(priv, *arg++, priv->dcfg->xTR + (i % num_tr) * 4);
|
|
for (; i < size; i++) {
|
|
ret = readl_poll_timeout(priv->base + priv->dcfg->xSR[IMX_MU_TSR],
|
|
xsr,
|
|
xsr & IMX_MU_xSR_TEn(priv->dcfg->type, i % num_tr),
|
|
0, 5 * USEC_PER_SEC);
|
|
if (ret) {
|
|
dev_err(priv->dev, "Send data index: %d timeout\n", i);
|
|
return ret;
|
|
}
|
|
imx_mu_write(priv, *arg++, priv->dcfg->xTR + (i % num_tr) * 4);
|
|
}
|
|
|
|
imx_mu_xcr_rmw(priv, IMX_MU_TCR, IMX_MU_xCR_TIEn(priv->dcfg->type, cp->idx), 0);
|
|
break;
|
|
default:
|
|
dev_warn_ratelimited(priv->dev, "Send data on wrong channel type: %d\n", cp->type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_mu_specific_rx(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp)
|
|
{
|
|
u32 *data;
|
|
int i, ret;
|
|
u32 xsr;
|
|
u32 size, max_size;
|
|
|
|
data = (u32 *)priv->msg;
|
|
|
|
imx_mu_xcr_rmw(priv, IMX_MU_RCR, 0, IMX_MU_xCR_RIEn(priv->dcfg->type, 0));
|
|
*data++ = imx_mu_read(priv, priv->dcfg->xRR);
|
|
|
|
if (priv->dcfg->type & IMX_MU_V2_S4) {
|
|
size = ((struct imx_s4_rpc_msg_max *)priv->msg)->hdr.size;
|
|
max_size = sizeof(struct imx_s4_rpc_msg_max);
|
|
} else {
|
|
size = ((struct imx_sc_rpc_msg_max *)priv->msg)->hdr.size;
|
|
max_size = sizeof(struct imx_sc_rpc_msg_max);
|
|
}
|
|
|
|
if (size > max_size / 4) {
|
|
dev_err(priv->dev, "Maximal message size (%u bytes) exceeded on RX; got: %i bytes\n", max_size, size << 2);
|
|
return -EINVAL;
|
|
}
|
|
|
|
for (i = 1; i < size; i++) {
|
|
ret = readl_poll_timeout(priv->base + priv->dcfg->xSR[IMX_MU_RSR], xsr,
|
|
xsr & IMX_MU_xSR_RFn(priv->dcfg->type, i % 4), 0,
|
|
5 * USEC_PER_SEC);
|
|
if (ret) {
|
|
dev_err(priv->dev, "timeout read idx %d\n", i);
|
|
return ret;
|
|
}
|
|
*data++ = imx_mu_read(priv, priv->dcfg->xRR + (i % 4) * 4);
|
|
}
|
|
|
|
imx_mu_xcr_rmw(priv, IMX_MU_RCR, IMX_MU_xCR_RIEn(priv->dcfg->type, 0), 0);
|
|
mbox_chan_received_data(cp->chan, (void *)priv->msg);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_mu_seco_tx(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp,
|
|
void *data)
|
|
{
|
|
struct imx_sc_rpc_msg_max *msg = data;
|
|
u32 *arg = data;
|
|
u32 byte_size;
|
|
int err;
|
|
int i;
|
|
|
|
dev_dbg(priv->dev, "Sending message\n");
|
|
|
|
switch (cp->type) {
|
|
case IMX_MU_TYPE_TXDB:
|
|
byte_size = msg->hdr.size * sizeof(u32);
|
|
if (byte_size > sizeof(*msg)) {
|
|
/*
|
|
* The real message size can be different to
|
|
* struct imx_sc_rpc_msg_max size
|
|
*/
|
|
dev_err(priv->dev,
|
|
"Exceed max msg size (%zu) on TX, got: %i\n",
|
|
sizeof(*msg), byte_size);
|
|
return -EINVAL;
|
|
}
|
|
|
|
print_hex_dump_debug("from client ", DUMP_PREFIX_OFFSET, 4, 4,
|
|
data, byte_size, false);
|
|
|
|
/* Send first word */
|
|
dev_dbg(priv->dev, "Sending header\n");
|
|
imx_mu_write(priv, *arg++, priv->dcfg->xTR);
|
|
|
|
/* Send signaling */
|
|
dev_dbg(priv->dev, "Sending signaling\n");
|
|
imx_mu_xcr_rmw(priv, IMX_MU_GCR,
|
|
IMX_MU_xCR_GIRn(priv->dcfg->type, cp->idx), 0);
|
|
|
|
/* Send words to fill the mailbox */
|
|
for (i = 1; i < 4 && i < msg->hdr.size; i++) {
|
|
dev_dbg(priv->dev, "Sending word %d\n", i);
|
|
imx_mu_write(priv, *arg++,
|
|
priv->dcfg->xTR + (i % 4) * 4);
|
|
}
|
|
|
|
/* Send rest of message waiting for remote read */
|
|
for (; i < msg->hdr.size; i++) {
|
|
dev_dbg(priv->dev, "Sending word %d\n", i);
|
|
err = imx_mu_tx_waiting_write(priv, *arg++, i);
|
|
if (err) {
|
|
dev_err(priv->dev, "Timeout tx %d\n", i);
|
|
return err;
|
|
}
|
|
}
|
|
|
|
/* Simulate hack for mbox framework */
|
|
tasklet_schedule(&cp->txdb_tasklet);
|
|
|
|
break;
|
|
default:
|
|
dev_warn_ratelimited(priv->dev,
|
|
"Send data on wrong channel type: %d\n",
|
|
cp->type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_mu_seco_rxdb(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp)
|
|
{
|
|
struct imx_sc_rpc_msg_max msg;
|
|
u32 *data = (u32 *)&msg;
|
|
u32 byte_size;
|
|
int err = 0;
|
|
int i;
|
|
|
|
dev_dbg(priv->dev, "Receiving message\n");
|
|
|
|
/* Read header */
|
|
dev_dbg(priv->dev, "Receiving header\n");
|
|
*data++ = imx_mu_read(priv, priv->dcfg->xRR);
|
|
byte_size = msg.hdr.size * sizeof(u32);
|
|
if (byte_size > sizeof(msg)) {
|
|
dev_err(priv->dev, "Exceed max msg size (%zu) on RX, got: %i\n",
|
|
sizeof(msg), byte_size);
|
|
err = -EINVAL;
|
|
goto error;
|
|
}
|
|
|
|
/* Read message waiting they are written */
|
|
for (i = 1; i < msg.hdr.size; i++) {
|
|
dev_dbg(priv->dev, "Receiving word %d\n", i);
|
|
err = imx_mu_rx_waiting_read(priv, data++, i);
|
|
if (err) {
|
|
dev_err(priv->dev, "Timeout rx %d\n", i);
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
/* Clear GIP */
|
|
imx_mu_write(priv, IMX_MU_xSR_GIPn(priv->dcfg->type, cp->idx),
|
|
priv->dcfg->xSR[IMX_MU_GSR]);
|
|
|
|
print_hex_dump_debug("to client ", DUMP_PREFIX_OFFSET, 4, 4,
|
|
&msg, byte_size, false);
|
|
|
|
/* send data to client */
|
|
dev_dbg(priv->dev, "Sending message to client\n");
|
|
mbox_chan_received_data(cp->chan, (void *)&msg);
|
|
|
|
goto exit;
|
|
|
|
error:
|
|
mbox_chan_received_data(cp->chan, ERR_PTR(err));
|
|
|
|
exit:
|
|
return err;
|
|
}
|
|
|
|
static void imx_mu_txdb_tasklet(unsigned long data)
|
|
{
|
|
struct imx_mu_con_priv *cp = (struct imx_mu_con_priv *)data;
|
|
|
|
mbox_chan_txdone(cp->chan, 0);
|
|
}
|
|
|
|
static irqreturn_t imx_mu_isr(int irq, void *p)
|
|
{
|
|
struct mbox_chan *chan = p;
|
|
struct imx_mu_priv *priv = to_imx_mu_priv(chan->mbox);
|
|
struct imx_mu_con_priv *cp = chan->con_priv;
|
|
u32 val, ctrl;
|
|
|
|
switch (cp->type) {
|
|
case IMX_MU_TYPE_TX:
|
|
ctrl = imx_mu_read(priv, priv->dcfg->xCR[IMX_MU_TCR]);
|
|
val = imx_mu_read(priv, priv->dcfg->xSR[IMX_MU_TSR]);
|
|
val &= IMX_MU_xSR_TEn(priv->dcfg->type, cp->idx) &
|
|
(ctrl & IMX_MU_xCR_TIEn(priv->dcfg->type, cp->idx));
|
|
break;
|
|
case IMX_MU_TYPE_RX:
|
|
ctrl = imx_mu_read(priv, priv->dcfg->xCR[IMX_MU_RCR]);
|
|
val = imx_mu_read(priv, priv->dcfg->xSR[IMX_MU_RSR]);
|
|
val &= IMX_MU_xSR_RFn(priv->dcfg->type, cp->idx) &
|
|
(ctrl & IMX_MU_xCR_RIEn(priv->dcfg->type, cp->idx));
|
|
break;
|
|
case IMX_MU_TYPE_RXDB:
|
|
ctrl = imx_mu_read(priv, priv->dcfg->xCR[IMX_MU_GIER]);
|
|
val = imx_mu_read(priv, priv->dcfg->xSR[IMX_MU_GSR]);
|
|
val &= IMX_MU_xSR_GIPn(priv->dcfg->type, cp->idx) &
|
|
(ctrl & IMX_MU_xCR_GIEn(priv->dcfg->type, cp->idx));
|
|
break;
|
|
case IMX_MU_TYPE_RST:
|
|
return IRQ_NONE;
|
|
default:
|
|
dev_warn_ratelimited(priv->dev, "Unhandled channel type %d\n",
|
|
cp->type);
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
if (!val)
|
|
return IRQ_NONE;
|
|
|
|
if ((val == IMX_MU_xSR_TEn(priv->dcfg->type, cp->idx)) &&
|
|
(cp->type == IMX_MU_TYPE_TX)) {
|
|
imx_mu_xcr_rmw(priv, IMX_MU_TCR, 0, IMX_MU_xCR_TIEn(priv->dcfg->type, cp->idx));
|
|
mbox_chan_txdone(chan, 0);
|
|
} else if ((val == IMX_MU_xSR_RFn(priv->dcfg->type, cp->idx)) &&
|
|
(cp->type == IMX_MU_TYPE_RX)) {
|
|
priv->dcfg->rx(priv, cp);
|
|
} else if ((val == IMX_MU_xSR_GIPn(priv->dcfg->type, cp->idx)) &&
|
|
(cp->type == IMX_MU_TYPE_RXDB)) {
|
|
priv->dcfg->rxdb(priv, cp);
|
|
} else {
|
|
dev_warn_ratelimited(priv->dev, "Not handled interrupt\n");
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
if (priv->suspend)
|
|
pm_system_wakeup();
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int imx_mu_send_data(struct mbox_chan *chan, void *data)
|
|
{
|
|
struct imx_mu_priv *priv = to_imx_mu_priv(chan->mbox);
|
|
struct imx_mu_con_priv *cp = chan->con_priv;
|
|
|
|
return priv->dcfg->tx(priv, cp, data);
|
|
}
|
|
|
|
static int imx_mu_startup(struct mbox_chan *chan)
|
|
{
|
|
struct imx_mu_priv *priv = to_imx_mu_priv(chan->mbox);
|
|
struct imx_mu_con_priv *cp = chan->con_priv;
|
|
unsigned long irq_flag = 0;
|
|
int ret;
|
|
|
|
pm_runtime_get_sync(priv->dev);
|
|
if (cp->type == IMX_MU_TYPE_TXDB_V2)
|
|
return 0;
|
|
|
|
if (cp->type == IMX_MU_TYPE_TXDB) {
|
|
/* Tx doorbell don't have ACK support */
|
|
tasklet_init(&cp->txdb_tasklet, imx_mu_txdb_tasklet,
|
|
(unsigned long)cp);
|
|
return 0;
|
|
}
|
|
|
|
/* IPC MU should be with IRQF_NO_SUSPEND set */
|
|
if (!priv->dev->pm_domain)
|
|
irq_flag |= IRQF_NO_SUSPEND;
|
|
|
|
if (!(priv->dcfg->type & IMX_MU_V2_IRQ))
|
|
irq_flag |= IRQF_SHARED;
|
|
|
|
ret = request_irq(priv->irq[cp->type], imx_mu_isr, irq_flag, cp->irq_desc, chan);
|
|
if (ret) {
|
|
dev_err(priv->dev, "Unable to acquire IRQ %d\n", priv->irq[cp->type]);
|
|
return ret;
|
|
}
|
|
|
|
switch (cp->type) {
|
|
case IMX_MU_TYPE_RX:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_RCR, IMX_MU_xCR_RIEn(priv->dcfg->type, cp->idx), 0);
|
|
break;
|
|
case IMX_MU_TYPE_RXDB:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_GIER, IMX_MU_xCR_GIEn(priv->dcfg->type, cp->idx), 0);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void imx_mu_shutdown(struct mbox_chan *chan)
|
|
{
|
|
struct imx_mu_priv *priv = to_imx_mu_priv(chan->mbox);
|
|
struct imx_mu_con_priv *cp = chan->con_priv;
|
|
int ret;
|
|
u32 sr;
|
|
|
|
if (cp->type == IMX_MU_TYPE_TXDB_V2) {
|
|
pm_runtime_put_sync(priv->dev);
|
|
return;
|
|
}
|
|
|
|
if (cp->type == IMX_MU_TYPE_TXDB) {
|
|
tasklet_kill(&cp->txdb_tasklet);
|
|
pm_runtime_put_sync(priv->dev);
|
|
return;
|
|
}
|
|
|
|
switch (cp->type) {
|
|
case IMX_MU_TYPE_TX:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_TCR, 0, IMX_MU_xCR_TIEn(priv->dcfg->type, cp->idx));
|
|
break;
|
|
case IMX_MU_TYPE_RX:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_RCR, 0, IMX_MU_xCR_RIEn(priv->dcfg->type, cp->idx));
|
|
break;
|
|
case IMX_MU_TYPE_RXDB:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_GIER, 0, IMX_MU_xCR_GIEn(priv->dcfg->type, cp->idx));
|
|
break;
|
|
case IMX_MU_TYPE_RST:
|
|
imx_mu_xcr_rmw(priv, IMX_MU_CR, IMX_MU_xCR_RST(priv->dcfg->type), 0);
|
|
ret = readl_poll_timeout(priv->base + priv->dcfg->xSR[IMX_MU_SR], sr,
|
|
!(sr & IMX_MU_xSR_RST(priv->dcfg->type)), 1, 5);
|
|
if (ret)
|
|
dev_warn(priv->dev, "RST channel timeout\n");
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
free_irq(priv->irq[cp->type], chan);
|
|
pm_runtime_put_sync(priv->dev);
|
|
}
|
|
|
|
static const struct mbox_chan_ops imx_mu_ops = {
|
|
.send_data = imx_mu_send_data,
|
|
.startup = imx_mu_startup,
|
|
.shutdown = imx_mu_shutdown,
|
|
};
|
|
|
|
static struct mbox_chan *imx_mu_specific_xlate(struct mbox_controller *mbox,
|
|
const struct of_phandle_args *sp)
|
|
{
|
|
u32 type, idx, chan;
|
|
|
|
if (sp->args_count != 2) {
|
|
dev_err(mbox->dev, "Invalid argument count %d\n", sp->args_count);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
type = sp->args[0]; /* channel type */
|
|
idx = sp->args[1]; /* index */
|
|
|
|
switch (type) {
|
|
case IMX_MU_TYPE_TX:
|
|
case IMX_MU_TYPE_RX:
|
|
if (idx != 0)
|
|
dev_err(mbox->dev, "Invalid chan idx: %d\n", idx);
|
|
chan = type;
|
|
break;
|
|
case IMX_MU_TYPE_RXDB:
|
|
chan = 2 + idx;
|
|
break;
|
|
default:
|
|
dev_err(mbox->dev, "Invalid chan type: %d\n", type);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
if (chan >= mbox->num_chans) {
|
|
dev_err(mbox->dev, "Not supported channel number: %d. (type: %d, idx: %d)\n", chan, type, idx);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
return &mbox->chans[chan];
|
|
}
|
|
|
|
static struct mbox_chan * imx_mu_xlate(struct mbox_controller *mbox,
|
|
const struct of_phandle_args *sp)
|
|
{
|
|
struct mbox_chan *p_chan;
|
|
u32 type, idx, chan;
|
|
|
|
if (sp->args_count != 2) {
|
|
dev_err(mbox->dev, "Invalid argument count %d\n", sp->args_count);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
type = sp->args[0]; /* channel type */
|
|
idx = sp->args[1]; /* index */
|
|
|
|
/* RST only supports 1 channel */
|
|
if ((type == IMX_MU_TYPE_RST) && idx) {
|
|
dev_err(mbox->dev, "Invalid RST channel %d\n", idx);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
chan = type * 4 + idx;
|
|
if (chan >= mbox->num_chans) {
|
|
dev_err(mbox->dev, "Not supported channel number: %d. (type: %d, idx: %d)\n", chan, type, idx);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
p_chan = &mbox->chans[chan];
|
|
|
|
if (type == IMX_MU_TYPE_TXDB_V2)
|
|
p_chan->txdone_method = TXDONE_BY_ACK;
|
|
|
|
return p_chan;
|
|
}
|
|
|
|
static struct mbox_chan *imx_mu_seco_xlate(struct mbox_controller *mbox,
|
|
const struct of_phandle_args *sp)
|
|
{
|
|
u32 type;
|
|
|
|
if (sp->args_count < 1) {
|
|
dev_err(mbox->dev, "Invalid argument count %d\n", sp->args_count);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
type = sp->args[0]; /* channel type */
|
|
|
|
/* Only supports TXDB and RXDB */
|
|
if (type == IMX_MU_TYPE_TX || type == IMX_MU_TYPE_RX) {
|
|
dev_err(mbox->dev, "Invalid type: %d\n", type);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
return imx_mu_xlate(mbox, sp);
|
|
}
|
|
|
|
static void imx_mu_init_generic(struct imx_mu_priv *priv)
|
|
{
|
|
unsigned int i;
|
|
unsigned int val;
|
|
|
|
for (i = 0; i < IMX_MU_CHANS; i++) {
|
|
struct imx_mu_con_priv *cp = &priv->con_priv[i];
|
|
|
|
cp->idx = i % 4;
|
|
cp->type = i >> 2;
|
|
cp->chan = &priv->mbox_chans[i];
|
|
priv->mbox_chans[i].con_priv = cp;
|
|
snprintf(cp->irq_desc, sizeof(cp->irq_desc),
|
|
"imx_mu_chan[%i-%i]", cp->type, cp->idx);
|
|
}
|
|
|
|
priv->mbox.num_chans = IMX_MU_CHANS;
|
|
priv->mbox.of_xlate = imx_mu_xlate;
|
|
|
|
if (priv->side_b)
|
|
return;
|
|
|
|
/* Set default MU configuration */
|
|
for (i = 0; i < IMX_MU_xCR_MAX; i++)
|
|
imx_mu_write(priv, 0, priv->dcfg->xCR[i]);
|
|
|
|
/* Clear any pending GIP */
|
|
val = imx_mu_read(priv, priv->dcfg->xSR[IMX_MU_GSR]);
|
|
imx_mu_write(priv, val, priv->dcfg->xSR[IMX_MU_GSR]);
|
|
|
|
/* Clear any pending RSR */
|
|
for (i = 0; i < IMX_MU_NUM_RR; i++)
|
|
imx_mu_read(priv, priv->dcfg->xRR + (i % 4) * 4);
|
|
}
|
|
|
|
static void imx_mu_init_specific(struct imx_mu_priv *priv)
|
|
{
|
|
unsigned int i;
|
|
int num_chans = priv->dcfg->type & IMX_MU_V2_S4 ? IMX_MU_S4_CHANS : IMX_MU_SCU_CHANS;
|
|
|
|
for (i = 0; i < num_chans; i++) {
|
|
struct imx_mu_con_priv *cp = &priv->con_priv[i];
|
|
|
|
cp->idx = i < 2 ? 0 : i - 2;
|
|
cp->type = i < 2 ? i : IMX_MU_TYPE_RXDB;
|
|
cp->chan = &priv->mbox_chans[i];
|
|
priv->mbox_chans[i].con_priv = cp;
|
|
snprintf(cp->irq_desc, sizeof(cp->irq_desc),
|
|
"imx_mu_chan[%i-%i]", cp->type, cp->idx);
|
|
}
|
|
|
|
priv->mbox.num_chans = num_chans;
|
|
priv->mbox.of_xlate = imx_mu_specific_xlate;
|
|
|
|
/* Set default MU configuration */
|
|
for (i = 0; i < IMX_MU_xCR_MAX; i++)
|
|
imx_mu_write(priv, 0, priv->dcfg->xCR[i]);
|
|
}
|
|
|
|
static void imx_mu_init_seco(struct imx_mu_priv *priv)
|
|
{
|
|
imx_mu_init_generic(priv);
|
|
priv->mbox.of_xlate = imx_mu_seco_xlate;
|
|
}
|
|
|
|
static int imx_mu_probe(struct platform_device *pdev)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct device_node *np = dev->of_node;
|
|
struct imx_mu_priv *priv;
|
|
const struct imx_mu_dcfg *dcfg;
|
|
int i, ret;
|
|
u32 size;
|
|
|
|
priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
|
|
if (!priv)
|
|
return -ENOMEM;
|
|
|
|
priv->dev = dev;
|
|
|
|
priv->base = devm_platform_ioremap_resource(pdev, 0);
|
|
if (IS_ERR(priv->base))
|
|
return PTR_ERR(priv->base);
|
|
|
|
dcfg = of_device_get_match_data(dev);
|
|
if (!dcfg)
|
|
return -EINVAL;
|
|
priv->dcfg = dcfg;
|
|
if (priv->dcfg->type & IMX_MU_V2_IRQ) {
|
|
priv->irq[IMX_MU_TYPE_TX] = platform_get_irq_byname(pdev, "tx");
|
|
if (priv->irq[IMX_MU_TYPE_TX] < 0)
|
|
return priv->irq[IMX_MU_TYPE_TX];
|
|
priv->irq[IMX_MU_TYPE_RX] = platform_get_irq_byname(pdev, "rx");
|
|
if (priv->irq[IMX_MU_TYPE_RX] < 0)
|
|
return priv->irq[IMX_MU_TYPE_RX];
|
|
} else {
|
|
ret = platform_get_irq(pdev, 0);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
for (i = 0; i < IMX_MU_CHANS; i++)
|
|
priv->irq[i] = ret;
|
|
}
|
|
|
|
if (priv->dcfg->type & IMX_MU_V2_S4)
|
|
size = sizeof(struct imx_s4_rpc_msg_max);
|
|
else
|
|
size = sizeof(struct imx_sc_rpc_msg_max);
|
|
|
|
priv->msg = devm_kzalloc(dev, size, GFP_KERNEL);
|
|
if (!priv->msg)
|
|
return -ENOMEM;
|
|
|
|
priv->clk = devm_clk_get(dev, NULL);
|
|
if (IS_ERR(priv->clk)) {
|
|
if (PTR_ERR(priv->clk) != -ENOENT)
|
|
return PTR_ERR(priv->clk);
|
|
|
|
priv->clk = NULL;
|
|
}
|
|
|
|
ret = clk_prepare_enable(priv->clk);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to enable clock\n");
|
|
return ret;
|
|
}
|
|
|
|
priv->side_b = of_property_read_bool(np, "fsl,mu-side-b");
|
|
|
|
priv->dcfg->init(priv);
|
|
|
|
spin_lock_init(&priv->xcr_lock);
|
|
|
|
priv->mbox.dev = dev;
|
|
priv->mbox.ops = &imx_mu_ops;
|
|
priv->mbox.chans = priv->mbox_chans;
|
|
priv->mbox.txdone_irq = true;
|
|
|
|
platform_set_drvdata(pdev, priv);
|
|
|
|
ret = devm_mbox_controller_register(dev, &priv->mbox);
|
|
if (ret) {
|
|
clk_disable_unprepare(priv->clk);
|
|
return ret;
|
|
}
|
|
|
|
pm_runtime_enable(dev);
|
|
|
|
ret = pm_runtime_resume_and_get(dev);
|
|
if (ret < 0)
|
|
goto disable_runtime_pm;
|
|
|
|
ret = pm_runtime_put_sync(dev);
|
|
if (ret < 0)
|
|
goto disable_runtime_pm;
|
|
|
|
clk_disable_unprepare(priv->clk);
|
|
|
|
return 0;
|
|
|
|
disable_runtime_pm:
|
|
pm_runtime_disable(dev);
|
|
clk_disable_unprepare(priv->clk);
|
|
return ret;
|
|
}
|
|
|
|
static void imx_mu_remove(struct platform_device *pdev)
|
|
{
|
|
struct imx_mu_priv *priv = platform_get_drvdata(pdev);
|
|
|
|
pm_runtime_disable(priv->dev);
|
|
}
|
|
|
|
static const struct imx_mu_dcfg imx_mu_cfg_imx6sx = {
|
|
.tx = imx_mu_generic_tx,
|
|
.rx = imx_mu_generic_rx,
|
|
.rxdb = imx_mu_generic_rxdb,
|
|
.init = imx_mu_init_generic,
|
|
.xTR = 0x0,
|
|
.xRR = 0x10,
|
|
.xSR = {0x20, 0x20, 0x20, 0x20},
|
|
.xCR = {0x24, 0x24, 0x24, 0x24, 0x24},
|
|
};
|
|
|
|
static const struct imx_mu_dcfg imx_mu_cfg_imx7ulp = {
|
|
.tx = imx_mu_generic_tx,
|
|
.rx = imx_mu_generic_rx,
|
|
.rxdb = imx_mu_generic_rxdb,
|
|
.init = imx_mu_init_generic,
|
|
.xTR = 0x20,
|
|
.xRR = 0x40,
|
|
.xSR = {0x60, 0x60, 0x60, 0x60},
|
|
.xCR = {0x64, 0x64, 0x64, 0x64, 0x64},
|
|
};
|
|
|
|
static const struct imx_mu_dcfg imx_mu_cfg_imx8ulp = {
|
|
.tx = imx_mu_generic_tx,
|
|
.rx = imx_mu_generic_rx,
|
|
.rxdb = imx_mu_generic_rxdb,
|
|
.init = imx_mu_init_generic,
|
|
.type = IMX_MU_V2,
|
|
.xTR = 0x200,
|
|
.xRR = 0x280,
|
|
.xSR = {0xC, 0x118, 0x124, 0x12C},
|
|
.xCR = {0x8, 0x110, 0x114, 0x120, 0x128},
|
|
};
|
|
|
|
static const struct imx_mu_dcfg imx_mu_cfg_imx8ulp_s4 = {
|
|
.tx = imx_mu_specific_tx,
|
|
.rx = imx_mu_specific_rx,
|
|
.init = imx_mu_init_specific,
|
|
.type = IMX_MU_V2 | IMX_MU_V2_S4,
|
|
.xTR = 0x200,
|
|
.xRR = 0x280,
|
|
.xSR = {0xC, 0x118, 0x124, 0x12C},
|
|
.xCR = {0x8, 0x110, 0x114, 0x120, 0x128},
|
|
};
|
|
|
|
static const struct imx_mu_dcfg imx_mu_cfg_imx93_s4 = {
|
|
.tx = imx_mu_specific_tx,
|
|
.rx = imx_mu_specific_rx,
|
|
.init = imx_mu_init_specific,
|
|
.type = IMX_MU_V2 | IMX_MU_V2_S4 | IMX_MU_V2_IRQ,
|
|
.xTR = 0x200,
|
|
.xRR = 0x280,
|
|
.xSR = {0xC, 0x118, 0x124, 0x12C},
|
|
.xCR = {0x8, 0x110, 0x114, 0x120, 0x128},
|
|
};
|
|
|
|
static const struct imx_mu_dcfg imx_mu_cfg_imx8_scu = {
|
|
.tx = imx_mu_specific_tx,
|
|
.rx = imx_mu_specific_rx,
|
|
.init = imx_mu_init_specific,
|
|
.rxdb = imx_mu_generic_rxdb,
|
|
.xTR = 0x0,
|
|
.xRR = 0x10,
|
|
.xSR = {0x20, 0x20, 0x20, 0x20},
|
|
.xCR = {0x24, 0x24, 0x24, 0x24, 0x24},
|
|
};
|
|
|
|
static const struct imx_mu_dcfg imx_mu_cfg_imx8_seco = {
|
|
.tx = imx_mu_seco_tx,
|
|
.rx = imx_mu_generic_rx,
|
|
.rxdb = imx_mu_seco_rxdb,
|
|
.init = imx_mu_init_seco,
|
|
.xTR = 0x0,
|
|
.xRR = 0x10,
|
|
.xSR = {0x20, 0x20, 0x20, 0x20},
|
|
.xCR = {0x24, 0x24, 0x24, 0x24, 0x24},
|
|
};
|
|
|
|
static const struct of_device_id imx_mu_dt_ids[] = {
|
|
{ .compatible = "fsl,imx7ulp-mu", .data = &imx_mu_cfg_imx7ulp },
|
|
{ .compatible = "fsl,imx6sx-mu", .data = &imx_mu_cfg_imx6sx },
|
|
{ .compatible = "fsl,imx8ulp-mu", .data = &imx_mu_cfg_imx8ulp },
|
|
{ .compatible = "fsl,imx8ulp-mu-s4", .data = &imx_mu_cfg_imx8ulp_s4 },
|
|
{ .compatible = "fsl,imx93-mu-s4", .data = &imx_mu_cfg_imx93_s4 },
|
|
{ .compatible = "fsl,imx8-mu-scu", .data = &imx_mu_cfg_imx8_scu },
|
|
{ .compatible = "fsl,imx8-mu-seco", .data = &imx_mu_cfg_imx8_seco },
|
|
{ },
|
|
};
|
|
MODULE_DEVICE_TABLE(of, imx_mu_dt_ids);
|
|
|
|
static int __maybe_unused imx_mu_suspend_noirq(struct device *dev)
|
|
{
|
|
struct imx_mu_priv *priv = dev_get_drvdata(dev);
|
|
int i;
|
|
|
|
if (!priv->clk) {
|
|
for (i = 0; i < IMX_MU_xCR_MAX; i++)
|
|
priv->xcr[i] = imx_mu_read(priv, priv->dcfg->xCR[i]);
|
|
}
|
|
|
|
priv->suspend = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __maybe_unused imx_mu_resume_noirq(struct device *dev)
|
|
{
|
|
struct imx_mu_priv *priv = dev_get_drvdata(dev);
|
|
int i;
|
|
|
|
/*
|
|
* ONLY restore MU when context lost, the TIE could
|
|
* be set during noirq resume as there is MU data
|
|
* communication going on, and restore the saved
|
|
* value will overwrite the TIE and cause MU data
|
|
* send failed, may lead to system freeze. This issue
|
|
* is observed by testing freeze mode suspend.
|
|
*/
|
|
if (!priv->clk && !imx_mu_read(priv, priv->dcfg->xCR[0])) {
|
|
for (i = 0; i < IMX_MU_xCR_MAX; i++)
|
|
imx_mu_write(priv, priv->xcr[i], priv->dcfg->xCR[i]);
|
|
}
|
|
|
|
priv->suspend = false;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __maybe_unused imx_mu_runtime_suspend(struct device *dev)
|
|
{
|
|
struct imx_mu_priv *priv = dev_get_drvdata(dev);
|
|
|
|
clk_disable_unprepare(priv->clk);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __maybe_unused imx_mu_runtime_resume(struct device *dev)
|
|
{
|
|
struct imx_mu_priv *priv = dev_get_drvdata(dev);
|
|
int ret;
|
|
|
|
ret = clk_prepare_enable(priv->clk);
|
|
if (ret)
|
|
dev_err(dev, "failed to enable clock\n");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static const struct dev_pm_ops imx_mu_pm_ops = {
|
|
SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(imx_mu_suspend_noirq,
|
|
imx_mu_resume_noirq)
|
|
SET_RUNTIME_PM_OPS(imx_mu_runtime_suspend,
|
|
imx_mu_runtime_resume, NULL)
|
|
};
|
|
|
|
static struct platform_driver imx_mu_driver = {
|
|
.probe = imx_mu_probe,
|
|
.remove_new = imx_mu_remove,
|
|
.driver = {
|
|
.name = "imx_mu",
|
|
.of_match_table = imx_mu_dt_ids,
|
|
.pm = &imx_mu_pm_ops,
|
|
},
|
|
};
|
|
module_platform_driver(imx_mu_driver);
|
|
|
|
MODULE_AUTHOR("Oleksij Rempel <o.rempel@pengutronix.de>");
|
|
MODULE_DESCRIPTION("Message Unit driver for i.MX");
|
|
MODULE_LICENSE("GPL v2");
|