mirror of
https://github.com/torvalds/linux.git
synced 2024-11-10 14:11:52 +00:00
e160d811c5
Use the device lifecycle managed allocation function. This helps prevent mistakes like freeing out of order in cleanup functions and forgetting to free on error paths. Signed-off-by: Andrew Davis <afd@ti.com> Link: https://lore.kernel.org/r/20240123184632.725054-9-afd@ti.com Signed-off-by: Mathieu Poirier <mathieu.poirier@linaro.org>
963 lines
23 KiB
C
963 lines
23 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) STMicroelectronics 2018 - All Rights Reserved
|
|
* Authors: Ludovic Barre <ludovic.barre@st.com> for STMicroelectronics.
|
|
* Fabien Dessenne <fabien.dessenne@st.com> for STMicroelectronics.
|
|
*/
|
|
|
|
#include <linux/arm-smccc.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/io.h>
|
|
#include <linux/mailbox_client.h>
|
|
#include <linux/mfd/syscon.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_reserved_mem.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/pm_wakeirq.h>
|
|
#include <linux/regmap.h>
|
|
#include <linux/remoteproc.h>
|
|
#include <linux/reset.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/workqueue.h>
|
|
|
|
#include "remoteproc_internal.h"
|
|
|
|
#define HOLD_BOOT 0
|
|
#define RELEASE_BOOT 1
|
|
|
|
#define MBOX_NB_VQ 2
|
|
#define MBOX_NB_MBX 4
|
|
|
|
#define STM32_SMC_RCC 0x82001000
|
|
#define STM32_SMC_REG_WRITE 0x1
|
|
|
|
#define STM32_MBX_VQ0 "vq0"
|
|
#define STM32_MBX_VQ0_ID 0
|
|
#define STM32_MBX_VQ1 "vq1"
|
|
#define STM32_MBX_VQ1_ID 1
|
|
#define STM32_MBX_SHUTDOWN "shutdown"
|
|
#define STM32_MBX_DETACH "detach"
|
|
|
|
#define RSC_TBL_SIZE 1024
|
|
|
|
#define M4_STATE_OFF 0
|
|
#define M4_STATE_INI 1
|
|
#define M4_STATE_CRUN 2
|
|
#define M4_STATE_CSTOP 3
|
|
#define M4_STATE_STANDBY 4
|
|
#define M4_STATE_CRASH 5
|
|
|
|
struct stm32_syscon {
|
|
struct regmap *map;
|
|
u32 reg;
|
|
u32 mask;
|
|
};
|
|
|
|
struct stm32_rproc_mem {
|
|
char name[20];
|
|
void __iomem *cpu_addr;
|
|
phys_addr_t bus_addr;
|
|
u32 dev_addr;
|
|
size_t size;
|
|
};
|
|
|
|
struct stm32_rproc_mem_ranges {
|
|
u32 dev_addr;
|
|
u32 bus_addr;
|
|
u32 size;
|
|
};
|
|
|
|
struct stm32_mbox {
|
|
const unsigned char name[10];
|
|
struct mbox_chan *chan;
|
|
struct mbox_client client;
|
|
struct work_struct vq_work;
|
|
int vq_id;
|
|
};
|
|
|
|
struct stm32_rproc {
|
|
struct reset_control *rst;
|
|
struct reset_control *hold_boot_rst;
|
|
struct stm32_syscon hold_boot;
|
|
struct stm32_syscon pdds;
|
|
struct stm32_syscon m4_state;
|
|
struct stm32_syscon rsctbl;
|
|
int wdg_irq;
|
|
u32 nb_rmems;
|
|
struct stm32_rproc_mem *rmems;
|
|
struct stm32_mbox mb[MBOX_NB_MBX];
|
|
struct workqueue_struct *workqueue;
|
|
bool hold_boot_smc;
|
|
void __iomem *rsc_va;
|
|
};
|
|
|
|
static int stm32_rproc_pa_to_da(struct rproc *rproc, phys_addr_t pa, u64 *da)
|
|
{
|
|
unsigned int i;
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
struct stm32_rproc_mem *p_mem;
|
|
|
|
for (i = 0; i < ddata->nb_rmems; i++) {
|
|
p_mem = &ddata->rmems[i];
|
|
|
|
if (pa < p_mem->bus_addr ||
|
|
pa >= p_mem->bus_addr + p_mem->size)
|
|
continue;
|
|
*da = pa - p_mem->bus_addr + p_mem->dev_addr;
|
|
dev_dbg(rproc->dev.parent, "pa %pa to da %llx\n", &pa, *da);
|
|
return 0;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static int stm32_rproc_mem_alloc(struct rproc *rproc,
|
|
struct rproc_mem_entry *mem)
|
|
{
|
|
struct device *dev = rproc->dev.parent;
|
|
void *va;
|
|
|
|
dev_dbg(dev, "map memory: %pad+%zx\n", &mem->dma, mem->len);
|
|
va = (__force void *)ioremap_wc(mem->dma, mem->len);
|
|
if (IS_ERR_OR_NULL(va)) {
|
|
dev_err(dev, "Unable to map memory region: %pad+0x%zx\n",
|
|
&mem->dma, mem->len);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Update memory entry va */
|
|
mem->va = va;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int stm32_rproc_mem_release(struct rproc *rproc,
|
|
struct rproc_mem_entry *mem)
|
|
{
|
|
dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma);
|
|
iounmap((__force __iomem void *)mem->va);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int stm32_rproc_of_memory_translations(struct platform_device *pdev,
|
|
struct stm32_rproc *ddata)
|
|
{
|
|
struct device *parent, *dev = &pdev->dev;
|
|
struct device_node *np;
|
|
struct stm32_rproc_mem *p_mems;
|
|
struct stm32_rproc_mem_ranges *mem_range;
|
|
int cnt, array_size, i, ret = 0;
|
|
|
|
parent = dev->parent;
|
|
np = parent->of_node;
|
|
|
|
cnt = of_property_count_elems_of_size(np, "dma-ranges",
|
|
sizeof(*mem_range));
|
|
if (cnt <= 0) {
|
|
dev_err(dev, "%s: dma-ranges property not defined\n", __func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
p_mems = devm_kcalloc(dev, cnt, sizeof(*p_mems), GFP_KERNEL);
|
|
if (!p_mems)
|
|
return -ENOMEM;
|
|
mem_range = kcalloc(cnt, sizeof(*mem_range), GFP_KERNEL);
|
|
if (!mem_range)
|
|
return -ENOMEM;
|
|
|
|
array_size = cnt * sizeof(struct stm32_rproc_mem_ranges) / sizeof(u32);
|
|
|
|
ret = of_property_read_u32_array(np, "dma-ranges",
|
|
(u32 *)mem_range, array_size);
|
|
if (ret) {
|
|
dev_err(dev, "error while get dma-ranges property: %x\n", ret);
|
|
goto free_mem;
|
|
}
|
|
|
|
for (i = 0; i < cnt; i++) {
|
|
p_mems[i].bus_addr = mem_range[i].bus_addr;
|
|
p_mems[i].dev_addr = mem_range[i].dev_addr;
|
|
p_mems[i].size = mem_range[i].size;
|
|
|
|
dev_dbg(dev, "memory range[%i]: da %#x, pa %pa, size %#zx:\n",
|
|
i, p_mems[i].dev_addr, &p_mems[i].bus_addr,
|
|
p_mems[i].size);
|
|
}
|
|
|
|
ddata->rmems = p_mems;
|
|
ddata->nb_rmems = cnt;
|
|
|
|
free_mem:
|
|
kfree(mem_range);
|
|
return ret;
|
|
}
|
|
|
|
static int stm32_rproc_mbox_idx(struct rproc *rproc, const unsigned char *name)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ddata->mb); i++) {
|
|
if (!strncmp(ddata->mb[i].name, name, strlen(name)))
|
|
return i;
|
|
}
|
|
dev_err(&rproc->dev, "mailbox %s not found\n", name);
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static int stm32_rproc_prepare(struct rproc *rproc)
|
|
{
|
|
struct device *dev = rproc->dev.parent;
|
|
struct device_node *np = dev->of_node;
|
|
struct of_phandle_iterator it;
|
|
struct rproc_mem_entry *mem;
|
|
struct reserved_mem *rmem;
|
|
u64 da;
|
|
int index = 0;
|
|
|
|
/* Register associated reserved memory regions */
|
|
of_phandle_iterator_init(&it, np, "memory-region", NULL, 0);
|
|
while (of_phandle_iterator_next(&it) == 0) {
|
|
rmem = of_reserved_mem_lookup(it.node);
|
|
if (!rmem) {
|
|
of_node_put(it.node);
|
|
dev_err(dev, "unable to acquire memory-region\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (stm32_rproc_pa_to_da(rproc, rmem->base, &da) < 0) {
|
|
of_node_put(it.node);
|
|
dev_err(dev, "memory region not valid %pa\n",
|
|
&rmem->base);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* No need to map vdev buffer */
|
|
if (strcmp(it.node->name, "vdev0buffer")) {
|
|
/* Register memory region */
|
|
mem = rproc_mem_entry_init(dev, NULL,
|
|
(dma_addr_t)rmem->base,
|
|
rmem->size, da,
|
|
stm32_rproc_mem_alloc,
|
|
stm32_rproc_mem_release,
|
|
it.node->name);
|
|
|
|
if (mem)
|
|
rproc_coredump_add_segment(rproc, da,
|
|
rmem->size);
|
|
} else {
|
|
/* Register reserved memory for vdev buffer alloc */
|
|
mem = rproc_of_resm_mem_entry_init(dev, index,
|
|
rmem->size,
|
|
rmem->base,
|
|
it.node->name);
|
|
}
|
|
|
|
if (!mem) {
|
|
of_node_put(it.node);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
rproc_add_carveout(rproc, mem);
|
|
index++;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int stm32_rproc_parse_fw(struct rproc *rproc, const struct firmware *fw)
|
|
{
|
|
if (rproc_elf_load_rsc_table(rproc, fw))
|
|
dev_warn(&rproc->dev, "no resource table found for this firmware\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static irqreturn_t stm32_rproc_wdg(int irq, void *data)
|
|
{
|
|
struct platform_device *pdev = data;
|
|
struct rproc *rproc = platform_get_drvdata(pdev);
|
|
|
|
rproc_report_crash(rproc, RPROC_WATCHDOG);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void stm32_rproc_mb_vq_work(struct work_struct *work)
|
|
{
|
|
struct stm32_mbox *mb = container_of(work, struct stm32_mbox, vq_work);
|
|
struct rproc *rproc = dev_get_drvdata(mb->client.dev);
|
|
|
|
mutex_lock(&rproc->lock);
|
|
|
|
if (rproc->state != RPROC_RUNNING)
|
|
goto unlock_mutex;
|
|
|
|
if (rproc_vq_interrupt(rproc, mb->vq_id) == IRQ_NONE)
|
|
dev_dbg(&rproc->dev, "no message found in vq%d\n", mb->vq_id);
|
|
|
|
unlock_mutex:
|
|
mutex_unlock(&rproc->lock);
|
|
}
|
|
|
|
static void stm32_rproc_mb_callback(struct mbox_client *cl, void *data)
|
|
{
|
|
struct rproc *rproc = dev_get_drvdata(cl->dev);
|
|
struct stm32_mbox *mb = container_of(cl, struct stm32_mbox, client);
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
|
|
queue_work(ddata->workqueue, &mb->vq_work);
|
|
}
|
|
|
|
static void stm32_rproc_free_mbox(struct rproc *rproc)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ddata->mb); i++) {
|
|
if (ddata->mb[i].chan)
|
|
mbox_free_channel(ddata->mb[i].chan);
|
|
ddata->mb[i].chan = NULL;
|
|
}
|
|
}
|
|
|
|
static const struct stm32_mbox stm32_rproc_mbox[MBOX_NB_MBX] = {
|
|
{
|
|
.name = STM32_MBX_VQ0,
|
|
.vq_id = STM32_MBX_VQ0_ID,
|
|
.client = {
|
|
.rx_callback = stm32_rproc_mb_callback,
|
|
.tx_block = false,
|
|
},
|
|
},
|
|
{
|
|
.name = STM32_MBX_VQ1,
|
|
.vq_id = STM32_MBX_VQ1_ID,
|
|
.client = {
|
|
.rx_callback = stm32_rproc_mb_callback,
|
|
.tx_block = false,
|
|
},
|
|
},
|
|
{
|
|
.name = STM32_MBX_SHUTDOWN,
|
|
.vq_id = -1,
|
|
.client = {
|
|
.tx_block = true,
|
|
.tx_done = NULL,
|
|
.tx_tout = 500, /* 500 ms time out */
|
|
},
|
|
},
|
|
{
|
|
.name = STM32_MBX_DETACH,
|
|
.vq_id = -1,
|
|
.client = {
|
|
.tx_block = true,
|
|
.tx_done = NULL,
|
|
.tx_tout = 200, /* 200 ms time out to detach should be fair enough */
|
|
},
|
|
}
|
|
};
|
|
|
|
static int stm32_rproc_request_mbox(struct rproc *rproc)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
struct device *dev = &rproc->dev;
|
|
unsigned int i;
|
|
int j;
|
|
const unsigned char *name;
|
|
struct mbox_client *cl;
|
|
|
|
/* Initialise mailbox structure table */
|
|
memcpy(ddata->mb, stm32_rproc_mbox, sizeof(stm32_rproc_mbox));
|
|
|
|
for (i = 0; i < MBOX_NB_MBX; i++) {
|
|
name = ddata->mb[i].name;
|
|
|
|
cl = &ddata->mb[i].client;
|
|
cl->dev = dev->parent;
|
|
|
|
ddata->mb[i].chan = mbox_request_channel_byname(cl, name);
|
|
if (IS_ERR(ddata->mb[i].chan)) {
|
|
if (PTR_ERR(ddata->mb[i].chan) == -EPROBE_DEFER) {
|
|
dev_err_probe(dev->parent,
|
|
PTR_ERR(ddata->mb[i].chan),
|
|
"failed to request mailbox %s\n",
|
|
name);
|
|
goto err_probe;
|
|
}
|
|
dev_warn(dev, "cannot get %s mbox\n", name);
|
|
ddata->mb[i].chan = NULL;
|
|
}
|
|
if (ddata->mb[i].vq_id >= 0) {
|
|
INIT_WORK(&ddata->mb[i].vq_work,
|
|
stm32_rproc_mb_vq_work);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_probe:
|
|
for (j = i - 1; j >= 0; j--)
|
|
if (ddata->mb[j].chan)
|
|
mbox_free_channel(ddata->mb[j].chan);
|
|
return -EPROBE_DEFER;
|
|
}
|
|
|
|
static int stm32_rproc_set_hold_boot(struct rproc *rproc, bool hold)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
struct stm32_syscon hold_boot = ddata->hold_boot;
|
|
struct arm_smccc_res smc_res;
|
|
int val, err;
|
|
|
|
/*
|
|
* Three ways to manage the hold boot
|
|
* - using SCMI: the hold boot is managed as a reset,
|
|
* - using Linux(no SCMI): the hold boot is managed as a syscon register
|
|
* - using SMC call (deprecated): use SMC reset interface
|
|
*/
|
|
|
|
val = hold ? HOLD_BOOT : RELEASE_BOOT;
|
|
|
|
if (ddata->hold_boot_rst) {
|
|
/* Use the SCMI reset controller */
|
|
if (!hold)
|
|
err = reset_control_deassert(ddata->hold_boot_rst);
|
|
else
|
|
err = reset_control_assert(ddata->hold_boot_rst);
|
|
} else if (IS_ENABLED(CONFIG_HAVE_ARM_SMCCC) && ddata->hold_boot_smc) {
|
|
/* Use the SMC call */
|
|
arm_smccc_smc(STM32_SMC_RCC, STM32_SMC_REG_WRITE,
|
|
hold_boot.reg, val, 0, 0, 0, 0, &smc_res);
|
|
err = smc_res.a0;
|
|
} else {
|
|
/* Use syscon */
|
|
err = regmap_update_bits(hold_boot.map, hold_boot.reg,
|
|
hold_boot.mask, val);
|
|
}
|
|
|
|
if (err)
|
|
dev_err(&rproc->dev, "failed to set hold boot\n");
|
|
|
|
return err;
|
|
}
|
|
|
|
static void stm32_rproc_add_coredump_trace(struct rproc *rproc)
|
|
{
|
|
struct rproc_debug_trace *trace;
|
|
struct rproc_dump_segment *segment;
|
|
bool already_added;
|
|
|
|
list_for_each_entry(trace, &rproc->traces, node) {
|
|
already_added = false;
|
|
|
|
list_for_each_entry(segment, &rproc->dump_segments, node) {
|
|
if (segment->da == trace->trace_mem.da) {
|
|
already_added = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!already_added)
|
|
rproc_coredump_add_segment(rproc, trace->trace_mem.da,
|
|
trace->trace_mem.len);
|
|
}
|
|
}
|
|
|
|
static int stm32_rproc_start(struct rproc *rproc)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
int err;
|
|
|
|
stm32_rproc_add_coredump_trace(rproc);
|
|
|
|
/* clear remote proc Deep Sleep */
|
|
if (ddata->pdds.map) {
|
|
err = regmap_update_bits(ddata->pdds.map, ddata->pdds.reg,
|
|
ddata->pdds.mask, 0);
|
|
if (err) {
|
|
dev_err(&rproc->dev, "failed to clear pdds\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
err = stm32_rproc_set_hold_boot(rproc, false);
|
|
if (err)
|
|
return err;
|
|
|
|
return stm32_rproc_set_hold_boot(rproc, true);
|
|
}
|
|
|
|
static int stm32_rproc_attach(struct rproc *rproc)
|
|
{
|
|
stm32_rproc_add_coredump_trace(rproc);
|
|
|
|
return stm32_rproc_set_hold_boot(rproc, true);
|
|
}
|
|
|
|
static int stm32_rproc_detach(struct rproc *rproc)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
int err, idx;
|
|
|
|
/* Inform the remote processor of the detach */
|
|
idx = stm32_rproc_mbox_idx(rproc, STM32_MBX_DETACH);
|
|
if (idx >= 0 && ddata->mb[idx].chan) {
|
|
err = mbox_send_message(ddata->mb[idx].chan, "stop");
|
|
if (err < 0)
|
|
dev_warn(&rproc->dev, "warning: remote FW detach without ack\n");
|
|
}
|
|
|
|
/* Allow remote processor to auto-reboot */
|
|
return stm32_rproc_set_hold_boot(rproc, false);
|
|
}
|
|
|
|
static int stm32_rproc_stop(struct rproc *rproc)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
int err, idx;
|
|
|
|
/* request shutdown of the remote processor */
|
|
if (rproc->state != RPROC_OFFLINE && rproc->state != RPROC_CRASHED) {
|
|
idx = stm32_rproc_mbox_idx(rproc, STM32_MBX_SHUTDOWN);
|
|
if (idx >= 0 && ddata->mb[idx].chan) {
|
|
err = mbox_send_message(ddata->mb[idx].chan, "detach");
|
|
if (err < 0)
|
|
dev_warn(&rproc->dev, "warning: remote FW shutdown without ack\n");
|
|
}
|
|
}
|
|
|
|
err = stm32_rproc_set_hold_boot(rproc, true);
|
|
if (err)
|
|
return err;
|
|
|
|
err = reset_control_assert(ddata->rst);
|
|
if (err) {
|
|
dev_err(&rproc->dev, "failed to assert the reset\n");
|
|
return err;
|
|
}
|
|
|
|
/* to allow platform Standby power mode, set remote proc Deep Sleep */
|
|
if (ddata->pdds.map) {
|
|
err = regmap_update_bits(ddata->pdds.map, ddata->pdds.reg,
|
|
ddata->pdds.mask, 1);
|
|
if (err) {
|
|
dev_err(&rproc->dev, "failed to set pdds\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
/* update coprocessor state to OFF if available */
|
|
if (ddata->m4_state.map) {
|
|
err = regmap_update_bits(ddata->m4_state.map,
|
|
ddata->m4_state.reg,
|
|
ddata->m4_state.mask,
|
|
M4_STATE_OFF);
|
|
if (err) {
|
|
dev_err(&rproc->dev, "failed to set copro state\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void stm32_rproc_kick(struct rproc *rproc, int vqid)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
unsigned int i;
|
|
int err;
|
|
|
|
if (WARN_ON(vqid >= MBOX_NB_VQ))
|
|
return;
|
|
|
|
for (i = 0; i < MBOX_NB_MBX; i++) {
|
|
if (vqid != ddata->mb[i].vq_id)
|
|
continue;
|
|
if (!ddata->mb[i].chan)
|
|
return;
|
|
err = mbox_send_message(ddata->mb[i].chan, "kick");
|
|
if (err < 0)
|
|
dev_err(&rproc->dev, "%s: failed (%s, err:%d)\n",
|
|
__func__, ddata->mb[i].name, err);
|
|
return;
|
|
}
|
|
}
|
|
|
|
static int stm32_rproc_da_to_pa(struct rproc *rproc,
|
|
u64 da, phys_addr_t *pa)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
struct device *dev = rproc->dev.parent;
|
|
struct stm32_rproc_mem *p_mem;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ddata->nb_rmems; i++) {
|
|
p_mem = &ddata->rmems[i];
|
|
|
|
if (da < p_mem->dev_addr ||
|
|
da >= p_mem->dev_addr + p_mem->size)
|
|
continue;
|
|
|
|
*pa = da - p_mem->dev_addr + p_mem->bus_addr;
|
|
dev_dbg(dev, "da %llx to pa %pap\n", da, pa);
|
|
|
|
return 0;
|
|
}
|
|
|
|
dev_err(dev, "can't translate da %llx\n", da);
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static struct resource_table *
|
|
stm32_rproc_get_loaded_rsc_table(struct rproc *rproc, size_t *table_sz)
|
|
{
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
struct device *dev = rproc->dev.parent;
|
|
phys_addr_t rsc_pa;
|
|
u32 rsc_da;
|
|
int err;
|
|
|
|
/* The resource table has already been mapped, nothing to do */
|
|
if (ddata->rsc_va)
|
|
goto done;
|
|
|
|
err = regmap_read(ddata->rsctbl.map, ddata->rsctbl.reg, &rsc_da);
|
|
if (err) {
|
|
dev_err(dev, "failed to read rsc tbl addr\n");
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
if (!rsc_da)
|
|
/* no rsc table */
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
err = stm32_rproc_da_to_pa(rproc, rsc_da, &rsc_pa);
|
|
if (err)
|
|
return ERR_PTR(err);
|
|
|
|
ddata->rsc_va = devm_ioremap_wc(dev, rsc_pa, RSC_TBL_SIZE);
|
|
if (IS_ERR_OR_NULL(ddata->rsc_va)) {
|
|
dev_err(dev, "Unable to map memory region: %pa+%x\n",
|
|
&rsc_pa, RSC_TBL_SIZE);
|
|
ddata->rsc_va = NULL;
|
|
return ERR_PTR(-ENOMEM);
|
|
}
|
|
|
|
done:
|
|
/*
|
|
* Assuming the resource table fits in 1kB is fair.
|
|
* Notice for the detach, that this 1 kB memory area has to be reserved in the coprocessor
|
|
* firmware for the resource table. On detach, the remoteproc core re-initializes this
|
|
* entire area by overwriting it with the initial values stored in rproc->clean_table.
|
|
*/
|
|
*table_sz = RSC_TBL_SIZE;
|
|
return (__force struct resource_table *)ddata->rsc_va;
|
|
}
|
|
|
|
static const struct rproc_ops st_rproc_ops = {
|
|
.prepare = stm32_rproc_prepare,
|
|
.start = stm32_rproc_start,
|
|
.stop = stm32_rproc_stop,
|
|
.attach = stm32_rproc_attach,
|
|
.detach = stm32_rproc_detach,
|
|
.kick = stm32_rproc_kick,
|
|
.load = rproc_elf_load_segments,
|
|
.parse_fw = stm32_rproc_parse_fw,
|
|
.find_loaded_rsc_table = rproc_elf_find_loaded_rsc_table,
|
|
.get_loaded_rsc_table = stm32_rproc_get_loaded_rsc_table,
|
|
.sanity_check = rproc_elf_sanity_check,
|
|
.get_boot_addr = rproc_elf_get_boot_addr,
|
|
};
|
|
|
|
static const struct of_device_id stm32_rproc_match[] = {
|
|
{ .compatible = "st,stm32mp1-m4" },
|
|
{},
|
|
};
|
|
MODULE_DEVICE_TABLE(of, stm32_rproc_match);
|
|
|
|
static int stm32_rproc_get_syscon(struct device_node *np, const char *prop,
|
|
struct stm32_syscon *syscon)
|
|
{
|
|
int err = 0;
|
|
|
|
syscon->map = syscon_regmap_lookup_by_phandle(np, prop);
|
|
if (IS_ERR(syscon->map)) {
|
|
err = PTR_ERR(syscon->map);
|
|
syscon->map = NULL;
|
|
goto out;
|
|
}
|
|
|
|
err = of_property_read_u32_index(np, prop, 1, &syscon->reg);
|
|
if (err)
|
|
goto out;
|
|
|
|
err = of_property_read_u32_index(np, prop, 2, &syscon->mask);
|
|
|
|
out:
|
|
return err;
|
|
}
|
|
|
|
static int stm32_rproc_parse_dt(struct platform_device *pdev,
|
|
struct stm32_rproc *ddata, bool *auto_boot)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct device_node *np = dev->of_node;
|
|
struct stm32_syscon tz;
|
|
unsigned int tzen;
|
|
int err, irq;
|
|
|
|
irq = platform_get_irq_optional(pdev, 0);
|
|
if (irq == -EPROBE_DEFER)
|
|
return irq;
|
|
|
|
if (irq > 0) {
|
|
err = devm_request_irq(dev, irq, stm32_rproc_wdg, 0,
|
|
dev_name(dev), pdev);
|
|
if (err)
|
|
return dev_err_probe(dev, err,
|
|
"failed to request wdg irq\n");
|
|
|
|
ddata->wdg_irq = irq;
|
|
|
|
if (of_property_read_bool(np, "wakeup-source")) {
|
|
device_init_wakeup(dev, true);
|
|
dev_pm_set_wake_irq(dev, irq);
|
|
}
|
|
|
|
dev_info(dev, "wdg irq registered\n");
|
|
}
|
|
|
|
ddata->rst = devm_reset_control_get_optional(dev, "mcu_rst");
|
|
if (!ddata->rst) {
|
|
/* Try legacy fallback method: get it by index */
|
|
ddata->rst = devm_reset_control_get_by_index(dev, 0);
|
|
}
|
|
if (IS_ERR(ddata->rst))
|
|
return dev_err_probe(dev, PTR_ERR(ddata->rst),
|
|
"failed to get mcu_reset\n");
|
|
|
|
/*
|
|
* Three ways to manage the hold boot
|
|
* - using SCMI: the hold boot is managed as a reset
|
|
* The DT "reset-mames" property should be defined with 2 items:
|
|
* reset-names = "mcu_rst", "hold_boot";
|
|
* - using SMC call (deprecated): use SMC reset interface
|
|
* The DT "reset-mames" property is optional, "st,syscfg-tz" is required
|
|
* - default(no SCMI, no SMC): the hold boot is managed as a syscon register
|
|
* The DT "reset-mames" property is optional, "st,syscfg-holdboot" is required
|
|
*/
|
|
|
|
ddata->hold_boot_rst = devm_reset_control_get_optional(dev, "hold_boot");
|
|
if (IS_ERR(ddata->hold_boot_rst))
|
|
return dev_err_probe(dev, PTR_ERR(ddata->hold_boot_rst),
|
|
"failed to get hold_boot reset\n");
|
|
|
|
if (!ddata->hold_boot_rst && IS_ENABLED(CONFIG_HAVE_ARM_SMCCC)) {
|
|
/* Manage the MCU_BOOT using SMC call */
|
|
err = stm32_rproc_get_syscon(np, "st,syscfg-tz", &tz);
|
|
if (!err) {
|
|
err = regmap_read(tz.map, tz.reg, &tzen);
|
|
if (err) {
|
|
dev_err(dev, "failed to read tzen\n");
|
|
return err;
|
|
}
|
|
ddata->hold_boot_smc = tzen & tz.mask;
|
|
}
|
|
}
|
|
|
|
if (!ddata->hold_boot_rst && !ddata->hold_boot_smc) {
|
|
/* Default: hold boot manage it through the syscon controller */
|
|
err = stm32_rproc_get_syscon(np, "st,syscfg-holdboot",
|
|
&ddata->hold_boot);
|
|
if (err) {
|
|
dev_err(dev, "failed to get hold boot\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
err = stm32_rproc_get_syscon(np, "st,syscfg-pdds", &ddata->pdds);
|
|
if (err)
|
|
dev_info(dev, "failed to get pdds\n");
|
|
|
|
*auto_boot = of_property_read_bool(np, "st,auto-boot");
|
|
|
|
/*
|
|
* See if we can check the M4 status, i.e if it was started
|
|
* from the boot loader or not.
|
|
*/
|
|
err = stm32_rproc_get_syscon(np, "st,syscfg-m4-state",
|
|
&ddata->m4_state);
|
|
if (err) {
|
|
/* remember this */
|
|
ddata->m4_state.map = NULL;
|
|
/* no coprocessor state syscon (optional) */
|
|
dev_warn(dev, "m4 state not supported\n");
|
|
|
|
/* no need to go further */
|
|
return 0;
|
|
}
|
|
|
|
/* See if we can get the resource table */
|
|
err = stm32_rproc_get_syscon(np, "st,syscfg-rsc-tbl",
|
|
&ddata->rsctbl);
|
|
if (err) {
|
|
/* no rsc table syscon (optional) */
|
|
dev_warn(dev, "rsc tbl syscon not supported\n");
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int stm32_rproc_get_m4_status(struct stm32_rproc *ddata,
|
|
unsigned int *state)
|
|
{
|
|
/* See stm32_rproc_parse_dt() */
|
|
if (!ddata->m4_state.map) {
|
|
/*
|
|
* We couldn't get the coprocessor's state, assume
|
|
* it is not running.
|
|
*/
|
|
*state = M4_STATE_OFF;
|
|
return 0;
|
|
}
|
|
|
|
return regmap_read(ddata->m4_state.map, ddata->m4_state.reg, state);
|
|
}
|
|
|
|
static int stm32_rproc_probe(struct platform_device *pdev)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct stm32_rproc *ddata;
|
|
struct device_node *np = dev->of_node;
|
|
struct rproc *rproc;
|
|
unsigned int state;
|
|
int ret;
|
|
|
|
ret = dma_coerce_mask_and_coherent(dev, DMA_BIT_MASK(32));
|
|
if (ret)
|
|
return ret;
|
|
|
|
rproc = devm_rproc_alloc(dev, np->name, &st_rproc_ops, NULL, sizeof(*ddata));
|
|
if (!rproc)
|
|
return -ENOMEM;
|
|
|
|
ddata = rproc->priv;
|
|
|
|
rproc_coredump_set_elf_info(rproc, ELFCLASS32, EM_NONE);
|
|
|
|
ret = stm32_rproc_parse_dt(pdev, ddata, &rproc->auto_boot);
|
|
if (ret)
|
|
goto free_rproc;
|
|
|
|
ret = stm32_rproc_of_memory_translations(pdev, ddata);
|
|
if (ret)
|
|
goto free_rproc;
|
|
|
|
ret = stm32_rproc_get_m4_status(ddata, &state);
|
|
if (ret)
|
|
goto free_rproc;
|
|
|
|
if (state == M4_STATE_CRUN)
|
|
rproc->state = RPROC_DETACHED;
|
|
|
|
rproc->has_iommu = false;
|
|
ddata->workqueue = create_workqueue(dev_name(dev));
|
|
if (!ddata->workqueue) {
|
|
dev_err(dev, "cannot create workqueue\n");
|
|
ret = -ENOMEM;
|
|
goto free_resources;
|
|
}
|
|
|
|
platform_set_drvdata(pdev, rproc);
|
|
|
|
ret = stm32_rproc_request_mbox(rproc);
|
|
if (ret)
|
|
goto free_wkq;
|
|
|
|
ret = rproc_add(rproc);
|
|
if (ret)
|
|
goto free_mb;
|
|
|
|
return 0;
|
|
|
|
free_mb:
|
|
stm32_rproc_free_mbox(rproc);
|
|
free_wkq:
|
|
destroy_workqueue(ddata->workqueue);
|
|
free_resources:
|
|
rproc_resource_cleanup(rproc);
|
|
free_rproc:
|
|
if (device_may_wakeup(dev)) {
|
|
dev_pm_clear_wake_irq(dev);
|
|
device_init_wakeup(dev, false);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static void stm32_rproc_remove(struct platform_device *pdev)
|
|
{
|
|
struct rproc *rproc = platform_get_drvdata(pdev);
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
struct device *dev = &pdev->dev;
|
|
|
|
if (atomic_read(&rproc->power) > 0)
|
|
rproc_shutdown(rproc);
|
|
|
|
rproc_del(rproc);
|
|
stm32_rproc_free_mbox(rproc);
|
|
destroy_workqueue(ddata->workqueue);
|
|
|
|
if (device_may_wakeup(dev)) {
|
|
dev_pm_clear_wake_irq(dev);
|
|
device_init_wakeup(dev, false);
|
|
}
|
|
}
|
|
|
|
static int stm32_rproc_suspend(struct device *dev)
|
|
{
|
|
struct rproc *rproc = dev_get_drvdata(dev);
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
|
|
if (device_may_wakeup(dev))
|
|
return enable_irq_wake(ddata->wdg_irq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int stm32_rproc_resume(struct device *dev)
|
|
{
|
|
struct rproc *rproc = dev_get_drvdata(dev);
|
|
struct stm32_rproc *ddata = rproc->priv;
|
|
|
|
if (device_may_wakeup(dev))
|
|
return disable_irq_wake(ddata->wdg_irq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static DEFINE_SIMPLE_DEV_PM_OPS(stm32_rproc_pm_ops,
|
|
stm32_rproc_suspend, stm32_rproc_resume);
|
|
|
|
static struct platform_driver stm32_rproc_driver = {
|
|
.probe = stm32_rproc_probe,
|
|
.remove_new = stm32_rproc_remove,
|
|
.driver = {
|
|
.name = "stm32-rproc",
|
|
.pm = pm_ptr(&stm32_rproc_pm_ops),
|
|
.of_match_table = stm32_rproc_match,
|
|
},
|
|
};
|
|
module_platform_driver(stm32_rproc_driver);
|
|
|
|
MODULE_DESCRIPTION("STM32 Remote Processor Control Driver");
|
|
MODULE_AUTHOR("Ludovic Barre <ludovic.barre@st.com>");
|
|
MODULE_AUTHOR("Fabien Dessenne <fabien.dessenne@st.com>");
|
|
MODULE_LICENSE("GPL v2");
|
|
|