mirror of
https://github.com/torvalds/linux.git
synced 2024-11-15 00:21:59 +00:00
9814c1447f
Commit9ee3f0d8c9
("ASOC: SOF: Intel: hda-loader: only wait for HDaudio IOC for IPC4 devices") removed DMA wait for IPC3 case. Proceed and remove the wait for IPC4 devices as well. There is no dependency to IPC version in the load logic and checking the firmware status is a sufficient check in case of errors. The removed code also had a bug in that -ETIMEDOUT is returned without stopping the DMA transfer. Cc: stable@vger.kernel.org Link: https://github.com/thesofproject/linux/issues/5135 Fixes:9ee3f0d8c9
("ASOC: SOF: Intel: hda-loader: only wait for HDaudio IOC for IPC4 devices") Suggested-by: Peter Ujfalusi <peter.ujfalusi@linux.intel.com> Signed-off-by: Kai Vehmanen <kai.vehmanen@linux.intel.com> Reviewed-by: Péter Ujfalusi <peter.ujfalusi@linux.intel.com> Reviewed-by: Pierre-Louis Bossart <pierre-louis.bossart@linux.intel.com> Reviewed-by: Ranjani Sridharan <ranjani.sridharan@linux.intel.com> Signed-off-by: Peter Ujfalusi <peter.ujfalusi@linux.intel.com> Link: https://patch.msgid.link/20241008060710.15409-1-peter.ujfalusi@linux.intel.com Signed-off-by: Mark Brown <broonie@kernel.org>
682 lines
20 KiB
C
682 lines
20 KiB
C
// SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
|
|
//
|
|
// This file is provided under a dual BSD/GPLv2 license. When using or
|
|
// redistributing this file, you may do so under either license.
|
|
//
|
|
// Copyright(c) 2018 Intel Corporation
|
|
//
|
|
// Authors: Liam Girdwood <liam.r.girdwood@linux.intel.com>
|
|
// Ranjani Sridharan <ranjani.sridharan@linux.intel.com>
|
|
// Rander Wang <rander.wang@intel.com>
|
|
// Keyon Jie <yang.jie@linux.intel.com>
|
|
//
|
|
|
|
/*
|
|
* Hardware interface for HDA DSP code loader
|
|
*/
|
|
|
|
#include <linux/firmware.h>
|
|
#include <sound/hdaudio_ext.h>
|
|
#include <sound/hda_register.h>
|
|
#include <sound/sof.h>
|
|
#include <sound/sof/ipc4/header.h>
|
|
#include "ext_manifest.h"
|
|
#include "../ipc4-priv.h"
|
|
#include "../ops.h"
|
|
#include "../sof-priv.h"
|
|
#include "hda.h"
|
|
|
|
static void hda_ssp_set_cbp_cfp(struct snd_sof_dev *sdev)
|
|
{
|
|
struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
|
|
const struct sof_intel_dsp_desc *chip = hda->desc;
|
|
int i;
|
|
|
|
/* DSP is powered up, set all SSPs to clock consumer/codec provider mode */
|
|
for (i = 0; i < chip->ssp_count; i++) {
|
|
snd_sof_dsp_update_bits_unlocked(sdev, HDA_DSP_BAR,
|
|
chip->ssp_base_offset
|
|
+ i * SSP_DEV_MEM_SIZE
|
|
+ SSP_SSC1_OFFSET,
|
|
SSP_SET_CBP_CFP,
|
|
SSP_SET_CBP_CFP);
|
|
}
|
|
}
|
|
|
|
struct hdac_ext_stream *hda_cl_prepare(struct device *dev, unsigned int format,
|
|
unsigned int size, struct snd_dma_buffer *dmab,
|
|
int direction, bool is_iccmax)
|
|
{
|
|
struct snd_sof_dev *sdev = dev_get_drvdata(dev);
|
|
struct hdac_ext_stream *hext_stream;
|
|
struct hdac_stream *hstream;
|
|
int ret;
|
|
|
|
hext_stream = hda_dsp_stream_get(sdev, direction, 0);
|
|
|
|
if (!hext_stream) {
|
|
dev_err(sdev->dev, "error: no stream available\n");
|
|
return ERR_PTR(-ENODEV);
|
|
}
|
|
hstream = &hext_stream->hstream;
|
|
hstream->substream = NULL;
|
|
|
|
/* allocate DMA buffer */
|
|
ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV_SG, dev, size, dmab);
|
|
if (ret < 0) {
|
|
dev_err(sdev->dev, "error: memory alloc failed: %d\n", ret);
|
|
goto out_put;
|
|
}
|
|
|
|
hstream->period_bytes = 0;/* initialize period_bytes */
|
|
hstream->format_val = format;
|
|
hstream->bufsize = size;
|
|
|
|
if (is_iccmax) {
|
|
ret = hda_dsp_iccmax_stream_hw_params(sdev, hext_stream, dmab, NULL);
|
|
if (ret < 0) {
|
|
dev_err(sdev->dev, "error: iccmax stream prepare failed: %d\n", ret);
|
|
goto out_free;
|
|
}
|
|
} else {
|
|
ret = hda_dsp_stream_hw_params(sdev, hext_stream, dmab, NULL);
|
|
if (ret < 0) {
|
|
dev_err(sdev->dev, "error: hdac prepare failed: %d\n", ret);
|
|
goto out_free;
|
|
}
|
|
hda_dsp_stream_spib_config(sdev, hext_stream, HDA_DSP_SPIB_ENABLE, size);
|
|
}
|
|
|
|
return hext_stream;
|
|
|
|
out_free:
|
|
snd_dma_free_pages(dmab);
|
|
out_put:
|
|
hda_dsp_stream_put(sdev, direction, hstream->stream_tag);
|
|
return ERR_PTR(ret);
|
|
}
|
|
EXPORT_SYMBOL_NS(hda_cl_prepare, SND_SOC_SOF_INTEL_HDA_COMMON);
|
|
|
|
/*
|
|
* first boot sequence has some extra steps.
|
|
* power on all host managed cores and only unstall/run the boot core to boot the
|
|
* DSP then turn off all non boot cores (if any) is powered on.
|
|
*/
|
|
int cl_dsp_init(struct snd_sof_dev *sdev, int stream_tag, bool imr_boot)
|
|
{
|
|
struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
|
|
const struct sof_intel_dsp_desc *chip = hda->desc;
|
|
unsigned int status, target_status;
|
|
u32 flags, ipc_hdr, j;
|
|
unsigned long mask;
|
|
char *dump_msg;
|
|
int ret;
|
|
|
|
/* step 1: power up corex */
|
|
ret = hda_dsp_core_power_up(sdev, chip->host_managed_cores_mask);
|
|
if (ret < 0) {
|
|
if (hda->boot_iteration == HDA_FW_BOOT_ATTEMPTS)
|
|
dev_err(sdev->dev, "error: dsp core 0/1 power up failed\n");
|
|
goto err;
|
|
}
|
|
|
|
hda_ssp_set_cbp_cfp(sdev);
|
|
|
|
/* step 2: Send ROM_CONTROL command (stream_tag is ignored for IMR boot) */
|
|
ipc_hdr = chip->ipc_req_mask | HDA_DSP_ROM_IPC_CONTROL;
|
|
if (!imr_boot)
|
|
ipc_hdr |= HDA_DSP_ROM_IPC_PURGE_FW | ((stream_tag - 1) << 9);
|
|
|
|
snd_sof_dsp_write(sdev, HDA_DSP_BAR, chip->ipc_req, ipc_hdr);
|
|
|
|
/* step 3: unset core 0 reset state & unstall/run core 0 */
|
|
ret = hda_dsp_core_run(sdev, chip->init_core_mask);
|
|
if (ret < 0) {
|
|
if (hda->boot_iteration == HDA_FW_BOOT_ATTEMPTS)
|
|
dev_err(sdev->dev,
|
|
"error: dsp core start failed %d\n", ret);
|
|
ret = -EIO;
|
|
goto err;
|
|
}
|
|
|
|
/* step 4: wait for IPC DONE bit from ROM */
|
|
ret = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_BAR,
|
|
chip->ipc_ack, status,
|
|
((status & chip->ipc_ack_mask)
|
|
== chip->ipc_ack_mask),
|
|
HDA_DSP_REG_POLL_INTERVAL_US,
|
|
HDA_DSP_INIT_TIMEOUT_US);
|
|
|
|
if (ret < 0) {
|
|
if (hda->boot_iteration == HDA_FW_BOOT_ATTEMPTS)
|
|
dev_err(sdev->dev,
|
|
"error: %s: timeout for HIPCIE done\n",
|
|
__func__);
|
|
goto err;
|
|
}
|
|
|
|
/* set DONE bit to clear the reply IPC message */
|
|
snd_sof_dsp_update_bits_forced(sdev, HDA_DSP_BAR,
|
|
chip->ipc_ack,
|
|
chip->ipc_ack_mask,
|
|
chip->ipc_ack_mask);
|
|
|
|
/* step 5: power down cores that are no longer needed */
|
|
ret = hda_dsp_core_reset_power_down(sdev, chip->host_managed_cores_mask &
|
|
~(chip->init_core_mask));
|
|
if (ret < 0) {
|
|
if (hda->boot_iteration == HDA_FW_BOOT_ATTEMPTS)
|
|
dev_err(sdev->dev,
|
|
"error: dsp core x power down failed\n");
|
|
goto err;
|
|
}
|
|
|
|
/* step 6: enable IPC interrupts */
|
|
hda_dsp_ipc_int_enable(sdev);
|
|
|
|
/*
|
|
* step 7:
|
|
* - Cold/Full boot: wait for ROM init to proceed to download the firmware
|
|
* - IMR boot: wait for ROM firmware entered (firmware booted up from IMR)
|
|
*/
|
|
if (imr_boot)
|
|
target_status = FSR_STATE_FW_ENTERED;
|
|
else
|
|
target_status = FSR_STATE_INIT_DONE;
|
|
|
|
ret = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_BAR,
|
|
chip->rom_status_reg, status,
|
|
(FSR_TO_STATE_CODE(status) == target_status),
|
|
HDA_DSP_REG_POLL_INTERVAL_US,
|
|
chip->rom_init_timeout *
|
|
USEC_PER_MSEC);
|
|
if (!ret) {
|
|
/* set enabled cores mask and increment ref count for cores in init_core_mask */
|
|
sdev->enabled_cores_mask |= chip->init_core_mask;
|
|
mask = sdev->enabled_cores_mask;
|
|
for_each_set_bit(j, &mask, SOF_MAX_DSP_NUM_CORES)
|
|
sdev->dsp_core_ref_count[j]++;
|
|
return 0;
|
|
}
|
|
|
|
if (hda->boot_iteration == HDA_FW_BOOT_ATTEMPTS)
|
|
dev_err(sdev->dev,
|
|
"%s: timeout with rom_status_reg (%#x) read\n",
|
|
__func__, chip->rom_status_reg);
|
|
|
|
err:
|
|
flags = SOF_DBG_DUMP_PCI | SOF_DBG_DUMP_MBOX | SOF_DBG_DUMP_OPTIONAL;
|
|
|
|
/* after max boot attempts make sure that the dump is printed */
|
|
if (hda->boot_iteration == HDA_FW_BOOT_ATTEMPTS)
|
|
flags &= ~SOF_DBG_DUMP_OPTIONAL;
|
|
|
|
dump_msg = kasprintf(GFP_KERNEL, "Boot iteration failed: %d/%d",
|
|
hda->boot_iteration, HDA_FW_BOOT_ATTEMPTS);
|
|
snd_sof_dsp_dbg_dump(sdev, dump_msg, flags);
|
|
hda_dsp_core_reset_power_down(sdev, chip->host_managed_cores_mask);
|
|
|
|
kfree(dump_msg);
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_NS(cl_dsp_init, SND_SOC_SOF_INTEL_HDA_COMMON);
|
|
|
|
int hda_cl_trigger(struct device *dev, struct hdac_ext_stream *hext_stream, int cmd)
|
|
{
|
|
struct snd_sof_dev *sdev = dev_get_drvdata(dev);
|
|
struct hdac_stream *hstream = &hext_stream->hstream;
|
|
int sd_offset = SOF_STREAM_SD_OFFSET(hstream);
|
|
struct sof_intel_hda_stream *hda_stream;
|
|
|
|
/* code loader is special case that reuses stream ops */
|
|
switch (cmd) {
|
|
case SNDRV_PCM_TRIGGER_START:
|
|
hda_stream = container_of(hext_stream, struct sof_intel_hda_stream,
|
|
hext_stream);
|
|
reinit_completion(&hda_stream->ioc);
|
|
|
|
snd_sof_dsp_update_bits(sdev, HDA_DSP_HDA_BAR, SOF_HDA_INTCTL,
|
|
1 << hstream->index,
|
|
1 << hstream->index);
|
|
|
|
snd_sof_dsp_update_bits(sdev, HDA_DSP_HDA_BAR,
|
|
sd_offset,
|
|
SOF_HDA_SD_CTL_DMA_START |
|
|
SOF_HDA_CL_DMA_SD_INT_MASK,
|
|
SOF_HDA_SD_CTL_DMA_START |
|
|
SOF_HDA_CL_DMA_SD_INT_MASK);
|
|
|
|
hstream->running = true;
|
|
return 0;
|
|
default:
|
|
return hda_dsp_stream_trigger(sdev, hext_stream, cmd);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL_NS(hda_cl_trigger, SND_SOC_SOF_INTEL_HDA_COMMON);
|
|
|
|
int hda_cl_cleanup(struct device *dev, struct snd_dma_buffer *dmab,
|
|
struct hdac_ext_stream *hext_stream)
|
|
{
|
|
struct snd_sof_dev *sdev = dev_get_drvdata(dev);
|
|
struct hdac_stream *hstream = &hext_stream->hstream;
|
|
int sd_offset = SOF_STREAM_SD_OFFSET(hstream);
|
|
int ret = 0;
|
|
|
|
if (hstream->direction == SNDRV_PCM_STREAM_PLAYBACK)
|
|
ret = hda_dsp_stream_spib_config(sdev, hext_stream, HDA_DSP_SPIB_DISABLE, 0);
|
|
else
|
|
snd_sof_dsp_update_bits(sdev, HDA_DSP_HDA_BAR, sd_offset,
|
|
SOF_HDA_SD_CTL_DMA_START, 0);
|
|
|
|
hda_dsp_stream_put(sdev, hstream->direction, hstream->stream_tag);
|
|
hstream->running = 0;
|
|
hstream->substream = NULL;
|
|
|
|
/* reset BDL address */
|
|
snd_sof_dsp_write(sdev, HDA_DSP_HDA_BAR,
|
|
sd_offset + SOF_HDA_ADSP_REG_SD_BDLPL, 0);
|
|
snd_sof_dsp_write(sdev, HDA_DSP_HDA_BAR,
|
|
sd_offset + SOF_HDA_ADSP_REG_SD_BDLPU, 0);
|
|
|
|
snd_sof_dsp_write(sdev, HDA_DSP_HDA_BAR, sd_offset, 0);
|
|
snd_dma_free_pages(dmab);
|
|
dmab->area = NULL;
|
|
hstream->bufsize = 0;
|
|
hstream->format_val = 0;
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_NS(hda_cl_cleanup, SND_SOC_SOF_INTEL_HDA_COMMON);
|
|
|
|
#define HDA_CL_DMA_IOC_TIMEOUT_MS 500
|
|
|
|
int hda_cl_copy_fw(struct snd_sof_dev *sdev, struct hdac_ext_stream *hext_stream)
|
|
{
|
|
struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
|
|
const struct sof_intel_dsp_desc *chip = hda->desc;
|
|
unsigned int reg;
|
|
int ret, status;
|
|
|
|
dev_dbg(sdev->dev, "Code loader DMA starting\n");
|
|
|
|
ret = hda_cl_trigger(sdev->dev, hext_stream, SNDRV_PCM_TRIGGER_START);
|
|
if (ret < 0) {
|
|
dev_err(sdev->dev, "error: DMA trigger start failed\n");
|
|
return ret;
|
|
}
|
|
|
|
dev_dbg(sdev->dev, "waiting for FW_ENTERED status\n");
|
|
|
|
status = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_BAR,
|
|
chip->rom_status_reg, reg,
|
|
(FSR_TO_STATE_CODE(reg) == FSR_STATE_FW_ENTERED),
|
|
HDA_DSP_REG_POLL_INTERVAL_US,
|
|
HDA_DSP_BASEFW_TIMEOUT_US);
|
|
|
|
/*
|
|
* even in case of errors we still need to stop the DMAs,
|
|
* but we return the initial error should the DMA stop also fail
|
|
*/
|
|
|
|
if (status < 0) {
|
|
dev_err(sdev->dev,
|
|
"%s: timeout with rom_status_reg (%#x) read\n",
|
|
__func__, chip->rom_status_reg);
|
|
} else {
|
|
dev_dbg(sdev->dev, "Code loader FW_ENTERED status\n");
|
|
}
|
|
|
|
ret = hda_cl_trigger(sdev->dev, hext_stream, SNDRV_PCM_TRIGGER_STOP);
|
|
if (ret < 0) {
|
|
dev_err(sdev->dev, "error: DMA trigger stop failed\n");
|
|
if (!status)
|
|
status = ret;
|
|
} else {
|
|
dev_dbg(sdev->dev, "Code loader DMA stopped\n");
|
|
}
|
|
|
|
return status;
|
|
}
|
|
|
|
int hda_dsp_cl_boot_firmware_iccmax(struct snd_sof_dev *sdev)
|
|
{
|
|
struct hdac_ext_stream *iccmax_stream;
|
|
struct snd_dma_buffer dmab_bdl;
|
|
int ret, ret1;
|
|
u8 original_gb;
|
|
|
|
/* save the original LTRP guardband value */
|
|
original_gb = snd_sof_dsp_read8(sdev, HDA_DSP_HDA_BAR, HDA_VS_INTEL_LTRP) &
|
|
HDA_VS_INTEL_LTRP_GB_MASK;
|
|
|
|
/*
|
|
* Prepare capture stream for ICCMAX. We do not need to store
|
|
* the data, so use a buffer of PAGE_SIZE for receiving.
|
|
*/
|
|
iccmax_stream = hda_cl_prepare(sdev->dev, HDA_CL_STREAM_FORMAT, PAGE_SIZE,
|
|
&dmab_bdl, SNDRV_PCM_STREAM_CAPTURE, true);
|
|
if (IS_ERR(iccmax_stream)) {
|
|
dev_err(sdev->dev, "error: dma prepare for ICCMAX stream failed\n");
|
|
return PTR_ERR(iccmax_stream);
|
|
}
|
|
|
|
ret = hda_dsp_cl_boot_firmware(sdev);
|
|
|
|
/*
|
|
* Perform iccmax stream cleanup. This should be done even if firmware loading fails.
|
|
* If the cleanup also fails, we return the initial error
|
|
*/
|
|
ret1 = hda_cl_cleanup(sdev->dev, &dmab_bdl, iccmax_stream);
|
|
if (ret1 < 0) {
|
|
dev_err(sdev->dev, "error: ICCMAX stream cleanup failed\n");
|
|
|
|
/* set return value to indicate cleanup failure */
|
|
if (!ret)
|
|
ret = ret1;
|
|
}
|
|
|
|
/* restore the original guardband value after FW boot */
|
|
snd_sof_dsp_update8(sdev, HDA_DSP_HDA_BAR, HDA_VS_INTEL_LTRP,
|
|
HDA_VS_INTEL_LTRP_GB_MASK, original_gb);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_NS(hda_dsp_cl_boot_firmware_iccmax, SND_SOC_SOF_INTEL_CNL);
|
|
|
|
static int hda_dsp_boot_imr(struct snd_sof_dev *sdev)
|
|
{
|
|
const struct sof_intel_dsp_desc *chip_info;
|
|
int ret;
|
|
|
|
chip_info = get_chip_info(sdev->pdata);
|
|
if (chip_info->cl_init)
|
|
ret = chip_info->cl_init(sdev, 0, true);
|
|
else
|
|
ret = -EINVAL;
|
|
|
|
if (!ret)
|
|
hda_sdw_process_wakeen(sdev);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int hda_dsp_cl_boot_firmware(struct snd_sof_dev *sdev)
|
|
{
|
|
struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
|
|
struct snd_sof_pdata *plat_data = sdev->pdata;
|
|
const struct sof_dev_desc *desc = plat_data->desc;
|
|
const struct sof_intel_dsp_desc *chip_info;
|
|
struct hdac_ext_stream *hext_stream;
|
|
struct firmware stripped_firmware;
|
|
struct snd_dma_buffer dmab;
|
|
int ret, ret1, i;
|
|
|
|
if (hda->imrboot_supported && !sdev->first_boot && !hda->skip_imr_boot) {
|
|
dev_dbg(sdev->dev, "IMR restore supported, booting from IMR directly\n");
|
|
hda->boot_iteration = 0;
|
|
ret = hda_dsp_boot_imr(sdev);
|
|
if (!ret) {
|
|
hda->booted_from_imr = true;
|
|
return 0;
|
|
}
|
|
|
|
dev_warn(sdev->dev, "IMR restore failed, trying to cold boot\n");
|
|
}
|
|
|
|
hda->booted_from_imr = false;
|
|
|
|
chip_info = desc->chip_info;
|
|
|
|
if (sdev->basefw.fw->size <= sdev->basefw.payload_offset) {
|
|
dev_err(sdev->dev, "error: firmware size must be greater than firmware offset\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
stripped_firmware.data = sdev->basefw.fw->data + sdev->basefw.payload_offset;
|
|
stripped_firmware.size = sdev->basefw.fw->size - sdev->basefw.payload_offset;
|
|
|
|
/* init for booting wait */
|
|
init_waitqueue_head(&sdev->boot_wait);
|
|
|
|
/* prepare DMA for code loader stream */
|
|
hext_stream = hda_cl_prepare(sdev->dev, HDA_CL_STREAM_FORMAT,
|
|
stripped_firmware.size,
|
|
&dmab, SNDRV_PCM_STREAM_PLAYBACK, false);
|
|
if (IS_ERR(hext_stream)) {
|
|
dev_err(sdev->dev, "error: dma prepare for fw loading failed\n");
|
|
return PTR_ERR(hext_stream);
|
|
}
|
|
|
|
memcpy(dmab.area, stripped_firmware.data,
|
|
stripped_firmware.size);
|
|
|
|
/* try ROM init a few times before giving up */
|
|
for (i = 0; i < HDA_FW_BOOT_ATTEMPTS; i++) {
|
|
dev_dbg(sdev->dev,
|
|
"Attempting iteration %d of Core En/ROM load...\n", i);
|
|
|
|
hda->boot_iteration = i + 1;
|
|
if (chip_info->cl_init)
|
|
ret = chip_info->cl_init(sdev, hext_stream->hstream.stream_tag, false);
|
|
else
|
|
ret = -EINVAL;
|
|
|
|
/* don't retry anymore if successful */
|
|
if (!ret)
|
|
break;
|
|
}
|
|
|
|
if (i == HDA_FW_BOOT_ATTEMPTS) {
|
|
dev_err(sdev->dev, "error: dsp init failed after %d attempts with err: %d\n",
|
|
i, ret);
|
|
goto cleanup;
|
|
}
|
|
|
|
/*
|
|
* When a SoundWire link is in clock stop state, a Slave
|
|
* device may trigger in-band wakes for events such as jack
|
|
* insertion or acoustic event detection. This event will lead
|
|
* to a WAKEEN interrupt, handled by the PCI device and routed
|
|
* to PME if the PCI device is in D3. The resume function in
|
|
* audio PCI driver will be invoked by ACPI for PME event and
|
|
* initialize the device and process WAKEEN interrupt.
|
|
*
|
|
* The WAKEEN interrupt should be processed ASAP to prevent an
|
|
* interrupt flood, otherwise other interrupts, such IPC,
|
|
* cannot work normally. The WAKEEN is handled after the ROM
|
|
* is initialized successfully, which ensures power rails are
|
|
* enabled before accessing the SoundWire SHIM registers
|
|
*/
|
|
if (!sdev->first_boot)
|
|
hda_sdw_process_wakeen(sdev);
|
|
|
|
/*
|
|
* Set the boot_iteration to the last attempt, indicating that the
|
|
* DSP ROM has been initialized and from this point there will be no
|
|
* retry done to boot.
|
|
*
|
|
* Continue with code loading and firmware boot
|
|
*/
|
|
hda->boot_iteration = HDA_FW_BOOT_ATTEMPTS;
|
|
ret = hda_cl_copy_fw(sdev, hext_stream);
|
|
if (!ret) {
|
|
dev_dbg(sdev->dev, "Firmware download successful, booting...\n");
|
|
hda->skip_imr_boot = false;
|
|
} else {
|
|
snd_sof_dsp_dbg_dump(sdev, "Firmware download failed",
|
|
SOF_DBG_DUMP_PCI | SOF_DBG_DUMP_MBOX);
|
|
hda->skip_imr_boot = true;
|
|
}
|
|
|
|
cleanup:
|
|
/*
|
|
* Perform codeloader stream cleanup.
|
|
* This should be done even if firmware loading fails.
|
|
* If the cleanup also fails, we return the initial error
|
|
*/
|
|
ret1 = hda_cl_cleanup(sdev->dev, &dmab, hext_stream);
|
|
if (ret1 < 0) {
|
|
dev_err(sdev->dev, "error: Code loader DSP cleanup failed\n");
|
|
|
|
/* set return value to indicate cleanup failure */
|
|
if (!ret)
|
|
ret = ret1;
|
|
}
|
|
|
|
/*
|
|
* return primary core id if both fw copy
|
|
* and stream clean up are successful
|
|
*/
|
|
if (!ret)
|
|
return chip_info->init_core_mask;
|
|
|
|
/* disable DSP */
|
|
hda_dsp_ctrl_ppcap_enable(sdev, false);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_NS(hda_dsp_cl_boot_firmware, SND_SOC_SOF_INTEL_HDA_COMMON);
|
|
|
|
int hda_dsp_ipc4_load_library(struct snd_sof_dev *sdev,
|
|
struct sof_ipc4_fw_library *fw_lib, bool reload)
|
|
{
|
|
struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
|
|
struct sof_ipc4_fw_data *ipc4_data = sdev->private;
|
|
struct hdac_ext_stream *hext_stream;
|
|
struct firmware stripped_firmware;
|
|
struct sof_ipc4_msg msg = {};
|
|
struct snd_dma_buffer dmab;
|
|
int ret, ret1;
|
|
|
|
/* if IMR booting is enabled and fw context is saved for D3 state, skip the loading */
|
|
if (reload && hda->booted_from_imr && ipc4_data->fw_context_save)
|
|
return 0;
|
|
|
|
/* the fw_lib has been verified during loading, we can trust the validity here */
|
|
stripped_firmware.data = fw_lib->sof_fw.fw->data + fw_lib->sof_fw.payload_offset;
|
|
stripped_firmware.size = fw_lib->sof_fw.fw->size - fw_lib->sof_fw.payload_offset;
|
|
|
|
/* prepare DMA for code loader stream */
|
|
hext_stream = hda_cl_prepare(sdev->dev, HDA_CL_STREAM_FORMAT,
|
|
stripped_firmware.size,
|
|
&dmab, SNDRV_PCM_STREAM_PLAYBACK, false);
|
|
if (IS_ERR(hext_stream)) {
|
|
dev_err(sdev->dev, "%s: DMA prepare failed\n", __func__);
|
|
return PTR_ERR(hext_stream);
|
|
}
|
|
|
|
memcpy(dmab.area, stripped_firmware.data, stripped_firmware.size);
|
|
|
|
/*
|
|
* 1st stage: SOF_IPC4_GLB_LOAD_LIBRARY_PREPARE
|
|
* Message includes the dma_id to be prepared for the library loading.
|
|
* If the firmware does not have support for the message, we will
|
|
* receive -EOPNOTSUPP. In this case we will use single step library
|
|
* loading and proceed to send the LOAD_LIBRARY message.
|
|
*/
|
|
msg.primary = hext_stream->hstream.stream_tag - 1;
|
|
msg.primary |= SOF_IPC4_MSG_TYPE_SET(SOF_IPC4_GLB_LOAD_LIBRARY_PREPARE);
|
|
msg.primary |= SOF_IPC4_MSG_DIR(SOF_IPC4_MSG_REQUEST);
|
|
msg.primary |= SOF_IPC4_MSG_TARGET(SOF_IPC4_FW_GEN_MSG);
|
|
ret = sof_ipc_tx_message_no_reply(sdev->ipc, &msg, 0);
|
|
if (!ret) {
|
|
int sd_offset = SOF_STREAM_SD_OFFSET(&hext_stream->hstream);
|
|
unsigned int status;
|
|
|
|
/*
|
|
* Make sure that the FIFOS value is not 0 in SDxFIFOS register
|
|
* which indicates that the firmware set the GEN bit and we can
|
|
* continue to start the DMA
|
|
*/
|
|
ret = snd_sof_dsp_read_poll_timeout(sdev, HDA_DSP_HDA_BAR,
|
|
sd_offset + SOF_HDA_ADSP_REG_SD_FIFOSIZE,
|
|
status,
|
|
status & SOF_HDA_SD_FIFOSIZE_FIFOS_MASK,
|
|
HDA_DSP_REG_POLL_INTERVAL_US,
|
|
HDA_DSP_BASEFW_TIMEOUT_US);
|
|
|
|
if (ret < 0)
|
|
dev_warn(sdev->dev,
|
|
"%s: timeout waiting for FIFOS\n", __func__);
|
|
} else if (ret != -EOPNOTSUPP) {
|
|
goto cleanup;
|
|
}
|
|
|
|
ret = hda_cl_trigger(sdev->dev, hext_stream, SNDRV_PCM_TRIGGER_START);
|
|
if (ret < 0) {
|
|
dev_err(sdev->dev, "%s: DMA trigger start failed\n", __func__);
|
|
goto cleanup;
|
|
}
|
|
|
|
/*
|
|
* 2nd stage: LOAD_LIBRARY
|
|
* Message includes the dma_id and the lib_id, the dma_id must be
|
|
* identical to the one sent via LOAD_LIBRARY_PREPARE
|
|
*/
|
|
msg.primary &= ~SOF_IPC4_MSG_TYPE_MASK;
|
|
msg.primary |= SOF_IPC4_MSG_TYPE_SET(SOF_IPC4_GLB_LOAD_LIBRARY);
|
|
msg.primary |= SOF_IPC4_GLB_LOAD_LIBRARY_LIB_ID(fw_lib->id);
|
|
ret = sof_ipc_tx_message_no_reply(sdev->ipc, &msg, 0);
|
|
|
|
/* Stop the DMA channel */
|
|
ret1 = hda_cl_trigger(sdev->dev, hext_stream, SNDRV_PCM_TRIGGER_STOP);
|
|
if (ret1 < 0) {
|
|
dev_err(sdev->dev, "%s: DMA trigger stop failed\n", __func__);
|
|
if (!ret)
|
|
ret = ret1;
|
|
}
|
|
|
|
cleanup:
|
|
/* clean up even in case of error and return the first error */
|
|
ret1 = hda_cl_cleanup(sdev->dev, &dmab, hext_stream);
|
|
if (ret1 < 0) {
|
|
dev_err(sdev->dev, "%s: Code loader DSP cleanup failed\n", __func__);
|
|
|
|
/* set return value to indicate cleanup failure */
|
|
if (!ret)
|
|
ret = ret1;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_NS(hda_dsp_ipc4_load_library, SND_SOC_SOF_INTEL_HDA_COMMON);
|
|
|
|
int hda_dsp_ext_man_get_cavs_config_data(struct snd_sof_dev *sdev,
|
|
const struct sof_ext_man_elem_header *hdr)
|
|
{
|
|
const struct sof_ext_man_cavs_config_data *config_data =
|
|
container_of(hdr, struct sof_ext_man_cavs_config_data, hdr);
|
|
struct sof_intel_hda_dev *hda = sdev->pdata->hw_pdata;
|
|
int i, elem_num;
|
|
|
|
/* calculate total number of config data elements */
|
|
elem_num = (hdr->size - sizeof(struct sof_ext_man_elem_header))
|
|
/ sizeof(struct sof_config_elem);
|
|
if (elem_num <= 0) {
|
|
dev_err(sdev->dev, "cavs config data is inconsistent: %d\n", elem_num);
|
|
return -EINVAL;
|
|
}
|
|
|
|
for (i = 0; i < elem_num; i++)
|
|
switch (config_data->elems[i].token) {
|
|
case SOF_EXT_MAN_CAVS_CONFIG_EMPTY:
|
|
/* skip empty token */
|
|
break;
|
|
case SOF_EXT_MAN_CAVS_CONFIG_CAVS_LPRO:
|
|
hda->clk_config_lpro = config_data->elems[i].value;
|
|
dev_dbg(sdev->dev, "FW clock config: %s\n",
|
|
hda->clk_config_lpro ? "LPRO" : "HPRO");
|
|
break;
|
|
case SOF_EXT_MAN_CAVS_CONFIG_OUTBOX_SIZE:
|
|
case SOF_EXT_MAN_CAVS_CONFIG_INBOX_SIZE:
|
|
/* These elements are defined but not being used yet. No warn is required */
|
|
break;
|
|
default:
|
|
dev_info(sdev->dev, "unsupported token type: %d\n",
|
|
config_data->elems[i].token);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_NS(hda_dsp_ext_man_get_cavs_config_data, SND_SOC_SOF_INTEL_HDA_COMMON);
|