mirror of
https://github.com/torvalds/linux.git
synced 2024-12-27 13:22:23 +00:00
Merge branch 'pci/controller/endpoint'
- Change "PCI Endpoint Virtual NTB driver" Kconfig prompt to be different from "PCI Endpoint NTB driver" (Shunsuke Mie) - Automatically create a function specific attributes group for endpoint drivers to avoid reference counting issues (Damien Le Moal) - Move and unexport pci_epf_type_add_cfs() (Damien Le Moal) - Reinitialize EPF test DMA transfer completion before submitting it to avoid losing the completion notification (Damien Le Moal) - Fix EPF test DMA transfer completion detection (Damien Le Moal) - Submit EPF test DMA transfers with dmaengine_submit(), not tx_submit() (Damien Le Moal) - Simplify EPF test read/write/copy functions (Damien Le Moal) - Simplify EPF test "raise IRQ" interface (Damien Le Moal) - Simplify EPF test IRQ command execution (Damien Le Moal) - Improve EPF test command/status register handling (Damien Le Moal) - Free IRQs before removing device (Damien Le Moal) - Reinitialize IRQ completions for every test (Damien Le Moal) - Don't write status in IRQ handler to avoid race (Damien Le Moal) - Fix dma_chan direction in data transfer test (Yoshihiro Shimoda) - Return pci_epf_type_add_cfs() error if EPF has no driver (Damien Le Moal) - Add kernel-doc for pci_epc_raise_irq() and pci_epc_map_msi_irq() MSI vector parameters (Manivannan Sadhasivam) - Pass EPF device ID to driver probe functions (Manivannan Sadhasivam) - Return -EALREADY if EPC has already been started/stopped (Manivannan Sadhasivam) - Add linkdown notifier support and use it in qcom-ep (Manivannan Sadhasivam) - Add Bus Master Enable event support and use it in qcom-ep (Manivannan Sadhasivam) - Add Qualcomm Modem Host Interface (MHI) endpoint driver (Manivannan Sadhasivam) - Add Layerscape PME interrupt handling to manage link-up notification (Frank Li) * pci/controller/endpoint: PCI: layerscape: Add the endpoint linkup notifier support PCI: endpoint: pci-epf-vntb: Fix typo in comments MAINTAINERS: Add PCI MHI endpoint function driver under MHI bus PCI: endpoint: Add PCI Endpoint function driver for MHI bus PCI: qcom-ep: Add support for BME notification PCI: qcom-ep: Add support for Link down notification PCI: endpoint: Add BME notifier support PCI: endpoint: Add linkdown notifier support PCI: endpoint: Return error if EPC is started/stopped multiple times PCI: endpoint: Pass EPF device ID to the probe function PCI: endpoint: Add missing documentation about the MSI/MSI-X range PCI: endpoint: Improve pci_epf_type_add_cfs() PCI: endpoint: functions/pci-epf-test: Fix dma_chan direction misc: pci_endpoint_test: Simplify pci_endpoint_test_msi_irq() misc: pci_endpoint_test: Do not write status in IRQ handler misc: pci_endpoint_test: Re-init completion for every test misc: pci_endpoint_test: Free IRQs before removing the device PCI: epf-test: Simplify transfers result print PCI: epf-test: Simplify DMA support checks PCI: epf-test: Cleanup request result handling PCI: epf-test: Cleanup pci_epf_test_cmd_handler() PCI: epf-test: Improve handling of command and status registers PCI: epf-test: Simplify IRQ test commands execution PCI: epf-test: Simplify pci_epf_test_raise_irq() PCI: epf-test: Simplify read/write/copy test functions PCI: epf-test: Use dmaengine_submit() to initiate DMA transfer PCI: epf-test: Fix DMA transfer completion detection PCI: epf-test: Fix DMA transfer completion initialization PCI: endpoint: Move pci_epf_type_add_cfs() code PCI: endpoint: Automatically create a function specific attributes group PCI: endpoint: Fix a Kconfig prompt of vNTB driver
This commit is contained in:
commit
d8c226ac1f
@ -88,13 +88,10 @@ commands can be used::
|
||||
# echo 0x104c > functions/pci_epf_ntb/func1/vendorid
|
||||
# echo 0xb00d > functions/pci_epf_ntb/func1/deviceid
|
||||
|
||||
In order to configure NTB specific attributes, a new sub-directory to func1
|
||||
should be created::
|
||||
|
||||
# mkdir functions/pci_epf_ntb/func1/pci_epf_ntb.0/
|
||||
|
||||
The NTB function driver will populate this directory with various attributes
|
||||
that can be configured by the user::
|
||||
The PCI endpoint framework also automatically creates a sub-directory in the
|
||||
function attribute directory. This sub-directory has the same name as the name
|
||||
of the function device and is populated with the following NTB specific
|
||||
attributes that can be configured by the user::
|
||||
|
||||
# ls functions/pci_epf_ntb/func1/pci_epf_ntb.0/
|
||||
db_count mw1 mw2 mw3 mw4 num_mws
|
||||
|
@ -84,13 +84,10 @@ commands can be used::
|
||||
# echo 0x1957 > functions/pci_epf_vntb/func1/vendorid
|
||||
# echo 0x0809 > functions/pci_epf_vntb/func1/deviceid
|
||||
|
||||
In order to configure NTB specific attributes, a new sub-directory to func1
|
||||
should be created::
|
||||
|
||||
# mkdir functions/pci_epf_vntb/func1/pci_epf_vntb.0/
|
||||
|
||||
The NTB function driver will populate this directory with various attributes
|
||||
that can be configured by the user::
|
||||
The PCI endpoint framework also automatically creates a sub-directory in the
|
||||
function attribute directory. This sub-directory has the same name as the name
|
||||
of the function device and is populated with the following NTB specific
|
||||
attributes that can be configured by the user::
|
||||
|
||||
# ls functions/pci_epf_vntb/func1/pci_epf_vntb.0/
|
||||
db_count mw1 mw2 mw3 mw4 num_mws
|
||||
|
@ -13629,6 +13629,7 @@ T: git git://git.kernel.org/pub/scm/linux/kernel/git/mani/mhi.git
|
||||
F: Documentation/ABI/stable/sysfs-bus-mhi
|
||||
F: Documentation/mhi/
|
||||
F: drivers/bus/mhi/
|
||||
F: drivers/pci/endpoint/functions/pci-epf-mhi.c
|
||||
F: include/linux/mhi.h
|
||||
|
||||
MICROBLAZE ARCHITECTURE
|
||||
|
@ -159,10 +159,7 @@ static irqreturn_t pci_endpoint_test_irqhandler(int irq, void *dev_id)
|
||||
if (reg & STATUS_IRQ_RAISED) {
|
||||
test->last_irq = irq;
|
||||
complete(&test->irq_raised);
|
||||
reg &= ~STATUS_IRQ_RAISED;
|
||||
}
|
||||
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_STATUS,
|
||||
reg);
|
||||
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
@ -316,21 +313,17 @@ static bool pci_endpoint_test_msi_irq(struct pci_endpoint_test *test,
|
||||
struct pci_dev *pdev = test->pdev;
|
||||
|
||||
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE,
|
||||
msix == false ? IRQ_TYPE_MSI :
|
||||
IRQ_TYPE_MSIX);
|
||||
msix ? IRQ_TYPE_MSIX : IRQ_TYPE_MSI);
|
||||
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, msi_num);
|
||||
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND,
|
||||
msix == false ? COMMAND_RAISE_MSI_IRQ :
|
||||
COMMAND_RAISE_MSIX_IRQ);
|
||||
msix ? COMMAND_RAISE_MSIX_IRQ :
|
||||
COMMAND_RAISE_MSI_IRQ);
|
||||
val = wait_for_completion_timeout(&test->irq_raised,
|
||||
msecs_to_jiffies(1000));
|
||||
if (!val)
|
||||
return false;
|
||||
|
||||
if (pci_irq_vector(pdev, msi_num - 1) == test->last_irq)
|
||||
return true;
|
||||
|
||||
return false;
|
||||
return pci_irq_vector(pdev, msi_num - 1) == test->last_irq;
|
||||
}
|
||||
|
||||
static int pci_endpoint_test_validate_xfer_params(struct device *dev,
|
||||
@ -729,6 +722,10 @@ static long pci_endpoint_test_ioctl(struct file *file, unsigned int cmd,
|
||||
struct pci_dev *pdev = test->pdev;
|
||||
|
||||
mutex_lock(&test->mutex);
|
||||
|
||||
reinit_completion(&test->irq_raised);
|
||||
test->last_irq = -ENODATA;
|
||||
|
||||
switch (cmd) {
|
||||
case PCITEST_BAR:
|
||||
bar = arg;
|
||||
@ -938,6 +935,9 @@ static void pci_endpoint_test_remove(struct pci_dev *pdev)
|
||||
if (id < 0)
|
||||
return;
|
||||
|
||||
pci_endpoint_test_release_irq(test);
|
||||
pci_endpoint_test_free_irq_vectors(test);
|
||||
|
||||
misc_deregister(&test->miscdev);
|
||||
kfree(misc_device->name);
|
||||
kfree(test->name);
|
||||
@ -947,9 +947,6 @@ static void pci_endpoint_test_remove(struct pci_dev *pdev)
|
||||
pci_iounmap(pdev, test->bar[bar]);
|
||||
}
|
||||
|
||||
pci_endpoint_test_release_irq(test);
|
||||
pci_endpoint_test_free_irq_vectors(test);
|
||||
|
||||
pci_release_regions(pdev);
|
||||
pci_disable_device(pdev);
|
||||
}
|
||||
|
@ -18,6 +18,20 @@
|
||||
|
||||
#include "pcie-designware.h"
|
||||
|
||||
#define PEX_PF0_CONFIG 0xC0014
|
||||
#define PEX_PF0_CFG_READY BIT(0)
|
||||
|
||||
/* PEX PFa PCIE PME and message interrupt registers*/
|
||||
#define PEX_PF0_PME_MES_DR 0xC0020
|
||||
#define PEX_PF0_PME_MES_DR_LUD BIT(7)
|
||||
#define PEX_PF0_PME_MES_DR_LDD BIT(9)
|
||||
#define PEX_PF0_PME_MES_DR_HRD BIT(10)
|
||||
|
||||
#define PEX_PF0_PME_MES_IER 0xC0028
|
||||
#define PEX_PF0_PME_MES_IER_LUDIE BIT(7)
|
||||
#define PEX_PF0_PME_MES_IER_LDDIE BIT(9)
|
||||
#define PEX_PF0_PME_MES_IER_HRDIE BIT(10)
|
||||
|
||||
#define to_ls_pcie_ep(x) dev_get_drvdata((x)->dev)
|
||||
|
||||
struct ls_pcie_ep_drvdata {
|
||||
@ -30,8 +44,84 @@ struct ls_pcie_ep {
|
||||
struct dw_pcie *pci;
|
||||
struct pci_epc_features *ls_epc;
|
||||
const struct ls_pcie_ep_drvdata *drvdata;
|
||||
int irq;
|
||||
bool big_endian;
|
||||
};
|
||||
|
||||
static u32 ls_lut_readl(struct ls_pcie_ep *pcie, u32 offset)
|
||||
{
|
||||
struct dw_pcie *pci = pcie->pci;
|
||||
|
||||
if (pcie->big_endian)
|
||||
return ioread32be(pci->dbi_base + offset);
|
||||
else
|
||||
return ioread32(pci->dbi_base + offset);
|
||||
}
|
||||
|
||||
static void ls_lut_writel(struct ls_pcie_ep *pcie, u32 offset, u32 value)
|
||||
{
|
||||
struct dw_pcie *pci = pcie->pci;
|
||||
|
||||
if (pcie->big_endian)
|
||||
iowrite32be(value, pci->dbi_base + offset);
|
||||
else
|
||||
iowrite32(value, pci->dbi_base + offset);
|
||||
}
|
||||
|
||||
static irqreturn_t ls_pcie_ep_event_handler(int irq, void *dev_id)
|
||||
{
|
||||
struct ls_pcie_ep *pcie = dev_id;
|
||||
struct dw_pcie *pci = pcie->pci;
|
||||
u32 val, cfg;
|
||||
|
||||
val = ls_lut_readl(pcie, PEX_PF0_PME_MES_DR);
|
||||
ls_lut_writel(pcie, PEX_PF0_PME_MES_DR, val);
|
||||
|
||||
if (!val)
|
||||
return IRQ_NONE;
|
||||
|
||||
if (val & PEX_PF0_PME_MES_DR_LUD) {
|
||||
cfg = ls_lut_readl(pcie, PEX_PF0_CONFIG);
|
||||
cfg |= PEX_PF0_CFG_READY;
|
||||
ls_lut_writel(pcie, PEX_PF0_CONFIG, cfg);
|
||||
dw_pcie_ep_linkup(&pci->ep);
|
||||
|
||||
dev_dbg(pci->dev, "Link up\n");
|
||||
} else if (val & PEX_PF0_PME_MES_DR_LDD) {
|
||||
dev_dbg(pci->dev, "Link down\n");
|
||||
} else if (val & PEX_PF0_PME_MES_DR_HRD) {
|
||||
dev_dbg(pci->dev, "Hot reset\n");
|
||||
}
|
||||
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
|
||||
static int ls_pcie_ep_interrupt_init(struct ls_pcie_ep *pcie,
|
||||
struct platform_device *pdev)
|
||||
{
|
||||
u32 val;
|
||||
int ret;
|
||||
|
||||
pcie->irq = platform_get_irq_byname(pdev, "pme");
|
||||
if (pcie->irq < 0)
|
||||
return pcie->irq;
|
||||
|
||||
ret = devm_request_irq(&pdev->dev, pcie->irq, ls_pcie_ep_event_handler,
|
||||
IRQF_SHARED, pdev->name, pcie);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev, "Can't register PCIe IRQ\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* Enable interrupts */
|
||||
val = ls_lut_readl(pcie, PEX_PF0_PME_MES_IER);
|
||||
val |= PEX_PF0_PME_MES_IER_LDDIE | PEX_PF0_PME_MES_IER_HRDIE |
|
||||
PEX_PF0_PME_MES_IER_LUDIE;
|
||||
ls_lut_writel(pcie, PEX_PF0_PME_MES_IER, val);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static const struct pci_epc_features*
|
||||
ls_pcie_ep_get_features(struct dw_pcie_ep *ep)
|
||||
{
|
||||
@ -125,6 +215,7 @@ static int __init ls_pcie_ep_probe(struct platform_device *pdev)
|
||||
struct ls_pcie_ep *pcie;
|
||||
struct pci_epc_features *ls_epc;
|
||||
struct resource *dbi_base;
|
||||
int ret;
|
||||
|
||||
pcie = devm_kzalloc(dev, sizeof(*pcie), GFP_KERNEL);
|
||||
if (!pcie)
|
||||
@ -144,6 +235,7 @@ static int __init ls_pcie_ep_probe(struct platform_device *pdev)
|
||||
pci->ops = pcie->drvdata->dw_pcie_ops;
|
||||
|
||||
ls_epc->bar_fixed_64bit = (1 << BAR_2) | (1 << BAR_4);
|
||||
ls_epc->linkup_notifier = true;
|
||||
|
||||
pcie->pci = pci;
|
||||
pcie->ls_epc = ls_epc;
|
||||
@ -155,9 +247,15 @@ static int __init ls_pcie_ep_probe(struct platform_device *pdev)
|
||||
|
||||
pci->ep.ops = &ls_pcie_ep_ops;
|
||||
|
||||
pcie->big_endian = of_property_read_bool(dev->of_node, "big-endian");
|
||||
|
||||
platform_set_drvdata(pdev, pcie);
|
||||
|
||||
return dw_pcie_ep_init(&pci->ep);
|
||||
ret = dw_pcie_ep_init(&pci->ep);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
return ls_pcie_ep_interrupt_init(pcie, pdev);
|
||||
}
|
||||
|
||||
static struct platform_driver ls_pcie_ep_driver = {
|
||||
|
@ -569,9 +569,11 @@ static irqreturn_t qcom_pcie_ep_global_irq_thread(int irq, void *data)
|
||||
if (FIELD_GET(PARF_INT_ALL_LINK_DOWN, status)) {
|
||||
dev_dbg(dev, "Received Linkdown event\n");
|
||||
pcie_ep->link_status = QCOM_PCIE_EP_LINK_DOWN;
|
||||
pci_epc_linkdown(pci->ep.epc);
|
||||
} else if (FIELD_GET(PARF_INT_ALL_BME, status)) {
|
||||
dev_dbg(dev, "Received BME event. Link is enabled!\n");
|
||||
pcie_ep->link_status = QCOM_PCIE_EP_LINK_ENABLED;
|
||||
pci_epc_bme_notify(pci->ep.epc);
|
||||
} else if (FIELD_GET(PARF_INT_ALL_PM_TURNOFF, status)) {
|
||||
dev_dbg(dev, "Received PM Turn-off event! Entering L23\n");
|
||||
val = readl_relaxed(pcie_ep->parf + PARF_PM_CTRL);
|
||||
|
@ -27,7 +27,7 @@ config PCI_EPF_NTB
|
||||
If in doubt, say "N" to disable Endpoint NTB driver.
|
||||
|
||||
config PCI_EPF_VNTB
|
||||
tristate "PCI Endpoint NTB driver"
|
||||
tristate "PCI Endpoint Virtual NTB driver"
|
||||
depends on PCI_ENDPOINT
|
||||
depends on NTB
|
||||
select CONFIGFS_FS
|
||||
@ -37,3 +37,13 @@ config PCI_EPF_VNTB
|
||||
between PCI Root Port and PCIe Endpoint.
|
||||
|
||||
If in doubt, say "N" to disable Endpoint NTB driver.
|
||||
|
||||
config PCI_EPF_MHI
|
||||
tristate "PCI Endpoint driver for MHI bus"
|
||||
depends on PCI_ENDPOINT && MHI_BUS_EP
|
||||
help
|
||||
Enable this configuration option to enable the PCI Endpoint
|
||||
driver for Modem Host Interface (MHI) bus in Qualcomm Endpoint
|
||||
devices such as SDX55.
|
||||
|
||||
If in doubt, say "N" to disable Endpoint driver for MHI bus.
|
||||
|
@ -6,3 +6,4 @@
|
||||
obj-$(CONFIG_PCI_EPF_TEST) += pci-epf-test.o
|
||||
obj-$(CONFIG_PCI_EPF_NTB) += pci-epf-ntb.o
|
||||
obj-$(CONFIG_PCI_EPF_VNTB) += pci-epf-vntb.o
|
||||
obj-$(CONFIG_PCI_EPF_MHI) += pci-epf-mhi.o
|
||||
|
458
drivers/pci/endpoint/functions/pci-epf-mhi.c
Normal file
458
drivers/pci/endpoint/functions/pci-epf-mhi.c
Normal file
@ -0,0 +1,458 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
/*
|
||||
* PCI EPF driver for MHI Endpoint devices
|
||||
*
|
||||
* Copyright (C) 2023 Linaro Ltd.
|
||||
* Author: Manivannan Sadhasivam <manivannan.sadhasivam@linaro.org>
|
||||
*/
|
||||
|
||||
#include <linux/mhi_ep.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/pci-epc.h>
|
||||
#include <linux/pci-epf.h>
|
||||
|
||||
#define MHI_VERSION_1_0 0x01000000
|
||||
|
||||
#define to_epf_mhi(cntrl) container_of(cntrl, struct pci_epf_mhi, cntrl)
|
||||
|
||||
struct pci_epf_mhi_ep_info {
|
||||
const struct mhi_ep_cntrl_config *config;
|
||||
struct pci_epf_header *epf_header;
|
||||
enum pci_barno bar_num;
|
||||
u32 epf_flags;
|
||||
u32 msi_count;
|
||||
u32 mru;
|
||||
};
|
||||
|
||||
#define MHI_EP_CHANNEL_CONFIG(ch_num, ch_name, direction) \
|
||||
{ \
|
||||
.num = ch_num, \
|
||||
.name = ch_name, \
|
||||
.dir = direction, \
|
||||
}
|
||||
|
||||
#define MHI_EP_CHANNEL_CONFIG_UL(ch_num, ch_name) \
|
||||
MHI_EP_CHANNEL_CONFIG(ch_num, ch_name, DMA_TO_DEVICE)
|
||||
|
||||
#define MHI_EP_CHANNEL_CONFIG_DL(ch_num, ch_name) \
|
||||
MHI_EP_CHANNEL_CONFIG(ch_num, ch_name, DMA_FROM_DEVICE)
|
||||
|
||||
static const struct mhi_ep_channel_config mhi_v1_channels[] = {
|
||||
MHI_EP_CHANNEL_CONFIG_UL(0, "LOOPBACK"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(1, "LOOPBACK"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(2, "SAHARA"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(3, "SAHARA"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(4, "DIAG"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(5, "DIAG"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(6, "SSR"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(7, "SSR"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(8, "QDSS"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(9, "QDSS"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(10, "EFS"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(11, "EFS"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(12, "MBIM"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(13, "MBIM"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(14, "QMI"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(15, "QMI"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(16, "QMI"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(17, "QMI"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(18, "IP-CTRL-1"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(19, "IP-CTRL-1"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(20, "IPCR"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(21, "IPCR"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(32, "DUN"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(33, "DUN"),
|
||||
MHI_EP_CHANNEL_CONFIG_UL(46, "IP_SW0"),
|
||||
MHI_EP_CHANNEL_CONFIG_DL(47, "IP_SW0"),
|
||||
};
|
||||
|
||||
static const struct mhi_ep_cntrl_config mhi_v1_config = {
|
||||
.max_channels = 128,
|
||||
.num_channels = ARRAY_SIZE(mhi_v1_channels),
|
||||
.ch_cfg = mhi_v1_channels,
|
||||
.mhi_version = MHI_VERSION_1_0,
|
||||
};
|
||||
|
||||
static struct pci_epf_header sdx55_header = {
|
||||
.vendorid = PCI_VENDOR_ID_QCOM,
|
||||
.deviceid = 0x0306,
|
||||
.baseclass_code = PCI_BASE_CLASS_COMMUNICATION,
|
||||
.subclass_code = PCI_CLASS_COMMUNICATION_MODEM & 0xff,
|
||||
.interrupt_pin = PCI_INTERRUPT_INTA,
|
||||
};
|
||||
|
||||
static const struct pci_epf_mhi_ep_info sdx55_info = {
|
||||
.config = &mhi_v1_config,
|
||||
.epf_header = &sdx55_header,
|
||||
.bar_num = BAR_0,
|
||||
.epf_flags = PCI_BASE_ADDRESS_MEM_TYPE_32,
|
||||
.msi_count = 32,
|
||||
.mru = 0x8000,
|
||||
};
|
||||
|
||||
struct pci_epf_mhi {
|
||||
const struct pci_epf_mhi_ep_info *info;
|
||||
struct mhi_ep_cntrl mhi_cntrl;
|
||||
struct pci_epf *epf;
|
||||
struct mutex lock;
|
||||
void __iomem *mmio;
|
||||
resource_size_t mmio_phys;
|
||||
u32 mmio_size;
|
||||
int irq;
|
||||
};
|
||||
|
||||
static int __pci_epf_mhi_alloc_map(struct mhi_ep_cntrl *mhi_cntrl, u64 pci_addr,
|
||||
phys_addr_t *paddr, void __iomem **vaddr,
|
||||
size_t offset, size_t size)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = to_epf_mhi(mhi_cntrl);
|
||||
struct pci_epf *epf = epf_mhi->epf;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
int ret;
|
||||
|
||||
*vaddr = pci_epc_mem_alloc_addr(epc, paddr, size + offset);
|
||||
if (!*vaddr)
|
||||
return -ENOMEM;
|
||||
|
||||
ret = pci_epc_map_addr(epc, epf->func_no, epf->vfunc_no, *paddr,
|
||||
pci_addr - offset, size + offset);
|
||||
if (ret) {
|
||||
pci_epc_mem_free_addr(epc, *paddr, *vaddr, size + offset);
|
||||
return ret;
|
||||
}
|
||||
|
||||
*paddr = *paddr + offset;
|
||||
*vaddr = *vaddr + offset;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_alloc_map(struct mhi_ep_cntrl *mhi_cntrl, u64 pci_addr,
|
||||
phys_addr_t *paddr, void __iomem **vaddr,
|
||||
size_t size)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = to_epf_mhi(mhi_cntrl);
|
||||
struct pci_epc *epc = epf_mhi->epf->epc;
|
||||
size_t offset = pci_addr & (epc->mem->window.page_size - 1);
|
||||
|
||||
return __pci_epf_mhi_alloc_map(mhi_cntrl, pci_addr, paddr, vaddr,
|
||||
offset, size);
|
||||
}
|
||||
|
||||
static void __pci_epf_mhi_unmap_free(struct mhi_ep_cntrl *mhi_cntrl,
|
||||
u64 pci_addr, phys_addr_t paddr,
|
||||
void __iomem *vaddr, size_t offset,
|
||||
size_t size)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = to_epf_mhi(mhi_cntrl);
|
||||
struct pci_epf *epf = epf_mhi->epf;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
|
||||
pci_epc_unmap_addr(epc, epf->func_no, epf->vfunc_no, paddr - offset);
|
||||
pci_epc_mem_free_addr(epc, paddr - offset, vaddr - offset,
|
||||
size + offset);
|
||||
}
|
||||
|
||||
static void pci_epf_mhi_unmap_free(struct mhi_ep_cntrl *mhi_cntrl, u64 pci_addr,
|
||||
phys_addr_t paddr, void __iomem *vaddr,
|
||||
size_t size)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = to_epf_mhi(mhi_cntrl);
|
||||
struct pci_epf *epf = epf_mhi->epf;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
size_t offset = pci_addr & (epc->mem->window.page_size - 1);
|
||||
|
||||
__pci_epf_mhi_unmap_free(mhi_cntrl, pci_addr, paddr, vaddr, offset,
|
||||
size);
|
||||
}
|
||||
|
||||
static void pci_epf_mhi_raise_irq(struct mhi_ep_cntrl *mhi_cntrl, u32 vector)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = to_epf_mhi(mhi_cntrl);
|
||||
struct pci_epf *epf = epf_mhi->epf;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
|
||||
/*
|
||||
* MHI supplies 0 based MSI vectors but the API expects the vector
|
||||
* number to start from 1, so we need to increment the vector by 1.
|
||||
*/
|
||||
pci_epc_raise_irq(epc, epf->func_no, epf->vfunc_no, PCI_EPC_IRQ_MSI,
|
||||
vector + 1);
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_read_from_host(struct mhi_ep_cntrl *mhi_cntrl, u64 from,
|
||||
void *to, size_t size)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = to_epf_mhi(mhi_cntrl);
|
||||
size_t offset = from % SZ_4K;
|
||||
void __iomem *tre_buf;
|
||||
phys_addr_t tre_phys;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&epf_mhi->lock);
|
||||
|
||||
ret = __pci_epf_mhi_alloc_map(mhi_cntrl, from, &tre_phys, &tre_buf,
|
||||
offset, size);
|
||||
if (ret) {
|
||||
mutex_unlock(&epf_mhi->lock);
|
||||
return ret;
|
||||
}
|
||||
|
||||
memcpy_fromio(to, tre_buf, size);
|
||||
|
||||
__pci_epf_mhi_unmap_free(mhi_cntrl, from, tre_phys, tre_buf, offset,
|
||||
size);
|
||||
|
||||
mutex_unlock(&epf_mhi->lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_write_to_host(struct mhi_ep_cntrl *mhi_cntrl,
|
||||
void *from, u64 to, size_t size)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = to_epf_mhi(mhi_cntrl);
|
||||
size_t offset = to % SZ_4K;
|
||||
void __iomem *tre_buf;
|
||||
phys_addr_t tre_phys;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&epf_mhi->lock);
|
||||
|
||||
ret = __pci_epf_mhi_alloc_map(mhi_cntrl, to, &tre_phys, &tre_buf,
|
||||
offset, size);
|
||||
if (ret) {
|
||||
mutex_unlock(&epf_mhi->lock);
|
||||
return ret;
|
||||
}
|
||||
|
||||
memcpy_toio(tre_buf, from, size);
|
||||
|
||||
__pci_epf_mhi_unmap_free(mhi_cntrl, to, tre_phys, tre_buf, offset,
|
||||
size);
|
||||
|
||||
mutex_unlock(&epf_mhi->lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_core_init(struct pci_epf *epf)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = epf_get_drvdata(epf);
|
||||
const struct pci_epf_mhi_ep_info *info = epf_mhi->info;
|
||||
struct pci_epf_bar *epf_bar = &epf->bar[info->bar_num];
|
||||
struct pci_epc *epc = epf->epc;
|
||||
struct device *dev = &epf->dev;
|
||||
int ret;
|
||||
|
||||
epf_bar->phys_addr = epf_mhi->mmio_phys;
|
||||
epf_bar->size = epf_mhi->mmio_size;
|
||||
epf_bar->barno = info->bar_num;
|
||||
epf_bar->flags = info->epf_flags;
|
||||
ret = pci_epc_set_bar(epc, epf->func_no, epf->vfunc_no, epf_bar);
|
||||
if (ret) {
|
||||
dev_err(dev, "Failed to set BAR: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = pci_epc_set_msi(epc, epf->func_no, epf->vfunc_no,
|
||||
order_base_2(info->msi_count));
|
||||
if (ret) {
|
||||
dev_err(dev, "Failed to set MSI configuration: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = pci_epc_write_header(epc, epf->func_no, epf->vfunc_no,
|
||||
epf->header);
|
||||
if (ret) {
|
||||
dev_err(dev, "Failed to set Configuration header: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_link_up(struct pci_epf *epf)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = epf_get_drvdata(epf);
|
||||
const struct pci_epf_mhi_ep_info *info = epf_mhi->info;
|
||||
struct mhi_ep_cntrl *mhi_cntrl = &epf_mhi->mhi_cntrl;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
struct device *dev = &epf->dev;
|
||||
int ret;
|
||||
|
||||
mhi_cntrl->mmio = epf_mhi->mmio;
|
||||
mhi_cntrl->irq = epf_mhi->irq;
|
||||
mhi_cntrl->mru = info->mru;
|
||||
|
||||
/* Assign the struct dev of PCI EP as MHI controller device */
|
||||
mhi_cntrl->cntrl_dev = epc->dev.parent;
|
||||
mhi_cntrl->raise_irq = pci_epf_mhi_raise_irq;
|
||||
mhi_cntrl->alloc_map = pci_epf_mhi_alloc_map;
|
||||
mhi_cntrl->unmap_free = pci_epf_mhi_unmap_free;
|
||||
mhi_cntrl->read_from_host = pci_epf_mhi_read_from_host;
|
||||
mhi_cntrl->write_to_host = pci_epf_mhi_write_to_host;
|
||||
|
||||
/* Register the MHI EP controller */
|
||||
ret = mhi_ep_register_controller(mhi_cntrl, info->config);
|
||||
if (ret) {
|
||||
dev_err(dev, "Failed to register MHI EP controller: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_link_down(struct pci_epf *epf)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = epf_get_drvdata(epf);
|
||||
struct mhi_ep_cntrl *mhi_cntrl = &epf_mhi->mhi_cntrl;
|
||||
|
||||
if (mhi_cntrl->mhi_dev) {
|
||||
mhi_ep_power_down(mhi_cntrl);
|
||||
mhi_ep_unregister_controller(mhi_cntrl);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_bme(struct pci_epf *epf)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = epf_get_drvdata(epf);
|
||||
struct mhi_ep_cntrl *mhi_cntrl = &epf_mhi->mhi_cntrl;
|
||||
struct device *dev = &epf->dev;
|
||||
int ret;
|
||||
|
||||
/*
|
||||
* Power up the MHI EP stack if link is up and stack is in power down
|
||||
* state.
|
||||
*/
|
||||
if (!mhi_cntrl->enabled && mhi_cntrl->mhi_dev) {
|
||||
ret = mhi_ep_power_up(mhi_cntrl);
|
||||
if (ret) {
|
||||
dev_err(dev, "Failed to power up MHI EP: %d\n", ret);
|
||||
mhi_ep_unregister_controller(mhi_cntrl);
|
||||
}
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pci_epf_mhi_bind(struct pci_epf *epf)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = epf_get_drvdata(epf);
|
||||
struct pci_epc *epc = epf->epc;
|
||||
struct platform_device *pdev = to_platform_device(epc->dev.parent);
|
||||
struct resource *res;
|
||||
int ret;
|
||||
|
||||
/* Get MMIO base address from Endpoint controller */
|
||||
res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "mmio");
|
||||
epf_mhi->mmio_phys = res->start;
|
||||
epf_mhi->mmio_size = resource_size(res);
|
||||
|
||||
epf_mhi->mmio = ioremap(epf_mhi->mmio_phys, epf_mhi->mmio_size);
|
||||
if (!epf_mhi->mmio)
|
||||
return -ENOMEM;
|
||||
|
||||
ret = platform_get_irq_byname(pdev, "doorbell");
|
||||
if (ret < 0) {
|
||||
iounmap(epf_mhi->mmio);
|
||||
return ret;
|
||||
}
|
||||
|
||||
epf_mhi->irq = ret;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void pci_epf_mhi_unbind(struct pci_epf *epf)
|
||||
{
|
||||
struct pci_epf_mhi *epf_mhi = epf_get_drvdata(epf);
|
||||
const struct pci_epf_mhi_ep_info *info = epf_mhi->info;
|
||||
struct pci_epf_bar *epf_bar = &epf->bar[info->bar_num];
|
||||
struct mhi_ep_cntrl *mhi_cntrl = &epf_mhi->mhi_cntrl;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
|
||||
/*
|
||||
* Forcefully power down the MHI EP stack. Only way to bring the MHI EP
|
||||
* stack back to working state after successive bind is by getting BME
|
||||
* from host.
|
||||
*/
|
||||
if (mhi_cntrl->mhi_dev) {
|
||||
mhi_ep_power_down(mhi_cntrl);
|
||||
mhi_ep_unregister_controller(mhi_cntrl);
|
||||
}
|
||||
|
||||
iounmap(epf_mhi->mmio);
|
||||
pci_epc_clear_bar(epc, epf->func_no, epf->vfunc_no, epf_bar);
|
||||
}
|
||||
|
||||
static struct pci_epc_event_ops pci_epf_mhi_event_ops = {
|
||||
.core_init = pci_epf_mhi_core_init,
|
||||
.link_up = pci_epf_mhi_link_up,
|
||||
.link_down = pci_epf_mhi_link_down,
|
||||
.bme = pci_epf_mhi_bme,
|
||||
};
|
||||
|
||||
static int pci_epf_mhi_probe(struct pci_epf *epf,
|
||||
const struct pci_epf_device_id *id)
|
||||
{
|
||||
struct pci_epf_mhi_ep_info *info =
|
||||
(struct pci_epf_mhi_ep_info *)id->driver_data;
|
||||
struct pci_epf_mhi *epf_mhi;
|
||||
struct device *dev = &epf->dev;
|
||||
|
||||
epf_mhi = devm_kzalloc(dev, sizeof(*epf_mhi), GFP_KERNEL);
|
||||
if (!epf_mhi)
|
||||
return -ENOMEM;
|
||||
|
||||
epf->header = info->epf_header;
|
||||
epf_mhi->info = info;
|
||||
epf_mhi->epf = epf;
|
||||
|
||||
epf->event_ops = &pci_epf_mhi_event_ops;
|
||||
|
||||
mutex_init(&epf_mhi->lock);
|
||||
|
||||
epf_set_drvdata(epf, epf_mhi);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static const struct pci_epf_device_id pci_epf_mhi_ids[] = {
|
||||
{
|
||||
.name = "sdx55", .driver_data = (kernel_ulong_t)&sdx55_info,
|
||||
},
|
||||
{},
|
||||
};
|
||||
|
||||
static struct pci_epf_ops pci_epf_mhi_ops = {
|
||||
.unbind = pci_epf_mhi_unbind,
|
||||
.bind = pci_epf_mhi_bind,
|
||||
};
|
||||
|
||||
static struct pci_epf_driver pci_epf_mhi_driver = {
|
||||
.driver.name = "pci_epf_mhi",
|
||||
.probe = pci_epf_mhi_probe,
|
||||
.id_table = pci_epf_mhi_ids,
|
||||
.ops = &pci_epf_mhi_ops,
|
||||
.owner = THIS_MODULE,
|
||||
};
|
||||
|
||||
static int __init pci_epf_mhi_init(void)
|
||||
{
|
||||
return pci_epf_register_driver(&pci_epf_mhi_driver);
|
||||
}
|
||||
module_init(pci_epf_mhi_init);
|
||||
|
||||
static void __exit pci_epf_mhi_exit(void)
|
||||
{
|
||||
pci_epf_unregister_driver(&pci_epf_mhi_driver);
|
||||
}
|
||||
module_exit(pci_epf_mhi_exit);
|
||||
|
||||
MODULE_DESCRIPTION("PCI EPF driver for MHI Endpoint devices");
|
||||
MODULE_AUTHOR("Manivannan Sadhasivam <manivannan.sadhasivam@linaro.org>");
|
||||
MODULE_LICENSE("GPL");
|
@ -2075,11 +2075,13 @@ static struct config_group *epf_ntb_add_cfs(struct pci_epf *epf,
|
||||
/**
|
||||
* epf_ntb_probe() - Probe NTB function driver
|
||||
* @epf: NTB endpoint function device
|
||||
* @id: NTB endpoint function device ID
|
||||
*
|
||||
* Probe NTB function driver when endpoint function bus detects a NTB
|
||||
* endpoint function.
|
||||
*/
|
||||
static int epf_ntb_probe(struct pci_epf *epf)
|
||||
static int epf_ntb_probe(struct pci_epf *epf,
|
||||
const struct pci_epf_device_id *id)
|
||||
{
|
||||
struct epf_ntb *ntb;
|
||||
struct device *dev;
|
||||
|
@ -54,6 +54,9 @@ struct pci_epf_test {
|
||||
struct delayed_work cmd_handler;
|
||||
struct dma_chan *dma_chan_tx;
|
||||
struct dma_chan *dma_chan_rx;
|
||||
struct dma_chan *transfer_chan;
|
||||
dma_cookie_t transfer_cookie;
|
||||
enum dma_status transfer_status;
|
||||
struct completion transfer_complete;
|
||||
bool dma_supported;
|
||||
bool dma_private;
|
||||
@ -85,8 +88,14 @@ static size_t bar_size[] = { 512, 512, 1024, 16384, 131072, 1048576 };
|
||||
static void pci_epf_test_dma_callback(void *param)
|
||||
{
|
||||
struct pci_epf_test *epf_test = param;
|
||||
struct dma_tx_state state;
|
||||
|
||||
complete(&epf_test->transfer_complete);
|
||||
epf_test->transfer_status =
|
||||
dmaengine_tx_status(epf_test->transfer_chan,
|
||||
epf_test->transfer_cookie, &state);
|
||||
if (epf_test->transfer_status == DMA_COMPLETE ||
|
||||
epf_test->transfer_status == DMA_ERROR)
|
||||
complete(&epf_test->transfer_complete);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -112,7 +121,7 @@ static int pci_epf_test_data_transfer(struct pci_epf_test *epf_test,
|
||||
size_t len, dma_addr_t dma_remote,
|
||||
enum dma_transfer_direction dir)
|
||||
{
|
||||
struct dma_chan *chan = (dir == DMA_DEV_TO_MEM) ?
|
||||
struct dma_chan *chan = (dir == DMA_MEM_TO_DEV) ?
|
||||
epf_test->dma_chan_tx : epf_test->dma_chan_rx;
|
||||
dma_addr_t dma_local = (dir == DMA_MEM_TO_DEV) ? dma_src : dma_dst;
|
||||
enum dma_ctrl_flags flags = DMA_CTRL_ACK | DMA_PREP_INTERRUPT;
|
||||
@ -120,7 +129,6 @@ static int pci_epf_test_data_transfer(struct pci_epf_test *epf_test,
|
||||
struct dma_async_tx_descriptor *tx;
|
||||
struct dma_slave_config sconf = {};
|
||||
struct device *dev = &epf->dev;
|
||||
dma_cookie_t cookie;
|
||||
int ret;
|
||||
|
||||
if (IS_ERR_OR_NULL(chan)) {
|
||||
@ -151,26 +159,34 @@ static int pci_epf_test_data_transfer(struct pci_epf_test *epf_test,
|
||||
return -EIO;
|
||||
}
|
||||
|
||||
reinit_completion(&epf_test->transfer_complete);
|
||||
epf_test->transfer_chan = chan;
|
||||
tx->callback = pci_epf_test_dma_callback;
|
||||
tx->callback_param = epf_test;
|
||||
cookie = tx->tx_submit(tx);
|
||||
reinit_completion(&epf_test->transfer_complete);
|
||||
epf_test->transfer_cookie = dmaengine_submit(tx);
|
||||
|
||||
ret = dma_submit_error(cookie);
|
||||
ret = dma_submit_error(epf_test->transfer_cookie);
|
||||
if (ret) {
|
||||
dev_err(dev, "Failed to do DMA tx_submit %d\n", cookie);
|
||||
return -EIO;
|
||||
dev_err(dev, "Failed to do DMA tx_submit %d\n", ret);
|
||||
goto terminate;
|
||||
}
|
||||
|
||||
dma_async_issue_pending(chan);
|
||||
ret = wait_for_completion_interruptible(&epf_test->transfer_complete);
|
||||
if (ret < 0) {
|
||||
dmaengine_terminate_sync(chan);
|
||||
dev_err(dev, "DMA wait_for_completion_timeout\n");
|
||||
return -ETIMEDOUT;
|
||||
dev_err(dev, "DMA wait_for_completion interrupted\n");
|
||||
goto terminate;
|
||||
}
|
||||
|
||||
return 0;
|
||||
if (epf_test->transfer_status == DMA_ERROR) {
|
||||
dev_err(dev, "DMA transfer failed\n");
|
||||
ret = -EIO;
|
||||
}
|
||||
|
||||
terminate:
|
||||
dmaengine_terminate_sync(chan);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
struct epf_dma_filter {
|
||||
@ -279,40 +295,29 @@ static void pci_epf_test_clean_dma_chan(struct pci_epf_test *epf_test)
|
||||
return;
|
||||
}
|
||||
|
||||
static void pci_epf_test_print_rate(const char *ops, u64 size,
|
||||
static void pci_epf_test_print_rate(struct pci_epf_test *epf_test,
|
||||
const char *op, u64 size,
|
||||
struct timespec64 *start,
|
||||
struct timespec64 *end, bool dma)
|
||||
{
|
||||
struct timespec64 ts;
|
||||
u64 rate, ns;
|
||||
|
||||
ts = timespec64_sub(*end, *start);
|
||||
|
||||
/* convert both size (stored in 'rate') and time in terms of 'ns' */
|
||||
ns = timespec64_to_ns(&ts);
|
||||
rate = size * NSEC_PER_SEC;
|
||||
|
||||
/* Divide both size (stored in 'rate') and ns by a common factor */
|
||||
while (ns > UINT_MAX) {
|
||||
rate >>= 1;
|
||||
ns >>= 1;
|
||||
}
|
||||
|
||||
if (!ns)
|
||||
return;
|
||||
struct timespec64 ts = timespec64_sub(*end, *start);
|
||||
u64 rate = 0, ns;
|
||||
|
||||
/* calculate the rate */
|
||||
do_div(rate, (uint32_t)ns);
|
||||
ns = timespec64_to_ns(&ts);
|
||||
if (ns)
|
||||
rate = div64_u64(size * NSEC_PER_SEC, ns * 1000);
|
||||
|
||||
pr_info("\n%s => Size: %llu bytes\t DMA: %s\t Time: %llu.%09u seconds\t"
|
||||
"Rate: %llu KB/s\n", ops, size, dma ? "YES" : "NO",
|
||||
(u64)ts.tv_sec, (u32)ts.tv_nsec, rate / 1024);
|
||||
dev_info(&epf_test->epf->dev,
|
||||
"%s => Size: %llu B, DMA: %s, Time: %llu.%09u s, Rate: %llu KB/s\n",
|
||||
op, size, dma ? "YES" : "NO",
|
||||
(u64)ts.tv_sec, (u32)ts.tv_nsec, rate);
|
||||
}
|
||||
|
||||
static int pci_epf_test_copy(struct pci_epf_test *epf_test)
|
||||
static void pci_epf_test_copy(struct pci_epf_test *epf_test,
|
||||
struct pci_epf_test_reg *reg)
|
||||
{
|
||||
int ret;
|
||||
bool use_dma;
|
||||
void __iomem *src_addr;
|
||||
void __iomem *dst_addr;
|
||||
phys_addr_t src_phys_addr;
|
||||
@ -321,8 +326,6 @@ static int pci_epf_test_copy(struct pci_epf_test *epf_test)
|
||||
struct pci_epf *epf = epf_test->epf;
|
||||
struct device *dev = &epf->dev;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
enum pci_barno test_reg_bar = epf_test->test_reg_bar;
|
||||
struct pci_epf_test_reg *reg = epf_test->reg[test_reg_bar];
|
||||
|
||||
src_addr = pci_epc_mem_alloc_addr(epc, &src_phys_addr, reg->size);
|
||||
if (!src_addr) {
|
||||
@ -357,14 +360,7 @@ static int pci_epf_test_copy(struct pci_epf_test *epf_test)
|
||||
}
|
||||
|
||||
ktime_get_ts64(&start);
|
||||
use_dma = !!(reg->flags & FLAG_USE_DMA);
|
||||
if (use_dma) {
|
||||
if (!epf_test->dma_supported) {
|
||||
dev_err(dev, "Cannot transfer data using DMA\n");
|
||||
ret = -EINVAL;
|
||||
goto err_map_addr;
|
||||
}
|
||||
|
||||
if (reg->flags & FLAG_USE_DMA) {
|
||||
if (epf_test->dma_private) {
|
||||
dev_err(dev, "Cannot transfer data using DMA\n");
|
||||
ret = -EINVAL;
|
||||
@ -390,7 +386,8 @@ static int pci_epf_test_copy(struct pci_epf_test *epf_test)
|
||||
kfree(buf);
|
||||
}
|
||||
ktime_get_ts64(&end);
|
||||
pci_epf_test_print_rate("COPY", reg->size, &start, &end, use_dma);
|
||||
pci_epf_test_print_rate(epf_test, "COPY", reg->size, &start, &end,
|
||||
reg->flags & FLAG_USE_DMA);
|
||||
|
||||
err_map_addr:
|
||||
pci_epc_unmap_addr(epc, epf->func_no, epf->vfunc_no, dst_phys_addr);
|
||||
@ -405,16 +402,19 @@ err_src_addr:
|
||||
pci_epc_mem_free_addr(epc, src_phys_addr, src_addr, reg->size);
|
||||
|
||||
err:
|
||||
return ret;
|
||||
if (!ret)
|
||||
reg->status |= STATUS_COPY_SUCCESS;
|
||||
else
|
||||
reg->status |= STATUS_COPY_FAIL;
|
||||
}
|
||||
|
||||
static int pci_epf_test_read(struct pci_epf_test *epf_test)
|
||||
static void pci_epf_test_read(struct pci_epf_test *epf_test,
|
||||
struct pci_epf_test_reg *reg)
|
||||
{
|
||||
int ret;
|
||||
void __iomem *src_addr;
|
||||
void *buf;
|
||||
u32 crc32;
|
||||
bool use_dma;
|
||||
phys_addr_t phys_addr;
|
||||
phys_addr_t dst_phys_addr;
|
||||
struct timespec64 start, end;
|
||||
@ -422,8 +422,6 @@ static int pci_epf_test_read(struct pci_epf_test *epf_test)
|
||||
struct device *dev = &epf->dev;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
struct device *dma_dev = epf->epc->dev.parent;
|
||||
enum pci_barno test_reg_bar = epf_test->test_reg_bar;
|
||||
struct pci_epf_test_reg *reg = epf_test->reg[test_reg_bar];
|
||||
|
||||
src_addr = pci_epc_mem_alloc_addr(epc, &phys_addr, reg->size);
|
||||
if (!src_addr) {
|
||||
@ -447,14 +445,7 @@ static int pci_epf_test_read(struct pci_epf_test *epf_test)
|
||||
goto err_map_addr;
|
||||
}
|
||||
|
||||
use_dma = !!(reg->flags & FLAG_USE_DMA);
|
||||
if (use_dma) {
|
||||
if (!epf_test->dma_supported) {
|
||||
dev_err(dev, "Cannot transfer data using DMA\n");
|
||||
ret = -EINVAL;
|
||||
goto err_dma_map;
|
||||
}
|
||||
|
||||
if (reg->flags & FLAG_USE_DMA) {
|
||||
dst_phys_addr = dma_map_single(dma_dev, buf, reg->size,
|
||||
DMA_FROM_DEVICE);
|
||||
if (dma_mapping_error(dma_dev, dst_phys_addr)) {
|
||||
@ -479,7 +470,8 @@ static int pci_epf_test_read(struct pci_epf_test *epf_test)
|
||||
ktime_get_ts64(&end);
|
||||
}
|
||||
|
||||
pci_epf_test_print_rate("READ", reg->size, &start, &end, use_dma);
|
||||
pci_epf_test_print_rate(epf_test, "READ", reg->size, &start, &end,
|
||||
reg->flags & FLAG_USE_DMA);
|
||||
|
||||
crc32 = crc32_le(~0, buf, reg->size);
|
||||
if (crc32 != reg->checksum)
|
||||
@ -495,15 +487,18 @@ err_addr:
|
||||
pci_epc_mem_free_addr(epc, phys_addr, src_addr, reg->size);
|
||||
|
||||
err:
|
||||
return ret;
|
||||
if (!ret)
|
||||
reg->status |= STATUS_READ_SUCCESS;
|
||||
else
|
||||
reg->status |= STATUS_READ_FAIL;
|
||||
}
|
||||
|
||||
static int pci_epf_test_write(struct pci_epf_test *epf_test)
|
||||
static void pci_epf_test_write(struct pci_epf_test *epf_test,
|
||||
struct pci_epf_test_reg *reg)
|
||||
{
|
||||
int ret;
|
||||
void __iomem *dst_addr;
|
||||
void *buf;
|
||||
bool use_dma;
|
||||
phys_addr_t phys_addr;
|
||||
phys_addr_t src_phys_addr;
|
||||
struct timespec64 start, end;
|
||||
@ -511,8 +506,6 @@ static int pci_epf_test_write(struct pci_epf_test *epf_test)
|
||||
struct device *dev = &epf->dev;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
struct device *dma_dev = epf->epc->dev.parent;
|
||||
enum pci_barno test_reg_bar = epf_test->test_reg_bar;
|
||||
struct pci_epf_test_reg *reg = epf_test->reg[test_reg_bar];
|
||||
|
||||
dst_addr = pci_epc_mem_alloc_addr(epc, &phys_addr, reg->size);
|
||||
if (!dst_addr) {
|
||||
@ -539,14 +532,7 @@ static int pci_epf_test_write(struct pci_epf_test *epf_test)
|
||||
get_random_bytes(buf, reg->size);
|
||||
reg->checksum = crc32_le(~0, buf, reg->size);
|
||||
|
||||
use_dma = !!(reg->flags & FLAG_USE_DMA);
|
||||
if (use_dma) {
|
||||
if (!epf_test->dma_supported) {
|
||||
dev_err(dev, "Cannot transfer data using DMA\n");
|
||||
ret = -EINVAL;
|
||||
goto err_dma_map;
|
||||
}
|
||||
|
||||
if (reg->flags & FLAG_USE_DMA) {
|
||||
src_phys_addr = dma_map_single(dma_dev, buf, reg->size,
|
||||
DMA_TO_DEVICE);
|
||||
if (dma_mapping_error(dma_dev, src_phys_addr)) {
|
||||
@ -573,7 +559,8 @@ static int pci_epf_test_write(struct pci_epf_test *epf_test)
|
||||
ktime_get_ts64(&end);
|
||||
}
|
||||
|
||||
pci_epf_test_print_rate("WRITE", reg->size, &start, &end, use_dma);
|
||||
pci_epf_test_print_rate(epf_test, "WRITE", reg->size, &start, &end,
|
||||
reg->flags & FLAG_USE_DMA);
|
||||
|
||||
/*
|
||||
* wait 1ms inorder for the write to complete. Without this delay L3
|
||||
@ -591,32 +578,51 @@ err_addr:
|
||||
pci_epc_mem_free_addr(epc, phys_addr, dst_addr, reg->size);
|
||||
|
||||
err:
|
||||
return ret;
|
||||
if (!ret)
|
||||
reg->status |= STATUS_WRITE_SUCCESS;
|
||||
else
|
||||
reg->status |= STATUS_WRITE_FAIL;
|
||||
}
|
||||
|
||||
static void pci_epf_test_raise_irq(struct pci_epf_test *epf_test, u8 irq_type,
|
||||
u16 irq)
|
||||
static void pci_epf_test_raise_irq(struct pci_epf_test *epf_test,
|
||||
struct pci_epf_test_reg *reg)
|
||||
{
|
||||
struct pci_epf *epf = epf_test->epf;
|
||||
struct device *dev = &epf->dev;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
enum pci_barno test_reg_bar = epf_test->test_reg_bar;
|
||||
struct pci_epf_test_reg *reg = epf_test->reg[test_reg_bar];
|
||||
u32 status = reg->status | STATUS_IRQ_RAISED;
|
||||
int count;
|
||||
|
||||
reg->status |= STATUS_IRQ_RAISED;
|
||||
/*
|
||||
* Set the status before raising the IRQ to ensure that the host sees
|
||||
* the updated value when it gets the IRQ.
|
||||
*/
|
||||
WRITE_ONCE(reg->status, status);
|
||||
|
||||
switch (irq_type) {
|
||||
switch (reg->irq_type) {
|
||||
case IRQ_TYPE_LEGACY:
|
||||
pci_epc_raise_irq(epc, epf->func_no, epf->vfunc_no,
|
||||
PCI_EPC_IRQ_LEGACY, 0);
|
||||
break;
|
||||
case IRQ_TYPE_MSI:
|
||||
count = pci_epc_get_msi(epc, epf->func_no, epf->vfunc_no);
|
||||
if (reg->irq_number > count || count <= 0) {
|
||||
dev_err(dev, "Invalid MSI IRQ number %d / %d\n",
|
||||
reg->irq_number, count);
|
||||
return;
|
||||
}
|
||||
pci_epc_raise_irq(epc, epf->func_no, epf->vfunc_no,
|
||||
PCI_EPC_IRQ_MSI, irq);
|
||||
PCI_EPC_IRQ_MSI, reg->irq_number);
|
||||
break;
|
||||
case IRQ_TYPE_MSIX:
|
||||
count = pci_epc_get_msix(epc, epf->func_no, epf->vfunc_no);
|
||||
if (reg->irq_number > count || count <= 0) {
|
||||
dev_err(dev, "Invalid MSIX IRQ number %d / %d\n",
|
||||
reg->irq_number, count);
|
||||
return;
|
||||
}
|
||||
pci_epc_raise_irq(epc, epf->func_no, epf->vfunc_no,
|
||||
PCI_EPC_IRQ_MSIX, irq);
|
||||
PCI_EPC_IRQ_MSIX, reg->irq_number);
|
||||
break;
|
||||
default:
|
||||
dev_err(dev, "Failed to raise IRQ, unknown type\n");
|
||||
@ -626,87 +632,53 @@ static void pci_epf_test_raise_irq(struct pci_epf_test *epf_test, u8 irq_type,
|
||||
|
||||
static void pci_epf_test_cmd_handler(struct work_struct *work)
|
||||
{
|
||||
int ret;
|
||||
int count;
|
||||
u32 command;
|
||||
struct pci_epf_test *epf_test = container_of(work, struct pci_epf_test,
|
||||
cmd_handler.work);
|
||||
struct pci_epf *epf = epf_test->epf;
|
||||
struct device *dev = &epf->dev;
|
||||
struct pci_epc *epc = epf->epc;
|
||||
enum pci_barno test_reg_bar = epf_test->test_reg_bar;
|
||||
struct pci_epf_test_reg *reg = epf_test->reg[test_reg_bar];
|
||||
|
||||
command = reg->command;
|
||||
command = READ_ONCE(reg->command);
|
||||
if (!command)
|
||||
goto reset_handler;
|
||||
|
||||
reg->command = 0;
|
||||
reg->status = 0;
|
||||
WRITE_ONCE(reg->command, 0);
|
||||
WRITE_ONCE(reg->status, 0);
|
||||
|
||||
if ((READ_ONCE(reg->flags) & FLAG_USE_DMA) &&
|
||||
!epf_test->dma_supported) {
|
||||
dev_err(dev, "Cannot transfer data using DMA\n");
|
||||
goto reset_handler;
|
||||
}
|
||||
|
||||
if (reg->irq_type > IRQ_TYPE_MSIX) {
|
||||
dev_err(dev, "Failed to detect IRQ type\n");
|
||||
goto reset_handler;
|
||||
}
|
||||
|
||||
if (command & COMMAND_RAISE_LEGACY_IRQ) {
|
||||
reg->status = STATUS_IRQ_RAISED;
|
||||
pci_epc_raise_irq(epc, epf->func_no, epf->vfunc_no,
|
||||
PCI_EPC_IRQ_LEGACY, 0);
|
||||
goto reset_handler;
|
||||
}
|
||||
|
||||
if (command & COMMAND_WRITE) {
|
||||
ret = pci_epf_test_write(epf_test);
|
||||
if (ret)
|
||||
reg->status |= STATUS_WRITE_FAIL;
|
||||
else
|
||||
reg->status |= STATUS_WRITE_SUCCESS;
|
||||
pci_epf_test_raise_irq(epf_test, reg->irq_type,
|
||||
reg->irq_number);
|
||||
goto reset_handler;
|
||||
}
|
||||
|
||||
if (command & COMMAND_READ) {
|
||||
ret = pci_epf_test_read(epf_test);
|
||||
if (!ret)
|
||||
reg->status |= STATUS_READ_SUCCESS;
|
||||
else
|
||||
reg->status |= STATUS_READ_FAIL;
|
||||
pci_epf_test_raise_irq(epf_test, reg->irq_type,
|
||||
reg->irq_number);
|
||||
goto reset_handler;
|
||||
}
|
||||
|
||||
if (command & COMMAND_COPY) {
|
||||
ret = pci_epf_test_copy(epf_test);
|
||||
if (!ret)
|
||||
reg->status |= STATUS_COPY_SUCCESS;
|
||||
else
|
||||
reg->status |= STATUS_COPY_FAIL;
|
||||
pci_epf_test_raise_irq(epf_test, reg->irq_type,
|
||||
reg->irq_number);
|
||||
goto reset_handler;
|
||||
}
|
||||
|
||||
if (command & COMMAND_RAISE_MSI_IRQ) {
|
||||
count = pci_epc_get_msi(epc, epf->func_no, epf->vfunc_no);
|
||||
if (reg->irq_number > count || count <= 0)
|
||||
goto reset_handler;
|
||||
reg->status = STATUS_IRQ_RAISED;
|
||||
pci_epc_raise_irq(epc, epf->func_no, epf->vfunc_no,
|
||||
PCI_EPC_IRQ_MSI, reg->irq_number);
|
||||
goto reset_handler;
|
||||
}
|
||||
|
||||
if (command & COMMAND_RAISE_MSIX_IRQ) {
|
||||
count = pci_epc_get_msix(epc, epf->func_no, epf->vfunc_no);
|
||||
if (reg->irq_number > count || count <= 0)
|
||||
goto reset_handler;
|
||||
reg->status = STATUS_IRQ_RAISED;
|
||||
pci_epc_raise_irq(epc, epf->func_no, epf->vfunc_no,
|
||||
PCI_EPC_IRQ_MSIX, reg->irq_number);
|
||||
goto reset_handler;
|
||||
switch (command) {
|
||||
case COMMAND_RAISE_LEGACY_IRQ:
|
||||
case COMMAND_RAISE_MSI_IRQ:
|
||||
case COMMAND_RAISE_MSIX_IRQ:
|
||||
pci_epf_test_raise_irq(epf_test, reg);
|
||||
break;
|
||||
case COMMAND_WRITE:
|
||||
pci_epf_test_write(epf_test, reg);
|
||||
pci_epf_test_raise_irq(epf_test, reg);
|
||||
break;
|
||||
case COMMAND_READ:
|
||||
pci_epf_test_read(epf_test, reg);
|
||||
pci_epf_test_raise_irq(epf_test, reg);
|
||||
break;
|
||||
case COMMAND_COPY:
|
||||
pci_epf_test_copy(epf_test, reg);
|
||||
pci_epf_test_raise_irq(epf_test, reg);
|
||||
break;
|
||||
default:
|
||||
dev_err(dev, "Invalid command 0x%x\n", command);
|
||||
break;
|
||||
}
|
||||
|
||||
reset_handler:
|
||||
@ -980,7 +952,8 @@ static const struct pci_epf_device_id pci_epf_test_ids[] = {
|
||||
{},
|
||||
};
|
||||
|
||||
static int pci_epf_test_probe(struct pci_epf *epf)
|
||||
static int pci_epf_test_probe(struct pci_epf *epf,
|
||||
const struct pci_epf_device_id *id)
|
||||
{
|
||||
struct pci_epf_test *epf_test;
|
||||
struct device *dev = &epf->dev;
|
||||
|
@ -84,15 +84,15 @@ enum epf_ntb_bar {
|
||||
* | |
|
||||
* | |
|
||||
* | |
|
||||
* +-----------------------+--------------------------+ Base+span_offset
|
||||
* +-----------------------+--------------------------+ Base+spad_offset
|
||||
* | | |
|
||||
* | Peer Span Space | Span Space |
|
||||
* | Peer Spad Space | Spad Space |
|
||||
* | | |
|
||||
* | | |
|
||||
* +-----------------------+--------------------------+ Base+span_offset
|
||||
* | | | +span_count * 4
|
||||
* +-----------------------+--------------------------+ Base+spad_offset
|
||||
* | | | +spad_count * 4
|
||||
* | | |
|
||||
* | Span Space | Peer Span Space |
|
||||
* | Spad Space | Peer Spad Space |
|
||||
* | | |
|
||||
* +-----------------------+--------------------------+
|
||||
* Virtual PCI PCIe Endpoint
|
||||
@ -1395,13 +1395,15 @@ static struct pci_epf_ops epf_ntb_ops = {
|
||||
/**
|
||||
* epf_ntb_probe() - Probe NTB function driver
|
||||
* @epf: NTB endpoint function device
|
||||
* @id: NTB endpoint function device ID
|
||||
*
|
||||
* Probe NTB function driver when endpoint function bus detects a NTB
|
||||
* endpoint function.
|
||||
*
|
||||
* Returns: Zero for success, or an error code in case of failure
|
||||
*/
|
||||
static int epf_ntb_probe(struct pci_epf *epf)
|
||||
static int epf_ntb_probe(struct pci_epf *epf,
|
||||
const struct pci_epf_device_id *id)
|
||||
{
|
||||
struct epf_ntb *ntb;
|
||||
struct device *dev;
|
||||
|
@ -23,6 +23,7 @@ struct pci_epf_group {
|
||||
struct config_group group;
|
||||
struct config_group primary_epc_group;
|
||||
struct config_group secondary_epc_group;
|
||||
struct config_group *type_group;
|
||||
struct delayed_work cfs_work;
|
||||
struct pci_epf *epf;
|
||||
int index;
|
||||
@ -178,6 +179,9 @@ static ssize_t pci_epc_start_store(struct config_item *item, const char *page,
|
||||
if (kstrtobool(page, &start) < 0)
|
||||
return -EINVAL;
|
||||
|
||||
if (start == epc_group->start)
|
||||
return -EALREADY;
|
||||
|
||||
if (!start) {
|
||||
pci_epc_stop(epc);
|
||||
epc_group->start = 0;
|
||||
@ -502,34 +506,65 @@ static struct configfs_item_operations pci_epf_ops = {
|
||||
.release = pci_epf_release,
|
||||
};
|
||||
|
||||
static struct config_group *pci_epf_type_make(struct config_group *group,
|
||||
const char *name)
|
||||
{
|
||||
struct pci_epf_group *epf_group = to_pci_epf_group(&group->cg_item);
|
||||
struct config_group *epf_type_group;
|
||||
|
||||
epf_type_group = pci_epf_type_add_cfs(epf_group->epf, group);
|
||||
return epf_type_group;
|
||||
}
|
||||
|
||||
static void pci_epf_type_drop(struct config_group *group,
|
||||
struct config_item *item)
|
||||
{
|
||||
config_item_put(item);
|
||||
}
|
||||
|
||||
static struct configfs_group_operations pci_epf_type_group_ops = {
|
||||
.make_group = &pci_epf_type_make,
|
||||
.drop_item = &pci_epf_type_drop,
|
||||
};
|
||||
|
||||
static const struct config_item_type pci_epf_type = {
|
||||
.ct_group_ops = &pci_epf_type_group_ops,
|
||||
.ct_item_ops = &pci_epf_ops,
|
||||
.ct_attrs = pci_epf_attrs,
|
||||
.ct_owner = THIS_MODULE,
|
||||
};
|
||||
|
||||
/**
|
||||
* pci_epf_type_add_cfs() - Help function drivers to expose function specific
|
||||
* attributes in configfs
|
||||
* @epf: the EPF device that has to be configured using configfs
|
||||
* @group: the parent configfs group (corresponding to entries in
|
||||
* pci_epf_device_id)
|
||||
*
|
||||
* Invoke to expose function specific attributes in configfs.
|
||||
*
|
||||
* Return: A pointer to a config_group structure or NULL if the function driver
|
||||
* does not have anything to expose (attributes configured by user) or if
|
||||
* the function driver does not implement the add_cfs() method.
|
||||
*
|
||||
* Returns an error pointer if this function is called for an unbound EPF device
|
||||
* or if the EPF driver add_cfs() method fails.
|
||||
*/
|
||||
static struct config_group *pci_epf_type_add_cfs(struct pci_epf *epf,
|
||||
struct config_group *group)
|
||||
{
|
||||
struct config_group *epf_type_group;
|
||||
|
||||
if (!epf->driver) {
|
||||
dev_err(&epf->dev, "epf device not bound to driver\n");
|
||||
return ERR_PTR(-ENODEV);
|
||||
}
|
||||
|
||||
if (!epf->driver->ops->add_cfs)
|
||||
return NULL;
|
||||
|
||||
mutex_lock(&epf->lock);
|
||||
epf_type_group = epf->driver->ops->add_cfs(epf, group);
|
||||
mutex_unlock(&epf->lock);
|
||||
|
||||
return epf_type_group;
|
||||
}
|
||||
|
||||
static void pci_ep_cfs_add_type_group(struct pci_epf_group *epf_group)
|
||||
{
|
||||
struct config_group *group;
|
||||
|
||||
group = pci_epf_type_add_cfs(epf_group->epf, &epf_group->group);
|
||||
if (!group)
|
||||
return;
|
||||
|
||||
if (IS_ERR(group)) {
|
||||
dev_err(&epf_group->epf->dev,
|
||||
"failed to create epf type specific attributes\n");
|
||||
return;
|
||||
}
|
||||
|
||||
configfs_register_group(&epf_group->group, group);
|
||||
}
|
||||
|
||||
static void pci_epf_cfs_work(struct work_struct *work)
|
||||
{
|
||||
struct pci_epf_group *epf_group;
|
||||
@ -547,6 +582,8 @@ static void pci_epf_cfs_work(struct work_struct *work)
|
||||
pr_err("failed to create 'secondary' EPC interface\n");
|
||||
return;
|
||||
}
|
||||
|
||||
pci_ep_cfs_add_type_group(epf_group);
|
||||
}
|
||||
|
||||
static struct config_group *pci_epf_make(struct config_group *group,
|
||||
|
@ -213,7 +213,7 @@ EXPORT_SYMBOL_GPL(pci_epc_start);
|
||||
* @func_no: the physical endpoint function number in the EPC device
|
||||
* @vfunc_no: the virtual endpoint function number in the physical function
|
||||
* @type: specify the type of interrupt; legacy, MSI or MSI-X
|
||||
* @interrupt_num: the MSI or MSI-X interrupt number
|
||||
* @interrupt_num: the MSI or MSI-X interrupt number with range (1-N)
|
||||
*
|
||||
* Invoke to raise an legacy, MSI or MSI-X interrupt
|
||||
*/
|
||||
@ -246,7 +246,7 @@ EXPORT_SYMBOL_GPL(pci_epc_raise_irq);
|
||||
* @func_no: the physical endpoint function number in the EPC device
|
||||
* @vfunc_no: the virtual endpoint function number in the physical function
|
||||
* @phys_addr: the physical address of the outbound region
|
||||
* @interrupt_num: the MSI interrupt number
|
||||
* @interrupt_num: the MSI interrupt number with range (1-N)
|
||||
* @entry_size: Size of Outbound address region for each interrupt
|
||||
* @msi_data: the data that should be written in order to raise MSI interrupt
|
||||
* with interrupt number as 'interrupt num'
|
||||
@ -706,6 +706,32 @@ void pci_epc_linkup(struct pci_epc *epc)
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(pci_epc_linkup);
|
||||
|
||||
/**
|
||||
* pci_epc_linkdown() - Notify the EPF device that EPC device has dropped the
|
||||
* connection with the Root Complex.
|
||||
* @epc: the EPC device which has dropped the link with the host
|
||||
*
|
||||
* Invoke to Notify the EPF device that the EPC device has dropped the
|
||||
* connection with the Root Complex.
|
||||
*/
|
||||
void pci_epc_linkdown(struct pci_epc *epc)
|
||||
{
|
||||
struct pci_epf *epf;
|
||||
|
||||
if (!epc || IS_ERR(epc))
|
||||
return;
|
||||
|
||||
mutex_lock(&epc->list_lock);
|
||||
list_for_each_entry(epf, &epc->pci_epf, list) {
|
||||
mutex_lock(&epf->lock);
|
||||
if (epf->event_ops && epf->event_ops->link_down)
|
||||
epf->event_ops->link_down(epf);
|
||||
mutex_unlock(&epf->lock);
|
||||
}
|
||||
mutex_unlock(&epc->list_lock);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(pci_epc_linkdown);
|
||||
|
||||
/**
|
||||
* pci_epc_init_notify() - Notify the EPF device that EPC device's core
|
||||
* initialization is completed.
|
||||
@ -732,6 +758,32 @@ void pci_epc_init_notify(struct pci_epc *epc)
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(pci_epc_init_notify);
|
||||
|
||||
/**
|
||||
* pci_epc_bme_notify() - Notify the EPF device that the EPC device has received
|
||||
* the BME event from the Root complex
|
||||
* @epc: the EPC device that received the BME event
|
||||
*
|
||||
* Invoke to Notify the EPF device that the EPC device has received the Bus
|
||||
* Master Enable (BME) event from the Root complex
|
||||
*/
|
||||
void pci_epc_bme_notify(struct pci_epc *epc)
|
||||
{
|
||||
struct pci_epf *epf;
|
||||
|
||||
if (!epc || IS_ERR(epc))
|
||||
return;
|
||||
|
||||
mutex_lock(&epc->list_lock);
|
||||
list_for_each_entry(epf, &epc->pci_epf, list) {
|
||||
mutex_lock(&epf->lock);
|
||||
if (epf->event_ops && epf->event_ops->bme)
|
||||
epf->event_ops->bme(epf);
|
||||
mutex_unlock(&epf->lock);
|
||||
}
|
||||
mutex_unlock(&epc->list_lock);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(pci_epc_bme_notify);
|
||||
|
||||
/**
|
||||
* pci_epc_destroy() - destroy the EPC device
|
||||
* @epc: the EPC device that has to be destroyed
|
||||
|
@ -20,38 +20,6 @@ static DEFINE_MUTEX(pci_epf_mutex);
|
||||
static struct bus_type pci_epf_bus_type;
|
||||
static const struct device_type pci_epf_type;
|
||||
|
||||
/**
|
||||
* pci_epf_type_add_cfs() - Help function drivers to expose function specific
|
||||
* attributes in configfs
|
||||
* @epf: the EPF device that has to be configured using configfs
|
||||
* @group: the parent configfs group (corresponding to entries in
|
||||
* pci_epf_device_id)
|
||||
*
|
||||
* Invoke to expose function specific attributes in configfs. If the function
|
||||
* driver does not have anything to expose (attributes configured by user),
|
||||
* return NULL.
|
||||
*/
|
||||
struct config_group *pci_epf_type_add_cfs(struct pci_epf *epf,
|
||||
struct config_group *group)
|
||||
{
|
||||
struct config_group *epf_type_group;
|
||||
|
||||
if (!epf->driver) {
|
||||
dev_err(&epf->dev, "epf device not bound to driver\n");
|
||||
return NULL;
|
||||
}
|
||||
|
||||
if (!epf->driver->ops->add_cfs)
|
||||
return NULL;
|
||||
|
||||
mutex_lock(&epf->lock);
|
||||
epf_type_group = epf->driver->ops->add_cfs(epf, group);
|
||||
mutex_unlock(&epf->lock);
|
||||
|
||||
return epf_type_group;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(pci_epf_type_add_cfs);
|
||||
|
||||
/**
|
||||
* pci_epf_unbind() - Notify the function driver that the binding between the
|
||||
* EPF device and EPC device has been lost
|
||||
@ -493,16 +461,16 @@ static const struct device_type pci_epf_type = {
|
||||
.release = pci_epf_dev_release,
|
||||
};
|
||||
|
||||
static int
|
||||
static const struct pci_epf_device_id *
|
||||
pci_epf_match_id(const struct pci_epf_device_id *id, const struct pci_epf *epf)
|
||||
{
|
||||
while (id->name[0]) {
|
||||
if (strcmp(epf->name, id->name) == 0)
|
||||
return true;
|
||||
return id;
|
||||
id++;
|
||||
}
|
||||
|
||||
return false;
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static int pci_epf_device_match(struct device *dev, struct device_driver *drv)
|
||||
@ -511,7 +479,7 @@ static int pci_epf_device_match(struct device *dev, struct device_driver *drv)
|
||||
struct pci_epf_driver *driver = to_pci_epf_driver(drv);
|
||||
|
||||
if (driver->id_table)
|
||||
return pci_epf_match_id(driver->id_table, epf);
|
||||
return !!pci_epf_match_id(driver->id_table, epf);
|
||||
|
||||
return !strcmp(epf->name, drv->name);
|
||||
}
|
||||
@ -526,7 +494,7 @@ static int pci_epf_device_probe(struct device *dev)
|
||||
|
||||
epf->driver = driver;
|
||||
|
||||
return driver->probe(epf);
|
||||
return driver->probe(epf, pci_epf_match_id(driver->id_table, epf));
|
||||
}
|
||||
|
||||
static void pci_epf_device_remove(struct device *dev)
|
||||
|
@ -203,7 +203,9 @@ void pci_epc_destroy(struct pci_epc *epc);
|
||||
int pci_epc_add_epf(struct pci_epc *epc, struct pci_epf *epf,
|
||||
enum pci_epc_interface_type type);
|
||||
void pci_epc_linkup(struct pci_epc *epc);
|
||||
void pci_epc_linkdown(struct pci_epc *epc);
|
||||
void pci_epc_init_notify(struct pci_epc *epc);
|
||||
void pci_epc_bme_notify(struct pci_epc *epc);
|
||||
void pci_epc_remove_epf(struct pci_epc *epc, struct pci_epf *epf,
|
||||
enum pci_epc_interface_type type);
|
||||
int pci_epc_write_header(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
||||
|
@ -71,10 +71,14 @@ struct pci_epf_ops {
|
||||
* struct pci_epf_event_ops - Callbacks for capturing the EPC events
|
||||
* @core_init: Callback for the EPC initialization complete event
|
||||
* @link_up: Callback for the EPC link up event
|
||||
* @link_down: Callback for the EPC link down event
|
||||
* @bme: Callback for the EPC BME (Bus Master Enable) event
|
||||
*/
|
||||
struct pci_epc_event_ops {
|
||||
int (*core_init)(struct pci_epf *epf);
|
||||
int (*link_up)(struct pci_epf *epf);
|
||||
int (*link_down)(struct pci_epf *epf);
|
||||
int (*bme)(struct pci_epf *epf);
|
||||
};
|
||||
|
||||
/**
|
||||
@ -89,7 +93,8 @@ struct pci_epc_event_ops {
|
||||
* @id_table: identifies EPF devices for probing
|
||||
*/
|
||||
struct pci_epf_driver {
|
||||
int (*probe)(struct pci_epf *epf);
|
||||
int (*probe)(struct pci_epf *epf,
|
||||
const struct pci_epf_device_id *id);
|
||||
void (*remove)(struct pci_epf *epf);
|
||||
|
||||
struct device_driver driver;
|
||||
@ -131,6 +136,7 @@ struct pci_epf_bar {
|
||||
* @epc: the EPC device to which this EPF device is bound
|
||||
* @epf_pf: the physical EPF device to which this virtual EPF device is bound
|
||||
* @driver: the EPF driver to which this EPF device is bound
|
||||
* @id: Pointer to the EPF device ID
|
||||
* @list: to add pci_epf as a list of PCI endpoint functions to pci_epc
|
||||
* @lock: mutex to protect pci_epf_ops
|
||||
* @sec_epc: the secondary EPC device to which this EPF device is bound
|
||||
@ -158,6 +164,7 @@ struct pci_epf {
|
||||
struct pci_epc *epc;
|
||||
struct pci_epf *epf_pf;
|
||||
struct pci_epf_driver *driver;
|
||||
const struct pci_epf_device_id *id;
|
||||
struct list_head list;
|
||||
/* mutex to protect against concurrent access of pci_epf_ops */
|
||||
struct mutex lock;
|
||||
@ -214,8 +221,6 @@ void pci_epf_free_space(struct pci_epf *epf, void *addr, enum pci_barno bar,
|
||||
enum pci_epc_interface_type type);
|
||||
int pci_epf_bind(struct pci_epf *epf);
|
||||
void pci_epf_unbind(struct pci_epf *epf);
|
||||
struct config_group *pci_epf_type_add_cfs(struct pci_epf *epf,
|
||||
struct config_group *group);
|
||||
int pci_epf_add_vepf(struct pci_epf *epf_pf, struct pci_epf *epf_vf);
|
||||
void pci_epf_remove_vepf(struct pci_epf *epf_pf, struct pci_epf *epf_vf);
|
||||
#endif /* __LINUX_PCI_EPF_H */
|
||||
|
Loading…
Reference in New Issue
Block a user