mirror of
https://github.com/torvalds/linux.git
synced 2025-11-06 03:29:11 +02:00
-----BEGIN PGP SIGNATURE-----
iQJIBAABCgAyFiEEgMe7l+5h9hnxdsnuWYigwDrT+vwFAmXw04sUHGJoZWxnYWFz
QGdvb2dsZS5jb20ACgkQWYigwDrT+vyT3xAAsp5+c2IcbrXpZZM7figwx4y9PPRp
jcQ4AYSGP41xqTUGXUTcVZYvRorSIAFEOz33U0SL1UNxoOZz8j/M6SD58k8a6XRr
9SSPuKja1OKJjONhS1bzrcbVtuzr71ISrECXfLkvW5hY5hvq+3+anMtu3/UIEHu6
M1vVc+basRjjPJNTixMWvVqS3R+4gDAFeBtdZl/D+U6v0v2xOK+81YZqjfZZCw9v
xmdHHK2dKNEdysNoRJ5cafY3b1NnSsrxlHbIhBnKt+7uRSWKD1dHcBQj7wDc/HrX
yBGca+BZBKitXEJM3p5KcWWs4ijaywGw0GSffUIKrN9i6RIfwnxBH9jUbwDngifU
2IR/kLEqdjYi/WnENxIHpQATLyXhXZ8uEnLS0xMlRIA96u3M5B0mrYOZxaN3bo12
A3aE+aPOTw0u1wf7G8dBX6IdYnjZ/ZuR9Q+fVoKpZBvsYUVaKyiKCtKMCNaVirn5
z7nxR1W71ee+35+37KthPXhiw+YtURGz1wBWt+wWUMjBcpIj2bpzU9wQDE9ZMdYt
XJoJcatrRhJzefO3uzd0egft+vwk0xrj5LQEDhMQyDrnBLC4EgI5niKPWqbay5Nx
Cnll01CI82xAnIF6eu7OOuI1nYGtoFcY8rP3hTC85cWN7Xi8SAOLTZZcVTpfBMUr
l2uEll8p+8dZ6IY=
=AP3I
-----END PGP SIGNATURE-----
Merge tag 'pci-v6.9-changes' of git://git.kernel.org/pub/scm/linux/kernel/git/pci/pci
Pull PCI updates from Bjorn Helgaas:
"Enumeration:
- Consolidate interrupt related code in irq.c (Ilpo Järvinen)
- Reduce kernel size by replacing sysfs resource macros with
functions (Ilpo Järvinen)
- Reduce kernel size by compiling sysfs support only when
CONFIG_SYSFS=y (Lukas Wunner)
- Avoid using Extended Tags on 3ware-9650SE Root Port to work around
an apparent hardware defect (Jörg Wedekind)
Resource management:
- Fix an MMIO mapping leak in pci_iounmap() (Philipp Stanner)
- Move pci_iomap.c and other PCI-specific devres code to drivers/pci
(Philipp Stanner)
- Consolidate PCI devres code in devres.c (Philipp Stanner)
Power management:
- Avoid D3cold on Asus B1400 PCI-NVMe bridge, where firmware doesn't
know how to return correctly to D0, and remove previous quirk that
wasn't as specific (Daniel Drake)
- Allow runtime PM when the driver enables it but doesn't need any
runtime PM callbacks (Raag Jadav)
- Drain runtime-idle callbacks before driver removal to avoid races
between .remove() and .runtime_idle(), which caused intermittent
page faults when the rtsx .runtime_idle() accessed registers that
its .remove() had already unmapped (Rafael J. Wysocki)
Virtualization:
- Avoid Secondary Bus Reset on LSI FW643 so it can be assigned to VMs
with VFIO, e.g., for professional audio software on many Apple
machines, at the cost of leaking state between VMs (Edmund Raile)
Error handling:
- Print all logged TLP Prefixes, not just the first, after AER or DPC
errors (Ilpo Järvinen)
- Quirk the DPC PIO log size for Intel Raptor Lake Root Ports, which
still don't advertise a legal size (Paul Menzel)
- Ignore expected DPC Surprise Down errors on hot removal (Smita
Koralahalli)
- Block runtime suspend while handling AER errors to avoid races that
prevent the device form being resumed from D3hot (Stanislaw
Gruszka)
Peer-to-peer DMA:
- Use atomic XA allocation in RCU read section (Christophe JAILLET)
ASPM:
- Collect bits of ASPM-related code that we need even without
CONFIG_PCIEASPM into aspm.c (David E. Box)
- Save/restore L1 PM Substates config for suspend/resume (David E.
Box)
- Update save_save when ASPM config is changed, so a .slot_reset()
during error recovery restores the changed config, not the
.probe()-time config (Vidya Sagar)
Endpoint framework:
- Refactor and improve pci_epf_alloc_space() API (Niklas Cassel)
- Clean up endpoint BAR descriptions (Niklas Cassel)
- Fix ntb_register_device() name leak in error path (Yang Yingliang)
- Return actual error code for pci_vntb_probe() failure (Yang
Yingliang)
Broadcom STB PCIe controller driver:
- Fix MDIO write polling, which previously never waited for
completion (Jonathan Bell)
Cadence PCIe endpoint driver:
- Clear the ARI "Next Function Number" of last function (Jasko-EXT
Wojciech)
Freescale i.MX6 PCIe controller driver:
- Simplify by replacing switch statements with function pointers for
different hardware variants (Frank Li)
- Simplify by using clk_bulk*() API (Frank Li)
- Remove redundant DT clock and reg/reg-name details (Frank Li)
- Add i.MX95 DT and driver support for both Root Complex and Endpoint
mode (Frank Li)
Microsoft Hyper-V host bridge driver:
- Reduce memory usage by limiting ring buffer size to 16KB instead of
4 pages (Michael Kelley)
Qualcomm PCIe controller driver:
- Add X1E80100 DT and driver support (Abel Vesa)
- Add DT 'required-opps' for SoCs that require a minimum performance
level (Johan Hovold)
- Make DT 'msi-map-mask' optional, depending on how MSI interrupts
are mapped (Johan Hovold)
- Disable ASPM L0s for sc8280xp, sa8540p and sa8295p because the PHY
configuration isn't tuned correctly for L0s (Johan Hovold)
- Split dt-binding qcom,pcie.yaml into qcom,pcie-common.yaml and
separate files for SA8775p, SC7280, SC8180X, SC8280XP, SM8150,
SM8250, SM8350, SM8450, SM8550 for easier reviewing (Krzysztof
Kozlowski)
- Enable BDF to SID translation by disabling bypass mode (Manivannan
Sadhasivam)
- Add endpoint MHI support for Snapdragon SA8775P SoC (Mrinmay
Sarkar)
Synopsys DesignWare PCIe controller driver:
- Allocate 64-bit MSI address if no 32-bit address is available (Ajay
Agarwal)
- Fix endpoint Resizable BAR to actually advertise the required 1MB
size (Niklas Cassel)
MicroSemi Switchtec management driver:
- Release resources if the .probe() fails (Christophe JAILLET)
Miscellaneous:
- Make pcie_port_bus_type const (Ricardo B. Marliere)"
* tag 'pci-v6.9-changes' of git://git.kernel.org/pub/scm/linux/kernel/git/pci/pci: (77 commits)
PCI/ASPM: Update save_state when configuration changes
PCI/ASPM: Disable L1 before configuring L1 Substates
PCI/ASPM: Call pci_save_ltr_state() from pci_save_pcie_state()
PCI/ASPM: Save L1 PM Substates Capability for suspend/resume
PCI: hv: Fix ring buffer size calculation
PCI: dwc: endpoint: Fix advertised resizable BAR size
PCI: cadence: Clear the ARI Capability Next Function Number of the last function
PCI: dwc: Strengthen the MSI address allocation logic
PCI: brcmstb: Fix broken brcm_pcie_mdio_write() polling
PCI: qcom: Add X1E80100 PCIe support
dt-bindings: PCI: qcom: Document the X1E80100 PCIe Controller
PCI: qcom: Enable BDF to SID translation properly
PCI/AER: Generalize TLP Header Log reading
PCI/AER: Use explicit register size for PCI_ERR_CAP
PCI: qcom: Disable ASPM L0s for sc8280xp, sa8540p and sa8295p
dt-bindings: PCI: qcom: Do not require 'msi-map-mask'
dt-bindings: PCI: qcom: Allow 'required-opps'
PCI/AER: Block runtime suspend when handling errors
PCI/ASPM: Move pci_save_ltr_state() to aspm.c
PCI/ASPM: Always build aspm.c
...
791 lines
20 KiB
C
791 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Synopsys DesignWare PCIe Endpoint controller driver
|
|
*
|
|
* Copyright (C) 2017 Texas Instruments
|
|
* Author: Kishon Vijay Abraham I <kishon@ti.com>
|
|
*/
|
|
|
|
#include <linux/align.h>
|
|
#include <linux/bitfield.h>
|
|
#include <linux/of.h>
|
|
#include <linux/platform_device.h>
|
|
|
|
#include "pcie-designware.h"
|
|
#include <linux/pci-epc.h>
|
|
#include <linux/pci-epf.h>
|
|
|
|
void dw_pcie_ep_linkup(struct dw_pcie_ep *ep)
|
|
{
|
|
struct pci_epc *epc = ep->epc;
|
|
|
|
pci_epc_linkup(epc);
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_linkup);
|
|
|
|
void dw_pcie_ep_init_notify(struct dw_pcie_ep *ep)
|
|
{
|
|
struct pci_epc *epc = ep->epc;
|
|
|
|
pci_epc_init_notify(epc);
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_init_notify);
|
|
|
|
struct dw_pcie_ep_func *
|
|
dw_pcie_ep_get_func_from_ep(struct dw_pcie_ep *ep, u8 func_no)
|
|
{
|
|
struct dw_pcie_ep_func *ep_func;
|
|
|
|
list_for_each_entry(ep_func, &ep->func_list, list) {
|
|
if (ep_func->func_no == func_no)
|
|
return ep_func;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void __dw_pcie_ep_reset_bar(struct dw_pcie *pci, u8 func_no,
|
|
enum pci_barno bar, int flags)
|
|
{
|
|
struct dw_pcie_ep *ep = &pci->ep;
|
|
u32 reg;
|
|
|
|
reg = PCI_BASE_ADDRESS_0 + (4 * bar);
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
dw_pcie_ep_writel_dbi2(ep, func_no, reg, 0x0);
|
|
dw_pcie_ep_writel_dbi(ep, func_no, reg, 0x0);
|
|
if (flags & PCI_BASE_ADDRESS_MEM_TYPE_64) {
|
|
dw_pcie_ep_writel_dbi2(ep, func_no, reg + 4, 0x0);
|
|
dw_pcie_ep_writel_dbi(ep, func_no, reg + 4, 0x0);
|
|
}
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
}
|
|
|
|
void dw_pcie_ep_reset_bar(struct dw_pcie *pci, enum pci_barno bar)
|
|
{
|
|
u8 func_no, funcs;
|
|
|
|
funcs = pci->ep.epc->max_functions;
|
|
|
|
for (func_no = 0; func_no < funcs; func_no++)
|
|
__dw_pcie_ep_reset_bar(pci, func_no, bar, 0);
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_reset_bar);
|
|
|
|
static u8 __dw_pcie_ep_find_next_cap(struct dw_pcie_ep *ep, u8 func_no,
|
|
u8 cap_ptr, u8 cap)
|
|
{
|
|
u8 cap_id, next_cap_ptr;
|
|
u16 reg;
|
|
|
|
if (!cap_ptr)
|
|
return 0;
|
|
|
|
reg = dw_pcie_ep_readw_dbi(ep, func_no, cap_ptr);
|
|
cap_id = (reg & 0x00ff);
|
|
|
|
if (cap_id > PCI_CAP_ID_MAX)
|
|
return 0;
|
|
|
|
if (cap_id == cap)
|
|
return cap_ptr;
|
|
|
|
next_cap_ptr = (reg & 0xff00) >> 8;
|
|
return __dw_pcie_ep_find_next_cap(ep, func_no, next_cap_ptr, cap);
|
|
}
|
|
|
|
static u8 dw_pcie_ep_find_capability(struct dw_pcie_ep *ep, u8 func_no, u8 cap)
|
|
{
|
|
u8 next_cap_ptr;
|
|
u16 reg;
|
|
|
|
reg = dw_pcie_ep_readw_dbi(ep, func_no, PCI_CAPABILITY_LIST);
|
|
next_cap_ptr = (reg & 0x00ff);
|
|
|
|
return __dw_pcie_ep_find_next_cap(ep, func_no, next_cap_ptr, cap);
|
|
}
|
|
|
|
static int dw_pcie_ep_write_header(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
struct pci_epf_header *hdr)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
dw_pcie_ep_writew_dbi(ep, func_no, PCI_VENDOR_ID, hdr->vendorid);
|
|
dw_pcie_ep_writew_dbi(ep, func_no, PCI_DEVICE_ID, hdr->deviceid);
|
|
dw_pcie_ep_writeb_dbi(ep, func_no, PCI_REVISION_ID, hdr->revid);
|
|
dw_pcie_ep_writeb_dbi(ep, func_no, PCI_CLASS_PROG, hdr->progif_code);
|
|
dw_pcie_ep_writew_dbi(ep, func_no, PCI_CLASS_DEVICE,
|
|
hdr->subclass_code | hdr->baseclass_code << 8);
|
|
dw_pcie_ep_writeb_dbi(ep, func_no, PCI_CACHE_LINE_SIZE,
|
|
hdr->cache_line_size);
|
|
dw_pcie_ep_writew_dbi(ep, func_no, PCI_SUBSYSTEM_VENDOR_ID,
|
|
hdr->subsys_vendor_id);
|
|
dw_pcie_ep_writew_dbi(ep, func_no, PCI_SUBSYSTEM_ID, hdr->subsys_id);
|
|
dw_pcie_ep_writeb_dbi(ep, func_no, PCI_INTERRUPT_PIN,
|
|
hdr->interrupt_pin);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dw_pcie_ep_inbound_atu(struct dw_pcie_ep *ep, u8 func_no, int type,
|
|
dma_addr_t cpu_addr, enum pci_barno bar)
|
|
{
|
|
int ret;
|
|
u32 free_win;
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
if (!ep->bar_to_atu[bar])
|
|
free_win = find_first_zero_bit(ep->ib_window_map, pci->num_ib_windows);
|
|
else
|
|
free_win = ep->bar_to_atu[bar];
|
|
|
|
if (free_win >= pci->num_ib_windows) {
|
|
dev_err(pci->dev, "No free inbound window\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ret = dw_pcie_prog_ep_inbound_atu(pci, func_no, free_win, type,
|
|
cpu_addr, bar);
|
|
if (ret < 0) {
|
|
dev_err(pci->dev, "Failed to program IB window\n");
|
|
return ret;
|
|
}
|
|
|
|
ep->bar_to_atu[bar] = free_win;
|
|
set_bit(free_win, ep->ib_window_map);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dw_pcie_ep_outbound_atu(struct dw_pcie_ep *ep, u8 func_no,
|
|
phys_addr_t phys_addr,
|
|
u64 pci_addr, size_t size)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
u32 free_win;
|
|
int ret;
|
|
|
|
free_win = find_first_zero_bit(ep->ob_window_map, pci->num_ob_windows);
|
|
if (free_win >= pci->num_ob_windows) {
|
|
dev_err(pci->dev, "No free outbound window\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ret = dw_pcie_prog_ep_outbound_atu(pci, func_no, free_win, PCIE_ATU_TYPE_MEM,
|
|
phys_addr, pci_addr, size);
|
|
if (ret)
|
|
return ret;
|
|
|
|
set_bit(free_win, ep->ob_window_map);
|
|
ep->outbound_addr[free_win] = phys_addr;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void dw_pcie_ep_clear_bar(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
struct pci_epf_bar *epf_bar)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
enum pci_barno bar = epf_bar->barno;
|
|
u32 atu_index = ep->bar_to_atu[bar];
|
|
|
|
__dw_pcie_ep_reset_bar(pci, func_no, bar, epf_bar->flags);
|
|
|
|
dw_pcie_disable_atu(pci, PCIE_ATU_REGION_DIR_IB, atu_index);
|
|
clear_bit(atu_index, ep->ib_window_map);
|
|
ep->epf_bar[bar] = NULL;
|
|
ep->bar_to_atu[bar] = 0;
|
|
}
|
|
|
|
static int dw_pcie_ep_set_bar(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
struct pci_epf_bar *epf_bar)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
enum pci_barno bar = epf_bar->barno;
|
|
size_t size = epf_bar->size;
|
|
int flags = epf_bar->flags;
|
|
int ret, type;
|
|
u32 reg;
|
|
|
|
reg = PCI_BASE_ADDRESS_0 + (4 * bar);
|
|
|
|
if (!(flags & PCI_BASE_ADDRESS_SPACE))
|
|
type = PCIE_ATU_TYPE_MEM;
|
|
else
|
|
type = PCIE_ATU_TYPE_IO;
|
|
|
|
ret = dw_pcie_ep_inbound_atu(ep, func_no, type, epf_bar->phys_addr, bar);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (ep->epf_bar[bar])
|
|
return 0;
|
|
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
|
|
dw_pcie_ep_writel_dbi2(ep, func_no, reg, lower_32_bits(size - 1));
|
|
dw_pcie_ep_writel_dbi(ep, func_no, reg, flags);
|
|
|
|
if (flags & PCI_BASE_ADDRESS_MEM_TYPE_64) {
|
|
dw_pcie_ep_writel_dbi2(ep, func_no, reg + 4, upper_32_bits(size - 1));
|
|
dw_pcie_ep_writel_dbi(ep, func_no, reg + 4, 0);
|
|
}
|
|
|
|
ep->epf_bar[bar] = epf_bar;
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dw_pcie_find_index(struct dw_pcie_ep *ep, phys_addr_t addr,
|
|
u32 *atu_index)
|
|
{
|
|
u32 index;
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
for (index = 0; index < pci->num_ob_windows; index++) {
|
|
if (ep->outbound_addr[index] != addr)
|
|
continue;
|
|
*atu_index = index;
|
|
return 0;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static void dw_pcie_ep_unmap_addr(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
phys_addr_t addr)
|
|
{
|
|
int ret;
|
|
u32 atu_index;
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
ret = dw_pcie_find_index(ep, addr, &atu_index);
|
|
if (ret < 0)
|
|
return;
|
|
|
|
dw_pcie_disable_atu(pci, PCIE_ATU_REGION_DIR_OB, atu_index);
|
|
clear_bit(atu_index, ep->ob_window_map);
|
|
}
|
|
|
|
static int dw_pcie_ep_map_addr(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
phys_addr_t addr, u64 pci_addr, size_t size)
|
|
{
|
|
int ret;
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
ret = dw_pcie_ep_outbound_atu(ep, func_no, addr, pci_addr, size);
|
|
if (ret) {
|
|
dev_err(pci->dev, "Failed to enable address\n");
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dw_pcie_ep_get_msi(struct pci_epc *epc, u8 func_no, u8 vfunc_no)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie_ep_func *ep_func;
|
|
u32 val, reg;
|
|
|
|
ep_func = dw_pcie_ep_get_func_from_ep(ep, func_no);
|
|
if (!ep_func || !ep_func->msi_cap)
|
|
return -EINVAL;
|
|
|
|
reg = ep_func->msi_cap + PCI_MSI_FLAGS;
|
|
val = dw_pcie_ep_readw_dbi(ep, func_no, reg);
|
|
if (!(val & PCI_MSI_FLAGS_ENABLE))
|
|
return -EINVAL;
|
|
|
|
val = FIELD_GET(PCI_MSI_FLAGS_QSIZE, val);
|
|
|
|
return val;
|
|
}
|
|
|
|
static int dw_pcie_ep_set_msi(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
u8 interrupts)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct dw_pcie_ep_func *ep_func;
|
|
u32 val, reg;
|
|
|
|
ep_func = dw_pcie_ep_get_func_from_ep(ep, func_no);
|
|
if (!ep_func || !ep_func->msi_cap)
|
|
return -EINVAL;
|
|
|
|
reg = ep_func->msi_cap + PCI_MSI_FLAGS;
|
|
val = dw_pcie_ep_readw_dbi(ep, func_no, reg);
|
|
val &= ~PCI_MSI_FLAGS_QMASK;
|
|
val |= FIELD_PREP(PCI_MSI_FLAGS_QMASK, interrupts);
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
dw_pcie_ep_writew_dbi(ep, func_no, reg, val);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dw_pcie_ep_get_msix(struct pci_epc *epc, u8 func_no, u8 vfunc_no)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie_ep_func *ep_func;
|
|
u32 val, reg;
|
|
|
|
ep_func = dw_pcie_ep_get_func_from_ep(ep, func_no);
|
|
if (!ep_func || !ep_func->msix_cap)
|
|
return -EINVAL;
|
|
|
|
reg = ep_func->msix_cap + PCI_MSIX_FLAGS;
|
|
val = dw_pcie_ep_readw_dbi(ep, func_no, reg);
|
|
if (!(val & PCI_MSIX_FLAGS_ENABLE))
|
|
return -EINVAL;
|
|
|
|
val &= PCI_MSIX_FLAGS_QSIZE;
|
|
|
|
return val;
|
|
}
|
|
|
|
static int dw_pcie_ep_set_msix(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
u16 interrupts, enum pci_barno bir, u32 offset)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct dw_pcie_ep_func *ep_func;
|
|
u32 val, reg;
|
|
|
|
ep_func = dw_pcie_ep_get_func_from_ep(ep, func_no);
|
|
if (!ep_func || !ep_func->msix_cap)
|
|
return -EINVAL;
|
|
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
|
|
reg = ep_func->msix_cap + PCI_MSIX_FLAGS;
|
|
val = dw_pcie_ep_readw_dbi(ep, func_no, reg);
|
|
val &= ~PCI_MSIX_FLAGS_QSIZE;
|
|
val |= interrupts;
|
|
dw_pcie_writew_dbi(pci, reg, val);
|
|
|
|
reg = ep_func->msix_cap + PCI_MSIX_TABLE;
|
|
val = offset | bir;
|
|
dw_pcie_ep_writel_dbi(ep, func_no, reg, val);
|
|
|
|
reg = ep_func->msix_cap + PCI_MSIX_PBA;
|
|
val = (offset + (interrupts * PCI_MSIX_ENTRY_SIZE)) | bir;
|
|
dw_pcie_ep_writel_dbi(ep, func_no, reg, val);
|
|
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dw_pcie_ep_raise_irq(struct pci_epc *epc, u8 func_no, u8 vfunc_no,
|
|
unsigned int type, u16 interrupt_num)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
|
|
if (!ep->ops->raise_irq)
|
|
return -EINVAL;
|
|
|
|
return ep->ops->raise_irq(ep, func_no, type, interrupt_num);
|
|
}
|
|
|
|
static void dw_pcie_ep_stop(struct pci_epc *epc)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
dw_pcie_stop_link(pci);
|
|
}
|
|
|
|
static int dw_pcie_ep_start(struct pci_epc *epc)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
return dw_pcie_start_link(pci);
|
|
}
|
|
|
|
static const struct pci_epc_features*
|
|
dw_pcie_ep_get_features(struct pci_epc *epc, u8 func_no, u8 vfunc_no)
|
|
{
|
|
struct dw_pcie_ep *ep = epc_get_drvdata(epc);
|
|
|
|
if (!ep->ops->get_features)
|
|
return NULL;
|
|
|
|
return ep->ops->get_features(ep);
|
|
}
|
|
|
|
static const struct pci_epc_ops epc_ops = {
|
|
.write_header = dw_pcie_ep_write_header,
|
|
.set_bar = dw_pcie_ep_set_bar,
|
|
.clear_bar = dw_pcie_ep_clear_bar,
|
|
.map_addr = dw_pcie_ep_map_addr,
|
|
.unmap_addr = dw_pcie_ep_unmap_addr,
|
|
.set_msi = dw_pcie_ep_set_msi,
|
|
.get_msi = dw_pcie_ep_get_msi,
|
|
.set_msix = dw_pcie_ep_set_msix,
|
|
.get_msix = dw_pcie_ep_get_msix,
|
|
.raise_irq = dw_pcie_ep_raise_irq,
|
|
.start = dw_pcie_ep_start,
|
|
.stop = dw_pcie_ep_stop,
|
|
.get_features = dw_pcie_ep_get_features,
|
|
};
|
|
|
|
int dw_pcie_ep_raise_intx_irq(struct dw_pcie_ep *ep, u8 func_no)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct device *dev = pci->dev;
|
|
|
|
dev_err(dev, "EP cannot raise INTX IRQs\n");
|
|
|
|
return -EINVAL;
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_raise_intx_irq);
|
|
|
|
int dw_pcie_ep_raise_msi_irq(struct dw_pcie_ep *ep, u8 func_no,
|
|
u8 interrupt_num)
|
|
{
|
|
u32 msg_addr_lower, msg_addr_upper, reg;
|
|
struct dw_pcie_ep_func *ep_func;
|
|
struct pci_epc *epc = ep->epc;
|
|
unsigned int aligned_offset;
|
|
u16 msg_ctrl, msg_data;
|
|
bool has_upper;
|
|
u64 msg_addr;
|
|
int ret;
|
|
|
|
ep_func = dw_pcie_ep_get_func_from_ep(ep, func_no);
|
|
if (!ep_func || !ep_func->msi_cap)
|
|
return -EINVAL;
|
|
|
|
/* Raise MSI per the PCI Local Bus Specification Revision 3.0, 6.8.1. */
|
|
reg = ep_func->msi_cap + PCI_MSI_FLAGS;
|
|
msg_ctrl = dw_pcie_ep_readw_dbi(ep, func_no, reg);
|
|
has_upper = !!(msg_ctrl & PCI_MSI_FLAGS_64BIT);
|
|
reg = ep_func->msi_cap + PCI_MSI_ADDRESS_LO;
|
|
msg_addr_lower = dw_pcie_ep_readl_dbi(ep, func_no, reg);
|
|
if (has_upper) {
|
|
reg = ep_func->msi_cap + PCI_MSI_ADDRESS_HI;
|
|
msg_addr_upper = dw_pcie_ep_readl_dbi(ep, func_no, reg);
|
|
reg = ep_func->msi_cap + PCI_MSI_DATA_64;
|
|
msg_data = dw_pcie_ep_readw_dbi(ep, func_no, reg);
|
|
} else {
|
|
msg_addr_upper = 0;
|
|
reg = ep_func->msi_cap + PCI_MSI_DATA_32;
|
|
msg_data = dw_pcie_ep_readw_dbi(ep, func_no, reg);
|
|
}
|
|
msg_addr = ((u64)msg_addr_upper) << 32 | msg_addr_lower;
|
|
|
|
aligned_offset = msg_addr & (epc->mem->window.page_size - 1);
|
|
msg_addr = ALIGN_DOWN(msg_addr, epc->mem->window.page_size);
|
|
ret = dw_pcie_ep_map_addr(epc, func_no, 0, ep->msi_mem_phys, msg_addr,
|
|
epc->mem->window.page_size);
|
|
if (ret)
|
|
return ret;
|
|
|
|
writel(msg_data | (interrupt_num - 1), ep->msi_mem + aligned_offset);
|
|
|
|
dw_pcie_ep_unmap_addr(epc, func_no, 0, ep->msi_mem_phys);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_raise_msi_irq);
|
|
|
|
int dw_pcie_ep_raise_msix_irq_doorbell(struct dw_pcie_ep *ep, u8 func_no,
|
|
u16 interrupt_num)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct dw_pcie_ep_func *ep_func;
|
|
u32 msg_data;
|
|
|
|
ep_func = dw_pcie_ep_get_func_from_ep(ep, func_no);
|
|
if (!ep_func || !ep_func->msix_cap)
|
|
return -EINVAL;
|
|
|
|
msg_data = (func_no << PCIE_MSIX_DOORBELL_PF_SHIFT) |
|
|
(interrupt_num - 1);
|
|
|
|
dw_pcie_writel_dbi(pci, PCIE_MSIX_DOORBELL, msg_data);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int dw_pcie_ep_raise_msix_irq(struct dw_pcie_ep *ep, u8 func_no,
|
|
u16 interrupt_num)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct pci_epf_msix_tbl *msix_tbl;
|
|
struct dw_pcie_ep_func *ep_func;
|
|
struct pci_epc *epc = ep->epc;
|
|
u32 reg, msg_data, vec_ctrl;
|
|
unsigned int aligned_offset;
|
|
u32 tbl_offset;
|
|
u64 msg_addr;
|
|
int ret;
|
|
u8 bir;
|
|
|
|
ep_func = dw_pcie_ep_get_func_from_ep(ep, func_no);
|
|
if (!ep_func || !ep_func->msix_cap)
|
|
return -EINVAL;
|
|
|
|
reg = ep_func->msix_cap + PCI_MSIX_TABLE;
|
|
tbl_offset = dw_pcie_ep_readl_dbi(ep, func_no, reg);
|
|
bir = FIELD_GET(PCI_MSIX_TABLE_BIR, tbl_offset);
|
|
tbl_offset &= PCI_MSIX_TABLE_OFFSET;
|
|
|
|
msix_tbl = ep->epf_bar[bir]->addr + tbl_offset;
|
|
msg_addr = msix_tbl[(interrupt_num - 1)].msg_addr;
|
|
msg_data = msix_tbl[(interrupt_num - 1)].msg_data;
|
|
vec_ctrl = msix_tbl[(interrupt_num - 1)].vector_ctrl;
|
|
|
|
if (vec_ctrl & PCI_MSIX_ENTRY_CTRL_MASKBIT) {
|
|
dev_dbg(pci->dev, "MSI-X entry ctrl set\n");
|
|
return -EPERM;
|
|
}
|
|
|
|
aligned_offset = msg_addr & (epc->mem->window.page_size - 1);
|
|
msg_addr = ALIGN_DOWN(msg_addr, epc->mem->window.page_size);
|
|
ret = dw_pcie_ep_map_addr(epc, func_no, 0, ep->msi_mem_phys, msg_addr,
|
|
epc->mem->window.page_size);
|
|
if (ret)
|
|
return ret;
|
|
|
|
writel(msg_data, ep->msi_mem + aligned_offset);
|
|
|
|
dw_pcie_ep_unmap_addr(epc, func_no, 0, ep->msi_mem_phys);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void dw_pcie_ep_exit(struct dw_pcie_ep *ep)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct pci_epc *epc = ep->epc;
|
|
|
|
dw_pcie_edma_remove(pci);
|
|
|
|
pci_epc_mem_free_addr(epc, ep->msi_mem_phys, ep->msi_mem,
|
|
epc->mem->window.page_size);
|
|
|
|
pci_epc_mem_exit(epc);
|
|
|
|
if (ep->ops->deinit)
|
|
ep->ops->deinit(ep);
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_exit);
|
|
|
|
static unsigned int dw_pcie_ep_find_ext_capability(struct dw_pcie *pci, int cap)
|
|
{
|
|
u32 header;
|
|
int pos = PCI_CFG_SPACE_SIZE;
|
|
|
|
while (pos) {
|
|
header = dw_pcie_readl_dbi(pci, pos);
|
|
if (PCI_EXT_CAP_ID(header) == cap)
|
|
return pos;
|
|
|
|
pos = PCI_EXT_CAP_NEXT(header);
|
|
if (!pos)
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int dw_pcie_ep_init_complete(struct dw_pcie_ep *ep)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
unsigned int offset, ptm_cap_base;
|
|
unsigned int nbars;
|
|
u8 hdr_type;
|
|
u32 reg;
|
|
int i;
|
|
|
|
hdr_type = dw_pcie_readb_dbi(pci, PCI_HEADER_TYPE) &
|
|
PCI_HEADER_TYPE_MASK;
|
|
if (hdr_type != PCI_HEADER_TYPE_NORMAL) {
|
|
dev_err(pci->dev,
|
|
"PCIe controller is not set to EP mode (hdr_type:0x%x)!\n",
|
|
hdr_type);
|
|
return -EIO;
|
|
}
|
|
|
|
offset = dw_pcie_ep_find_ext_capability(pci, PCI_EXT_CAP_ID_REBAR);
|
|
ptm_cap_base = dw_pcie_ep_find_ext_capability(pci, PCI_EXT_CAP_ID_PTM);
|
|
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
|
|
if (offset) {
|
|
reg = dw_pcie_readl_dbi(pci, offset + PCI_REBAR_CTRL);
|
|
nbars = (reg & PCI_REBAR_CTRL_NBAR_MASK) >>
|
|
PCI_REBAR_CTRL_NBAR_SHIFT;
|
|
|
|
/*
|
|
* PCIe r6.0, sec 7.8.6.2 require us to support at least one
|
|
* size in the range from 1 MB to 512 GB. Advertise support
|
|
* for 1 MB BAR size only.
|
|
*/
|
|
for (i = 0; i < nbars; i++, offset += PCI_REBAR_CTRL)
|
|
dw_pcie_writel_dbi(pci, offset + PCI_REBAR_CAP, BIT(4));
|
|
}
|
|
|
|
/*
|
|
* PTM responder capability can be disabled only after disabling
|
|
* PTM root capability.
|
|
*/
|
|
if (ptm_cap_base) {
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
reg = dw_pcie_readl_dbi(pci, ptm_cap_base + PCI_PTM_CAP);
|
|
reg &= ~PCI_PTM_CAP_ROOT;
|
|
dw_pcie_writel_dbi(pci, ptm_cap_base + PCI_PTM_CAP, reg);
|
|
|
|
reg = dw_pcie_readl_dbi(pci, ptm_cap_base + PCI_PTM_CAP);
|
|
reg &= ~(PCI_PTM_CAP_RES | PCI_PTM_GRANULARITY_MASK);
|
|
dw_pcie_writel_dbi(pci, ptm_cap_base + PCI_PTM_CAP, reg);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
}
|
|
|
|
dw_pcie_setup(pci);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_init_complete);
|
|
|
|
int dw_pcie_ep_init(struct dw_pcie_ep *ep)
|
|
{
|
|
int ret;
|
|
void *addr;
|
|
u8 func_no;
|
|
struct resource *res;
|
|
struct pci_epc *epc;
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct device *dev = pci->dev;
|
|
struct platform_device *pdev = to_platform_device(dev);
|
|
struct device_node *np = dev->of_node;
|
|
const struct pci_epc_features *epc_features;
|
|
struct dw_pcie_ep_func *ep_func;
|
|
|
|
INIT_LIST_HEAD(&ep->func_list);
|
|
|
|
ret = dw_pcie_get_resources(pci);
|
|
if (ret)
|
|
return ret;
|
|
|
|
res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "addr_space");
|
|
if (!res)
|
|
return -EINVAL;
|
|
|
|
ep->phys_base = res->start;
|
|
ep->addr_size = resource_size(res);
|
|
|
|
if (ep->ops->pre_init)
|
|
ep->ops->pre_init(ep);
|
|
|
|
dw_pcie_version_detect(pci);
|
|
|
|
dw_pcie_iatu_detect(pci);
|
|
|
|
ep->ib_window_map = devm_bitmap_zalloc(dev, pci->num_ib_windows,
|
|
GFP_KERNEL);
|
|
if (!ep->ib_window_map)
|
|
return -ENOMEM;
|
|
|
|
ep->ob_window_map = devm_bitmap_zalloc(dev, pci->num_ob_windows,
|
|
GFP_KERNEL);
|
|
if (!ep->ob_window_map)
|
|
return -ENOMEM;
|
|
|
|
addr = devm_kcalloc(dev, pci->num_ob_windows, sizeof(phys_addr_t),
|
|
GFP_KERNEL);
|
|
if (!addr)
|
|
return -ENOMEM;
|
|
ep->outbound_addr = addr;
|
|
|
|
epc = devm_pci_epc_create(dev, &epc_ops);
|
|
if (IS_ERR(epc)) {
|
|
dev_err(dev, "Failed to create epc device\n");
|
|
return PTR_ERR(epc);
|
|
}
|
|
|
|
ep->epc = epc;
|
|
epc_set_drvdata(epc, ep);
|
|
|
|
ret = of_property_read_u8(np, "max-functions", &epc->max_functions);
|
|
if (ret < 0)
|
|
epc->max_functions = 1;
|
|
|
|
for (func_no = 0; func_no < epc->max_functions; func_no++) {
|
|
ep_func = devm_kzalloc(dev, sizeof(*ep_func), GFP_KERNEL);
|
|
if (!ep_func)
|
|
return -ENOMEM;
|
|
|
|
ep_func->func_no = func_no;
|
|
ep_func->msi_cap = dw_pcie_ep_find_capability(ep, func_no,
|
|
PCI_CAP_ID_MSI);
|
|
ep_func->msix_cap = dw_pcie_ep_find_capability(ep, func_no,
|
|
PCI_CAP_ID_MSIX);
|
|
|
|
list_add_tail(&ep_func->list, &ep->func_list);
|
|
}
|
|
|
|
if (ep->ops->init)
|
|
ep->ops->init(ep);
|
|
|
|
ret = pci_epc_mem_init(epc, ep->phys_base, ep->addr_size,
|
|
ep->page_size);
|
|
if (ret < 0) {
|
|
dev_err(dev, "Failed to initialize address space\n");
|
|
goto err_ep_deinit;
|
|
}
|
|
|
|
ep->msi_mem = pci_epc_mem_alloc_addr(epc, &ep->msi_mem_phys,
|
|
epc->mem->window.page_size);
|
|
if (!ep->msi_mem) {
|
|
ret = -ENOMEM;
|
|
dev_err(dev, "Failed to reserve memory for MSI/MSI-X\n");
|
|
goto err_exit_epc_mem;
|
|
}
|
|
|
|
ret = dw_pcie_edma_detect(pci);
|
|
if (ret)
|
|
goto err_free_epc_mem;
|
|
|
|
if (ep->ops->get_features) {
|
|
epc_features = ep->ops->get_features(ep);
|
|
if (epc_features->core_init_notifier)
|
|
return 0;
|
|
}
|
|
|
|
ret = dw_pcie_ep_init_complete(ep);
|
|
if (ret)
|
|
goto err_remove_edma;
|
|
|
|
return 0;
|
|
|
|
err_remove_edma:
|
|
dw_pcie_edma_remove(pci);
|
|
|
|
err_free_epc_mem:
|
|
pci_epc_mem_free_addr(epc, ep->msi_mem_phys, ep->msi_mem,
|
|
epc->mem->window.page_size);
|
|
|
|
err_exit_epc_mem:
|
|
pci_epc_mem_exit(epc);
|
|
|
|
err_ep_deinit:
|
|
if (ep->ops->deinit)
|
|
ep->ops->deinit(ep);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_GPL(dw_pcie_ep_init);
|