|  | /* | 
|  | * Broadcom specific AMBA | 
|  | * PCI Host | 
|  | * | 
|  | * Licensed under the GNU/GPL. See COPYING for details. | 
|  | */ | 
|  |  | 
|  | #include "bcma_private.h" | 
|  | #include <linux/slab.h> | 
|  | #include <linux/bcma/bcma.h> | 
|  | #include <linux/pci.h> | 
|  | #include <linux/module.h> | 
|  |  | 
|  | static void bcma_host_pci_switch_core(struct bcma_device *core) | 
|  | { | 
|  | int win2 = core->bus->host_is_pcie2 ? | 
|  | BCMA_PCIE2_BAR0_WIN2 : BCMA_PCI_BAR0_WIN2; | 
|  |  | 
|  | pci_write_config_dword(core->bus->host_pci, BCMA_PCI_BAR0_WIN, | 
|  | core->addr); | 
|  | pci_write_config_dword(core->bus->host_pci, win2, core->wrap); | 
|  | core->bus->mapped_core = core; | 
|  | bcma_debug(core->bus, "Switched to core: 0x%X\n", core->id.id); | 
|  | } | 
|  |  | 
|  | /* Provides access to the requested core. Returns base offset that has to be | 
|  | * used. It makes use of fixed windows when possible. */ | 
|  | static u16 bcma_host_pci_provide_access_to_core(struct bcma_device *core) | 
|  | { | 
|  | switch (core->id.id) { | 
|  | case BCMA_CORE_CHIPCOMMON: | 
|  | return 3 * BCMA_CORE_SIZE; | 
|  | case BCMA_CORE_PCIE: | 
|  | return 2 * BCMA_CORE_SIZE; | 
|  | } | 
|  |  | 
|  | if (core->bus->mapped_core != core) | 
|  | bcma_host_pci_switch_core(core); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static u8 bcma_host_pci_read8(struct bcma_device *core, u16 offset) | 
|  | { | 
|  | offset += bcma_host_pci_provide_access_to_core(core); | 
|  | return ioread8(core->bus->mmio + offset); | 
|  | } | 
|  |  | 
|  | static u16 bcma_host_pci_read16(struct bcma_device *core, u16 offset) | 
|  | { | 
|  | offset += bcma_host_pci_provide_access_to_core(core); | 
|  | return ioread16(core->bus->mmio + offset); | 
|  | } | 
|  |  | 
|  | static u32 bcma_host_pci_read32(struct bcma_device *core, u16 offset) | 
|  | { | 
|  | offset += bcma_host_pci_provide_access_to_core(core); | 
|  | return ioread32(core->bus->mmio + offset); | 
|  | } | 
|  |  | 
|  | static void bcma_host_pci_write8(struct bcma_device *core, u16 offset, | 
|  | u8 value) | 
|  | { | 
|  | offset += bcma_host_pci_provide_access_to_core(core); | 
|  | iowrite8(value, core->bus->mmio + offset); | 
|  | } | 
|  |  | 
|  | static void bcma_host_pci_write16(struct bcma_device *core, u16 offset, | 
|  | u16 value) | 
|  | { | 
|  | offset += bcma_host_pci_provide_access_to_core(core); | 
|  | iowrite16(value, core->bus->mmio + offset); | 
|  | } | 
|  |  | 
|  | static void bcma_host_pci_write32(struct bcma_device *core, u16 offset, | 
|  | u32 value) | 
|  | { | 
|  | offset += bcma_host_pci_provide_access_to_core(core); | 
|  | iowrite32(value, core->bus->mmio + offset); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_BCMA_BLOCKIO | 
|  | static void bcma_host_pci_block_read(struct bcma_device *core, void *buffer, | 
|  | size_t count, u16 offset, u8 reg_width) | 
|  | { | 
|  | void __iomem *addr = core->bus->mmio + offset; | 
|  | if (core->bus->mapped_core != core) | 
|  | bcma_host_pci_switch_core(core); | 
|  | switch (reg_width) { | 
|  | case sizeof(u8): | 
|  | ioread8_rep(addr, buffer, count); | 
|  | break; | 
|  | case sizeof(u16): | 
|  | WARN_ON(count & 1); | 
|  | ioread16_rep(addr, buffer, count >> 1); | 
|  | break; | 
|  | case sizeof(u32): | 
|  | WARN_ON(count & 3); | 
|  | ioread32_rep(addr, buffer, count >> 2); | 
|  | break; | 
|  | default: | 
|  | WARN_ON(1); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void bcma_host_pci_block_write(struct bcma_device *core, | 
|  | const void *buffer, size_t count, | 
|  | u16 offset, u8 reg_width) | 
|  | { | 
|  | void __iomem *addr = core->bus->mmio + offset; | 
|  | if (core->bus->mapped_core != core) | 
|  | bcma_host_pci_switch_core(core); | 
|  | switch (reg_width) { | 
|  | case sizeof(u8): | 
|  | iowrite8_rep(addr, buffer, count); | 
|  | break; | 
|  | case sizeof(u16): | 
|  | WARN_ON(count & 1); | 
|  | iowrite16_rep(addr, buffer, count >> 1); | 
|  | break; | 
|  | case sizeof(u32): | 
|  | WARN_ON(count & 3); | 
|  | iowrite32_rep(addr, buffer, count >> 2); | 
|  | break; | 
|  | default: | 
|  | WARN_ON(1); | 
|  | } | 
|  | } | 
|  | #endif | 
|  |  | 
|  | static u32 bcma_host_pci_aread32(struct bcma_device *core, u16 offset) | 
|  | { | 
|  | if (core->bus->mapped_core != core) | 
|  | bcma_host_pci_switch_core(core); | 
|  | return ioread32(core->bus->mmio + (1 * BCMA_CORE_SIZE) + offset); | 
|  | } | 
|  |  | 
|  | static void bcma_host_pci_awrite32(struct bcma_device *core, u16 offset, | 
|  | u32 value) | 
|  | { | 
|  | if (core->bus->mapped_core != core) | 
|  | bcma_host_pci_switch_core(core); | 
|  | iowrite32(value, core->bus->mmio + (1 * BCMA_CORE_SIZE) + offset); | 
|  | } | 
|  |  | 
|  | static const struct bcma_host_ops bcma_host_pci_ops = { | 
|  | .read8		= bcma_host_pci_read8, | 
|  | .read16		= bcma_host_pci_read16, | 
|  | .read32		= bcma_host_pci_read32, | 
|  | .write8		= bcma_host_pci_write8, | 
|  | .write16	= bcma_host_pci_write16, | 
|  | .write32	= bcma_host_pci_write32, | 
|  | #ifdef CONFIG_BCMA_BLOCKIO | 
|  | .block_read	= bcma_host_pci_block_read, | 
|  | .block_write	= bcma_host_pci_block_write, | 
|  | #endif | 
|  | .aread32	= bcma_host_pci_aread32, | 
|  | .awrite32	= bcma_host_pci_awrite32, | 
|  | }; | 
|  |  | 
|  | static int bcma_host_pci_probe(struct pci_dev *dev, | 
|  | const struct pci_device_id *id) | 
|  | { | 
|  | struct bcma_bus *bus; | 
|  | int err = -ENOMEM; | 
|  | const char *name; | 
|  | u32 val; | 
|  |  | 
|  | /* Alloc */ | 
|  | bus = kzalloc(sizeof(*bus), GFP_KERNEL); | 
|  | if (!bus) | 
|  | goto out; | 
|  |  | 
|  | /* Basic PCI configuration */ | 
|  | err = pci_enable_device(dev); | 
|  | if (err) | 
|  | goto err_kfree_bus; | 
|  |  | 
|  | name = dev_name(&dev->dev); | 
|  | if (dev->driver && dev->driver->name) | 
|  | name = dev->driver->name; | 
|  | err = pci_request_regions(dev, name); | 
|  | if (err) | 
|  | goto err_pci_disable; | 
|  | pci_set_master(dev); | 
|  |  | 
|  | /* Disable the RETRY_TIMEOUT register (0x41) to keep | 
|  | * PCI Tx retries from interfering with C3 CPU state */ | 
|  | pci_read_config_dword(dev, 0x40, &val); | 
|  | if ((val & 0x0000ff00) != 0) | 
|  | pci_write_config_dword(dev, 0x40, val & 0xffff00ff); | 
|  |  | 
|  | /* SSB needed additional powering up, do we have any AMBA PCI cards? */ | 
|  | if (!pci_is_pcie(dev)) { | 
|  | bcma_err(bus, "PCI card detected, they are not supported.\n"); | 
|  | err = -ENXIO; | 
|  | goto err_pci_release_regions; | 
|  | } | 
|  |  | 
|  | bus->dev = &dev->dev; | 
|  |  | 
|  | /* Map MMIO */ | 
|  | err = -ENOMEM; | 
|  | bus->mmio = pci_iomap(dev, 0, ~0UL); | 
|  | if (!bus->mmio) | 
|  | goto err_pci_release_regions; | 
|  |  | 
|  | /* Host specific */ | 
|  | bus->host_pci = dev; | 
|  | bus->hosttype = BCMA_HOSTTYPE_PCI; | 
|  | bus->ops = &bcma_host_pci_ops; | 
|  |  | 
|  | bus->boardinfo.vendor = bus->host_pci->subsystem_vendor; | 
|  | bus->boardinfo.type = bus->host_pci->subsystem_device; | 
|  |  | 
|  | /* Initialize struct, detect chip */ | 
|  | bcma_init_bus(bus); | 
|  |  | 
|  | /* Scan bus to find out generation of PCIe core */ | 
|  | err = bcma_bus_scan(bus); | 
|  | if (err) | 
|  | goto err_pci_unmap_mmio; | 
|  |  | 
|  | if (bcma_find_core(bus, BCMA_CORE_PCIE2)) | 
|  | bus->host_is_pcie2 = true; | 
|  |  | 
|  | /* Register */ | 
|  | err = bcma_bus_register(bus); | 
|  | if (err) | 
|  | goto err_unregister_cores; | 
|  |  | 
|  | pci_set_drvdata(dev, bus); | 
|  |  | 
|  | out: | 
|  | return err; | 
|  |  | 
|  | err_unregister_cores: | 
|  | bcma_unregister_cores(bus); | 
|  | err_pci_unmap_mmio: | 
|  | pci_iounmap(dev, bus->mmio); | 
|  | err_pci_release_regions: | 
|  | pci_release_regions(dev); | 
|  | err_pci_disable: | 
|  | pci_disable_device(dev); | 
|  | err_kfree_bus: | 
|  | kfree(bus); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | static void bcma_host_pci_remove(struct pci_dev *dev) | 
|  | { | 
|  | struct bcma_bus *bus = pci_get_drvdata(dev); | 
|  |  | 
|  | bcma_bus_unregister(bus); | 
|  | pci_iounmap(dev, bus->mmio); | 
|  | pci_release_regions(dev); | 
|  | pci_disable_device(dev); | 
|  | kfree(bus); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_PM_SLEEP | 
|  | static int bcma_host_pci_suspend(struct device *dev) | 
|  | { | 
|  | struct pci_dev *pdev = to_pci_dev(dev); | 
|  | struct bcma_bus *bus = pci_get_drvdata(pdev); | 
|  |  | 
|  | bus->mapped_core = NULL; | 
|  |  | 
|  | return bcma_bus_suspend(bus); | 
|  | } | 
|  |  | 
|  | static int bcma_host_pci_resume(struct device *dev) | 
|  | { | 
|  | struct pci_dev *pdev = to_pci_dev(dev); | 
|  | struct bcma_bus *bus = pci_get_drvdata(pdev); | 
|  |  | 
|  | return bcma_bus_resume(bus); | 
|  | } | 
|  |  | 
|  | static SIMPLE_DEV_PM_OPS(bcma_pm_ops, bcma_host_pci_suspend, | 
|  | bcma_host_pci_resume); | 
|  | #define BCMA_PM_OPS	(&bcma_pm_ops) | 
|  |  | 
|  | #else /* CONFIG_PM_SLEEP */ | 
|  |  | 
|  | #define BCMA_PM_OPS     NULL | 
|  |  | 
|  | #endif /* CONFIG_PM_SLEEP */ | 
|  |  | 
|  | static const struct pci_device_id bcma_pci_bridge_tbl[] = { | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x0576) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4313) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43224) },	/* 0xa8d8 */ | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4331) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4353) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4357) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4358) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4359) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4360) }, | 
|  | { PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_DELL, 0x0016) }, | 
|  | { PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_DELL, 0x0018) }, | 
|  | { PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_FOXCONN, 0xe092) }, | 
|  | { PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_HP, 0x804a) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43a0) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43a9) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43aa) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43b1) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4727) }, | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43227) },	/* 0xa8db, BCM43217 (sic!) */ | 
|  | { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43228) },	/* 0xa8dc */ | 
|  | { 0, }, | 
|  | }; | 
|  | MODULE_DEVICE_TABLE(pci, bcma_pci_bridge_tbl); | 
|  |  | 
|  | static struct pci_driver bcma_pci_bridge_driver = { | 
|  | .name = "bcma-pci-bridge", | 
|  | .id_table = bcma_pci_bridge_tbl, | 
|  | .probe = bcma_host_pci_probe, | 
|  | .remove = bcma_host_pci_remove, | 
|  | .driver.pm = BCMA_PM_OPS, | 
|  | }; | 
|  |  | 
|  | int __init bcma_host_pci_init(void) | 
|  | { | 
|  | return pci_register_driver(&bcma_pci_bridge_driver); | 
|  | } | 
|  |  | 
|  | void __exit bcma_host_pci_exit(void) | 
|  | { | 
|  | pci_unregister_driver(&bcma_pci_bridge_driver); | 
|  | } | 
|  |  | 
|  | /************************************************** | 
|  | * Runtime ops for drivers. | 
|  | **************************************************/ | 
|  |  | 
|  | /* See also pcicore_up */ | 
|  | void bcma_host_pci_up(struct bcma_bus *bus) | 
|  | { | 
|  | if (bus->hosttype != BCMA_HOSTTYPE_PCI) | 
|  | return; | 
|  |  | 
|  | if (bus->host_is_pcie2) | 
|  | bcma_core_pcie2_up(&bus->drv_pcie2); | 
|  | else | 
|  | bcma_core_pci_up(&bus->drv_pci[0]); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(bcma_host_pci_up); | 
|  |  | 
|  | /* See also pcicore_down */ | 
|  | void bcma_host_pci_down(struct bcma_bus *bus) | 
|  | { | 
|  | if (bus->hosttype != BCMA_HOSTTYPE_PCI) | 
|  | return; | 
|  |  | 
|  | if (!bus->host_is_pcie2) | 
|  | bcma_core_pci_down(&bus->drv_pci[0]); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(bcma_host_pci_down); | 
|  |  | 
|  | /* See also si_pci_setup */ | 
|  | int bcma_host_pci_irq_ctl(struct bcma_bus *bus, struct bcma_device *core, | 
|  | bool enable) | 
|  | { | 
|  | struct pci_dev *pdev; | 
|  | u32 coremask, tmp; | 
|  | int err = 0; | 
|  |  | 
|  | if (bus->hosttype != BCMA_HOSTTYPE_PCI) { | 
|  | /* This bcma device is not on a PCI host-bus. So the IRQs are | 
|  | * not routed through the PCI core. | 
|  | * So we must not enable routing through the PCI core. */ | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | pdev = bus->host_pci; | 
|  |  | 
|  | err = pci_read_config_dword(pdev, BCMA_PCI_IRQMASK, &tmp); | 
|  | if (err) | 
|  | goto out; | 
|  |  | 
|  | coremask = BIT(core->core_index) << 8; | 
|  | if (enable) | 
|  | tmp |= coremask; | 
|  | else | 
|  | tmp &= ~coremask; | 
|  |  | 
|  | err = pci_write_config_dword(pdev, BCMA_PCI_IRQMASK, tmp); | 
|  |  | 
|  | out: | 
|  | return err; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(bcma_host_pci_irq_ctl); |