| /* | 
 |  * Broadcom specific AMBA | 
 |  * PCI Host | 
 |  * | 
 |  * Licensed under the GNU/GPL. See COPYING for details. | 
 |  */ | 
 |  | 
 | #include "bcma_private.h" | 
 | #include <linux/slab.h> | 
 | #include <linux/bcma/bcma.h> | 
 | #include <linux/pci.h> | 
 | #include <linux/module.h> | 
 |  | 
 | static void bcma_host_pci_switch_core(struct bcma_device *core) | 
 | { | 
 | 	int win2 = core->bus->host_is_pcie2 ? | 
 | 		BCMA_PCIE2_BAR0_WIN2 : BCMA_PCI_BAR0_WIN2; | 
 |  | 
 | 	pci_write_config_dword(core->bus->host_pci, BCMA_PCI_BAR0_WIN, | 
 | 			       core->addr); | 
 | 	pci_write_config_dword(core->bus->host_pci, win2, core->wrap); | 
 | 	core->bus->mapped_core = core; | 
 | 	bcma_debug(core->bus, "Switched to core: 0x%X\n", core->id.id); | 
 | } | 
 |  | 
 | /* Provides access to the requested core. Returns base offset that has to be | 
 |  * used. It makes use of fixed windows when possible. */ | 
 | static u16 bcma_host_pci_provide_access_to_core(struct bcma_device *core) | 
 | { | 
 | 	switch (core->id.id) { | 
 | 	case BCMA_CORE_CHIPCOMMON: | 
 | 		return 3 * BCMA_CORE_SIZE; | 
 | 	case BCMA_CORE_PCIE: | 
 | 		return 2 * BCMA_CORE_SIZE; | 
 | 	} | 
 |  | 
 | 	if (core->bus->mapped_core != core) | 
 | 		bcma_host_pci_switch_core(core); | 
 | 	return 0; | 
 | } | 
 |  | 
 | static u8 bcma_host_pci_read8(struct bcma_device *core, u16 offset) | 
 | { | 
 | 	offset += bcma_host_pci_provide_access_to_core(core); | 
 | 	return ioread8(core->bus->mmio + offset); | 
 | } | 
 |  | 
 | static u16 bcma_host_pci_read16(struct bcma_device *core, u16 offset) | 
 | { | 
 | 	offset += bcma_host_pci_provide_access_to_core(core); | 
 | 	return ioread16(core->bus->mmio + offset); | 
 | } | 
 |  | 
 | static u32 bcma_host_pci_read32(struct bcma_device *core, u16 offset) | 
 | { | 
 | 	offset += bcma_host_pci_provide_access_to_core(core); | 
 | 	return ioread32(core->bus->mmio + offset); | 
 | } | 
 |  | 
 | static void bcma_host_pci_write8(struct bcma_device *core, u16 offset, | 
 | 				 u8 value) | 
 | { | 
 | 	offset += bcma_host_pci_provide_access_to_core(core); | 
 | 	iowrite8(value, core->bus->mmio + offset); | 
 | } | 
 |  | 
 | static void bcma_host_pci_write16(struct bcma_device *core, u16 offset, | 
 | 				 u16 value) | 
 | { | 
 | 	offset += bcma_host_pci_provide_access_to_core(core); | 
 | 	iowrite16(value, core->bus->mmio + offset); | 
 | } | 
 |  | 
 | static void bcma_host_pci_write32(struct bcma_device *core, u16 offset, | 
 | 				 u32 value) | 
 | { | 
 | 	offset += bcma_host_pci_provide_access_to_core(core); | 
 | 	iowrite32(value, core->bus->mmio + offset); | 
 | } | 
 |  | 
 | #ifdef CONFIG_BCMA_BLOCKIO | 
 | static void bcma_host_pci_block_read(struct bcma_device *core, void *buffer, | 
 | 				     size_t count, u16 offset, u8 reg_width) | 
 | { | 
 | 	void __iomem *addr = core->bus->mmio + offset; | 
 | 	if (core->bus->mapped_core != core) | 
 | 		bcma_host_pci_switch_core(core); | 
 | 	switch (reg_width) { | 
 | 	case sizeof(u8): | 
 | 		ioread8_rep(addr, buffer, count); | 
 | 		break; | 
 | 	case sizeof(u16): | 
 | 		WARN_ON(count & 1); | 
 | 		ioread16_rep(addr, buffer, count >> 1); | 
 | 		break; | 
 | 	case sizeof(u32): | 
 | 		WARN_ON(count & 3); | 
 | 		ioread32_rep(addr, buffer, count >> 2); | 
 | 		break; | 
 | 	default: | 
 | 		WARN_ON(1); | 
 | 	} | 
 | } | 
 |  | 
 | static void bcma_host_pci_block_write(struct bcma_device *core, | 
 | 				      const void *buffer, size_t count, | 
 | 				      u16 offset, u8 reg_width) | 
 | { | 
 | 	void __iomem *addr = core->bus->mmio + offset; | 
 | 	if (core->bus->mapped_core != core) | 
 | 		bcma_host_pci_switch_core(core); | 
 | 	switch (reg_width) { | 
 | 	case sizeof(u8): | 
 | 		iowrite8_rep(addr, buffer, count); | 
 | 		break; | 
 | 	case sizeof(u16): | 
 | 		WARN_ON(count & 1); | 
 | 		iowrite16_rep(addr, buffer, count >> 1); | 
 | 		break; | 
 | 	case sizeof(u32): | 
 | 		WARN_ON(count & 3); | 
 | 		iowrite32_rep(addr, buffer, count >> 2); | 
 | 		break; | 
 | 	default: | 
 | 		WARN_ON(1); | 
 | 	} | 
 | } | 
 | #endif | 
 |  | 
 | static u32 bcma_host_pci_aread32(struct bcma_device *core, u16 offset) | 
 | { | 
 | 	if (core->bus->mapped_core != core) | 
 | 		bcma_host_pci_switch_core(core); | 
 | 	return ioread32(core->bus->mmio + (1 * BCMA_CORE_SIZE) + offset); | 
 | } | 
 |  | 
 | static void bcma_host_pci_awrite32(struct bcma_device *core, u16 offset, | 
 | 				  u32 value) | 
 | { | 
 | 	if (core->bus->mapped_core != core) | 
 | 		bcma_host_pci_switch_core(core); | 
 | 	iowrite32(value, core->bus->mmio + (1 * BCMA_CORE_SIZE) + offset); | 
 | } | 
 |  | 
 | static const struct bcma_host_ops bcma_host_pci_ops = { | 
 | 	.read8		= bcma_host_pci_read8, | 
 | 	.read16		= bcma_host_pci_read16, | 
 | 	.read32		= bcma_host_pci_read32, | 
 | 	.write8		= bcma_host_pci_write8, | 
 | 	.write16	= bcma_host_pci_write16, | 
 | 	.write32	= bcma_host_pci_write32, | 
 | #ifdef CONFIG_BCMA_BLOCKIO | 
 | 	.block_read	= bcma_host_pci_block_read, | 
 | 	.block_write	= bcma_host_pci_block_write, | 
 | #endif | 
 | 	.aread32	= bcma_host_pci_aread32, | 
 | 	.awrite32	= bcma_host_pci_awrite32, | 
 | }; | 
 |  | 
 | static int bcma_host_pci_probe(struct pci_dev *dev, | 
 | 			       const struct pci_device_id *id) | 
 | { | 
 | 	struct bcma_bus *bus; | 
 | 	int err = -ENOMEM; | 
 | 	const char *name; | 
 | 	u32 val; | 
 |  | 
 | 	/* Alloc */ | 
 | 	bus = kzalloc(sizeof(*bus), GFP_KERNEL); | 
 | 	if (!bus) | 
 | 		goto out; | 
 |  | 
 | 	/* Basic PCI configuration */ | 
 | 	err = pci_enable_device(dev); | 
 | 	if (err) | 
 | 		goto err_kfree_bus; | 
 |  | 
 | 	name = dev_name(&dev->dev); | 
 | 	if (dev->driver && dev->driver->name) | 
 | 		name = dev->driver->name; | 
 | 	err = pci_request_regions(dev, name); | 
 | 	if (err) | 
 | 		goto err_pci_disable; | 
 | 	pci_set_master(dev); | 
 |  | 
 | 	/* Disable the RETRY_TIMEOUT register (0x41) to keep | 
 | 	 * PCI Tx retries from interfering with C3 CPU state */ | 
 | 	pci_read_config_dword(dev, 0x40, &val); | 
 | 	if ((val & 0x0000ff00) != 0) | 
 | 		pci_write_config_dword(dev, 0x40, val & 0xffff00ff); | 
 |  | 
 | 	/* SSB needed additional powering up, do we have any AMBA PCI cards? */ | 
 | 	if (!pci_is_pcie(dev)) { | 
 | 		bcma_err(bus, "PCI card detected, they are not supported.\n"); | 
 | 		err = -ENXIO; | 
 | 		goto err_pci_release_regions; | 
 | 	} | 
 |  | 
 | 	bus->dev = &dev->dev; | 
 |  | 
 | 	/* Map MMIO */ | 
 | 	err = -ENOMEM; | 
 | 	bus->mmio = pci_iomap(dev, 0, ~0UL); | 
 | 	if (!bus->mmio) | 
 | 		goto err_pci_release_regions; | 
 |  | 
 | 	/* Host specific */ | 
 | 	bus->host_pci = dev; | 
 | 	bus->hosttype = BCMA_HOSTTYPE_PCI; | 
 | 	bus->ops = &bcma_host_pci_ops; | 
 |  | 
 | 	bus->boardinfo.vendor = bus->host_pci->subsystem_vendor; | 
 | 	bus->boardinfo.type = bus->host_pci->subsystem_device; | 
 |  | 
 | 	/* Initialize struct, detect chip */ | 
 | 	bcma_init_bus(bus); | 
 |  | 
 | 	/* Scan bus to find out generation of PCIe core */ | 
 | 	err = bcma_bus_scan(bus); | 
 | 	if (err) | 
 | 		goto err_pci_unmap_mmio; | 
 |  | 
 | 	if (bcma_find_core(bus, BCMA_CORE_PCIE2)) | 
 | 		bus->host_is_pcie2 = true; | 
 |  | 
 | 	/* Register */ | 
 | 	err = bcma_bus_register(bus); | 
 | 	if (err) | 
 | 		goto err_unregister_cores; | 
 |  | 
 | 	pci_set_drvdata(dev, bus); | 
 |  | 
 | out: | 
 | 	return err; | 
 |  | 
 | err_unregister_cores: | 
 | 	bcma_unregister_cores(bus); | 
 | err_pci_unmap_mmio: | 
 | 	pci_iounmap(dev, bus->mmio); | 
 | err_pci_release_regions: | 
 | 	pci_release_regions(dev); | 
 | err_pci_disable: | 
 | 	pci_disable_device(dev); | 
 | err_kfree_bus: | 
 | 	kfree(bus); | 
 | 	return err; | 
 | } | 
 |  | 
 | static void bcma_host_pci_remove(struct pci_dev *dev) | 
 | { | 
 | 	struct bcma_bus *bus = pci_get_drvdata(dev); | 
 |  | 
 | 	bcma_bus_unregister(bus); | 
 | 	pci_iounmap(dev, bus->mmio); | 
 | 	pci_release_regions(dev); | 
 | 	pci_disable_device(dev); | 
 | 	kfree(bus); | 
 | } | 
 |  | 
 | #ifdef CONFIG_PM_SLEEP | 
 | static int bcma_host_pci_suspend(struct device *dev) | 
 | { | 
 | 	struct pci_dev *pdev = to_pci_dev(dev); | 
 | 	struct bcma_bus *bus = pci_get_drvdata(pdev); | 
 |  | 
 | 	bus->mapped_core = NULL; | 
 |  | 
 | 	return bcma_bus_suspend(bus); | 
 | } | 
 |  | 
 | static int bcma_host_pci_resume(struct device *dev) | 
 | { | 
 | 	struct pci_dev *pdev = to_pci_dev(dev); | 
 | 	struct bcma_bus *bus = pci_get_drvdata(pdev); | 
 |  | 
 | 	return bcma_bus_resume(bus); | 
 | } | 
 |  | 
 | static SIMPLE_DEV_PM_OPS(bcma_pm_ops, bcma_host_pci_suspend, | 
 | 			 bcma_host_pci_resume); | 
 | #define BCMA_PM_OPS	(&bcma_pm_ops) | 
 |  | 
 | #else /* CONFIG_PM_SLEEP */ | 
 |  | 
 | #define BCMA_PM_OPS     NULL | 
 |  | 
 | #endif /* CONFIG_PM_SLEEP */ | 
 |  | 
 | static const struct pci_device_id bcma_pci_bridge_tbl[] = { | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x0576) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4313) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43224) },	/* 0xa8d8 */ | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4331) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4353) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4357) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4358) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4359) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4360) }, | 
 | 	{ PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_DELL, 0x0016) }, | 
 | 	{ PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_DELL, 0x0018) }, | 
 | 	{ PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_FOXCONN, 0xe092) }, | 
 | 	{ PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_HP, 0x804a) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43a0) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43a9) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43aa) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43b1) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4727) }, | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43227) },	/* 0xa8db, BCM43217 (sic!) */ | 
 | 	{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43228) },	/* 0xa8dc */ | 
 | 	{ 0, }, | 
 | }; | 
 | MODULE_DEVICE_TABLE(pci, bcma_pci_bridge_tbl); | 
 |  | 
 | static struct pci_driver bcma_pci_bridge_driver = { | 
 | 	.name = "bcma-pci-bridge", | 
 | 	.id_table = bcma_pci_bridge_tbl, | 
 | 	.probe = bcma_host_pci_probe, | 
 | 	.remove = bcma_host_pci_remove, | 
 | 	.driver.pm = BCMA_PM_OPS, | 
 | }; | 
 |  | 
 | int __init bcma_host_pci_init(void) | 
 | { | 
 | 	return pci_register_driver(&bcma_pci_bridge_driver); | 
 | } | 
 |  | 
 | void __exit bcma_host_pci_exit(void) | 
 | { | 
 | 	pci_unregister_driver(&bcma_pci_bridge_driver); | 
 | } | 
 |  | 
 | /************************************************** | 
 |  * Runtime ops for drivers. | 
 |  **************************************************/ | 
 |  | 
 | /* See also pcicore_up */ | 
 | void bcma_host_pci_up(struct bcma_bus *bus) | 
 | { | 
 | 	if (bus->hosttype != BCMA_HOSTTYPE_PCI) | 
 | 		return; | 
 |  | 
 | 	if (bus->host_is_pcie2) | 
 | 		bcma_core_pcie2_up(&bus->drv_pcie2); | 
 | 	else | 
 | 		bcma_core_pci_up(&bus->drv_pci[0]); | 
 | } | 
 | EXPORT_SYMBOL_GPL(bcma_host_pci_up); | 
 |  | 
 | /* See also pcicore_down */ | 
 | void bcma_host_pci_down(struct bcma_bus *bus) | 
 | { | 
 | 	if (bus->hosttype != BCMA_HOSTTYPE_PCI) | 
 | 		return; | 
 |  | 
 | 	if (!bus->host_is_pcie2) | 
 | 		bcma_core_pci_down(&bus->drv_pci[0]); | 
 | } | 
 | EXPORT_SYMBOL_GPL(bcma_host_pci_down); | 
 |  | 
 | /* See also si_pci_setup */ | 
 | int bcma_host_pci_irq_ctl(struct bcma_bus *bus, struct bcma_device *core, | 
 | 			  bool enable) | 
 | { | 
 | 	struct pci_dev *pdev; | 
 | 	u32 coremask, tmp; | 
 | 	int err = 0; | 
 |  | 
 | 	if (bus->hosttype != BCMA_HOSTTYPE_PCI) { | 
 | 		/* This bcma device is not on a PCI host-bus. So the IRQs are | 
 | 		 * not routed through the PCI core. | 
 | 		 * So we must not enable routing through the PCI core. */ | 
 | 		goto out; | 
 | 	} | 
 |  | 
 | 	pdev = bus->host_pci; | 
 |  | 
 | 	err = pci_read_config_dword(pdev, BCMA_PCI_IRQMASK, &tmp); | 
 | 	if (err) | 
 | 		goto out; | 
 |  | 
 | 	coremask = BIT(core->core_index) << 8; | 
 | 	if (enable) | 
 | 		tmp |= coremask; | 
 | 	else | 
 | 		tmp &= ~coremask; | 
 |  | 
 | 	err = pci_write_config_dword(pdev, BCMA_PCI_IRQMASK, tmp); | 
 |  | 
 | out: | 
 | 	return err; | 
 | } | 
 | EXPORT_SYMBOL_GPL(bcma_host_pci_irq_ctl); |