|  | /* | 
|  | * Copyright 2014 IBM Corp. | 
|  | * | 
|  | * This program is free software; you can redistribute it and/or | 
|  | * modify it under the terms of the GNU General Public License | 
|  | * as published by the Free Software Foundation; either version | 
|  | * 2 of the License, or (at your option) any later version. | 
|  | */ | 
|  |  | 
|  | #include <linux/module.h> | 
|  | #include <linux/rcupdate.h> | 
|  | #include <asm/errno.h> | 
|  | #include <misc/cxl-base.h> | 
|  | #include <linux/of_platform.h> | 
|  | #include "cxl.h" | 
|  |  | 
|  | /* protected by rcu */ | 
|  | static struct cxl_calls *cxl_calls; | 
|  |  | 
|  | atomic_t cxl_use_count = ATOMIC_INIT(0); | 
|  | EXPORT_SYMBOL(cxl_use_count); | 
|  |  | 
|  | #ifdef CONFIG_CXL_MODULE | 
|  |  | 
|  | static inline struct cxl_calls *cxl_calls_get(void) | 
|  | { | 
|  | struct cxl_calls *calls = NULL; | 
|  |  | 
|  | rcu_read_lock(); | 
|  | calls = rcu_dereference(cxl_calls); | 
|  | if (calls && !try_module_get(calls->owner)) | 
|  | calls = NULL; | 
|  | rcu_read_unlock(); | 
|  |  | 
|  | return calls; | 
|  | } | 
|  |  | 
|  | static inline void cxl_calls_put(struct cxl_calls *calls) | 
|  | { | 
|  | BUG_ON(calls != cxl_calls); | 
|  |  | 
|  | /* we don't need to rcu this, as we hold a reference to the module */ | 
|  | module_put(cxl_calls->owner); | 
|  | } | 
|  |  | 
|  | #else /* !defined CONFIG_CXL_MODULE */ | 
|  |  | 
|  | static inline struct cxl_calls *cxl_calls_get(void) | 
|  | { | 
|  | return cxl_calls; | 
|  | } | 
|  |  | 
|  | static inline void cxl_calls_put(struct cxl_calls *calls) { } | 
|  |  | 
|  | #endif /* CONFIG_CXL_MODULE */ | 
|  |  | 
|  | /* AFU refcount management */ | 
|  | struct cxl_afu *cxl_afu_get(struct cxl_afu *afu) | 
|  | { | 
|  | return (get_device(&afu->dev) == NULL) ? NULL : afu; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_afu_get); | 
|  |  | 
|  | void cxl_afu_put(struct cxl_afu *afu) | 
|  | { | 
|  | put_device(&afu->dev); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_afu_put); | 
|  |  | 
|  | void cxl_slbia(struct mm_struct *mm) | 
|  | { | 
|  | struct cxl_calls *calls; | 
|  |  | 
|  | calls = cxl_calls_get(); | 
|  | if (!calls) | 
|  | return; | 
|  |  | 
|  | if (cxl_ctx_in_use()) | 
|  | calls->cxl_slbia(mm); | 
|  |  | 
|  | cxl_calls_put(calls); | 
|  | } | 
|  |  | 
|  | int register_cxl_calls(struct cxl_calls *calls) | 
|  | { | 
|  | if (cxl_calls) | 
|  | return -EBUSY; | 
|  |  | 
|  | rcu_assign_pointer(cxl_calls, calls); | 
|  | return 0; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(register_cxl_calls); | 
|  |  | 
|  | void unregister_cxl_calls(struct cxl_calls *calls) | 
|  | { | 
|  | BUG_ON(cxl_calls->owner != calls->owner); | 
|  | RCU_INIT_POINTER(cxl_calls, NULL); | 
|  | synchronize_rcu(); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(unregister_cxl_calls); | 
|  |  | 
|  | int cxl_update_properties(struct device_node *dn, | 
|  | struct property *new_prop) | 
|  | { | 
|  | return of_update_property(dn, new_prop); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_update_properties); | 
|  |  | 
|  | /* | 
|  | * API calls into the driver that may be called from the PHB code and must be | 
|  | * built in. | 
|  | */ | 
|  | bool cxl_pci_associate_default_context(struct pci_dev *dev, struct cxl_afu *afu) | 
|  | { | 
|  | bool ret; | 
|  | struct cxl_calls *calls; | 
|  |  | 
|  | calls = cxl_calls_get(); | 
|  | if (!calls) | 
|  | return false; | 
|  |  | 
|  | ret = calls->cxl_pci_associate_default_context(dev, afu); | 
|  |  | 
|  | cxl_calls_put(calls); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_pci_associate_default_context); | 
|  |  | 
|  | void cxl_pci_disable_device(struct pci_dev *dev) | 
|  | { | 
|  | struct cxl_calls *calls; | 
|  |  | 
|  | calls = cxl_calls_get(); | 
|  | if (!calls) | 
|  | return; | 
|  |  | 
|  | calls->cxl_pci_disable_device(dev); | 
|  |  | 
|  | cxl_calls_put(calls); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_pci_disable_device); | 
|  |  | 
|  | int cxl_next_msi_hwirq(struct pci_dev *pdev, struct cxl_context **ctx, int *afu_irq) | 
|  | { | 
|  | int ret; | 
|  | struct cxl_calls *calls; | 
|  |  | 
|  | calls = cxl_calls_get(); | 
|  | if (!calls) | 
|  | return -EBUSY; | 
|  |  | 
|  | ret = calls->cxl_next_msi_hwirq(pdev, ctx, afu_irq); | 
|  |  | 
|  | cxl_calls_put(calls); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_next_msi_hwirq); | 
|  |  | 
|  | int cxl_cx4_setup_msi_irqs(struct pci_dev *pdev, int nvec, int type) | 
|  | { | 
|  | int ret; | 
|  | struct cxl_calls *calls; | 
|  |  | 
|  | calls = cxl_calls_get(); | 
|  | if (!calls) | 
|  | return false; | 
|  |  | 
|  | ret = calls->cxl_cx4_setup_msi_irqs(pdev, nvec, type); | 
|  |  | 
|  | cxl_calls_put(calls); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_cx4_setup_msi_irqs); | 
|  |  | 
|  | void cxl_cx4_teardown_msi_irqs(struct pci_dev *pdev) | 
|  | { | 
|  | struct cxl_calls *calls; | 
|  |  | 
|  | calls = cxl_calls_get(); | 
|  | if (!calls) | 
|  | return; | 
|  |  | 
|  | calls->cxl_cx4_teardown_msi_irqs(pdev); | 
|  |  | 
|  | cxl_calls_put(calls); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(cxl_cx4_teardown_msi_irqs); | 
|  |  | 
|  | static int __init cxl_base_init(void) | 
|  | { | 
|  | struct device_node *np; | 
|  | struct platform_device *dev; | 
|  | int count = 0; | 
|  |  | 
|  | /* | 
|  | * Scan for compatible devices in guest only | 
|  | */ | 
|  | if (cpu_has_feature(CPU_FTR_HVMODE)) | 
|  | return 0; | 
|  |  | 
|  | for_each_compatible_node(np, NULL, "ibm,coherent-platform-facility") { | 
|  | dev = of_platform_device_create(np, NULL, NULL); | 
|  | if (dev) | 
|  | count++; | 
|  | } | 
|  | pr_devel("Found %d cxl device(s)\n", count); | 
|  | return 0; | 
|  | } | 
|  | device_initcall(cxl_base_init); |