at v5.10 15 kB view raw
1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef LINUX_MSI_H 3#define LINUX_MSI_H 4 5#include <linux/kobject.h> 6#include <linux/list.h> 7 8struct msi_msg { 9 u32 address_lo; /* low 32 bits of msi message address */ 10 u32 address_hi; /* high 32 bits of msi message address */ 11 u32 data; /* 16 bits of msi message data */ 12}; 13 14extern int pci_msi_ignore_mask; 15/* Helper functions */ 16struct irq_data; 17struct msi_desc; 18struct pci_dev; 19struct platform_msi_priv_data; 20void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg); 21#ifdef CONFIG_GENERIC_MSI_IRQ 22void get_cached_msi_msg(unsigned int irq, struct msi_msg *msg); 23#else 24static inline void get_cached_msi_msg(unsigned int irq, struct msi_msg *msg) 25{ 26} 27#endif 28 29typedef void (*irq_write_msi_msg_t)(struct msi_desc *desc, 30 struct msi_msg *msg); 31 32/** 33 * platform_msi_desc - Platform device specific msi descriptor data 34 * @msi_priv_data: Pointer to platform private data 35 * @msi_index: The index of the MSI descriptor for multi MSI 36 */ 37struct platform_msi_desc { 38 struct platform_msi_priv_data *msi_priv_data; 39 u16 msi_index; 40}; 41 42/** 43 * fsl_mc_msi_desc - FSL-MC device specific msi descriptor data 44 * @msi_index: The index of the MSI descriptor 45 */ 46struct fsl_mc_msi_desc { 47 u16 msi_index; 48}; 49 50/** 51 * ti_sci_inta_msi_desc - TISCI based INTA specific msi descriptor data 52 * @dev_index: TISCI device index 53 */ 54struct ti_sci_inta_msi_desc { 55 u16 dev_index; 56}; 57 58/** 59 * struct msi_desc - Descriptor structure for MSI based interrupts 60 * @list: List head for management 61 * @irq: The base interrupt number 62 * @nvec_used: The number of vectors used 63 * @dev: Pointer to the device which uses this descriptor 64 * @msg: The last set MSI message cached for reuse 65 * @affinity: Optional pointer to a cpu affinity mask for this descriptor 66 * 67 * @write_msi_msg: Callback that may be called when the MSI message 68 * address or data changes 69 * @write_msi_msg_data: Data parameter for the callback. 70 * 71 * @masked: [PCI MSI/X] Mask bits 72 * @is_msix: [PCI MSI/X] True if MSI-X 73 * @multiple: [PCI MSI/X] log2 num of messages allocated 74 * @multi_cap: [PCI MSI/X] log2 num of messages supported 75 * @maskbit: [PCI MSI/X] Mask-Pending bit supported? 76 * @is_64: [PCI MSI/X] Address size: 0=32bit 1=64bit 77 * @entry_nr: [PCI MSI/X] Entry which is described by this descriptor 78 * @default_irq:[PCI MSI/X] The default pre-assigned non-MSI irq 79 * @mask_pos: [PCI MSI] Mask register position 80 * @mask_base: [PCI MSI-X] Mask register base address 81 * @platform: [platform] Platform device specific msi descriptor data 82 * @fsl_mc: [fsl-mc] FSL MC device specific msi descriptor data 83 * @inta: [INTA] TISCI based INTA specific msi descriptor data 84 */ 85struct msi_desc { 86 /* Shared device/bus type independent data */ 87 struct list_head list; 88 unsigned int irq; 89 unsigned int nvec_used; 90 struct device *dev; 91 struct msi_msg msg; 92 struct irq_affinity_desc *affinity; 93#ifdef CONFIG_IRQ_MSI_IOMMU 94 const void *iommu_cookie; 95#endif 96 97 void (*write_msi_msg)(struct msi_desc *entry, void *data); 98 void *write_msi_msg_data; 99 100 union { 101 /* PCI MSI/X specific data */ 102 struct { 103 u32 masked; 104 struct { 105 u8 is_msix : 1; 106 u8 multiple : 3; 107 u8 multi_cap : 3; 108 u8 maskbit : 1; 109 u8 is_64 : 1; 110 u8 is_virtual : 1; 111 u16 entry_nr; 112 unsigned default_irq; 113 } msi_attrib; 114 union { 115 u8 mask_pos; 116 void __iomem *mask_base; 117 }; 118 }; 119 120 /* 121 * Non PCI variants add their data structure here. New 122 * entries need to use a named structure. We want 123 * proper name spaces for this. The PCI part is 124 * anonymous for now as it would require an immediate 125 * tree wide cleanup. 126 */ 127 struct platform_msi_desc platform; 128 struct fsl_mc_msi_desc fsl_mc; 129 struct ti_sci_inta_msi_desc inta; 130 }; 131}; 132 133/* Helpers to hide struct msi_desc implementation details */ 134#define msi_desc_to_dev(desc) ((desc)->dev) 135#define dev_to_msi_list(dev) (&(dev)->msi_list) 136#define first_msi_entry(dev) \ 137 list_first_entry(dev_to_msi_list((dev)), struct msi_desc, list) 138#define for_each_msi_entry(desc, dev) \ 139 list_for_each_entry((desc), dev_to_msi_list((dev)), list) 140#define for_each_msi_entry_safe(desc, tmp, dev) \ 141 list_for_each_entry_safe((desc), (tmp), dev_to_msi_list((dev)), list) 142 143#ifdef CONFIG_IRQ_MSI_IOMMU 144static inline const void *msi_desc_get_iommu_cookie(struct msi_desc *desc) 145{ 146 return desc->iommu_cookie; 147} 148 149static inline void msi_desc_set_iommu_cookie(struct msi_desc *desc, 150 const void *iommu_cookie) 151{ 152 desc->iommu_cookie = iommu_cookie; 153} 154#else 155static inline const void *msi_desc_get_iommu_cookie(struct msi_desc *desc) 156{ 157 return NULL; 158} 159 160static inline void msi_desc_set_iommu_cookie(struct msi_desc *desc, 161 const void *iommu_cookie) 162{ 163} 164#endif 165 166#ifdef CONFIG_PCI_MSI 167#define first_pci_msi_entry(pdev) first_msi_entry(&(pdev)->dev) 168#define for_each_pci_msi_entry(desc, pdev) \ 169 for_each_msi_entry((desc), &(pdev)->dev) 170 171struct pci_dev *msi_desc_to_pci_dev(struct msi_desc *desc); 172void *msi_desc_to_pci_sysdata(struct msi_desc *desc); 173void pci_write_msi_msg(unsigned int irq, struct msi_msg *msg); 174#else /* CONFIG_PCI_MSI */ 175static inline void *msi_desc_to_pci_sysdata(struct msi_desc *desc) 176{ 177 return NULL; 178} 179static inline void pci_write_msi_msg(unsigned int irq, struct msi_msg *msg) 180{ 181} 182#endif /* CONFIG_PCI_MSI */ 183 184struct msi_desc *alloc_msi_entry(struct device *dev, int nvec, 185 const struct irq_affinity_desc *affinity); 186void free_msi_entry(struct msi_desc *entry); 187void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg); 188void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg); 189 190u32 __pci_msix_desc_mask_irq(struct msi_desc *desc, u32 flag); 191u32 __pci_msi_desc_mask_irq(struct msi_desc *desc, u32 mask, u32 flag); 192void pci_msi_mask_irq(struct irq_data *data); 193void pci_msi_unmask_irq(struct irq_data *data); 194 195/* 196 * The arch hooks to setup up msi irqs. Default functions are implemented 197 * as weak symbols so that they /can/ be overriden by architecture specific 198 * code if needed. These hooks must be enabled by the architecture or by 199 * drivers which depend on them via msi_controller based MSI handling. 200 * 201 * If CONFIG_PCI_MSI_ARCH_FALLBACKS is not selected they are replaced by 202 * stubs with warnings. 203 */ 204#ifdef CONFIG_PCI_MSI_ARCH_FALLBACKS 205int arch_setup_msi_irq(struct pci_dev *dev, struct msi_desc *desc); 206void arch_teardown_msi_irq(unsigned int irq); 207int arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type); 208void arch_teardown_msi_irqs(struct pci_dev *dev); 209void default_teardown_msi_irqs(struct pci_dev *dev); 210#else 211static inline int arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type) 212{ 213 WARN_ON_ONCE(1); 214 return -ENODEV; 215} 216 217static inline void arch_teardown_msi_irqs(struct pci_dev *dev) 218{ 219 WARN_ON_ONCE(1); 220} 221#endif 222 223/* 224 * The restore hooks are still available as they are useful even 225 * for fully irq domain based setups. Courtesy to XEN/X86. 226 */ 227void arch_restore_msi_irqs(struct pci_dev *dev); 228void default_restore_msi_irqs(struct pci_dev *dev); 229 230struct msi_controller { 231 struct module *owner; 232 struct device *dev; 233 struct device_node *of_node; 234 struct list_head list; 235 236 int (*setup_irq)(struct msi_controller *chip, struct pci_dev *dev, 237 struct msi_desc *desc); 238 int (*setup_irqs)(struct msi_controller *chip, struct pci_dev *dev, 239 int nvec, int type); 240 void (*teardown_irq)(struct msi_controller *chip, unsigned int irq); 241}; 242 243#ifdef CONFIG_GENERIC_MSI_IRQ_DOMAIN 244 245#include <linux/irqhandler.h> 246#include <asm/msi.h> 247 248struct irq_domain; 249struct irq_domain_ops; 250struct irq_chip; 251struct device_node; 252struct fwnode_handle; 253struct msi_domain_info; 254 255/** 256 * struct msi_domain_ops - MSI interrupt domain callbacks 257 * @get_hwirq: Retrieve the resulting hw irq number 258 * @msi_init: Domain specific init function for MSI interrupts 259 * @msi_free: Domain specific function to free a MSI interrupts 260 * @msi_check: Callback for verification of the domain/info/dev data 261 * @msi_prepare: Prepare the allocation of the interrupts in the domain 262 * @msi_finish: Optional callback to finalize the allocation 263 * @set_desc: Set the msi descriptor for an interrupt 264 * @handle_error: Optional error handler if the allocation fails 265 * @domain_alloc_irqs: Optional function to override the default allocation 266 * function. 267 * @domain_free_irqs: Optional function to override the default free 268 * function. 269 * 270 * @get_hwirq, @msi_init and @msi_free are callbacks used by 271 * msi_create_irq_domain() and related interfaces 272 * 273 * @msi_check, @msi_prepare, @msi_finish, @set_desc and @handle_error 274 * are callbacks used by msi_domain_alloc_irqs() and related 275 * interfaces which are based on msi_desc. 276 * 277 * @domain_alloc_irqs, @domain_free_irqs can be used to override the 278 * default allocation/free functions (__msi_domain_alloc/free_irqs). This 279 * is initially for a wrapper around XENs seperate MSI universe which can't 280 * be wrapped into the regular irq domains concepts by mere mortals. This 281 * allows to universally use msi_domain_alloc/free_irqs without having to 282 * special case XEN all over the place. 283 * 284 * Contrary to other operations @domain_alloc_irqs and @domain_free_irqs 285 * are set to the default implementation if NULL and even when 286 * MSI_FLAG_USE_DEF_DOM_OPS is not set to avoid breaking existing users and 287 * because these callbacks are obviously mandatory. 288 * 289 * This is NOT meant to be abused, but it can be useful to build wrappers 290 * for specialized MSI irq domains which need extra work before and after 291 * calling __msi_domain_alloc_irqs()/__msi_domain_free_irqs(). 292 */ 293struct msi_domain_ops { 294 irq_hw_number_t (*get_hwirq)(struct msi_domain_info *info, 295 msi_alloc_info_t *arg); 296 int (*msi_init)(struct irq_domain *domain, 297 struct msi_domain_info *info, 298 unsigned int virq, irq_hw_number_t hwirq, 299 msi_alloc_info_t *arg); 300 void (*msi_free)(struct irq_domain *domain, 301 struct msi_domain_info *info, 302 unsigned int virq); 303 int (*msi_check)(struct irq_domain *domain, 304 struct msi_domain_info *info, 305 struct device *dev); 306 int (*msi_prepare)(struct irq_domain *domain, 307 struct device *dev, int nvec, 308 msi_alloc_info_t *arg); 309 void (*msi_finish)(msi_alloc_info_t *arg, int retval); 310 void (*set_desc)(msi_alloc_info_t *arg, 311 struct msi_desc *desc); 312 int (*handle_error)(struct irq_domain *domain, 313 struct msi_desc *desc, int error); 314 int (*domain_alloc_irqs)(struct irq_domain *domain, 315 struct device *dev, int nvec); 316 void (*domain_free_irqs)(struct irq_domain *domain, 317 struct device *dev); 318}; 319 320/** 321 * struct msi_domain_info - MSI interrupt domain data 322 * @flags: Flags to decribe features and capabilities 323 * @ops: The callback data structure 324 * @chip: Optional: associated interrupt chip 325 * @chip_data: Optional: associated interrupt chip data 326 * @handler: Optional: associated interrupt flow handler 327 * @handler_data: Optional: associated interrupt flow handler data 328 * @handler_name: Optional: associated interrupt flow handler name 329 * @data: Optional: domain specific data 330 */ 331struct msi_domain_info { 332 u32 flags; 333 struct msi_domain_ops *ops; 334 struct irq_chip *chip; 335 void *chip_data; 336 irq_flow_handler_t handler; 337 void *handler_data; 338 const char *handler_name; 339 void *data; 340}; 341 342/* Flags for msi_domain_info */ 343enum { 344 /* 345 * Init non implemented ops callbacks with default MSI domain 346 * callbacks. 347 */ 348 MSI_FLAG_USE_DEF_DOM_OPS = (1 << 0), 349 /* 350 * Init non implemented chip callbacks with default MSI chip 351 * callbacks. 352 */ 353 MSI_FLAG_USE_DEF_CHIP_OPS = (1 << 1), 354 /* Support multiple PCI MSI interrupts */ 355 MSI_FLAG_MULTI_PCI_MSI = (1 << 2), 356 /* Support PCI MSIX interrupts */ 357 MSI_FLAG_PCI_MSIX = (1 << 3), 358 /* Needs early activate, required for PCI */ 359 MSI_FLAG_ACTIVATE_EARLY = (1 << 4), 360 /* 361 * Must reactivate when irq is started even when 362 * MSI_FLAG_ACTIVATE_EARLY has been set. 363 */ 364 MSI_FLAG_MUST_REACTIVATE = (1 << 5), 365 /* Is level-triggered capable, using two messages */ 366 MSI_FLAG_LEVEL_CAPABLE = (1 << 6), 367}; 368 369int msi_domain_set_affinity(struct irq_data *data, const struct cpumask *mask, 370 bool force); 371 372struct irq_domain *msi_create_irq_domain(struct fwnode_handle *fwnode, 373 struct msi_domain_info *info, 374 struct irq_domain *parent); 375int __msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev, 376 int nvec); 377int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev, 378 int nvec); 379void __msi_domain_free_irqs(struct irq_domain *domain, struct device *dev); 380void msi_domain_free_irqs(struct irq_domain *domain, struct device *dev); 381struct msi_domain_info *msi_get_domain_info(struct irq_domain *domain); 382 383struct irq_domain *platform_msi_create_irq_domain(struct fwnode_handle *fwnode, 384 struct msi_domain_info *info, 385 struct irq_domain *parent); 386int platform_msi_domain_alloc_irqs(struct device *dev, unsigned int nvec, 387 irq_write_msi_msg_t write_msi_msg); 388void platform_msi_domain_free_irqs(struct device *dev); 389 390/* When an MSI domain is used as an intermediate domain */ 391int msi_domain_prepare_irqs(struct irq_domain *domain, struct device *dev, 392 int nvec, msi_alloc_info_t *args); 393int msi_domain_populate_irqs(struct irq_domain *domain, struct device *dev, 394 int virq, int nvec, msi_alloc_info_t *args); 395struct irq_domain * 396__platform_msi_create_device_domain(struct device *dev, 397 unsigned int nvec, 398 bool is_tree, 399 irq_write_msi_msg_t write_msi_msg, 400 const struct irq_domain_ops *ops, 401 void *host_data); 402 403#define platform_msi_create_device_domain(dev, nvec, write, ops, data) \ 404 __platform_msi_create_device_domain(dev, nvec, false, write, ops, data) 405#define platform_msi_create_device_tree_domain(dev, nvec, write, ops, data) \ 406 __platform_msi_create_device_domain(dev, nvec, true, write, ops, data) 407 408int platform_msi_domain_alloc(struct irq_domain *domain, unsigned int virq, 409 unsigned int nr_irqs); 410void platform_msi_domain_free(struct irq_domain *domain, unsigned int virq, 411 unsigned int nvec); 412void *platform_msi_get_host_data(struct irq_domain *domain); 413#endif /* CONFIG_GENERIC_MSI_IRQ_DOMAIN */ 414 415#ifdef CONFIG_PCI_MSI_IRQ_DOMAIN 416void pci_msi_domain_write_msg(struct irq_data *irq_data, struct msi_msg *msg); 417struct irq_domain *pci_msi_create_irq_domain(struct fwnode_handle *fwnode, 418 struct msi_domain_info *info, 419 struct irq_domain *parent); 420int pci_msi_domain_check_cap(struct irq_domain *domain, 421 struct msi_domain_info *info, struct device *dev); 422u32 pci_msi_domain_get_msi_rid(struct irq_domain *domain, struct pci_dev *pdev); 423struct irq_domain *pci_msi_get_device_domain(struct pci_dev *pdev); 424bool pci_dev_has_special_msi_domain(struct pci_dev *pdev); 425#else 426static inline struct irq_domain *pci_msi_get_device_domain(struct pci_dev *pdev) 427{ 428 return NULL; 429} 430#endif /* CONFIG_PCI_MSI_IRQ_DOMAIN */ 431 432#endif /* LINUX_MSI_H */