1 /* SPDX-License-Identifier: GPL-2.0 */
5 #include <linux/kobject.h>
6 #include <linux/list.h>
9 u32 address_lo; /* low 32 bits of msi message address */
10 u32 address_hi; /* high 32 bits of msi message address */
11 u32 data; /* 16 bits of msi message data */
14 extern int pci_msi_ignore_mask;
15 /* Helper functions */
19 struct platform_msi_priv_data;
20 void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg);
21 #ifdef CONFIG_GENERIC_MSI_IRQ
22 void get_cached_msi_msg(unsigned int irq, struct msi_msg *msg);
24 static inline void get_cached_msi_msg(unsigned int irq, struct msi_msg *msg)
29 typedef void (*irq_write_msi_msg_t)(struct msi_desc *desc,
33 * platform_msi_desc - Platform device specific msi descriptor data
34 * @msi_priv_data: Pointer to platform private data
35 * @msi_index: The index of the MSI descriptor for multi MSI
37 struct platform_msi_desc {
38 struct platform_msi_priv_data *msi_priv_data;
43 * fsl_mc_msi_desc - FSL-MC device specific msi descriptor data
44 * @msi_index: The index of the MSI descriptor
46 struct fsl_mc_msi_desc {
51 * ti_sci_inta_msi_desc - TISCI based INTA specific msi descriptor data
52 * @dev_index: TISCI device index
54 struct ti_sci_inta_msi_desc {
59 * struct msi_desc - Descriptor structure for MSI based interrupts
60 * @list: List head for management
61 * @irq: The base interrupt number
62 * @nvec_used: The number of vectors used
63 * @dev: Pointer to the device which uses this descriptor
64 * @msg: The last set MSI message cached for reuse
65 * @affinity: Optional pointer to a cpu affinity mask for this descriptor
67 * @write_msi_msg: Callback that may be called when the MSI message
68 * address or data changes
69 * @write_msi_msg_data: Data parameter for the callback.
71 * @masked: [PCI MSI/X] Mask bits
72 * @is_msix: [PCI MSI/X] True if MSI-X
73 * @multiple: [PCI MSI/X] log2 num of messages allocated
74 * @multi_cap: [PCI MSI/X] log2 num of messages supported
75 * @maskbit: [PCI MSI/X] Mask-Pending bit supported?
76 * @is_64: [PCI MSI/X] Address size: 0=32bit 1=64bit
77 * @entry_nr: [PCI MSI/X] Entry which is described by this descriptor
78 * @default_irq:[PCI MSI/X] The default pre-assigned non-MSI irq
79 * @mask_pos: [PCI MSI] Mask register position
80 * @mask_base: [PCI MSI-X] Mask register base address
81 * @platform: [platform] Platform device specific msi descriptor data
82 * @fsl_mc: [fsl-mc] FSL MC device specific msi descriptor data
83 * @inta: [INTA] TISCI based INTA specific msi descriptor data
86 /* Shared device/bus type independent data */
87 struct list_head list;
89 unsigned int nvec_used;
92 struct irq_affinity_desc *affinity;
93 #ifdef CONFIG_IRQ_MSI_IOMMU
94 const void *iommu_cookie;
97 void (*write_msi_msg)(struct msi_desc *entry, void *data);
98 void *write_msi_msg_data;
101 /* PCI MSI/X specific data */
112 unsigned default_irq;
116 void __iomem *mask_base;
121 * Non PCI variants add their data structure here. New
122 * entries need to use a named structure. We want
123 * proper name spaces for this. The PCI part is
124 * anonymous for now as it would require an immediate
127 struct platform_msi_desc platform;
128 struct fsl_mc_msi_desc fsl_mc;
129 struct ti_sci_inta_msi_desc inta;
133 /* Helpers to hide struct msi_desc implementation details */
134 #define msi_desc_to_dev(desc) ((desc)->dev)
135 #define dev_to_msi_list(dev) (&(dev)->msi_list)
136 #define first_msi_entry(dev) \
137 list_first_entry(dev_to_msi_list((dev)), struct msi_desc, list)
138 #define for_each_msi_entry(desc, dev) \
139 list_for_each_entry((desc), dev_to_msi_list((dev)), list)
140 #define for_each_msi_entry_safe(desc, tmp, dev) \
141 list_for_each_entry_safe((desc), (tmp), dev_to_msi_list((dev)), list)
143 #ifdef CONFIG_IRQ_MSI_IOMMU
144 static inline const void *msi_desc_get_iommu_cookie(struct msi_desc *desc)
146 return desc->iommu_cookie;
149 static inline void msi_desc_set_iommu_cookie(struct msi_desc *desc,
150 const void *iommu_cookie)
152 desc->iommu_cookie = iommu_cookie;
155 static inline const void *msi_desc_get_iommu_cookie(struct msi_desc *desc)
160 static inline void msi_desc_set_iommu_cookie(struct msi_desc *desc,
161 const void *iommu_cookie)
166 #ifdef CONFIG_PCI_MSI
167 #define first_pci_msi_entry(pdev) first_msi_entry(&(pdev)->dev)
168 #define for_each_pci_msi_entry(desc, pdev) \
169 for_each_msi_entry((desc), &(pdev)->dev)
171 struct pci_dev *msi_desc_to_pci_dev(struct msi_desc *desc);
172 void *msi_desc_to_pci_sysdata(struct msi_desc *desc);
173 void pci_write_msi_msg(unsigned int irq, struct msi_msg *msg);
174 #else /* CONFIG_PCI_MSI */
175 static inline void *msi_desc_to_pci_sysdata(struct msi_desc *desc)
179 static inline void pci_write_msi_msg(unsigned int irq, struct msi_msg *msg)
182 #endif /* CONFIG_PCI_MSI */
184 struct msi_desc *alloc_msi_entry(struct device *dev, int nvec,
185 const struct irq_affinity_desc *affinity);
186 void free_msi_entry(struct msi_desc *entry);
187 void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg);
188 void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg);
190 u32 __pci_msix_desc_mask_irq(struct msi_desc *desc, u32 flag);
191 u32 __pci_msi_desc_mask_irq(struct msi_desc *desc, u32 mask, u32 flag);
192 void pci_msi_mask_irq(struct irq_data *data);
193 void pci_msi_unmask_irq(struct irq_data *data);
196 * The arch hooks to setup up msi irqs. Default functions are implemented
197 * as weak symbols so that they /can/ be overriden by architecture specific
198 * code if needed. These hooks must be enabled by the architecture or by
199 * drivers which depend on them via msi_controller based MSI handling.
201 * If CONFIG_PCI_MSI_ARCH_FALLBACKS is not selected they are replaced by
202 * stubs with warnings.
204 #ifdef CONFIG_PCI_MSI_ARCH_FALLBACKS
205 int arch_setup_msi_irq(struct pci_dev *dev, struct msi_desc *desc);
206 void arch_teardown_msi_irq(unsigned int irq);
207 int arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type);
208 void arch_teardown_msi_irqs(struct pci_dev *dev);
209 void default_teardown_msi_irqs(struct pci_dev *dev);
211 static inline int arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
217 static inline void arch_teardown_msi_irqs(struct pci_dev *dev)
224 * The restore hooks are still available as they are useful even
225 * for fully irq domain based setups. Courtesy to XEN/X86.
227 void arch_restore_msi_irqs(struct pci_dev *dev);
228 void default_restore_msi_irqs(struct pci_dev *dev);
230 struct msi_controller {
231 struct module *owner;
233 struct device_node *of_node;
234 struct list_head list;
236 int (*setup_irq)(struct msi_controller *chip, struct pci_dev *dev,
237 struct msi_desc *desc);
238 int (*setup_irqs)(struct msi_controller *chip, struct pci_dev *dev,
240 void (*teardown_irq)(struct msi_controller *chip, unsigned int irq);
243 #ifdef CONFIG_GENERIC_MSI_IRQ_DOMAIN
245 #include <linux/irqhandler.h>
249 struct irq_domain_ops;
252 struct fwnode_handle;
253 struct msi_domain_info;
256 * struct msi_domain_ops - MSI interrupt domain callbacks
257 * @get_hwirq: Retrieve the resulting hw irq number
258 * @msi_init: Domain specific init function for MSI interrupts
259 * @msi_free: Domain specific function to free a MSI interrupts
260 * @msi_check: Callback for verification of the domain/info/dev data
261 * @msi_prepare: Prepare the allocation of the interrupts in the domain
262 * @msi_finish: Optional callback to finalize the allocation
263 * @set_desc: Set the msi descriptor for an interrupt
264 * @handle_error: Optional error handler if the allocation fails
265 * @domain_alloc_irqs: Optional function to override the default allocation
267 * @domain_free_irqs: Optional function to override the default free
270 * @get_hwirq, @msi_init and @msi_free are callbacks used by
271 * msi_create_irq_domain() and related interfaces
273 * @msi_check, @msi_prepare, @msi_finish, @set_desc and @handle_error
274 * are callbacks used by msi_domain_alloc_irqs() and related
275 * interfaces which are based on msi_desc.
277 * @domain_alloc_irqs, @domain_free_irqs can be used to override the
278 * default allocation/free functions (__msi_domain_alloc/free_irqs). This
279 * is initially for a wrapper around XENs seperate MSI universe which can't
280 * be wrapped into the regular irq domains concepts by mere mortals. This
281 * allows to universally use msi_domain_alloc/free_irqs without having to
282 * special case XEN all over the place.
284 * Contrary to other operations @domain_alloc_irqs and @domain_free_irqs
285 * are set to the default implementation if NULL and even when
286 * MSI_FLAG_USE_DEF_DOM_OPS is not set to avoid breaking existing users and
287 * because these callbacks are obviously mandatory.
289 * This is NOT meant to be abused, but it can be useful to build wrappers
290 * for specialized MSI irq domains which need extra work before and after
291 * calling __msi_domain_alloc_irqs()/__msi_domain_free_irqs().
293 struct msi_domain_ops {
294 irq_hw_number_t (*get_hwirq)(struct msi_domain_info *info,
295 msi_alloc_info_t *arg);
296 int (*msi_init)(struct irq_domain *domain,
297 struct msi_domain_info *info,
298 unsigned int virq, irq_hw_number_t hwirq,
299 msi_alloc_info_t *arg);
300 void (*msi_free)(struct irq_domain *domain,
301 struct msi_domain_info *info,
303 int (*msi_check)(struct irq_domain *domain,
304 struct msi_domain_info *info,
306 int (*msi_prepare)(struct irq_domain *domain,
307 struct device *dev, int nvec,
308 msi_alloc_info_t *arg);
309 void (*msi_finish)(msi_alloc_info_t *arg, int retval);
310 void (*set_desc)(msi_alloc_info_t *arg,
311 struct msi_desc *desc);
312 int (*handle_error)(struct irq_domain *domain,
313 struct msi_desc *desc, int error);
314 int (*domain_alloc_irqs)(struct irq_domain *domain,
315 struct device *dev, int nvec);
316 void (*domain_free_irqs)(struct irq_domain *domain,
321 * struct msi_domain_info - MSI interrupt domain data
322 * @flags: Flags to decribe features and capabilities
323 * @ops: The callback data structure
324 * @chip: Optional: associated interrupt chip
325 * @chip_data: Optional: associated interrupt chip data
326 * @handler: Optional: associated interrupt flow handler
327 * @handler_data: Optional: associated interrupt flow handler data
328 * @handler_name: Optional: associated interrupt flow handler name
329 * @data: Optional: domain specific data
331 struct msi_domain_info {
333 struct msi_domain_ops *ops;
334 struct irq_chip *chip;
336 irq_flow_handler_t handler;
338 const char *handler_name;
342 /* Flags for msi_domain_info */
345 * Init non implemented ops callbacks with default MSI domain
348 MSI_FLAG_USE_DEF_DOM_OPS = (1 << 0),
350 * Init non implemented chip callbacks with default MSI chip
353 MSI_FLAG_USE_DEF_CHIP_OPS = (1 << 1),
354 /* Support multiple PCI MSI interrupts */
355 MSI_FLAG_MULTI_PCI_MSI = (1 << 2),
356 /* Support PCI MSIX interrupts */
357 MSI_FLAG_PCI_MSIX = (1 << 3),
358 /* Needs early activate, required for PCI */
359 MSI_FLAG_ACTIVATE_EARLY = (1 << 4),
361 * Must reactivate when irq is started even when
362 * MSI_FLAG_ACTIVATE_EARLY has been set.
364 MSI_FLAG_MUST_REACTIVATE = (1 << 5),
365 /* Is level-triggered capable, using two messages */
366 MSI_FLAG_LEVEL_CAPABLE = (1 << 6),
369 int msi_domain_set_affinity(struct irq_data *data, const struct cpumask *mask,
372 struct irq_domain *msi_create_irq_domain(struct fwnode_handle *fwnode,
373 struct msi_domain_info *info,
374 struct irq_domain *parent);
375 int __msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev,
377 int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev,
379 void __msi_domain_free_irqs(struct irq_domain *domain, struct device *dev);
380 void msi_domain_free_irqs(struct irq_domain *domain, struct device *dev);
381 struct msi_domain_info *msi_get_domain_info(struct irq_domain *domain);
383 struct irq_domain *platform_msi_create_irq_domain(struct fwnode_handle *fwnode,
384 struct msi_domain_info *info,
385 struct irq_domain *parent);
386 int platform_msi_domain_alloc_irqs(struct device *dev, unsigned int nvec,
387 irq_write_msi_msg_t write_msi_msg);
388 void platform_msi_domain_free_irqs(struct device *dev);
390 /* When an MSI domain is used as an intermediate domain */
391 int msi_domain_prepare_irqs(struct irq_domain *domain, struct device *dev,
392 int nvec, msi_alloc_info_t *args);
393 int msi_domain_populate_irqs(struct irq_domain *domain, struct device *dev,
394 int virq, int nvec, msi_alloc_info_t *args);
396 __platform_msi_create_device_domain(struct device *dev,
399 irq_write_msi_msg_t write_msi_msg,
400 const struct irq_domain_ops *ops,
403 #define platform_msi_create_device_domain(dev, nvec, write, ops, data) \
404 __platform_msi_create_device_domain(dev, nvec, false, write, ops, data)
405 #define platform_msi_create_device_tree_domain(dev, nvec, write, ops, data) \
406 __platform_msi_create_device_domain(dev, nvec, true, write, ops, data)
408 int platform_msi_domain_alloc(struct irq_domain *domain, unsigned int virq,
409 unsigned int nr_irqs);
410 void platform_msi_domain_free(struct irq_domain *domain, unsigned int virq,
412 void *platform_msi_get_host_data(struct irq_domain *domain);
413 #endif /* CONFIG_GENERIC_MSI_IRQ_DOMAIN */
415 #ifdef CONFIG_PCI_MSI_IRQ_DOMAIN
416 void pci_msi_domain_write_msg(struct irq_data *irq_data, struct msi_msg *msg);
417 struct irq_domain *pci_msi_create_irq_domain(struct fwnode_handle *fwnode,
418 struct msi_domain_info *info,
419 struct irq_domain *parent);
420 int pci_msi_domain_check_cap(struct irq_domain *domain,
421 struct msi_domain_info *info, struct device *dev);
422 u32 pci_msi_domain_get_msi_rid(struct irq_domain *domain, struct pci_dev *pdev);
423 struct irq_domain *pci_msi_get_device_domain(struct pci_dev *pdev);
424 bool pci_dev_has_special_msi_domain(struct pci_dev *pdev);
426 static inline struct irq_domain *pci_msi_get_device_domain(struct pci_dev *pdev)
430 #endif /* CONFIG_PCI_MSI_IRQ_DOMAIN */
432 #endif /* LINUX_MSI_H */