cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

irq-mvebu-odmi.c (5801B)


      1/*
      2 * Copyright (C) 2016 Marvell
      3 *
      4 * Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
      5 *
      6 * This file is licensed under the terms of the GNU General Public
      7 * License version 2.  This program is licensed "as is" without any
      8 * warranty of any kind, whether express or implied.
      9 */
     10
     11#define pr_fmt(fmt) "GIC-ODMI: " fmt
     12
     13#include <linux/irq.h>
     14#include <linux/irqchip.h>
     15#include <linux/irqdomain.h>
     16#include <linux/kernel.h>
     17#include <linux/msi.h>
     18#include <linux/of_address.h>
     19#include <linux/slab.h>
     20#include <dt-bindings/interrupt-controller/arm-gic.h>
     21
     22#define GICP_ODMIN_SET			0x40
     23#define   GICP_ODMI_INT_NUM_SHIFT	12
     24#define GICP_ODMIN_GM_EP_R0		0x110
     25#define GICP_ODMIN_GM_EP_R1		0x114
     26#define GICP_ODMIN_GM_EA_R0		0x108
     27#define GICP_ODMIN_GM_EA_R1		0x118
     28
     29/*
     30 * We don't support the group events, so we simply have 8 interrupts
     31 * per frame.
     32 */
     33#define NODMIS_SHIFT		3
     34#define NODMIS_PER_FRAME	(1 << NODMIS_SHIFT)
     35#define NODMIS_MASK		(NODMIS_PER_FRAME - 1)
     36
     37struct odmi_data {
     38	struct resource res;
     39	void __iomem *base;
     40	unsigned int spi_base;
     41};
     42
     43static struct odmi_data *odmis;
     44static unsigned long *odmis_bm;
     45static unsigned int odmis_count;
     46
     47/* Protects odmis_bm */
     48static DEFINE_SPINLOCK(odmis_bm_lock);
     49
     50static void odmi_compose_msi_msg(struct irq_data *d, struct msi_msg *msg)
     51{
     52	struct odmi_data *odmi;
     53	phys_addr_t addr;
     54	unsigned int odmin;
     55
     56	if (WARN_ON(d->hwirq >= odmis_count * NODMIS_PER_FRAME))
     57		return;
     58
     59	odmi = &odmis[d->hwirq >> NODMIS_SHIFT];
     60	odmin = d->hwirq & NODMIS_MASK;
     61
     62	addr = odmi->res.start + GICP_ODMIN_SET;
     63
     64	msg->address_hi = upper_32_bits(addr);
     65	msg->address_lo = lower_32_bits(addr);
     66	msg->data = odmin << GICP_ODMI_INT_NUM_SHIFT;
     67}
     68
     69static struct irq_chip odmi_irq_chip = {
     70	.name			= "ODMI",
     71	.irq_mask		= irq_chip_mask_parent,
     72	.irq_unmask		= irq_chip_unmask_parent,
     73	.irq_eoi		= irq_chip_eoi_parent,
     74	.irq_set_affinity	= irq_chip_set_affinity_parent,
     75	.irq_compose_msi_msg	= odmi_compose_msi_msg,
     76};
     77
     78static int odmi_irq_domain_alloc(struct irq_domain *domain, unsigned int virq,
     79				 unsigned int nr_irqs, void *args)
     80{
     81	struct odmi_data *odmi = NULL;
     82	struct irq_fwspec fwspec;
     83	struct irq_data *d;
     84	unsigned int hwirq, odmin;
     85	int ret;
     86
     87	spin_lock(&odmis_bm_lock);
     88	hwirq = find_first_zero_bit(odmis_bm, NODMIS_PER_FRAME * odmis_count);
     89	if (hwirq >= NODMIS_PER_FRAME * odmis_count) {
     90		spin_unlock(&odmis_bm_lock);
     91		return -ENOSPC;
     92	}
     93
     94	__set_bit(hwirq, odmis_bm);
     95	spin_unlock(&odmis_bm_lock);
     96
     97	odmi = &odmis[hwirq >> NODMIS_SHIFT];
     98	odmin = hwirq & NODMIS_MASK;
     99
    100	fwspec.fwnode = domain->parent->fwnode;
    101	fwspec.param_count = 3;
    102	fwspec.param[0] = GIC_SPI;
    103	fwspec.param[1] = odmi->spi_base - 32 + odmin;
    104	fwspec.param[2] = IRQ_TYPE_EDGE_RISING;
    105
    106	ret = irq_domain_alloc_irqs_parent(domain, virq, 1, &fwspec);
    107	if (ret) {
    108		pr_err("Cannot allocate parent IRQ\n");
    109		spin_lock(&odmis_bm_lock);
    110		__clear_bit(odmin, odmis_bm);
    111		spin_unlock(&odmis_bm_lock);
    112		return ret;
    113	}
    114
    115	/* Configure the interrupt line to be edge */
    116	d = irq_domain_get_irq_data(domain->parent, virq);
    117	d->chip->irq_set_type(d, IRQ_TYPE_EDGE_RISING);
    118
    119	irq_domain_set_hwirq_and_chip(domain, virq, hwirq,
    120				      &odmi_irq_chip, NULL);
    121
    122	return 0;
    123}
    124
    125static void odmi_irq_domain_free(struct irq_domain *domain,
    126				 unsigned int virq, unsigned int nr_irqs)
    127{
    128	struct irq_data *d = irq_domain_get_irq_data(domain, virq);
    129
    130	if (d->hwirq >= odmis_count * NODMIS_PER_FRAME) {
    131		pr_err("Failed to teardown msi. Invalid hwirq %lu\n", d->hwirq);
    132		return;
    133	}
    134
    135	irq_domain_free_irqs_parent(domain, virq, nr_irqs);
    136
    137	/* Actually free the MSI */
    138	spin_lock(&odmis_bm_lock);
    139	__clear_bit(d->hwirq, odmis_bm);
    140	spin_unlock(&odmis_bm_lock);
    141}
    142
    143static const struct irq_domain_ops odmi_domain_ops = {
    144	.alloc	= odmi_irq_domain_alloc,
    145	.free	= odmi_irq_domain_free,
    146};
    147
    148static struct irq_chip odmi_msi_irq_chip = {
    149	.name	= "ODMI",
    150};
    151
    152static struct msi_domain_ops odmi_msi_ops = {
    153};
    154
    155static struct msi_domain_info odmi_msi_domain_info = {
    156	.flags	= (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS),
    157	.ops	= &odmi_msi_ops,
    158	.chip	= &odmi_msi_irq_chip,
    159};
    160
    161static int __init mvebu_odmi_init(struct device_node *node,
    162				  struct device_node *parent)
    163{
    164	struct irq_domain *inner_domain, *plat_domain;
    165	int ret, i;
    166
    167	if (of_property_read_u32(node, "marvell,odmi-frames", &odmis_count))
    168		return -EINVAL;
    169
    170	odmis = kcalloc(odmis_count, sizeof(struct odmi_data), GFP_KERNEL);
    171	if (!odmis)
    172		return -ENOMEM;
    173
    174	odmis_bm = bitmap_zalloc(odmis_count * NODMIS_PER_FRAME, GFP_KERNEL);
    175	if (!odmis_bm) {
    176		ret = -ENOMEM;
    177		goto err_alloc;
    178	}
    179
    180	for (i = 0; i < odmis_count; i++) {
    181		struct odmi_data *odmi = &odmis[i];
    182
    183		ret = of_address_to_resource(node, i, &odmi->res);
    184		if (ret)
    185			goto err_unmap;
    186
    187		odmi->base = of_io_request_and_map(node, i, "odmi");
    188		if (IS_ERR(odmi->base)) {
    189			ret = PTR_ERR(odmi->base);
    190			goto err_unmap;
    191		}
    192
    193		if (of_property_read_u32_index(node, "marvell,spi-base",
    194					       i, &odmi->spi_base)) {
    195			ret = -EINVAL;
    196			goto err_unmap;
    197		}
    198	}
    199
    200	inner_domain = irq_domain_create_linear(of_node_to_fwnode(node),
    201						odmis_count * NODMIS_PER_FRAME,
    202						&odmi_domain_ops, NULL);
    203	if (!inner_domain) {
    204		ret = -ENOMEM;
    205		goto err_unmap;
    206	}
    207
    208	inner_domain->parent = irq_find_host(parent);
    209
    210	plat_domain = platform_msi_create_irq_domain(of_node_to_fwnode(node),
    211						     &odmi_msi_domain_info,
    212						     inner_domain);
    213	if (!plat_domain) {
    214		ret = -ENOMEM;
    215		goto err_remove_inner;
    216	}
    217
    218	return 0;
    219
    220err_remove_inner:
    221	irq_domain_remove(inner_domain);
    222err_unmap:
    223	for (i = 0; i < odmis_count; i++) {
    224		struct odmi_data *odmi = &odmis[i];
    225
    226		if (odmi->base && !IS_ERR(odmi->base))
    227			iounmap(odmis[i].base);
    228	}
    229	bitmap_free(odmis_bm);
    230err_alloc:
    231	kfree(odmis);
    232	return ret;
    233}
    234
    235IRQCHIP_DECLARE(mvebu_odmi, "marvell,odmi-controller", mvebu_odmi_init);