/* * RISC-V IMSIC (Incoming Message Signaled Interrupt Controller) * * Copyright (c) 2021 Western Digital Corporation or its affiliates. * * This program is free software; you can redistribute it and/or modify it * under the terms and conditions of the GNU General Public License, * version 2 or later, as published by the Free Software Foundation. * * This program is distributed in the hope it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for * more details. * * You should have received a copy of the GNU General Public License along with * this program. If not, see . */ #include "qemu/osdep.h" #include "qapi/error.h" #include "qemu/log.h" #include "qemu/module.h" #include "qemu/error-report.h" #include "qemu/bswap.h" #include "exec/address-spaces.h" #include "hw/sysbus.h" #include "hw/pci/msi.h" #include "hw/boards.h" #include "hw/qdev-properties.h" #include "hw/intc/riscv_imsic.h" #include "hw/irq.h" #include "target/riscv/cpu.h" #include "target/riscv/cpu_bits.h" #include "sysemu/sysemu.h" #include "sysemu/kvm.h" #include "migration/vmstate.h" #define IMSIC_MMIO_PAGE_LE 0x00 #define IMSIC_MMIO_PAGE_BE 0x04 #define IMSIC_MIN_ID ((IMSIC_EIPx_BITS * 2) - 1) #define IMSIC_MAX_ID (IMSIC_TOPEI_IID_MASK) #define IMSIC_EISTATE_PENDING (1U << 0) #define IMSIC_EISTATE_ENABLED (1U << 1) #define IMSIC_EISTATE_ENPEND (IMSIC_EISTATE_ENABLED | \ IMSIC_EISTATE_PENDING) static uint32_t riscv_imsic_topei(RISCVIMSICState *imsic, uint32_t page) { uint32_t i, max_irq, base; base = page * imsic->num_irqs; max_irq = (imsic->eithreshold[page] && (imsic->eithreshold[page] <= imsic->num_irqs)) ? imsic->eithreshold[page] : imsic->num_irqs; for (i = 1; i < max_irq; i++) { if ((qatomic_read(&imsic->eistate[base + i]) & IMSIC_EISTATE_ENPEND) == IMSIC_EISTATE_ENPEND) { return (i << IMSIC_TOPEI_IID_SHIFT) | i; } } return 0; } static void riscv_imsic_update(RISCVIMSICState *imsic, uint32_t page) { uint32_t base = page * imsic->num_irqs; /* * Lower the interrupt line if necessary, then evaluate the current * IMSIC state. * This sequence ensures that any race between evaluating the eistate and * updating the interrupt line will not result in an incorrectly * deactivated connected CPU IRQ line. * If multiple interrupts are pending, this sequence functions identically * to qemu_irq_pulse. */ if (qatomic_fetch_and(&imsic->eistate[base], ~IMSIC_EISTATE_ENPEND)) { qemu_irq_lower(imsic->external_irqs[page]); } if (imsic->eidelivery[page] && riscv_imsic_topei(imsic, page)) { qemu_irq_raise(imsic->external_irqs[page]); qatomic_or(&imsic->eistate[base], IMSIC_EISTATE_ENPEND); } } static int riscv_imsic_eidelivery_rmw(RISCVIMSICState *imsic, uint32_t page, target_ulong *val, target_ulong new_val, target_ulong wr_mask) { target_ulong old_val = imsic->eidelivery[page]; if (val) { *val = old_val; } wr_mask &= 0x1; imsic->eidelivery[page] = (old_val & ~wr_mask) | (new_val & wr_mask); riscv_imsic_update(imsic, page); return 0; } static int riscv_imsic_eithreshold_rmw(RISCVIMSICState *imsic, uint32_t page, target_ulong *val, target_ulong new_val, target_ulong wr_mask) { target_ulong old_val = imsic->eithreshold[page]; if (val) { *val = old_val; } wr_mask &= IMSIC_MAX_ID; imsic->eithreshold[page] = (old_val & ~wr_mask) | (new_val & wr_mask); riscv_imsic_update(imsic, page); return 0; } static int riscv_imsic_topei_rmw(RISCVIMSICState *imsic, uint32_t page, target_ulong *val, target_ulong new_val, target_ulong wr_mask) { uint32_t base, topei = riscv_imsic_topei(imsic, page); /* Read pending and enabled interrupt with highest priority */ if (val) { *val = topei; } /* Writes ignore value and clear top pending interrupt */ if (topei && wr_mask) { topei >>= IMSIC_TOPEI_IID_SHIFT; base = page * imsic->num_irqs; if (topei) { qatomic_and(&imsic->eistate[base + topei], ~IMSIC_EISTATE_PENDING); } } riscv_imsic_update(imsic, page); return 0; } static int riscv_imsic_eix_rmw(RISCVIMSICState *imsic, uint32_t xlen, uint32_t page, uint32_t num, bool pend, target_ulong *val, target_ulong new_val, target_ulong wr_mask) { uint32_t i, base, prev; target_ulong mask; uint32_t state = (pend) ? IMSIC_EISTATE_PENDING : IMSIC_EISTATE_ENABLED; if (xlen != 32) { if (num & 0x1) { return -EINVAL; } num >>= 1; } if (num >= (imsic->num_irqs / xlen)) { return -EINVAL; } base = (page * imsic->num_irqs) + (num * xlen); if (val) { *val = 0; } for (i = 0; i < xlen; i++) { /* Bit0 of eip0 and eie0 are read-only zero */ if (!num && !i) { continue; } mask = (target_ulong)1 << i; if (wr_mask & mask) { if (new_val & mask) { prev = qatomic_fetch_or(&imsic->eistate[base + i], state); } else { prev = qatomic_fetch_and(&imsic->eistate[base + i], ~state); } } else { prev = qatomic_read(&imsic->eistate[base + i]); } if (val && (prev & state)) { *val |= mask; } } riscv_imsic_update(imsic, page); return 0; } static int riscv_imsic_rmw(void *arg, target_ulong reg, target_ulong *val, target_ulong new_val, target_ulong wr_mask) { RISCVIMSICState *imsic = arg; uint32_t isel, priv, virt, vgein, xlen, page; priv = AIA_IREG_PRIV(reg); virt = AIA_IREG_VIRT(reg); isel = AIA_IREG_ISEL(reg); vgein = AIA_IREG_VGEIN(reg); xlen = AIA_IREG_XLEN(reg); if (imsic->mmode) { if (priv == PRV_M && !virt) { page = 0; } else { goto err; } } else { if (priv == PRV_S) { if (virt) { if (vgein && vgein < imsic->num_pages) { page = vgein; } else { goto err; } } else { page = 0; } } else { goto err; } } switch (isel) { case ISELECT_IMSIC_EIDELIVERY: return riscv_imsic_eidelivery_rmw(imsic, page, val, new_val, wr_mask); case ISELECT_IMSIC_EITHRESHOLD: return riscv_imsic_eithreshold_rmw(imsic, page, val, new_val, wr_mask); case ISELECT_IMSIC_TOPEI: return riscv_imsic_topei_rmw(imsic, page, val, new_val, wr_mask); case ISELECT_IMSIC_EIP0 ... ISELECT_IMSIC_EIP63: return riscv_imsic_eix_rmw(imsic, xlen, page, isel - ISELECT_IMSIC_EIP0, true, val, new_val, wr_mask); case ISELECT_IMSIC_EIE0 ... ISELECT_IMSIC_EIE63: return riscv_imsic_eix_rmw(imsic, xlen, page, isel - ISELECT_IMSIC_EIE0, false, val, new_val, wr_mask); default: break; }; err: qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid register priv=%d virt=%d isel=%d vgein=%d\n", __func__, priv, virt, isel, vgein); return -EINVAL; } static uint64_t riscv_imsic_read(void *opaque, hwaddr addr, unsigned size) { RISCVIMSICState *imsic = opaque; /* Reads must be 4 byte words */ if ((addr & 0x3) != 0) { goto err; } /* Reads cannot be out of range */ if (addr > IMSIC_MMIO_SIZE(imsic->num_pages)) { goto err; } return 0; err: qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid register read 0x%" HWADDR_PRIx "\n", __func__, addr); return 0; } static void riscv_imsic_write(void *opaque, hwaddr addr, uint64_t value, unsigned size) { RISCVIMSICState *imsic = opaque; uint32_t page; /* Writes must be 4 byte words */ if ((addr & 0x3) != 0) { goto err; } /* Writes cannot be out of range */ if (addr > IMSIC_MMIO_SIZE(imsic->num_pages)) { goto err; } #if defined(CONFIG_KVM) if (kvm_irqchip_in_kernel()) { struct kvm_msi msi; msi.address_lo = extract64(imsic->mmio.addr + addr, 0, 32); msi.address_hi = extract64(imsic->mmio.addr + addr, 32, 32); msi.data = le32_to_cpu(value); kvm_vm_ioctl(kvm_state, KVM_SIGNAL_MSI, &msi); return; } #endif /* Writes only supported for MSI little-endian registers */ page = addr >> IMSIC_MMIO_PAGE_SHIFT; if ((addr & (IMSIC_MMIO_PAGE_SZ - 1)) == IMSIC_MMIO_PAGE_LE) { if (value && (value < imsic->num_irqs)) { qatomic_or(&imsic->eistate[(page * imsic->num_irqs) + value], IMSIC_EISTATE_PENDING); /* Update CPU external interrupt status */ riscv_imsic_update(imsic, page); } } return; err: qemu_log_mask(LOG_GUEST_ERROR, "%s: Invalid register write 0x%" HWADDR_PRIx "\n", __func__, addr); } static const MemoryRegionOps riscv_imsic_ops = { .read = riscv_imsic_read, .write = riscv_imsic_write, .endianness = DEVICE_LITTLE_ENDIAN, .valid = { .min_access_size = 4, .max_access_size = 4 } }; static void riscv_imsic_realize(DeviceState *dev, Error **errp) { RISCVIMSICState *imsic = RISCV_IMSIC(dev); RISCVCPU *rcpu = RISCV_CPU(cpu_by_arch_id(imsic->hartid)); CPUState *cpu = cpu_by_arch_id(imsic->hartid); CPURISCVState *env = cpu ? cpu_env(cpu) : NULL; if (!kvm_irqchip_in_kernel()) { imsic->num_eistate = imsic->num_pages * imsic->num_irqs; imsic->eidelivery = g_new0(uint32_t, imsic->num_pages); imsic->eithreshold = g_new0(uint32_t, imsic->num_pages); imsic->eistate = g_new0(uint32_t, imsic->num_eistate); } memory_region_init_io(&imsic->mmio, OBJECT(dev), &riscv_imsic_ops, imsic, TYPE_RISCV_IMSIC, IMSIC_MMIO_SIZE(imsic->num_pages)); sysbus_init_mmio(SYS_BUS_DEVICE(dev), &imsic->mmio); /* Claim the CPU interrupt to be triggered by this IMSIC */ if (riscv_cpu_claim_interrupts(rcpu, (imsic->mmode) ? MIP_MEIP : MIP_SEIP) < 0) { error_setg(errp, "%s already claimed", (imsic->mmode) ? "MEIP" : "SEIP"); return; } /* Create output IRQ lines */ imsic->external_irqs = g_malloc(sizeof(qemu_irq) * imsic->num_pages); qdev_init_gpio_out(dev, imsic->external_irqs, imsic->num_pages); /* Force select AIA feature and setup CSR read-modify-write callback */ if (env) { if (!imsic->mmode) { rcpu->cfg.ext_ssaia = true; riscv_cpu_set_geilen(env, imsic->num_pages - 1); } else { rcpu->cfg.ext_smaia = true; } riscv_cpu_set_aia_ireg_rmw_fn(env, (imsic->mmode) ? PRV_M : PRV_S, riscv_imsic_rmw, imsic); } msi_nonbroken = true; } static Property riscv_imsic_properties[] = { DEFINE_PROP_BOOL("mmode", RISCVIMSICState, mmode, 0), DEFINE_PROP_UINT32("hartid", RISCVIMSICState, hartid, 0), DEFINE_PROP_UINT32("num-pages", RISCVIMSICState, num_pages, 0), DEFINE_PROP_UINT32("num-irqs", RISCVIMSICState, num_irqs, 0), DEFINE_PROP_END_OF_LIST(), }; static const VMStateDescription vmstate_riscv_imsic = { .name = "riscv_imsic", .version_id = 1, .minimum_version_id = 1, .fields = (const VMStateField[]) { VMSTATE_VARRAY_UINT32(eidelivery, RISCVIMSICState, num_pages, 0, vmstate_info_uint32, uint32_t), VMSTATE_VARRAY_UINT32(eithreshold, RISCVIMSICState, num_pages, 0, vmstate_info_uint32, uint32_t), VMSTATE_VARRAY_UINT32(eistate, RISCVIMSICState, num_eistate, 0, vmstate_info_uint32, uint32_t), VMSTATE_END_OF_LIST() } }; static void riscv_imsic_class_init(ObjectClass *klass, void *data) { DeviceClass *dc = DEVICE_CLASS(klass); device_class_set_props(dc, riscv_imsic_properties); dc->realize = riscv_imsic_realize; dc->vmsd = &vmstate_riscv_imsic; } static const TypeInfo riscv_imsic_info = { .name = TYPE_RISCV_IMSIC, .parent = TYPE_SYS_BUS_DEVICE, .instance_size = sizeof(RISCVIMSICState), .class_init = riscv_imsic_class_init, }; static void riscv_imsic_register_types(void) { type_register_static(&riscv_imsic_info); } type_init(riscv_imsic_register_types) /* * Create IMSIC device. */ DeviceState *riscv_imsic_create(hwaddr addr, uint32_t hartid, bool mmode, uint32_t num_pages, uint32_t num_ids) { DeviceState *dev = qdev_new(TYPE_RISCV_IMSIC); CPUState *cpu = cpu_by_arch_id(hartid); uint32_t i; assert(!(addr & (IMSIC_MMIO_PAGE_SZ - 1))); if (mmode) { assert(num_pages == 1); } else { assert(num_pages >= 1 && num_pages <= (IRQ_LOCAL_GUEST_MAX + 1)); } assert(IMSIC_MIN_ID <= num_ids); assert(num_ids <= IMSIC_MAX_ID); assert((num_ids & IMSIC_MIN_ID) == IMSIC_MIN_ID); qdev_prop_set_bit(dev, "mmode", mmode); qdev_prop_set_uint32(dev, "hartid", hartid); qdev_prop_set_uint32(dev, "num-pages", num_pages); qdev_prop_set_uint32(dev, "num-irqs", num_ids + 1); sysbus_realize_and_unref(SYS_BUS_DEVICE(dev), &error_fatal); sysbus_mmio_map(SYS_BUS_DEVICE(dev), 0, addr); for (i = 0; i < num_pages; i++) { if (!i) { qdev_connect_gpio_out_named(dev, NULL, i, qdev_get_gpio_in(DEVICE(cpu), (mmode) ? IRQ_M_EXT : IRQ_S_EXT)); } else { qdev_connect_gpio_out_named(dev, NULL, i, qdev_get_gpio_in(DEVICE(cpu), IRQ_LOCAL_MAX + i - 1)); } } return dev; }