2 * VGICv2 MMIO handling functions
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License version 2 as
6 * published by the Free Software Foundation.
8 * This program is distributed in the hope that it will be useful,
9 * but WITHOUT ANY WARRANTY; without even the implied warranty of
10 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
11 * GNU General Public License for more details.
14 #include <linux/irqchip/arm-gic.h>
15 #include <linux/kvm.h>
16 #include <linux/kvm_host.h>
17 #include <linux/nospec.h>
19 #include <kvm/iodev.h>
20 #include <kvm/arm_vgic.h>
23 #include "vgic-mmio.h"
25 static unsigned long vgic_mmio_read_v2_misc(struct kvm_vcpu *vcpu,
26 gpa_t addr, unsigned int len)
30 switch (addr & 0x0c) {
32 value = vcpu->kvm->arch.vgic.enabled ? GICD_ENABLE : 0;
35 value = vcpu->kvm->arch.vgic.nr_spis + VGIC_NR_PRIVATE_IRQS;
36 value = (value >> 5) - 1;
37 value |= (atomic_read(&vcpu->kvm->online_vcpus) - 1) << 5;
40 value = (PRODUCT_ID_KVM << 24) | (IMPLEMENTER_ARM << 0);
49 static void vgic_mmio_write_v2_misc(struct kvm_vcpu *vcpu,
50 gpa_t addr, unsigned int len,
53 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
54 bool was_enabled = dist->enabled;
56 switch (addr & 0x0c) {
58 dist->enabled = val & GICD_ENABLE;
59 if (!was_enabled && dist->enabled)
60 vgic_kick_vcpus(vcpu->kvm);
69 static void vgic_mmio_write_sgir(struct kvm_vcpu *source_vcpu,
70 gpa_t addr, unsigned int len,
73 int nr_vcpus = atomic_read(&source_vcpu->kvm->online_vcpus);
74 int intid = val & 0xf;
75 int targets = (val >> 16) & 0xff;
76 int mode = (val >> 24) & 0x03;
78 struct kvm_vcpu *vcpu;
81 case 0x0: /* as specified by targets */
84 targets = (1U << nr_vcpus) - 1; /* all, ... */
85 targets &= ~(1U << source_vcpu->vcpu_id); /* but self */
87 case 0x2: /* this very vCPU only */
88 targets = (1U << source_vcpu->vcpu_id);
90 case 0x3: /* reserved */
94 kvm_for_each_vcpu(c, vcpu, source_vcpu->kvm) {
97 if (!(targets & (1U << c)))
100 irq = vgic_get_irq(source_vcpu->kvm, vcpu, intid);
102 spin_lock(&irq->irq_lock);
103 irq->pending_latch = true;
104 irq->source |= 1U << source_vcpu->vcpu_id;
106 vgic_queue_irq_unlock(source_vcpu->kvm, irq);
107 vgic_put_irq(source_vcpu->kvm, irq);
111 static unsigned long vgic_mmio_read_target(struct kvm_vcpu *vcpu,
112 gpa_t addr, unsigned int len)
114 u32 intid = VGIC_ADDR_TO_INTID(addr, 8);
118 for (i = 0; i < len; i++) {
119 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
121 val |= (u64)irq->targets << (i * 8);
123 vgic_put_irq(vcpu->kvm, irq);
129 static void vgic_mmio_write_target(struct kvm_vcpu *vcpu,
130 gpa_t addr, unsigned int len,
133 u32 intid = VGIC_ADDR_TO_INTID(addr, 8);
134 u8 cpu_mask = GENMASK(atomic_read(&vcpu->kvm->online_vcpus) - 1, 0);
137 /* GICD_ITARGETSR[0-7] are read-only */
138 if (intid < VGIC_NR_PRIVATE_IRQS)
141 for (i = 0; i < len; i++) {
142 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, NULL, intid + i);
145 spin_lock(&irq->irq_lock);
147 irq->targets = (val >> (i * 8)) & cpu_mask;
148 target = irq->targets ? __ffs(irq->targets) : 0;
149 irq->target_vcpu = kvm_get_vcpu(vcpu->kvm, target);
151 spin_unlock(&irq->irq_lock);
152 vgic_put_irq(vcpu->kvm, irq);
156 static unsigned long vgic_mmio_read_sgipend(struct kvm_vcpu *vcpu,
157 gpa_t addr, unsigned int len)
159 u32 intid = addr & 0x0f;
163 for (i = 0; i < len; i++) {
164 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
166 val |= (u64)irq->source << (i * 8);
168 vgic_put_irq(vcpu->kvm, irq);
173 static void vgic_mmio_write_sgipendc(struct kvm_vcpu *vcpu,
174 gpa_t addr, unsigned int len,
177 u32 intid = addr & 0x0f;
180 for (i = 0; i < len; i++) {
181 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
183 spin_lock(&irq->irq_lock);
185 irq->source &= ~((val >> (i * 8)) & 0xff);
187 irq->pending_latch = false;
189 spin_unlock(&irq->irq_lock);
190 vgic_put_irq(vcpu->kvm, irq);
194 static void vgic_mmio_write_sgipends(struct kvm_vcpu *vcpu,
195 gpa_t addr, unsigned int len,
198 u32 intid = addr & 0x0f;
201 for (i = 0; i < len; i++) {
202 struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, intid + i);
204 spin_lock(&irq->irq_lock);
206 irq->source |= (val >> (i * 8)) & 0xff;
209 irq->pending_latch = true;
210 vgic_queue_irq_unlock(vcpu->kvm, irq);
212 spin_unlock(&irq->irq_lock);
214 vgic_put_irq(vcpu->kvm, irq);
218 #define GICC_ARCH_VERSION_V2 0x2
220 /* These are for userland accesses only, there is no guest-facing emulation. */
221 static unsigned long vgic_mmio_read_vcpuif(struct kvm_vcpu *vcpu,
222 gpa_t addr, unsigned int len)
224 struct vgic_vmcr vmcr;
227 vgic_get_vmcr(vcpu, &vmcr);
229 switch (addr & 0xff) {
231 val = vmcr.grpen0 << GIC_CPU_CTRL_EnableGrp0_SHIFT;
232 val |= vmcr.grpen1 << GIC_CPU_CTRL_EnableGrp1_SHIFT;
233 val |= vmcr.ackctl << GIC_CPU_CTRL_AckCtl_SHIFT;
234 val |= vmcr.fiqen << GIC_CPU_CTRL_FIQEn_SHIFT;
235 val |= vmcr.cbpr << GIC_CPU_CTRL_CBPR_SHIFT;
236 val |= vmcr.eoim << GIC_CPU_CTRL_EOImodeNS_SHIFT;
239 case GIC_CPU_PRIMASK:
241 * Our KVM_DEV_TYPE_ARM_VGIC_V2 device ABI exports the
242 * the PMR field as GICH_VMCR.VMPriMask rather than
243 * GICC_PMR.Priority, so we expose the upper five bits of
244 * priority mask to userspace using the lower bits in the
247 val = (vmcr.pmr & GICV_PMR_PRIORITY_MASK) >>
248 GICV_PMR_PRIORITY_SHIFT;
250 case GIC_CPU_BINPOINT:
253 case GIC_CPU_ALIAS_BINPOINT:
257 val = ((PRODUCT_ID_KVM << 20) |
258 (GICC_ARCH_VERSION_V2 << 16) |
268 static void vgic_mmio_write_vcpuif(struct kvm_vcpu *vcpu,
269 gpa_t addr, unsigned int len,
272 struct vgic_vmcr vmcr;
274 vgic_get_vmcr(vcpu, &vmcr);
276 switch (addr & 0xff) {
278 vmcr.grpen0 = !!(val & GIC_CPU_CTRL_EnableGrp0);
279 vmcr.grpen1 = !!(val & GIC_CPU_CTRL_EnableGrp1);
280 vmcr.ackctl = !!(val & GIC_CPU_CTRL_AckCtl);
281 vmcr.fiqen = !!(val & GIC_CPU_CTRL_FIQEn);
282 vmcr.cbpr = !!(val & GIC_CPU_CTRL_CBPR);
283 vmcr.eoim = !!(val & GIC_CPU_CTRL_EOImodeNS);
286 case GIC_CPU_PRIMASK:
288 * Our KVM_DEV_TYPE_ARM_VGIC_V2 device ABI exports the
289 * the PMR field as GICH_VMCR.VMPriMask rather than
290 * GICC_PMR.Priority, so we expose the upper five bits of
291 * priority mask to userspace using the lower bits in the
294 vmcr.pmr = (val << GICV_PMR_PRIORITY_SHIFT) &
295 GICV_PMR_PRIORITY_MASK;
297 case GIC_CPU_BINPOINT:
300 case GIC_CPU_ALIAS_BINPOINT:
305 vgic_set_vmcr(vcpu, &vmcr);
308 static unsigned long vgic_mmio_read_apr(struct kvm_vcpu *vcpu,
309 gpa_t addr, unsigned int len)
311 int n; /* which APRn is this */
313 n = (addr >> 2) & 0x3;
315 if (kvm_vgic_global_state.type == VGIC_V2) {
316 /* GICv2 hardware systems support max. 32 groups */
319 return vcpu->arch.vgic_cpu.vgic_v2.vgic_apr;
321 struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3;
323 if (n > vgic_v3_max_apr_idx(vcpu))
326 n = array_index_nospec(n, 4);
328 /* GICv3 only uses ICH_AP1Rn for memory mapped (GICv2) guests */
329 return vgicv3->vgic_ap1r[n];
333 static void vgic_mmio_write_apr(struct kvm_vcpu *vcpu,
334 gpa_t addr, unsigned int len,
337 int n; /* which APRn is this */
339 n = (addr >> 2) & 0x3;
341 if (kvm_vgic_global_state.type == VGIC_V2) {
342 /* GICv2 hardware systems support max. 32 groups */
345 vcpu->arch.vgic_cpu.vgic_v2.vgic_apr = val;
347 struct vgic_v3_cpu_if *vgicv3 = &vcpu->arch.vgic_cpu.vgic_v3;
349 if (n > vgic_v3_max_apr_idx(vcpu))
352 n = array_index_nospec(n, 4);
354 /* GICv3 only uses ICH_AP1Rn for memory mapped (GICv2) guests */
355 vgicv3->vgic_ap1r[n] = val;
359 static const struct vgic_register_region vgic_v2_dist_registers[] = {
360 REGISTER_DESC_WITH_LENGTH(GIC_DIST_CTRL,
361 vgic_mmio_read_v2_misc, vgic_mmio_write_v2_misc, 12,
363 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_IGROUP,
364 vgic_mmio_read_rao, vgic_mmio_write_wi, NULL, NULL, 1,
366 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_ENABLE_SET,
367 vgic_mmio_read_enable, vgic_mmio_write_senable, NULL, NULL, 1,
369 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_ENABLE_CLEAR,
370 vgic_mmio_read_enable, vgic_mmio_write_cenable, NULL, NULL, 1,
372 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_PENDING_SET,
373 vgic_mmio_read_pending, vgic_mmio_write_spending, NULL, NULL, 1,
375 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_PENDING_CLEAR,
376 vgic_mmio_read_pending, vgic_mmio_write_cpending, NULL, NULL, 1,
378 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_ACTIVE_SET,
379 vgic_mmio_read_active, vgic_mmio_write_sactive,
380 NULL, vgic_mmio_uaccess_write_sactive, 1,
382 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_ACTIVE_CLEAR,
383 vgic_mmio_read_active, vgic_mmio_write_cactive,
384 NULL, vgic_mmio_uaccess_write_cactive, 1,
386 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_PRI,
387 vgic_mmio_read_priority, vgic_mmio_write_priority, NULL, NULL,
388 8, VGIC_ACCESS_32bit | VGIC_ACCESS_8bit),
389 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_TARGET,
390 vgic_mmio_read_target, vgic_mmio_write_target, NULL, NULL, 8,
391 VGIC_ACCESS_32bit | VGIC_ACCESS_8bit),
392 REGISTER_DESC_WITH_BITS_PER_IRQ(GIC_DIST_CONFIG,
393 vgic_mmio_read_config, vgic_mmio_write_config, NULL, NULL, 2,
395 REGISTER_DESC_WITH_LENGTH(GIC_DIST_SOFTINT,
396 vgic_mmio_read_raz, vgic_mmio_write_sgir, 4,
398 REGISTER_DESC_WITH_LENGTH(GIC_DIST_SGI_PENDING_CLEAR,
399 vgic_mmio_read_sgipend, vgic_mmio_write_sgipendc, 16,
400 VGIC_ACCESS_32bit | VGIC_ACCESS_8bit),
401 REGISTER_DESC_WITH_LENGTH(GIC_DIST_SGI_PENDING_SET,
402 vgic_mmio_read_sgipend, vgic_mmio_write_sgipends, 16,
403 VGIC_ACCESS_32bit | VGIC_ACCESS_8bit),
406 static const struct vgic_register_region vgic_v2_cpu_registers[] = {
407 REGISTER_DESC_WITH_LENGTH(GIC_CPU_CTRL,
408 vgic_mmio_read_vcpuif, vgic_mmio_write_vcpuif, 4,
410 REGISTER_DESC_WITH_LENGTH(GIC_CPU_PRIMASK,
411 vgic_mmio_read_vcpuif, vgic_mmio_write_vcpuif, 4,
413 REGISTER_DESC_WITH_LENGTH(GIC_CPU_BINPOINT,
414 vgic_mmio_read_vcpuif, vgic_mmio_write_vcpuif, 4,
416 REGISTER_DESC_WITH_LENGTH(GIC_CPU_ALIAS_BINPOINT,
417 vgic_mmio_read_vcpuif, vgic_mmio_write_vcpuif, 4,
419 REGISTER_DESC_WITH_LENGTH(GIC_CPU_ACTIVEPRIO,
420 vgic_mmio_read_apr, vgic_mmio_write_apr, 16,
422 REGISTER_DESC_WITH_LENGTH(GIC_CPU_IDENT,
423 vgic_mmio_read_vcpuif, vgic_mmio_write_vcpuif, 4,
427 unsigned int vgic_v2_init_dist_iodev(struct vgic_io_device *dev)
429 dev->regions = vgic_v2_dist_registers;
430 dev->nr_regions = ARRAY_SIZE(vgic_v2_dist_registers);
432 kvm_iodevice_init(&dev->dev, &kvm_io_gic_ops);
437 int vgic_v2_has_attr_regs(struct kvm_device *dev, struct kvm_device_attr *attr)
439 const struct vgic_register_region *region;
440 struct vgic_io_device iodev;
441 struct vgic_reg_attr reg_attr;
442 struct kvm_vcpu *vcpu;
446 ret = vgic_v2_parse_attr(dev, attr, ®_attr);
450 vcpu = reg_attr.vcpu;
451 addr = reg_attr.addr;
453 switch (attr->group) {
454 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
455 iodev.regions = vgic_v2_dist_registers;
456 iodev.nr_regions = ARRAY_SIZE(vgic_v2_dist_registers);
459 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
460 iodev.regions = vgic_v2_cpu_registers;
461 iodev.nr_regions = ARRAY_SIZE(vgic_v2_cpu_registers);
468 /* We only support aligned 32-bit accesses. */
472 region = vgic_get_mmio_region(vcpu, &iodev, addr, sizeof(u32));
479 int vgic_v2_cpuif_uaccess(struct kvm_vcpu *vcpu, bool is_write,
480 int offset, u32 *val)
482 struct vgic_io_device dev = {
483 .regions = vgic_v2_cpu_registers,
484 .nr_regions = ARRAY_SIZE(vgic_v2_cpu_registers),
485 .iodev_type = IODEV_CPUIF,
488 return vgic_uaccess(vcpu, &dev, is_write, offset, val);
491 int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
492 int offset, u32 *val)
494 struct vgic_io_device dev = {
495 .regions = vgic_v2_dist_registers,
496 .nr_regions = ARRAY_SIZE(vgic_v2_dist_registers),
497 .iodev_type = IODEV_DIST,
500 return vgic_uaccess(vcpu, &dev, is_write, offset, val);