750 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			750 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * VGIC: KVM DEVICE API
 | |
|  *
 | |
|  * Copyright (C) 2015 ARM Ltd.
 | |
|  * Author: Marc Zyngier <marc.zyngier@arm.com>
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License version 2 as
 | |
|  * published by the Free Software Foundation.
 | |
|  *
 | |
|  * This program is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
|  * GNU General Public License for more details.
 | |
|  */
 | |
| #include <linux/kvm_host.h>
 | |
| #include <kvm/arm_vgic.h>
 | |
| #include <linux/uaccess.h>
 | |
| #include <asm/kvm_mmu.h>
 | |
| #include <asm/cputype.h>
 | |
| #include "vgic.h"
 | |
| 
 | |
| /* common helpers */
 | |
| 
 | |
| int vgic_check_ioaddr(struct kvm *kvm, phys_addr_t *ioaddr,
 | |
| 		      phys_addr_t addr, phys_addr_t alignment)
 | |
| {
 | |
| 	if (addr & ~KVM_PHYS_MASK)
 | |
| 		return -E2BIG;
 | |
| 
 | |
| 	if (!IS_ALIGNED(addr, alignment))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (!IS_VGIC_ADDR_UNDEF(*ioaddr))
 | |
| 		return -EEXIST;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int vgic_check_type(struct kvm *kvm, int type_needed)
 | |
| {
 | |
| 	if (kvm->arch.vgic.vgic_model != type_needed)
 | |
| 		return -ENODEV;
 | |
| 	else
 | |
| 		return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * kvm_vgic_addr - set or get vgic VM base addresses
 | |
|  * @kvm:   pointer to the vm struct
 | |
|  * @type:  the VGIC addr type, one of KVM_VGIC_V[23]_ADDR_TYPE_XXX
 | |
|  * @addr:  pointer to address value
 | |
|  * @write: if true set the address in the VM address space, if false read the
 | |
|  *          address
 | |
|  *
 | |
|  * Set or get the vgic base addresses for the distributor and the virtual CPU
 | |
|  * interface in the VM physical address space.  These addresses are properties
 | |
|  * of the emulated core/SoC and therefore user space initially knows this
 | |
|  * information.
 | |
|  * Check them for sanity (alignment, double assignment). We can't check for
 | |
|  * overlapping regions in case of a virtual GICv3 here, since we don't know
 | |
|  * the number of VCPUs yet, so we defer this check to map_resources().
 | |
|  */
 | |
| int kvm_vgic_addr(struct kvm *kvm, unsigned long type, u64 *addr, bool write)
 | |
| {
 | |
| 	int r = 0;
 | |
| 	struct vgic_dist *vgic = &kvm->arch.vgic;
 | |
| 	phys_addr_t *addr_ptr, alignment;
 | |
| 	u64 undef_value = VGIC_ADDR_UNDEF;
 | |
| 
 | |
| 	mutex_lock(&kvm->lock);
 | |
| 	switch (type) {
 | |
| 	case KVM_VGIC_V2_ADDR_TYPE_DIST:
 | |
| 		r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V2);
 | |
| 		addr_ptr = &vgic->vgic_dist_base;
 | |
| 		alignment = SZ_4K;
 | |
| 		break;
 | |
| 	case KVM_VGIC_V2_ADDR_TYPE_CPU:
 | |
| 		r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V2);
 | |
| 		addr_ptr = &vgic->vgic_cpu_base;
 | |
| 		alignment = SZ_4K;
 | |
| 		break;
 | |
| 	case KVM_VGIC_V3_ADDR_TYPE_DIST:
 | |
| 		r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V3);
 | |
| 		addr_ptr = &vgic->vgic_dist_base;
 | |
| 		alignment = SZ_64K;
 | |
| 		break;
 | |
| 	case KVM_VGIC_V3_ADDR_TYPE_REDIST: {
 | |
| 		struct vgic_redist_region *rdreg;
 | |
| 
 | |
| 		r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V3);
 | |
| 		if (r)
 | |
| 			break;
 | |
| 		if (write) {
 | |
| 			r = vgic_v3_set_redist_base(kvm, 0, *addr, 0);
 | |
| 			goto out;
 | |
| 		}
 | |
| 		rdreg = list_first_entry(&vgic->rd_regions,
 | |
| 					 struct vgic_redist_region, list);
 | |
| 		if (!rdreg)
 | |
| 			addr_ptr = &undef_value;
 | |
| 		else
 | |
| 			addr_ptr = &rdreg->base;
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_VGIC_V3_ADDR_TYPE_REDIST_REGION:
 | |
| 	{
 | |
| 		struct vgic_redist_region *rdreg;
 | |
| 		u8 index;
 | |
| 
 | |
| 		r = vgic_check_type(kvm, KVM_DEV_TYPE_ARM_VGIC_V3);
 | |
| 		if (r)
 | |
| 			break;
 | |
| 
 | |
| 		index = *addr & KVM_VGIC_V3_RDIST_INDEX_MASK;
 | |
| 
 | |
| 		if (write) {
 | |
| 			gpa_t base = *addr & KVM_VGIC_V3_RDIST_BASE_MASK;
 | |
| 			u32 count = (*addr & KVM_VGIC_V3_RDIST_COUNT_MASK)
 | |
| 					>> KVM_VGIC_V3_RDIST_COUNT_SHIFT;
 | |
| 			u8 flags = (*addr & KVM_VGIC_V3_RDIST_FLAGS_MASK)
 | |
| 					>> KVM_VGIC_V3_RDIST_FLAGS_SHIFT;
 | |
| 
 | |
| 			if (!count || flags)
 | |
| 				r = -EINVAL;
 | |
| 			else
 | |
| 				r = vgic_v3_set_redist_base(kvm, index,
 | |
| 							    base, count);
 | |
| 			goto out;
 | |
| 		}
 | |
| 
 | |
| 		rdreg = vgic_v3_rdist_region_from_index(kvm, index);
 | |
| 		if (!rdreg) {
 | |
| 			r = -ENOENT;
 | |
| 			goto out;
 | |
| 		}
 | |
| 
 | |
| 		*addr = index;
 | |
| 		*addr |= rdreg->base;
 | |
| 		*addr |= (u64)rdreg->count << KVM_VGIC_V3_RDIST_COUNT_SHIFT;
 | |
| 		goto out;
 | |
| 	}
 | |
| 	default:
 | |
| 		r = -ENODEV;
 | |
| 	}
 | |
| 
 | |
| 	if (r)
 | |
| 		goto out;
 | |
| 
 | |
| 	if (write) {
 | |
| 		r = vgic_check_ioaddr(kvm, addr_ptr, *addr, alignment);
 | |
| 		if (!r)
 | |
| 			*addr_ptr = *addr;
 | |
| 	} else {
 | |
| 		*addr = *addr_ptr;
 | |
| 	}
 | |
| 
 | |
| out:
 | |
| 	mutex_unlock(&kvm->lock);
 | |
| 	return r;
 | |
| }
 | |
| 
 | |
| static int vgic_set_common_attr(struct kvm_device *dev,
 | |
| 				struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int r;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_ADDR: {
 | |
| 		u64 __user *uaddr = (u64 __user *)(long)attr->addr;
 | |
| 		u64 addr;
 | |
| 		unsigned long type = (unsigned long)attr->attr;
 | |
| 
 | |
| 		if (copy_from_user(&addr, uaddr, sizeof(addr)))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		r = kvm_vgic_addr(dev->kvm, type, &addr, true);
 | |
| 		return (r == -ENODEV) ? -ENXIO : r;
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_NR_IRQS: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 		u32 val;
 | |
| 		int ret = 0;
 | |
| 
 | |
| 		if (get_user(val, uaddr))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		/*
 | |
| 		 * We require:
 | |
| 		 * - at least 32 SPIs on top of the 16 SGIs and 16 PPIs
 | |
| 		 * - at most 1024 interrupts
 | |
| 		 * - a multiple of 32 interrupts
 | |
| 		 */
 | |
| 		if (val < (VGIC_NR_PRIVATE_IRQS + 32) ||
 | |
| 		    val > VGIC_MAX_RESERVED ||
 | |
| 		    (val & 31))
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		mutex_lock(&dev->kvm->lock);
 | |
| 
 | |
| 		if (vgic_ready(dev->kvm) || dev->kvm->arch.vgic.nr_spis)
 | |
| 			ret = -EBUSY;
 | |
| 		else
 | |
| 			dev->kvm->arch.vgic.nr_spis =
 | |
| 				val - VGIC_NR_PRIVATE_IRQS;
 | |
| 
 | |
| 		mutex_unlock(&dev->kvm->lock);
 | |
| 
 | |
| 		return ret;
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CTRL: {
 | |
| 		switch (attr->attr) {
 | |
| 		case KVM_DEV_ARM_VGIC_CTRL_INIT:
 | |
| 			mutex_lock(&dev->kvm->lock);
 | |
| 			r = vgic_init(dev->kvm);
 | |
| 			mutex_unlock(&dev->kvm->lock);
 | |
| 			return r;
 | |
| 		}
 | |
| 		break;
 | |
| 	}
 | |
| 	}
 | |
| 
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| static int vgic_get_common_attr(struct kvm_device *dev,
 | |
| 				struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int r = -ENXIO;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_ADDR: {
 | |
| 		u64 __user *uaddr = (u64 __user *)(long)attr->addr;
 | |
| 		u64 addr;
 | |
| 		unsigned long type = (unsigned long)attr->attr;
 | |
| 
 | |
| 		r = kvm_vgic_addr(dev->kvm, type, &addr, false);
 | |
| 		if (r)
 | |
| 			return (r == -ENODEV) ? -ENXIO : r;
 | |
| 
 | |
| 		if (copy_to_user(uaddr, &addr, sizeof(addr)))
 | |
| 			return -EFAULT;
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_NR_IRQS: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 
 | |
| 		r = put_user(dev->kvm->arch.vgic.nr_spis +
 | |
| 			     VGIC_NR_PRIVATE_IRQS, uaddr);
 | |
| 		break;
 | |
| 	}
 | |
| 	}
 | |
| 
 | |
| 	return r;
 | |
| }
 | |
| 
 | |
| static int vgic_create(struct kvm_device *dev, u32 type)
 | |
| {
 | |
| 	return kvm_vgic_create(dev->kvm, type);
 | |
| }
 | |
| 
 | |
| static void vgic_destroy(struct kvm_device *dev)
 | |
| {
 | |
| 	kfree(dev);
 | |
| }
 | |
| 
 | |
| int kvm_register_vgic_device(unsigned long type)
 | |
| {
 | |
| 	int ret = -ENODEV;
 | |
| 
 | |
| 	switch (type) {
 | |
| 	case KVM_DEV_TYPE_ARM_VGIC_V2:
 | |
| 		ret = kvm_register_device_ops(&kvm_arm_vgic_v2_ops,
 | |
| 					      KVM_DEV_TYPE_ARM_VGIC_V2);
 | |
| 		break;
 | |
| 	case KVM_DEV_TYPE_ARM_VGIC_V3:
 | |
| 		ret = kvm_register_device_ops(&kvm_arm_vgic_v3_ops,
 | |
| 					      KVM_DEV_TYPE_ARM_VGIC_V3);
 | |
| 
 | |
| 		if (ret)
 | |
| 			break;
 | |
| 		ret = kvm_vgic_register_its_device();
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int vgic_v2_parse_attr(struct kvm_device *dev, struct kvm_device_attr *attr,
 | |
| 		       struct vgic_reg_attr *reg_attr)
 | |
| {
 | |
| 	int cpuid;
 | |
| 
 | |
| 	cpuid = (attr->attr & KVM_DEV_ARM_VGIC_CPUID_MASK) >>
 | |
| 		 KVM_DEV_ARM_VGIC_CPUID_SHIFT;
 | |
| 
 | |
| 	if (cpuid >= atomic_read(&dev->kvm->online_vcpus))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	reg_attr->vcpu = kvm_get_vcpu(dev->kvm, cpuid);
 | |
| 	reg_attr->addr = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* unlocks vcpus from @vcpu_lock_idx and smaller */
 | |
| static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx)
 | |
| {
 | |
| 	struct kvm_vcpu *tmp_vcpu;
 | |
| 
 | |
| 	for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) {
 | |
| 		tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx);
 | |
| 		mutex_unlock(&tmp_vcpu->mutex);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void unlock_all_vcpus(struct kvm *kvm)
 | |
| {
 | |
| 	unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1);
 | |
| }
 | |
| 
 | |
| /* Returns true if all vcpus were locked, false otherwise */
 | |
| bool lock_all_vcpus(struct kvm *kvm)
 | |
| {
 | |
| 	struct kvm_vcpu *tmp_vcpu;
 | |
| 	int c;
 | |
| 
 | |
| 	/*
 | |
| 	 * Any time a vcpu is run, vcpu_load is called which tries to grab the
 | |
| 	 * vcpu->mutex.  By grabbing the vcpu->mutex of all VCPUs we ensure
 | |
| 	 * that no other VCPUs are run and fiddle with the vgic state while we
 | |
| 	 * access it.
 | |
| 	 */
 | |
| 	kvm_for_each_vcpu(c, tmp_vcpu, kvm) {
 | |
| 		if (!mutex_trylock(&tmp_vcpu->mutex)) {
 | |
| 			unlock_vcpus(kvm, c - 1);
 | |
| 			return false;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * vgic_v2_attr_regs_access - allows user space to access VGIC v2 state
 | |
|  *
 | |
|  * @dev:      kvm device handle
 | |
|  * @attr:     kvm device attribute
 | |
|  * @reg:      address the value is read or written
 | |
|  * @is_write: true if userspace is writing a register
 | |
|  */
 | |
| static int vgic_v2_attr_regs_access(struct kvm_device *dev,
 | |
| 				    struct kvm_device_attr *attr,
 | |
| 				    u32 *reg, bool is_write)
 | |
| {
 | |
| 	struct vgic_reg_attr reg_attr;
 | |
| 	gpa_t addr;
 | |
| 	struct kvm_vcpu *vcpu;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = vgic_v2_parse_attr(dev, attr, ®_attr);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	vcpu = reg_attr.vcpu;
 | |
| 	addr = reg_attr.addr;
 | |
| 
 | |
| 	mutex_lock(&dev->kvm->lock);
 | |
| 
 | |
| 	ret = vgic_init(dev->kvm);
 | |
| 	if (ret)
 | |
| 		goto out;
 | |
| 
 | |
| 	if (!lock_all_vcpus(dev->kvm)) {
 | |
| 		ret = -EBUSY;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
 | |
| 		ret = vgic_v2_cpuif_uaccess(vcpu, is_write, addr, reg);
 | |
| 		break;
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 		ret = vgic_v2_dist_uaccess(vcpu, is_write, addr, reg);
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = -EINVAL;
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	unlock_all_vcpus(dev->kvm);
 | |
| out:
 | |
| 	mutex_unlock(&dev->kvm->lock);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int vgic_v2_set_attr(struct kvm_device *dev,
 | |
| 			    struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = vgic_set_common_attr(dev, attr);
 | |
| 	if (ret != -ENXIO)
 | |
| 		return ret;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 		u32 reg;
 | |
| 
 | |
| 		if (get_user(reg, uaddr))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		return vgic_v2_attr_regs_access(dev, attr, ®, true);
 | |
| 	}
 | |
| 	}
 | |
| 
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| static int vgic_v2_get_attr(struct kvm_device *dev,
 | |
| 			    struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = vgic_get_common_attr(dev, attr);
 | |
| 	if (ret != -ENXIO)
 | |
| 		return ret;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 		u32 reg = 0;
 | |
| 
 | |
| 		ret = vgic_v2_attr_regs_access(dev, attr, ®, false);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		return put_user(reg, uaddr);
 | |
| 	}
 | |
| 	}
 | |
| 
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| static int vgic_v2_has_attr(struct kvm_device *dev,
 | |
| 			    struct kvm_device_attr *attr)
 | |
| {
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_ADDR:
 | |
| 		switch (attr->attr) {
 | |
| 		case KVM_VGIC_V2_ADDR_TYPE_DIST:
 | |
| 		case KVM_VGIC_V2_ADDR_TYPE_CPU:
 | |
| 			return 0;
 | |
| 		}
 | |
| 		break;
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
 | |
| 		return vgic_v2_has_attr_regs(dev, attr);
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_NR_IRQS:
 | |
| 		return 0;
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CTRL:
 | |
| 		switch (attr->attr) {
 | |
| 		case KVM_DEV_ARM_VGIC_CTRL_INIT:
 | |
| 			return 0;
 | |
| 		}
 | |
| 	}
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| struct kvm_device_ops kvm_arm_vgic_v2_ops = {
 | |
| 	.name = "kvm-arm-vgic-v2",
 | |
| 	.create = vgic_create,
 | |
| 	.destroy = vgic_destroy,
 | |
| 	.set_attr = vgic_v2_set_attr,
 | |
| 	.get_attr = vgic_v2_get_attr,
 | |
| 	.has_attr = vgic_v2_has_attr,
 | |
| };
 | |
| 
 | |
| int vgic_v3_parse_attr(struct kvm_device *dev, struct kvm_device_attr *attr,
 | |
| 		       struct vgic_reg_attr *reg_attr)
 | |
| {
 | |
| 	unsigned long vgic_mpidr, mpidr_reg;
 | |
| 
 | |
| 	/*
 | |
| 	 * For KVM_DEV_ARM_VGIC_GRP_DIST_REGS group,
 | |
| 	 * attr might not hold MPIDR. Hence assume vcpu0.
 | |
| 	 */
 | |
| 	if (attr->group != KVM_DEV_ARM_VGIC_GRP_DIST_REGS) {
 | |
| 		vgic_mpidr = (attr->attr & KVM_DEV_ARM_VGIC_V3_MPIDR_MASK) >>
 | |
| 			      KVM_DEV_ARM_VGIC_V3_MPIDR_SHIFT;
 | |
| 
 | |
| 		mpidr_reg = VGIC_TO_MPIDR(vgic_mpidr);
 | |
| 		reg_attr->vcpu = kvm_mpidr_to_vcpu(dev->kvm, mpidr_reg);
 | |
| 	} else {
 | |
| 		reg_attr->vcpu = kvm_get_vcpu(dev->kvm, 0);
 | |
| 	}
 | |
| 
 | |
| 	if (!reg_attr->vcpu)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	reg_attr->addr = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * vgic_v3_attr_regs_access - allows user space to access VGIC v3 state
 | |
|  *
 | |
|  * @dev:      kvm device handle
 | |
|  * @attr:     kvm device attribute
 | |
|  * @reg:      address the value is read or written
 | |
|  * @is_write: true if userspace is writing a register
 | |
|  */
 | |
| static int vgic_v3_attr_regs_access(struct kvm_device *dev,
 | |
| 				    struct kvm_device_attr *attr,
 | |
| 				    u64 *reg, bool is_write)
 | |
| {
 | |
| 	struct vgic_reg_attr reg_attr;
 | |
| 	gpa_t addr;
 | |
| 	struct kvm_vcpu *vcpu;
 | |
| 	int ret;
 | |
| 	u32 tmp32;
 | |
| 
 | |
| 	ret = vgic_v3_parse_attr(dev, attr, ®_attr);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	vcpu = reg_attr.vcpu;
 | |
| 	addr = reg_attr.addr;
 | |
| 
 | |
| 	mutex_lock(&dev->kvm->lock);
 | |
| 
 | |
| 	if (unlikely(!vgic_initialized(dev->kvm))) {
 | |
| 		ret = -EBUSY;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	if (!lock_all_vcpus(dev->kvm)) {
 | |
| 		ret = -EBUSY;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 		if (is_write)
 | |
| 			tmp32 = *reg;
 | |
| 
 | |
| 		ret = vgic_v3_dist_uaccess(vcpu, is_write, addr, &tmp32);
 | |
| 		if (!is_write)
 | |
| 			*reg = tmp32;
 | |
| 		break;
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS:
 | |
| 		if (is_write)
 | |
| 			tmp32 = *reg;
 | |
| 
 | |
| 		ret = vgic_v3_redist_uaccess(vcpu, is_write, addr, &tmp32);
 | |
| 		if (!is_write)
 | |
| 			*reg = tmp32;
 | |
| 		break;
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS: {
 | |
| 		u64 regid;
 | |
| 
 | |
| 		regid = (attr->attr & KVM_DEV_ARM_VGIC_SYSREG_INSTR_MASK);
 | |
| 		ret = vgic_v3_cpu_sysregs_uaccess(vcpu, is_write,
 | |
| 						  regid, reg);
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: {
 | |
| 		unsigned int info, intid;
 | |
| 
 | |
| 		info = (attr->attr & KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_MASK) >>
 | |
| 			KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_SHIFT;
 | |
| 		if (info == VGIC_LEVEL_INFO_LINE_LEVEL) {
 | |
| 			intid = attr->attr &
 | |
| 				KVM_DEV_ARM_VGIC_LINE_LEVEL_INTID_MASK;
 | |
| 			ret = vgic_v3_line_level_info_uaccess(vcpu, is_write,
 | |
| 							      intid, reg);
 | |
| 		} else {
 | |
| 			ret = -EINVAL;
 | |
| 		}
 | |
| 		break;
 | |
| 	}
 | |
| 	default:
 | |
| 		ret = -EINVAL;
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	unlock_all_vcpus(dev->kvm);
 | |
| out:
 | |
| 	mutex_unlock(&dev->kvm->lock);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int vgic_v3_set_attr(struct kvm_device *dev,
 | |
| 			    struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = vgic_set_common_attr(dev, attr);
 | |
| 	if (ret != -ENXIO)
 | |
| 		return ret;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 		u32 tmp32;
 | |
| 		u64 reg;
 | |
| 
 | |
| 		if (get_user(tmp32, uaddr))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		reg = tmp32;
 | |
| 		return vgic_v3_attr_regs_access(dev, attr, ®, true);
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS: {
 | |
| 		u64 __user *uaddr = (u64 __user *)(long)attr->addr;
 | |
| 		u64 reg;
 | |
| 
 | |
| 		if (get_user(reg, uaddr))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		return vgic_v3_attr_regs_access(dev, attr, ®, true);
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 		u64 reg;
 | |
| 		u32 tmp32;
 | |
| 
 | |
| 		if (get_user(tmp32, uaddr))
 | |
| 			return -EFAULT;
 | |
| 
 | |
| 		reg = tmp32;
 | |
| 		return vgic_v3_attr_regs_access(dev, attr, ®, true);
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CTRL: {
 | |
| 		int ret;
 | |
| 
 | |
| 		switch (attr->attr) {
 | |
| 		case KVM_DEV_ARM_VGIC_SAVE_PENDING_TABLES:
 | |
| 			mutex_lock(&dev->kvm->lock);
 | |
| 
 | |
| 			if (!lock_all_vcpus(dev->kvm)) {
 | |
| 				mutex_unlock(&dev->kvm->lock);
 | |
| 				return -EBUSY;
 | |
| 			}
 | |
| 			ret = vgic_v3_save_pending_tables(dev->kvm);
 | |
| 			unlock_all_vcpus(dev->kvm);
 | |
| 			mutex_unlock(&dev->kvm->lock);
 | |
| 			return ret;
 | |
| 		}
 | |
| 		break;
 | |
| 	}
 | |
| 	}
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| static int vgic_v3_get_attr(struct kvm_device *dev,
 | |
| 			    struct kvm_device_attr *attr)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = vgic_get_common_attr(dev, attr);
 | |
| 	if (ret != -ENXIO)
 | |
| 		return ret;
 | |
| 
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 		u64 reg;
 | |
| 		u32 tmp32;
 | |
| 
 | |
| 		ret = vgic_v3_attr_regs_access(dev, attr, ®, false);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		tmp32 = reg;
 | |
| 		return put_user(tmp32, uaddr);
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS: {
 | |
| 		u64 __user *uaddr = (u64 __user *)(long)attr->addr;
 | |
| 		u64 reg;
 | |
| 
 | |
| 		ret = vgic_v3_attr_regs_access(dev, attr, ®, false);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		return put_user(reg, uaddr);
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: {
 | |
| 		u32 __user *uaddr = (u32 __user *)(long)attr->addr;
 | |
| 		u64 reg;
 | |
| 		u32 tmp32;
 | |
| 
 | |
| 		ret = vgic_v3_attr_regs_access(dev, attr, ®, false);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		tmp32 = reg;
 | |
| 		return put_user(tmp32, uaddr);
 | |
| 	}
 | |
| 	}
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| static int vgic_v3_has_attr(struct kvm_device *dev,
 | |
| 			    struct kvm_device_attr *attr)
 | |
| {
 | |
| 	switch (attr->group) {
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_ADDR:
 | |
| 		switch (attr->attr) {
 | |
| 		case KVM_VGIC_V3_ADDR_TYPE_DIST:
 | |
| 		case KVM_VGIC_V3_ADDR_TYPE_REDIST:
 | |
| 		case KVM_VGIC_V3_ADDR_TYPE_REDIST_REGION:
 | |
| 			return 0;
 | |
| 		}
 | |
| 		break;
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_REDIST_REGS:
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CPU_SYSREGS:
 | |
| 		return vgic_v3_has_attr_regs(dev, attr);
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_NR_IRQS:
 | |
| 		return 0;
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_LEVEL_INFO: {
 | |
| 		if (((attr->attr & KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_MASK) >>
 | |
| 		      KVM_DEV_ARM_VGIC_LINE_LEVEL_INFO_SHIFT) ==
 | |
| 		      VGIC_LEVEL_INFO_LINE_LEVEL)
 | |
| 			return 0;
 | |
| 		break;
 | |
| 	}
 | |
| 	case KVM_DEV_ARM_VGIC_GRP_CTRL:
 | |
| 		switch (attr->attr) {
 | |
| 		case KVM_DEV_ARM_VGIC_CTRL_INIT:
 | |
| 			return 0;
 | |
| 		case KVM_DEV_ARM_VGIC_SAVE_PENDING_TABLES:
 | |
| 			return 0;
 | |
| 		}
 | |
| 	}
 | |
| 	return -ENXIO;
 | |
| }
 | |
| 
 | |
| struct kvm_device_ops kvm_arm_vgic_v3_ops = {
 | |
| 	.name = "kvm-arm-vgic-v3",
 | |
| 	.create = vgic_create,
 | |
| 	.destroy = vgic_destroy,
 | |
| 	.set_attr = vgic_v3_set_attr,
 | |
| 	.get_attr = vgic_v3_get_attr,
 | |
| 	.has_attr = vgic_v3_has_attr,
 | |
| };
 | 
