Merge tag 'for-linus' of git://git.kernel.org/pub/scm/virt/kvm/kvm
[linux-2.6-microblaze.git] / arch / arm64 / kvm / arm.c
index 4ea6c22..e5f75f1 100644 (file)
@@ -205,6 +205,7 @@ void kvm_arch_destroy_vm(struct kvm *kvm)
        if (is_protected_kvm_enabled())
                pkvm_destroy_hyp_vm(kvm);
 
+       kfree(kvm->arch.mpidr_data);
        kvm_destroy_vcpus(kvm);
 
        kvm_unshare_hyp(kvm, kvm + 1);
@@ -317,6 +318,9 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext)
        case KVM_CAP_ARM_SUPPORTED_BLOCK_SIZES:
                r = kvm_supported_block_sizes();
                break;
+       case KVM_CAP_ARM_SUPPORTED_REG_MASK_RANGES:
+               r = BIT(0);
+               break;
        default:
                r = 0;
        }
@@ -367,7 +371,6 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
 
        /* Force users to call KVM_ARM_VCPU_INIT */
        vcpu_clear_flag(vcpu, VCPU_INITIALIZED);
-       bitmap_zero(vcpu->arch.features, KVM_VCPU_MAX_FEATURES);
 
        vcpu->arch.mmu_page_cache.gfp_zero = __GFP_ZERO;
 
@@ -438,9 +441,9 @@ void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
         * We might get preempted before the vCPU actually runs, but
         * over-invalidation doesn't affect correctness.
         */
-       if (*last_ran != vcpu->vcpu_id) {
+       if (*last_ran != vcpu->vcpu_idx) {
                kvm_call_hyp(__kvm_flush_cpu_context, mmu);
-               *last_ran = vcpu->vcpu_id;
+               *last_ran = vcpu->vcpu_idx;
        }
 
        vcpu->cpu = cpu;
@@ -448,7 +451,7 @@ void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
        kvm_vgic_load(vcpu);
        kvm_timer_vcpu_load(vcpu);
        if (has_vhe())
-               kvm_vcpu_load_sysregs_vhe(vcpu);
+               kvm_vcpu_load_vhe(vcpu);
        kvm_arch_vcpu_load_fp(vcpu);
        kvm_vcpu_pmu_restore_guest(vcpu);
        if (kvm_arm_is_pvtime_enabled(&vcpu->arch))
@@ -472,7 +475,7 @@ void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
        kvm_arch_vcpu_put_debug_state_flags(vcpu);
        kvm_arch_vcpu_put_fp(vcpu);
        if (has_vhe())
-               kvm_vcpu_put_sysregs_vhe(vcpu);
+               kvm_vcpu_put_vhe(vcpu);
        kvm_timer_vcpu_put(vcpu);
        kvm_vgic_put(vcpu);
        kvm_vcpu_pmu_restore_host(vcpu);
@@ -578,6 +581,57 @@ static int kvm_vcpu_initialized(struct kvm_vcpu *vcpu)
        return vcpu_get_flag(vcpu, VCPU_INITIALIZED);
 }
 
+static void kvm_init_mpidr_data(struct kvm *kvm)
+{
+       struct kvm_mpidr_data *data = NULL;
+       unsigned long c, mask, nr_entries;
+       u64 aff_set = 0, aff_clr = ~0UL;
+       struct kvm_vcpu *vcpu;
+
+       mutex_lock(&kvm->arch.config_lock);
+
+       if (kvm->arch.mpidr_data || atomic_read(&kvm->online_vcpus) == 1)
+               goto out;
+
+       kvm_for_each_vcpu(c, vcpu, kvm) {
+               u64 aff = kvm_vcpu_get_mpidr_aff(vcpu);
+               aff_set |= aff;
+               aff_clr &= aff;
+       }
+
+       /*
+        * A significant bit can be either 0 or 1, and will only appear in
+        * aff_set. Use aff_clr to weed out the useless stuff.
+        */
+       mask = aff_set ^ aff_clr;
+       nr_entries = BIT_ULL(hweight_long(mask));
+
+       /*
+        * Don't let userspace fool us. If we need more than a single page
+        * to describe the compressed MPIDR array, just fall back to the
+        * iterative method. Single vcpu VMs do not need this either.
+        */
+       if (struct_size(data, cmpidr_to_idx, nr_entries) <= PAGE_SIZE)
+               data = kzalloc(struct_size(data, cmpidr_to_idx, nr_entries),
+                              GFP_KERNEL_ACCOUNT);
+
+       if (!data)
+               goto out;
+
+       data->mpidr_mask = mask;
+
+       kvm_for_each_vcpu(c, vcpu, kvm) {
+               u64 aff = kvm_vcpu_get_mpidr_aff(vcpu);
+               u16 index = kvm_mpidr_index(data, aff);
+
+               data->cmpidr_to_idx[index] = c;
+       }
+
+       kvm->arch.mpidr_data = data;
+out:
+       mutex_unlock(&kvm->arch.config_lock);
+}
+
 /*
  * Handle both the initialisation that is being done when the vcpu is
  * run for the first time, as well as the updates that must be
@@ -601,6 +655,8 @@ int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu)
        if (likely(vcpu_has_run_once(vcpu)))
                return 0;
 
+       kvm_init_mpidr_data(kvm);
+
        kvm_arm_vcpu_init_debug(vcpu);
 
        if (likely(irqchip_in_kernel(kvm))) {
@@ -801,8 +857,7 @@ static int check_vcpu_requests(struct kvm_vcpu *vcpu)
                }
 
                if (kvm_check_request(KVM_REQ_RELOAD_PMU, vcpu))
-                       kvm_pmu_handle_pmcr(vcpu,
-                                           __vcpu_sys_reg(vcpu, PMCR_EL0));
+                       kvm_vcpu_reload_pmu(vcpu);
 
                if (kvm_check_request(KVM_REQ_RESYNC_PMU_EL0, vcpu))
                        kvm_vcpu_pmu_restore_guest(vcpu);
@@ -950,7 +1005,10 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
                 * making a thread's VMID inactive. So we need to call
                 * kvm_arm_vmid_update() in non-premptible context.
                 */
-               kvm_arm_vmid_update(&vcpu->arch.hw_mmu->vmid);
+               if (kvm_arm_vmid_update(&vcpu->arch.hw_mmu->vmid) &&
+                   has_vhe())
+                       __load_stage2(vcpu->arch.hw_mmu,
+                                     vcpu->arch.hw_mmu->arch);
 
                kvm_pmu_flush_hwstate(vcpu);
 
@@ -1134,27 +1192,23 @@ int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irq_level,
                          bool line_status)
 {
        u32 irq = irq_level->irq;
-       unsigned int irq_type, vcpu_idx, irq_num;
-       int nrcpus = atomic_read(&kvm->online_vcpus);
+       unsigned int irq_type, vcpu_id, irq_num;
        struct kvm_vcpu *vcpu = NULL;
        bool level = irq_level->level;
 
        irq_type = (irq >> KVM_ARM_IRQ_TYPE_SHIFT) & KVM_ARM_IRQ_TYPE_MASK;
-       vcpu_idx = (irq >> KVM_ARM_IRQ_VCPU_SHIFT) & KVM_ARM_IRQ_VCPU_MASK;
-       vcpu_idx += ((irq >> KVM_ARM_IRQ_VCPU2_SHIFT) & KVM_ARM_IRQ_VCPU2_MASK) * (KVM_ARM_IRQ_VCPU_MASK + 1);
+       vcpu_id = (irq >> KVM_ARM_IRQ_VCPU_SHIFT) & KVM_ARM_IRQ_VCPU_MASK;
+       vcpu_id += ((irq >> KVM_ARM_IRQ_VCPU2_SHIFT) & KVM_ARM_IRQ_VCPU2_MASK) * (KVM_ARM_IRQ_VCPU_MASK + 1);
        irq_num = (irq >> KVM_ARM_IRQ_NUM_SHIFT) & KVM_ARM_IRQ_NUM_MASK;
 
-       trace_kvm_irq_line(irq_type, vcpu_idx, irq_num, irq_level->level);
+       trace_kvm_irq_line(irq_type, vcpu_id, irq_num, irq_level->level);
 
        switch (irq_type) {
        case KVM_ARM_IRQ_TYPE_CPU:
                if (irqchip_in_kernel(kvm))
                        return -ENXIO;
 
-               if (vcpu_idx >= nrcpus)
-                       return -EINVAL;
-
-               vcpu = kvm_get_vcpu(kvm, vcpu_idx);
+               vcpu = kvm_get_vcpu_by_id(kvm, vcpu_id);
                if (!vcpu)
                        return -EINVAL;
 
@@ -1166,17 +1220,14 @@ int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irq_level,
                if (!irqchip_in_kernel(kvm))
                        return -ENXIO;
 
-               if (vcpu_idx >= nrcpus)
-                       return -EINVAL;
-
-               vcpu = kvm_get_vcpu(kvm, vcpu_idx);
+               vcpu = kvm_get_vcpu_by_id(kvm, vcpu_id);
                if (!vcpu)
                        return -EINVAL;
 
                if (irq_num < VGIC_NR_SGIS || irq_num >= VGIC_NR_PRIVATE_IRQS)
                        return -EINVAL;
 
-               return kvm_vgic_inject_irq(kvm, vcpu->vcpu_id, irq_num, level, NULL);
+               return kvm_vgic_inject_irq(kvm, vcpu, irq_num, level, NULL);
        case KVM_ARM_IRQ_TYPE_SPI:
                if (!irqchip_in_kernel(kvm))
                        return -ENXIO;
@@ -1184,12 +1235,36 @@ int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irq_level,
                if (irq_num < VGIC_NR_PRIVATE_IRQS)
                        return -EINVAL;
 
-               return kvm_vgic_inject_irq(kvm, 0, irq_num, level, NULL);
+               return kvm_vgic_inject_irq(kvm, NULL, irq_num, level, NULL);
        }
 
        return -EINVAL;
 }
 
+static unsigned long system_supported_vcpu_features(void)
+{
+       unsigned long features = KVM_VCPU_VALID_FEATURES;
+
+       if (!cpus_have_final_cap(ARM64_HAS_32BIT_EL1))
+               clear_bit(KVM_ARM_VCPU_EL1_32BIT, &features);
+
+       if (!kvm_arm_support_pmu_v3())
+               clear_bit(KVM_ARM_VCPU_PMU_V3, &features);
+
+       if (!system_supports_sve())
+               clear_bit(KVM_ARM_VCPU_SVE, &features);
+
+       if (!system_has_full_ptr_auth()) {
+               clear_bit(KVM_ARM_VCPU_PTRAUTH_ADDRESS, &features);
+               clear_bit(KVM_ARM_VCPU_PTRAUTH_GENERIC, &features);
+       }
+
+       if (!cpus_have_final_cap(ARM64_HAS_NESTED_VIRT))
+               clear_bit(KVM_ARM_VCPU_HAS_EL2, &features);
+
+       return features;
+}
+
 static int kvm_vcpu_init_check_features(struct kvm_vcpu *vcpu,
                                        const struct kvm_vcpu_init *init)
 {
@@ -1204,12 +1279,25 @@ static int kvm_vcpu_init_check_features(struct kvm_vcpu *vcpu,
                        return -ENOENT;
        }
 
-       if (!test_bit(KVM_ARM_VCPU_EL1_32BIT, &features))
-               return 0;
+       if (features & ~system_supported_vcpu_features())
+               return -EINVAL;
 
-       if (!cpus_have_final_cap(ARM64_HAS_32BIT_EL1))
+       /*
+        * For now make sure that both address/generic pointer authentication
+        * features are requested by the userspace together.
+        */
+       if (test_bit(KVM_ARM_VCPU_PTRAUTH_ADDRESS, &features) !=
+           test_bit(KVM_ARM_VCPU_PTRAUTH_GENERIC, &features))
+               return -EINVAL;
+
+       /* Disallow NV+SVE for the time being */
+       if (test_bit(KVM_ARM_VCPU_HAS_EL2, &features) &&
+           test_bit(KVM_ARM_VCPU_SVE, &features))
                return -EINVAL;
 
+       if (!test_bit(KVM_ARM_VCPU_EL1_32BIT, &features))
+               return 0;
+
        /* MTE is incompatible with AArch32 */
        if (kvm_has_mte(vcpu->kvm))
                return -EINVAL;
@@ -1226,7 +1314,23 @@ static bool kvm_vcpu_init_changed(struct kvm_vcpu *vcpu,
 {
        unsigned long features = init->features[0];
 
-       return !bitmap_equal(vcpu->arch.features, &features, KVM_VCPU_MAX_FEATURES);
+       return !bitmap_equal(vcpu->kvm->arch.vcpu_features, &features,
+                            KVM_VCPU_MAX_FEATURES);
+}
+
+static int kvm_setup_vcpu(struct kvm_vcpu *vcpu)
+{
+       struct kvm *kvm = vcpu->kvm;
+       int ret = 0;
+
+       /*
+        * When the vCPU has a PMU, but no PMU is set for the guest
+        * yet, set the default one.
+        */
+       if (kvm_vcpu_has_pmu(vcpu) && !kvm->arch.arm_pmu)
+               ret = kvm_arm_set_default_pmu(kvm);
+
+       return ret;
 }
 
 static int __kvm_vcpu_set_target(struct kvm_vcpu *vcpu,
@@ -1239,21 +1343,21 @@ static int __kvm_vcpu_set_target(struct kvm_vcpu *vcpu,
        mutex_lock(&kvm->arch.config_lock);
 
        if (test_bit(KVM_ARCH_FLAG_VCPU_FEATURES_CONFIGURED, &kvm->arch.flags) &&
-           !bitmap_equal(kvm->arch.vcpu_features, &features, KVM_VCPU_MAX_FEATURES))
+           kvm_vcpu_init_changed(vcpu, init))
                goto out_unlock;
 
-       bitmap_copy(vcpu->arch.features, &features, KVM_VCPU_MAX_FEATURES);
+       bitmap_copy(kvm->arch.vcpu_features, &features, KVM_VCPU_MAX_FEATURES);
 
-       /* Now we know what it is, we can reset it. */
-       ret = kvm_reset_vcpu(vcpu);
-       if (ret) {
-               bitmap_zero(vcpu->arch.features, KVM_VCPU_MAX_FEATURES);
+       ret = kvm_setup_vcpu(vcpu);
+       if (ret)
                goto out_unlock;
-       }
 
-       bitmap_copy(kvm->arch.vcpu_features, &features, KVM_VCPU_MAX_FEATURES);
+       /* Now we know what it is, we can reset it. */
+       kvm_reset_vcpu(vcpu);
+
        set_bit(KVM_ARCH_FLAG_VCPU_FEATURES_CONFIGURED, &kvm->arch.flags);
        vcpu_set_flag(vcpu, VCPU_INITIALIZED);
+       ret = 0;
 out_unlock:
        mutex_unlock(&kvm->arch.config_lock);
        return ret;
@@ -1278,7 +1382,8 @@ static int kvm_vcpu_set_target(struct kvm_vcpu *vcpu,
        if (kvm_vcpu_init_changed(vcpu, init))
                return -EINVAL;
 
-       return kvm_reset_vcpu(vcpu);
+       kvm_reset_vcpu(vcpu);
+       return 0;
 }
 
 static int kvm_arch_vcpu_ioctl_vcpu_init(struct kvm_vcpu *vcpu,
@@ -1629,6 +1734,13 @@ int kvm_arch_vm_ioctl(struct file *filp, unsigned int ioctl, unsigned long arg)
 
                return kvm_vm_set_attr(kvm, &attr);
        }
+       case KVM_ARM_GET_REG_WRITABLE_MASKS: {
+               struct reg_mask_range range;
+
+               if (copy_from_user(&range, argp, sizeof(range)))
+                       return -EFAULT;
+               return kvm_vm_ioctl_get_reg_writable_masks(kvm, &range);
+       }
        default:
                return -EINVAL;
        }
@@ -2341,6 +2453,18 @@ struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr)
        unsigned long i;
 
        mpidr &= MPIDR_HWID_BITMASK;
+
+       if (kvm->arch.mpidr_data) {
+               u16 idx = kvm_mpidr_index(kvm->arch.mpidr_data, mpidr);
+
+               vcpu = kvm_get_vcpu(kvm,
+                                   kvm->arch.mpidr_data->cmpidr_to_idx[idx]);
+               if (mpidr != kvm_vcpu_get_mpidr_aff(vcpu))
+                       vcpu = NULL;
+
+               return vcpu;
+       }
+
        kvm_for_each_vcpu(i, vcpu, kvm) {
                if (mpidr == kvm_vcpu_get_mpidr_aff(vcpu))
                        return vcpu;