KVM: x86/pmu: do not mask the value that is written to fixed PMUs
[sfrench/cifs-2.6.git] / arch / x86 / kvm / vmx / pmu_intel.c
index 5ab4a364348e3c10987c33203be4ff6fa97e1e73..a99613a060dd8cee8790a3213cfc826c506fc784 100644 (file)
@@ -126,7 +126,7 @@ static int intel_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx)
 }
 
 static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu,
-                                           unsigned idx)
+                                           unsigned idx, u64 *mask)
 {
        struct kvm_pmu *pmu = vcpu_to_pmu(vcpu);
        bool fixed = idx & (1u << 30);
@@ -138,6 +138,7 @@ static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu,
        if (fixed && idx >= pmu->nr_arch_fixed_counters)
                return NULL;
        counters = fixed ? pmu->fixed_counters : pmu->gp_counters;
+       *mask &= pmu->counter_bitmask[fixed ? KVM_PMC_FIXED : KVM_PMC_GP];
 
        return &counters[idx];
 }
@@ -183,9 +184,13 @@ static int intel_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data)
                *data = pmu->global_ovf_ctrl;
                return 0;
        default:
-               if ((pmc = get_gp_pmc(pmu, msr, MSR_IA32_PERFCTR0)) ||
-                   (pmc = get_fixed_pmc(pmu, msr))) {
-                       *data = pmc_read_counter(pmc);
+               if ((pmc = get_gp_pmc(pmu, msr, MSR_IA32_PERFCTR0))) {
+                       u64 val = pmc_read_counter(pmc);
+                       *data = val & pmu->counter_bitmask[KVM_PMC_GP];
+                       return 0;
+               } else if ((pmc = get_fixed_pmc(pmu, msr))) {
+                       u64 val = pmc_read_counter(pmc);
+                       *data = val & pmu->counter_bitmask[KVM_PMC_FIXED];
                        return 0;
                } else if ((pmc = get_gp_pmc(pmu, msr, MSR_P6_EVNTSEL0))) {
                        *data = pmc->eventsel;
@@ -227,7 +232,7 @@ static int intel_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
                }
                break;
        case MSR_CORE_PERF_GLOBAL_OVF_CTRL:
-               if (!(data & (pmu->global_ctrl_mask & ~(3ull<<62)))) {
+               if (!(data & pmu->global_ovf_ctrl_mask)) {
                        if (!msr_info->host_initiated)
                                pmu->global_status &= ~data;
                        pmu->global_ovf_ctrl = data;
@@ -235,11 +240,14 @@ static int intel_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
                }
                break;
        default:
-               if ((pmc = get_gp_pmc(pmu, msr, MSR_IA32_PERFCTR0)) ||
-                   (pmc = get_fixed_pmc(pmu, msr))) {
-                       if (!msr_info->host_initiated)
-                               data = (s64)(s32)data;
-                       pmc->counter += data - pmc_read_counter(pmc);
+               if ((pmc = get_gp_pmc(pmu, msr, MSR_IA32_PERFCTR0))) {
+                       if (msr_info->host_initiated)
+                               pmc->counter = data;
+                       else
+                               pmc->counter = (s32)data;
+                       return 0;
+               } else if ((pmc = get_fixed_pmc(pmu, msr))) {
+                       pmc->counter = data;
                        return 0;
                } else if ((pmc = get_gp_pmc(pmu, msr, MSR_P6_EVNTSEL0))) {
                        if (data == pmc->eventsel)
@@ -297,6 +305,12 @@ static void intel_pmu_refresh(struct kvm_vcpu *vcpu)
        pmu->global_ctrl = ((1ull << pmu->nr_arch_gp_counters) - 1) |
                (((1ull << pmu->nr_arch_fixed_counters) - 1) << INTEL_PMC_IDX_FIXED);
        pmu->global_ctrl_mask = ~pmu->global_ctrl;
+       pmu->global_ovf_ctrl_mask = pmu->global_ctrl_mask
+                       & ~(MSR_CORE_PERF_GLOBAL_OVF_CTRL_OVF_BUF |
+                           MSR_CORE_PERF_GLOBAL_OVF_CTRL_COND_CHGD);
+       if (kvm_x86_ops->pt_supported())
+               pmu->global_ovf_ctrl_mask &=
+                               ~MSR_CORE_PERF_GLOBAL_OVF_CTRL_TRACE_TOPA_PMI;
 
        entry = kvm_find_cpuid_entry(vcpu, 7, 0);
        if (entry &&