Revert "KVM: arm64: Snapshot all non-zero RES0/RES1 sysreg fields for later checking"
[sfrench/cifs-2.6.git] / arch / powerpc / perf / power10-pmu.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Performance counter support for POWER10 processors.
4  *
5  * Copyright 2020 Madhavan Srinivasan, IBM Corporation.
6  * Copyright 2020 Athira Rajeev, IBM Corporation.
7  */
8
9 #define pr_fmt(fmt)     "power10-pmu: " fmt
10
11 #include "isa207-common.h"
12
13 /*
14  * Raw event encoding for Power10:
15  *
16  *        60        56        52        48        44        40        36        32
17  * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
18  *   | | [ ]   [ src_match ] [  src_mask ]   | [ ] [ l2l3_sel ]  [  thresh_ctl   ]
19  *   | |  |                                  |  |                         |
20  *   | |  *- IFM (Linux)                     |  |        thresh start/stop -*
21  *   | *- BHRB (Linux)                       |  src_sel
22  *   *- EBB (Linux)                          *invert_bit
23  *
24  *        28        24        20        16        12         8         4         0
25  * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
26  *   [   ] [  sample ]   [ ] [ ]   [ pmc ]   [unit ]   [ ] |  m   [    pmcxsel    ]
27  *     |        |        |    |                        |   |  |
28  *     |        |        |    |                        |   |  *- mark
29  *     |        |        |    *- L1/L2/L3 cache_sel    |   |*-radix_scope_qual
30  *     |        |        sdar_mode                     |
31  *     |        *- sampling mode for marked events     *- combine
32  *     |
33  *     *- thresh_sel
34  *
35  * Below uses IBM bit numbering.
36  *
37  * MMCR1[x:y] = unit    (PMCxUNIT)
38  * MMCR1[24]   = pmc1combine[0]
39  * MMCR1[25]   = pmc1combine[1]
40  * MMCR1[26]   = pmc2combine[0]
41  * MMCR1[27]   = pmc2combine[1]
42  * MMCR1[28]   = pmc3combine[0]
43  * MMCR1[29]   = pmc3combine[1]
44  * MMCR1[30]   = pmc4combine[0]
45  * MMCR1[31]   = pmc4combine[1]
46  *
47  * if pmc == 3 and unit == 0 and pmcxsel[0:6] == 0b0101011
48  *      MMCR1[20:27] = thresh_ctl
49  * else if pmc == 4 and unit == 0xf and pmcxsel[0:6] == 0b0101001
50  *      MMCR1[20:27] = thresh_ctl
51  * else
52  *      MMCRA[48:55] = thresh_ctl   (THRESH START/END)
53  *
54  * if thresh_sel:
55  *      MMCRA[45:47] = thresh_sel
56  *
57  * if l2l3_sel:
58  * MMCR2[56:60] = l2l3_sel[0:4]
59  *
60  * MMCR1[16] = cache_sel[0]
61  * MMCR1[17] = cache_sel[1]
62  * MMCR1[18] = radix_scope_qual
63  *
64  * if mark:
65  *      MMCRA[63]    = 1                (SAMPLE_ENABLE)
66  *      MMCRA[57:59] = sample[0:2]      (RAND_SAMP_ELIG)
67  *      MMCRA[61:62] = sample[3:4]      (RAND_SAMP_MODE)
68  *
69  * if EBB and BHRB:
70  *      MMCRA[32:33] = IFM
71  *
72  * MMCRA[SDAR_MODE]  = sdar_mode[0:1]
73  */
74
75 /*
76  * Some power10 event codes.
77  */
78 #define EVENT(_name, _code)     enum{_name = _code}
79
80 #include "power10-events-list.h"
81
82 #undef EVENT
83
84 /* MMCRA IFM bits - POWER10 */
85 #define POWER10_MMCRA_IFM1              0x0000000040000000UL
86 #define POWER10_MMCRA_IFM2              0x0000000080000000UL
87 #define POWER10_MMCRA_IFM3              0x00000000C0000000UL
88 #define POWER10_MMCRA_BHRB_MASK         0x00000000C0000000UL
89
90 extern u64 PERF_REG_EXTENDED_MASK;
91
92 /* Table of alternatives, sorted by column 0 */
93 static const unsigned int power10_event_alternatives[][MAX_ALT] = {
94         { PM_INST_CMPL_ALT,             PM_INST_CMPL },
95         { PM_CYC_ALT,                   PM_CYC },
96 };
97
98 static int power10_get_alternatives(u64 event, unsigned int flags, u64 alt[])
99 {
100         int num_alt = 0;
101
102         num_alt = isa207_get_alternatives(event, alt,
103                                           ARRAY_SIZE(power10_event_alternatives), flags,
104                                           power10_event_alternatives);
105
106         return num_alt;
107 }
108
109 static int power10_check_attr_config(struct perf_event *ev)
110 {
111         u64 val;
112         u64 event = ev->attr.config;
113
114         val = (event >> EVENT_SAMPLE_SHIFT) & EVENT_SAMPLE_MASK;
115         if (val == 0x10 || isa3XX_check_attr_config(ev))
116                 return -EINVAL;
117
118         return 0;
119 }
120
121 GENERIC_EVENT_ATTR(cpu-cycles,                  PM_CYC);
122 GENERIC_EVENT_ATTR(instructions,                PM_INST_CMPL);
123 GENERIC_EVENT_ATTR(branch-instructions,         PM_BR_CMPL);
124 GENERIC_EVENT_ATTR(branch-misses,               PM_BR_MPRED_CMPL);
125 GENERIC_EVENT_ATTR(cache-references,            PM_LD_REF_L1);
126 GENERIC_EVENT_ATTR(cache-misses,                PM_LD_MISS_L1);
127 GENERIC_EVENT_ATTR(mem-loads,                   MEM_LOADS);
128 GENERIC_EVENT_ATTR(mem-stores,                  MEM_STORES);
129 GENERIC_EVENT_ATTR(branch-instructions,         PM_BR_FIN);
130 GENERIC_EVENT_ATTR(branch-misses,               PM_MPRED_BR_FIN);
131 GENERIC_EVENT_ATTR(cache-misses,                PM_LD_DEMAND_MISS_L1_FIN);
132
133 CACHE_EVENT_ATTR(L1-dcache-load-misses,         PM_LD_MISS_L1);
134 CACHE_EVENT_ATTR(L1-dcache-loads,               PM_LD_REF_L1);
135 CACHE_EVENT_ATTR(L1-dcache-prefetches,          PM_LD_PREFETCH_CACHE_LINE_MISS);
136 CACHE_EVENT_ATTR(L1-dcache-store-misses,        PM_ST_MISS_L1);
137 CACHE_EVENT_ATTR(L1-icache-load-misses,         PM_L1_ICACHE_MISS);
138 CACHE_EVENT_ATTR(L1-icache-loads,               PM_INST_FROM_L1);
139 CACHE_EVENT_ATTR(L1-icache-prefetches,          PM_IC_PREF_REQ);
140 CACHE_EVENT_ATTR(LLC-load-misses,               PM_DATA_FROM_L3MISS);
141 CACHE_EVENT_ATTR(LLC-loads,                     PM_DATA_FROM_L3);
142 CACHE_EVENT_ATTR(LLC-prefetches,                PM_L3_PF_MISS_L3);
143 CACHE_EVENT_ATTR(LLC-store-misses,              PM_L2_ST_MISS);
144 CACHE_EVENT_ATTR(LLC-stores,                    PM_L2_ST);
145 CACHE_EVENT_ATTR(branch-load-misses,            PM_BR_MPRED_CMPL);
146 CACHE_EVENT_ATTR(branch-loads,                  PM_BR_CMPL);
147 CACHE_EVENT_ATTR(dTLB-load-misses,              PM_DTLB_MISS);
148 CACHE_EVENT_ATTR(iTLB-load-misses,              PM_ITLB_MISS);
149
150 static struct attribute *power10_events_attr_dd1[] = {
151         GENERIC_EVENT_PTR(PM_CYC),
152         GENERIC_EVENT_PTR(PM_INST_CMPL),
153         GENERIC_EVENT_PTR(PM_BR_CMPL),
154         GENERIC_EVENT_PTR(PM_BR_MPRED_CMPL),
155         GENERIC_EVENT_PTR(PM_LD_REF_L1),
156         GENERIC_EVENT_PTR(PM_LD_MISS_L1),
157         GENERIC_EVENT_PTR(MEM_LOADS),
158         GENERIC_EVENT_PTR(MEM_STORES),
159         CACHE_EVENT_PTR(PM_LD_MISS_L1),
160         CACHE_EVENT_PTR(PM_LD_REF_L1),
161         CACHE_EVENT_PTR(PM_LD_PREFETCH_CACHE_LINE_MISS),
162         CACHE_EVENT_PTR(PM_ST_MISS_L1),
163         CACHE_EVENT_PTR(PM_L1_ICACHE_MISS),
164         CACHE_EVENT_PTR(PM_INST_FROM_L1),
165         CACHE_EVENT_PTR(PM_IC_PREF_REQ),
166         CACHE_EVENT_PTR(PM_DATA_FROM_L3MISS),
167         CACHE_EVENT_PTR(PM_DATA_FROM_L3),
168         CACHE_EVENT_PTR(PM_BR_MPRED_CMPL),
169         CACHE_EVENT_PTR(PM_BR_CMPL),
170         CACHE_EVENT_PTR(PM_DTLB_MISS),
171         CACHE_EVENT_PTR(PM_ITLB_MISS),
172         NULL
173 };
174
175 static struct attribute *power10_events_attr[] = {
176         GENERIC_EVENT_PTR(PM_CYC),
177         GENERIC_EVENT_PTR(PM_INST_CMPL),
178         GENERIC_EVENT_PTR(PM_BR_FIN),
179         GENERIC_EVENT_PTR(PM_MPRED_BR_FIN),
180         GENERIC_EVENT_PTR(PM_LD_REF_L1),
181         GENERIC_EVENT_PTR(PM_LD_DEMAND_MISS_L1_FIN),
182         GENERIC_EVENT_PTR(MEM_LOADS),
183         GENERIC_EVENT_PTR(MEM_STORES),
184         CACHE_EVENT_PTR(PM_LD_MISS_L1),
185         CACHE_EVENT_PTR(PM_LD_REF_L1),
186         CACHE_EVENT_PTR(PM_LD_PREFETCH_CACHE_LINE_MISS),
187         CACHE_EVENT_PTR(PM_ST_MISS_L1),
188         CACHE_EVENT_PTR(PM_L1_ICACHE_MISS),
189         CACHE_EVENT_PTR(PM_INST_FROM_L1),
190         CACHE_EVENT_PTR(PM_IC_PREF_REQ),
191         CACHE_EVENT_PTR(PM_DATA_FROM_L3MISS),
192         CACHE_EVENT_PTR(PM_DATA_FROM_L3),
193         CACHE_EVENT_PTR(PM_L3_PF_MISS_L3),
194         CACHE_EVENT_PTR(PM_L2_ST_MISS),
195         CACHE_EVENT_PTR(PM_L2_ST),
196         CACHE_EVENT_PTR(PM_BR_MPRED_CMPL),
197         CACHE_EVENT_PTR(PM_BR_CMPL),
198         CACHE_EVENT_PTR(PM_DTLB_MISS),
199         CACHE_EVENT_PTR(PM_ITLB_MISS),
200         NULL
201 };
202
203 static const struct attribute_group power10_pmu_events_group_dd1 = {
204         .name = "events",
205         .attrs = power10_events_attr_dd1,
206 };
207
208 static const struct attribute_group power10_pmu_events_group = {
209         .name = "events",
210         .attrs = power10_events_attr,
211 };
212
213 PMU_FORMAT_ATTR(event,          "config:0-59");
214 PMU_FORMAT_ATTR(pmcxsel,        "config:0-7");
215 PMU_FORMAT_ATTR(mark,           "config:8");
216 PMU_FORMAT_ATTR(combine,        "config:10-11");
217 PMU_FORMAT_ATTR(unit,           "config:12-15");
218 PMU_FORMAT_ATTR(pmc,            "config:16-19");
219 PMU_FORMAT_ATTR(cache_sel,      "config:20-21");
220 PMU_FORMAT_ATTR(sdar_mode,      "config:22-23");
221 PMU_FORMAT_ATTR(sample_mode,    "config:24-28");
222 PMU_FORMAT_ATTR(thresh_sel,     "config:29-31");
223 PMU_FORMAT_ATTR(thresh_stop,    "config:32-35");
224 PMU_FORMAT_ATTR(thresh_start,   "config:36-39");
225 PMU_FORMAT_ATTR(l2l3_sel,       "config:40-44");
226 PMU_FORMAT_ATTR(src_sel,        "config:45-46");
227 PMU_FORMAT_ATTR(invert_bit,     "config:47");
228 PMU_FORMAT_ATTR(src_mask,       "config:48-53");
229 PMU_FORMAT_ATTR(src_match,      "config:54-59");
230 PMU_FORMAT_ATTR(radix_scope,    "config:9");
231 PMU_FORMAT_ATTR(thresh_cmp,     "config1:0-17");
232
233 static struct attribute *power10_pmu_format_attr[] = {
234         &format_attr_event.attr,
235         &format_attr_pmcxsel.attr,
236         &format_attr_mark.attr,
237         &format_attr_combine.attr,
238         &format_attr_unit.attr,
239         &format_attr_pmc.attr,
240         &format_attr_cache_sel.attr,
241         &format_attr_sdar_mode.attr,
242         &format_attr_sample_mode.attr,
243         &format_attr_thresh_sel.attr,
244         &format_attr_thresh_stop.attr,
245         &format_attr_thresh_start.attr,
246         &format_attr_l2l3_sel.attr,
247         &format_attr_src_sel.attr,
248         &format_attr_invert_bit.attr,
249         &format_attr_src_mask.attr,
250         &format_attr_src_match.attr,
251         &format_attr_radix_scope.attr,
252         &format_attr_thresh_cmp.attr,
253         NULL,
254 };
255
256 static const struct attribute_group power10_pmu_format_group = {
257         .name = "format",
258         .attrs = power10_pmu_format_attr,
259 };
260
261 static struct attribute *power10_pmu_caps_attrs[] = {
262         NULL
263 };
264
265 static struct attribute_group power10_pmu_caps_group = {
266         .name  = "caps",
267         .attrs = power10_pmu_caps_attrs,
268 };
269
270 static const struct attribute_group *power10_pmu_attr_groups_dd1[] = {
271         &power10_pmu_format_group,
272         &power10_pmu_events_group_dd1,
273         &power10_pmu_caps_group,
274         NULL,
275 };
276
277 static const struct attribute_group *power10_pmu_attr_groups[] = {
278         &power10_pmu_format_group,
279         &power10_pmu_events_group,
280         &power10_pmu_caps_group,
281         NULL,
282 };
283
284 static int power10_generic_events_dd1[] = {
285         [PERF_COUNT_HW_CPU_CYCLES] =                    PM_CYC,
286         [PERF_COUNT_HW_INSTRUCTIONS] =                  PM_INST_CMPL,
287         [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] =           PM_BR_CMPL,
288         [PERF_COUNT_HW_BRANCH_MISSES] =                 PM_BR_MPRED_CMPL,
289         [PERF_COUNT_HW_CACHE_REFERENCES] =              PM_LD_REF_L1,
290         [PERF_COUNT_HW_CACHE_MISSES] =                  PM_LD_MISS_L1,
291 };
292
293 static int power10_generic_events[] = {
294         [PERF_COUNT_HW_CPU_CYCLES] =                    PM_CYC,
295         [PERF_COUNT_HW_INSTRUCTIONS] =                  PM_INST_CMPL,
296         [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] =           PM_BR_FIN,
297         [PERF_COUNT_HW_BRANCH_MISSES] =                 PM_MPRED_BR_FIN,
298         [PERF_COUNT_HW_CACHE_REFERENCES] =              PM_LD_REF_L1,
299         [PERF_COUNT_HW_CACHE_MISSES] =                  PM_LD_DEMAND_MISS_L1_FIN,
300 };
301
302 static u64 power10_bhrb_filter_map(u64 branch_sample_type)
303 {
304         u64 pmu_bhrb_filter = 0;
305
306         /* BHRB and regular PMU events share the same privilege state
307          * filter configuration. BHRB is always recorded along with a
308          * regular PMU event. As the privilege state filter is handled
309          * in the basic PMC configuration of the accompanying regular
310          * PMU event, we ignore any separate BHRB specific request.
311          */
312
313         /* No branch filter requested */
314         if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY)
315                 return pmu_bhrb_filter;
316
317         /* Invalid branch filter options - HW does not support */
318         if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY_RETURN)
319                 return -1;
320
321         if (branch_sample_type & PERF_SAMPLE_BRANCH_IND_CALL) {
322                 pmu_bhrb_filter |= POWER10_MMCRA_IFM2;
323                 return pmu_bhrb_filter;
324         }
325
326         if (branch_sample_type & PERF_SAMPLE_BRANCH_COND) {
327                 pmu_bhrb_filter |= POWER10_MMCRA_IFM3;
328                 return pmu_bhrb_filter;
329         }
330
331         if (branch_sample_type & PERF_SAMPLE_BRANCH_CALL)
332                 return -1;
333
334         if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY_CALL) {
335                 pmu_bhrb_filter |= POWER10_MMCRA_IFM1;
336                 return pmu_bhrb_filter;
337         }
338
339         /* Every thing else is unsupported */
340         return -1;
341 }
342
343 static void power10_config_bhrb(u64 pmu_bhrb_filter)
344 {
345         pmu_bhrb_filter &= POWER10_MMCRA_BHRB_MASK;
346
347         /* Enable BHRB filter in PMU */
348         mtspr(SPRN_MMCRA, (mfspr(SPRN_MMCRA) | pmu_bhrb_filter));
349 }
350
351 #define C(x)    PERF_COUNT_HW_CACHE_##x
352
353 /*
354  * Table of generalized cache-related events.
355  * 0 means not supported, -1 means nonsensical, other values
356  * are event codes.
357  */
358 static u64 power10_cache_events_dd1[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = {
359         [C(L1D)] = {
360                 [C(OP_READ)] = {
361                         [C(RESULT_ACCESS)] = PM_LD_REF_L1,
362                         [C(RESULT_MISS)] = PM_LD_MISS_L1,
363                 },
364                 [C(OP_WRITE)] = {
365                         [C(RESULT_ACCESS)] = 0,
366                         [C(RESULT_MISS)] = PM_ST_MISS_L1,
367                 },
368                 [C(OP_PREFETCH)] = {
369                         [C(RESULT_ACCESS)] = PM_LD_PREFETCH_CACHE_LINE_MISS,
370                         [C(RESULT_MISS)] = 0,
371                 },
372         },
373         [C(L1I)] = {
374                 [C(OP_READ)] = {
375                         [C(RESULT_ACCESS)] = PM_INST_FROM_L1,
376                         [C(RESULT_MISS)] = PM_L1_ICACHE_MISS,
377                 },
378                 [C(OP_WRITE)] = {
379                         [C(RESULT_ACCESS)] = PM_INST_FROM_L1MISS,
380                         [C(RESULT_MISS)] = -1,
381                 },
382                 [C(OP_PREFETCH)] = {
383                         [C(RESULT_ACCESS)] = PM_IC_PREF_REQ,
384                         [C(RESULT_MISS)] = 0,
385                 },
386         },
387         [C(LL)] = {
388                 [C(OP_READ)] = {
389                         [C(RESULT_ACCESS)] = PM_DATA_FROM_L3,
390                         [C(RESULT_MISS)] = PM_DATA_FROM_L3MISS,
391                 },
392                 [C(OP_WRITE)] = {
393                         [C(RESULT_ACCESS)] = -1,
394                         [C(RESULT_MISS)] = -1,
395                 },
396                 [C(OP_PREFETCH)] = {
397                         [C(RESULT_ACCESS)] = -1,
398                         [C(RESULT_MISS)] = 0,
399                 },
400         },
401          [C(DTLB)] = {
402                 [C(OP_READ)] = {
403                         [C(RESULT_ACCESS)] = 0,
404                         [C(RESULT_MISS)] = PM_DTLB_MISS,
405                 },
406                 [C(OP_WRITE)] = {
407                         [C(RESULT_ACCESS)] = -1,
408                         [C(RESULT_MISS)] = -1,
409                 },
410                 [C(OP_PREFETCH)] = {
411                         [C(RESULT_ACCESS)] = -1,
412                         [C(RESULT_MISS)] = -1,
413                 },
414         },
415         [C(ITLB)] = {
416                 [C(OP_READ)] = {
417                         [C(RESULT_ACCESS)] = 0,
418                         [C(RESULT_MISS)] = PM_ITLB_MISS,
419                 },
420                 [C(OP_WRITE)] = {
421                         [C(RESULT_ACCESS)] = -1,
422                         [C(RESULT_MISS)] = -1,
423                 },
424                 [C(OP_PREFETCH)] = {
425                         [C(RESULT_ACCESS)] = -1,
426                         [C(RESULT_MISS)] = -1,
427                 },
428         },
429         [C(BPU)] = {
430                 [C(OP_READ)] = {
431                         [C(RESULT_ACCESS)] = PM_BR_CMPL,
432                         [C(RESULT_MISS)] = PM_BR_MPRED_CMPL,
433                 },
434                 [C(OP_WRITE)] = {
435                         [C(RESULT_ACCESS)] = -1,
436                         [C(RESULT_MISS)] = -1,
437                 },
438                 [C(OP_PREFETCH)] = {
439                         [C(RESULT_ACCESS)] = -1,
440                         [C(RESULT_MISS)] = -1,
441                 },
442         },
443         [C(NODE)] = {
444                 [C(OP_READ)] = {
445                         [C(RESULT_ACCESS)] = -1,
446                         [C(RESULT_MISS)] = -1,
447                 },
448                 [C(OP_WRITE)] = {
449                         [C(RESULT_ACCESS)] = -1,
450                         [C(RESULT_MISS)] = -1,
451                 },
452                 [C(OP_PREFETCH)] = {
453                         [C(RESULT_ACCESS)] = -1,
454                         [C(RESULT_MISS)] = -1,
455                 },
456         },
457 };
458
459 static u64 power10_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = {
460         [C(L1D)] = {
461                 [C(OP_READ)] = {
462                         [C(RESULT_ACCESS)] = PM_LD_REF_L1,
463                         [C(RESULT_MISS)] = PM_LD_MISS_L1,
464                 },
465                 [C(OP_WRITE)] = {
466                         [C(RESULT_ACCESS)] = 0,
467                         [C(RESULT_MISS)] = PM_ST_MISS_L1,
468                 },
469                 [C(OP_PREFETCH)] = {
470                         [C(RESULT_ACCESS)] = PM_LD_PREFETCH_CACHE_LINE_MISS,
471                         [C(RESULT_MISS)] = 0,
472                 },
473         },
474         [C(L1I)] = {
475                 [C(OP_READ)] = {
476                         [C(RESULT_ACCESS)] = PM_INST_FROM_L1,
477                         [C(RESULT_MISS)] = PM_L1_ICACHE_MISS,
478                 },
479                 [C(OP_WRITE)] = {
480                         [C(RESULT_ACCESS)] = PM_INST_FROM_L1MISS,
481                         [C(RESULT_MISS)] = -1,
482                 },
483                 [C(OP_PREFETCH)] = {
484                         [C(RESULT_ACCESS)] = PM_IC_PREF_REQ,
485                         [C(RESULT_MISS)] = 0,
486                 },
487         },
488         [C(LL)] = {
489                 [C(OP_READ)] = {
490                         [C(RESULT_ACCESS)] = PM_DATA_FROM_L3,
491                         [C(RESULT_MISS)] = PM_DATA_FROM_L3MISS,
492                 },
493                 [C(OP_WRITE)] = {
494                         [C(RESULT_ACCESS)] = PM_L2_ST,
495                         [C(RESULT_MISS)] = PM_L2_ST_MISS,
496                 },
497                 [C(OP_PREFETCH)] = {
498                         [C(RESULT_ACCESS)] = PM_L3_PF_MISS_L3,
499                         [C(RESULT_MISS)] = 0,
500                 },
501         },
502          [C(DTLB)] = {
503                 [C(OP_READ)] = {
504                         [C(RESULT_ACCESS)] = 0,
505                         [C(RESULT_MISS)] = PM_DTLB_MISS,
506                 },
507                 [C(OP_WRITE)] = {
508                         [C(RESULT_ACCESS)] = -1,
509                         [C(RESULT_MISS)] = -1,
510                 },
511                 [C(OP_PREFETCH)] = {
512                         [C(RESULT_ACCESS)] = -1,
513                         [C(RESULT_MISS)] = -1,
514                 },
515         },
516         [C(ITLB)] = {
517                 [C(OP_READ)] = {
518                         [C(RESULT_ACCESS)] = 0,
519                         [C(RESULT_MISS)] = PM_ITLB_MISS,
520                 },
521                 [C(OP_WRITE)] = {
522                         [C(RESULT_ACCESS)] = -1,
523                         [C(RESULT_MISS)] = -1,
524                 },
525                 [C(OP_PREFETCH)] = {
526                         [C(RESULT_ACCESS)] = -1,
527                         [C(RESULT_MISS)] = -1,
528                 },
529         },
530         [C(BPU)] = {
531                 [C(OP_READ)] = {
532                         [C(RESULT_ACCESS)] = PM_BR_CMPL,
533                         [C(RESULT_MISS)] = PM_BR_MPRED_CMPL,
534                 },
535                 [C(OP_WRITE)] = {
536                         [C(RESULT_ACCESS)] = -1,
537                         [C(RESULT_MISS)] = -1,
538                 },
539                 [C(OP_PREFETCH)] = {
540                         [C(RESULT_ACCESS)] = -1,
541                         [C(RESULT_MISS)] = -1,
542                 },
543         },
544         [C(NODE)] = {
545                 [C(OP_READ)] = {
546                         [C(RESULT_ACCESS)] = -1,
547                         [C(RESULT_MISS)] = -1,
548                 },
549                 [C(OP_WRITE)] = {
550                         [C(RESULT_ACCESS)] = -1,
551                         [C(RESULT_MISS)] = -1,
552                 },
553                 [C(OP_PREFETCH)] = {
554                         [C(RESULT_ACCESS)] = -1,
555                         [C(RESULT_MISS)] = -1,
556                 },
557         },
558 };
559
560 #undef C
561
562 /*
563  * Set the MMCR0[CC56RUN] bit to enable counting for
564  * PMC5 and PMC6 regardless of the state of CTRL[RUN],
565  * so that we can use counters 5 and 6 as PM_INST_CMPL and
566  * PM_CYC.
567  */
568 static int power10_compute_mmcr(u64 event[], int n_ev,
569                                 unsigned int hwc[], struct mmcr_regs *mmcr,
570                                 struct perf_event *pevents[], u32 flags)
571 {
572         int ret;
573
574         ret = isa207_compute_mmcr(event, n_ev, hwc, mmcr, pevents, flags);
575         if (!ret)
576                 mmcr->mmcr0 |= MMCR0_C56RUN;
577         return ret;
578 }
579
580 static struct power_pmu power10_pmu = {
581         .name                   = "POWER10",
582         .n_counter              = MAX_PMU_COUNTERS,
583         .add_fields             = ISA207_ADD_FIELDS,
584         .test_adder             = ISA207_TEST_ADDER,
585         .group_constraint_mask  = CNST_CACHE_PMC4_MASK,
586         .group_constraint_val   = CNST_CACHE_PMC4_VAL,
587         .compute_mmcr           = power10_compute_mmcr,
588         .config_bhrb            = power10_config_bhrb,
589         .bhrb_filter_map        = power10_bhrb_filter_map,
590         .get_constraint         = isa207_get_constraint,
591         .get_alternatives       = power10_get_alternatives,
592         .get_mem_data_src       = isa207_get_mem_data_src,
593         .get_mem_weight         = isa207_get_mem_weight,
594         .disable_pmc            = isa207_disable_pmc,
595         .flags                  = PPMU_HAS_SIER | PPMU_ARCH_207S |
596                                   PPMU_ARCH_31 | PPMU_HAS_ATTR_CONFIG1,
597         .n_generic              = ARRAY_SIZE(power10_generic_events),
598         .generic_events         = power10_generic_events,
599         .cache_events           = &power10_cache_events,
600         .attr_groups            = power10_pmu_attr_groups,
601         .bhrb_nr                = 32,
602         .capabilities           = PERF_PMU_CAP_EXTENDED_REGS,
603         .check_attr_config      = power10_check_attr_config,
604 };
605
606 int __init init_power10_pmu(void)
607 {
608         unsigned int pvr;
609         int rc;
610
611         pvr = mfspr(SPRN_PVR);
612         if (PVR_VER(pvr) != PVR_POWER10)
613                 return -ENODEV;
614
615         /* Add the ppmu flag for power10 DD1 */
616         if ((PVR_CFG(pvr) == 1))
617                 power10_pmu.flags |= PPMU_P10_DD1;
618
619         /* Set the PERF_REG_EXTENDED_MASK here */
620         PERF_REG_EXTENDED_MASK = PERF_REG_PMU_MASK_31;
621
622         if ((PVR_CFG(pvr) == 1)) {
623                 power10_pmu.generic_events = power10_generic_events_dd1;
624                 power10_pmu.attr_groups = power10_pmu_attr_groups_dd1;
625                 power10_pmu.cache_events = &power10_cache_events_dd1;
626         }
627
628         rc = register_power_pmu(&power10_pmu);
629         if (rc)
630                 return rc;
631
632         /* Tell userspace that EBB is supported */
633         cur_cpu_spec->cpu_user_features2 |= PPC_FEATURE2_EBB;
634
635         return 0;
636 }