Merge tag 'mfd-next-5.4' of git://git.kernel.org/pub/scm/linux/kernel/git/lee/mfd
[sfrench/cifs-2.6.git] / arch / x86 / kernel / quirks.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * This file contains work-arounds for x86 and x86_64 platform bugs.
4  */
5 #include <linux/dmi.h>
6 #include <linux/pci.h>
7 #include <linux/irq.h>
8
9 #include <asm/hpet.h>
10 #include <asm/setup.h>
11
12 #if defined(CONFIG_X86_IO_APIC) && defined(CONFIG_SMP) && defined(CONFIG_PCI)
13
14 static void quirk_intel_irqbalance(struct pci_dev *dev)
15 {
16         u8 config;
17         u16 word;
18
19         /* BIOS may enable hardware IRQ balancing for
20          * E7520/E7320/E7525(revision ID 0x9 and below)
21          * based platforms.
22          * Disable SW irqbalance/affinity on those platforms.
23          */
24         if (dev->revision > 0x9)
25                 return;
26
27         /* enable access to config space*/
28         pci_read_config_byte(dev, 0xf4, &config);
29         pci_write_config_byte(dev, 0xf4, config|0x2);
30
31         /*
32          * read xTPR register.  We may not have a pci_dev for device 8
33          * because it might be hidden until the above write.
34          */
35         pci_bus_read_config_word(dev->bus, PCI_DEVFN(8, 0), 0x4c, &word);
36
37         if (!(word & (1 << 13))) {
38                 dev_info(&dev->dev, "Intel E7520/7320/7525 detected; "
39                         "disabling irq balancing and affinity\n");
40                 noirqdebug_setup("");
41 #ifdef CONFIG_PROC_FS
42                 no_irq_affinity = 1;
43 #endif
44         }
45
46         /* put back the original value for config space*/
47         if (!(config & 0x2))
48                 pci_write_config_byte(dev, 0xf4, config);
49 }
50 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7320_MCH,
51                         quirk_intel_irqbalance);
52 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7525_MCH,
53                         quirk_intel_irqbalance);
54 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7520_MCH,
55                         quirk_intel_irqbalance);
56 #endif
57
58 #if defined(CONFIG_HPET_TIMER)
59 unsigned long force_hpet_address;
60
61 static enum {
62         NONE_FORCE_HPET_RESUME,
63         OLD_ICH_FORCE_HPET_RESUME,
64         ICH_FORCE_HPET_RESUME,
65         VT8237_FORCE_HPET_RESUME,
66         NVIDIA_FORCE_HPET_RESUME,
67         ATI_FORCE_HPET_RESUME,
68 } force_hpet_resume_type;
69
70 static void __iomem *rcba_base;
71
72 static void ich_force_hpet_resume(void)
73 {
74         u32 val;
75
76         if (!force_hpet_address)
77                 return;
78
79         BUG_ON(rcba_base == NULL);
80
81         /* read the Function Disable register, dword mode only */
82         val = readl(rcba_base + 0x3404);
83         if (!(val & 0x80)) {
84                 /* HPET disabled in HPTC. Trying to enable */
85                 writel(val | 0x80, rcba_base + 0x3404);
86         }
87
88         val = readl(rcba_base + 0x3404);
89         if (!(val & 0x80))
90                 BUG();
91         else
92                 printk(KERN_DEBUG "Force enabled HPET at resume\n");
93 }
94
95 static void ich_force_enable_hpet(struct pci_dev *dev)
96 {
97         u32 val;
98         u32 uninitialized_var(rcba);
99         int err = 0;
100
101         if (hpet_address || force_hpet_address)
102                 return;
103
104         pci_read_config_dword(dev, 0xF0, &rcba);
105         rcba &= 0xFFFFC000;
106         if (rcba == 0) {
107                 dev_printk(KERN_DEBUG, &dev->dev, "RCBA disabled; "
108                         "cannot force enable HPET\n");
109                 return;
110         }
111
112         /* use bits 31:14, 16 kB aligned */
113         rcba_base = ioremap_nocache(rcba, 0x4000);
114         if (rcba_base == NULL) {
115                 dev_printk(KERN_DEBUG, &dev->dev, "ioremap failed; "
116                         "cannot force enable HPET\n");
117                 return;
118         }
119
120         /* read the Function Disable register, dword mode only */
121         val = readl(rcba_base + 0x3404);
122
123         if (val & 0x80) {
124                 /* HPET is enabled in HPTC. Just not reported by BIOS */
125                 val = val & 0x3;
126                 force_hpet_address = 0xFED00000 | (val << 12);
127                 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
128                         "0x%lx\n", force_hpet_address);
129                 iounmap(rcba_base);
130                 return;
131         }
132
133         /* HPET disabled in HPTC. Trying to enable */
134         writel(val | 0x80, rcba_base + 0x3404);
135
136         val = readl(rcba_base + 0x3404);
137         if (!(val & 0x80)) {
138                 err = 1;
139         } else {
140                 val = val & 0x3;
141                 force_hpet_address = 0xFED00000 | (val << 12);
142         }
143
144         if (err) {
145                 force_hpet_address = 0;
146                 iounmap(rcba_base);
147                 dev_printk(KERN_DEBUG, &dev->dev,
148                         "Failed to force enable HPET\n");
149         } else {
150                 force_hpet_resume_type = ICH_FORCE_HPET_RESUME;
151                 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
152                         "0x%lx\n", force_hpet_address);
153         }
154 }
155
156 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB2_0,
157                          ich_force_enable_hpet);
158 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_0,
159                          ich_force_enable_hpet);
160 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_1,
161                          ich_force_enable_hpet);
162 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_0,
163                          ich_force_enable_hpet);
164 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_1,
165                          ich_force_enable_hpet);
166 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_31,
167                          ich_force_enable_hpet);
168 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_1,
169                          ich_force_enable_hpet);
170 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_4,
171                          ich_force_enable_hpet);
172 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH9_7,
173                          ich_force_enable_hpet);
174 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, 0x3a16,   /* ICH10 */
175                          ich_force_enable_hpet);
176
177 static struct pci_dev *cached_dev;
178
179 static void hpet_print_force_info(void)
180 {
181         printk(KERN_INFO "HPET not enabled in BIOS. "
182                "You might try hpet=force boot option\n");
183 }
184
185 static void old_ich_force_hpet_resume(void)
186 {
187         u32 val;
188         u32 uninitialized_var(gen_cntl);
189
190         if (!force_hpet_address || !cached_dev)
191                 return;
192
193         pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
194         gen_cntl &= (~(0x7 << 15));
195         gen_cntl |= (0x4 << 15);
196
197         pci_write_config_dword(cached_dev, 0xD0, gen_cntl);
198         pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
199         val = gen_cntl >> 15;
200         val &= 0x7;
201         if (val == 0x4)
202                 printk(KERN_DEBUG "Force enabled HPET at resume\n");
203         else
204                 BUG();
205 }
206
207 static void old_ich_force_enable_hpet(struct pci_dev *dev)
208 {
209         u32 val;
210         u32 uninitialized_var(gen_cntl);
211
212         if (hpet_address || force_hpet_address)
213                 return;
214
215         pci_read_config_dword(dev, 0xD0, &gen_cntl);
216         /*
217          * Bit 17 is HPET enable bit.
218          * Bit 16:15 control the HPET base address.
219          */
220         val = gen_cntl >> 15;
221         val &= 0x7;
222         if (val & 0x4) {
223                 val &= 0x3;
224                 force_hpet_address = 0xFED00000 | (val << 12);
225                 dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
226                         force_hpet_address);
227                 return;
228         }
229
230         /*
231          * HPET is disabled. Trying enabling at FED00000 and check
232          * whether it sticks
233          */
234         gen_cntl &= (~(0x7 << 15));
235         gen_cntl |= (0x4 << 15);
236         pci_write_config_dword(dev, 0xD0, gen_cntl);
237
238         pci_read_config_dword(dev, 0xD0, &gen_cntl);
239
240         val = gen_cntl >> 15;
241         val &= 0x7;
242         if (val & 0x4) {
243                 /* HPET is enabled in HPTC. Just not reported by BIOS */
244                 val &= 0x3;
245                 force_hpet_address = 0xFED00000 | (val << 12);
246                 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
247                         "0x%lx\n", force_hpet_address);
248                 cached_dev = dev;
249                 force_hpet_resume_type = OLD_ICH_FORCE_HPET_RESUME;
250                 return;
251         }
252
253         dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
254 }
255
256 /*
257  * Undocumented chipset features. Make sure that the user enforced
258  * this.
259  */
260 static void old_ich_force_enable_hpet_user(struct pci_dev *dev)
261 {
262         if (hpet_force_user)
263                 old_ich_force_enable_hpet(dev);
264 }
265
266 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB_1,
267                          old_ich_force_enable_hpet_user);
268 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_0,
269                          old_ich_force_enable_hpet_user);
270 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_12,
271                          old_ich_force_enable_hpet_user);
272 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_0,
273                          old_ich_force_enable_hpet_user);
274 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_12,
275                          old_ich_force_enable_hpet_user);
276 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_0,
277                          old_ich_force_enable_hpet);
278 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_12,
279                          old_ich_force_enable_hpet);
280
281
282 static void vt8237_force_hpet_resume(void)
283 {
284         u32 val;
285
286         if (!force_hpet_address || !cached_dev)
287                 return;
288
289         val = 0xfed00000 | 0x80;
290         pci_write_config_dword(cached_dev, 0x68, val);
291
292         pci_read_config_dword(cached_dev, 0x68, &val);
293         if (val & 0x80)
294                 printk(KERN_DEBUG "Force enabled HPET at resume\n");
295         else
296                 BUG();
297 }
298
299 static void vt8237_force_enable_hpet(struct pci_dev *dev)
300 {
301         u32 uninitialized_var(val);
302
303         if (hpet_address || force_hpet_address)
304                 return;
305
306         if (!hpet_force_user) {
307                 hpet_print_force_info();
308                 return;
309         }
310
311         pci_read_config_dword(dev, 0x68, &val);
312         /*
313          * Bit 7 is HPET enable bit.
314          * Bit 31:10 is HPET base address (contrary to what datasheet claims)
315          */
316         if (val & 0x80) {
317                 force_hpet_address = (val & ~0x3ff);
318                 dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
319                         force_hpet_address);
320                 return;
321         }
322
323         /*
324          * HPET is disabled. Trying enabling at FED00000 and check
325          * whether it sticks
326          */
327         val = 0xfed00000 | 0x80;
328         pci_write_config_dword(dev, 0x68, val);
329
330         pci_read_config_dword(dev, 0x68, &val);
331         if (val & 0x80) {
332                 force_hpet_address = (val & ~0x3ff);
333                 dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
334                         "0x%lx\n", force_hpet_address);
335                 cached_dev = dev;
336                 force_hpet_resume_type = VT8237_FORCE_HPET_RESUME;
337                 return;
338         }
339
340         dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
341 }
342
343 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235,
344                          vt8237_force_enable_hpet);
345 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8237,
346                          vt8237_force_enable_hpet);
347 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_CX700,
348                          vt8237_force_enable_hpet);
349
350 static void ati_force_hpet_resume(void)
351 {
352         pci_write_config_dword(cached_dev, 0x14, 0xfed00000);
353         printk(KERN_DEBUG "Force enabled HPET at resume\n");
354 }
355
356 static u32 ati_ixp4x0_rev(struct pci_dev *dev)
357 {
358         int err = 0;
359         u32 d = 0;
360         u8  b = 0;
361
362         err = pci_read_config_byte(dev, 0xac, &b);
363         b &= ~(1<<5);
364         err |= pci_write_config_byte(dev, 0xac, b);
365         err |= pci_read_config_dword(dev, 0x70, &d);
366         d |= 1<<8;
367         err |= pci_write_config_dword(dev, 0x70, d);
368         err |= pci_read_config_dword(dev, 0x8, &d);
369         d &= 0xff;
370         dev_printk(KERN_DEBUG, &dev->dev, "SB4X0 revision 0x%x\n", d);
371
372         WARN_ON_ONCE(err);
373
374         return d;
375 }
376
377 static void ati_force_enable_hpet(struct pci_dev *dev)
378 {
379         u32 d, val;
380         u8  b;
381
382         if (hpet_address || force_hpet_address)
383                 return;
384
385         if (!hpet_force_user) {
386                 hpet_print_force_info();
387                 return;
388         }
389
390         d = ati_ixp4x0_rev(dev);
391         if (d  < 0x82)
392                 return;
393
394         /* base address */
395         pci_write_config_dword(dev, 0x14, 0xfed00000);
396         pci_read_config_dword(dev, 0x14, &val);
397
398         /* enable interrupt */
399         outb(0x72, 0xcd6); b = inb(0xcd7);
400         b |= 0x1;
401         outb(0x72, 0xcd6); outb(b, 0xcd7);
402         outb(0x72, 0xcd6); b = inb(0xcd7);
403         if (!(b & 0x1))
404                 return;
405         pci_read_config_dword(dev, 0x64, &d);
406         d |= (1<<10);
407         pci_write_config_dword(dev, 0x64, d);
408         pci_read_config_dword(dev, 0x64, &d);
409         if (!(d & (1<<10)))
410                 return;
411
412         force_hpet_address = val;
413         force_hpet_resume_type = ATI_FORCE_HPET_RESUME;
414         dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
415                    force_hpet_address);
416         cached_dev = dev;
417 }
418 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_IXP400_SMBUS,
419                          ati_force_enable_hpet);
420
421 /*
422  * Undocumented chipset feature taken from LinuxBIOS.
423  */
424 static void nvidia_force_hpet_resume(void)
425 {
426         pci_write_config_dword(cached_dev, 0x44, 0xfed00001);
427         printk(KERN_DEBUG "Force enabled HPET at resume\n");
428 }
429
430 static void nvidia_force_enable_hpet(struct pci_dev *dev)
431 {
432         u32 uninitialized_var(val);
433
434         if (hpet_address || force_hpet_address)
435                 return;
436
437         if (!hpet_force_user) {
438                 hpet_print_force_info();
439                 return;
440         }
441
442         pci_write_config_dword(dev, 0x44, 0xfed00001);
443         pci_read_config_dword(dev, 0x44, &val);
444         force_hpet_address = val & 0xfffffffe;
445         force_hpet_resume_type = NVIDIA_FORCE_HPET_RESUME;
446         dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
447                 force_hpet_address);
448         cached_dev = dev;
449 }
450
451 /* ISA Bridges */
452 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0050,
453                         nvidia_force_enable_hpet);
454 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0051,
455                         nvidia_force_enable_hpet);
456
457 /* LPC bridges */
458 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0260,
459                         nvidia_force_enable_hpet);
460 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0360,
461                         nvidia_force_enable_hpet);
462 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0361,
463                         nvidia_force_enable_hpet);
464 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0362,
465                         nvidia_force_enable_hpet);
466 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0363,
467                         nvidia_force_enable_hpet);
468 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0364,
469                         nvidia_force_enable_hpet);
470 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0365,
471                         nvidia_force_enable_hpet);
472 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0366,
473                         nvidia_force_enable_hpet);
474 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0367,
475                         nvidia_force_enable_hpet);
476
477 void force_hpet_resume(void)
478 {
479         switch (force_hpet_resume_type) {
480         case ICH_FORCE_HPET_RESUME:
481                 ich_force_hpet_resume();
482                 return;
483         case OLD_ICH_FORCE_HPET_RESUME:
484                 old_ich_force_hpet_resume();
485                 return;
486         case VT8237_FORCE_HPET_RESUME:
487                 vt8237_force_hpet_resume();
488                 return;
489         case NVIDIA_FORCE_HPET_RESUME:
490                 nvidia_force_hpet_resume();
491                 return;
492         case ATI_FORCE_HPET_RESUME:
493                 ati_force_hpet_resume();
494                 return;
495         default:
496                 break;
497         }
498 }
499
500 /*
501  * According to the datasheet e6xx systems have the HPET hardwired to
502  * 0xfed00000
503  */
504 static void e6xx_force_enable_hpet(struct pci_dev *dev)
505 {
506         if (hpet_address || force_hpet_address)
507                 return;
508
509         force_hpet_address = 0xFED00000;
510         force_hpet_resume_type = NONE_FORCE_HPET_RESUME;
511         dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
512                 "0x%lx\n", force_hpet_address);
513 }
514 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E6XX_CU,
515                          e6xx_force_enable_hpet);
516
517 /*
518  * HPET MSI on some boards (ATI SB700/SB800) has side effect on
519  * floppy DMA. Disable HPET MSI on such platforms.
520  * See erratum #27 (Misinterpreted MSI Requests May Result in
521  * Corrupted LPC DMA Data) in AMD Publication #46837,
522  * "SB700 Family Product Errata", Rev. 1.0, March 2010.
523  */
524 static void force_disable_hpet_msi(struct pci_dev *unused)
525 {
526         hpet_msi_disable = true;
527 }
528
529 DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_SBX00_SMBUS,
530                          force_disable_hpet_msi);
531
532 #endif
533
534 #if defined(CONFIG_PCI) && defined(CONFIG_NUMA)
535 /* Set correct numa_node information for AMD NB functions */
536 static void quirk_amd_nb_node(struct pci_dev *dev)
537 {
538         struct pci_dev *nb_ht;
539         unsigned int devfn;
540         u32 node;
541         u32 val;
542
543         devfn = PCI_DEVFN(PCI_SLOT(dev->devfn), 0);
544         nb_ht = pci_get_slot(dev->bus, devfn);
545         if (!nb_ht)
546                 return;
547
548         pci_read_config_dword(nb_ht, 0x60, &val);
549         node = pcibus_to_node(dev->bus) | (val & 7);
550         /*
551          * Some hardware may return an invalid node ID,
552          * so check it first:
553          */
554         if (node_online(node))
555                 set_dev_node(&dev->dev, node);
556         pci_dev_put(nb_ht);
557 }
558
559 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB,
560                         quirk_amd_nb_node);
561 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP,
562                         quirk_amd_nb_node);
563 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MEMCTL,
564                         quirk_amd_nb_node);
565 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MISC,
566                         quirk_amd_nb_node);
567 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_HT,
568                         quirk_amd_nb_node);
569 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MAP,
570                         quirk_amd_nb_node);
571 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_DRAM,
572                         quirk_amd_nb_node);
573 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MISC,
574                         quirk_amd_nb_node);
575 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_LINK,
576                         quirk_amd_nb_node);
577 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F0,
578                         quirk_amd_nb_node);
579 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F1,
580                         quirk_amd_nb_node);
581 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F2,
582                         quirk_amd_nb_node);
583 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F3,
584                         quirk_amd_nb_node);
585 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F4,
586                         quirk_amd_nb_node);
587 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F5,
588                         quirk_amd_nb_node);
589
590 #endif
591
592 #ifdef CONFIG_PCI
593 /*
594  * Processor does not ensure DRAM scrub read/write sequence
595  * is atomic wrt accesses to CC6 save state area. Therefore
596  * if a concurrent scrub read/write access is to same address
597  * the entry may appear as if it is not written. This quirk
598  * applies to Fam16h models 00h-0Fh
599  *
600  * See "Revision Guide" for AMD F16h models 00h-0fh,
601  * document 51810 rev. 3.04, Nov 2013
602  */
603 static void amd_disable_seq_and_redirect_scrub(struct pci_dev *dev)
604 {
605         u32 val;
606
607         /*
608          * Suggested workaround:
609          * set D18F3x58[4:0] = 00h and set D18F3x5C[0] = 0b
610          */
611         pci_read_config_dword(dev, 0x58, &val);
612         if (val & 0x1F) {
613                 val &= ~(0x1F);
614                 pci_write_config_dword(dev, 0x58, val);
615         }
616
617         pci_read_config_dword(dev, 0x5C, &val);
618         if (val & BIT(0)) {
619                 val &= ~BIT(0);
620                 pci_write_config_dword(dev, 0x5c, val);
621         }
622 }
623
624 DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_16H_NB_F3,
625                         amd_disable_seq_and_redirect_scrub);
626
627 #if defined(CONFIG_X86_64) && defined(CONFIG_X86_MCE)
628 #include <linux/jump_label.h>
629 #include <asm/string_64.h>
630
631 /* Ivy Bridge, Haswell, Broadwell */
632 static void quirk_intel_brickland_xeon_ras_cap(struct pci_dev *pdev)
633 {
634         u32 capid0;
635
636         pci_read_config_dword(pdev, 0x84, &capid0);
637
638         if (capid0 & 0x10)
639                 static_branch_inc(&mcsafe_key);
640 }
641
642 /* Skylake */
643 static void quirk_intel_purley_xeon_ras_cap(struct pci_dev *pdev)
644 {
645         u32 capid0, capid5;
646
647         pci_read_config_dword(pdev, 0x84, &capid0);
648         pci_read_config_dword(pdev, 0x98, &capid5);
649
650         /*
651          * CAPID0{7:6} indicate whether this is an advanced RAS SKU
652          * CAPID5{8:5} indicate that various NVDIMM usage modes are
653          * enabled, so memory machine check recovery is also enabled.
654          */
655         if ((capid0 & 0xc0) == 0xc0 || (capid5 & 0x1e0))
656                 static_branch_inc(&mcsafe_key);
657
658 }
659 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x0ec3, quirk_intel_brickland_xeon_ras_cap);
660 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2fc0, quirk_intel_brickland_xeon_ras_cap);
661 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x6fc0, quirk_intel_brickland_xeon_ras_cap);
662 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2083, quirk_intel_purley_xeon_ras_cap);
663 #endif
664 #endif
665
666 bool x86_apple_machine;
667 EXPORT_SYMBOL(x86_apple_machine);
668
669 void __init early_platform_quirks(void)
670 {
671         x86_apple_machine = dmi_match(DMI_SYS_VENDOR, "Apple Inc.") ||
672                             dmi_match(DMI_SYS_VENDOR, "Apple Computer, Inc.");
673 }