quirks.c 18 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * This file contains work-arounds for x86 and x86_64 platform bugs.
  4. */
  5. #include <linux/dmi.h>
  6. #include <linux/pci.h>
  7. #include <linux/irq.h>
  8. #include <asm/hpet.h>
  9. #if defined(CONFIG_X86_IO_APIC) && defined(CONFIG_SMP) && defined(CONFIG_PCI)
  10. static void quirk_intel_irqbalance(struct pci_dev *dev)
  11. {
  12. u8 config;
  13. u16 word;
  14. /* BIOS may enable hardware IRQ balancing for
  15. * E7520/E7320/E7525(revision ID 0x9 and below)
  16. * based platforms.
  17. * Disable SW irqbalance/affinity on those platforms.
  18. */
  19. if (dev->revision > 0x9)
  20. return;
  21. /* enable access to config space*/
  22. pci_read_config_byte(dev, 0xf4, &config);
  23. pci_write_config_byte(dev, 0xf4, config|0x2);
  24. /*
  25. * read xTPR register. We may not have a pci_dev for device 8
  26. * because it might be hidden until the above write.
  27. */
  28. pci_bus_read_config_word(dev->bus, PCI_DEVFN(8, 0), 0x4c, &word);
  29. if (!(word & (1 << 13))) {
  30. dev_info(&dev->dev, "Intel E7520/7320/7525 detected; "
  31. "disabling irq balancing and affinity\n");
  32. noirqdebug_setup("");
  33. #ifdef CONFIG_PROC_FS
  34. no_irq_affinity = 1;
  35. #endif
  36. }
  37. /* put back the original value for config space*/
  38. if (!(config & 0x2))
  39. pci_write_config_byte(dev, 0xf4, config);
  40. }
  41. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7320_MCH,
  42. quirk_intel_irqbalance);
  43. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7525_MCH,
  44. quirk_intel_irqbalance);
  45. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7520_MCH,
  46. quirk_intel_irqbalance);
  47. #endif
  48. #if defined(CONFIG_HPET_TIMER)
  49. unsigned long force_hpet_address;
  50. static enum {
  51. NONE_FORCE_HPET_RESUME,
  52. OLD_ICH_FORCE_HPET_RESUME,
  53. ICH_FORCE_HPET_RESUME,
  54. VT8237_FORCE_HPET_RESUME,
  55. NVIDIA_FORCE_HPET_RESUME,
  56. ATI_FORCE_HPET_RESUME,
  57. } force_hpet_resume_type;
  58. static void __iomem *rcba_base;
  59. static void ich_force_hpet_resume(void)
  60. {
  61. u32 val;
  62. if (!force_hpet_address)
  63. return;
  64. BUG_ON(rcba_base == NULL);
  65. /* read the Function Disable register, dword mode only */
  66. val = readl(rcba_base + 0x3404);
  67. if (!(val & 0x80)) {
  68. /* HPET disabled in HPTC. Trying to enable */
  69. writel(val | 0x80, rcba_base + 0x3404);
  70. }
  71. val = readl(rcba_base + 0x3404);
  72. if (!(val & 0x80))
  73. BUG();
  74. else
  75. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  76. return;
  77. }
  78. static void ich_force_enable_hpet(struct pci_dev *dev)
  79. {
  80. u32 val;
  81. u32 uninitialized_var(rcba);
  82. int err = 0;
  83. if (hpet_address || force_hpet_address)
  84. return;
  85. pci_read_config_dword(dev, 0xF0, &rcba);
  86. rcba &= 0xFFFFC000;
  87. if (rcba == 0) {
  88. dev_printk(KERN_DEBUG, &dev->dev, "RCBA disabled; "
  89. "cannot force enable HPET\n");
  90. return;
  91. }
  92. /* use bits 31:14, 16 kB aligned */
  93. rcba_base = ioremap_nocache(rcba, 0x4000);
  94. if (rcba_base == NULL) {
  95. dev_printk(KERN_DEBUG, &dev->dev, "ioremap failed; "
  96. "cannot force enable HPET\n");
  97. return;
  98. }
  99. /* read the Function Disable register, dword mode only */
  100. val = readl(rcba_base + 0x3404);
  101. if (val & 0x80) {
  102. /* HPET is enabled in HPTC. Just not reported by BIOS */
  103. val = val & 0x3;
  104. force_hpet_address = 0xFED00000 | (val << 12);
  105. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  106. "0x%lx\n", force_hpet_address);
  107. iounmap(rcba_base);
  108. return;
  109. }
  110. /* HPET disabled in HPTC. Trying to enable */
  111. writel(val | 0x80, rcba_base + 0x3404);
  112. val = readl(rcba_base + 0x3404);
  113. if (!(val & 0x80)) {
  114. err = 1;
  115. } else {
  116. val = val & 0x3;
  117. force_hpet_address = 0xFED00000 | (val << 12);
  118. }
  119. if (err) {
  120. force_hpet_address = 0;
  121. iounmap(rcba_base);
  122. dev_printk(KERN_DEBUG, &dev->dev,
  123. "Failed to force enable HPET\n");
  124. } else {
  125. force_hpet_resume_type = ICH_FORCE_HPET_RESUME;
  126. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  127. "0x%lx\n", force_hpet_address);
  128. }
  129. }
  130. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB2_0,
  131. ich_force_enable_hpet);
  132. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_0,
  133. ich_force_enable_hpet);
  134. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_1,
  135. ich_force_enable_hpet);
  136. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_0,
  137. ich_force_enable_hpet);
  138. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_1,
  139. ich_force_enable_hpet);
  140. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_31,
  141. ich_force_enable_hpet);
  142. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_1,
  143. ich_force_enable_hpet);
  144. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_4,
  145. ich_force_enable_hpet);
  146. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH9_7,
  147. ich_force_enable_hpet);
  148. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, 0x3a16, /* ICH10 */
  149. ich_force_enable_hpet);
  150. static struct pci_dev *cached_dev;
  151. static void hpet_print_force_info(void)
  152. {
  153. printk(KERN_INFO "HPET not enabled in BIOS. "
  154. "You might try hpet=force boot option\n");
  155. }
  156. static void old_ich_force_hpet_resume(void)
  157. {
  158. u32 val;
  159. u32 uninitialized_var(gen_cntl);
  160. if (!force_hpet_address || !cached_dev)
  161. return;
  162. pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
  163. gen_cntl &= (~(0x7 << 15));
  164. gen_cntl |= (0x4 << 15);
  165. pci_write_config_dword(cached_dev, 0xD0, gen_cntl);
  166. pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
  167. val = gen_cntl >> 15;
  168. val &= 0x7;
  169. if (val == 0x4)
  170. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  171. else
  172. BUG();
  173. }
  174. static void old_ich_force_enable_hpet(struct pci_dev *dev)
  175. {
  176. u32 val;
  177. u32 uninitialized_var(gen_cntl);
  178. if (hpet_address || force_hpet_address)
  179. return;
  180. pci_read_config_dword(dev, 0xD0, &gen_cntl);
  181. /*
  182. * Bit 17 is HPET enable bit.
  183. * Bit 16:15 control the HPET base address.
  184. */
  185. val = gen_cntl >> 15;
  186. val &= 0x7;
  187. if (val & 0x4) {
  188. val &= 0x3;
  189. force_hpet_address = 0xFED00000 | (val << 12);
  190. dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
  191. force_hpet_address);
  192. return;
  193. }
  194. /*
  195. * HPET is disabled. Trying enabling at FED00000 and check
  196. * whether it sticks
  197. */
  198. gen_cntl &= (~(0x7 << 15));
  199. gen_cntl |= (0x4 << 15);
  200. pci_write_config_dword(dev, 0xD0, gen_cntl);
  201. pci_read_config_dword(dev, 0xD0, &gen_cntl);
  202. val = gen_cntl >> 15;
  203. val &= 0x7;
  204. if (val & 0x4) {
  205. /* HPET is enabled in HPTC. Just not reported by BIOS */
  206. val &= 0x3;
  207. force_hpet_address = 0xFED00000 | (val << 12);
  208. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  209. "0x%lx\n", force_hpet_address);
  210. cached_dev = dev;
  211. force_hpet_resume_type = OLD_ICH_FORCE_HPET_RESUME;
  212. return;
  213. }
  214. dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
  215. }
  216. /*
  217. * Undocumented chipset features. Make sure that the user enforced
  218. * this.
  219. */
  220. static void old_ich_force_enable_hpet_user(struct pci_dev *dev)
  221. {
  222. if (hpet_force_user)
  223. old_ich_force_enable_hpet(dev);
  224. }
  225. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB_1,
  226. old_ich_force_enable_hpet_user);
  227. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_0,
  228. old_ich_force_enable_hpet_user);
  229. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_12,
  230. old_ich_force_enable_hpet_user);
  231. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_0,
  232. old_ich_force_enable_hpet_user);
  233. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_12,
  234. old_ich_force_enable_hpet_user);
  235. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_0,
  236. old_ich_force_enable_hpet);
  237. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_12,
  238. old_ich_force_enable_hpet);
  239. static void vt8237_force_hpet_resume(void)
  240. {
  241. u32 val;
  242. if (!force_hpet_address || !cached_dev)
  243. return;
  244. val = 0xfed00000 | 0x80;
  245. pci_write_config_dword(cached_dev, 0x68, val);
  246. pci_read_config_dword(cached_dev, 0x68, &val);
  247. if (val & 0x80)
  248. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  249. else
  250. BUG();
  251. }
  252. static void vt8237_force_enable_hpet(struct pci_dev *dev)
  253. {
  254. u32 uninitialized_var(val);
  255. if (hpet_address || force_hpet_address)
  256. return;
  257. if (!hpet_force_user) {
  258. hpet_print_force_info();
  259. return;
  260. }
  261. pci_read_config_dword(dev, 0x68, &val);
  262. /*
  263. * Bit 7 is HPET enable bit.
  264. * Bit 31:10 is HPET base address (contrary to what datasheet claims)
  265. */
  266. if (val & 0x80) {
  267. force_hpet_address = (val & ~0x3ff);
  268. dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
  269. force_hpet_address);
  270. return;
  271. }
  272. /*
  273. * HPET is disabled. Trying enabling at FED00000 and check
  274. * whether it sticks
  275. */
  276. val = 0xfed00000 | 0x80;
  277. pci_write_config_dword(dev, 0x68, val);
  278. pci_read_config_dword(dev, 0x68, &val);
  279. if (val & 0x80) {
  280. force_hpet_address = (val & ~0x3ff);
  281. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  282. "0x%lx\n", force_hpet_address);
  283. cached_dev = dev;
  284. force_hpet_resume_type = VT8237_FORCE_HPET_RESUME;
  285. return;
  286. }
  287. dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
  288. }
  289. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235,
  290. vt8237_force_enable_hpet);
  291. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8237,
  292. vt8237_force_enable_hpet);
  293. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_CX700,
  294. vt8237_force_enable_hpet);
  295. static void ati_force_hpet_resume(void)
  296. {
  297. pci_write_config_dword(cached_dev, 0x14, 0xfed00000);
  298. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  299. }
  300. static u32 ati_ixp4x0_rev(struct pci_dev *dev)
  301. {
  302. int err = 0;
  303. u32 d = 0;
  304. u8 b = 0;
  305. err = pci_read_config_byte(dev, 0xac, &b);
  306. b &= ~(1<<5);
  307. err |= pci_write_config_byte(dev, 0xac, b);
  308. err |= pci_read_config_dword(dev, 0x70, &d);
  309. d |= 1<<8;
  310. err |= pci_write_config_dword(dev, 0x70, d);
  311. err |= pci_read_config_dword(dev, 0x8, &d);
  312. d &= 0xff;
  313. dev_printk(KERN_DEBUG, &dev->dev, "SB4X0 revision 0x%x\n", d);
  314. WARN_ON_ONCE(err);
  315. return d;
  316. }
  317. static void ati_force_enable_hpet(struct pci_dev *dev)
  318. {
  319. u32 d, val;
  320. u8 b;
  321. if (hpet_address || force_hpet_address)
  322. return;
  323. if (!hpet_force_user) {
  324. hpet_print_force_info();
  325. return;
  326. }
  327. d = ati_ixp4x0_rev(dev);
  328. if (d < 0x82)
  329. return;
  330. /* base address */
  331. pci_write_config_dword(dev, 0x14, 0xfed00000);
  332. pci_read_config_dword(dev, 0x14, &val);
  333. /* enable interrupt */
  334. outb(0x72, 0xcd6); b = inb(0xcd7);
  335. b |= 0x1;
  336. outb(0x72, 0xcd6); outb(b, 0xcd7);
  337. outb(0x72, 0xcd6); b = inb(0xcd7);
  338. if (!(b & 0x1))
  339. return;
  340. pci_read_config_dword(dev, 0x64, &d);
  341. d |= (1<<10);
  342. pci_write_config_dword(dev, 0x64, d);
  343. pci_read_config_dword(dev, 0x64, &d);
  344. if (!(d & (1<<10)))
  345. return;
  346. force_hpet_address = val;
  347. force_hpet_resume_type = ATI_FORCE_HPET_RESUME;
  348. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
  349. force_hpet_address);
  350. cached_dev = dev;
  351. }
  352. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_IXP400_SMBUS,
  353. ati_force_enable_hpet);
  354. /*
  355. * Undocumented chipset feature taken from LinuxBIOS.
  356. */
  357. static void nvidia_force_hpet_resume(void)
  358. {
  359. pci_write_config_dword(cached_dev, 0x44, 0xfed00001);
  360. printk(KERN_DEBUG "Force enabled HPET at resume\n");
  361. }
  362. static void nvidia_force_enable_hpet(struct pci_dev *dev)
  363. {
  364. u32 uninitialized_var(val);
  365. if (hpet_address || force_hpet_address)
  366. return;
  367. if (!hpet_force_user) {
  368. hpet_print_force_info();
  369. return;
  370. }
  371. pci_write_config_dword(dev, 0x44, 0xfed00001);
  372. pci_read_config_dword(dev, 0x44, &val);
  373. force_hpet_address = val & 0xfffffffe;
  374. force_hpet_resume_type = NVIDIA_FORCE_HPET_RESUME;
  375. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
  376. force_hpet_address);
  377. cached_dev = dev;
  378. return;
  379. }
  380. /* ISA Bridges */
  381. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0050,
  382. nvidia_force_enable_hpet);
  383. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0051,
  384. nvidia_force_enable_hpet);
  385. /* LPC bridges */
  386. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0260,
  387. nvidia_force_enable_hpet);
  388. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0360,
  389. nvidia_force_enable_hpet);
  390. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0361,
  391. nvidia_force_enable_hpet);
  392. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0362,
  393. nvidia_force_enable_hpet);
  394. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0363,
  395. nvidia_force_enable_hpet);
  396. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0364,
  397. nvidia_force_enable_hpet);
  398. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0365,
  399. nvidia_force_enable_hpet);
  400. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0366,
  401. nvidia_force_enable_hpet);
  402. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0367,
  403. nvidia_force_enable_hpet);
  404. void force_hpet_resume(void)
  405. {
  406. switch (force_hpet_resume_type) {
  407. case ICH_FORCE_HPET_RESUME:
  408. ich_force_hpet_resume();
  409. return;
  410. case OLD_ICH_FORCE_HPET_RESUME:
  411. old_ich_force_hpet_resume();
  412. return;
  413. case VT8237_FORCE_HPET_RESUME:
  414. vt8237_force_hpet_resume();
  415. return;
  416. case NVIDIA_FORCE_HPET_RESUME:
  417. nvidia_force_hpet_resume();
  418. return;
  419. case ATI_FORCE_HPET_RESUME:
  420. ati_force_hpet_resume();
  421. return;
  422. default:
  423. break;
  424. }
  425. }
  426. /*
  427. * According to the datasheet e6xx systems have the HPET hardwired to
  428. * 0xfed00000
  429. */
  430. static void e6xx_force_enable_hpet(struct pci_dev *dev)
  431. {
  432. if (hpet_address || force_hpet_address)
  433. return;
  434. force_hpet_address = 0xFED00000;
  435. force_hpet_resume_type = NONE_FORCE_HPET_RESUME;
  436. dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
  437. "0x%lx\n", force_hpet_address);
  438. return;
  439. }
  440. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E6XX_CU,
  441. e6xx_force_enable_hpet);
  442. /*
  443. * HPET MSI on some boards (ATI SB700/SB800) has side effect on
  444. * floppy DMA. Disable HPET MSI on such platforms.
  445. * See erratum #27 (Misinterpreted MSI Requests May Result in
  446. * Corrupted LPC DMA Data) in AMD Publication #46837,
  447. * "SB700 Family Product Errata", Rev. 1.0, March 2010.
  448. */
  449. static void force_disable_hpet_msi(struct pci_dev *unused)
  450. {
  451. hpet_msi_disable = true;
  452. }
  453. DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_SBX00_SMBUS,
  454. force_disable_hpet_msi);
  455. #endif
  456. #if defined(CONFIG_PCI) && defined(CONFIG_NUMA)
  457. /* Set correct numa_node information for AMD NB functions */
  458. static void quirk_amd_nb_node(struct pci_dev *dev)
  459. {
  460. struct pci_dev *nb_ht;
  461. unsigned int devfn;
  462. u32 node;
  463. u32 val;
  464. devfn = PCI_DEVFN(PCI_SLOT(dev->devfn), 0);
  465. nb_ht = pci_get_slot(dev->bus, devfn);
  466. if (!nb_ht)
  467. return;
  468. pci_read_config_dword(nb_ht, 0x60, &val);
  469. node = pcibus_to_node(dev->bus) | (val & 7);
  470. /*
  471. * Some hardware may return an invalid node ID,
  472. * so check it first:
  473. */
  474. if (node_online(node))
  475. set_dev_node(&dev->dev, node);
  476. pci_dev_put(nb_ht);
  477. }
  478. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB,
  479. quirk_amd_nb_node);
  480. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP,
  481. quirk_amd_nb_node);
  482. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MEMCTL,
  483. quirk_amd_nb_node);
  484. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MISC,
  485. quirk_amd_nb_node);
  486. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_HT,
  487. quirk_amd_nb_node);
  488. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MAP,
  489. quirk_amd_nb_node);
  490. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_DRAM,
  491. quirk_amd_nb_node);
  492. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MISC,
  493. quirk_amd_nb_node);
  494. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_LINK,
  495. quirk_amd_nb_node);
  496. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F0,
  497. quirk_amd_nb_node);
  498. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F1,
  499. quirk_amd_nb_node);
  500. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F2,
  501. quirk_amd_nb_node);
  502. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F3,
  503. quirk_amd_nb_node);
  504. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F4,
  505. quirk_amd_nb_node);
  506. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F5,
  507. quirk_amd_nb_node);
  508. #endif
  509. #ifdef CONFIG_PCI
  510. /*
  511. * Processor does not ensure DRAM scrub read/write sequence
  512. * is atomic wrt accesses to CC6 save state area. Therefore
  513. * if a concurrent scrub read/write access is to same address
  514. * the entry may appear as if it is not written. This quirk
  515. * applies to Fam16h models 00h-0Fh
  516. *
  517. * See "Revision Guide" for AMD F16h models 00h-0fh,
  518. * document 51810 rev. 3.04, Nov 2013
  519. */
  520. static void amd_disable_seq_and_redirect_scrub(struct pci_dev *dev)
  521. {
  522. u32 val;
  523. /*
  524. * Suggested workaround:
  525. * set D18F3x58[4:0] = 00h and set D18F3x5C[0] = 0b
  526. */
  527. pci_read_config_dword(dev, 0x58, &val);
  528. if (val & 0x1F) {
  529. val &= ~(0x1F);
  530. pci_write_config_dword(dev, 0x58, val);
  531. }
  532. pci_read_config_dword(dev, 0x5C, &val);
  533. if (val & BIT(0)) {
  534. val &= ~BIT(0);
  535. pci_write_config_dword(dev, 0x5c, val);
  536. }
  537. }
  538. DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_16H_NB_F3,
  539. amd_disable_seq_and_redirect_scrub);
  540. #if defined(CONFIG_X86_64) && defined(CONFIG_X86_MCE)
  541. #include <linux/jump_label.h>
  542. #include <asm/string_64.h>
  543. /* Ivy Bridge, Haswell, Broadwell */
  544. static void quirk_intel_brickland_xeon_ras_cap(struct pci_dev *pdev)
  545. {
  546. u32 capid0;
  547. pci_read_config_dword(pdev, 0x84, &capid0);
  548. if (capid0 & 0x10)
  549. static_branch_inc(&mcsafe_key);
  550. }
  551. /* Skylake */
  552. static void quirk_intel_purley_xeon_ras_cap(struct pci_dev *pdev)
  553. {
  554. u32 capid0, capid5;
  555. pci_read_config_dword(pdev, 0x84, &capid0);
  556. pci_read_config_dword(pdev, 0x98, &capid5);
  557. /*
  558. * CAPID0{7:6} indicate whether this is an advanced RAS SKU
  559. * CAPID5{8:5} indicate that various NVDIMM usage modes are
  560. * enabled, so memory machine check recovery is also enabled.
  561. */
  562. if ((capid0 & 0xc0) == 0xc0 || (capid5 & 0x1e0))
  563. static_branch_inc(&mcsafe_key);
  564. }
  565. DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x0ec3, quirk_intel_brickland_xeon_ras_cap);
  566. DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2fc0, quirk_intel_brickland_xeon_ras_cap);
  567. DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x6fc0, quirk_intel_brickland_xeon_ras_cap);
  568. DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2083, quirk_intel_purley_xeon_ras_cap);
  569. #endif
  570. #endif
  571. bool x86_apple_machine;
  572. EXPORT_SYMBOL(x86_apple_machine);
  573. void __init early_platform_quirks(void)
  574. {
  575. x86_apple_machine = dmi_match(DMI_SYS_VENDOR, "Apple Inc.") ||
  576. dmi_match(DMI_SYS_VENDOR, "Apple Computer, Inc.");
  577. }