Merge tag 'gpio-for-v3.11-3' of git://git.kernel.org/pub/scm/linux/kernel/git/linusw...
[linux-2.6.git] / drivers / edac / amd64_edac.c
blob8b6a0343c2208121cf858c6a22aea1816f427f16
1 #include "amd64_edac.h"
2 #include <asm/amd_nb.h>
4 static struct edac_pci_ctl_info *amd64_ctl_pci;
6 static int report_gart_errors;
7 module_param(report_gart_errors, int, 0644);
9 /*
10 * Set by command line parameter. If BIOS has enabled the ECC, this override is
11 * cleared to prevent re-enabling the hardware by this driver.
13 static int ecc_enable_override;
14 module_param(ecc_enable_override, int, 0644);
16 static struct msr __percpu *msrs;
19 * count successfully initialized driver instances for setup_pci_device()
21 static atomic_t drv_instances = ATOMIC_INIT(0);
23 /* Per-node driver instances */
24 static struct mem_ctl_info **mcis;
25 static struct ecc_settings **ecc_stngs;
28 * Valid scrub rates for the K8 hardware memory scrubber. We map the scrubbing
29 * bandwidth to a valid bit pattern. The 'set' operation finds the 'matching-
30 * or higher value'.
32 *FIXME: Produce a better mapping/linearisation.
34 static const struct scrubrate {
35 u32 scrubval; /* bit pattern for scrub rate */
36 u32 bandwidth; /* bandwidth consumed (bytes/sec) */
37 } scrubrates[] = {
38 { 0x01, 1600000000UL},
39 { 0x02, 800000000UL},
40 { 0x03, 400000000UL},
41 { 0x04, 200000000UL},
42 { 0x05, 100000000UL},
43 { 0x06, 50000000UL},
44 { 0x07, 25000000UL},
45 { 0x08, 12284069UL},
46 { 0x09, 6274509UL},
47 { 0x0A, 3121951UL},
48 { 0x0B, 1560975UL},
49 { 0x0C, 781440UL},
50 { 0x0D, 390720UL},
51 { 0x0E, 195300UL},
52 { 0x0F, 97650UL},
53 { 0x10, 48854UL},
54 { 0x11, 24427UL},
55 { 0x12, 12213UL},
56 { 0x13, 6101UL},
57 { 0x14, 3051UL},
58 { 0x15, 1523UL},
59 { 0x16, 761UL},
60 { 0x00, 0UL}, /* scrubbing off */
63 int __amd64_read_pci_cfg_dword(struct pci_dev *pdev, int offset,
64 u32 *val, const char *func)
66 int err = 0;
68 err = pci_read_config_dword(pdev, offset, val);
69 if (err)
70 amd64_warn("%s: error reading F%dx%03x.\n",
71 func, PCI_FUNC(pdev->devfn), offset);
73 return err;
76 int __amd64_write_pci_cfg_dword(struct pci_dev *pdev, int offset,
77 u32 val, const char *func)
79 int err = 0;
81 err = pci_write_config_dword(pdev, offset, val);
82 if (err)
83 amd64_warn("%s: error writing to F%dx%03x.\n",
84 func, PCI_FUNC(pdev->devfn), offset);
86 return err;
91 * Depending on the family, F2 DCT reads need special handling:
93 * K8: has a single DCT only
95 * F10h: each DCT has its own set of regs
96 * DCT0 -> F2x040..
97 * DCT1 -> F2x140..
99 * F15h: we select which DCT we access using F1x10C[DctCfgSel]
101 * F16h: has only 1 DCT
103 static int k8_read_dct_pci_cfg(struct amd64_pvt *pvt, int addr, u32 *val,
104 const char *func)
106 if (addr >= 0x100)
107 return -EINVAL;
109 return __amd64_read_pci_cfg_dword(pvt->F2, addr, val, func);
112 static int f10_read_dct_pci_cfg(struct amd64_pvt *pvt, int addr, u32 *val,
113 const char *func)
115 return __amd64_read_pci_cfg_dword(pvt->F2, addr, val, func);
119 * Select DCT to which PCI cfg accesses are routed
121 static void f15h_select_dct(struct amd64_pvt *pvt, u8 dct)
123 u32 reg = 0;
125 amd64_read_pci_cfg(pvt->F1, DCT_CFG_SEL, &reg);
126 reg &= 0xfffffffe;
127 reg |= dct;
128 amd64_write_pci_cfg(pvt->F1, DCT_CFG_SEL, reg);
131 static int f15_read_dct_pci_cfg(struct amd64_pvt *pvt, int addr, u32 *val,
132 const char *func)
134 u8 dct = 0;
136 if (addr >= 0x140 && addr <= 0x1a0) {
137 dct = 1;
138 addr -= 0x100;
141 f15h_select_dct(pvt, dct);
143 return __amd64_read_pci_cfg_dword(pvt->F2, addr, val, func);
147 * Memory scrubber control interface. For K8, memory scrubbing is handled by
148 * hardware and can involve L2 cache, dcache as well as the main memory. With
149 * F10, this is extended to L3 cache scrubbing on CPU models sporting that
150 * functionality.
152 * This causes the "units" for the scrubbing speed to vary from 64 byte blocks
153 * (dram) over to cache lines. This is nasty, so we will use bandwidth in
154 * bytes/sec for the setting.
156 * Currently, we only do dram scrubbing. If the scrubbing is done in software on
157 * other archs, we might not have access to the caches directly.
161 * scan the scrub rate mapping table for a close or matching bandwidth value to
162 * issue. If requested is too big, then use last maximum value found.
164 static int __amd64_set_scrub_rate(struct pci_dev *ctl, u32 new_bw, u32 min_rate)
166 u32 scrubval;
167 int i;
170 * map the configured rate (new_bw) to a value specific to the AMD64
171 * memory controller and apply to register. Search for the first
172 * bandwidth entry that is greater or equal than the setting requested
173 * and program that. If at last entry, turn off DRAM scrubbing.
175 * If no suitable bandwidth is found, turn off DRAM scrubbing entirely
176 * by falling back to the last element in scrubrates[].
178 for (i = 0; i < ARRAY_SIZE(scrubrates) - 1; i++) {
180 * skip scrub rates which aren't recommended
181 * (see F10 BKDG, F3x58)
183 if (scrubrates[i].scrubval < min_rate)
184 continue;
186 if (scrubrates[i].bandwidth <= new_bw)
187 break;
190 scrubval = scrubrates[i].scrubval;
192 pci_write_bits32(ctl, SCRCTRL, scrubval, 0x001F);
194 if (scrubval)
195 return scrubrates[i].bandwidth;
197 return 0;
200 static int amd64_set_scrub_rate(struct mem_ctl_info *mci, u32 bw)
202 struct amd64_pvt *pvt = mci->pvt_info;
203 u32 min_scrubrate = 0x5;
205 if (boot_cpu_data.x86 == 0xf)
206 min_scrubrate = 0x0;
208 /* F15h Erratum #505 */
209 if (boot_cpu_data.x86 == 0x15)
210 f15h_select_dct(pvt, 0);
212 return __amd64_set_scrub_rate(pvt->F3, bw, min_scrubrate);
215 static int amd64_get_scrub_rate(struct mem_ctl_info *mci)
217 struct amd64_pvt *pvt = mci->pvt_info;
218 u32 scrubval = 0;
219 int i, retval = -EINVAL;
221 /* F15h Erratum #505 */
222 if (boot_cpu_data.x86 == 0x15)
223 f15h_select_dct(pvt, 0);
225 amd64_read_pci_cfg(pvt->F3, SCRCTRL, &scrubval);
227 scrubval = scrubval & 0x001F;
229 for (i = 0; i < ARRAY_SIZE(scrubrates); i++) {
230 if (scrubrates[i].scrubval == scrubval) {
231 retval = scrubrates[i].bandwidth;
232 break;
235 return retval;
239 * returns true if the SysAddr given by sys_addr matches the
240 * DRAM base/limit associated with node_id
242 static bool amd64_base_limit_match(struct amd64_pvt *pvt, u64 sys_addr,
243 u8 nid)
245 u64 addr;
247 /* The K8 treats this as a 40-bit value. However, bits 63-40 will be
248 * all ones if the most significant implemented address bit is 1.
249 * Here we discard bits 63-40. See section 3.4.2 of AMD publication
250 * 24592: AMD x86-64 Architecture Programmer's Manual Volume 1
251 * Application Programming.
253 addr = sys_addr & 0x000000ffffffffffull;
255 return ((addr >= get_dram_base(pvt, nid)) &&
256 (addr <= get_dram_limit(pvt, nid)));
260 * Attempt to map a SysAddr to a node. On success, return a pointer to the
261 * mem_ctl_info structure for the node that the SysAddr maps to.
263 * On failure, return NULL.
265 static struct mem_ctl_info *find_mc_by_sys_addr(struct mem_ctl_info *mci,
266 u64 sys_addr)
268 struct amd64_pvt *pvt;
269 u8 node_id;
270 u32 intlv_en, bits;
273 * Here we use the DRAM Base (section 3.4.4.1) and DRAM Limit (section
274 * 3.4.4.2) registers to map the SysAddr to a node ID.
276 pvt = mci->pvt_info;
279 * The value of this field should be the same for all DRAM Base
280 * registers. Therefore we arbitrarily choose to read it from the
281 * register for node 0.
283 intlv_en = dram_intlv_en(pvt, 0);
285 if (intlv_en == 0) {
286 for (node_id = 0; node_id < DRAM_RANGES; node_id++) {
287 if (amd64_base_limit_match(pvt, sys_addr, node_id))
288 goto found;
290 goto err_no_match;
293 if (unlikely((intlv_en != 0x01) &&
294 (intlv_en != 0x03) &&
295 (intlv_en != 0x07))) {
296 amd64_warn("DRAM Base[IntlvEn] junk value: 0x%x, BIOS bug?\n", intlv_en);
297 return NULL;
300 bits = (((u32) sys_addr) >> 12) & intlv_en;
302 for (node_id = 0; ; ) {
303 if ((dram_intlv_sel(pvt, node_id) & intlv_en) == bits)
304 break; /* intlv_sel field matches */
306 if (++node_id >= DRAM_RANGES)
307 goto err_no_match;
310 /* sanity test for sys_addr */
311 if (unlikely(!amd64_base_limit_match(pvt, sys_addr, node_id))) {
312 amd64_warn("%s: sys_addr 0x%llx falls outside base/limit address"
313 "range for node %d with node interleaving enabled.\n",
314 __func__, sys_addr, node_id);
315 return NULL;
318 found:
319 return edac_mc_find((int)node_id);
321 err_no_match:
322 edac_dbg(2, "sys_addr 0x%lx doesn't match any node\n",
323 (unsigned long)sys_addr);
325 return NULL;
329 * compute the CS base address of the @csrow on the DRAM controller @dct.
330 * For details see F2x[5C:40] in the processor's BKDG
332 static void get_cs_base_and_mask(struct amd64_pvt *pvt, int csrow, u8 dct,
333 u64 *base, u64 *mask)
335 u64 csbase, csmask, base_bits, mask_bits;
336 u8 addr_shift;
338 if (boot_cpu_data.x86 == 0xf && pvt->ext_model < K8_REV_F) {
339 csbase = pvt->csels[dct].csbases[csrow];
340 csmask = pvt->csels[dct].csmasks[csrow];
341 base_bits = GENMASK(21, 31) | GENMASK(9, 15);
342 mask_bits = GENMASK(21, 29) | GENMASK(9, 15);
343 addr_shift = 4;
346 * F16h needs two addr_shift values: 8 for high and 6 for low
347 * (cf. F16h BKDG).
349 } else if (boot_cpu_data.x86 == 0x16) {
350 csbase = pvt->csels[dct].csbases[csrow];
351 csmask = pvt->csels[dct].csmasks[csrow >> 1];
353 *base = (csbase & GENMASK(5, 15)) << 6;
354 *base |= (csbase & GENMASK(19, 30)) << 8;
356 *mask = ~0ULL;
357 /* poke holes for the csmask */
358 *mask &= ~((GENMASK(5, 15) << 6) |
359 (GENMASK(19, 30) << 8));
361 *mask |= (csmask & GENMASK(5, 15)) << 6;
362 *mask |= (csmask & GENMASK(19, 30)) << 8;
364 return;
365 } else {
366 csbase = pvt->csels[dct].csbases[csrow];
367 csmask = pvt->csels[dct].csmasks[csrow >> 1];
368 addr_shift = 8;
370 if (boot_cpu_data.x86 == 0x15)
371 base_bits = mask_bits = GENMASK(19,30) | GENMASK(5,13);
372 else
373 base_bits = mask_bits = GENMASK(19,28) | GENMASK(5,13);
376 *base = (csbase & base_bits) << addr_shift;
378 *mask = ~0ULL;
379 /* poke holes for the csmask */
380 *mask &= ~(mask_bits << addr_shift);
381 /* OR them in */
382 *mask |= (csmask & mask_bits) << addr_shift;
385 #define for_each_chip_select(i, dct, pvt) \
386 for (i = 0; i < pvt->csels[dct].b_cnt; i++)
388 #define chip_select_base(i, dct, pvt) \
389 pvt->csels[dct].csbases[i]
391 #define for_each_chip_select_mask(i, dct, pvt) \
392 for (i = 0; i < pvt->csels[dct].m_cnt; i++)
395 * @input_addr is an InputAddr associated with the node given by mci. Return the
396 * csrow that input_addr maps to, or -1 on failure (no csrow claims input_addr).
398 static int input_addr_to_csrow(struct mem_ctl_info *mci, u64 input_addr)
400 struct amd64_pvt *pvt;
401 int csrow;
402 u64 base, mask;
404 pvt = mci->pvt_info;
406 for_each_chip_select(csrow, 0, pvt) {
407 if (!csrow_enabled(csrow, 0, pvt))
408 continue;
410 get_cs_base_and_mask(pvt, csrow, 0, &base, &mask);
412 mask = ~mask;
414 if ((input_addr & mask) == (base & mask)) {
415 edac_dbg(2, "InputAddr 0x%lx matches csrow %d (node %d)\n",
416 (unsigned long)input_addr, csrow,
417 pvt->mc_node_id);
419 return csrow;
422 edac_dbg(2, "no matching csrow for InputAddr 0x%lx (MC node %d)\n",
423 (unsigned long)input_addr, pvt->mc_node_id);
425 return -1;
429 * Obtain info from the DRAM Hole Address Register (section 3.4.8, pub #26094)
430 * for the node represented by mci. Info is passed back in *hole_base,
431 * *hole_offset, and *hole_size. Function returns 0 if info is valid or 1 if
432 * info is invalid. Info may be invalid for either of the following reasons:
434 * - The revision of the node is not E or greater. In this case, the DRAM Hole
435 * Address Register does not exist.
437 * - The DramHoleValid bit is cleared in the DRAM Hole Address Register,
438 * indicating that its contents are not valid.
440 * The values passed back in *hole_base, *hole_offset, and *hole_size are
441 * complete 32-bit values despite the fact that the bitfields in the DHAR
442 * only represent bits 31-24 of the base and offset values.
444 int amd64_get_dram_hole_info(struct mem_ctl_info *mci, u64 *hole_base,
445 u64 *hole_offset, u64 *hole_size)
447 struct amd64_pvt *pvt = mci->pvt_info;
449 /* only revE and later have the DRAM Hole Address Register */
450 if (boot_cpu_data.x86 == 0xf && pvt->ext_model < K8_REV_E) {
451 edac_dbg(1, " revision %d for node %d does not support DHAR\n",
452 pvt->ext_model, pvt->mc_node_id);
453 return 1;
456 /* valid for Fam10h and above */
457 if (boot_cpu_data.x86 >= 0x10 && !dhar_mem_hoist_valid(pvt)) {
458 edac_dbg(1, " Dram Memory Hoisting is DISABLED on this system\n");
459 return 1;
462 if (!dhar_valid(pvt)) {
463 edac_dbg(1, " Dram Memory Hoisting is DISABLED on this node %d\n",
464 pvt->mc_node_id);
465 return 1;
468 /* This node has Memory Hoisting */
470 /* +------------------+--------------------+--------------------+-----
471 * | memory | DRAM hole | relocated |
472 * | [0, (x - 1)] | [x, 0xffffffff] | addresses from |
473 * | | | DRAM hole |
474 * | | | [0x100000000, |
475 * | | | (0x100000000+ |
476 * | | | (0xffffffff-x))] |
477 * +------------------+--------------------+--------------------+-----
479 * Above is a diagram of physical memory showing the DRAM hole and the
480 * relocated addresses from the DRAM hole. As shown, the DRAM hole
481 * starts at address x (the base address) and extends through address
482 * 0xffffffff. The DRAM Hole Address Register (DHAR) relocates the
483 * addresses in the hole so that they start at 0x100000000.
486 *hole_base = dhar_base(pvt);
487 *hole_size = (1ULL << 32) - *hole_base;
489 if (boot_cpu_data.x86 > 0xf)
490 *hole_offset = f10_dhar_offset(pvt);
491 else
492 *hole_offset = k8_dhar_offset(pvt);
494 edac_dbg(1, " DHAR info for node %d base 0x%lx offset 0x%lx size 0x%lx\n",
495 pvt->mc_node_id, (unsigned long)*hole_base,
496 (unsigned long)*hole_offset, (unsigned long)*hole_size);
498 return 0;
500 EXPORT_SYMBOL_GPL(amd64_get_dram_hole_info);
503 * Return the DramAddr that the SysAddr given by @sys_addr maps to. It is
504 * assumed that sys_addr maps to the node given by mci.
506 * The first part of section 3.4.4 (p. 70) shows how the DRAM Base (section
507 * 3.4.4.1) and DRAM Limit (section 3.4.4.2) registers are used to translate a
508 * SysAddr to a DramAddr. If the DRAM Hole Address Register (DHAR) is enabled,
509 * then it is also involved in translating a SysAddr to a DramAddr. Sections
510 * 3.4.8 and 3.5.8.2 describe the DHAR and how it is used for memory hoisting.
511 * These parts of the documentation are unclear. I interpret them as follows:
513 * When node n receives a SysAddr, it processes the SysAddr as follows:
515 * 1. It extracts the DRAMBase and DRAMLimit values from the DRAM Base and DRAM
516 * Limit registers for node n. If the SysAddr is not within the range
517 * specified by the base and limit values, then node n ignores the Sysaddr
518 * (since it does not map to node n). Otherwise continue to step 2 below.
520 * 2. If the DramHoleValid bit of the DHAR for node n is clear, the DHAR is
521 * disabled so skip to step 3 below. Otherwise see if the SysAddr is within
522 * the range of relocated addresses (starting at 0x100000000) from the DRAM
523 * hole. If not, skip to step 3 below. Else get the value of the
524 * DramHoleOffset field from the DHAR. To obtain the DramAddr, subtract the
525 * offset defined by this value from the SysAddr.
527 * 3. Obtain the base address for node n from the DRAMBase field of the DRAM
528 * Base register for node n. To obtain the DramAddr, subtract the base
529 * address from the SysAddr, as shown near the start of section 3.4.4 (p.70).
531 static u64 sys_addr_to_dram_addr(struct mem_ctl_info *mci, u64 sys_addr)
533 struct amd64_pvt *pvt = mci->pvt_info;
534 u64 dram_base, hole_base, hole_offset, hole_size, dram_addr;
535 int ret;
537 dram_base = get_dram_base(pvt, pvt->mc_node_id);
539 ret = amd64_get_dram_hole_info(mci, &hole_base, &hole_offset,
540 &hole_size);
541 if (!ret) {
542 if ((sys_addr >= (1ULL << 32)) &&
543 (sys_addr < ((1ULL << 32) + hole_size))) {
544 /* use DHAR to translate SysAddr to DramAddr */
545 dram_addr = sys_addr - hole_offset;
547 edac_dbg(2, "using DHAR to translate SysAddr 0x%lx to DramAddr 0x%lx\n",
548 (unsigned long)sys_addr,
549 (unsigned long)dram_addr);
551 return dram_addr;
556 * Translate the SysAddr to a DramAddr as shown near the start of
557 * section 3.4.4 (p. 70). Although sys_addr is a 64-bit value, the k8
558 * only deals with 40-bit values. Therefore we discard bits 63-40 of
559 * sys_addr below. If bit 39 of sys_addr is 1 then the bits we
560 * discard are all 1s. Otherwise the bits we discard are all 0s. See
561 * section 3.4.2 of AMD publication 24592: AMD x86-64 Architecture
562 * Programmer's Manual Volume 1 Application Programming.
564 dram_addr = (sys_addr & GENMASK(0, 39)) - dram_base;
566 edac_dbg(2, "using DRAM Base register to translate SysAddr 0x%lx to DramAddr 0x%lx\n",
567 (unsigned long)sys_addr, (unsigned long)dram_addr);
568 return dram_addr;
572 * @intlv_en is the value of the IntlvEn field from a DRAM Base register
573 * (section 3.4.4.1). Return the number of bits from a SysAddr that are used
574 * for node interleaving.
576 static int num_node_interleave_bits(unsigned intlv_en)
578 static const int intlv_shift_table[] = { 0, 1, 0, 2, 0, 0, 0, 3 };
579 int n;
581 BUG_ON(intlv_en > 7);
582 n = intlv_shift_table[intlv_en];
583 return n;
586 /* Translate the DramAddr given by @dram_addr to an InputAddr. */
587 static u64 dram_addr_to_input_addr(struct mem_ctl_info *mci, u64 dram_addr)
589 struct amd64_pvt *pvt;
590 int intlv_shift;
591 u64 input_addr;
593 pvt = mci->pvt_info;
596 * See the start of section 3.4.4 (p. 70, BKDG #26094, K8, revA-E)
597 * concerning translating a DramAddr to an InputAddr.
599 intlv_shift = num_node_interleave_bits(dram_intlv_en(pvt, 0));
600 input_addr = ((dram_addr >> intlv_shift) & GENMASK(12, 35)) +
601 (dram_addr & 0xfff);
603 edac_dbg(2, " Intlv Shift=%d DramAddr=0x%lx maps to InputAddr=0x%lx\n",
604 intlv_shift, (unsigned long)dram_addr,
605 (unsigned long)input_addr);
607 return input_addr;
611 * Translate the SysAddr represented by @sys_addr to an InputAddr. It is
612 * assumed that @sys_addr maps to the node given by mci.
614 static u64 sys_addr_to_input_addr(struct mem_ctl_info *mci, u64 sys_addr)
616 u64 input_addr;
618 input_addr =
619 dram_addr_to_input_addr(mci, sys_addr_to_dram_addr(mci, sys_addr));
621 edac_dbg(2, "SysAdddr 0x%lx translates to InputAddr 0x%lx\n",
622 (unsigned long)sys_addr, (unsigned long)input_addr);
624 return input_addr;
627 /* Map the Error address to a PAGE and PAGE OFFSET. */
628 static inline void error_address_to_page_and_offset(u64 error_address,
629 struct err_info *err)
631 err->page = (u32) (error_address >> PAGE_SHIFT);
632 err->offset = ((u32) error_address) & ~PAGE_MASK;
636 * @sys_addr is an error address (a SysAddr) extracted from the MCA NB Address
637 * Low (section 3.6.4.5) and MCA NB Address High (section 3.6.4.6) registers
638 * of a node that detected an ECC memory error. mci represents the node that
639 * the error address maps to (possibly different from the node that detected
640 * the error). Return the number of the csrow that sys_addr maps to, or -1 on
641 * error.
643 static int sys_addr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr)
645 int csrow;
647 csrow = input_addr_to_csrow(mci, sys_addr_to_input_addr(mci, sys_addr));
649 if (csrow == -1)
650 amd64_mc_err(mci, "Failed to translate InputAddr to csrow for "
651 "address 0x%lx\n", (unsigned long)sys_addr);
652 return csrow;
655 static int get_channel_from_ecc_syndrome(struct mem_ctl_info *, u16);
658 * Determine if the DIMMs have ECC enabled. ECC is enabled ONLY if all the DIMMs
659 * are ECC capable.
661 static unsigned long amd64_determine_edac_cap(struct amd64_pvt *pvt)
663 u8 bit;
664 unsigned long edac_cap = EDAC_FLAG_NONE;
666 bit = (boot_cpu_data.x86 > 0xf || pvt->ext_model >= K8_REV_F)
667 ? 19
668 : 17;
670 if (pvt->dclr0 & BIT(bit))
671 edac_cap = EDAC_FLAG_SECDED;
673 return edac_cap;
676 static void amd64_debug_display_dimm_sizes(struct amd64_pvt *, u8);
678 static void amd64_dump_dramcfg_low(u32 dclr, int chan)
680 edac_dbg(1, "F2x%d90 (DRAM Cfg Low): 0x%08x\n", chan, dclr);
682 edac_dbg(1, " DIMM type: %sbuffered; all DIMMs support ECC: %s\n",
683 (dclr & BIT(16)) ? "un" : "",
684 (dclr & BIT(19)) ? "yes" : "no");
686 edac_dbg(1, " PAR/ERR parity: %s\n",
687 (dclr & BIT(8)) ? "enabled" : "disabled");
689 if (boot_cpu_data.x86 == 0x10)
690 edac_dbg(1, " DCT 128bit mode width: %s\n",
691 (dclr & BIT(11)) ? "128b" : "64b");
693 edac_dbg(1, " x4 logical DIMMs present: L0: %s L1: %s L2: %s L3: %s\n",
694 (dclr & BIT(12)) ? "yes" : "no",
695 (dclr & BIT(13)) ? "yes" : "no",
696 (dclr & BIT(14)) ? "yes" : "no",
697 (dclr & BIT(15)) ? "yes" : "no");
700 /* Display and decode various NB registers for debug purposes. */
701 static void dump_misc_regs(struct amd64_pvt *pvt)
703 edac_dbg(1, "F3xE8 (NB Cap): 0x%08x\n", pvt->nbcap);
705 edac_dbg(1, " NB two channel DRAM capable: %s\n",
706 (pvt->nbcap & NBCAP_DCT_DUAL) ? "yes" : "no");
708 edac_dbg(1, " ECC capable: %s, ChipKill ECC capable: %s\n",
709 (pvt->nbcap & NBCAP_SECDED) ? "yes" : "no",
710 (pvt->nbcap & NBCAP_CHIPKILL) ? "yes" : "no");
712 amd64_dump_dramcfg_low(pvt->dclr0, 0);
714 edac_dbg(1, "F3xB0 (Online Spare): 0x%08x\n", pvt->online_spare);
716 edac_dbg(1, "F1xF0 (DRAM Hole Address): 0x%08x, base: 0x%08x, offset: 0x%08x\n",
717 pvt->dhar, dhar_base(pvt),
718 (boot_cpu_data.x86 == 0xf) ? k8_dhar_offset(pvt)
719 : f10_dhar_offset(pvt));
721 edac_dbg(1, " DramHoleValid: %s\n", dhar_valid(pvt) ? "yes" : "no");
723 amd64_debug_display_dimm_sizes(pvt, 0);
725 /* everything below this point is Fam10h and above */
726 if (boot_cpu_data.x86 == 0xf)
727 return;
729 amd64_debug_display_dimm_sizes(pvt, 1);
731 amd64_info("using %s syndromes.\n", ((pvt->ecc_sym_sz == 8) ? "x8" : "x4"));
733 /* Only if NOT ganged does dclr1 have valid info */
734 if (!dct_ganging_enabled(pvt))
735 amd64_dump_dramcfg_low(pvt->dclr1, 1);
739 * see BKDG, F2x[1,0][5C:40], F2[1,0][6C:60]
741 static void prep_chip_selects(struct amd64_pvt *pvt)
743 if (boot_cpu_data.x86 == 0xf && pvt->ext_model < K8_REV_F) {
744 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 8;
745 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 8;
746 } else {
747 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 8;
748 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 4;
753 * Function 2 Offset F10_DCSB0; read in the DCS Base and DCS Mask registers
755 static void read_dct_base_mask(struct amd64_pvt *pvt)
757 int cs;
759 prep_chip_selects(pvt);
761 for_each_chip_select(cs, 0, pvt) {
762 int reg0 = DCSB0 + (cs * 4);
763 int reg1 = DCSB1 + (cs * 4);
764 u32 *base0 = &pvt->csels[0].csbases[cs];
765 u32 *base1 = &pvt->csels[1].csbases[cs];
767 if (!amd64_read_dct_pci_cfg(pvt, reg0, base0))
768 edac_dbg(0, " DCSB0[%d]=0x%08x reg: F2x%x\n",
769 cs, *base0, reg0);
771 if (boot_cpu_data.x86 == 0xf || dct_ganging_enabled(pvt))
772 continue;
774 if (!amd64_read_dct_pci_cfg(pvt, reg1, base1))
775 edac_dbg(0, " DCSB1[%d]=0x%08x reg: F2x%x\n",
776 cs, *base1, reg1);
779 for_each_chip_select_mask(cs, 0, pvt) {
780 int reg0 = DCSM0 + (cs * 4);
781 int reg1 = DCSM1 + (cs * 4);
782 u32 *mask0 = &pvt->csels[0].csmasks[cs];
783 u32 *mask1 = &pvt->csels[1].csmasks[cs];
785 if (!amd64_read_dct_pci_cfg(pvt, reg0, mask0))
786 edac_dbg(0, " DCSM0[%d]=0x%08x reg: F2x%x\n",
787 cs, *mask0, reg0);
789 if (boot_cpu_data.x86 == 0xf || dct_ganging_enabled(pvt))
790 continue;
792 if (!amd64_read_dct_pci_cfg(pvt, reg1, mask1))
793 edac_dbg(0, " DCSM1[%d]=0x%08x reg: F2x%x\n",
794 cs, *mask1, reg1);
798 static enum mem_type amd64_determine_memory_type(struct amd64_pvt *pvt, int cs)
800 enum mem_type type;
802 /* F15h supports only DDR3 */
803 if (boot_cpu_data.x86 >= 0x15)
804 type = (pvt->dclr0 & BIT(16)) ? MEM_DDR3 : MEM_RDDR3;
805 else if (boot_cpu_data.x86 == 0x10 || pvt->ext_model >= K8_REV_F) {
806 if (pvt->dchr0 & DDR3_MODE)
807 type = (pvt->dclr0 & BIT(16)) ? MEM_DDR3 : MEM_RDDR3;
808 else
809 type = (pvt->dclr0 & BIT(16)) ? MEM_DDR2 : MEM_RDDR2;
810 } else {
811 type = (pvt->dclr0 & BIT(18)) ? MEM_DDR : MEM_RDDR;
814 amd64_info("CS%d: %s\n", cs, edac_mem_types[type]);
816 return type;
819 /* Get the number of DCT channels the memory controller is using. */
820 static int k8_early_channel_count(struct amd64_pvt *pvt)
822 int flag;
824 if (pvt->ext_model >= K8_REV_F)
825 /* RevF (NPT) and later */
826 flag = pvt->dclr0 & WIDTH_128;
827 else
828 /* RevE and earlier */
829 flag = pvt->dclr0 & REVE_WIDTH_128;
831 /* not used */
832 pvt->dclr1 = 0;
834 return (flag) ? 2 : 1;
837 /* On F10h and later ErrAddr is MC4_ADDR[47:1] */
838 static u64 get_error_address(struct mce *m)
840 struct cpuinfo_x86 *c = &boot_cpu_data;
841 u64 addr;
842 u8 start_bit = 1;
843 u8 end_bit = 47;
845 if (c->x86 == 0xf) {
846 start_bit = 3;
847 end_bit = 39;
850 addr = m->addr & GENMASK(start_bit, end_bit);
853 * Erratum 637 workaround
855 if (c->x86 == 0x15) {
856 struct amd64_pvt *pvt;
857 u64 cc6_base, tmp_addr;
858 u32 tmp;
859 u16 mce_nid;
860 u8 intlv_en;
862 if ((addr & GENMASK(24, 47)) >> 24 != 0x00fdf7)
863 return addr;
865 mce_nid = amd_get_nb_id(m->extcpu);
866 pvt = mcis[mce_nid]->pvt_info;
868 amd64_read_pci_cfg(pvt->F1, DRAM_LOCAL_NODE_LIM, &tmp);
869 intlv_en = tmp >> 21 & 0x7;
871 /* add [47:27] + 3 trailing bits */
872 cc6_base = (tmp & GENMASK(0, 20)) << 3;
874 /* reverse and add DramIntlvEn */
875 cc6_base |= intlv_en ^ 0x7;
877 /* pin at [47:24] */
878 cc6_base <<= 24;
880 if (!intlv_en)
881 return cc6_base | (addr & GENMASK(0, 23));
883 amd64_read_pci_cfg(pvt->F1, DRAM_LOCAL_NODE_BASE, &tmp);
885 /* faster log2 */
886 tmp_addr = (addr & GENMASK(12, 23)) << __fls(intlv_en + 1);
888 /* OR DramIntlvSel into bits [14:12] */
889 tmp_addr |= (tmp & GENMASK(21, 23)) >> 9;
891 /* add remaining [11:0] bits from original MC4_ADDR */
892 tmp_addr |= addr & GENMASK(0, 11);
894 return cc6_base | tmp_addr;
897 return addr;
900 static struct pci_dev *pci_get_related_function(unsigned int vendor,
901 unsigned int device,
902 struct pci_dev *related)
904 struct pci_dev *dev = NULL;
906 while ((dev = pci_get_device(vendor, device, dev))) {
907 if (pci_domain_nr(dev->bus) == pci_domain_nr(related->bus) &&
908 (dev->bus->number == related->bus->number) &&
909 (PCI_SLOT(dev->devfn) == PCI_SLOT(related->devfn)))
910 break;
913 return dev;
916 static void read_dram_base_limit_regs(struct amd64_pvt *pvt, unsigned range)
918 struct amd_northbridge *nb;
919 struct pci_dev *misc, *f1 = NULL;
920 struct cpuinfo_x86 *c = &boot_cpu_data;
921 int off = range << 3;
922 u32 llim;
924 amd64_read_pci_cfg(pvt->F1, DRAM_BASE_LO + off, &pvt->ranges[range].base.lo);
925 amd64_read_pci_cfg(pvt->F1, DRAM_LIMIT_LO + off, &pvt->ranges[range].lim.lo);
927 if (c->x86 == 0xf)
928 return;
930 if (!dram_rw(pvt, range))
931 return;
933 amd64_read_pci_cfg(pvt->F1, DRAM_BASE_HI + off, &pvt->ranges[range].base.hi);
934 amd64_read_pci_cfg(pvt->F1, DRAM_LIMIT_HI + off, &pvt->ranges[range].lim.hi);
936 /* F15h: factor in CC6 save area by reading dst node's limit reg */
937 if (c->x86 != 0x15)
938 return;
940 nb = node_to_amd_nb(dram_dst_node(pvt, range));
941 if (WARN_ON(!nb))
942 return;
944 misc = nb->misc;
945 f1 = pci_get_related_function(misc->vendor, PCI_DEVICE_ID_AMD_15H_NB_F1, misc);
946 if (WARN_ON(!f1))
947 return;
949 amd64_read_pci_cfg(f1, DRAM_LOCAL_NODE_LIM, &llim);
951 pvt->ranges[range].lim.lo &= GENMASK(0, 15);
953 /* {[39:27],111b} */
954 pvt->ranges[range].lim.lo |= ((llim & 0x1fff) << 3 | 0x7) << 16;
956 pvt->ranges[range].lim.hi &= GENMASK(0, 7);
958 /* [47:40] */
959 pvt->ranges[range].lim.hi |= llim >> 13;
961 pci_dev_put(f1);
964 static void k8_map_sysaddr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr,
965 struct err_info *err)
967 struct amd64_pvt *pvt = mci->pvt_info;
969 error_address_to_page_and_offset(sys_addr, err);
972 * Find out which node the error address belongs to. This may be
973 * different from the node that detected the error.
975 err->src_mci = find_mc_by_sys_addr(mci, sys_addr);
976 if (!err->src_mci) {
977 amd64_mc_err(mci, "failed to map error addr 0x%lx to a node\n",
978 (unsigned long)sys_addr);
979 err->err_code = ERR_NODE;
980 return;
983 /* Now map the sys_addr to a CSROW */
984 err->csrow = sys_addr_to_csrow(err->src_mci, sys_addr);
985 if (err->csrow < 0) {
986 err->err_code = ERR_CSROW;
987 return;
990 /* CHIPKILL enabled */
991 if (pvt->nbcfg & NBCFG_CHIPKILL) {
992 err->channel = get_channel_from_ecc_syndrome(mci, err->syndrome);
993 if (err->channel < 0) {
995 * Syndrome didn't map, so we don't know which of the
996 * 2 DIMMs is in error. So we need to ID 'both' of them
997 * as suspect.
999 amd64_mc_warn(err->src_mci, "unknown syndrome 0x%04x - "
1000 "possible error reporting race\n",
1001 err->syndrome);
1002 err->err_code = ERR_CHANNEL;
1003 return;
1005 } else {
1007 * non-chipkill ecc mode
1009 * The k8 documentation is unclear about how to determine the
1010 * channel number when using non-chipkill memory. This method
1011 * was obtained from email communication with someone at AMD.
1012 * (Wish the email was placed in this comment - norsk)
1014 err->channel = ((sys_addr & BIT(3)) != 0);
1018 static int ddr2_cs_size(unsigned i, bool dct_width)
1020 unsigned shift = 0;
1022 if (i <= 2)
1023 shift = i;
1024 else if (!(i & 0x1))
1025 shift = i >> 1;
1026 else
1027 shift = (i + 1) >> 1;
1029 return 128 << (shift + !!dct_width);
1032 static int k8_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
1033 unsigned cs_mode)
1035 u32 dclr = dct ? pvt->dclr1 : pvt->dclr0;
1037 if (pvt->ext_model >= K8_REV_F) {
1038 WARN_ON(cs_mode > 11);
1039 return ddr2_cs_size(cs_mode, dclr & WIDTH_128);
1041 else if (pvt->ext_model >= K8_REV_D) {
1042 unsigned diff;
1043 WARN_ON(cs_mode > 10);
1046 * the below calculation, besides trying to win an obfuscated C
1047 * contest, maps cs_mode values to DIMM chip select sizes. The
1048 * mappings are:
1050 * cs_mode CS size (mb)
1051 * ======= ============
1052 * 0 32
1053 * 1 64
1054 * 2 128
1055 * 3 128
1056 * 4 256
1057 * 5 512
1058 * 6 256
1059 * 7 512
1060 * 8 1024
1061 * 9 1024
1062 * 10 2048
1064 * Basically, it calculates a value with which to shift the
1065 * smallest CS size of 32MB.
1067 * ddr[23]_cs_size have a similar purpose.
1069 diff = cs_mode/3 + (unsigned)(cs_mode > 5);
1071 return 32 << (cs_mode - diff);
1073 else {
1074 WARN_ON(cs_mode > 6);
1075 return 32 << cs_mode;
1080 * Get the number of DCT channels in use.
1082 * Return:
1083 * number of Memory Channels in operation
1084 * Pass back:
1085 * contents of the DCL0_LOW register
1087 static int f1x_early_channel_count(struct amd64_pvt *pvt)
1089 int i, j, channels = 0;
1091 /* On F10h, if we are in 128 bit mode, then we are using 2 channels */
1092 if (boot_cpu_data.x86 == 0x10 && (pvt->dclr0 & WIDTH_128))
1093 return 2;
1096 * Need to check if in unganged mode: In such, there are 2 channels,
1097 * but they are not in 128 bit mode and thus the above 'dclr0' status
1098 * bit will be OFF.
1100 * Need to check DCT0[0] and DCT1[0] to see if only one of them has
1101 * their CSEnable bit on. If so, then SINGLE DIMM case.
1103 edac_dbg(0, "Data width is not 128 bits - need more decoding\n");
1106 * Check DRAM Bank Address Mapping values for each DIMM to see if there
1107 * is more than just one DIMM present in unganged mode. Need to check
1108 * both controllers since DIMMs can be placed in either one.
1110 for (i = 0; i < 2; i++) {
1111 u32 dbam = (i ? pvt->dbam1 : pvt->dbam0);
1113 for (j = 0; j < 4; j++) {
1114 if (DBAM_DIMM(j, dbam) > 0) {
1115 channels++;
1116 break;
1121 if (channels > 2)
1122 channels = 2;
1124 amd64_info("MCT channel count: %d\n", channels);
1126 return channels;
1129 static int ddr3_cs_size(unsigned i, bool dct_width)
1131 unsigned shift = 0;
1132 int cs_size = 0;
1134 if (i == 0 || i == 3 || i == 4)
1135 cs_size = -1;
1136 else if (i <= 2)
1137 shift = i;
1138 else if (i == 12)
1139 shift = 7;
1140 else if (!(i & 0x1))
1141 shift = i >> 1;
1142 else
1143 shift = (i + 1) >> 1;
1145 if (cs_size != -1)
1146 cs_size = (128 * (1 << !!dct_width)) << shift;
1148 return cs_size;
1151 static int f10_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
1152 unsigned cs_mode)
1154 u32 dclr = dct ? pvt->dclr1 : pvt->dclr0;
1156 WARN_ON(cs_mode > 11);
1158 if (pvt->dchr0 & DDR3_MODE || pvt->dchr1 & DDR3_MODE)
1159 return ddr3_cs_size(cs_mode, dclr & WIDTH_128);
1160 else
1161 return ddr2_cs_size(cs_mode, dclr & WIDTH_128);
1165 * F15h supports only 64bit DCT interfaces
1167 static int f15_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
1168 unsigned cs_mode)
1170 WARN_ON(cs_mode > 12);
1172 return ddr3_cs_size(cs_mode, false);
1176 * F16h has only limited cs_modes
1178 static int f16_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct,
1179 unsigned cs_mode)
1181 WARN_ON(cs_mode > 12);
1183 if (cs_mode == 6 || cs_mode == 8 ||
1184 cs_mode == 9 || cs_mode == 12)
1185 return -1;
1186 else
1187 return ddr3_cs_size(cs_mode, false);
1190 static void read_dram_ctl_register(struct amd64_pvt *pvt)
1193 if (boot_cpu_data.x86 == 0xf)
1194 return;
1196 if (!amd64_read_dct_pci_cfg(pvt, DCT_SEL_LO, &pvt->dct_sel_lo)) {
1197 edac_dbg(0, "F2x110 (DCTSelLow): 0x%08x, High range addrs at: 0x%x\n",
1198 pvt->dct_sel_lo, dct_sel_baseaddr(pvt));
1200 edac_dbg(0, " DCTs operate in %s mode\n",
1201 (dct_ganging_enabled(pvt) ? "ganged" : "unganged"));
1203 if (!dct_ganging_enabled(pvt))
1204 edac_dbg(0, " Address range split per DCT: %s\n",
1205 (dct_high_range_enabled(pvt) ? "yes" : "no"));
1207 edac_dbg(0, " data interleave for ECC: %s, DRAM cleared since last warm reset: %s\n",
1208 (dct_data_intlv_enabled(pvt) ? "enabled" : "disabled"),
1209 (dct_memory_cleared(pvt) ? "yes" : "no"));
1211 edac_dbg(0, " channel interleave: %s, "
1212 "interleave bits selector: 0x%x\n",
1213 (dct_interleave_enabled(pvt) ? "enabled" : "disabled"),
1214 dct_sel_interleave_addr(pvt));
1217 amd64_read_dct_pci_cfg(pvt, DCT_SEL_HI, &pvt->dct_sel_hi);
1221 * Determine channel (DCT) based on the interleaving mode: F10h BKDG, 2.8.9 Memory
1222 * Interleaving Modes.
1224 static u8 f1x_determine_channel(struct amd64_pvt *pvt, u64 sys_addr,
1225 bool hi_range_sel, u8 intlv_en)
1227 u8 dct_sel_high = (pvt->dct_sel_lo >> 1) & 1;
1229 if (dct_ganging_enabled(pvt))
1230 return 0;
1232 if (hi_range_sel)
1233 return dct_sel_high;
1236 * see F2x110[DctSelIntLvAddr] - channel interleave mode
1238 if (dct_interleave_enabled(pvt)) {
1239 u8 intlv_addr = dct_sel_interleave_addr(pvt);
1241 /* return DCT select function: 0=DCT0, 1=DCT1 */
1242 if (!intlv_addr)
1243 return sys_addr >> 6 & 1;
1245 if (intlv_addr & 0x2) {
1246 u8 shift = intlv_addr & 0x1 ? 9 : 6;
1247 u32 temp = hweight_long((u32) ((sys_addr >> 16) & 0x1F)) % 2;
1249 return ((sys_addr >> shift) & 1) ^ temp;
1252 return (sys_addr >> (12 + hweight8(intlv_en))) & 1;
1255 if (dct_high_range_enabled(pvt))
1256 return ~dct_sel_high & 1;
1258 return 0;
1261 /* Convert the sys_addr to the normalized DCT address */
1262 static u64 f1x_get_norm_dct_addr(struct amd64_pvt *pvt, u8 range,
1263 u64 sys_addr, bool hi_rng,
1264 u32 dct_sel_base_addr)
1266 u64 chan_off;
1267 u64 dram_base = get_dram_base(pvt, range);
1268 u64 hole_off = f10_dhar_offset(pvt);
1269 u64 dct_sel_base_off = (pvt->dct_sel_hi & 0xFFFFFC00) << 16;
1271 if (hi_rng) {
1273 * if
1274 * base address of high range is below 4Gb
1275 * (bits [47:27] at [31:11])
1276 * DRAM address space on this DCT is hoisted above 4Gb &&
1277 * sys_addr > 4Gb
1279 * remove hole offset from sys_addr
1280 * else
1281 * remove high range offset from sys_addr
1283 if ((!(dct_sel_base_addr >> 16) ||
1284 dct_sel_base_addr < dhar_base(pvt)) &&
1285 dhar_valid(pvt) &&
1286 (sys_addr >= BIT_64(32)))
1287 chan_off = hole_off;
1288 else
1289 chan_off = dct_sel_base_off;
1290 } else {
1292 * if
1293 * we have a valid hole &&
1294 * sys_addr > 4Gb
1296 * remove hole
1297 * else
1298 * remove dram base to normalize to DCT address
1300 if (dhar_valid(pvt) && (sys_addr >= BIT_64(32)))
1301 chan_off = hole_off;
1302 else
1303 chan_off = dram_base;
1306 return (sys_addr & GENMASK(6,47)) - (chan_off & GENMASK(23,47));
1310 * checks if the csrow passed in is marked as SPARED, if so returns the new
1311 * spare row
1313 static int f10_process_possible_spare(struct amd64_pvt *pvt, u8 dct, int csrow)
1315 int tmp_cs;
1317 if (online_spare_swap_done(pvt, dct) &&
1318 csrow == online_spare_bad_dramcs(pvt, dct)) {
1320 for_each_chip_select(tmp_cs, dct, pvt) {
1321 if (chip_select_base(tmp_cs, dct, pvt) & 0x2) {
1322 csrow = tmp_cs;
1323 break;
1327 return csrow;
1331 * Iterate over the DRAM DCT "base" and "mask" registers looking for a
1332 * SystemAddr match on the specified 'ChannelSelect' and 'NodeID'
1334 * Return:
1335 * -EINVAL: NOT FOUND
1336 * 0..csrow = Chip-Select Row
1338 static int f1x_lookup_addr_in_dct(u64 in_addr, u8 nid, u8 dct)
1340 struct mem_ctl_info *mci;
1341 struct amd64_pvt *pvt;
1342 u64 cs_base, cs_mask;
1343 int cs_found = -EINVAL;
1344 int csrow;
1346 mci = mcis[nid];
1347 if (!mci)
1348 return cs_found;
1350 pvt = mci->pvt_info;
1352 edac_dbg(1, "input addr: 0x%llx, DCT: %d\n", in_addr, dct);
1354 for_each_chip_select(csrow, dct, pvt) {
1355 if (!csrow_enabled(csrow, dct, pvt))
1356 continue;
1358 get_cs_base_and_mask(pvt, csrow, dct, &cs_base, &cs_mask);
1360 edac_dbg(1, " CSROW=%d CSBase=0x%llx CSMask=0x%llx\n",
1361 csrow, cs_base, cs_mask);
1363 cs_mask = ~cs_mask;
1365 edac_dbg(1, " (InputAddr & ~CSMask)=0x%llx (CSBase & ~CSMask)=0x%llx\n",
1366 (in_addr & cs_mask), (cs_base & cs_mask));
1368 if ((in_addr & cs_mask) == (cs_base & cs_mask)) {
1369 cs_found = f10_process_possible_spare(pvt, dct, csrow);
1371 edac_dbg(1, " MATCH csrow=%d\n", cs_found);
1372 break;
1375 return cs_found;
1379 * See F2x10C. Non-interleaved graphics framebuffer memory under the 16G is
1380 * swapped with a region located at the bottom of memory so that the GPU can use
1381 * the interleaved region and thus two channels.
1383 static u64 f1x_swap_interleaved_region(struct amd64_pvt *pvt, u64 sys_addr)
1385 u32 swap_reg, swap_base, swap_limit, rgn_size, tmp_addr;
1387 if (boot_cpu_data.x86 == 0x10) {
1388 /* only revC3 and revE have that feature */
1389 if (boot_cpu_data.x86_model < 4 ||
1390 (boot_cpu_data.x86_model < 0xa &&
1391 boot_cpu_data.x86_mask < 3))
1392 return sys_addr;
1395 amd64_read_dct_pci_cfg(pvt, SWAP_INTLV_REG, &swap_reg);
1397 if (!(swap_reg & 0x1))
1398 return sys_addr;
1400 swap_base = (swap_reg >> 3) & 0x7f;
1401 swap_limit = (swap_reg >> 11) & 0x7f;
1402 rgn_size = (swap_reg >> 20) & 0x7f;
1403 tmp_addr = sys_addr >> 27;
1405 if (!(sys_addr >> 34) &&
1406 (((tmp_addr >= swap_base) &&
1407 (tmp_addr <= swap_limit)) ||
1408 (tmp_addr < rgn_size)))
1409 return sys_addr ^ (u64)swap_base << 27;
1411 return sys_addr;
1414 /* For a given @dram_range, check if @sys_addr falls within it. */
1415 static int f1x_match_to_this_node(struct amd64_pvt *pvt, unsigned range,
1416 u64 sys_addr, int *chan_sel)
1418 int cs_found = -EINVAL;
1419 u64 chan_addr;
1420 u32 dct_sel_base;
1421 u8 channel;
1422 bool high_range = false;
1424 u8 node_id = dram_dst_node(pvt, range);
1425 u8 intlv_en = dram_intlv_en(pvt, range);
1426 u32 intlv_sel = dram_intlv_sel(pvt, range);
1428 edac_dbg(1, "(range %d) SystemAddr= 0x%llx Limit=0x%llx\n",
1429 range, sys_addr, get_dram_limit(pvt, range));
1431 if (dhar_valid(pvt) &&
1432 dhar_base(pvt) <= sys_addr &&
1433 sys_addr < BIT_64(32)) {
1434 amd64_warn("Huh? Address is in the MMIO hole: 0x%016llx\n",
1435 sys_addr);
1436 return -EINVAL;
1439 if (intlv_en && (intlv_sel != ((sys_addr >> 12) & intlv_en)))
1440 return -EINVAL;
1442 sys_addr = f1x_swap_interleaved_region(pvt, sys_addr);
1444 dct_sel_base = dct_sel_baseaddr(pvt);
1447 * check whether addresses >= DctSelBaseAddr[47:27] are to be used to
1448 * select between DCT0 and DCT1.
1450 if (dct_high_range_enabled(pvt) &&
1451 !dct_ganging_enabled(pvt) &&
1452 ((sys_addr >> 27) >= (dct_sel_base >> 11)))
1453 high_range = true;
1455 channel = f1x_determine_channel(pvt, sys_addr, high_range, intlv_en);
1457 chan_addr = f1x_get_norm_dct_addr(pvt, range, sys_addr,
1458 high_range, dct_sel_base);
1460 /* Remove node interleaving, see F1x120 */
1461 if (intlv_en)
1462 chan_addr = ((chan_addr >> (12 + hweight8(intlv_en))) << 12) |
1463 (chan_addr & 0xfff);
1465 /* remove channel interleave */
1466 if (dct_interleave_enabled(pvt) &&
1467 !dct_high_range_enabled(pvt) &&
1468 !dct_ganging_enabled(pvt)) {
1470 if (dct_sel_interleave_addr(pvt) != 1) {
1471 if (dct_sel_interleave_addr(pvt) == 0x3)
1472 /* hash 9 */
1473 chan_addr = ((chan_addr >> 10) << 9) |
1474 (chan_addr & 0x1ff);
1475 else
1476 /* A[6] or hash 6 */
1477 chan_addr = ((chan_addr >> 7) << 6) |
1478 (chan_addr & 0x3f);
1479 } else
1480 /* A[12] */
1481 chan_addr = ((chan_addr >> 13) << 12) |
1482 (chan_addr & 0xfff);
1485 edac_dbg(1, " Normalized DCT addr: 0x%llx\n", chan_addr);
1487 cs_found = f1x_lookup_addr_in_dct(chan_addr, node_id, channel);
1489 if (cs_found >= 0)
1490 *chan_sel = channel;
1492 return cs_found;
1495 static int f1x_translate_sysaddr_to_cs(struct amd64_pvt *pvt, u64 sys_addr,
1496 int *chan_sel)
1498 int cs_found = -EINVAL;
1499 unsigned range;
1501 for (range = 0; range < DRAM_RANGES; range++) {
1503 if (!dram_rw(pvt, range))
1504 continue;
1506 if ((get_dram_base(pvt, range) <= sys_addr) &&
1507 (get_dram_limit(pvt, range) >= sys_addr)) {
1509 cs_found = f1x_match_to_this_node(pvt, range,
1510 sys_addr, chan_sel);
1511 if (cs_found >= 0)
1512 break;
1515 return cs_found;
1519 * For reference see "2.8.5 Routing DRAM Requests" in F10 BKDG. This code maps
1520 * a @sys_addr to NodeID, DCT (channel) and chip select (CSROW).
1522 * The @sys_addr is usually an error address received from the hardware
1523 * (MCX_ADDR).
1525 static void f1x_map_sysaddr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr,
1526 struct err_info *err)
1528 struct amd64_pvt *pvt = mci->pvt_info;
1530 error_address_to_page_and_offset(sys_addr, err);
1532 err->csrow = f1x_translate_sysaddr_to_cs(pvt, sys_addr, &err->channel);
1533 if (err->csrow < 0) {
1534 err->err_code = ERR_CSROW;
1535 return;
1539 * We need the syndromes for channel detection only when we're
1540 * ganged. Otherwise @chan should already contain the channel at
1541 * this point.
1543 if (dct_ganging_enabled(pvt))
1544 err->channel = get_channel_from_ecc_syndrome(mci, err->syndrome);
1548 * debug routine to display the memory sizes of all logical DIMMs and its
1549 * CSROWs
1551 static void amd64_debug_display_dimm_sizes(struct amd64_pvt *pvt, u8 ctrl)
1553 int dimm, size0, size1;
1554 u32 *dcsb = ctrl ? pvt->csels[1].csbases : pvt->csels[0].csbases;
1555 u32 dbam = ctrl ? pvt->dbam1 : pvt->dbam0;
1557 if (boot_cpu_data.x86 == 0xf) {
1558 /* K8 families < revF not supported yet */
1559 if (pvt->ext_model < K8_REV_F)
1560 return;
1561 else
1562 WARN_ON(ctrl != 0);
1565 dbam = (ctrl && !dct_ganging_enabled(pvt)) ? pvt->dbam1 : pvt->dbam0;
1566 dcsb = (ctrl && !dct_ganging_enabled(pvt)) ? pvt->csels[1].csbases
1567 : pvt->csels[0].csbases;
1569 edac_dbg(1, "F2x%d80 (DRAM Bank Address Mapping): 0x%08x\n",
1570 ctrl, dbam);
1572 edac_printk(KERN_DEBUG, EDAC_MC, "DCT%d chip selects:\n", ctrl);
1574 /* Dump memory sizes for DIMM and its CSROWs */
1575 for (dimm = 0; dimm < 4; dimm++) {
1577 size0 = 0;
1578 if (dcsb[dimm*2] & DCSB_CS_ENABLE)
1579 size0 = pvt->ops->dbam_to_cs(pvt, ctrl,
1580 DBAM_DIMM(dimm, dbam));
1582 size1 = 0;
1583 if (dcsb[dimm*2 + 1] & DCSB_CS_ENABLE)
1584 size1 = pvt->ops->dbam_to_cs(pvt, ctrl,
1585 DBAM_DIMM(dimm, dbam));
1587 amd64_info(EDAC_MC ": %d: %5dMB %d: %5dMB\n",
1588 dimm * 2, size0,
1589 dimm * 2 + 1, size1);
1593 static struct amd64_family_type amd64_family_types[] = {
1594 [K8_CPUS] = {
1595 .ctl_name = "K8",
1596 .f1_id = PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP,
1597 .f3_id = PCI_DEVICE_ID_AMD_K8_NB_MISC,
1598 .ops = {
1599 .early_channel_count = k8_early_channel_count,
1600 .map_sysaddr_to_csrow = k8_map_sysaddr_to_csrow,
1601 .dbam_to_cs = k8_dbam_to_chip_select,
1602 .read_dct_pci_cfg = k8_read_dct_pci_cfg,
1605 [F10_CPUS] = {
1606 .ctl_name = "F10h",
1607 .f1_id = PCI_DEVICE_ID_AMD_10H_NB_MAP,
1608 .f3_id = PCI_DEVICE_ID_AMD_10H_NB_MISC,
1609 .ops = {
1610 .early_channel_count = f1x_early_channel_count,
1611 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
1612 .dbam_to_cs = f10_dbam_to_chip_select,
1613 .read_dct_pci_cfg = f10_read_dct_pci_cfg,
1616 [F15_CPUS] = {
1617 .ctl_name = "F15h",
1618 .f1_id = PCI_DEVICE_ID_AMD_15H_NB_F1,
1619 .f3_id = PCI_DEVICE_ID_AMD_15H_NB_F3,
1620 .ops = {
1621 .early_channel_count = f1x_early_channel_count,
1622 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
1623 .dbam_to_cs = f15_dbam_to_chip_select,
1624 .read_dct_pci_cfg = f15_read_dct_pci_cfg,
1627 [F16_CPUS] = {
1628 .ctl_name = "F16h",
1629 .f1_id = PCI_DEVICE_ID_AMD_16H_NB_F1,
1630 .f3_id = PCI_DEVICE_ID_AMD_16H_NB_F3,
1631 .ops = {
1632 .early_channel_count = f1x_early_channel_count,
1633 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow,
1634 .dbam_to_cs = f16_dbam_to_chip_select,
1635 .read_dct_pci_cfg = f10_read_dct_pci_cfg,
1641 * These are tables of eigenvectors (one per line) which can be used for the
1642 * construction of the syndrome tables. The modified syndrome search algorithm
1643 * uses those to find the symbol in error and thus the DIMM.
1645 * Algorithm courtesy of Ross LaFetra from AMD.
1647 static const u16 x4_vectors[] = {
1648 0x2f57, 0x1afe, 0x66cc, 0xdd88,
1649 0x11eb, 0x3396, 0x7f4c, 0xeac8,
1650 0x0001, 0x0002, 0x0004, 0x0008,
1651 0x1013, 0x3032, 0x4044, 0x8088,
1652 0x106b, 0x30d6, 0x70fc, 0xe0a8,
1653 0x4857, 0xc4fe, 0x13cc, 0x3288,
1654 0x1ac5, 0x2f4a, 0x5394, 0xa1e8,
1655 0x1f39, 0x251e, 0xbd6c, 0x6bd8,
1656 0x15c1, 0x2a42, 0x89ac, 0x4758,
1657 0x2b03, 0x1602, 0x4f0c, 0xca08,
1658 0x1f07, 0x3a0e, 0x6b04, 0xbd08,
1659 0x8ba7, 0x465e, 0x244c, 0x1cc8,
1660 0x2b87, 0x164e, 0x642c, 0xdc18,
1661 0x40b9, 0x80de, 0x1094, 0x20e8,
1662 0x27db, 0x1eb6, 0x9dac, 0x7b58,
1663 0x11c1, 0x2242, 0x84ac, 0x4c58,
1664 0x1be5, 0x2d7a, 0x5e34, 0xa718,
1665 0x4b39, 0x8d1e, 0x14b4, 0x28d8,
1666 0x4c97, 0xc87e, 0x11fc, 0x33a8,
1667 0x8e97, 0x497e, 0x2ffc, 0x1aa8,
1668 0x16b3, 0x3d62, 0x4f34, 0x8518,
1669 0x1e2f, 0x391a, 0x5cac, 0xf858,
1670 0x1d9f, 0x3b7a, 0x572c, 0xfe18,
1671 0x15f5, 0x2a5a, 0x5264, 0xa3b8,
1672 0x1dbb, 0x3b66, 0x715c, 0xe3f8,
1673 0x4397, 0xc27e, 0x17fc, 0x3ea8,
1674 0x1617, 0x3d3e, 0x6464, 0xb8b8,
1675 0x23ff, 0x12aa, 0xab6c, 0x56d8,
1676 0x2dfb, 0x1ba6, 0x913c, 0x7328,
1677 0x185d, 0x2ca6, 0x7914, 0x9e28,
1678 0x171b, 0x3e36, 0x7d7c, 0xebe8,
1679 0x4199, 0x82ee, 0x19f4, 0x2e58,
1680 0x4807, 0xc40e, 0x130c, 0x3208,
1681 0x1905, 0x2e0a, 0x5804, 0xac08,
1682 0x213f, 0x132a, 0xadfc, 0x5ba8,
1683 0x19a9, 0x2efe, 0xb5cc, 0x6f88,
1686 static const u16 x8_vectors[] = {
1687 0x0145, 0x028a, 0x2374, 0x43c8, 0xa1f0, 0x0520, 0x0a40, 0x1480,
1688 0x0211, 0x0422, 0x0844, 0x1088, 0x01b0, 0x44e0, 0x23c0, 0xed80,
1689 0x1011, 0x0116, 0x022c, 0x0458, 0x08b0, 0x8c60, 0x2740, 0x4e80,
1690 0x0411, 0x0822, 0x1044, 0x0158, 0x02b0, 0x2360, 0x46c0, 0xab80,
1691 0x0811, 0x1022, 0x012c, 0x0258, 0x04b0, 0x4660, 0x8cc0, 0x2780,
1692 0x2071, 0x40e2, 0xa0c4, 0x0108, 0x0210, 0x0420, 0x0840, 0x1080,
1693 0x4071, 0x80e2, 0x0104, 0x0208, 0x0410, 0x0820, 0x1040, 0x2080,
1694 0x8071, 0x0102, 0x0204, 0x0408, 0x0810, 0x1020, 0x2040, 0x4080,
1695 0x019d, 0x03d6, 0x136c, 0x2198, 0x50b0, 0xb2e0, 0x0740, 0x0e80,
1696 0x0189, 0x03ea, 0x072c, 0x0e58, 0x1cb0, 0x56e0, 0x37c0, 0xf580,
1697 0x01fd, 0x0376, 0x06ec, 0x0bb8, 0x1110, 0x2220, 0x4440, 0x8880,
1698 0x0163, 0x02c6, 0x1104, 0x0758, 0x0eb0, 0x2be0, 0x6140, 0xc280,
1699 0x02fd, 0x01c6, 0x0b5c, 0x1108, 0x07b0, 0x25a0, 0x8840, 0x6180,
1700 0x0801, 0x012e, 0x025c, 0x04b8, 0x1370, 0x26e0, 0x57c0, 0xb580,
1701 0x0401, 0x0802, 0x015c, 0x02b8, 0x22b0, 0x13e0, 0x7140, 0xe280,
1702 0x0201, 0x0402, 0x0804, 0x01b8, 0x11b0, 0x31a0, 0x8040, 0x7180,
1703 0x0101, 0x0202, 0x0404, 0x0808, 0x1010, 0x2020, 0x4040, 0x8080,
1704 0x0001, 0x0002, 0x0004, 0x0008, 0x0010, 0x0020, 0x0040, 0x0080,
1705 0x0100, 0x0200, 0x0400, 0x0800, 0x1000, 0x2000, 0x4000, 0x8000,
1708 static int decode_syndrome(u16 syndrome, const u16 *vectors, unsigned num_vecs,
1709 unsigned v_dim)
1711 unsigned int i, err_sym;
1713 for (err_sym = 0; err_sym < num_vecs / v_dim; err_sym++) {
1714 u16 s = syndrome;
1715 unsigned v_idx = err_sym * v_dim;
1716 unsigned v_end = (err_sym + 1) * v_dim;
1718 /* walk over all 16 bits of the syndrome */
1719 for (i = 1; i < (1U << 16); i <<= 1) {
1721 /* if bit is set in that eigenvector... */
1722 if (v_idx < v_end && vectors[v_idx] & i) {
1723 u16 ev_comp = vectors[v_idx++];
1725 /* ... and bit set in the modified syndrome, */
1726 if (s & i) {
1727 /* remove it. */
1728 s ^= ev_comp;
1730 if (!s)
1731 return err_sym;
1734 } else if (s & i)
1735 /* can't get to zero, move to next symbol */
1736 break;
1740 edac_dbg(0, "syndrome(%x) not found\n", syndrome);
1741 return -1;
1744 static int map_err_sym_to_channel(int err_sym, int sym_size)
1746 if (sym_size == 4)
1747 switch (err_sym) {
1748 case 0x20:
1749 case 0x21:
1750 return 0;
1751 break;
1752 case 0x22:
1753 case 0x23:
1754 return 1;
1755 break;
1756 default:
1757 return err_sym >> 4;
1758 break;
1760 /* x8 symbols */
1761 else
1762 switch (err_sym) {
1763 /* imaginary bits not in a DIMM */
1764 case 0x10:
1765 WARN(1, KERN_ERR "Invalid error symbol: 0x%x\n",
1766 err_sym);
1767 return -1;
1768 break;
1770 case 0x11:
1771 return 0;
1772 break;
1773 case 0x12:
1774 return 1;
1775 break;
1776 default:
1777 return err_sym >> 3;
1778 break;
1780 return -1;
1783 static int get_channel_from_ecc_syndrome(struct mem_ctl_info *mci, u16 syndrome)
1785 struct amd64_pvt *pvt = mci->pvt_info;
1786 int err_sym = -1;
1788 if (pvt->ecc_sym_sz == 8)
1789 err_sym = decode_syndrome(syndrome, x8_vectors,
1790 ARRAY_SIZE(x8_vectors),
1791 pvt->ecc_sym_sz);
1792 else if (pvt->ecc_sym_sz == 4)
1793 err_sym = decode_syndrome(syndrome, x4_vectors,
1794 ARRAY_SIZE(x4_vectors),
1795 pvt->ecc_sym_sz);
1796 else {
1797 amd64_warn("Illegal syndrome type: %u\n", pvt->ecc_sym_sz);
1798 return err_sym;
1801 return map_err_sym_to_channel(err_sym, pvt->ecc_sym_sz);
1804 static void __log_bus_error(struct mem_ctl_info *mci, struct err_info *err,
1805 u8 ecc_type)
1807 enum hw_event_mc_err_type err_type;
1808 const char *string;
1810 if (ecc_type == 2)
1811 err_type = HW_EVENT_ERR_CORRECTED;
1812 else if (ecc_type == 1)
1813 err_type = HW_EVENT_ERR_UNCORRECTED;
1814 else {
1815 WARN(1, "Something is rotten in the state of Denmark.\n");
1816 return;
1819 switch (err->err_code) {
1820 case DECODE_OK:
1821 string = "";
1822 break;
1823 case ERR_NODE:
1824 string = "Failed to map error addr to a node";
1825 break;
1826 case ERR_CSROW:
1827 string = "Failed to map error addr to a csrow";
1828 break;
1829 case ERR_CHANNEL:
1830 string = "unknown syndrome - possible error reporting race";
1831 break;
1832 default:
1833 string = "WTF error";
1834 break;
1837 edac_mc_handle_error(err_type, mci, 1,
1838 err->page, err->offset, err->syndrome,
1839 err->csrow, err->channel, -1,
1840 string, "");
1843 static inline void __amd64_decode_bus_error(struct mem_ctl_info *mci,
1844 struct mce *m)
1846 struct amd64_pvt *pvt = mci->pvt_info;
1847 u8 ecc_type = (m->status >> 45) & 0x3;
1848 u8 xec = XEC(m->status, 0x1f);
1849 u16 ec = EC(m->status);
1850 u64 sys_addr;
1851 struct err_info err;
1853 /* Bail out early if this was an 'observed' error */
1854 if (PP(ec) == NBSL_PP_OBS)
1855 return;
1857 /* Do only ECC errors */
1858 if (xec && xec != F10_NBSL_EXT_ERR_ECC)
1859 return;
1861 memset(&err, 0, sizeof(err));
1863 sys_addr = get_error_address(m);
1865 if (ecc_type == 2)
1866 err.syndrome = extract_syndrome(m->status);
1868 pvt->ops->map_sysaddr_to_csrow(mci, sys_addr, &err);
1870 __log_bus_error(mci, &err, ecc_type);
1873 void amd64_decode_bus_error(int node_id, struct mce *m)
1875 __amd64_decode_bus_error(mcis[node_id], m);
1879 * Use pvt->F2 which contains the F2 CPU PCI device to get the related
1880 * F1 (AddrMap) and F3 (Misc) devices. Return negative value on error.
1882 static int reserve_mc_sibling_devs(struct amd64_pvt *pvt, u16 f1_id, u16 f3_id)
1884 /* Reserve the ADDRESS MAP Device */
1885 pvt->F1 = pci_get_related_function(pvt->F2->vendor, f1_id, pvt->F2);
1886 if (!pvt->F1) {
1887 amd64_err("error address map device not found: "
1888 "vendor %x device 0x%x (broken BIOS?)\n",
1889 PCI_VENDOR_ID_AMD, f1_id);
1890 return -ENODEV;
1893 /* Reserve the MISC Device */
1894 pvt->F3 = pci_get_related_function(pvt->F2->vendor, f3_id, pvt->F2);
1895 if (!pvt->F3) {
1896 pci_dev_put(pvt->F1);
1897 pvt->F1 = NULL;
1899 amd64_err("error F3 device not found: "
1900 "vendor %x device 0x%x (broken BIOS?)\n",
1901 PCI_VENDOR_ID_AMD, f3_id);
1903 return -ENODEV;
1905 edac_dbg(1, "F1: %s\n", pci_name(pvt->F1));
1906 edac_dbg(1, "F2: %s\n", pci_name(pvt->F2));
1907 edac_dbg(1, "F3: %s\n", pci_name(pvt->F3));
1909 return 0;
1912 static void free_mc_sibling_devs(struct amd64_pvt *pvt)
1914 pci_dev_put(pvt->F1);
1915 pci_dev_put(pvt->F3);
1919 * Retrieve the hardware registers of the memory controller (this includes the
1920 * 'Address Map' and 'Misc' device regs)
1922 static void read_mc_regs(struct amd64_pvt *pvt)
1924 struct cpuinfo_x86 *c = &boot_cpu_data;
1925 u64 msr_val;
1926 u32 tmp;
1927 unsigned range;
1930 * Retrieve TOP_MEM and TOP_MEM2; no masking off of reserved bits since
1931 * those are Read-As-Zero
1933 rdmsrl(MSR_K8_TOP_MEM1, pvt->top_mem);
1934 edac_dbg(0, " TOP_MEM: 0x%016llx\n", pvt->top_mem);
1936 /* check first whether TOP_MEM2 is enabled */
1937 rdmsrl(MSR_K8_SYSCFG, msr_val);
1938 if (msr_val & (1U << 21)) {
1939 rdmsrl(MSR_K8_TOP_MEM2, pvt->top_mem2);
1940 edac_dbg(0, " TOP_MEM2: 0x%016llx\n", pvt->top_mem2);
1941 } else
1942 edac_dbg(0, " TOP_MEM2 disabled\n");
1944 amd64_read_pci_cfg(pvt->F3, NBCAP, &pvt->nbcap);
1946 read_dram_ctl_register(pvt);
1948 for (range = 0; range < DRAM_RANGES; range++) {
1949 u8 rw;
1951 /* read settings for this DRAM range */
1952 read_dram_base_limit_regs(pvt, range);
1954 rw = dram_rw(pvt, range);
1955 if (!rw)
1956 continue;
1958 edac_dbg(1, " DRAM range[%d], base: 0x%016llx; limit: 0x%016llx\n",
1959 range,
1960 get_dram_base(pvt, range),
1961 get_dram_limit(pvt, range));
1963 edac_dbg(1, " IntlvEn=%s; Range access: %s%s IntlvSel=%d DstNode=%d\n",
1964 dram_intlv_en(pvt, range) ? "Enabled" : "Disabled",
1965 (rw & 0x1) ? "R" : "-",
1966 (rw & 0x2) ? "W" : "-",
1967 dram_intlv_sel(pvt, range),
1968 dram_dst_node(pvt, range));
1971 read_dct_base_mask(pvt);
1973 amd64_read_pci_cfg(pvt->F1, DHAR, &pvt->dhar);
1974 amd64_read_dct_pci_cfg(pvt, DBAM0, &pvt->dbam0);
1976 amd64_read_pci_cfg(pvt->F3, F10_ONLINE_SPARE, &pvt->online_spare);
1978 amd64_read_dct_pci_cfg(pvt, DCLR0, &pvt->dclr0);
1979 amd64_read_dct_pci_cfg(pvt, DCHR0, &pvt->dchr0);
1981 if (!dct_ganging_enabled(pvt)) {
1982 amd64_read_dct_pci_cfg(pvt, DCLR1, &pvt->dclr1);
1983 amd64_read_dct_pci_cfg(pvt, DCHR1, &pvt->dchr1);
1986 pvt->ecc_sym_sz = 4;
1988 if (c->x86 >= 0x10) {
1989 amd64_read_pci_cfg(pvt->F3, EXT_NB_MCA_CFG, &tmp);
1990 if (c->x86 != 0x16)
1991 /* F16h has only DCT0 */
1992 amd64_read_dct_pci_cfg(pvt, DBAM1, &pvt->dbam1);
1994 /* F10h, revD and later can do x8 ECC too */
1995 if ((c->x86 > 0x10 || c->x86_model > 7) && tmp & BIT(25))
1996 pvt->ecc_sym_sz = 8;
1998 dump_misc_regs(pvt);
2002 * NOTE: CPU Revision Dependent code
2004 * Input:
2005 * @csrow_nr ChipSelect Row Number (0..NUM_CHIPSELECTS-1)
2006 * k8 private pointer to -->
2007 * DRAM Bank Address mapping register
2008 * node_id
2009 * DCL register where dual_channel_active is
2011 * The DBAM register consists of 4 sets of 4 bits each definitions:
2013 * Bits: CSROWs
2014 * 0-3 CSROWs 0 and 1
2015 * 4-7 CSROWs 2 and 3
2016 * 8-11 CSROWs 4 and 5
2017 * 12-15 CSROWs 6 and 7
2019 * Values range from: 0 to 15
2020 * The meaning of the values depends on CPU revision and dual-channel state,
2021 * see relevant BKDG more info.
2023 * The memory controller provides for total of only 8 CSROWs in its current
2024 * architecture. Each "pair" of CSROWs normally represents just one DIMM in
2025 * single channel or two (2) DIMMs in dual channel mode.
2027 * The following code logic collapses the various tables for CSROW based on CPU
2028 * revision.
2030 * Returns:
2031 * The number of PAGE_SIZE pages on the specified CSROW number it
2032 * encompasses
2035 static u32 amd64_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr)
2037 u32 cs_mode, nr_pages;
2038 u32 dbam = dct ? pvt->dbam1 : pvt->dbam0;
2042 * The math on this doesn't look right on the surface because x/2*4 can
2043 * be simplified to x*2 but this expression makes use of the fact that
2044 * it is integral math where 1/2=0. This intermediate value becomes the
2045 * number of bits to shift the DBAM register to extract the proper CSROW
2046 * field.
2048 cs_mode = DBAM_DIMM(csrow_nr / 2, dbam);
2050 nr_pages = pvt->ops->dbam_to_cs(pvt, dct, cs_mode) << (20 - PAGE_SHIFT);
2052 edac_dbg(0, "csrow: %d, channel: %d, DBAM idx: %d\n",
2053 csrow_nr, dct, cs_mode);
2054 edac_dbg(0, "nr_pages/channel: %u\n", nr_pages);
2056 return nr_pages;
2060 * Initialize the array of csrow attribute instances, based on the values
2061 * from pci config hardware registers.
2063 static int init_csrows(struct mem_ctl_info *mci)
2065 struct amd64_pvt *pvt = mci->pvt_info;
2066 struct csrow_info *csrow;
2067 struct dimm_info *dimm;
2068 enum edac_type edac_mode;
2069 enum mem_type mtype;
2070 int i, j, empty = 1;
2071 int nr_pages = 0;
2072 u32 val;
2074 amd64_read_pci_cfg(pvt->F3, NBCFG, &val);
2076 pvt->nbcfg = val;
2078 edac_dbg(0, "node %d, NBCFG=0x%08x[ChipKillEccCap: %d|DramEccEn: %d]\n",
2079 pvt->mc_node_id, val,
2080 !!(val & NBCFG_CHIPKILL), !!(val & NBCFG_ECC_ENABLE));
2083 * We iterate over DCT0 here but we look at DCT1 in parallel, if needed.
2085 for_each_chip_select(i, 0, pvt) {
2086 bool row_dct0 = !!csrow_enabled(i, 0, pvt);
2087 bool row_dct1 = false;
2089 if (boot_cpu_data.x86 != 0xf)
2090 row_dct1 = !!csrow_enabled(i, 1, pvt);
2092 if (!row_dct0 && !row_dct1)
2093 continue;
2095 csrow = mci->csrows[i];
2096 empty = 0;
2098 edac_dbg(1, "MC node: %d, csrow: %d\n",
2099 pvt->mc_node_id, i);
2101 if (row_dct0) {
2102 nr_pages = amd64_csrow_nr_pages(pvt, 0, i);
2103 csrow->channels[0]->dimm->nr_pages = nr_pages;
2106 /* K8 has only one DCT */
2107 if (boot_cpu_data.x86 != 0xf && row_dct1) {
2108 int row_dct1_pages = amd64_csrow_nr_pages(pvt, 1, i);
2110 csrow->channels[1]->dimm->nr_pages = row_dct1_pages;
2111 nr_pages += row_dct1_pages;
2114 mtype = amd64_determine_memory_type(pvt, i);
2116 edac_dbg(1, "Total csrow%d pages: %u\n", i, nr_pages);
2119 * determine whether CHIPKILL or JUST ECC or NO ECC is operating
2121 if (pvt->nbcfg & NBCFG_ECC_ENABLE)
2122 edac_mode = (pvt->nbcfg & NBCFG_CHIPKILL) ?
2123 EDAC_S4ECD4ED : EDAC_SECDED;
2124 else
2125 edac_mode = EDAC_NONE;
2127 for (j = 0; j < pvt->channel_count; j++) {
2128 dimm = csrow->channels[j]->dimm;
2129 dimm->mtype = mtype;
2130 dimm->edac_mode = edac_mode;
2134 return empty;
2137 /* get all cores on this DCT */
2138 static void get_cpus_on_this_dct_cpumask(struct cpumask *mask, u16 nid)
2140 int cpu;
2142 for_each_online_cpu(cpu)
2143 if (amd_get_nb_id(cpu) == nid)
2144 cpumask_set_cpu(cpu, mask);
2147 /* check MCG_CTL on all the cpus on this node */
2148 static bool amd64_nb_mce_bank_enabled_on_node(u16 nid)
2150 cpumask_var_t mask;
2151 int cpu, nbe;
2152 bool ret = false;
2154 if (!zalloc_cpumask_var(&mask, GFP_KERNEL)) {
2155 amd64_warn("%s: Error allocating mask\n", __func__);
2156 return false;
2159 get_cpus_on_this_dct_cpumask(mask, nid);
2161 rdmsr_on_cpus(mask, MSR_IA32_MCG_CTL, msrs);
2163 for_each_cpu(cpu, mask) {
2164 struct msr *reg = per_cpu_ptr(msrs, cpu);
2165 nbe = reg->l & MSR_MCGCTL_NBE;
2167 edac_dbg(0, "core: %u, MCG_CTL: 0x%llx, NB MSR is %s\n",
2168 cpu, reg->q,
2169 (nbe ? "enabled" : "disabled"));
2171 if (!nbe)
2172 goto out;
2174 ret = true;
2176 out:
2177 free_cpumask_var(mask);
2178 return ret;
2181 static int toggle_ecc_err_reporting(struct ecc_settings *s, u16 nid, bool on)
2183 cpumask_var_t cmask;
2184 int cpu;
2186 if (!zalloc_cpumask_var(&cmask, GFP_KERNEL)) {
2187 amd64_warn("%s: error allocating mask\n", __func__);
2188 return false;
2191 get_cpus_on_this_dct_cpumask(cmask, nid);
2193 rdmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs);
2195 for_each_cpu(cpu, cmask) {
2197 struct msr *reg = per_cpu_ptr(msrs, cpu);
2199 if (on) {
2200 if (reg->l & MSR_MCGCTL_NBE)
2201 s->flags.nb_mce_enable = 1;
2203 reg->l |= MSR_MCGCTL_NBE;
2204 } else {
2206 * Turn off NB MCE reporting only when it was off before
2208 if (!s->flags.nb_mce_enable)
2209 reg->l &= ~MSR_MCGCTL_NBE;
2212 wrmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs);
2214 free_cpumask_var(cmask);
2216 return 0;
2219 static bool enable_ecc_error_reporting(struct ecc_settings *s, u16 nid,
2220 struct pci_dev *F3)
2222 bool ret = true;
2223 u32 value, mask = 0x3; /* UECC/CECC enable */
2225 if (toggle_ecc_err_reporting(s, nid, ON)) {
2226 amd64_warn("Error enabling ECC reporting over MCGCTL!\n");
2227 return false;
2230 amd64_read_pci_cfg(F3, NBCTL, &value);
2232 s->old_nbctl = value & mask;
2233 s->nbctl_valid = true;
2235 value |= mask;
2236 amd64_write_pci_cfg(F3, NBCTL, value);
2238 amd64_read_pci_cfg(F3, NBCFG, &value);
2240 edac_dbg(0, "1: node %d, NBCFG=0x%08x[DramEccEn: %d]\n",
2241 nid, value, !!(value & NBCFG_ECC_ENABLE));
2243 if (!(value & NBCFG_ECC_ENABLE)) {
2244 amd64_warn("DRAM ECC disabled on this node, enabling...\n");
2246 s->flags.nb_ecc_prev = 0;
2248 /* Attempt to turn on DRAM ECC Enable */
2249 value |= NBCFG_ECC_ENABLE;
2250 amd64_write_pci_cfg(F3, NBCFG, value);
2252 amd64_read_pci_cfg(F3, NBCFG, &value);
2254 if (!(value & NBCFG_ECC_ENABLE)) {
2255 amd64_warn("Hardware rejected DRAM ECC enable,"
2256 "check memory DIMM configuration.\n");
2257 ret = false;
2258 } else {
2259 amd64_info("Hardware accepted DRAM ECC Enable\n");
2261 } else {
2262 s->flags.nb_ecc_prev = 1;
2265 edac_dbg(0, "2: node %d, NBCFG=0x%08x[DramEccEn: %d]\n",
2266 nid, value, !!(value & NBCFG_ECC_ENABLE));
2268 return ret;
2271 static void restore_ecc_error_reporting(struct ecc_settings *s, u16 nid,
2272 struct pci_dev *F3)
2274 u32 value, mask = 0x3; /* UECC/CECC enable */
2277 if (!s->nbctl_valid)
2278 return;
2280 amd64_read_pci_cfg(F3, NBCTL, &value);
2281 value &= ~mask;
2282 value |= s->old_nbctl;
2284 amd64_write_pci_cfg(F3, NBCTL, value);
2286 /* restore previous BIOS DRAM ECC "off" setting we force-enabled */
2287 if (!s->flags.nb_ecc_prev) {
2288 amd64_read_pci_cfg(F3, NBCFG, &value);
2289 value &= ~NBCFG_ECC_ENABLE;
2290 amd64_write_pci_cfg(F3, NBCFG, value);
2293 /* restore the NB Enable MCGCTL bit */
2294 if (toggle_ecc_err_reporting(s, nid, OFF))
2295 amd64_warn("Error restoring NB MCGCTL settings!\n");
2299 * EDAC requires that the BIOS have ECC enabled before
2300 * taking over the processing of ECC errors. A command line
2301 * option allows to force-enable hardware ECC later in
2302 * enable_ecc_error_reporting().
2304 static const char *ecc_msg =
2305 "ECC disabled in the BIOS or no ECC capability, module will not load.\n"
2306 " Either enable ECC checking or force module loading by setting "
2307 "'ecc_enable_override'.\n"
2308 " (Note that use of the override may cause unknown side effects.)\n";
2310 static bool ecc_enabled(struct pci_dev *F3, u16 nid)
2312 u32 value;
2313 u8 ecc_en = 0;
2314 bool nb_mce_en = false;
2316 amd64_read_pci_cfg(F3, NBCFG, &value);
2318 ecc_en = !!(value & NBCFG_ECC_ENABLE);
2319 amd64_info("DRAM ECC %s.\n", (ecc_en ? "enabled" : "disabled"));
2321 nb_mce_en = amd64_nb_mce_bank_enabled_on_node(nid);
2322 if (!nb_mce_en)
2323 amd64_notice("NB MCE bank disabled, set MSR "
2324 "0x%08x[4] on node %d to enable.\n",
2325 MSR_IA32_MCG_CTL, nid);
2327 if (!ecc_en || !nb_mce_en) {
2328 amd64_notice("%s", ecc_msg);
2329 return false;
2331 return true;
2334 static int set_mc_sysfs_attrs(struct mem_ctl_info *mci)
2336 int rc;
2338 rc = amd64_create_sysfs_dbg_files(mci);
2339 if (rc < 0)
2340 return rc;
2342 if (boot_cpu_data.x86 >= 0x10) {
2343 rc = amd64_create_sysfs_inject_files(mci);
2344 if (rc < 0)
2345 return rc;
2348 return 0;
2351 static void del_mc_sysfs_attrs(struct mem_ctl_info *mci)
2353 amd64_remove_sysfs_dbg_files(mci);
2355 if (boot_cpu_data.x86 >= 0x10)
2356 amd64_remove_sysfs_inject_files(mci);
2359 static void setup_mci_misc_attrs(struct mem_ctl_info *mci,
2360 struct amd64_family_type *fam)
2362 struct amd64_pvt *pvt = mci->pvt_info;
2364 mci->mtype_cap = MEM_FLAG_DDR2 | MEM_FLAG_RDDR2;
2365 mci->edac_ctl_cap = EDAC_FLAG_NONE;
2367 if (pvt->nbcap & NBCAP_SECDED)
2368 mci->edac_ctl_cap |= EDAC_FLAG_SECDED;
2370 if (pvt->nbcap & NBCAP_CHIPKILL)
2371 mci->edac_ctl_cap |= EDAC_FLAG_S4ECD4ED;
2373 mci->edac_cap = amd64_determine_edac_cap(pvt);
2374 mci->mod_name = EDAC_MOD_STR;
2375 mci->mod_ver = EDAC_AMD64_VERSION;
2376 mci->ctl_name = fam->ctl_name;
2377 mci->dev_name = pci_name(pvt->F2);
2378 mci->ctl_page_to_phys = NULL;
2380 /* memory scrubber interface */
2381 mci->set_sdram_scrub_rate = amd64_set_scrub_rate;
2382 mci->get_sdram_scrub_rate = amd64_get_scrub_rate;
2386 * returns a pointer to the family descriptor on success, NULL otherwise.
2388 static struct amd64_family_type *amd64_per_family_init(struct amd64_pvt *pvt)
2390 u8 fam = boot_cpu_data.x86;
2391 struct amd64_family_type *fam_type = NULL;
2393 switch (fam) {
2394 case 0xf:
2395 fam_type = &amd64_family_types[K8_CPUS];
2396 pvt->ops = &amd64_family_types[K8_CPUS].ops;
2397 break;
2399 case 0x10:
2400 fam_type = &amd64_family_types[F10_CPUS];
2401 pvt->ops = &amd64_family_types[F10_CPUS].ops;
2402 break;
2404 case 0x15:
2405 fam_type = &amd64_family_types[F15_CPUS];
2406 pvt->ops = &amd64_family_types[F15_CPUS].ops;
2407 break;
2409 case 0x16:
2410 fam_type = &amd64_family_types[F16_CPUS];
2411 pvt->ops = &amd64_family_types[F16_CPUS].ops;
2412 break;
2414 default:
2415 amd64_err("Unsupported family!\n");
2416 return NULL;
2419 pvt->ext_model = boot_cpu_data.x86_model >> 4;
2421 amd64_info("%s %sdetected (node %d).\n", fam_type->ctl_name,
2422 (fam == 0xf ?
2423 (pvt->ext_model >= K8_REV_F ? "revF or later "
2424 : "revE or earlier ")
2425 : ""), pvt->mc_node_id);
2426 return fam_type;
2429 static int amd64_init_one_instance(struct pci_dev *F2)
2431 struct amd64_pvt *pvt = NULL;
2432 struct amd64_family_type *fam_type = NULL;
2433 struct mem_ctl_info *mci = NULL;
2434 struct edac_mc_layer layers[2];
2435 int err = 0, ret;
2436 u16 nid = amd_get_node_id(F2);
2438 ret = -ENOMEM;
2439 pvt = kzalloc(sizeof(struct amd64_pvt), GFP_KERNEL);
2440 if (!pvt)
2441 goto err_ret;
2443 pvt->mc_node_id = nid;
2444 pvt->F2 = F2;
2446 ret = -EINVAL;
2447 fam_type = amd64_per_family_init(pvt);
2448 if (!fam_type)
2449 goto err_free;
2451 ret = -ENODEV;
2452 err = reserve_mc_sibling_devs(pvt, fam_type->f1_id, fam_type->f3_id);
2453 if (err)
2454 goto err_free;
2456 read_mc_regs(pvt);
2459 * We need to determine how many memory channels there are. Then use
2460 * that information for calculating the size of the dynamic instance
2461 * tables in the 'mci' structure.
2463 ret = -EINVAL;
2464 pvt->channel_count = pvt->ops->early_channel_count(pvt);
2465 if (pvt->channel_count < 0)
2466 goto err_siblings;
2468 ret = -ENOMEM;
2469 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
2470 layers[0].size = pvt->csels[0].b_cnt;
2471 layers[0].is_virt_csrow = true;
2472 layers[1].type = EDAC_MC_LAYER_CHANNEL;
2473 layers[1].size = pvt->channel_count;
2474 layers[1].is_virt_csrow = false;
2475 mci = edac_mc_alloc(nid, ARRAY_SIZE(layers), layers, 0);
2476 if (!mci)
2477 goto err_siblings;
2479 mci->pvt_info = pvt;
2480 mci->pdev = &pvt->F2->dev;
2482 setup_mci_misc_attrs(mci, fam_type);
2484 if (init_csrows(mci))
2485 mci->edac_cap = EDAC_FLAG_NONE;
2487 ret = -ENODEV;
2488 if (edac_mc_add_mc(mci)) {
2489 edac_dbg(1, "failed edac_mc_add_mc()\n");
2490 goto err_add_mc;
2492 if (set_mc_sysfs_attrs(mci)) {
2493 edac_dbg(1, "failed edac_mc_add_mc()\n");
2494 goto err_add_sysfs;
2497 /* register stuff with EDAC MCE */
2498 if (report_gart_errors)
2499 amd_report_gart_errors(true);
2501 amd_register_ecc_decoder(amd64_decode_bus_error);
2503 mcis[nid] = mci;
2505 atomic_inc(&drv_instances);
2507 return 0;
2509 err_add_sysfs:
2510 edac_mc_del_mc(mci->pdev);
2511 err_add_mc:
2512 edac_mc_free(mci);
2514 err_siblings:
2515 free_mc_sibling_devs(pvt);
2517 err_free:
2518 kfree(pvt);
2520 err_ret:
2521 return ret;
2524 static int amd64_probe_one_instance(struct pci_dev *pdev,
2525 const struct pci_device_id *mc_type)
2527 u16 nid = amd_get_node_id(pdev);
2528 struct pci_dev *F3 = node_to_amd_nb(nid)->misc;
2529 struct ecc_settings *s;
2530 int ret = 0;
2532 ret = pci_enable_device(pdev);
2533 if (ret < 0) {
2534 edac_dbg(0, "ret=%d\n", ret);
2535 return -EIO;
2538 ret = -ENOMEM;
2539 s = kzalloc(sizeof(struct ecc_settings), GFP_KERNEL);
2540 if (!s)
2541 goto err_out;
2543 ecc_stngs[nid] = s;
2545 if (!ecc_enabled(F3, nid)) {
2546 ret = -ENODEV;
2548 if (!ecc_enable_override)
2549 goto err_enable;
2551 amd64_warn("Forcing ECC on!\n");
2553 if (!enable_ecc_error_reporting(s, nid, F3))
2554 goto err_enable;
2557 ret = amd64_init_one_instance(pdev);
2558 if (ret < 0) {
2559 amd64_err("Error probing instance: %d\n", nid);
2560 restore_ecc_error_reporting(s, nid, F3);
2563 return ret;
2565 err_enable:
2566 kfree(s);
2567 ecc_stngs[nid] = NULL;
2569 err_out:
2570 return ret;
2573 static void amd64_remove_one_instance(struct pci_dev *pdev)
2575 struct mem_ctl_info *mci;
2576 struct amd64_pvt *pvt;
2577 u16 nid = amd_get_node_id(pdev);
2578 struct pci_dev *F3 = node_to_amd_nb(nid)->misc;
2579 struct ecc_settings *s = ecc_stngs[nid];
2581 mci = find_mci_by_dev(&pdev->dev);
2582 del_mc_sysfs_attrs(mci);
2583 /* Remove from EDAC CORE tracking list */
2584 mci = edac_mc_del_mc(&pdev->dev);
2585 if (!mci)
2586 return;
2588 pvt = mci->pvt_info;
2590 restore_ecc_error_reporting(s, nid, F3);
2592 free_mc_sibling_devs(pvt);
2594 /* unregister from EDAC MCE */
2595 amd_report_gart_errors(false);
2596 amd_unregister_ecc_decoder(amd64_decode_bus_error);
2598 kfree(ecc_stngs[nid]);
2599 ecc_stngs[nid] = NULL;
2601 /* Free the EDAC CORE resources */
2602 mci->pvt_info = NULL;
2603 mcis[nid] = NULL;
2605 kfree(pvt);
2606 edac_mc_free(mci);
2610 * This table is part of the interface for loading drivers for PCI devices. The
2611 * PCI core identifies what devices are on a system during boot, and then
2612 * inquiry this table to see if this driver is for a given device found.
2614 static DEFINE_PCI_DEVICE_TABLE(amd64_pci_table) = {
2616 .vendor = PCI_VENDOR_ID_AMD,
2617 .device = PCI_DEVICE_ID_AMD_K8_NB_MEMCTL,
2618 .subvendor = PCI_ANY_ID,
2619 .subdevice = PCI_ANY_ID,
2620 .class = 0,
2621 .class_mask = 0,
2624 .vendor = PCI_VENDOR_ID_AMD,
2625 .device = PCI_DEVICE_ID_AMD_10H_NB_DRAM,
2626 .subvendor = PCI_ANY_ID,
2627 .subdevice = PCI_ANY_ID,
2628 .class = 0,
2629 .class_mask = 0,
2632 .vendor = PCI_VENDOR_ID_AMD,
2633 .device = PCI_DEVICE_ID_AMD_15H_NB_F2,
2634 .subvendor = PCI_ANY_ID,
2635 .subdevice = PCI_ANY_ID,
2636 .class = 0,
2637 .class_mask = 0,
2640 .vendor = PCI_VENDOR_ID_AMD,
2641 .device = PCI_DEVICE_ID_AMD_16H_NB_F2,
2642 .subvendor = PCI_ANY_ID,
2643 .subdevice = PCI_ANY_ID,
2644 .class = 0,
2645 .class_mask = 0,
2648 {0, }
2650 MODULE_DEVICE_TABLE(pci, amd64_pci_table);
2652 static struct pci_driver amd64_pci_driver = {
2653 .name = EDAC_MOD_STR,
2654 .probe = amd64_probe_one_instance,
2655 .remove = amd64_remove_one_instance,
2656 .id_table = amd64_pci_table,
2659 static void setup_pci_device(void)
2661 struct mem_ctl_info *mci;
2662 struct amd64_pvt *pvt;
2664 if (amd64_ctl_pci)
2665 return;
2667 mci = mcis[0];
2668 if (mci) {
2670 pvt = mci->pvt_info;
2671 amd64_ctl_pci =
2672 edac_pci_create_generic_ctl(&pvt->F2->dev, EDAC_MOD_STR);
2674 if (!amd64_ctl_pci) {
2675 pr_warning("%s(): Unable to create PCI control\n",
2676 __func__);
2678 pr_warning("%s(): PCI error report via EDAC not set\n",
2679 __func__);
2684 static int __init amd64_edac_init(void)
2686 int err = -ENODEV;
2688 printk(KERN_INFO "AMD64 EDAC driver v%s\n", EDAC_AMD64_VERSION);
2690 opstate_init();
2692 if (amd_cache_northbridges() < 0)
2693 goto err_ret;
2695 err = -ENOMEM;
2696 mcis = kzalloc(amd_nb_num() * sizeof(mcis[0]), GFP_KERNEL);
2697 ecc_stngs = kzalloc(amd_nb_num() * sizeof(ecc_stngs[0]), GFP_KERNEL);
2698 if (!(mcis && ecc_stngs))
2699 goto err_free;
2701 msrs = msrs_alloc();
2702 if (!msrs)
2703 goto err_free;
2705 err = pci_register_driver(&amd64_pci_driver);
2706 if (err)
2707 goto err_pci;
2709 err = -ENODEV;
2710 if (!atomic_read(&drv_instances))
2711 goto err_no_instances;
2713 setup_pci_device();
2714 return 0;
2716 err_no_instances:
2717 pci_unregister_driver(&amd64_pci_driver);
2719 err_pci:
2720 msrs_free(msrs);
2721 msrs = NULL;
2723 err_free:
2724 kfree(mcis);
2725 mcis = NULL;
2727 kfree(ecc_stngs);
2728 ecc_stngs = NULL;
2730 err_ret:
2731 return err;
2734 static void __exit amd64_edac_exit(void)
2736 if (amd64_ctl_pci)
2737 edac_pci_release_generic_ctl(amd64_ctl_pci);
2739 pci_unregister_driver(&amd64_pci_driver);
2741 kfree(ecc_stngs);
2742 ecc_stngs = NULL;
2744 kfree(mcis);
2745 mcis = NULL;
2747 msrs_free(msrs);
2748 msrs = NULL;
2751 module_init(amd64_edac_init);
2752 module_exit(amd64_edac_exit);
2754 MODULE_LICENSE("GPL");
2755 MODULE_AUTHOR("SoftwareBitMaker: Doug Thompson, "
2756 "Dave Peterson, Thayne Harbaugh");
2757 MODULE_DESCRIPTION("MC support for AMD64 memory controllers - "
2758 EDAC_AMD64_VERSION);
2760 module_param(edac_op_state, int, 0444);
2761 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");