2 * edac_mc kernel module
3 * (C) 2005, 2006 Linux Networx (http://lnxi.com)
4 * This file may be distributed under the terms of the
5 * GNU General Public License.
7 * Written by Thayne Harbaugh
8 * Based on work by Dan Hollis <goemon at anime dot net> and others.
9 * http://www.anime.net/~goemon/linux-ecc/
11 * Modified by Dave Peterson and Doug Thompson
15 #include <linux/module.h>
16 #include <linux/proc_fs.h>
17 #include <linux/kernel.h>
18 #include <linux/types.h>
19 #include <linux/smp.h>
20 #include <linux/init.h>
21 #include <linux/sysctl.h>
22 #include <linux/highmem.h>
23 #include <linux/timer.h>
24 #include <linux/slab.h>
25 #include <linux/jiffies.h>
26 #include <linux/spinlock.h>
27 #include <linux/list.h>
28 #include <linux/sysdev.h>
29 #include <linux/ctype.h>
30 #include <linux/edac.h>
31 #include <asm/uaccess.h>
34 #include "edac_core.h"
35 #include "edac_module.h"
37 /* lock to memory controller's control array */
38 static DEFINE_MUTEX(mem_ctls_mutex
);
39 static LIST_HEAD(mc_devices
);
41 #ifdef CONFIG_EDAC_DEBUG
43 static void edac_mc_dump_channel(struct channel_info
*chan
)
45 debugf4("\tchannel = %p\n", chan
);
46 debugf4("\tchannel->chan_idx = %d\n", chan
->chan_idx
);
47 debugf4("\tchannel->ce_count = %d\n", chan
->ce_count
);
48 debugf4("\tchannel->label = '%s'\n", chan
->label
);
49 debugf4("\tchannel->csrow = %p\n\n", chan
->csrow
);
52 static void edac_mc_dump_csrow(struct csrow_info
*csrow
)
54 debugf4("\tcsrow = %p\n", csrow
);
55 debugf4("\tcsrow->csrow_idx = %d\n", csrow
->csrow_idx
);
56 debugf4("\tcsrow->first_page = 0x%lx\n", csrow
->first_page
);
57 debugf4("\tcsrow->last_page = 0x%lx\n", csrow
->last_page
);
58 debugf4("\tcsrow->page_mask = 0x%lx\n", csrow
->page_mask
);
59 debugf4("\tcsrow->nr_pages = 0x%x\n", csrow
->nr_pages
);
60 debugf4("\tcsrow->nr_channels = %d\n", csrow
->nr_channels
);
61 debugf4("\tcsrow->channels = %p\n", csrow
->channels
);
62 debugf4("\tcsrow->mci = %p\n\n", csrow
->mci
);
65 static void edac_mc_dump_mci(struct mem_ctl_info
*mci
)
67 debugf3("\tmci = %p\n", mci
);
68 debugf3("\tmci->mtype_cap = %lx\n", mci
->mtype_cap
);
69 debugf3("\tmci->edac_ctl_cap = %lx\n", mci
->edac_ctl_cap
);
70 debugf3("\tmci->edac_cap = %lx\n", mci
->edac_cap
);
71 debugf4("\tmci->edac_check = %p\n", mci
->edac_check
);
72 debugf3("\tmci->nr_csrows = %d, csrows = %p\n",
73 mci
->nr_csrows
, mci
->csrows
);
74 debugf3("\tdev = %p\n", mci
->dev
);
75 debugf3("\tmod_name:ctl_name = %s:%s\n", mci
->mod_name
, mci
->ctl_name
);
76 debugf3("\tpvt_info = %p\n\n", mci
->pvt_info
);
79 #endif /* CONFIG_EDAC_DEBUG */
82 * keep those in sync with the enum mem_type
84 const char *edac_mem_types
[] = {
86 "Reserved csrow type",
89 "Extended data out RAM",
90 "Burst Extended data out RAM",
91 "Single data rate SDRAM",
92 "Registered single data rate SDRAM",
93 "Double data rate SDRAM",
94 "Registered Double data rate SDRAM",
96 "Unbuffered DDR2 RAM",
97 "Fully buffered DDR2",
98 "Registered DDR2 RAM",
100 "Unbuffered DDR3 RAM",
101 "Registered DDR3 RAM",
103 EXPORT_SYMBOL_GPL(edac_mem_types
);
105 /* 'ptr' points to a possibly unaligned item X such that sizeof(X) is 'size'.
106 * Adjust 'ptr' so that its alignment is at least as stringent as what the
107 * compiler would provide for X and return the aligned result.
109 * If 'size' is a constant, the compiler will optimize this whole function
110 * down to either a no-op or the addition of a constant to the value of 'ptr'.
112 void *edac_align_ptr(void *ptr
, unsigned size
)
116 /* Here we assume that the alignment of a "long long" is the most
117 * stringent alignment that the compiler will ever provide by default.
118 * As far as I know, this is a reasonable assumption.
120 if (size
> sizeof(long))
121 align
= sizeof(long long);
122 else if (size
> sizeof(int))
123 align
= sizeof(long);
124 else if (size
> sizeof(short))
126 else if (size
> sizeof(char))
127 align
= sizeof(short);
136 return (void *)(((unsigned long)ptr
) + align
- r
);
140 * edac_mc_alloc: Allocate a struct mem_ctl_info structure
141 * @size_pvt: size of private storage needed
142 * @nr_csrows: Number of CWROWS needed for this MC
143 * @nr_chans: Number of channels for the MC
145 * Everything is kmalloc'ed as one big chunk - more efficient.
146 * Only can be used if all structures have the same lifetime - otherwise
147 * you have to allocate and initialize your own structures.
149 * Use edac_mc_free() to free mc structures allocated by this function.
152 * NULL allocation failed
153 * struct mem_ctl_info pointer
155 struct mem_ctl_info
*edac_mc_alloc(unsigned sz_pvt
, unsigned nr_csrows
,
156 unsigned nr_chans
, int edac_index
)
158 struct mem_ctl_info
*mci
;
159 struct csrow_info
*csi
, *csrow
;
160 struct channel_info
*chi
, *chp
, *chan
;
166 /* Figure out the offsets of the various items from the start of an mc
167 * structure. We want the alignment of each item to be at least as
168 * stringent as what the compiler would provide if we could simply
169 * hardcode everything into a single struct.
171 mci
= (struct mem_ctl_info
*)0;
172 csi
= edac_align_ptr(&mci
[1], sizeof(*csi
));
173 chi
= edac_align_ptr(&csi
[nr_csrows
], sizeof(*chi
));
174 pvt
= edac_align_ptr(&chi
[nr_chans
* nr_csrows
], sz_pvt
);
175 size
= ((unsigned long)pvt
) + sz_pvt
;
177 mci
= kzalloc(size
, GFP_KERNEL
);
181 /* Adjust pointers so they point within the memory we just allocated
182 * rather than an imaginary chunk of memory located at address 0.
184 csi
= (struct csrow_info
*)(((char *)mci
) + ((unsigned long)csi
));
185 chi
= (struct channel_info
*)(((char *)mci
) + ((unsigned long)chi
));
186 pvt
= sz_pvt
? (((char *)mci
) + ((unsigned long)pvt
)) : NULL
;
188 /* setup index and various internal pointers */
189 mci
->mc_idx
= edac_index
;
192 mci
->nr_csrows
= nr_csrows
;
194 for (row
= 0; row
< nr_csrows
; row
++) {
196 csrow
->csrow_idx
= row
;
198 csrow
->nr_channels
= nr_chans
;
199 chp
= &chi
[row
* nr_chans
];
200 csrow
->channels
= chp
;
202 for (chn
= 0; chn
< nr_chans
; chn
++) {
204 chan
->chan_idx
= chn
;
209 mci
->op_state
= OP_ALLOC
;
210 INIT_LIST_HEAD(&mci
->grp_kobj_list
);
213 * Initialize the 'root' kobj for the edac_mc controller
215 err
= edac_mc_register_sysfs_main_kobj(mci
);
221 /* at this point, the root kobj is valid, and in order to
222 * 'free' the object, then the function:
223 * edac_mc_unregister_sysfs_main_kobj() must be called
224 * which will perform kobj unregistration and the actual free
225 * will occur during the kobject callback operation
229 EXPORT_SYMBOL_GPL(edac_mc_alloc
);
233 * 'Free' a previously allocated 'mci' structure
234 * @mci: pointer to a struct mem_ctl_info structure
236 void edac_mc_free(struct mem_ctl_info
*mci
)
238 debugf1("%s()\n", __func__
);
240 edac_mc_unregister_sysfs_main_kobj(mci
);
242 /* free the mci instance memory here */
245 EXPORT_SYMBOL_GPL(edac_mc_free
);
251 * scan list of controllers looking for the one that manages
253 * @dev: pointer to a struct device related with the MCI
255 struct mem_ctl_info
*find_mci_by_dev(struct device
*dev
)
257 struct mem_ctl_info
*mci
;
258 struct list_head
*item
;
260 debugf3("%s()\n", __func__
);
262 list_for_each(item
, &mc_devices
) {
263 mci
= list_entry(item
, struct mem_ctl_info
, link
);
271 EXPORT_SYMBOL_GPL(find_mci_by_dev
);
274 * handler for EDAC to check if NMI type handler has asserted interrupt
276 static int edac_mc_assert_error_check_and_clear(void)
280 if (edac_op_state
== EDAC_OPSTATE_POLL
)
283 old_state
= edac_err_assert
;
290 * edac_mc_workq_function
291 * performs the operation scheduled by a workq request
293 static void edac_mc_workq_function(struct work_struct
*work_req
)
295 struct delayed_work
*d_work
= to_delayed_work(work_req
);
296 struct mem_ctl_info
*mci
= to_edac_mem_ctl_work(d_work
);
298 mutex_lock(&mem_ctls_mutex
);
300 /* if this control struct has movd to offline state, we are done */
301 if (mci
->op_state
== OP_OFFLINE
) {
302 mutex_unlock(&mem_ctls_mutex
);
306 /* Only poll controllers that are running polled and have a check */
307 if (edac_mc_assert_error_check_and_clear() && (mci
->edac_check
!= NULL
))
308 mci
->edac_check(mci
);
310 mutex_unlock(&mem_ctls_mutex
);
313 queue_delayed_work(edac_workqueue
, &mci
->work
,
314 msecs_to_jiffies(edac_mc_get_poll_msec()));
318 * edac_mc_workq_setup
319 * initialize a workq item for this mci
320 * passing in the new delay period in msec
324 * called with the mem_ctls_mutex held
326 static void edac_mc_workq_setup(struct mem_ctl_info
*mci
, unsigned msec
)
328 debugf0("%s()\n", __func__
);
330 /* if this instance is not in the POLL state, then simply return */
331 if (mci
->op_state
!= OP_RUNNING_POLL
)
334 INIT_DELAYED_WORK(&mci
->work
, edac_mc_workq_function
);
335 queue_delayed_work(edac_workqueue
, &mci
->work
, msecs_to_jiffies(msec
));
339 * edac_mc_workq_teardown
340 * stop the workq processing on this mci
344 * called WITHOUT lock held
346 static void edac_mc_workq_teardown(struct mem_ctl_info
*mci
)
350 if (mci
->op_state
!= OP_RUNNING_POLL
)
353 status
= cancel_delayed_work(&mci
->work
);
355 debugf0("%s() not canceled, flush the queue\n",
358 /* workq instance might be running, wait for it */
359 flush_workqueue(edac_workqueue
);
364 * edac_mc_reset_delay_period(unsigned long value)
366 * user space has updated our poll period value, need to
367 * reset our workq delays
369 void edac_mc_reset_delay_period(int value
)
371 struct mem_ctl_info
*mci
;
372 struct list_head
*item
;
374 mutex_lock(&mem_ctls_mutex
);
376 /* scan the list and turn off all workq timers, doing so under lock
378 list_for_each(item
, &mc_devices
) {
379 mci
= list_entry(item
, struct mem_ctl_info
, link
);
381 if (mci
->op_state
== OP_RUNNING_POLL
)
382 cancel_delayed_work(&mci
->work
);
385 mutex_unlock(&mem_ctls_mutex
);
388 /* re-walk the list, and reset the poll delay */
389 mutex_lock(&mem_ctls_mutex
);
391 list_for_each(item
, &mc_devices
) {
392 mci
= list_entry(item
, struct mem_ctl_info
, link
);
394 edac_mc_workq_setup(mci
, (unsigned long) value
);
397 mutex_unlock(&mem_ctls_mutex
);
402 /* Return 0 on success, 1 on failure.
403 * Before calling this function, caller must
404 * assign a unique value to mci->mc_idx.
408 * called with the mem_ctls_mutex lock held
410 static int add_mc_to_global_list(struct mem_ctl_info
*mci
)
412 struct list_head
*item
, *insert_before
;
413 struct mem_ctl_info
*p
;
415 insert_before
= &mc_devices
;
417 p
= find_mci_by_dev(mci
->dev
);
418 if (unlikely(p
!= NULL
))
421 list_for_each(item
, &mc_devices
) {
422 p
= list_entry(item
, struct mem_ctl_info
, link
);
424 if (p
->mc_idx
>= mci
->mc_idx
) {
425 if (unlikely(p
->mc_idx
== mci
->mc_idx
))
428 insert_before
= item
;
433 list_add_tail_rcu(&mci
->link
, insert_before
);
434 atomic_inc(&edac_handlers
);
438 edac_printk(KERN_WARNING
, EDAC_MC
,
439 "%s (%s) %s %s already assigned %d\n", dev_name(p
->dev
),
440 edac_dev_name(mci
), p
->mod_name
, p
->ctl_name
, p
->mc_idx
);
444 edac_printk(KERN_WARNING
, EDAC_MC
,
445 "bug in low-level driver: attempt to assign\n"
446 " duplicate mc_idx %d in %s()\n", p
->mc_idx
, __func__
);
450 static void del_mc_from_global_list(struct mem_ctl_info
*mci
)
452 atomic_dec(&edac_handlers
);
453 list_del_rcu(&mci
->link
);
455 /* these are for safe removal of devices from global list while
456 * NMI handlers may be traversing list
459 INIT_LIST_HEAD(&mci
->link
);
463 * edac_mc_find: Search for a mem_ctl_info structure whose index is 'idx'.
465 * If found, return a pointer to the structure.
468 * Caller must hold mem_ctls_mutex.
470 struct mem_ctl_info
*edac_mc_find(int idx
)
472 struct list_head
*item
;
473 struct mem_ctl_info
*mci
;
475 list_for_each(item
, &mc_devices
) {
476 mci
= list_entry(item
, struct mem_ctl_info
, link
);
478 if (mci
->mc_idx
>= idx
) {
479 if (mci
->mc_idx
== idx
)
488 EXPORT_SYMBOL(edac_mc_find
);
491 * edac_mc_add_mc: Insert the 'mci' structure into the mci global list and
492 * create sysfs entries associated with mci structure
493 * @mci: pointer to the mci structure to be added to the list
494 * @mc_idx: A unique numeric identifier to be assigned to the 'mci' structure.
501 /* FIXME - should a warning be printed if no error detection? correction? */
502 int edac_mc_add_mc(struct mem_ctl_info
*mci
)
504 debugf0("%s()\n", __func__
);
506 #ifdef CONFIG_EDAC_DEBUG
507 if (edac_debug_level
>= 3)
508 edac_mc_dump_mci(mci
);
510 if (edac_debug_level
>= 4) {
513 for (i
= 0; i
< mci
->nr_csrows
; i
++) {
516 edac_mc_dump_csrow(&mci
->csrows
[i
]);
517 for (j
= 0; j
< mci
->csrows
[i
].nr_channels
; j
++)
518 edac_mc_dump_channel(&mci
->csrows
[i
].
523 mutex_lock(&mem_ctls_mutex
);
525 if (add_mc_to_global_list(mci
))
528 /* set load time so that error rate can be tracked */
529 mci
->start_time
= jiffies
;
531 if (edac_create_sysfs_mci_device(mci
)) {
532 edac_mc_printk(mci
, KERN_WARNING
,
533 "failed to create sysfs device\n");
537 /* If there IS a check routine, then we are running POLLED */
538 if (mci
->edac_check
!= NULL
) {
539 /* This instance is NOW RUNNING */
540 mci
->op_state
= OP_RUNNING_POLL
;
542 edac_mc_workq_setup(mci
, edac_mc_get_poll_msec());
544 mci
->op_state
= OP_RUNNING_INTERRUPT
;
547 /* Report action taken */
548 edac_mc_printk(mci
, KERN_INFO
, "Giving out device to '%s' '%s':"
549 " DEV %s\n", mci
->mod_name
, mci
->ctl_name
, edac_dev_name(mci
));
551 mutex_unlock(&mem_ctls_mutex
);
555 del_mc_from_global_list(mci
);
558 mutex_unlock(&mem_ctls_mutex
);
561 EXPORT_SYMBOL_GPL(edac_mc_add_mc
);
564 * edac_mc_del_mc: Remove sysfs entries for specified mci structure and
565 * remove mci structure from global list
566 * @pdev: Pointer to 'struct device' representing mci structure to remove.
568 * Return pointer to removed mci structure, or NULL if device not found.
570 struct mem_ctl_info
*edac_mc_del_mc(struct device
*dev
)
572 struct mem_ctl_info
*mci
;
574 debugf0("%s()\n", __func__
);
576 mutex_lock(&mem_ctls_mutex
);
578 /* find the requested mci struct in the global list */
579 mci
= find_mci_by_dev(dev
);
581 mutex_unlock(&mem_ctls_mutex
);
585 del_mc_from_global_list(mci
);
586 mutex_unlock(&mem_ctls_mutex
);
588 /* flush workq processes */
589 edac_mc_workq_teardown(mci
);
591 /* marking MCI offline */
592 mci
->op_state
= OP_OFFLINE
;
594 /* remove from sysfs */
595 edac_remove_sysfs_mci_device(mci
);
597 edac_printk(KERN_INFO
, EDAC_MC
,
598 "Removed device %d for %s %s: DEV %s\n", mci
->mc_idx
,
599 mci
->mod_name
, mci
->ctl_name
, edac_dev_name(mci
));
603 EXPORT_SYMBOL_GPL(edac_mc_del_mc
);
605 static void edac_mc_scrub_block(unsigned long page
, unsigned long offset
,
610 unsigned long flags
= 0;
612 debugf3("%s()\n", __func__
);
614 /* ECC error page was not in our memory. Ignore it. */
615 if (!pfn_valid(page
))
618 /* Find the actual page structure then map it and fix */
619 pg
= pfn_to_page(page
);
622 local_irq_save(flags
);
624 virt_addr
= kmap_atomic(pg
, KM_BOUNCE_READ
);
626 /* Perform architecture specific atomic scrub operation */
627 atomic_scrub(virt_addr
+ offset
, size
);
629 /* Unmap and complete */
630 kunmap_atomic(virt_addr
, KM_BOUNCE_READ
);
633 local_irq_restore(flags
);
636 /* FIXME - should return -1 */
637 int edac_mc_find_csrow_by_page(struct mem_ctl_info
*mci
, unsigned long page
)
639 struct csrow_info
*csrows
= mci
->csrows
;
642 debugf1("MC%d: %s(): 0x%lx\n", mci
->mc_idx
, __func__
, page
);
645 for (i
= 0; i
< mci
->nr_csrows
; i
++) {
646 struct csrow_info
*csrow
= &csrows
[i
];
648 if (csrow
->nr_pages
== 0)
651 debugf3("MC%d: %s(): first(0x%lx) page(0x%lx) last(0x%lx) "
652 "mask(0x%lx)\n", mci
->mc_idx
, __func__
,
653 csrow
->first_page
, page
, csrow
->last_page
,
656 if ((page
>= csrow
->first_page
) &&
657 (page
<= csrow
->last_page
) &&
658 ((page
& csrow
->page_mask
) ==
659 (csrow
->first_page
& csrow
->page_mask
))) {
666 edac_mc_printk(mci
, KERN_ERR
,
667 "could not look up page error address %lx\n",
668 (unsigned long)page
);
672 EXPORT_SYMBOL_GPL(edac_mc_find_csrow_by_page
);
674 /* FIXME - setable log (warning/emerg) levels */
675 /* FIXME - integrate with evlog: http://evlog.sourceforge.net/ */
676 void edac_mc_handle_ce(struct mem_ctl_info
*mci
,
677 unsigned long page_frame_number
,
678 unsigned long offset_in_page
, unsigned long syndrome
,
679 int row
, int channel
, const char *msg
)
681 unsigned long remapped_page
;
683 debugf3("MC%d: %s()\n", mci
->mc_idx
, __func__
);
685 /* FIXME - maybe make panic on INTERNAL ERROR an option */
686 if (row
>= mci
->nr_csrows
|| row
< 0) {
687 /* something is wrong */
688 edac_mc_printk(mci
, KERN_ERR
,
689 "INTERNAL ERROR: row out of range "
690 "(%d >= %d)\n", row
, mci
->nr_csrows
);
691 edac_mc_handle_ce_no_info(mci
, "INTERNAL ERROR");
695 if (channel
>= mci
->csrows
[row
].nr_channels
|| channel
< 0) {
696 /* something is wrong */
697 edac_mc_printk(mci
, KERN_ERR
,
698 "INTERNAL ERROR: channel out of range "
699 "(%d >= %d)\n", channel
,
700 mci
->csrows
[row
].nr_channels
);
701 edac_mc_handle_ce_no_info(mci
, "INTERNAL ERROR");
705 if (edac_mc_get_log_ce())
706 /* FIXME - put in DIMM location */
707 edac_mc_printk(mci
, KERN_WARNING
,
708 "CE page 0x%lx, offset 0x%lx, grain %d, syndrome "
709 "0x%lx, row %d, channel %d, label \"%s\": %s\n",
710 page_frame_number
, offset_in_page
,
711 mci
->csrows
[row
].grain
, syndrome
, row
, channel
,
712 mci
->csrows
[row
].channels
[channel
].label
, msg
);
715 mci
->csrows
[row
].ce_count
++;
716 mci
->csrows
[row
].channels
[channel
].ce_count
++;
718 if (mci
->scrub_mode
& SCRUB_SW_SRC
) {
720 * Some MC's can remap memory so that it is still available
721 * at a different address when PCI devices map into memory.
722 * MC's that can't do this lose the memory where PCI devices
723 * are mapped. This mapping is MC dependent and so we call
724 * back into the MC driver for it to map the MC page to
725 * a physical (CPU) page which can then be mapped to a virtual
726 * page - which can then be scrubbed.
728 remapped_page
= mci
->ctl_page_to_phys
?
729 mci
->ctl_page_to_phys(mci
, page_frame_number
) :
732 edac_mc_scrub_block(remapped_page
, offset_in_page
,
733 mci
->csrows
[row
].grain
);
736 EXPORT_SYMBOL_GPL(edac_mc_handle_ce
);
738 void edac_mc_handle_ce_no_info(struct mem_ctl_info
*mci
, const char *msg
)
740 if (edac_mc_get_log_ce())
741 edac_mc_printk(mci
, KERN_WARNING
,
742 "CE - no information available: %s\n", msg
);
744 mci
->ce_noinfo_count
++;
747 EXPORT_SYMBOL_GPL(edac_mc_handle_ce_no_info
);
749 void edac_mc_handle_ue(struct mem_ctl_info
*mci
,
750 unsigned long page_frame_number
,
751 unsigned long offset_in_page
, int row
, const char *msg
)
753 int len
= EDAC_MC_LABEL_LEN
* 4;
754 char labels
[len
+ 1];
759 debugf3("MC%d: %s()\n", mci
->mc_idx
, __func__
);
761 /* FIXME - maybe make panic on INTERNAL ERROR an option */
762 if (row
>= mci
->nr_csrows
|| row
< 0) {
763 /* something is wrong */
764 edac_mc_printk(mci
, KERN_ERR
,
765 "INTERNAL ERROR: row out of range "
766 "(%d >= %d)\n", row
, mci
->nr_csrows
);
767 edac_mc_handle_ue_no_info(mci
, "INTERNAL ERROR");
771 chars
= snprintf(pos
, len
+ 1, "%s",
772 mci
->csrows
[row
].channels
[0].label
);
776 for (chan
= 1; (chan
< mci
->csrows
[row
].nr_channels
) && (len
> 0);
778 chars
= snprintf(pos
, len
+ 1, ":%s",
779 mci
->csrows
[row
].channels
[chan
].label
);
784 if (edac_mc_get_log_ue())
785 edac_mc_printk(mci
, KERN_EMERG
,
786 "UE page 0x%lx, offset 0x%lx, grain %d, row %d, "
787 "labels \"%s\": %s\n", page_frame_number
,
788 offset_in_page
, mci
->csrows
[row
].grain
, row
,
791 if (edac_mc_get_panic_on_ue())
792 panic("EDAC MC%d: UE page 0x%lx, offset 0x%lx, grain %d, "
793 "row %d, labels \"%s\": %s\n", mci
->mc_idx
,
794 page_frame_number
, offset_in_page
,
795 mci
->csrows
[row
].grain
, row
, labels
, msg
);
798 mci
->csrows
[row
].ue_count
++;
800 EXPORT_SYMBOL_GPL(edac_mc_handle_ue
);
802 void edac_mc_handle_ue_no_info(struct mem_ctl_info
*mci
, const char *msg
)
804 if (edac_mc_get_panic_on_ue())
805 panic("EDAC MC%d: Uncorrected Error", mci
->mc_idx
);
807 if (edac_mc_get_log_ue())
808 edac_mc_printk(mci
, KERN_WARNING
,
809 "UE - no information available: %s\n", msg
);
810 mci
->ue_noinfo_count
++;
813 EXPORT_SYMBOL_GPL(edac_mc_handle_ue_no_info
);
815 /*************************************************************
816 * On Fully Buffered DIMM modules, this help function is
817 * called to process UE events
819 void edac_mc_handle_fbd_ue(struct mem_ctl_info
*mci
,
821 unsigned int channela
,
822 unsigned int channelb
, char *msg
)
824 int len
= EDAC_MC_LABEL_LEN
* 4;
825 char labels
[len
+ 1];
829 if (csrow
>= mci
->nr_csrows
) {
830 /* something is wrong */
831 edac_mc_printk(mci
, KERN_ERR
,
832 "INTERNAL ERROR: row out of range (%d >= %d)\n",
833 csrow
, mci
->nr_csrows
);
834 edac_mc_handle_ue_no_info(mci
, "INTERNAL ERROR");
838 if (channela
>= mci
->csrows
[csrow
].nr_channels
) {
839 /* something is wrong */
840 edac_mc_printk(mci
, KERN_ERR
,
841 "INTERNAL ERROR: channel-a out of range "
843 channela
, mci
->csrows
[csrow
].nr_channels
);
844 edac_mc_handle_ue_no_info(mci
, "INTERNAL ERROR");
848 if (channelb
>= mci
->csrows
[csrow
].nr_channels
) {
849 /* something is wrong */
850 edac_mc_printk(mci
, KERN_ERR
,
851 "INTERNAL ERROR: channel-b out of range "
853 channelb
, mci
->csrows
[csrow
].nr_channels
);
854 edac_mc_handle_ue_no_info(mci
, "INTERNAL ERROR");
859 mci
->csrows
[csrow
].ue_count
++;
861 /* Generate the DIMM labels from the specified channels */
862 chars
= snprintf(pos
, len
+ 1, "%s",
863 mci
->csrows
[csrow
].channels
[channela
].label
);
866 chars
= snprintf(pos
, len
+ 1, "-%s",
867 mci
->csrows
[csrow
].channels
[channelb
].label
);
869 if (edac_mc_get_log_ue())
870 edac_mc_printk(mci
, KERN_EMERG
,
871 "UE row %d, channel-a= %d channel-b= %d "
872 "labels \"%s\": %s\n", csrow
, channela
, channelb
,
875 if (edac_mc_get_panic_on_ue())
876 panic("UE row %d, channel-a= %d channel-b= %d "
877 "labels \"%s\": %s\n", csrow
, channela
,
878 channelb
, labels
, msg
);
880 EXPORT_SYMBOL(edac_mc_handle_fbd_ue
);
882 /*************************************************************
883 * On Fully Buffered DIMM modules, this help function is
884 * called to process CE events
886 void edac_mc_handle_fbd_ce(struct mem_ctl_info
*mci
,
887 unsigned int csrow
, unsigned int channel
, char *msg
)
890 /* Ensure boundary values */
891 if (csrow
>= mci
->nr_csrows
) {
892 /* something is wrong */
893 edac_mc_printk(mci
, KERN_ERR
,
894 "INTERNAL ERROR: row out of range (%d >= %d)\n",
895 csrow
, mci
->nr_csrows
);
896 edac_mc_handle_ce_no_info(mci
, "INTERNAL ERROR");
899 if (channel
>= mci
->csrows
[csrow
].nr_channels
) {
900 /* something is wrong */
901 edac_mc_printk(mci
, KERN_ERR
,
902 "INTERNAL ERROR: channel out of range (%d >= %d)\n",
903 channel
, mci
->csrows
[csrow
].nr_channels
);
904 edac_mc_handle_ce_no_info(mci
, "INTERNAL ERROR");
908 if (edac_mc_get_log_ce())
909 /* FIXME - put in DIMM location */
910 edac_mc_printk(mci
, KERN_WARNING
,
911 "CE row %d, channel %d, label \"%s\": %s\n",
913 mci
->csrows
[csrow
].channels
[channel
].label
, msg
);
916 mci
->csrows
[csrow
].ce_count
++;
917 mci
->csrows
[csrow
].channels
[channel
].ce_count
++;
919 EXPORT_SYMBOL(edac_mc_handle_fbd_ce
);