2 * edac_mc kernel module
3 * (C) 2005, 2006 Linux Networx (http://lnxi.com)
4 * This file may be distributed under the terms of the
5 * GNU General Public License.
7 * Written by Thayne Harbaugh
8 * Based on work by Dan Hollis <goemon at anime dot net> and others.
9 * http://www.anime.net/~goemon/linux-ecc/
11 * Modified by Dave Peterson and Doug Thompson
15 #include <linux/module.h>
16 #include <linux/proc_fs.h>
17 #include <linux/kernel.h>
18 #include <linux/types.h>
19 #include <linux/smp.h>
20 #include <linux/init.h>
21 #include <linux/sysctl.h>
22 #include <linux/highmem.h>
23 #include <linux/timer.h>
24 #include <linux/slab.h>
25 #include <linux/jiffies.h>
26 #include <linux/spinlock.h>
27 #include <linux/list.h>
28 #include <linux/ctype.h>
29 #include <linux/edac.h>
30 #include <asm/uaccess.h>
33 #include "edac_core.h"
34 #include "edac_module.h"
36 /* lock to memory controller's control array */
37 static DEFINE_MUTEX(mem_ctls_mutex);
38 static LIST_HEAD(mc_devices);
40 #ifdef CONFIG_EDAC_DEBUG
42 static void edac_mc_dump_channel(struct channel_info *chan)
44 debugf4("\tchannel = %p\n", chan);
45 debugf4("\tchannel->chan_idx = %d\n", chan->chan_idx);
46 debugf4("\tchannel->ce_count = %d\n", chan->ce_count);
47 debugf4("\tchannel->label = '%s'\n", chan->label);
48 debugf4("\tchannel->csrow = %p\n\n", chan->csrow);
51 static void edac_mc_dump_csrow(struct csrow_info *csrow)
53 debugf4("\tcsrow = %p\n", csrow);
54 debugf4("\tcsrow->csrow_idx = %d\n", csrow->csrow_idx);
55 debugf4("\tcsrow->first_page = 0x%lx\n", csrow->first_page);
56 debugf4("\tcsrow->last_page = 0x%lx\n", csrow->last_page);
57 debugf4("\tcsrow->page_mask = 0x%lx\n", csrow->page_mask);
58 debugf4("\tcsrow->nr_pages = 0x%x\n", csrow->nr_pages);
59 debugf4("\tcsrow->nr_channels = %d\n", csrow->nr_channels);
60 debugf4("\tcsrow->channels = %p\n", csrow->channels);
61 debugf4("\tcsrow->mci = %p\n\n", csrow->mci);
64 static void edac_mc_dump_mci(struct mem_ctl_info *mci)
66 debugf3("\tmci = %p\n", mci);
67 debugf3("\tmci->mtype_cap = %lx\n", mci->mtype_cap);
68 debugf3("\tmci->edac_ctl_cap = %lx\n", mci->edac_ctl_cap);
69 debugf3("\tmci->edac_cap = %lx\n", mci->edac_cap);
70 debugf4("\tmci->edac_check = %p\n", mci->edac_check);
71 debugf3("\tmci->nr_csrows = %d, csrows = %p\n",
72 mci->nr_csrows, mci->csrows);
73 debugf3("\tdev = %p\n", mci->dev);
74 debugf3("\tmod_name:ctl_name = %s:%s\n", mci->mod_name, mci->ctl_name);
75 debugf3("\tpvt_info = %p\n\n", mci->pvt_info);
78 #endif /* CONFIG_EDAC_DEBUG */
81 * keep those in sync with the enum mem_type
83 const char *edac_mem_types[] = {
85 "Reserved csrow type",
88 "Extended data out RAM",
89 "Burst Extended data out RAM",
90 "Single data rate SDRAM",
91 "Registered single data rate SDRAM",
92 "Double data rate SDRAM",
93 "Registered Double data rate SDRAM",
95 "Unbuffered DDR2 RAM",
96 "Fully buffered DDR2",
97 "Registered DDR2 RAM",
99 "Unbuffered DDR3 RAM",
100 "Registered DDR3 RAM",
102 EXPORT_SYMBOL_GPL(edac_mem_types);
104 /* 'ptr' points to a possibly unaligned item X such that sizeof(X) is 'size'.
105 * Adjust 'ptr' so that its alignment is at least as stringent as what the
106 * compiler would provide for X and return the aligned result.
108 * If 'size' is a constant, the compiler will optimize this whole function
109 * down to either a no-op or the addition of a constant to the value of 'ptr'.
111 void *edac_align_ptr(void *ptr, unsigned size)
115 /* Here we assume that the alignment of a "long long" is the most
116 * stringent alignment that the compiler will ever provide by default.
117 * As far as I know, this is a reasonable assumption.
119 if (size > sizeof(long))
120 align = sizeof(long long);
121 else if (size > sizeof(int))
122 align = sizeof(long);
123 else if (size > sizeof(short))
125 else if (size > sizeof(char))
126 align = sizeof(short);
135 return (void *)(((unsigned long)ptr) + align - r);
139 * edac_mc_alloc: Allocate a struct mem_ctl_info structure
140 * @size_pvt: size of private storage needed
141 * @nr_csrows: Number of CWROWS needed for this MC
142 * @nr_chans: Number of channels for the MC
144 * Everything is kmalloc'ed as one big chunk - more efficient.
145 * Only can be used if all structures have the same lifetime - otherwise
146 * you have to allocate and initialize your own structures.
148 * Use edac_mc_free() to free mc structures allocated by this function.
151 * NULL allocation failed
152 * struct mem_ctl_info pointer
154 struct mem_ctl_info *edac_mc_alloc(unsigned sz_pvt, unsigned nr_csrows,
155 unsigned nr_chans, int edac_index)
157 struct mem_ctl_info *mci;
158 struct csrow_info *csi, *csrow;
159 struct channel_info *chi, *chp, *chan;
165 /* Figure out the offsets of the various items from the start of an mc
166 * structure. We want the alignment of each item to be at least as
167 * stringent as what the compiler would provide if we could simply
168 * hardcode everything into a single struct.
170 mci = (struct mem_ctl_info *)0;
171 csi = edac_align_ptr(&mci[1], sizeof(*csi));
172 chi = edac_align_ptr(&csi[nr_csrows], sizeof(*chi));
173 pvt = edac_align_ptr(&chi[nr_chans * nr_csrows], sz_pvt);
174 size = ((unsigned long)pvt) + sz_pvt;
176 mci = kzalloc(size, GFP_KERNEL);
180 /* Adjust pointers so they point within the memory we just allocated
181 * rather than an imaginary chunk of memory located at address 0.
183 csi = (struct csrow_info *)(((char *)mci) + ((unsigned long)csi));
184 chi = (struct channel_info *)(((char *)mci) + ((unsigned long)chi));
185 pvt = sz_pvt ? (((char *)mci) + ((unsigned long)pvt)) : NULL;
187 /* setup index and various internal pointers */
188 mci->mc_idx = edac_index;
191 mci->nr_csrows = nr_csrows;
193 for (row = 0; row < nr_csrows; row++) {
195 csrow->csrow_idx = row;
197 csrow->nr_channels = nr_chans;
198 chp = &chi[row * nr_chans];
199 csrow->channels = chp;
201 for (chn = 0; chn < nr_chans; chn++) {
203 chan->chan_idx = chn;
208 mci->op_state = OP_ALLOC;
209 INIT_LIST_HEAD(&mci->grp_kobj_list);
212 * Initialize the 'root' kobj for the edac_mc controller
214 err = edac_mc_register_sysfs_main_kobj(mci);
220 /* at this point, the root kobj is valid, and in order to
221 * 'free' the object, then the function:
222 * edac_mc_unregister_sysfs_main_kobj() must be called
223 * which will perform kobj unregistration and the actual free
224 * will occur during the kobject callback operation
228 EXPORT_SYMBOL_GPL(edac_mc_alloc);
232 * 'Free' a previously allocated 'mci' structure
233 * @mci: pointer to a struct mem_ctl_info structure
235 void edac_mc_free(struct mem_ctl_info *mci)
237 debugf1("%s()\n", __func__);
239 edac_mc_unregister_sysfs_main_kobj(mci);
241 /* free the mci instance memory here */
244 EXPORT_SYMBOL_GPL(edac_mc_free);
250 * scan list of controllers looking for the one that manages
252 * @dev: pointer to a struct device related with the MCI
254 struct mem_ctl_info *find_mci_by_dev(struct device *dev)
256 struct mem_ctl_info *mci;
257 struct list_head *item;
259 debugf3("%s()\n", __func__);
261 list_for_each(item, &mc_devices) {
262 mci = list_entry(item, struct mem_ctl_info, link);
270 EXPORT_SYMBOL_GPL(find_mci_by_dev);
273 * handler for EDAC to check if NMI type handler has asserted interrupt
275 static int edac_mc_assert_error_check_and_clear(void)
279 if (edac_op_state == EDAC_OPSTATE_POLL)
282 old_state = edac_err_assert;
289 * edac_mc_workq_function
290 * performs the operation scheduled by a workq request
292 static void edac_mc_workq_function(struct work_struct *work_req)
294 struct delayed_work *d_work = to_delayed_work(work_req);
295 struct mem_ctl_info *mci = to_edac_mem_ctl_work(d_work);
297 mutex_lock(&mem_ctls_mutex);
299 /* if this control struct has movd to offline state, we are done */
300 if (mci->op_state == OP_OFFLINE) {
301 mutex_unlock(&mem_ctls_mutex);
305 /* Only poll controllers that are running polled and have a check */
306 if (edac_mc_assert_error_check_and_clear() && (mci->edac_check != NULL))
307 mci->edac_check(mci);
309 mutex_unlock(&mem_ctls_mutex);
312 queue_delayed_work(edac_workqueue, &mci->work,
313 msecs_to_jiffies(edac_mc_get_poll_msec()));
317 * edac_mc_workq_setup
318 * initialize a workq item for this mci
319 * passing in the new delay period in msec
323 * called with the mem_ctls_mutex held
325 static void edac_mc_workq_setup(struct mem_ctl_info *mci, unsigned msec)
327 debugf0("%s()\n", __func__);
329 /* if this instance is not in the POLL state, then simply return */
330 if (mci->op_state != OP_RUNNING_POLL)
333 INIT_DELAYED_WORK(&mci->work, edac_mc_workq_function);
334 queue_delayed_work(edac_workqueue, &mci->work, msecs_to_jiffies(msec));
338 * edac_mc_workq_teardown
339 * stop the workq processing on this mci
343 * called WITHOUT lock held
345 static void edac_mc_workq_teardown(struct mem_ctl_info *mci)
349 if (mci->op_state != OP_RUNNING_POLL)
352 status = cancel_delayed_work(&mci->work);
354 debugf0("%s() not canceled, flush the queue\n",
357 /* workq instance might be running, wait for it */
358 flush_workqueue(edac_workqueue);
363 * edac_mc_reset_delay_period(unsigned long value)
365 * user space has updated our poll period value, need to
366 * reset our workq delays
368 void edac_mc_reset_delay_period(int value)
370 struct mem_ctl_info *mci;
371 struct list_head *item;
373 mutex_lock(&mem_ctls_mutex);
375 /* scan the list and turn off all workq timers, doing so under lock
377 list_for_each(item, &mc_devices) {
378 mci = list_entry(item, struct mem_ctl_info, link);
380 if (mci->op_state == OP_RUNNING_POLL)
381 cancel_delayed_work(&mci->work);
384 mutex_unlock(&mem_ctls_mutex);
387 /* re-walk the list, and reset the poll delay */
388 mutex_lock(&mem_ctls_mutex);
390 list_for_each(item, &mc_devices) {
391 mci = list_entry(item, struct mem_ctl_info, link);
393 edac_mc_workq_setup(mci, (unsigned long) value);
396 mutex_unlock(&mem_ctls_mutex);
401 /* Return 0 on success, 1 on failure.
402 * Before calling this function, caller must
403 * assign a unique value to mci->mc_idx.
407 * called with the mem_ctls_mutex lock held
409 static int add_mc_to_global_list(struct mem_ctl_info *mci)
411 struct list_head *item, *insert_before;
412 struct mem_ctl_info *p;
414 insert_before = &mc_devices;
416 p = find_mci_by_dev(mci->dev);
417 if (unlikely(p != NULL))
420 list_for_each(item, &mc_devices) {
421 p = list_entry(item, struct mem_ctl_info, link);
423 if (p->mc_idx >= mci->mc_idx) {
424 if (unlikely(p->mc_idx == mci->mc_idx))
427 insert_before = item;
432 list_add_tail_rcu(&mci->link, insert_before);
433 atomic_inc(&edac_handlers);
437 edac_printk(KERN_WARNING, EDAC_MC,
438 "%s (%s) %s %s already assigned %d\n", dev_name(p->dev),
439 edac_dev_name(mci), p->mod_name, p->ctl_name, p->mc_idx);
443 edac_printk(KERN_WARNING, EDAC_MC,
444 "bug in low-level driver: attempt to assign\n"
445 " duplicate mc_idx %d in %s()\n", p->mc_idx, __func__);
449 static void del_mc_from_global_list(struct mem_ctl_info *mci)
451 atomic_dec(&edac_handlers);
452 list_del_rcu(&mci->link);
454 /* these are for safe removal of devices from global list while
455 * NMI handlers may be traversing list
458 INIT_LIST_HEAD(&mci->link);
462 * edac_mc_find: Search for a mem_ctl_info structure whose index is 'idx'.
464 * If found, return a pointer to the structure.
467 * Caller must hold mem_ctls_mutex.
469 struct mem_ctl_info *edac_mc_find(int idx)
471 struct list_head *item;
472 struct mem_ctl_info *mci;
474 list_for_each(item, &mc_devices) {
475 mci = list_entry(item, struct mem_ctl_info, link);
477 if (mci->mc_idx >= idx) {
478 if (mci->mc_idx == idx)
487 EXPORT_SYMBOL(edac_mc_find);
490 * edac_mc_add_mc: Insert the 'mci' structure into the mci global list and
491 * create sysfs entries associated with mci structure
492 * @mci: pointer to the mci structure to be added to the list
493 * @mc_idx: A unique numeric identifier to be assigned to the 'mci' structure.
500 /* FIXME - should a warning be printed if no error detection? correction? */
501 int edac_mc_add_mc(struct mem_ctl_info *mci)
503 debugf0("%s()\n", __func__);
505 #ifdef CONFIG_EDAC_DEBUG
506 if (edac_debug_level >= 3)
507 edac_mc_dump_mci(mci);
509 if (edac_debug_level >= 4) {
512 for (i = 0; i < mci->nr_csrows; i++) {
515 edac_mc_dump_csrow(&mci->csrows[i]);
516 for (j = 0; j < mci->csrows[i].nr_channels; j++)
517 edac_mc_dump_channel(&mci->csrows[i].
522 mutex_lock(&mem_ctls_mutex);
524 if (add_mc_to_global_list(mci))
527 /* set load time so that error rate can be tracked */
528 mci->start_time = jiffies;
530 if (edac_create_sysfs_mci_device(mci)) {
531 edac_mc_printk(mci, KERN_WARNING,
532 "failed to create sysfs device\n");
536 /* If there IS a check routine, then we are running POLLED */
537 if (mci->edac_check != NULL) {
538 /* This instance is NOW RUNNING */
539 mci->op_state = OP_RUNNING_POLL;
541 edac_mc_workq_setup(mci, edac_mc_get_poll_msec());
543 mci->op_state = OP_RUNNING_INTERRUPT;
546 /* Report action taken */
547 edac_mc_printk(mci, KERN_INFO, "Giving out device to '%s' '%s':"
548 " DEV %s\n", mci->mod_name, mci->ctl_name, edac_dev_name(mci));
550 mutex_unlock(&mem_ctls_mutex);
554 del_mc_from_global_list(mci);
557 mutex_unlock(&mem_ctls_mutex);
560 EXPORT_SYMBOL_GPL(edac_mc_add_mc);
563 * edac_mc_del_mc: Remove sysfs entries for specified mci structure and
564 * remove mci structure from global list
565 * @pdev: Pointer to 'struct device' representing mci structure to remove.
567 * Return pointer to removed mci structure, or NULL if device not found.
569 struct mem_ctl_info *edac_mc_del_mc(struct device *dev)
571 struct mem_ctl_info *mci;
573 debugf0("%s()\n", __func__);
575 mutex_lock(&mem_ctls_mutex);
577 /* find the requested mci struct in the global list */
578 mci = find_mci_by_dev(dev);
580 mutex_unlock(&mem_ctls_mutex);
584 del_mc_from_global_list(mci);
585 mutex_unlock(&mem_ctls_mutex);
587 /* flush workq processes */
588 edac_mc_workq_teardown(mci);
590 /* marking MCI offline */
591 mci->op_state = OP_OFFLINE;
593 /* remove from sysfs */
594 edac_remove_sysfs_mci_device(mci);
596 edac_printk(KERN_INFO, EDAC_MC,
597 "Removed device %d for %s %s: DEV %s\n", mci->mc_idx,
598 mci->mod_name, mci->ctl_name, edac_dev_name(mci));
602 EXPORT_SYMBOL_GPL(edac_mc_del_mc);
604 static void edac_mc_scrub_block(unsigned long page, unsigned long offset,
609 unsigned long flags = 0;
611 debugf3("%s()\n", __func__);
614 page = mfn_to_local_pfn(page);
617 /* ECC error page was not in our memory. Ignore it. */
618 if (!pfn_valid(page))
621 /* Find the actual page structure then map it and fix */
622 pg = pfn_to_page(page);
625 local_irq_save(flags);
627 virt_addr = kmap_atomic(pg, KM_BOUNCE_READ);
629 /* Perform architecture specific atomic scrub operation */
630 atomic_scrub(virt_addr + offset, size);
632 /* Unmap and complete */
633 kunmap_atomic(virt_addr, KM_BOUNCE_READ);
636 local_irq_restore(flags);
639 /* FIXME - should return -1 */
640 int edac_mc_find_csrow_by_page(struct mem_ctl_info *mci, unsigned long page)
642 struct csrow_info *csrows = mci->csrows;
645 debugf1("MC%d: %s(): 0x%lx\n", mci->mc_idx, __func__, page);
648 for (i = 0; i < mci->nr_csrows; i++) {
649 struct csrow_info *csrow = &csrows[i];
651 if (csrow->nr_pages == 0)
654 debugf3("MC%d: %s(): first(0x%lx) page(0x%lx) last(0x%lx) "
655 "mask(0x%lx)\n", mci->mc_idx, __func__,
656 csrow->first_page, page, csrow->last_page,
659 if ((page >= csrow->first_page) &&
660 (page <= csrow->last_page) &&
661 ((page & csrow->page_mask) ==
662 (csrow->first_page & csrow->page_mask))) {
669 edac_mc_printk(mci, KERN_ERR,
670 "could not look up page error address %lx\n",
671 (unsigned long)page);
675 EXPORT_SYMBOL_GPL(edac_mc_find_csrow_by_page);
677 /* FIXME - setable log (warning/emerg) levels */
678 /* FIXME - integrate with evlog: http://evlog.sourceforge.net/ */
679 void edac_mc_handle_ce(struct mem_ctl_info *mci,
680 unsigned long page_frame_number,
681 unsigned long offset_in_page, unsigned long syndrome,
682 int row, int channel, const char *msg)
684 unsigned long remapped_page;
686 debugf3("MC%d: %s()\n", mci->mc_idx, __func__);
688 /* FIXME - maybe make panic on INTERNAL ERROR an option */
689 if (row >= mci->nr_csrows || row < 0) {
690 /* something is wrong */
691 edac_mc_printk(mci, KERN_ERR,
692 "INTERNAL ERROR: row out of range "
693 "(%d >= %d)\n", row, mci->nr_csrows);
694 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
698 if (channel >= mci->csrows[row].nr_channels || channel < 0) {
699 /* something is wrong */
700 edac_mc_printk(mci, KERN_ERR,
701 "INTERNAL ERROR: channel out of range "
702 "(%d >= %d)\n", channel,
703 mci->csrows[row].nr_channels);
704 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
708 if (edac_mc_get_log_ce())
709 /* FIXME - put in DIMM location */
710 edac_mc_printk(mci, KERN_WARNING,
711 "CE page 0x%lx, offset 0x%lx, grain %d, syndrome "
712 "0x%lx, row %d, channel %d, label \"%s\": %s\n",
713 page_frame_number, offset_in_page,
714 mci->csrows[row].grain, syndrome, row, channel,
715 mci->csrows[row].channels[channel].label, msg);
718 mci->csrows[row].ce_count++;
719 mci->csrows[row].channels[channel].ce_count++;
721 if (mci->scrub_mode & SCRUB_SW_SRC) {
723 * Some MC's can remap memory so that it is still available
724 * at a different address when PCI devices map into memory.
725 * MC's that can't do this lose the memory where PCI devices
726 * are mapped. This mapping is MC dependent and so we call
727 * back into the MC driver for it to map the MC page to
728 * a physical (CPU) page which can then be mapped to a virtual
729 * page - which can then be scrubbed.
731 remapped_page = mci->ctl_page_to_phys ?
732 mci->ctl_page_to_phys(mci, page_frame_number) :
735 edac_mc_scrub_block(remapped_page, offset_in_page,
736 mci->csrows[row].grain);
739 EXPORT_SYMBOL_GPL(edac_mc_handle_ce);
741 void edac_mc_handle_ce_no_info(struct mem_ctl_info *mci, const char *msg)
743 if (edac_mc_get_log_ce())
744 edac_mc_printk(mci, KERN_WARNING,
745 "CE - no information available: %s\n", msg);
747 mci->ce_noinfo_count++;
750 EXPORT_SYMBOL_GPL(edac_mc_handle_ce_no_info);
752 void edac_mc_handle_ue(struct mem_ctl_info *mci,
753 unsigned long page_frame_number,
754 unsigned long offset_in_page, int row, const char *msg)
756 int len = EDAC_MC_LABEL_LEN * 4;
757 char labels[len + 1];
762 debugf3("MC%d: %s()\n", mci->mc_idx, __func__);
764 /* FIXME - maybe make panic on INTERNAL ERROR an option */
765 if (row >= mci->nr_csrows || row < 0) {
766 /* something is wrong */
767 edac_mc_printk(mci, KERN_ERR,
768 "INTERNAL ERROR: row out of range "
769 "(%d >= %d)\n", row, mci->nr_csrows);
770 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
774 chars = snprintf(pos, len + 1, "%s",
775 mci->csrows[row].channels[0].label);
779 for (chan = 1; (chan < mci->csrows[row].nr_channels) && (len > 0);
781 chars = snprintf(pos, len + 1, ":%s",
782 mci->csrows[row].channels[chan].label);
787 if (edac_mc_get_log_ue())
788 edac_mc_printk(mci, KERN_EMERG,
789 "UE page 0x%lx, offset 0x%lx, grain %d, row %d, "
790 "labels \"%s\": %s\n", page_frame_number,
791 offset_in_page, mci->csrows[row].grain, row,
794 if (edac_mc_get_panic_on_ue())
795 panic("EDAC MC%d: UE page 0x%lx, offset 0x%lx, grain %d, "
796 "row %d, labels \"%s\": %s\n", mci->mc_idx,
797 page_frame_number, offset_in_page,
798 mci->csrows[row].grain, row, labels, msg);
801 mci->csrows[row].ue_count++;
803 EXPORT_SYMBOL_GPL(edac_mc_handle_ue);
805 void edac_mc_handle_ue_no_info(struct mem_ctl_info *mci, const char *msg)
807 if (edac_mc_get_panic_on_ue())
808 panic("EDAC MC%d: Uncorrected Error", mci->mc_idx);
810 if (edac_mc_get_log_ue())
811 edac_mc_printk(mci, KERN_WARNING,
812 "UE - no information available: %s\n", msg);
813 mci->ue_noinfo_count++;
816 EXPORT_SYMBOL_GPL(edac_mc_handle_ue_no_info);
818 /*************************************************************
819 * On Fully Buffered DIMM modules, this help function is
820 * called to process UE events
822 void edac_mc_handle_fbd_ue(struct mem_ctl_info *mci,
824 unsigned int channela,
825 unsigned int channelb, char *msg)
827 int len = EDAC_MC_LABEL_LEN * 4;
828 char labels[len + 1];
832 if (csrow >= mci->nr_csrows) {
833 /* something is wrong */
834 edac_mc_printk(mci, KERN_ERR,
835 "INTERNAL ERROR: row out of range (%d >= %d)\n",
836 csrow, mci->nr_csrows);
837 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
841 if (channela >= mci->csrows[csrow].nr_channels) {
842 /* something is wrong */
843 edac_mc_printk(mci, KERN_ERR,
844 "INTERNAL ERROR: channel-a out of range "
846 channela, mci->csrows[csrow].nr_channels);
847 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
851 if (channelb >= mci->csrows[csrow].nr_channels) {
852 /* something is wrong */
853 edac_mc_printk(mci, KERN_ERR,
854 "INTERNAL ERROR: channel-b out of range "
856 channelb, mci->csrows[csrow].nr_channels);
857 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
862 mci->csrows[csrow].ue_count++;
864 /* Generate the DIMM labels from the specified channels */
865 chars = snprintf(pos, len + 1, "%s",
866 mci->csrows[csrow].channels[channela].label);
869 chars = snprintf(pos, len + 1, "-%s",
870 mci->csrows[csrow].channels[channelb].label);
872 if (edac_mc_get_log_ue())
873 edac_mc_printk(mci, KERN_EMERG,
874 "UE row %d, channel-a= %d channel-b= %d "
875 "labels \"%s\": %s\n", csrow, channela, channelb,
878 if (edac_mc_get_panic_on_ue())
879 panic("UE row %d, channel-a= %d channel-b= %d "
880 "labels \"%s\": %s\n", csrow, channela,
881 channelb, labels, msg);
883 EXPORT_SYMBOL(edac_mc_handle_fbd_ue);
885 /*************************************************************
886 * On Fully Buffered DIMM modules, this help function is
887 * called to process CE events
889 void edac_mc_handle_fbd_ce(struct mem_ctl_info *mci,
890 unsigned int csrow, unsigned int channel, char *msg)
893 /* Ensure boundary values */
894 if (csrow >= mci->nr_csrows) {
895 /* something is wrong */
896 edac_mc_printk(mci, KERN_ERR,
897 "INTERNAL ERROR: row out of range (%d >= %d)\n",
898 csrow, mci->nr_csrows);
899 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
902 if (channel >= mci->csrows[csrow].nr_channels) {
903 /* something is wrong */
904 edac_mc_printk(mci, KERN_ERR,
905 "INTERNAL ERROR: channel out of range (%d >= %d)\n",
906 channel, mci->csrows[csrow].nr_channels);
907 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
911 if (edac_mc_get_log_ce())
912 /* FIXME - put in DIMM location */
913 edac_mc_printk(mci, KERN_WARNING,
914 "CE row %d, channel %d, label \"%s\": %s\n",
916 mci->csrows[csrow].channels[channel].label, msg);
919 mci->csrows[csrow].ce_count++;
920 mci->csrows[csrow].channels[channel].ce_count++;
922 EXPORT_SYMBOL(edac_mc_handle_fbd_ce);