2 * cpc925_edac.c, EDAC driver for IBM CPC925 Bridge and Memory Controller.
4 * Copyright (c) 2008 Wind River Systems, Inc.
6 * Authors: Cao Qingtao <qingtao.cao@windriver.com>
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License version 2 as
10 * published by the Free Software Foundation.
12 * This program is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
15 * See the GNU General Public License for more details.
17 * You should have received a copy of the GNU General Public License
18 * along with this program; if not, write to the Free Software
19 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
22 #include <linux/module.h>
23 #include <linux/init.h>
25 #include <linux/edac.h>
27 #include <linux/platform_device.h>
28 #include <linux/gfp.h>
30 #include "edac_core.h"
31 #include "edac_module.h"
33 #define CPC925_EDAC_REVISION " Ver: 1.0.0"
34 #define CPC925_EDAC_MOD_STR "cpc925_edac"
36 #define cpc925_printk(level, fmt, arg...) \
37 edac_printk(level, "CPC925", fmt, ##arg)
39 #define cpc925_mc_printk(mci, level, fmt, arg...) \
40 edac_mc_chipset_printk(mci, level, "CPC925", fmt, ##arg)
43 * CPC925 registers are of 32 bits with bit0 defined at the
44 * most significant bit and bit31 at that of least significant.
46 #define CPC925_BITS_PER_REG 32
47 #define CPC925_BIT(nr) (1UL << (CPC925_BITS_PER_REG - 1 - nr))
50 * EDAC device names for the error detections of
51 * CPU Interface and Hypertransport Link.
53 #define CPC925_CPU_ERR_DEV "cpu"
54 #define CPC925_HT_LINK_DEV "htlink"
56 /* Suppose DDR Refresh cycle is 15.6 microsecond */
57 #define CPC925_REF_FREQ 0xFA69
58 #define CPC925_SCRUB_BLOCK_SIZE 64 /* bytes */
59 #define CPC925_NR_CSROWS 8
62 * All registers and bits definitions are taken from
63 * "CPC925 Bridge and Memory Controller User Manual, SA14-2761-02".
67 * CPU and Memory Controller Registers
69 /************************************************************
70 * Processor Interface Exception Mask Register (APIMASK)
71 ************************************************************/
72 #define REG_APIMASK_OFFSET 0x30070
74 APIMASK_DART = CPC925_BIT(0), /* DART Exception */
75 APIMASK_ADI0 = CPC925_BIT(1), /* Handshake Error on PI0_ADI */
76 APIMASK_ADI1 = CPC925_BIT(2), /* Handshake Error on PI1_ADI */
77 APIMASK_STAT = CPC925_BIT(3), /* Status Exception */
78 APIMASK_DERR = CPC925_BIT(4), /* Data Error Exception */
79 APIMASK_ADRS0 = CPC925_BIT(5), /* Addressing Exception on PI0 */
80 APIMASK_ADRS1 = CPC925_BIT(6), /* Addressing Exception on PI1 */
82 APIMASK_ECC_UE_H = CPC925_BIT(8), /* UECC upper */
83 APIMASK_ECC_CE_H = CPC925_BIT(9), /* CECC upper */
84 APIMASK_ECC_UE_L = CPC925_BIT(10), /* UECC lower */
85 APIMASK_ECC_CE_L = CPC925_BIT(11), /* CECC lower */
87 CPU_MASK_ENABLE = (APIMASK_DART | APIMASK_ADI0 | APIMASK_ADI1 |
88 APIMASK_STAT | APIMASK_DERR | APIMASK_ADRS0 |
90 ECC_MASK_ENABLE = (APIMASK_ECC_UE_H | APIMASK_ECC_CE_H |
91 APIMASK_ECC_UE_L | APIMASK_ECC_CE_L),
93 #define APIMASK_ADI(n) CPC925_BIT(((n)+1))
95 /************************************************************
96 * Processor Interface Exception Register (APIEXCP)
97 ************************************************************/
98 #define REG_APIEXCP_OFFSET 0x30060
100 APIEXCP_DART = CPC925_BIT(0), /* DART Exception */
101 APIEXCP_ADI0 = CPC925_BIT(1), /* Handshake Error on PI0_ADI */
102 APIEXCP_ADI1 = CPC925_BIT(2), /* Handshake Error on PI1_ADI */
103 APIEXCP_STAT = CPC925_BIT(3), /* Status Exception */
104 APIEXCP_DERR = CPC925_BIT(4), /* Data Error Exception */
105 APIEXCP_ADRS0 = CPC925_BIT(5), /* Addressing Exception on PI0 */
106 APIEXCP_ADRS1 = CPC925_BIT(6), /* Addressing Exception on PI1 */
107 /* BIT(7) Reserved */
108 APIEXCP_ECC_UE_H = CPC925_BIT(8), /* UECC upper */
109 APIEXCP_ECC_CE_H = CPC925_BIT(9), /* CECC upper */
110 APIEXCP_ECC_UE_L = CPC925_BIT(10), /* UECC lower */
111 APIEXCP_ECC_CE_L = CPC925_BIT(11), /* CECC lower */
113 CPU_EXCP_DETECTED = (APIEXCP_DART | APIEXCP_ADI0 | APIEXCP_ADI1 |
114 APIEXCP_STAT | APIEXCP_DERR | APIEXCP_ADRS0 |
116 UECC_EXCP_DETECTED = (APIEXCP_ECC_UE_H | APIEXCP_ECC_UE_L),
117 CECC_EXCP_DETECTED = (APIEXCP_ECC_CE_H | APIEXCP_ECC_CE_L),
118 ECC_EXCP_DETECTED = (UECC_EXCP_DETECTED | CECC_EXCP_DETECTED),
121 /************************************************************
122 * Memory Bus Configuration Register (MBCR)
123 ************************************************************/
124 #define REG_MBCR_OFFSET 0x2190
125 #define MBCR_64BITCFG_SHIFT 23
126 #define MBCR_64BITCFG_MASK (1UL << MBCR_64BITCFG_SHIFT)
127 #define MBCR_64BITBUS_SHIFT 22
128 #define MBCR_64BITBUS_MASK (1UL << MBCR_64BITBUS_SHIFT)
130 /************************************************************
131 * Memory Bank Mode Register (MBMR)
132 ************************************************************/
133 #define REG_MBMR_OFFSET 0x21C0
134 #define MBMR_MODE_MAX_VALUE 0xF
135 #define MBMR_MODE_SHIFT 25
136 #define MBMR_MODE_MASK (MBMR_MODE_MAX_VALUE << MBMR_MODE_SHIFT)
137 #define MBMR_BBA_SHIFT 24
138 #define MBMR_BBA_MASK (1UL << MBMR_BBA_SHIFT)
140 /************************************************************
141 * Memory Bank Boundary Address Register (MBBAR)
142 ************************************************************/
143 #define REG_MBBAR_OFFSET 0x21D0
144 #define MBBAR_BBA_MAX_VALUE 0xFF
145 #define MBBAR_BBA_SHIFT 24
146 #define MBBAR_BBA_MASK (MBBAR_BBA_MAX_VALUE << MBBAR_BBA_SHIFT)
148 /************************************************************
149 * Memory Scrub Control Register (MSCR)
150 ************************************************************/
151 #define REG_MSCR_OFFSET 0x2400
152 #define MSCR_SCRUB_MOD_MASK 0xC0000000 /* scrub_mod - bit0:1*/
153 #define MSCR_BACKGR_SCRUB 0x40000000 /* 01 */
154 #define MSCR_SI_SHIFT 16 /* si - bit8:15*/
155 #define MSCR_SI_MAX_VALUE 0xFF
156 #define MSCR_SI_MASK (MSCR_SI_MAX_VALUE << MSCR_SI_SHIFT)
158 /************************************************************
159 * Memory Scrub Range Start Register (MSRSR)
160 ************************************************************/
161 #define REG_MSRSR_OFFSET 0x2410
163 /************************************************************
164 * Memory Scrub Range End Register (MSRER)
165 ************************************************************/
166 #define REG_MSRER_OFFSET 0x2420
168 /************************************************************
169 * Memory Scrub Pattern Register (MSPR)
170 ************************************************************/
171 #define REG_MSPR_OFFSET 0x2430
173 /************************************************************
174 * Memory Check Control Register (MCCR)
175 ************************************************************/
176 #define REG_MCCR_OFFSET 0x2440
178 MCCR_ECC_EN = CPC925_BIT(0), /* ECC high and low check */
181 /************************************************************
182 * Memory Check Range End Register (MCRER)
183 ************************************************************/
184 #define REG_MCRER_OFFSET 0x2450
186 /************************************************************
187 * Memory Error Address Register (MEAR)
188 ************************************************************/
189 #define REG_MEAR_OFFSET 0x2460
190 #define MEAR_BCNT_MAX_VALUE 0x3
191 #define MEAR_BCNT_SHIFT 30
192 #define MEAR_BCNT_MASK (MEAR_BCNT_MAX_VALUE << MEAR_BCNT_SHIFT)
193 #define MEAR_RANK_MAX_VALUE 0x7
194 #define MEAR_RANK_SHIFT 27
195 #define MEAR_RANK_MASK (MEAR_RANK_MAX_VALUE << MEAR_RANK_SHIFT)
196 #define MEAR_COL_MAX_VALUE 0x7FF
197 #define MEAR_COL_SHIFT 16
198 #define MEAR_COL_MASK (MEAR_COL_MAX_VALUE << MEAR_COL_SHIFT)
199 #define MEAR_BANK_MAX_VALUE 0x3
200 #define MEAR_BANK_SHIFT 14
201 #define MEAR_BANK_MASK (MEAR_BANK_MAX_VALUE << MEAR_BANK_SHIFT)
202 #define MEAR_ROW_MASK 0x00003FFF
204 /************************************************************
205 * Memory Error Syndrome Register (MESR)
206 ************************************************************/
207 #define REG_MESR_OFFSET 0x2470
208 #define MESR_ECC_SYN_H_MASK 0xFF00
209 #define MESR_ECC_SYN_L_MASK 0x00FF
211 /************************************************************
212 * Memory Mode Control Register (MMCR)
213 ************************************************************/
214 #define REG_MMCR_OFFSET 0x2500
216 MMCR_REG_DIMM_MODE = CPC925_BIT(3),
220 * HyperTransport Link Registers
222 /************************************************************
223 * Error Handling/Enumeration Scratch Pad Register (ERRCTRL)
224 ************************************************************/
225 #define REG_ERRCTRL_OFFSET 0x70140
226 enum errctrl_bits { /* nonfatal interrupts for */
227 ERRCTRL_SERR_NF = CPC925_BIT(0), /* system error */
228 ERRCTRL_CRC_NF = CPC925_BIT(1), /* CRC error */
229 ERRCTRL_RSP_NF = CPC925_BIT(2), /* Response error */
230 ERRCTRL_EOC_NF = CPC925_BIT(3), /* End-Of-Chain error */
231 ERRCTRL_OVF_NF = CPC925_BIT(4), /* Overflow error */
232 ERRCTRL_PROT_NF = CPC925_BIT(5), /* Protocol error */
234 ERRCTRL_RSP_ERR = CPC925_BIT(6), /* Response error received */
235 ERRCTRL_CHN_FAL = CPC925_BIT(7), /* Sync flooding detected */
237 HT_ERRCTRL_ENABLE = (ERRCTRL_SERR_NF | ERRCTRL_CRC_NF |
238 ERRCTRL_RSP_NF | ERRCTRL_EOC_NF |
239 ERRCTRL_OVF_NF | ERRCTRL_PROT_NF),
240 HT_ERRCTRL_DETECTED = (ERRCTRL_RSP_ERR | ERRCTRL_CHN_FAL),
243 /************************************************************
244 * Link Configuration and Link Control Register (LINKCTRL)
245 ************************************************************/
246 #define REG_LINKCTRL_OFFSET 0x70110
248 LINKCTRL_CRC_ERR = (CPC925_BIT(22) | CPC925_BIT(23)),
249 LINKCTRL_LINK_FAIL = CPC925_BIT(27),
251 HT_LINKCTRL_DETECTED = (LINKCTRL_CRC_ERR | LINKCTRL_LINK_FAIL),
254 /************************************************************
255 * Link FreqCap/Error/Freq/Revision ID Register (LINKERR)
256 ************************************************************/
257 #define REG_LINKERR_OFFSET 0x70120
259 LINKERR_EOC_ERR = CPC925_BIT(17), /* End-Of-Chain error */
260 LINKERR_OVF_ERR = CPC925_BIT(18), /* Receive Buffer Overflow */
261 LINKERR_PROT_ERR = CPC925_BIT(19), /* Protocol error */
263 HT_LINKERR_DETECTED = (LINKERR_EOC_ERR | LINKERR_OVF_ERR |
267 /************************************************************
268 * Bridge Control Register (BRGCTRL)
269 ************************************************************/
270 #define REG_BRGCTRL_OFFSET 0x70300
272 BRGCTRL_DETSERR = CPC925_BIT(0), /* SERR on Secondary Bus */
273 BRGCTRL_SECBUSRESET = CPC925_BIT(9), /* Secondary Bus Reset */
276 /* Private structure for edac memory controller */
277 struct cpc925_mc_pdata {
279 unsigned long total_mem;
284 /* Private structure for common edac device */
285 struct cpc925_dev_info {
287 struct platform_device *pdev;
290 struct edac_device_ctl_info *edac_dev;
291 void (*init)(struct cpc925_dev_info *dev_info);
292 void (*exit)(struct cpc925_dev_info *dev_info);
293 void (*check)(struct edac_device_ctl_info *edac_dev);
296 /* Get total memory size from Open Firmware DTB */
297 static void get_total_mem(struct cpc925_mc_pdata *pdata)
299 struct device_node *np = NULL;
300 const unsigned int *reg, *reg_end;
302 unsigned long start, size;
304 np = of_find_node_by_type(NULL, "memory");
308 aw = of_n_addr_cells(np);
309 sw = of_n_size_cells(np);
310 reg = (const unsigned int *)of_get_property(np, "reg", &len);
311 reg_end = reg + len/4;
313 pdata->total_mem = 0;
315 start = of_read_number(reg, aw);
317 size = of_read_number(reg, sw);
319 edac_dbg(1, "start 0x%lx, size 0x%lx\n", start, size);
320 pdata->total_mem += size;
321 } while (reg < reg_end);
324 edac_dbg(0, "total_mem 0x%lx\n", pdata->total_mem);
327 static void cpc925_init_csrows(struct mem_ctl_info *mci)
329 struct cpc925_mc_pdata *pdata = mci->pvt_info;
330 struct csrow_info *csrow;
331 struct dimm_info *dimm;
334 u32 mbmr, mbbar, bba, grain;
335 unsigned long row_size, nr_pages, last_nr_pages = 0;
337 get_total_mem(pdata);
339 for (index = 0; index < mci->nr_csrows; index++) {
340 mbmr = __raw_readl(pdata->vbase + REG_MBMR_OFFSET +
342 mbbar = __raw_readl(pdata->vbase + REG_MBBAR_OFFSET +
344 bba = (((mbmr & MBMR_BBA_MASK) >> MBMR_BBA_SHIFT) << 8) |
345 ((mbbar & MBBAR_BBA_MASK) >> MBBAR_BBA_SHIFT);
348 continue; /* not populated */
350 csrow = mci->csrows[index];
352 row_size = bba * (1UL << 28); /* 256M */
353 csrow->first_page = last_nr_pages;
354 nr_pages = row_size >> PAGE_SHIFT;
355 csrow->last_page = csrow->first_page + nr_pages - 1;
356 last_nr_pages = csrow->last_page + 1;
358 switch (csrow->nr_channels) {
359 case 1: /* Single channel */
360 grain = 32; /* four-beat burst of 32 bytes */
362 case 2: /* Dual channel */
364 grain = 64; /* four-beat burst of 64 bytes */
367 switch ((mbmr & MBMR_MODE_MASK) >> MBMR_MODE_SHIFT) {
368 case 6: /* 0110, no way to differentiate X8 VS X16 */
381 for (j = 0; j < csrow->nr_channels; j++) {
382 dimm = csrow->channels[j]->dimm;
383 dimm->nr_pages = nr_pages / csrow->nr_channels;
384 dimm->mtype = MEM_RDDR;
385 dimm->edac_mode = EDAC_SECDED;
392 /* Enable memory controller ECC detection */
393 static void cpc925_mc_init(struct mem_ctl_info *mci)
395 struct cpc925_mc_pdata *pdata = mci->pvt_info;
399 /* Enable various ECC error exceptions */
400 apimask = __raw_readl(pdata->vbase + REG_APIMASK_OFFSET);
401 if ((apimask & ECC_MASK_ENABLE) == 0) {
402 apimask |= ECC_MASK_ENABLE;
403 __raw_writel(apimask, pdata->vbase + REG_APIMASK_OFFSET);
406 /* Enable ECC detection */
407 mccr = __raw_readl(pdata->vbase + REG_MCCR_OFFSET);
408 if ((mccr & MCCR_ECC_EN) == 0) {
410 __raw_writel(mccr, pdata->vbase + REG_MCCR_OFFSET);
414 /* Disable memory controller ECC detection */
415 static void cpc925_mc_exit(struct mem_ctl_info *mci)
419 * We are supposed to clear the ECC error detection bits,
420 * and it will be no problem to do so. However, once they
421 * are cleared here if we want to re-install CPC925 EDAC
422 * module later, setting them up in cpc925_mc_init() will
423 * trigger machine check exception.
424 * Also, it's ok to leave ECC error detection bits enabled,
425 * since they are reset to 1 by default or by boot loader.
432 * Revert DDR column/row/bank addresses into page frame number and
435 * Suppose memory mode is 0x0111(128-bit mode, identical DIMM pairs),
436 * physical address(PA) bits to column address(CA) bits mappings are:
437 * CA 0 1 2 3 4 5 6 7 8 9 10
438 * PA 59 58 57 56 55 54 53 52 51 50 49
440 * physical address(PA) bits to bank address(BA) bits mappings are:
444 * physical address(PA) bits to row address(RA) bits mappings are:
445 * RA 0 1 2 3 4 5 6 7 8 9 10 11 12
446 * PA 36 35 34 48 47 46 45 40 41 42 39 38 37
448 static void cpc925_mc_get_pfn(struct mem_ctl_info *mci, u32 mear,
449 unsigned long *pfn, unsigned long *offset, int *csrow)
451 u32 bcnt, rank, col, bank, row;
456 bcnt = (mear & MEAR_BCNT_MASK) >> MEAR_BCNT_SHIFT;
457 rank = (mear & MEAR_RANK_MASK) >> MEAR_RANK_SHIFT;
458 col = (mear & MEAR_COL_MASK) >> MEAR_COL_SHIFT;
459 bank = (mear & MEAR_BANK_MASK) >> MEAR_BANK_SHIFT;
460 row = mear & MEAR_ROW_MASK;
464 #ifdef CONFIG_EDAC_DEBUG
465 if (mci->csrows[rank]->first_page == 0) {
466 cpc925_mc_printk(mci, KERN_ERR, "ECC occurs in a "
467 "non-populated csrow, broken hardware?\n");
472 /* Revert csrow number */
473 pa = mci->csrows[rank]->first_page << PAGE_SHIFT;
475 /* Revert column address */
477 for (i = 0; i < 11; i++) {
483 /* Revert bank address */
486 /* Revert row address, in 4 steps */
487 for (i = 0; i < 3; i++) {
493 for (i = 0; i < 3; i++) {
499 for (i = 0; i < 4; i++) {
505 for (i = 0; i < 3; i++) {
511 *offset = pa & (PAGE_SIZE - 1);
512 *pfn = pa >> PAGE_SHIFT;
514 edac_dbg(0, "ECC physical address 0x%lx\n", pa);
517 static int cpc925_mc_find_channel(struct mem_ctl_info *mci, u16 syndrome)
519 if ((syndrome & MESR_ECC_SYN_H_MASK) == 0)
522 if ((syndrome & MESR_ECC_SYN_L_MASK) == 0)
525 cpc925_mc_printk(mci, KERN_INFO, "Unexpected syndrome value: 0x%x\n",
530 /* Check memory controller registers for ECC errors */
531 static void cpc925_mc_check(struct mem_ctl_info *mci)
533 struct cpc925_mc_pdata *pdata = mci->pvt_info;
538 unsigned long pfn = 0, offset = 0;
539 int csrow = 0, channel = 0;
541 /* APIEXCP is cleared when read */
542 apiexcp = __raw_readl(pdata->vbase + REG_APIEXCP_OFFSET);
543 if ((apiexcp & ECC_EXCP_DETECTED) == 0)
546 mesr = __raw_readl(pdata->vbase + REG_MESR_OFFSET);
547 syndrome = mesr | (MESR_ECC_SYN_H_MASK | MESR_ECC_SYN_L_MASK);
549 mear = __raw_readl(pdata->vbase + REG_MEAR_OFFSET);
551 /* Revert column/row addresses into page frame number, etc */
552 cpc925_mc_get_pfn(mci, mear, &pfn, &offset, &csrow);
554 if (apiexcp & CECC_EXCP_DETECTED) {
555 cpc925_mc_printk(mci, KERN_INFO, "DRAM CECC Fault\n");
556 channel = cpc925_mc_find_channel(mci, syndrome);
557 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1,
558 pfn, offset, syndrome,
563 if (apiexcp & UECC_EXCP_DETECTED) {
564 cpc925_mc_printk(mci, KERN_INFO, "DRAM UECC Fault\n");
565 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
571 cpc925_mc_printk(mci, KERN_INFO, "Dump registers:\n");
572 cpc925_mc_printk(mci, KERN_INFO, "APIMASK 0x%08x\n",
573 __raw_readl(pdata->vbase + REG_APIMASK_OFFSET));
574 cpc925_mc_printk(mci, KERN_INFO, "APIEXCP 0x%08x\n",
576 cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Ctrl 0x%08x\n",
577 __raw_readl(pdata->vbase + REG_MSCR_OFFSET));
578 cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Rge Start 0x%08x\n",
579 __raw_readl(pdata->vbase + REG_MSRSR_OFFSET));
580 cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Rge End 0x%08x\n",
581 __raw_readl(pdata->vbase + REG_MSRER_OFFSET));
582 cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Pattern 0x%08x\n",
583 __raw_readl(pdata->vbase + REG_MSPR_OFFSET));
584 cpc925_mc_printk(mci, KERN_INFO, "Mem Chk Ctrl 0x%08x\n",
585 __raw_readl(pdata->vbase + REG_MCCR_OFFSET));
586 cpc925_mc_printk(mci, KERN_INFO, "Mem Chk Rge End 0x%08x\n",
587 __raw_readl(pdata->vbase + REG_MCRER_OFFSET));
588 cpc925_mc_printk(mci, KERN_INFO, "Mem Err Address 0x%08x\n",
590 cpc925_mc_printk(mci, KERN_INFO, "Mem Err Syndrome 0x%08x\n",
594 /******************** CPU err device********************************/
595 static u32 cpc925_cpu_mask_disabled(void)
597 struct device_node *cpus;
598 struct device_node *cpunode = NULL;
601 /* use cached value if available */
605 mask = APIMASK_ADI0 | APIMASK_ADI1;
607 cpus = of_find_node_by_path("/cpus");
609 cpc925_printk(KERN_DEBUG, "No /cpus node !\n");
613 while ((cpunode = of_get_next_child(cpus, cpunode)) != NULL) {
614 const u32 *reg = of_get_property(cpunode, "reg", NULL);
616 if (strcmp(cpunode->type, "cpu")) {
617 cpc925_printk(KERN_ERR, "Not a cpu node in /cpus: %s\n", cpunode->name);
621 if (reg == NULL || *reg > 2) {
622 cpc925_printk(KERN_ERR, "Bad reg value at %s\n", cpunode->full_name);
626 mask &= ~APIMASK_ADI(*reg);
629 if (mask != (APIMASK_ADI0 | APIMASK_ADI1)) {
630 /* We assume that each CPU sits on it's own PI and that
631 * for present CPUs the reg property equals to the PI
633 cpc925_printk(KERN_WARNING,
634 "Assuming PI id is equal to CPU MPIC id!\n");
637 of_node_put(cpunode);
643 /* Enable CPU Errors detection */
644 static void cpc925_cpu_init(struct cpc925_dev_info *dev_info)
649 apimask = __raw_readl(dev_info->vbase + REG_APIMASK_OFFSET);
651 cpumask = cpc925_cpu_mask_disabled();
652 if (apimask & cpumask) {
653 cpc925_printk(KERN_WARNING, "CPU(s) not present, "
654 "but enabled in APIMASK, disabling\n");
658 if ((apimask & CPU_MASK_ENABLE) == 0)
659 apimask |= CPU_MASK_ENABLE;
661 __raw_writel(apimask, dev_info->vbase + REG_APIMASK_OFFSET);
664 /* Disable CPU Errors detection */
665 static void cpc925_cpu_exit(struct cpc925_dev_info *dev_info)
669 * We are supposed to clear the CPU error detection bits,
670 * and it will be no problem to do so. However, once they
671 * are cleared here if we want to re-install CPC925 EDAC
672 * module later, setting them up in cpc925_cpu_init() will
673 * trigger machine check exception.
674 * Also, it's ok to leave CPU error detection bits enabled,
675 * since they are reset to 1 by default.
681 /* Check for CPU Errors */
682 static void cpc925_cpu_check(struct edac_device_ctl_info *edac_dev)
684 struct cpc925_dev_info *dev_info = edac_dev->pvt_info;
688 /* APIEXCP is cleared when read */
689 apiexcp = __raw_readl(dev_info->vbase + REG_APIEXCP_OFFSET);
690 if ((apiexcp & CPU_EXCP_DETECTED) == 0)
693 if ((apiexcp & ~cpc925_cpu_mask_disabled()) == 0)
696 apimask = __raw_readl(dev_info->vbase + REG_APIMASK_OFFSET);
697 cpc925_printk(KERN_INFO, "Processor Interface Fault\n"
698 "Processor Interface register dump:\n");
699 cpc925_printk(KERN_INFO, "APIMASK 0x%08x\n", apimask);
700 cpc925_printk(KERN_INFO, "APIEXCP 0x%08x\n", apiexcp);
702 edac_device_handle_ue(edac_dev, 0, 0, edac_dev->ctl_name);
705 /******************** HT Link err device****************************/
706 /* Enable HyperTransport Link Error detection */
707 static void cpc925_htlink_init(struct cpc925_dev_info *dev_info)
711 ht_errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
712 if ((ht_errctrl & HT_ERRCTRL_ENABLE) == 0) {
713 ht_errctrl |= HT_ERRCTRL_ENABLE;
714 __raw_writel(ht_errctrl, dev_info->vbase + REG_ERRCTRL_OFFSET);
718 /* Disable HyperTransport Link Error detection */
719 static void cpc925_htlink_exit(struct cpc925_dev_info *dev_info)
723 ht_errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
724 ht_errctrl &= ~HT_ERRCTRL_ENABLE;
725 __raw_writel(ht_errctrl, dev_info->vbase + REG_ERRCTRL_OFFSET);
728 /* Check for HyperTransport Link errors */
729 static void cpc925_htlink_check(struct edac_device_ctl_info *edac_dev)
731 struct cpc925_dev_info *dev_info = edac_dev->pvt_info;
732 u32 brgctrl = __raw_readl(dev_info->vbase + REG_BRGCTRL_OFFSET);
733 u32 linkctrl = __raw_readl(dev_info->vbase + REG_LINKCTRL_OFFSET);
734 u32 errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
735 u32 linkerr = __raw_readl(dev_info->vbase + REG_LINKERR_OFFSET);
737 if (!((brgctrl & BRGCTRL_DETSERR) ||
738 (linkctrl & HT_LINKCTRL_DETECTED) ||
739 (errctrl & HT_ERRCTRL_DETECTED) ||
740 (linkerr & HT_LINKERR_DETECTED)))
743 cpc925_printk(KERN_INFO, "HT Link Fault\n"
744 "HT register dump:\n");
745 cpc925_printk(KERN_INFO, "Bridge Ctrl 0x%08x\n",
747 cpc925_printk(KERN_INFO, "Link Config Ctrl 0x%08x\n",
749 cpc925_printk(KERN_INFO, "Error Enum and Ctrl 0x%08x\n",
751 cpc925_printk(KERN_INFO, "Link Error 0x%08x\n",
754 /* Clear by write 1 */
755 if (brgctrl & BRGCTRL_DETSERR)
756 __raw_writel(BRGCTRL_DETSERR,
757 dev_info->vbase + REG_BRGCTRL_OFFSET);
759 if (linkctrl & HT_LINKCTRL_DETECTED)
760 __raw_writel(HT_LINKCTRL_DETECTED,
761 dev_info->vbase + REG_LINKCTRL_OFFSET);
763 /* Initiate Secondary Bus Reset to clear the chain failure */
764 if (errctrl & ERRCTRL_CHN_FAL)
765 __raw_writel(BRGCTRL_SECBUSRESET,
766 dev_info->vbase + REG_BRGCTRL_OFFSET);
768 if (errctrl & ERRCTRL_RSP_ERR)
769 __raw_writel(ERRCTRL_RSP_ERR,
770 dev_info->vbase + REG_ERRCTRL_OFFSET);
772 if (linkerr & HT_LINKERR_DETECTED)
773 __raw_writel(HT_LINKERR_DETECTED,
774 dev_info->vbase + REG_LINKERR_OFFSET);
776 edac_device_handle_ce(edac_dev, 0, 0, edac_dev->ctl_name);
779 static struct cpc925_dev_info cpc925_devs[] = {
781 .ctl_name = CPC925_CPU_ERR_DEV,
782 .init = cpc925_cpu_init,
783 .exit = cpc925_cpu_exit,
784 .check = cpc925_cpu_check,
787 .ctl_name = CPC925_HT_LINK_DEV,
788 .init = cpc925_htlink_init,
789 .exit = cpc925_htlink_exit,
790 .check = cpc925_htlink_check,
796 * Add CPU Err detection and HyperTransport Link Err detection
797 * as common "edac_device", they have no corresponding device
798 * nodes in the Open Firmware DTB and we have to add platform
799 * devices for them. Also, they will share the MMIO with that
800 * of memory controller.
802 static void cpc925_add_edac_devices(void __iomem *vbase)
804 struct cpc925_dev_info *dev_info;
807 cpc925_printk(KERN_ERR, "MMIO not established yet\n");
811 for (dev_info = &cpc925_devs[0]; dev_info->init; dev_info++) {
812 dev_info->vbase = vbase;
813 dev_info->pdev = platform_device_register_simple(
814 dev_info->ctl_name, 0, NULL, 0);
815 if (IS_ERR(dev_info->pdev)) {
816 cpc925_printk(KERN_ERR,
817 "Can't register platform device for %s\n",
823 * Don't have to allocate private structure but
824 * make use of cpc925_devs[] instead.
826 dev_info->edac_idx = edac_device_alloc_index();
828 edac_device_alloc_ctl_info(0, dev_info->ctl_name,
829 1, NULL, 0, 0, NULL, 0, dev_info->edac_idx);
830 if (!dev_info->edac_dev) {
831 cpc925_printk(KERN_ERR, "No memory for edac device\n");
835 dev_info->edac_dev->pvt_info = dev_info;
836 dev_info->edac_dev->dev = &dev_info->pdev->dev;
837 dev_info->edac_dev->ctl_name = dev_info->ctl_name;
838 dev_info->edac_dev->mod_name = CPC925_EDAC_MOD_STR;
839 dev_info->edac_dev->dev_name = dev_name(&dev_info->pdev->dev);
841 if (edac_op_state == EDAC_OPSTATE_POLL)
842 dev_info->edac_dev->edac_check = dev_info->check;
845 dev_info->init(dev_info);
847 if (edac_device_add_device(dev_info->edac_dev) > 0) {
848 cpc925_printk(KERN_ERR,
849 "Unable to add edac device for %s\n",
854 edac_dbg(0, "Successfully added edac device for %s\n",
861 dev_info->exit(dev_info);
862 edac_device_free_ctl_info(dev_info->edac_dev);
864 platform_device_unregister(dev_info->pdev);
869 * Delete the common "edac_device" for CPU Err Detection
870 * and HyperTransport Link Err Detection
872 static void cpc925_del_edac_devices(void)
874 struct cpc925_dev_info *dev_info;
876 for (dev_info = &cpc925_devs[0]; dev_info->init; dev_info++) {
877 if (dev_info->edac_dev) {
878 edac_device_del_device(dev_info->edac_dev->dev);
879 edac_device_free_ctl_info(dev_info->edac_dev);
880 platform_device_unregister(dev_info->pdev);
884 dev_info->exit(dev_info);
886 edac_dbg(0, "Successfully deleted edac device for %s\n",
891 /* Convert current back-ground scrub rate into byte/sec bandwidth */
892 static int cpc925_get_sdram_scrub_rate(struct mem_ctl_info *mci)
894 struct cpc925_mc_pdata *pdata = mci->pvt_info;
899 mscr = __raw_readl(pdata->vbase + REG_MSCR_OFFSET);
900 si = (mscr & MSCR_SI_MASK) >> MSCR_SI_SHIFT;
902 edac_dbg(0, "Mem Scrub Ctrl Register 0x%x\n", mscr);
904 if (((mscr & MSCR_SCRUB_MOD_MASK) != MSCR_BACKGR_SCRUB) ||
906 cpc925_mc_printk(mci, KERN_INFO, "Scrub mode not enabled\n");
909 bw = CPC925_SCRUB_BLOCK_SIZE * 0xFA67 / si;
914 /* Return 0 for single channel; 1 for dual channel */
915 static int cpc925_mc_get_channels(void __iomem *vbase)
920 mbcr = __raw_readl(vbase + REG_MBCR_OFFSET);
923 * Dual channel only when 128-bit wide physical bus
924 * and 128-bit configuration.
926 if (((mbcr & MBCR_64BITCFG_MASK) == 0) &&
927 ((mbcr & MBCR_64BITBUS_MASK) == 0))
930 edac_dbg(0, "%s channel\n", (dual > 0) ? "Dual" : "Single");
935 static int cpc925_probe(struct platform_device *pdev)
937 static int edac_mc_idx;
938 struct mem_ctl_info *mci;
939 struct edac_mc_layer layers[2];
941 struct cpc925_mc_pdata *pdata;
943 int res = 0, nr_channels;
945 edac_dbg(0, "%s platform device found!\n", pdev->name);
947 if (!devres_open_group(&pdev->dev, cpc925_probe, GFP_KERNEL)) {
952 r = platform_get_resource(pdev, IORESOURCE_MEM, 0);
954 cpc925_printk(KERN_ERR, "Unable to get resource\n");
959 if (!devm_request_mem_region(&pdev->dev,
963 cpc925_printk(KERN_ERR, "Unable to request mem region\n");
968 vbase = devm_ioremap(&pdev->dev, r->start, resource_size(r));
970 cpc925_printk(KERN_ERR, "Unable to ioremap device\n");
975 nr_channels = cpc925_mc_get_channels(vbase) + 1;
977 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
978 layers[0].size = CPC925_NR_CSROWS;
979 layers[0].is_virt_csrow = true;
980 layers[1].type = EDAC_MC_LAYER_CHANNEL;
981 layers[1].size = nr_channels;
982 layers[1].is_virt_csrow = false;
983 mci = edac_mc_alloc(edac_mc_idx, ARRAY_SIZE(layers), layers,
984 sizeof(struct cpc925_mc_pdata));
986 cpc925_printk(KERN_ERR, "No memory for mem_ctl_info\n");
991 pdata = mci->pvt_info;
992 pdata->vbase = vbase;
993 pdata->edac_idx = edac_mc_idx++;
994 pdata->name = pdev->name;
996 mci->pdev = &pdev->dev;
997 platform_set_drvdata(pdev, mci);
998 mci->dev_name = dev_name(&pdev->dev);
999 mci->mtype_cap = MEM_FLAG_RDDR | MEM_FLAG_DDR;
1000 mci->edac_ctl_cap = EDAC_FLAG_NONE | EDAC_FLAG_SECDED;
1001 mci->edac_cap = EDAC_FLAG_SECDED;
1002 mci->mod_name = CPC925_EDAC_MOD_STR;
1003 mci->mod_ver = CPC925_EDAC_REVISION;
1004 mci->ctl_name = pdev->name;
1006 if (edac_op_state == EDAC_OPSTATE_POLL)
1007 mci->edac_check = cpc925_mc_check;
1009 mci->ctl_page_to_phys = NULL;
1010 mci->scrub_mode = SCRUB_SW_SRC;
1011 mci->set_sdram_scrub_rate = NULL;
1012 mci->get_sdram_scrub_rate = cpc925_get_sdram_scrub_rate;
1014 cpc925_init_csrows(mci);
1016 /* Setup memory controller registers */
1017 cpc925_mc_init(mci);
1019 if (edac_mc_add_mc(mci) > 0) {
1020 cpc925_mc_printk(mci, KERN_ERR, "Failed edac_mc_add_mc()\n");
1024 cpc925_add_edac_devices(vbase);
1026 /* get this far and it's successful */
1027 edac_dbg(0, "success\n");
1033 cpc925_mc_exit(mci);
1036 devm_release_mem_region(&pdev->dev, r->start, resource_size(r));
1038 devres_release_group(&pdev->dev, cpc925_probe);
1043 static int cpc925_remove(struct platform_device *pdev)
1045 struct mem_ctl_info *mci = platform_get_drvdata(pdev);
1048 * Delete common edac devices before edac mc, because
1049 * the former share the MMIO of the latter.
1051 cpc925_del_edac_devices();
1052 cpc925_mc_exit(mci);
1054 edac_mc_del_mc(&pdev->dev);
1060 static struct platform_driver cpc925_edac_driver = {
1061 .probe = cpc925_probe,
1062 .remove = cpc925_remove,
1064 .name = "cpc925_edac",
1068 static int __init cpc925_edac_init(void)
1072 printk(KERN_INFO "IBM CPC925 EDAC driver " CPC925_EDAC_REVISION "\n");
1073 printk(KERN_INFO "\t(c) 2008 Wind River Systems, Inc\n");
1075 /* Only support POLL mode so far */
1076 edac_op_state = EDAC_OPSTATE_POLL;
1078 ret = platform_driver_register(&cpc925_edac_driver);
1080 printk(KERN_WARNING "Failed to register %s\n",
1081 CPC925_EDAC_MOD_STR);
1087 static void __exit cpc925_edac_exit(void)
1089 platform_driver_unregister(&cpc925_edac_driver);
1092 module_init(cpc925_edac_init);
1093 module_exit(cpc925_edac_exit);
1095 MODULE_LICENSE("GPL");
1096 MODULE_AUTHOR("Cao Qingtao <qingtao.cao@windriver.com>");
1097 MODULE_DESCRIPTION("IBM CPC925 Bridge and MC EDAC kernel module");