edac: Rename the parent dev to pdev
[linux-3.10.git] / drivers / edac / e752x_edac.c
1 /*
2  * Intel e752x Memory Controller kernel module
3  * (C) 2004 Linux Networx (http://lnxi.com)
4  * This file may be distributed under the terms of the
5  * GNU General Public License.
6  *
7  * Implement support for the e7520, E7525, e7320 and i3100 memory controllers.
8  *
9  * Datasheets:
10  *      http://www.intel.in/content/www/in/en/chipsets/e7525-memory-controller-hub-datasheet.html
11  *      ftp://download.intel.com/design/intarch/datashts/31345803.pdf
12  *
13  * Written by Tom Zimmerman
14  *
15  * Contributors:
16  *      Thayne Harbaugh at realmsys.com (?)
17  *      Wang Zhenyu at intel.com
18  *      Dave Jiang at mvista.com
19  *
20  */
21
22 #include <linux/module.h>
23 #include <linux/init.h>
24 #include <linux/pci.h>
25 #include <linux/pci_ids.h>
26 #include <linux/edac.h>
27 #include "edac_core.h"
28
29 #define E752X_REVISION  " Ver: 2.0.2"
30 #define EDAC_MOD_STR    "e752x_edac"
31
32 static int report_non_memory_errors;
33 static int force_function_unhide;
34 static int sysbus_parity = -1;
35
36 static struct edac_pci_ctl_info *e752x_pci;
37
38 #define e752x_printk(level, fmt, arg...) \
39         edac_printk(level, "e752x", fmt, ##arg)
40
41 #define e752x_mc_printk(mci, level, fmt, arg...) \
42         edac_mc_chipset_printk(mci, level, "e752x", fmt, ##arg)
43
44 #ifndef PCI_DEVICE_ID_INTEL_7520_0
45 #define PCI_DEVICE_ID_INTEL_7520_0      0x3590
46 #endif                          /* PCI_DEVICE_ID_INTEL_7520_0      */
47
48 #ifndef PCI_DEVICE_ID_INTEL_7520_1_ERR
49 #define PCI_DEVICE_ID_INTEL_7520_1_ERR  0x3591
50 #endif                          /* PCI_DEVICE_ID_INTEL_7520_1_ERR  */
51
52 #ifndef PCI_DEVICE_ID_INTEL_7525_0
53 #define PCI_DEVICE_ID_INTEL_7525_0      0x359E
54 #endif                          /* PCI_DEVICE_ID_INTEL_7525_0      */
55
56 #ifndef PCI_DEVICE_ID_INTEL_7525_1_ERR
57 #define PCI_DEVICE_ID_INTEL_7525_1_ERR  0x3593
58 #endif                          /* PCI_DEVICE_ID_INTEL_7525_1_ERR  */
59
60 #ifndef PCI_DEVICE_ID_INTEL_7320_0
61 #define PCI_DEVICE_ID_INTEL_7320_0      0x3592
62 #endif                          /* PCI_DEVICE_ID_INTEL_7320_0 */
63
64 #ifndef PCI_DEVICE_ID_INTEL_7320_1_ERR
65 #define PCI_DEVICE_ID_INTEL_7320_1_ERR  0x3593
66 #endif                          /* PCI_DEVICE_ID_INTEL_7320_1_ERR */
67
68 #ifndef PCI_DEVICE_ID_INTEL_3100_0
69 #define PCI_DEVICE_ID_INTEL_3100_0      0x35B0
70 #endif                          /* PCI_DEVICE_ID_INTEL_3100_0 */
71
72 #ifndef PCI_DEVICE_ID_INTEL_3100_1_ERR
73 #define PCI_DEVICE_ID_INTEL_3100_1_ERR  0x35B1
74 #endif                          /* PCI_DEVICE_ID_INTEL_3100_1_ERR */
75
76 #define E752X_NR_CSROWS         8       /* number of csrows */
77
78 /* E752X register addresses - device 0 function 0 */
79 #define E752X_MCHSCRB           0x52    /* Memory Scrub register (16b) */
80                                         /*
81                                          * 6:5     Scrub Completion Count
82                                          * 3:2     Scrub Rate (i3100 only)
83                                          *      01=fast 10=normal
84                                          * 1:0     Scrub Mode enable
85                                          *      00=off 10=on
86                                          */
87 #define E752X_DRB               0x60    /* DRAM row boundary register (8b) */
88 #define E752X_DRA               0x70    /* DRAM row attribute register (8b) */
89                                         /*
90                                          * 31:30   Device width row 7
91                                          *      01=x8 10=x4 11=x8 DDR2
92                                          * 27:26   Device width row 6
93                                          * 23:22   Device width row 5
94                                          * 19:20   Device width row 4
95                                          * 15:14   Device width row 3
96                                          * 11:10   Device width row 2
97                                          *  7:6    Device width row 1
98                                          *  3:2    Device width row 0
99                                          */
100 #define E752X_DRC               0x7C    /* DRAM controller mode reg (32b) */
101                                         /* FIXME:IS THIS RIGHT? */
102                                         /*
103                                          * 22    Number channels 0=1,1=2
104                                          * 19:18 DRB Granularity 32/64MB
105                                          */
106 #define E752X_DRM               0x80    /* Dimm mapping register */
107 #define E752X_DDRCSR            0x9A    /* DDR control and status reg (16b) */
108                                         /*
109                                          * 14:12 1 single A, 2 single B, 3 dual
110                                          */
111 #define E752X_TOLM              0xC4    /* DRAM top of low memory reg (16b) */
112 #define E752X_REMAPBASE         0xC6    /* DRAM remap base address reg (16b) */
113 #define E752X_REMAPLIMIT        0xC8    /* DRAM remap limit address reg (16b) */
114 #define E752X_REMAPOFFSET       0xCA    /* DRAM remap limit offset reg (16b) */
115
116 /* E752X register addresses - device 0 function 1 */
117 #define E752X_FERR_GLOBAL       0x40    /* Global first error register (32b) */
118 #define E752X_NERR_GLOBAL       0x44    /* Global next error register (32b) */
119 #define E752X_HI_FERR           0x50    /* Hub interface first error reg (8b) */
120 #define E752X_HI_NERR           0x52    /* Hub interface next error reg (8b) */
121 #define E752X_HI_ERRMASK        0x54    /* Hub interface error mask reg (8b) */
122 #define E752X_HI_SMICMD         0x5A    /* Hub interface SMI command reg (8b) */
123 #define E752X_SYSBUS_FERR       0x60    /* System buss first error reg (16b) */
124 #define E752X_SYSBUS_NERR       0x62    /* System buss next error reg (16b) */
125 #define E752X_SYSBUS_ERRMASK    0x64    /* System buss error mask reg (16b) */
126 #define E752X_SYSBUS_SMICMD     0x6A    /* System buss SMI command reg (16b) */
127 #define E752X_BUF_FERR          0x70    /* Memory buffer first error reg (8b) */
128 #define E752X_BUF_NERR          0x72    /* Memory buffer next error reg (8b) */
129 #define E752X_BUF_ERRMASK       0x74    /* Memory buffer error mask reg (8b) */
130 #define E752X_BUF_SMICMD        0x7A    /* Memory buffer SMI cmd reg (8b) */
131 #define E752X_DRAM_FERR         0x80    /* DRAM first error register (16b) */
132 #define E752X_DRAM_NERR         0x82    /* DRAM next error register (16b) */
133 #define E752X_DRAM_ERRMASK      0x84    /* DRAM error mask register (8b) */
134 #define E752X_DRAM_SMICMD       0x8A    /* DRAM SMI command register (8b) */
135 #define E752X_DRAM_RETR_ADD     0xAC    /* DRAM Retry address register (32b) */
136 #define E752X_DRAM_SEC1_ADD     0xA0    /* DRAM first correctable memory */
137                                         /*     error address register (32b) */
138                                         /*
139                                          * 31    Reserved
140                                          * 30:2  CE address (64 byte block 34:6
141                                          * 1     Reserved
142                                          * 0     HiLoCS
143                                          */
144 #define E752X_DRAM_SEC2_ADD     0xC8    /* DRAM first correctable memory */
145                                         /*     error address register (32b) */
146                                         /*
147                                          * 31    Reserved
148                                          * 30:2  CE address (64 byte block 34:6)
149                                          * 1     Reserved
150                                          * 0     HiLoCS
151                                          */
152 #define E752X_DRAM_DED_ADD      0xA4    /* DRAM first uncorrectable memory */
153                                         /*     error address register (32b) */
154                                         /*
155                                          * 31    Reserved
156                                          * 30:2  CE address (64 byte block 34:6)
157                                          * 1     Reserved
158                                          * 0     HiLoCS
159                                          */
160 #define E752X_DRAM_SCRB_ADD     0xA8    /* DRAM 1st uncorrectable scrub mem */
161                                         /*     error address register (32b) */
162                                         /*
163                                          * 31    Reserved
164                                          * 30:2  CE address (64 byte block 34:6
165                                          * 1     Reserved
166                                          * 0     HiLoCS
167                                          */
168 #define E752X_DRAM_SEC1_SYNDROME 0xC4   /* DRAM first correctable memory */
169                                         /*     error syndrome register (16b) */
170 #define E752X_DRAM_SEC2_SYNDROME 0xC6   /* DRAM second correctable memory */
171                                         /*     error syndrome register (16b) */
172 #define E752X_DEVPRES1          0xF4    /* Device Present 1 register (8b) */
173
174 /* 3100 IMCH specific register addresses - device 0 function 1 */
175 #define I3100_NSI_FERR          0x48    /* NSI first error reg (32b) */
176 #define I3100_NSI_NERR          0x4C    /* NSI next error reg (32b) */
177 #define I3100_NSI_SMICMD        0x54    /* NSI SMI command register (32b) */
178 #define I3100_NSI_EMASK         0x90    /* NSI error mask register (32b) */
179
180 /* ICH5R register addresses - device 30 function 0 */
181 #define ICH5R_PCI_STAT          0x06    /* PCI status register (16b) */
182 #define ICH5R_PCI_2ND_STAT      0x1E    /* PCI status secondary reg (16b) */
183 #define ICH5R_PCI_BRIDGE_CTL    0x3E    /* PCI bridge control register (16b) */
184
185 enum e752x_chips {
186         E7520 = 0,
187         E7525 = 1,
188         E7320 = 2,
189         I3100 = 3
190 };
191
192 /*
193  * Those chips Support single-rank and dual-rank memories only.
194  *
195  * On e752x chips, the odd rows are present only on dual-rank memories.
196  * Dividing the rank by two will provide the dimm#
197  *
198  * i3100 MC has a different mapping: it supports only 4 ranks.
199  *
200  * The mapping is (from 1 to n):
201  *      slot       single-ranked        double-ranked
202  *      dimm #1 -> rank #4              NA
203  *      dimm #2 -> rank #3              NA
204  *      dimm #3 -> rank #2              Ranks 2 and 3
205  *      dimm #4 -> rank $1              Ranks 1 and 4
206  *
207  * FIXME: The current mapping for i3100 considers that it supports up to 8
208  *        ranks/chanel, but datasheet says that the MC supports only 4 ranks.
209  */
210
211 struct e752x_pvt {
212         struct pci_dev *bridge_ck;
213         struct pci_dev *dev_d0f0;
214         struct pci_dev *dev_d0f1;
215         u32 tolm;
216         u32 remapbase;
217         u32 remaplimit;
218         int mc_symmetric;
219         u8 map[8];
220         int map_type;
221         const struct e752x_dev_info *dev_info;
222 };
223
224 struct e752x_dev_info {
225         u16 err_dev;
226         u16 ctl_dev;
227         const char *ctl_name;
228 };
229
230 struct e752x_error_info {
231         u32 ferr_global;
232         u32 nerr_global;
233         u32 nsi_ferr;   /* 3100 only */
234         u32 nsi_nerr;   /* 3100 only */
235         u8 hi_ferr;     /* all but 3100 */
236         u8 hi_nerr;     /* all but 3100 */
237         u16 sysbus_ferr;
238         u16 sysbus_nerr;
239         u8 buf_ferr;
240         u8 buf_nerr;
241         u16 dram_ferr;
242         u16 dram_nerr;
243         u32 dram_sec1_add;
244         u32 dram_sec2_add;
245         u16 dram_sec1_syndrome;
246         u16 dram_sec2_syndrome;
247         u32 dram_ded_add;
248         u32 dram_scrb_add;
249         u32 dram_retr_add;
250 };
251
252 static const struct e752x_dev_info e752x_devs[] = {
253         [E7520] = {
254                 .err_dev = PCI_DEVICE_ID_INTEL_7520_1_ERR,
255                 .ctl_dev = PCI_DEVICE_ID_INTEL_7520_0,
256                 .ctl_name = "E7520"},
257         [E7525] = {
258                 .err_dev = PCI_DEVICE_ID_INTEL_7525_1_ERR,
259                 .ctl_dev = PCI_DEVICE_ID_INTEL_7525_0,
260                 .ctl_name = "E7525"},
261         [E7320] = {
262                 .err_dev = PCI_DEVICE_ID_INTEL_7320_1_ERR,
263                 .ctl_dev = PCI_DEVICE_ID_INTEL_7320_0,
264                 .ctl_name = "E7320"},
265         [I3100] = {
266                 .err_dev = PCI_DEVICE_ID_INTEL_3100_1_ERR,
267                 .ctl_dev = PCI_DEVICE_ID_INTEL_3100_0,
268                 .ctl_name = "3100"},
269 };
270
271 /* Valid scrub rates for the e752x/3100 hardware memory scrubber. We
272  * map the scrubbing bandwidth to a hardware register value. The 'set'
273  * operation finds the 'matching or higher value'.  Note that scrubbing
274  * on the e752x can only be enabled/disabled.  The 3100 supports
275  * a normal and fast mode.
276  */
277
278 #define SDRATE_EOT 0xFFFFFFFF
279
280 struct scrubrate {
281         u32 bandwidth;  /* bandwidth consumed by scrubbing in bytes/sec */
282         u16 scrubval;   /* register value for scrub rate */
283 };
284
285 /* Rate below assumes same performance as i3100 using PC3200 DDR2 in
286  * normal mode.  e752x bridges don't support choosing normal or fast mode,
287  * so the scrubbing bandwidth value isn't all that important - scrubbing is
288  * either on or off.
289  */
290 static const struct scrubrate scrubrates_e752x[] = {
291         {0,             0x00},  /* Scrubbing Off */
292         {500000,        0x02},  /* Scrubbing On */
293         {SDRATE_EOT,    0x00}   /* End of Table */
294 };
295
296 /* Fast mode: 2 GByte PC3200 DDR2 scrubbed in 33s = 63161283 bytes/s
297  * Normal mode: 125 (32000 / 256) times slower than fast mode.
298  */
299 static const struct scrubrate scrubrates_i3100[] = {
300         {0,             0x00},  /* Scrubbing Off */
301         {500000,        0x0a},  /* Normal mode - 32k clocks */
302         {62500000,      0x06},  /* Fast mode - 256 clocks */
303         {SDRATE_EOT,    0x00}   /* End of Table */
304 };
305
306 static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci,
307                                 unsigned long page)
308 {
309         u32 remap;
310         struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
311
312         debugf3("%s()\n", __func__);
313
314         if (page < pvt->tolm)
315                 return page;
316
317         if ((page >= 0x100000) && (page < pvt->remapbase))
318                 return page;
319
320         remap = (page - pvt->tolm) + pvt->remapbase;
321
322         if (remap < pvt->remaplimit)
323                 return remap;
324
325         e752x_printk(KERN_ERR, "Invalid page %lx - out of range\n", page);
326         return pvt->tolm - 1;
327 }
328
329 static void do_process_ce(struct mem_ctl_info *mci, u16 error_one,
330                         u32 sec1_add, u16 sec1_syndrome)
331 {
332         u32 page;
333         int row;
334         int channel;
335         int i;
336         struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
337
338         debugf3("%s()\n", __func__);
339
340         /* convert the addr to 4k page */
341         page = sec1_add >> (PAGE_SHIFT - 4);
342
343         /* FIXME - check for -1 */
344         if (pvt->mc_symmetric) {
345                 /* chip select are bits 14 & 13 */
346                 row = ((page >> 1) & 3);
347                 e752x_printk(KERN_WARNING,
348                         "Test row %d Table %d %d %d %d %d %d %d %d\n", row,
349                         pvt->map[0], pvt->map[1], pvt->map[2], pvt->map[3],
350                         pvt->map[4], pvt->map[5], pvt->map[6],
351                         pvt->map[7]);
352
353                 /* test for channel remapping */
354                 for (i = 0; i < 8; i++) {
355                         if (pvt->map[i] == row)
356                                 break;
357                 }
358
359                 e752x_printk(KERN_WARNING, "Test computed row %d\n", i);
360
361                 if (i < 8)
362                         row = i;
363                 else
364                         e752x_mc_printk(mci, KERN_WARNING,
365                                         "row %d not found in remap table\n",
366                                         row);
367         } else
368                 row = edac_mc_find_csrow_by_page(mci, page);
369
370         /* 0 = channel A, 1 = channel B */
371         channel = !(error_one & 1);
372
373         /* e752x mc reads 34:6 of the DRAM linear address */
374         edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci,
375                              page, offset_in_page(sec1_add << 4), sec1_syndrome,
376                              row, channel, -1,
377                              "e752x CE", "", NULL);
378 }
379
380 static inline void process_ce(struct mem_ctl_info *mci, u16 error_one,
381                         u32 sec1_add, u16 sec1_syndrome, int *error_found,
382                         int handle_error)
383 {
384         *error_found = 1;
385
386         if (handle_error)
387                 do_process_ce(mci, error_one, sec1_add, sec1_syndrome);
388 }
389
390 static void do_process_ue(struct mem_ctl_info *mci, u16 error_one,
391                         u32 ded_add, u32 scrb_add)
392 {
393         u32 error_2b, block_page;
394         int row;
395         struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
396
397         debugf3("%s()\n", __func__);
398
399         if (error_one & 0x0202) {
400                 error_2b = ded_add;
401
402                 /* convert to 4k address */
403                 block_page = error_2b >> (PAGE_SHIFT - 4);
404
405                 row = pvt->mc_symmetric ?
406                 /* chip select are bits 14 & 13 */
407                         ((block_page >> 1) & 3) :
408                         edac_mc_find_csrow_by_page(mci, block_page);
409
410                 /* e752x mc reads 34:6 of the DRAM linear address */
411                 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci,
412                                         block_page,
413                                         offset_in_page(error_2b << 4), 0,
414                                          row, -1, -1,
415                                         "e752x UE from Read", "", NULL);
416
417         }
418         if (error_one & 0x0404) {
419                 error_2b = scrb_add;
420
421                 /* convert to 4k address */
422                 block_page = error_2b >> (PAGE_SHIFT - 4);
423
424                 row = pvt->mc_symmetric ?
425                 /* chip select are bits 14 & 13 */
426                         ((block_page >> 1) & 3) :
427                         edac_mc_find_csrow_by_page(mci, block_page);
428
429                 /* e752x mc reads 34:6 of the DRAM linear address */
430                 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci,
431                                         block_page,
432                                         offset_in_page(error_2b << 4), 0,
433                                         row, -1, -1,
434                                         "e752x UE from Scruber", "", NULL);
435         }
436 }
437
438 static inline void process_ue(struct mem_ctl_info *mci, u16 error_one,
439                         u32 ded_add, u32 scrb_add, int *error_found,
440                         int handle_error)
441 {
442         *error_found = 1;
443
444         if (handle_error)
445                 do_process_ue(mci, error_one, ded_add, scrb_add);
446 }
447
448 static inline void process_ue_no_info_wr(struct mem_ctl_info *mci,
449                                          int *error_found, int handle_error)
450 {
451         *error_found = 1;
452
453         if (!handle_error)
454                 return;
455
456         debugf3("%s()\n", __func__);
457         edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 0, 0, 0,
458                              -1, -1, -1,
459                              "e752x UE log memory write", "", NULL);
460 }
461
462 static void do_process_ded_retry(struct mem_ctl_info *mci, u16 error,
463                                  u32 retry_add)
464 {
465         u32 error_1b, page;
466         int row;
467         struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
468
469         error_1b = retry_add;
470         page = error_1b >> (PAGE_SHIFT - 4);  /* convert the addr to 4k page */
471
472         /* chip select are bits 14 & 13 */
473         row = pvt->mc_symmetric ? ((page >> 1) & 3) :
474                 edac_mc_find_csrow_by_page(mci, page);
475
476         e752x_mc_printk(mci, KERN_WARNING,
477                         "CE page 0x%lx, row %d : Memory read retry\n",
478                         (long unsigned int)page, row);
479 }
480
481 static inline void process_ded_retry(struct mem_ctl_info *mci, u16 error,
482                                 u32 retry_add, int *error_found,
483                                 int handle_error)
484 {
485         *error_found = 1;
486
487         if (handle_error)
488                 do_process_ded_retry(mci, error, retry_add);
489 }
490
491 static inline void process_threshold_ce(struct mem_ctl_info *mci, u16 error,
492                                         int *error_found, int handle_error)
493 {
494         *error_found = 1;
495
496         if (handle_error)
497                 e752x_mc_printk(mci, KERN_WARNING, "Memory threshold CE\n");
498 }
499
500 static char *global_message[11] = {
501         "PCI Express C1",
502         "PCI Express C",
503         "PCI Express B1",
504         "PCI Express B",
505         "PCI Express A1",
506         "PCI Express A",
507         "DMA Controller",
508         "HUB or NS Interface",
509         "System Bus",
510         "DRAM Controller",  /* 9th entry */
511         "Internal Buffer"
512 };
513
514 #define DRAM_ENTRY      9
515
516 static char *fatal_message[2] = { "Non-Fatal ", "Fatal " };
517
518 static void do_global_error(int fatal, u32 errors)
519 {
520         int i;
521
522         for (i = 0; i < 11; i++) {
523                 if (errors & (1 << i)) {
524                         /* If the error is from DRAM Controller OR
525                          * we are to report ALL errors, then
526                          * report the error
527                          */
528                         if ((i == DRAM_ENTRY) || report_non_memory_errors)
529                                 e752x_printk(KERN_WARNING, "%sError %s\n",
530                                         fatal_message[fatal],
531                                         global_message[i]);
532                 }
533         }
534 }
535
536 static inline void global_error(int fatal, u32 errors, int *error_found,
537                                 int handle_error)
538 {
539         *error_found = 1;
540
541         if (handle_error)
542                 do_global_error(fatal, errors);
543 }
544
545 static char *hub_message[7] = {
546         "HI Address or Command Parity", "HI Illegal Access",
547         "HI Internal Parity", "Out of Range Access",
548         "HI Data Parity", "Enhanced Config Access",
549         "Hub Interface Target Abort"
550 };
551
552 static void do_hub_error(int fatal, u8 errors)
553 {
554         int i;
555
556         for (i = 0; i < 7; i++) {
557                 if (errors & (1 << i))
558                         e752x_printk(KERN_WARNING, "%sError %s\n",
559                                 fatal_message[fatal], hub_message[i]);
560         }
561 }
562
563 static inline void hub_error(int fatal, u8 errors, int *error_found,
564                         int handle_error)
565 {
566         *error_found = 1;
567
568         if (handle_error)
569                 do_hub_error(fatal, errors);
570 }
571
572 #define NSI_FATAL_MASK          0x0c080081
573 #define NSI_NON_FATAL_MASK      0x23a0ba64
574 #define NSI_ERR_MASK            (NSI_FATAL_MASK | NSI_NON_FATAL_MASK)
575
576 static char *nsi_message[30] = {
577         "NSI Link Down",        /* NSI_FERR/NSI_NERR bit 0, fatal error */
578         "",                                             /* reserved */
579         "NSI Parity Error",                             /* bit 2, non-fatal */
580         "",                                             /* reserved */
581         "",                                             /* reserved */
582         "Correctable Error Message",                    /* bit 5, non-fatal */
583         "Non-Fatal Error Message",                      /* bit 6, non-fatal */
584         "Fatal Error Message",                          /* bit 7, fatal */
585         "",                                             /* reserved */
586         "Receiver Error",                               /* bit 9, non-fatal */
587         "",                                             /* reserved */
588         "Bad TLP",                                      /* bit 11, non-fatal */
589         "Bad DLLP",                                     /* bit 12, non-fatal */
590         "REPLAY_NUM Rollover",                          /* bit 13, non-fatal */
591         "",                                             /* reserved */
592         "Replay Timer Timeout",                         /* bit 15, non-fatal */
593         "",                                             /* reserved */
594         "",                                             /* reserved */
595         "",                                             /* reserved */
596         "Data Link Protocol Error",                     /* bit 19, fatal */
597         "",                                             /* reserved */
598         "Poisoned TLP",                                 /* bit 21, non-fatal */
599         "",                                             /* reserved */
600         "Completion Timeout",                           /* bit 23, non-fatal */
601         "Completer Abort",                              /* bit 24, non-fatal */
602         "Unexpected Completion",                        /* bit 25, non-fatal */
603         "Receiver Overflow",                            /* bit 26, fatal */
604         "Malformed TLP",                                /* bit 27, fatal */
605         "",                                             /* reserved */
606         "Unsupported Request"                           /* bit 29, non-fatal */
607 };
608
609 static void do_nsi_error(int fatal, u32 errors)
610 {
611         int i;
612
613         for (i = 0; i < 30; i++) {
614                 if (errors & (1 << i))
615                         printk(KERN_WARNING "%sError %s\n",
616                                fatal_message[fatal], nsi_message[i]);
617         }
618 }
619
620 static inline void nsi_error(int fatal, u32 errors, int *error_found,
621                 int handle_error)
622 {
623         *error_found = 1;
624
625         if (handle_error)
626                 do_nsi_error(fatal, errors);
627 }
628
629 static char *membuf_message[4] = {
630         "Internal PMWB to DRAM parity",
631         "Internal PMWB to System Bus Parity",
632         "Internal System Bus or IO to PMWB Parity",
633         "Internal DRAM to PMWB Parity"
634 };
635
636 static void do_membuf_error(u8 errors)
637 {
638         int i;
639
640         for (i = 0; i < 4; i++) {
641                 if (errors & (1 << i))
642                         e752x_printk(KERN_WARNING, "Non-Fatal Error %s\n",
643                                 membuf_message[i]);
644         }
645 }
646
647 static inline void membuf_error(u8 errors, int *error_found, int handle_error)
648 {
649         *error_found = 1;
650
651         if (handle_error)
652                 do_membuf_error(errors);
653 }
654
655 static char *sysbus_message[10] = {
656         "Addr or Request Parity",
657         "Data Strobe Glitch",
658         "Addr Strobe Glitch",
659         "Data Parity",
660         "Addr Above TOM",
661         "Non DRAM Lock Error",
662         "MCERR", "BINIT",
663         "Memory Parity",
664         "IO Subsystem Parity"
665 };
666
667 static void do_sysbus_error(int fatal, u32 errors)
668 {
669         int i;
670
671         for (i = 0; i < 10; i++) {
672                 if (errors & (1 << i))
673                         e752x_printk(KERN_WARNING, "%sError System Bus %s\n",
674                                 fatal_message[fatal], sysbus_message[i]);
675         }
676 }
677
678 static inline void sysbus_error(int fatal, u32 errors, int *error_found,
679                                 int handle_error)
680 {
681         *error_found = 1;
682
683         if (handle_error)
684                 do_sysbus_error(fatal, errors);
685 }
686
687 static void e752x_check_hub_interface(struct e752x_error_info *info,
688                                 int *error_found, int handle_error)
689 {
690         u8 stat8;
691
692         //pci_read_config_byte(dev,E752X_HI_FERR,&stat8);
693
694         stat8 = info->hi_ferr;
695
696         if (stat8 & 0x7f) {     /* Error, so process */
697                 stat8 &= 0x7f;
698
699                 if (stat8 & 0x2b)
700                         hub_error(1, stat8 & 0x2b, error_found, handle_error);
701
702                 if (stat8 & 0x54)
703                         hub_error(0, stat8 & 0x54, error_found, handle_error);
704         }
705         //pci_read_config_byte(dev,E752X_HI_NERR,&stat8);
706
707         stat8 = info->hi_nerr;
708
709         if (stat8 & 0x7f) {     /* Error, so process */
710                 stat8 &= 0x7f;
711
712                 if (stat8 & 0x2b)
713                         hub_error(1, stat8 & 0x2b, error_found, handle_error);
714
715                 if (stat8 & 0x54)
716                         hub_error(0, stat8 & 0x54, error_found, handle_error);
717         }
718 }
719
720 static void e752x_check_ns_interface(struct e752x_error_info *info,
721                                 int *error_found, int handle_error)
722 {
723         u32 stat32;
724
725         stat32 = info->nsi_ferr;
726         if (stat32 & NSI_ERR_MASK) { /* Error, so process */
727                 if (stat32 & NSI_FATAL_MASK)    /* check for fatal errors */
728                         nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
729                                   handle_error);
730                 if (stat32 & NSI_NON_FATAL_MASK) /* check for non-fatal ones */
731                         nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
732                                   handle_error);
733         }
734         stat32 = info->nsi_nerr;
735         if (stat32 & NSI_ERR_MASK) {
736                 if (stat32 & NSI_FATAL_MASK)
737                         nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
738                                   handle_error);
739                 if (stat32 & NSI_NON_FATAL_MASK)
740                         nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
741                                   handle_error);
742         }
743 }
744
745 static void e752x_check_sysbus(struct e752x_error_info *info,
746                         int *error_found, int handle_error)
747 {
748         u32 stat32, error32;
749
750         //pci_read_config_dword(dev,E752X_SYSBUS_FERR,&stat32);
751         stat32 = info->sysbus_ferr + (info->sysbus_nerr << 16);
752
753         if (stat32 == 0)
754                 return;         /* no errors */
755
756         error32 = (stat32 >> 16) & 0x3ff;
757         stat32 = stat32 & 0x3ff;
758
759         if (stat32 & 0x087)
760                 sysbus_error(1, stat32 & 0x087, error_found, handle_error);
761
762         if (stat32 & 0x378)
763                 sysbus_error(0, stat32 & 0x378, error_found, handle_error);
764
765         if (error32 & 0x087)
766                 sysbus_error(1, error32 & 0x087, error_found, handle_error);
767
768         if (error32 & 0x378)
769                 sysbus_error(0, error32 & 0x378, error_found, handle_error);
770 }
771
772 static void e752x_check_membuf(struct e752x_error_info *info,
773                         int *error_found, int handle_error)
774 {
775         u8 stat8;
776
777         stat8 = info->buf_ferr;
778
779         if (stat8 & 0x0f) {     /* Error, so process */
780                 stat8 &= 0x0f;
781                 membuf_error(stat8, error_found, handle_error);
782         }
783
784         stat8 = info->buf_nerr;
785
786         if (stat8 & 0x0f) {     /* Error, so process */
787                 stat8 &= 0x0f;
788                 membuf_error(stat8, error_found, handle_error);
789         }
790 }
791
792 static void e752x_check_dram(struct mem_ctl_info *mci,
793                         struct e752x_error_info *info, int *error_found,
794                         int handle_error)
795 {
796         u16 error_one, error_next;
797
798         error_one = info->dram_ferr;
799         error_next = info->dram_nerr;
800
801         /* decode and report errors */
802         if (error_one & 0x0101) /* check first error correctable */
803                 process_ce(mci, error_one, info->dram_sec1_add,
804                         info->dram_sec1_syndrome, error_found, handle_error);
805
806         if (error_next & 0x0101)        /* check next error correctable */
807                 process_ce(mci, error_next, info->dram_sec2_add,
808                         info->dram_sec2_syndrome, error_found, handle_error);
809
810         if (error_one & 0x4040)
811                 process_ue_no_info_wr(mci, error_found, handle_error);
812
813         if (error_next & 0x4040)
814                 process_ue_no_info_wr(mci, error_found, handle_error);
815
816         if (error_one & 0x2020)
817                 process_ded_retry(mci, error_one, info->dram_retr_add,
818                                 error_found, handle_error);
819
820         if (error_next & 0x2020)
821                 process_ded_retry(mci, error_next, info->dram_retr_add,
822                                 error_found, handle_error);
823
824         if (error_one & 0x0808)
825                 process_threshold_ce(mci, error_one, error_found, handle_error);
826
827         if (error_next & 0x0808)
828                 process_threshold_ce(mci, error_next, error_found,
829                                 handle_error);
830
831         if (error_one & 0x0606)
832                 process_ue(mci, error_one, info->dram_ded_add,
833                         info->dram_scrb_add, error_found, handle_error);
834
835         if (error_next & 0x0606)
836                 process_ue(mci, error_next, info->dram_ded_add,
837                         info->dram_scrb_add, error_found, handle_error);
838 }
839
840 static void e752x_get_error_info(struct mem_ctl_info *mci,
841                                  struct e752x_error_info *info)
842 {
843         struct pci_dev *dev;
844         struct e752x_pvt *pvt;
845
846         memset(info, 0, sizeof(*info));
847         pvt = (struct e752x_pvt *)mci->pvt_info;
848         dev = pvt->dev_d0f1;
849         pci_read_config_dword(dev, E752X_FERR_GLOBAL, &info->ferr_global);
850
851         if (info->ferr_global) {
852                 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
853                         pci_read_config_dword(dev, I3100_NSI_FERR,
854                                              &info->nsi_ferr);
855                         info->hi_ferr = 0;
856                 } else {
857                         pci_read_config_byte(dev, E752X_HI_FERR,
858                                              &info->hi_ferr);
859                         info->nsi_ferr = 0;
860                 }
861                 pci_read_config_word(dev, E752X_SYSBUS_FERR,
862                                 &info->sysbus_ferr);
863                 pci_read_config_byte(dev, E752X_BUF_FERR, &info->buf_ferr);
864                 pci_read_config_word(dev, E752X_DRAM_FERR, &info->dram_ferr);
865                 pci_read_config_dword(dev, E752X_DRAM_SEC1_ADD,
866                                 &info->dram_sec1_add);
867                 pci_read_config_word(dev, E752X_DRAM_SEC1_SYNDROME,
868                                 &info->dram_sec1_syndrome);
869                 pci_read_config_dword(dev, E752X_DRAM_DED_ADD,
870                                 &info->dram_ded_add);
871                 pci_read_config_dword(dev, E752X_DRAM_SCRB_ADD,
872                                 &info->dram_scrb_add);
873                 pci_read_config_dword(dev, E752X_DRAM_RETR_ADD,
874                                 &info->dram_retr_add);
875
876                 /* ignore the reserved bits just in case */
877                 if (info->hi_ferr & 0x7f)
878                         pci_write_config_byte(dev, E752X_HI_FERR,
879                                         info->hi_ferr);
880
881                 if (info->nsi_ferr & NSI_ERR_MASK)
882                         pci_write_config_dword(dev, I3100_NSI_FERR,
883                                         info->nsi_ferr);
884
885                 if (info->sysbus_ferr)
886                         pci_write_config_word(dev, E752X_SYSBUS_FERR,
887                                         info->sysbus_ferr);
888
889                 if (info->buf_ferr & 0x0f)
890                         pci_write_config_byte(dev, E752X_BUF_FERR,
891                                         info->buf_ferr);
892
893                 if (info->dram_ferr)
894                         pci_write_bits16(pvt->bridge_ck, E752X_DRAM_FERR,
895                                          info->dram_ferr, info->dram_ferr);
896
897                 pci_write_config_dword(dev, E752X_FERR_GLOBAL,
898                                 info->ferr_global);
899         }
900
901         pci_read_config_dword(dev, E752X_NERR_GLOBAL, &info->nerr_global);
902
903         if (info->nerr_global) {
904                 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
905                         pci_read_config_dword(dev, I3100_NSI_NERR,
906                                              &info->nsi_nerr);
907                         info->hi_nerr = 0;
908                 } else {
909                         pci_read_config_byte(dev, E752X_HI_NERR,
910                                              &info->hi_nerr);
911                         info->nsi_nerr = 0;
912                 }
913                 pci_read_config_word(dev, E752X_SYSBUS_NERR,
914                                 &info->sysbus_nerr);
915                 pci_read_config_byte(dev, E752X_BUF_NERR, &info->buf_nerr);
916                 pci_read_config_word(dev, E752X_DRAM_NERR, &info->dram_nerr);
917                 pci_read_config_dword(dev, E752X_DRAM_SEC2_ADD,
918                                 &info->dram_sec2_add);
919                 pci_read_config_word(dev, E752X_DRAM_SEC2_SYNDROME,
920                                 &info->dram_sec2_syndrome);
921
922                 if (info->hi_nerr & 0x7f)
923                         pci_write_config_byte(dev, E752X_HI_NERR,
924                                         info->hi_nerr);
925
926                 if (info->nsi_nerr & NSI_ERR_MASK)
927                         pci_write_config_dword(dev, I3100_NSI_NERR,
928                                         info->nsi_nerr);
929
930                 if (info->sysbus_nerr)
931                         pci_write_config_word(dev, E752X_SYSBUS_NERR,
932                                         info->sysbus_nerr);
933
934                 if (info->buf_nerr & 0x0f)
935                         pci_write_config_byte(dev, E752X_BUF_NERR,
936                                         info->buf_nerr);
937
938                 if (info->dram_nerr)
939                         pci_write_bits16(pvt->bridge_ck, E752X_DRAM_NERR,
940                                          info->dram_nerr, info->dram_nerr);
941
942                 pci_write_config_dword(dev, E752X_NERR_GLOBAL,
943                                 info->nerr_global);
944         }
945 }
946
947 static int e752x_process_error_info(struct mem_ctl_info *mci,
948                                 struct e752x_error_info *info,
949                                 int handle_errors)
950 {
951         u32 error32, stat32;
952         int error_found;
953
954         error_found = 0;
955         error32 = (info->ferr_global >> 18) & 0x3ff;
956         stat32 = (info->ferr_global >> 4) & 0x7ff;
957
958         if (error32)
959                 global_error(1, error32, &error_found, handle_errors);
960
961         if (stat32)
962                 global_error(0, stat32, &error_found, handle_errors);
963
964         error32 = (info->nerr_global >> 18) & 0x3ff;
965         stat32 = (info->nerr_global >> 4) & 0x7ff;
966
967         if (error32)
968                 global_error(1, error32, &error_found, handle_errors);
969
970         if (stat32)
971                 global_error(0, stat32, &error_found, handle_errors);
972
973         e752x_check_hub_interface(info, &error_found, handle_errors);
974         e752x_check_ns_interface(info, &error_found, handle_errors);
975         e752x_check_sysbus(info, &error_found, handle_errors);
976         e752x_check_membuf(info, &error_found, handle_errors);
977         e752x_check_dram(mci, info, &error_found, handle_errors);
978         return error_found;
979 }
980
981 static void e752x_check(struct mem_ctl_info *mci)
982 {
983         struct e752x_error_info info;
984
985         debugf3("%s()\n", __func__);
986         e752x_get_error_info(mci, &info);
987         e752x_process_error_info(mci, &info, 1);
988 }
989
990 /* Program byte/sec bandwidth scrub rate to hardware */
991 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
992 {
993         const struct scrubrate *scrubrates;
994         struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
995         struct pci_dev *pdev = pvt->dev_d0f0;
996         int i;
997
998         if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
999                 scrubrates = scrubrates_i3100;
1000         else
1001                 scrubrates = scrubrates_e752x;
1002
1003         /* Translate the desired scrub rate to a e752x/3100 register value.
1004          * Search for the bandwidth that is equal or greater than the
1005          * desired rate and program the cooresponding register value.
1006          */
1007         for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1008                 if (scrubrates[i].bandwidth >= new_bw)
1009                         break;
1010
1011         if (scrubrates[i].bandwidth == SDRATE_EOT)
1012                 return -1;
1013
1014         pci_write_config_word(pdev, E752X_MCHSCRB, scrubrates[i].scrubval);
1015
1016         return scrubrates[i].bandwidth;
1017 }
1018
1019 /* Convert current scrub rate value into byte/sec bandwidth */
1020 static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
1021 {
1022         const struct scrubrate *scrubrates;
1023         struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
1024         struct pci_dev *pdev = pvt->dev_d0f0;
1025         u16 scrubval;
1026         int i;
1027
1028         if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
1029                 scrubrates = scrubrates_i3100;
1030         else
1031                 scrubrates = scrubrates_e752x;
1032
1033         /* Find the bandwidth matching the memory scrubber configuration */
1034         pci_read_config_word(pdev, E752X_MCHSCRB, &scrubval);
1035         scrubval = scrubval & 0x0f;
1036
1037         for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1038                 if (scrubrates[i].scrubval == scrubval)
1039                         break;
1040
1041         if (scrubrates[i].bandwidth == SDRATE_EOT) {
1042                 e752x_printk(KERN_WARNING,
1043                         "Invalid sdram scrub control value: 0x%x\n", scrubval);
1044                 return -1;
1045         }
1046         return scrubrates[i].bandwidth;
1047
1048 }
1049
1050 /* Return 1 if dual channel mode is active.  Else return 0. */
1051 static inline int dual_channel_active(u16 ddrcsr)
1052 {
1053         return (((ddrcsr >> 12) & 3) == 3);
1054 }
1055
1056 /* Remap csrow index numbers if map_type is "reverse"
1057  */
1058 static inline int remap_csrow_index(struct mem_ctl_info *mci, int index)
1059 {
1060         struct e752x_pvt *pvt = mci->pvt_info;
1061
1062         if (!pvt->map_type)
1063                 return (7 - index);
1064
1065         return (index);
1066 }
1067
1068 static void e752x_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev,
1069                         u16 ddrcsr)
1070 {
1071         struct csrow_info *csrow;
1072         unsigned long last_cumul_size;
1073         int index, mem_dev, drc_chan;
1074         int drc_drbg;           /* DRB granularity 0=64mb, 1=128mb */
1075         int drc_ddim;           /* DRAM Data Integrity Mode 0=none, 2=edac */
1076         u8 value;
1077         u32 dra, drc, cumul_size, i, nr_pages;
1078
1079         dra = 0;
1080         for (index = 0; index < 4; index++) {
1081                 u8 dra_reg;
1082                 pci_read_config_byte(pdev, E752X_DRA + index, &dra_reg);
1083                 dra |= dra_reg << (index * 8);
1084         }
1085         pci_read_config_dword(pdev, E752X_DRC, &drc);
1086         drc_chan = dual_channel_active(ddrcsr) ? 1 : 0;
1087         drc_drbg = drc_chan + 1;        /* 128 in dual mode, 64 in single */
1088         drc_ddim = (drc >> 20) & 0x3;
1089
1090         /* The dram row boundary (DRB) reg values are boundary address for
1091          * each DRAM row with a granularity of 64 or 128MB (single/dual
1092          * channel operation).  DRB regs are cumulative; therefore DRB7 will
1093          * contain the total memory contained in all eight rows.
1094          */
1095         for (last_cumul_size = index = 0; index < mci->nr_csrows; index++) {
1096                 /* mem_dev 0=x8, 1=x4 */
1097                 mem_dev = (dra >> (index * 4 + 2)) & 0x3;
1098                 csrow = &mci->csrows[remap_csrow_index(mci, index)];
1099
1100                 mem_dev = (mem_dev == 2);
1101                 pci_read_config_byte(pdev, E752X_DRB + index, &value);
1102                 /* convert a 128 or 64 MiB DRB to a page size. */
1103                 cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
1104                 debugf3("%s(): (%d) cumul_size 0x%x\n", __func__, index,
1105                         cumul_size);
1106                 if (cumul_size == last_cumul_size)
1107                         continue;       /* not populated */
1108
1109                 csrow->first_page = last_cumul_size;
1110                 csrow->last_page = cumul_size - 1;
1111                 nr_pages = cumul_size - last_cumul_size;
1112                 last_cumul_size = cumul_size;
1113
1114                 for (i = 0; i < csrow->nr_channels; i++) {
1115                         struct dimm_info *dimm = csrow->channels[i].dimm;
1116
1117                         debugf3("Initializing rank at (%i,%i)\n", index, i);
1118                         dimm->nr_pages = nr_pages / csrow->nr_channels;
1119                         dimm->grain = 1 << 12;  /* 4KiB - resolution of CELOG */
1120                         dimm->mtype = MEM_RDDR; /* only one type supported */
1121                         dimm->dtype = mem_dev ? DEV_X4 : DEV_X8;
1122
1123                         /*
1124                         * if single channel or x8 devices then SECDED
1125                         * if dual channel and x4 then S4ECD4ED
1126                         */
1127                         if (drc_ddim) {
1128                                 if (drc_chan && mem_dev) {
1129                                         dimm->edac_mode = EDAC_S4ECD4ED;
1130                                         mci->edac_cap |= EDAC_FLAG_S4ECD4ED;
1131                                 } else {
1132                                         dimm->edac_mode = EDAC_SECDED;
1133                                         mci->edac_cap |= EDAC_FLAG_SECDED;
1134                                 }
1135                         } else
1136                                 dimm->edac_mode = EDAC_NONE;
1137                 }
1138         }
1139 }
1140
1141 static void e752x_init_mem_map_table(struct pci_dev *pdev,
1142                                 struct e752x_pvt *pvt)
1143 {
1144         int index;
1145         u8 value, last, row;
1146
1147         last = 0;
1148         row = 0;
1149
1150         for (index = 0; index < 8; index += 2) {
1151                 pci_read_config_byte(pdev, E752X_DRB + index, &value);
1152                 /* test if there is a dimm in this slot */
1153                 if (value == last) {
1154                         /* no dimm in the slot, so flag it as empty */
1155                         pvt->map[index] = 0xff;
1156                         pvt->map[index + 1] = 0xff;
1157                 } else {        /* there is a dimm in the slot */
1158                         pvt->map[index] = row;
1159                         row++;
1160                         last = value;
1161                         /* test the next value to see if the dimm is double
1162                          * sided
1163                          */
1164                         pci_read_config_byte(pdev, E752X_DRB + index + 1,
1165                                         &value);
1166
1167                         /* the dimm is single sided, so flag as empty */
1168                         /* this is a double sided dimm to save the next row #*/
1169                         pvt->map[index + 1] = (value == last) ? 0xff :  row;
1170                         row++;
1171                         last = value;
1172                 }
1173         }
1174 }
1175
1176 /* Return 0 on success or 1 on failure. */
1177 static int e752x_get_devs(struct pci_dev *pdev, int dev_idx,
1178                         struct e752x_pvt *pvt)
1179 {
1180         struct pci_dev *dev;
1181
1182         pvt->bridge_ck = pci_get_device(PCI_VENDOR_ID_INTEL,
1183                                 pvt->dev_info->err_dev, pvt->bridge_ck);
1184
1185         if (pvt->bridge_ck == NULL)
1186                 pvt->bridge_ck = pci_scan_single_device(pdev->bus,
1187                                                         PCI_DEVFN(0, 1));
1188
1189         if (pvt->bridge_ck == NULL) {
1190                 e752x_printk(KERN_ERR, "error reporting device not found:"
1191                         "vendor %x device 0x%x (broken BIOS?)\n",
1192                         PCI_VENDOR_ID_INTEL, e752x_devs[dev_idx].err_dev);
1193                 return 1;
1194         }
1195
1196         dev = pci_get_device(PCI_VENDOR_ID_INTEL,
1197                                 e752x_devs[dev_idx].ctl_dev,
1198                                 NULL);
1199
1200         if (dev == NULL)
1201                 goto fail;
1202
1203         pvt->dev_d0f0 = dev;
1204         pvt->dev_d0f1 = pci_dev_get(pvt->bridge_ck);
1205
1206         return 0;
1207
1208 fail:
1209         pci_dev_put(pvt->bridge_ck);
1210         return 1;
1211 }
1212
1213 /* Setup system bus parity mask register.
1214  * Sysbus parity supported on:
1215  * e7320/e7520/e7525 + Xeon
1216  */
1217 static void e752x_init_sysbus_parity_mask(struct e752x_pvt *pvt)
1218 {
1219         char *cpu_id = cpu_data(0).x86_model_id;
1220         struct pci_dev *dev = pvt->dev_d0f1;
1221         int enable = 1;
1222
1223         /* Allow module parameter override, else see if CPU supports parity */
1224         if (sysbus_parity != -1) {
1225                 enable = sysbus_parity;
1226         } else if (cpu_id[0] && !strstr(cpu_id, "Xeon")) {
1227                 e752x_printk(KERN_INFO, "System Bus Parity not "
1228                              "supported by CPU, disabling\n");
1229                 enable = 0;
1230         }
1231
1232         if (enable)
1233                 pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0000);
1234         else
1235                 pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0309);
1236 }
1237
1238 static void e752x_init_error_reporting_regs(struct e752x_pvt *pvt)
1239 {
1240         struct pci_dev *dev;
1241
1242         dev = pvt->dev_d0f1;
1243         /* Turn off error disable & SMI in case the BIOS turned it on */
1244         if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
1245                 pci_write_config_dword(dev, I3100_NSI_EMASK, 0);
1246                 pci_write_config_dword(dev, I3100_NSI_SMICMD, 0);
1247         } else {
1248                 pci_write_config_byte(dev, E752X_HI_ERRMASK, 0x00);
1249                 pci_write_config_byte(dev, E752X_HI_SMICMD, 0x00);
1250         }
1251
1252         e752x_init_sysbus_parity_mask(pvt);
1253
1254         pci_write_config_word(dev, E752X_SYSBUS_SMICMD, 0x00);
1255         pci_write_config_byte(dev, E752X_BUF_ERRMASK, 0x00);
1256         pci_write_config_byte(dev, E752X_BUF_SMICMD, 0x00);
1257         pci_write_config_byte(dev, E752X_DRAM_ERRMASK, 0x00);
1258         pci_write_config_byte(dev, E752X_DRAM_SMICMD, 0x00);
1259 }
1260
1261 static int e752x_probe1(struct pci_dev *pdev, int dev_idx)
1262 {
1263         u16 pci_data;
1264         u8 stat8;
1265         struct mem_ctl_info *mci;
1266         struct edac_mc_layer layers[2];
1267         struct e752x_pvt *pvt;
1268         u16 ddrcsr;
1269         int drc_chan;           /* Number of channels 0=1chan,1=2chan */
1270         struct e752x_error_info discard;
1271
1272         debugf0("%s(): mci\n", __func__);
1273         debugf0("Starting Probe1\n");
1274
1275         /* check to see if device 0 function 1 is enabled; if it isn't, we
1276          * assume the BIOS has reserved it for a reason and is expecting
1277          * exclusive access, we take care not to violate that assumption and
1278          * fail the probe. */
1279         pci_read_config_byte(pdev, E752X_DEVPRES1, &stat8);
1280         if (!force_function_unhide && !(stat8 & (1 << 5))) {
1281                 printk(KERN_INFO "Contact your BIOS vendor to see if the "
1282                         "E752x error registers can be safely un-hidden\n");
1283                 return -ENODEV;
1284         }
1285         stat8 |= (1 << 5);
1286         pci_write_config_byte(pdev, E752X_DEVPRES1, stat8);
1287
1288         pci_read_config_word(pdev, E752X_DDRCSR, &ddrcsr);
1289         /* FIXME: should check >>12 or 0xf, true for all? */
1290         /* Dual channel = 1, Single channel = 0 */
1291         drc_chan = dual_channel_active(ddrcsr);
1292
1293         layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
1294         layers[0].size = E752X_NR_CSROWS;
1295         layers[0].is_virt_csrow = true;
1296         layers[1].type = EDAC_MC_LAYER_CHANNEL;
1297         layers[1].size = drc_chan + 1;
1298         layers[1].is_virt_csrow = false;
1299         mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt));
1300         if (mci == NULL)
1301                 return -ENOMEM;
1302
1303         debugf3("%s(): init mci\n", __func__);
1304         mci->mtype_cap = MEM_FLAG_RDDR;
1305         /* 3100 IMCH supports SECDEC only */
1306         mci->edac_ctl_cap = (dev_idx == I3100) ? EDAC_FLAG_SECDED :
1307                 (EDAC_FLAG_NONE | EDAC_FLAG_SECDED | EDAC_FLAG_S4ECD4ED);
1308         /* FIXME - what if different memory types are in different csrows? */
1309         mci->mod_name = EDAC_MOD_STR;
1310         mci->mod_ver = E752X_REVISION;
1311         mci->pdev = &pdev->dev;
1312
1313         debugf3("%s(): init pvt\n", __func__);
1314         pvt = (struct e752x_pvt *)mci->pvt_info;
1315         pvt->dev_info = &e752x_devs[dev_idx];
1316         pvt->mc_symmetric = ((ddrcsr & 0x10) != 0);
1317
1318         if (e752x_get_devs(pdev, dev_idx, pvt)) {
1319                 edac_mc_free(mci);
1320                 return -ENODEV;
1321         }
1322
1323         debugf3("%s(): more mci init\n", __func__);
1324         mci->ctl_name = pvt->dev_info->ctl_name;
1325         mci->dev_name = pci_name(pdev);
1326         mci->edac_check = e752x_check;
1327         mci->ctl_page_to_phys = ctl_page_to_phys;
1328         mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
1329         mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
1330
1331         /* set the map type.  1 = normal, 0 = reversed
1332          * Must be set before e752x_init_csrows in case csrow mapping
1333          * is reversed.
1334          */
1335         pci_read_config_byte(pdev, E752X_DRM, &stat8);
1336         pvt->map_type = ((stat8 & 0x0f) > ((stat8 >> 4) & 0x0f));
1337
1338         e752x_init_csrows(mci, pdev, ddrcsr);
1339         e752x_init_mem_map_table(pdev, pvt);
1340
1341         if (dev_idx == I3100)
1342                 mci->edac_cap = EDAC_FLAG_SECDED; /* the only mode supported */
1343         else
1344                 mci->edac_cap |= EDAC_FLAG_NONE;
1345         debugf3("%s(): tolm, remapbase, remaplimit\n", __func__);
1346
1347         /* load the top of low memory, remap base, and remap limit vars */
1348         pci_read_config_word(pdev, E752X_TOLM, &pci_data);
1349         pvt->tolm = ((u32) pci_data) << 4;
1350         pci_read_config_word(pdev, E752X_REMAPBASE, &pci_data);
1351         pvt->remapbase = ((u32) pci_data) << 14;
1352         pci_read_config_word(pdev, E752X_REMAPLIMIT, &pci_data);
1353         pvt->remaplimit = ((u32) pci_data) << 14;
1354         e752x_printk(KERN_INFO,
1355                         "tolm = %x, remapbase = %x, remaplimit = %x\n",
1356                         pvt->tolm, pvt->remapbase, pvt->remaplimit);
1357
1358         /* Here we assume that we will never see multiple instances of this
1359          * type of memory controller.  The ID is therefore hardcoded to 0.
1360          */
1361         if (edac_mc_add_mc(mci)) {
1362                 debugf3("%s(): failed edac_mc_add_mc()\n", __func__);
1363                 goto fail;
1364         }
1365
1366         e752x_init_error_reporting_regs(pvt);
1367         e752x_get_error_info(mci, &discard);    /* clear other MCH errors */
1368
1369         /* allocating generic PCI control info */
1370         e752x_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
1371         if (!e752x_pci) {
1372                 printk(KERN_WARNING
1373                         "%s(): Unable to create PCI control\n", __func__);
1374                 printk(KERN_WARNING
1375                         "%s(): PCI error report via EDAC not setup\n",
1376                         __func__);
1377         }
1378
1379         /* get this far and it's successful */
1380         debugf3("%s(): success\n", __func__);
1381         return 0;
1382
1383 fail:
1384         pci_dev_put(pvt->dev_d0f0);
1385         pci_dev_put(pvt->dev_d0f1);
1386         pci_dev_put(pvt->bridge_ck);
1387         edac_mc_free(mci);
1388
1389         return -ENODEV;
1390 }
1391
1392 /* returns count (>= 0), or negative on error */
1393 static int __devinit e752x_init_one(struct pci_dev *pdev,
1394                                 const struct pci_device_id *ent)
1395 {
1396         debugf0("%s()\n", __func__);
1397
1398         /* wake up and enable device */
1399         if (pci_enable_device(pdev) < 0)
1400                 return -EIO;
1401
1402         return e752x_probe1(pdev, ent->driver_data);
1403 }
1404
1405 static void __devexit e752x_remove_one(struct pci_dev *pdev)
1406 {
1407         struct mem_ctl_info *mci;
1408         struct e752x_pvt *pvt;
1409
1410         debugf0("%s()\n", __func__);
1411
1412         if (e752x_pci)
1413                 edac_pci_release_generic_ctl(e752x_pci);
1414
1415         if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL)
1416                 return;
1417
1418         pvt = (struct e752x_pvt *)mci->pvt_info;
1419         pci_dev_put(pvt->dev_d0f0);
1420         pci_dev_put(pvt->dev_d0f1);
1421         pci_dev_put(pvt->bridge_ck);
1422         edac_mc_free(mci);
1423 }
1424
1425 static DEFINE_PCI_DEVICE_TABLE(e752x_pci_tbl) = {
1426         {
1427          PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1428          E7520},
1429         {
1430          PCI_VEND_DEV(INTEL, 7525_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1431          E7525},
1432         {
1433          PCI_VEND_DEV(INTEL, 7320_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1434          E7320},
1435         {
1436          PCI_VEND_DEV(INTEL, 3100_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1437          I3100},
1438         {
1439          0,
1440          }                      /* 0 terminated list. */
1441 };
1442
1443 MODULE_DEVICE_TABLE(pci, e752x_pci_tbl);
1444
1445 static struct pci_driver e752x_driver = {
1446         .name = EDAC_MOD_STR,
1447         .probe = e752x_init_one,
1448         .remove = __devexit_p(e752x_remove_one),
1449         .id_table = e752x_pci_tbl,
1450 };
1451
1452 static int __init e752x_init(void)
1453 {
1454         int pci_rc;
1455
1456         debugf3("%s()\n", __func__);
1457
1458        /* Ensure that the OPSTATE is set correctly for POLL or NMI */
1459        opstate_init();
1460
1461         pci_rc = pci_register_driver(&e752x_driver);
1462         return (pci_rc < 0) ? pci_rc : 0;
1463 }
1464
1465 static void __exit e752x_exit(void)
1466 {
1467         debugf3("%s()\n", __func__);
1468         pci_unregister_driver(&e752x_driver);
1469 }
1470
1471 module_init(e752x_init);
1472 module_exit(e752x_exit);
1473
1474 MODULE_LICENSE("GPL");
1475 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Tom Zimmerman\n");
1476 MODULE_DESCRIPTION("MC support for Intel e752x/3100 memory controllers");
1477
1478 module_param(force_function_unhide, int, 0444);
1479 MODULE_PARM_DESC(force_function_unhide, "if BIOS sets Dev0:Fun1 up as hidden:"
1480                  " 1=force unhide and hope BIOS doesn't fight driver for "
1481                 "Dev0:Fun1 access");
1482
1483 module_param(edac_op_state, int, 0444);
1484 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
1485
1486 module_param(sysbus_parity, int, 0444);
1487 MODULE_PARM_DESC(sysbus_parity, "0=disable system bus parity checking,"
1488                 " 1=enable system bus parity checking, default=auto-detect");
1489 module_param(report_non_memory_errors, int, 0644);
1490 MODULE_PARM_DESC(report_non_memory_errors, "0=disable non-memory error "
1491                 "reporting, 1=enable non-memory error reporting");