1 /*
2 * Intel e752x Memory Controller kernel module
3 * (C) 2004 Linux Networx (http://lnxi.com)
4 * This file may be distributed under the terms of the
5 * GNU General Public License.
6 *
7 * Implement support for the e7520, E7525, e7320 and i3100 memory controllers.
8 *
9 * Datasheets:
10 * http://www.intel.in/content/www/in/en/chipsets/e7525-memory-controller-hub-datasheet.html
11 * ftp://download.intel.com/design/intarch/datashts/31345803.pdf
12 *
13 * Written by Tom Zimmerman
14 *
15 * Contributors:
16 * Thayne Harbaugh at realmsys.com (?)
17 * Wang Zhenyu at intel.com
18 * Dave Jiang at mvista.com
19 *
20 */
21
22 #include <linux/module.h>
23 #include <linux/init.h>
24 #include <linux/pci.h>
25 #include <linux/pci_ids.h>
26 #include <linux/edac.h>
27 #include "edac_module.h"
28
29 #define EDAC_MOD_STR "e752x_edac"
30
31 static int report_non_memory_errors;
32 static int force_function_unhide;
33 static int sysbus_parity = -1;
34
35 static struct edac_pci_ctl_info *e752x_pci;
36
37 #define e752x_printk(level, fmt, arg...) \
38 edac_printk(level, "e752x", fmt, ##arg)
39
40 #define e752x_mc_printk(mci, level, fmt, arg...) \
41 edac_mc_chipset_printk(mci, level, "e752x", fmt, ##arg)
42
43 #ifndef PCI_DEVICE_ID_INTEL_7520_0
44 #define PCI_DEVICE_ID_INTEL_7520_0 0x3590
45 #endif /* PCI_DEVICE_ID_INTEL_7520_0 */
46
47 #ifndef PCI_DEVICE_ID_INTEL_7520_1_ERR
48 #define PCI_DEVICE_ID_INTEL_7520_1_ERR 0x3591
49 #endif /* PCI_DEVICE_ID_INTEL_7520_1_ERR */
50
51 #ifndef PCI_DEVICE_ID_INTEL_7525_0
52 #define PCI_DEVICE_ID_INTEL_7525_0 0x359E
53 #endif /* PCI_DEVICE_ID_INTEL_7525_0 */
54
55 #ifndef PCI_DEVICE_ID_INTEL_7525_1_ERR
56 #define PCI_DEVICE_ID_INTEL_7525_1_ERR 0x3593
57 #endif /* PCI_DEVICE_ID_INTEL_7525_1_ERR */
58
59 #ifndef PCI_DEVICE_ID_INTEL_7320_0
60 #define PCI_DEVICE_ID_INTEL_7320_0 0x3592
61 #endif /* PCI_DEVICE_ID_INTEL_7320_0 */
62
63 #ifndef PCI_DEVICE_ID_INTEL_7320_1_ERR
64 #define PCI_DEVICE_ID_INTEL_7320_1_ERR 0x3593
65 #endif /* PCI_DEVICE_ID_INTEL_7320_1_ERR */
66
67 #ifndef PCI_DEVICE_ID_INTEL_3100_0
68 #define PCI_DEVICE_ID_INTEL_3100_0 0x35B0
69 #endif /* PCI_DEVICE_ID_INTEL_3100_0 */
70
71 #ifndef PCI_DEVICE_ID_INTEL_3100_1_ERR
72 #define PCI_DEVICE_ID_INTEL_3100_1_ERR 0x35B1
73 #endif /* PCI_DEVICE_ID_INTEL_3100_1_ERR */
74
75 #define E752X_NR_CSROWS 8 /* number of csrows */
76
77 /* E752X register addresses - device 0 function 0 */
78 #define E752X_MCHSCRB 0x52 /* Memory Scrub register (16b) */
79 /*
80 * 6:5 Scrub Completion Count
81 * 3:2 Scrub Rate (i3100 only)
82 * 01=fast 10=normal
83 * 1:0 Scrub Mode enable
84 * 00=off 10=on
85 */
86 #define E752X_DRB 0x60 /* DRAM row boundary register (8b) */
87 #define E752X_DRA 0x70 /* DRAM row attribute register (8b) */
88 /*
89 * 31:30 Device width row 7
90 * 01=x8 10=x4 11=x8 DDR2
91 * 27:26 Device width row 6
92 * 23:22 Device width row 5
93 * 19:20 Device width row 4
94 * 15:14 Device width row 3
95 * 11:10 Device width row 2
96 * 7:6 Device width row 1
97 * 3:2 Device width row 0
98 */
99 #define E752X_DRC 0x7C /* DRAM controller mode reg (32b) */
100 /* FIXME:IS THIS RIGHT? */
101 /*
102 * 22 Number channels 0=1,1=2
103 * 19:18 DRB Granularity 32/64MB
104 */
105 #define E752X_DRM 0x80 /* Dimm mapping register */
106 #define E752X_DDRCSR 0x9A /* DDR control and status reg (16b) */
107 /*
108 * 14:12 1 single A, 2 single B, 3 dual
109 */
110 #define E752X_TOLM 0xC4 /* DRAM top of low memory reg (16b) */
111 #define E752X_REMAPBASE 0xC6 /* DRAM remap base address reg (16b) */
112 #define E752X_REMAPLIMIT 0xC8 /* DRAM remap limit address reg (16b) */
113 #define E752X_REMAPOFFSET 0xCA /* DRAM remap limit offset reg (16b) */
114
115 /* E752X register addresses - device 0 function 1 */
116 #define E752X_FERR_GLOBAL 0x40 /* Global first error register (32b) */
117 #define E752X_NERR_GLOBAL 0x44 /* Global next error register (32b) */
118 #define E752X_HI_FERR 0x50 /* Hub interface first error reg (8b) */
119 #define E752X_HI_NERR 0x52 /* Hub interface next error reg (8b) */
120 #define E752X_HI_ERRMASK 0x54 /* Hub interface error mask reg (8b) */
121 #define E752X_HI_SMICMD 0x5A /* Hub interface SMI command reg (8b) */
122 #define E752X_SYSBUS_FERR 0x60 /* System buss first error reg (16b) */
123 #define E752X_SYSBUS_NERR 0x62 /* System buss next error reg (16b) */
124 #define E752X_SYSBUS_ERRMASK 0x64 /* System buss error mask reg (16b) */
125 #define E752X_SYSBUS_SMICMD 0x6A /* System buss SMI command reg (16b) */
126 #define E752X_BUF_FERR 0x70 /* Memory buffer first error reg (8b) */
127 #define E752X_BUF_NERR 0x72 /* Memory buffer next error reg (8b) */
128 #define E752X_BUF_ERRMASK 0x74 /* Memory buffer error mask reg (8b) */
129 #define E752X_BUF_SMICMD 0x7A /* Memory buffer SMI cmd reg (8b) */
130 #define E752X_DRAM_FERR 0x80 /* DRAM first error register (16b) */
131 #define E752X_DRAM_NERR 0x82 /* DRAM next error register (16b) */
132 #define E752X_DRAM_ERRMASK 0x84 /* DRAM error mask register (8b) */
133 #define E752X_DRAM_SMICMD 0x8A /* DRAM SMI command register (8b) */
134 #define E752X_DRAM_RETR_ADD 0xAC /* DRAM Retry address register (32b) */
135 #define E752X_DRAM_SEC1_ADD 0xA0 /* DRAM first correctable memory */
136 /* error address register (32b) */
137 /*
138 * 31 Reserved
139 * 30:2 CE address (64 byte block 34:6
140 * 1 Reserved
141 * 0 HiLoCS
142 */
143 #define E752X_DRAM_SEC2_ADD 0xC8 /* DRAM first correctable memory */
144 /* error address register (32b) */
145 /*
146 * 31 Reserved
147 * 30:2 CE address (64 byte block 34:6)
148 * 1 Reserved
149 * 0 HiLoCS
150 */
151 #define E752X_DRAM_DED_ADD 0xA4 /* DRAM first uncorrectable memory */
152 /* error address register (32b) */
153 /*
154 * 31 Reserved
155 * 30:2 CE address (64 byte block 34:6)
156 * 1 Reserved
157 * 0 HiLoCS
158 */
159 #define E752X_DRAM_SCRB_ADD 0xA8 /* DRAM 1st uncorrectable scrub mem */
160 /* error address register (32b) */
161 /*
162 * 31 Reserved
163 * 30:2 CE address (64 byte block 34:6
164 * 1 Reserved
165 * 0 HiLoCS
166 */
167 #define E752X_DRAM_SEC1_SYNDROME 0xC4 /* DRAM first correctable memory */
168 /* error syndrome register (16b) */
169 #define E752X_DRAM_SEC2_SYNDROME 0xC6 /* DRAM second correctable memory */
170 /* error syndrome register (16b) */
171 #define E752X_DEVPRES1 0xF4 /* Device Present 1 register (8b) */
172
173 /* 3100 IMCH specific register addresses - device 0 function 1 */
174 #define I3100_NSI_FERR 0x48 /* NSI first error reg (32b) */
175 #define I3100_NSI_NERR 0x4C /* NSI next error reg (32b) */
176 #define I3100_NSI_SMICMD 0x54 /* NSI SMI command register (32b) */
177 #define I3100_NSI_EMASK 0x90 /* NSI error mask register (32b) */
178
179 /* ICH5R register addresses - device 30 function 0 */
180 #define ICH5R_PCI_STAT 0x06 /* PCI status register (16b) */
181 #define ICH5R_PCI_2ND_STAT 0x1E /* PCI status secondary reg (16b) */
182 #define ICH5R_PCI_BRIDGE_CTL 0x3E /* PCI bridge control register (16b) */
183
184 enum e752x_chips {
185 E7520 = 0,
186 E7525 = 1,
187 E7320 = 2,
188 I3100 = 3
189 };
190
191 /*
192 * Those chips Support single-rank and dual-rank memories only.
193 *
194 * On e752x chips, the odd rows are present only on dual-rank memories.
195 * Dividing the rank by two will provide the dimm#
196 *
197 * i3100 MC has a different mapping: it supports only 4 ranks.
198 *
199 * The mapping is (from 1 to n):
200 * slot single-ranked double-ranked
201 * dimm #1 -> rank #4 NA
202 * dimm #2 -> rank #3 NA
203 * dimm #3 -> rank #2 Ranks 2 and 3
204 * dimm #4 -> rank $1 Ranks 1 and 4
205 *
206 * FIXME: The current mapping for i3100 considers that it supports up to 8
207 * ranks/chanel, but datasheet says that the MC supports only 4 ranks.
208 */
209
210 struct e752x_pvt {
211 struct pci_dev *dev_d0f0;
212 struct pci_dev *dev_d0f1;
213 u32 tolm;
214 u32 remapbase;
215 u32 remaplimit;
216 int mc_symmetric;
217 u8 map[8];
218 int map_type;
219 const struct e752x_dev_info *dev_info;
220 };
221
222 struct e752x_dev_info {
223 u16 err_dev;
224 u16 ctl_dev;
225 const char *ctl_name;
226 };
227
228 struct e752x_error_info {
229 u32 ferr_global;
230 u32 nerr_global;
231 u32 nsi_ferr; /* 3100 only */
232 u32 nsi_nerr; /* 3100 only */
233 u8 hi_ferr; /* all but 3100 */
234 u8 hi_nerr; /* all but 3100 */
235 u16 sysbus_ferr;
236 u16 sysbus_nerr;
237 u8 buf_ferr;
238 u8 buf_nerr;
239 u16 dram_ferr;
240 u16 dram_nerr;
241 u32 dram_sec1_add;
242 u32 dram_sec2_add;
243 u16 dram_sec1_syndrome;
244 u16 dram_sec2_syndrome;
245 u32 dram_ded_add;
246 u32 dram_scrb_add;
247 u32 dram_retr_add;
248 };
249
250 static const struct e752x_dev_info e752x_devs[] = {
251 [E7520] = {
252 .err_dev = PCI_DEVICE_ID_INTEL_7520_1_ERR,
253 .ctl_dev = PCI_DEVICE_ID_INTEL_7520_0,
254 .ctl_name = "E7520"},
255 [E7525] = {
256 .err_dev = PCI_DEVICE_ID_INTEL_7525_1_ERR,
257 .ctl_dev = PCI_DEVICE_ID_INTEL_7525_0,
258 .ctl_name = "E7525"},
259 [E7320] = {
260 .err_dev = PCI_DEVICE_ID_INTEL_7320_1_ERR,
261 .ctl_dev = PCI_DEVICE_ID_INTEL_7320_0,
262 .ctl_name = "E7320"},
263 [I3100] = {
264 .err_dev = PCI_DEVICE_ID_INTEL_3100_1_ERR,
265 .ctl_dev = PCI_DEVICE_ID_INTEL_3100_0,
266 .ctl_name = "3100"},
267 };
268
269 /* Valid scrub rates for the e752x/3100 hardware memory scrubber. We
270 * map the scrubbing bandwidth to a hardware register value. The 'set'
271 * operation finds the 'matching or higher value'. Note that scrubbing
272 * on the e752x can only be enabled/disabled. The 3100 supports
273 * a normal and fast mode.
274 */
275
276 #define SDRATE_EOT 0xFFFFFFFF
277
278 struct scrubrate {
279 u32 bandwidth; /* bandwidth consumed by scrubbing in bytes/sec */
280 u16 scrubval; /* register value for scrub rate */
281 };
282
283 /* Rate below assumes same performance as i3100 using PC3200 DDR2 in
284 * normal mode. e752x bridges don't support choosing normal or fast mode,
285 * so the scrubbing bandwidth value isn't all that important - scrubbing is
286 * either on or off.
287 */
288 static const struct scrubrate scrubrates_e752x[] = {
289 {0, 0x00}, /* Scrubbing Off */
290 {500000, 0x02}, /* Scrubbing On */
291 {SDRATE_EOT, 0x00} /* End of Table */
292 };
293
294 /* Fast mode: 2 GByte PC3200 DDR2 scrubbed in 33s = 63161283 bytes/s
295 * Normal mode: 125 (32000 / 256) times slower than fast mode.
296 */
297 static const struct scrubrate scrubrates_i3100[] = {
298 {0, 0x00}, /* Scrubbing Off */
299 {500000, 0x0a}, /* Normal mode - 32k clocks */
300 {62500000, 0x06}, /* Fast mode - 256 clocks */
301 {SDRATE_EOT, 0x00} /* End of Table */
302 };
303
ctl_page_to_phys(struct mem_ctl_info * mci,unsigned long page)304 static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci,
305 unsigned long page)
306 {
307 u32 remap;
308 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
309
310 edac_dbg(3, "\n");
311
312 if (page < pvt->tolm)
313 return page;
314
315 if ((page >= 0x100000) && (page < pvt->remapbase))
316 return page;
317
318 remap = (page - pvt->tolm) + pvt->remapbase;
319
320 if (remap < pvt->remaplimit)
321 return remap;
322
323 e752x_printk(KERN_ERR, "Invalid page %lx - out of range\n", page);
324 return pvt->tolm - 1;
325 }
326
do_process_ce(struct mem_ctl_info * mci,u16 error_one,u32 sec1_add,u16 sec1_syndrome)327 static void do_process_ce(struct mem_ctl_info *mci, u16 error_one,
328 u32 sec1_add, u16 sec1_syndrome)
329 {
330 u32 page;
331 int row;
332 int channel;
333 int i;
334 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
335
336 edac_dbg(3, "\n");
337
338 /* convert the addr to 4k page */
339 page = sec1_add >> (PAGE_SHIFT - 4);
340
341 /* FIXME - check for -1 */
342 if (pvt->mc_symmetric) {
343 /* chip select are bits 14 & 13 */
344 row = ((page >> 1) & 3);
345 e752x_printk(KERN_WARNING,
346 "Test row %d Table %d %d %d %d %d %d %d %d\n", row,
347 pvt->map[0], pvt->map[1], pvt->map[2], pvt->map[3],
348 pvt->map[4], pvt->map[5], pvt->map[6],
349 pvt->map[7]);
350
351 /* test for channel remapping */
352 for (i = 0; i < 8; i++) {
353 if (pvt->map[i] == row)
354 break;
355 }
356
357 e752x_printk(KERN_WARNING, "Test computed row %d\n", i);
358
359 if (i < 8)
360 row = i;
361 else
362 e752x_mc_printk(mci, KERN_WARNING,
363 "row %d not found in remap table\n",
364 row);
365 } else
366 row = edac_mc_find_csrow_by_page(mci, page);
367
368 /* 0 = channel A, 1 = channel B */
369 channel = !(error_one & 1);
370
371 /* e752x mc reads 34:6 of the DRAM linear address */
372 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1,
373 page, offset_in_page(sec1_add << 4), sec1_syndrome,
374 row, channel, -1,
375 "e752x CE", "");
376 }
377
process_ce(struct mem_ctl_info * mci,u16 error_one,u32 sec1_add,u16 sec1_syndrome,int * error_found,int handle_error)378 static inline void process_ce(struct mem_ctl_info *mci, u16 error_one,
379 u32 sec1_add, u16 sec1_syndrome, int *error_found,
380 int handle_error)
381 {
382 *error_found = 1;
383
384 if (handle_error)
385 do_process_ce(mci, error_one, sec1_add, sec1_syndrome);
386 }
387
do_process_ue(struct mem_ctl_info * mci,u16 error_one,u32 ded_add,u32 scrb_add)388 static void do_process_ue(struct mem_ctl_info *mci, u16 error_one,
389 u32 ded_add, u32 scrb_add)
390 {
391 u32 error_2b, block_page;
392 int row;
393 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
394
395 edac_dbg(3, "\n");
396
397 if (error_one & 0x0202) {
398 error_2b = ded_add;
399
400 /* convert to 4k address */
401 block_page = error_2b >> (PAGE_SHIFT - 4);
402
403 row = pvt->mc_symmetric ?
404 /* chip select are bits 14 & 13 */
405 ((block_page >> 1) & 3) :
406 edac_mc_find_csrow_by_page(mci, block_page);
407
408 /* e752x mc reads 34:6 of the DRAM linear address */
409 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
410 block_page,
411 offset_in_page(error_2b << 4), 0,
412 row, -1, -1,
413 "e752x UE from Read", "");
414
415 }
416 if (error_one & 0x0404) {
417 error_2b = scrb_add;
418
419 /* convert to 4k address */
420 block_page = error_2b >> (PAGE_SHIFT - 4);
421
422 row = pvt->mc_symmetric ?
423 /* chip select are bits 14 & 13 */
424 ((block_page >> 1) & 3) :
425 edac_mc_find_csrow_by_page(mci, block_page);
426
427 /* e752x mc reads 34:6 of the DRAM linear address */
428 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
429 block_page,
430 offset_in_page(error_2b << 4), 0,
431 row, -1, -1,
432 "e752x UE from Scruber", "");
433 }
434 }
435
process_ue(struct mem_ctl_info * mci,u16 error_one,u32 ded_add,u32 scrb_add,int * error_found,int handle_error)436 static inline void process_ue(struct mem_ctl_info *mci, u16 error_one,
437 u32 ded_add, u32 scrb_add, int *error_found,
438 int handle_error)
439 {
440 *error_found = 1;
441
442 if (handle_error)
443 do_process_ue(mci, error_one, ded_add, scrb_add);
444 }
445
process_ue_no_info_wr(struct mem_ctl_info * mci,int * error_found,int handle_error)446 static inline void process_ue_no_info_wr(struct mem_ctl_info *mci,
447 int *error_found, int handle_error)
448 {
449 *error_found = 1;
450
451 if (!handle_error)
452 return;
453
454 edac_dbg(3, "\n");
455 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, 0, 0, 0,
456 -1, -1, -1,
457 "e752x UE log memory write", "");
458 }
459
do_process_ded_retry(struct mem_ctl_info * mci,u16 error,u32 retry_add)460 static void do_process_ded_retry(struct mem_ctl_info *mci, u16 error,
461 u32 retry_add)
462 {
463 u32 error_1b, page;
464 int row;
465 struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
466
467 error_1b = retry_add;
468 page = error_1b >> (PAGE_SHIFT - 4); /* convert the addr to 4k page */
469
470 /* chip select are bits 14 & 13 */
471 row = pvt->mc_symmetric ? ((page >> 1) & 3) :
472 edac_mc_find_csrow_by_page(mci, page);
473
474 e752x_mc_printk(mci, KERN_WARNING,
475 "CE page 0x%lx, row %d : Memory read retry\n",
476 (long unsigned int)page, row);
477 }
478
process_ded_retry(struct mem_ctl_info * mci,u16 error,u32 retry_add,int * error_found,int handle_error)479 static inline void process_ded_retry(struct mem_ctl_info *mci, u16 error,
480 u32 retry_add, int *error_found,
481 int handle_error)
482 {
483 *error_found = 1;
484
485 if (handle_error)
486 do_process_ded_retry(mci, error, retry_add);
487 }
488
process_threshold_ce(struct mem_ctl_info * mci,u16 error,int * error_found,int handle_error)489 static inline void process_threshold_ce(struct mem_ctl_info *mci, u16 error,
490 int *error_found, int handle_error)
491 {
492 *error_found = 1;
493
494 if (handle_error)
495 e752x_mc_printk(mci, KERN_WARNING, "Memory threshold CE\n");
496 }
497
498 static char *global_message[11] = {
499 "PCI Express C1",
500 "PCI Express C",
501 "PCI Express B1",
502 "PCI Express B",
503 "PCI Express A1",
504 "PCI Express A",
505 "DMA Controller",
506 "HUB or NS Interface",
507 "System Bus",
508 "DRAM Controller", /* 9th entry */
509 "Internal Buffer"
510 };
511
512 #define DRAM_ENTRY 9
513
514 static char *fatal_message[2] = { "Non-Fatal ", "Fatal " };
515
do_global_error(int fatal,u32 errors)516 static void do_global_error(int fatal, u32 errors)
517 {
518 int i;
519
520 for (i = 0; i < 11; i++) {
521 if (errors & (1 << i)) {
522 /* If the error is from DRAM Controller OR
523 * we are to report ALL errors, then
524 * report the error
525 */
526 if ((i == DRAM_ENTRY) || report_non_memory_errors)
527 e752x_printk(KERN_WARNING, "%sError %s\n",
528 fatal_message[fatal],
529 global_message[i]);
530 }
531 }
532 }
533
global_error(int fatal,u32 errors,int * error_found,int handle_error)534 static inline void global_error(int fatal, u32 errors, int *error_found,
535 int handle_error)
536 {
537 *error_found = 1;
538
539 if (handle_error)
540 do_global_error(fatal, errors);
541 }
542
543 static char *hub_message[7] = {
544 "HI Address or Command Parity", "HI Illegal Access",
545 "HI Internal Parity", "Out of Range Access",
546 "HI Data Parity", "Enhanced Config Access",
547 "Hub Interface Target Abort"
548 };
549
do_hub_error(int fatal,u8 errors)550 static void do_hub_error(int fatal, u8 errors)
551 {
552 int i;
553
554 for (i = 0; i < 7; i++) {
555 if (errors & (1 << i))
556 e752x_printk(KERN_WARNING, "%sError %s\n",
557 fatal_message[fatal], hub_message[i]);
558 }
559 }
560
hub_error(int fatal,u8 errors,int * error_found,int handle_error)561 static inline void hub_error(int fatal, u8 errors, int *error_found,
562 int handle_error)
563 {
564 *error_found = 1;
565
566 if (handle_error)
567 do_hub_error(fatal, errors);
568 }
569
570 #define NSI_FATAL_MASK 0x0c080081
571 #define NSI_NON_FATAL_MASK 0x23a0ba64
572 #define NSI_ERR_MASK (NSI_FATAL_MASK | NSI_NON_FATAL_MASK)
573
574 static char *nsi_message[30] = {
575 "NSI Link Down", /* NSI_FERR/NSI_NERR bit 0, fatal error */
576 "", /* reserved */
577 "NSI Parity Error", /* bit 2, non-fatal */
578 "", /* reserved */
579 "", /* reserved */
580 "Correctable Error Message", /* bit 5, non-fatal */
581 "Non-Fatal Error Message", /* bit 6, non-fatal */
582 "Fatal Error Message", /* bit 7, fatal */
583 "", /* reserved */
584 "Receiver Error", /* bit 9, non-fatal */
585 "", /* reserved */
586 "Bad TLP", /* bit 11, non-fatal */
587 "Bad DLLP", /* bit 12, non-fatal */
588 "REPLAY_NUM Rollover", /* bit 13, non-fatal */
589 "", /* reserved */
590 "Replay Timer Timeout", /* bit 15, non-fatal */
591 "", /* reserved */
592 "", /* reserved */
593 "", /* reserved */
594 "Data Link Protocol Error", /* bit 19, fatal */
595 "", /* reserved */
596 "Poisoned TLP", /* bit 21, non-fatal */
597 "", /* reserved */
598 "Completion Timeout", /* bit 23, non-fatal */
599 "Completer Abort", /* bit 24, non-fatal */
600 "Unexpected Completion", /* bit 25, non-fatal */
601 "Receiver Overflow", /* bit 26, fatal */
602 "Malformed TLP", /* bit 27, fatal */
603 "", /* reserved */
604 "Unsupported Request" /* bit 29, non-fatal */
605 };
606
do_nsi_error(int fatal,u32 errors)607 static void do_nsi_error(int fatal, u32 errors)
608 {
609 int i;
610
611 for (i = 0; i < 30; i++) {
612 if (errors & (1 << i))
613 printk(KERN_WARNING "%sError %s\n",
614 fatal_message[fatal], nsi_message[i]);
615 }
616 }
617
nsi_error(int fatal,u32 errors,int * error_found,int handle_error)618 static inline void nsi_error(int fatal, u32 errors, int *error_found,
619 int handle_error)
620 {
621 *error_found = 1;
622
623 if (handle_error)
624 do_nsi_error(fatal, errors);
625 }
626
627 static char *membuf_message[4] = {
628 "Internal PMWB to DRAM parity",
629 "Internal PMWB to System Bus Parity",
630 "Internal System Bus or IO to PMWB Parity",
631 "Internal DRAM to PMWB Parity"
632 };
633
do_membuf_error(u8 errors)634 static void do_membuf_error(u8 errors)
635 {
636 int i;
637
638 for (i = 0; i < 4; i++) {
639 if (errors & (1 << i))
640 e752x_printk(KERN_WARNING, "Non-Fatal Error %s\n",
641 membuf_message[i]);
642 }
643 }
644
membuf_error(u8 errors,int * error_found,int handle_error)645 static inline void membuf_error(u8 errors, int *error_found, int handle_error)
646 {
647 *error_found = 1;
648
649 if (handle_error)
650 do_membuf_error(errors);
651 }
652
653 static char *sysbus_message[10] = {
654 "Addr or Request Parity",
655 "Data Strobe Glitch",
656 "Addr Strobe Glitch",
657 "Data Parity",
658 "Addr Above TOM",
659 "Non DRAM Lock Error",
660 "MCERR", "BINIT",
661 "Memory Parity",
662 "IO Subsystem Parity"
663 };
664
do_sysbus_error(int fatal,u32 errors)665 static void do_sysbus_error(int fatal, u32 errors)
666 {
667 int i;
668
669 for (i = 0; i < 10; i++) {
670 if (errors & (1 << i))
671 e752x_printk(KERN_WARNING, "%sError System Bus %s\n",
672 fatal_message[fatal], sysbus_message[i]);
673 }
674 }
675
sysbus_error(int fatal,u32 errors,int * error_found,int handle_error)676 static inline void sysbus_error(int fatal, u32 errors, int *error_found,
677 int handle_error)
678 {
679 *error_found = 1;
680
681 if (handle_error)
682 do_sysbus_error(fatal, errors);
683 }
684
e752x_check_hub_interface(struct e752x_error_info * info,int * error_found,int handle_error)685 static void e752x_check_hub_interface(struct e752x_error_info *info,
686 int *error_found, int handle_error)
687 {
688 u8 stat8;
689
690 //pci_read_config_byte(dev,E752X_HI_FERR,&stat8);
691
692 stat8 = info->hi_ferr;
693
694 if (stat8 & 0x7f) { /* Error, so process */
695 stat8 &= 0x7f;
696
697 if (stat8 & 0x2b)
698 hub_error(1, stat8 & 0x2b, error_found, handle_error);
699
700 if (stat8 & 0x54)
701 hub_error(0, stat8 & 0x54, error_found, handle_error);
702 }
703 //pci_read_config_byte(dev,E752X_HI_NERR,&stat8);
704
705 stat8 = info->hi_nerr;
706
707 if (stat8 & 0x7f) { /* Error, so process */
708 stat8 &= 0x7f;
709
710 if (stat8 & 0x2b)
711 hub_error(1, stat8 & 0x2b, error_found, handle_error);
712
713 if (stat8 & 0x54)
714 hub_error(0, stat8 & 0x54, error_found, handle_error);
715 }
716 }
717
e752x_check_ns_interface(struct e752x_error_info * info,int * error_found,int handle_error)718 static void e752x_check_ns_interface(struct e752x_error_info *info,
719 int *error_found, int handle_error)
720 {
721 u32 stat32;
722
723 stat32 = info->nsi_ferr;
724 if (stat32 & NSI_ERR_MASK) { /* Error, so process */
725 if (stat32 & NSI_FATAL_MASK) /* check for fatal errors */
726 nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
727 handle_error);
728 if (stat32 & NSI_NON_FATAL_MASK) /* check for non-fatal ones */
729 nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
730 handle_error);
731 }
732 stat32 = info->nsi_nerr;
733 if (stat32 & NSI_ERR_MASK) {
734 if (stat32 & NSI_FATAL_MASK)
735 nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
736 handle_error);
737 if (stat32 & NSI_NON_FATAL_MASK)
738 nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
739 handle_error);
740 }
741 }
742
e752x_check_sysbus(struct e752x_error_info * info,int * error_found,int handle_error)743 static void e752x_check_sysbus(struct e752x_error_info *info,
744 int *error_found, int handle_error)
745 {
746 u32 stat32, error32;
747
748 //pci_read_config_dword(dev,E752X_SYSBUS_FERR,&stat32);
749 stat32 = info->sysbus_ferr + (info->sysbus_nerr << 16);
750
751 if (stat32 == 0)
752 return; /* no errors */
753
754 error32 = (stat32 >> 16) & 0x3ff;
755 stat32 = stat32 & 0x3ff;
756
757 if (stat32 & 0x087)
758 sysbus_error(1, stat32 & 0x087, error_found, handle_error);
759
760 if (stat32 & 0x378)
761 sysbus_error(0, stat32 & 0x378, error_found, handle_error);
762
763 if (error32 & 0x087)
764 sysbus_error(1, error32 & 0x087, error_found, handle_error);
765
766 if (error32 & 0x378)
767 sysbus_error(0, error32 & 0x378, error_found, handle_error);
768 }
769
e752x_check_membuf(struct e752x_error_info * info,int * error_found,int handle_error)770 static void e752x_check_membuf(struct e752x_error_info *info,
771 int *error_found, int handle_error)
772 {
773 u8 stat8;
774
775 stat8 = info->buf_ferr;
776
777 if (stat8 & 0x0f) { /* Error, so process */
778 stat8 &= 0x0f;
779 membuf_error(stat8, error_found, handle_error);
780 }
781
782 stat8 = info->buf_nerr;
783
784 if (stat8 & 0x0f) { /* Error, so process */
785 stat8 &= 0x0f;
786 membuf_error(stat8, error_found, handle_error);
787 }
788 }
789
e752x_check_dram(struct mem_ctl_info * mci,struct e752x_error_info * info,int * error_found,int handle_error)790 static void e752x_check_dram(struct mem_ctl_info *mci,
791 struct e752x_error_info *info, int *error_found,
792 int handle_error)
793 {
794 u16 error_one, error_next;
795
796 error_one = info->dram_ferr;
797 error_next = info->dram_nerr;
798
799 /* decode and report errors */
800 if (error_one & 0x0101) /* check first error correctable */
801 process_ce(mci, error_one, info->dram_sec1_add,
802 info->dram_sec1_syndrome, error_found, handle_error);
803
804 if (error_next & 0x0101) /* check next error correctable */
805 process_ce(mci, error_next, info->dram_sec2_add,
806 info->dram_sec2_syndrome, error_found, handle_error);
807
808 if (error_one & 0x4040)
809 process_ue_no_info_wr(mci, error_found, handle_error);
810
811 if (error_next & 0x4040)
812 process_ue_no_info_wr(mci, error_found, handle_error);
813
814 if (error_one & 0x2020)
815 process_ded_retry(mci, error_one, info->dram_retr_add,
816 error_found, handle_error);
817
818 if (error_next & 0x2020)
819 process_ded_retry(mci, error_next, info->dram_retr_add,
820 error_found, handle_error);
821
822 if (error_one & 0x0808)
823 process_threshold_ce(mci, error_one, error_found, handle_error);
824
825 if (error_next & 0x0808)
826 process_threshold_ce(mci, error_next, error_found,
827 handle_error);
828
829 if (error_one & 0x0606)
830 process_ue(mci, error_one, info->dram_ded_add,
831 info->dram_scrb_add, error_found, handle_error);
832
833 if (error_next & 0x0606)
834 process_ue(mci, error_next, info->dram_ded_add,
835 info->dram_scrb_add, error_found, handle_error);
836 }
837
e752x_get_error_info(struct mem_ctl_info * mci,struct e752x_error_info * info)838 static void e752x_get_error_info(struct mem_ctl_info *mci,
839 struct e752x_error_info *info)
840 {
841 struct pci_dev *dev;
842 struct e752x_pvt *pvt;
843
844 memset(info, 0, sizeof(*info));
845 pvt = (struct e752x_pvt *)mci->pvt_info;
846 dev = pvt->dev_d0f1;
847 pci_read_config_dword(dev, E752X_FERR_GLOBAL, &info->ferr_global);
848
849 if (info->ferr_global) {
850 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
851 pci_read_config_dword(dev, I3100_NSI_FERR,
852 &info->nsi_ferr);
853 info->hi_ferr = 0;
854 } else {
855 pci_read_config_byte(dev, E752X_HI_FERR,
856 &info->hi_ferr);
857 info->nsi_ferr = 0;
858 }
859 pci_read_config_word(dev, E752X_SYSBUS_FERR,
860 &info->sysbus_ferr);
861 pci_read_config_byte(dev, E752X_BUF_FERR, &info->buf_ferr);
862 pci_read_config_word(dev, E752X_DRAM_FERR, &info->dram_ferr);
863 pci_read_config_dword(dev, E752X_DRAM_SEC1_ADD,
864 &info->dram_sec1_add);
865 pci_read_config_word(dev, E752X_DRAM_SEC1_SYNDROME,
866 &info->dram_sec1_syndrome);
867 pci_read_config_dword(dev, E752X_DRAM_DED_ADD,
868 &info->dram_ded_add);
869 pci_read_config_dword(dev, E752X_DRAM_SCRB_ADD,
870 &info->dram_scrb_add);
871 pci_read_config_dword(dev, E752X_DRAM_RETR_ADD,
872 &info->dram_retr_add);
873
874 /* ignore the reserved bits just in case */
875 if (info->hi_ferr & 0x7f)
876 pci_write_config_byte(dev, E752X_HI_FERR,
877 info->hi_ferr);
878
879 if (info->nsi_ferr & NSI_ERR_MASK)
880 pci_write_config_dword(dev, I3100_NSI_FERR,
881 info->nsi_ferr);
882
883 if (info->sysbus_ferr)
884 pci_write_config_word(dev, E752X_SYSBUS_FERR,
885 info->sysbus_ferr);
886
887 if (info->buf_ferr & 0x0f)
888 pci_write_config_byte(dev, E752X_BUF_FERR,
889 info->buf_ferr);
890
891 if (info->dram_ferr)
892 pci_write_bits16(pvt->dev_d0f1, E752X_DRAM_FERR,
893 info->dram_ferr, info->dram_ferr);
894
895 pci_write_config_dword(dev, E752X_FERR_GLOBAL,
896 info->ferr_global);
897 }
898
899 pci_read_config_dword(dev, E752X_NERR_GLOBAL, &info->nerr_global);
900
901 if (info->nerr_global) {
902 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
903 pci_read_config_dword(dev, I3100_NSI_NERR,
904 &info->nsi_nerr);
905 info->hi_nerr = 0;
906 } else {
907 pci_read_config_byte(dev, E752X_HI_NERR,
908 &info->hi_nerr);
909 info->nsi_nerr = 0;
910 }
911 pci_read_config_word(dev, E752X_SYSBUS_NERR,
912 &info->sysbus_nerr);
913 pci_read_config_byte(dev, E752X_BUF_NERR, &info->buf_nerr);
914 pci_read_config_word(dev, E752X_DRAM_NERR, &info->dram_nerr);
915 pci_read_config_dword(dev, E752X_DRAM_SEC2_ADD,
916 &info->dram_sec2_add);
917 pci_read_config_word(dev, E752X_DRAM_SEC2_SYNDROME,
918 &info->dram_sec2_syndrome);
919
920 if (info->hi_nerr & 0x7f)
921 pci_write_config_byte(dev, E752X_HI_NERR,
922 info->hi_nerr);
923
924 if (info->nsi_nerr & NSI_ERR_MASK)
925 pci_write_config_dword(dev, I3100_NSI_NERR,
926 info->nsi_nerr);
927
928 if (info->sysbus_nerr)
929 pci_write_config_word(dev, E752X_SYSBUS_NERR,
930 info->sysbus_nerr);
931
932 if (info->buf_nerr & 0x0f)
933 pci_write_config_byte(dev, E752X_BUF_NERR,
934 info->buf_nerr);
935
936 if (info->dram_nerr)
937 pci_write_bits16(pvt->dev_d0f1, E752X_DRAM_NERR,
938 info->dram_nerr, info->dram_nerr);
939
940 pci_write_config_dword(dev, E752X_NERR_GLOBAL,
941 info->nerr_global);
942 }
943 }
944
e752x_process_error_info(struct mem_ctl_info * mci,struct e752x_error_info * info,int handle_errors)945 static int e752x_process_error_info(struct mem_ctl_info *mci,
946 struct e752x_error_info *info,
947 int handle_errors)
948 {
949 u32 error32, stat32;
950 int error_found;
951
952 error_found = 0;
953 error32 = (info->ferr_global >> 18) & 0x3ff;
954 stat32 = (info->ferr_global >> 4) & 0x7ff;
955
956 if (error32)
957 global_error(1, error32, &error_found, handle_errors);
958
959 if (stat32)
960 global_error(0, stat32, &error_found, handle_errors);
961
962 error32 = (info->nerr_global >> 18) & 0x3ff;
963 stat32 = (info->nerr_global >> 4) & 0x7ff;
964
965 if (error32)
966 global_error(1, error32, &error_found, handle_errors);
967
968 if (stat32)
969 global_error(0, stat32, &error_found, handle_errors);
970
971 e752x_check_hub_interface(info, &error_found, handle_errors);
972 e752x_check_ns_interface(info, &error_found, handle_errors);
973 e752x_check_sysbus(info, &error_found, handle_errors);
974 e752x_check_membuf(info, &error_found, handle_errors);
975 e752x_check_dram(mci, info, &error_found, handle_errors);
976 return error_found;
977 }
978
e752x_check(struct mem_ctl_info * mci)979 static void e752x_check(struct mem_ctl_info *mci)
980 {
981 struct e752x_error_info info;
982
983 edac_dbg(3, "\n");
984 e752x_get_error_info(mci, &info);
985 e752x_process_error_info(mci, &info, 1);
986 }
987
988 /* Program byte/sec bandwidth scrub rate to hardware */
set_sdram_scrub_rate(struct mem_ctl_info * mci,u32 new_bw)989 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
990 {
991 const struct scrubrate *scrubrates;
992 struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
993 struct pci_dev *pdev = pvt->dev_d0f0;
994 int i;
995
996 if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
997 scrubrates = scrubrates_i3100;
998 else
999 scrubrates = scrubrates_e752x;
1000
1001 /* Translate the desired scrub rate to a e752x/3100 register value.
1002 * Search for the bandwidth that is equal or greater than the
1003 * desired rate and program the cooresponding register value.
1004 */
1005 for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1006 if (scrubrates[i].bandwidth >= new_bw)
1007 break;
1008
1009 if (scrubrates[i].bandwidth == SDRATE_EOT)
1010 return -1;
1011
1012 pci_write_config_word(pdev, E752X_MCHSCRB, scrubrates[i].scrubval);
1013
1014 return scrubrates[i].bandwidth;
1015 }
1016
1017 /* Convert current scrub rate value into byte/sec bandwidth */
get_sdram_scrub_rate(struct mem_ctl_info * mci)1018 static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
1019 {
1020 const struct scrubrate *scrubrates;
1021 struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
1022 struct pci_dev *pdev = pvt->dev_d0f0;
1023 u16 scrubval;
1024 int i;
1025
1026 if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
1027 scrubrates = scrubrates_i3100;
1028 else
1029 scrubrates = scrubrates_e752x;
1030
1031 /* Find the bandwidth matching the memory scrubber configuration */
1032 pci_read_config_word(pdev, E752X_MCHSCRB, &scrubval);
1033 scrubval = scrubval & 0x0f;
1034
1035 for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1036 if (scrubrates[i].scrubval == scrubval)
1037 break;
1038
1039 if (scrubrates[i].bandwidth == SDRATE_EOT) {
1040 e752x_printk(KERN_WARNING,
1041 "Invalid sdram scrub control value: 0x%x\n", scrubval);
1042 return -1;
1043 }
1044 return scrubrates[i].bandwidth;
1045
1046 }
1047
1048 /* Return 1 if dual channel mode is active. Else return 0. */
dual_channel_active(u16 ddrcsr)1049 static inline int dual_channel_active(u16 ddrcsr)
1050 {
1051 return (((ddrcsr >> 12) & 3) == 3);
1052 }
1053
1054 /* Remap csrow index numbers if map_type is "reverse"
1055 */
remap_csrow_index(struct mem_ctl_info * mci,int index)1056 static inline int remap_csrow_index(struct mem_ctl_info *mci, int index)
1057 {
1058 struct e752x_pvt *pvt = mci->pvt_info;
1059
1060 if (!pvt->map_type)
1061 return (7 - index);
1062
1063 return (index);
1064 }
1065
e752x_init_csrows(struct mem_ctl_info * mci,struct pci_dev * pdev,u16 ddrcsr)1066 static void e752x_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev,
1067 u16 ddrcsr)
1068 {
1069 struct csrow_info *csrow;
1070 enum edac_type edac_mode;
1071 unsigned long last_cumul_size;
1072 int index, mem_dev, drc_chan;
1073 int drc_drbg; /* DRB granularity 0=64mb, 1=128mb */
1074 int drc_ddim; /* DRAM Data Integrity Mode 0=none, 2=edac */
1075 u8 value;
1076 u32 dra, drc, cumul_size, i, nr_pages;
1077
1078 dra = 0;
1079 for (index = 0; index < 4; index++) {
1080 u8 dra_reg;
1081 pci_read_config_byte(pdev, E752X_DRA + index, &dra_reg);
1082 dra |= dra_reg << (index * 8);
1083 }
1084 pci_read_config_dword(pdev, E752X_DRC, &drc);
1085 drc_chan = dual_channel_active(ddrcsr) ? 1 : 0;
1086 drc_drbg = drc_chan + 1; /* 128 in dual mode, 64 in single */
1087 drc_ddim = (drc >> 20) & 0x3;
1088
1089 /* The dram row boundary (DRB) reg values are boundary address for
1090 * each DRAM row with a granularity of 64 or 128MB (single/dual
1091 * channel operation). DRB regs are cumulative; therefore DRB7 will
1092 * contain the total memory contained in all eight rows.
1093 */
1094 for (last_cumul_size = index = 0; index < mci->nr_csrows; index++) {
1095 /* mem_dev 0=x8, 1=x4 */
1096 mem_dev = (dra >> (index * 4 + 2)) & 0x3;
1097 csrow = mci->csrows[remap_csrow_index(mci, index)];
1098
1099 mem_dev = (mem_dev == 2);
1100 pci_read_config_byte(pdev, E752X_DRB + index, &value);
1101 /* convert a 128 or 64 MiB DRB to a page size. */
1102 cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
1103 edac_dbg(3, "(%d) cumul_size 0x%x\n", index, cumul_size);
1104 if (cumul_size == last_cumul_size)
1105 continue; /* not populated */
1106
1107 csrow->first_page = last_cumul_size;
1108 csrow->last_page = cumul_size - 1;
1109 nr_pages = cumul_size - last_cumul_size;
1110 last_cumul_size = cumul_size;
1111
1112 /*
1113 * if single channel or x8 devices then SECDED
1114 * if dual channel and x4 then S4ECD4ED
1115 */
1116 if (drc_ddim) {
1117 if (drc_chan && mem_dev) {
1118 edac_mode = EDAC_S4ECD4ED;
1119 mci->edac_cap |= EDAC_FLAG_S4ECD4ED;
1120 } else {
1121 edac_mode = EDAC_SECDED;
1122 mci->edac_cap |= EDAC_FLAG_SECDED;
1123 }
1124 } else
1125 edac_mode = EDAC_NONE;
1126 for (i = 0; i < csrow->nr_channels; i++) {
1127 struct dimm_info *dimm = csrow->channels[i]->dimm;
1128
1129 edac_dbg(3, "Initializing rank at (%i,%i)\n", index, i);
1130 dimm->nr_pages = nr_pages / csrow->nr_channels;
1131 dimm->grain = 1 << 12; /* 4KiB - resolution of CELOG */
1132 dimm->mtype = MEM_RDDR; /* only one type supported */
1133 dimm->dtype = mem_dev ? DEV_X4 : DEV_X8;
1134 dimm->edac_mode = edac_mode;
1135 }
1136 }
1137 }
1138
e752x_init_mem_map_table(struct pci_dev * pdev,struct e752x_pvt * pvt)1139 static void e752x_init_mem_map_table(struct pci_dev *pdev,
1140 struct e752x_pvt *pvt)
1141 {
1142 int index;
1143 u8 value, last, row;
1144
1145 last = 0;
1146 row = 0;
1147
1148 for (index = 0; index < 8; index += 2) {
1149 pci_read_config_byte(pdev, E752X_DRB + index, &value);
1150 /* test if there is a dimm in this slot */
1151 if (value == last) {
1152 /* no dimm in the slot, so flag it as empty */
1153 pvt->map[index] = 0xff;
1154 pvt->map[index + 1] = 0xff;
1155 } else { /* there is a dimm in the slot */
1156 pvt->map[index] = row;
1157 row++;
1158 last = value;
1159 /* test the next value to see if the dimm is double
1160 * sided
1161 */
1162 pci_read_config_byte(pdev, E752X_DRB + index + 1,
1163 &value);
1164
1165 /* the dimm is single sided, so flag as empty */
1166 /* this is a double sided dimm to save the next row #*/
1167 pvt->map[index + 1] = (value == last) ? 0xff : row;
1168 row++;
1169 last = value;
1170 }
1171 }
1172 }
1173
1174 /* Return 0 on success or 1 on failure. */
e752x_get_devs(struct pci_dev * pdev,int dev_idx,struct e752x_pvt * pvt)1175 static int e752x_get_devs(struct pci_dev *pdev, int dev_idx,
1176 struct e752x_pvt *pvt)
1177 {
1178 pvt->dev_d0f1 = pci_get_device(PCI_VENDOR_ID_INTEL,
1179 pvt->dev_info->err_dev, NULL);
1180
1181 if (pvt->dev_d0f1 == NULL) {
1182 pvt->dev_d0f1 = pci_scan_single_device(pdev->bus,
1183 PCI_DEVFN(0, 1));
1184 pci_dev_get(pvt->dev_d0f1);
1185 }
1186
1187 if (pvt->dev_d0f1 == NULL) {
1188 e752x_printk(KERN_ERR, "error reporting device not found:"
1189 "vendor %x device 0x%x (broken BIOS?)\n",
1190 PCI_VENDOR_ID_INTEL, e752x_devs[dev_idx].err_dev);
1191 return 1;
1192 }
1193
1194 pvt->dev_d0f0 = pci_get_device(PCI_VENDOR_ID_INTEL,
1195 e752x_devs[dev_idx].ctl_dev,
1196 NULL);
1197
1198 if (pvt->dev_d0f0 == NULL)
1199 goto fail;
1200
1201 return 0;
1202
1203 fail:
1204 pci_dev_put(pvt->dev_d0f1);
1205 return 1;
1206 }
1207
1208 /* Setup system bus parity mask register.
1209 * Sysbus parity supported on:
1210 * e7320/e7520/e7525 + Xeon
1211 */
e752x_init_sysbus_parity_mask(struct e752x_pvt * pvt)1212 static void e752x_init_sysbus_parity_mask(struct e752x_pvt *pvt)
1213 {
1214 char *cpu_id = cpu_data(0).x86_model_id;
1215 struct pci_dev *dev = pvt->dev_d0f1;
1216 int enable = 1;
1217
1218 /* Allow module parameter override, else see if CPU supports parity */
1219 if (sysbus_parity != -1) {
1220 enable = sysbus_parity;
1221 } else if (cpu_id[0] && !strstr(cpu_id, "Xeon")) {
1222 e752x_printk(KERN_INFO, "System Bus Parity not "
1223 "supported by CPU, disabling\n");
1224 enable = 0;
1225 }
1226
1227 if (enable)
1228 pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0000);
1229 else
1230 pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0309);
1231 }
1232
e752x_init_error_reporting_regs(struct e752x_pvt * pvt)1233 static void e752x_init_error_reporting_regs(struct e752x_pvt *pvt)
1234 {
1235 struct pci_dev *dev;
1236
1237 dev = pvt->dev_d0f1;
1238 /* Turn off error disable & SMI in case the BIOS turned it on */
1239 if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
1240 pci_write_config_dword(dev, I3100_NSI_EMASK, 0);
1241 pci_write_config_dword(dev, I3100_NSI_SMICMD, 0);
1242 } else {
1243 pci_write_config_byte(dev, E752X_HI_ERRMASK, 0x00);
1244 pci_write_config_byte(dev, E752X_HI_SMICMD, 0x00);
1245 }
1246
1247 e752x_init_sysbus_parity_mask(pvt);
1248
1249 pci_write_config_word(dev, E752X_SYSBUS_SMICMD, 0x00);
1250 pci_write_config_byte(dev, E752X_BUF_ERRMASK, 0x00);
1251 pci_write_config_byte(dev, E752X_BUF_SMICMD, 0x00);
1252 pci_write_config_byte(dev, E752X_DRAM_ERRMASK, 0x00);
1253 pci_write_config_byte(dev, E752X_DRAM_SMICMD, 0x00);
1254 }
1255
e752x_probe1(struct pci_dev * pdev,int dev_idx)1256 static int e752x_probe1(struct pci_dev *pdev, int dev_idx)
1257 {
1258 u16 pci_data;
1259 u8 stat8;
1260 struct mem_ctl_info *mci;
1261 struct edac_mc_layer layers[2];
1262 struct e752x_pvt *pvt;
1263 u16 ddrcsr;
1264 int drc_chan; /* Number of channels 0=1chan,1=2chan */
1265 struct e752x_error_info discard;
1266
1267 edac_dbg(0, "mci\n");
1268 edac_dbg(0, "Starting Probe1\n");
1269
1270 /* check to see if device 0 function 1 is enabled; if it isn't, we
1271 * assume the BIOS has reserved it for a reason and is expecting
1272 * exclusive access, we take care not to violate that assumption and
1273 * fail the probe. */
1274 pci_read_config_byte(pdev, E752X_DEVPRES1, &stat8);
1275 if (!force_function_unhide && !(stat8 & (1 << 5))) {
1276 printk(KERN_INFO "Contact your BIOS vendor to see if the "
1277 "E752x error registers can be safely un-hidden\n");
1278 return -ENODEV;
1279 }
1280 stat8 |= (1 << 5);
1281 pci_write_config_byte(pdev, E752X_DEVPRES1, stat8);
1282
1283 pci_read_config_word(pdev, E752X_DDRCSR, &ddrcsr);
1284 /* FIXME: should check >>12 or 0xf, true for all? */
1285 /* Dual channel = 1, Single channel = 0 */
1286 drc_chan = dual_channel_active(ddrcsr);
1287
1288 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
1289 layers[0].size = E752X_NR_CSROWS;
1290 layers[0].is_virt_csrow = true;
1291 layers[1].type = EDAC_MC_LAYER_CHANNEL;
1292 layers[1].size = drc_chan + 1;
1293 layers[1].is_virt_csrow = false;
1294 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt));
1295 if (mci == NULL)
1296 return -ENOMEM;
1297
1298 edac_dbg(3, "init mci\n");
1299 mci->mtype_cap = MEM_FLAG_RDDR;
1300 /* 3100 IMCH supports SECDEC only */
1301 mci->edac_ctl_cap = (dev_idx == I3100) ? EDAC_FLAG_SECDED :
1302 (EDAC_FLAG_NONE | EDAC_FLAG_SECDED | EDAC_FLAG_S4ECD4ED);
1303 /* FIXME - what if different memory types are in different csrows? */
1304 mci->mod_name = EDAC_MOD_STR;
1305 mci->pdev = &pdev->dev;
1306
1307 edac_dbg(3, "init pvt\n");
1308 pvt = (struct e752x_pvt *)mci->pvt_info;
1309 pvt->dev_info = &e752x_devs[dev_idx];
1310 pvt->mc_symmetric = ((ddrcsr & 0x10) != 0);
1311
1312 if (e752x_get_devs(pdev, dev_idx, pvt)) {
1313 edac_mc_free(mci);
1314 return -ENODEV;
1315 }
1316
1317 edac_dbg(3, "more mci init\n");
1318 mci->ctl_name = pvt->dev_info->ctl_name;
1319 mci->dev_name = pci_name(pdev);
1320 mci->edac_check = e752x_check;
1321 mci->ctl_page_to_phys = ctl_page_to_phys;
1322 mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
1323 mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
1324
1325 /* set the map type. 1 = normal, 0 = reversed
1326 * Must be set before e752x_init_csrows in case csrow mapping
1327 * is reversed.
1328 */
1329 pci_read_config_byte(pdev, E752X_DRM, &stat8);
1330 pvt->map_type = ((stat8 & 0x0f) > ((stat8 >> 4) & 0x0f));
1331
1332 e752x_init_csrows(mci, pdev, ddrcsr);
1333 e752x_init_mem_map_table(pdev, pvt);
1334
1335 if (dev_idx == I3100)
1336 mci->edac_cap = EDAC_FLAG_SECDED; /* the only mode supported */
1337 else
1338 mci->edac_cap |= EDAC_FLAG_NONE;
1339 edac_dbg(3, "tolm, remapbase, remaplimit\n");
1340
1341 /* load the top of low memory, remap base, and remap limit vars */
1342 pci_read_config_word(pdev, E752X_TOLM, &pci_data);
1343 pvt->tolm = ((u32) pci_data) << 4;
1344 pci_read_config_word(pdev, E752X_REMAPBASE, &pci_data);
1345 pvt->remapbase = ((u32) pci_data) << 14;
1346 pci_read_config_word(pdev, E752X_REMAPLIMIT, &pci_data);
1347 pvt->remaplimit = ((u32) pci_data) << 14;
1348 e752x_printk(KERN_INFO,
1349 "tolm = %x, remapbase = %x, remaplimit = %x\n",
1350 pvt->tolm, pvt->remapbase, pvt->remaplimit);
1351
1352 /* Here we assume that we will never see multiple instances of this
1353 * type of memory controller. The ID is therefore hardcoded to 0.
1354 */
1355 if (edac_mc_add_mc(mci)) {
1356 edac_dbg(3, "failed edac_mc_add_mc()\n");
1357 goto fail;
1358 }
1359
1360 e752x_init_error_reporting_regs(pvt);
1361 e752x_get_error_info(mci, &discard); /* clear other MCH errors */
1362
1363 /* allocating generic PCI control info */
1364 e752x_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
1365 if (!e752x_pci) {
1366 printk(KERN_WARNING
1367 "%s(): Unable to create PCI control\n", __func__);
1368 printk(KERN_WARNING
1369 "%s(): PCI error report via EDAC not setup\n",
1370 __func__);
1371 }
1372
1373 /* get this far and it's successful */
1374 edac_dbg(3, "success\n");
1375 return 0;
1376
1377 fail:
1378 pci_dev_put(pvt->dev_d0f0);
1379 pci_dev_put(pvt->dev_d0f1);
1380 edac_mc_free(mci);
1381
1382 return -ENODEV;
1383 }
1384
1385 /* returns count (>= 0), or negative on error */
e752x_init_one(struct pci_dev * pdev,const struct pci_device_id * ent)1386 static int e752x_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
1387 {
1388 edac_dbg(0, "\n");
1389
1390 /* wake up and enable device */
1391 if (pci_enable_device(pdev) < 0)
1392 return -EIO;
1393
1394 return e752x_probe1(pdev, ent->driver_data);
1395 }
1396
e752x_remove_one(struct pci_dev * pdev)1397 static void e752x_remove_one(struct pci_dev *pdev)
1398 {
1399 struct mem_ctl_info *mci;
1400 struct e752x_pvt *pvt;
1401
1402 edac_dbg(0, "\n");
1403
1404 if (e752x_pci)
1405 edac_pci_release_generic_ctl(e752x_pci);
1406
1407 if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL)
1408 return;
1409
1410 pvt = (struct e752x_pvt *)mci->pvt_info;
1411 pci_dev_put(pvt->dev_d0f0);
1412 pci_dev_put(pvt->dev_d0f1);
1413 edac_mc_free(mci);
1414 }
1415
1416 static const struct pci_device_id e752x_pci_tbl[] = {
1417 {
1418 PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1419 E7520},
1420 {
1421 PCI_VEND_DEV(INTEL, 7525_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1422 E7525},
1423 {
1424 PCI_VEND_DEV(INTEL, 7320_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1425 E7320},
1426 {
1427 PCI_VEND_DEV(INTEL, 3100_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1428 I3100},
1429 {
1430 0,
1431 } /* 0 terminated list. */
1432 };
1433
1434 MODULE_DEVICE_TABLE(pci, e752x_pci_tbl);
1435
1436 static struct pci_driver e752x_driver = {
1437 .name = EDAC_MOD_STR,
1438 .probe = e752x_init_one,
1439 .remove = e752x_remove_one,
1440 .id_table = e752x_pci_tbl,
1441 };
1442
e752x_init(void)1443 static int __init e752x_init(void)
1444 {
1445 int pci_rc;
1446
1447 edac_dbg(3, "\n");
1448
1449 /* Ensure that the OPSTATE is set correctly for POLL or NMI */
1450 opstate_init();
1451
1452 pci_rc = pci_register_driver(&e752x_driver);
1453 return (pci_rc < 0) ? pci_rc : 0;
1454 }
1455
e752x_exit(void)1456 static void __exit e752x_exit(void)
1457 {
1458 edac_dbg(3, "\n");
1459 pci_unregister_driver(&e752x_driver);
1460 }
1461
1462 module_init(e752x_init);
1463 module_exit(e752x_exit);
1464
1465 MODULE_LICENSE("GPL");
1466 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Tom Zimmerman\n");
1467 MODULE_DESCRIPTION("MC support for Intel e752x/3100 memory controllers");
1468
1469 module_param(force_function_unhide, int, 0444);
1470 MODULE_PARM_DESC(force_function_unhide, "if BIOS sets Dev0:Fun1 up as hidden:"
1471 " 1=force unhide and hope BIOS doesn't fight driver for "
1472 "Dev0:Fun1 access");
1473
1474 module_param(edac_op_state, int, 0444);
1475 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
1476
1477 module_param(sysbus_parity, int, 0444);
1478 MODULE_PARM_DESC(sysbus_parity, "0=disable system bus parity checking,"
1479 " 1=enable system bus parity checking, default=auto-detect");
1480 module_param(report_non_memory_errors, int, 0644);
1481 MODULE_PARM_DESC(report_non_memory_errors, "0=disable non-memory error "
1482 "reporting, 1=enable non-memory error reporting");
1483