1 /*
2  * Intel e752x Memory Controller kernel module
3  * (C) 2004 Linux Networx (http://lnxi.com)
4  * This file may be distributed under the terms of the
5  * GNU General Public License.
6  *
7  * Implement support for the e7520, E7525, e7320 and i3100 memory controllers.
8  *
9  * Datasheets:
10  *	http://www.intel.in/content/www/in/en/chipsets/e7525-memory-controller-hub-datasheet.html
11  *	ftp://download.intel.com/design/intarch/datashts/31345803.pdf
12  *
13  * Written by Tom Zimmerman
14  *
15  * Contributors:
16  * 	Thayne Harbaugh at realmsys.com (?)
17  * 	Wang Zhenyu at intel.com
18  * 	Dave Jiang at mvista.com
19  *
20  */
21 
22 #include <linux/module.h>
23 #include <linux/init.h>
24 #include <linux/pci.h>
25 #include <linux/pci_ids.h>
26 #include <linux/edac.h>
27 #include "edac_module.h"
28 
29 #define EDAC_MOD_STR	"e752x_edac"
30 
31 static int report_non_memory_errors;
32 static int force_function_unhide;
33 static int sysbus_parity = -1;
34 
35 static struct edac_pci_ctl_info *e752x_pci;
36 
37 #define e752x_printk(level, fmt, arg...) \
38 	edac_printk(level, "e752x", fmt, ##arg)
39 
40 #define e752x_mc_printk(mci, level, fmt, arg...) \
41 	edac_mc_chipset_printk(mci, level, "e752x", fmt, ##arg)
42 
43 #ifndef PCI_DEVICE_ID_INTEL_7520_0
44 #define PCI_DEVICE_ID_INTEL_7520_0      0x3590
45 #endif				/* PCI_DEVICE_ID_INTEL_7520_0      */
46 
47 #ifndef PCI_DEVICE_ID_INTEL_7520_1_ERR
48 #define PCI_DEVICE_ID_INTEL_7520_1_ERR  0x3591
49 #endif				/* PCI_DEVICE_ID_INTEL_7520_1_ERR  */
50 
51 #ifndef PCI_DEVICE_ID_INTEL_7525_0
52 #define PCI_DEVICE_ID_INTEL_7525_0      0x359E
53 #endif				/* PCI_DEVICE_ID_INTEL_7525_0      */
54 
55 #ifndef PCI_DEVICE_ID_INTEL_7525_1_ERR
56 #define PCI_DEVICE_ID_INTEL_7525_1_ERR  0x3593
57 #endif				/* PCI_DEVICE_ID_INTEL_7525_1_ERR  */
58 
59 #ifndef PCI_DEVICE_ID_INTEL_7320_0
60 #define PCI_DEVICE_ID_INTEL_7320_0	0x3592
61 #endif				/* PCI_DEVICE_ID_INTEL_7320_0 */
62 
63 #ifndef PCI_DEVICE_ID_INTEL_7320_1_ERR
64 #define PCI_DEVICE_ID_INTEL_7320_1_ERR	0x3593
65 #endif				/* PCI_DEVICE_ID_INTEL_7320_1_ERR */
66 
67 #ifndef PCI_DEVICE_ID_INTEL_3100_0
68 #define PCI_DEVICE_ID_INTEL_3100_0	0x35B0
69 #endif				/* PCI_DEVICE_ID_INTEL_3100_0 */
70 
71 #ifndef PCI_DEVICE_ID_INTEL_3100_1_ERR
72 #define PCI_DEVICE_ID_INTEL_3100_1_ERR	0x35B1
73 #endif				/* PCI_DEVICE_ID_INTEL_3100_1_ERR */
74 
75 #define E752X_NR_CSROWS		8	/* number of csrows */
76 
77 /* E752X register addresses - device 0 function 0 */
78 #define E752X_MCHSCRB		0x52	/* Memory Scrub register (16b) */
79 					/*
80 					 * 6:5     Scrub Completion Count
81 					 * 3:2     Scrub Rate (i3100 only)
82 					 *      01=fast 10=normal
83 					 * 1:0     Scrub Mode enable
84 					 *      00=off 10=on
85 					 */
86 #define E752X_DRB		0x60	/* DRAM row boundary register (8b) */
87 #define E752X_DRA		0x70	/* DRAM row attribute register (8b) */
88 					/*
89 					 * 31:30   Device width row 7
90 					 *      01=x8 10=x4 11=x8 DDR2
91 					 * 27:26   Device width row 6
92 					 * 23:22   Device width row 5
93 					 * 19:20   Device width row 4
94 					 * 15:14   Device width row 3
95 					 * 11:10   Device width row 2
96 					 *  7:6    Device width row 1
97 					 *  3:2    Device width row 0
98 					 */
99 #define E752X_DRC		0x7C	/* DRAM controller mode reg (32b) */
100 					/* FIXME:IS THIS RIGHT? */
101 					/*
102 					 * 22    Number channels 0=1,1=2
103 					 * 19:18 DRB Granularity 32/64MB
104 					 */
105 #define E752X_DRM		0x80	/* Dimm mapping register */
106 #define E752X_DDRCSR		0x9A	/* DDR control and status reg (16b) */
107 					/*
108 					 * 14:12 1 single A, 2 single B, 3 dual
109 					 */
110 #define E752X_TOLM		0xC4	/* DRAM top of low memory reg (16b) */
111 #define E752X_REMAPBASE		0xC6	/* DRAM remap base address reg (16b) */
112 #define E752X_REMAPLIMIT	0xC8	/* DRAM remap limit address reg (16b) */
113 #define E752X_REMAPOFFSET	0xCA	/* DRAM remap limit offset reg (16b) */
114 
115 /* E752X register addresses - device 0 function 1 */
116 #define E752X_FERR_GLOBAL	0x40	/* Global first error register (32b) */
117 #define E752X_NERR_GLOBAL	0x44	/* Global next error register (32b) */
118 #define E752X_HI_FERR		0x50	/* Hub interface first error reg (8b) */
119 #define E752X_HI_NERR		0x52	/* Hub interface next error reg (8b) */
120 #define E752X_HI_ERRMASK	0x54	/* Hub interface error mask reg (8b) */
121 #define E752X_HI_SMICMD		0x5A	/* Hub interface SMI command reg (8b) */
122 #define E752X_SYSBUS_FERR	0x60	/* System buss first error reg (16b) */
123 #define E752X_SYSBUS_NERR	0x62	/* System buss next error reg (16b) */
124 #define E752X_SYSBUS_ERRMASK	0x64	/* System buss error mask reg (16b) */
125 #define E752X_SYSBUS_SMICMD	0x6A	/* System buss SMI command reg (16b) */
126 #define E752X_BUF_FERR		0x70	/* Memory buffer first error reg (8b) */
127 #define E752X_BUF_NERR		0x72	/* Memory buffer next error reg (8b) */
128 #define E752X_BUF_ERRMASK	0x74	/* Memory buffer error mask reg (8b) */
129 #define E752X_BUF_SMICMD	0x7A	/* Memory buffer SMI cmd reg (8b) */
130 #define E752X_DRAM_FERR		0x80	/* DRAM first error register (16b) */
131 #define E752X_DRAM_NERR		0x82	/* DRAM next error register (16b) */
132 #define E752X_DRAM_ERRMASK	0x84	/* DRAM error mask register (8b) */
133 #define E752X_DRAM_SMICMD	0x8A	/* DRAM SMI command register (8b) */
134 #define E752X_DRAM_RETR_ADD	0xAC	/* DRAM Retry address register (32b) */
135 #define E752X_DRAM_SEC1_ADD	0xA0	/* DRAM first correctable memory */
136 					/*     error address register (32b) */
137 					/*
138 					 * 31    Reserved
139 					 * 30:2  CE address (64 byte block 34:6
140 					 * 1     Reserved
141 					 * 0     HiLoCS
142 					 */
143 #define E752X_DRAM_SEC2_ADD	0xC8	/* DRAM first correctable memory */
144 					/*     error address register (32b) */
145 					/*
146 					 * 31    Reserved
147 					 * 30:2  CE address (64 byte block 34:6)
148 					 * 1     Reserved
149 					 * 0     HiLoCS
150 					 */
151 #define E752X_DRAM_DED_ADD	0xA4	/* DRAM first uncorrectable memory */
152 					/*     error address register (32b) */
153 					/*
154 					 * 31    Reserved
155 					 * 30:2  CE address (64 byte block 34:6)
156 					 * 1     Reserved
157 					 * 0     HiLoCS
158 					 */
159 #define E752X_DRAM_SCRB_ADD	0xA8	/* DRAM 1st uncorrectable scrub mem */
160 					/*     error address register (32b) */
161 					/*
162 					 * 31    Reserved
163 					 * 30:2  CE address (64 byte block 34:6
164 					 * 1     Reserved
165 					 * 0     HiLoCS
166 					 */
167 #define E752X_DRAM_SEC1_SYNDROME 0xC4	/* DRAM first correctable memory */
168 					/*     error syndrome register (16b) */
169 #define E752X_DRAM_SEC2_SYNDROME 0xC6	/* DRAM second correctable memory */
170 					/*     error syndrome register (16b) */
171 #define E752X_DEVPRES1		0xF4	/* Device Present 1 register (8b) */
172 
173 /* 3100 IMCH specific register addresses - device 0 function 1 */
174 #define I3100_NSI_FERR		0x48	/* NSI first error reg (32b) */
175 #define I3100_NSI_NERR		0x4C	/* NSI next error reg (32b) */
176 #define I3100_NSI_SMICMD	0x54	/* NSI SMI command register (32b) */
177 #define I3100_NSI_EMASK		0x90	/* NSI error mask register (32b) */
178 
179 /* ICH5R register addresses - device 30 function 0 */
180 #define ICH5R_PCI_STAT		0x06	/* PCI status register (16b) */
181 #define ICH5R_PCI_2ND_STAT	0x1E	/* PCI status secondary reg (16b) */
182 #define ICH5R_PCI_BRIDGE_CTL	0x3E	/* PCI bridge control register (16b) */
183 
184 enum e752x_chips {
185 	E7520 = 0,
186 	E7525 = 1,
187 	E7320 = 2,
188 	I3100 = 3
189 };
190 
191 /*
192  * Those chips Support single-rank and dual-rank memories only.
193  *
194  * On e752x chips, the odd rows are present only on dual-rank memories.
195  * Dividing the rank by two will provide the dimm#
196  *
197  * i3100 MC has a different mapping: it supports only 4 ranks.
198  *
199  * The mapping is (from 1 to n):
200  *	slot	   single-ranked	double-ranked
201  *	dimm #1 -> rank #4		NA
202  *	dimm #2 -> rank #3		NA
203  *	dimm #3 -> rank #2		Ranks 2 and 3
204  *	dimm #4 -> rank $1		Ranks 1 and 4
205  *
206  * FIXME: The current mapping for i3100 considers that it supports up to 8
207  *	  ranks/chanel, but datasheet says that the MC supports only 4 ranks.
208  */
209 
210 struct e752x_pvt {
211 	struct pci_dev *dev_d0f0;
212 	struct pci_dev *dev_d0f1;
213 	u32 tolm;
214 	u32 remapbase;
215 	u32 remaplimit;
216 	int mc_symmetric;
217 	u8 map[8];
218 	int map_type;
219 	const struct e752x_dev_info *dev_info;
220 };
221 
222 struct e752x_dev_info {
223 	u16 err_dev;
224 	u16 ctl_dev;
225 	const char *ctl_name;
226 };
227 
228 struct e752x_error_info {
229 	u32 ferr_global;
230 	u32 nerr_global;
231 	u32 nsi_ferr;	/* 3100 only */
232 	u32 nsi_nerr;	/* 3100 only */
233 	u8 hi_ferr;	/* all but 3100 */
234 	u8 hi_nerr;	/* all but 3100 */
235 	u16 sysbus_ferr;
236 	u16 sysbus_nerr;
237 	u8 buf_ferr;
238 	u8 buf_nerr;
239 	u16 dram_ferr;
240 	u16 dram_nerr;
241 	u32 dram_sec1_add;
242 	u32 dram_sec2_add;
243 	u16 dram_sec1_syndrome;
244 	u16 dram_sec2_syndrome;
245 	u32 dram_ded_add;
246 	u32 dram_scrb_add;
247 	u32 dram_retr_add;
248 };
249 
250 static const struct e752x_dev_info e752x_devs[] = {
251 	[E7520] = {
252 		.err_dev = PCI_DEVICE_ID_INTEL_7520_1_ERR,
253 		.ctl_dev = PCI_DEVICE_ID_INTEL_7520_0,
254 		.ctl_name = "E7520"},
255 	[E7525] = {
256 		.err_dev = PCI_DEVICE_ID_INTEL_7525_1_ERR,
257 		.ctl_dev = PCI_DEVICE_ID_INTEL_7525_0,
258 		.ctl_name = "E7525"},
259 	[E7320] = {
260 		.err_dev = PCI_DEVICE_ID_INTEL_7320_1_ERR,
261 		.ctl_dev = PCI_DEVICE_ID_INTEL_7320_0,
262 		.ctl_name = "E7320"},
263 	[I3100] = {
264 		.err_dev = PCI_DEVICE_ID_INTEL_3100_1_ERR,
265 		.ctl_dev = PCI_DEVICE_ID_INTEL_3100_0,
266 		.ctl_name = "3100"},
267 };
268 
269 /* Valid scrub rates for the e752x/3100 hardware memory scrubber. We
270  * map the scrubbing bandwidth to a hardware register value. The 'set'
271  * operation finds the 'matching or higher value'.  Note that scrubbing
272  * on the e752x can only be enabled/disabled.  The 3100 supports
273  * a normal and fast mode.
274  */
275 
276 #define SDRATE_EOT 0xFFFFFFFF
277 
278 struct scrubrate {
279 	u32 bandwidth;	/* bandwidth consumed by scrubbing in bytes/sec */
280 	u16 scrubval;	/* register value for scrub rate */
281 };
282 
283 /* Rate below assumes same performance as i3100 using PC3200 DDR2 in
284  * normal mode.  e752x bridges don't support choosing normal or fast mode,
285  * so the scrubbing bandwidth value isn't all that important - scrubbing is
286  * either on or off.
287  */
288 static const struct scrubrate scrubrates_e752x[] = {
289 	{0,		0x00},	/* Scrubbing Off */
290 	{500000,	0x02},	/* Scrubbing On */
291 	{SDRATE_EOT,	0x00}	/* End of Table */
292 };
293 
294 /* Fast mode: 2 GByte PC3200 DDR2 scrubbed in 33s = 63161283 bytes/s
295  * Normal mode: 125 (32000 / 256) times slower than fast mode.
296  */
297 static const struct scrubrate scrubrates_i3100[] = {
298 	{0,		0x00},	/* Scrubbing Off */
299 	{500000,	0x0a},	/* Normal mode - 32k clocks */
300 	{62500000,	0x06},	/* Fast mode - 256 clocks */
301 	{SDRATE_EOT,	0x00}	/* End of Table */
302 };
303 
ctl_page_to_phys(struct mem_ctl_info * mci,unsigned long page)304 static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci,
305 				unsigned long page)
306 {
307 	u32 remap;
308 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
309 
310 	edac_dbg(3, "\n");
311 
312 	if (page < pvt->tolm)
313 		return page;
314 
315 	if ((page >= 0x100000) && (page < pvt->remapbase))
316 		return page;
317 
318 	remap = (page - pvt->tolm) + pvt->remapbase;
319 
320 	if (remap < pvt->remaplimit)
321 		return remap;
322 
323 	e752x_printk(KERN_ERR, "Invalid page %lx - out of range\n", page);
324 	return pvt->tolm - 1;
325 }
326 
do_process_ce(struct mem_ctl_info * mci,u16 error_one,u32 sec1_add,u16 sec1_syndrome)327 static void do_process_ce(struct mem_ctl_info *mci, u16 error_one,
328 			u32 sec1_add, u16 sec1_syndrome)
329 {
330 	u32 page;
331 	int row;
332 	int channel;
333 	int i;
334 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
335 
336 	edac_dbg(3, "\n");
337 
338 	/* convert the addr to 4k page */
339 	page = sec1_add >> (PAGE_SHIFT - 4);
340 
341 	/* FIXME - check for -1 */
342 	if (pvt->mc_symmetric) {
343 		/* chip select are bits 14 & 13 */
344 		row = ((page >> 1) & 3);
345 		e752x_printk(KERN_WARNING,
346 			"Test row %d Table %d %d %d %d %d %d %d %d\n", row,
347 			pvt->map[0], pvt->map[1], pvt->map[2], pvt->map[3],
348 			pvt->map[4], pvt->map[5], pvt->map[6],
349 			pvt->map[7]);
350 
351 		/* test for channel remapping */
352 		for (i = 0; i < 8; i++) {
353 			if (pvt->map[i] == row)
354 				break;
355 		}
356 
357 		e752x_printk(KERN_WARNING, "Test computed row %d\n", i);
358 
359 		if (i < 8)
360 			row = i;
361 		else
362 			e752x_mc_printk(mci, KERN_WARNING,
363 					"row %d not found in remap table\n",
364 					row);
365 	} else
366 		row = edac_mc_find_csrow_by_page(mci, page);
367 
368 	/* 0 = channel A, 1 = channel B */
369 	channel = !(error_one & 1);
370 
371 	/* e752x mc reads 34:6 of the DRAM linear address */
372 	edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1,
373 			     page, offset_in_page(sec1_add << 4), sec1_syndrome,
374 			     row, channel, -1,
375 			     "e752x CE", "");
376 }
377 
process_ce(struct mem_ctl_info * mci,u16 error_one,u32 sec1_add,u16 sec1_syndrome,int * error_found,int handle_error)378 static inline void process_ce(struct mem_ctl_info *mci, u16 error_one,
379 			u32 sec1_add, u16 sec1_syndrome, int *error_found,
380 			int handle_error)
381 {
382 	*error_found = 1;
383 
384 	if (handle_error)
385 		do_process_ce(mci, error_one, sec1_add, sec1_syndrome);
386 }
387 
do_process_ue(struct mem_ctl_info * mci,u16 error_one,u32 ded_add,u32 scrb_add)388 static void do_process_ue(struct mem_ctl_info *mci, u16 error_one,
389 			u32 ded_add, u32 scrb_add)
390 {
391 	u32 error_2b, block_page;
392 	int row;
393 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
394 
395 	edac_dbg(3, "\n");
396 
397 	if (error_one & 0x0202) {
398 		error_2b = ded_add;
399 
400 		/* convert to 4k address */
401 		block_page = error_2b >> (PAGE_SHIFT - 4);
402 
403 		row = pvt->mc_symmetric ?
404 		/* chip select are bits 14 & 13 */
405 			((block_page >> 1) & 3) :
406 			edac_mc_find_csrow_by_page(mci, block_page);
407 
408 		/* e752x mc reads 34:6 of the DRAM linear address */
409 		edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
410 					block_page,
411 					offset_in_page(error_2b << 4), 0,
412 					 row, -1, -1,
413 					"e752x UE from Read", "");
414 
415 	}
416 	if (error_one & 0x0404) {
417 		error_2b = scrb_add;
418 
419 		/* convert to 4k address */
420 		block_page = error_2b >> (PAGE_SHIFT - 4);
421 
422 		row = pvt->mc_symmetric ?
423 		/* chip select are bits 14 & 13 */
424 			((block_page >> 1) & 3) :
425 			edac_mc_find_csrow_by_page(mci, block_page);
426 
427 		/* e752x mc reads 34:6 of the DRAM linear address */
428 		edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
429 					block_page,
430 					offset_in_page(error_2b << 4), 0,
431 					row, -1, -1,
432 					"e752x UE from Scruber", "");
433 	}
434 }
435 
process_ue(struct mem_ctl_info * mci,u16 error_one,u32 ded_add,u32 scrb_add,int * error_found,int handle_error)436 static inline void process_ue(struct mem_ctl_info *mci, u16 error_one,
437 			u32 ded_add, u32 scrb_add, int *error_found,
438 			int handle_error)
439 {
440 	*error_found = 1;
441 
442 	if (handle_error)
443 		do_process_ue(mci, error_one, ded_add, scrb_add);
444 }
445 
process_ue_no_info_wr(struct mem_ctl_info * mci,int * error_found,int handle_error)446 static inline void process_ue_no_info_wr(struct mem_ctl_info *mci,
447 					 int *error_found, int handle_error)
448 {
449 	*error_found = 1;
450 
451 	if (!handle_error)
452 		return;
453 
454 	edac_dbg(3, "\n");
455 	edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, 0, 0, 0,
456 			     -1, -1, -1,
457 			     "e752x UE log memory write", "");
458 }
459 
do_process_ded_retry(struct mem_ctl_info * mci,u16 error,u32 retry_add)460 static void do_process_ded_retry(struct mem_ctl_info *mci, u16 error,
461 				 u32 retry_add)
462 {
463 	u32 error_1b, page;
464 	int row;
465 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
466 
467 	error_1b = retry_add;
468 	page = error_1b >> (PAGE_SHIFT - 4);  /* convert the addr to 4k page */
469 
470 	/* chip select are bits 14 & 13 */
471 	row = pvt->mc_symmetric ? ((page >> 1) & 3) :
472 		edac_mc_find_csrow_by_page(mci, page);
473 
474 	e752x_mc_printk(mci, KERN_WARNING,
475 			"CE page 0x%lx, row %d : Memory read retry\n",
476 			(long unsigned int)page, row);
477 }
478 
process_ded_retry(struct mem_ctl_info * mci,u16 error,u32 retry_add,int * error_found,int handle_error)479 static inline void process_ded_retry(struct mem_ctl_info *mci, u16 error,
480 				u32 retry_add, int *error_found,
481 				int handle_error)
482 {
483 	*error_found = 1;
484 
485 	if (handle_error)
486 		do_process_ded_retry(mci, error, retry_add);
487 }
488 
process_threshold_ce(struct mem_ctl_info * mci,u16 error,int * error_found,int handle_error)489 static inline void process_threshold_ce(struct mem_ctl_info *mci, u16 error,
490 					int *error_found, int handle_error)
491 {
492 	*error_found = 1;
493 
494 	if (handle_error)
495 		e752x_mc_printk(mci, KERN_WARNING, "Memory threshold CE\n");
496 }
497 
498 static char *global_message[11] = {
499 	"PCI Express C1",
500 	"PCI Express C",
501 	"PCI Express B1",
502 	"PCI Express B",
503 	"PCI Express A1",
504 	"PCI Express A",
505 	"DMA Controller",
506 	"HUB or NS Interface",
507 	"System Bus",
508 	"DRAM Controller",  /* 9th entry */
509 	"Internal Buffer"
510 };
511 
512 #define DRAM_ENTRY	9
513 
514 static char *fatal_message[2] = { "Non-Fatal ", "Fatal " };
515 
do_global_error(int fatal,u32 errors)516 static void do_global_error(int fatal, u32 errors)
517 {
518 	int i;
519 
520 	for (i = 0; i < 11; i++) {
521 		if (errors & (1 << i)) {
522 			/* If the error is from DRAM Controller OR
523 			 * we are to report ALL errors, then
524 			 * report the error
525 			 */
526 			if ((i == DRAM_ENTRY) || report_non_memory_errors)
527 				e752x_printk(KERN_WARNING, "%sError %s\n",
528 					fatal_message[fatal],
529 					global_message[i]);
530 		}
531 	}
532 }
533 
global_error(int fatal,u32 errors,int * error_found,int handle_error)534 static inline void global_error(int fatal, u32 errors, int *error_found,
535 				int handle_error)
536 {
537 	*error_found = 1;
538 
539 	if (handle_error)
540 		do_global_error(fatal, errors);
541 }
542 
543 static char *hub_message[7] = {
544 	"HI Address or Command Parity", "HI Illegal Access",
545 	"HI Internal Parity", "Out of Range Access",
546 	"HI Data Parity", "Enhanced Config Access",
547 	"Hub Interface Target Abort"
548 };
549 
do_hub_error(int fatal,u8 errors)550 static void do_hub_error(int fatal, u8 errors)
551 {
552 	int i;
553 
554 	for (i = 0; i < 7; i++) {
555 		if (errors & (1 << i))
556 			e752x_printk(KERN_WARNING, "%sError %s\n",
557 				fatal_message[fatal], hub_message[i]);
558 	}
559 }
560 
hub_error(int fatal,u8 errors,int * error_found,int handle_error)561 static inline void hub_error(int fatal, u8 errors, int *error_found,
562 			int handle_error)
563 {
564 	*error_found = 1;
565 
566 	if (handle_error)
567 		do_hub_error(fatal, errors);
568 }
569 
570 #define NSI_FATAL_MASK		0x0c080081
571 #define NSI_NON_FATAL_MASK	0x23a0ba64
572 #define NSI_ERR_MASK		(NSI_FATAL_MASK | NSI_NON_FATAL_MASK)
573 
574 static char *nsi_message[30] = {
575 	"NSI Link Down",	/* NSI_FERR/NSI_NERR bit 0, fatal error */
576 	"",						/* reserved */
577 	"NSI Parity Error",				/* bit 2, non-fatal */
578 	"",						/* reserved */
579 	"",						/* reserved */
580 	"Correctable Error Message",			/* bit 5, non-fatal */
581 	"Non-Fatal Error Message",			/* bit 6, non-fatal */
582 	"Fatal Error Message",				/* bit 7, fatal */
583 	"",						/* reserved */
584 	"Receiver Error",				/* bit 9, non-fatal */
585 	"",						/* reserved */
586 	"Bad TLP",					/* bit 11, non-fatal */
587 	"Bad DLLP",					/* bit 12, non-fatal */
588 	"REPLAY_NUM Rollover",				/* bit 13, non-fatal */
589 	"",						/* reserved */
590 	"Replay Timer Timeout",				/* bit 15, non-fatal */
591 	"",						/* reserved */
592 	"",						/* reserved */
593 	"",						/* reserved */
594 	"Data Link Protocol Error",			/* bit 19, fatal */
595 	"",						/* reserved */
596 	"Poisoned TLP",					/* bit 21, non-fatal */
597 	"",						/* reserved */
598 	"Completion Timeout",				/* bit 23, non-fatal */
599 	"Completer Abort",				/* bit 24, non-fatal */
600 	"Unexpected Completion",			/* bit 25, non-fatal */
601 	"Receiver Overflow",				/* bit 26, fatal */
602 	"Malformed TLP",				/* bit 27, fatal */
603 	"",						/* reserved */
604 	"Unsupported Request"				/* bit 29, non-fatal */
605 };
606 
do_nsi_error(int fatal,u32 errors)607 static void do_nsi_error(int fatal, u32 errors)
608 {
609 	int i;
610 
611 	for (i = 0; i < 30; i++) {
612 		if (errors & (1 << i))
613 			printk(KERN_WARNING "%sError %s\n",
614 			       fatal_message[fatal], nsi_message[i]);
615 	}
616 }
617 
nsi_error(int fatal,u32 errors,int * error_found,int handle_error)618 static inline void nsi_error(int fatal, u32 errors, int *error_found,
619 		int handle_error)
620 {
621 	*error_found = 1;
622 
623 	if (handle_error)
624 		do_nsi_error(fatal, errors);
625 }
626 
627 static char *membuf_message[4] = {
628 	"Internal PMWB to DRAM parity",
629 	"Internal PMWB to System Bus Parity",
630 	"Internal System Bus or IO to PMWB Parity",
631 	"Internal DRAM to PMWB Parity"
632 };
633 
do_membuf_error(u8 errors)634 static void do_membuf_error(u8 errors)
635 {
636 	int i;
637 
638 	for (i = 0; i < 4; i++) {
639 		if (errors & (1 << i))
640 			e752x_printk(KERN_WARNING, "Non-Fatal Error %s\n",
641 				membuf_message[i]);
642 	}
643 }
644 
membuf_error(u8 errors,int * error_found,int handle_error)645 static inline void membuf_error(u8 errors, int *error_found, int handle_error)
646 {
647 	*error_found = 1;
648 
649 	if (handle_error)
650 		do_membuf_error(errors);
651 }
652 
653 static char *sysbus_message[10] = {
654 	"Addr or Request Parity",
655 	"Data Strobe Glitch",
656 	"Addr Strobe Glitch",
657 	"Data Parity",
658 	"Addr Above TOM",
659 	"Non DRAM Lock Error",
660 	"MCERR", "BINIT",
661 	"Memory Parity",
662 	"IO Subsystem Parity"
663 };
664 
do_sysbus_error(int fatal,u32 errors)665 static void do_sysbus_error(int fatal, u32 errors)
666 {
667 	int i;
668 
669 	for (i = 0; i < 10; i++) {
670 		if (errors & (1 << i))
671 			e752x_printk(KERN_WARNING, "%sError System Bus %s\n",
672 				fatal_message[fatal], sysbus_message[i]);
673 	}
674 }
675 
sysbus_error(int fatal,u32 errors,int * error_found,int handle_error)676 static inline void sysbus_error(int fatal, u32 errors, int *error_found,
677 				int handle_error)
678 {
679 	*error_found = 1;
680 
681 	if (handle_error)
682 		do_sysbus_error(fatal, errors);
683 }
684 
e752x_check_hub_interface(struct e752x_error_info * info,int * error_found,int handle_error)685 static void e752x_check_hub_interface(struct e752x_error_info *info,
686 				int *error_found, int handle_error)
687 {
688 	u8 stat8;
689 
690 	//pci_read_config_byte(dev,E752X_HI_FERR,&stat8);
691 
692 	stat8 = info->hi_ferr;
693 
694 	if (stat8 & 0x7f) {	/* Error, so process */
695 		stat8 &= 0x7f;
696 
697 		if (stat8 & 0x2b)
698 			hub_error(1, stat8 & 0x2b, error_found, handle_error);
699 
700 		if (stat8 & 0x54)
701 			hub_error(0, stat8 & 0x54, error_found, handle_error);
702 	}
703 	//pci_read_config_byte(dev,E752X_HI_NERR,&stat8);
704 
705 	stat8 = info->hi_nerr;
706 
707 	if (stat8 & 0x7f) {	/* Error, so process */
708 		stat8 &= 0x7f;
709 
710 		if (stat8 & 0x2b)
711 			hub_error(1, stat8 & 0x2b, error_found, handle_error);
712 
713 		if (stat8 & 0x54)
714 			hub_error(0, stat8 & 0x54, error_found, handle_error);
715 	}
716 }
717 
e752x_check_ns_interface(struct e752x_error_info * info,int * error_found,int handle_error)718 static void e752x_check_ns_interface(struct e752x_error_info *info,
719 				int *error_found, int handle_error)
720 {
721 	u32 stat32;
722 
723 	stat32 = info->nsi_ferr;
724 	if (stat32 & NSI_ERR_MASK) { /* Error, so process */
725 		if (stat32 & NSI_FATAL_MASK)	/* check for fatal errors */
726 			nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
727 				  handle_error);
728 		if (stat32 & NSI_NON_FATAL_MASK) /* check for non-fatal ones */
729 			nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
730 				  handle_error);
731 	}
732 	stat32 = info->nsi_nerr;
733 	if (stat32 & NSI_ERR_MASK) {
734 		if (stat32 & NSI_FATAL_MASK)
735 			nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
736 				  handle_error);
737 		if (stat32 & NSI_NON_FATAL_MASK)
738 			nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
739 				  handle_error);
740 	}
741 }
742 
e752x_check_sysbus(struct e752x_error_info * info,int * error_found,int handle_error)743 static void e752x_check_sysbus(struct e752x_error_info *info,
744 			int *error_found, int handle_error)
745 {
746 	u32 stat32, error32;
747 
748 	//pci_read_config_dword(dev,E752X_SYSBUS_FERR,&stat32);
749 	stat32 = info->sysbus_ferr + (info->sysbus_nerr << 16);
750 
751 	if (stat32 == 0)
752 		return;		/* no errors */
753 
754 	error32 = (stat32 >> 16) & 0x3ff;
755 	stat32 = stat32 & 0x3ff;
756 
757 	if (stat32 & 0x087)
758 		sysbus_error(1, stat32 & 0x087, error_found, handle_error);
759 
760 	if (stat32 & 0x378)
761 		sysbus_error(0, stat32 & 0x378, error_found, handle_error);
762 
763 	if (error32 & 0x087)
764 		sysbus_error(1, error32 & 0x087, error_found, handle_error);
765 
766 	if (error32 & 0x378)
767 		sysbus_error(0, error32 & 0x378, error_found, handle_error);
768 }
769 
e752x_check_membuf(struct e752x_error_info * info,int * error_found,int handle_error)770 static void e752x_check_membuf(struct e752x_error_info *info,
771 			int *error_found, int handle_error)
772 {
773 	u8 stat8;
774 
775 	stat8 = info->buf_ferr;
776 
777 	if (stat8 & 0x0f) {	/* Error, so process */
778 		stat8 &= 0x0f;
779 		membuf_error(stat8, error_found, handle_error);
780 	}
781 
782 	stat8 = info->buf_nerr;
783 
784 	if (stat8 & 0x0f) {	/* Error, so process */
785 		stat8 &= 0x0f;
786 		membuf_error(stat8, error_found, handle_error);
787 	}
788 }
789 
e752x_check_dram(struct mem_ctl_info * mci,struct e752x_error_info * info,int * error_found,int handle_error)790 static void e752x_check_dram(struct mem_ctl_info *mci,
791 			struct e752x_error_info *info, int *error_found,
792 			int handle_error)
793 {
794 	u16 error_one, error_next;
795 
796 	error_one = info->dram_ferr;
797 	error_next = info->dram_nerr;
798 
799 	/* decode and report errors */
800 	if (error_one & 0x0101)	/* check first error correctable */
801 		process_ce(mci, error_one, info->dram_sec1_add,
802 			info->dram_sec1_syndrome, error_found, handle_error);
803 
804 	if (error_next & 0x0101)	/* check next error correctable */
805 		process_ce(mci, error_next, info->dram_sec2_add,
806 			info->dram_sec2_syndrome, error_found, handle_error);
807 
808 	if (error_one & 0x4040)
809 		process_ue_no_info_wr(mci, error_found, handle_error);
810 
811 	if (error_next & 0x4040)
812 		process_ue_no_info_wr(mci, error_found, handle_error);
813 
814 	if (error_one & 0x2020)
815 		process_ded_retry(mci, error_one, info->dram_retr_add,
816 				error_found, handle_error);
817 
818 	if (error_next & 0x2020)
819 		process_ded_retry(mci, error_next, info->dram_retr_add,
820 				error_found, handle_error);
821 
822 	if (error_one & 0x0808)
823 		process_threshold_ce(mci, error_one, error_found, handle_error);
824 
825 	if (error_next & 0x0808)
826 		process_threshold_ce(mci, error_next, error_found,
827 				handle_error);
828 
829 	if (error_one & 0x0606)
830 		process_ue(mci, error_one, info->dram_ded_add,
831 			info->dram_scrb_add, error_found, handle_error);
832 
833 	if (error_next & 0x0606)
834 		process_ue(mci, error_next, info->dram_ded_add,
835 			info->dram_scrb_add, error_found, handle_error);
836 }
837 
e752x_get_error_info(struct mem_ctl_info * mci,struct e752x_error_info * info)838 static void e752x_get_error_info(struct mem_ctl_info *mci,
839 				 struct e752x_error_info *info)
840 {
841 	struct pci_dev *dev;
842 	struct e752x_pvt *pvt;
843 
844 	memset(info, 0, sizeof(*info));
845 	pvt = (struct e752x_pvt *)mci->pvt_info;
846 	dev = pvt->dev_d0f1;
847 	pci_read_config_dword(dev, E752X_FERR_GLOBAL, &info->ferr_global);
848 
849 	if (info->ferr_global) {
850 		if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
851 			pci_read_config_dword(dev, I3100_NSI_FERR,
852 					     &info->nsi_ferr);
853 			info->hi_ferr = 0;
854 		} else {
855 			pci_read_config_byte(dev, E752X_HI_FERR,
856 					     &info->hi_ferr);
857 			info->nsi_ferr = 0;
858 		}
859 		pci_read_config_word(dev, E752X_SYSBUS_FERR,
860 				&info->sysbus_ferr);
861 		pci_read_config_byte(dev, E752X_BUF_FERR, &info->buf_ferr);
862 		pci_read_config_word(dev, E752X_DRAM_FERR, &info->dram_ferr);
863 		pci_read_config_dword(dev, E752X_DRAM_SEC1_ADD,
864 				&info->dram_sec1_add);
865 		pci_read_config_word(dev, E752X_DRAM_SEC1_SYNDROME,
866 				&info->dram_sec1_syndrome);
867 		pci_read_config_dword(dev, E752X_DRAM_DED_ADD,
868 				&info->dram_ded_add);
869 		pci_read_config_dword(dev, E752X_DRAM_SCRB_ADD,
870 				&info->dram_scrb_add);
871 		pci_read_config_dword(dev, E752X_DRAM_RETR_ADD,
872 				&info->dram_retr_add);
873 
874 		/* ignore the reserved bits just in case */
875 		if (info->hi_ferr & 0x7f)
876 			pci_write_config_byte(dev, E752X_HI_FERR,
877 					info->hi_ferr);
878 
879 		if (info->nsi_ferr & NSI_ERR_MASK)
880 			pci_write_config_dword(dev, I3100_NSI_FERR,
881 					info->nsi_ferr);
882 
883 		if (info->sysbus_ferr)
884 			pci_write_config_word(dev, E752X_SYSBUS_FERR,
885 					info->sysbus_ferr);
886 
887 		if (info->buf_ferr & 0x0f)
888 			pci_write_config_byte(dev, E752X_BUF_FERR,
889 					info->buf_ferr);
890 
891 		if (info->dram_ferr)
892 			pci_write_bits16(pvt->dev_d0f1, E752X_DRAM_FERR,
893 					 info->dram_ferr, info->dram_ferr);
894 
895 		pci_write_config_dword(dev, E752X_FERR_GLOBAL,
896 				info->ferr_global);
897 	}
898 
899 	pci_read_config_dword(dev, E752X_NERR_GLOBAL, &info->nerr_global);
900 
901 	if (info->nerr_global) {
902 		if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
903 			pci_read_config_dword(dev, I3100_NSI_NERR,
904 					     &info->nsi_nerr);
905 			info->hi_nerr = 0;
906 		} else {
907 			pci_read_config_byte(dev, E752X_HI_NERR,
908 					     &info->hi_nerr);
909 			info->nsi_nerr = 0;
910 		}
911 		pci_read_config_word(dev, E752X_SYSBUS_NERR,
912 				&info->sysbus_nerr);
913 		pci_read_config_byte(dev, E752X_BUF_NERR, &info->buf_nerr);
914 		pci_read_config_word(dev, E752X_DRAM_NERR, &info->dram_nerr);
915 		pci_read_config_dword(dev, E752X_DRAM_SEC2_ADD,
916 				&info->dram_sec2_add);
917 		pci_read_config_word(dev, E752X_DRAM_SEC2_SYNDROME,
918 				&info->dram_sec2_syndrome);
919 
920 		if (info->hi_nerr & 0x7f)
921 			pci_write_config_byte(dev, E752X_HI_NERR,
922 					info->hi_nerr);
923 
924 		if (info->nsi_nerr & NSI_ERR_MASK)
925 			pci_write_config_dword(dev, I3100_NSI_NERR,
926 					info->nsi_nerr);
927 
928 		if (info->sysbus_nerr)
929 			pci_write_config_word(dev, E752X_SYSBUS_NERR,
930 					info->sysbus_nerr);
931 
932 		if (info->buf_nerr & 0x0f)
933 			pci_write_config_byte(dev, E752X_BUF_NERR,
934 					info->buf_nerr);
935 
936 		if (info->dram_nerr)
937 			pci_write_bits16(pvt->dev_d0f1, E752X_DRAM_NERR,
938 					 info->dram_nerr, info->dram_nerr);
939 
940 		pci_write_config_dword(dev, E752X_NERR_GLOBAL,
941 				info->nerr_global);
942 	}
943 }
944 
e752x_process_error_info(struct mem_ctl_info * mci,struct e752x_error_info * info,int handle_errors)945 static int e752x_process_error_info(struct mem_ctl_info *mci,
946 				struct e752x_error_info *info,
947 				int handle_errors)
948 {
949 	u32 error32, stat32;
950 	int error_found;
951 
952 	error_found = 0;
953 	error32 = (info->ferr_global >> 18) & 0x3ff;
954 	stat32 = (info->ferr_global >> 4) & 0x7ff;
955 
956 	if (error32)
957 		global_error(1, error32, &error_found, handle_errors);
958 
959 	if (stat32)
960 		global_error(0, stat32, &error_found, handle_errors);
961 
962 	error32 = (info->nerr_global >> 18) & 0x3ff;
963 	stat32 = (info->nerr_global >> 4) & 0x7ff;
964 
965 	if (error32)
966 		global_error(1, error32, &error_found, handle_errors);
967 
968 	if (stat32)
969 		global_error(0, stat32, &error_found, handle_errors);
970 
971 	e752x_check_hub_interface(info, &error_found, handle_errors);
972 	e752x_check_ns_interface(info, &error_found, handle_errors);
973 	e752x_check_sysbus(info, &error_found, handle_errors);
974 	e752x_check_membuf(info, &error_found, handle_errors);
975 	e752x_check_dram(mci, info, &error_found, handle_errors);
976 	return error_found;
977 }
978 
e752x_check(struct mem_ctl_info * mci)979 static void e752x_check(struct mem_ctl_info *mci)
980 {
981 	struct e752x_error_info info;
982 
983 	edac_dbg(3, "\n");
984 	e752x_get_error_info(mci, &info);
985 	e752x_process_error_info(mci, &info, 1);
986 }
987 
988 /* Program byte/sec bandwidth scrub rate to hardware */
set_sdram_scrub_rate(struct mem_ctl_info * mci,u32 new_bw)989 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
990 {
991 	const struct scrubrate *scrubrates;
992 	struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
993 	struct pci_dev *pdev = pvt->dev_d0f0;
994 	int i;
995 
996 	if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
997 		scrubrates = scrubrates_i3100;
998 	else
999 		scrubrates = scrubrates_e752x;
1000 
1001 	/* Translate the desired scrub rate to a e752x/3100 register value.
1002 	 * Search for the bandwidth that is equal or greater than the
1003 	 * desired rate and program the cooresponding register value.
1004 	 */
1005 	for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1006 		if (scrubrates[i].bandwidth >= new_bw)
1007 			break;
1008 
1009 	if (scrubrates[i].bandwidth == SDRATE_EOT)
1010 		return -1;
1011 
1012 	pci_write_config_word(pdev, E752X_MCHSCRB, scrubrates[i].scrubval);
1013 
1014 	return scrubrates[i].bandwidth;
1015 }
1016 
1017 /* Convert current scrub rate value into byte/sec bandwidth */
get_sdram_scrub_rate(struct mem_ctl_info * mci)1018 static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
1019 {
1020 	const struct scrubrate *scrubrates;
1021 	struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
1022 	struct pci_dev *pdev = pvt->dev_d0f0;
1023 	u16 scrubval;
1024 	int i;
1025 
1026 	if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
1027 		scrubrates = scrubrates_i3100;
1028 	else
1029 		scrubrates = scrubrates_e752x;
1030 
1031 	/* Find the bandwidth matching the memory scrubber configuration */
1032 	pci_read_config_word(pdev, E752X_MCHSCRB, &scrubval);
1033 	scrubval = scrubval & 0x0f;
1034 
1035 	for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1036 		if (scrubrates[i].scrubval == scrubval)
1037 			break;
1038 
1039 	if (scrubrates[i].bandwidth == SDRATE_EOT) {
1040 		e752x_printk(KERN_WARNING,
1041 			"Invalid sdram scrub control value: 0x%x\n", scrubval);
1042 		return -1;
1043 	}
1044 	return scrubrates[i].bandwidth;
1045 
1046 }
1047 
1048 /* Return 1 if dual channel mode is active.  Else return 0. */
dual_channel_active(u16 ddrcsr)1049 static inline int dual_channel_active(u16 ddrcsr)
1050 {
1051 	return (((ddrcsr >> 12) & 3) == 3);
1052 }
1053 
1054 /* Remap csrow index numbers if map_type is "reverse"
1055  */
remap_csrow_index(struct mem_ctl_info * mci,int index)1056 static inline int remap_csrow_index(struct mem_ctl_info *mci, int index)
1057 {
1058 	struct e752x_pvt *pvt = mci->pvt_info;
1059 
1060 	if (!pvt->map_type)
1061 		return (7 - index);
1062 
1063 	return (index);
1064 }
1065 
e752x_init_csrows(struct mem_ctl_info * mci,struct pci_dev * pdev,u16 ddrcsr)1066 static void e752x_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev,
1067 			u16 ddrcsr)
1068 {
1069 	struct csrow_info *csrow;
1070 	enum edac_type edac_mode;
1071 	unsigned long last_cumul_size;
1072 	int index, mem_dev, drc_chan;
1073 	int drc_drbg;		/* DRB granularity 0=64mb, 1=128mb */
1074 	int drc_ddim;		/* DRAM Data Integrity Mode 0=none, 2=edac */
1075 	u8 value;
1076 	u32 dra, drc, cumul_size, i, nr_pages;
1077 
1078 	dra = 0;
1079 	for (index = 0; index < 4; index++) {
1080 		u8 dra_reg;
1081 		pci_read_config_byte(pdev, E752X_DRA + index, &dra_reg);
1082 		dra |= dra_reg << (index * 8);
1083 	}
1084 	pci_read_config_dword(pdev, E752X_DRC, &drc);
1085 	drc_chan = dual_channel_active(ddrcsr) ? 1 : 0;
1086 	drc_drbg = drc_chan + 1;	/* 128 in dual mode, 64 in single */
1087 	drc_ddim = (drc >> 20) & 0x3;
1088 
1089 	/* The dram row boundary (DRB) reg values are boundary address for
1090 	 * each DRAM row with a granularity of 64 or 128MB (single/dual
1091 	 * channel operation).  DRB regs are cumulative; therefore DRB7 will
1092 	 * contain the total memory contained in all eight rows.
1093 	 */
1094 	for (last_cumul_size = index = 0; index < mci->nr_csrows; index++) {
1095 		/* mem_dev 0=x8, 1=x4 */
1096 		mem_dev = (dra >> (index * 4 + 2)) & 0x3;
1097 		csrow = mci->csrows[remap_csrow_index(mci, index)];
1098 
1099 		mem_dev = (mem_dev == 2);
1100 		pci_read_config_byte(pdev, E752X_DRB + index, &value);
1101 		/* convert a 128 or 64 MiB DRB to a page size. */
1102 		cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
1103 		edac_dbg(3, "(%d) cumul_size 0x%x\n", index, cumul_size);
1104 		if (cumul_size == last_cumul_size)
1105 			continue;	/* not populated */
1106 
1107 		csrow->first_page = last_cumul_size;
1108 		csrow->last_page = cumul_size - 1;
1109 		nr_pages = cumul_size - last_cumul_size;
1110 		last_cumul_size = cumul_size;
1111 
1112 		/*
1113 		* if single channel or x8 devices then SECDED
1114 		* if dual channel and x4 then S4ECD4ED
1115 		*/
1116 		if (drc_ddim) {
1117 			if (drc_chan && mem_dev) {
1118 				edac_mode = EDAC_S4ECD4ED;
1119 				mci->edac_cap |= EDAC_FLAG_S4ECD4ED;
1120 			} else {
1121 				edac_mode = EDAC_SECDED;
1122 				mci->edac_cap |= EDAC_FLAG_SECDED;
1123 			}
1124 		} else
1125 			edac_mode = EDAC_NONE;
1126 		for (i = 0; i < csrow->nr_channels; i++) {
1127 			struct dimm_info *dimm = csrow->channels[i]->dimm;
1128 
1129 			edac_dbg(3, "Initializing rank at (%i,%i)\n", index, i);
1130 			dimm->nr_pages = nr_pages / csrow->nr_channels;
1131 			dimm->grain = 1 << 12;	/* 4KiB - resolution of CELOG */
1132 			dimm->mtype = MEM_RDDR;	/* only one type supported */
1133 			dimm->dtype = mem_dev ? DEV_X4 : DEV_X8;
1134 			dimm->edac_mode = edac_mode;
1135 		}
1136 	}
1137 }
1138 
e752x_init_mem_map_table(struct pci_dev * pdev,struct e752x_pvt * pvt)1139 static void e752x_init_mem_map_table(struct pci_dev *pdev,
1140 				struct e752x_pvt *pvt)
1141 {
1142 	int index;
1143 	u8 value, last, row;
1144 
1145 	last = 0;
1146 	row = 0;
1147 
1148 	for (index = 0; index < 8; index += 2) {
1149 		pci_read_config_byte(pdev, E752X_DRB + index, &value);
1150 		/* test if there is a dimm in this slot */
1151 		if (value == last) {
1152 			/* no dimm in the slot, so flag it as empty */
1153 			pvt->map[index] = 0xff;
1154 			pvt->map[index + 1] = 0xff;
1155 		} else {	/* there is a dimm in the slot */
1156 			pvt->map[index] = row;
1157 			row++;
1158 			last = value;
1159 			/* test the next value to see if the dimm is double
1160 			 * sided
1161 			 */
1162 			pci_read_config_byte(pdev, E752X_DRB + index + 1,
1163 					&value);
1164 
1165 			/* the dimm is single sided, so flag as empty */
1166 			/* this is a double sided dimm to save the next row #*/
1167 			pvt->map[index + 1] = (value == last) ? 0xff :	row;
1168 			row++;
1169 			last = value;
1170 		}
1171 	}
1172 }
1173 
1174 /* Return 0 on success or 1 on failure. */
e752x_get_devs(struct pci_dev * pdev,int dev_idx,struct e752x_pvt * pvt)1175 static int e752x_get_devs(struct pci_dev *pdev, int dev_idx,
1176 			struct e752x_pvt *pvt)
1177 {
1178 	pvt->dev_d0f1 = pci_get_device(PCI_VENDOR_ID_INTEL,
1179 				pvt->dev_info->err_dev, NULL);
1180 
1181 	if (pvt->dev_d0f1 == NULL) {
1182 		pvt->dev_d0f1 = pci_scan_single_device(pdev->bus,
1183 							PCI_DEVFN(0, 1));
1184 		pci_dev_get(pvt->dev_d0f1);
1185 	}
1186 
1187 	if (pvt->dev_d0f1 == NULL) {
1188 		e752x_printk(KERN_ERR, "error reporting device not found:"
1189 			"vendor %x device 0x%x (broken BIOS?)\n",
1190 			PCI_VENDOR_ID_INTEL, e752x_devs[dev_idx].err_dev);
1191 		return 1;
1192 	}
1193 
1194 	pvt->dev_d0f0 = pci_get_device(PCI_VENDOR_ID_INTEL,
1195 				e752x_devs[dev_idx].ctl_dev,
1196 				NULL);
1197 
1198 	if (pvt->dev_d0f0 == NULL)
1199 		goto fail;
1200 
1201 	return 0;
1202 
1203 fail:
1204 	pci_dev_put(pvt->dev_d0f1);
1205 	return 1;
1206 }
1207 
1208 /* Setup system bus parity mask register.
1209  * Sysbus parity supported on:
1210  * e7320/e7520/e7525 + Xeon
1211  */
e752x_init_sysbus_parity_mask(struct e752x_pvt * pvt)1212 static void e752x_init_sysbus_parity_mask(struct e752x_pvt *pvt)
1213 {
1214 	char *cpu_id = cpu_data(0).x86_model_id;
1215 	struct pci_dev *dev = pvt->dev_d0f1;
1216 	int enable = 1;
1217 
1218 	/* Allow module parameter override, else see if CPU supports parity */
1219 	if (sysbus_parity != -1) {
1220 		enable = sysbus_parity;
1221 	} else if (cpu_id[0] && !strstr(cpu_id, "Xeon")) {
1222 		e752x_printk(KERN_INFO, "System Bus Parity not "
1223 			     "supported by CPU, disabling\n");
1224 		enable = 0;
1225 	}
1226 
1227 	if (enable)
1228 		pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0000);
1229 	else
1230 		pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0309);
1231 }
1232 
e752x_init_error_reporting_regs(struct e752x_pvt * pvt)1233 static void e752x_init_error_reporting_regs(struct e752x_pvt *pvt)
1234 {
1235 	struct pci_dev *dev;
1236 
1237 	dev = pvt->dev_d0f1;
1238 	/* Turn off error disable & SMI in case the BIOS turned it on */
1239 	if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
1240 		pci_write_config_dword(dev, I3100_NSI_EMASK, 0);
1241 		pci_write_config_dword(dev, I3100_NSI_SMICMD, 0);
1242 	} else {
1243 		pci_write_config_byte(dev, E752X_HI_ERRMASK, 0x00);
1244 		pci_write_config_byte(dev, E752X_HI_SMICMD, 0x00);
1245 	}
1246 
1247 	e752x_init_sysbus_parity_mask(pvt);
1248 
1249 	pci_write_config_word(dev, E752X_SYSBUS_SMICMD, 0x00);
1250 	pci_write_config_byte(dev, E752X_BUF_ERRMASK, 0x00);
1251 	pci_write_config_byte(dev, E752X_BUF_SMICMD, 0x00);
1252 	pci_write_config_byte(dev, E752X_DRAM_ERRMASK, 0x00);
1253 	pci_write_config_byte(dev, E752X_DRAM_SMICMD, 0x00);
1254 }
1255 
e752x_probe1(struct pci_dev * pdev,int dev_idx)1256 static int e752x_probe1(struct pci_dev *pdev, int dev_idx)
1257 {
1258 	u16 pci_data;
1259 	u8 stat8;
1260 	struct mem_ctl_info *mci;
1261 	struct edac_mc_layer layers[2];
1262 	struct e752x_pvt *pvt;
1263 	u16 ddrcsr;
1264 	int drc_chan;		/* Number of channels 0=1chan,1=2chan */
1265 	struct e752x_error_info discard;
1266 
1267 	edac_dbg(0, "mci\n");
1268 	edac_dbg(0, "Starting Probe1\n");
1269 
1270 	/* check to see if device 0 function 1 is enabled; if it isn't, we
1271 	 * assume the BIOS has reserved it for a reason and is expecting
1272 	 * exclusive access, we take care not to violate that assumption and
1273 	 * fail the probe. */
1274 	pci_read_config_byte(pdev, E752X_DEVPRES1, &stat8);
1275 	if (!force_function_unhide && !(stat8 & (1 << 5))) {
1276 		printk(KERN_INFO "Contact your BIOS vendor to see if the "
1277 			"E752x error registers can be safely un-hidden\n");
1278 		return -ENODEV;
1279 	}
1280 	stat8 |= (1 << 5);
1281 	pci_write_config_byte(pdev, E752X_DEVPRES1, stat8);
1282 
1283 	pci_read_config_word(pdev, E752X_DDRCSR, &ddrcsr);
1284 	/* FIXME: should check >>12 or 0xf, true for all? */
1285 	/* Dual channel = 1, Single channel = 0 */
1286 	drc_chan = dual_channel_active(ddrcsr);
1287 
1288 	layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
1289 	layers[0].size = E752X_NR_CSROWS;
1290 	layers[0].is_virt_csrow = true;
1291 	layers[1].type = EDAC_MC_LAYER_CHANNEL;
1292 	layers[1].size = drc_chan + 1;
1293 	layers[1].is_virt_csrow = false;
1294 	mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt));
1295 	if (mci == NULL)
1296 		return -ENOMEM;
1297 
1298 	edac_dbg(3, "init mci\n");
1299 	mci->mtype_cap = MEM_FLAG_RDDR;
1300 	/* 3100 IMCH supports SECDEC only */
1301 	mci->edac_ctl_cap = (dev_idx == I3100) ? EDAC_FLAG_SECDED :
1302 		(EDAC_FLAG_NONE | EDAC_FLAG_SECDED | EDAC_FLAG_S4ECD4ED);
1303 	/* FIXME - what if different memory types are in different csrows? */
1304 	mci->mod_name = EDAC_MOD_STR;
1305 	mci->pdev = &pdev->dev;
1306 
1307 	edac_dbg(3, "init pvt\n");
1308 	pvt = (struct e752x_pvt *)mci->pvt_info;
1309 	pvt->dev_info = &e752x_devs[dev_idx];
1310 	pvt->mc_symmetric = ((ddrcsr & 0x10) != 0);
1311 
1312 	if (e752x_get_devs(pdev, dev_idx, pvt)) {
1313 		edac_mc_free(mci);
1314 		return -ENODEV;
1315 	}
1316 
1317 	edac_dbg(3, "more mci init\n");
1318 	mci->ctl_name = pvt->dev_info->ctl_name;
1319 	mci->dev_name = pci_name(pdev);
1320 	mci->edac_check = e752x_check;
1321 	mci->ctl_page_to_phys = ctl_page_to_phys;
1322 	mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
1323 	mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
1324 
1325 	/* set the map type.  1 = normal, 0 = reversed
1326 	 * Must be set before e752x_init_csrows in case csrow mapping
1327 	 * is reversed.
1328 	 */
1329 	pci_read_config_byte(pdev, E752X_DRM, &stat8);
1330 	pvt->map_type = ((stat8 & 0x0f) > ((stat8 >> 4) & 0x0f));
1331 
1332 	e752x_init_csrows(mci, pdev, ddrcsr);
1333 	e752x_init_mem_map_table(pdev, pvt);
1334 
1335 	if (dev_idx == I3100)
1336 		mci->edac_cap = EDAC_FLAG_SECDED; /* the only mode supported */
1337 	else
1338 		mci->edac_cap |= EDAC_FLAG_NONE;
1339 	edac_dbg(3, "tolm, remapbase, remaplimit\n");
1340 
1341 	/* load the top of low memory, remap base, and remap limit vars */
1342 	pci_read_config_word(pdev, E752X_TOLM, &pci_data);
1343 	pvt->tolm = ((u32) pci_data) << 4;
1344 	pci_read_config_word(pdev, E752X_REMAPBASE, &pci_data);
1345 	pvt->remapbase = ((u32) pci_data) << 14;
1346 	pci_read_config_word(pdev, E752X_REMAPLIMIT, &pci_data);
1347 	pvt->remaplimit = ((u32) pci_data) << 14;
1348 	e752x_printk(KERN_INFO,
1349 			"tolm = %x, remapbase = %x, remaplimit = %x\n",
1350 			pvt->tolm, pvt->remapbase, pvt->remaplimit);
1351 
1352 	/* Here we assume that we will never see multiple instances of this
1353 	 * type of memory controller.  The ID is therefore hardcoded to 0.
1354 	 */
1355 	if (edac_mc_add_mc(mci)) {
1356 		edac_dbg(3, "failed edac_mc_add_mc()\n");
1357 		goto fail;
1358 	}
1359 
1360 	e752x_init_error_reporting_regs(pvt);
1361 	e752x_get_error_info(mci, &discard);	/* clear other MCH errors */
1362 
1363 	/* allocating generic PCI control info */
1364 	e752x_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
1365 	if (!e752x_pci) {
1366 		printk(KERN_WARNING
1367 			"%s(): Unable to create PCI control\n", __func__);
1368 		printk(KERN_WARNING
1369 			"%s(): PCI error report via EDAC not setup\n",
1370 			__func__);
1371 	}
1372 
1373 	/* get this far and it's successful */
1374 	edac_dbg(3, "success\n");
1375 	return 0;
1376 
1377 fail:
1378 	pci_dev_put(pvt->dev_d0f0);
1379 	pci_dev_put(pvt->dev_d0f1);
1380 	edac_mc_free(mci);
1381 
1382 	return -ENODEV;
1383 }
1384 
1385 /* returns count (>= 0), or negative on error */
e752x_init_one(struct pci_dev * pdev,const struct pci_device_id * ent)1386 static int e752x_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
1387 {
1388 	edac_dbg(0, "\n");
1389 
1390 	/* wake up and enable device */
1391 	if (pci_enable_device(pdev) < 0)
1392 		return -EIO;
1393 
1394 	return e752x_probe1(pdev, ent->driver_data);
1395 }
1396 
e752x_remove_one(struct pci_dev * pdev)1397 static void e752x_remove_one(struct pci_dev *pdev)
1398 {
1399 	struct mem_ctl_info *mci;
1400 	struct e752x_pvt *pvt;
1401 
1402 	edac_dbg(0, "\n");
1403 
1404 	if (e752x_pci)
1405 		edac_pci_release_generic_ctl(e752x_pci);
1406 
1407 	if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL)
1408 		return;
1409 
1410 	pvt = (struct e752x_pvt *)mci->pvt_info;
1411 	pci_dev_put(pvt->dev_d0f0);
1412 	pci_dev_put(pvt->dev_d0f1);
1413 	edac_mc_free(mci);
1414 }
1415 
1416 static const struct pci_device_id e752x_pci_tbl[] = {
1417 	{
1418 	 PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1419 	 E7520},
1420 	{
1421 	 PCI_VEND_DEV(INTEL, 7525_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1422 	 E7525},
1423 	{
1424 	 PCI_VEND_DEV(INTEL, 7320_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1425 	 E7320},
1426 	{
1427 	 PCI_VEND_DEV(INTEL, 3100_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1428 	 I3100},
1429 	{
1430 	 0,
1431 	 }			/* 0 terminated list. */
1432 };
1433 
1434 MODULE_DEVICE_TABLE(pci, e752x_pci_tbl);
1435 
1436 static struct pci_driver e752x_driver = {
1437 	.name = EDAC_MOD_STR,
1438 	.probe = e752x_init_one,
1439 	.remove = e752x_remove_one,
1440 	.id_table = e752x_pci_tbl,
1441 };
1442 
e752x_init(void)1443 static int __init e752x_init(void)
1444 {
1445 	int pci_rc;
1446 
1447 	edac_dbg(3, "\n");
1448 
1449        /* Ensure that the OPSTATE is set correctly for POLL or NMI */
1450        opstate_init();
1451 
1452 	pci_rc = pci_register_driver(&e752x_driver);
1453 	return (pci_rc < 0) ? pci_rc : 0;
1454 }
1455 
e752x_exit(void)1456 static void __exit e752x_exit(void)
1457 {
1458 	edac_dbg(3, "\n");
1459 	pci_unregister_driver(&e752x_driver);
1460 }
1461 
1462 module_init(e752x_init);
1463 module_exit(e752x_exit);
1464 
1465 MODULE_LICENSE("GPL");
1466 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Tom Zimmerman\n");
1467 MODULE_DESCRIPTION("MC support for Intel e752x/3100 memory controllers");
1468 
1469 module_param(force_function_unhide, int, 0444);
1470 MODULE_PARM_DESC(force_function_unhide, "if BIOS sets Dev0:Fun1 up as hidden:"
1471 		 " 1=force unhide and hope BIOS doesn't fight driver for "
1472 		"Dev0:Fun1 access");
1473 
1474 module_param(edac_op_state, int, 0444);
1475 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
1476 
1477 module_param(sysbus_parity, int, 0444);
1478 MODULE_PARM_DESC(sysbus_parity, "0=disable system bus parity checking,"
1479 		" 1=enable system bus parity checking, default=auto-detect");
1480 module_param(report_non_memory_errors, int, 0644);
1481 MODULE_PARM_DESC(report_non_memory_errors, "0=disable non-memory error "
1482 		"reporting, 1=enable non-memory error reporting");
1483