1 /*
2  * Derived from arch/powerpc/kernel/iommu.c
3  *
4  * Copyright IBM Corporation, 2006-2007
5  * Copyright (C) 2006  Jon Mason <jdmason@kudzu.us>
6  *
7  * Author: Jon Mason <jdmason@kudzu.us>
8  * Author: Muli Ben-Yehuda <muli@il.ibm.com>
9 
10  * This program is free software; you can redistribute it and/or modify
11  * it under the terms of the GNU General Public License as published by
12  * the Free Software Foundation; either version 2 of the License, or
13  * (at your option) any later version.
14  *
15  * This program is distributed in the hope that it will be useful,
16  * but WITHOUT ANY WARRANTY; without even the implied warranty of
17  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
18  * GNU General Public License for more details.
19  *
20  * You should have received a copy of the GNU General Public License
21  * along with this program; if not, write to the Free Software
22  * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307 USA
23  */
24 
25 #define pr_fmt(fmt) "Calgary: " fmt
26 
27 #include <linux/kernel.h>
28 #include <linux/init.h>
29 #include <linux/types.h>
30 #include <linux/slab.h>
31 #include <linux/mm.h>
32 #include <linux/spinlock.h>
33 #include <linux/string.h>
34 #include <linux/crash_dump.h>
35 #include <linux/dma-mapping.h>
36 #include <linux/dma-direct.h>
37 #include <linux/bitmap.h>
38 #include <linux/pci_ids.h>
39 #include <linux/pci.h>
40 #include <linux/delay.h>
41 #include <linux/scatterlist.h>
42 #include <linux/iommu-helper.h>
43 
44 #include <asm/iommu.h>
45 #include <asm/calgary.h>
46 #include <asm/tce.h>
47 #include <asm/pci-direct.h>
48 #include <asm/dma.h>
49 #include <asm/rio.h>
50 #include <asm/bios_ebda.h>
51 #include <asm/x86_init.h>
52 #include <asm/iommu_table.h>
53 
54 #define CALGARY_MAPPING_ERROR	0
55 
56 #ifdef CONFIG_CALGARY_IOMMU_ENABLED_BY_DEFAULT
57 int use_calgary __read_mostly = 1;
58 #else
59 int use_calgary __read_mostly = 0;
60 #endif /* CONFIG_CALGARY_DEFAULT_ENABLED */
61 
62 #define PCI_DEVICE_ID_IBM_CALGARY 0x02a1
63 #define PCI_DEVICE_ID_IBM_CALIOC2 0x0308
64 
65 /* register offsets inside the host bridge space */
66 #define CALGARY_CONFIG_REG	0x0108
67 #define PHB_CSR_OFFSET		0x0110 /* Channel Status */
68 #define PHB_PLSSR_OFFSET	0x0120
69 #define PHB_CONFIG_RW_OFFSET	0x0160
70 #define PHB_IOBASE_BAR_LOW	0x0170
71 #define PHB_IOBASE_BAR_HIGH	0x0180
72 #define PHB_MEM_1_LOW		0x0190
73 #define PHB_MEM_1_HIGH		0x01A0
74 #define PHB_IO_ADDR_SIZE	0x01B0
75 #define PHB_MEM_1_SIZE		0x01C0
76 #define PHB_MEM_ST_OFFSET	0x01D0
77 #define PHB_AER_OFFSET		0x0200
78 #define PHB_CONFIG_0_HIGH	0x0220
79 #define PHB_CONFIG_0_LOW	0x0230
80 #define PHB_CONFIG_0_END	0x0240
81 #define PHB_MEM_2_LOW		0x02B0
82 #define PHB_MEM_2_HIGH		0x02C0
83 #define PHB_MEM_2_SIZE_HIGH	0x02D0
84 #define PHB_MEM_2_SIZE_LOW	0x02E0
85 #define PHB_DOSHOLE_OFFSET	0x08E0
86 
87 /* CalIOC2 specific */
88 #define PHB_SAVIOR_L2		0x0DB0
89 #define PHB_PAGE_MIG_CTRL	0x0DA8
90 #define PHB_PAGE_MIG_DEBUG	0x0DA0
91 #define PHB_ROOT_COMPLEX_STATUS 0x0CB0
92 
93 /* PHB_CONFIG_RW */
94 #define PHB_TCE_ENABLE		0x20000000
95 #define PHB_SLOT_DISABLE	0x1C000000
96 #define PHB_DAC_DISABLE		0x01000000
97 #define PHB_MEM2_ENABLE		0x00400000
98 #define PHB_MCSR_ENABLE		0x00100000
99 /* TAR (Table Address Register) */
100 #define TAR_SW_BITS		0x0000ffffffff800fUL
101 #define TAR_VALID		0x0000000000000008UL
102 /* CSR (Channel/DMA Status Register) */
103 #define CSR_AGENT_MASK		0xffe0ffff
104 /* CCR (Calgary Configuration Register) */
105 #define CCR_2SEC_TIMEOUT	0x000000000000000EUL
106 /* PMCR/PMDR (Page Migration Control/Debug Registers */
107 #define PMR_SOFTSTOP		0x80000000
108 #define PMR_SOFTSTOPFAULT	0x40000000
109 #define PMR_HARDSTOP		0x20000000
110 
111 /*
112  * The maximum PHB bus number.
113  * x3950M2 (rare): 8 chassis, 48 PHBs per chassis = 384
114  * x3950M2: 4 chassis, 48 PHBs per chassis        = 192
115  * x3950 (PCIE): 8 chassis, 32 PHBs per chassis   = 256
116  * x3950 (PCIX): 8 chassis, 16 PHBs per chassis   = 128
117  */
118 #define MAX_PHB_BUS_NUM		256
119 
120 #define PHBS_PER_CALGARY	  4
121 
122 /* register offsets in Calgary's internal register space */
123 static const unsigned long tar_offsets[] = {
124 	0x0580 /* TAR0 */,
125 	0x0588 /* TAR1 */,
126 	0x0590 /* TAR2 */,
127 	0x0598 /* TAR3 */
128 };
129 
130 static const unsigned long split_queue_offsets[] = {
131 	0x4870 /* SPLIT QUEUE 0 */,
132 	0x5870 /* SPLIT QUEUE 1 */,
133 	0x6870 /* SPLIT QUEUE 2 */,
134 	0x7870 /* SPLIT QUEUE 3 */
135 };
136 
137 static const unsigned long phb_offsets[] = {
138 	0x8000 /* PHB0 */,
139 	0x9000 /* PHB1 */,
140 	0xA000 /* PHB2 */,
141 	0xB000 /* PHB3 */
142 };
143 
144 /* PHB debug registers */
145 
146 static const unsigned long phb_debug_offsets[] = {
147 	0x4000	/* PHB 0 DEBUG */,
148 	0x5000	/* PHB 1 DEBUG */,
149 	0x6000	/* PHB 2 DEBUG */,
150 	0x7000	/* PHB 3 DEBUG */
151 };
152 
153 /*
154  * STUFF register for each debug PHB,
155  * byte 1 = start bus number, byte 2 = end bus number
156  */
157 
158 #define PHB_DEBUG_STUFF_OFFSET	0x0020
159 
160 #define EMERGENCY_PAGES 32 /* = 128KB */
161 
162 unsigned int specified_table_size = TCE_TABLE_SIZE_UNSPECIFIED;
163 static int translate_empty_slots __read_mostly = 0;
164 static int calgary_detected __read_mostly = 0;
165 
166 static struct rio_table_hdr	*rio_table_hdr __initdata;
167 static struct scal_detail	*scal_devs[MAX_NUMNODES] __initdata;
168 static struct rio_detail	*rio_devs[MAX_NUMNODES * 4] __initdata;
169 
170 struct calgary_bus_info {
171 	void *tce_space;
172 	unsigned char translation_disabled;
173 	signed char phbid;
174 	void __iomem *bbar;
175 };
176 
177 static void calgary_handle_quirks(struct iommu_table *tbl, struct pci_dev *dev);
178 static void calgary_tce_cache_blast(struct iommu_table *tbl);
179 static void calgary_dump_error_regs(struct iommu_table *tbl);
180 static void calioc2_handle_quirks(struct iommu_table *tbl, struct pci_dev *dev);
181 static void calioc2_tce_cache_blast(struct iommu_table *tbl);
182 static void calioc2_dump_error_regs(struct iommu_table *tbl);
183 static void calgary_init_bitmap_from_tce_table(struct iommu_table *tbl);
184 static void get_tce_space_from_tar(void);
185 
186 static const struct cal_chipset_ops calgary_chip_ops = {
187 	.handle_quirks = calgary_handle_quirks,
188 	.tce_cache_blast = calgary_tce_cache_blast,
189 	.dump_error_regs = calgary_dump_error_regs
190 };
191 
192 static const struct cal_chipset_ops calioc2_chip_ops = {
193 	.handle_quirks = calioc2_handle_quirks,
194 	.tce_cache_blast = calioc2_tce_cache_blast,
195 	.dump_error_regs = calioc2_dump_error_regs
196 };
197 
198 static struct calgary_bus_info bus_info[MAX_PHB_BUS_NUM] = { { NULL, 0, 0 }, };
199 
translation_enabled(struct iommu_table * tbl)200 static inline int translation_enabled(struct iommu_table *tbl)
201 {
202 	/* only PHBs with translation enabled have an IOMMU table */
203 	return (tbl != NULL);
204 }
205 
iommu_range_reserve(struct iommu_table * tbl,unsigned long start_addr,unsigned int npages)206 static void iommu_range_reserve(struct iommu_table *tbl,
207 	unsigned long start_addr, unsigned int npages)
208 {
209 	unsigned long index;
210 	unsigned long end;
211 	unsigned long flags;
212 
213 	index = start_addr >> PAGE_SHIFT;
214 
215 	/* bail out if we're asked to reserve a region we don't cover */
216 	if (index >= tbl->it_size)
217 		return;
218 
219 	end = index + npages;
220 	if (end > tbl->it_size) /* don't go off the table */
221 		end = tbl->it_size;
222 
223 	spin_lock_irqsave(&tbl->it_lock, flags);
224 
225 	bitmap_set(tbl->it_map, index, npages);
226 
227 	spin_unlock_irqrestore(&tbl->it_lock, flags);
228 }
229 
iommu_range_alloc(struct device * dev,struct iommu_table * tbl,unsigned int npages)230 static unsigned long iommu_range_alloc(struct device *dev,
231 				       struct iommu_table *tbl,
232 				       unsigned int npages)
233 {
234 	unsigned long flags;
235 	unsigned long offset;
236 	unsigned long boundary_size;
237 
238 	boundary_size = ALIGN(dma_get_seg_boundary(dev) + 1,
239 			      PAGE_SIZE) >> PAGE_SHIFT;
240 
241 	BUG_ON(npages == 0);
242 
243 	spin_lock_irqsave(&tbl->it_lock, flags);
244 
245 	offset = iommu_area_alloc(tbl->it_map, tbl->it_size, tbl->it_hint,
246 				  npages, 0, boundary_size, 0);
247 	if (offset == ~0UL) {
248 		tbl->chip_ops->tce_cache_blast(tbl);
249 
250 		offset = iommu_area_alloc(tbl->it_map, tbl->it_size, 0,
251 					  npages, 0, boundary_size, 0);
252 		if (offset == ~0UL) {
253 			pr_warn("IOMMU full\n");
254 			spin_unlock_irqrestore(&tbl->it_lock, flags);
255 			if (panic_on_overflow)
256 				panic("Calgary: fix the allocator.\n");
257 			else
258 				return CALGARY_MAPPING_ERROR;
259 		}
260 	}
261 
262 	tbl->it_hint = offset + npages;
263 	BUG_ON(tbl->it_hint > tbl->it_size);
264 
265 	spin_unlock_irqrestore(&tbl->it_lock, flags);
266 
267 	return offset;
268 }
269 
iommu_alloc(struct device * dev,struct iommu_table * tbl,void * vaddr,unsigned int npages,int direction)270 static dma_addr_t iommu_alloc(struct device *dev, struct iommu_table *tbl,
271 			      void *vaddr, unsigned int npages, int direction)
272 {
273 	unsigned long entry;
274 	dma_addr_t ret;
275 
276 	entry = iommu_range_alloc(dev, tbl, npages);
277 
278 	if (unlikely(entry == CALGARY_MAPPING_ERROR)) {
279 		pr_warn("failed to allocate %u pages in iommu %p\n",
280 			npages, tbl);
281 		return CALGARY_MAPPING_ERROR;
282 	}
283 
284 	/* set the return dma address */
285 	ret = (entry << PAGE_SHIFT) | ((unsigned long)vaddr & ~PAGE_MASK);
286 
287 	/* put the TCEs in the HW table */
288 	tce_build(tbl, entry, npages, (unsigned long)vaddr & PAGE_MASK,
289 		  direction);
290 	return ret;
291 }
292 
iommu_free(struct iommu_table * tbl,dma_addr_t dma_addr,unsigned int npages)293 static void iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr,
294 	unsigned int npages)
295 {
296 	unsigned long entry;
297 	unsigned long badend;
298 	unsigned long flags;
299 
300 	/* were we called with bad_dma_address? */
301 	badend = CALGARY_MAPPING_ERROR + (EMERGENCY_PAGES * PAGE_SIZE);
302 	if (unlikely(dma_addr < badend)) {
303 		WARN(1, KERN_ERR "Calgary: driver tried unmapping bad DMA "
304 		       "address 0x%Lx\n", dma_addr);
305 		return;
306 	}
307 
308 	entry = dma_addr >> PAGE_SHIFT;
309 
310 	BUG_ON(entry + npages > tbl->it_size);
311 
312 	tce_free(tbl, entry, npages);
313 
314 	spin_lock_irqsave(&tbl->it_lock, flags);
315 
316 	bitmap_clear(tbl->it_map, entry, npages);
317 
318 	spin_unlock_irqrestore(&tbl->it_lock, flags);
319 }
320 
find_iommu_table(struct device * dev)321 static inline struct iommu_table *find_iommu_table(struct device *dev)
322 {
323 	struct pci_dev *pdev;
324 	struct pci_bus *pbus;
325 	struct iommu_table *tbl;
326 
327 	pdev = to_pci_dev(dev);
328 
329 	/* search up the device tree for an iommu */
330 	pbus = pdev->bus;
331 	do {
332 		tbl = pci_iommu(pbus);
333 		if (tbl && tbl->it_busno == pbus->number)
334 			break;
335 		tbl = NULL;
336 		pbus = pbus->parent;
337 	} while (pbus);
338 
339 	BUG_ON(tbl && (tbl->it_busno != pbus->number));
340 
341 	return tbl;
342 }
343 
calgary_unmap_sg(struct device * dev,struct scatterlist * sglist,int nelems,enum dma_data_direction dir,unsigned long attrs)344 static void calgary_unmap_sg(struct device *dev, struct scatterlist *sglist,
345 			     int nelems,enum dma_data_direction dir,
346 			     unsigned long attrs)
347 {
348 	struct iommu_table *tbl = find_iommu_table(dev);
349 	struct scatterlist *s;
350 	int i;
351 
352 	if (!translation_enabled(tbl))
353 		return;
354 
355 	for_each_sg(sglist, s, nelems, i) {
356 		unsigned int npages;
357 		dma_addr_t dma = s->dma_address;
358 		unsigned int dmalen = s->dma_length;
359 
360 		if (dmalen == 0)
361 			break;
362 
363 		npages = iommu_num_pages(dma, dmalen, PAGE_SIZE);
364 		iommu_free(tbl, dma, npages);
365 	}
366 }
367 
calgary_map_sg(struct device * dev,struct scatterlist * sg,int nelems,enum dma_data_direction dir,unsigned long attrs)368 static int calgary_map_sg(struct device *dev, struct scatterlist *sg,
369 			  int nelems, enum dma_data_direction dir,
370 			  unsigned long attrs)
371 {
372 	struct iommu_table *tbl = find_iommu_table(dev);
373 	struct scatterlist *s;
374 	unsigned long vaddr;
375 	unsigned int npages;
376 	unsigned long entry;
377 	int i;
378 
379 	for_each_sg(sg, s, nelems, i) {
380 		BUG_ON(!sg_page(s));
381 
382 		vaddr = (unsigned long) sg_virt(s);
383 		npages = iommu_num_pages(vaddr, s->length, PAGE_SIZE);
384 
385 		entry = iommu_range_alloc(dev, tbl, npages);
386 		if (entry == CALGARY_MAPPING_ERROR) {
387 			/* makes sure unmap knows to stop */
388 			s->dma_length = 0;
389 			goto error;
390 		}
391 
392 		s->dma_address = (entry << PAGE_SHIFT) | s->offset;
393 
394 		/* insert into HW table */
395 		tce_build(tbl, entry, npages, vaddr & PAGE_MASK, dir);
396 
397 		s->dma_length = s->length;
398 	}
399 
400 	return nelems;
401 error:
402 	calgary_unmap_sg(dev, sg, nelems, dir, 0);
403 	for_each_sg(sg, s, nelems, i) {
404 		sg->dma_address = CALGARY_MAPPING_ERROR;
405 		sg->dma_length = 0;
406 	}
407 	return 0;
408 }
409 
calgary_map_page(struct device * dev,struct page * page,unsigned long offset,size_t size,enum dma_data_direction dir,unsigned long attrs)410 static dma_addr_t calgary_map_page(struct device *dev, struct page *page,
411 				   unsigned long offset, size_t size,
412 				   enum dma_data_direction dir,
413 				   unsigned long attrs)
414 {
415 	void *vaddr = page_address(page) + offset;
416 	unsigned long uaddr;
417 	unsigned int npages;
418 	struct iommu_table *tbl = find_iommu_table(dev);
419 
420 	uaddr = (unsigned long)vaddr;
421 	npages = iommu_num_pages(uaddr, size, PAGE_SIZE);
422 
423 	return iommu_alloc(dev, tbl, vaddr, npages, dir);
424 }
425 
calgary_unmap_page(struct device * dev,dma_addr_t dma_addr,size_t size,enum dma_data_direction dir,unsigned long attrs)426 static void calgary_unmap_page(struct device *dev, dma_addr_t dma_addr,
427 			       size_t size, enum dma_data_direction dir,
428 			       unsigned long attrs)
429 {
430 	struct iommu_table *tbl = find_iommu_table(dev);
431 	unsigned int npages;
432 
433 	npages = iommu_num_pages(dma_addr, size, PAGE_SIZE);
434 	iommu_free(tbl, dma_addr, npages);
435 }
436 
calgary_alloc_coherent(struct device * dev,size_t size,dma_addr_t * dma_handle,gfp_t flag,unsigned long attrs)437 static void* calgary_alloc_coherent(struct device *dev, size_t size,
438 	dma_addr_t *dma_handle, gfp_t flag, unsigned long attrs)
439 {
440 	void *ret = NULL;
441 	dma_addr_t mapping;
442 	unsigned int npages, order;
443 	struct iommu_table *tbl = find_iommu_table(dev);
444 
445 	size = PAGE_ALIGN(size); /* size rounded up to full pages */
446 	npages = size >> PAGE_SHIFT;
447 	order = get_order(size);
448 
449 	/* alloc enough pages (and possibly more) */
450 	ret = (void *)__get_free_pages(flag, order);
451 	if (!ret)
452 		goto error;
453 	memset(ret, 0, size);
454 
455 	/* set up tces to cover the allocated range */
456 	mapping = iommu_alloc(dev, tbl, ret, npages, DMA_BIDIRECTIONAL);
457 	if (mapping == CALGARY_MAPPING_ERROR)
458 		goto free;
459 	*dma_handle = mapping;
460 	return ret;
461 free:
462 	free_pages((unsigned long)ret, get_order(size));
463 	ret = NULL;
464 error:
465 	return ret;
466 }
467 
calgary_free_coherent(struct device * dev,size_t size,void * vaddr,dma_addr_t dma_handle,unsigned long attrs)468 static void calgary_free_coherent(struct device *dev, size_t size,
469 				  void *vaddr, dma_addr_t dma_handle,
470 				  unsigned long attrs)
471 {
472 	unsigned int npages;
473 	struct iommu_table *tbl = find_iommu_table(dev);
474 
475 	size = PAGE_ALIGN(size);
476 	npages = size >> PAGE_SHIFT;
477 
478 	iommu_free(tbl, dma_handle, npages);
479 	free_pages((unsigned long)vaddr, get_order(size));
480 }
481 
calgary_mapping_error(struct device * dev,dma_addr_t dma_addr)482 static int calgary_mapping_error(struct device *dev, dma_addr_t dma_addr)
483 {
484 	return dma_addr == CALGARY_MAPPING_ERROR;
485 }
486 
487 static const struct dma_map_ops calgary_dma_ops = {
488 	.alloc = calgary_alloc_coherent,
489 	.free = calgary_free_coherent,
490 	.map_sg = calgary_map_sg,
491 	.unmap_sg = calgary_unmap_sg,
492 	.map_page = calgary_map_page,
493 	.unmap_page = calgary_unmap_page,
494 	.mapping_error = calgary_mapping_error,
495 	.dma_supported = dma_direct_supported,
496 };
497 
busno_to_bbar(unsigned char num)498 static inline void __iomem * busno_to_bbar(unsigned char num)
499 {
500 	return bus_info[num].bbar;
501 }
502 
busno_to_phbid(unsigned char num)503 static inline int busno_to_phbid(unsigned char num)
504 {
505 	return bus_info[num].phbid;
506 }
507 
split_queue_offset(unsigned char num)508 static inline unsigned long split_queue_offset(unsigned char num)
509 {
510 	size_t idx = busno_to_phbid(num);
511 
512 	return split_queue_offsets[idx];
513 }
514 
tar_offset(unsigned char num)515 static inline unsigned long tar_offset(unsigned char num)
516 {
517 	size_t idx = busno_to_phbid(num);
518 
519 	return tar_offsets[idx];
520 }
521 
phb_offset(unsigned char num)522 static inline unsigned long phb_offset(unsigned char num)
523 {
524 	size_t idx = busno_to_phbid(num);
525 
526 	return phb_offsets[idx];
527 }
528 
calgary_reg(void __iomem * bar,unsigned long offset)529 static inline void __iomem* calgary_reg(void __iomem *bar, unsigned long offset)
530 {
531 	unsigned long target = ((unsigned long)bar) | offset;
532 	return (void __iomem*)target;
533 }
534 
is_calioc2(unsigned short device)535 static inline int is_calioc2(unsigned short device)
536 {
537 	return (device == PCI_DEVICE_ID_IBM_CALIOC2);
538 }
539 
is_calgary(unsigned short device)540 static inline int is_calgary(unsigned short device)
541 {
542 	return (device == PCI_DEVICE_ID_IBM_CALGARY);
543 }
544 
is_cal_pci_dev(unsigned short device)545 static inline int is_cal_pci_dev(unsigned short device)
546 {
547 	return (is_calgary(device) || is_calioc2(device));
548 }
549 
calgary_tce_cache_blast(struct iommu_table * tbl)550 static void calgary_tce_cache_blast(struct iommu_table *tbl)
551 {
552 	u64 val;
553 	u32 aer;
554 	int i = 0;
555 	void __iomem *bbar = tbl->bbar;
556 	void __iomem *target;
557 
558 	/* disable arbitration on the bus */
559 	target = calgary_reg(bbar, phb_offset(tbl->it_busno) | PHB_AER_OFFSET);
560 	aer = readl(target);
561 	writel(0, target);
562 
563 	/* read plssr to ensure it got there */
564 	target = calgary_reg(bbar, phb_offset(tbl->it_busno) | PHB_PLSSR_OFFSET);
565 	val = readl(target);
566 
567 	/* poll split queues until all DMA activity is done */
568 	target = calgary_reg(bbar, split_queue_offset(tbl->it_busno));
569 	do {
570 		val = readq(target);
571 		i++;
572 	} while ((val & 0xff) != 0xff && i < 100);
573 	if (i == 100)
574 		pr_warn("PCI bus not quiesced, continuing anyway\n");
575 
576 	/* invalidate TCE cache */
577 	target = calgary_reg(bbar, tar_offset(tbl->it_busno));
578 	writeq(tbl->tar_val, target);
579 
580 	/* enable arbitration */
581 	target = calgary_reg(bbar, phb_offset(tbl->it_busno) | PHB_AER_OFFSET);
582 	writel(aer, target);
583 	(void)readl(target); /* flush */
584 }
585 
calioc2_tce_cache_blast(struct iommu_table * tbl)586 static void calioc2_tce_cache_blast(struct iommu_table *tbl)
587 {
588 	void __iomem *bbar = tbl->bbar;
589 	void __iomem *target;
590 	u64 val64;
591 	u32 val;
592 	int i = 0;
593 	int count = 1;
594 	unsigned char bus = tbl->it_busno;
595 
596 begin:
597 	printk(KERN_DEBUG "Calgary: CalIOC2 bus 0x%x entering tce cache blast "
598 	       "sequence - count %d\n", bus, count);
599 
600 	/* 1. using the Page Migration Control reg set SoftStop */
601 	target = calgary_reg(bbar, phb_offset(bus) | PHB_PAGE_MIG_CTRL);
602 	val = be32_to_cpu(readl(target));
603 	printk(KERN_DEBUG "1a. read 0x%x [LE] from %p\n", val, target);
604 	val |= PMR_SOFTSTOP;
605 	printk(KERN_DEBUG "1b. writing 0x%x [LE] to %p\n", val, target);
606 	writel(cpu_to_be32(val), target);
607 
608 	/* 2. poll split queues until all DMA activity is done */
609 	printk(KERN_DEBUG "2a. starting to poll split queues\n");
610 	target = calgary_reg(bbar, split_queue_offset(bus));
611 	do {
612 		val64 = readq(target);
613 		i++;
614 	} while ((val64 & 0xff) != 0xff && i < 100);
615 	if (i == 100)
616 		pr_warn("CalIOC2: PCI bus not quiesced, continuing anyway\n");
617 
618 	/* 3. poll Page Migration DEBUG for SoftStopFault */
619 	target = calgary_reg(bbar, phb_offset(bus) | PHB_PAGE_MIG_DEBUG);
620 	val = be32_to_cpu(readl(target));
621 	printk(KERN_DEBUG "3. read 0x%x [LE] from %p\n", val, target);
622 
623 	/* 4. if SoftStopFault - goto (1) */
624 	if (val & PMR_SOFTSTOPFAULT) {
625 		if (++count < 100)
626 			goto begin;
627 		else {
628 			pr_warn("CalIOC2: too many SoftStopFaults, aborting TCE cache flush sequence!\n");
629 			return; /* pray for the best */
630 		}
631 	}
632 
633 	/* 5. Slam into HardStop by reading PHB_PAGE_MIG_CTRL */
634 	target = calgary_reg(bbar, phb_offset(bus) | PHB_PAGE_MIG_CTRL);
635 	printk(KERN_DEBUG "5a. slamming into HardStop by reading %p\n", target);
636 	val = be32_to_cpu(readl(target));
637 	printk(KERN_DEBUG "5b. read 0x%x [LE] from %p\n", val, target);
638 	target = calgary_reg(bbar, phb_offset(bus) | PHB_PAGE_MIG_DEBUG);
639 	val = be32_to_cpu(readl(target));
640 	printk(KERN_DEBUG "5c. read 0x%x [LE] from %p (debug)\n", val, target);
641 
642 	/* 6. invalidate TCE cache */
643 	printk(KERN_DEBUG "6. invalidating TCE cache\n");
644 	target = calgary_reg(bbar, tar_offset(bus));
645 	writeq(tbl->tar_val, target);
646 
647 	/* 7. Re-read PMCR */
648 	printk(KERN_DEBUG "7a. Re-reading PMCR\n");
649 	target = calgary_reg(bbar, phb_offset(bus) | PHB_PAGE_MIG_CTRL);
650 	val = be32_to_cpu(readl(target));
651 	printk(KERN_DEBUG "7b. read 0x%x [LE] from %p\n", val, target);
652 
653 	/* 8. Remove HardStop */
654 	printk(KERN_DEBUG "8a. removing HardStop from PMCR\n");
655 	target = calgary_reg(bbar, phb_offset(bus) | PHB_PAGE_MIG_CTRL);
656 	val = 0;
657 	printk(KERN_DEBUG "8b. writing 0x%x [LE] to %p\n", val, target);
658 	writel(cpu_to_be32(val), target);
659 	val = be32_to_cpu(readl(target));
660 	printk(KERN_DEBUG "8c. read 0x%x [LE] from %p\n", val, target);
661 }
662 
calgary_reserve_mem_region(struct pci_dev * dev,u64 start,u64 limit)663 static void __init calgary_reserve_mem_region(struct pci_dev *dev, u64 start,
664 	u64 limit)
665 {
666 	unsigned int numpages;
667 
668 	limit = limit | 0xfffff;
669 	limit++;
670 
671 	numpages = ((limit - start) >> PAGE_SHIFT);
672 	iommu_range_reserve(pci_iommu(dev->bus), start, numpages);
673 }
674 
calgary_reserve_peripheral_mem_1(struct pci_dev * dev)675 static void __init calgary_reserve_peripheral_mem_1(struct pci_dev *dev)
676 {
677 	void __iomem *target;
678 	u64 low, high, sizelow;
679 	u64 start, limit;
680 	struct iommu_table *tbl = pci_iommu(dev->bus);
681 	unsigned char busnum = dev->bus->number;
682 	void __iomem *bbar = tbl->bbar;
683 
684 	/* peripheral MEM_1 region */
685 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_MEM_1_LOW);
686 	low = be32_to_cpu(readl(target));
687 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_MEM_1_HIGH);
688 	high = be32_to_cpu(readl(target));
689 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_MEM_1_SIZE);
690 	sizelow = be32_to_cpu(readl(target));
691 
692 	start = (high << 32) | low;
693 	limit = sizelow;
694 
695 	calgary_reserve_mem_region(dev, start, limit);
696 }
697 
calgary_reserve_peripheral_mem_2(struct pci_dev * dev)698 static void __init calgary_reserve_peripheral_mem_2(struct pci_dev *dev)
699 {
700 	void __iomem *target;
701 	u32 val32;
702 	u64 low, high, sizelow, sizehigh;
703 	u64 start, limit;
704 	struct iommu_table *tbl = pci_iommu(dev->bus);
705 	unsigned char busnum = dev->bus->number;
706 	void __iomem *bbar = tbl->bbar;
707 
708 	/* is it enabled? */
709 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_CONFIG_RW_OFFSET);
710 	val32 = be32_to_cpu(readl(target));
711 	if (!(val32 & PHB_MEM2_ENABLE))
712 		return;
713 
714 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_MEM_2_LOW);
715 	low = be32_to_cpu(readl(target));
716 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_MEM_2_HIGH);
717 	high = be32_to_cpu(readl(target));
718 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_MEM_2_SIZE_LOW);
719 	sizelow = be32_to_cpu(readl(target));
720 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_MEM_2_SIZE_HIGH);
721 	sizehigh = be32_to_cpu(readl(target));
722 
723 	start = (high << 32) | low;
724 	limit = (sizehigh << 32) | sizelow;
725 
726 	calgary_reserve_mem_region(dev, start, limit);
727 }
728 
729 /*
730  * some regions of the IO address space do not get translated, so we
731  * must not give devices IO addresses in those regions. The regions
732  * are the 640KB-1MB region and the two PCI peripheral memory holes.
733  * Reserve all of them in the IOMMU bitmap to avoid giving them out
734  * later.
735  */
calgary_reserve_regions(struct pci_dev * dev)736 static void __init calgary_reserve_regions(struct pci_dev *dev)
737 {
738 	unsigned int npages;
739 	u64 start;
740 	struct iommu_table *tbl = pci_iommu(dev->bus);
741 
742 	/* reserve EMERGENCY_PAGES from bad_dma_address and up */
743 	iommu_range_reserve(tbl, CALGARY_MAPPING_ERROR, EMERGENCY_PAGES);
744 
745 	/* avoid the BIOS/VGA first 640KB-1MB region */
746 	/* for CalIOC2 - avoid the entire first MB */
747 	if (is_calgary(dev->device)) {
748 		start = (640 * 1024);
749 		npages = ((1024 - 640) * 1024) >> PAGE_SHIFT;
750 	} else { /* calioc2 */
751 		start = 0;
752 		npages = (1 * 1024 * 1024) >> PAGE_SHIFT;
753 	}
754 	iommu_range_reserve(tbl, start, npages);
755 
756 	/* reserve the two PCI peripheral memory regions in IO space */
757 	calgary_reserve_peripheral_mem_1(dev);
758 	calgary_reserve_peripheral_mem_2(dev);
759 }
760 
calgary_setup_tar(struct pci_dev * dev,void __iomem * bbar)761 static int __init calgary_setup_tar(struct pci_dev *dev, void __iomem *bbar)
762 {
763 	u64 val64;
764 	u64 table_phys;
765 	void __iomem *target;
766 	int ret;
767 	struct iommu_table *tbl;
768 
769 	/* build TCE tables for each PHB */
770 	ret = build_tce_table(dev, bbar);
771 	if (ret)
772 		return ret;
773 
774 	tbl = pci_iommu(dev->bus);
775 	tbl->it_base = (unsigned long)bus_info[dev->bus->number].tce_space;
776 
777 	if (is_kdump_kernel())
778 		calgary_init_bitmap_from_tce_table(tbl);
779 	else
780 		tce_free(tbl, 0, tbl->it_size);
781 
782 	if (is_calgary(dev->device))
783 		tbl->chip_ops = &calgary_chip_ops;
784 	else if (is_calioc2(dev->device))
785 		tbl->chip_ops = &calioc2_chip_ops;
786 	else
787 		BUG();
788 
789 	calgary_reserve_regions(dev);
790 
791 	/* set TARs for each PHB */
792 	target = calgary_reg(bbar, tar_offset(dev->bus->number));
793 	val64 = be64_to_cpu(readq(target));
794 
795 	/* zero out all TAR bits under sw control */
796 	val64 &= ~TAR_SW_BITS;
797 	table_phys = (u64)__pa(tbl->it_base);
798 
799 	val64 |= table_phys;
800 
801 	BUG_ON(specified_table_size > TCE_TABLE_SIZE_8M);
802 	val64 |= (u64) specified_table_size;
803 
804 	tbl->tar_val = cpu_to_be64(val64);
805 
806 	writeq(tbl->tar_val, target);
807 	readq(target); /* flush */
808 
809 	return 0;
810 }
811 
calgary_free_bus(struct pci_dev * dev)812 static void __init calgary_free_bus(struct pci_dev *dev)
813 {
814 	u64 val64;
815 	struct iommu_table *tbl = pci_iommu(dev->bus);
816 	void __iomem *target;
817 	unsigned int bitmapsz;
818 
819 	target = calgary_reg(tbl->bbar, tar_offset(dev->bus->number));
820 	val64 = be64_to_cpu(readq(target));
821 	val64 &= ~TAR_SW_BITS;
822 	writeq(cpu_to_be64(val64), target);
823 	readq(target); /* flush */
824 
825 	bitmapsz = tbl->it_size / BITS_PER_BYTE;
826 	free_pages((unsigned long)tbl->it_map, get_order(bitmapsz));
827 	tbl->it_map = NULL;
828 
829 	kfree(tbl);
830 
831 	set_pci_iommu(dev->bus, NULL);
832 
833 	/* Can't free bootmem allocated memory after system is up :-( */
834 	bus_info[dev->bus->number].tce_space = NULL;
835 }
836 
calgary_dump_error_regs(struct iommu_table * tbl)837 static void calgary_dump_error_regs(struct iommu_table *tbl)
838 {
839 	void __iomem *bbar = tbl->bbar;
840 	void __iomem *target;
841 	u32 csr, plssr;
842 
843 	target = calgary_reg(bbar, phb_offset(tbl->it_busno) | PHB_CSR_OFFSET);
844 	csr = be32_to_cpu(readl(target));
845 
846 	target = calgary_reg(bbar, phb_offset(tbl->it_busno) | PHB_PLSSR_OFFSET);
847 	plssr = be32_to_cpu(readl(target));
848 
849 	/* If no error, the agent ID in the CSR is not valid */
850 	pr_emerg("DMA error on Calgary PHB 0x%x, 0x%08x@CSR 0x%08x@PLSSR\n",
851 		 tbl->it_busno, csr, plssr);
852 }
853 
calioc2_dump_error_regs(struct iommu_table * tbl)854 static void calioc2_dump_error_regs(struct iommu_table *tbl)
855 {
856 	void __iomem *bbar = tbl->bbar;
857 	u32 csr, csmr, plssr, mck, rcstat;
858 	void __iomem *target;
859 	unsigned long phboff = phb_offset(tbl->it_busno);
860 	unsigned long erroff;
861 	u32 errregs[7];
862 	int i;
863 
864 	/* dump CSR */
865 	target = calgary_reg(bbar, phboff | PHB_CSR_OFFSET);
866 	csr = be32_to_cpu(readl(target));
867 	/* dump PLSSR */
868 	target = calgary_reg(bbar, phboff | PHB_PLSSR_OFFSET);
869 	plssr = be32_to_cpu(readl(target));
870 	/* dump CSMR */
871 	target = calgary_reg(bbar, phboff | 0x290);
872 	csmr = be32_to_cpu(readl(target));
873 	/* dump mck */
874 	target = calgary_reg(bbar, phboff | 0x800);
875 	mck = be32_to_cpu(readl(target));
876 
877 	pr_emerg("DMA error on CalIOC2 PHB 0x%x\n", tbl->it_busno);
878 
879 	pr_emerg("0x%08x@CSR 0x%08x@PLSSR 0x%08x@CSMR 0x%08x@MCK\n",
880 		 csr, plssr, csmr, mck);
881 
882 	/* dump rest of error regs */
883 	pr_emerg("");
884 	for (i = 0; i < ARRAY_SIZE(errregs); i++) {
885 		/* err regs are at 0x810 - 0x870 */
886 		erroff = (0x810 + (i * 0x10));
887 		target = calgary_reg(bbar, phboff | erroff);
888 		errregs[i] = be32_to_cpu(readl(target));
889 		pr_cont("0x%08x@0x%lx ", errregs[i], erroff);
890 	}
891 	pr_cont("\n");
892 
893 	/* root complex status */
894 	target = calgary_reg(bbar, phboff | PHB_ROOT_COMPLEX_STATUS);
895 	rcstat = be32_to_cpu(readl(target));
896 	printk(KERN_EMERG "Calgary: 0x%08x@0x%x\n", rcstat,
897 	       PHB_ROOT_COMPLEX_STATUS);
898 }
899 
calgary_watchdog(struct timer_list * t)900 static void calgary_watchdog(struct timer_list *t)
901 {
902 	struct iommu_table *tbl = from_timer(tbl, t, watchdog_timer);
903 	void __iomem *bbar = tbl->bbar;
904 	u32 val32;
905 	void __iomem *target;
906 
907 	target = calgary_reg(bbar, phb_offset(tbl->it_busno) | PHB_CSR_OFFSET);
908 	val32 = be32_to_cpu(readl(target));
909 
910 	/* If no error, the agent ID in the CSR is not valid */
911 	if (val32 & CSR_AGENT_MASK) {
912 		tbl->chip_ops->dump_error_regs(tbl);
913 
914 		/* reset error */
915 		writel(0, target);
916 
917 		/* Disable bus that caused the error */
918 		target = calgary_reg(bbar, phb_offset(tbl->it_busno) |
919 				     PHB_CONFIG_RW_OFFSET);
920 		val32 = be32_to_cpu(readl(target));
921 		val32 |= PHB_SLOT_DISABLE;
922 		writel(cpu_to_be32(val32), target);
923 		readl(target); /* flush */
924 	} else {
925 		/* Reset the timer */
926 		mod_timer(&tbl->watchdog_timer, jiffies + 2 * HZ);
927 	}
928 }
929 
calgary_set_split_completion_timeout(void __iomem * bbar,unsigned char busnum,unsigned long timeout)930 static void __init calgary_set_split_completion_timeout(void __iomem *bbar,
931 	unsigned char busnum, unsigned long timeout)
932 {
933 	u64 val64;
934 	void __iomem *target;
935 	unsigned int phb_shift = ~0; /* silence gcc */
936 	u64 mask;
937 
938 	switch (busno_to_phbid(busnum)) {
939 	case 0: phb_shift = (63 - 19);
940 		break;
941 	case 1: phb_shift = (63 - 23);
942 		break;
943 	case 2: phb_shift = (63 - 27);
944 		break;
945 	case 3: phb_shift = (63 - 35);
946 		break;
947 	default:
948 		BUG_ON(busno_to_phbid(busnum));
949 	}
950 
951 	target = calgary_reg(bbar, CALGARY_CONFIG_REG);
952 	val64 = be64_to_cpu(readq(target));
953 
954 	/* zero out this PHB's timer bits */
955 	mask = ~(0xFUL << phb_shift);
956 	val64 &= mask;
957 	val64 |= (timeout << phb_shift);
958 	writeq(cpu_to_be64(val64), target);
959 	readq(target); /* flush */
960 }
961 
calioc2_handle_quirks(struct iommu_table * tbl,struct pci_dev * dev)962 static void __init calioc2_handle_quirks(struct iommu_table *tbl, struct pci_dev *dev)
963 {
964 	unsigned char busnum = dev->bus->number;
965 	void __iomem *bbar = tbl->bbar;
966 	void __iomem *target;
967 	u32 val;
968 
969 	/*
970 	 * CalIOC2 designers recommend setting bit 8 in 0xnDB0 to 1
971 	 */
972 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_SAVIOR_L2);
973 	val = cpu_to_be32(readl(target));
974 	val |= 0x00800000;
975 	writel(cpu_to_be32(val), target);
976 }
977 
calgary_handle_quirks(struct iommu_table * tbl,struct pci_dev * dev)978 static void __init calgary_handle_quirks(struct iommu_table *tbl, struct pci_dev *dev)
979 {
980 	unsigned char busnum = dev->bus->number;
981 
982 	/*
983 	 * Give split completion a longer timeout on bus 1 for aic94xx
984 	 * http://bugzilla.kernel.org/show_bug.cgi?id=7180
985 	 */
986 	if (is_calgary(dev->device) && (busnum == 1))
987 		calgary_set_split_completion_timeout(tbl->bbar, busnum,
988 						     CCR_2SEC_TIMEOUT);
989 }
990 
calgary_enable_translation(struct pci_dev * dev)991 static void __init calgary_enable_translation(struct pci_dev *dev)
992 {
993 	u32 val32;
994 	unsigned char busnum;
995 	void __iomem *target;
996 	void __iomem *bbar;
997 	struct iommu_table *tbl;
998 
999 	busnum = dev->bus->number;
1000 	tbl = pci_iommu(dev->bus);
1001 	bbar = tbl->bbar;
1002 
1003 	/* enable TCE in PHB Config Register */
1004 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_CONFIG_RW_OFFSET);
1005 	val32 = be32_to_cpu(readl(target));
1006 	val32 |= PHB_TCE_ENABLE | PHB_DAC_DISABLE | PHB_MCSR_ENABLE;
1007 
1008 	printk(KERN_INFO "Calgary: enabling translation on %s PHB %#x\n",
1009 	       (dev->device == PCI_DEVICE_ID_IBM_CALGARY) ?
1010 	       "Calgary" : "CalIOC2", busnum);
1011 	printk(KERN_INFO "Calgary: errant DMAs will now be prevented on this "
1012 	       "bus.\n");
1013 
1014 	writel(cpu_to_be32(val32), target);
1015 	readl(target); /* flush */
1016 
1017 	timer_setup(&tbl->watchdog_timer, calgary_watchdog, 0);
1018 	mod_timer(&tbl->watchdog_timer, jiffies);
1019 }
1020 
calgary_disable_translation(struct pci_dev * dev)1021 static void __init calgary_disable_translation(struct pci_dev *dev)
1022 {
1023 	u32 val32;
1024 	unsigned char busnum;
1025 	void __iomem *target;
1026 	void __iomem *bbar;
1027 	struct iommu_table *tbl;
1028 
1029 	busnum = dev->bus->number;
1030 	tbl = pci_iommu(dev->bus);
1031 	bbar = tbl->bbar;
1032 
1033 	/* disable TCE in PHB Config Register */
1034 	target = calgary_reg(bbar, phb_offset(busnum) | PHB_CONFIG_RW_OFFSET);
1035 	val32 = be32_to_cpu(readl(target));
1036 	val32 &= ~(PHB_TCE_ENABLE | PHB_DAC_DISABLE | PHB_MCSR_ENABLE);
1037 
1038 	printk(KERN_INFO "Calgary: disabling translation on PHB %#x!\n", busnum);
1039 	writel(cpu_to_be32(val32), target);
1040 	readl(target); /* flush */
1041 
1042 	del_timer_sync(&tbl->watchdog_timer);
1043 }
1044 
calgary_init_one_nontraslated(struct pci_dev * dev)1045 static void __init calgary_init_one_nontraslated(struct pci_dev *dev)
1046 {
1047 	pci_dev_get(dev);
1048 	set_pci_iommu(dev->bus, NULL);
1049 
1050 	/* is the device behind a bridge? */
1051 	if (dev->bus->parent)
1052 		dev->bus->parent->self = dev;
1053 	else
1054 		dev->bus->self = dev;
1055 }
1056 
calgary_init_one(struct pci_dev * dev)1057 static int __init calgary_init_one(struct pci_dev *dev)
1058 {
1059 	void __iomem *bbar;
1060 	struct iommu_table *tbl;
1061 	int ret;
1062 
1063 	bbar = busno_to_bbar(dev->bus->number);
1064 	ret = calgary_setup_tar(dev, bbar);
1065 	if (ret)
1066 		goto done;
1067 
1068 	pci_dev_get(dev);
1069 
1070 	if (dev->bus->parent) {
1071 		if (dev->bus->parent->self)
1072 			printk(KERN_WARNING "Calgary: IEEEE, dev %p has "
1073 			       "bus->parent->self!\n", dev);
1074 		dev->bus->parent->self = dev;
1075 	} else
1076 		dev->bus->self = dev;
1077 
1078 	tbl = pci_iommu(dev->bus);
1079 	tbl->chip_ops->handle_quirks(tbl, dev);
1080 
1081 	calgary_enable_translation(dev);
1082 
1083 	return 0;
1084 
1085 done:
1086 	return ret;
1087 }
1088 
calgary_locate_bbars(void)1089 static int __init calgary_locate_bbars(void)
1090 {
1091 	int ret;
1092 	int rioidx, phb, bus;
1093 	void __iomem *bbar;
1094 	void __iomem *target;
1095 	unsigned long offset;
1096 	u8 start_bus, end_bus;
1097 	u32 val;
1098 
1099 	ret = -ENODATA;
1100 	for (rioidx = 0; rioidx < rio_table_hdr->num_rio_dev; rioidx++) {
1101 		struct rio_detail *rio = rio_devs[rioidx];
1102 
1103 		if ((rio->type != COMPAT_CALGARY) && (rio->type != ALT_CALGARY))
1104 			continue;
1105 
1106 		/* map entire 1MB of Calgary config space */
1107 		bbar = ioremap_nocache(rio->BBAR, 1024 * 1024);
1108 		if (!bbar)
1109 			goto error;
1110 
1111 		for (phb = 0; phb < PHBS_PER_CALGARY; phb++) {
1112 			offset = phb_debug_offsets[phb] | PHB_DEBUG_STUFF_OFFSET;
1113 			target = calgary_reg(bbar, offset);
1114 
1115 			val = be32_to_cpu(readl(target));
1116 
1117 			start_bus = (u8)((val & 0x00FF0000) >> 16);
1118 			end_bus = (u8)((val & 0x0000FF00) >> 8);
1119 
1120 			if (end_bus) {
1121 				for (bus = start_bus; bus <= end_bus; bus++) {
1122 					bus_info[bus].bbar = bbar;
1123 					bus_info[bus].phbid = phb;
1124 				}
1125 			} else {
1126 				bus_info[start_bus].bbar = bbar;
1127 				bus_info[start_bus].phbid = phb;
1128 			}
1129 		}
1130 	}
1131 
1132 	return 0;
1133 
1134 error:
1135 	/* scan bus_info and iounmap any bbars we previously ioremap'd */
1136 	for (bus = 0; bus < ARRAY_SIZE(bus_info); bus++)
1137 		if (bus_info[bus].bbar)
1138 			iounmap(bus_info[bus].bbar);
1139 
1140 	return ret;
1141 }
1142 
calgary_init(void)1143 static int __init calgary_init(void)
1144 {
1145 	int ret;
1146 	struct pci_dev *dev = NULL;
1147 	struct calgary_bus_info *info;
1148 
1149 	ret = calgary_locate_bbars();
1150 	if (ret)
1151 		return ret;
1152 
1153 	/* Purely for kdump kernel case */
1154 	if (is_kdump_kernel())
1155 		get_tce_space_from_tar();
1156 
1157 	do {
1158 		dev = pci_get_device(PCI_VENDOR_ID_IBM, PCI_ANY_ID, dev);
1159 		if (!dev)
1160 			break;
1161 		if (!is_cal_pci_dev(dev->device))
1162 			continue;
1163 
1164 		info = &bus_info[dev->bus->number];
1165 		if (info->translation_disabled) {
1166 			calgary_init_one_nontraslated(dev);
1167 			continue;
1168 		}
1169 
1170 		if (!info->tce_space && !translate_empty_slots)
1171 			continue;
1172 
1173 		ret = calgary_init_one(dev);
1174 		if (ret)
1175 			goto error;
1176 	} while (1);
1177 
1178 	dev = NULL;
1179 	for_each_pci_dev(dev) {
1180 		struct iommu_table *tbl;
1181 
1182 		tbl = find_iommu_table(&dev->dev);
1183 
1184 		if (translation_enabled(tbl))
1185 			dev->dev.dma_ops = &calgary_dma_ops;
1186 	}
1187 
1188 	return ret;
1189 
1190 error:
1191 	do {
1192 		dev = pci_get_device(PCI_VENDOR_ID_IBM, PCI_ANY_ID, dev);
1193 		if (!dev)
1194 			break;
1195 		if (!is_cal_pci_dev(dev->device))
1196 			continue;
1197 
1198 		info = &bus_info[dev->bus->number];
1199 		if (info->translation_disabled) {
1200 			pci_dev_put(dev);
1201 			continue;
1202 		}
1203 		if (!info->tce_space && !translate_empty_slots)
1204 			continue;
1205 
1206 		calgary_disable_translation(dev);
1207 		calgary_free_bus(dev);
1208 		pci_dev_put(dev); /* Undo calgary_init_one()'s pci_dev_get() */
1209 		dev->dev.dma_ops = NULL;
1210 	} while (1);
1211 
1212 	return ret;
1213 }
1214 
determine_tce_table_size(void)1215 static inline int __init determine_tce_table_size(void)
1216 {
1217 	int ret;
1218 
1219 	if (specified_table_size != TCE_TABLE_SIZE_UNSPECIFIED)
1220 		return specified_table_size;
1221 
1222 	if (is_kdump_kernel() && saved_max_pfn) {
1223 		/*
1224 		 * Table sizes are from 0 to 7 (TCE_TABLE_SIZE_64K to
1225 		 * TCE_TABLE_SIZE_8M). Table size 0 has 8K entries and each
1226 		 * larger table size has twice as many entries, so shift the
1227 		 * max ram address by 13 to divide by 8K and then look at the
1228 		 * order of the result to choose between 0-7.
1229 		 */
1230 		ret = get_order((saved_max_pfn * PAGE_SIZE) >> 13);
1231 		if (ret > TCE_TABLE_SIZE_8M)
1232 			ret = TCE_TABLE_SIZE_8M;
1233 	} else {
1234 		/*
1235 		 * Use 8M by default (suggested by Muli) if it's not
1236 		 * kdump kernel and saved_max_pfn isn't set.
1237 		 */
1238 		ret = TCE_TABLE_SIZE_8M;
1239 	}
1240 
1241 	return ret;
1242 }
1243 
build_detail_arrays(void)1244 static int __init build_detail_arrays(void)
1245 {
1246 	unsigned long ptr;
1247 	unsigned numnodes, i;
1248 	int scal_detail_size, rio_detail_size;
1249 
1250 	numnodes = rio_table_hdr->num_scal_dev;
1251 	if (numnodes > MAX_NUMNODES){
1252 		printk(KERN_WARNING
1253 			"Calgary: MAX_NUMNODES too low! Defined as %d, "
1254 			"but system has %d nodes.\n",
1255 			MAX_NUMNODES, numnodes);
1256 		return -ENODEV;
1257 	}
1258 
1259 	switch (rio_table_hdr->version){
1260 	case 2:
1261 		scal_detail_size = 11;
1262 		rio_detail_size = 13;
1263 		break;
1264 	case 3:
1265 		scal_detail_size = 12;
1266 		rio_detail_size = 15;
1267 		break;
1268 	default:
1269 		printk(KERN_WARNING
1270 		       "Calgary: Invalid Rio Grande Table Version: %d\n",
1271 		       rio_table_hdr->version);
1272 		return -EPROTO;
1273 	}
1274 
1275 	ptr = ((unsigned long)rio_table_hdr) + 3;
1276 	for (i = 0; i < numnodes; i++, ptr += scal_detail_size)
1277 		scal_devs[i] = (struct scal_detail *)ptr;
1278 
1279 	for (i = 0; i < rio_table_hdr->num_rio_dev;
1280 		    i++, ptr += rio_detail_size)
1281 		rio_devs[i] = (struct rio_detail *)ptr;
1282 
1283 	return 0;
1284 }
1285 
calgary_bus_has_devices(int bus,unsigned short pci_dev)1286 static int __init calgary_bus_has_devices(int bus, unsigned short pci_dev)
1287 {
1288 	int dev;
1289 	u32 val;
1290 
1291 	if (pci_dev == PCI_DEVICE_ID_IBM_CALIOC2) {
1292 		/*
1293 		 * FIXME: properly scan for devices across the
1294 		 * PCI-to-PCI bridge on every CalIOC2 port.
1295 		 */
1296 		return 1;
1297 	}
1298 
1299 	for (dev = 1; dev < 8; dev++) {
1300 		val = read_pci_config(bus, dev, 0, 0);
1301 		if (val != 0xffffffff)
1302 			break;
1303 	}
1304 	return (val != 0xffffffff);
1305 }
1306 
1307 /*
1308  * calgary_init_bitmap_from_tce_table():
1309  * Function for kdump case. In the second/kdump kernel initialize
1310  * the bitmap based on the tce table entries obtained from first kernel
1311  */
calgary_init_bitmap_from_tce_table(struct iommu_table * tbl)1312 static void calgary_init_bitmap_from_tce_table(struct iommu_table *tbl)
1313 {
1314 	u64 *tp;
1315 	unsigned int index;
1316 	tp = ((u64 *)tbl->it_base);
1317 	for (index = 0 ; index < tbl->it_size; index++) {
1318 		if (*tp != 0x0)
1319 			set_bit(index, tbl->it_map);
1320 		tp++;
1321 	}
1322 }
1323 
1324 /*
1325  * get_tce_space_from_tar():
1326  * Function for kdump case. Get the tce tables from first kernel
1327  * by reading the contents of the base address register of calgary iommu
1328  */
get_tce_space_from_tar(void)1329 static void __init get_tce_space_from_tar(void)
1330 {
1331 	int bus;
1332 	void __iomem *target;
1333 	unsigned long tce_space;
1334 
1335 	for (bus = 0; bus < MAX_PHB_BUS_NUM; bus++) {
1336 		struct calgary_bus_info *info = &bus_info[bus];
1337 		unsigned short pci_device;
1338 		u32 val;
1339 
1340 		val = read_pci_config(bus, 0, 0, 0);
1341 		pci_device = (val & 0xFFFF0000) >> 16;
1342 
1343 		if (!is_cal_pci_dev(pci_device))
1344 			continue;
1345 		if (info->translation_disabled)
1346 			continue;
1347 
1348 		if (calgary_bus_has_devices(bus, pci_device) ||
1349 						translate_empty_slots) {
1350 			target = calgary_reg(bus_info[bus].bbar,
1351 						tar_offset(bus));
1352 			tce_space = be64_to_cpu(readq(target));
1353 			tce_space = tce_space & TAR_SW_BITS;
1354 
1355 			tce_space = tce_space & (~specified_table_size);
1356 			info->tce_space = (u64 *)__va(tce_space);
1357 		}
1358 	}
1359 	return;
1360 }
1361 
calgary_iommu_init(void)1362 static int __init calgary_iommu_init(void)
1363 {
1364 	int ret;
1365 
1366 	/* ok, we're trying to use Calgary - let's roll */
1367 	printk(KERN_INFO "PCI-DMA: Using Calgary IOMMU\n");
1368 
1369 	ret = calgary_init();
1370 	if (ret) {
1371 		printk(KERN_ERR "PCI-DMA: Calgary init failed %d, "
1372 		       "falling back to no_iommu\n", ret);
1373 		return ret;
1374 	}
1375 
1376 	return 0;
1377 }
1378 
detect_calgary(void)1379 int __init detect_calgary(void)
1380 {
1381 	int bus;
1382 	void *tbl;
1383 	int calgary_found = 0;
1384 	unsigned long ptr;
1385 	unsigned int offset, prev_offset;
1386 	int ret;
1387 
1388 	/*
1389 	 * if the user specified iommu=off or iommu=soft or we found
1390 	 * another HW IOMMU already, bail out.
1391 	 */
1392 	if (no_iommu || iommu_detected)
1393 		return -ENODEV;
1394 
1395 	if (!use_calgary)
1396 		return -ENODEV;
1397 
1398 	if (!early_pci_allowed())
1399 		return -ENODEV;
1400 
1401 	printk(KERN_DEBUG "Calgary: detecting Calgary via BIOS EBDA area\n");
1402 
1403 	ptr = (unsigned long)phys_to_virt(get_bios_ebda());
1404 
1405 	rio_table_hdr = NULL;
1406 	prev_offset = 0;
1407 	offset = 0x180;
1408 	/*
1409 	 * The next offset is stored in the 1st word.
1410 	 * Only parse up until the offset increases:
1411 	 */
1412 	while (offset > prev_offset) {
1413 		/* The block id is stored in the 2nd word */
1414 		if (*((unsigned short *)(ptr + offset + 2)) == 0x4752){
1415 			/* set the pointer past the offset & block id */
1416 			rio_table_hdr = (struct rio_table_hdr *)(ptr + offset + 4);
1417 			break;
1418 		}
1419 		prev_offset = offset;
1420 		offset = *((unsigned short *)(ptr + offset));
1421 	}
1422 	if (!rio_table_hdr) {
1423 		printk(KERN_DEBUG "Calgary: Unable to locate Rio Grande table "
1424 		       "in EBDA - bailing!\n");
1425 		return -ENODEV;
1426 	}
1427 
1428 	ret = build_detail_arrays();
1429 	if (ret) {
1430 		printk(KERN_DEBUG "Calgary: build_detail_arrays ret %d\n", ret);
1431 		return -ENOMEM;
1432 	}
1433 
1434 	specified_table_size = determine_tce_table_size();
1435 
1436 	for (bus = 0; bus < MAX_PHB_BUS_NUM; bus++) {
1437 		struct calgary_bus_info *info = &bus_info[bus];
1438 		unsigned short pci_device;
1439 		u32 val;
1440 
1441 		val = read_pci_config(bus, 0, 0, 0);
1442 		pci_device = (val & 0xFFFF0000) >> 16;
1443 
1444 		if (!is_cal_pci_dev(pci_device))
1445 			continue;
1446 
1447 		if (info->translation_disabled)
1448 			continue;
1449 
1450 		if (calgary_bus_has_devices(bus, pci_device) ||
1451 		    translate_empty_slots) {
1452 			/*
1453 			 * If it is kdump kernel, find and use tce tables
1454 			 * from first kernel, else allocate tce tables here
1455 			 */
1456 			if (!is_kdump_kernel()) {
1457 				tbl = alloc_tce_table();
1458 				if (!tbl)
1459 					goto cleanup;
1460 				info->tce_space = tbl;
1461 			}
1462 			calgary_found = 1;
1463 		}
1464 	}
1465 
1466 	printk(KERN_DEBUG "Calgary: finished detection, Calgary %s\n",
1467 	       calgary_found ? "found" : "not found");
1468 
1469 	if (calgary_found) {
1470 		iommu_detected = 1;
1471 		calgary_detected = 1;
1472 		printk(KERN_INFO "PCI-DMA: Calgary IOMMU detected.\n");
1473 		printk(KERN_INFO "PCI-DMA: Calgary TCE table spec is %d\n",
1474 		       specified_table_size);
1475 
1476 		x86_init.iommu.iommu_init = calgary_iommu_init;
1477 	}
1478 	return calgary_found;
1479 
1480 cleanup:
1481 	for (--bus; bus >= 0; --bus) {
1482 		struct calgary_bus_info *info = &bus_info[bus];
1483 
1484 		if (info->tce_space)
1485 			free_tce_table(info->tce_space);
1486 	}
1487 	return -ENOMEM;
1488 }
1489 
calgary_parse_options(char * p)1490 static int __init calgary_parse_options(char *p)
1491 {
1492 	unsigned int bridge;
1493 	unsigned long val;
1494 	size_t len;
1495 	ssize_t ret;
1496 
1497 	while (*p) {
1498 		if (!strncmp(p, "64k", 3))
1499 			specified_table_size = TCE_TABLE_SIZE_64K;
1500 		else if (!strncmp(p, "128k", 4))
1501 			specified_table_size = TCE_TABLE_SIZE_128K;
1502 		else if (!strncmp(p, "256k", 4))
1503 			specified_table_size = TCE_TABLE_SIZE_256K;
1504 		else if (!strncmp(p, "512k", 4))
1505 			specified_table_size = TCE_TABLE_SIZE_512K;
1506 		else if (!strncmp(p, "1M", 2))
1507 			specified_table_size = TCE_TABLE_SIZE_1M;
1508 		else if (!strncmp(p, "2M", 2))
1509 			specified_table_size = TCE_TABLE_SIZE_2M;
1510 		else if (!strncmp(p, "4M", 2))
1511 			specified_table_size = TCE_TABLE_SIZE_4M;
1512 		else if (!strncmp(p, "8M", 2))
1513 			specified_table_size = TCE_TABLE_SIZE_8M;
1514 
1515 		len = strlen("translate_empty_slots");
1516 		if (!strncmp(p, "translate_empty_slots", len))
1517 			translate_empty_slots = 1;
1518 
1519 		len = strlen("disable");
1520 		if (!strncmp(p, "disable", len)) {
1521 			p += len;
1522 			if (*p == '=')
1523 				++p;
1524 			if (*p == '\0')
1525 				break;
1526 			ret = kstrtoul(p, 0, &val);
1527 			if (ret)
1528 				break;
1529 
1530 			bridge = val;
1531 			if (bridge < MAX_PHB_BUS_NUM) {
1532 				printk(KERN_INFO "Calgary: disabling "
1533 				       "translation for PHB %#x\n", bridge);
1534 				bus_info[bridge].translation_disabled = 1;
1535 			}
1536 		}
1537 
1538 		p = strpbrk(p, ",");
1539 		if (!p)
1540 			break;
1541 
1542 		p++; /* skip ',' */
1543 	}
1544 	return 1;
1545 }
1546 __setup("calgary=", calgary_parse_options);
1547 
calgary_fixup_one_tce_space(struct pci_dev * dev)1548 static void __init calgary_fixup_one_tce_space(struct pci_dev *dev)
1549 {
1550 	struct iommu_table *tbl;
1551 	unsigned int npages;
1552 	int i;
1553 
1554 	tbl = pci_iommu(dev->bus);
1555 
1556 	for (i = 0; i < 4; i++) {
1557 		struct resource *r = &dev->resource[PCI_BRIDGE_RESOURCES + i];
1558 
1559 		/* Don't give out TCEs that map MEM resources */
1560 		if (!(r->flags & IORESOURCE_MEM))
1561 			continue;
1562 
1563 		/* 0-based? we reserve the whole 1st MB anyway */
1564 		if (!r->start)
1565 			continue;
1566 
1567 		/* cover the whole region */
1568 		npages = resource_size(r) >> PAGE_SHIFT;
1569 		npages++;
1570 
1571 		iommu_range_reserve(tbl, r->start, npages);
1572 	}
1573 }
1574 
calgary_fixup_tce_spaces(void)1575 static int __init calgary_fixup_tce_spaces(void)
1576 {
1577 	struct pci_dev *dev = NULL;
1578 	struct calgary_bus_info *info;
1579 
1580 	if (no_iommu || swiotlb || !calgary_detected)
1581 		return -ENODEV;
1582 
1583 	printk(KERN_DEBUG "Calgary: fixing up tce spaces\n");
1584 
1585 	do {
1586 		dev = pci_get_device(PCI_VENDOR_ID_IBM, PCI_ANY_ID, dev);
1587 		if (!dev)
1588 			break;
1589 		if (!is_cal_pci_dev(dev->device))
1590 			continue;
1591 
1592 		info = &bus_info[dev->bus->number];
1593 		if (info->translation_disabled)
1594 			continue;
1595 
1596 		if (!info->tce_space)
1597 			continue;
1598 
1599 		calgary_fixup_one_tce_space(dev);
1600 
1601 	} while (1);
1602 
1603 	return 0;
1604 }
1605 
1606 /*
1607  * We need to be call after pcibios_assign_resources (fs_initcall level)
1608  * and before device_initcall.
1609  */
1610 rootfs_initcall(calgary_fixup_tce_spaces);
1611 
1612 IOMMU_INIT_POST(detect_calgary);
1613