Lines Matching refs:r

226 static int ps3_mm_get_repository_highmem(struct mem_region *r)  in ps3_mm_get_repository_highmem()  argument
232 result = ps3_repository_read_highmem_info(0, &r->base, &r->size); in ps3_mm_get_repository_highmem()
237 if (!r->base || !r->size) { in ps3_mm_get_repository_highmem()
242 r->offset = r->base - map.rm.size; in ps3_mm_get_repository_highmem()
245 __func__, __LINE__, r->base, r->size); in ps3_mm_get_repository_highmem()
252 r->size = r->base = r->offset = 0; in ps3_mm_get_repository_highmem()
256 static int ps3_mm_set_repository_highmem(const struct mem_region *r) in ps3_mm_set_repository_highmem() argument
260 return r ? ps3_repository_write_highmem_info(0, r->base, r->size) : in ps3_mm_set_repository_highmem()
273 static int ps3_mm_region_create(struct mem_region *r, unsigned long size) in ps3_mm_region_create() argument
278 r->size = _ALIGN_DOWN(size, 1 << PAGE_SHIFT_16M); in ps3_mm_region_create()
281 DBG("%s:%d actual %llxh\n", __func__, __LINE__, r->size); in ps3_mm_region_create()
283 size - r->size, (size - r->size) / 1024 / 1024); in ps3_mm_region_create()
285 if (r->size == 0) { in ps3_mm_region_create()
291 result = lv1_allocate_memory(r->size, PAGE_SHIFT_16M, 0, in ps3_mm_region_create()
292 ALLOCATE_MEMORY_TRY_ALT_UNIT, &r->base, &muid); in ps3_mm_region_create()
294 if (result || r->base < map.rm.size) { in ps3_mm_region_create()
300 r->destroy = 1; in ps3_mm_region_create()
301 r->offset = r->base - map.rm.size; in ps3_mm_region_create()
305 r->size = r->base = r->offset = 0; in ps3_mm_region_create()
314 static void ps3_mm_region_destroy(struct mem_region *r) in ps3_mm_region_destroy() argument
318 if (!r->destroy) { in ps3_mm_region_destroy()
320 __func__, __LINE__, r->base, r->size); in ps3_mm_region_destroy()
324 DBG("%s:%d: r->base = %llxh\n", __func__, __LINE__, r->base); in ps3_mm_region_destroy()
326 if (r->base) { in ps3_mm_region_destroy()
327 result = lv1_release_memory(r->base); in ps3_mm_region_destroy()
329 r->size = r->base = r->offset = 0; in ps3_mm_region_destroy()
345 static unsigned long dma_sb_lpar_to_bus(struct ps3_dma_region *r, in dma_sb_lpar_to_bus() argument
350 BUG_ON(lpar_addr < r->offset); in dma_sb_lpar_to_bus()
351 BUG_ON(lpar_addr >= r->offset + r->len); in dma_sb_lpar_to_bus()
352 return r->bus_addr + lpar_addr - r->offset; in dma_sb_lpar_to_bus()
356 static void __maybe_unused _dma_dump_region(const struct ps3_dma_region *r, in _dma_dump_region() argument
359 DBG("%s:%d: dev %llu:%llu\n", func, line, r->dev->bus_id, in _dma_dump_region()
360 r->dev->dev_id); in _dma_dump_region()
361 DBG("%s:%d: page_size %u\n", func, line, r->page_size); in _dma_dump_region()
362 DBG("%s:%d: bus_addr %lxh\n", func, line, r->bus_addr); in _dma_dump_region()
363 DBG("%s:%d: len %lxh\n", func, line, r->len); in _dma_dump_region()
364 DBG("%s:%d: offset %lxh\n", func, line, r->offset); in _dma_dump_region()
405 static struct dma_chunk * dma_find_chunk(struct ps3_dma_region *r, in dma_find_chunk() argument
409 unsigned long aligned_bus = _ALIGN_DOWN(bus_addr, 1 << r->page_size); in dma_find_chunk()
411 1 << r->page_size); in dma_find_chunk()
413 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk()
434 static struct dma_chunk *dma_find_chunk_lpar(struct ps3_dma_region *r, in dma_find_chunk_lpar() argument
438 unsigned long aligned_lpar = _ALIGN_DOWN(lpar_addr, 1 << r->page_size); in dma_find_chunk_lpar()
440 1 << r->page_size); in dma_find_chunk_lpar()
442 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk_lpar()
484 struct ps3_dma_region *r = c->region; in dma_ioc0_free_chunk() local
487 for (iopage = 0; iopage < (c->len >> r->page_size); iopage++) { in dma_ioc0_free_chunk()
488 offset = (1 << r->page_size) * iopage; in dma_ioc0_free_chunk()
493 r->ioid, in dma_ioc0_free_chunk()
498 r->ioid); in dma_ioc0_free_chunk()
521 static int dma_sb_map_pages(struct ps3_dma_region *r, unsigned long phys_addr, in dma_sb_map_pages() argument
533 c->region = r; in dma_sb_map_pages()
535 c->bus_addr = dma_sb_lpar_to_bus(r, c->lpar_addr); in dma_sb_map_pages()
548 list_add(&c->link, &r->chunk_list.head); in dma_sb_map_pages()
561 static int dma_ioc0_map_pages(struct ps3_dma_region *r, unsigned long phys_addr, in dma_ioc0_map_pages() argument
578 c->region = r; in dma_ioc0_map_pages()
582 if (list_empty(&r->chunk_list.head)) { in dma_ioc0_map_pages()
584 c->bus_addr = r->bus_addr; in dma_ioc0_map_pages()
587 last = list_entry(r->chunk_list.head.next, in dma_ioc0_map_pages()
597 pages = len >> r->page_size; in dma_ioc0_map_pages()
599 r->page_size, r->len, pages, iopte_flag); in dma_ioc0_map_pages()
601 offset = (1 << r->page_size) * iopage; in dma_ioc0_map_pages()
605 r->ioid, in dma_ioc0_map_pages()
614 r->ioid); in dma_ioc0_map_pages()
618 list_add(&c->link, &r->chunk_list.head); in dma_ioc0_map_pages()
629 r->ioid, in dma_ioc0_map_pages()
646 static int dma_sb_region_create(struct ps3_dma_region *r) in dma_sb_region_create() argument
653 BUG_ON(!r); in dma_sb_region_create()
655 if (!r->dev->bus_id) { in dma_sb_region_create()
657 r->dev->bus_id, r->dev->dev_id); in dma_sb_region_create()
662 __LINE__, r->len, r->page_size, r->offset); in dma_sb_region_create()
664 BUG_ON(!r->len); in dma_sb_region_create()
665 BUG_ON(!r->page_size); in dma_sb_region_create()
666 BUG_ON(!r->region_ops); in dma_sb_region_create()
668 INIT_LIST_HEAD(&r->chunk_list.head); in dma_sb_region_create()
669 spin_lock_init(&r->chunk_list.lock); in dma_sb_region_create()
671 result = lv1_allocate_device_dma_region(r->dev->bus_id, r->dev->dev_id, in dma_sb_region_create()
672 roundup_pow_of_two(r->len), r->page_size, r->region_type, in dma_sb_region_create()
674 r->bus_addr = bus_addr; in dma_sb_region_create()
679 r->len = r->bus_addr = 0; in dma_sb_region_create()
685 static int dma_ioc0_region_create(struct ps3_dma_region *r) in dma_ioc0_region_create() argument
690 INIT_LIST_HEAD(&r->chunk_list.head); in dma_ioc0_region_create()
691 spin_lock_init(&r->chunk_list.lock); in dma_ioc0_region_create()
694 r->len, in dma_ioc0_region_create()
695 r->page_size, in dma_ioc0_region_create()
697 r->bus_addr = bus_addr; in dma_ioc0_region_create()
701 r->len = r->bus_addr = 0; in dma_ioc0_region_create()
704 r->len, r->page_size, r->bus_addr); in dma_ioc0_region_create()
716 static int dma_sb_region_free(struct ps3_dma_region *r) in dma_sb_region_free() argument
722 BUG_ON(!r); in dma_sb_region_free()
724 if (!r->dev->bus_id) { in dma_sb_region_free()
726 r->dev->bus_id, r->dev->dev_id); in dma_sb_region_free()
730 list_for_each_entry_safe(c, tmp, &r->chunk_list.head, link) { in dma_sb_region_free()
735 result = lv1_free_device_dma_region(r->dev->bus_id, r->dev->dev_id, in dma_sb_region_free()
736 r->bus_addr); in dma_sb_region_free()
742 r->bus_addr = 0; in dma_sb_region_free()
747 static int dma_ioc0_region_free(struct ps3_dma_region *r) in dma_ioc0_region_free() argument
753 list_for_each_entry_safe(c, n, &r->chunk_list.head, link) { in dma_ioc0_region_free()
758 result = lv1_release_io_segment(0, r->bus_addr); in dma_ioc0_region_free()
764 r->bus_addr = 0; in dma_ioc0_region_free()
781 static int dma_sb_map_area(struct ps3_dma_region *r, unsigned long virt_addr, in dma_sb_map_area() argument
790 unsigned long aligned_phys = _ALIGN_DOWN(phys_addr, 1 << r->page_size); in dma_sb_map_area()
792 1 << r->page_size); in dma_sb_map_area()
793 *bus_addr = dma_sb_lpar_to_bus(r, ps3_mm_phys_to_lpar(phys_addr)); in dma_sb_map_area()
809 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_sb_map_area()
810 c = dma_find_chunk(r, *bus_addr, len); in dma_sb_map_area()
816 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_map_area()
820 result = dma_sb_map_pages(r, aligned_phys, aligned_len, &c, iopte_flag); in dma_sb_map_area()
826 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_map_area()
832 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_map_area()
836 static int dma_ioc0_map_area(struct ps3_dma_region *r, unsigned long virt_addr, in dma_ioc0_map_area() argument
845 unsigned long aligned_phys = _ALIGN_DOWN(phys_addr, 1 << r->page_size); in dma_ioc0_map_area()
847 1 << r->page_size); in dma_ioc0_map_area()
854 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
855 c = dma_find_chunk_lpar(r, ps3_mm_phys_to_lpar(phys_addr), len); in dma_ioc0_map_area()
862 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
866 result = dma_ioc0_map_pages(r, aligned_phys, aligned_len, &c, in dma_ioc0_map_area()
873 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
881 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
894 static int dma_sb_unmap_area(struct ps3_dma_region *r, dma_addr_t bus_addr, in dma_sb_unmap_area() argument
900 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_sb_unmap_area()
901 c = dma_find_chunk(r, bus_addr, len); in dma_sb_unmap_area()
905 1 << r->page_size); in dma_sb_unmap_area()
907 - aligned_bus, 1 << r->page_size); in dma_sb_unmap_area()
926 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_unmap_area()
930 static int dma_ioc0_unmap_area(struct ps3_dma_region *r, in dma_ioc0_unmap_area() argument
937 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_ioc0_unmap_area()
938 c = dma_find_chunk(r, bus_addr, len); in dma_ioc0_unmap_area()
942 1 << r->page_size); in dma_ioc0_unmap_area()
945 1 << r->page_size); in dma_ioc0_unmap_area()
964 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_unmap_area()
977 static int dma_sb_region_create_linear(struct ps3_dma_region *r) in dma_sb_region_create_linear() argument
983 if (r->len > 16*1024*1024) { /* FIXME: need proper fix */ in dma_sb_region_create_linear()
985 if (r->page_size != PS3_DMA_16M) { in dma_sb_region_create_linear()
988 r->page_size = PS3_DMA_16M; in dma_sb_region_create_linear()
989 r->len = _ALIGN_UP(r->len, 1 << r->page_size); in dma_sb_region_create_linear()
993 result = dma_sb_region_create(r); in dma_sb_region_create_linear()
996 if (r->offset < map.rm.size) { in dma_sb_region_create_linear()
998 virt_addr = map.rm.base + r->offset; in dma_sb_region_create_linear()
999 len = map.rm.size - r->offset; in dma_sb_region_create_linear()
1000 if (len > r->len) in dma_sb_region_create_linear()
1001 len = r->len; in dma_sb_region_create_linear()
1002 result = dma_sb_map_area(r, virt_addr, len, &tmp, in dma_sb_region_create_linear()
1008 if (r->offset + r->len > map.rm.size) { in dma_sb_region_create_linear()
1011 len = r->len; in dma_sb_region_create_linear()
1012 if (r->offset >= map.rm.size) in dma_sb_region_create_linear()
1013 virt_addr += r->offset - map.rm.size; in dma_sb_region_create_linear()
1015 len -= map.rm.size - r->offset; in dma_sb_region_create_linear()
1016 result = dma_sb_map_area(r, virt_addr, len, &tmp, in dma_sb_region_create_linear()
1032 static int dma_sb_region_free_linear(struct ps3_dma_region *r) in dma_sb_region_free_linear() argument
1038 if (r->offset < map.rm.size) { in dma_sb_region_free_linear()
1040 lpar_addr = map.rm.base + r->offset; in dma_sb_region_free_linear()
1041 len = map.rm.size - r->offset; in dma_sb_region_free_linear()
1042 if (len > r->len) in dma_sb_region_free_linear()
1043 len = r->len; in dma_sb_region_free_linear()
1044 bus_addr = dma_sb_lpar_to_bus(r, lpar_addr); in dma_sb_region_free_linear()
1045 result = dma_sb_unmap_area(r, bus_addr, len); in dma_sb_region_free_linear()
1049 if (r->offset + r->len > map.rm.size) { in dma_sb_region_free_linear()
1052 len = r->len; in dma_sb_region_free_linear()
1053 if (r->offset >= map.rm.size) in dma_sb_region_free_linear()
1054 lpar_addr += r->offset - map.rm.size; in dma_sb_region_free_linear()
1056 len -= map.rm.size - r->offset; in dma_sb_region_free_linear()
1057 bus_addr = dma_sb_lpar_to_bus(r, lpar_addr); in dma_sb_region_free_linear()
1058 result = dma_sb_unmap_area(r, bus_addr, len); in dma_sb_region_free_linear()
1062 result = dma_sb_region_free(r); in dma_sb_region_free_linear()
1080 static int dma_sb_map_area_linear(struct ps3_dma_region *r, in dma_sb_map_area_linear() argument
1086 *bus_addr = dma_sb_lpar_to_bus(r, ps3_mm_phys_to_lpar(phys_addr)); in dma_sb_map_area_linear()
1099 static int dma_sb_unmap_area_linear(struct ps3_dma_region *r, in dma_sb_unmap_area_linear() argument
1127 struct ps3_dma_region *r, enum ps3_dma_page_size page_size, in ps3_dma_region_init() argument
1134 r->dev = dev; in ps3_dma_region_init()
1135 r->page_size = page_size; in ps3_dma_region_init()
1136 r->region_type = region_type; in ps3_dma_region_init()
1137 r->offset = lpar_addr; in ps3_dma_region_init()
1138 if (r->offset >= map.rm.size) in ps3_dma_region_init()
1139 r->offset -= map.r1.offset; in ps3_dma_region_init()
1140 r->len = len ? len : _ALIGN_UP(map.total, 1 << r->page_size); in ps3_dma_region_init()
1144 r->region_ops = (USE_DYNAMIC_DMA) in ps3_dma_region_init()
1149 r->region_ops = &ps3_dma_ioc0_region_ops; in ps3_dma_region_init()
1159 int ps3_dma_region_create(struct ps3_dma_region *r) in ps3_dma_region_create() argument
1161 BUG_ON(!r); in ps3_dma_region_create()
1162 BUG_ON(!r->region_ops); in ps3_dma_region_create()
1163 BUG_ON(!r->region_ops->create); in ps3_dma_region_create()
1164 return r->region_ops->create(r); in ps3_dma_region_create()
1168 int ps3_dma_region_free(struct ps3_dma_region *r) in ps3_dma_region_free() argument
1170 BUG_ON(!r); in ps3_dma_region_free()
1171 BUG_ON(!r->region_ops); in ps3_dma_region_free()
1172 BUG_ON(!r->region_ops->free); in ps3_dma_region_free()
1173 return r->region_ops->free(r); in ps3_dma_region_free()
1177 int ps3_dma_map(struct ps3_dma_region *r, unsigned long virt_addr, in ps3_dma_map() argument
1181 return r->region_ops->map(r, virt_addr, len, bus_addr, iopte_flag); in ps3_dma_map()
1184 int ps3_dma_unmap(struct ps3_dma_region *r, dma_addr_t bus_addr, in ps3_dma_unmap() argument
1187 return r->region_ops->unmap(r, bus_addr, len); in ps3_dma_unmap()