/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | mem.c | 58 unsigned long page_shift = umem->page_shift; in mlx5_ib_cont_pages() local 62 *count = *ncont << (page_shift - PAGE_SHIFT); in mlx5_ib_cont_pages() 63 *shift = page_shift; in mlx5_ib_cont_pages() 70 addr = addr >> page_shift; in mlx5_ib_cont_pages() 74 m = min_t(unsigned long, max_page_shift - page_shift, m); in mlx5_ib_cont_pages() 77 len = sg_dma_len(sg) >> page_shift; in mlx5_ib_cont_pages() 78 pfn = sg_dma_address(sg) >> page_shift; in mlx5_ib_cont_pages() 110 *shift = page_shift + m; in mlx5_ib_cont_pages() 142 int page_shift, size_t offset, size_t num_pages, in __mlx5_ib_populate_pas() argument 145 unsigned long umem_page_shift = umem->page_shift; in __mlx5_ib_populate_pas() [all …]
|
D | mr.c | 832 static int get_octo_len(u64 addr, u64 len, int page_shift) in get_octo_len() argument 834 u64 page_size = 1ULL << page_shift; in get_octo_len() 839 npages = ALIGN(len + offset, page_size) >> page_shift; in get_octo_len() 852 int *npages, int *page_shift, int *ncont, in mr_umem_get() argument 869 page_shift, ncont, order); in mr_umem_get() 879 *npages, *ncont, *order, *page_shift); in mr_umem_get() 930 int page_shift, int order, int access_flags) in alloc_mr_from_cache() argument 964 void *xlt, int page_shift, size_t size, in populate_xlt() argument 980 __mlx5_ib_populate_pas(dev, umem, page_shift, in populate_xlt() 998 int page_shift, int flags) in mlx5_ib_update_xlt() argument [all …]
|
D | srq.c | 84 int page_shift; in create_srq_user() local 122 &page_shift, &ncont, NULL); in create_srq_user() 123 err = mlx5_ib_get_buf_offset(ucmd.buf_addr, page_shift, in create_srq_user() 136 mlx5_ib_populate_pas(dev, srq->umem, page_shift, in->pas, 0); in create_srq_user() 145 in->log_page_size = page_shift - MLX5_ADAPTER_PAGE_SHIFT; in create_srq_user() 191 mlx5_ib_dbg(dev, "srq->buf.page_shift = %d\n", srq->buf.page_shift); in create_srq_kernel() 206 in->log_page_size = srq->buf.page_shift - MLX5_ADAPTER_PAGE_SHIFT; in create_srq_kernel()
|
D | odp.c | 203 for (addr = start; addr < end; addr += BIT(umem->page_shift)) { in mlx5_ib_invalidate_range() 204 idx = (addr - ib_umem_start(umem)) >> umem->page_shift; in mlx5_ib_invalidate_range() 501 int npages = 0, page_shift, np; in pagefault_mr() local 522 page_shift = mr->umem->page_shift; in pagefault_mr() 523 page_mask = ~(BIT(page_shift) - 1); in pagefault_mr() 524 start_idx = (io_virt - (mr->mmkey.iova & page_mask)) >> page_shift; in pagefault_mr() 552 page_shift, MLX5_IB_UPD_XLT_ATOMIC); in pagefault_mr() 565 u32 new_mappings = (np << page_shift) - in pagefault_mr() 566 (io_virt - round_down(io_virt, 1 << page_shift)); in pagefault_mr() 570 npages += np << (page_shift - PAGE_SHIFT); in pagefault_mr()
|
D | cq.c | 783 int page_shift; in create_cq_user() local 818 mlx5_ib_cont_pages(cq->buf.umem, ucmd.buf_addr, 0, &npages, &page_shift, in create_cq_user() 821 ucmd.buf_addr, entries * ucmd.cqe_size, npages, page_shift, ncont); in create_cq_user() 832 mlx5_ib_populate_pas(dev, cq->buf.umem, page_shift, pas, 0); in create_cq_user() 836 page_shift - MLX5_ADAPTER_PAGE_SHIFT); in create_cq_user() 949 cq->buf.fbc.frag_buf.page_shift - in create_cq_kernel() 1202 int *page_shift, int *cqe_size) in resize_user() argument 1229 mlx5_ib_cont_pages(umem, ucmd.buf_addr, 0, &npages, page_shift, in resize_user() 1335 int page_shift; in mlx5_ib_resize_cq() local 1362 err = resize_user(dev, cq, entries, udata, &npas, &page_shift, in mlx5_ib_resize_cq() [all …]
|
/Linux-v4.19/drivers/pci/endpoint/ |
D | pci-epc-mem.c | 26 unsigned int page_shift = ilog2(mem->page_size); in pci_epc_mem_get_order() local 29 size >>= page_shift; in pci_epc_mem_get_order() 54 unsigned int page_shift; in __pci_epc_mem_init() local 61 page_shift = ilog2(page_size); in __pci_epc_mem_init() 62 pages = size >> page_shift; in __pci_epc_mem_init() 127 unsigned int page_shift = ilog2(mem->page_size); in pci_epc_mem_alloc_addr() local 137 *phys_addr = mem->phys_base + (pageno << page_shift); in pci_epc_mem_alloc_addr() 160 unsigned int page_shift = ilog2(mem->page_size); in pci_epc_mem_free_addr() local 164 pageno = (phys_addr - mem->phys_base) >> page_shift; in pci_epc_mem_free_addr()
|
/Linux-v4.19/drivers/infiniband/hw/hns/ |
D | hns_roce_alloc.c | 171 dma_free_coherent(dev, 1 << buf->page_shift, in hns_roce_buf_free() 180 struct hns_roce_buf *buf, u32 page_shift) in hns_roce_buf_alloc() argument 185 u32 page_size = 1 << page_shift; in hns_roce_buf_alloc() 193 if (order <= page_shift - PAGE_SHIFT) in hns_roce_buf_alloc() 196 order -= page_shift - PAGE_SHIFT; in hns_roce_buf_alloc() 198 buf->page_shift = page_shift; in hns_roce_buf_alloc() 207 while (t & ((1 << buf->page_shift) - 1)) { in hns_roce_buf_alloc() 208 --buf->page_shift; in hns_roce_buf_alloc() 214 buf->page_shift = page_shift; in hns_roce_buf_alloc()
|
D | hns_roce_cq.c | 223 u32 page_shift; in hns_roce_ib_get_cq_umem() local 240 page_shift = PAGE_SHIFT + hr_dev->caps.cqe_buf_pg_sz; in hns_roce_ib_get_cq_umem() 241 ret = hns_roce_mtt_init(hr_dev, npages, page_shift, in hns_roce_ib_get_cq_umem() 245 (*umem)->page_shift, in hns_roce_ib_get_cq_umem() 269 u32 page_shift = PAGE_SHIFT + hr_dev->caps.cqe_buf_pg_sz; in hns_roce_ib_alloc_cq_buf() local 272 (1 << page_shift) * 2, &buf->hr_buf, in hns_roce_ib_alloc_cq_buf() 273 page_shift); in hns_roce_ib_alloc_cq_buf() 283 buf->hr_buf.page_shift, &buf->hr_mtt); in hns_roce_ib_alloc_cq_buf()
|
D | hns_roce_mr.c | 208 int hns_roce_mtt_init(struct hns_roce_dev *hr_dev, int npages, int page_shift, in hns_roce_mtt_init() argument 217 mtt->page_shift = HNS_ROCE_HEM_PAGE_SHIFT; in hns_roce_mtt_init() 222 mtt->page_shift = page_shift; in hns_roce_mtt_init() 794 page_list[i] = buf->direct.map + (i << buf->page_shift); in hns_roce_buf_write_mtt() 910 sg_dma_address(sg) + (k << umem->page_shift); in hns_roce_ib_umem_write_mtt() 911 if (!(npage % (1 << (mtt->page_shift - PAGE_SHIFT)))) { in hns_roce_ib_umem_write_mtt() 912 if (page_addr & ((1 << mtt->page_shift) - 1)) { in hns_roce_ib_umem_write_mtt() 914 page_addr, mtt->page_shift); in hns_roce_ib_umem_write_mtt() 959 (k << umem->page_shift); in hns_roce_ib_umem_write_mr()
|
/Linux-v4.19/include/rdma/ |
D | ib_umem.h | 47 int page_shift; member 62 return umem->address & (BIT(umem->page_shift) - 1); in ib_umem_offset() 74 return ALIGN(umem->address + umem->length, BIT(umem->page_shift)); in ib_umem_end() 79 return (ib_umem_end(umem) - ib_umem_start(umem)) >> umem->page_shift; in ib_umem_num_pages()
|
/Linux-v4.19/arch/powerpc/kvm/ |
D | book3s_64_vio.c | 179 if ((tbltmp->it_page_shift <= stt->page_shift) && in kvm_spapr_tce_attach_iommu_group() 181 stt->offset << stt->page_shift) && in kvm_spapr_tce_attach_iommu_group() 183 stt->size << stt->page_shift)) { in kvm_spapr_tce_attach_iommu_group() 302 if (!args->size || args->page_shift < 12 || args->page_shift > 34 || in kvm_vm_ioctl_create_spapr_tce() 303 (args->offset + args->size > (ULLONG_MAX >> args->page_shift))) in kvm_vm_ioctl_create_spapr_tce() 318 stt->page_shift = args->page_shift; in kvm_vm_ioctl_create_spapr_tce() 421 unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); in kvmppc_tce_iommu_unmap() 477 unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); in kvmppc_tce_iommu_map() 526 entry = ioba >> stt->page_shift; in kvmppc_h_put_tce() 570 entry = ioba >> stt->page_shift; in kvmppc_h_put_tce_indirect() [all …]
|
D | book3s_64_vio_hv.c | 110 if (iommu_tce_check_gpa(stt->page_shift, gpa)) in kvmppc_tce_validate() 272 unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); in kvmppc_rm_tce_iommu_unmap() 332 unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); in kvmppc_rm_tce_iommu_map() 380 entry = ioba >> stt->page_shift; in kvmppc_rm_h_put_tce() 460 entry = ioba >> stt->page_shift; in kvmppc_rm_h_put_tce_indirect() 584 unsigned long entry = ioba >> stt->page_shift; in kvmppc_rm_h_stuff_tce() 601 for (i = 0; i < npages; ++i, ioba += (1ULL << stt->page_shift)) in kvmppc_rm_h_stuff_tce() 602 kvmppc_tce_put(stt, ioba >> stt->page_shift, tce_value); in kvmppc_rm_h_stuff_tce() 625 idx = (ioba >> stt->page_shift) - stt->offset; in kvmppc_h_get_tce()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/ |
D | mr.c | 194 int mlx4_mtt_init(struct mlx4_dev *dev, int npages, int page_shift, in mlx4_mtt_init() argument 201 mtt->page_shift = MLX4_ICM_PAGE_SHIFT; in mlx4_mtt_init() 204 mtt->page_shift = page_shift; in mlx4_mtt_init() 418 int page_shift, struct mlx4_mr *mr) in mlx4_mr_alloc_reserved() argument 427 return mlx4_mtt_init(dev, npages, page_shift, &mr->mtt); in mlx4_mr_alloc_reserved() 527 int npages, int page_shift, struct mlx4_mr *mr) in mlx4_mr_alloc() argument 537 access, npages, page_shift, mr); in mlx4_mr_alloc() 590 int page_shift, struct mlx4_mpt_entry *mpt_entry) in mlx4_mr_rereg_mem_write() argument 594 err = mlx4_mtt_init(dev, npages, page_shift, &mr->mtt); in mlx4_mr_rereg_mem_write() 600 mpt_entry->entity_size = cpu_to_be32(page_shift); in mlx4_mr_rereg_mem_write() [all …]
|
/Linux-v4.19/drivers/infiniband/core/ |
D | umem_odp.c | 290 umem->page_shift = PAGE_SHIFT; in ib_alloc_odp_umem() 362 umem->page_shift = huge_page_shift(h); in ib_umem_odp_get() 581 0, BIT(umem->page_shift), in ib_umem_odp_map_dma_single_page() 609 ib_umem_start(umem) + (page_index >> umem->page_shift), in ib_umem_odp_map_dma_single_page() 611 umem->page_shift), in ib_umem_odp_map_dma_single_page() 651 int j, k, ret = 0, start_idx, npages = 0, page_shift; in ib_umem_odp_map_dma_pages() local 666 page_shift = umem->page_shift; in ib_umem_odp_map_dma_pages() 667 page_mask = ~(BIT(page_shift) - 1); in ib_umem_odp_map_dma_pages() 687 start_idx = (user_virt - ib_umem_start(umem)) >> page_shift; in ib_umem_odp_map_dma_pages() 692 (bcnt + BIT(page_shift) - 1) >> page_shift, in ib_umem_odp_map_dma_pages() [all …]
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ |
D | alloc.c | 80 buf->page_shift = (u8)get_order(size) + PAGE_SHIFT; in mlx5_buf_alloc_node() 93 while (t & ((1 << buf->page_shift) - 1)) { in mlx5_buf_alloc_node() 94 --buf->page_shift; in mlx5_buf_alloc_node() 127 buf->page_shift = PAGE_SHIFT; in mlx5_frag_buf_alloc_node() 141 if (frag->map & ((1 << buf->page_shift) - 1)) { in mlx5_frag_buf_alloc_node() 145 &frag->map, buf->page_shift); in mlx5_frag_buf_alloc_node() 297 addr = buf->frags->map + (i << buf->page_shift); in mlx5_fill_page_array()
|
/Linux-v4.19/tools/testing/selftests/kvm/lib/ |
D | kvm_util.c | 135 vm->page_shift = 12; in vm_create() 140 0, (1ULL << (48 - 1)) >> vm->page_shift); in vm_create() 142 (~((1ULL << (48 - 1)) - 1)) >> vm->page_shift, in vm_create() 143 (1ULL << (48 - 1)) >> vm->page_shift); in vm_create() 146 vm->max_gfn = ((1ULL << 52) >> vm->page_shift) - 1; in vm_create() 428 if ((ptr1 >> vm->page_shift) != ((ptr1 + amt) >> vm->page_shift)) in kvm_memcmp_hva_gva() 430 if ((ptr2 >> vm->page_shift) != ((ptr2 + amt) >> vm->page_shift)) in kvm_memcmp_hva_gva() 433 assert((ptr1 >> vm->page_shift) == ((ptr1 + amt - 1) >> vm->page_shift)); in kvm_memcmp_hva_gva() 434 assert((ptr2 >> vm->page_shift) == ((ptr2 + amt - 1) >> vm->page_shift)); in kvm_memcmp_hva_gva() 582 TEST_ASSERT((((guest_paddr >> vm->page_shift) + npages) - 1) in vm_userspace_mem_region_add() [all …]
|
/Linux-v4.19/drivers/infiniband/sw/rdmavt/ |
D | mr.c | 411 mr->mr.page_shift = umem->page_shift; in rvt_reg_user_mr() 423 mr->mr.map[m]->segs[n].length = BIT(umem->page_shift); in rvt_reg_user_mr() 425 BIT(umem->page_shift)); in rvt_reg_user_mr() 607 u32 ps = 1 << mr->mr.page_shift; in rvt_set_page() 608 u32 mapped_segs = mr->mr.length >> mr->mr.page_shift; in rvt_set_page() 644 mr->mr.page_shift = PAGE_SHIFT; in rvt_map_mr_sg() 757 fmr->mr.page_shift = fmr_attr->page_shift; in rvt_alloc_fmr() 805 ps = 1 << fmr->mr.page_shift; in rvt_map_phys_fmr() 969 if (mr->page_shift) { in rvt_lkey_ok() 977 entries_spanned_by_off = off >> mr->page_shift; in rvt_lkey_ok() [all …]
|
/Linux-v4.19/arch/powerpc/platforms/powernv/ |
D | pci-ioda-tce.c | 22 u64 dma_offset, unsigned int page_shift) in pnv_pci_setup_iommu_table() argument 26 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table() 252 __u32 page_shift, __u64 window_size, __u32 levels, in pnv_pci_ioda2_table_alloc_pages() argument 259 unsigned int entries_shift = window_shift - page_shift; in pnv_pci_ioda2_table_alloc_pages() 279 if ((level_shift - 3) * levels + page_shift >= 55) in pnv_pci_ioda2_table_alloc_pages() 313 page_shift); in pnv_pci_ioda2_table_alloc_pages()
|
/Linux-v4.19/drivers/infiniband/sw/rxe/ |
D | rxe_mr.c | 194 mem->page_shift = umem->page_shift; in rxe_mem_init_user() 195 mem->page_mask = BIT(umem->page_shift) - 1; in rxe_mem_init_user() 211 buf->size = BIT(umem->page_shift); in rxe_mem_init_user() 276 if (likely(mem->page_shift)) { in lookup_iova() 278 offset >>= mem->page_shift; in lookup_iova() 621 page_size = 1 << mem->page_shift; in rxe_mem_map_pages() 640 mem->length = num_pages << mem->page_shift; in rxe_mem_map_pages()
|
/Linux-v4.19/drivers/mtd/nand/raw/ |
D | nand_bbt.c | 185 from = ((loff_t)page) << this->page_shift; in read_bbt() 395 scan_read(mtd, buf, (loff_t)td->pages[0] << this->page_shift, in read_abs_bbts() 404 scan_read(mtd, buf, (loff_t)md->pages[0] << this->page_shift, in read_abs_bbts() 531 int blocktopage = this->bbt_erase_shift - this->page_shift; in search_bbt() 632 (this->bbt_erase_shift - this->page_shift); in get_bbt_block() 660 page = block << (this->bbt_erase_shift - this->page_shift); in get_bbt_block() 763 page = block << (this->bbt_erase_shift - this->page_shift); in write_bbt() 784 to = ((loff_t)page) << this->page_shift; in write_bbt() 800 ops.ooblen = (len >> this->page_shift) * mtd->oobsize; in write_bbt() 807 pageoffs = page - (int)(to >> this->page_shift); in write_bbt() [all …]
|
/Linux-v4.19/arch/powerpc/include/asm/ |
D | iommu.h | 175 __u32 page_shift, 180 __u32 page_shift, 305 extern int iommu_tce_check_ioba(unsigned long page_shift, 308 extern int iommu_tce_check_gpa(unsigned long page_shift,
|
/Linux-v4.19/arch/powerpc/platforms/pseries/ |
D | iommu.c | 925 struct ddw_create_response *create, int page_shift, in create_ddw() argument 949 page_shift, window_shift); in create_ddw() 954 cfg_addr, BUID_HI(buid), BUID_LO(buid), page_shift, in create_ddw() 983 int page_shift; in enable_ddw() local 1044 page_shift = 24; /* 16MB */ in enable_ddw() 1046 page_shift = 16; /* 64kB */ in enable_ddw() 1048 page_shift = 12; /* 4kB */ in enable_ddw() 1057 if (query.largest_available_block < (max_addr >> page_shift)) { in enable_ddw() 1060 1ULL << page_shift); in enable_ddw() 1079 ret = create_ddw(dev, ddw_avail, &create, page_shift, len); in enable_ddw() [all …]
|
/Linux-v4.19/drivers/vfio/ |
D | vfio_iommu_spapr_tce.c | 214 static bool tce_page_is_contained(struct page *page, unsigned page_shift) in tce_page_is_contained() argument 221 return (PAGE_SHIFT + compound_order(compound_head(page))) >= page_shift; in tce_page_is_contained() 614 __u32 page_shift, in tce_iommu_create_table() argument 621 table_size = table_group->ops->get_table_size(page_shift, window_size, in tce_iommu_create_table() 631 page_shift, window_size, levels, ptbl); in tce_iommu_create_table() 649 __u32 page_shift, __u64 window_size, __u32 levels, in tce_iommu_create_window() argument 668 if (!(table_group->pgsizes & (1ULL << page_shift))) in tce_iommu_create_window() 678 page_shift, window_size, levels, &tbl); in tce_iommu_create_window() 1092 create.page_shift, in tce_iommu_ioctl()
|
/Linux-v4.19/kernel/events/ |
D | ring_buffer.c | 124 int have_lost, page_shift; in __perf_output_begin() local 201 page_shift = PAGE_SHIFT + page_order(rb); in __perf_output_begin() 203 handle->page = (offset >> page_shift) & (rb->nr_pages - 1); in __perf_output_begin() 204 offset &= (1UL << page_shift) - 1; in __perf_output_begin() 206 handle->size = (1UL << page_shift) - offset; in __perf_output_begin()
|
/Linux-v4.19/arch/powerpc/mm/ |
D | tlb-radix.c | 698 unsigned int page_shift = mmu_psize_defs[mmu_virtual_psize].shift; in __radix__flush_tlb_range() local 699 unsigned long page_size = 1UL << page_shift; in __radix__flush_tlb_range() 700 unsigned long nr_pages = (end - start) >> page_shift; in __radix__flush_tlb_range() 926 unsigned int page_shift = mmu_psize_defs[psize].shift; in __radix__flush_tlb_range_psize() local 927 unsigned long page_size = 1UL << page_shift; in __radix__flush_tlb_range_psize() 928 unsigned long nr_pages = (end - start) >> page_shift; in __radix__flush_tlb_range_psize()
|