Home
last modified time | relevance | path

Searched refs:page_offset (Results 1 – 25 of 218) sorted by relevance

123456789

/Linux-v4.19/drivers/net/ethernet/cavium/liquidio/
Docteon_network.h301 pg_info->page_offset = 0; in recv_buffer_alloc()
303 skb_pg_info->page_offset = 0; in recv_buffer_alloc()
327 skb_pg_info->page_offset = 0; in recv_buffer_fast_alloc()
351 pg_info->page_offset = 0; in recv_buffer_recycle()
356 if (pg_info->page_offset == 0) in recv_buffer_recycle()
357 pg_info->page_offset = LIO_RXBUFFER_SZ; in recv_buffer_recycle()
359 pg_info->page_offset = 0; in recv_buffer_recycle()
387 skb_pg_info->page_offset = pg_info->page_offset; in recv_buffer_reuse()
401 pg_info->page_offset = 0; in recv_buffer_destroy()
418 pg_info->page_offset = 0; in recv_buffer_free()
[all …]
/Linux-v4.19/drivers/scsi/fnic/
Dfnic_trace.c71 fnic_trace_entries.page_offset[fnic_trace_entries.wr_idx]; in fnic_trace_get_buf()
124 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data()
166 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data()
480 fnic_trace_entries.page_offset = in fnic_trace_buf_init()
483 if (!fnic_trace_entries.page_offset) { in fnic_trace_buf_init()
493 memset((void *)fnic_trace_entries.page_offset, 0, in fnic_trace_buf_init()
504 fnic_trace_entries.page_offset[i] = fnic_buf_head; in fnic_trace_buf_init()
527 if (fnic_trace_entries.page_offset) { in fnic_trace_free()
528 vfree((void *)fnic_trace_entries.page_offset); in fnic_trace_free()
529 fnic_trace_entries.page_offset = NULL; in fnic_trace_free()
[all …]
/Linux-v4.19/drivers/gpu/drm/vkms/
Dvkms_gem.c51 pgoff_t page_offset; in vkms_gem_fault() local
55 page_offset = (vaddr - vma->vm_start) >> PAGE_SHIFT; in vkms_gem_fault()
58 if (page_offset > num_pages) in vkms_gem_fault()
64 get_page(obj->pages[page_offset]); in vkms_gem_fault()
65 vmf->page = obj->pages[page_offset]; in vkms_gem_fault()
74 page = shmem_read_mapping_page(mapping, page_offset); in vkms_gem_fault()
/Linux-v4.19/drivers/gpu/drm/ttm/
Dttm_bo_vm.c98 unsigned long page_offset) in ttm_bo_io_mem_pfn() argument
103 return bdev->driver->io_mem_pfn(bo, page_offset); in ttm_bo_io_mem_pfn()
106 + page_offset; in ttm_bo_io_mem_pfn()
115 unsigned long page_offset; in ttm_bo_vm_fault() local
208 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + in ttm_bo_vm_fault()
213 if (unlikely(page_offset >= bo->num_pages)) { in ttm_bo_vm_fault()
256 pfn = ttm_bo_io_mem_pfn(bo, page_offset); in ttm_bo_vm_fault()
258 page = ttm->pages[page_offset]; in ttm_bo_vm_fault()
266 page_offset; in ttm_bo_vm_fault()
287 if (unlikely(++page_offset >= page_last)) in ttm_bo_vm_fault()
/Linux-v4.19/scripts/
Dleaking_addresses.pl305 state $page_offset = get_page_offset();
311 if (hex($match) < $page_offset) {
321 my $page_offset;
329 $page_offset = get_kernel_config_option('CONFIG_PAGE_OFFSET');
330 if (!$page_offset) {
333 return $page_offset;
/Linux-v4.19/fs/hfs/
Dbnode.c23 off += node->page_offset; in hfs_bnode_read()
65 off += node->page_offset; in hfs_bnode_write()
90 off += node->page_offset; in hfs_bnode_clear()
106 src += src_node->page_offset; in hfs_bnode_copy()
107 dst += dst_node->page_offset; in hfs_bnode_copy()
125 src += node->page_offset; in hfs_bnode_move()
126 dst += node->page_offset; in hfs_bnode_move()
279 node->page_offset = off & ~PAGE_MASK; in __hfs_bnode_create()
338 desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset); in hfs_bnode_find()
428 memset(kmap(*pagep) + node->page_offset, 0, in hfs_bnode_create()
/Linux-v4.19/drivers/mtd/tests/
Dnandbiterrs.c58 static unsigned page_offset; variable
59 module_param(page_offset, uint, S_IRUGO);
60 MODULE_PARM_DESC(page_offset, "Page number relative to dev start");
371 offset = (loff_t)page_offset * mtd->writesize; in mtd_nandbiterrs_init()
375 page_offset, offset, eraseblock); in mtd_nandbiterrs_init()
/Linux-v4.19/drivers/gpu/drm/gma500/
Dgem.c145 pgoff_t page_offset; in psb_gem_fault() local
173 page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in psb_gem_fault()
179 pfn = page_to_pfn(r->pages[page_offset]); in psb_gem_fault()
/Linux-v4.19/drivers/nvmem/
Drave-sp-eeprom.c161 const unsigned int page_offset = offset % RAVE_SP_EEPROM_PAGE_SIZE; in rave_sp_eeprom_page_access() local
172 if (WARN_ON(data_len > sizeof(page.data) - page_offset)) in rave_sp_eeprom_page_access()
187 memcpy(&page.data[page_offset], data, data_len); in rave_sp_eeprom_page_access()
199 memcpy(data, &page.data[page_offset], data_len); in rave_sp_eeprom_page_access()
/Linux-v4.19/drivers/gpu/drm/qxl/
Dqxl_image.c165 unsigned page_base, page_offset, out_offset; in qxl_image_init_helper() local
173 page_offset = offset_in_page(out_offset); in qxl_image_init_helper()
174 size = min((int)(PAGE_SIZE - page_offset), remain); in qxl_image_init_helper()
177 k_data = ptr + page_offset; in qxl_image_init_helper()
Dqxl_object.c145 struct qxl_bo *bo, int page_offset) in qxl_bo_kmap_atomic_page() argument
163 return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset); in qxl_bo_kmap_atomic_page()
166 rptr = bo->kptr + (page_offset * PAGE_SIZE); in qxl_bo_kmap_atomic_page()
174 rptr += page_offset * PAGE_SIZE; in qxl_bo_kmap_atomic_page()
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/
Dsrq.c71 u32 page_offset = in->page_offset; in get_pas_size() local
75 u32 rq_sz_po = rq_sz + (page_offset * po_quanta); in get_pas_size()
88 MLX5_SET(wq, wq, page_offset, in->page_offset); in set_wq()
101 MLX5_SET(srqc, srqc, page_offset, in->page_offset); in set_srqc()
116 in->page_offset = MLX5_GET(wq, wq, page_offset); in get_wq()
129 in->page_offset = MLX5_GET(srqc, srqc, page_offset); in get_srqc()
/Linux-v4.19/fs/hfsplus/
Dbnode.c27 off += node->page_offset; in hfs_bnode_read()
80 off += node->page_offset; in hfs_bnode_write()
110 off += node->page_offset; in hfs_bnode_clear()
136 src += src_node->page_offset; in hfs_bnode_copy()
137 dst += dst_node->page_offset; in hfs_bnode_copy()
193 src += node->page_offset; in hfs_bnode_move()
194 dst += node->page_offset; in hfs_bnode_move()
445 node->page_offset = off & ~PAGE_MASK; in __hfs_bnode_create()
506 node->page_offset); in hfs_bnode_find()
596 memset(kmap(*pagep) + node->page_offset, 0, in hfs_bnode_create()
Dwrapper.c76 unsigned int page_offset = offset_in_page(buf); in hfsplus_submit_bio() local
77 unsigned int len = min_t(unsigned int, PAGE_SIZE - page_offset, in hfsplus_submit_bio()
80 ret = bio_add_page(bio, virt_to_page(buf), len, page_offset); in hfsplus_submit_bio()
/Linux-v4.19/drivers/net/ethernet/intel/i40evf/
Di40e_txrx.c682 rx_bi->page_offset, in i40evf_clean_rx_ring()
695 rx_bi->page_offset = 0; in i40evf_clean_rx_ring()
847 bi->page_offset = i40e_rx_offset(rx_ring); in i40e_alloc_mapped_page()
899 bi->page_offset, in i40evf_alloc_rx_buffers()
906 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in i40evf_alloc_rx_buffers()
1141 new_buff->page_offset = old_buff->page_offset; in i40e_reuse_rx_page()
1201 if (rx_buffer->page_offset > I40E_LAST_OFFSET) in i40e_can_reuse_rx_page()
1241 rx_buffer->page_offset, size, truesize); in i40e_add_rx_frag()
1245 rx_buffer->page_offset ^= truesize; in i40e_add_rx_frag()
1247 rx_buffer->page_offset += truesize; in i40e_add_rx_frag()
[all …]
/Linux-v4.19/net/ceph/
Dmessenger.c537 int page_offset, size_t length) in ceph_tcp_recvpage() argument
541 .bv_offset = page_offset, in ceph_tcp_recvpage()
547 BUG_ON(page_offset + length > PAGE_SIZE); in ceph_tcp_recvpage()
852 size_t *page_offset, in ceph_msg_data_bio_next() argument
858 *page_offset = bv.bv_offset; in ceph_msg_data_bio_next()
911 size_t *page_offset, in ceph_msg_data_bvecs_next() argument
917 *page_offset = bv.bv_offset; in ceph_msg_data_bvecs_next()
964 cursor->page_offset = data->alignment & ~PAGE_MASK; in ceph_msg_data_pages_cursor_init()
968 BUG_ON(length > SIZE_MAX - cursor->page_offset); in ceph_msg_data_pages_cursor_init()
969 cursor->last_piece = cursor->page_offset + cursor->resid <= PAGE_SIZE; in ceph_msg_data_pages_cursor_init()
[all …]
/Linux-v4.19/drivers/gpu/drm/vgem/
Dvgem_drv.c72 pgoff_t page_offset; in vgem_gem_fault() local
73 page_offset = (vaddr - vma->vm_start) >> PAGE_SHIFT; in vgem_gem_fault()
77 if (page_offset >= num_pages) in vgem_gem_fault()
82 get_page(obj->pages[page_offset]); in vgem_gem_fault()
83 vmf->page = obj->pages[page_offset]; in vgem_gem_fault()
92 page_offset); in vgem_gem_fault()
/Linux-v4.19/drivers/net/ethernet/sfc/falcon/
Drx.c62 return page_address(buf->page) + buf->page_offset; in ef4_rx_buf_va()
157 unsigned int page_offset; in ef4_init_rx_buffers() local
188 page_offset = sizeof(struct ef4_rx_page_state); in ef4_init_rx_buffers()
195 rx_buf->page_offset = page_offset + efx->rx_ip_align; in ef4_init_rx_buffers()
201 page_offset += efx->rx_page_buf_step; in ef4_init_rx_buffers()
202 } while (page_offset + efx->rx_page_buf_step <= PAGE_SIZE); in ef4_init_rx_buffers()
451 rx_buf->page, rx_buf->page_offset, in ef4_rx_packet_gro()
498 rx_buf->page_offset += hdr_len; in ef4_rx_mk_skb()
503 rx_buf->page, rx_buf->page_offset, in ef4_rx_mk_skb()
585 rx_buf->page_offset += efx->rx_prefix_size; in ef4_rx_packet()
/Linux-v4.19/drivers/net/ethernet/sfc/
Drx.c62 return page_address(buf->page) + buf->page_offset; in efx_rx_buf_va()
157 unsigned int page_offset; in efx_init_rx_buffers() local
188 page_offset = sizeof(struct efx_rx_page_state); in efx_init_rx_buffers()
195 rx_buf->page_offset = page_offset + efx->rx_ip_align; in efx_init_rx_buffers()
201 page_offset += efx->rx_page_buf_step; in efx_init_rx_buffers()
202 } while (page_offset + efx->rx_page_buf_step <= PAGE_SIZE); in efx_init_rx_buffers()
441 rx_buf->page, rx_buf->page_offset, in efx_rx_packet_gro()
488 rx_buf->page_offset += hdr_len; in efx_rx_mk_skb()
493 rx_buf->page, rx_buf->page_offset, in efx_rx_mk_skb()
575 rx_buf->page_offset += efx->rx_prefix_size; in efx_rx_packet()
/Linux-v4.19/drivers/gpu/drm/udl/
Dudl_gem.c108 unsigned int page_offset; in udl_gem_fault() local
110 page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in udl_gem_fault()
115 page = obj->pages[page_offset]; in udl_gem_fault()
/Linux-v4.19/tools/testing/selftests/powerpc/primitives/
Dload_unaligned_zeropad.c114 static int do_one_test(char *p, int page_offset) in do_one_test() argument
126 …printf("offset %u load_unaligned_zeropad returned 0x%lx, should be 0x%lx\n", page_offset, got, sho… in do_one_test()
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c70 frag->page_offset = priv->rx_headroom; in mlx4_alloc_page()
89 frags->page_offset); in mlx4_en_alloc_frags()
148 frags->page_offset = XDP_PACKET_HEADROOM; in mlx4_en_prepare_rx_desc()
491 dma_sync_single_range_for_cpu(priv->ddev, dma, frags->page_offset, in mlx4_en_complete_rx_desc()
494 __skb_fill_page_desc(skb, nr, page, frags->page_offset, in mlx4_en_complete_rx_desc()
499 frags->page_offset ^= PAGE_SIZE / 2; in mlx4_en_complete_rx_desc()
510 frags->page_offset += sz_align; in mlx4_en_complete_rx_desc()
511 release = frags->page_offset + frag_info->frag_size > PAGE_SIZE; in mlx4_en_complete_rx_desc()
696 va = page_address(frags[0].page) + frags[0].page_offset; in mlx4_en_process_rx_cq()
725 dma = frags[0].dma + frags[0].page_offset; in mlx4_en_process_rx_cq()
[all …]
/Linux-v4.19/net/core/
Dtso.c59 tso->data = page_address(frag->page.p) + frag->page_offset; in tso_build_data()
83 tso->data = page_address(frag->page.p) + frag->page_offset; in tso_start()
/Linux-v4.19/drivers/net/ethernet/qlogic/qede/
Dqede_fp.c82 sw_rx_data->page_offset = 0; in qede_alloc_rx_buffer()
511 new_mapping = curr_prod->mapping + curr_prod->page_offset; in qede_reuse_page()
539 curr_cons->page_offset += rxq->rx_buf_seg_size; in qede_realloc_rx_buffer()
541 if (curr_cons->page_offset == PAGE_SIZE) { in qede_realloc_rx_buffer()
546 curr_cons->page_offset -= rxq->rx_buf_seg_size; in qede_realloc_rx_buffer()
664 current_bd->page_offset + rxq->rx_headroom, in qede_fill_frag_skb()
731 buf = page_address(bd->data) + bd->page_offset; in qede_build_skb()
749 bd->page_offset += rxq->rx_buf_seg_size; in qede_tpa_rx_build_skb()
751 if (bd->page_offset == PAGE_SIZE) { in qede_tpa_rx_build_skb()
755 bd->page_offset -= rxq->rx_buf_seg_size; in qede_tpa_rx_build_skb()
[all …]
/Linux-v4.19/drivers/net/ethernet/intel/ice/
Dice_txrx.c291 rx_buf->page_offset = 0; in ice_clean_rx_ring()
428 bi->page_offset = 0; in ice_alloc_mapped_page()
461 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in ice_alloc_rx_bufs()
543 unsigned char *va = page_address(page) + rx_buf->page_offset; in ice_add_rx_frag()
557 rx_buf->page_offset, size, truesize); in ice_add_rx_frag()
569 rx_buf->page_offset ^= truesize; in ice_add_rx_frag()
572 rx_buf->page_offset += truesize; in ice_add_rx_frag()
574 if (rx_buf->page_offset > last_offset) in ice_add_rx_frag()
633 u8 *page_addr = page_address(page) + rx_buf->page_offset; in ice_fetch_rx_buf()
660 rx_buf->page_offset, in ice_fetch_rx_buf()
[all …]

123456789