Home
last modified time | relevance | path

Searched refs:XEN_PAGE_SIZE (Results 1 – 25 of 32) sorted by relevance

12

/Linux-v5.4/include/xen/
Dpage.h9 #define XEN_PAGE_SIZE (_AC(1, UL) << XEN_PAGE_SHIFT) macro
10 #define XEN_PAGE_MASK (~(XEN_PAGE_SIZE-1))
23 #define XEN_PFN_PER_PAGE (PAGE_SIZE / XEN_PAGE_SIZE)
26 #define XEN_PFN_UP(x) (((x) + XEN_PAGE_SIZE-1) >> XEN_PAGE_SHIFT)
Dgrant_table.h285 len = min_t(unsigned int, XEN_PAGE_SIZE - (offset & ~XEN_PAGE_MASK), in gnttab_for_one_grant()
/Linux-v5.4/drivers/xen/
Dmem-reservation.c45 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_update()
76 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_reset()
Dbiomerge.c11 #if XEN_PAGE_SIZE == PAGE_SIZE in xen_biovec_phys_mergeable()
Darm-device.c24 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_unmap_device_mmio()
58 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_map_device_mmio()
Dgrant-table.c741 vaddr = xen_remap(addr, XEN_PAGE_SIZE * max_nr_gframes); in gnttab_setup_auto_xlat_frames()
1000 glen = min_t(unsigned int, XEN_PAGE_SIZE - goffset, len); in gnttab_foreach_grant_in_range()
1025 fn(pfn_to_gfn(xen_pfn), goffset, XEN_PAGE_SIZE, data); in gnttab_foreach_grant()
1027 goffset += XEN_PAGE_SIZE; in gnttab_foreach_grant()
1293 .grefs_per_grant_frame = XEN_PAGE_SIZE /
1305 .grefs_per_grant_frame = XEN_PAGE_SIZE /
Dballoon.c317 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in reserve_additional_memory()
615 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in alloc_xenballooned_pages()
Dgntdev.c913 if (seg->source.foreign.offset + seg->len > XEN_PAGE_SIZE) in gntdev_grant_copy_seg()
917 if (seg->dest.foreign.offset + seg->len > XEN_PAGE_SIZE) in gntdev_grant_copy_seg()
950 len = min(len, (size_t)XEN_PAGE_SIZE - off); in gntdev_grant_copy_seg()
969 len = min(len, (size_t)XEN_PAGE_SIZE - off); in gntdev_grant_copy_seg()
/Linux-v5.4/drivers/net/xen-netback/
Dcommon.h69 #define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, XEN_PAGE_SIZE)
70 #define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, XEN_PAGE_SIZE)
85 #define MAX_BUFFER_OFFSET XEN_PAGE_SIZE
92 #define MAX_XEN_SKB_FRAGS (65536 / XEN_PAGE_SIZE + 1)
Drx.c46 needed = DIV_ROUND_UP(skb->len, XEN_PAGE_SIZE); in xenvif_rx_ring_slots_available()
340 chunk_len = min_t(size_t, frag_len, XEN_PAGE_SIZE - offset); in xenvif_rx_next_chunk()
341 chunk_len = min_t(size_t, chunk_len, XEN_PAGE_SIZE - in xenvif_rx_next_chunk()
372 } while (offset < XEN_PAGE_SIZE && pkt->remaining_len > 0); in xenvif_rx_data_slot()
Dhash.c347 len > XEN_PAGE_SIZE / sizeof(*mapping)) in xenvif_set_hash_mapping()
352 if (copy_op[0].dest.offset + copy_op[0].len > XEN_PAGE_SIZE) { in xenvif_set_hash_mapping()
354 copy_op[1].source.offset = XEN_PAGE_SIZE - copy_op[0].dest.offset; in xenvif_set_hash_mapping()
Dnetback.c298 if (unlikely((txp->offset + txp->size) > XEN_PAGE_SIZE)) { in xenvif_count_requests()
890 if (unlikely((txreq.offset + txreq.size) > XEN_PAGE_SIZE)) { in xenvif_tx_build_gops()
1465 BACK_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE); in xenvif_map_frontend_data_rings()
1473 BACK_RING_INIT(&queue->rx, rxs, XEN_PAGE_SIZE); in xenvif_map_frontend_data_rings()
/Linux-v5.4/arch/arm/xen/
Dmm.c53 if (size + cflush.offset > XEN_PAGE_SIZE) in dma_cache_maint()
54 cflush.length = XEN_PAGE_SIZE - cflush.offset; in dma_cache_maint()
/Linux-v5.4/include/xen/interface/io/
Ddisplif.h853 #define XENDISPL_EVENT_PAGE_SIZE XEN_PAGE_SIZE
Dsndif.h1072 #define XENSND_EVENT_PAGE_SIZE XEN_PAGE_SIZE
/Linux-v5.4/sound/xen/
Dxen_snd_front.c376 if (XEN_PAGE_SIZE != PAGE_SIZE) { in xen_drv_init()
378 XEN_PAGE_SIZE, PAGE_SIZE); in xen_drv_init()
/Linux-v5.4/drivers/xen/xenbus/
Dxenbus_client.c375 vaddr = vaddr + XEN_PAGE_SIZE; in xenbus_grant_ring()
691 area = alloc_vm_area(XEN_PAGE_SIZE * nr_grefs, ptes); in xenbus_map_ring_valloc_pv()
757 addr = (unsigned long)vaddr + (XEN_PAGE_SIZE * i); in xenbus_unmap_ring_vfree_pv()
/Linux-v5.4/drivers/block/xen-blkback/
Dcommon.h60 #define XEN_PAGES_PER_SEGMENT (PAGE_SIZE / XEN_PAGE_SIZE)
63 (XEN_PAGE_SIZE/sizeof(struct blkif_request_segment))
Dxenbus.c200 XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
208 XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
216 XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
/Linux-v5.4/drivers/block/
Dxen-blkfront.c152 __CONST_RING_SIZE(blkif, XEN_PAGE_SIZE * (info)->nr_ring_pages)
155 __CONST_RING_SIZE(blkif, XEN_PAGE_SIZE * XENBUS_MAX_RING_GRANTS)
253 #define GRANTS_PER_PSEG (PAGE_SIZE / XEN_PAGE_SIZE)
256 (XEN_PAGE_SIZE / sizeof(struct blkif_request_segment))
685 (BLKIF_MAX_SEGMENTS_PER_REQUEST * XEN_PAGE_SIZE) / 512; in blkif_setup_extra_req()
947 blk_queue_max_hw_sectors(rq, (segments * XEN_PAGE_SIZE) / 512); in blkif_set_queue_limits()
1334 free_pages((unsigned long)rinfo->ring.sring, get_order(info->nr_ring_pages * XEN_PAGE_SIZE)); in blkif_free_ring()
1675 unsigned long ring_size = info->nr_ring_pages * XEN_PAGE_SIZE; in setup_blkring()
/Linux-v5.4/drivers/tty/hvc/
Dhvc_xen.c221 info->intf = xen_remap(gfn << XEN_PAGE_SHIFT, XEN_PAGE_SIZE); in xen_hvm_console_init()
470 memset(info->intf, 0, XEN_PAGE_SIZE); in xencons_resume()
/Linux-v5.4/drivers/net/
Dxen-netfront.c78 #define NET_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, XEN_PAGE_SIZE)
79 #define NET_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, XEN_PAGE_SIZE)
563 #define MAX_XEN_SKB_FRAGS (65536 / XEN_PAGE_SIZE + 1)
803 rx->offset + rx->status > XEN_PAGE_SIZE)) { in xennet_get_responses()
1552 FRONT_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE); in setup_netfront()
1566 FRONT_RING_INIT(&queue->rx, rxs, XEN_PAGE_SIZE); in setup_netfront()
/Linux-v5.4/drivers/gpu/drm/xen/
Dxen_drm_front.c792 if (XEN_PAGE_SIZE != PAGE_SIZE) { in xen_drv_init()
794 XEN_PAGE_SIZE, PAGE_SIZE); in xen_drv_init()
Dxen_drm_front_evtchnl.c186 FRONT_RING_INIT(&evtchnl->u.req.ring, sring, XEN_PAGE_SIZE); in evtchnl_alloc()
/Linux-v5.4/drivers/xen/events/
Devents_fifo.c57 #define EVENT_WORDS_PER_PAGE (XEN_PAGE_SIZE / sizeof(event_word_t))

12