Home
last modified time | relevance | path

Searched refs:chunk_mask (Results 1 – 16 of 16) sorted by relevance

/Linux-v5.4/net/xdp/
Dxsk_queue.c12 void xskq_set_umem(struct xsk_queue *q, u64 size, u64 chunk_mask) in xskq_set_umem() argument
18 q->chunk_mask = chunk_mask; in xskq_set_umem()
Dxsk_queue.h35 u64 chunk_mask; member
181 *addr = READ_ONCE(ring->desc[idx]) & q->chunk_mask; in xskq_validate_addr()
290 if (((d->addr + d->len) & q->chunk_mask) != (d->addr & q->chunk_mask) || in xskq_is_valid_desc()
374 void xskq_set_umem(struct xsk_queue *q, u64 size, u64 chunk_mask);
Dxsk.c686 xs->umem->chunk_mask); in xsk_bind()
688 xs->umem->chunk_mask); in xsk_bind()
700 xskq_set_umem(xs->rx, xs->umem->size, xs->umem->chunk_mask); in xsk_bind()
701 xskq_set_umem(xs->tx, xs->umem->size, xs->umem->chunk_mask); in xsk_bind()
Dxdp_umem.c390 umem->chunk_mask = unaligned_chunks ? XSK_UNALIGNED_BUF_ADDR_MASK in xdp_umem_reg()
/Linux-v5.4/drivers/md/
Ddm-exception-store.c153 store->chunk_size = store->chunk_mask = store->chunk_shift = 0; in set_chunk_size()
185 store->chunk_mask = chunk_size - 1; in dm_exception_store_set_chunk_size()
Draid10.h56 sector_t chunk_mask; member
Ddm-exception-store.h122 unsigned chunk_mask; member
Draid10.c579 sector = r10bio->sector & geo->chunk_mask; in __raid10_find_phys()
619 sector += (geo->chunk_mask + 1); in __raid10_find_phys()
661 offset = sector & geo->chunk_mask; in raid10_find_virt()
1524 sector_t chunk_mask = (conf->geo.chunk_mask & conf->prev.chunk_mask); in raid10_make_request() local
1525 int chunk_sects = chunk_mask + 1; in raid10_make_request()
1540 if (unlikely((bio->bi_iter.bi_sector & chunk_mask) + in raid10_make_request()
2914 sector_t chunk_mask = conf->geo.chunk_mask; in raid10_sync_request() local
3013 max_sector > (sector_nr | chunk_mask)) in raid10_sync_request()
3014 max_sector = (sector_nr | chunk_mask) + 1; in raid10_sync_request()
3333 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()
[all …]
Ddm-snap-persistent.c326 ps->store->chunk_mask = ps->store->chunk_size - 1; in read_header()
Ddm-snap.c1907 (bio->bi_iter.bi_sector & s->store->chunk_mask); in remap_exception()
/Linux-v5.4/fs/ufs/
Ddir.c117 const unsigned chunk_mask = UFS_SB(sb)->s_uspi->s_dirblksize - 1; in ufs_check_page() local
123 if (limit & chunk_mask) in ufs_check_page()
138 if (((offs + rec_len - 1) ^ offs) & ~chunk_mask) in ufs_check_page()
431 unsigned chunk_mask = ~(UFS_SB(sb)->s_uspi->s_dirblksize - 1); in ufs_readdir() local
456 offset = ufs_validate_entry(sb, kaddr, offset, chunk_mask); in ufs_readdir()
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
Drx.c48 xsk_umem_fq_reuse(rq->umem, handle & rq->umem->chunk_mask); in mlx5e_xsk_recycle_frame()
/Linux-v5.4/include/net/
Dxdp_sock.h48 u64 chunk_mask; member
/Linux-v5.4/drivers/net/ethernet/intel/ixgbe/
Dixgbe_xsk.c235 mask = rx_ring->xsk_umem->chunk_mask; in ixgbe_zca_free()
295 handle &= rx_ring->xsk_umem->chunk_mask; in ixgbe_alloc_buffer_slow_zc()
/Linux-v5.4/drivers/net/ethernet/intel/i40e/
Di40e_xsk.c297 handle &= rx_ring->xsk_umem->chunk_mask; in i40e_alloc_buffer_slow_zc()
456 mask = rx_ring->xsk_umem->chunk_mask; in i40e_zca_free()
/Linux-v5.4/fs/ext2/
Ddir.c272 unsigned chunk_mask = ~(ext2_chunk_size(inode)-1); in ext2_readdir() local
297 offset = ext2_validate_entry(kaddr, offset, chunk_mask); in ext2_readdir()