Searched refs:chunk_mask (Results 1 – 16 of 16) sorted by relevance
| /Linux-v5.4/net/xdp/ |
| D | xsk_queue.c | 12 void xskq_set_umem(struct xsk_queue *q, u64 size, u64 chunk_mask) in xskq_set_umem() argument 18 q->chunk_mask = chunk_mask; in xskq_set_umem()
|
| D | xsk_queue.h | 35 u64 chunk_mask; member 181 *addr = READ_ONCE(ring->desc[idx]) & q->chunk_mask; in xskq_validate_addr() 290 if (((d->addr + d->len) & q->chunk_mask) != (d->addr & q->chunk_mask) || in xskq_is_valid_desc() 374 void xskq_set_umem(struct xsk_queue *q, u64 size, u64 chunk_mask);
|
| D | xsk.c | 686 xs->umem->chunk_mask); in xsk_bind() 688 xs->umem->chunk_mask); in xsk_bind() 700 xskq_set_umem(xs->rx, xs->umem->size, xs->umem->chunk_mask); in xsk_bind() 701 xskq_set_umem(xs->tx, xs->umem->size, xs->umem->chunk_mask); in xsk_bind()
|
| D | xdp_umem.c | 390 umem->chunk_mask = unaligned_chunks ? XSK_UNALIGNED_BUF_ADDR_MASK in xdp_umem_reg()
|
| /Linux-v5.4/drivers/md/ |
| D | dm-exception-store.c | 153 store->chunk_size = store->chunk_mask = store->chunk_shift = 0; in set_chunk_size() 185 store->chunk_mask = chunk_size - 1; in dm_exception_store_set_chunk_size()
|
| D | raid10.h | 56 sector_t chunk_mask; member
|
| D | dm-exception-store.h | 122 unsigned chunk_mask; member
|
| D | raid10.c | 579 sector = r10bio->sector & geo->chunk_mask; in __raid10_find_phys() 619 sector += (geo->chunk_mask + 1); in __raid10_find_phys() 661 offset = sector & geo->chunk_mask; in raid10_find_virt() 1524 sector_t chunk_mask = (conf->geo.chunk_mask & conf->prev.chunk_mask); in raid10_make_request() local 1525 int chunk_sects = chunk_mask + 1; in raid10_make_request() 1540 if (unlikely((bio->bi_iter.bi_sector & chunk_mask) + in raid10_make_request() 2914 sector_t chunk_mask = conf->geo.chunk_mask; in raid10_sync_request() local 3013 max_sector > (sector_nr | chunk_mask)) in raid10_sync_request() 3014 max_sector = (sector_nr | chunk_mask) + 1; in raid10_sync_request() 3333 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request() [all …]
|
| D | dm-snap-persistent.c | 326 ps->store->chunk_mask = ps->store->chunk_size - 1; in read_header()
|
| D | dm-snap.c | 1907 (bio->bi_iter.bi_sector & s->store->chunk_mask); in remap_exception()
|
| /Linux-v5.4/fs/ufs/ |
| D | dir.c | 117 const unsigned chunk_mask = UFS_SB(sb)->s_uspi->s_dirblksize - 1; in ufs_check_page() local 123 if (limit & chunk_mask) in ufs_check_page() 138 if (((offs + rec_len - 1) ^ offs) & ~chunk_mask) in ufs_check_page() 431 unsigned chunk_mask = ~(UFS_SB(sb)->s_uspi->s_dirblksize - 1); in ufs_readdir() local 456 offset = ufs_validate_entry(sb, kaddr, offset, chunk_mask); in ufs_readdir()
|
| /Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
| D | rx.c | 48 xsk_umem_fq_reuse(rq->umem, handle & rq->umem->chunk_mask); in mlx5e_xsk_recycle_frame()
|
| /Linux-v5.4/include/net/ |
| D | xdp_sock.h | 48 u64 chunk_mask; member
|
| /Linux-v5.4/drivers/net/ethernet/intel/ixgbe/ |
| D | ixgbe_xsk.c | 235 mask = rx_ring->xsk_umem->chunk_mask; in ixgbe_zca_free() 295 handle &= rx_ring->xsk_umem->chunk_mask; in ixgbe_alloc_buffer_slow_zc()
|
| /Linux-v5.4/drivers/net/ethernet/intel/i40e/ |
| D | i40e_xsk.c | 297 handle &= rx_ring->xsk_umem->chunk_mask; in i40e_alloc_buffer_slow_zc() 456 mask = rx_ring->xsk_umem->chunk_mask; in i40e_zca_free()
|
| /Linux-v5.4/fs/ext2/ |
| D | dir.c | 272 unsigned chunk_mask = ~(ext2_chunk_size(inode)-1); in ext2_readdir() local 297 offset = ext2_validate_entry(kaddr, offset, chunk_mask); in ext2_readdir()
|