Lines Matching refs:lba
876 unsigned long long lba) in lba2fake_store() argument
880 lba = do_div(lba, sdebug_store_sectors); in lba2fake_store()
885 return lsip->storep + lba * sdebug_sector_size; in lba2fake_store()
2666 unsigned long long lba) in zbc_zone() argument
2668 return &devip->zstate[lba >> devip->zsize_shift]; in zbc_zone()
2746 unsigned long long lba, unsigned int num) in zbc_inc_wp() argument
2748 struct sdeb_zone_state *zsp = zbc_zone(devip, lba); in zbc_inc_wp()
2762 if (lba != zsp->z_wp) in zbc_inc_wp()
2765 end = lba + num; in zbc_inc_wp()
2767 n = zend - lba; in zbc_inc_wp()
2779 lba += n; in zbc_inc_wp()
2788 unsigned long long lba, unsigned int num, bool write) in check_zbc_access_params() argument
2792 struct sdeb_zone_state *zsp = zbc_zone(devip, lba); in check_zbc_access_params()
2793 struct sdeb_zone_state *zsp_end = zbc_zone(devip, lba + num - 1); in check_zbc_access_params()
2836 if (lba != zsp->z_wp) { in check_zbc_access_params()
2860 (struct scsi_cmnd *scp, unsigned long long lba, in check_device_access_params() argument
2866 if (lba + num > sdebug_capacity) { in check_device_access_params()
2881 return check_zbc_access_params(scp, lba, num, write); in check_device_access_params()
2904 u32 sg_skip, u64 lba, u32 num, bool do_write) in do_device_access() argument
2925 block = do_div(lba, sdebug_store_sectors); in do_device_access()
2961 static bool comp_write_worker(struct sdeb_store_info *sip, u64 lba, u32 num, in comp_write_worker() argument
2970 block = do_div(lba, store_blks); in comp_write_worker()
3122 u64 lba; in resp_read_dt0() local
3130 lba = get_unaligned_be64(cmd + 2); in resp_read_dt0()
3136 lba = get_unaligned_be32(cmd + 2); in resp_read_dt0()
3142 lba = (u32)cmd[3] | (u32)cmd[2] << 8 | in resp_read_dt0()
3149 lba = get_unaligned_be32(cmd + 2); in resp_read_dt0()
3155 lba = get_unaligned_be32(cmd + 2); in resp_read_dt0()
3160 lba = get_unaligned_be64(cmd + 12); in resp_read_dt0()
3184 ret = check_device_access_params(scp, lba, num, false); in resp_read_dt0()
3188 (lba <= (sdebug_medium_error_start + sdebug_medium_error_count - 1)) && in resp_read_dt0()
3189 ((lba + num) > sdebug_medium_error_start))) { in resp_read_dt0()
3195 ret = (lba < OPT_MEDIUM_ERR_ADDR) in resp_read_dt0()
3196 ? OPT_MEDIUM_ERR_ADDR : (int)lba; in resp_read_dt0()
3207 switch (prot_verify_read(scp, lba, num, ei_lba)) { in resp_read_dt0()
3233 ret = do_device_access(sip, scp, 0, lba, num, false); in resp_read_dt0()
3332 static unsigned long lba_to_map_index(sector_t lba) in lba_to_map_index() argument
3335 lba += sdebug_unmap_granularity - sdebug_unmap_alignment; in lba_to_map_index()
3336 sector_div(lba, sdebug_unmap_granularity); in lba_to_map_index()
3337 return lba; in lba_to_map_index()
3342 sector_t lba = index * sdebug_unmap_granularity; in map_index_to_lba() local
3345 lba -= sdebug_unmap_granularity - sdebug_unmap_alignment; in map_index_to_lba()
3346 return lba; in map_index_to_lba()
3349 static unsigned int map_state(struct sdeb_store_info *sip, sector_t lba, in map_state() argument
3357 index = lba_to_map_index(lba); in map_state()
3366 *num = end - lba; in map_state()
3370 static void map_region(struct sdeb_store_info *sip, sector_t lba, in map_region() argument
3373 sector_t end = lba + len; in map_region()
3375 while (lba < end) { in map_region()
3376 unsigned long index = lba_to_map_index(lba); in map_region()
3381 lba = map_index_to_lba(index + 1); in map_region()
3385 static void unmap_region(struct sdeb_store_info *sip, sector_t lba, in unmap_region() argument
3388 sector_t end = lba + len; in unmap_region()
3391 while (lba < end) { in unmap_region()
3392 unsigned long index = lba_to_map_index(lba); in unmap_region()
3394 if (lba == map_index_to_lba(index) && in unmap_region()
3395 lba + sdebug_unmap_granularity <= end && in unmap_region()
3399 memset(fsp + lba * sdebug_sector_size, in unmap_region()
3405 memset(sip->dif_storep + lba, 0xff, in unmap_region()
3410 lba = map_index_to_lba(index + 1); in unmap_region()
3420 u64 lba; in resp_write_dt0() local
3428 lba = get_unaligned_be64(cmd + 2); in resp_write_dt0()
3434 lba = get_unaligned_be32(cmd + 2); in resp_write_dt0()
3440 lba = (u32)cmd[3] | (u32)cmd[2] << 8 | in resp_write_dt0()
3447 lba = get_unaligned_be32(cmd + 2); in resp_write_dt0()
3453 lba = get_unaligned_be32(cmd + 2); in resp_write_dt0()
3458 lba = get_unaligned_be64(cmd + 12); in resp_write_dt0()
3478 ret = check_device_access_params(scp, lba, num, true); in resp_write_dt0()
3486 switch (prot_verify_write(scp, lba, num, ei_lba)) { in resp_write_dt0()
3512 ret = do_device_access(sip, scp, 0, lba, num, true); in resp_write_dt0()
3514 map_region(sip, lba, num); in resp_write_dt0()
3517 zbc_inc_wp(devip, lba, num); in resp_write_dt0()
3564 u64 lba; in resp_write_scat() local
3631 lba = get_unaligned_be64(up + 0); in resp_write_scat()
3636 my_name, __func__, k, lba, num, sg_off); in resp_write_scat()
3639 ret = check_device_access_params(scp, lba, num, true); in resp_write_scat()
3658 int prot_ret = prot_verify_write(scp, lba, num, in resp_write_scat()
3669 ret = do_device_access(sip, scp, sg_off, lba, num, true); in resp_write_scat()
3672 zbc_inc_wp(devip, lba, num); in resp_write_scat()
3674 map_region(sip, lba, num); in resp_write_scat()
3714 static int resp_write_same(struct scsi_cmnd *scp, u64 lba, u32 num, in resp_write_same() argument
3731 ret = check_device_access_params(scp, lba, num, true); in resp_write_same()
3738 unmap_region(sip, lba, num); in resp_write_same()
3741 lbaa = lba; in resp_write_same()
3762 lbaa = lba + i; in resp_write_same()
3767 map_region(sip, lba, num); in resp_write_same()
3770 zbc_inc_wp(devip, lba, num); in resp_write_same()
3781 u32 lba; in resp_write_same_10() local
3793 lba = get_unaligned_be32(cmd + 2); in resp_write_same_10()
3799 return resp_write_same(scp, lba, num, ei_lba, unmap, false); in resp_write_same_10()
3806 u64 lba; in resp_write_same_16() local
3821 lba = get_unaligned_be64(cmd + 2); in resp_write_same_16()
3827 return resp_write_same(scp, lba, num, ei_lba, unmap, ndob); in resp_write_same_16()
3886 u64 lba; in resp_comp_write() local
3893 lba = get_unaligned_be64(cmd + 2); in resp_comp_write()
3907 ret = check_device_access_params(scp, lba, num, false); in resp_comp_write()
3928 if (!comp_write_worker(sip, lba, num, arr, false)) { in resp_comp_write()
3934 map_region(sip, lba, num); in resp_comp_write()
3942 __be64 lba; member
3984 unsigned long long lba = get_unaligned_be64(&desc[i].lba); in resp_unmap() local
3987 ret = check_device_access_params(scp, lba, num, true); in resp_unmap()
3991 unmap_region(sip, lba, num); in resp_unmap()
4009 u64 lba; in resp_get_lba_status() local
4014 lba = get_unaligned_be64(cmd + 2); in resp_get_lba_status()
4020 ret = check_device_access_params(scp, lba, 1, false); in resp_get_lba_status()
4027 mapped = map_state(sip, lba, &num); in resp_get_lba_status()
4032 if (sdebug_capacity - lba <= 0xffffffff) in resp_get_lba_status()
4033 num = sdebug_capacity - lba; in resp_get_lba_status()
4040 put_unaligned_be64(lba, arr + 8); /* LBA */ in resp_get_lba_status()
4051 u64 lba; in resp_sync_cache() local
4056 lba = get_unaligned_be32(cmd + 2); in resp_sync_cache()
4059 lba = get_unaligned_be64(cmd + 2); in resp_sync_cache()
4062 if (lba + num_blocks > sdebug_capacity) { in resp_sync_cache()
4084 u64 lba; in resp_pre_fetch() local
4093 lba = get_unaligned_be32(cmd + 2); in resp_pre_fetch()
4096 lba = get_unaligned_be64(cmd + 2); in resp_pre_fetch()
4099 if (lba + nblks > sdebug_capacity) { in resp_pre_fetch()
4106 block = do_div(lba, sdebug_store_sectors); in resp_pre_fetch()
4233 u64 lba; in resp_verify() local
4250 lba = get_unaligned_be64(cmd + 2); in resp_verify()
4254 lba = get_unaligned_be32(cmd + 2); in resp_verify()
4263 ret = check_device_access_params(scp, lba, a_num, false); in resp_verify()
4290 if (!comp_write_worker(sip, lba, vnum, arr, true)) { in resp_verify()
4311 u64 lba, zs_lba; in resp_report_zones() local
4347 lba = zs_lba + devip->zsize * i; in resp_report_zones()
4348 if (lba > sdebug_capacity) in resp_report_zones()
4350 zsp = zbc_zone(devip, lba); in resp_report_zones()