Lines Matching refs:sector
126 drbd_set_out_of_sync(device, peer_req->i.sector, peer_req->i.size); in drbd_endio_write_sec_final()
156 drbd_rs_complete_io(device, i.sector); in drbd_endio_write_sec_final()
182 (unsigned long long)peer_req->i.sector); in drbd_peer_request_endio()
360 sector_t sector = peer_req->i.sector; in w_e_send_csum() local
371 err = drbd_send_drequest_csum(peer_device, sector, size, in w_e_send_csum()
391 static int read_for_csum(struct drbd_peer_device *peer_device, sector_t sector, int size) in read_for_csum() argument
401 peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER /* unused */, sector, in read_for_csum()
592 sector_t sector; in make_resync_request() local
658 sector = BM_BIT_TO_SECT(bit); in make_resync_request()
660 if (drbd_try_rs_begin_io(device, sector)) { in make_resync_request()
667 drbd_rs_complete_io(device, sector); in make_resync_request()
685 if (sector & ((1<<(align+3))-1)) in make_resync_request()
714 if (sector + (size>>9) > capacity) in make_resync_request()
715 size = (capacity-sector)<<9; in make_resync_request()
718 switch (read_for_csum(peer_device, sector, size)) { in make_resync_request()
723 drbd_rs_complete_io(device, sector); in make_resync_request()
724 device->bm_resync_fo = BM_SECT_TO_BIT(sector); in make_resync_request()
739 sector, size, ID_SYNCER); in make_resync_request()
770 sector_t sector; in make_ov_request() local
779 sector = device->ov_position; in make_ov_request()
781 if (sector >= capacity) in make_ov_request()
789 && sector >= device->ov_stop_sector; in make_ov_request()
795 if (drbd_try_rs_begin_io(device, sector)) { in make_ov_request()
796 device->ov_position = sector; in make_ov_request()
800 if (sector + (size>>9) > capacity) in make_ov_request()
801 size = (capacity-sector)<<9; in make_ov_request()
804 if (drbd_send_ov_request(first_peer_device(device), sector, size)) { in make_ov_request()
808 sector += BM_SECT_PER_BIT; in make_ov_request()
810 device->ov_position = sector; in make_ov_request()
1071 (unsigned long long)peer_req->i.sector); in w_e_end_data_req()
1128 drbd_rs_complete_io(device, peer_req->i.sector); in w_e_end_rsdata_req()
1150 (unsigned long long)peer_req->i.sector); in w_e_end_rsdata_req()
1155 drbd_rs_failed_io(device, peer_req->i.sector, peer_req->i.size); in w_e_end_rsdata_req()
1184 drbd_rs_complete_io(device, peer_req->i.sector); in w_e_end_csum_rs_req()
1206 drbd_set_in_sync(device, peer_req->i.sector, peer_req->i.size); in w_e_end_csum_rs_req()
1236 sector_t sector = peer_req->i.sector; in w_e_end_ov_req() local
1265 err = drbd_send_drequest_csum(peer_device, sector, size, digest, digest_size, P_OV_REPLY); in w_e_end_ov_req()
1277 void drbd_ov_out_of_sync_found(struct drbd_device *device, sector_t sector, int size) in drbd_ov_out_of_sync_found() argument
1279 if (device->ov_last_oos_start + device->ov_last_oos_size == sector) { in drbd_ov_out_of_sync_found()
1282 device->ov_last_oos_start = sector; in drbd_ov_out_of_sync_found()
1285 drbd_set_out_of_sync(device, sector, size); in drbd_ov_out_of_sync_found()
1295 sector_t sector = peer_req->i.sector; in w_e_end_ov_reply() local
1310 drbd_rs_complete_io(device, peer_req->i.sector); in w_e_end_ov_reply()
1335 drbd_ov_out_of_sync_found(device, sector, size); in w_e_end_ov_reply()
1339 err = drbd_send_ack_ex(peer_device, P_OV_RESULT, sector, size, in w_e_end_ov_reply()
1351 (sector + (size>>9)) >= device->ov_stop_sector; in w_e_end_ov_reply()
1506 err = drbd_send_drequest(peer_device, P_DATA_REQUEST, req->i.sector, req->i.size, in w_send_read_req()