Lines Matching refs:peer_req
92 static void drbd_endio_read_sec_final(struct drbd_peer_request *peer_req) __releases(local) in drbd_endio_read_sec_final() argument
95 struct drbd_peer_device *peer_device = peer_req->peer_device; in drbd_endio_read_sec_final()
99 device->read_cnt += peer_req->i.size >> 9; in drbd_endio_read_sec_final()
100 list_del(&peer_req->w.list); in drbd_endio_read_sec_final()
103 if (test_bit(__EE_WAS_ERROR, &peer_req->flags)) in drbd_endio_read_sec_final()
107 drbd_queue_work(&peer_device->connection->sender_work, &peer_req->w); in drbd_endio_read_sec_final()
113 void drbd_endio_write_sec_final(struct drbd_peer_request *peer_req) __releases(local) in drbd_endio_write_sec_final() argument
116 struct drbd_peer_device *peer_device = peer_req->peer_device; in drbd_endio_write_sec_final()
128 i = peer_req->i; in drbd_endio_write_sec_final()
129 do_al_complete_io = peer_req->flags & EE_CALL_AL_COMPLETE_IO; in drbd_endio_write_sec_final()
130 block_id = peer_req->block_id; in drbd_endio_write_sec_final()
131 peer_req->flags &= ~EE_CALL_AL_COMPLETE_IO; in drbd_endio_write_sec_final()
133 if (peer_req->flags & EE_WAS_ERROR) { in drbd_endio_write_sec_final()
136 if (!__test_and_set_bit(__EE_SEND_WRITE_ACK, &peer_req->flags)) in drbd_endio_write_sec_final()
138 drbd_set_out_of_sync(device, peer_req->i.sector, peer_req->i.size); in drbd_endio_write_sec_final()
142 device->writ_cnt += peer_req->i.size >> 9; in drbd_endio_write_sec_final()
143 list_move_tail(&peer_req->w.list, &device->done_ee); in drbd_endio_write_sec_final()
157 if (peer_req->flags & EE_WAS_ERROR) in drbd_endio_write_sec_final()
184 struct drbd_peer_request *peer_req = bio->bi_private; in drbd_peer_request_endio() local
185 struct drbd_device *device = peer_req->peer_device->device; in drbd_peer_request_endio()
194 (unsigned long long)peer_req->i.sector); in drbd_peer_request_endio()
197 set_bit(__EE_WAS_ERROR, &peer_req->flags); in drbd_peer_request_endio()
200 if (atomic_dec_and_test(&peer_req->pending_bios)) { in drbd_peer_request_endio()
202 drbd_endio_write_sec_final(peer_req); in drbd_peer_request_endio()
204 drbd_endio_read_sec_final(peer_req); in drbd_peer_request_endio()
298 void drbd_csum_ee(struct crypto_ahash *tfm, struct drbd_peer_request *peer_req, void *digest) in drbd_csum_ee() argument
302 struct page *page = peer_req->pages; in drbd_csum_ee()
320 len = peer_req->i.size & (PAGE_SIZE - 1); in drbd_csum_ee()
357 struct drbd_peer_request *peer_req = container_of(w, struct drbd_peer_request, w); in w_e_send_csum() local
358 struct drbd_peer_device *peer_device = peer_req->peer_device; in w_e_send_csum()
367 if (unlikely((peer_req->flags & EE_WAS_ERROR) != 0)) in w_e_send_csum()
373 sector_t sector = peer_req->i.sector; in w_e_send_csum()
374 unsigned int size = peer_req->i.size; in w_e_send_csum()
375 drbd_csum_ee(peer_device->connection->csums_tfm, peer_req, digest); in w_e_send_csum()
381 drbd_free_peer_req(device, peer_req); in w_e_send_csum()
382 peer_req = NULL; in w_e_send_csum()
394 if (peer_req) in w_e_send_csum()
395 drbd_free_peer_req(device, peer_req); in w_e_send_csum()
407 struct drbd_peer_request *peer_req; in read_for_csum() local
414 peer_req = drbd_alloc_peer_req(peer_device, ID_SYNCER /* unused */, sector, in read_for_csum()
416 if (!peer_req) in read_for_csum()
419 peer_req->w.cb = w_e_send_csum; in read_for_csum()
421 list_add_tail(&peer_req->w.list, &device->read_ee); in read_for_csum()
425 if (drbd_submit_peer_request(device, peer_req, REQ_OP_READ, 0, in read_for_csum()
434 list_del(&peer_req->w.list); in read_for_csum()
437 drbd_free_peer_req(device, peer_req); in read_for_csum()
1046 static void move_to_net_ee_or_free(struct drbd_device *device, struct drbd_peer_request *peer_req) in move_to_net_ee_or_free() argument
1048 if (drbd_peer_req_has_active_page(peer_req)) { in move_to_net_ee_or_free()
1050 int i = (peer_req->i.size + PAGE_SIZE -1) >> PAGE_SHIFT; in move_to_net_ee_or_free()
1054 list_add_tail(&peer_req->w.list, &device->net_ee); in move_to_net_ee_or_free()
1058 drbd_free_peer_req(device, peer_req); in move_to_net_ee_or_free()
1068 struct drbd_peer_request *peer_req = container_of(w, struct drbd_peer_request, w); in w_e_end_data_req() local
1069 struct drbd_peer_device *peer_device = peer_req->peer_device; in w_e_end_data_req()
1074 drbd_free_peer_req(device, peer_req); in w_e_end_data_req()
1079 if (likely((peer_req->flags & EE_WAS_ERROR) == 0)) { in w_e_end_data_req()
1080 err = drbd_send_block(peer_device, P_DATA_REPLY, peer_req); in w_e_end_data_req()
1084 (unsigned long long)peer_req->i.sector); in w_e_end_data_req()
1086 err = drbd_send_ack(peer_device, P_NEG_DREPLY, peer_req); in w_e_end_data_req()
1091 move_to_net_ee_or_free(device, peer_req); in w_e_end_data_req()
1098 static bool all_zero(struct drbd_peer_request *peer_req) in all_zero() argument
1100 struct page *page = peer_req->pages; in all_zero()
1101 unsigned int len = peer_req->i.size; in all_zero()
1129 struct drbd_peer_request *peer_req = container_of(w, struct drbd_peer_request, w); in w_e_end_rsdata_req() local
1130 struct drbd_peer_device *peer_device = peer_req->peer_device; in w_e_end_rsdata_req()
1135 drbd_free_peer_req(device, peer_req); in w_e_end_rsdata_req()
1141 drbd_rs_complete_io(device, peer_req->i.sector); in w_e_end_rsdata_req()
1146 err = drbd_send_ack(peer_device, P_RS_CANCEL, peer_req); in w_e_end_rsdata_req()
1147 } else if (likely((peer_req->flags & EE_WAS_ERROR) == 0)) { in w_e_end_rsdata_req()
1150 if (peer_req->flags & EE_RS_THIN_REQ && all_zero(peer_req)) in w_e_end_rsdata_req()
1151 err = drbd_send_rs_deallocated(peer_device, peer_req); in w_e_end_rsdata_req()
1153 err = drbd_send_block(peer_device, P_RS_DATA_REPLY, peer_req); in w_e_end_rsdata_req()
1163 (unsigned long long)peer_req->i.sector); in w_e_end_rsdata_req()
1165 err = drbd_send_ack(peer_device, P_NEG_RS_DREPLY, peer_req); in w_e_end_rsdata_req()
1168 drbd_rs_failed_io(device, peer_req->i.sector, peer_req->i.size); in w_e_end_rsdata_req()
1173 move_to_net_ee_or_free(device, peer_req); in w_e_end_rsdata_req()
1182 struct drbd_peer_request *peer_req = container_of(w, struct drbd_peer_request, w); in w_e_end_csum_rs_req() local
1183 struct drbd_peer_device *peer_device = peer_req->peer_device; in w_e_end_csum_rs_req()
1191 drbd_free_peer_req(device, peer_req); in w_e_end_csum_rs_req()
1197 drbd_rs_complete_io(device, peer_req->i.sector); in w_e_end_csum_rs_req()
1201 di = peer_req->digest; in w_e_end_csum_rs_req()
1203 if (likely((peer_req->flags & EE_WAS_ERROR) == 0)) { in w_e_end_csum_rs_req()
1213 drbd_csum_ee(peer_device->connection->csums_tfm, peer_req, digest); in w_e_end_csum_rs_req()
1219 drbd_set_in_sync(device, peer_req->i.sector, peer_req->i.size); in w_e_end_csum_rs_req()
1221 device->rs_same_csum += peer_req->i.size >> BM_BLOCK_SHIFT; in w_e_end_csum_rs_req()
1222 err = drbd_send_ack(peer_device, P_RS_IS_IN_SYNC, peer_req); in w_e_end_csum_rs_req()
1225 peer_req->block_id = ID_SYNCER; /* By setting block_id, digest pointer becomes invalid! */ in w_e_end_csum_rs_req()
1226 peer_req->flags &= ~EE_HAS_DIGEST; /* This peer request no longer has a digest pointer */ in w_e_end_csum_rs_req()
1228 err = drbd_send_block(peer_device, P_RS_DATA_REPLY, peer_req); in w_e_end_csum_rs_req()
1231 err = drbd_send_ack(peer_device, P_NEG_RS_DREPLY, peer_req); in w_e_end_csum_rs_req()
1237 move_to_net_ee_or_free(device, peer_req); in w_e_end_csum_rs_req()
1246 struct drbd_peer_request *peer_req = container_of(w, struct drbd_peer_request, w); in w_e_end_ov_req() local
1247 struct drbd_peer_device *peer_device = peer_req->peer_device; in w_e_end_ov_req()
1249 sector_t sector = peer_req->i.sector; in w_e_end_ov_req()
1250 unsigned int size = peer_req->i.size; in w_e_end_ov_req()
1265 if (likely(!(peer_req->flags & EE_WAS_ERROR))) in w_e_end_ov_req()
1266 drbd_csum_ee(peer_device->connection->verify_tfm, peer_req, digest); in w_e_end_ov_req()
1275 drbd_free_peer_req(device, peer_req); in w_e_end_ov_req()
1276 peer_req = NULL; in w_e_end_ov_req()
1284 if (peer_req) in w_e_end_ov_req()
1285 drbd_free_peer_req(device, peer_req); in w_e_end_ov_req()
1303 struct drbd_peer_request *peer_req = container_of(w, struct drbd_peer_request, w); in w_e_end_ov_reply() local
1304 struct drbd_peer_device *peer_device = peer_req->peer_device; in w_e_end_ov_reply()
1308 sector_t sector = peer_req->i.sector; in w_e_end_ov_reply()
1309 unsigned int size = peer_req->i.size; in w_e_end_ov_reply()
1315 drbd_free_peer_req(device, peer_req); in w_e_end_ov_reply()
1323 drbd_rs_complete_io(device, peer_req->i.sector); in w_e_end_ov_reply()
1327 di = peer_req->digest; in w_e_end_ov_reply()
1329 if (likely((peer_req->flags & EE_WAS_ERROR) == 0)) { in w_e_end_ov_reply()
1333 drbd_csum_ee(peer_device->connection->verify_tfm, peer_req, digest); in w_e_end_ov_reply()
1346 drbd_free_peer_req(device, peer_req); in w_e_end_ov_reply()