Lines Matching refs:raid_map

1241 	struct pqi_scsi_dev *device, struct raid_map *raid_map)  in pqi_validate_raid_map()  argument
1247 raid_map_size = get_unaligned_le32(&raid_map->structure_size); in pqi_validate_raid_map()
1249 if (raid_map_size < offsetof(struct raid_map, disk_data)) { in pqi_validate_raid_map()
1255 if (get_unaligned_le16(&raid_map->layout_map_count) != 2) { in pqi_validate_raid_map()
1260 if (get_unaligned_le16(&raid_map->layout_map_count) != 3) { in pqi_validate_raid_map()
1266 get_unaligned_le16(&raid_map->layout_map_count) > 1) { in pqi_validate_raid_map()
1269 get_unaligned_le16(&raid_map->strip_size) * in pqi_validate_raid_map()
1270 get_unaligned_le16(&raid_map->data_disks_per_row); in pqi_validate_raid_map()
1293 struct raid_map *raid_map; in pqi_get_raid_map() local
1295 raid_map = kmalloc(sizeof(*raid_map), GFP_KERNEL); in pqi_get_raid_map()
1296 if (!raid_map) in pqi_get_raid_map()
1300 device->scsi3addr, raid_map, sizeof(*raid_map), 0, NULL); in pqi_get_raid_map()
1304 raid_map_size = get_unaligned_le32(&raid_map->structure_size); in pqi_get_raid_map()
1306 if (raid_map_size > sizeof(*raid_map)) { in pqi_get_raid_map()
1308 kfree(raid_map); in pqi_get_raid_map()
1310 raid_map = kmalloc(raid_map_size, GFP_KERNEL); in pqi_get_raid_map()
1311 if (!raid_map) in pqi_get_raid_map()
1315 device->scsi3addr, raid_map, raid_map_size, 0, NULL); in pqi_get_raid_map()
1319 if (get_unaligned_le32(&raid_map->structure_size) in pqi_get_raid_map()
1324 get_unaligned_le32(&raid_map->structure_size)); in pqi_get_raid_map()
1330 rc = pqi_validate_raid_map(ctrl_info, device, raid_map); in pqi_get_raid_map()
1334 device->raid_map = raid_map; in pqi_get_raid_map()
1339 kfree(raid_map); in pqi_get_raid_map()
1404 if (get_unaligned_le16(&device->raid_map->flags) & in pqi_get_raid_bypass_status()
1880 kfree(existing_device->raid_map); in pqi_scsi_update_device()
1881 existing_device->raid_map = new_device->raid_map; in pqi_scsi_update_device()
1889 new_device->raid_map = NULL; in pqi_scsi_update_device()
1895 kfree(device->raid_map); in pqi_free_device()
2379 struct raid_map *raid_map, u64 first_block) in pqi_set_encryption_info() argument
2388 volume_blk_size = get_unaligned_le32(&raid_map->volume_blk_size); in pqi_set_encryption_info()
2393 get_unaligned_le16(&raid_map->data_encryption_key_index); in pqi_set_encryption_info()
2487 struct pqi_scsi_dev_raid_map_data *rmd, struct raid_map *raid_map) in pci_get_aio_common_raid_map_values() argument
2497 get_unaligned_le64(&raid_map->volume_blk_cnt) || in pci_get_aio_common_raid_map_values()
2502 get_unaligned_le16(&raid_map->data_disks_per_row); in pci_get_aio_common_raid_map_values()
2503 rmd->strip_size = get_unaligned_le16(&raid_map->strip_size); in pci_get_aio_common_raid_map_values()
2504 rmd->layout_map_count = get_unaligned_le16(&raid_map->layout_map_count); in pci_get_aio_common_raid_map_values()
2543 get_unaligned_le16(&raid_map->metadata_disks_per_row); in pci_get_aio_common_raid_map_values()
2545 raid_map->parity_rotation_shift)) % in pci_get_aio_common_raid_map_values()
2546 get_unaligned_le16(&raid_map->row_cnt); in pci_get_aio_common_raid_map_values()
2554 struct raid_map *raid_map) in pqi_calc_aio_r5_or_r6() argument
2637 ((u32)(rmd->first_row >> raid_map->parity_rotation_shift)) % in pqi_calc_aio_r5_or_r6()
2638 get_unaligned_le16(&raid_map->row_cnt); in pqi_calc_aio_r5_or_r6()
2641 (get_unaligned_le16(&raid_map->row_cnt) * in pqi_calc_aio_r5_or_r6()
2659 index -= get_unaligned_le16(&raid_map->metadata_disks_per_row); in pqi_calc_aio_r5_or_r6()
2661 rmd->p_parity_it_nexus = raid_map->disk_data[index].aio_handle; in pqi_calc_aio_r5_or_r6()
2663 rmd->q_parity_it_nexus = raid_map->disk_data[index + 1].aio_handle; in pqi_calc_aio_r5_or_r6()
2664 rmd->xor_mult = raid_map->disk_data[rmd->map_index].xor_mult[1]; in pqi_calc_aio_r5_or_r6()
2700 static void pqi_calc_aio_r1_nexus(struct raid_map *raid_map, in pqi_calc_aio_r1_nexus() argument
2709 rmd->it_nexus[0] = raid_map->disk_data[index].aio_handle; in pqi_calc_aio_r1_nexus()
2711 rmd->it_nexus[1] = raid_map->disk_data[index].aio_handle; in pqi_calc_aio_r1_nexus()
2714 rmd->it_nexus[2] = raid_map->disk_data[index].aio_handle; in pqi_calc_aio_r1_nexus()
2725 struct raid_map *raid_map; in pqi_raid_bypass_submit_scsi_cmd() local
2744 raid_map = device->raid_map; in pqi_raid_bypass_submit_scsi_cmd()
2746 rc = pci_get_aio_common_raid_map_values(ctrl_info, &rmd, raid_map); in pqi_raid_bypass_submit_scsi_cmd()
2753 pqi_calc_aio_r1_nexus(raid_map, &rmd); in pqi_raid_bypass_submit_scsi_cmd()
2765 rc = pqi_calc_aio_r5_or_r6(&rmd, raid_map); in pqi_raid_bypass_submit_scsi_cmd()
2773 rmd.aio_handle = raid_map->disk_data[rmd.map_index].aio_handle; in pqi_raid_bypass_submit_scsi_cmd()
2774 rmd.disk_block = get_unaligned_le64(&raid_map->disk_starting_blk) + in pqi_raid_bypass_submit_scsi_cmd()
2780 if (raid_map->phys_blk_shift) { in pqi_raid_bypass_submit_scsi_cmd()
2781 rmd.disk_block <<= raid_map->phys_blk_shift; in pqi_raid_bypass_submit_scsi_cmd()
2782 rmd.disk_block_cnt <<= raid_map->phys_blk_shift; in pqi_raid_bypass_submit_scsi_cmd()
2790 if (get_unaligned_le16(&raid_map->flags) & RAID_MAP_ENCRYPTION_ENABLED) { in pqi_raid_bypass_submit_scsi_cmd()
2793 pqi_set_encryption_info(&encryption_info, raid_map, rmd.first_block); in pqi_raid_bypass_submit_scsi_cmd()