Searched refs:mqd_backup (Results 1 – 10 of 10) sorted by relevance
/Linux-v6.6/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_gfx.c | 410 kiq->mqd_backup = kmalloc(mqd_size, GFP_KERNEL); in amdgpu_gfx_mqd_sw_init() 411 if (!kiq->mqd_backup) { in amdgpu_gfx_mqd_sw_init() 433 adev->gfx.me.mqd_backup[i] = kmalloc(mqd_size, GFP_KERNEL); in amdgpu_gfx_mqd_sw_init() 434 if (!adev->gfx.me.mqd_backup[i]) { in amdgpu_gfx_mqd_sw_init() 457 adev->gfx.mec.mqd_backup[j] = kmalloc(mqd_size, GFP_KERNEL); in amdgpu_gfx_mqd_sw_init() 458 if (!adev->gfx.mec.mqd_backup[j]) { in amdgpu_gfx_mqd_sw_init() 477 kfree(adev->gfx.me.mqd_backup[i]); in amdgpu_gfx_mqd_sw_fini() 487 kfree(adev->gfx.mec.mqd_backup[j]); in amdgpu_gfx_mqd_sw_fini() 494 kfree(kiq->mqd_backup); in amdgpu_gfx_mqd_sw_fini()
|
D | amdgpu_gfx.h | 114 void *mqd_backup[AMDGPU_MAX_COMPUTE_RINGS * AMDGPU_MAX_GC_INSTANCES]; member 161 void *mqd_backup; member 341 void *mqd_backup[AMDGPU_MAX_GFX_RINGS]; member
|
D | amdgpu_mes.h | 106 void *mqd_backup[AMDGPU_MAX_MES_PIPES]; member
|
D | mes_v10_1.c | 911 adev->mes.mqd_backup[pipe] = kmalloc(mqd_size, GFP_KERNEL); in mes_v10_1_mqd_sw_init() 912 if (!adev->mes.mqd_backup[pipe]) { in mes_v10_1_mqd_sw_init() 969 kfree(adev->mes.mqd_backup[pipe]); in mes_v10_1_sw_fini()
|
D | gfx_v9_4_3.c | 1765 tmp_mqd = (struct v9_mqd *)adev->gfx.kiq[xcc_id].mqd_backup; in gfx_v9_4_3_xcc_kiq_init_queue() 1768 if (adev->gfx.kiq[xcc_id].mqd_backup) in gfx_v9_4_3_xcc_kiq_init_queue() 1769 memcpy(mqd, adev->gfx.kiq[xcc_id].mqd_backup, sizeof(struct v9_mqd_allocation)); in gfx_v9_4_3_xcc_kiq_init_queue() 1792 if (adev->gfx.kiq[xcc_id].mqd_backup) in gfx_v9_4_3_xcc_kiq_init_queue() 1793 memcpy(adev->gfx.kiq[xcc_id].mqd_backup, mqd, sizeof(struct v9_mqd_allocation)); in gfx_v9_4_3_xcc_kiq_init_queue() 1809 tmp_mqd = (struct v9_mqd *)adev->gfx.mec.mqd_backup[mqd_idx]; in gfx_v9_4_3_xcc_kcq_init_queue() 1822 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v9_4_3_xcc_kcq_init_queue() 1823 memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct v9_mqd_allocation)); in gfx_v9_4_3_xcc_kcq_init_queue() 1826 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v9_4_3_xcc_kcq_init_queue() 1827 memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct v9_mqd_allocation)); in gfx_v9_4_3_xcc_kcq_init_queue()
|
D | mes_v11_0.c | 1018 adev->mes.mqd_backup[pipe] = kmalloc(mqd_size, GFP_KERNEL); in mes_v11_0_mqd_sw_init() 1019 if (!adev->mes.mqd_backup[pipe]) { in mes_v11_0_mqd_sw_init() 1077 kfree(adev->mes.mqd_backup[pipe]); in mes_v11_0_sw_fini()
|
D | gfx_v11_0.c | 3686 if (adev->gfx.me.mqd_backup[mqd_idx]) in gfx_v11_0_gfx_init_queue() 3687 memcpy(adev->gfx.me.mqd_backup[mqd_idx], mqd, sizeof(*mqd)); in gfx_v11_0_gfx_init_queue() 3690 if (adev->gfx.me.mqd_backup[mqd_idx]) in gfx_v11_0_gfx_init_queue() 3691 memcpy(mqd, adev->gfx.me.mqd_backup[mqd_idx], sizeof(*mqd)); in gfx_v11_0_gfx_init_queue() 3979 if (adev->gfx.kiq[0].mqd_backup) in gfx_v11_0_kiq_init_queue() 3980 memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(*mqd)); in gfx_v11_0_kiq_init_queue() 4002 if (adev->gfx.kiq[0].mqd_backup) in gfx_v11_0_kiq_init_queue() 4003 memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(*mqd)); in gfx_v11_0_kiq_init_queue() 4023 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v11_0_kcq_init_queue() 4024 memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd)); in gfx_v11_0_kcq_init_queue() [all …]
|
D | gfx_v9_0.c | 3534 tmp_mqd = (struct v9_mqd *)adev->gfx.kiq[0].mqd_backup; in gfx_v9_0_kiq_init_queue() 3537 if (adev->gfx.kiq[0].mqd_backup) in gfx_v9_0_kiq_init_queue() 3538 memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(struct v9_mqd_allocation)); in gfx_v9_0_kiq_init_queue() 3562 if (adev->gfx.kiq[0].mqd_backup) in gfx_v9_0_kiq_init_queue() 3563 memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(struct v9_mqd_allocation)); in gfx_v9_0_kiq_init_queue() 3579 tmp_mqd = (struct v9_mqd *)adev->gfx.mec.mqd_backup[mqd_idx]; in gfx_v9_0_kcq_init_queue() 3592 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v9_0_kcq_init_queue() 3593 memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct v9_mqd_allocation)); in gfx_v9_0_kcq_init_queue() 3596 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v9_0_kcq_init_queue() 3597 memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct v9_mqd_allocation)); in gfx_v9_0_kcq_init_queue()
|
D | gfx_v10_0.c | 6459 if (adev->gfx.me.mqd_backup[mqd_idx]) in gfx_v10_0_gfx_init_queue() 6460 memcpy(adev->gfx.me.mqd_backup[mqd_idx], mqd, sizeof(*mqd)); in gfx_v10_0_gfx_init_queue() 6463 if (adev->gfx.me.mqd_backup[mqd_idx]) in gfx_v10_0_gfx_init_queue() 6464 memcpy(mqd, adev->gfx.me.mqd_backup[mqd_idx], sizeof(*mqd)); in gfx_v10_0_gfx_init_queue() 6737 if (adev->gfx.kiq[0].mqd_backup) in gfx_v10_0_kiq_init_queue() 6738 memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(*mqd)); in gfx_v10_0_kiq_init_queue() 6760 if (adev->gfx.kiq[0].mqd_backup) in gfx_v10_0_kiq_init_queue() 6761 memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(*mqd)); in gfx_v10_0_kiq_init_queue() 6781 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v10_0_kcq_init_queue() 6782 memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd)); in gfx_v10_0_kcq_init_queue() [all …]
|
D | gfx_v8_0.c | 4604 if (adev->gfx.kiq[0].mqd_backup) in gfx_v8_0_kiq_init_queue() 4605 memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(struct vi_mqd_allocation)); in gfx_v8_0_kiq_init_queue() 4628 if (adev->gfx.kiq[0].mqd_backup) in gfx_v8_0_kiq_init_queue() 4629 memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(struct vi_mqd_allocation)); in gfx_v8_0_kiq_init_queue() 4651 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v8_0_kcq_init_queue() 4652 memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct vi_mqd_allocation)); in gfx_v8_0_kcq_init_queue() 4655 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v8_0_kcq_init_queue() 4656 memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct vi_mqd_allocation)); in gfx_v8_0_kcq_init_queue()
|