mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	drm/amdgpu: add multi-xcc support to amdgpu_gfx interfaces (v4)
v1: Modify kiq_init/fini, mqd_sw_init/fini and enable/disable_kcq to adapt to multi-die case. Pass 0 as default to all asics with single xcc (Le) v2: squash commits to avoid breaking the build (Le) v3: unify naming style (Le) v4: apply the changes to gc v11_0 (Hawking) Signed-off-by: Le Ma <le.ma@amd.com> Reviewed-by: Hawking Zhang <Hawking.Zhang@amd.com> Signed-off-by: Hawking Zhang <Hawking.Zhang@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
		
							parent
							
								
									c38be07035
								
							
						
					
					
						commit
						def799c659
					
				
					 6 changed files with 93 additions and 88 deletions
				
			
		| 
						 | 
					@ -267,7 +267,7 @@ void amdgpu_gfx_graphics_queue_acquire(struct amdgpu_device *adev)
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int amdgpu_gfx_kiq_acquire(struct amdgpu_device *adev,
 | 
					static int amdgpu_gfx_kiq_acquire(struct amdgpu_device *adev,
 | 
				
			||||||
				  struct amdgpu_ring *ring)
 | 
									  struct amdgpu_ring *ring, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	int queue_bit;
 | 
						int queue_bit;
 | 
				
			||||||
	int mec, pipe, queue;
 | 
						int mec, pipe, queue;
 | 
				
			||||||
| 
						 | 
					@ -277,7 +277,7 @@ static int amdgpu_gfx_kiq_acquire(struct amdgpu_device *adev,
 | 
				
			||||||
		    * adev->gfx.mec.num_queue_per_pipe;
 | 
							    * adev->gfx.mec.num_queue_per_pipe;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	while (--queue_bit >= 0) {
 | 
						while (--queue_bit >= 0) {
 | 
				
			||||||
		if (test_bit(queue_bit, adev->gfx.mec_bitmap[0].queue_bitmap))
 | 
							if (test_bit(queue_bit, adev->gfx.mec_bitmap[xcc_id].queue_bitmap))
 | 
				
			||||||
			continue;
 | 
								continue;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		amdgpu_queue_mask_bit_to_mec_queue(adev, queue_bit, &mec, &pipe, &queue);
 | 
							amdgpu_queue_mask_bit_to_mec_queue(adev, queue_bit, &mec, &pipe, &queue);
 | 
				
			||||||
| 
						 | 
					@ -303,9 +303,9 @@ static int amdgpu_gfx_kiq_acquire(struct amdgpu_device *adev,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int amdgpu_gfx_kiq_init_ring(struct amdgpu_device *adev,
 | 
					int amdgpu_gfx_kiq_init_ring(struct amdgpu_device *adev,
 | 
				
			||||||
			     struct amdgpu_ring *ring,
 | 
								     struct amdgpu_ring *ring,
 | 
				
			||||||
			     struct amdgpu_irq_src *irq)
 | 
								     struct amdgpu_irq_src *irq, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_kiq *kiq = &adev->gfx.kiq[0];
 | 
						struct amdgpu_kiq *kiq = &adev->gfx.kiq[xcc_id];
 | 
				
			||||||
	int r = 0;
 | 
						int r = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	spin_lock_init(&kiq->ring_lock);
 | 
						spin_lock_init(&kiq->ring_lock);
 | 
				
			||||||
| 
						 | 
					@ -314,15 +314,16 @@ int amdgpu_gfx_kiq_init_ring(struct amdgpu_device *adev,
 | 
				
			||||||
	ring->ring_obj = NULL;
 | 
						ring->ring_obj = NULL;
 | 
				
			||||||
	ring->use_doorbell = true;
 | 
						ring->use_doorbell = true;
 | 
				
			||||||
	ring->doorbell_index = adev->doorbell_index.kiq;
 | 
						ring->doorbell_index = adev->doorbell_index.kiq;
 | 
				
			||||||
 | 
						ring->xcc_id = xcc_id;
 | 
				
			||||||
	ring->vm_hub = AMDGPU_GFXHUB_0;
 | 
						ring->vm_hub = AMDGPU_GFXHUB_0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_kiq_acquire(adev, ring);
 | 
						r = amdgpu_gfx_kiq_acquire(adev, ring, xcc_id);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	ring->eop_gpu_addr = kiq->eop_gpu_addr;
 | 
						ring->eop_gpu_addr = kiq->eop_gpu_addr;
 | 
				
			||||||
	ring->no_scheduler = true;
 | 
						ring->no_scheduler = true;
 | 
				
			||||||
	sprintf(ring->name, "kiq_%d.%d.%d", ring->me, ring->pipe, ring->queue);
 | 
						sprintf(ring->name, "kiq_%d.%d.%d.%d", xcc_id, ring->me, ring->pipe, ring->queue);
 | 
				
			||||||
	r = amdgpu_ring_init(adev, ring, 1024, irq, AMDGPU_CP_KIQ_IRQ_DRIVER0,
 | 
						r = amdgpu_ring_init(adev, ring, 1024, irq, AMDGPU_CP_KIQ_IRQ_DRIVER0,
 | 
				
			||||||
			     AMDGPU_RING_PRIO_DEFAULT, NULL);
 | 
								     AMDGPU_RING_PRIO_DEFAULT, NULL);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
| 
						 | 
					@ -336,19 +337,19 @@ void amdgpu_gfx_kiq_free_ring(struct amdgpu_ring *ring)
 | 
				
			||||||
	amdgpu_ring_fini(ring);
 | 
						amdgpu_ring_fini(ring);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void amdgpu_gfx_kiq_fini(struct amdgpu_device *adev)
 | 
					void amdgpu_gfx_kiq_fini(struct amdgpu_device *adev, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_kiq *kiq = &adev->gfx.kiq[0];
 | 
						struct amdgpu_kiq *kiq = &adev->gfx.kiq[xcc_id];
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	amdgpu_bo_free_kernel(&kiq->eop_obj, &kiq->eop_gpu_addr, NULL);
 | 
						amdgpu_bo_free_kernel(&kiq->eop_obj, &kiq->eop_gpu_addr, NULL);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int amdgpu_gfx_kiq_init(struct amdgpu_device *adev,
 | 
					int amdgpu_gfx_kiq_init(struct amdgpu_device *adev,
 | 
				
			||||||
			unsigned hpd_size)
 | 
								unsigned hpd_size, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	int r;
 | 
						int r;
 | 
				
			||||||
	u32 *hpd;
 | 
						u32 *hpd;
 | 
				
			||||||
	struct amdgpu_kiq *kiq = &adev->gfx.kiq[0];
 | 
						struct amdgpu_kiq *kiq = &adev->gfx.kiq[xcc_id];
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_bo_create_kernel(adev, hpd_size, PAGE_SIZE,
 | 
						r = amdgpu_bo_create_kernel(adev, hpd_size, PAGE_SIZE,
 | 
				
			||||||
				    AMDGPU_GEM_DOMAIN_GTT, &kiq->eop_obj,
 | 
									    AMDGPU_GEM_DOMAIN_GTT, &kiq->eop_obj,
 | 
				
			||||||
| 
						 | 
					@ -371,13 +372,13 @@ int amdgpu_gfx_kiq_init(struct amdgpu_device *adev,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* create MQD for each compute/gfx queue */
 | 
					/* create MQD for each compute/gfx queue */
 | 
				
			||||||
int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
					int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
				
			||||||
			   unsigned mqd_size)
 | 
								   unsigned mqd_size, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_ring *ring = NULL;
 | 
					 | 
				
			||||||
	int r, i;
 | 
						int r, i;
 | 
				
			||||||
 | 
						struct amdgpu_kiq *kiq = &adev->gfx.kiq[xcc_id];
 | 
				
			||||||
 | 
						struct amdgpu_ring *ring = &kiq->ring;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* create MQD for KIQ */
 | 
						/* create MQD for KIQ */
 | 
				
			||||||
	ring = &adev->gfx.kiq[0].ring;
 | 
					 | 
				
			||||||
	if (!adev->enable_mes_kiq && !ring->mqd_obj) {
 | 
						if (!adev->enable_mes_kiq && !ring->mqd_obj) {
 | 
				
			||||||
		/* originaly the KIQ MQD is put in GTT domain, but for SRIOV VRAM domain is a must
 | 
							/* originaly the KIQ MQD is put in GTT domain, but for SRIOV VRAM domain is a must
 | 
				
			||||||
		 * otherwise hypervisor trigger SAVE_VF fail after driver unloaded which mean MQD
 | 
							 * otherwise hypervisor trigger SAVE_VF fail after driver unloaded which mean MQD
 | 
				
			||||||
| 
						 | 
					@ -396,8 +397,8 @@ int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* prepare MQD backup */
 | 
							/* prepare MQD backup */
 | 
				
			||||||
		adev->gfx.mec.mqd_backup[AMDGPU_MAX_COMPUTE_RINGS] = kmalloc(mqd_size, GFP_KERNEL);
 | 
							kiq->mqd_backup = kmalloc(mqd_size, GFP_KERNEL);
 | 
				
			||||||
		if (!adev->gfx.mec.mqd_backup[AMDGPU_MAX_COMPUTE_RINGS])
 | 
							if (!kiq->mqd_backup)
 | 
				
			||||||
				dev_warn(adev->dev, "no memory to create MQD backup for ring %s\n", ring->name);
 | 
									dev_warn(adev->dev, "no memory to create MQD backup for ring %s\n", ring->name);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -424,7 +425,7 @@ int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* create MQD for each KCQ */
 | 
						/* create MQD for each KCQ */
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++) {
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++) {
 | 
				
			||||||
		ring = &adev->gfx.compute_ring[i];
 | 
							ring = &adev->gfx.compute_ring[i + xcc_id * adev->gfx.num_compute_rings];
 | 
				
			||||||
		if (!ring->mqd_obj) {
 | 
							if (!ring->mqd_obj) {
 | 
				
			||||||
			r = amdgpu_bo_create_kernel(adev, mqd_size, PAGE_SIZE,
 | 
								r = amdgpu_bo_create_kernel(adev, mqd_size, PAGE_SIZE,
 | 
				
			||||||
						    AMDGPU_GEM_DOMAIN_GTT, &ring->mqd_obj,
 | 
											    AMDGPU_GEM_DOMAIN_GTT, &ring->mqd_obj,
 | 
				
			||||||
| 
						 | 
					@ -435,7 +436,7 @@ int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			/* prepare MQD backup */
 | 
								/* prepare MQD backup */
 | 
				
			||||||
			adev->gfx.mec.mqd_backup[i] = kmalloc(mqd_size, GFP_KERNEL);
 | 
								adev->gfx.mec.mqd_backup[i + xcc_id * adev->gfx.num_compute_rings] = kmalloc(mqd_size, GFP_KERNEL);
 | 
				
			||||||
			if (!adev->gfx.mec.mqd_backup[i])
 | 
								if (!adev->gfx.mec.mqd_backup[i])
 | 
				
			||||||
				dev_warn(adev->dev, "no memory to create MQD backup for ring %s\n", ring->name);
 | 
									dev_warn(adev->dev, "no memory to create MQD backup for ring %s\n", ring->name);
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
| 
						 | 
					@ -444,10 +445,11 @@ int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void amdgpu_gfx_mqd_sw_fini(struct amdgpu_device *adev)
 | 
					void amdgpu_gfx_mqd_sw_fini(struct amdgpu_device *adev, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_ring *ring = NULL;
 | 
						struct amdgpu_ring *ring = NULL;
 | 
				
			||||||
	int i;
 | 
						int i, j;
 | 
				
			||||||
 | 
						struct amdgpu_kiq *kiq = &adev->gfx.kiq[xcc_id];
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (adev->asic_type >= CHIP_NAVI10 && amdgpu_async_gfx_ring) {
 | 
						if (adev->asic_type >= CHIP_NAVI10 && amdgpu_async_gfx_ring) {
 | 
				
			||||||
		for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
 | 
							for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
 | 
				
			||||||
| 
						 | 
					@ -460,6 +462,7 @@ void amdgpu_gfx_mqd_sw_fini(struct amdgpu_device *adev)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++) {
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++) {
 | 
				
			||||||
 | 
							j = i + xcc_id * adev->gfx.num_compute_rings;
 | 
				
			||||||
		ring = &adev->gfx.compute_ring[i];
 | 
							ring = &adev->gfx.compute_ring[i];
 | 
				
			||||||
		kfree(adev->gfx.mec.mqd_backup[i]);
 | 
							kfree(adev->gfx.mec.mqd_backup[i]);
 | 
				
			||||||
		amdgpu_bo_free_kernel(&ring->mqd_obj,
 | 
							amdgpu_bo_free_kernel(&ring->mqd_obj,
 | 
				
			||||||
| 
						 | 
					@ -467,36 +470,40 @@ void amdgpu_gfx_mqd_sw_fini(struct amdgpu_device *adev)
 | 
				
			||||||
				      &ring->mqd_ptr);
 | 
									      &ring->mqd_ptr);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	ring = &adev->gfx.kiq[0].ring;
 | 
						ring = &kiq->ring;
 | 
				
			||||||
 | 
						kfree(kiq->mqd_backup);
 | 
				
			||||||
	kfree(adev->gfx.mec.mqd_backup[AMDGPU_MAX_COMPUTE_RINGS]);
 | 
						kfree(adev->gfx.mec.mqd_backup[AMDGPU_MAX_COMPUTE_RINGS]);
 | 
				
			||||||
	amdgpu_bo_free_kernel(&ring->mqd_obj,
 | 
						amdgpu_bo_free_kernel(&ring->mqd_obj,
 | 
				
			||||||
			      &ring->mqd_gpu_addr,
 | 
								      &ring->mqd_gpu_addr,
 | 
				
			||||||
			      &ring->mqd_ptr);
 | 
								      &ring->mqd_ptr);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int amdgpu_gfx_disable_kcq(struct amdgpu_device *adev)
 | 
					int amdgpu_gfx_disable_kcq(struct amdgpu_device *adev, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_kiq *kiq = &adev->gfx.kiq[0];
 | 
						struct amdgpu_kiq *kiq = &adev->gfx.kiq[xcc_id];
 | 
				
			||||||
	struct amdgpu_ring *kiq_ring = &kiq->ring;
 | 
						struct amdgpu_ring *kiq_ring = &kiq->ring;
 | 
				
			||||||
	int i, r = 0;
 | 
						int i, r = 0;
 | 
				
			||||||
 | 
						int j;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!kiq->pmf || !kiq->pmf->kiq_unmap_queues)
 | 
						if (!kiq->pmf || !kiq->pmf->kiq_unmap_queues)
 | 
				
			||||||
		return -EINVAL;
 | 
							return -EINVAL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	spin_lock(&adev->gfx.kiq[0].ring_lock);
 | 
						spin_lock(&kiq->ring_lock);
 | 
				
			||||||
	if (amdgpu_ring_alloc(kiq_ring, kiq->pmf->unmap_queues_size *
 | 
						if (amdgpu_ring_alloc(kiq_ring, kiq->pmf->unmap_queues_size *
 | 
				
			||||||
					adev->gfx.num_compute_rings)) {
 | 
										adev->gfx.num_compute_rings)) {
 | 
				
			||||||
		spin_unlock(&adev->gfx.kiq[0].ring_lock);
 | 
							spin_unlock(&adev->gfx.kiq[0].ring_lock);
 | 
				
			||||||
		return -ENOMEM;
 | 
							return -ENOMEM;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++) {
 | 
				
			||||||
 | 
							j = i + xcc_id * adev->gfx.num_compute_rings;
 | 
				
			||||||
		kiq->pmf->kiq_unmap_queues(kiq_ring, &adev->gfx.compute_ring[i],
 | 
							kiq->pmf->kiq_unmap_queues(kiq_ring, &adev->gfx.compute_ring[i],
 | 
				
			||||||
					   RESET_QUEUES, 0, 0);
 | 
										   RESET_QUEUES, 0, 0);
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (adev->gfx.kiq[0].ring.sched.ready && !adev->job_hang)
 | 
						if (adev->gfx.kiq[0].ring.sched.ready && !adev->job_hang)
 | 
				
			||||||
		r = amdgpu_ring_test_helper(kiq_ring);
 | 
							r = amdgpu_ring_test_helper(kiq_ring);
 | 
				
			||||||
	spin_unlock(&adev->gfx.kiq[0].ring_lock);
 | 
						spin_unlock(&kiq->ring_lock);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return r;
 | 
						return r;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -514,18 +521,18 @@ int amdgpu_queue_mask_bit_to_set_resource_bit(struct amdgpu_device *adev,
 | 
				
			||||||
	return set_resource_bit;
 | 
						return set_resource_bit;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int amdgpu_gfx_enable_kcq(struct amdgpu_device *adev)
 | 
					int amdgpu_gfx_enable_kcq(struct amdgpu_device *adev, int xcc_id)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_kiq *kiq = &adev->gfx.kiq[0];
 | 
						struct amdgpu_kiq *kiq = &adev->gfx.kiq[xcc_id];
 | 
				
			||||||
	struct amdgpu_ring *kiq_ring = &adev->gfx.kiq[0].ring;
 | 
						struct amdgpu_ring *kiq_ring = &kiq->ring;
 | 
				
			||||||
	uint64_t queue_mask = 0;
 | 
						uint64_t queue_mask = 0;
 | 
				
			||||||
	int r, i;
 | 
						int r, i, j;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!kiq->pmf || !kiq->pmf->kiq_map_queues || !kiq->pmf->kiq_set_resources)
 | 
						if (!kiq->pmf || !kiq->pmf->kiq_map_queues || !kiq->pmf->kiq_set_resources)
 | 
				
			||||||
		return -EINVAL;
 | 
							return -EINVAL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < AMDGPU_MAX_COMPUTE_QUEUES; ++i) {
 | 
						for (i = 0; i < AMDGPU_MAX_COMPUTE_QUEUES; ++i) {
 | 
				
			||||||
		if (!test_bit(i, adev->gfx.mec_bitmap[0].queue_bitmap))
 | 
							if (!test_bit(i, adev->gfx.mec_bitmap[xcc_id].queue_bitmap))
 | 
				
			||||||
			continue;
 | 
								continue;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* This situation may be hit in the future if a new HW
 | 
							/* This situation may be hit in the future if a new HW
 | 
				
			||||||
| 
						 | 
					@ -541,7 +548,7 @@ int amdgpu_gfx_enable_kcq(struct amdgpu_device *adev)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	DRM_INFO("kiq ring mec %d pipe %d q %d\n", kiq_ring->me, kiq_ring->pipe,
 | 
						DRM_INFO("kiq ring mec %d pipe %d q %d\n", kiq_ring->me, kiq_ring->pipe,
 | 
				
			||||||
							kiq_ring->queue);
 | 
												kiq_ring->queue);
 | 
				
			||||||
	spin_lock(&adev->gfx.kiq[0].ring_lock);
 | 
						spin_lock(&kiq->ring_lock);
 | 
				
			||||||
	r = amdgpu_ring_alloc(kiq_ring, kiq->pmf->map_queues_size *
 | 
						r = amdgpu_ring_alloc(kiq_ring, kiq->pmf->map_queues_size *
 | 
				
			||||||
					adev->gfx.num_compute_rings +
 | 
										adev->gfx.num_compute_rings +
 | 
				
			||||||
					kiq->pmf->set_resources_size);
 | 
										kiq->pmf->set_resources_size);
 | 
				
			||||||
| 
						 | 
					@ -555,11 +562,13 @@ int amdgpu_gfx_enable_kcq(struct amdgpu_device *adev)
 | 
				
			||||||
		queue_mask = ~0ULL;
 | 
							queue_mask = ~0ULL;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	kiq->pmf->kiq_set_resources(kiq_ring, queue_mask);
 | 
						kiq->pmf->kiq_set_resources(kiq_ring, queue_mask);
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++) {
 | 
				
			||||||
 | 
							j = i + xcc_id * adev->gfx.num_compute_rings;
 | 
				
			||||||
		kiq->pmf->kiq_map_queues(kiq_ring, &adev->gfx.compute_ring[i]);
 | 
							kiq->pmf->kiq_map_queues(kiq_ring, &adev->gfx.compute_ring[i]);
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_ring_test_helper(kiq_ring);
 | 
						r = amdgpu_ring_test_helper(kiq_ring);
 | 
				
			||||||
	spin_unlock(&adev->gfx.kiq[0].ring_lock);
 | 
						spin_unlock(&kiq->ring_lock);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		DRM_ERROR("KCQ enable failed\n");
 | 
							DRM_ERROR("KCQ enable failed\n");
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -408,19 +408,19 @@ void amdgpu_gfx_parse_disable_cu(unsigned *mask, unsigned max_se,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int amdgpu_gfx_kiq_init_ring(struct amdgpu_device *adev,
 | 
					int amdgpu_gfx_kiq_init_ring(struct amdgpu_device *adev,
 | 
				
			||||||
			     struct amdgpu_ring *ring,
 | 
								     struct amdgpu_ring *ring,
 | 
				
			||||||
			     struct amdgpu_irq_src *irq);
 | 
								     struct amdgpu_irq_src *irq, int xcc_id);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void amdgpu_gfx_kiq_free_ring(struct amdgpu_ring *ring);
 | 
					void amdgpu_gfx_kiq_free_ring(struct amdgpu_ring *ring);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void amdgpu_gfx_kiq_fini(struct amdgpu_device *adev);
 | 
					void amdgpu_gfx_kiq_fini(struct amdgpu_device *adev, int xcc_id);
 | 
				
			||||||
int amdgpu_gfx_kiq_init(struct amdgpu_device *adev,
 | 
					int amdgpu_gfx_kiq_init(struct amdgpu_device *adev,
 | 
				
			||||||
			unsigned hpd_size);
 | 
								unsigned hpd_size, int xcc_id);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
					int amdgpu_gfx_mqd_sw_init(struct amdgpu_device *adev,
 | 
				
			||||||
			   unsigned mqd_size);
 | 
								   unsigned mqd_size, int xcc_id);
 | 
				
			||||||
void amdgpu_gfx_mqd_sw_fini(struct amdgpu_device *adev);
 | 
					void amdgpu_gfx_mqd_sw_fini(struct amdgpu_device *adev, int xcc_id);
 | 
				
			||||||
int amdgpu_gfx_disable_kcq(struct amdgpu_device *adev);
 | 
					int amdgpu_gfx_disable_kcq(struct amdgpu_device *adev, int xcc_id);
 | 
				
			||||||
int amdgpu_gfx_enable_kcq(struct amdgpu_device *adev);
 | 
					int amdgpu_gfx_enable_kcq(struct amdgpu_device *adev, int xcc_id);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
void amdgpu_gfx_compute_queue_acquire(struct amdgpu_device *adev);
 | 
					void amdgpu_gfx_compute_queue_acquire(struct amdgpu_device *adev);
 | 
				
			||||||
void amdgpu_gfx_graphics_queue_acquire(struct amdgpu_device *adev);
 | 
					void amdgpu_gfx_graphics_queue_acquire(struct amdgpu_device *adev);
 | 
				
			||||||
| 
						 | 
					@ -429,7 +429,7 @@ int amdgpu_gfx_mec_queue_to_bit(struct amdgpu_device *adev, int mec,
 | 
				
			||||||
				int pipe, int queue);
 | 
									int pipe, int queue);
 | 
				
			||||||
void amdgpu_queue_mask_bit_to_mec_queue(struct amdgpu_device *adev, int bit,
 | 
					void amdgpu_queue_mask_bit_to_mec_queue(struct amdgpu_device *adev, int bit,
 | 
				
			||||||
				 int *mec, int *pipe, int *queue);
 | 
									 int *mec, int *pipe, int *queue);
 | 
				
			||||||
bool amdgpu_gfx_is_mec_queue_enabled(struct amdgpu_device *adev, int inst,
 | 
					bool amdgpu_gfx_is_mec_queue_enabled(struct amdgpu_device *adev, int xcc_id,
 | 
				
			||||||
				     int mec, int pipe, int queue);
 | 
									     int mec, int pipe, int queue);
 | 
				
			||||||
bool amdgpu_gfx_is_high_priority_compute_queue(struct amdgpu_device *adev,
 | 
					bool amdgpu_gfx_is_high_priority_compute_queue(struct amdgpu_device *adev,
 | 
				
			||||||
					       struct amdgpu_ring *ring);
 | 
										       struct amdgpu_ring *ring);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -4629,19 +4629,19 @@ static int gfx_v10_0_sw_init(void *handle)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!adev->enable_mes_kiq) {
 | 
						if (!adev->enable_mes_kiq) {
 | 
				
			||||||
		r = amdgpu_gfx_kiq_init(adev, GFX10_MEC_HPD_SIZE);
 | 
							r = amdgpu_gfx_kiq_init(adev, GFX10_MEC_HPD_SIZE, 0);
 | 
				
			||||||
		if (r) {
 | 
							if (r) {
 | 
				
			||||||
			DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
								DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
				
			||||||
			return r;
 | 
								return r;
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		kiq = &adev->gfx.kiq[0];
 | 
							kiq = &adev->gfx.kiq[0];
 | 
				
			||||||
		r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq);
 | 
							r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq, 0);
 | 
				
			||||||
		if (r)
 | 
							if (r)
 | 
				
			||||||
			return r;
 | 
								return r;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct v10_compute_mqd));
 | 
						r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct v10_compute_mqd), 0);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -4690,11 +4690,11 @@ static int gfx_v10_0_sw_fini(void *handle)
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
				
			||||||
		amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
							amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	amdgpu_gfx_mqd_sw_fini(adev);
 | 
						amdgpu_gfx_mqd_sw_fini(adev, 0);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!adev->enable_mes_kiq) {
 | 
						if (!adev->enable_mes_kiq) {
 | 
				
			||||||
		amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
							amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
				
			||||||
		amdgpu_gfx_kiq_fini(adev);
 | 
							amdgpu_gfx_kiq_fini(adev, 0);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v10_0_pfp_fini(adev);
 | 
						gfx_v10_0_pfp_fini(adev);
 | 
				
			||||||
| 
						 | 
					@ -6812,14 +6812,13 @@ static int gfx_v10_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_device *adev = ring->adev;
 | 
						struct amdgpu_device *adev = ring->adev;
 | 
				
			||||||
	struct v10_compute_mqd *mqd = ring->mqd_ptr;
 | 
						struct v10_compute_mqd *mqd = ring->mqd_ptr;
 | 
				
			||||||
	int mqd_idx = AMDGPU_MAX_COMPUTE_RINGS;
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v10_0_kiq_setting(ring);
 | 
						gfx_v10_0_kiq_setting(ring);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (amdgpu_in_reset(adev)) { /* for GPU_RESET case */
 | 
						if (amdgpu_in_reset(adev)) { /* for GPU_RESET case */
 | 
				
			||||||
		/* reset MQD to a clean status */
 | 
							/* reset MQD to a clean status */
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(*mqd));
 | 
								memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(*mqd));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* reset ring buffer */
 | 
							/* reset ring buffer */
 | 
				
			||||||
		ring->wptr = 0;
 | 
							ring->wptr = 0;
 | 
				
			||||||
| 
						 | 
					@ -6841,8 +6840,8 @@ static int gfx_v10_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
		nv_grbm_select(adev, 0, 0, 0, 0);
 | 
							nv_grbm_select(adev, 0, 0, 0, 0);
 | 
				
			||||||
		mutex_unlock(&adev->srbm_mutex);
 | 
							mutex_unlock(&adev->srbm_mutex);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd));
 | 
								memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(*mqd));
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
| 
						 | 
					@ -6927,7 +6926,7 @@ static int gfx_v10_0_kcq_resume(struct amdgpu_device *adev)
 | 
				
			||||||
			goto done;
 | 
								goto done;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_enable_kcq(adev);
 | 
						r = amdgpu_gfx_enable_kcq(adev, 0);
 | 
				
			||||||
done:
 | 
					done:
 | 
				
			||||||
	return r;
 | 
						return r;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -7280,7 +7279,7 @@ static int gfx_v10_0_hw_fini(void *handle)
 | 
				
			||||||
				DRM_ERROR("KGQ disable failed\n");
 | 
									DRM_ERROR("KGQ disable failed\n");
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
		if (amdgpu_gfx_disable_kcq(adev))
 | 
							if (amdgpu_gfx_disable_kcq(adev, 0))
 | 
				
			||||||
			DRM_ERROR("KCQ disable failed\n");
 | 
								DRM_ERROR("KCQ disable failed\n");
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1389,19 +1389,19 @@ static int gfx_v11_0_sw_init(void *handle)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!adev->enable_mes_kiq) {
 | 
						if (!adev->enable_mes_kiq) {
 | 
				
			||||||
		r = amdgpu_gfx_kiq_init(adev, GFX11_MEC_HPD_SIZE);
 | 
							r = amdgpu_gfx_kiq_init(adev, GFX11_MEC_HPD_SIZE, 0);
 | 
				
			||||||
		if (r) {
 | 
							if (r) {
 | 
				
			||||||
			DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
								DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
				
			||||||
			return r;
 | 
								return r;
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		kiq = &adev->gfx.kiq[0];
 | 
							kiq = &adev->gfx.kiq[0];
 | 
				
			||||||
		r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq);
 | 
							r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq, 0);
 | 
				
			||||||
		if (r)
 | 
							if (r)
 | 
				
			||||||
			return r;
 | 
								return r;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct v11_compute_mqd));
 | 
						r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct v11_compute_mqd), 0);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1463,11 +1463,11 @@ static int gfx_v11_0_sw_fini(void *handle)
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
				
			||||||
		amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
							amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	amdgpu_gfx_mqd_sw_fini(adev);
 | 
						amdgpu_gfx_mqd_sw_fini(adev, 0);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!adev->enable_mes_kiq) {
 | 
						if (!adev->enable_mes_kiq) {
 | 
				
			||||||
		amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
							amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
				
			||||||
		amdgpu_gfx_kiq_fini(adev);
 | 
							amdgpu_gfx_kiq_fini(adev, 0);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v11_0_pfp_fini(adev);
 | 
						gfx_v11_0_pfp_fini(adev);
 | 
				
			||||||
| 
						 | 
					@ -4035,14 +4035,13 @@ static int gfx_v11_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_device *adev = ring->adev;
 | 
						struct amdgpu_device *adev = ring->adev;
 | 
				
			||||||
	struct v11_compute_mqd *mqd = ring->mqd_ptr;
 | 
						struct v11_compute_mqd *mqd = ring->mqd_ptr;
 | 
				
			||||||
	int mqd_idx = AMDGPU_MAX_COMPUTE_RINGS;
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v11_0_kiq_setting(ring);
 | 
						gfx_v11_0_kiq_setting(ring);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (amdgpu_in_reset(adev)) { /* for GPU_RESET case */
 | 
						if (amdgpu_in_reset(adev)) { /* for GPU_RESET case */
 | 
				
			||||||
		/* reset MQD to a clean status */
 | 
							/* reset MQD to a clean status */
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(*mqd));
 | 
								memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(*mqd));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* reset ring buffer */
 | 
							/* reset ring buffer */
 | 
				
			||||||
		ring->wptr = 0;
 | 
							ring->wptr = 0;
 | 
				
			||||||
| 
						 | 
					@ -4064,8 +4063,8 @@ static int gfx_v11_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
		soc21_grbm_select(adev, 0, 0, 0, 0);
 | 
							soc21_grbm_select(adev, 0, 0, 0, 0);
 | 
				
			||||||
		mutex_unlock(&adev->srbm_mutex);
 | 
							mutex_unlock(&adev->srbm_mutex);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(*mqd));
 | 
								memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(*mqd));
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
| 
						 | 
					@ -4153,7 +4152,7 @@ static int gfx_v11_0_kcq_resume(struct amdgpu_device *adev)
 | 
				
			||||||
			goto done;
 | 
								goto done;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_enable_kcq(adev);
 | 
						r = amdgpu_gfx_enable_kcq(adev, 0);
 | 
				
			||||||
done:
 | 
					done:
 | 
				
			||||||
	return r;
 | 
						return r;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -4456,7 +4455,7 @@ static int gfx_v11_0_hw_fini(void *handle)
 | 
				
			||||||
				DRM_ERROR("KGQ disable failed\n");
 | 
									DRM_ERROR("KGQ disable failed\n");
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
#endif
 | 
					#endif
 | 
				
			||||||
		if (amdgpu_gfx_disable_kcq(adev))
 | 
							if (amdgpu_gfx_disable_kcq(adev, 0))
 | 
				
			||||||
			DRM_ERROR("KCQ disable failed\n");
 | 
								DRM_ERROR("KCQ disable failed\n");
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		amdgpu_mes_kiq_hw_fini(adev);
 | 
							amdgpu_mes_kiq_hw_fini(adev);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -2016,19 +2016,19 @@ static int gfx_v8_0_sw_init(void *handle)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_kiq_init(adev, GFX8_MEC_HPD_SIZE);
 | 
						r = amdgpu_gfx_kiq_init(adev, GFX8_MEC_HPD_SIZE, 0);
 | 
				
			||||||
	if (r) {
 | 
						if (r) {
 | 
				
			||||||
		DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
							DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	kiq = &adev->gfx.kiq[0];
 | 
						kiq = &adev->gfx.kiq[0];
 | 
				
			||||||
	r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq);
 | 
						r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq, 0);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* create MQD for all compute queues as well as KIQ for SRIOV case */
 | 
						/* create MQD for all compute queues as well as KIQ for SRIOV case */
 | 
				
			||||||
	r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct vi_mqd_allocation));
 | 
						r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct vi_mqd_allocation), 0);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -2051,9 +2051,9 @@ static int gfx_v8_0_sw_fini(void *handle)
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
				
			||||||
		amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
							amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	amdgpu_gfx_mqd_sw_fini(adev);
 | 
						amdgpu_gfx_mqd_sw_fini(adev, 0);
 | 
				
			||||||
	amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
						amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
				
			||||||
	amdgpu_gfx_kiq_fini(adev);
 | 
						amdgpu_gfx_kiq_fini(adev, 0);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v8_0_mec_fini(adev);
 | 
						gfx_v8_0_mec_fini(adev);
 | 
				
			||||||
	amdgpu_gfx_rlc_fini(adev);
 | 
						amdgpu_gfx_rlc_fini(adev);
 | 
				
			||||||
| 
						 | 
					@ -4596,14 +4596,13 @@ static int gfx_v8_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_device *adev = ring->adev;
 | 
						struct amdgpu_device *adev = ring->adev;
 | 
				
			||||||
	struct vi_mqd *mqd = ring->mqd_ptr;
 | 
						struct vi_mqd *mqd = ring->mqd_ptr;
 | 
				
			||||||
	int mqd_idx = AMDGPU_MAX_COMPUTE_RINGS;
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v8_0_kiq_setting(ring);
 | 
						gfx_v8_0_kiq_setting(ring);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (amdgpu_in_reset(adev)) { /* for GPU_RESET case */
 | 
						if (amdgpu_in_reset(adev)) { /* for GPU_RESET case */
 | 
				
			||||||
		/* reset MQD to a clean status */
 | 
							/* reset MQD to a clean status */
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct vi_mqd_allocation));
 | 
								memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(struct vi_mqd_allocation));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* reset ring buffer */
 | 
							/* reset ring buffer */
 | 
				
			||||||
		ring->wptr = 0;
 | 
							ring->wptr = 0;
 | 
				
			||||||
| 
						 | 
					@ -4626,8 +4625,8 @@ static int gfx_v8_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
		vi_srbm_select(adev, 0, 0, 0, 0);
 | 
							vi_srbm_select(adev, 0, 0, 0, 0);
 | 
				
			||||||
		mutex_unlock(&adev->srbm_mutex);
 | 
							mutex_unlock(&adev->srbm_mutex);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct vi_mqd_allocation));
 | 
								memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(struct vi_mqd_allocation));
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -2169,19 +2169,19 @@ static int gfx_v9_0_sw_init(void *handle)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_kiq_init(adev, GFX9_MEC_HPD_SIZE);
 | 
						r = amdgpu_gfx_kiq_init(adev, GFX9_MEC_HPD_SIZE, 0);
 | 
				
			||||||
	if (r) {
 | 
						if (r) {
 | 
				
			||||||
		DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
							DRM_ERROR("Failed to init KIQ BOs!\n");
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	kiq = &adev->gfx.kiq[0];
 | 
						kiq = &adev->gfx.kiq[0];
 | 
				
			||||||
	r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq);
 | 
						r = amdgpu_gfx_kiq_init_ring(adev, &kiq->ring, &kiq->irq, 0);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* create MQD for all compute queues as wel as KIQ for SRIOV case */
 | 
						/* create MQD for all compute queues as wel as KIQ for SRIOV case */
 | 
				
			||||||
	r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct v9_mqd_allocation));
 | 
						r = amdgpu_gfx_mqd_sw_init(adev, sizeof(struct v9_mqd_allocation), 0);
 | 
				
			||||||
	if (r)
 | 
						if (r)
 | 
				
			||||||
		return r;
 | 
							return r;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -2216,9 +2216,9 @@ static int gfx_v9_0_sw_fini(void *handle)
 | 
				
			||||||
	for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
						for (i = 0; i < adev->gfx.num_compute_rings; i++)
 | 
				
			||||||
		amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
							amdgpu_ring_fini(&adev->gfx.compute_ring[i]);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	amdgpu_gfx_mqd_sw_fini(adev);
 | 
						amdgpu_gfx_mqd_sw_fini(adev, 0);
 | 
				
			||||||
	amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
						amdgpu_gfx_kiq_free_ring(&adev->gfx.kiq[0].ring);
 | 
				
			||||||
	amdgpu_gfx_kiq_fini(adev);
 | 
						amdgpu_gfx_kiq_fini(adev, 0);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v9_0_mec_fini(adev);
 | 
						gfx_v9_0_mec_fini(adev);
 | 
				
			||||||
	amdgpu_bo_free_kernel(&adev->gfx.rlc.clear_state_obj,
 | 
						amdgpu_bo_free_kernel(&adev->gfx.rlc.clear_state_obj,
 | 
				
			||||||
| 
						 | 
					@ -3520,7 +3520,6 @@ static int gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct amdgpu_device *adev = ring->adev;
 | 
						struct amdgpu_device *adev = ring->adev;
 | 
				
			||||||
	struct v9_mqd *mqd = ring->mqd_ptr;
 | 
						struct v9_mqd *mqd = ring->mqd_ptr;
 | 
				
			||||||
	int mqd_idx = AMDGPU_MAX_COMPUTE_RINGS;
 | 
					 | 
				
			||||||
	struct v9_mqd *tmp_mqd;
 | 
						struct v9_mqd *tmp_mqd;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	gfx_v9_0_kiq_setting(ring);
 | 
						gfx_v9_0_kiq_setting(ring);
 | 
				
			||||||
| 
						 | 
					@ -3530,11 +3529,11 @@ static int gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
	 * driver need to re-init the mqd.
 | 
						 * driver need to re-init the mqd.
 | 
				
			||||||
	 * check mqd->cp_hqd_pq_control since this value should not be 0
 | 
						 * check mqd->cp_hqd_pq_control since this value should not be 0
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	tmp_mqd = (struct v9_mqd *)adev->gfx.mec.mqd_backup[mqd_idx];
 | 
						tmp_mqd = (struct v9_mqd *)adev->gfx.kiq[0].mqd_backup;
 | 
				
			||||||
	if (amdgpu_in_reset(adev) && tmp_mqd->cp_hqd_pq_control){
 | 
						if (amdgpu_in_reset(adev) && tmp_mqd->cp_hqd_pq_control){
 | 
				
			||||||
		/* for GPU_RESET case , reset MQD to a clean status */
 | 
							/* for GPU_RESET case , reset MQD to a clean status */
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(mqd, adev->gfx.mec.mqd_backup[mqd_idx], sizeof(struct v9_mqd_allocation));
 | 
								memcpy(mqd, adev->gfx.kiq[0].mqd_backup, sizeof(struct v9_mqd_allocation));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* reset ring buffer */
 | 
							/* reset ring buffer */
 | 
				
			||||||
		ring->wptr = 0;
 | 
							ring->wptr = 0;
 | 
				
			||||||
| 
						 | 
					@ -3558,8 +3557,8 @@ static int gfx_v9_0_kiq_init_queue(struct amdgpu_ring *ring)
 | 
				
			||||||
		soc15_grbm_select(adev, 0, 0, 0, 0);
 | 
							soc15_grbm_select(adev, 0, 0, 0, 0);
 | 
				
			||||||
		mutex_unlock(&adev->srbm_mutex);
 | 
							mutex_unlock(&adev->srbm_mutex);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		if (adev->gfx.mec.mqd_backup[mqd_idx])
 | 
							if (adev->gfx.kiq[0].mqd_backup)
 | 
				
			||||||
			memcpy(adev->gfx.mec.mqd_backup[mqd_idx], mqd, sizeof(struct v9_mqd_allocation));
 | 
								memcpy(adev->gfx.kiq[0].mqd_backup, mqd, sizeof(struct v9_mqd_allocation));
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
| 
						 | 
					@ -3653,7 +3652,7 @@ static int gfx_v9_0_kcq_resume(struct amdgpu_device *adev)
 | 
				
			||||||
			goto done;
 | 
								goto done;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	r = amdgpu_gfx_enable_kcq(adev);
 | 
						r = amdgpu_gfx_enable_kcq(adev, 0);
 | 
				
			||||||
done:
 | 
					done:
 | 
				
			||||||
	return r;
 | 
						return r;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					@ -3772,7 +3771,7 @@ static int gfx_v9_0_hw_fini(void *handle)
 | 
				
			||||||
	/* DF freeze and kcq disable will fail */
 | 
						/* DF freeze and kcq disable will fail */
 | 
				
			||||||
	if (!amdgpu_ras_intr_triggered())
 | 
						if (!amdgpu_ras_intr_triggered())
 | 
				
			||||||
		/* disable KCQ to avoid CPC touch memory not valid anymore */
 | 
							/* disable KCQ to avoid CPC touch memory not valid anymore */
 | 
				
			||||||
		amdgpu_gfx_disable_kcq(adev);
 | 
							amdgpu_gfx_disable_kcq(adev, 0);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (amdgpu_sriov_vf(adev)) {
 | 
						if (amdgpu_sriov_vf(adev)) {
 | 
				
			||||||
		gfx_v9_0_cp_gfx_enable(adev, false);
 | 
							gfx_v9_0_cp_gfx_enable(adev, false);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue