drm/radeon: allocate semaphore from the ib pool
This allow to share the ib pool with semaphore and avoid having more bo around. Signed-off-by: Jerome Glisse <jglisse@redhat.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com> Signed-off-by: Dave Airlie <airlied@redhat.com>
This commit is contained in:
committed by
Dave Airlie
parent
69e130a6a4
commit
c1341e5280
@@ -242,32 +242,6 @@ struct radeon_fence *radeon_fence_ref(struct radeon_fence *fence);
|
|||||||
void radeon_fence_unref(struct radeon_fence **fence);
|
void radeon_fence_unref(struct radeon_fence **fence);
|
||||||
int radeon_fence_count_emitted(struct radeon_device *rdev, int ring);
|
int radeon_fence_count_emitted(struct radeon_device *rdev, int ring);
|
||||||
|
|
||||||
/*
|
|
||||||
* Semaphores.
|
|
||||||
*/
|
|
||||||
struct radeon_ring;
|
|
||||||
|
|
||||||
struct radeon_semaphore_driver {
|
|
||||||
rwlock_t lock;
|
|
||||||
struct list_head free;
|
|
||||||
};
|
|
||||||
|
|
||||||
struct radeon_semaphore {
|
|
||||||
struct radeon_bo *robj;
|
|
||||||
struct list_head list;
|
|
||||||
uint64_t gpu_addr;
|
|
||||||
};
|
|
||||||
|
|
||||||
void radeon_semaphore_driver_fini(struct radeon_device *rdev);
|
|
||||||
int radeon_semaphore_create(struct radeon_device *rdev,
|
|
||||||
struct radeon_semaphore **semaphore);
|
|
||||||
void radeon_semaphore_emit_signal(struct radeon_device *rdev, int ring,
|
|
||||||
struct radeon_semaphore *semaphore);
|
|
||||||
void radeon_semaphore_emit_wait(struct radeon_device *rdev, int ring,
|
|
||||||
struct radeon_semaphore *semaphore);
|
|
||||||
void radeon_semaphore_free(struct radeon_device *rdev,
|
|
||||||
struct radeon_semaphore *semaphore);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Tiling registers
|
* Tiling registers
|
||||||
*/
|
*/
|
||||||
@@ -386,6 +360,46 @@ int radeon_mode_dumb_destroy(struct drm_file *file_priv,
|
|||||||
struct drm_device *dev,
|
struct drm_device *dev,
|
||||||
uint32_t handle);
|
uint32_t handle);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Semaphores.
|
||||||
|
*/
|
||||||
|
struct radeon_ring;
|
||||||
|
|
||||||
|
#define RADEON_SEMAPHORE_BO_SIZE 256
|
||||||
|
|
||||||
|
struct radeon_semaphore_driver {
|
||||||
|
rwlock_t lock;
|
||||||
|
struct list_head bo;
|
||||||
|
};
|
||||||
|
|
||||||
|
struct radeon_semaphore_bo;
|
||||||
|
|
||||||
|
/* everything here is constant */
|
||||||
|
struct radeon_semaphore {
|
||||||
|
struct list_head list;
|
||||||
|
uint64_t gpu_addr;
|
||||||
|
uint32_t *cpu_ptr;
|
||||||
|
struct radeon_semaphore_bo *bo;
|
||||||
|
};
|
||||||
|
|
||||||
|
struct radeon_semaphore_bo {
|
||||||
|
struct list_head list;
|
||||||
|
struct radeon_ib *ib;
|
||||||
|
struct list_head free;
|
||||||
|
struct radeon_semaphore semaphores[RADEON_SEMAPHORE_BO_SIZE/8];
|
||||||
|
unsigned nused;
|
||||||
|
};
|
||||||
|
|
||||||
|
void radeon_semaphore_driver_fini(struct radeon_device *rdev);
|
||||||
|
int radeon_semaphore_create(struct radeon_device *rdev,
|
||||||
|
struct radeon_semaphore **semaphore);
|
||||||
|
void radeon_semaphore_emit_signal(struct radeon_device *rdev, int ring,
|
||||||
|
struct radeon_semaphore *semaphore);
|
||||||
|
void radeon_semaphore_emit_wait(struct radeon_device *rdev, int ring,
|
||||||
|
struct radeon_semaphore *semaphore);
|
||||||
|
void radeon_semaphore_free(struct radeon_device *rdev,
|
||||||
|
struct radeon_semaphore *semaphore);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* GART structures, functions & helpers
|
* GART structures, functions & helpers
|
||||||
*/
|
*/
|
||||||
@@ -641,6 +655,7 @@ void r600_blit_suspend(struct radeon_device *rdev);
|
|||||||
int radeon_ib_get(struct radeon_device *rdev, int ring,
|
int radeon_ib_get(struct radeon_device *rdev, int ring,
|
||||||
struct radeon_ib **ib, unsigned size);
|
struct radeon_ib **ib, unsigned size);
|
||||||
void radeon_ib_free(struct radeon_device *rdev, struct radeon_ib **ib);
|
void radeon_ib_free(struct radeon_device *rdev, struct radeon_ib **ib);
|
||||||
|
bool radeon_ib_try_free(struct radeon_device *rdev, struct radeon_ib *ib);
|
||||||
int radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib);
|
int radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib);
|
||||||
int radeon_ib_pool_init(struct radeon_device *rdev);
|
int radeon_ib_pool_init(struct radeon_device *rdev);
|
||||||
void radeon_ib_pool_fini(struct radeon_device *rdev);
|
void radeon_ib_pool_fini(struct radeon_device *rdev);
|
||||||
|
@@ -731,7 +731,7 @@ int radeon_device_init(struct radeon_device *rdev,
|
|||||||
INIT_LIST_HEAD(&rdev->gem.objects);
|
INIT_LIST_HEAD(&rdev->gem.objects);
|
||||||
init_waitqueue_head(&rdev->irq.vblank_queue);
|
init_waitqueue_head(&rdev->irq.vblank_queue);
|
||||||
init_waitqueue_head(&rdev->irq.idle_queue);
|
init_waitqueue_head(&rdev->irq.idle_queue);
|
||||||
INIT_LIST_HEAD(&rdev->semaphore_drv.free);
|
INIT_LIST_HEAD(&rdev->semaphore_drv.bo);
|
||||||
|
|
||||||
/* Set asic functions */
|
/* Set asic functions */
|
||||||
r = radeon_asic_init(rdev);
|
r = radeon_asic_init(rdev);
|
||||||
|
@@ -77,8 +77,7 @@ void radeon_ring_write(struct radeon_ring *ring, uint32_t v)
|
|||||||
/*
|
/*
|
||||||
* IB.
|
* IB.
|
||||||
*/
|
*/
|
||||||
static bool radeon_ib_try_free(struct radeon_device *rdev,
|
bool radeon_ib_try_free(struct radeon_device *rdev, struct radeon_ib *ib)
|
||||||
struct radeon_ib *ib)
|
|
||||||
{
|
{
|
||||||
bool done = false;
|
bool done = false;
|
||||||
|
|
||||||
@@ -148,7 +147,7 @@ retry:
|
|||||||
/* this should be rare event, ie all ib scheduled none signaled yet.
|
/* this should be rare event, ie all ib scheduled none signaled yet.
|
||||||
*/
|
*/
|
||||||
for (i = 0; i < RADEON_IB_POOL_SIZE; i++) {
|
for (i = 0; i < RADEON_IB_POOL_SIZE; i++) {
|
||||||
if (rdev->ib_pool.ibs[idx].fence) {
|
if (rdev->ib_pool.ibs[idx].fence && rdev->ib_pool.ibs[idx].fence->emitted) {
|
||||||
r = radeon_fence_wait(rdev->ib_pool.ibs[idx].fence, false);
|
r = radeon_fence_wait(rdev->ib_pool.ibs[idx].fence, false);
|
||||||
if (!r) {
|
if (!r) {
|
||||||
goto retry;
|
goto retry;
|
||||||
|
@@ -31,90 +31,109 @@
|
|||||||
#include "drm.h"
|
#include "drm.h"
|
||||||
#include "radeon.h"
|
#include "radeon.h"
|
||||||
|
|
||||||
static int allocate_semaphores(struct radeon_device *rdev)
|
static int radeon_semaphore_add_bo(struct radeon_device *rdev)
|
||||||
{
|
{
|
||||||
const unsigned long bo_size = PAGE_SIZE * 4;
|
struct radeon_semaphore_bo *bo;
|
||||||
|
|
||||||
struct radeon_bo *bo;
|
|
||||||
struct list_head new_entrys;
|
|
||||||
unsigned long irq_flags;
|
unsigned long irq_flags;
|
||||||
uint64_t gpu_addr;
|
uint64_t gpu_addr;
|
||||||
void *map;
|
uint32_t *cpu_ptr;
|
||||||
int i, r;
|
int r, i;
|
||||||
|
|
||||||
r = radeon_bo_create(rdev, bo_size, RADEON_GPU_PAGE_SIZE, true,
|
|
||||||
RADEON_GEM_DOMAIN_GTT, &bo);
|
bo = kmalloc(sizeof(struct radeon_semaphore_bo), GFP_KERNEL);
|
||||||
|
if (bo == NULL) {
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
INIT_LIST_HEAD(&bo->free);
|
||||||
|
INIT_LIST_HEAD(&bo->list);
|
||||||
|
bo->nused = 0;
|
||||||
|
|
||||||
|
r = radeon_ib_get(rdev, 0, &bo->ib, RADEON_SEMAPHORE_BO_SIZE);
|
||||||
if (r) {
|
if (r) {
|
||||||
dev_err(rdev->dev, "(%d) failed to allocate semaphore bo\n", r);
|
dev_err(rdev->dev, "failed to get a bo after 5 retry\n");
|
||||||
|
kfree(bo);
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
gpu_addr = rdev->ib_pool.sa_manager.gpu_addr;
|
||||||
r = radeon_bo_reserve(bo, false);
|
gpu_addr += bo->ib->sa_bo.offset;
|
||||||
if (r) {
|
cpu_ptr = rdev->ib_pool.sa_manager.cpu_ptr;
|
||||||
radeon_bo_unref(&bo);
|
cpu_ptr += (bo->ib->sa_bo.offset >> 2);
|
||||||
dev_err(rdev->dev, "(%d) failed to reserve semaphore bo\n", r);
|
for (i = 0; i < (RADEON_SEMAPHORE_BO_SIZE/8); i++) {
|
||||||
return r;
|
bo->semaphores[i].gpu_addr = gpu_addr;
|
||||||
}
|
bo->semaphores[i].cpu_ptr = cpu_ptr;
|
||||||
|
bo->semaphores[i].bo = bo;
|
||||||
r = radeon_bo_kmap(bo, &map);
|
list_add_tail(&bo->semaphores[i].list, &bo->free);
|
||||||
if (r) {
|
|
||||||
radeon_bo_unreserve(bo);
|
|
||||||
radeon_bo_unref(&bo);
|
|
||||||
dev_err(rdev->dev, "(%d) semaphore map failed\n", r);
|
|
||||||
return r;
|
|
||||||
}
|
|
||||||
memset(map, 0, bo_size);
|
|
||||||
radeon_bo_kunmap(bo);
|
|
||||||
|
|
||||||
r = radeon_bo_pin(bo, RADEON_GEM_DOMAIN_VRAM, &gpu_addr);
|
|
||||||
if (r) {
|
|
||||||
radeon_bo_unreserve(bo);
|
|
||||||
radeon_bo_unref(&bo);
|
|
||||||
dev_err(rdev->dev, "(%d) semaphore pin failed\n", r);
|
|
||||||
return r;
|
|
||||||
}
|
|
||||||
|
|
||||||
INIT_LIST_HEAD(&new_entrys);
|
|
||||||
for (i = 0; i < bo_size/8; ++i) {
|
|
||||||
struct radeon_semaphore *sem = kmalloc(sizeof(struct radeon_semaphore), GFP_KERNEL);
|
|
||||||
ttm_bo_reference(&bo->tbo);
|
|
||||||
sem->robj = bo;
|
|
||||||
sem->gpu_addr = gpu_addr;
|
|
||||||
gpu_addr += 8;
|
gpu_addr += 8;
|
||||||
list_add_tail(&sem->list, &new_entrys);
|
cpu_ptr += 2;
|
||||||
|
}
|
||||||
|
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
|
list_add_tail(&bo->list, &rdev->semaphore_drv.bo);
|
||||||
|
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
radeon_bo_unreserve(bo);
|
static void radeon_semaphore_del_bo_locked(struct radeon_device *rdev,
|
||||||
radeon_bo_unref(&bo);
|
struct radeon_semaphore_bo *bo)
|
||||||
|
{
|
||||||
|
radeon_sa_bo_free(rdev, &bo->ib->sa_bo);
|
||||||
|
radeon_fence_unref(&bo->ib->fence);
|
||||||
|
list_del(&bo->list);
|
||||||
|
kfree(bo);
|
||||||
|
}
|
||||||
|
|
||||||
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
void radeon_semaphore_shrink_locked(struct radeon_device *rdev)
|
||||||
list_splice_tail(&new_entrys, &rdev->semaphore_drv.free);
|
{
|
||||||
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
struct radeon_semaphore_bo *bo, *n;
|
||||||
|
|
||||||
DRM_INFO("%d new semaphores allocated\n", (int)(bo_size/8));
|
if (list_empty(&rdev->semaphore_drv.bo)) {
|
||||||
|
return;
|
||||||
return 0;
|
}
|
||||||
|
/* only shrink if first bo has free semaphore */
|
||||||
|
bo = list_first_entry(&rdev->semaphore_drv.bo, struct radeon_semaphore_bo, list);
|
||||||
|
if (list_empty(&bo->free)) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
list_for_each_entry_safe_continue(bo, n, &rdev->semaphore_drv.bo, list) {
|
||||||
|
if (bo->nused)
|
||||||
|
continue;
|
||||||
|
radeon_semaphore_del_bo_locked(rdev, bo);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
int radeon_semaphore_create(struct radeon_device *rdev,
|
int radeon_semaphore_create(struct radeon_device *rdev,
|
||||||
struct radeon_semaphore **semaphore)
|
struct radeon_semaphore **semaphore)
|
||||||
{
|
{
|
||||||
|
struct radeon_semaphore_bo *bo;
|
||||||
unsigned long irq_flags;
|
unsigned long irq_flags;
|
||||||
|
bool do_retry = true;
|
||||||
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
|
||||||
if (list_empty(&rdev->semaphore_drv.free)) {
|
|
||||||
int r;
|
int r;
|
||||||
|
|
||||||
|
retry:
|
||||||
|
*semaphore = NULL;
|
||||||
|
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
|
list_for_each_entry(bo, &rdev->semaphore_drv.bo, list) {
|
||||||
|
if (list_empty(&bo->free))
|
||||||
|
continue;
|
||||||
|
*semaphore = list_first_entry(&bo->free, struct radeon_semaphore, list);
|
||||||
|
(*semaphore)->cpu_ptr[0] = 0;
|
||||||
|
(*semaphore)->cpu_ptr[1] = 0;
|
||||||
|
list_del(&(*semaphore)->list);
|
||||||
|
bo->nused++;
|
||||||
|
break;
|
||||||
|
}
|
||||||
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
r = allocate_semaphores(rdev);
|
|
||||||
|
if (*semaphore == NULL) {
|
||||||
|
if (do_retry) {
|
||||||
|
do_retry = false;
|
||||||
|
r = radeon_semaphore_add_bo(rdev);
|
||||||
if (r)
|
if (r)
|
||||||
return r;
|
return r;
|
||||||
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
goto retry;
|
||||||
|
}
|
||||||
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
|
|
||||||
*semaphore = list_first_entry(&rdev->semaphore_drv.free, struct radeon_semaphore, list);
|
|
||||||
list_del(&(*semaphore)->list);
|
|
||||||
|
|
||||||
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -136,26 +155,24 @@ void radeon_semaphore_free(struct radeon_device *rdev,
|
|||||||
unsigned long irq_flags;
|
unsigned long irq_flags;
|
||||||
|
|
||||||
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
list_add_tail(&semaphore->list, &rdev->semaphore_drv.free);
|
semaphore->bo->nused--;
|
||||||
|
list_add_tail(&semaphore->list, &semaphore->bo->free);
|
||||||
|
radeon_semaphore_shrink_locked(rdev);
|
||||||
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
void radeon_semaphore_driver_fini(struct radeon_device *rdev)
|
void radeon_semaphore_driver_fini(struct radeon_device *rdev)
|
||||||
{
|
{
|
||||||
struct radeon_semaphore *i, *n;
|
struct radeon_semaphore_bo *bo, *n;
|
||||||
struct list_head entrys;
|
|
||||||
unsigned long irq_flags;
|
unsigned long irq_flags;
|
||||||
|
|
||||||
INIT_LIST_HEAD(&entrys);
|
|
||||||
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
write_lock_irqsave(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
if (!list_empty(&rdev->semaphore_drv.free)) {
|
/* we force to free everything */
|
||||||
list_splice(&rdev->semaphore_drv.free, &entrys);
|
list_for_each_entry_safe(bo, n, &rdev->semaphore_drv.bo, list) {
|
||||||
|
if (!list_empty(&bo->free)) {
|
||||||
|
dev_err(rdev->dev, "still in use semaphore\n");
|
||||||
|
}
|
||||||
|
radeon_semaphore_del_bo_locked(rdev, bo);
|
||||||
}
|
}
|
||||||
INIT_LIST_HEAD(&rdev->semaphore_drv.free);
|
|
||||||
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
write_unlock_irqrestore(&rdev->semaphore_drv.lock, irq_flags);
|
||||||
|
|
||||||
list_for_each_entry_safe(i, n, &entrys, list) {
|
|
||||||
radeon_bo_unref(&i->robj);
|
|
||||||
kfree(i);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
Reference in New Issue
Block a user