drm/xe/bo: Introduce xe_bo_put_async

Introduce xe_bo_put_async to put a bo where the context is such that
the bo destructor can't run due to lockdep problems or atomic context.

If the put is the final put, freeing will be done from a work item.

v5:
 - Kerenl doc for xe_bo_put_async (Thomas)
v7:
 - Fix kernel doc (CI)

Signed-off-by: Matthew Brost <matthew.brost@intel.com>
Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Reviewed-by: Matthew Brost <matthew.brost@intel.com>
Reviewed-by: Himal Prasad Ghimiray <himal.prasad.ghimiray@intel.com>
Tested-by: Gwan-gyeong Mun <gwan-gyeong.mun@intel.com>
Reviewed-by: Gwan-gyeong Mun <gwan-gyeong.mun@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20250306012657.3505757-6-matthew.brost@intel.com
This commit is contained in:
Thomas Hellström
2025-03-05 17:26:30 -08:00
committed by Matthew Brost
parent 73463dac9b
commit 5473f4d4e2
4 changed files with 55 additions and 0 deletions
+25
View File
@@ -2660,6 +2660,31 @@ void xe_bo_put_commit(struct llist_head *deferred)
drm_gem_object_free(&bo->ttm.base.refcount);
}
static void xe_bo_dev_work_func(struct work_struct *work)
{
struct xe_bo_dev *bo_dev = container_of(work, typeof(*bo_dev), async_free);
xe_bo_put_commit(&bo_dev->async_list);
}
/**
* xe_bo_dev_init() - Initialize BO dev to manage async BO freeing
* @bo_dev: The BO dev structure
*/
void xe_bo_dev_init(struct xe_bo_dev *bo_dev)
{
INIT_WORK(&bo_dev->async_free, xe_bo_dev_work_func);
}
/**
* xe_bo_dev_fini() - Finalize BO dev managing async BO freeing
* @bo_dev: The BO dev structure
*/
void xe_bo_dev_fini(struct xe_bo_dev *bo_dev)
{
flush_work(&bo_dev->async_free);
}
void xe_bo_put(struct xe_bo *bo)
{
struct xe_tile *tile;
+19
View File
@@ -323,6 +323,25 @@ xe_bo_put_deferred(struct xe_bo *bo, struct llist_head *deferred)
void xe_bo_put_commit(struct llist_head *deferred);
/**
* xe_bo_put_async() - Put BO async
* @bo: The bo to put.
*
* Put BO async, the final put is deferred to a worker to exit an IRQ context.
*/
static inline void
xe_bo_put_async(struct xe_bo *bo)
{
struct xe_bo_dev *bo_device = &xe_bo_device(bo)->bo_device;
if (xe_bo_put_deferred(bo, &bo_device->async_list))
schedule_work(&bo_device->async_free);
}
void xe_bo_dev_init(struct xe_bo_dev *bo_device);
void xe_bo_dev_fini(struct xe_bo_dev *bo_device);
struct sg_table *xe_bo_sg(struct xe_bo *bo);
/*
+3
View File
@@ -387,6 +387,8 @@ static void xe_device_destroy(struct drm_device *dev, void *dummy)
{
struct xe_device *xe = to_xe_device(dev);
xe_bo_dev_fini(&xe->bo_device);
if (xe->preempt_fence_wq)
destroy_workqueue(xe->preempt_fence_wq);
@@ -424,6 +426,7 @@ struct xe_device *xe_device_create(struct pci_dev *pdev,
if (WARN_ON(err))
goto err;
xe_bo_dev_init(&xe->bo_device);
err = drmm_add_action_or_reset(&xe->drm, xe_device_destroy, NULL);
if (err)
goto err;
+8
View File
@@ -525,6 +525,14 @@ struct xe_device {
int mode;
} wedged;
/** @bo_device: Struct to control async free of BOs */
struct xe_bo_dev {
/** @bo_device.async_free: Free worker */
struct work_struct async_free;
/** @bo_device.async_list: List of BOs to be freed */
struct llist_head async_list;
} bo_device;
/** @pmu: performance monitoring unit */
struct xe_pmu pmu;