We use some of the lower bits of the retire function pointer for potential flags, which is quite thorny, since the caller needs to remember to give the function the correct alignment with __i915_active_call, otherwise we might incorrectly unpack the pointer and jump to some garbage address later. Instead of all this let's just pass the flags along as a separate parameter. Suggested-by: Ville Syrjälä <ville.syrjala@linux.intel.com> Suggested-by: Daniel Vetter <daniel@ffwll.ch> References: ca419f407b43 ("drm/i915: Fix crash in auto_retire") References: d8e44e4dd221 ("drm/i915/overlay: Fix active retire callback alignment") References: fd5f262db118 ("drm/i915/selftests: Fix active retire callback alignment") Signed-off-by: Matthew Auld <matthew.auld@intel.com> Reviewed-by: Matthew Brost <matthew.brost@intel.com> Acked-by: Daniel Vetter <daniel.vetter@ffwll.ch> Link: https://patchwork.freedesktop.org/patch/msgid/20210504164136.96456-1-matthew.auld@intel.com
352 lines
7.1 KiB
C
352 lines
7.1 KiB
C
/*
|
|
* SPDX-License-Identifier: MIT
|
|
*
|
|
* Copyright © 2018 Intel Corporation
|
|
*/
|
|
|
|
#include <linux/kref.h>
|
|
|
|
#include "gem/i915_gem_pm.h"
|
|
#include "gt/intel_gt.h"
|
|
|
|
#include "i915_selftest.h"
|
|
|
|
#include "igt_flush_test.h"
|
|
#include "lib_sw_fence.h"
|
|
|
|
struct live_active {
|
|
struct i915_active base;
|
|
struct kref ref;
|
|
bool retired;
|
|
};
|
|
|
|
static void __live_get(struct live_active *active)
|
|
{
|
|
kref_get(&active->ref);
|
|
}
|
|
|
|
static void __live_free(struct live_active *active)
|
|
{
|
|
i915_active_fini(&active->base);
|
|
kfree(active);
|
|
}
|
|
|
|
static void __live_release(struct kref *ref)
|
|
{
|
|
struct live_active *active = container_of(ref, typeof(*active), ref);
|
|
|
|
__live_free(active);
|
|
}
|
|
|
|
static void __live_put(struct live_active *active)
|
|
{
|
|
kref_put(&active->ref, __live_release);
|
|
}
|
|
|
|
static int __live_active(struct i915_active *base)
|
|
{
|
|
struct live_active *active = container_of(base, typeof(*active), base);
|
|
|
|
__live_get(active);
|
|
return 0;
|
|
}
|
|
|
|
static void __live_retire(struct i915_active *base)
|
|
{
|
|
struct live_active *active = container_of(base, typeof(*active), base);
|
|
|
|
active->retired = true;
|
|
__live_put(active);
|
|
}
|
|
|
|
static struct live_active *__live_alloc(struct drm_i915_private *i915)
|
|
{
|
|
struct live_active *active;
|
|
|
|
active = kzalloc(sizeof(*active), GFP_KERNEL);
|
|
if (!active)
|
|
return NULL;
|
|
|
|
kref_init(&active->ref);
|
|
i915_active_init(&active->base, __live_active, __live_retire, 0);
|
|
|
|
return active;
|
|
}
|
|
|
|
static struct live_active *
|
|
__live_active_setup(struct drm_i915_private *i915)
|
|
{
|
|
struct intel_engine_cs *engine;
|
|
struct i915_sw_fence *submit;
|
|
struct live_active *active;
|
|
unsigned int count = 0;
|
|
int err = 0;
|
|
|
|
active = __live_alloc(i915);
|
|
if (!active)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
submit = heap_fence_create(GFP_KERNEL);
|
|
if (!submit) {
|
|
kfree(active);
|
|
return ERR_PTR(-ENOMEM);
|
|
}
|
|
|
|
err = i915_active_acquire(&active->base);
|
|
if (err)
|
|
goto out;
|
|
|
|
for_each_uabi_engine(engine, i915) {
|
|
struct i915_request *rq;
|
|
|
|
rq = intel_engine_create_kernel_request(engine);
|
|
if (IS_ERR(rq)) {
|
|
err = PTR_ERR(rq);
|
|
break;
|
|
}
|
|
|
|
err = i915_sw_fence_await_sw_fence_gfp(&rq->submit,
|
|
submit,
|
|
GFP_KERNEL);
|
|
if (err >= 0)
|
|
err = i915_active_add_request(&active->base, rq);
|
|
i915_request_add(rq);
|
|
if (err) {
|
|
pr_err("Failed to track active ref!\n");
|
|
break;
|
|
}
|
|
|
|
count++;
|
|
}
|
|
|
|
i915_active_release(&active->base);
|
|
if (READ_ONCE(active->retired) && count) {
|
|
pr_err("i915_active retired before submission!\n");
|
|
err = -EINVAL;
|
|
}
|
|
if (atomic_read(&active->base.count) != count) {
|
|
pr_err("i915_active not tracking all requests, found %d, expected %d\n",
|
|
atomic_read(&active->base.count), count);
|
|
err = -EINVAL;
|
|
}
|
|
|
|
out:
|
|
i915_sw_fence_commit(submit);
|
|
heap_fence_put(submit);
|
|
if (err) {
|
|
__live_put(active);
|
|
active = ERR_PTR(err);
|
|
}
|
|
|
|
return active;
|
|
}
|
|
|
|
static int live_active_wait(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct live_active *active;
|
|
int err = 0;
|
|
|
|
/* Check that we get a callback when requests retire upon waiting */
|
|
|
|
active = __live_active_setup(i915);
|
|
if (IS_ERR(active))
|
|
return PTR_ERR(active);
|
|
|
|
__i915_active_wait(&active->base, TASK_UNINTERRUPTIBLE);
|
|
if (!READ_ONCE(active->retired)) {
|
|
struct drm_printer p = drm_err_printer(__func__);
|
|
|
|
pr_err("i915_active not retired after waiting!\n");
|
|
i915_active_print(&active->base, &p);
|
|
|
|
err = -EINVAL;
|
|
}
|
|
|
|
__live_put(active);
|
|
|
|
if (igt_flush_test(i915))
|
|
err = -EIO;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int live_active_retire(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct live_active *active;
|
|
int err = 0;
|
|
|
|
/* Check that we get a callback when requests are indirectly retired */
|
|
|
|
active = __live_active_setup(i915);
|
|
if (IS_ERR(active))
|
|
return PTR_ERR(active);
|
|
|
|
/* waits for & retires all requests */
|
|
if (igt_flush_test(i915))
|
|
err = -EIO;
|
|
|
|
if (!READ_ONCE(active->retired)) {
|
|
struct drm_printer p = drm_err_printer(__func__);
|
|
|
|
pr_err("i915_active not retired after flushing!\n");
|
|
i915_active_print(&active->base, &p);
|
|
|
|
err = -EINVAL;
|
|
}
|
|
|
|
__live_put(active);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int live_active_barrier(void *arg)
|
|
{
|
|
struct drm_i915_private *i915 = arg;
|
|
struct intel_engine_cs *engine;
|
|
struct live_active *active;
|
|
int err = 0;
|
|
|
|
/* Check that we get a callback when requests retire upon waiting */
|
|
|
|
active = __live_alloc(i915);
|
|
if (!active)
|
|
return -ENOMEM;
|
|
|
|
err = i915_active_acquire(&active->base);
|
|
if (err)
|
|
goto out;
|
|
|
|
for_each_uabi_engine(engine, i915) {
|
|
err = i915_active_acquire_preallocate_barrier(&active->base,
|
|
engine);
|
|
if (err)
|
|
break;
|
|
|
|
i915_active_acquire_barrier(&active->base);
|
|
}
|
|
|
|
i915_active_release(&active->base);
|
|
if (err)
|
|
goto out;
|
|
|
|
__i915_active_wait(&active->base, TASK_UNINTERRUPTIBLE);
|
|
if (!READ_ONCE(active->retired)) {
|
|
pr_err("i915_active not retired after flushing barriers!\n");
|
|
err = -EINVAL;
|
|
}
|
|
|
|
out:
|
|
__live_put(active);
|
|
|
|
if (igt_flush_test(i915))
|
|
err = -EIO;
|
|
|
|
return err;
|
|
}
|
|
|
|
int i915_active_live_selftests(struct drm_i915_private *i915)
|
|
{
|
|
static const struct i915_subtest tests[] = {
|
|
SUBTEST(live_active_wait),
|
|
SUBTEST(live_active_retire),
|
|
SUBTEST(live_active_barrier),
|
|
};
|
|
|
|
if (intel_gt_is_wedged(&i915->gt))
|
|
return 0;
|
|
|
|
return i915_subtests(tests, i915);
|
|
}
|
|
|
|
static struct intel_engine_cs *node_to_barrier(struct active_node *it)
|
|
{
|
|
struct intel_engine_cs *engine;
|
|
|
|
if (!is_barrier(&it->base))
|
|
return NULL;
|
|
|
|
engine = __barrier_to_engine(it);
|
|
smp_rmb(); /* serialise with add_active_barriers */
|
|
if (!is_barrier(&it->base))
|
|
return NULL;
|
|
|
|
return engine;
|
|
}
|
|
|
|
void i915_active_print(struct i915_active *ref, struct drm_printer *m)
|
|
{
|
|
drm_printf(m, "active %ps:%ps\n", ref->active, ref->retire);
|
|
drm_printf(m, "\tcount: %d\n", atomic_read(&ref->count));
|
|
drm_printf(m, "\tpreallocated barriers? %s\n",
|
|
yesno(!llist_empty(&ref->preallocated_barriers)));
|
|
|
|
if (i915_active_acquire_if_busy(ref)) {
|
|
struct active_node *it, *n;
|
|
|
|
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node) {
|
|
struct intel_engine_cs *engine;
|
|
|
|
engine = node_to_barrier(it);
|
|
if (engine) {
|
|
drm_printf(m, "\tbarrier: %s\n", engine->name);
|
|
continue;
|
|
}
|
|
|
|
if (i915_active_fence_isset(&it->base)) {
|
|
drm_printf(m,
|
|
"\ttimeline: %llx\n", it->timeline);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
i915_active_release(ref);
|
|
}
|
|
}
|
|
|
|
static void spin_unlock_wait(spinlock_t *lock)
|
|
{
|
|
spin_lock_irq(lock);
|
|
spin_unlock_irq(lock);
|
|
}
|
|
|
|
static void active_flush(struct i915_active *ref,
|
|
struct i915_active_fence *active)
|
|
{
|
|
struct dma_fence *fence;
|
|
|
|
fence = xchg(__active_fence_slot(active), NULL);
|
|
if (!fence)
|
|
return;
|
|
|
|
spin_lock_irq(fence->lock);
|
|
__list_del_entry(&active->cb.node);
|
|
spin_unlock_irq(fence->lock); /* serialise with fence->cb_list */
|
|
atomic_dec(&ref->count);
|
|
|
|
GEM_BUG_ON(!test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags));
|
|
}
|
|
|
|
void i915_active_unlock_wait(struct i915_active *ref)
|
|
{
|
|
if (i915_active_acquire_if_busy(ref)) {
|
|
struct active_node *it, *n;
|
|
|
|
/* Wait for all active callbacks */
|
|
rcu_read_lock();
|
|
active_flush(ref, &ref->excl);
|
|
rbtree_postorder_for_each_entry_safe(it, n, &ref->tree, node)
|
|
active_flush(ref, &it->base);
|
|
rcu_read_unlock();
|
|
|
|
i915_active_release(ref);
|
|
}
|
|
|
|
/* And wait for the retire callback */
|
|
spin_unlock_wait(&ref->tree_lock);
|
|
|
|
/* ... which may have been on a thread instead */
|
|
flush_work(&ref->work);
|
|
}
|