blob: 96aed0ee700af753ee1b21d6002445a2a3b6a7f5 [file] [log] [blame]
/*
* SPDX-License-Identifier: MIT
*
* Copyright © 2019 Intel Corporation
*/
#ifndef _I915_ACTIVE_TYPES_H_
#define _I915_ACTIVE_TYPES_H_
#include <linux/atomic.h>
#include <linux/dma-fence.h>
#include <linux/llist.h>
#include <linux/mutex.h>
#include <linux/rbtree.h>
#include <linux/rcupdate.h>
#include <linux/workqueue.h>
#include "i915_utils.h"
struct i915_active_fence {
struct dma_fence __rcu *fence;
struct dma_fence_cb cb;
#if IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM)
/*
* Incorporeal!
*
* Updates to the i915_active_request must be serialised under a lock
* to ensure that the timeline is ordered. Normally, this is the
* timeline->mutex, but another mutex may be used so long as it is
* done so consistently.
*
* For lockdep tracking of the above, we store the lock we intend
* to always use for updates of this i915_active_request during
* construction and assert that is held on every update.
*/
struct mutex *lock;
#endif
};
struct active_node;
#define I915_ACTIVE_MAY_SLEEP BIT(0)
#define __i915_active_call __aligned(4)
#define i915_active_may_sleep(fn) ptr_pack_bits(&(fn), I915_ACTIVE_MAY_SLEEP, 2)
struct i915_active {
atomic_t count;
struct mutex mutex;
spinlock_t tree_lock;
struct active_node *cache;
struct rb_root tree;
/* Preallocated "exclusive" node */
struct i915_active_fence excl;
unsigned long flags;
#define I915_ACTIVE_RETIRE_SLEEPS BIT(0)
int (*active)(struct i915_active *ref);
void (*retire)(struct i915_active *ref);
struct work_struct work;
struct llist_head preallocated_barriers;
};
#endif /* _I915_ACTIVE_TYPES_H_ */