Lines Matching defs:active

21  * they idle (when we know the active requests are inactive) and allocate the
36 node_from_active(struct i915_active_fence *active)
38 return container_of(active, struct active_node, base);
43 static inline bool is_barrier(const struct i915_active_fence *active)
45 return IS_ERR(rcu_access_pointer(active->fence));
79 return (void *)ref->active ?: (void *)ref->retire ?: (void *)ref;
204 __active_fence_slot(struct i915_active_fence *active)
206 return (struct dma_fence ** __force)&active->fence;
212 struct i915_active_fence *active =
213 container_of(cb, typeof(*active), cb);
215 return try_cmpxchg(__active_fence_slot(active), &fence, NULL);
269 /* While active, the tree can only be built; not destroyed */
339 int (*active)(struct i915_active *ref),
348 ref->active = active;
414 replace_barrier(struct i915_active *ref, struct i915_active_fence *active)
416 if (!is_barrier(active)) /* proto-node used by our idle barrier? */
424 return __active_del_barrier(ref, node_from_active(active));
431 struct i915_active_fence *active;
440 active = active_instance(ref, idx);
441 if (!active) {
446 if (replace_barrier(ref, active)) {
447 RCU_INIT_POINTER(active->fence, NULL);
450 } while (unlikely(is_barrier(active)));
452 fence = __i915_active_fence_set(active, fence);
465 struct i915_active_fence *active,
470 if (replace_barrier(ref, active)) {
471 RCU_INIT_POINTER(active->fence, fence);
475 prev = __i915_active_fence_set(active, fence);
510 if (!ref->active) {
520 err = ref->active(ref);
536 static void enable_signaling(struct i915_active_fence *active)
540 if (unlikely(is_barrier(active)))
543 fence = i915_active_fence_get(active);
602 * After the wait is complete, the caller may free the active.
609 static int __await_active(struct i915_active_fence *active,
615 if (is_barrier(active)) /* XXX flush the barrier? */
618 fence = i915_active_fence_get(active);
767 * i915_active, due to overlapping active phases there is likely a
997 * __i915_active_fence_set: Update the last active fence along its timeline
998 * @active: the active tracker
1001 * Records the new @fence as the last active fence along its timeline in
1002 * this active tracker, moving the tracking callbacks from the previous
1010 __i915_active_fence_set(struct i915_active_fence *active,
1021 * while tracked under a different active tracker. Combined with i915
1026 * As a countermeasure, we try to get a reference to the active->fence
1031 prev = i915_active_fence_get(active);
1039 * C already resident as the active->fence.
1058 * active->fence. Meanwhile, B follows the same path as A.
1060 * active->fence, locks it as soon as A completes, and possibly
1063 while (cmpxchg(__active_fence_slot(active), prev, fence) != prev) {
1070 prev = i915_active_fence_get(active);
1091 __list_del_entry(&active->cb.node);
1094 list_add_tail(&active->cb.node, &fence->cb_list);
1100 int i915_active_fence_set(struct i915_active_fence *active,
1106 /* Must maintain timeline ordering wrt previous active requests */
1107 fence = __i915_active_fence_set(active, &rq->fence);