intel_context.h 6.2 KB
Newer Older
1 2 3 4 5 6 7 8 9
/*
 * SPDX-License-Identifier: MIT
 *
 * Copyright © 2019 Intel Corporation
 */

#ifndef __INTEL_CONTEXT_H__
#define __INTEL_CONTEXT_H__

10
#include <linux/bitops.h>
11
#include <linux/lockdep.h>
12
#include <linux/types.h>
13

14
#include "i915_active.h"
15
#include "i915_drv.h"
16 17
#include "intel_context_types.h"
#include "intel_engine_types.h"
18
#include "intel_ring_types.h"
19
#include "intel_timeline_types.h"
20

21 22
#define CE_TRACE(ce, fmt, ...) do {					\
	const struct intel_context *ce__ = (ce);			\
23
	ENGINE_TRACE(ce__->engine, "context:%llx " fmt,			\
24 25 26 27
		     ce__->timeline->fence_context,			\
		     ##__VA_ARGS__);					\
} while (0)

28 29
void intel_context_init(struct intel_context *ce,
			struct intel_engine_cs *engine);
30
void intel_context_fini(struct intel_context *ce);
31

32
struct intel_context *
33
intel_context_create(struct intel_engine_cs *engine);
34

35 36
int intel_context_alloc_state(struct intel_context *ce);

37 38
void intel_context_free(struct intel_context *ce);

39 40 41
int intel_context_reconfigure_sseu(struct intel_context *ce,
				   const struct intel_sseu sseu);

42
/**
43 44
 * intel_context_lock_pinned - Stablises the 'pinned' status of the HW context
 * @ce - the context
45
 *
46 47 48
 * Acquire a lock on the pinned status of the HW context, such that the context
 * can neither be bound to the GPU or unbound whilst the lock is held, i.e.
 * intel_context_is_pinned() remains stable.
49
 */
50 51 52 53 54
static inline int intel_context_lock_pinned(struct intel_context *ce)
	__acquires(ce->pin_mutex)
{
	return mutex_lock_interruptible(&ce->pin_mutex);
}
55

56 57 58 59 60 61 62 63 64
/**
 * intel_context_is_pinned - Reports the 'pinned' status
 * @ce - the context
 *
 * While in use by the GPU, the context, along with its ring and page
 * tables is pinned into memory and the GTT.
 *
 * Returns: true if the context is currently pinned for use by the GPU.
 */
65 66 67 68 69 70
static inline bool
intel_context_is_pinned(struct intel_context *ce)
{
	return atomic_read(&ce->pin_count);
}

71 72 73 74 75 76 77 78 79 80 81
/**
 * intel_context_unlock_pinned - Releases the earlier locking of 'pinned' status
 * @ce - the context
 *
 * Releases the lock earlier acquired by intel_context_unlock_pinned().
 */
static inline void intel_context_unlock_pinned(struct intel_context *ce)
	__releases(ce->pin_mutex)
{
	mutex_unlock(&ce->pin_mutex);
}
82

83 84
int __intel_context_do_pin(struct intel_context *ce);

85 86 87 88 89
static inline bool intel_context_pin_if_active(struct intel_context *ce)
{
	return atomic_inc_not_zero(&ce->pin_count);
}

90 91
static inline int intel_context_pin(struct intel_context *ce)
{
92
	if (likely(intel_context_pin_if_active(ce)))
93 94 95 96
		return 0;

	return __intel_context_do_pin(ce);
}
97 98 99

static inline void __intel_context_pin(struct intel_context *ce)
{
100 101
	GEM_BUG_ON(!intel_context_is_pinned(ce));
	atomic_inc(&ce->pin_count);
102 103
}

104
void intel_context_unpin(struct intel_context *ce);
105

106 107 108 109 110
void intel_context_enter_engine(struct intel_context *ce);
void intel_context_exit_engine(struct intel_context *ce);

static inline void intel_context_enter(struct intel_context *ce)
{
111
	lockdep_assert_held(&ce->timeline->mutex);
112 113 114 115 116 117
	if (!ce->active_count++)
		ce->ops->enter(ce);
}

static inline void intel_context_mark_active(struct intel_context *ce)
{
118
	lockdep_assert_held(&ce->timeline->mutex);
119 120 121 122 123
	++ce->active_count;
}

static inline void intel_context_exit(struct intel_context *ce)
{
124
	lockdep_assert_held(&ce->timeline->mutex);
125 126 127 128 129
	GEM_BUG_ON(!ce->active_count);
	if (!--ce->active_count)
		ce->ops->exit(ce);
}

130 131 132 133 134 135 136 137 138 139 140
static inline struct intel_context *intel_context_get(struct intel_context *ce)
{
	kref_get(&ce->ref);
	return ce;
}

static inline void intel_context_put(struct intel_context *ce)
{
	kref_put(&ce->ref, ce->ops->destroy);
}

141
static inline struct intel_timeline *__must_check
142
intel_context_timeline_lock(struct intel_context *ce)
143
	__acquires(&ce->timeline->mutex)
144
{
145 146 147 148 149 150 151 152
	struct intel_timeline *tl = ce->timeline;
	int err;

	err = mutex_lock_interruptible(&tl->mutex);
	if (err)
		return ERR_PTR(err);

	return tl;
153 154
}

155 156
static inline void intel_context_timeline_unlock(struct intel_timeline *tl)
	__releases(&tl->mutex)
157
{
158
	mutex_unlock(&tl->mutex);
159 160
}

161 162 163
int intel_context_prepare_remote_request(struct intel_context *ce,
					 struct i915_request *rq);

164 165
struct i915_request *intel_context_create_request(struct intel_context *ce);

166 167 168 169 170
static inline struct intel_ring *__intel_context_ring_size(u64 sz)
{
	return u64_to_ptr(struct intel_ring, sz);
}

171 172 173 174 175
static inline bool intel_context_is_barrier(const struct intel_context *ce)
{
	return test_bit(CONTEXT_BARRIER_BIT, &ce->flags);
}

176 177 178 179 180
static inline bool intel_context_is_closed(const struct intel_context *ce)
{
	return test_bit(CONTEXT_CLOSED_BIT, &ce->flags);
}

181 182 183 184 185 186 187 188 189 190 191 192 193 194 195
static inline bool intel_context_use_semaphores(const struct intel_context *ce)
{
	return test_bit(CONTEXT_USE_SEMAPHORES, &ce->flags);
}

static inline void intel_context_set_use_semaphores(struct intel_context *ce)
{
	set_bit(CONTEXT_USE_SEMAPHORES, &ce->flags);
}

static inline void intel_context_clear_use_semaphores(struct intel_context *ce)
{
	clear_bit(CONTEXT_USE_SEMAPHORES, &ce->flags);
}

196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235
static inline bool intel_context_is_banned(const struct intel_context *ce)
{
	return test_bit(CONTEXT_BANNED, &ce->flags);
}

static inline bool intel_context_set_banned(struct intel_context *ce)
{
	return test_and_set_bit(CONTEXT_BANNED, &ce->flags);
}

static inline bool
intel_context_force_single_submission(const struct intel_context *ce)
{
	return test_bit(CONTEXT_FORCE_SINGLE_SUBMISSION, &ce->flags);
}

static inline void
intel_context_set_single_submission(struct intel_context *ce)
{
	__set_bit(CONTEXT_FORCE_SINGLE_SUBMISSION, &ce->flags);
}

static inline bool
intel_context_nopreempt(const struct intel_context *ce)
{
	return test_bit(CONTEXT_NOPREEMPT, &ce->flags);
}

static inline void
intel_context_set_nopreempt(struct intel_context *ce)
{
	set_bit(CONTEXT_NOPREEMPT, &ce->flags);
}

static inline void
intel_context_clear_nopreempt(struct intel_context *ce)
{
	clear_bit(CONTEXT_NOPREEMPT, &ce->flags);
}

236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251
static inline u64 intel_context_get_total_runtime_ns(struct intel_context *ce)
{
	const u32 period =
		RUNTIME_INFO(ce->engine->i915)->cs_timestamp_period_ns;

	return READ_ONCE(ce->runtime.total) * period;
}

static inline u64 intel_context_get_avg_runtime_ns(struct intel_context *ce)
{
	const u32 period =
		RUNTIME_INFO(ce->engine->i915)->cs_timestamp_period_ns;

	return mul_u32_u32(ewma_runtime_read(&ce->runtime.avg), period);
}

252
#endif /* __INTEL_CONTEXT_H__ */