@@ -105,61 +105,67 @@ struct intel_guc {
105
105
*/
106
106
struct {
107
107
/**
108
- * @lock: protects everything in submission_state,
109
- * ce->guc_id.id, and ce->guc_id.ref when transitioning in and
110
- * out of zero
108
+ * @submission_state. lock: protects everything in
109
+ * submission_state, ce->guc_id.id, and ce->guc_id.ref
110
+ * when transitioning in and out of zero
111
111
*/
112
112
spinlock_t lock ;
113
113
/**
114
- * @guc_ids: used to allocate new guc_ids, single-lrc
114
+ * @submission_state.guc_ids: used to allocate new
115
+ * guc_ids, single-lrc
115
116
*/
116
117
struct ida guc_ids ;
117
118
/**
118
- * @num_guc_ids: Number of guc_ids, selftest feature to be able
119
- * to reduce this number while testing.
119
+ * @submission_state. num_guc_ids: Number of guc_ids, selftest
120
+ * feature to be able to reduce this number while testing.
120
121
*/
121
122
int num_guc_ids ;
122
123
/**
123
- * @guc_ids_bitmap: used to allocate new guc_ids, multi-lrc
124
+ * @submission_state.guc_ids_bitmap: used to allocate
125
+ * new guc_ids, multi-lrc
124
126
*/
125
127
unsigned long * guc_ids_bitmap ;
126
128
/**
127
- * @guc_id_list: list of intel_context with valid guc_ids but no
128
- * refs
129
+ * @submission_state. guc_id_list: list of intel_context
130
+ * with valid guc_ids but no refs
129
131
*/
130
132
struct list_head guc_id_list ;
131
133
/**
132
- * @guc_ids_in_use: Number single-lrc guc_ids in use
134
+ * @submission_state.guc_ids_in_use: Number single-lrc
135
+ * guc_ids in use
133
136
*/
134
137
unsigned int guc_ids_in_use ;
135
138
/**
136
- * @destroyed_contexts: list of contexts waiting to be destroyed
137
- * (deregistered with the GuC)
139
+ * @submission_state. destroyed_contexts: list of contexts
140
+ * waiting to be destroyed (deregistered with the GuC)
138
141
*/
139
142
struct list_head destroyed_contexts ;
140
143
/**
141
- * @destroyed_worker: worker to deregister contexts, need as we
142
- * need to take a GT PM reference and can't from destroy
143
- * function as it might be in an atomic context (no sleeping)
144
+ * @submission_state.destroyed_worker: worker to deregister
145
+ * contexts, need as we need to take a GT PM reference and
146
+ * can't from destroy function as it might be in an atomic
147
+ * context (no sleeping)
144
148
*/
145
149
struct work_struct destroyed_worker ;
146
150
/**
147
- * @reset_fail_worker: worker to trigger a GT reset after an
148
- * engine reset fails
151
+ * @submission_state. reset_fail_worker: worker to trigger
152
+ * a GT reset after an engine reset fails
149
153
*/
150
154
struct work_struct reset_fail_worker ;
151
155
/**
152
- * @reset_fail_mask: mask of engines that failed to reset
156
+ * @submission_state.reset_fail_mask: mask of engines that
157
+ * failed to reset
153
158
*/
154
159
intel_engine_mask_t reset_fail_mask ;
155
160
/**
156
- * @sched_disable_delay_ms: schedule disable delay, in ms, for
157
- * contexts
161
+ * @submission_state. sched_disable_delay_ms: schedule
162
+ * disable delay, in ms, for contexts
158
163
*/
159
164
unsigned int sched_disable_delay_ms ;
160
165
/**
161
- * @sched_disable_gucid_threshold: threshold of min remaining available
162
- * guc_ids before we start bypassing the schedule disable delay
166
+ * @submission_state.sched_disable_gucid_threshold:
167
+ * threshold of min remaining available guc_ids before
168
+ * we start bypassing the schedule disable delay
163
169
*/
164
170
unsigned int sched_disable_gucid_threshold ;
165
171
} submission_state ;
@@ -243,37 +249,40 @@ struct intel_guc {
243
249
*/
244
250
struct {
245
251
/**
246
- * @lock: Lock protecting the below fields and the engine stats.
252
+ * @timestamp.lock: Lock protecting the below fields and
253
+ * the engine stats.
247
254
*/
248
255
spinlock_t lock ;
249
256
250
257
/**
251
- * @gt_stamp: 64 bit extended value of the GT timestamp.
258
+ * @timestamp.gt_stamp: 64-bit extended value of the GT
259
+ * timestamp.
252
260
*/
253
261
u64 gt_stamp ;
254
262
255
263
/**
256
- * @ping_delay: Period for polling the GT timestamp for
257
- * overflow.
264
+ * @timestamp. ping_delay: Period for polling the GT
265
+ * timestamp for overflow.
258
266
*/
259
267
unsigned long ping_delay ;
260
268
261
269
/**
262
- * @work: Periodic work to adjust GT timestamp, engine and
263
- * context usage for overflows.
270
+ * @timestamp. work: Periodic work to adjust GT timestamp,
271
+ * engine and context usage for overflows.
264
272
*/
265
273
struct delayed_work work ;
266
274
267
275
/**
268
- * @shift: Right shift value for the gpm timestamp
276
+ * @timestamp. shift: Right shift value for the gpm timestamp
269
277
*/
270
278
u32 shift ;
271
279
272
280
/**
273
- * @last_stat_jiffies: jiffies at last actual stats collection time
274
- * We use this timestamp to ensure we don't oversample the
275
- * stats because runtime power management events can trigger
276
- * stats collection at much higher rates than required.
281
+ * @timestamp.last_stat_jiffies: jiffies at last actual
282
+ * stats collection time. We use this timestamp to ensure
283
+ * we don't oversample the stats because runtime power
284
+ * management events can trigger stats collection at much
285
+ * higher rates than required.
277
286
*/
278
287
unsigned long last_stat_jiffies ;
279
288
} timestamp ;
0 commit comments