Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/drivers/accel/ivpu/ivpu_jsm_msg.c
26428 views
1
// SPDX-License-Identifier: GPL-2.0-only
2
/*
3
* Copyright (C) 2020-2024 Intel Corporation
4
*/
5
6
#include "ivpu_drv.h"
7
#include "ivpu_hw.h"
8
#include "ivpu_ipc.h"
9
#include "ivpu_jsm_msg.h"
10
#include "ivpu_pm.h"
11
#include "vpu_jsm_api.h"
12
13
const char *ivpu_jsm_msg_type_to_str(enum vpu_ipc_msg_type type)
14
{
15
#define IVPU_CASE_TO_STR(x) case x: return #x
16
switch (type) {
17
IVPU_CASE_TO_STR(VPU_JSM_MSG_UNKNOWN);
18
IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_RESET);
19
IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_PREEMPT);
20
IVPU_CASE_TO_STR(VPU_JSM_MSG_REGISTER_DB);
21
IVPU_CASE_TO_STR(VPU_JSM_MSG_UNREGISTER_DB);
22
IVPU_CASE_TO_STR(VPU_JSM_MSG_QUERY_ENGINE_HB);
23
IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL_COUNT);
24
IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL);
25
IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_POWER_LEVEL);
26
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_OPEN);
27
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_CLOSE);
28
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_SET_CONFIG);
29
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CONFIG);
30
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CAPABILITY);
31
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_NAME);
32
IVPU_CASE_TO_STR(VPU_JSM_MSG_SSID_RELEASE);
33
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_START);
34
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_STOP);
35
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_UPDATE);
36
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_INFO);
37
IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP);
38
IVPU_CASE_TO_STR(VPU_JSM_MSG_CREATE_CMD_QUEUE);
39
IVPU_CASE_TO_STR(VPU_JSM_MSG_DESTROY_CMD_QUEUE);
40
IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES);
41
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_REGISTER_DB);
42
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_RESUME_CMDQ);
43
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SUSPEND_CMDQ);
44
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_RESUME_CMDQ_RSP);
45
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SUSPEND_CMDQ_DONE);
46
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG);
47
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG_RSP);
48
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_SCHEDULING_LOG_NOTIFICATION);
49
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_ENGINE_RESUME);
50
IVPU_CASE_TO_STR(VPU_JSM_MSG_HWS_RESUME_ENGINE_DONE);
51
IVPU_CASE_TO_STR(VPU_JSM_MSG_STATE_DUMP);
52
IVPU_CASE_TO_STR(VPU_JSM_MSG_STATE_DUMP_RSP);
53
IVPU_CASE_TO_STR(VPU_JSM_MSG_BLOB_DEINIT_DEPRECATED);
54
IVPU_CASE_TO_STR(VPU_JSM_MSG_DYNDBG_CONTROL);
55
IVPU_CASE_TO_STR(VPU_JSM_MSG_JOB_DONE);
56
IVPU_CASE_TO_STR(VPU_JSM_MSG_NATIVE_FENCE_SIGNALLED);
57
IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_RESET_DONE);
58
IVPU_CASE_TO_STR(VPU_JSM_MSG_ENGINE_PREEMPT_DONE);
59
IVPU_CASE_TO_STR(VPU_JSM_MSG_REGISTER_DB_DONE);
60
IVPU_CASE_TO_STR(VPU_JSM_MSG_UNREGISTER_DB_DONE);
61
IVPU_CASE_TO_STR(VPU_JSM_MSG_QUERY_ENGINE_HB_DONE);
62
IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL_COUNT_DONE);
63
IVPU_CASE_TO_STR(VPU_JSM_MSG_GET_POWER_LEVEL_DONE);
64
IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_POWER_LEVEL_DONE);
65
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_OPEN_DONE);
66
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_CLOSE_DONE);
67
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_SET_CONFIG_RSP);
68
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CONFIG_RSP);
69
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_CAPABILITY_RSP);
70
IVPU_CASE_TO_STR(VPU_JSM_MSG_TRACE_GET_NAME_RSP);
71
IVPU_CASE_TO_STR(VPU_JSM_MSG_SSID_RELEASE_DONE);
72
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_START_DONE);
73
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_STOP_DONE);
74
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_UPDATE_DONE);
75
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_INFO_DONE);
76
IVPU_CASE_TO_STR(VPU_JSM_MSG_METRIC_STREAMER_NOTIFICATION);
77
IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP_RSP);
78
IVPU_CASE_TO_STR(VPU_JSM_MSG_CREATE_CMD_QUEUE_RSP);
79
IVPU_CASE_TO_STR(VPU_JSM_MSG_DESTROY_CMD_QUEUE_RSP);
80
IVPU_CASE_TO_STR(VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES_RSP);
81
IVPU_CASE_TO_STR(VPU_JSM_MSG_BLOB_DEINIT_DONE);
82
IVPU_CASE_TO_STR(VPU_JSM_MSG_DYNDBG_CONTROL_RSP);
83
IVPU_CASE_TO_STR(VPU_JSM_MSG_PWR_D0I3_ENTER);
84
IVPU_CASE_TO_STR(VPU_JSM_MSG_PWR_D0I3_ENTER_DONE);
85
IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_ENABLE);
86
IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_ENABLE_DONE);
87
IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_DISABLE);
88
IVPU_CASE_TO_STR(VPU_JSM_MSG_DCT_DISABLE_DONE);
89
}
90
#undef IVPU_CASE_TO_STR
91
92
return "Unknown JSM message type";
93
}
94
95
int ivpu_jsm_register_db(struct ivpu_device *vdev, u32 ctx_id, u32 db_id,
96
u64 jobq_base, u32 jobq_size)
97
{
98
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_REGISTER_DB };
99
struct vpu_jsm_msg resp;
100
int ret = 0;
101
102
req.payload.register_db.db_idx = db_id;
103
req.payload.register_db.jobq_base = jobq_base;
104
req.payload.register_db.jobq_size = jobq_size;
105
req.payload.register_db.host_ssid = ctx_id;
106
107
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_REGISTER_DB_DONE, &resp,
108
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
109
if (ret)
110
ivpu_err_ratelimited(vdev, "Failed to register doorbell %u: %d\n", db_id, ret);
111
112
return ret;
113
}
114
115
int ivpu_jsm_unregister_db(struct ivpu_device *vdev, u32 db_id)
116
{
117
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_UNREGISTER_DB };
118
struct vpu_jsm_msg resp;
119
int ret = 0;
120
121
req.payload.unregister_db.db_idx = db_id;
122
123
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_UNREGISTER_DB_DONE, &resp,
124
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
125
if (ret)
126
ivpu_warn_ratelimited(vdev, "Failed to unregister doorbell %u: %d\n", db_id, ret);
127
128
return ret;
129
}
130
131
int ivpu_jsm_get_heartbeat(struct ivpu_device *vdev, u32 engine, u64 *heartbeat)
132
{
133
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_QUERY_ENGINE_HB };
134
struct vpu_jsm_msg resp;
135
int ret;
136
137
if (engine != VPU_ENGINE_COMPUTE)
138
return -EINVAL;
139
140
req.payload.query_engine_hb.engine_idx = engine;
141
142
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_QUERY_ENGINE_HB_DONE, &resp,
143
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
144
if (ret) {
145
ivpu_err_ratelimited(vdev, "Failed to get heartbeat from engine %d: %d\n",
146
engine, ret);
147
return ret;
148
}
149
150
*heartbeat = resp.payload.query_engine_hb_done.heartbeat;
151
return ret;
152
}
153
154
int ivpu_jsm_reset_engine(struct ivpu_device *vdev, u32 engine)
155
{
156
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_ENGINE_RESET };
157
struct vpu_jsm_msg resp;
158
int ret;
159
160
if (engine != VPU_ENGINE_COMPUTE)
161
return -EINVAL;
162
163
req.payload.engine_reset.engine_idx = engine;
164
165
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_ENGINE_RESET_DONE, &resp,
166
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
167
if (ret) {
168
ivpu_err_ratelimited(vdev, "Failed to reset engine %d: %d\n", engine, ret);
169
ivpu_pm_trigger_recovery(vdev, "Engine reset failed");
170
}
171
172
return ret;
173
}
174
175
int ivpu_jsm_preempt_engine(struct ivpu_device *vdev, u32 engine, u32 preempt_id)
176
{
177
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_ENGINE_PREEMPT };
178
struct vpu_jsm_msg resp;
179
int ret;
180
181
if (engine != VPU_ENGINE_COMPUTE)
182
return -EINVAL;
183
184
req.payload.engine_preempt.engine_idx = engine;
185
req.payload.engine_preempt.preempt_id = preempt_id;
186
187
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_ENGINE_PREEMPT_DONE, &resp,
188
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
189
if (ret)
190
ivpu_err_ratelimited(vdev, "Failed to preempt engine %d: %d\n", engine, ret);
191
192
return ret;
193
}
194
195
int ivpu_jsm_dyndbg_control(struct ivpu_device *vdev, char *command, size_t size)
196
{
197
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DYNDBG_CONTROL };
198
struct vpu_jsm_msg resp;
199
int ret;
200
201
strscpy(req.payload.dyndbg_control.dyndbg_cmd, command, VPU_DYNDBG_CMD_MAX_LEN);
202
203
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_DYNDBG_CONTROL_RSP, &resp,
204
VPU_IPC_CHAN_GEN_CMD, vdev->timeout.jsm);
205
if (ret)
206
ivpu_warn_ratelimited(vdev, "Failed to send command \"%s\": ret %d\n",
207
command, ret);
208
209
return ret;
210
}
211
212
int ivpu_jsm_trace_get_capability(struct ivpu_device *vdev, u32 *trace_destination_mask,
213
u64 *trace_hw_component_mask)
214
{
215
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_TRACE_GET_CAPABILITY };
216
struct vpu_jsm_msg resp;
217
int ret;
218
219
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_TRACE_GET_CAPABILITY_RSP, &resp,
220
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
221
if (ret) {
222
ivpu_warn_ratelimited(vdev, "Failed to get trace capability: %d\n", ret);
223
return ret;
224
}
225
226
*trace_destination_mask = resp.payload.trace_capability.trace_destination_mask;
227
*trace_hw_component_mask = resp.payload.trace_capability.trace_hw_component_mask;
228
229
return ret;
230
}
231
232
int ivpu_jsm_trace_set_config(struct ivpu_device *vdev, u32 trace_level, u32 trace_destination_mask,
233
u64 trace_hw_component_mask)
234
{
235
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_TRACE_SET_CONFIG };
236
struct vpu_jsm_msg resp;
237
int ret;
238
239
req.payload.trace_config.trace_level = trace_level;
240
req.payload.trace_config.trace_destination_mask = trace_destination_mask;
241
req.payload.trace_config.trace_hw_component_mask = trace_hw_component_mask;
242
243
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_TRACE_SET_CONFIG_RSP, &resp,
244
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
245
if (ret)
246
ivpu_warn_ratelimited(vdev, "Failed to set config: %d\n", ret);
247
248
return ret;
249
}
250
251
int ivpu_jsm_context_release(struct ivpu_device *vdev, u32 host_ssid)
252
{
253
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_SSID_RELEASE };
254
struct vpu_jsm_msg resp;
255
int ret;
256
257
req.payload.ssid_release.host_ssid = host_ssid;
258
259
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_SSID_RELEASE_DONE, &resp,
260
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
261
if (ret)
262
ivpu_warn_ratelimited(vdev, "Failed to release context: %d\n", ret);
263
264
return ret;
265
}
266
267
int ivpu_jsm_pwr_d0i3_enter(struct ivpu_device *vdev)
268
{
269
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_PWR_D0I3_ENTER };
270
struct vpu_jsm_msg resp;
271
int ret;
272
273
if (IVPU_WA(disable_d0i3_msg))
274
return 0;
275
276
req.payload.pwr_d0i3_enter.send_response = 1;
277
278
ret = ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_PWR_D0I3_ENTER_DONE, &resp,
279
VPU_IPC_CHAN_GEN_CMD, vdev->timeout.d0i3_entry_msg);
280
if (ret)
281
return ret;
282
283
return ivpu_hw_wait_for_idle(vdev);
284
}
285
286
int ivpu_jsm_hws_create_cmdq(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_group, u32 cmdq_id,
287
u32 pid, u32 engine, u64 cmdq_base, u32 cmdq_size)
288
{
289
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_CREATE_CMD_QUEUE };
290
struct vpu_jsm_msg resp;
291
int ret;
292
293
req.payload.hws_create_cmdq.host_ssid = ctx_id;
294
req.payload.hws_create_cmdq.process_id = pid;
295
req.payload.hws_create_cmdq.engine_idx = engine;
296
req.payload.hws_create_cmdq.cmdq_group = cmdq_group;
297
req.payload.hws_create_cmdq.cmdq_id = cmdq_id;
298
req.payload.hws_create_cmdq.cmdq_base = cmdq_base;
299
req.payload.hws_create_cmdq.cmdq_size = cmdq_size;
300
301
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_CREATE_CMD_QUEUE_RSP, &resp,
302
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
303
if (ret)
304
ivpu_warn_ratelimited(vdev, "Failed to create command queue: %d\n", ret);
305
306
return ret;
307
}
308
309
int ivpu_jsm_hws_destroy_cmdq(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id)
310
{
311
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DESTROY_CMD_QUEUE };
312
struct vpu_jsm_msg resp;
313
int ret;
314
315
req.payload.hws_destroy_cmdq.host_ssid = ctx_id;
316
req.payload.hws_destroy_cmdq.cmdq_id = cmdq_id;
317
318
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_DESTROY_CMD_QUEUE_RSP, &resp,
319
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
320
if (ret)
321
ivpu_warn_ratelimited(vdev, "Failed to destroy command queue: %d\n", ret);
322
323
return ret;
324
}
325
326
int ivpu_jsm_hws_register_db(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id, u32 db_id,
327
u64 cmdq_base, u32 cmdq_size)
328
{
329
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_HWS_REGISTER_DB };
330
struct vpu_jsm_msg resp;
331
int ret = 0;
332
333
req.payload.hws_register_db.db_id = db_id;
334
req.payload.hws_register_db.host_ssid = ctx_id;
335
req.payload.hws_register_db.cmdq_id = cmdq_id;
336
req.payload.hws_register_db.cmdq_base = cmdq_base;
337
req.payload.hws_register_db.cmdq_size = cmdq_size;
338
339
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_REGISTER_DB_DONE, &resp,
340
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
341
if (ret)
342
ivpu_err_ratelimited(vdev, "Failed to register doorbell %u: %d\n", db_id, ret);
343
344
return ret;
345
}
346
347
int ivpu_jsm_hws_resume_engine(struct ivpu_device *vdev, u32 engine)
348
{
349
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_HWS_ENGINE_RESUME };
350
struct vpu_jsm_msg resp;
351
int ret;
352
353
if (engine != VPU_ENGINE_COMPUTE)
354
return -EINVAL;
355
356
req.payload.hws_resume_engine.engine_idx = engine;
357
358
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_HWS_RESUME_ENGINE_DONE, &resp,
359
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
360
if (ret) {
361
ivpu_err_ratelimited(vdev, "Failed to resume engine %d: %d\n", engine, ret);
362
ivpu_pm_trigger_recovery(vdev, "Engine resume failed");
363
}
364
365
return ret;
366
}
367
368
int ivpu_jsm_hws_set_context_sched_properties(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id,
369
u32 priority)
370
{
371
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES };
372
struct vpu_jsm_msg resp;
373
int ret;
374
375
req.payload.hws_set_context_sched_properties.host_ssid = ctx_id;
376
req.payload.hws_set_context_sched_properties.cmdq_id = cmdq_id;
377
req.payload.hws_set_context_sched_properties.priority_band = priority;
378
req.payload.hws_set_context_sched_properties.realtime_priority_level = 0;
379
req.payload.hws_set_context_sched_properties.in_process_priority = 0;
380
req.payload.hws_set_context_sched_properties.context_quantum = 20000;
381
req.payload.hws_set_context_sched_properties.grace_period_same_priority = 10000;
382
req.payload.hws_set_context_sched_properties.grace_period_lower_priority = 0;
383
384
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_SET_CONTEXT_SCHED_PROPERTIES_RSP, &resp,
385
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
386
if (ret)
387
ivpu_warn_ratelimited(vdev, "Failed to set context sched properties: %d\n", ret);
388
389
return ret;
390
}
391
392
int ivpu_jsm_hws_set_scheduling_log(struct ivpu_device *vdev, u32 engine_idx, u32 host_ssid,
393
u64 vpu_log_buffer_va)
394
{
395
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG };
396
struct vpu_jsm_msg resp;
397
int ret;
398
399
req.payload.hws_set_scheduling_log.engine_idx = engine_idx;
400
req.payload.hws_set_scheduling_log.host_ssid = host_ssid;
401
req.payload.hws_set_scheduling_log.vpu_log_buffer_va = vpu_log_buffer_va;
402
req.payload.hws_set_scheduling_log.notify_index = 0;
403
404
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_HWS_SET_SCHEDULING_LOG_RSP, &resp,
405
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
406
if (ret)
407
ivpu_warn_ratelimited(vdev, "Failed to set scheduling log: %d\n", ret);
408
409
return ret;
410
}
411
412
int ivpu_jsm_hws_setup_priority_bands(struct ivpu_device *vdev)
413
{
414
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP };
415
struct vpu_jsm_msg resp;
416
struct ivpu_hw_info *hw = vdev->hw;
417
struct vpu_ipc_msg_payload_hws_priority_band_setup *setup =
418
&req.payload.hws_priority_band_setup;
419
int ret;
420
421
for (int band = VPU_JOB_SCHEDULING_PRIORITY_BAND_IDLE;
422
band < VPU_JOB_SCHEDULING_PRIORITY_BAND_COUNT; band++) {
423
setup->grace_period[band] = hw->hws.grace_period[band];
424
setup->process_grace_period[band] = hw->hws.process_grace_period[band];
425
setup->process_quantum[band] = hw->hws.process_quantum[band];
426
}
427
setup->normal_band_percentage = 10;
428
429
ret = ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_SET_PRIORITY_BAND_SETUP_RSP,
430
&resp, VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
431
if (ret)
432
ivpu_warn_ratelimited(vdev, "Failed to set priority bands: %d\n", ret);
433
434
return ret;
435
}
436
437
int ivpu_jsm_metric_streamer_start(struct ivpu_device *vdev, u64 metric_group_mask,
438
u64 sampling_rate, u64 buffer_addr, u64 buffer_size)
439
{
440
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_START };
441
struct vpu_jsm_msg resp;
442
int ret;
443
444
req.payload.metric_streamer_start.metric_group_mask = metric_group_mask;
445
req.payload.metric_streamer_start.sampling_rate = sampling_rate;
446
req.payload.metric_streamer_start.buffer_addr = buffer_addr;
447
req.payload.metric_streamer_start.buffer_size = buffer_size;
448
449
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_START_DONE, &resp,
450
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
451
if (ret) {
452
ivpu_warn_ratelimited(vdev, "Failed to start metric streamer: ret %d\n", ret);
453
return ret;
454
}
455
456
return ret;
457
}
458
459
int ivpu_jsm_metric_streamer_stop(struct ivpu_device *vdev, u64 metric_group_mask)
460
{
461
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_STOP };
462
struct vpu_jsm_msg resp;
463
int ret;
464
465
req.payload.metric_streamer_stop.metric_group_mask = metric_group_mask;
466
467
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_STOP_DONE, &resp,
468
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
469
if (ret)
470
ivpu_warn_ratelimited(vdev, "Failed to stop metric streamer: ret %d\n", ret);
471
472
return ret;
473
}
474
475
int ivpu_jsm_metric_streamer_update(struct ivpu_device *vdev, u64 metric_group_mask,
476
u64 buffer_addr, u64 buffer_size, u64 *bytes_written)
477
{
478
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_UPDATE };
479
struct vpu_jsm_msg resp;
480
int ret;
481
482
req.payload.metric_streamer_update.metric_group_mask = metric_group_mask;
483
req.payload.metric_streamer_update.buffer_addr = buffer_addr;
484
req.payload.metric_streamer_update.buffer_size = buffer_size;
485
486
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_UPDATE_DONE, &resp,
487
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
488
if (ret) {
489
ivpu_warn_ratelimited(vdev, "Failed to update metric streamer: ret %d\n", ret);
490
return ret;
491
}
492
493
if (buffer_size && resp.payload.metric_streamer_done.bytes_written > buffer_size) {
494
ivpu_warn_ratelimited(vdev, "MS buffer overflow: bytes_written %#llx > buffer_size %#llx\n",
495
resp.payload.metric_streamer_done.bytes_written, buffer_size);
496
return -EOVERFLOW;
497
}
498
499
*bytes_written = resp.payload.metric_streamer_done.bytes_written;
500
501
return ret;
502
}
503
504
int ivpu_jsm_metric_streamer_info(struct ivpu_device *vdev, u64 metric_group_mask, u64 buffer_addr,
505
u64 buffer_size, u32 *sample_size, u64 *info_size)
506
{
507
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_METRIC_STREAMER_INFO };
508
struct vpu_jsm_msg resp;
509
int ret;
510
511
req.payload.metric_streamer_start.metric_group_mask = metric_group_mask;
512
req.payload.metric_streamer_start.buffer_addr = buffer_addr;
513
req.payload.metric_streamer_start.buffer_size = buffer_size;
514
515
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_METRIC_STREAMER_INFO_DONE, &resp,
516
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
517
if (ret) {
518
ivpu_warn_ratelimited(vdev, "Failed to get metric streamer info: ret %d\n", ret);
519
return ret;
520
}
521
522
if (!resp.payload.metric_streamer_done.sample_size) {
523
ivpu_warn_ratelimited(vdev, "Invalid sample size\n");
524
return -EBADMSG;
525
}
526
527
if (sample_size)
528
*sample_size = resp.payload.metric_streamer_done.sample_size;
529
if (info_size)
530
*info_size = resp.payload.metric_streamer_done.bytes_written;
531
532
return ret;
533
}
534
535
int ivpu_jsm_dct_enable(struct ivpu_device *vdev, u32 active_us, u32 inactive_us)
536
{
537
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DCT_ENABLE };
538
struct vpu_jsm_msg resp;
539
540
req.payload.pwr_dct_control.dct_active_us = active_us;
541
req.payload.pwr_dct_control.dct_inactive_us = inactive_us;
542
543
return ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_DCT_ENABLE_DONE, &resp,
544
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
545
}
546
547
int ivpu_jsm_dct_disable(struct ivpu_device *vdev)
548
{
549
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_DCT_DISABLE };
550
struct vpu_jsm_msg resp;
551
552
return ivpu_ipc_send_receive_internal(vdev, &req, VPU_JSM_MSG_DCT_DISABLE_DONE, &resp,
553
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
554
}
555
556
int ivpu_jsm_state_dump(struct ivpu_device *vdev)
557
{
558
struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_STATE_DUMP };
559
560
return ivpu_ipc_send_and_wait(vdev, &req, VPU_IPC_CHAN_ASYNC_CMD,
561
vdev->timeout.state_dump_msg);
562
}
563
564