Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/sound/soc/intel/atom/sst/sst_ipc.c
26516 views
1
// SPDX-License-Identifier: GPL-2.0-only
2
/*
3
* sst_ipc.c - Intel SST Driver for audio engine
4
*
5
* Copyright (C) 2008-14 Intel Corporation
6
* Authors: Vinod Koul <[email protected]>
7
* Harsha Priya <[email protected]>
8
* Dharageswari R <[email protected]>
9
* KP Jeeja <[email protected]>
10
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
11
*
12
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
13
*/
14
#include <linux/pci.h>
15
#include <linux/firmware.h>
16
#include <linux/sched.h>
17
#include <linux/delay.h>
18
#include <sound/core.h>
19
#include <sound/pcm.h>
20
#include <sound/soc.h>
21
#include <sound/compress_driver.h>
22
23
#include <asm/platform_sst_audio.h>
24
25
#include "../sst-mfld-platform.h"
26
#include "sst.h"
27
28
struct sst_block *sst_create_block(struct intel_sst_drv *ctx,
29
u32 msg_id, u32 drv_id)
30
{
31
struct sst_block *msg;
32
33
dev_dbg(ctx->dev, "Enter\n");
34
msg = kzalloc(sizeof(*msg), GFP_KERNEL);
35
if (!msg)
36
return NULL;
37
msg->condition = false;
38
msg->on = true;
39
msg->msg_id = msg_id;
40
msg->drv_id = drv_id;
41
spin_lock_bh(&ctx->block_lock);
42
list_add_tail(&msg->node, &ctx->block_list);
43
spin_unlock_bh(&ctx->block_lock);
44
45
return msg;
46
}
47
48
/*
49
* while handling the interrupts, we need to check for message status and
50
* then if we are blocking for a message
51
*
52
* here we are unblocking the blocked ones, this is based on id we have
53
* passed and search that for block threads.
54
* We will not find block in two cases
55
* a) when its small message and block in not there, so silently ignore
56
* them
57
* b) when we are actually not able to find the block (bug perhaps)
58
*
59
* Since we have bit of small messages we can spam kernel log with err
60
* print on above so need to keep as debug prints which should be enabled
61
* via dynamic debug while debugging IPC issues
62
*/
63
int sst_wake_up_block(struct intel_sst_drv *ctx, int result,
64
u32 drv_id, u32 ipc, void *data, u32 size)
65
{
66
struct sst_block *block;
67
68
dev_dbg(ctx->dev, "Enter\n");
69
70
spin_lock_bh(&ctx->block_lock);
71
list_for_each_entry(block, &ctx->block_list, node) {
72
dev_dbg(ctx->dev, "Block ipc %d, drv_id %d\n", block->msg_id,
73
block->drv_id);
74
if (block->msg_id == ipc && block->drv_id == drv_id) {
75
dev_dbg(ctx->dev, "free up the block\n");
76
block->ret_code = result;
77
block->data = data;
78
block->size = size;
79
block->condition = true;
80
spin_unlock_bh(&ctx->block_lock);
81
wake_up(&ctx->wait_queue);
82
return 0;
83
}
84
}
85
spin_unlock_bh(&ctx->block_lock);
86
dev_dbg(ctx->dev,
87
"Block not found or a response received for a short msg for ipc %d, drv_id %d\n",
88
ipc, drv_id);
89
return -EINVAL;
90
}
91
92
int sst_free_block(struct intel_sst_drv *ctx, struct sst_block *freed)
93
{
94
struct sst_block *block, *__block;
95
96
dev_dbg(ctx->dev, "Enter\n");
97
spin_lock_bh(&ctx->block_lock);
98
list_for_each_entry_safe(block, __block, &ctx->block_list, node) {
99
if (block == freed) {
100
pr_debug("pvt_id freed --> %d\n", freed->drv_id);
101
/* toggle the index position of pvt_id */
102
list_del(&freed->node);
103
spin_unlock_bh(&ctx->block_lock);
104
kfree(freed->data);
105
freed->data = NULL;
106
kfree(freed);
107
return 0;
108
}
109
}
110
spin_unlock_bh(&ctx->block_lock);
111
dev_err(ctx->dev, "block is already freed!!!\n");
112
return -EINVAL;
113
}
114
115
int sst_post_message_mrfld(struct intel_sst_drv *sst_drv_ctx,
116
struct ipc_post *ipc_msg, bool sync)
117
{
118
struct ipc_post *msg = ipc_msg;
119
union ipc_header_mrfld header;
120
unsigned int loop_count = 0;
121
int retval = 0;
122
unsigned long irq_flags;
123
124
dev_dbg(sst_drv_ctx->dev, "Enter: sync: %d\n", sync);
125
spin_lock_irqsave(&sst_drv_ctx->ipc_spin_lock, irq_flags);
126
header.full = sst_shim_read64(sst_drv_ctx->shim, SST_IPCX);
127
if (sync) {
128
while (header.p.header_high.part.busy) {
129
if (loop_count > 25) {
130
dev_err(sst_drv_ctx->dev,
131
"sst: Busy wait failed, can't send this msg\n");
132
retval = -EBUSY;
133
goto out;
134
}
135
cpu_relax();
136
loop_count++;
137
header.full = sst_shim_read64(sst_drv_ctx->shim, SST_IPCX);
138
}
139
} else {
140
if (list_empty(&sst_drv_ctx->ipc_dispatch_list)) {
141
/* queue is empty, nothing to send */
142
spin_unlock_irqrestore(&sst_drv_ctx->ipc_spin_lock, irq_flags);
143
dev_dbg(sst_drv_ctx->dev,
144
"Empty msg queue... NO Action\n");
145
return 0;
146
}
147
148
if (header.p.header_high.part.busy) {
149
spin_unlock_irqrestore(&sst_drv_ctx->ipc_spin_lock, irq_flags);
150
dev_dbg(sst_drv_ctx->dev, "Busy not free... post later\n");
151
return 0;
152
}
153
154
/* copy msg from list */
155
msg = list_entry(sst_drv_ctx->ipc_dispatch_list.next,
156
struct ipc_post, node);
157
list_del(&msg->node);
158
}
159
dev_dbg(sst_drv_ctx->dev, "sst: Post message: header = %x\n",
160
msg->mrfld_header.p.header_high.full);
161
dev_dbg(sst_drv_ctx->dev, "sst: size = 0x%x\n",
162
msg->mrfld_header.p.header_low_payload);
163
164
if (msg->mrfld_header.p.header_high.part.large)
165
memcpy_toio(sst_drv_ctx->mailbox + SST_MAILBOX_SEND,
166
msg->mailbox_data,
167
msg->mrfld_header.p.header_low_payload);
168
169
sst_shim_write64(sst_drv_ctx->shim, SST_IPCX, msg->mrfld_header.full);
170
171
out:
172
spin_unlock_irqrestore(&sst_drv_ctx->ipc_spin_lock, irq_flags);
173
kfree(msg->mailbox_data);
174
kfree(msg);
175
return retval;
176
}
177
178
void intel_sst_clear_intr_mrfld(struct intel_sst_drv *sst_drv_ctx)
179
{
180
union interrupt_reg_mrfld isr;
181
union interrupt_reg_mrfld imr;
182
union ipc_header_mrfld clear_ipc;
183
unsigned long irq_flags;
184
185
spin_lock_irqsave(&sst_drv_ctx->ipc_spin_lock, irq_flags);
186
imr.full = sst_shim_read64(sst_drv_ctx->shim, SST_IMRX);
187
isr.full = sst_shim_read64(sst_drv_ctx->shim, SST_ISRX);
188
189
/* write 1 to clear*/
190
isr.part.busy_interrupt = 1;
191
sst_shim_write64(sst_drv_ctx->shim, SST_ISRX, isr.full);
192
193
/* Set IA done bit */
194
clear_ipc.full = sst_shim_read64(sst_drv_ctx->shim, SST_IPCD);
195
196
clear_ipc.p.header_high.part.busy = 0;
197
clear_ipc.p.header_high.part.done = 1;
198
clear_ipc.p.header_low_payload = IPC_ACK_SUCCESS;
199
sst_shim_write64(sst_drv_ctx->shim, SST_IPCD, clear_ipc.full);
200
/* un mask busy interrupt */
201
imr.part.busy_interrupt = 0;
202
sst_shim_write64(sst_drv_ctx->shim, SST_IMRX, imr.full);
203
spin_unlock_irqrestore(&sst_drv_ctx->ipc_spin_lock, irq_flags);
204
}
205
206
207
/*
208
* process_fw_init - process the FW init msg
209
*
210
* @msg: IPC message mailbox data from FW
211
*
212
* This function processes the FW init msg from FW
213
* marks FW state and prints debug info of loaded FW
214
*/
215
static void process_fw_init(struct intel_sst_drv *sst_drv_ctx,
216
void *msg)
217
{
218
struct ipc_header_fw_init *init =
219
(struct ipc_header_fw_init *)msg;
220
int retval = 0;
221
222
dev_dbg(sst_drv_ctx->dev, "*** FW Init msg came***\n");
223
if (init->result) {
224
sst_set_fw_state_locked(sst_drv_ctx, SST_RESET);
225
dev_err(sst_drv_ctx->dev, "FW Init failed, Error %x\n",
226
init->result);
227
retval = init->result;
228
goto ret;
229
}
230
if (memcmp(&sst_drv_ctx->fw_version, &init->fw_version,
231
sizeof(init->fw_version)))
232
dev_info(sst_drv_ctx->dev, "FW Version %02x.%02x.%02x.%02x\n",
233
init->fw_version.type, init->fw_version.major,
234
init->fw_version.minor, init->fw_version.build);
235
dev_dbg(sst_drv_ctx->dev, "Build date %s Time %s\n",
236
init->build_info.date, init->build_info.time);
237
238
/* Save FW version */
239
sst_drv_ctx->fw_version.type = init->fw_version.type;
240
sst_drv_ctx->fw_version.major = init->fw_version.major;
241
sst_drv_ctx->fw_version.minor = init->fw_version.minor;
242
sst_drv_ctx->fw_version.build = init->fw_version.build;
243
244
ret:
245
sst_wake_up_block(sst_drv_ctx, retval, FW_DWNL_ID, 0 , NULL, 0);
246
}
247
248
static void process_fw_async_msg(struct intel_sst_drv *sst_drv_ctx,
249
struct ipc_post *msg)
250
{
251
u32 msg_id;
252
int str_id;
253
u32 data_size, i;
254
void *data_offset;
255
struct stream_info *stream;
256
u32 msg_low, pipe_id;
257
258
msg_low = msg->mrfld_header.p.header_low_payload;
259
msg_id = ((struct ipc_dsp_hdr *)msg->mailbox_data)->cmd_id;
260
data_offset = (msg->mailbox_data + sizeof(struct ipc_dsp_hdr));
261
data_size = msg_low - (sizeof(struct ipc_dsp_hdr));
262
263
switch (msg_id) {
264
case IPC_SST_PERIOD_ELAPSED_MRFLD:
265
pipe_id = ((struct ipc_dsp_hdr *)msg->mailbox_data)->pipe_id;
266
str_id = get_stream_id_mrfld(sst_drv_ctx, pipe_id);
267
if (str_id > 0) {
268
dev_dbg(sst_drv_ctx->dev,
269
"Period elapsed rcvd for pipe id 0x%x\n",
270
pipe_id);
271
stream = &sst_drv_ctx->streams[str_id];
272
/* If stream is dropped, skip processing this message*/
273
if (stream->status == STREAM_INIT)
274
break;
275
if (stream->period_elapsed)
276
stream->period_elapsed(stream->pcm_substream);
277
if (stream->compr_cb)
278
stream->compr_cb(stream->compr_cb_param);
279
}
280
break;
281
282
case IPC_IA_DRAIN_STREAM_MRFLD:
283
pipe_id = ((struct ipc_dsp_hdr *)msg->mailbox_data)->pipe_id;
284
str_id = get_stream_id_mrfld(sst_drv_ctx, pipe_id);
285
if (str_id > 0) {
286
stream = &sst_drv_ctx->streams[str_id];
287
if (stream->drain_notify)
288
stream->drain_notify(stream->drain_cb_param);
289
}
290
break;
291
292
case IPC_IA_FW_ASYNC_ERR_MRFLD:
293
dev_err(sst_drv_ctx->dev, "FW sent async error msg:\n");
294
for (i = 0; i < (data_size/4); i++)
295
print_hex_dump(KERN_DEBUG, NULL, DUMP_PREFIX_NONE,
296
16, 4, data_offset, data_size, false);
297
break;
298
299
case IPC_IA_FW_INIT_CMPLT_MRFLD:
300
process_fw_init(sst_drv_ctx, data_offset);
301
break;
302
303
case IPC_IA_BUF_UNDER_RUN_MRFLD:
304
pipe_id = ((struct ipc_dsp_hdr *)msg->mailbox_data)->pipe_id;
305
str_id = get_stream_id_mrfld(sst_drv_ctx, pipe_id);
306
if (str_id > 0)
307
dev_err(sst_drv_ctx->dev,
308
"Buffer under-run for pipe:%#x str_id:%d\n",
309
pipe_id, str_id);
310
break;
311
312
default:
313
dev_err(sst_drv_ctx->dev,
314
"Unrecognized async msg from FW msg_id %#x\n", msg_id);
315
}
316
}
317
318
void sst_process_reply_mrfld(struct intel_sst_drv *sst_drv_ctx,
319
struct ipc_post *msg)
320
{
321
unsigned int drv_id;
322
void *data;
323
union ipc_header_high msg_high;
324
u32 msg_low;
325
struct ipc_dsp_hdr *dsp_hdr;
326
327
msg_high = msg->mrfld_header.p.header_high;
328
msg_low = msg->mrfld_header.p.header_low_payload;
329
330
dev_dbg(sst_drv_ctx->dev, "IPC process message header %x payload %x\n",
331
msg->mrfld_header.p.header_high.full,
332
msg->mrfld_header.p.header_low_payload);
333
334
drv_id = msg_high.part.drv_id;
335
336
/* Check for async messages first */
337
if (drv_id == SST_ASYNC_DRV_ID) {
338
/*FW sent async large message*/
339
process_fw_async_msg(sst_drv_ctx, msg);
340
return;
341
}
342
343
/* FW sent short error response for an IPC */
344
if (msg_high.part.result && !msg_high.part.large) {
345
/* 32-bit FW error code in msg_low */
346
dev_err(sst_drv_ctx->dev, "FW sent error response 0x%x", msg_low);
347
sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
348
msg_high.part.drv_id,
349
msg_high.part.msg_id, NULL, 0);
350
return;
351
}
352
353
/*
354
* Process all valid responses
355
* if it is a large message, the payload contains the size to
356
* copy from mailbox
357
**/
358
if (msg_high.part.large) {
359
data = kmemdup((void *)msg->mailbox_data, msg_low, GFP_KERNEL);
360
if (!data)
361
return;
362
/* Copy command id so that we can use to put sst to reset */
363
dsp_hdr = (struct ipc_dsp_hdr *)data;
364
dev_dbg(sst_drv_ctx->dev, "cmd_id %d\n", dsp_hdr->cmd_id);
365
if (sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
366
msg_high.part.drv_id,
367
msg_high.part.msg_id, data, msg_low))
368
kfree(data);
369
} else {
370
sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
371
msg_high.part.drv_id,
372
msg_high.part.msg_id, NULL, 0);
373
}
374
375
}
376
377