Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/sound/soc/soc-generic-dmaengine-pcm.c
26381 views
1
// SPDX-License-Identifier: GPL-2.0+
2
//
3
// Copyright (C) 2013, Analog Devices Inc.
4
// Author: Lars-Peter Clausen <[email protected]>
5
6
#include <linux/module.h>
7
#include <linux/init.h>
8
#include <linux/dmaengine.h>
9
#include <linux/slab.h>
10
#include <sound/pcm.h>
11
#include <sound/pcm_params.h>
12
#include <sound/soc.h>
13
#include <linux/dma-mapping.h>
14
#include <linux/of.h>
15
16
#include <sound/dmaengine_pcm.h>
17
18
static unsigned int prealloc_buffer_size_kbytes = 512;
19
module_param(prealloc_buffer_size_kbytes, uint, 0444);
20
MODULE_PARM_DESC(prealloc_buffer_size_kbytes, "Preallocate DMA buffer size (KB).");
21
22
/*
23
* The platforms dmaengine driver does not support reporting the amount of
24
* bytes that are still left to transfer.
25
*/
26
#define SND_DMAENGINE_PCM_FLAG_NO_RESIDUE BIT(31)
27
28
static struct device *dmaengine_dma_dev(struct dmaengine_pcm *pcm,
29
struct snd_pcm_substream *substream)
30
{
31
if (!pcm->chan[substream->stream])
32
return NULL;
33
34
return pcm->chan[substream->stream]->device->dev;
35
}
36
37
/**
38
* snd_dmaengine_pcm_prepare_slave_config() - Generic prepare_slave_config callback
39
* @substream: PCM substream
40
* @params: hw_params
41
* @slave_config: DMA slave config to prepare
42
*
43
* This function can be used as a generic prepare_slave_config callback for
44
* platforms which make use of the snd_dmaengine_dai_dma_data struct for their
45
* DAI DMA data. Internally the function will first call
46
* snd_hwparams_to_dma_slave_config to fill in the slave config based on the
47
* hw_params, followed by snd_dmaengine_pcm_set_config_from_dai_data to fill in
48
* the remaining fields based on the DAI DMA data.
49
*/
50
int snd_dmaengine_pcm_prepare_slave_config(struct snd_pcm_substream *substream,
51
struct snd_pcm_hw_params *params, struct dma_slave_config *slave_config)
52
{
53
struct snd_soc_pcm_runtime *rtd = snd_soc_substream_to_rtd(substream);
54
struct snd_dmaengine_dai_dma_data *dma_data;
55
int ret;
56
57
if (rtd->dai_link->num_cpus > 1) {
58
dev_err(rtd->dev,
59
"%s doesn't support Multi CPU yet\n", __func__);
60
return -EINVAL;
61
}
62
63
dma_data = snd_soc_dai_get_dma_data(snd_soc_rtd_to_cpu(rtd, 0), substream);
64
65
ret = snd_hwparams_to_dma_slave_config(substream, params, slave_config);
66
if (ret)
67
return ret;
68
69
snd_dmaengine_pcm_set_config_from_dai_data(substream, dma_data,
70
slave_config);
71
72
return 0;
73
}
74
EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_prepare_slave_config);
75
76
static int dmaengine_pcm_hw_params(struct snd_soc_component *component,
77
struct snd_pcm_substream *substream,
78
struct snd_pcm_hw_params *params)
79
{
80
struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
81
struct dma_chan *chan = snd_dmaengine_pcm_get_chan(substream);
82
struct dma_slave_config slave_config;
83
int ret;
84
85
if (!pcm->config->prepare_slave_config)
86
return 0;
87
88
memset(&slave_config, 0, sizeof(slave_config));
89
90
ret = pcm->config->prepare_slave_config(substream, params, &slave_config);
91
if (ret)
92
return ret;
93
94
return dmaengine_slave_config(chan, &slave_config);
95
}
96
97
static int
98
dmaengine_pcm_set_runtime_hwparams(struct snd_soc_component *component,
99
struct snd_pcm_substream *substream)
100
{
101
struct snd_soc_pcm_runtime *rtd = snd_soc_substream_to_rtd(substream);
102
struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
103
struct device *dma_dev = dmaengine_dma_dev(pcm, substream);
104
struct dma_chan *chan = pcm->chan[substream->stream];
105
struct snd_dmaengine_dai_dma_data *dma_data;
106
struct snd_pcm_hardware hw;
107
108
if (rtd->dai_link->num_cpus > 1) {
109
dev_err(rtd->dev,
110
"%s doesn't support Multi CPU yet\n", __func__);
111
return -EINVAL;
112
}
113
114
if (pcm->config->pcm_hardware)
115
return snd_soc_set_runtime_hwparams(substream,
116
pcm->config->pcm_hardware);
117
118
dma_data = snd_soc_dai_get_dma_data(snd_soc_rtd_to_cpu(rtd, 0), substream);
119
120
memset(&hw, 0, sizeof(hw));
121
hw.info = SNDRV_PCM_INFO_MMAP | SNDRV_PCM_INFO_MMAP_VALID |
122
SNDRV_PCM_INFO_INTERLEAVED;
123
hw.periods_min = 2;
124
hw.periods_max = UINT_MAX;
125
hw.period_bytes_min = dma_data->maxburst * DMA_SLAVE_BUSWIDTH_8_BYTES;
126
if (!hw.period_bytes_min)
127
hw.period_bytes_min = 256;
128
hw.period_bytes_max = dma_get_max_seg_size(dma_dev);
129
hw.buffer_bytes_max = SIZE_MAX;
130
hw.fifo_size = dma_data->fifo_size;
131
132
if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE)
133
hw.info |= SNDRV_PCM_INFO_BATCH;
134
135
/**
136
* FIXME: Remove the return value check to align with the code
137
* before adding snd_dmaengine_pcm_refine_runtime_hwparams
138
* function.
139
*/
140
snd_dmaengine_pcm_refine_runtime_hwparams(substream,
141
dma_data,
142
&hw,
143
chan);
144
145
return snd_soc_set_runtime_hwparams(substream, &hw);
146
}
147
148
static int dmaengine_pcm_open(struct snd_soc_component *component,
149
struct snd_pcm_substream *substream)
150
{
151
struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
152
struct dma_chan *chan = pcm->chan[substream->stream];
153
int ret;
154
155
ret = dmaengine_pcm_set_runtime_hwparams(component, substream);
156
if (ret)
157
return ret;
158
159
return snd_dmaengine_pcm_open(substream, chan);
160
}
161
162
static int dmaengine_pcm_close(struct snd_soc_component *component,
163
struct snd_pcm_substream *substream)
164
{
165
return snd_dmaengine_pcm_close(substream);
166
}
167
168
static int dmaengine_pcm_trigger(struct snd_soc_component *component,
169
struct snd_pcm_substream *substream, int cmd)
170
{
171
return snd_dmaengine_pcm_trigger(substream, cmd);
172
}
173
174
static struct dma_chan *dmaengine_pcm_compat_request_channel(
175
struct snd_soc_component *component,
176
struct snd_soc_pcm_runtime *rtd,
177
struct snd_pcm_substream *substream)
178
{
179
struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
180
struct snd_dmaengine_dai_dma_data *dma_data;
181
182
if (rtd->dai_link->num_cpus > 1) {
183
dev_err(rtd->dev,
184
"%s doesn't support Multi CPU yet\n", __func__);
185
return NULL;
186
}
187
188
dma_data = snd_soc_dai_get_dma_data(snd_soc_rtd_to_cpu(rtd, 0), substream);
189
190
if ((pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) && pcm->chan[0])
191
return pcm->chan[0];
192
193
if (pcm->config->compat_request_channel)
194
return pcm->config->compat_request_channel(rtd, substream);
195
196
return snd_dmaengine_pcm_request_channel(pcm->config->compat_filter_fn,
197
dma_data->filter_data);
198
}
199
200
static bool dmaengine_pcm_can_report_residue(struct device *dev,
201
struct dma_chan *chan)
202
{
203
struct dma_slave_caps dma_caps;
204
int ret;
205
206
ret = dma_get_slave_caps(chan, &dma_caps);
207
if (ret != 0) {
208
dev_warn(dev, "Failed to get DMA channel capabilities, falling back to period counting: %d\n",
209
ret);
210
return false;
211
}
212
213
if (dma_caps.residue_granularity == DMA_RESIDUE_GRANULARITY_DESCRIPTOR)
214
return false;
215
216
return true;
217
}
218
219
static int dmaengine_pcm_new(struct snd_soc_component *component,
220
struct snd_soc_pcm_runtime *rtd)
221
{
222
struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
223
const struct snd_dmaengine_pcm_config *config = pcm->config;
224
struct device *dev = component->dev;
225
size_t prealloc_buffer_size;
226
size_t max_buffer_size;
227
unsigned int i;
228
229
if (config->prealloc_buffer_size)
230
prealloc_buffer_size = config->prealloc_buffer_size;
231
else
232
prealloc_buffer_size = prealloc_buffer_size_kbytes * 1024;
233
234
if (config->pcm_hardware && config->pcm_hardware->buffer_bytes_max)
235
max_buffer_size = config->pcm_hardware->buffer_bytes_max;
236
else
237
max_buffer_size = SIZE_MAX;
238
239
for_each_pcm_streams(i) {
240
struct snd_pcm_substream *substream = rtd->pcm->streams[i].substream;
241
if (!substream)
242
continue;
243
244
if (!pcm->chan[i] && config->chan_names[i])
245
pcm->chan[i] = dma_request_slave_channel(dev,
246
config->chan_names[i]);
247
248
if (!pcm->chan[i] && (pcm->flags & SND_DMAENGINE_PCM_FLAG_COMPAT)) {
249
pcm->chan[i] = dmaengine_pcm_compat_request_channel(
250
component, rtd, substream);
251
}
252
253
if (!pcm->chan[i]) {
254
dev_err(component->dev,
255
"Missing dma channel for stream: %d\n", i);
256
return -EINVAL;
257
}
258
259
snd_pcm_set_managed_buffer(substream,
260
SNDRV_DMA_TYPE_DEV_IRAM,
261
dmaengine_dma_dev(pcm, substream),
262
prealloc_buffer_size,
263
max_buffer_size);
264
265
if (!dmaengine_pcm_can_report_residue(dev, pcm->chan[i]))
266
pcm->flags |= SND_DMAENGINE_PCM_FLAG_NO_RESIDUE;
267
268
if (rtd->pcm->streams[i].pcm->name[0] == '\0') {
269
strscpy_pad(rtd->pcm->streams[i].pcm->name,
270
rtd->pcm->streams[i].pcm->id,
271
sizeof(rtd->pcm->streams[i].pcm->name));
272
}
273
}
274
275
return 0;
276
}
277
278
static snd_pcm_uframes_t dmaengine_pcm_pointer(
279
struct snd_soc_component *component,
280
struct snd_pcm_substream *substream)
281
{
282
struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
283
284
if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE)
285
return snd_dmaengine_pcm_pointer_no_residue(substream);
286
else
287
return snd_dmaengine_pcm_pointer(substream);
288
}
289
290
static int dmaengine_copy(struct snd_soc_component *component,
291
struct snd_pcm_substream *substream,
292
int channel, unsigned long hwoff,
293
struct iov_iter *iter, unsigned long bytes)
294
{
295
struct snd_pcm_runtime *runtime = substream->runtime;
296
struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
297
int (*process)(struct snd_pcm_substream *substream,
298
int channel, unsigned long hwoff,
299
unsigned long bytes) = pcm->config->process;
300
bool is_playback = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
301
void *dma_ptr = runtime->dma_area + hwoff +
302
channel * (runtime->dma_bytes / runtime->channels);
303
304
if (is_playback)
305
if (copy_from_iter(dma_ptr, bytes, iter) != bytes)
306
return -EFAULT;
307
308
if (process) {
309
int ret = process(substream, channel, hwoff, bytes);
310
if (ret < 0)
311
return ret;
312
}
313
314
if (!is_playback)
315
if (copy_to_iter(dma_ptr, bytes, iter) != bytes)
316
return -EFAULT;
317
318
return 0;
319
}
320
321
static int dmaengine_pcm_sync_stop(struct snd_soc_component *component,
322
struct snd_pcm_substream *substream)
323
{
324
return snd_dmaengine_pcm_sync_stop(substream);
325
}
326
327
static const struct snd_soc_component_driver dmaengine_pcm_component = {
328
.name = SND_DMAENGINE_PCM_DRV_NAME,
329
.probe_order = SND_SOC_COMP_ORDER_LATE,
330
.open = dmaengine_pcm_open,
331
.close = dmaengine_pcm_close,
332
.hw_params = dmaengine_pcm_hw_params,
333
.trigger = dmaengine_pcm_trigger,
334
.pointer = dmaengine_pcm_pointer,
335
.pcm_construct = dmaengine_pcm_new,
336
.sync_stop = dmaengine_pcm_sync_stop,
337
};
338
339
static const struct snd_soc_component_driver dmaengine_pcm_component_process = {
340
.name = SND_DMAENGINE_PCM_DRV_NAME,
341
.probe_order = SND_SOC_COMP_ORDER_LATE,
342
.open = dmaengine_pcm_open,
343
.close = dmaengine_pcm_close,
344
.hw_params = dmaengine_pcm_hw_params,
345
.trigger = dmaengine_pcm_trigger,
346
.pointer = dmaengine_pcm_pointer,
347
.copy = dmaengine_copy,
348
.pcm_construct = dmaengine_pcm_new,
349
.sync_stop = dmaengine_pcm_sync_stop,
350
};
351
352
static const char * const dmaengine_pcm_dma_channel_names[] = {
353
[SNDRV_PCM_STREAM_PLAYBACK] = "tx",
354
[SNDRV_PCM_STREAM_CAPTURE] = "rx",
355
};
356
357
static int dmaengine_pcm_request_chan_of(struct dmaengine_pcm *pcm,
358
struct device *dev, const struct snd_dmaengine_pcm_config *config)
359
{
360
unsigned int i;
361
const char *name;
362
struct dma_chan *chan;
363
364
if ((pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_DT) || (!dev->of_node &&
365
!(config->dma_dev && config->dma_dev->of_node)))
366
return 0;
367
368
if (config->dma_dev) {
369
/*
370
* If this warning is seen, it probably means that your Linux
371
* device structure does not match your HW device structure.
372
* It would be best to refactor the Linux device structure to
373
* correctly match the HW structure.
374
*/
375
dev_warn(dev, "DMA channels sourced from device %s",
376
dev_name(config->dma_dev));
377
dev = config->dma_dev;
378
}
379
380
for_each_pcm_streams(i) {
381
if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
382
name = "rx-tx";
383
else
384
name = dmaengine_pcm_dma_channel_names[i];
385
if (config->chan_names[i])
386
name = config->chan_names[i];
387
chan = dma_request_chan(dev, name);
388
if (IS_ERR(chan)) {
389
/*
390
* Only report probe deferral errors, channels
391
* might not be present for devices that
392
* support only TX or only RX.
393
*/
394
if (PTR_ERR(chan) == -EPROBE_DEFER)
395
return -EPROBE_DEFER;
396
pcm->chan[i] = NULL;
397
} else {
398
pcm->chan[i] = chan;
399
}
400
if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
401
break;
402
}
403
404
if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
405
pcm->chan[1] = pcm->chan[0];
406
407
return 0;
408
}
409
410
static void dmaengine_pcm_release_chan(struct dmaengine_pcm *pcm)
411
{
412
unsigned int i;
413
414
for_each_pcm_streams(i) {
415
if (!pcm->chan[i])
416
continue;
417
dma_release_channel(pcm->chan[i]);
418
if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
419
break;
420
}
421
}
422
423
static const struct snd_dmaengine_pcm_config snd_dmaengine_pcm_default_config = {
424
.prepare_slave_config = snd_dmaengine_pcm_prepare_slave_config,
425
};
426
427
/**
428
* snd_dmaengine_pcm_register - Register a dmaengine based PCM device
429
* @dev: The parent device for the PCM device
430
* @config: Platform specific PCM configuration
431
* @flags: Platform specific quirks
432
*/
433
int snd_dmaengine_pcm_register(struct device *dev,
434
const struct snd_dmaengine_pcm_config *config, unsigned int flags)
435
{
436
const struct snd_soc_component_driver *driver;
437
struct dmaengine_pcm *pcm;
438
int ret;
439
440
pcm = kzalloc(sizeof(*pcm), GFP_KERNEL);
441
if (!pcm)
442
return -ENOMEM;
443
444
#ifdef CONFIG_DEBUG_FS
445
pcm->component.debugfs_prefix = "dma";
446
#endif
447
if (!config)
448
config = &snd_dmaengine_pcm_default_config;
449
pcm->config = config;
450
pcm->flags = flags;
451
452
if (config->name)
453
pcm->component.name = config->name;
454
455
ret = dmaengine_pcm_request_chan_of(pcm, dev, config);
456
if (ret)
457
goto err_free_dma;
458
459
if (config->process)
460
driver = &dmaengine_pcm_component_process;
461
else
462
driver = &dmaengine_pcm_component;
463
464
ret = snd_soc_component_initialize(&pcm->component, driver, dev);
465
if (ret)
466
goto err_free_dma;
467
468
ret = snd_soc_add_component(&pcm->component, NULL, 0);
469
if (ret)
470
goto err_free_dma;
471
472
return 0;
473
474
err_free_dma:
475
dmaengine_pcm_release_chan(pcm);
476
kfree(pcm);
477
return ret;
478
}
479
EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_register);
480
481
/**
482
* snd_dmaengine_pcm_unregister - Removes a dmaengine based PCM device
483
* @dev: Parent device the PCM was register with
484
*
485
* Removes a dmaengine based PCM device previously registered with
486
* snd_dmaengine_pcm_register.
487
*/
488
void snd_dmaengine_pcm_unregister(struct device *dev)
489
{
490
struct snd_soc_component *component;
491
struct dmaengine_pcm *pcm;
492
493
component = snd_soc_lookup_component(dev, SND_DMAENGINE_PCM_DRV_NAME);
494
if (!component)
495
return;
496
497
pcm = soc_component_to_pcm(component);
498
499
snd_soc_unregister_component_by_driver(dev, component->driver);
500
dmaengine_pcm_release_chan(pcm);
501
kfree(pcm);
502
}
503
EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_unregister);
504
505
MODULE_DESCRIPTION("ASoC helpers for generic PCM dmaengine API");
506
MODULE_LICENSE("GPL");
507
508