Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
awilliam
GitHub Repository: awilliam/linux-vfio
Path: blob/master/sound/core/seq/seq_clientmgr.c
10817 views
1
/*
2
* ALSA sequencer Client Manager
3
* Copyright (c) 1998-2001 by Frank van de Pol <[email protected]>
4
* Jaroslav Kysela <[email protected]>
5
* Takashi Iwai <[email protected]>
6
*
7
*
8
* This program is free software; you can redistribute it and/or modify
9
* it under the terms of the GNU General Public License as published by
10
* the Free Software Foundation; either version 2 of the License, or
11
* (at your option) any later version.
12
*
13
* This program is distributed in the hope that it will be useful,
14
* but WITHOUT ANY WARRANTY; without even the implied warranty of
15
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16
* GNU General Public License for more details.
17
*
18
* You should have received a copy of the GNU General Public License
19
* along with this program; if not, write to the Free Software
20
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
21
*
22
*/
23
24
#include <linux/init.h>
25
#include <linux/slab.h>
26
#include <sound/core.h>
27
#include <sound/minors.h>
28
#include <linux/kmod.h>
29
30
#include <sound/seq_kernel.h>
31
#include "seq_clientmgr.h"
32
#include "seq_memory.h"
33
#include "seq_queue.h"
34
#include "seq_timer.h"
35
#include "seq_info.h"
36
#include "seq_system.h"
37
#include <sound/seq_device.h>
38
#ifdef CONFIG_COMPAT
39
#include <linux/compat.h>
40
#endif
41
42
/* Client Manager
43
44
* this module handles the connections of userland and kernel clients
45
*
46
*/
47
48
/*
49
* There are four ranges of client numbers (last two shared):
50
* 0..15: global clients
51
* 16..127: statically allocated client numbers for cards 0..27
52
* 128..191: dynamically allocated client numbers for cards 28..31
53
* 128..191: dynamically allocated client numbers for applications
54
*/
55
56
/* number of kernel non-card clients */
57
#define SNDRV_SEQ_GLOBAL_CLIENTS 16
58
/* clients per cards, for static clients */
59
#define SNDRV_SEQ_CLIENTS_PER_CARD 4
60
/* dynamically allocated client numbers (both kernel drivers and user space) */
61
#define SNDRV_SEQ_DYNAMIC_CLIENTS_BEGIN 128
62
63
#define SNDRV_SEQ_LFLG_INPUT 0x0001
64
#define SNDRV_SEQ_LFLG_OUTPUT 0x0002
65
#define SNDRV_SEQ_LFLG_OPEN (SNDRV_SEQ_LFLG_INPUT|SNDRV_SEQ_LFLG_OUTPUT)
66
67
static DEFINE_SPINLOCK(clients_lock);
68
static DEFINE_MUTEX(register_mutex);
69
70
/*
71
* client table
72
*/
73
static char clienttablock[SNDRV_SEQ_MAX_CLIENTS];
74
static struct snd_seq_client *clienttab[SNDRV_SEQ_MAX_CLIENTS];
75
static struct snd_seq_usage client_usage;
76
77
/*
78
* prototypes
79
*/
80
static int bounce_error_event(struct snd_seq_client *client,
81
struct snd_seq_event *event,
82
int err, int atomic, int hop);
83
static int snd_seq_deliver_single_event(struct snd_seq_client *client,
84
struct snd_seq_event *event,
85
int filter, int atomic, int hop);
86
87
/*
88
*/
89
90
static inline mm_segment_t snd_enter_user(void)
91
{
92
mm_segment_t fs = get_fs();
93
set_fs(get_ds());
94
return fs;
95
}
96
97
static inline void snd_leave_user(mm_segment_t fs)
98
{
99
set_fs(fs);
100
}
101
102
/*
103
*/
104
static inline unsigned short snd_seq_file_flags(struct file *file)
105
{
106
switch (file->f_mode & (FMODE_READ | FMODE_WRITE)) {
107
case FMODE_WRITE:
108
return SNDRV_SEQ_LFLG_OUTPUT;
109
case FMODE_READ:
110
return SNDRV_SEQ_LFLG_INPUT;
111
default:
112
return SNDRV_SEQ_LFLG_OPEN;
113
}
114
}
115
116
static inline int snd_seq_write_pool_allocated(struct snd_seq_client *client)
117
{
118
return snd_seq_total_cells(client->pool) > 0;
119
}
120
121
/* return pointer to client structure for specified id */
122
static struct snd_seq_client *clientptr(int clientid)
123
{
124
if (clientid < 0 || clientid >= SNDRV_SEQ_MAX_CLIENTS) {
125
snd_printd("Seq: oops. Trying to get pointer to client %d\n",
126
clientid);
127
return NULL;
128
}
129
return clienttab[clientid];
130
}
131
132
struct snd_seq_client *snd_seq_client_use_ptr(int clientid)
133
{
134
unsigned long flags;
135
struct snd_seq_client *client;
136
137
if (clientid < 0 || clientid >= SNDRV_SEQ_MAX_CLIENTS) {
138
snd_printd("Seq: oops. Trying to get pointer to client %d\n",
139
clientid);
140
return NULL;
141
}
142
spin_lock_irqsave(&clients_lock, flags);
143
client = clientptr(clientid);
144
if (client)
145
goto __lock;
146
if (clienttablock[clientid]) {
147
spin_unlock_irqrestore(&clients_lock, flags);
148
return NULL;
149
}
150
spin_unlock_irqrestore(&clients_lock, flags);
151
#ifdef CONFIG_MODULES
152
if (!in_interrupt()) {
153
static char client_requested[SNDRV_SEQ_GLOBAL_CLIENTS];
154
static char card_requested[SNDRV_CARDS];
155
if (clientid < SNDRV_SEQ_GLOBAL_CLIENTS) {
156
int idx;
157
158
if (!client_requested[clientid]) {
159
client_requested[clientid] = 1;
160
for (idx = 0; idx < 15; idx++) {
161
if (seq_client_load[idx] < 0)
162
break;
163
if (seq_client_load[idx] == clientid) {
164
request_module("snd-seq-client-%i",
165
clientid);
166
break;
167
}
168
}
169
}
170
} else if (clientid < SNDRV_SEQ_DYNAMIC_CLIENTS_BEGIN) {
171
int card = (clientid - SNDRV_SEQ_GLOBAL_CLIENTS) /
172
SNDRV_SEQ_CLIENTS_PER_CARD;
173
if (card < snd_ecards_limit) {
174
if (! card_requested[card]) {
175
card_requested[card] = 1;
176
snd_request_card(card);
177
}
178
snd_seq_device_load_drivers();
179
}
180
}
181
spin_lock_irqsave(&clients_lock, flags);
182
client = clientptr(clientid);
183
if (client)
184
goto __lock;
185
spin_unlock_irqrestore(&clients_lock, flags);
186
}
187
#endif
188
return NULL;
189
190
__lock:
191
snd_use_lock_use(&client->use_lock);
192
spin_unlock_irqrestore(&clients_lock, flags);
193
return client;
194
}
195
196
static void usage_alloc(struct snd_seq_usage *res, int num)
197
{
198
res->cur += num;
199
if (res->cur > res->peak)
200
res->peak = res->cur;
201
}
202
203
static void usage_free(struct snd_seq_usage *res, int num)
204
{
205
res->cur -= num;
206
}
207
208
/* initialise data structures */
209
int __init client_init_data(void)
210
{
211
/* zap out the client table */
212
memset(&clienttablock, 0, sizeof(clienttablock));
213
memset(&clienttab, 0, sizeof(clienttab));
214
return 0;
215
}
216
217
218
static struct snd_seq_client *seq_create_client1(int client_index, int poolsize)
219
{
220
unsigned long flags;
221
int c;
222
struct snd_seq_client *client;
223
224
/* init client data */
225
client = kzalloc(sizeof(*client), GFP_KERNEL);
226
if (client == NULL)
227
return NULL;
228
client->pool = snd_seq_pool_new(poolsize);
229
if (client->pool == NULL) {
230
kfree(client);
231
return NULL;
232
}
233
client->type = NO_CLIENT;
234
snd_use_lock_init(&client->use_lock);
235
rwlock_init(&client->ports_lock);
236
mutex_init(&client->ports_mutex);
237
INIT_LIST_HEAD(&client->ports_list_head);
238
239
/* find free slot in the client table */
240
spin_lock_irqsave(&clients_lock, flags);
241
if (client_index < 0) {
242
for (c = SNDRV_SEQ_DYNAMIC_CLIENTS_BEGIN;
243
c < SNDRV_SEQ_MAX_CLIENTS;
244
c++) {
245
if (clienttab[c] || clienttablock[c])
246
continue;
247
clienttab[client->number = c] = client;
248
spin_unlock_irqrestore(&clients_lock, flags);
249
return client;
250
}
251
} else {
252
if (clienttab[client_index] == NULL && !clienttablock[client_index]) {
253
clienttab[client->number = client_index] = client;
254
spin_unlock_irqrestore(&clients_lock, flags);
255
return client;
256
}
257
}
258
spin_unlock_irqrestore(&clients_lock, flags);
259
snd_seq_pool_delete(&client->pool);
260
kfree(client);
261
return NULL; /* no free slot found or busy, return failure code */
262
}
263
264
265
static int seq_free_client1(struct snd_seq_client *client)
266
{
267
unsigned long flags;
268
269
if (!client)
270
return 0;
271
snd_seq_delete_all_ports(client);
272
snd_seq_queue_client_leave(client->number);
273
spin_lock_irqsave(&clients_lock, flags);
274
clienttablock[client->number] = 1;
275
clienttab[client->number] = NULL;
276
spin_unlock_irqrestore(&clients_lock, flags);
277
snd_use_lock_sync(&client->use_lock);
278
snd_seq_queue_client_termination(client->number);
279
if (client->pool)
280
snd_seq_pool_delete(&client->pool);
281
spin_lock_irqsave(&clients_lock, flags);
282
clienttablock[client->number] = 0;
283
spin_unlock_irqrestore(&clients_lock, flags);
284
return 0;
285
}
286
287
288
static void seq_free_client(struct snd_seq_client * client)
289
{
290
mutex_lock(&register_mutex);
291
switch (client->type) {
292
case NO_CLIENT:
293
snd_printk(KERN_WARNING "Seq: Trying to free unused client %d\n",
294
client->number);
295
break;
296
case USER_CLIENT:
297
case KERNEL_CLIENT:
298
seq_free_client1(client);
299
usage_free(&client_usage, 1);
300
break;
301
302
default:
303
snd_printk(KERN_ERR "Seq: Trying to free client %d with undefined type = %d\n",
304
client->number, client->type);
305
}
306
mutex_unlock(&register_mutex);
307
308
snd_seq_system_client_ev_client_exit(client->number);
309
}
310
311
312
313
/* -------------------------------------------------------- */
314
315
/* create a user client */
316
static int snd_seq_open(struct inode *inode, struct file *file)
317
{
318
int c, mode; /* client id */
319
struct snd_seq_client *client;
320
struct snd_seq_user_client *user;
321
int err;
322
323
err = nonseekable_open(inode, file);
324
if (err < 0)
325
return err;
326
327
if (mutex_lock_interruptible(&register_mutex))
328
return -ERESTARTSYS;
329
client = seq_create_client1(-1, SNDRV_SEQ_DEFAULT_EVENTS);
330
if (client == NULL) {
331
mutex_unlock(&register_mutex);
332
return -ENOMEM; /* failure code */
333
}
334
335
mode = snd_seq_file_flags(file);
336
if (mode & SNDRV_SEQ_LFLG_INPUT)
337
client->accept_input = 1;
338
if (mode & SNDRV_SEQ_LFLG_OUTPUT)
339
client->accept_output = 1;
340
341
user = &client->data.user;
342
user->fifo = NULL;
343
user->fifo_pool_size = 0;
344
345
if (mode & SNDRV_SEQ_LFLG_INPUT) {
346
user->fifo_pool_size = SNDRV_SEQ_DEFAULT_CLIENT_EVENTS;
347
user->fifo = snd_seq_fifo_new(user->fifo_pool_size);
348
if (user->fifo == NULL) {
349
seq_free_client1(client);
350
kfree(client);
351
mutex_unlock(&register_mutex);
352
return -ENOMEM;
353
}
354
}
355
356
usage_alloc(&client_usage, 1);
357
client->type = USER_CLIENT;
358
mutex_unlock(&register_mutex);
359
360
c = client->number;
361
file->private_data = client;
362
363
/* fill client data */
364
user->file = file;
365
sprintf(client->name, "Client-%d", c);
366
367
/* make others aware this new client */
368
snd_seq_system_client_ev_client_start(c);
369
370
return 0;
371
}
372
373
/* delete a user client */
374
static int snd_seq_release(struct inode *inode, struct file *file)
375
{
376
struct snd_seq_client *client = file->private_data;
377
378
if (client) {
379
seq_free_client(client);
380
if (client->data.user.fifo)
381
snd_seq_fifo_delete(&client->data.user.fifo);
382
kfree(client);
383
}
384
385
return 0;
386
}
387
388
389
/* handle client read() */
390
/* possible error values:
391
* -ENXIO invalid client or file open mode
392
* -ENOSPC FIFO overflow (the flag is cleared after this error report)
393
* -EINVAL no enough user-space buffer to write the whole event
394
* -EFAULT seg. fault during copy to user space
395
*/
396
static ssize_t snd_seq_read(struct file *file, char __user *buf, size_t count,
397
loff_t *offset)
398
{
399
struct snd_seq_client *client = file->private_data;
400
struct snd_seq_fifo *fifo;
401
int err;
402
long result = 0;
403
struct snd_seq_event_cell *cell;
404
405
if (!(snd_seq_file_flags(file) & SNDRV_SEQ_LFLG_INPUT))
406
return -ENXIO;
407
408
if (!access_ok(VERIFY_WRITE, buf, count))
409
return -EFAULT;
410
411
/* check client structures are in place */
412
if (snd_BUG_ON(!client))
413
return -ENXIO;
414
415
if (!client->accept_input || (fifo = client->data.user.fifo) == NULL)
416
return -ENXIO;
417
418
if (atomic_read(&fifo->overflow) > 0) {
419
/* buffer overflow is detected */
420
snd_seq_fifo_clear(fifo);
421
/* return error code */
422
return -ENOSPC;
423
}
424
425
cell = NULL;
426
err = 0;
427
snd_seq_fifo_lock(fifo);
428
429
/* while data available in queue */
430
while (count >= sizeof(struct snd_seq_event)) {
431
int nonblock;
432
433
nonblock = (file->f_flags & O_NONBLOCK) || result > 0;
434
if ((err = snd_seq_fifo_cell_out(fifo, &cell, nonblock)) < 0) {
435
break;
436
}
437
if (snd_seq_ev_is_variable(&cell->event)) {
438
struct snd_seq_event tmpev;
439
tmpev = cell->event;
440
tmpev.data.ext.len &= ~SNDRV_SEQ_EXT_MASK;
441
if (copy_to_user(buf, &tmpev, sizeof(struct snd_seq_event))) {
442
err = -EFAULT;
443
break;
444
}
445
count -= sizeof(struct snd_seq_event);
446
buf += sizeof(struct snd_seq_event);
447
err = snd_seq_expand_var_event(&cell->event, count,
448
(char __force *)buf, 0,
449
sizeof(struct snd_seq_event));
450
if (err < 0)
451
break;
452
result += err;
453
count -= err;
454
buf += err;
455
} else {
456
if (copy_to_user(buf, &cell->event, sizeof(struct snd_seq_event))) {
457
err = -EFAULT;
458
break;
459
}
460
count -= sizeof(struct snd_seq_event);
461
buf += sizeof(struct snd_seq_event);
462
}
463
snd_seq_cell_free(cell);
464
cell = NULL; /* to be sure */
465
result += sizeof(struct snd_seq_event);
466
}
467
468
if (err < 0) {
469
if (cell)
470
snd_seq_fifo_cell_putback(fifo, cell);
471
if (err == -EAGAIN && result > 0)
472
err = 0;
473
}
474
snd_seq_fifo_unlock(fifo);
475
476
return (err < 0) ? err : result;
477
}
478
479
480
/*
481
* check access permission to the port
482
*/
483
static int check_port_perm(struct snd_seq_client_port *port, unsigned int flags)
484
{
485
if ((port->capability & flags) != flags)
486
return 0;
487
return flags;
488
}
489
490
/*
491
* check if the destination client is available, and return the pointer
492
* if filter is non-zero, client filter bitmap is tested.
493
*/
494
static struct snd_seq_client *get_event_dest_client(struct snd_seq_event *event,
495
int filter)
496
{
497
struct snd_seq_client *dest;
498
499
dest = snd_seq_client_use_ptr(event->dest.client);
500
if (dest == NULL)
501
return NULL;
502
if (! dest->accept_input)
503
goto __not_avail;
504
if ((dest->filter & SNDRV_SEQ_FILTER_USE_EVENT) &&
505
! test_bit(event->type, dest->event_filter))
506
goto __not_avail;
507
if (filter && !(dest->filter & filter))
508
goto __not_avail;
509
510
return dest; /* ok - accessible */
511
__not_avail:
512
snd_seq_client_unlock(dest);
513
return NULL;
514
}
515
516
517
/*
518
* Return the error event.
519
*
520
* If the receiver client is a user client, the original event is
521
* encapsulated in SNDRV_SEQ_EVENT_BOUNCE as variable length event. If
522
* the original event is also variable length, the external data is
523
* copied after the event record.
524
* If the receiver client is a kernel client, the original event is
525
* quoted in SNDRV_SEQ_EVENT_KERNEL_ERROR, since this requires no extra
526
* kmalloc.
527
*/
528
static int bounce_error_event(struct snd_seq_client *client,
529
struct snd_seq_event *event,
530
int err, int atomic, int hop)
531
{
532
struct snd_seq_event bounce_ev;
533
int result;
534
535
if (client == NULL ||
536
! (client->filter & SNDRV_SEQ_FILTER_BOUNCE) ||
537
! client->accept_input)
538
return 0; /* ignored */
539
540
/* set up quoted error */
541
memset(&bounce_ev, 0, sizeof(bounce_ev));
542
bounce_ev.type = SNDRV_SEQ_EVENT_KERNEL_ERROR;
543
bounce_ev.flags = SNDRV_SEQ_EVENT_LENGTH_FIXED;
544
bounce_ev.queue = SNDRV_SEQ_QUEUE_DIRECT;
545
bounce_ev.source.client = SNDRV_SEQ_CLIENT_SYSTEM;
546
bounce_ev.source.port = SNDRV_SEQ_PORT_SYSTEM_ANNOUNCE;
547
bounce_ev.dest.client = client->number;
548
bounce_ev.dest.port = event->source.port;
549
bounce_ev.data.quote.origin = event->dest;
550
bounce_ev.data.quote.event = event;
551
bounce_ev.data.quote.value = -err; /* use positive value */
552
result = snd_seq_deliver_single_event(NULL, &bounce_ev, 0, atomic, hop + 1);
553
if (result < 0) {
554
client->event_lost++;
555
return result;
556
}
557
558
return result;
559
}
560
561
562
/*
563
* rewrite the time-stamp of the event record with the curren time
564
* of the given queue.
565
* return non-zero if updated.
566
*/
567
static int update_timestamp_of_queue(struct snd_seq_event *event,
568
int queue, int real_time)
569
{
570
struct snd_seq_queue *q;
571
572
q = queueptr(queue);
573
if (! q)
574
return 0;
575
event->queue = queue;
576
event->flags &= ~SNDRV_SEQ_TIME_STAMP_MASK;
577
if (real_time) {
578
event->time.time = snd_seq_timer_get_cur_time(q->timer);
579
event->flags |= SNDRV_SEQ_TIME_STAMP_REAL;
580
} else {
581
event->time.tick = snd_seq_timer_get_cur_tick(q->timer);
582
event->flags |= SNDRV_SEQ_TIME_STAMP_TICK;
583
}
584
queuefree(q);
585
return 1;
586
}
587
588
589
/*
590
* deliver an event to the specified destination.
591
* if filter is non-zero, client filter bitmap is tested.
592
*
593
* RETURN VALUE: 0 : if succeeded
594
* <0 : error
595
*/
596
static int snd_seq_deliver_single_event(struct snd_seq_client *client,
597
struct snd_seq_event *event,
598
int filter, int atomic, int hop)
599
{
600
struct snd_seq_client *dest = NULL;
601
struct snd_seq_client_port *dest_port = NULL;
602
int result = -ENOENT;
603
int direct;
604
605
direct = snd_seq_ev_is_direct(event);
606
607
dest = get_event_dest_client(event, filter);
608
if (dest == NULL)
609
goto __skip;
610
dest_port = snd_seq_port_use_ptr(dest, event->dest.port);
611
if (dest_port == NULL)
612
goto __skip;
613
614
/* check permission */
615
if (! check_port_perm(dest_port, SNDRV_SEQ_PORT_CAP_WRITE)) {
616
result = -EPERM;
617
goto __skip;
618
}
619
620
if (dest_port->timestamping)
621
update_timestamp_of_queue(event, dest_port->time_queue,
622
dest_port->time_real);
623
624
switch (dest->type) {
625
case USER_CLIENT:
626
if (dest->data.user.fifo)
627
result = snd_seq_fifo_event_in(dest->data.user.fifo, event);
628
break;
629
630
case KERNEL_CLIENT:
631
if (dest_port->event_input == NULL)
632
break;
633
result = dest_port->event_input(event, direct,
634
dest_port->private_data,
635
atomic, hop);
636
break;
637
default:
638
break;
639
}
640
641
__skip:
642
if (dest_port)
643
snd_seq_port_unlock(dest_port);
644
if (dest)
645
snd_seq_client_unlock(dest);
646
647
if (result < 0 && !direct) {
648
result = bounce_error_event(client, event, result, atomic, hop);
649
}
650
return result;
651
}
652
653
654
/*
655
* send the event to all subscribers:
656
*/
657
static int deliver_to_subscribers(struct snd_seq_client *client,
658
struct snd_seq_event *event,
659
int atomic, int hop)
660
{
661
struct snd_seq_subscribers *subs;
662
int err = 0, num_ev = 0;
663
struct snd_seq_event event_saved;
664
struct snd_seq_client_port *src_port;
665
struct snd_seq_port_subs_info *grp;
666
667
src_port = snd_seq_port_use_ptr(client, event->source.port);
668
if (src_port == NULL)
669
return -EINVAL; /* invalid source port */
670
/* save original event record */
671
event_saved = *event;
672
grp = &src_port->c_src;
673
674
/* lock list */
675
if (atomic)
676
read_lock(&grp->list_lock);
677
else
678
down_read(&grp->list_mutex);
679
list_for_each_entry(subs, &grp->list_head, src_list) {
680
event->dest = subs->info.dest;
681
if (subs->info.flags & SNDRV_SEQ_PORT_SUBS_TIMESTAMP)
682
/* convert time according to flag with subscription */
683
update_timestamp_of_queue(event, subs->info.queue,
684
subs->info.flags & SNDRV_SEQ_PORT_SUBS_TIME_REAL);
685
err = snd_seq_deliver_single_event(client, event,
686
0, atomic, hop);
687
if (err < 0)
688
break;
689
num_ev++;
690
/* restore original event record */
691
*event = event_saved;
692
}
693
if (atomic)
694
read_unlock(&grp->list_lock);
695
else
696
up_read(&grp->list_mutex);
697
*event = event_saved; /* restore */
698
snd_seq_port_unlock(src_port);
699
return (err < 0) ? err : num_ev;
700
}
701
702
703
#ifdef SUPPORT_BROADCAST
704
/*
705
* broadcast to all ports:
706
*/
707
static int port_broadcast_event(struct snd_seq_client *client,
708
struct snd_seq_event *event,
709
int atomic, int hop)
710
{
711
int num_ev = 0, err = 0;
712
struct snd_seq_client *dest_client;
713
struct snd_seq_client_port *port;
714
715
dest_client = get_event_dest_client(event, SNDRV_SEQ_FILTER_BROADCAST);
716
if (dest_client == NULL)
717
return 0; /* no matching destination */
718
719
read_lock(&dest_client->ports_lock);
720
list_for_each_entry(port, &dest_client->ports_list_head, list) {
721
event->dest.port = port->addr.port;
722
/* pass NULL as source client to avoid error bounce */
723
err = snd_seq_deliver_single_event(NULL, event,
724
SNDRV_SEQ_FILTER_BROADCAST,
725
atomic, hop);
726
if (err < 0)
727
break;
728
num_ev++;
729
}
730
read_unlock(&dest_client->ports_lock);
731
snd_seq_client_unlock(dest_client);
732
event->dest.port = SNDRV_SEQ_ADDRESS_BROADCAST; /* restore */
733
return (err < 0) ? err : num_ev;
734
}
735
736
/*
737
* send the event to all clients:
738
* if destination port is also ADDRESS_BROADCAST, deliver to all ports.
739
*/
740
static int broadcast_event(struct snd_seq_client *client,
741
struct snd_seq_event *event, int atomic, int hop)
742
{
743
int err = 0, num_ev = 0;
744
int dest;
745
struct snd_seq_addr addr;
746
747
addr = event->dest; /* save */
748
749
for (dest = 0; dest < SNDRV_SEQ_MAX_CLIENTS; dest++) {
750
/* don't send to itself */
751
if (dest == client->number)
752
continue;
753
event->dest.client = dest;
754
event->dest.port = addr.port;
755
if (addr.port == SNDRV_SEQ_ADDRESS_BROADCAST)
756
err = port_broadcast_event(client, event, atomic, hop);
757
else
758
/* pass NULL as source client to avoid error bounce */
759
err = snd_seq_deliver_single_event(NULL, event,
760
SNDRV_SEQ_FILTER_BROADCAST,
761
atomic, hop);
762
if (err < 0)
763
break;
764
num_ev += err;
765
}
766
event->dest = addr; /* restore */
767
return (err < 0) ? err : num_ev;
768
}
769
770
771
/* multicast - not supported yet */
772
static int multicast_event(struct snd_seq_client *client, struct snd_seq_event *event,
773
int atomic, int hop)
774
{
775
snd_printd("seq: multicast not supported yet.\n");
776
return 0; /* ignored */
777
}
778
#endif /* SUPPORT_BROADCAST */
779
780
781
/* deliver an event to the destination port(s).
782
* if the event is to subscribers or broadcast, the event is dispatched
783
* to multiple targets.
784
*
785
* RETURN VALUE: n > 0 : the number of delivered events.
786
* n == 0 : the event was not passed to any client.
787
* n < 0 : error - event was not processed.
788
*/
789
static int snd_seq_deliver_event(struct snd_seq_client *client, struct snd_seq_event *event,
790
int atomic, int hop)
791
{
792
int result;
793
794
hop++;
795
if (hop >= SNDRV_SEQ_MAX_HOPS) {
796
snd_printd("too long delivery path (%d:%d->%d:%d)\n",
797
event->source.client, event->source.port,
798
event->dest.client, event->dest.port);
799
return -EMLINK;
800
}
801
802
if (event->queue == SNDRV_SEQ_ADDRESS_SUBSCRIBERS ||
803
event->dest.client == SNDRV_SEQ_ADDRESS_SUBSCRIBERS)
804
result = deliver_to_subscribers(client, event, atomic, hop);
805
#ifdef SUPPORT_BROADCAST
806
else if (event->queue == SNDRV_SEQ_ADDRESS_BROADCAST ||
807
event->dest.client == SNDRV_SEQ_ADDRESS_BROADCAST)
808
result = broadcast_event(client, event, atomic, hop);
809
else if (event->dest.client >= SNDRV_SEQ_MAX_CLIENTS)
810
result = multicast_event(client, event, atomic, hop);
811
else if (event->dest.port == SNDRV_SEQ_ADDRESS_BROADCAST)
812
result = port_broadcast_event(client, event, atomic, hop);
813
#endif
814
else
815
result = snd_seq_deliver_single_event(client, event, 0, atomic, hop);
816
817
return result;
818
}
819
820
/*
821
* dispatch an event cell:
822
* This function is called only from queue check routines in timer
823
* interrupts or after enqueued.
824
* The event cell shall be released or re-queued in this function.
825
*
826
* RETURN VALUE: n > 0 : the number of delivered events.
827
* n == 0 : the event was not passed to any client.
828
* n < 0 : error - event was not processed.
829
*/
830
int snd_seq_dispatch_event(struct snd_seq_event_cell *cell, int atomic, int hop)
831
{
832
struct snd_seq_client *client;
833
int result;
834
835
if (snd_BUG_ON(!cell))
836
return -EINVAL;
837
838
client = snd_seq_client_use_ptr(cell->event.source.client);
839
if (client == NULL) {
840
snd_seq_cell_free(cell); /* release this cell */
841
return -EINVAL;
842
}
843
844
if (cell->event.type == SNDRV_SEQ_EVENT_NOTE) {
845
/* NOTE event:
846
* the event cell is re-used as a NOTE-OFF event and
847
* enqueued again.
848
*/
849
struct snd_seq_event tmpev, *ev;
850
851
/* reserve this event to enqueue note-off later */
852
tmpev = cell->event;
853
tmpev.type = SNDRV_SEQ_EVENT_NOTEON;
854
result = snd_seq_deliver_event(client, &tmpev, atomic, hop);
855
856
/*
857
* This was originally a note event. We now re-use the
858
* cell for the note-off event.
859
*/
860
861
ev = &cell->event;
862
ev->type = SNDRV_SEQ_EVENT_NOTEOFF;
863
ev->flags |= SNDRV_SEQ_PRIORITY_HIGH;
864
865
/* add the duration time */
866
switch (ev->flags & SNDRV_SEQ_TIME_STAMP_MASK) {
867
case SNDRV_SEQ_TIME_STAMP_TICK:
868
ev->time.tick += ev->data.note.duration;
869
break;
870
case SNDRV_SEQ_TIME_STAMP_REAL:
871
/* unit for duration is ms */
872
ev->time.time.tv_nsec += 1000000 * (ev->data.note.duration % 1000);
873
ev->time.time.tv_sec += ev->data.note.duration / 1000 +
874
ev->time.time.tv_nsec / 1000000000;
875
ev->time.time.tv_nsec %= 1000000000;
876
break;
877
}
878
ev->data.note.velocity = ev->data.note.off_velocity;
879
880
/* Now queue this cell as the note off event */
881
if (snd_seq_enqueue_event(cell, atomic, hop) < 0)
882
snd_seq_cell_free(cell); /* release this cell */
883
884
} else {
885
/* Normal events:
886
* event cell is freed after processing the event
887
*/
888
889
result = snd_seq_deliver_event(client, &cell->event, atomic, hop);
890
snd_seq_cell_free(cell);
891
}
892
893
snd_seq_client_unlock(client);
894
return result;
895
}
896
897
898
/* Allocate a cell from client pool and enqueue it to queue:
899
* if pool is empty and blocking is TRUE, sleep until a new cell is
900
* available.
901
*/
902
static int snd_seq_client_enqueue_event(struct snd_seq_client *client,
903
struct snd_seq_event *event,
904
struct file *file, int blocking,
905
int atomic, int hop)
906
{
907
struct snd_seq_event_cell *cell;
908
int err;
909
910
/* special queue values - force direct passing */
911
if (event->queue == SNDRV_SEQ_ADDRESS_SUBSCRIBERS) {
912
event->dest.client = SNDRV_SEQ_ADDRESS_SUBSCRIBERS;
913
event->queue = SNDRV_SEQ_QUEUE_DIRECT;
914
} else
915
#ifdef SUPPORT_BROADCAST
916
if (event->queue == SNDRV_SEQ_ADDRESS_BROADCAST) {
917
event->dest.client = SNDRV_SEQ_ADDRESS_BROADCAST;
918
event->queue = SNDRV_SEQ_QUEUE_DIRECT;
919
}
920
#endif
921
if (event->dest.client == SNDRV_SEQ_ADDRESS_SUBSCRIBERS) {
922
/* check presence of source port */
923
struct snd_seq_client_port *src_port = snd_seq_port_use_ptr(client, event->source.port);
924
if (src_port == NULL)
925
return -EINVAL;
926
snd_seq_port_unlock(src_port);
927
}
928
929
/* direct event processing without enqueued */
930
if (snd_seq_ev_is_direct(event)) {
931
if (event->type == SNDRV_SEQ_EVENT_NOTE)
932
return -EINVAL; /* this event must be enqueued! */
933
return snd_seq_deliver_event(client, event, atomic, hop);
934
}
935
936
/* Not direct, normal queuing */
937
if (snd_seq_queue_is_used(event->queue, client->number) <= 0)
938
return -EINVAL; /* invalid queue */
939
if (! snd_seq_write_pool_allocated(client))
940
return -ENXIO; /* queue is not allocated */
941
942
/* allocate an event cell */
943
err = snd_seq_event_dup(client->pool, event, &cell, !blocking || atomic, file);
944
if (err < 0)
945
return err;
946
947
/* we got a cell. enqueue it. */
948
if ((err = snd_seq_enqueue_event(cell, atomic, hop)) < 0) {
949
snd_seq_cell_free(cell);
950
return err;
951
}
952
953
return 0;
954
}
955
956
957
/*
958
* check validity of event type and data length.
959
* return non-zero if invalid.
960
*/
961
static int check_event_type_and_length(struct snd_seq_event *ev)
962
{
963
switch (snd_seq_ev_length_type(ev)) {
964
case SNDRV_SEQ_EVENT_LENGTH_FIXED:
965
if (snd_seq_ev_is_variable_type(ev))
966
return -EINVAL;
967
break;
968
case SNDRV_SEQ_EVENT_LENGTH_VARIABLE:
969
if (! snd_seq_ev_is_variable_type(ev) ||
970
(ev->data.ext.len & ~SNDRV_SEQ_EXT_MASK) >= SNDRV_SEQ_MAX_EVENT_LEN)
971
return -EINVAL;
972
break;
973
case SNDRV_SEQ_EVENT_LENGTH_VARUSR:
974
if (! snd_seq_ev_is_direct(ev))
975
return -EINVAL;
976
break;
977
}
978
return 0;
979
}
980
981
982
/* handle write() */
983
/* possible error values:
984
* -ENXIO invalid client or file open mode
985
* -ENOMEM malloc failed
986
* -EFAULT seg. fault during copy from user space
987
* -EINVAL invalid event
988
* -EAGAIN no space in output pool
989
* -EINTR interrupts while sleep
990
* -EMLINK too many hops
991
* others depends on return value from driver callback
992
*/
993
static ssize_t snd_seq_write(struct file *file, const char __user *buf,
994
size_t count, loff_t *offset)
995
{
996
struct snd_seq_client *client = file->private_data;
997
int written = 0, len;
998
int err = -EINVAL;
999
struct snd_seq_event event;
1000
1001
if (!(snd_seq_file_flags(file) & SNDRV_SEQ_LFLG_OUTPUT))
1002
return -ENXIO;
1003
1004
/* check client structures are in place */
1005
if (snd_BUG_ON(!client))
1006
return -ENXIO;
1007
1008
if (!client->accept_output || client->pool == NULL)
1009
return -ENXIO;
1010
1011
/* allocate the pool now if the pool is not allocated yet */
1012
if (client->pool->size > 0 && !snd_seq_write_pool_allocated(client)) {
1013
if (snd_seq_pool_init(client->pool) < 0)
1014
return -ENOMEM;
1015
}
1016
1017
/* only process whole events */
1018
while (count >= sizeof(struct snd_seq_event)) {
1019
/* Read in the event header from the user */
1020
len = sizeof(event);
1021
if (copy_from_user(&event, buf, len)) {
1022
err = -EFAULT;
1023
break;
1024
}
1025
event.source.client = client->number; /* fill in client number */
1026
/* Check for extension data length */
1027
if (check_event_type_and_length(&event)) {
1028
err = -EINVAL;
1029
break;
1030
}
1031
1032
/* check for special events */
1033
if (event.type == SNDRV_SEQ_EVENT_NONE)
1034
goto __skip_event;
1035
else if (snd_seq_ev_is_reserved(&event)) {
1036
err = -EINVAL;
1037
break;
1038
}
1039
1040
if (snd_seq_ev_is_variable(&event)) {
1041
int extlen = event.data.ext.len & ~SNDRV_SEQ_EXT_MASK;
1042
if ((size_t)(extlen + len) > count) {
1043
/* back out, will get an error this time or next */
1044
err = -EINVAL;
1045
break;
1046
}
1047
/* set user space pointer */
1048
event.data.ext.len = extlen | SNDRV_SEQ_EXT_USRPTR;
1049
event.data.ext.ptr = (char __force *)buf
1050
+ sizeof(struct snd_seq_event);
1051
len += extlen; /* increment data length */
1052
} else {
1053
#ifdef CONFIG_COMPAT
1054
if (client->convert32 && snd_seq_ev_is_varusr(&event)) {
1055
void *ptr = (void __force *)compat_ptr(event.data.raw32.d[1]);
1056
event.data.ext.ptr = ptr;
1057
}
1058
#endif
1059
}
1060
1061
/* ok, enqueue it */
1062
err = snd_seq_client_enqueue_event(client, &event, file,
1063
!(file->f_flags & O_NONBLOCK),
1064
0, 0);
1065
if (err < 0)
1066
break;
1067
1068
__skip_event:
1069
/* Update pointers and counts */
1070
count -= len;
1071
buf += len;
1072
written += len;
1073
}
1074
1075
return written ? written : err;
1076
}
1077
1078
1079
/*
1080
* handle polling
1081
*/
1082
static unsigned int snd_seq_poll(struct file *file, poll_table * wait)
1083
{
1084
struct snd_seq_client *client = file->private_data;
1085
unsigned int mask = 0;
1086
1087
/* check client structures are in place */
1088
if (snd_BUG_ON(!client))
1089
return -ENXIO;
1090
1091
if ((snd_seq_file_flags(file) & SNDRV_SEQ_LFLG_INPUT) &&
1092
client->data.user.fifo) {
1093
1094
/* check if data is available in the outqueue */
1095
if (snd_seq_fifo_poll_wait(client->data.user.fifo, file, wait))
1096
mask |= POLLIN | POLLRDNORM;
1097
}
1098
1099
if (snd_seq_file_flags(file) & SNDRV_SEQ_LFLG_OUTPUT) {
1100
1101
/* check if data is available in the pool */
1102
if (!snd_seq_write_pool_allocated(client) ||
1103
snd_seq_pool_poll_wait(client->pool, file, wait))
1104
mask |= POLLOUT | POLLWRNORM;
1105
}
1106
1107
return mask;
1108
}
1109
1110
1111
/*-----------------------------------------------------*/
1112
1113
1114
/* SYSTEM_INFO ioctl() */
1115
static int snd_seq_ioctl_system_info(struct snd_seq_client *client, void __user *arg)
1116
{
1117
struct snd_seq_system_info info;
1118
1119
memset(&info, 0, sizeof(info));
1120
/* fill the info fields */
1121
info.queues = SNDRV_SEQ_MAX_QUEUES;
1122
info.clients = SNDRV_SEQ_MAX_CLIENTS;
1123
info.ports = 256; /* fixed limit */
1124
info.channels = 256; /* fixed limit */
1125
info.cur_clients = client_usage.cur;
1126
info.cur_queues = snd_seq_queue_get_cur_queues();
1127
1128
if (copy_to_user(arg, &info, sizeof(info)))
1129
return -EFAULT;
1130
return 0;
1131
}
1132
1133
1134
/* RUNNING_MODE ioctl() */
1135
static int snd_seq_ioctl_running_mode(struct snd_seq_client *client, void __user *arg)
1136
{
1137
struct snd_seq_running_info info;
1138
struct snd_seq_client *cptr;
1139
int err = 0;
1140
1141
if (copy_from_user(&info, arg, sizeof(info)))
1142
return -EFAULT;
1143
1144
/* requested client number */
1145
cptr = snd_seq_client_use_ptr(info.client);
1146
if (cptr == NULL)
1147
return -ENOENT; /* don't change !!! */
1148
1149
#ifdef SNDRV_BIG_ENDIAN
1150
if (! info.big_endian) {
1151
err = -EINVAL;
1152
goto __err;
1153
}
1154
#else
1155
if (info.big_endian) {
1156
err = -EINVAL;
1157
goto __err;
1158
}
1159
1160
#endif
1161
if (info.cpu_mode > sizeof(long)) {
1162
err = -EINVAL;
1163
goto __err;
1164
}
1165
cptr->convert32 = (info.cpu_mode < sizeof(long));
1166
__err:
1167
snd_seq_client_unlock(cptr);
1168
return err;
1169
}
1170
1171
/* CLIENT_INFO ioctl() */
1172
static void get_client_info(struct snd_seq_client *cptr,
1173
struct snd_seq_client_info *info)
1174
{
1175
info->client = cptr->number;
1176
1177
/* fill the info fields */
1178
info->type = cptr->type;
1179
strcpy(info->name, cptr->name);
1180
info->filter = cptr->filter;
1181
info->event_lost = cptr->event_lost;
1182
memcpy(info->event_filter, cptr->event_filter, 32);
1183
info->num_ports = cptr->num_ports;
1184
memset(info->reserved, 0, sizeof(info->reserved));
1185
}
1186
1187
static int snd_seq_ioctl_get_client_info(struct snd_seq_client *client,
1188
void __user *arg)
1189
{
1190
struct snd_seq_client *cptr;
1191
struct snd_seq_client_info client_info;
1192
1193
if (copy_from_user(&client_info, arg, sizeof(client_info)))
1194
return -EFAULT;
1195
1196
/* requested client number */
1197
cptr = snd_seq_client_use_ptr(client_info.client);
1198
if (cptr == NULL)
1199
return -ENOENT; /* don't change !!! */
1200
1201
get_client_info(cptr, &client_info);
1202
snd_seq_client_unlock(cptr);
1203
1204
if (copy_to_user(arg, &client_info, sizeof(client_info)))
1205
return -EFAULT;
1206
return 0;
1207
}
1208
1209
1210
/* CLIENT_INFO ioctl() */
1211
static int snd_seq_ioctl_set_client_info(struct snd_seq_client *client,
1212
void __user *arg)
1213
{
1214
struct snd_seq_client_info client_info;
1215
1216
if (copy_from_user(&client_info, arg, sizeof(client_info)))
1217
return -EFAULT;
1218
1219
/* it is not allowed to set the info fields for an another client */
1220
if (client->number != client_info.client)
1221
return -EPERM;
1222
/* also client type must be set now */
1223
if (client->type != client_info.type)
1224
return -EINVAL;
1225
1226
/* fill the info fields */
1227
if (client_info.name[0])
1228
strlcpy(client->name, client_info.name, sizeof(client->name));
1229
1230
client->filter = client_info.filter;
1231
client->event_lost = client_info.event_lost;
1232
memcpy(client->event_filter, client_info.event_filter, 32);
1233
1234
return 0;
1235
}
1236
1237
1238
/*
1239
* CREATE PORT ioctl()
1240
*/
1241
static int snd_seq_ioctl_create_port(struct snd_seq_client *client,
1242
void __user *arg)
1243
{
1244
struct snd_seq_client_port *port;
1245
struct snd_seq_port_info info;
1246
struct snd_seq_port_callback *callback;
1247
1248
if (copy_from_user(&info, arg, sizeof(info)))
1249
return -EFAULT;
1250
1251
/* it is not allowed to create the port for an another client */
1252
if (info.addr.client != client->number)
1253
return -EPERM;
1254
1255
port = snd_seq_create_port(client, (info.flags & SNDRV_SEQ_PORT_FLG_GIVEN_PORT) ? info.addr.port : -1);
1256
if (port == NULL)
1257
return -ENOMEM;
1258
1259
if (client->type == USER_CLIENT && info.kernel) {
1260
snd_seq_delete_port(client, port->addr.port);
1261
return -EINVAL;
1262
}
1263
if (client->type == KERNEL_CLIENT) {
1264
if ((callback = info.kernel) != NULL) {
1265
if (callback->owner)
1266
port->owner = callback->owner;
1267
port->private_data = callback->private_data;
1268
port->private_free = callback->private_free;
1269
port->callback_all = callback->callback_all;
1270
port->event_input = callback->event_input;
1271
port->c_src.open = callback->subscribe;
1272
port->c_src.close = callback->unsubscribe;
1273
port->c_dest.open = callback->use;
1274
port->c_dest.close = callback->unuse;
1275
}
1276
}
1277
1278
info.addr = port->addr;
1279
1280
snd_seq_set_port_info(port, &info);
1281
snd_seq_system_client_ev_port_start(port->addr.client, port->addr.port);
1282
1283
if (copy_to_user(arg, &info, sizeof(info)))
1284
return -EFAULT;
1285
1286
return 0;
1287
}
1288
1289
/*
1290
* DELETE PORT ioctl()
1291
*/
1292
static int snd_seq_ioctl_delete_port(struct snd_seq_client *client,
1293
void __user *arg)
1294
{
1295
struct snd_seq_port_info info;
1296
int err;
1297
1298
/* set passed parameters */
1299
if (copy_from_user(&info, arg, sizeof(info)))
1300
return -EFAULT;
1301
1302
/* it is not allowed to remove the port for an another client */
1303
if (info.addr.client != client->number)
1304
return -EPERM;
1305
1306
err = snd_seq_delete_port(client, info.addr.port);
1307
if (err >= 0)
1308
snd_seq_system_client_ev_port_exit(client->number, info.addr.port);
1309
return err;
1310
}
1311
1312
1313
/*
1314
* GET_PORT_INFO ioctl() (on any client)
1315
*/
1316
static int snd_seq_ioctl_get_port_info(struct snd_seq_client *client,
1317
void __user *arg)
1318
{
1319
struct snd_seq_client *cptr;
1320
struct snd_seq_client_port *port;
1321
struct snd_seq_port_info info;
1322
1323
if (copy_from_user(&info, arg, sizeof(info)))
1324
return -EFAULT;
1325
cptr = snd_seq_client_use_ptr(info.addr.client);
1326
if (cptr == NULL)
1327
return -ENXIO;
1328
1329
port = snd_seq_port_use_ptr(cptr, info.addr.port);
1330
if (port == NULL) {
1331
snd_seq_client_unlock(cptr);
1332
return -ENOENT; /* don't change */
1333
}
1334
1335
/* get port info */
1336
snd_seq_get_port_info(port, &info);
1337
snd_seq_port_unlock(port);
1338
snd_seq_client_unlock(cptr);
1339
1340
if (copy_to_user(arg, &info, sizeof(info)))
1341
return -EFAULT;
1342
return 0;
1343
}
1344
1345
1346
/*
1347
* SET_PORT_INFO ioctl() (only ports on this/own client)
1348
*/
1349
static int snd_seq_ioctl_set_port_info(struct snd_seq_client *client,
1350
void __user *arg)
1351
{
1352
struct snd_seq_client_port *port;
1353
struct snd_seq_port_info info;
1354
1355
if (copy_from_user(&info, arg, sizeof(info)))
1356
return -EFAULT;
1357
1358
if (info.addr.client != client->number) /* only set our own ports ! */
1359
return -EPERM;
1360
port = snd_seq_port_use_ptr(client, info.addr.port);
1361
if (port) {
1362
snd_seq_set_port_info(port, &info);
1363
snd_seq_port_unlock(port);
1364
}
1365
return 0;
1366
}
1367
1368
1369
/*
1370
* port subscription (connection)
1371
*/
1372
#define PERM_RD (SNDRV_SEQ_PORT_CAP_READ|SNDRV_SEQ_PORT_CAP_SUBS_READ)
1373
#define PERM_WR (SNDRV_SEQ_PORT_CAP_WRITE|SNDRV_SEQ_PORT_CAP_SUBS_WRITE)
1374
1375
static int check_subscription_permission(struct snd_seq_client *client,
1376
struct snd_seq_client_port *sport,
1377
struct snd_seq_client_port *dport,
1378
struct snd_seq_port_subscribe *subs)
1379
{
1380
if (client->number != subs->sender.client &&
1381
client->number != subs->dest.client) {
1382
/* connection by third client - check export permission */
1383
if (check_port_perm(sport, SNDRV_SEQ_PORT_CAP_NO_EXPORT))
1384
return -EPERM;
1385
if (check_port_perm(dport, SNDRV_SEQ_PORT_CAP_NO_EXPORT))
1386
return -EPERM;
1387
}
1388
1389
/* check read permission */
1390
/* if sender or receiver is the subscribing client itself,
1391
* no permission check is necessary
1392
*/
1393
if (client->number != subs->sender.client) {
1394
if (! check_port_perm(sport, PERM_RD))
1395
return -EPERM;
1396
}
1397
/* check write permission */
1398
if (client->number != subs->dest.client) {
1399
if (! check_port_perm(dport, PERM_WR))
1400
return -EPERM;
1401
}
1402
return 0;
1403
}
1404
1405
/*
1406
* send an subscription notify event to user client:
1407
* client must be user client.
1408
*/
1409
int snd_seq_client_notify_subscription(int client, int port,
1410
struct snd_seq_port_subscribe *info,
1411
int evtype)
1412
{
1413
struct snd_seq_event event;
1414
1415
memset(&event, 0, sizeof(event));
1416
event.type = evtype;
1417
event.data.connect.dest = info->dest;
1418
event.data.connect.sender = info->sender;
1419
1420
return snd_seq_system_notify(client, port, &event); /* non-atomic */
1421
}
1422
1423
1424
/*
1425
* add to port's subscription list IOCTL interface
1426
*/
1427
static int snd_seq_ioctl_subscribe_port(struct snd_seq_client *client,
1428
void __user *arg)
1429
{
1430
int result = -EINVAL;
1431
struct snd_seq_client *receiver = NULL, *sender = NULL;
1432
struct snd_seq_client_port *sport = NULL, *dport = NULL;
1433
struct snd_seq_port_subscribe subs;
1434
1435
if (copy_from_user(&subs, arg, sizeof(subs)))
1436
return -EFAULT;
1437
1438
if ((receiver = snd_seq_client_use_ptr(subs.dest.client)) == NULL)
1439
goto __end;
1440
if ((sender = snd_seq_client_use_ptr(subs.sender.client)) == NULL)
1441
goto __end;
1442
if ((sport = snd_seq_port_use_ptr(sender, subs.sender.port)) == NULL)
1443
goto __end;
1444
if ((dport = snd_seq_port_use_ptr(receiver, subs.dest.port)) == NULL)
1445
goto __end;
1446
1447
result = check_subscription_permission(client, sport, dport, &subs);
1448
if (result < 0)
1449
goto __end;
1450
1451
/* connect them */
1452
result = snd_seq_port_connect(client, sender, sport, receiver, dport, &subs);
1453
if (! result) /* broadcast announce */
1454
snd_seq_client_notify_subscription(SNDRV_SEQ_ADDRESS_SUBSCRIBERS, 0,
1455
&subs, SNDRV_SEQ_EVENT_PORT_SUBSCRIBED);
1456
__end:
1457
if (sport)
1458
snd_seq_port_unlock(sport);
1459
if (dport)
1460
snd_seq_port_unlock(dport);
1461
if (sender)
1462
snd_seq_client_unlock(sender);
1463
if (receiver)
1464
snd_seq_client_unlock(receiver);
1465
return result;
1466
}
1467
1468
1469
/*
1470
* remove from port's subscription list
1471
*/
1472
static int snd_seq_ioctl_unsubscribe_port(struct snd_seq_client *client,
1473
void __user *arg)
1474
{
1475
int result = -ENXIO;
1476
struct snd_seq_client *receiver = NULL, *sender = NULL;
1477
struct snd_seq_client_port *sport = NULL, *dport = NULL;
1478
struct snd_seq_port_subscribe subs;
1479
1480
if (copy_from_user(&subs, arg, sizeof(subs)))
1481
return -EFAULT;
1482
1483
if ((receiver = snd_seq_client_use_ptr(subs.dest.client)) == NULL)
1484
goto __end;
1485
if ((sender = snd_seq_client_use_ptr(subs.sender.client)) == NULL)
1486
goto __end;
1487
if ((sport = snd_seq_port_use_ptr(sender, subs.sender.port)) == NULL)
1488
goto __end;
1489
if ((dport = snd_seq_port_use_ptr(receiver, subs.dest.port)) == NULL)
1490
goto __end;
1491
1492
result = check_subscription_permission(client, sport, dport, &subs);
1493
if (result < 0)
1494
goto __end;
1495
1496
result = snd_seq_port_disconnect(client, sender, sport, receiver, dport, &subs);
1497
if (! result) /* broadcast announce */
1498
snd_seq_client_notify_subscription(SNDRV_SEQ_ADDRESS_SUBSCRIBERS, 0,
1499
&subs, SNDRV_SEQ_EVENT_PORT_UNSUBSCRIBED);
1500
__end:
1501
if (sport)
1502
snd_seq_port_unlock(sport);
1503
if (dport)
1504
snd_seq_port_unlock(dport);
1505
if (sender)
1506
snd_seq_client_unlock(sender);
1507
if (receiver)
1508
snd_seq_client_unlock(receiver);
1509
return result;
1510
}
1511
1512
1513
/* CREATE_QUEUE ioctl() */
1514
static int snd_seq_ioctl_create_queue(struct snd_seq_client *client,
1515
void __user *arg)
1516
{
1517
struct snd_seq_queue_info info;
1518
int result;
1519
struct snd_seq_queue *q;
1520
1521
if (copy_from_user(&info, arg, sizeof(info)))
1522
return -EFAULT;
1523
1524
result = snd_seq_queue_alloc(client->number, info.locked, info.flags);
1525
if (result < 0)
1526
return result;
1527
1528
q = queueptr(result);
1529
if (q == NULL)
1530
return -EINVAL;
1531
1532
info.queue = q->queue;
1533
info.locked = q->locked;
1534
info.owner = q->owner;
1535
1536
/* set queue name */
1537
if (! info.name[0])
1538
snprintf(info.name, sizeof(info.name), "Queue-%d", q->queue);
1539
strlcpy(q->name, info.name, sizeof(q->name));
1540
queuefree(q);
1541
1542
if (copy_to_user(arg, &info, sizeof(info)))
1543
return -EFAULT;
1544
1545
return 0;
1546
}
1547
1548
/* DELETE_QUEUE ioctl() */
1549
static int snd_seq_ioctl_delete_queue(struct snd_seq_client *client,
1550
void __user *arg)
1551
{
1552
struct snd_seq_queue_info info;
1553
1554
if (copy_from_user(&info, arg, sizeof(info)))
1555
return -EFAULT;
1556
1557
return snd_seq_queue_delete(client->number, info.queue);
1558
}
1559
1560
/* GET_QUEUE_INFO ioctl() */
1561
static int snd_seq_ioctl_get_queue_info(struct snd_seq_client *client,
1562
void __user *arg)
1563
{
1564
struct snd_seq_queue_info info;
1565
struct snd_seq_queue *q;
1566
1567
if (copy_from_user(&info, arg, sizeof(info)))
1568
return -EFAULT;
1569
1570
q = queueptr(info.queue);
1571
if (q == NULL)
1572
return -EINVAL;
1573
1574
memset(&info, 0, sizeof(info));
1575
info.queue = q->queue;
1576
info.owner = q->owner;
1577
info.locked = q->locked;
1578
strlcpy(info.name, q->name, sizeof(info.name));
1579
queuefree(q);
1580
1581
if (copy_to_user(arg, &info, sizeof(info)))
1582
return -EFAULT;
1583
1584
return 0;
1585
}
1586
1587
/* SET_QUEUE_INFO ioctl() */
1588
static int snd_seq_ioctl_set_queue_info(struct snd_seq_client *client,
1589
void __user *arg)
1590
{
1591
struct snd_seq_queue_info info;
1592
struct snd_seq_queue *q;
1593
1594
if (copy_from_user(&info, arg, sizeof(info)))
1595
return -EFAULT;
1596
1597
if (info.owner != client->number)
1598
return -EINVAL;
1599
1600
/* change owner/locked permission */
1601
if (snd_seq_queue_check_access(info.queue, client->number)) {
1602
if (snd_seq_queue_set_owner(info.queue, client->number, info.locked) < 0)
1603
return -EPERM;
1604
if (info.locked)
1605
snd_seq_queue_use(info.queue, client->number, 1);
1606
} else {
1607
return -EPERM;
1608
}
1609
1610
q = queueptr(info.queue);
1611
if (! q)
1612
return -EINVAL;
1613
if (q->owner != client->number) {
1614
queuefree(q);
1615
return -EPERM;
1616
}
1617
strlcpy(q->name, info.name, sizeof(q->name));
1618
queuefree(q);
1619
1620
return 0;
1621
}
1622
1623
/* GET_NAMED_QUEUE ioctl() */
1624
static int snd_seq_ioctl_get_named_queue(struct snd_seq_client *client, void __user *arg)
1625
{
1626
struct snd_seq_queue_info info;
1627
struct snd_seq_queue *q;
1628
1629
if (copy_from_user(&info, arg, sizeof(info)))
1630
return -EFAULT;
1631
1632
q = snd_seq_queue_find_name(info.name);
1633
if (q == NULL)
1634
return -EINVAL;
1635
info.queue = q->queue;
1636
info.owner = q->owner;
1637
info.locked = q->locked;
1638
queuefree(q);
1639
1640
if (copy_to_user(arg, &info, sizeof(info)))
1641
return -EFAULT;
1642
1643
return 0;
1644
}
1645
1646
/* GET_QUEUE_STATUS ioctl() */
1647
static int snd_seq_ioctl_get_queue_status(struct snd_seq_client *client,
1648
void __user *arg)
1649
{
1650
struct snd_seq_queue_status status;
1651
struct snd_seq_queue *queue;
1652
struct snd_seq_timer *tmr;
1653
1654
if (copy_from_user(&status, arg, sizeof(status)))
1655
return -EFAULT;
1656
1657
queue = queueptr(status.queue);
1658
if (queue == NULL)
1659
return -EINVAL;
1660
memset(&status, 0, sizeof(status));
1661
status.queue = queue->queue;
1662
1663
tmr = queue->timer;
1664
status.events = queue->tickq->cells + queue->timeq->cells;
1665
1666
status.time = snd_seq_timer_get_cur_time(tmr);
1667
status.tick = snd_seq_timer_get_cur_tick(tmr);
1668
1669
status.running = tmr->running;
1670
1671
status.flags = queue->flags;
1672
queuefree(queue);
1673
1674
if (copy_to_user(arg, &status, sizeof(status)))
1675
return -EFAULT;
1676
return 0;
1677
}
1678
1679
1680
/* GET_QUEUE_TEMPO ioctl() */
1681
static int snd_seq_ioctl_get_queue_tempo(struct snd_seq_client *client,
1682
void __user *arg)
1683
{
1684
struct snd_seq_queue_tempo tempo;
1685
struct snd_seq_queue *queue;
1686
struct snd_seq_timer *tmr;
1687
1688
if (copy_from_user(&tempo, arg, sizeof(tempo)))
1689
return -EFAULT;
1690
1691
queue = queueptr(tempo.queue);
1692
if (queue == NULL)
1693
return -EINVAL;
1694
memset(&tempo, 0, sizeof(tempo));
1695
tempo.queue = queue->queue;
1696
1697
tmr = queue->timer;
1698
1699
tempo.tempo = tmr->tempo;
1700
tempo.ppq = tmr->ppq;
1701
tempo.skew_value = tmr->skew;
1702
tempo.skew_base = tmr->skew_base;
1703
queuefree(queue);
1704
1705
if (copy_to_user(arg, &tempo, sizeof(tempo)))
1706
return -EFAULT;
1707
return 0;
1708
}
1709
1710
1711
/* SET_QUEUE_TEMPO ioctl() */
1712
int snd_seq_set_queue_tempo(int client, struct snd_seq_queue_tempo *tempo)
1713
{
1714
if (!snd_seq_queue_check_access(tempo->queue, client))
1715
return -EPERM;
1716
return snd_seq_queue_timer_set_tempo(tempo->queue, client, tempo);
1717
}
1718
1719
EXPORT_SYMBOL(snd_seq_set_queue_tempo);
1720
1721
static int snd_seq_ioctl_set_queue_tempo(struct snd_seq_client *client,
1722
void __user *arg)
1723
{
1724
int result;
1725
struct snd_seq_queue_tempo tempo;
1726
1727
if (copy_from_user(&tempo, arg, sizeof(tempo)))
1728
return -EFAULT;
1729
1730
result = snd_seq_set_queue_tempo(client->number, &tempo);
1731
return result < 0 ? result : 0;
1732
}
1733
1734
1735
/* GET_QUEUE_TIMER ioctl() */
1736
static int snd_seq_ioctl_get_queue_timer(struct snd_seq_client *client,
1737
void __user *arg)
1738
{
1739
struct snd_seq_queue_timer timer;
1740
struct snd_seq_queue *queue;
1741
struct snd_seq_timer *tmr;
1742
1743
if (copy_from_user(&timer, arg, sizeof(timer)))
1744
return -EFAULT;
1745
1746
queue = queueptr(timer.queue);
1747
if (queue == NULL)
1748
return -EINVAL;
1749
1750
if (mutex_lock_interruptible(&queue->timer_mutex)) {
1751
queuefree(queue);
1752
return -ERESTARTSYS;
1753
}
1754
tmr = queue->timer;
1755
memset(&timer, 0, sizeof(timer));
1756
timer.queue = queue->queue;
1757
1758
timer.type = tmr->type;
1759
if (tmr->type == SNDRV_SEQ_TIMER_ALSA) {
1760
timer.u.alsa.id = tmr->alsa_id;
1761
timer.u.alsa.resolution = tmr->preferred_resolution;
1762
}
1763
mutex_unlock(&queue->timer_mutex);
1764
queuefree(queue);
1765
1766
if (copy_to_user(arg, &timer, sizeof(timer)))
1767
return -EFAULT;
1768
return 0;
1769
}
1770
1771
1772
/* SET_QUEUE_TIMER ioctl() */
1773
static int snd_seq_ioctl_set_queue_timer(struct snd_seq_client *client,
1774
void __user *arg)
1775
{
1776
int result = 0;
1777
struct snd_seq_queue_timer timer;
1778
1779
if (copy_from_user(&timer, arg, sizeof(timer)))
1780
return -EFAULT;
1781
1782
if (timer.type != SNDRV_SEQ_TIMER_ALSA)
1783
return -EINVAL;
1784
1785
if (snd_seq_queue_check_access(timer.queue, client->number)) {
1786
struct snd_seq_queue *q;
1787
struct snd_seq_timer *tmr;
1788
1789
q = queueptr(timer.queue);
1790
if (q == NULL)
1791
return -ENXIO;
1792
if (mutex_lock_interruptible(&q->timer_mutex)) {
1793
queuefree(q);
1794
return -ERESTARTSYS;
1795
}
1796
tmr = q->timer;
1797
snd_seq_queue_timer_close(timer.queue);
1798
tmr->type = timer.type;
1799
if (tmr->type == SNDRV_SEQ_TIMER_ALSA) {
1800
tmr->alsa_id = timer.u.alsa.id;
1801
tmr->preferred_resolution = timer.u.alsa.resolution;
1802
}
1803
result = snd_seq_queue_timer_open(timer.queue);
1804
mutex_unlock(&q->timer_mutex);
1805
queuefree(q);
1806
} else {
1807
return -EPERM;
1808
}
1809
1810
return result;
1811
}
1812
1813
1814
/* GET_QUEUE_CLIENT ioctl() */
1815
static int snd_seq_ioctl_get_queue_client(struct snd_seq_client *client,
1816
void __user *arg)
1817
{
1818
struct snd_seq_queue_client info;
1819
int used;
1820
1821
if (copy_from_user(&info, arg, sizeof(info)))
1822
return -EFAULT;
1823
1824
used = snd_seq_queue_is_used(info.queue, client->number);
1825
if (used < 0)
1826
return -EINVAL;
1827
info.used = used;
1828
info.client = client->number;
1829
1830
if (copy_to_user(arg, &info, sizeof(info)))
1831
return -EFAULT;
1832
return 0;
1833
}
1834
1835
1836
/* SET_QUEUE_CLIENT ioctl() */
1837
static int snd_seq_ioctl_set_queue_client(struct snd_seq_client *client,
1838
void __user *arg)
1839
{
1840
int err;
1841
struct snd_seq_queue_client info;
1842
1843
if (copy_from_user(&info, arg, sizeof(info)))
1844
return -EFAULT;
1845
1846
if (info.used >= 0) {
1847
err = snd_seq_queue_use(info.queue, client->number, info.used);
1848
if (err < 0)
1849
return err;
1850
}
1851
1852
return snd_seq_ioctl_get_queue_client(client, arg);
1853
}
1854
1855
1856
/* GET_CLIENT_POOL ioctl() */
1857
static int snd_seq_ioctl_get_client_pool(struct snd_seq_client *client,
1858
void __user *arg)
1859
{
1860
struct snd_seq_client_pool info;
1861
struct snd_seq_client *cptr;
1862
1863
if (copy_from_user(&info, arg, sizeof(info)))
1864
return -EFAULT;
1865
1866
cptr = snd_seq_client_use_ptr(info.client);
1867
if (cptr == NULL)
1868
return -ENOENT;
1869
memset(&info, 0, sizeof(info));
1870
info.output_pool = cptr->pool->size;
1871
info.output_room = cptr->pool->room;
1872
info.output_free = info.output_pool;
1873
info.output_free = snd_seq_unused_cells(cptr->pool);
1874
if (cptr->type == USER_CLIENT) {
1875
info.input_pool = cptr->data.user.fifo_pool_size;
1876
info.input_free = info.input_pool;
1877
if (cptr->data.user.fifo)
1878
info.input_free = snd_seq_unused_cells(cptr->data.user.fifo->pool);
1879
} else {
1880
info.input_pool = 0;
1881
info.input_free = 0;
1882
}
1883
snd_seq_client_unlock(cptr);
1884
1885
if (copy_to_user(arg, &info, sizeof(info)))
1886
return -EFAULT;
1887
return 0;
1888
}
1889
1890
/* SET_CLIENT_POOL ioctl() */
1891
static int snd_seq_ioctl_set_client_pool(struct snd_seq_client *client,
1892
void __user *arg)
1893
{
1894
struct snd_seq_client_pool info;
1895
int rc;
1896
1897
if (copy_from_user(&info, arg, sizeof(info)))
1898
return -EFAULT;
1899
1900
if (client->number != info.client)
1901
return -EINVAL; /* can't change other clients */
1902
1903
if (info.output_pool >= 1 && info.output_pool <= SNDRV_SEQ_MAX_EVENTS &&
1904
(! snd_seq_write_pool_allocated(client) ||
1905
info.output_pool != client->pool->size)) {
1906
if (snd_seq_write_pool_allocated(client)) {
1907
/* remove all existing cells */
1908
snd_seq_queue_client_leave_cells(client->number);
1909
snd_seq_pool_done(client->pool);
1910
}
1911
client->pool->size = info.output_pool;
1912
rc = snd_seq_pool_init(client->pool);
1913
if (rc < 0)
1914
return rc;
1915
}
1916
if (client->type == USER_CLIENT && client->data.user.fifo != NULL &&
1917
info.input_pool >= 1 &&
1918
info.input_pool <= SNDRV_SEQ_MAX_CLIENT_EVENTS &&
1919
info.input_pool != client->data.user.fifo_pool_size) {
1920
/* change pool size */
1921
rc = snd_seq_fifo_resize(client->data.user.fifo, info.input_pool);
1922
if (rc < 0)
1923
return rc;
1924
client->data.user.fifo_pool_size = info.input_pool;
1925
}
1926
if (info.output_room >= 1 &&
1927
info.output_room <= client->pool->size) {
1928
client->pool->room = info.output_room;
1929
}
1930
1931
return snd_seq_ioctl_get_client_pool(client, arg);
1932
}
1933
1934
1935
/* REMOVE_EVENTS ioctl() */
1936
static int snd_seq_ioctl_remove_events(struct snd_seq_client *client,
1937
void __user *arg)
1938
{
1939
struct snd_seq_remove_events info;
1940
1941
if (copy_from_user(&info, arg, sizeof(info)))
1942
return -EFAULT;
1943
1944
/*
1945
* Input mostly not implemented XXX.
1946
*/
1947
if (info.remove_mode & SNDRV_SEQ_REMOVE_INPUT) {
1948
/*
1949
* No restrictions so for a user client we can clear
1950
* the whole fifo
1951
*/
1952
if (client->type == USER_CLIENT)
1953
snd_seq_fifo_clear(client->data.user.fifo);
1954
}
1955
1956
if (info.remove_mode & SNDRV_SEQ_REMOVE_OUTPUT)
1957
snd_seq_queue_remove_cells(client->number, &info);
1958
1959
return 0;
1960
}
1961
1962
1963
/*
1964
* get subscription info
1965
*/
1966
static int snd_seq_ioctl_get_subscription(struct snd_seq_client *client,
1967
void __user *arg)
1968
{
1969
int result;
1970
struct snd_seq_client *sender = NULL;
1971
struct snd_seq_client_port *sport = NULL;
1972
struct snd_seq_port_subscribe subs;
1973
struct snd_seq_subscribers *p;
1974
1975
if (copy_from_user(&subs, arg, sizeof(subs)))
1976
return -EFAULT;
1977
1978
result = -EINVAL;
1979
if ((sender = snd_seq_client_use_ptr(subs.sender.client)) == NULL)
1980
goto __end;
1981
if ((sport = snd_seq_port_use_ptr(sender, subs.sender.port)) == NULL)
1982
goto __end;
1983
p = snd_seq_port_get_subscription(&sport->c_src, &subs.dest);
1984
if (p) {
1985
result = 0;
1986
subs = p->info;
1987
} else
1988
result = -ENOENT;
1989
1990
__end:
1991
if (sport)
1992
snd_seq_port_unlock(sport);
1993
if (sender)
1994
snd_seq_client_unlock(sender);
1995
if (result >= 0) {
1996
if (copy_to_user(arg, &subs, sizeof(subs)))
1997
return -EFAULT;
1998
}
1999
return result;
2000
}
2001
2002
2003
/*
2004
* get subscription info - check only its presence
2005
*/
2006
static int snd_seq_ioctl_query_subs(struct snd_seq_client *client,
2007
void __user *arg)
2008
{
2009
int result = -ENXIO;
2010
struct snd_seq_client *cptr = NULL;
2011
struct snd_seq_client_port *port = NULL;
2012
struct snd_seq_query_subs subs;
2013
struct snd_seq_port_subs_info *group;
2014
struct list_head *p;
2015
int i;
2016
2017
if (copy_from_user(&subs, arg, sizeof(subs)))
2018
return -EFAULT;
2019
2020
if ((cptr = snd_seq_client_use_ptr(subs.root.client)) == NULL)
2021
goto __end;
2022
if ((port = snd_seq_port_use_ptr(cptr, subs.root.port)) == NULL)
2023
goto __end;
2024
2025
switch (subs.type) {
2026
case SNDRV_SEQ_QUERY_SUBS_READ:
2027
group = &port->c_src;
2028
break;
2029
case SNDRV_SEQ_QUERY_SUBS_WRITE:
2030
group = &port->c_dest;
2031
break;
2032
default:
2033
goto __end;
2034
}
2035
2036
down_read(&group->list_mutex);
2037
/* search for the subscriber */
2038
subs.num_subs = group->count;
2039
i = 0;
2040
result = -ENOENT;
2041
list_for_each(p, &group->list_head) {
2042
if (i++ == subs.index) {
2043
/* found! */
2044
struct snd_seq_subscribers *s;
2045
if (subs.type == SNDRV_SEQ_QUERY_SUBS_READ) {
2046
s = list_entry(p, struct snd_seq_subscribers, src_list);
2047
subs.addr = s->info.dest;
2048
} else {
2049
s = list_entry(p, struct snd_seq_subscribers, dest_list);
2050
subs.addr = s->info.sender;
2051
}
2052
subs.flags = s->info.flags;
2053
subs.queue = s->info.queue;
2054
result = 0;
2055
break;
2056
}
2057
}
2058
up_read(&group->list_mutex);
2059
2060
__end:
2061
if (port)
2062
snd_seq_port_unlock(port);
2063
if (cptr)
2064
snd_seq_client_unlock(cptr);
2065
if (result >= 0) {
2066
if (copy_to_user(arg, &subs, sizeof(subs)))
2067
return -EFAULT;
2068
}
2069
return result;
2070
}
2071
2072
2073
/*
2074
* query next client
2075
*/
2076
static int snd_seq_ioctl_query_next_client(struct snd_seq_client *client,
2077
void __user *arg)
2078
{
2079
struct snd_seq_client *cptr = NULL;
2080
struct snd_seq_client_info info;
2081
2082
if (copy_from_user(&info, arg, sizeof(info)))
2083
return -EFAULT;
2084
2085
/* search for next client */
2086
info.client++;
2087
if (info.client < 0)
2088
info.client = 0;
2089
for (; info.client < SNDRV_SEQ_MAX_CLIENTS; info.client++) {
2090
cptr = snd_seq_client_use_ptr(info.client);
2091
if (cptr)
2092
break; /* found */
2093
}
2094
if (cptr == NULL)
2095
return -ENOENT;
2096
2097
get_client_info(cptr, &info);
2098
snd_seq_client_unlock(cptr);
2099
2100
if (copy_to_user(arg, &info, sizeof(info)))
2101
return -EFAULT;
2102
return 0;
2103
}
2104
2105
/*
2106
* query next port
2107
*/
2108
static int snd_seq_ioctl_query_next_port(struct snd_seq_client *client,
2109
void __user *arg)
2110
{
2111
struct snd_seq_client *cptr;
2112
struct snd_seq_client_port *port = NULL;
2113
struct snd_seq_port_info info;
2114
2115
if (copy_from_user(&info, arg, sizeof(info)))
2116
return -EFAULT;
2117
cptr = snd_seq_client_use_ptr(info.addr.client);
2118
if (cptr == NULL)
2119
return -ENXIO;
2120
2121
/* search for next port */
2122
info.addr.port++;
2123
port = snd_seq_port_query_nearest(cptr, &info);
2124
if (port == NULL) {
2125
snd_seq_client_unlock(cptr);
2126
return -ENOENT;
2127
}
2128
2129
/* get port info */
2130
info.addr = port->addr;
2131
snd_seq_get_port_info(port, &info);
2132
snd_seq_port_unlock(port);
2133
snd_seq_client_unlock(cptr);
2134
2135
if (copy_to_user(arg, &info, sizeof(info)))
2136
return -EFAULT;
2137
return 0;
2138
}
2139
2140
/* -------------------------------------------------------- */
2141
2142
static struct seq_ioctl_table {
2143
unsigned int cmd;
2144
int (*func)(struct snd_seq_client *client, void __user * arg);
2145
} ioctl_tables[] = {
2146
{ SNDRV_SEQ_IOCTL_SYSTEM_INFO, snd_seq_ioctl_system_info },
2147
{ SNDRV_SEQ_IOCTL_RUNNING_MODE, snd_seq_ioctl_running_mode },
2148
{ SNDRV_SEQ_IOCTL_GET_CLIENT_INFO, snd_seq_ioctl_get_client_info },
2149
{ SNDRV_SEQ_IOCTL_SET_CLIENT_INFO, snd_seq_ioctl_set_client_info },
2150
{ SNDRV_SEQ_IOCTL_CREATE_PORT, snd_seq_ioctl_create_port },
2151
{ SNDRV_SEQ_IOCTL_DELETE_PORT, snd_seq_ioctl_delete_port },
2152
{ SNDRV_SEQ_IOCTL_GET_PORT_INFO, snd_seq_ioctl_get_port_info },
2153
{ SNDRV_SEQ_IOCTL_SET_PORT_INFO, snd_seq_ioctl_set_port_info },
2154
{ SNDRV_SEQ_IOCTL_SUBSCRIBE_PORT, snd_seq_ioctl_subscribe_port },
2155
{ SNDRV_SEQ_IOCTL_UNSUBSCRIBE_PORT, snd_seq_ioctl_unsubscribe_port },
2156
{ SNDRV_SEQ_IOCTL_CREATE_QUEUE, snd_seq_ioctl_create_queue },
2157
{ SNDRV_SEQ_IOCTL_DELETE_QUEUE, snd_seq_ioctl_delete_queue },
2158
{ SNDRV_SEQ_IOCTL_GET_QUEUE_INFO, snd_seq_ioctl_get_queue_info },
2159
{ SNDRV_SEQ_IOCTL_SET_QUEUE_INFO, snd_seq_ioctl_set_queue_info },
2160
{ SNDRV_SEQ_IOCTL_GET_NAMED_QUEUE, snd_seq_ioctl_get_named_queue },
2161
{ SNDRV_SEQ_IOCTL_GET_QUEUE_STATUS, snd_seq_ioctl_get_queue_status },
2162
{ SNDRV_SEQ_IOCTL_GET_QUEUE_TEMPO, snd_seq_ioctl_get_queue_tempo },
2163
{ SNDRV_SEQ_IOCTL_SET_QUEUE_TEMPO, snd_seq_ioctl_set_queue_tempo },
2164
{ SNDRV_SEQ_IOCTL_GET_QUEUE_TIMER, snd_seq_ioctl_get_queue_timer },
2165
{ SNDRV_SEQ_IOCTL_SET_QUEUE_TIMER, snd_seq_ioctl_set_queue_timer },
2166
{ SNDRV_SEQ_IOCTL_GET_QUEUE_CLIENT, snd_seq_ioctl_get_queue_client },
2167
{ SNDRV_SEQ_IOCTL_SET_QUEUE_CLIENT, snd_seq_ioctl_set_queue_client },
2168
{ SNDRV_SEQ_IOCTL_GET_CLIENT_POOL, snd_seq_ioctl_get_client_pool },
2169
{ SNDRV_SEQ_IOCTL_SET_CLIENT_POOL, snd_seq_ioctl_set_client_pool },
2170
{ SNDRV_SEQ_IOCTL_GET_SUBSCRIPTION, snd_seq_ioctl_get_subscription },
2171
{ SNDRV_SEQ_IOCTL_QUERY_NEXT_CLIENT, snd_seq_ioctl_query_next_client },
2172
{ SNDRV_SEQ_IOCTL_QUERY_NEXT_PORT, snd_seq_ioctl_query_next_port },
2173
{ SNDRV_SEQ_IOCTL_REMOVE_EVENTS, snd_seq_ioctl_remove_events },
2174
{ SNDRV_SEQ_IOCTL_QUERY_SUBS, snd_seq_ioctl_query_subs },
2175
{ 0, NULL },
2176
};
2177
2178
static int snd_seq_do_ioctl(struct snd_seq_client *client, unsigned int cmd,
2179
void __user *arg)
2180
{
2181
struct seq_ioctl_table *p;
2182
2183
switch (cmd) {
2184
case SNDRV_SEQ_IOCTL_PVERSION:
2185
/* return sequencer version number */
2186
return put_user(SNDRV_SEQ_VERSION, (int __user *)arg) ? -EFAULT : 0;
2187
case SNDRV_SEQ_IOCTL_CLIENT_ID:
2188
/* return the id of this client */
2189
return put_user(client->number, (int __user *)arg) ? -EFAULT : 0;
2190
}
2191
2192
if (! arg)
2193
return -EFAULT;
2194
for (p = ioctl_tables; p->cmd; p++) {
2195
if (p->cmd == cmd)
2196
return p->func(client, arg);
2197
}
2198
snd_printd("seq unknown ioctl() 0x%x (type='%c', number=0x%02x)\n",
2199
cmd, _IOC_TYPE(cmd), _IOC_NR(cmd));
2200
return -ENOTTY;
2201
}
2202
2203
2204
static long snd_seq_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
2205
{
2206
struct snd_seq_client *client = file->private_data;
2207
2208
if (snd_BUG_ON(!client))
2209
return -ENXIO;
2210
2211
return snd_seq_do_ioctl(client, cmd, (void __user *) arg);
2212
}
2213
2214
#ifdef CONFIG_COMPAT
2215
#include "seq_compat.c"
2216
#else
2217
#define snd_seq_ioctl_compat NULL
2218
#endif
2219
2220
/* -------------------------------------------------------- */
2221
2222
2223
/* exported to kernel modules */
2224
int snd_seq_create_kernel_client(struct snd_card *card, int client_index,
2225
const char *name_fmt, ...)
2226
{
2227
struct snd_seq_client *client;
2228
va_list args;
2229
2230
if (snd_BUG_ON(in_interrupt()))
2231
return -EBUSY;
2232
2233
if (card && client_index >= SNDRV_SEQ_CLIENTS_PER_CARD)
2234
return -EINVAL;
2235
if (card == NULL && client_index >= SNDRV_SEQ_GLOBAL_CLIENTS)
2236
return -EINVAL;
2237
2238
if (mutex_lock_interruptible(&register_mutex))
2239
return -ERESTARTSYS;
2240
2241
if (card) {
2242
client_index += SNDRV_SEQ_GLOBAL_CLIENTS
2243
+ card->number * SNDRV_SEQ_CLIENTS_PER_CARD;
2244
if (client_index >= SNDRV_SEQ_DYNAMIC_CLIENTS_BEGIN)
2245
client_index = -1;
2246
}
2247
2248
/* empty write queue as default */
2249
client = seq_create_client1(client_index, 0);
2250
if (client == NULL) {
2251
mutex_unlock(&register_mutex);
2252
return -EBUSY; /* failure code */
2253
}
2254
usage_alloc(&client_usage, 1);
2255
2256
client->accept_input = 1;
2257
client->accept_output = 1;
2258
2259
va_start(args, name_fmt);
2260
vsnprintf(client->name, sizeof(client->name), name_fmt, args);
2261
va_end(args);
2262
2263
client->type = KERNEL_CLIENT;
2264
mutex_unlock(&register_mutex);
2265
2266
/* make others aware this new client */
2267
snd_seq_system_client_ev_client_start(client->number);
2268
2269
/* return client number to caller */
2270
return client->number;
2271
}
2272
2273
EXPORT_SYMBOL(snd_seq_create_kernel_client);
2274
2275
/* exported to kernel modules */
2276
int snd_seq_delete_kernel_client(int client)
2277
{
2278
struct snd_seq_client *ptr;
2279
2280
if (snd_BUG_ON(in_interrupt()))
2281
return -EBUSY;
2282
2283
ptr = clientptr(client);
2284
if (ptr == NULL)
2285
return -EINVAL;
2286
2287
seq_free_client(ptr);
2288
kfree(ptr);
2289
return 0;
2290
}
2291
2292
EXPORT_SYMBOL(snd_seq_delete_kernel_client);
2293
2294
/* skeleton to enqueue event, called from snd_seq_kernel_client_enqueue
2295
* and snd_seq_kernel_client_enqueue_blocking
2296
*/
2297
static int kernel_client_enqueue(int client, struct snd_seq_event *ev,
2298
struct file *file, int blocking,
2299
int atomic, int hop)
2300
{
2301
struct snd_seq_client *cptr;
2302
int result;
2303
2304
if (snd_BUG_ON(!ev))
2305
return -EINVAL;
2306
2307
if (ev->type == SNDRV_SEQ_EVENT_NONE)
2308
return 0; /* ignore this */
2309
if (ev->type == SNDRV_SEQ_EVENT_KERNEL_ERROR)
2310
return -EINVAL; /* quoted events can't be enqueued */
2311
2312
/* fill in client number */
2313
ev->source.client = client;
2314
2315
if (check_event_type_and_length(ev))
2316
return -EINVAL;
2317
2318
cptr = snd_seq_client_use_ptr(client);
2319
if (cptr == NULL)
2320
return -EINVAL;
2321
2322
if (! cptr->accept_output)
2323
result = -EPERM;
2324
else /* send it */
2325
result = snd_seq_client_enqueue_event(cptr, ev, file, blocking, atomic, hop);
2326
2327
snd_seq_client_unlock(cptr);
2328
return result;
2329
}
2330
2331
/*
2332
* exported, called by kernel clients to enqueue events (w/o blocking)
2333
*
2334
* RETURN VALUE: zero if succeed, negative if error
2335
*/
2336
int snd_seq_kernel_client_enqueue(int client, struct snd_seq_event * ev,
2337
int atomic, int hop)
2338
{
2339
return kernel_client_enqueue(client, ev, NULL, 0, atomic, hop);
2340
}
2341
2342
EXPORT_SYMBOL(snd_seq_kernel_client_enqueue);
2343
2344
/*
2345
* exported, called by kernel clients to enqueue events (with blocking)
2346
*
2347
* RETURN VALUE: zero if succeed, negative if error
2348
*/
2349
int snd_seq_kernel_client_enqueue_blocking(int client, struct snd_seq_event * ev,
2350
struct file *file,
2351
int atomic, int hop)
2352
{
2353
return kernel_client_enqueue(client, ev, file, 1, atomic, hop);
2354
}
2355
2356
EXPORT_SYMBOL(snd_seq_kernel_client_enqueue_blocking);
2357
2358
/*
2359
* exported, called by kernel clients to dispatch events directly to other
2360
* clients, bypassing the queues. Event time-stamp will be updated.
2361
*
2362
* RETURN VALUE: negative = delivery failed,
2363
* zero, or positive: the number of delivered events
2364
*/
2365
int snd_seq_kernel_client_dispatch(int client, struct snd_seq_event * ev,
2366
int atomic, int hop)
2367
{
2368
struct snd_seq_client *cptr;
2369
int result;
2370
2371
if (snd_BUG_ON(!ev))
2372
return -EINVAL;
2373
2374
/* fill in client number */
2375
ev->queue = SNDRV_SEQ_QUEUE_DIRECT;
2376
ev->source.client = client;
2377
2378
if (check_event_type_and_length(ev))
2379
return -EINVAL;
2380
2381
cptr = snd_seq_client_use_ptr(client);
2382
if (cptr == NULL)
2383
return -EINVAL;
2384
2385
if (!cptr->accept_output)
2386
result = -EPERM;
2387
else
2388
result = snd_seq_deliver_event(cptr, ev, atomic, hop);
2389
2390
snd_seq_client_unlock(cptr);
2391
return result;
2392
}
2393
2394
EXPORT_SYMBOL(snd_seq_kernel_client_dispatch);
2395
2396
/*
2397
* exported, called by kernel clients to perform same functions as with
2398
* userland ioctl()
2399
*/
2400
int snd_seq_kernel_client_ctl(int clientid, unsigned int cmd, void *arg)
2401
{
2402
struct snd_seq_client *client;
2403
mm_segment_t fs;
2404
int result;
2405
2406
client = clientptr(clientid);
2407
if (client == NULL)
2408
return -ENXIO;
2409
fs = snd_enter_user();
2410
result = snd_seq_do_ioctl(client, cmd, (void __force __user *)arg);
2411
snd_leave_user(fs);
2412
return result;
2413
}
2414
2415
EXPORT_SYMBOL(snd_seq_kernel_client_ctl);
2416
2417
/* exported (for OSS emulator) */
2418
int snd_seq_kernel_client_write_poll(int clientid, struct file *file, poll_table *wait)
2419
{
2420
struct snd_seq_client *client;
2421
2422
client = clientptr(clientid);
2423
if (client == NULL)
2424
return -ENXIO;
2425
2426
if (! snd_seq_write_pool_allocated(client))
2427
return 1;
2428
if (snd_seq_pool_poll_wait(client->pool, file, wait))
2429
return 1;
2430
return 0;
2431
}
2432
2433
EXPORT_SYMBOL(snd_seq_kernel_client_write_poll);
2434
2435
/*---------------------------------------------------------------------------*/
2436
2437
#ifdef CONFIG_PROC_FS
2438
/*
2439
* /proc interface
2440
*/
2441
static void snd_seq_info_dump_subscribers(struct snd_info_buffer *buffer,
2442
struct snd_seq_port_subs_info *group,
2443
int is_src, char *msg)
2444
{
2445
struct list_head *p;
2446
struct snd_seq_subscribers *s;
2447
int count = 0;
2448
2449
down_read(&group->list_mutex);
2450
if (list_empty(&group->list_head)) {
2451
up_read(&group->list_mutex);
2452
return;
2453
}
2454
snd_iprintf(buffer, msg);
2455
list_for_each(p, &group->list_head) {
2456
if (is_src)
2457
s = list_entry(p, struct snd_seq_subscribers, src_list);
2458
else
2459
s = list_entry(p, struct snd_seq_subscribers, dest_list);
2460
if (count++)
2461
snd_iprintf(buffer, ", ");
2462
snd_iprintf(buffer, "%d:%d",
2463
is_src ? s->info.dest.client : s->info.sender.client,
2464
is_src ? s->info.dest.port : s->info.sender.port);
2465
if (s->info.flags & SNDRV_SEQ_PORT_SUBS_TIMESTAMP)
2466
snd_iprintf(buffer, "[%c:%d]", ((s->info.flags & SNDRV_SEQ_PORT_SUBS_TIME_REAL) ? 'r' : 't'), s->info.queue);
2467
if (group->exclusive)
2468
snd_iprintf(buffer, "[ex]");
2469
}
2470
up_read(&group->list_mutex);
2471
snd_iprintf(buffer, "\n");
2472
}
2473
2474
#define FLAG_PERM_RD(perm) ((perm) & SNDRV_SEQ_PORT_CAP_READ ? ((perm) & SNDRV_SEQ_PORT_CAP_SUBS_READ ? 'R' : 'r') : '-')
2475
#define FLAG_PERM_WR(perm) ((perm) & SNDRV_SEQ_PORT_CAP_WRITE ? ((perm) & SNDRV_SEQ_PORT_CAP_SUBS_WRITE ? 'W' : 'w') : '-')
2476
#define FLAG_PERM_EX(perm) ((perm) & SNDRV_SEQ_PORT_CAP_NO_EXPORT ? '-' : 'e')
2477
2478
#define FLAG_PERM_DUPLEX(perm) ((perm) & SNDRV_SEQ_PORT_CAP_DUPLEX ? 'X' : '-')
2479
2480
static void snd_seq_info_dump_ports(struct snd_info_buffer *buffer,
2481
struct snd_seq_client *client)
2482
{
2483
struct snd_seq_client_port *p;
2484
2485
mutex_lock(&client->ports_mutex);
2486
list_for_each_entry(p, &client->ports_list_head, list) {
2487
snd_iprintf(buffer, " Port %3d : \"%s\" (%c%c%c%c)\n",
2488
p->addr.port, p->name,
2489
FLAG_PERM_RD(p->capability),
2490
FLAG_PERM_WR(p->capability),
2491
FLAG_PERM_EX(p->capability),
2492
FLAG_PERM_DUPLEX(p->capability));
2493
snd_seq_info_dump_subscribers(buffer, &p->c_src, 1, " Connecting To: ");
2494
snd_seq_info_dump_subscribers(buffer, &p->c_dest, 0, " Connected From: ");
2495
}
2496
mutex_unlock(&client->ports_mutex);
2497
}
2498
2499
2500
/* exported to seq_info.c */
2501
void snd_seq_info_clients_read(struct snd_info_entry *entry,
2502
struct snd_info_buffer *buffer)
2503
{
2504
int c;
2505
struct snd_seq_client *client;
2506
2507
snd_iprintf(buffer, "Client info\n");
2508
snd_iprintf(buffer, " cur clients : %d\n", client_usage.cur);
2509
snd_iprintf(buffer, " peak clients : %d\n", client_usage.peak);
2510
snd_iprintf(buffer, " max clients : %d\n", SNDRV_SEQ_MAX_CLIENTS);
2511
snd_iprintf(buffer, "\n");
2512
2513
/* list the client table */
2514
for (c = 0; c < SNDRV_SEQ_MAX_CLIENTS; c++) {
2515
client = snd_seq_client_use_ptr(c);
2516
if (client == NULL)
2517
continue;
2518
if (client->type == NO_CLIENT) {
2519
snd_seq_client_unlock(client);
2520
continue;
2521
}
2522
2523
snd_iprintf(buffer, "Client %3d : \"%s\" [%s]\n",
2524
c, client->name,
2525
client->type == USER_CLIENT ? "User" : "Kernel");
2526
snd_seq_info_dump_ports(buffer, client);
2527
if (snd_seq_write_pool_allocated(client)) {
2528
snd_iprintf(buffer, " Output pool :\n");
2529
snd_seq_info_pool(buffer, client->pool, " ");
2530
}
2531
if (client->type == USER_CLIENT && client->data.user.fifo &&
2532
client->data.user.fifo->pool) {
2533
snd_iprintf(buffer, " Input pool :\n");
2534
snd_seq_info_pool(buffer, client->data.user.fifo->pool, " ");
2535
}
2536
snd_seq_client_unlock(client);
2537
}
2538
}
2539
#endif /* CONFIG_PROC_FS */
2540
2541
/*---------------------------------------------------------------------------*/
2542
2543
2544
/*
2545
* REGISTRATION PART
2546
*/
2547
2548
static const struct file_operations snd_seq_f_ops =
2549
{
2550
.owner = THIS_MODULE,
2551
.read = snd_seq_read,
2552
.write = snd_seq_write,
2553
.open = snd_seq_open,
2554
.release = snd_seq_release,
2555
.llseek = no_llseek,
2556
.poll = snd_seq_poll,
2557
.unlocked_ioctl = snd_seq_ioctl,
2558
.compat_ioctl = snd_seq_ioctl_compat,
2559
};
2560
2561
/*
2562
* register sequencer device
2563
*/
2564
int __init snd_sequencer_device_init(void)
2565
{
2566
int err;
2567
2568
if (mutex_lock_interruptible(&register_mutex))
2569
return -ERESTARTSYS;
2570
2571
if ((err = snd_register_device(SNDRV_DEVICE_TYPE_SEQUENCER, NULL, 0,
2572
&snd_seq_f_ops, NULL, "seq")) < 0) {
2573
mutex_unlock(&register_mutex);
2574
return err;
2575
}
2576
2577
mutex_unlock(&register_mutex);
2578
2579
return 0;
2580
}
2581
2582
2583
2584
/*
2585
* unregister sequencer device
2586
*/
2587
void __exit snd_sequencer_device_done(void)
2588
{
2589
snd_unregister_device(SNDRV_DEVICE_TYPE_SEQUENCER, NULL, 0);
2590
}
2591
2592