Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/drivers/gpio/gpio-aggregator.c
50360 views
1
// SPDX-License-Identifier: GPL-2.0-only
2
//
3
// GPIO Aggregator
4
//
5
// Copyright (C) 2019-2020 Glider bv
6
7
#define DRV_NAME "gpio-aggregator"
8
#define pr_fmt(fmt) DRV_NAME ": " fmt
9
10
#include <linux/bitmap.h>
11
#include <linux/bitops.h>
12
#include <linux/configfs.h>
13
#include <linux/ctype.h>
14
#include <linux/delay.h>
15
#include <linux/export.h>
16
#include <linux/idr.h>
17
#include <linux/kernel.h>
18
#include <linux/list.h>
19
#include <linux/lockdep.h>
20
#include <linux/mod_devicetable.h>
21
#include <linux/module.h>
22
#include <linux/mutex.h>
23
#include <linux/overflow.h>
24
#include <linux/platform_device.h>
25
#include <linux/property.h>
26
#include <linux/slab.h>
27
#include <linux/spinlock.h>
28
#include <linux/string.h>
29
30
#include <linux/gpio/consumer.h>
31
#include <linux/gpio/driver.h>
32
#include <linux/gpio/forwarder.h>
33
#include <linux/gpio/machine.h>
34
35
#include "dev-sync-probe.h"
36
37
#define AGGREGATOR_MAX_GPIOS 512
38
#define AGGREGATOR_LEGACY_PREFIX "_sysfs"
39
40
/*
41
* GPIO Aggregator sysfs interface
42
*/
43
44
struct gpio_aggregator {
45
struct dev_sync_probe_data probe_data;
46
struct config_group group;
47
struct gpiod_lookup_table *lookups;
48
struct mutex lock;
49
int id;
50
51
/* List of gpio_aggregator_line. Always added in order */
52
struct list_head list_head;
53
54
/* used by legacy sysfs interface only */
55
bool init_via_sysfs;
56
char args[];
57
};
58
59
struct gpio_aggregator_line {
60
struct config_group group;
61
struct gpio_aggregator *parent;
62
struct list_head entry;
63
64
/* Line index within the aggregator device */
65
unsigned int idx;
66
67
/* Custom name for the virtual line */
68
const char *name;
69
/* GPIO chip label or line name */
70
const char *key;
71
/* Can be negative to indicate lookup by line name */
72
int offset;
73
74
enum gpio_lookup_flags flags;
75
};
76
77
struct gpio_aggregator_pdev_meta {
78
bool init_via_sysfs;
79
};
80
81
static DEFINE_MUTEX(gpio_aggregator_lock); /* protects idr */
82
static DEFINE_IDR(gpio_aggregator_idr);
83
84
static int gpio_aggregator_alloc(struct gpio_aggregator **aggr, size_t arg_size)
85
{
86
int ret;
87
88
struct gpio_aggregator *new __free(kfree) = kzalloc(
89
sizeof(*new) + arg_size, GFP_KERNEL);
90
if (!new)
91
return -ENOMEM;
92
93
scoped_guard(mutex, &gpio_aggregator_lock)
94
ret = idr_alloc(&gpio_aggregator_idr, new, 0, 0, GFP_KERNEL);
95
96
if (ret < 0)
97
return ret;
98
99
new->id = ret;
100
INIT_LIST_HEAD(&new->list_head);
101
mutex_init(&new->lock);
102
*aggr = no_free_ptr(new);
103
return 0;
104
}
105
106
static void gpio_aggregator_free(struct gpio_aggregator *aggr)
107
{
108
scoped_guard(mutex, &gpio_aggregator_lock)
109
idr_remove(&gpio_aggregator_idr, aggr->id);
110
111
mutex_destroy(&aggr->lock);
112
kfree(aggr);
113
}
114
115
static int gpio_aggregator_add_gpio(struct gpio_aggregator *aggr,
116
const char *key, int hwnum, unsigned int *n)
117
{
118
struct gpiod_lookup_table *lookups;
119
120
lookups = krealloc(aggr->lookups, struct_size(lookups, table, *n + 2),
121
GFP_KERNEL);
122
if (!lookups)
123
return -ENOMEM;
124
125
lookups->table[*n] = GPIO_LOOKUP_IDX(key, hwnum, NULL, *n, 0);
126
127
(*n)++;
128
memset(&lookups->table[*n], 0, sizeof(lookups->table[*n]));
129
130
aggr->lookups = lookups;
131
return 0;
132
}
133
134
static bool gpio_aggregator_is_active(struct gpio_aggregator *aggr)
135
{
136
lockdep_assert_held(&aggr->lock);
137
138
return aggr->probe_data.pdev && platform_get_drvdata(aggr->probe_data.pdev);
139
}
140
141
/* Only aggregators created via legacy sysfs can be "activating". */
142
static bool gpio_aggregator_is_activating(struct gpio_aggregator *aggr)
143
{
144
lockdep_assert_held(&aggr->lock);
145
146
return aggr->probe_data.pdev && !platform_get_drvdata(aggr->probe_data.pdev);
147
}
148
149
static size_t gpio_aggregator_count_lines(struct gpio_aggregator *aggr)
150
{
151
lockdep_assert_held(&aggr->lock);
152
153
return list_count_nodes(&aggr->list_head);
154
}
155
156
static struct gpio_aggregator_line *
157
gpio_aggregator_line_alloc(struct gpio_aggregator *parent, unsigned int idx,
158
char *key, int offset)
159
{
160
struct gpio_aggregator_line *line;
161
162
line = kzalloc(sizeof(*line), GFP_KERNEL);
163
if (!line)
164
return ERR_PTR(-ENOMEM);
165
166
if (key) {
167
line->key = kstrdup(key, GFP_KERNEL);
168
if (!line->key) {
169
kfree(line);
170
return ERR_PTR(-ENOMEM);
171
}
172
}
173
174
line->flags = GPIO_LOOKUP_FLAGS_DEFAULT;
175
line->parent = parent;
176
line->idx = idx;
177
line->offset = offset;
178
INIT_LIST_HEAD(&line->entry);
179
180
return line;
181
}
182
183
static void gpio_aggregator_line_add(struct gpio_aggregator *aggr,
184
struct gpio_aggregator_line *line)
185
{
186
struct gpio_aggregator_line *tmp;
187
188
lockdep_assert_held(&aggr->lock);
189
190
list_for_each_entry(tmp, &aggr->list_head, entry) {
191
if (tmp->idx > line->idx) {
192
list_add_tail(&line->entry, &tmp->entry);
193
return;
194
}
195
}
196
list_add_tail(&line->entry, &aggr->list_head);
197
}
198
199
static void gpio_aggregator_line_del(struct gpio_aggregator *aggr,
200
struct gpio_aggregator_line *line)
201
{
202
lockdep_assert_held(&aggr->lock);
203
204
list_del(&line->entry);
205
}
206
207
static void gpio_aggregator_free_lines(struct gpio_aggregator *aggr)
208
{
209
struct gpio_aggregator_line *line, *tmp;
210
211
list_for_each_entry_safe(line, tmp, &aggr->list_head, entry) {
212
configfs_unregister_group(&line->group);
213
/*
214
* Normally, we acquire aggr->lock within the configfs
215
* callback. However, in the legacy sysfs interface case,
216
* calling configfs_(un)register_group while holding
217
* aggr->lock could cause a deadlock. Fortunately, this is
218
* unnecessary because the new_device/delete_device path
219
* and the module unload path are mutually exclusive,
220
* thanks to an explicit try_module_get. That's why this
221
* minimal scoped_guard suffices.
222
*/
223
scoped_guard(mutex, &aggr->lock)
224
gpio_aggregator_line_del(aggr, line);
225
kfree(line->key);
226
kfree(line->name);
227
kfree(line);
228
}
229
}
230
231
232
/*
233
* GPIO Forwarder
234
*/
235
236
struct gpiochip_fwd_timing {
237
u32 ramp_up_us;
238
u32 ramp_down_us;
239
};
240
241
struct gpiochip_fwd {
242
struct gpio_chip chip;
243
struct gpio_desc **descs;
244
union {
245
struct mutex mlock; /* protects tmp[] if can_sleep */
246
spinlock_t slock; /* protects tmp[] if !can_sleep */
247
};
248
struct gpiochip_fwd_timing *delay_timings;
249
void *data;
250
unsigned long *valid_mask;
251
unsigned long tmp[]; /* values and descs for multiple ops */
252
};
253
254
#define fwd_tmp_values(fwd) (&(fwd)->tmp[0])
255
#define fwd_tmp_descs(fwd) ((void *)&(fwd)->tmp[BITS_TO_LONGS((fwd)->chip.ngpio)])
256
257
#define fwd_tmp_size(ngpios) (BITS_TO_LONGS((ngpios)) + (ngpios))
258
259
static int gpio_fwd_request(struct gpio_chip *chip, unsigned int offset)
260
{
261
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
262
263
return test_bit(offset, fwd->valid_mask) ? 0 : -ENODEV;
264
}
265
266
static int gpio_fwd_get_direction(struct gpio_chip *chip, unsigned int offset)
267
{
268
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
269
270
/*
271
* get_direction() is called during gpiochip registration, return
272
* -ENODEV if there is no GPIO desc for the line.
273
*/
274
if (!test_bit(offset, fwd->valid_mask))
275
return -ENODEV;
276
277
return gpiod_get_direction(fwd->descs[offset]);
278
}
279
280
static int gpio_fwd_direction_input(struct gpio_chip *chip, unsigned int offset)
281
{
282
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
283
284
return gpiod_direction_input(fwd->descs[offset]);
285
}
286
287
static int gpio_fwd_direction_output(struct gpio_chip *chip,
288
unsigned int offset, int value)
289
{
290
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
291
292
return gpiod_direction_output(fwd->descs[offset], value);
293
}
294
295
static int gpio_fwd_get(struct gpio_chip *chip, unsigned int offset)
296
{
297
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
298
299
return chip->can_sleep ? gpiod_get_value_cansleep(fwd->descs[offset])
300
: gpiod_get_value(fwd->descs[offset]);
301
}
302
303
static int gpio_fwd_get_multiple(struct gpiochip_fwd *fwd, unsigned long *mask,
304
unsigned long *bits)
305
{
306
struct gpio_desc **descs = fwd_tmp_descs(fwd);
307
unsigned long *values = fwd_tmp_values(fwd);
308
unsigned int i, j = 0;
309
int error;
310
311
bitmap_clear(values, 0, fwd->chip.ngpio);
312
for_each_set_bit(i, mask, fwd->chip.ngpio)
313
descs[j++] = fwd->descs[i];
314
315
if (fwd->chip.can_sleep)
316
error = gpiod_get_array_value_cansleep(j, descs, NULL, values);
317
else
318
error = gpiod_get_array_value(j, descs, NULL, values);
319
if (error)
320
return error;
321
322
j = 0;
323
for_each_set_bit(i, mask, fwd->chip.ngpio)
324
__assign_bit(i, bits, test_bit(j++, values));
325
326
return 0;
327
}
328
329
static int gpio_fwd_get_multiple_locked(struct gpio_chip *chip,
330
unsigned long *mask, unsigned long *bits)
331
{
332
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
333
unsigned long flags;
334
int error;
335
336
if (chip->can_sleep) {
337
mutex_lock(&fwd->mlock);
338
error = gpio_fwd_get_multiple(fwd, mask, bits);
339
mutex_unlock(&fwd->mlock);
340
} else {
341
spin_lock_irqsave(&fwd->slock, flags);
342
error = gpio_fwd_get_multiple(fwd, mask, bits);
343
spin_unlock_irqrestore(&fwd->slock, flags);
344
}
345
346
return error;
347
}
348
349
static void gpio_fwd_delay(struct gpio_chip *chip, unsigned int offset, int value)
350
{
351
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
352
const struct gpiochip_fwd_timing *delay_timings;
353
bool is_active_low = gpiod_is_active_low(fwd->descs[offset]);
354
u32 delay_us;
355
356
delay_timings = &fwd->delay_timings[offset];
357
if ((!is_active_low && value) || (is_active_low && !value))
358
delay_us = delay_timings->ramp_up_us;
359
else
360
delay_us = delay_timings->ramp_down_us;
361
if (!delay_us)
362
return;
363
364
if (chip->can_sleep)
365
fsleep(delay_us);
366
else
367
udelay(delay_us);
368
}
369
370
static int gpio_fwd_set(struct gpio_chip *chip, unsigned int offset, int value)
371
{
372
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
373
int ret;
374
375
if (chip->can_sleep)
376
ret = gpiod_set_value_cansleep(fwd->descs[offset], value);
377
else
378
ret = gpiod_set_value(fwd->descs[offset], value);
379
if (ret)
380
return ret;
381
382
if (fwd->delay_timings)
383
gpio_fwd_delay(chip, offset, value);
384
385
return ret;
386
}
387
388
static int gpio_fwd_set_multiple(struct gpiochip_fwd *fwd, unsigned long *mask,
389
unsigned long *bits)
390
{
391
struct gpio_desc **descs = fwd_tmp_descs(fwd);
392
unsigned long *values = fwd_tmp_values(fwd);
393
unsigned int i, j = 0, ret;
394
395
for_each_set_bit(i, mask, fwd->chip.ngpio) {
396
__assign_bit(j, values, test_bit(i, bits));
397
descs[j++] = fwd->descs[i];
398
}
399
400
if (fwd->chip.can_sleep)
401
ret = gpiod_set_array_value_cansleep(j, descs, NULL, values);
402
else
403
ret = gpiod_set_array_value(j, descs, NULL, values);
404
405
return ret;
406
}
407
408
static int gpio_fwd_set_multiple_locked(struct gpio_chip *chip,
409
unsigned long *mask, unsigned long *bits)
410
{
411
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
412
unsigned long flags;
413
int ret;
414
415
if (chip->can_sleep) {
416
mutex_lock(&fwd->mlock);
417
ret = gpio_fwd_set_multiple(fwd, mask, bits);
418
mutex_unlock(&fwd->mlock);
419
} else {
420
spin_lock_irqsave(&fwd->slock, flags);
421
ret = gpio_fwd_set_multiple(fwd, mask, bits);
422
spin_unlock_irqrestore(&fwd->slock, flags);
423
}
424
425
return ret;
426
}
427
428
static int gpio_fwd_set_config(struct gpio_chip *chip, unsigned int offset,
429
unsigned long config)
430
{
431
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
432
433
return gpiod_set_config(fwd->descs[offset], config);
434
}
435
436
static int gpio_fwd_to_irq(struct gpio_chip *chip, unsigned int offset)
437
{
438
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
439
440
return gpiod_to_irq(fwd->descs[offset]);
441
}
442
443
/*
444
* The GPIO delay provides a way to configure platform specific delays
445
* for the GPIO ramp-up or ramp-down delays. This can serve the following
446
* purposes:
447
* - Open-drain output using an RC filter
448
*/
449
#define FWD_FEATURE_DELAY BIT(0)
450
451
#ifdef CONFIG_OF_GPIO
452
static int gpiochip_fwd_delay_of_xlate(struct gpio_chip *chip,
453
const struct of_phandle_args *gpiospec,
454
u32 *flags)
455
{
456
struct gpiochip_fwd *fwd = gpiochip_get_data(chip);
457
struct gpiochip_fwd_timing *timings;
458
u32 line;
459
460
if (gpiospec->args_count != chip->of_gpio_n_cells)
461
return -EINVAL;
462
463
line = gpiospec->args[0];
464
if (line >= chip->ngpio)
465
return -EINVAL;
466
467
timings = &fwd->delay_timings[line];
468
timings->ramp_up_us = gpiospec->args[1];
469
timings->ramp_down_us = gpiospec->args[2];
470
471
return line;
472
}
473
474
static int gpiochip_fwd_setup_delay_line(struct gpiochip_fwd *fwd)
475
{
476
struct gpio_chip *chip = &fwd->chip;
477
478
fwd->delay_timings = devm_kcalloc(chip->parent, chip->ngpio,
479
sizeof(*fwd->delay_timings),
480
GFP_KERNEL);
481
if (!fwd->delay_timings)
482
return -ENOMEM;
483
484
chip->of_xlate = gpiochip_fwd_delay_of_xlate;
485
chip->of_gpio_n_cells = 3;
486
487
return 0;
488
}
489
#else
490
static int gpiochip_fwd_setup_delay_line(struct gpiochip_fwd *fwd)
491
{
492
return 0;
493
}
494
#endif /* !CONFIG_OF_GPIO */
495
496
/**
497
* gpiochip_fwd_get_gpiochip - Get the GPIO chip for the GPIO forwarder
498
* @fwd: GPIO forwarder
499
*
500
* Returns: The GPIO chip for the GPIO forwarder
501
*/
502
struct gpio_chip *gpiochip_fwd_get_gpiochip(struct gpiochip_fwd *fwd)
503
{
504
return &fwd->chip;
505
}
506
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_get_gpiochip, "GPIO_FORWARDER");
507
508
/**
509
* gpiochip_fwd_get_data - Get driver-private data for the GPIO forwarder
510
* @fwd: GPIO forwarder
511
*
512
* Returns: The driver-private data for the GPIO forwarder
513
*/
514
void *gpiochip_fwd_get_data(struct gpiochip_fwd *fwd)
515
{
516
return fwd->data;
517
}
518
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_get_data, "GPIO_FORWARDER");
519
520
/**
521
* gpiochip_fwd_gpio_request - Request a line of the GPIO forwarder
522
* @fwd: GPIO forwarder
523
* @offset: the offset of the line to request
524
*
525
* Returns: 0 on success, or negative errno on failure.
526
*/
527
int gpiochip_fwd_gpio_request(struct gpiochip_fwd *fwd, unsigned int offset)
528
{
529
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
530
531
return gpio_fwd_request(gc, offset);
532
}
533
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_request, "GPIO_FORWARDER");
534
535
/**
536
* gpiochip_fwd_gpio_get_direction - Return the current direction of a GPIO forwarder line
537
* @fwd: GPIO forwarder
538
* @offset: the offset of the line
539
*
540
* Returns: 0 for output, 1 for input, or an error code in case of error.
541
*/
542
int gpiochip_fwd_gpio_get_direction(struct gpiochip_fwd *fwd, unsigned int offset)
543
{
544
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
545
546
return gpio_fwd_get_direction(gc, offset);
547
}
548
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_get_direction, "GPIO_FORWARDER");
549
550
/**
551
* gpiochip_fwd_gpio_direction_output - Set a GPIO forwarder line direction to
552
* output
553
* @fwd: GPIO forwarder
554
* @offset: the offset of the line
555
* @value: value to set
556
*
557
* Returns: 0 on success, or negative errno on failure.
558
*/
559
int gpiochip_fwd_gpio_direction_output(struct gpiochip_fwd *fwd, unsigned int offset,
560
int value)
561
{
562
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
563
564
return gpio_fwd_direction_output(gc, offset, value);
565
}
566
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_direction_output, "GPIO_FORWARDER");
567
568
/**
569
* gpiochip_fwd_gpio_direction_input - Set a GPIO forwarder line direction to input
570
* @fwd: GPIO forwarder
571
* @offset: the offset of the line
572
*
573
* Returns: 0 on success, or negative errno on failure.
574
*/
575
int gpiochip_fwd_gpio_direction_input(struct gpiochip_fwd *fwd, unsigned int offset)
576
{
577
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
578
579
return gpio_fwd_direction_input(gc, offset);
580
}
581
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_direction_input, "GPIO_FORWARDER");
582
583
/**
584
* gpiochip_fwd_gpio_get - Return a GPIO forwarder line's value
585
* @fwd: GPIO forwarder
586
* @offset: the offset of the line
587
*
588
* Returns: The GPIO's logical value, i.e. taking the ACTIVE_LOW status into
589
* account, or negative errno on failure.
590
*/
591
int gpiochip_fwd_gpio_get(struct gpiochip_fwd *fwd, unsigned int offset)
592
{
593
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
594
595
return gpio_fwd_get(gc, offset);
596
}
597
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_get, "GPIO_FORWARDER");
598
599
/**
600
* gpiochip_fwd_gpio_get_multiple - Get values for multiple GPIO forwarder lines
601
* @fwd: GPIO forwarder
602
* @mask: bit mask array; one bit per line; BITS_PER_LONG bits per word defines
603
* which lines are to be read
604
* @bits: bit value array; one bit per line; BITS_PER_LONG bits per word will
605
* contains the read values for the lines specified by mask
606
*
607
* Returns: 0 on success, or negative errno on failure.
608
*/
609
int gpiochip_fwd_gpio_get_multiple(struct gpiochip_fwd *fwd, unsigned long *mask,
610
unsigned long *bits)
611
{
612
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
613
614
return gpio_fwd_get_multiple_locked(gc, mask, bits);
615
}
616
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_get_multiple, "GPIO_FORWARDER");
617
618
/**
619
* gpiochip_fwd_gpio_set - Assign value to a GPIO forwarder line.
620
* @fwd: GPIO forwarder
621
* @offset: the offset of the line
622
* @value: value to set
623
*
624
* Returns: 0 on success, or negative errno on failure.
625
*/
626
int gpiochip_fwd_gpio_set(struct gpiochip_fwd *fwd, unsigned int offset, int value)
627
{
628
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
629
630
return gpio_fwd_set(gc, offset, value);
631
}
632
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_set, "GPIO_FORWARDER");
633
634
/**
635
* gpiochip_fwd_gpio_set_multiple - Assign values to multiple GPIO forwarder lines
636
* @fwd: GPIO forwarder
637
* @mask: bit mask array; one bit per output; BITS_PER_LONG bits per word
638
* defines which outputs are to be changed
639
* @bits: bit value array; one bit per output; BITS_PER_LONG bits per word
640
* defines the values the outputs specified by mask are to be set to
641
*
642
* Returns: 0 on success, or negative errno on failure.
643
*/
644
int gpiochip_fwd_gpio_set_multiple(struct gpiochip_fwd *fwd, unsigned long *mask,
645
unsigned long *bits)
646
{
647
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
648
649
return gpio_fwd_set_multiple_locked(gc, mask, bits);
650
}
651
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_set_multiple, "GPIO_FORWARDER");
652
653
/**
654
* gpiochip_fwd_gpio_set_config - Set @config for a GPIO forwarder line
655
* @fwd: GPIO forwarder
656
* @offset: the offset of the line
657
* @config: Same packed config format as generic pinconf
658
*
659
* Returns: 0 on success, %-ENOTSUPP if the controller doesn't support setting
660
* the configuration.
661
*/
662
int gpiochip_fwd_gpio_set_config(struct gpiochip_fwd *fwd, unsigned int offset,
663
unsigned long config)
664
{
665
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
666
667
return gpio_fwd_set_config(gc, offset, config);
668
}
669
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_set_config, "GPIO_FORWARDER");
670
671
/**
672
* gpiochip_fwd_gpio_to_irq - Return the IRQ corresponding to a GPIO forwarder line
673
* @fwd: GPIO forwarder
674
* @offset: the offset of the line
675
*
676
* Returns: The Linux IRQ corresponding to the passed line, or an error code in
677
* case of error.
678
*/
679
int gpiochip_fwd_gpio_to_irq(struct gpiochip_fwd *fwd, unsigned int offset)
680
{
681
struct gpio_chip *gc = gpiochip_fwd_get_gpiochip(fwd);
682
683
return gpio_fwd_to_irq(gc, offset);
684
}
685
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_gpio_to_irq, "GPIO_FORWARDER");
686
687
/**
688
* devm_gpiochip_fwd_alloc - Allocate and initialize a new GPIO forwarder
689
* @dev: Parent device pointer
690
* @ngpios: Number of GPIOs in the forwarder
691
*
692
* Returns: An opaque object pointer, or an ERR_PTR()-encoded negative error
693
* code on failure.
694
*/
695
struct gpiochip_fwd *devm_gpiochip_fwd_alloc(struct device *dev,
696
unsigned int ngpios)
697
{
698
struct gpiochip_fwd *fwd;
699
struct gpio_chip *chip;
700
701
fwd = devm_kzalloc(dev, struct_size(fwd, tmp, fwd_tmp_size(ngpios)), GFP_KERNEL);
702
if (!fwd)
703
return ERR_PTR(-ENOMEM);
704
705
fwd->descs = devm_kcalloc(dev, ngpios, sizeof(*fwd->descs), GFP_KERNEL);
706
if (!fwd->descs)
707
return ERR_PTR(-ENOMEM);
708
709
fwd->valid_mask = devm_bitmap_zalloc(dev, ngpios, GFP_KERNEL);
710
if (!fwd->valid_mask)
711
return ERR_PTR(-ENOMEM);
712
713
chip = &fwd->chip;
714
715
chip->label = dev_name(dev);
716
chip->parent = dev;
717
chip->owner = THIS_MODULE;
718
chip->request = gpio_fwd_request;
719
chip->get_direction = gpio_fwd_get_direction;
720
chip->direction_input = gpio_fwd_direction_input;
721
chip->direction_output = gpio_fwd_direction_output;
722
chip->get = gpio_fwd_get;
723
chip->get_multiple = gpio_fwd_get_multiple_locked;
724
chip->set = gpio_fwd_set;
725
chip->set_multiple = gpio_fwd_set_multiple_locked;
726
chip->set_config = gpio_fwd_set_config;
727
chip->to_irq = gpio_fwd_to_irq;
728
chip->base = -1;
729
chip->ngpio = ngpios;
730
731
return fwd;
732
}
733
EXPORT_SYMBOL_NS_GPL(devm_gpiochip_fwd_alloc, "GPIO_FORWARDER");
734
735
/**
736
* gpiochip_fwd_desc_add - Add a GPIO desc in the forwarder
737
* @fwd: GPIO forwarder
738
* @desc: GPIO descriptor to register
739
* @offset: offset for the GPIO in the forwarder
740
*
741
* Returns: 0 on success, or negative errno on failure.
742
*/
743
int gpiochip_fwd_desc_add(struct gpiochip_fwd *fwd, struct gpio_desc *desc,
744
unsigned int offset)
745
{
746
struct gpio_chip *chip = &fwd->chip;
747
748
if (offset >= chip->ngpio)
749
return -EINVAL;
750
751
if (test_and_set_bit(offset, fwd->valid_mask))
752
return -EEXIST;
753
754
/*
755
* If any of the GPIO lines are sleeping, then the entire forwarder
756
* will be sleeping.
757
*/
758
if (gpiod_cansleep(desc))
759
chip->can_sleep = true;
760
761
fwd->descs[offset] = desc;
762
763
dev_dbg(chip->parent, "%u => gpio %d irq %d\n", offset,
764
desc_to_gpio(desc), gpiod_to_irq(desc));
765
766
return 0;
767
}
768
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_desc_add, "GPIO_FORWARDER");
769
770
/**
771
* gpiochip_fwd_desc_free - Remove a GPIO desc from the forwarder
772
* @fwd: GPIO forwarder
773
* @offset: offset of GPIO desc to remove
774
*/
775
void gpiochip_fwd_desc_free(struct gpiochip_fwd *fwd, unsigned int offset)
776
{
777
if (test_and_clear_bit(offset, fwd->valid_mask))
778
gpiod_put(fwd->descs[offset]);
779
}
780
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_desc_free, "GPIO_FORWARDER");
781
782
/**
783
* gpiochip_fwd_register - Register a GPIO forwarder
784
* @fwd: GPIO forwarder
785
* @data: driver-private data associated with this forwarder
786
*
787
* Returns: 0 on success, or negative errno on failure.
788
*/
789
int gpiochip_fwd_register(struct gpiochip_fwd *fwd, void *data)
790
{
791
struct gpio_chip *chip = &fwd->chip;
792
793
/*
794
* Some gpio_desc were not registered. They will be registered at runtime
795
* but we have to suppose they can sleep.
796
*/
797
if (!bitmap_full(fwd->valid_mask, chip->ngpio))
798
chip->can_sleep = true;
799
800
if (chip->can_sleep)
801
mutex_init(&fwd->mlock);
802
else
803
spin_lock_init(&fwd->slock);
804
805
fwd->data = data;
806
807
return devm_gpiochip_add_data(chip->parent, chip, fwd);
808
}
809
EXPORT_SYMBOL_NS_GPL(gpiochip_fwd_register, "GPIO_FORWARDER");
810
811
/**
812
* gpiochip_fwd_create() - Create a new GPIO forwarder
813
* @dev: Parent device pointer
814
* @ngpios: Number of GPIOs in the forwarder.
815
* @descs: Array containing the GPIO descriptors to forward to.
816
* This array must contain @ngpios entries, and can be deallocated
817
* as the forwarder has its own array.
818
* @features: Bitwise ORed features as defined with FWD_FEATURE_*.
819
*
820
* This function creates a new gpiochip, which forwards all GPIO operations to
821
* the passed GPIO descriptors.
822
*
823
* Return: An opaque object pointer, or an ERR_PTR()-encoded negative error
824
* code on failure.
825
*/
826
static struct gpiochip_fwd *gpiochip_fwd_create(struct device *dev,
827
unsigned int ngpios,
828
struct gpio_desc *descs[],
829
unsigned long features)
830
{
831
struct gpiochip_fwd *fwd;
832
unsigned int i;
833
int error;
834
835
fwd = devm_gpiochip_fwd_alloc(dev, ngpios);
836
if (IS_ERR(fwd))
837
return fwd;
838
839
for (i = 0; i < ngpios; i++) {
840
error = gpiochip_fwd_desc_add(fwd, descs[i], i);
841
if (error)
842
return ERR_PTR(error);
843
}
844
845
if (features & FWD_FEATURE_DELAY) {
846
error = gpiochip_fwd_setup_delay_line(fwd);
847
if (error)
848
return ERR_PTR(error);
849
}
850
851
error = gpiochip_fwd_register(fwd, NULL);
852
if (error)
853
return ERR_PTR(error);
854
855
return fwd;
856
}
857
858
/*
859
* Configfs interface
860
*/
861
862
static struct gpio_aggregator *
863
to_gpio_aggregator(struct config_item *item)
864
{
865
struct config_group *group = to_config_group(item);
866
867
return container_of(group, struct gpio_aggregator, group);
868
}
869
870
static struct gpio_aggregator_line *
871
to_gpio_aggregator_line(struct config_item *item)
872
{
873
struct config_group *group = to_config_group(item);
874
875
return container_of(group, struct gpio_aggregator_line, group);
876
}
877
878
static struct fwnode_handle *
879
gpio_aggregator_make_device_sw_node(struct gpio_aggregator *aggr)
880
{
881
struct property_entry properties[2];
882
struct gpio_aggregator_line *line;
883
size_t num_lines;
884
int n = 0;
885
886
memset(properties, 0, sizeof(properties));
887
888
num_lines = gpio_aggregator_count_lines(aggr);
889
if (num_lines == 0)
890
return NULL;
891
892
const char **line_names __free(kfree) = kcalloc(
893
num_lines, sizeof(*line_names), GFP_KERNEL);
894
if (!line_names)
895
return ERR_PTR(-ENOMEM);
896
897
/* The list is always sorted as new elements are inserted in order. */
898
list_for_each_entry(line, &aggr->list_head, entry)
899
line_names[n++] = line->name ?: "";
900
901
properties[0] = PROPERTY_ENTRY_STRING_ARRAY_LEN(
902
"gpio-line-names",
903
line_names, num_lines);
904
905
return fwnode_create_software_node(properties, NULL);
906
}
907
908
static int gpio_aggregator_activate(struct gpio_aggregator *aggr)
909
{
910
struct platform_device_info pdevinfo;
911
struct gpio_aggregator_line *line;
912
struct fwnode_handle *swnode;
913
unsigned int n = 0;
914
int ret = 0;
915
916
if (gpio_aggregator_count_lines(aggr) == 0)
917
return -EINVAL;
918
919
aggr->lookups = kzalloc(struct_size(aggr->lookups, table, 1),
920
GFP_KERNEL);
921
if (!aggr->lookups)
922
return -ENOMEM;
923
924
swnode = gpio_aggregator_make_device_sw_node(aggr);
925
if (IS_ERR(swnode)) {
926
ret = PTR_ERR(swnode);
927
goto err_remove_lookups;
928
}
929
930
memset(&pdevinfo, 0, sizeof(pdevinfo));
931
pdevinfo.name = DRV_NAME;
932
pdevinfo.id = aggr->id;
933
pdevinfo.fwnode = swnode;
934
935
/* The list is always sorted as new elements are inserted in order. */
936
list_for_each_entry(line, &aggr->list_head, entry) {
937
/*
938
* - Either GPIO chip label or line name must be configured
939
* (i.e. line->key must be non-NULL)
940
* - Line directories must be named with sequential numeric
941
* suffixes starting from 0. (i.e. ./line0, ./line1, ...)
942
*/
943
if (!line->key || line->idx != n) {
944
ret = -EINVAL;
945
goto err_remove_swnode;
946
}
947
948
if (line->offset < 0)
949
ret = gpio_aggregator_add_gpio(aggr, line->key,
950
U16_MAX, &n);
951
else
952
ret = gpio_aggregator_add_gpio(aggr, line->key,
953
line->offset, &n);
954
if (ret)
955
goto err_remove_swnode;
956
}
957
958
aggr->lookups->dev_id = kasprintf(GFP_KERNEL, "%s.%d", DRV_NAME, aggr->id);
959
if (!aggr->lookups->dev_id) {
960
ret = -ENOMEM;
961
goto err_remove_swnode;
962
}
963
964
gpiod_add_lookup_table(aggr->lookups);
965
966
ret = dev_sync_probe_register(&aggr->probe_data, &pdevinfo);
967
if (ret)
968
goto err_remove_lookup_table;
969
970
return 0;
971
972
err_remove_lookup_table:
973
kfree(aggr->lookups->dev_id);
974
gpiod_remove_lookup_table(aggr->lookups);
975
err_remove_swnode:
976
fwnode_remove_software_node(swnode);
977
err_remove_lookups:
978
kfree(aggr->lookups);
979
980
return ret;
981
}
982
983
static void gpio_aggregator_deactivate(struct gpio_aggregator *aggr)
984
{
985
dev_sync_probe_unregister(&aggr->probe_data);
986
gpiod_remove_lookup_table(aggr->lookups);
987
kfree(aggr->lookups->dev_id);
988
kfree(aggr->lookups);
989
}
990
991
static void gpio_aggregator_lockup_configfs(struct gpio_aggregator *aggr,
992
bool lock)
993
{
994
struct configfs_subsystem *subsys = aggr->group.cg_subsys;
995
struct gpio_aggregator_line *line;
996
997
/*
998
* The device only needs to depend on leaf lines. This is
999
* sufficient to lock up all the configfs entries that the
1000
* instantiated, alive device depends on.
1001
*/
1002
list_for_each_entry(line, &aggr->list_head, entry) {
1003
if (lock)
1004
configfs_depend_item_unlocked(
1005
subsys, &line->group.cg_item);
1006
else
1007
configfs_undepend_item_unlocked(
1008
&line->group.cg_item);
1009
}
1010
}
1011
1012
static ssize_t
1013
gpio_aggregator_line_key_show(struct config_item *item, char *page)
1014
{
1015
struct gpio_aggregator_line *line = to_gpio_aggregator_line(item);
1016
struct gpio_aggregator *aggr = line->parent;
1017
1018
guard(mutex)(&aggr->lock);
1019
1020
return sysfs_emit(page, "%s\n", line->key ?: "");
1021
}
1022
1023
static ssize_t
1024
gpio_aggregator_line_key_store(struct config_item *item, const char *page,
1025
size_t count)
1026
{
1027
struct gpio_aggregator_line *line = to_gpio_aggregator_line(item);
1028
struct gpio_aggregator *aggr = line->parent;
1029
1030
char *key __free(kfree) = kstrndup(skip_spaces(page), count,
1031
GFP_KERNEL);
1032
if (!key)
1033
return -ENOMEM;
1034
1035
strim(key);
1036
1037
guard(mutex)(&aggr->lock);
1038
1039
if (gpio_aggregator_is_activating(aggr) ||
1040
gpio_aggregator_is_active(aggr))
1041
return -EBUSY;
1042
1043
kfree(line->key);
1044
line->key = no_free_ptr(key);
1045
1046
return count;
1047
}
1048
CONFIGFS_ATTR(gpio_aggregator_line_, key);
1049
1050
static ssize_t
1051
gpio_aggregator_line_name_show(struct config_item *item, char *page)
1052
{
1053
struct gpio_aggregator_line *line = to_gpio_aggregator_line(item);
1054
struct gpio_aggregator *aggr = line->parent;
1055
1056
guard(mutex)(&aggr->lock);
1057
1058
return sysfs_emit(page, "%s\n", line->name ?: "");
1059
}
1060
1061
static ssize_t
1062
gpio_aggregator_line_name_store(struct config_item *item, const char *page,
1063
size_t count)
1064
{
1065
struct gpio_aggregator_line *line = to_gpio_aggregator_line(item);
1066
struct gpio_aggregator *aggr = line->parent;
1067
1068
char *name __free(kfree) = kstrndup(skip_spaces(page), count,
1069
GFP_KERNEL);
1070
if (!name)
1071
return -ENOMEM;
1072
1073
strim(name);
1074
1075
guard(mutex)(&aggr->lock);
1076
1077
if (gpio_aggregator_is_activating(aggr) ||
1078
gpio_aggregator_is_active(aggr))
1079
return -EBUSY;
1080
1081
kfree(line->name);
1082
line->name = no_free_ptr(name);
1083
1084
return count;
1085
}
1086
CONFIGFS_ATTR(gpio_aggregator_line_, name);
1087
1088
static ssize_t
1089
gpio_aggregator_line_offset_show(struct config_item *item, char *page)
1090
{
1091
struct gpio_aggregator_line *line = to_gpio_aggregator_line(item);
1092
struct gpio_aggregator *aggr = line->parent;
1093
1094
guard(mutex)(&aggr->lock);
1095
1096
return sysfs_emit(page, "%d\n", line->offset);
1097
}
1098
1099
static ssize_t
1100
gpio_aggregator_line_offset_store(struct config_item *item, const char *page,
1101
size_t count)
1102
{
1103
struct gpio_aggregator_line *line = to_gpio_aggregator_line(item);
1104
struct gpio_aggregator *aggr = line->parent;
1105
int offset, ret;
1106
1107
ret = kstrtoint(page, 0, &offset);
1108
if (ret)
1109
return ret;
1110
1111
/*
1112
* When offset == -1, 'key' represents a line name to lookup.
1113
* When 0 <= offset < 65535, 'key' represents the label of the chip with
1114
* the 'offset' value representing the line within that chip.
1115
*
1116
* GPIOLIB uses the U16_MAX value to indicate lookup by line name so
1117
* the greatest offset we can accept is (U16_MAX - 1).
1118
*/
1119
if (offset > (U16_MAX - 1) || offset < -1)
1120
return -EINVAL;
1121
1122
guard(mutex)(&aggr->lock);
1123
1124
if (gpio_aggregator_is_activating(aggr) ||
1125
gpio_aggregator_is_active(aggr))
1126
return -EBUSY;
1127
1128
line->offset = offset;
1129
1130
return count;
1131
}
1132
CONFIGFS_ATTR(gpio_aggregator_line_, offset);
1133
1134
static struct configfs_attribute *gpio_aggregator_line_attrs[] = {
1135
&gpio_aggregator_line_attr_key,
1136
&gpio_aggregator_line_attr_name,
1137
&gpio_aggregator_line_attr_offset,
1138
NULL
1139
};
1140
1141
static ssize_t
1142
gpio_aggregator_device_dev_name_show(struct config_item *item, char *page)
1143
{
1144
struct gpio_aggregator *aggr = to_gpio_aggregator(item);
1145
struct platform_device *pdev;
1146
1147
guard(mutex)(&aggr->lock);
1148
1149
pdev = aggr->probe_data.pdev;
1150
if (pdev)
1151
return sysfs_emit(page, "%s\n", dev_name(&pdev->dev));
1152
1153
return sysfs_emit(page, "%s.%d\n", DRV_NAME, aggr->id);
1154
}
1155
CONFIGFS_ATTR_RO(gpio_aggregator_device_, dev_name);
1156
1157
static ssize_t
1158
gpio_aggregator_device_live_show(struct config_item *item, char *page)
1159
{
1160
struct gpio_aggregator *aggr = to_gpio_aggregator(item);
1161
1162
guard(mutex)(&aggr->lock);
1163
1164
return sysfs_emit(page, "%c\n",
1165
gpio_aggregator_is_active(aggr) ? '1' : '0');
1166
}
1167
1168
static ssize_t
1169
gpio_aggregator_device_live_store(struct config_item *item, const char *page,
1170
size_t count)
1171
{
1172
struct gpio_aggregator *aggr = to_gpio_aggregator(item);
1173
int ret = 0;
1174
bool live;
1175
1176
ret = kstrtobool(page, &live);
1177
if (ret)
1178
return ret;
1179
1180
if (!try_module_get(THIS_MODULE))
1181
return -ENOENT;
1182
1183
if (live && !aggr->init_via_sysfs)
1184
gpio_aggregator_lockup_configfs(aggr, true);
1185
1186
scoped_guard(mutex, &aggr->lock) {
1187
if (gpio_aggregator_is_activating(aggr) ||
1188
(live == gpio_aggregator_is_active(aggr)))
1189
ret = -EPERM;
1190
else if (live)
1191
ret = gpio_aggregator_activate(aggr);
1192
else
1193
gpio_aggregator_deactivate(aggr);
1194
}
1195
1196
/*
1197
* Undepend is required only if device disablement (live == 0)
1198
* succeeds or if device enablement (live == 1) fails.
1199
*/
1200
if (live == !!ret && !aggr->init_via_sysfs)
1201
gpio_aggregator_lockup_configfs(aggr, false);
1202
1203
module_put(THIS_MODULE);
1204
1205
return ret ?: count;
1206
}
1207
CONFIGFS_ATTR(gpio_aggregator_device_, live);
1208
1209
static struct configfs_attribute *gpio_aggregator_device_attrs[] = {
1210
&gpio_aggregator_device_attr_dev_name,
1211
&gpio_aggregator_device_attr_live,
1212
NULL
1213
};
1214
1215
static void
1216
gpio_aggregator_line_release(struct config_item *item)
1217
{
1218
struct gpio_aggregator_line *line = to_gpio_aggregator_line(item);
1219
struct gpio_aggregator *aggr = line->parent;
1220
1221
guard(mutex)(&aggr->lock);
1222
1223
gpio_aggregator_line_del(aggr, line);
1224
kfree(line->key);
1225
kfree(line->name);
1226
kfree(line);
1227
}
1228
1229
static struct configfs_item_operations gpio_aggregator_line_item_ops = {
1230
.release = gpio_aggregator_line_release,
1231
};
1232
1233
static const struct config_item_type gpio_aggregator_line_type = {
1234
.ct_item_ops = &gpio_aggregator_line_item_ops,
1235
.ct_attrs = gpio_aggregator_line_attrs,
1236
.ct_owner = THIS_MODULE,
1237
};
1238
1239
static void gpio_aggregator_device_release(struct config_item *item)
1240
{
1241
struct gpio_aggregator *aggr = to_gpio_aggregator(item);
1242
1243
/*
1244
* At this point, aggr is neither active nor activating,
1245
* so calling gpio_aggregator_deactivate() is always unnecessary.
1246
*/
1247
gpio_aggregator_free(aggr);
1248
}
1249
1250
static struct configfs_item_operations gpio_aggregator_device_item_ops = {
1251
.release = gpio_aggregator_device_release,
1252
};
1253
1254
static struct config_group *
1255
gpio_aggregator_device_make_group(struct config_group *group, const char *name)
1256
{
1257
struct gpio_aggregator *aggr = to_gpio_aggregator(&group->cg_item);
1258
struct gpio_aggregator_line *line;
1259
unsigned int idx;
1260
int ret, nchar;
1261
1262
ret = sscanf(name, "line%u%n", &idx, &nchar);
1263
if (ret != 1 || nchar != strlen(name))
1264
return ERR_PTR(-EINVAL);
1265
1266
if (aggr->init_via_sysfs)
1267
/*
1268
* Aggregators created via legacy sysfs interface are exposed as
1269
* default groups, which means rmdir(2) is prohibited for them.
1270
* For simplicity, and to avoid confusion, we also prohibit
1271
* mkdir(2).
1272
*/
1273
return ERR_PTR(-EPERM);
1274
1275
guard(mutex)(&aggr->lock);
1276
1277
if (gpio_aggregator_is_active(aggr))
1278
return ERR_PTR(-EBUSY);
1279
1280
list_for_each_entry(line, &aggr->list_head, entry)
1281
if (line->idx == idx)
1282
return ERR_PTR(-EINVAL);
1283
1284
line = gpio_aggregator_line_alloc(aggr, idx, NULL, -1);
1285
if (IS_ERR(line))
1286
return ERR_CAST(line);
1287
1288
config_group_init_type_name(&line->group, name, &gpio_aggregator_line_type);
1289
1290
gpio_aggregator_line_add(aggr, line);
1291
1292
return &line->group;
1293
}
1294
1295
static struct configfs_group_operations gpio_aggregator_device_group_ops = {
1296
.make_group = gpio_aggregator_device_make_group,
1297
};
1298
1299
static const struct config_item_type gpio_aggregator_device_type = {
1300
.ct_group_ops = &gpio_aggregator_device_group_ops,
1301
.ct_item_ops = &gpio_aggregator_device_item_ops,
1302
.ct_attrs = gpio_aggregator_device_attrs,
1303
.ct_owner = THIS_MODULE,
1304
};
1305
1306
static struct config_group *
1307
gpio_aggregator_make_group(struct config_group *group, const char *name)
1308
{
1309
struct gpio_aggregator *aggr;
1310
int ret;
1311
1312
/*
1313
* "_sysfs" prefix is reserved for auto-generated config group
1314
* for devices create via legacy sysfs interface.
1315
*/
1316
if (strncmp(name, AGGREGATOR_LEGACY_PREFIX,
1317
sizeof(AGGREGATOR_LEGACY_PREFIX) - 1) == 0)
1318
return ERR_PTR(-EINVAL);
1319
1320
/* arg space is unneeded */
1321
ret = gpio_aggregator_alloc(&aggr, 0);
1322
if (ret)
1323
return ERR_PTR(ret);
1324
1325
config_group_init_type_name(&aggr->group, name, &gpio_aggregator_device_type);
1326
dev_sync_probe_init(&aggr->probe_data);
1327
1328
return &aggr->group;
1329
}
1330
1331
static struct configfs_group_operations gpio_aggregator_group_ops = {
1332
.make_group = gpio_aggregator_make_group,
1333
};
1334
1335
static const struct config_item_type gpio_aggregator_type = {
1336
.ct_group_ops = &gpio_aggregator_group_ops,
1337
.ct_owner = THIS_MODULE,
1338
};
1339
1340
static struct configfs_subsystem gpio_aggregator_subsys = {
1341
.su_group = {
1342
.cg_item = {
1343
.ci_namebuf = DRV_NAME,
1344
.ci_type = &gpio_aggregator_type,
1345
},
1346
},
1347
};
1348
1349
/*
1350
* Sysfs interface
1351
*/
1352
static int gpio_aggregator_parse(struct gpio_aggregator *aggr)
1353
{
1354
char *args = skip_spaces(aggr->args);
1355
struct gpio_aggregator_line *line;
1356
char name[CONFIGFS_ITEM_NAME_LEN];
1357
char *key, *offsets, *p;
1358
unsigned int i, n = 0;
1359
int error = 0;
1360
1361
unsigned long *bitmap __free(bitmap) =
1362
bitmap_alloc(AGGREGATOR_MAX_GPIOS, GFP_KERNEL);
1363
if (!bitmap)
1364
return -ENOMEM;
1365
1366
args = next_arg(args, &key, &p);
1367
while (*args) {
1368
args = next_arg(args, &offsets, &p);
1369
1370
p = get_options(offsets, 0, &error);
1371
if (error == 0 || *p) {
1372
/* Named GPIO line */
1373
scnprintf(name, sizeof(name), "line%u", n);
1374
line = gpio_aggregator_line_alloc(aggr, n, key, -1);
1375
if (IS_ERR(line)) {
1376
error = PTR_ERR(line);
1377
goto err;
1378
}
1379
config_group_init_type_name(&line->group, name,
1380
&gpio_aggregator_line_type);
1381
error = configfs_register_group(&aggr->group,
1382
&line->group);
1383
if (error)
1384
goto err;
1385
scoped_guard(mutex, &aggr->lock)
1386
gpio_aggregator_line_add(aggr, line);
1387
1388
error = gpio_aggregator_add_gpio(aggr, key, U16_MAX, &n);
1389
if (error)
1390
goto err;
1391
1392
key = offsets;
1393
continue;
1394
}
1395
1396
/* GPIO chip + offset(s) */
1397
error = bitmap_parselist(offsets, bitmap, AGGREGATOR_MAX_GPIOS);
1398
if (error) {
1399
pr_err("Cannot parse %s: %d\n", offsets, error);
1400
goto err;
1401
}
1402
1403
for_each_set_bit(i, bitmap, AGGREGATOR_MAX_GPIOS) {
1404
scnprintf(name, sizeof(name), "line%u", n);
1405
line = gpio_aggregator_line_alloc(aggr, n, key, i);
1406
if (IS_ERR(line)) {
1407
error = PTR_ERR(line);
1408
goto err;
1409
}
1410
config_group_init_type_name(&line->group, name,
1411
&gpio_aggregator_line_type);
1412
error = configfs_register_group(&aggr->group,
1413
&line->group);
1414
if (error)
1415
goto err;
1416
scoped_guard(mutex, &aggr->lock)
1417
gpio_aggregator_line_add(aggr, line);
1418
1419
error = gpio_aggregator_add_gpio(aggr, key, i, &n);
1420
if (error)
1421
goto err;
1422
}
1423
1424
args = next_arg(args, &key, &p);
1425
}
1426
1427
if (!n) {
1428
pr_err("No GPIOs specified\n");
1429
error = -EINVAL;
1430
goto err;
1431
}
1432
1433
return 0;
1434
1435
err:
1436
gpio_aggregator_free_lines(aggr);
1437
return error;
1438
}
1439
1440
static ssize_t gpio_aggregator_new_device_store(struct device_driver *driver,
1441
const char *buf, size_t count)
1442
{
1443
struct gpio_aggregator_pdev_meta meta = { .init_via_sysfs = true };
1444
char name[CONFIGFS_ITEM_NAME_LEN];
1445
struct gpio_aggregator *aggr;
1446
struct platform_device *pdev;
1447
int res;
1448
1449
if (!try_module_get(THIS_MODULE))
1450
return -ENOENT;
1451
1452
/* kernfs guarantees string termination, so count + 1 is safe */
1453
res = gpio_aggregator_alloc(&aggr, count + 1);
1454
if (res)
1455
goto put_module;
1456
1457
memcpy(aggr->args, buf, count + 1);
1458
1459
aggr->init_via_sysfs = true;
1460
aggr->lookups = kzalloc(struct_size(aggr->lookups, table, 1),
1461
GFP_KERNEL);
1462
if (!aggr->lookups) {
1463
res = -ENOMEM;
1464
goto free_ga;
1465
}
1466
1467
aggr->lookups->dev_id = kasprintf(GFP_KERNEL, "%s.%d", DRV_NAME, aggr->id);
1468
if (!aggr->lookups->dev_id) {
1469
res = -ENOMEM;
1470
goto free_table;
1471
}
1472
1473
scnprintf(name, sizeof(name), "%s.%d", AGGREGATOR_LEGACY_PREFIX, aggr->id);
1474
config_group_init_type_name(&aggr->group, name, &gpio_aggregator_device_type);
1475
1476
/*
1477
* Since the device created by sysfs might be toggled via configfs
1478
* 'live' attribute later, this initialization is needed.
1479
*/
1480
dev_sync_probe_init(&aggr->probe_data);
1481
1482
/* Expose to configfs */
1483
res = configfs_register_group(&gpio_aggregator_subsys.su_group,
1484
&aggr->group);
1485
if (res)
1486
goto free_dev_id;
1487
1488
res = gpio_aggregator_parse(aggr);
1489
if (res)
1490
goto unregister_group;
1491
1492
gpiod_add_lookup_table(aggr->lookups);
1493
1494
pdev = platform_device_register_data(NULL, DRV_NAME, aggr->id, &meta, sizeof(meta));
1495
if (IS_ERR(pdev)) {
1496
res = PTR_ERR(pdev);
1497
goto remove_table;
1498
}
1499
1500
aggr->probe_data.pdev = pdev;
1501
module_put(THIS_MODULE);
1502
return count;
1503
1504
remove_table:
1505
gpiod_remove_lookup_table(aggr->lookups);
1506
unregister_group:
1507
configfs_unregister_group(&aggr->group);
1508
free_dev_id:
1509
kfree(aggr->lookups->dev_id);
1510
free_table:
1511
kfree(aggr->lookups);
1512
free_ga:
1513
gpio_aggregator_free(aggr);
1514
put_module:
1515
module_put(THIS_MODULE);
1516
return res;
1517
}
1518
1519
static struct driver_attribute driver_attr_gpio_aggregator_new_device =
1520
__ATTR(new_device, 0200, NULL, gpio_aggregator_new_device_store);
1521
1522
static void gpio_aggregator_destroy(struct gpio_aggregator *aggr)
1523
{
1524
scoped_guard(mutex, &aggr->lock) {
1525
if (gpio_aggregator_is_activating(aggr) ||
1526
gpio_aggregator_is_active(aggr))
1527
gpio_aggregator_deactivate(aggr);
1528
}
1529
gpio_aggregator_free_lines(aggr);
1530
configfs_unregister_group(&aggr->group);
1531
kfree(aggr);
1532
}
1533
1534
static ssize_t gpio_aggregator_delete_device_store(struct device_driver *driver,
1535
const char *buf, size_t count)
1536
{
1537
struct gpio_aggregator *aggr;
1538
unsigned int id;
1539
int error;
1540
1541
if (!str_has_prefix(buf, DRV_NAME "."))
1542
return -EINVAL;
1543
1544
error = kstrtouint(buf + strlen(DRV_NAME "."), 10, &id);
1545
if (error)
1546
return error;
1547
1548
if (!try_module_get(THIS_MODULE))
1549
return -ENOENT;
1550
1551
mutex_lock(&gpio_aggregator_lock);
1552
aggr = idr_find(&gpio_aggregator_idr, id);
1553
/*
1554
* For simplicity, devices created via configfs cannot be deleted
1555
* via sysfs.
1556
*/
1557
if (aggr && aggr->init_via_sysfs)
1558
idr_remove(&gpio_aggregator_idr, id);
1559
else {
1560
mutex_unlock(&gpio_aggregator_lock);
1561
module_put(THIS_MODULE);
1562
return -ENOENT;
1563
}
1564
mutex_unlock(&gpio_aggregator_lock);
1565
1566
gpio_aggregator_destroy(aggr);
1567
module_put(THIS_MODULE);
1568
return count;
1569
}
1570
1571
static struct driver_attribute driver_attr_gpio_aggregator_delete_device =
1572
__ATTR(delete_device, 0200, NULL, gpio_aggregator_delete_device_store);
1573
1574
static struct attribute *gpio_aggregator_attrs[] = {
1575
&driver_attr_gpio_aggregator_new_device.attr,
1576
&driver_attr_gpio_aggregator_delete_device.attr,
1577
NULL
1578
};
1579
ATTRIBUTE_GROUPS(gpio_aggregator);
1580
1581
/*
1582
* GPIO Aggregator platform device
1583
*/
1584
1585
static int gpio_aggregator_probe(struct platform_device *pdev)
1586
{
1587
struct gpio_aggregator_pdev_meta *meta;
1588
struct device *dev = &pdev->dev;
1589
bool init_via_sysfs = false;
1590
struct gpio_desc **descs;
1591
struct gpiochip_fwd *fwd;
1592
unsigned long features;
1593
int i, n;
1594
1595
n = gpiod_count(dev, NULL);
1596
if (n < 0)
1597
return n;
1598
1599
descs = devm_kmalloc_array(dev, n, sizeof(*descs), GFP_KERNEL);
1600
if (!descs)
1601
return -ENOMEM;
1602
1603
meta = dev_get_platdata(&pdev->dev);
1604
if (meta && meta->init_via_sysfs)
1605
init_via_sysfs = true;
1606
1607
for (i = 0; i < n; i++) {
1608
descs[i] = devm_gpiod_get_index(dev, NULL, i, GPIOD_ASIS);
1609
if (IS_ERR(descs[i])) {
1610
/*
1611
* Deferred probing is not suitable when the aggregator
1612
* is created via configfs. They should just retry later
1613
* whenever they like. For device creation via sysfs,
1614
* error is propagated without overriding for backward
1615
* compatibility. .prevent_deferred_probe is kept unset
1616
* for other cases.
1617
*/
1618
if (!init_via_sysfs && !dev_of_node(dev) &&
1619
descs[i] == ERR_PTR(-EPROBE_DEFER)) {
1620
pr_warn("Deferred probe canceled for creation via configfs.\n");
1621
return -ENODEV;
1622
}
1623
return PTR_ERR(descs[i]);
1624
}
1625
}
1626
1627
features = (uintptr_t)device_get_match_data(dev);
1628
fwd = gpiochip_fwd_create(dev, n, descs, features);
1629
if (IS_ERR(fwd))
1630
return PTR_ERR(fwd);
1631
1632
platform_set_drvdata(pdev, fwd);
1633
devm_kfree(dev, descs);
1634
return 0;
1635
}
1636
1637
static const struct of_device_id gpio_aggregator_dt_ids[] = {
1638
{
1639
.compatible = "gpio-delay",
1640
.data = (void *)FWD_FEATURE_DELAY,
1641
},
1642
/*
1643
* Add GPIO-operated devices controlled from userspace below,
1644
* or use "driver_override" in sysfs.
1645
*/
1646
{}
1647
};
1648
MODULE_DEVICE_TABLE(of, gpio_aggregator_dt_ids);
1649
1650
static struct platform_driver gpio_aggregator_driver = {
1651
.probe = gpio_aggregator_probe,
1652
.driver = {
1653
.name = DRV_NAME,
1654
.groups = gpio_aggregator_groups,
1655
.of_match_table = gpio_aggregator_dt_ids,
1656
},
1657
};
1658
1659
static int __exit gpio_aggregator_idr_remove(int id, void *p, void *data)
1660
{
1661
/*
1662
* There should be no aggregator created via configfs, as their
1663
* presence would prevent module unloading.
1664
*/
1665
gpio_aggregator_destroy(p);
1666
return 0;
1667
}
1668
1669
static void __exit gpio_aggregator_remove_all(void)
1670
{
1671
/*
1672
* Configfs callbacks acquire gpio_aggregator_lock when accessing
1673
* gpio_aggregator_idr, so to prevent lock inversion deadlock, we
1674
* cannot protect idr_for_each invocation here with
1675
* gpio_aggregator_lock, as gpio_aggregator_idr_remove() accesses
1676
* configfs groups. Fortunately, the new_device/delete_device path
1677
* and the module unload path are mutually exclusive, thanks to an
1678
* explicit try_module_get inside of those driver attr handlers.
1679
* Also, when we reach here, no configfs entries present or being
1680
* created. Therefore, no need to protect with gpio_aggregator_lock
1681
* below.
1682
*/
1683
idr_for_each(&gpio_aggregator_idr, gpio_aggregator_idr_remove, NULL);
1684
idr_destroy(&gpio_aggregator_idr);
1685
}
1686
1687
static int __init gpio_aggregator_init(void)
1688
{
1689
int ret = 0;
1690
1691
config_group_init(&gpio_aggregator_subsys.su_group);
1692
mutex_init(&gpio_aggregator_subsys.su_mutex);
1693
ret = configfs_register_subsystem(&gpio_aggregator_subsys);
1694
if (ret) {
1695
pr_err("Failed to register the '%s' configfs subsystem: %d\n",
1696
gpio_aggregator_subsys.su_group.cg_item.ci_namebuf, ret);
1697
mutex_destroy(&gpio_aggregator_subsys.su_mutex);
1698
return ret;
1699
}
1700
1701
/*
1702
* CAVEAT: This must occur after configfs registration. Otherwise,
1703
* a race condition could arise: driver attribute groups might be
1704
* exposed and accessed by users before configfs registration
1705
* completes. new_device_store() does not expect a partially
1706
* initialized configfs state.
1707
*/
1708
ret = platform_driver_register(&gpio_aggregator_driver);
1709
if (ret) {
1710
pr_err("Failed to register the platform driver: %d\n", ret);
1711
mutex_destroy(&gpio_aggregator_subsys.su_mutex);
1712
configfs_unregister_subsystem(&gpio_aggregator_subsys);
1713
}
1714
1715
return ret;
1716
}
1717
module_init(gpio_aggregator_init);
1718
1719
static void __exit gpio_aggregator_exit(void)
1720
{
1721
gpio_aggregator_remove_all();
1722
platform_driver_unregister(&gpio_aggregator_driver);
1723
configfs_unregister_subsystem(&gpio_aggregator_subsys);
1724
}
1725
module_exit(gpio_aggregator_exit);
1726
1727
MODULE_AUTHOR("Geert Uytterhoeven <[email protected]>");
1728
MODULE_DESCRIPTION("GPIO Aggregator");
1729
MODULE_LICENSE("GPL v2");
1730
1731