Loading...
1/* The industrial I/O core
2 *
3 * Copyright (c) 2008 Jonathan Cameron
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
8 *
9 * Based on elements of hwmon and input subsystems.
10 */
11
12#define pr_fmt(fmt) "iio-core: " fmt
13
14#include <linux/kernel.h>
15#include <linux/module.h>
16#include <linux/idr.h>
17#include <linux/kdev_t.h>
18#include <linux/err.h>
19#include <linux/device.h>
20#include <linux/fs.h>
21#include <linux/poll.h>
22#include <linux/sched.h>
23#include <linux/wait.h>
24#include <linux/cdev.h>
25#include <linux/slab.h>
26#include <linux/anon_inodes.h>
27#include <linux/debugfs.h>
28#include <linux/iio/iio.h>
29#include "iio_core.h"
30#include "iio_core_trigger.h"
31#include <linux/iio/sysfs.h>
32#include <linux/iio/events.h>
33#include <linux/iio/buffer.h>
34
35/* IDA to assign each registered device a unique id */
36static DEFINE_IDA(iio_ida);
37
38static dev_t iio_devt;
39
40#define IIO_DEV_MAX 256
41struct bus_type iio_bus_type = {
42 .name = "iio",
43};
44EXPORT_SYMBOL(iio_bus_type);
45
46static struct dentry *iio_debugfs_dentry;
47
48static const char * const iio_direction[] = {
49 [0] = "in",
50 [1] = "out",
51};
52
53static const char * const iio_chan_type_name_spec[] = {
54 [IIO_VOLTAGE] = "voltage",
55 [IIO_CURRENT] = "current",
56 [IIO_POWER] = "power",
57 [IIO_ACCEL] = "accel",
58 [IIO_ANGL_VEL] = "anglvel",
59 [IIO_MAGN] = "magn",
60 [IIO_LIGHT] = "illuminance",
61 [IIO_INTENSITY] = "intensity",
62 [IIO_PROXIMITY] = "proximity",
63 [IIO_TEMP] = "temp",
64 [IIO_INCLI] = "incli",
65 [IIO_ROT] = "rot",
66 [IIO_ANGL] = "angl",
67 [IIO_TIMESTAMP] = "timestamp",
68 [IIO_CAPACITANCE] = "capacitance",
69 [IIO_ALTVOLTAGE] = "altvoltage",
70 [IIO_CCT] = "cct",
71 [IIO_PRESSURE] = "pressure",
72 [IIO_HUMIDITYRELATIVE] = "humidityrelative",
73 [IIO_ACTIVITY] = "activity",
74 [IIO_STEPS] = "steps",
75 [IIO_ENERGY] = "energy",
76 [IIO_DISTANCE] = "distance",
77 [IIO_VELOCITY] = "velocity",
78 [IIO_CONCENTRATION] = "concentration",
79 [IIO_RESISTANCE] = "resistance",
80 [IIO_PH] = "ph",
81};
82
83static const char * const iio_modifier_names[] = {
84 [IIO_MOD_X] = "x",
85 [IIO_MOD_Y] = "y",
86 [IIO_MOD_Z] = "z",
87 [IIO_MOD_X_AND_Y] = "x&y",
88 [IIO_MOD_X_AND_Z] = "x&z",
89 [IIO_MOD_Y_AND_Z] = "y&z",
90 [IIO_MOD_X_AND_Y_AND_Z] = "x&y&z",
91 [IIO_MOD_X_OR_Y] = "x|y",
92 [IIO_MOD_X_OR_Z] = "x|z",
93 [IIO_MOD_Y_OR_Z] = "y|z",
94 [IIO_MOD_X_OR_Y_OR_Z] = "x|y|z",
95 [IIO_MOD_ROOT_SUM_SQUARED_X_Y] = "sqrt(x^2+y^2)",
96 [IIO_MOD_SUM_SQUARED_X_Y_Z] = "x^2+y^2+z^2",
97 [IIO_MOD_LIGHT_BOTH] = "both",
98 [IIO_MOD_LIGHT_IR] = "ir",
99 [IIO_MOD_LIGHT_CLEAR] = "clear",
100 [IIO_MOD_LIGHT_RED] = "red",
101 [IIO_MOD_LIGHT_GREEN] = "green",
102 [IIO_MOD_LIGHT_BLUE] = "blue",
103 [IIO_MOD_QUATERNION] = "quaternion",
104 [IIO_MOD_TEMP_AMBIENT] = "ambient",
105 [IIO_MOD_TEMP_OBJECT] = "object",
106 [IIO_MOD_NORTH_MAGN] = "from_north_magnetic",
107 [IIO_MOD_NORTH_TRUE] = "from_north_true",
108 [IIO_MOD_NORTH_MAGN_TILT_COMP] = "from_north_magnetic_tilt_comp",
109 [IIO_MOD_NORTH_TRUE_TILT_COMP] = "from_north_true_tilt_comp",
110 [IIO_MOD_RUNNING] = "running",
111 [IIO_MOD_JOGGING] = "jogging",
112 [IIO_MOD_WALKING] = "walking",
113 [IIO_MOD_STILL] = "still",
114 [IIO_MOD_ROOT_SUM_SQUARED_X_Y_Z] = "sqrt(x^2+y^2+z^2)",
115 [IIO_MOD_I] = "i",
116 [IIO_MOD_Q] = "q",
117 [IIO_MOD_CO2] = "co2",
118 [IIO_MOD_VOC] = "voc",
119};
120
121/* relies on pairs of these shared then separate */
122static const char * const iio_chan_info_postfix[] = {
123 [IIO_CHAN_INFO_RAW] = "raw",
124 [IIO_CHAN_INFO_PROCESSED] = "input",
125 [IIO_CHAN_INFO_SCALE] = "scale",
126 [IIO_CHAN_INFO_OFFSET] = "offset",
127 [IIO_CHAN_INFO_CALIBSCALE] = "calibscale",
128 [IIO_CHAN_INFO_CALIBBIAS] = "calibbias",
129 [IIO_CHAN_INFO_PEAK] = "peak_raw",
130 [IIO_CHAN_INFO_PEAK_SCALE] = "peak_scale",
131 [IIO_CHAN_INFO_QUADRATURE_CORRECTION_RAW] = "quadrature_correction_raw",
132 [IIO_CHAN_INFO_AVERAGE_RAW] = "mean_raw",
133 [IIO_CHAN_INFO_LOW_PASS_FILTER_3DB_FREQUENCY]
134 = "filter_low_pass_3db_frequency",
135 [IIO_CHAN_INFO_HIGH_PASS_FILTER_3DB_FREQUENCY]
136 = "filter_high_pass_3db_frequency",
137 [IIO_CHAN_INFO_SAMP_FREQ] = "sampling_frequency",
138 [IIO_CHAN_INFO_FREQUENCY] = "frequency",
139 [IIO_CHAN_INFO_PHASE] = "phase",
140 [IIO_CHAN_INFO_HARDWAREGAIN] = "hardwaregain",
141 [IIO_CHAN_INFO_HYSTERESIS] = "hysteresis",
142 [IIO_CHAN_INFO_INT_TIME] = "integration_time",
143 [IIO_CHAN_INFO_ENABLE] = "en",
144 [IIO_CHAN_INFO_CALIBHEIGHT] = "calibheight",
145 [IIO_CHAN_INFO_CALIBWEIGHT] = "calibweight",
146 [IIO_CHAN_INFO_DEBOUNCE_COUNT] = "debounce_count",
147 [IIO_CHAN_INFO_DEBOUNCE_TIME] = "debounce_time",
148 [IIO_CHAN_INFO_CALIBEMISSIVITY] = "calibemissivity",
149 [IIO_CHAN_INFO_OVERSAMPLING_RATIO] = "oversampling_ratio",
150};
151
152/**
153 * iio_find_channel_from_si() - get channel from its scan index
154 * @indio_dev: device
155 * @si: scan index to match
156 */
157const struct iio_chan_spec
158*iio_find_channel_from_si(struct iio_dev *indio_dev, int si)
159{
160 int i;
161
162 for (i = 0; i < indio_dev->num_channels; i++)
163 if (indio_dev->channels[i].scan_index == si)
164 return &indio_dev->channels[i];
165 return NULL;
166}
167
168/* This turns up an awful lot */
169ssize_t iio_read_const_attr(struct device *dev,
170 struct device_attribute *attr,
171 char *buf)
172{
173 return sprintf(buf, "%s\n", to_iio_const_attr(attr)->string);
174}
175EXPORT_SYMBOL(iio_read_const_attr);
176
177static int __init iio_init(void)
178{
179 int ret;
180
181 /* Register sysfs bus */
182 ret = bus_register(&iio_bus_type);
183 if (ret < 0) {
184 pr_err("could not register bus type\n");
185 goto error_nothing;
186 }
187
188 ret = alloc_chrdev_region(&iio_devt, 0, IIO_DEV_MAX, "iio");
189 if (ret < 0) {
190 pr_err("failed to allocate char dev region\n");
191 goto error_unregister_bus_type;
192 }
193
194 iio_debugfs_dentry = debugfs_create_dir("iio", NULL);
195
196 return 0;
197
198error_unregister_bus_type:
199 bus_unregister(&iio_bus_type);
200error_nothing:
201 return ret;
202}
203
204static void __exit iio_exit(void)
205{
206 if (iio_devt)
207 unregister_chrdev_region(iio_devt, IIO_DEV_MAX);
208 bus_unregister(&iio_bus_type);
209 debugfs_remove(iio_debugfs_dentry);
210}
211
212#if defined(CONFIG_DEBUG_FS)
213static ssize_t iio_debugfs_read_reg(struct file *file, char __user *userbuf,
214 size_t count, loff_t *ppos)
215{
216 struct iio_dev *indio_dev = file->private_data;
217 char buf[20];
218 unsigned val = 0;
219 ssize_t len;
220 int ret;
221
222 ret = indio_dev->info->debugfs_reg_access(indio_dev,
223 indio_dev->cached_reg_addr,
224 0, &val);
225 if (ret)
226 dev_err(indio_dev->dev.parent, "%s: read failed\n", __func__);
227
228 len = snprintf(buf, sizeof(buf), "0x%X\n", val);
229
230 return simple_read_from_buffer(userbuf, count, ppos, buf, len);
231}
232
233static ssize_t iio_debugfs_write_reg(struct file *file,
234 const char __user *userbuf, size_t count, loff_t *ppos)
235{
236 struct iio_dev *indio_dev = file->private_data;
237 unsigned reg, val;
238 char buf[80];
239 int ret;
240
241 count = min_t(size_t, count, (sizeof(buf)-1));
242 if (copy_from_user(buf, userbuf, count))
243 return -EFAULT;
244
245 buf[count] = 0;
246
247 ret = sscanf(buf, "%i %i", ®, &val);
248
249 switch (ret) {
250 case 1:
251 indio_dev->cached_reg_addr = reg;
252 break;
253 case 2:
254 indio_dev->cached_reg_addr = reg;
255 ret = indio_dev->info->debugfs_reg_access(indio_dev, reg,
256 val, NULL);
257 if (ret) {
258 dev_err(indio_dev->dev.parent, "%s: write failed\n",
259 __func__);
260 return ret;
261 }
262 break;
263 default:
264 return -EINVAL;
265 }
266
267 return count;
268}
269
270static const struct file_operations iio_debugfs_reg_fops = {
271 .open = simple_open,
272 .read = iio_debugfs_read_reg,
273 .write = iio_debugfs_write_reg,
274};
275
276static void iio_device_unregister_debugfs(struct iio_dev *indio_dev)
277{
278 debugfs_remove_recursive(indio_dev->debugfs_dentry);
279}
280
281static int iio_device_register_debugfs(struct iio_dev *indio_dev)
282{
283 struct dentry *d;
284
285 if (indio_dev->info->debugfs_reg_access == NULL)
286 return 0;
287
288 if (!iio_debugfs_dentry)
289 return 0;
290
291 indio_dev->debugfs_dentry =
292 debugfs_create_dir(dev_name(&indio_dev->dev),
293 iio_debugfs_dentry);
294 if (indio_dev->debugfs_dentry == NULL) {
295 dev_warn(indio_dev->dev.parent,
296 "Failed to create debugfs directory\n");
297 return -EFAULT;
298 }
299
300 d = debugfs_create_file("direct_reg_access", 0644,
301 indio_dev->debugfs_dentry,
302 indio_dev, &iio_debugfs_reg_fops);
303 if (!d) {
304 iio_device_unregister_debugfs(indio_dev);
305 return -ENOMEM;
306 }
307
308 return 0;
309}
310#else
311static int iio_device_register_debugfs(struct iio_dev *indio_dev)
312{
313 return 0;
314}
315
316static void iio_device_unregister_debugfs(struct iio_dev *indio_dev)
317{
318}
319#endif /* CONFIG_DEBUG_FS */
320
321static ssize_t iio_read_channel_ext_info(struct device *dev,
322 struct device_attribute *attr,
323 char *buf)
324{
325 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
326 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
327 const struct iio_chan_spec_ext_info *ext_info;
328
329 ext_info = &this_attr->c->ext_info[this_attr->address];
330
331 return ext_info->read(indio_dev, ext_info->private, this_attr->c, buf);
332}
333
334static ssize_t iio_write_channel_ext_info(struct device *dev,
335 struct device_attribute *attr,
336 const char *buf,
337 size_t len)
338{
339 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
340 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
341 const struct iio_chan_spec_ext_info *ext_info;
342
343 ext_info = &this_attr->c->ext_info[this_attr->address];
344
345 return ext_info->write(indio_dev, ext_info->private,
346 this_attr->c, buf, len);
347}
348
349ssize_t iio_enum_available_read(struct iio_dev *indio_dev,
350 uintptr_t priv, const struct iio_chan_spec *chan, char *buf)
351{
352 const struct iio_enum *e = (const struct iio_enum *)priv;
353 unsigned int i;
354 size_t len = 0;
355
356 if (!e->num_items)
357 return 0;
358
359 for (i = 0; i < e->num_items; ++i)
360 len += scnprintf(buf + len, PAGE_SIZE - len, "%s ", e->items[i]);
361
362 /* replace last space with a newline */
363 buf[len - 1] = '\n';
364
365 return len;
366}
367EXPORT_SYMBOL_GPL(iio_enum_available_read);
368
369ssize_t iio_enum_read(struct iio_dev *indio_dev,
370 uintptr_t priv, const struct iio_chan_spec *chan, char *buf)
371{
372 const struct iio_enum *e = (const struct iio_enum *)priv;
373 int i;
374
375 if (!e->get)
376 return -EINVAL;
377
378 i = e->get(indio_dev, chan);
379 if (i < 0)
380 return i;
381 else if (i >= e->num_items)
382 return -EINVAL;
383
384 return snprintf(buf, PAGE_SIZE, "%s\n", e->items[i]);
385}
386EXPORT_SYMBOL_GPL(iio_enum_read);
387
388ssize_t iio_enum_write(struct iio_dev *indio_dev,
389 uintptr_t priv, const struct iio_chan_spec *chan, const char *buf,
390 size_t len)
391{
392 const struct iio_enum *e = (const struct iio_enum *)priv;
393 unsigned int i;
394 int ret;
395
396 if (!e->set)
397 return -EINVAL;
398
399 for (i = 0; i < e->num_items; i++) {
400 if (sysfs_streq(buf, e->items[i]))
401 break;
402 }
403
404 if (i == e->num_items)
405 return -EINVAL;
406
407 ret = e->set(indio_dev, chan, i);
408 return ret ? ret : len;
409}
410EXPORT_SYMBOL_GPL(iio_enum_write);
411
412/**
413 * iio_format_value() - Formats a IIO value into its string representation
414 * @buf: The buffer to which the formatted value gets written
415 * @type: One of the IIO_VAL_... constants. This decides how the val
416 * and val2 parameters are formatted.
417 * @size: Number of IIO value entries contained in vals
418 * @vals: Pointer to the values, exact meaning depends on the
419 * type parameter.
420 *
421 * Return: 0 by default, a negative number on failure or the
422 * total number of characters written for a type that belongs
423 * to the IIO_VAL_... constant.
424 */
425ssize_t iio_format_value(char *buf, unsigned int type, int size, int *vals)
426{
427 unsigned long long tmp;
428 bool scale_db = false;
429
430 switch (type) {
431 case IIO_VAL_INT:
432 return sprintf(buf, "%d\n", vals[0]);
433 case IIO_VAL_INT_PLUS_MICRO_DB:
434 scale_db = true;
435 case IIO_VAL_INT_PLUS_MICRO:
436 if (vals[1] < 0)
437 return sprintf(buf, "-%d.%06u%s\n", abs(vals[0]),
438 -vals[1], scale_db ? " dB" : "");
439 else
440 return sprintf(buf, "%d.%06u%s\n", vals[0], vals[1],
441 scale_db ? " dB" : "");
442 case IIO_VAL_INT_PLUS_NANO:
443 if (vals[1] < 0)
444 return sprintf(buf, "-%d.%09u\n", abs(vals[0]),
445 -vals[1]);
446 else
447 return sprintf(buf, "%d.%09u\n", vals[0], vals[1]);
448 case IIO_VAL_FRACTIONAL:
449 tmp = div_s64((s64)vals[0] * 1000000000LL, vals[1]);
450 vals[1] = do_div(tmp, 1000000000LL);
451 vals[0] = tmp;
452 return sprintf(buf, "%d.%09u\n", vals[0], vals[1]);
453 case IIO_VAL_FRACTIONAL_LOG2:
454 tmp = (s64)vals[0] * 1000000000LL >> vals[1];
455 vals[1] = do_div(tmp, 1000000000LL);
456 vals[0] = tmp;
457 return sprintf(buf, "%d.%09u\n", vals[0], vals[1]);
458 case IIO_VAL_INT_MULTIPLE:
459 {
460 int i;
461 int len = 0;
462
463 for (i = 0; i < size; ++i)
464 len += snprintf(&buf[len], PAGE_SIZE - len, "%d ",
465 vals[i]);
466 len += snprintf(&buf[len], PAGE_SIZE - len, "\n");
467 return len;
468 }
469 default:
470 return 0;
471 }
472}
473EXPORT_SYMBOL_GPL(iio_format_value);
474
475static ssize_t iio_read_channel_info(struct device *dev,
476 struct device_attribute *attr,
477 char *buf)
478{
479 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
480 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
481 int vals[INDIO_MAX_RAW_ELEMENTS];
482 int ret;
483 int val_len = 2;
484
485 if (indio_dev->info->read_raw_multi)
486 ret = indio_dev->info->read_raw_multi(indio_dev, this_attr->c,
487 INDIO_MAX_RAW_ELEMENTS,
488 vals, &val_len,
489 this_attr->address);
490 else
491 ret = indio_dev->info->read_raw(indio_dev, this_attr->c,
492 &vals[0], &vals[1], this_attr->address);
493
494 if (ret < 0)
495 return ret;
496
497 return iio_format_value(buf, ret, val_len, vals);
498}
499
500/**
501 * iio_str_to_fixpoint() - Parse a fixed-point number from a string
502 * @str: The string to parse
503 * @fract_mult: Multiplier for the first decimal place, should be a power of 10
504 * @integer: The integer part of the number
505 * @fract: The fractional part of the number
506 *
507 * Returns 0 on success, or a negative error code if the string could not be
508 * parsed.
509 */
510int iio_str_to_fixpoint(const char *str, int fract_mult,
511 int *integer, int *fract)
512{
513 int i = 0, f = 0;
514 bool integer_part = true, negative = false;
515
516 if (fract_mult == 0) {
517 *fract = 0;
518
519 return kstrtoint(str, 0, integer);
520 }
521
522 if (str[0] == '-') {
523 negative = true;
524 str++;
525 } else if (str[0] == '+') {
526 str++;
527 }
528
529 while (*str) {
530 if ('0' <= *str && *str <= '9') {
531 if (integer_part) {
532 i = i * 10 + *str - '0';
533 } else {
534 f += fract_mult * (*str - '0');
535 fract_mult /= 10;
536 }
537 } else if (*str == '\n') {
538 if (*(str + 1) == '\0')
539 break;
540 else
541 return -EINVAL;
542 } else if (*str == '.' && integer_part) {
543 integer_part = false;
544 } else {
545 return -EINVAL;
546 }
547 str++;
548 }
549
550 if (negative) {
551 if (i)
552 i = -i;
553 else
554 f = -f;
555 }
556
557 *integer = i;
558 *fract = f;
559
560 return 0;
561}
562EXPORT_SYMBOL_GPL(iio_str_to_fixpoint);
563
564static ssize_t iio_write_channel_info(struct device *dev,
565 struct device_attribute *attr,
566 const char *buf,
567 size_t len)
568{
569 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
570 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
571 int ret, fract_mult = 100000;
572 int integer, fract;
573
574 /* Assumes decimal - precision based on number of digits */
575 if (!indio_dev->info->write_raw)
576 return -EINVAL;
577
578 if (indio_dev->info->write_raw_get_fmt)
579 switch (indio_dev->info->write_raw_get_fmt(indio_dev,
580 this_attr->c, this_attr->address)) {
581 case IIO_VAL_INT:
582 fract_mult = 0;
583 break;
584 case IIO_VAL_INT_PLUS_MICRO:
585 fract_mult = 100000;
586 break;
587 case IIO_VAL_INT_PLUS_NANO:
588 fract_mult = 100000000;
589 break;
590 default:
591 return -EINVAL;
592 }
593
594 ret = iio_str_to_fixpoint(buf, fract_mult, &integer, &fract);
595 if (ret)
596 return ret;
597
598 ret = indio_dev->info->write_raw(indio_dev, this_attr->c,
599 integer, fract, this_attr->address);
600 if (ret)
601 return ret;
602
603 return len;
604}
605
606static
607int __iio_device_attr_init(struct device_attribute *dev_attr,
608 const char *postfix,
609 struct iio_chan_spec const *chan,
610 ssize_t (*readfunc)(struct device *dev,
611 struct device_attribute *attr,
612 char *buf),
613 ssize_t (*writefunc)(struct device *dev,
614 struct device_attribute *attr,
615 const char *buf,
616 size_t len),
617 enum iio_shared_by shared_by)
618{
619 int ret = 0;
620 char *name = NULL;
621 char *full_postfix;
622 sysfs_attr_init(&dev_attr->attr);
623
624 /* Build up postfix of <extend_name>_<modifier>_postfix */
625 if (chan->modified && (shared_by == IIO_SEPARATE)) {
626 if (chan->extend_name)
627 full_postfix = kasprintf(GFP_KERNEL, "%s_%s_%s",
628 iio_modifier_names[chan
629 ->channel2],
630 chan->extend_name,
631 postfix);
632 else
633 full_postfix = kasprintf(GFP_KERNEL, "%s_%s",
634 iio_modifier_names[chan
635 ->channel2],
636 postfix);
637 } else {
638 if (chan->extend_name == NULL || shared_by != IIO_SEPARATE)
639 full_postfix = kstrdup(postfix, GFP_KERNEL);
640 else
641 full_postfix = kasprintf(GFP_KERNEL,
642 "%s_%s",
643 chan->extend_name,
644 postfix);
645 }
646 if (full_postfix == NULL)
647 return -ENOMEM;
648
649 if (chan->differential) { /* Differential can not have modifier */
650 switch (shared_by) {
651 case IIO_SHARED_BY_ALL:
652 name = kasprintf(GFP_KERNEL, "%s", full_postfix);
653 break;
654 case IIO_SHARED_BY_DIR:
655 name = kasprintf(GFP_KERNEL, "%s_%s",
656 iio_direction[chan->output],
657 full_postfix);
658 break;
659 case IIO_SHARED_BY_TYPE:
660 name = kasprintf(GFP_KERNEL, "%s_%s-%s_%s",
661 iio_direction[chan->output],
662 iio_chan_type_name_spec[chan->type],
663 iio_chan_type_name_spec[chan->type],
664 full_postfix);
665 break;
666 case IIO_SEPARATE:
667 if (!chan->indexed) {
668 WARN(1, "Differential channels must be indexed\n");
669 ret = -EINVAL;
670 goto error_free_full_postfix;
671 }
672 name = kasprintf(GFP_KERNEL,
673 "%s_%s%d-%s%d_%s",
674 iio_direction[chan->output],
675 iio_chan_type_name_spec[chan->type],
676 chan->channel,
677 iio_chan_type_name_spec[chan->type],
678 chan->channel2,
679 full_postfix);
680 break;
681 }
682 } else { /* Single ended */
683 switch (shared_by) {
684 case IIO_SHARED_BY_ALL:
685 name = kasprintf(GFP_KERNEL, "%s", full_postfix);
686 break;
687 case IIO_SHARED_BY_DIR:
688 name = kasprintf(GFP_KERNEL, "%s_%s",
689 iio_direction[chan->output],
690 full_postfix);
691 break;
692 case IIO_SHARED_BY_TYPE:
693 name = kasprintf(GFP_KERNEL, "%s_%s_%s",
694 iio_direction[chan->output],
695 iio_chan_type_name_spec[chan->type],
696 full_postfix);
697 break;
698
699 case IIO_SEPARATE:
700 if (chan->indexed)
701 name = kasprintf(GFP_KERNEL, "%s_%s%d_%s",
702 iio_direction[chan->output],
703 iio_chan_type_name_spec[chan->type],
704 chan->channel,
705 full_postfix);
706 else
707 name = kasprintf(GFP_KERNEL, "%s_%s_%s",
708 iio_direction[chan->output],
709 iio_chan_type_name_spec[chan->type],
710 full_postfix);
711 break;
712 }
713 }
714 if (name == NULL) {
715 ret = -ENOMEM;
716 goto error_free_full_postfix;
717 }
718 dev_attr->attr.name = name;
719
720 if (readfunc) {
721 dev_attr->attr.mode |= S_IRUGO;
722 dev_attr->show = readfunc;
723 }
724
725 if (writefunc) {
726 dev_attr->attr.mode |= S_IWUSR;
727 dev_attr->store = writefunc;
728 }
729
730error_free_full_postfix:
731 kfree(full_postfix);
732
733 return ret;
734}
735
736static void __iio_device_attr_deinit(struct device_attribute *dev_attr)
737{
738 kfree(dev_attr->attr.name);
739}
740
741int __iio_add_chan_devattr(const char *postfix,
742 struct iio_chan_spec const *chan,
743 ssize_t (*readfunc)(struct device *dev,
744 struct device_attribute *attr,
745 char *buf),
746 ssize_t (*writefunc)(struct device *dev,
747 struct device_attribute *attr,
748 const char *buf,
749 size_t len),
750 u64 mask,
751 enum iio_shared_by shared_by,
752 struct device *dev,
753 struct list_head *attr_list)
754{
755 int ret;
756 struct iio_dev_attr *iio_attr, *t;
757
758 iio_attr = kzalloc(sizeof(*iio_attr), GFP_KERNEL);
759 if (iio_attr == NULL)
760 return -ENOMEM;
761 ret = __iio_device_attr_init(&iio_attr->dev_attr,
762 postfix, chan,
763 readfunc, writefunc, shared_by);
764 if (ret)
765 goto error_iio_dev_attr_free;
766 iio_attr->c = chan;
767 iio_attr->address = mask;
768 list_for_each_entry(t, attr_list, l)
769 if (strcmp(t->dev_attr.attr.name,
770 iio_attr->dev_attr.attr.name) == 0) {
771 if (shared_by == IIO_SEPARATE)
772 dev_err(dev, "tried to double register : %s\n",
773 t->dev_attr.attr.name);
774 ret = -EBUSY;
775 goto error_device_attr_deinit;
776 }
777 list_add(&iio_attr->l, attr_list);
778
779 return 0;
780
781error_device_attr_deinit:
782 __iio_device_attr_deinit(&iio_attr->dev_attr);
783error_iio_dev_attr_free:
784 kfree(iio_attr);
785 return ret;
786}
787
788static int iio_device_add_info_mask_type(struct iio_dev *indio_dev,
789 struct iio_chan_spec const *chan,
790 enum iio_shared_by shared_by,
791 const long *infomask)
792{
793 int i, ret, attrcount = 0;
794
795 for_each_set_bit(i, infomask, sizeof(infomask)*8) {
796 if (i >= ARRAY_SIZE(iio_chan_info_postfix))
797 return -EINVAL;
798 ret = __iio_add_chan_devattr(iio_chan_info_postfix[i],
799 chan,
800 &iio_read_channel_info,
801 &iio_write_channel_info,
802 i,
803 shared_by,
804 &indio_dev->dev,
805 &indio_dev->channel_attr_list);
806 if ((ret == -EBUSY) && (shared_by != IIO_SEPARATE))
807 continue;
808 else if (ret < 0)
809 return ret;
810 attrcount++;
811 }
812
813 return attrcount;
814}
815
816static int iio_device_add_channel_sysfs(struct iio_dev *indio_dev,
817 struct iio_chan_spec const *chan)
818{
819 int ret, attrcount = 0;
820 const struct iio_chan_spec_ext_info *ext_info;
821
822 if (chan->channel < 0)
823 return 0;
824 ret = iio_device_add_info_mask_type(indio_dev, chan,
825 IIO_SEPARATE,
826 &chan->info_mask_separate);
827 if (ret < 0)
828 return ret;
829 attrcount += ret;
830
831 ret = iio_device_add_info_mask_type(indio_dev, chan,
832 IIO_SHARED_BY_TYPE,
833 &chan->info_mask_shared_by_type);
834 if (ret < 0)
835 return ret;
836 attrcount += ret;
837
838 ret = iio_device_add_info_mask_type(indio_dev, chan,
839 IIO_SHARED_BY_DIR,
840 &chan->info_mask_shared_by_dir);
841 if (ret < 0)
842 return ret;
843 attrcount += ret;
844
845 ret = iio_device_add_info_mask_type(indio_dev, chan,
846 IIO_SHARED_BY_ALL,
847 &chan->info_mask_shared_by_all);
848 if (ret < 0)
849 return ret;
850 attrcount += ret;
851
852 if (chan->ext_info) {
853 unsigned int i = 0;
854 for (ext_info = chan->ext_info; ext_info->name; ext_info++) {
855 ret = __iio_add_chan_devattr(ext_info->name,
856 chan,
857 ext_info->read ?
858 &iio_read_channel_ext_info : NULL,
859 ext_info->write ?
860 &iio_write_channel_ext_info : NULL,
861 i,
862 ext_info->shared,
863 &indio_dev->dev,
864 &indio_dev->channel_attr_list);
865 i++;
866 if (ret == -EBUSY && ext_info->shared)
867 continue;
868
869 if (ret)
870 return ret;
871
872 attrcount++;
873 }
874 }
875
876 return attrcount;
877}
878
879/**
880 * iio_free_chan_devattr_list() - Free a list of IIO device attributes
881 * @attr_list: List of IIO device attributes
882 *
883 * This function frees the memory allocated for each of the IIO device
884 * attributes in the list.
885 */
886void iio_free_chan_devattr_list(struct list_head *attr_list)
887{
888 struct iio_dev_attr *p, *n;
889
890 list_for_each_entry_safe(p, n, attr_list, l) {
891 kfree(p->dev_attr.attr.name);
892 list_del(&p->l);
893 kfree(p);
894 }
895}
896
897static ssize_t iio_show_dev_name(struct device *dev,
898 struct device_attribute *attr,
899 char *buf)
900{
901 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
902 return snprintf(buf, PAGE_SIZE, "%s\n", indio_dev->name);
903}
904
905static DEVICE_ATTR(name, S_IRUGO, iio_show_dev_name, NULL);
906
907static int iio_device_register_sysfs(struct iio_dev *indio_dev)
908{
909 int i, ret = 0, attrcount, attrn, attrcount_orig = 0;
910 struct iio_dev_attr *p;
911 struct attribute **attr;
912
913 /* First count elements in any existing group */
914 if (indio_dev->info->attrs) {
915 attr = indio_dev->info->attrs->attrs;
916 while (*attr++ != NULL)
917 attrcount_orig++;
918 }
919 attrcount = attrcount_orig;
920 /*
921 * New channel registration method - relies on the fact a group does
922 * not need to be initialized if its name is NULL.
923 */
924 if (indio_dev->channels)
925 for (i = 0; i < indio_dev->num_channels; i++) {
926 ret = iio_device_add_channel_sysfs(indio_dev,
927 &indio_dev
928 ->channels[i]);
929 if (ret < 0)
930 goto error_clear_attrs;
931 attrcount += ret;
932 }
933
934 if (indio_dev->name)
935 attrcount++;
936
937 indio_dev->chan_attr_group.attrs = kcalloc(attrcount + 1,
938 sizeof(indio_dev->chan_attr_group.attrs[0]),
939 GFP_KERNEL);
940 if (indio_dev->chan_attr_group.attrs == NULL) {
941 ret = -ENOMEM;
942 goto error_clear_attrs;
943 }
944 /* Copy across original attributes */
945 if (indio_dev->info->attrs)
946 memcpy(indio_dev->chan_attr_group.attrs,
947 indio_dev->info->attrs->attrs,
948 sizeof(indio_dev->chan_attr_group.attrs[0])
949 *attrcount_orig);
950 attrn = attrcount_orig;
951 /* Add all elements from the list. */
952 list_for_each_entry(p, &indio_dev->channel_attr_list, l)
953 indio_dev->chan_attr_group.attrs[attrn++] = &p->dev_attr.attr;
954 if (indio_dev->name)
955 indio_dev->chan_attr_group.attrs[attrn++] = &dev_attr_name.attr;
956
957 indio_dev->groups[indio_dev->groupcounter++] =
958 &indio_dev->chan_attr_group;
959
960 return 0;
961
962error_clear_attrs:
963 iio_free_chan_devattr_list(&indio_dev->channel_attr_list);
964
965 return ret;
966}
967
968static void iio_device_unregister_sysfs(struct iio_dev *indio_dev)
969{
970
971 iio_free_chan_devattr_list(&indio_dev->channel_attr_list);
972 kfree(indio_dev->chan_attr_group.attrs);
973 indio_dev->chan_attr_group.attrs = NULL;
974}
975
976static void iio_dev_release(struct device *device)
977{
978 struct iio_dev *indio_dev = dev_to_iio_dev(device);
979 if (indio_dev->modes & (INDIO_BUFFER_TRIGGERED | INDIO_EVENT_TRIGGERED))
980 iio_device_unregister_trigger_consumer(indio_dev);
981 iio_device_unregister_eventset(indio_dev);
982 iio_device_unregister_sysfs(indio_dev);
983
984 iio_buffer_put(indio_dev->buffer);
985
986 ida_simple_remove(&iio_ida, indio_dev->id);
987 kfree(indio_dev);
988}
989
990struct device_type iio_device_type = {
991 .name = "iio_device",
992 .release = iio_dev_release,
993};
994
995/**
996 * iio_device_alloc() - allocate an iio_dev from a driver
997 * @sizeof_priv: Space to allocate for private structure.
998 **/
999struct iio_dev *iio_device_alloc(int sizeof_priv)
1000{
1001 struct iio_dev *dev;
1002 size_t alloc_size;
1003
1004 alloc_size = sizeof(struct iio_dev);
1005 if (sizeof_priv) {
1006 alloc_size = ALIGN(alloc_size, IIO_ALIGN);
1007 alloc_size += sizeof_priv;
1008 }
1009 /* ensure 32-byte alignment of whole construct ? */
1010 alloc_size += IIO_ALIGN - 1;
1011
1012 dev = kzalloc(alloc_size, GFP_KERNEL);
1013
1014 if (dev) {
1015 dev->dev.groups = dev->groups;
1016 dev->dev.type = &iio_device_type;
1017 dev->dev.bus = &iio_bus_type;
1018 device_initialize(&dev->dev);
1019 dev_set_drvdata(&dev->dev, (void *)dev);
1020 mutex_init(&dev->mlock);
1021 mutex_init(&dev->info_exist_lock);
1022 INIT_LIST_HEAD(&dev->channel_attr_list);
1023
1024 dev->id = ida_simple_get(&iio_ida, 0, 0, GFP_KERNEL);
1025 if (dev->id < 0) {
1026 /* cannot use a dev_err as the name isn't available */
1027 pr_err("failed to get device id\n");
1028 kfree(dev);
1029 return NULL;
1030 }
1031 dev_set_name(&dev->dev, "iio:device%d", dev->id);
1032 INIT_LIST_HEAD(&dev->buffer_list);
1033 }
1034
1035 return dev;
1036}
1037EXPORT_SYMBOL(iio_device_alloc);
1038
1039/**
1040 * iio_device_free() - free an iio_dev from a driver
1041 * @dev: the iio_dev associated with the device
1042 **/
1043void iio_device_free(struct iio_dev *dev)
1044{
1045 if (dev)
1046 put_device(&dev->dev);
1047}
1048EXPORT_SYMBOL(iio_device_free);
1049
1050static void devm_iio_device_release(struct device *dev, void *res)
1051{
1052 iio_device_free(*(struct iio_dev **)res);
1053}
1054
1055static int devm_iio_device_match(struct device *dev, void *res, void *data)
1056{
1057 struct iio_dev **r = res;
1058 if (!r || !*r) {
1059 WARN_ON(!r || !*r);
1060 return 0;
1061 }
1062 return *r == data;
1063}
1064
1065/**
1066 * devm_iio_device_alloc - Resource-managed iio_device_alloc()
1067 * @dev: Device to allocate iio_dev for
1068 * @sizeof_priv: Space to allocate for private structure.
1069 *
1070 * Managed iio_device_alloc. iio_dev allocated with this function is
1071 * automatically freed on driver detach.
1072 *
1073 * If an iio_dev allocated with this function needs to be freed separately,
1074 * devm_iio_device_free() must be used.
1075 *
1076 * RETURNS:
1077 * Pointer to allocated iio_dev on success, NULL on failure.
1078 */
1079struct iio_dev *devm_iio_device_alloc(struct device *dev, int sizeof_priv)
1080{
1081 struct iio_dev **ptr, *iio_dev;
1082
1083 ptr = devres_alloc(devm_iio_device_release, sizeof(*ptr),
1084 GFP_KERNEL);
1085 if (!ptr)
1086 return NULL;
1087
1088 iio_dev = iio_device_alloc(sizeof_priv);
1089 if (iio_dev) {
1090 *ptr = iio_dev;
1091 devres_add(dev, ptr);
1092 } else {
1093 devres_free(ptr);
1094 }
1095
1096 return iio_dev;
1097}
1098EXPORT_SYMBOL_GPL(devm_iio_device_alloc);
1099
1100/**
1101 * devm_iio_device_free - Resource-managed iio_device_free()
1102 * @dev: Device this iio_dev belongs to
1103 * @iio_dev: the iio_dev associated with the device
1104 *
1105 * Free iio_dev allocated with devm_iio_device_alloc().
1106 */
1107void devm_iio_device_free(struct device *dev, struct iio_dev *iio_dev)
1108{
1109 int rc;
1110
1111 rc = devres_release(dev, devm_iio_device_release,
1112 devm_iio_device_match, iio_dev);
1113 WARN_ON(rc);
1114}
1115EXPORT_SYMBOL_GPL(devm_iio_device_free);
1116
1117/**
1118 * iio_chrdev_open() - chrdev file open for buffer access and ioctls
1119 * @inode: Inode structure for identifying the device in the file system
1120 * @filp: File structure for iio device used to keep and later access
1121 * private data
1122 *
1123 * Return: 0 on success or -EBUSY if the device is already opened
1124 **/
1125static int iio_chrdev_open(struct inode *inode, struct file *filp)
1126{
1127 struct iio_dev *indio_dev = container_of(inode->i_cdev,
1128 struct iio_dev, chrdev);
1129
1130 if (test_and_set_bit(IIO_BUSY_BIT_POS, &indio_dev->flags))
1131 return -EBUSY;
1132
1133 iio_device_get(indio_dev);
1134
1135 filp->private_data = indio_dev;
1136
1137 return 0;
1138}
1139
1140/**
1141 * iio_chrdev_release() - chrdev file close buffer access and ioctls
1142 * @inode: Inode structure pointer for the char device
1143 * @filp: File structure pointer for the char device
1144 *
1145 * Return: 0 for successful release
1146 */
1147static int iio_chrdev_release(struct inode *inode, struct file *filp)
1148{
1149 struct iio_dev *indio_dev = container_of(inode->i_cdev,
1150 struct iio_dev, chrdev);
1151 clear_bit(IIO_BUSY_BIT_POS, &indio_dev->flags);
1152 iio_device_put(indio_dev);
1153
1154 return 0;
1155}
1156
1157/* Somewhat of a cross file organization violation - ioctls here are actually
1158 * event related */
1159static long iio_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
1160{
1161 struct iio_dev *indio_dev = filp->private_data;
1162 int __user *ip = (int __user *)arg;
1163 int fd;
1164
1165 if (!indio_dev->info)
1166 return -ENODEV;
1167
1168 if (cmd == IIO_GET_EVENT_FD_IOCTL) {
1169 fd = iio_event_getfd(indio_dev);
1170 if (fd < 0)
1171 return fd;
1172 if (copy_to_user(ip, &fd, sizeof(fd)))
1173 return -EFAULT;
1174 return 0;
1175 }
1176 return -EINVAL;
1177}
1178
1179static const struct file_operations iio_buffer_fileops = {
1180 .read = iio_buffer_read_first_n_outer_addr,
1181 .release = iio_chrdev_release,
1182 .open = iio_chrdev_open,
1183 .poll = iio_buffer_poll_addr,
1184 .owner = THIS_MODULE,
1185 .llseek = noop_llseek,
1186 .unlocked_ioctl = iio_ioctl,
1187 .compat_ioctl = iio_ioctl,
1188};
1189
1190static int iio_check_unique_scan_index(struct iio_dev *indio_dev)
1191{
1192 int i, j;
1193 const struct iio_chan_spec *channels = indio_dev->channels;
1194
1195 if (!(indio_dev->modes & INDIO_ALL_BUFFER_MODES))
1196 return 0;
1197
1198 for (i = 0; i < indio_dev->num_channels - 1; i++) {
1199 if (channels[i].scan_index < 0)
1200 continue;
1201 for (j = i + 1; j < indio_dev->num_channels; j++)
1202 if (channels[i].scan_index == channels[j].scan_index) {
1203 dev_err(&indio_dev->dev,
1204 "Duplicate scan index %d\n",
1205 channels[i].scan_index);
1206 return -EINVAL;
1207 }
1208 }
1209
1210 return 0;
1211}
1212
1213static const struct iio_buffer_setup_ops noop_ring_setup_ops;
1214
1215/**
1216 * iio_device_register() - register a device with the IIO subsystem
1217 * @indio_dev: Device structure filled by the device driver
1218 **/
1219int iio_device_register(struct iio_dev *indio_dev)
1220{
1221 int ret;
1222
1223 /* If the calling driver did not initialize of_node, do it here */
1224 if (!indio_dev->dev.of_node && indio_dev->dev.parent)
1225 indio_dev->dev.of_node = indio_dev->dev.parent->of_node;
1226
1227 ret = iio_check_unique_scan_index(indio_dev);
1228 if (ret < 0)
1229 return ret;
1230
1231 /* configure elements for the chrdev */
1232 indio_dev->dev.devt = MKDEV(MAJOR(iio_devt), indio_dev->id);
1233
1234 ret = iio_device_register_debugfs(indio_dev);
1235 if (ret) {
1236 dev_err(indio_dev->dev.parent,
1237 "Failed to register debugfs interfaces\n");
1238 return ret;
1239 }
1240
1241 ret = iio_buffer_alloc_sysfs_and_mask(indio_dev);
1242 if (ret) {
1243 dev_err(indio_dev->dev.parent,
1244 "Failed to create buffer sysfs interfaces\n");
1245 goto error_unreg_debugfs;
1246 }
1247
1248 ret = iio_device_register_sysfs(indio_dev);
1249 if (ret) {
1250 dev_err(indio_dev->dev.parent,
1251 "Failed to register sysfs interfaces\n");
1252 goto error_buffer_free_sysfs;
1253 }
1254 ret = iio_device_register_eventset(indio_dev);
1255 if (ret) {
1256 dev_err(indio_dev->dev.parent,
1257 "Failed to register event set\n");
1258 goto error_free_sysfs;
1259 }
1260 if (indio_dev->modes & (INDIO_BUFFER_TRIGGERED | INDIO_EVENT_TRIGGERED))
1261 iio_device_register_trigger_consumer(indio_dev);
1262
1263 if ((indio_dev->modes & INDIO_ALL_BUFFER_MODES) &&
1264 indio_dev->setup_ops == NULL)
1265 indio_dev->setup_ops = &noop_ring_setup_ops;
1266
1267 cdev_init(&indio_dev->chrdev, &iio_buffer_fileops);
1268 indio_dev->chrdev.owner = indio_dev->info->driver_module;
1269 indio_dev->chrdev.kobj.parent = &indio_dev->dev.kobj;
1270 ret = cdev_add(&indio_dev->chrdev, indio_dev->dev.devt, 1);
1271 if (ret < 0)
1272 goto error_unreg_eventset;
1273
1274 ret = device_add(&indio_dev->dev);
1275 if (ret < 0)
1276 goto error_cdev_del;
1277
1278 return 0;
1279error_cdev_del:
1280 cdev_del(&indio_dev->chrdev);
1281error_unreg_eventset:
1282 iio_device_unregister_eventset(indio_dev);
1283error_free_sysfs:
1284 iio_device_unregister_sysfs(indio_dev);
1285error_buffer_free_sysfs:
1286 iio_buffer_free_sysfs_and_mask(indio_dev);
1287error_unreg_debugfs:
1288 iio_device_unregister_debugfs(indio_dev);
1289 return ret;
1290}
1291EXPORT_SYMBOL(iio_device_register);
1292
1293/**
1294 * iio_device_unregister() - unregister a device from the IIO subsystem
1295 * @indio_dev: Device structure representing the device.
1296 **/
1297void iio_device_unregister(struct iio_dev *indio_dev)
1298{
1299 mutex_lock(&indio_dev->info_exist_lock);
1300
1301 device_del(&indio_dev->dev);
1302
1303 if (indio_dev->chrdev.dev)
1304 cdev_del(&indio_dev->chrdev);
1305 iio_device_unregister_debugfs(indio_dev);
1306
1307 iio_disable_all_buffers(indio_dev);
1308
1309 indio_dev->info = NULL;
1310
1311 iio_device_wakeup_eventset(indio_dev);
1312 iio_buffer_wakeup_poll(indio_dev);
1313
1314 mutex_unlock(&indio_dev->info_exist_lock);
1315
1316 iio_buffer_free_sysfs_and_mask(indio_dev);
1317}
1318EXPORT_SYMBOL(iio_device_unregister);
1319
1320static void devm_iio_device_unreg(struct device *dev, void *res)
1321{
1322 iio_device_unregister(*(struct iio_dev **)res);
1323}
1324
1325/**
1326 * devm_iio_device_register - Resource-managed iio_device_register()
1327 * @dev: Device to allocate iio_dev for
1328 * @indio_dev: Device structure filled by the device driver
1329 *
1330 * Managed iio_device_register. The IIO device registered with this
1331 * function is automatically unregistered on driver detach. This function
1332 * calls iio_device_register() internally. Refer to that function for more
1333 * information.
1334 *
1335 * If an iio_dev registered with this function needs to be unregistered
1336 * separately, devm_iio_device_unregister() must be used.
1337 *
1338 * RETURNS:
1339 * 0 on success, negative error number on failure.
1340 */
1341int devm_iio_device_register(struct device *dev, struct iio_dev *indio_dev)
1342{
1343 struct iio_dev **ptr;
1344 int ret;
1345
1346 ptr = devres_alloc(devm_iio_device_unreg, sizeof(*ptr), GFP_KERNEL);
1347 if (!ptr)
1348 return -ENOMEM;
1349
1350 *ptr = indio_dev;
1351 ret = iio_device_register(indio_dev);
1352 if (!ret)
1353 devres_add(dev, ptr);
1354 else
1355 devres_free(ptr);
1356
1357 return ret;
1358}
1359EXPORT_SYMBOL_GPL(devm_iio_device_register);
1360
1361/**
1362 * devm_iio_device_unregister - Resource-managed iio_device_unregister()
1363 * @dev: Device this iio_dev belongs to
1364 * @indio_dev: the iio_dev associated with the device
1365 *
1366 * Unregister iio_dev registered with devm_iio_device_register().
1367 */
1368void devm_iio_device_unregister(struct device *dev, struct iio_dev *indio_dev)
1369{
1370 int rc;
1371
1372 rc = devres_release(dev, devm_iio_device_unreg,
1373 devm_iio_device_match, indio_dev);
1374 WARN_ON(rc);
1375}
1376EXPORT_SYMBOL_GPL(devm_iio_device_unregister);
1377
1378subsys_initcall(iio_init);
1379module_exit(iio_exit);
1380
1381MODULE_AUTHOR("Jonathan Cameron <jic23@kernel.org>");
1382MODULE_DESCRIPTION("Industrial I/O core");
1383MODULE_LICENSE("GPL");
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * The industrial I/O core
4 *
5 * Copyright (c) 2008 Jonathan Cameron
6 *
7 * Based on elements of hwmon and input subsystems.
8 */
9
10#define pr_fmt(fmt) "iio-core: " fmt
11
12#include <linux/anon_inodes.h>
13#include <linux/cdev.h>
14#include <linux/debugfs.h>
15#include <linux/device.h>
16#include <linux/err.h>
17#include <linux/fs.h>
18#include <linux/idr.h>
19#include <linux/kdev_t.h>
20#include <linux/kernel.h>
21#include <linux/module.h>
22#include <linux/mutex.h>
23#include <linux/poll.h>
24#include <linux/property.h>
25#include <linux/sched.h>
26#include <linux/slab.h>
27#include <linux/wait.h>
28
29#include <linux/iio/buffer.h>
30#include <linux/iio/buffer_impl.h>
31#include <linux/iio/events.h>
32#include <linux/iio/iio-opaque.h>
33#include <linux/iio/iio.h>
34#include <linux/iio/sysfs.h>
35
36#include "iio_core.h"
37#include "iio_core_trigger.h"
38
39/* IDA to assign each registered device a unique id */
40static DEFINE_IDA(iio_ida);
41
42static dev_t iio_devt;
43
44#define IIO_DEV_MAX 256
45struct bus_type iio_bus_type = {
46 .name = "iio",
47};
48EXPORT_SYMBOL(iio_bus_type);
49
50static struct dentry *iio_debugfs_dentry;
51
52static const char * const iio_direction[] = {
53 [0] = "in",
54 [1] = "out",
55};
56
57static const char * const iio_chan_type_name_spec[] = {
58 [IIO_VOLTAGE] = "voltage",
59 [IIO_CURRENT] = "current",
60 [IIO_POWER] = "power",
61 [IIO_ACCEL] = "accel",
62 [IIO_ANGL_VEL] = "anglvel",
63 [IIO_MAGN] = "magn",
64 [IIO_LIGHT] = "illuminance",
65 [IIO_INTENSITY] = "intensity",
66 [IIO_PROXIMITY] = "proximity",
67 [IIO_TEMP] = "temp",
68 [IIO_INCLI] = "incli",
69 [IIO_ROT] = "rot",
70 [IIO_ANGL] = "angl",
71 [IIO_TIMESTAMP] = "timestamp",
72 [IIO_CAPACITANCE] = "capacitance",
73 [IIO_ALTVOLTAGE] = "altvoltage",
74 [IIO_CCT] = "cct",
75 [IIO_PRESSURE] = "pressure",
76 [IIO_HUMIDITYRELATIVE] = "humidityrelative",
77 [IIO_ACTIVITY] = "activity",
78 [IIO_STEPS] = "steps",
79 [IIO_ENERGY] = "energy",
80 [IIO_DISTANCE] = "distance",
81 [IIO_VELOCITY] = "velocity",
82 [IIO_CONCENTRATION] = "concentration",
83 [IIO_RESISTANCE] = "resistance",
84 [IIO_PH] = "ph",
85 [IIO_UVINDEX] = "uvindex",
86 [IIO_ELECTRICALCONDUCTIVITY] = "electricalconductivity",
87 [IIO_COUNT] = "count",
88 [IIO_INDEX] = "index",
89 [IIO_GRAVITY] = "gravity",
90 [IIO_POSITIONRELATIVE] = "positionrelative",
91 [IIO_PHASE] = "phase",
92 [IIO_MASSCONCENTRATION] = "massconcentration",
93 [IIO_DELTA_ANGL] = "deltaangl",
94 [IIO_DELTA_VELOCITY] = "deltavelocity",
95 [IIO_COLORTEMP] = "colortemp",
96 [IIO_CHROMATICITY] = "chromaticity",
97};
98
99static const char * const iio_modifier_names[] = {
100 [IIO_MOD_X] = "x",
101 [IIO_MOD_Y] = "y",
102 [IIO_MOD_Z] = "z",
103 [IIO_MOD_X_AND_Y] = "x&y",
104 [IIO_MOD_X_AND_Z] = "x&z",
105 [IIO_MOD_Y_AND_Z] = "y&z",
106 [IIO_MOD_X_AND_Y_AND_Z] = "x&y&z",
107 [IIO_MOD_X_OR_Y] = "x|y",
108 [IIO_MOD_X_OR_Z] = "x|z",
109 [IIO_MOD_Y_OR_Z] = "y|z",
110 [IIO_MOD_X_OR_Y_OR_Z] = "x|y|z",
111 [IIO_MOD_ROOT_SUM_SQUARED_X_Y] = "sqrt(x^2+y^2)",
112 [IIO_MOD_SUM_SQUARED_X_Y_Z] = "x^2+y^2+z^2",
113 [IIO_MOD_LIGHT_BOTH] = "both",
114 [IIO_MOD_LIGHT_IR] = "ir",
115 [IIO_MOD_LIGHT_CLEAR] = "clear",
116 [IIO_MOD_LIGHT_RED] = "red",
117 [IIO_MOD_LIGHT_GREEN] = "green",
118 [IIO_MOD_LIGHT_BLUE] = "blue",
119 [IIO_MOD_LIGHT_UV] = "uv",
120 [IIO_MOD_LIGHT_UVA] = "uva",
121 [IIO_MOD_LIGHT_UVB] = "uvb",
122 [IIO_MOD_LIGHT_DUV] = "duv",
123 [IIO_MOD_QUATERNION] = "quaternion",
124 [IIO_MOD_TEMP_AMBIENT] = "ambient",
125 [IIO_MOD_TEMP_OBJECT] = "object",
126 [IIO_MOD_NORTH_MAGN] = "from_north_magnetic",
127 [IIO_MOD_NORTH_TRUE] = "from_north_true",
128 [IIO_MOD_NORTH_MAGN_TILT_COMP] = "from_north_magnetic_tilt_comp",
129 [IIO_MOD_NORTH_TRUE_TILT_COMP] = "from_north_true_tilt_comp",
130 [IIO_MOD_RUNNING] = "running",
131 [IIO_MOD_JOGGING] = "jogging",
132 [IIO_MOD_WALKING] = "walking",
133 [IIO_MOD_STILL] = "still",
134 [IIO_MOD_ROOT_SUM_SQUARED_X_Y_Z] = "sqrt(x^2+y^2+z^2)",
135 [IIO_MOD_I] = "i",
136 [IIO_MOD_Q] = "q",
137 [IIO_MOD_CO2] = "co2",
138 [IIO_MOD_VOC] = "voc",
139 [IIO_MOD_PM1] = "pm1",
140 [IIO_MOD_PM2P5] = "pm2p5",
141 [IIO_MOD_PM4] = "pm4",
142 [IIO_MOD_PM10] = "pm10",
143 [IIO_MOD_ETHANOL] = "ethanol",
144 [IIO_MOD_H2] = "h2",
145 [IIO_MOD_O2] = "o2",
146 [IIO_MOD_LINEAR_X] = "linear_x",
147 [IIO_MOD_LINEAR_Y] = "linear_y",
148 [IIO_MOD_LINEAR_Z] = "linear_z",
149 [IIO_MOD_PITCH] = "pitch",
150 [IIO_MOD_YAW] = "yaw",
151 [IIO_MOD_ROLL] = "roll",
152};
153
154/* relies on pairs of these shared then separate */
155static const char * const iio_chan_info_postfix[] = {
156 [IIO_CHAN_INFO_RAW] = "raw",
157 [IIO_CHAN_INFO_PROCESSED] = "input",
158 [IIO_CHAN_INFO_SCALE] = "scale",
159 [IIO_CHAN_INFO_OFFSET] = "offset",
160 [IIO_CHAN_INFO_CALIBSCALE] = "calibscale",
161 [IIO_CHAN_INFO_CALIBBIAS] = "calibbias",
162 [IIO_CHAN_INFO_PEAK] = "peak_raw",
163 [IIO_CHAN_INFO_PEAK_SCALE] = "peak_scale",
164 [IIO_CHAN_INFO_QUADRATURE_CORRECTION_RAW] = "quadrature_correction_raw",
165 [IIO_CHAN_INFO_AVERAGE_RAW] = "mean_raw",
166 [IIO_CHAN_INFO_LOW_PASS_FILTER_3DB_FREQUENCY]
167 = "filter_low_pass_3db_frequency",
168 [IIO_CHAN_INFO_HIGH_PASS_FILTER_3DB_FREQUENCY]
169 = "filter_high_pass_3db_frequency",
170 [IIO_CHAN_INFO_SAMP_FREQ] = "sampling_frequency",
171 [IIO_CHAN_INFO_FREQUENCY] = "frequency",
172 [IIO_CHAN_INFO_PHASE] = "phase",
173 [IIO_CHAN_INFO_HARDWAREGAIN] = "hardwaregain",
174 [IIO_CHAN_INFO_HYSTERESIS] = "hysteresis",
175 [IIO_CHAN_INFO_HYSTERESIS_RELATIVE] = "hysteresis_relative",
176 [IIO_CHAN_INFO_INT_TIME] = "integration_time",
177 [IIO_CHAN_INFO_ENABLE] = "en",
178 [IIO_CHAN_INFO_CALIBHEIGHT] = "calibheight",
179 [IIO_CHAN_INFO_CALIBWEIGHT] = "calibweight",
180 [IIO_CHAN_INFO_DEBOUNCE_COUNT] = "debounce_count",
181 [IIO_CHAN_INFO_DEBOUNCE_TIME] = "debounce_time",
182 [IIO_CHAN_INFO_CALIBEMISSIVITY] = "calibemissivity",
183 [IIO_CHAN_INFO_OVERSAMPLING_RATIO] = "oversampling_ratio",
184 [IIO_CHAN_INFO_THERMOCOUPLE_TYPE] = "thermocouple_type",
185 [IIO_CHAN_INFO_CALIBAMBIENT] = "calibambient",
186 [IIO_CHAN_INFO_ZEROPOINT] = "zeropoint",
187 [IIO_CHAN_INFO_TROUGH] = "trough_raw",
188};
189/**
190 * iio_device_id() - query the unique ID for the device
191 * @indio_dev: Device structure whose ID is being queried
192 *
193 * The IIO device ID is a unique index used for example for the naming
194 * of the character device /dev/iio\:device[ID].
195 *
196 * Returns: Unique ID for the device.
197 */
198int iio_device_id(struct iio_dev *indio_dev)
199{
200 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
201
202 return iio_dev_opaque->id;
203}
204EXPORT_SYMBOL_GPL(iio_device_id);
205
206/**
207 * iio_buffer_enabled() - helper function to test if the buffer is enabled
208 * @indio_dev: IIO device structure for device
209 *
210 * Returns: True, if the buffer is enabled.
211 */
212bool iio_buffer_enabled(struct iio_dev *indio_dev)
213{
214 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
215
216 return iio_dev_opaque->currentmode &
217 (INDIO_BUFFER_HARDWARE | INDIO_BUFFER_SOFTWARE |
218 INDIO_BUFFER_TRIGGERED);
219}
220EXPORT_SYMBOL_GPL(iio_buffer_enabled);
221
222#if defined(CONFIG_DEBUG_FS)
223/*
224 * There's also a CONFIG_DEBUG_FS guard in include/linux/iio/iio.h for
225 * iio_get_debugfs_dentry() to make it inline if CONFIG_DEBUG_FS is undefined
226 */
227struct dentry *iio_get_debugfs_dentry(struct iio_dev *indio_dev)
228{
229 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
230
231 return iio_dev_opaque->debugfs_dentry;
232}
233EXPORT_SYMBOL_GPL(iio_get_debugfs_dentry);
234#endif
235
236/**
237 * iio_find_channel_from_si() - get channel from its scan index
238 * @indio_dev: device
239 * @si: scan index to match
240 *
241 * Returns:
242 * Constant pointer to iio_chan_spec, if scan index matches, NULL on failure.
243 */
244const struct iio_chan_spec
245*iio_find_channel_from_si(struct iio_dev *indio_dev, int si)
246{
247 int i;
248
249 for (i = 0; i < indio_dev->num_channels; i++)
250 if (indio_dev->channels[i].scan_index == si)
251 return &indio_dev->channels[i];
252 return NULL;
253}
254
255/* This turns up an awful lot */
256ssize_t iio_read_const_attr(struct device *dev,
257 struct device_attribute *attr,
258 char *buf)
259{
260 return sysfs_emit(buf, "%s\n", to_iio_const_attr(attr)->string);
261}
262EXPORT_SYMBOL(iio_read_const_attr);
263
264/**
265 * iio_device_set_clock() - Set current timestamping clock for the device
266 * @indio_dev: IIO device structure containing the device
267 * @clock_id: timestamping clock POSIX identifier to set.
268 *
269 * Returns: 0 on success, or a negative error code.
270 */
271int iio_device_set_clock(struct iio_dev *indio_dev, clockid_t clock_id)
272{
273 int ret;
274 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
275 const struct iio_event_interface *ev_int = iio_dev_opaque->event_interface;
276
277 ret = mutex_lock_interruptible(&iio_dev_opaque->mlock);
278 if (ret)
279 return ret;
280 if ((ev_int && iio_event_enabled(ev_int)) ||
281 iio_buffer_enabled(indio_dev)) {
282 mutex_unlock(&iio_dev_opaque->mlock);
283 return -EBUSY;
284 }
285 iio_dev_opaque->clock_id = clock_id;
286 mutex_unlock(&iio_dev_opaque->mlock);
287
288 return 0;
289}
290EXPORT_SYMBOL(iio_device_set_clock);
291
292/**
293 * iio_device_get_clock() - Retrieve current timestamping clock for the device
294 * @indio_dev: IIO device structure containing the device
295 *
296 * Returns: Clock ID of the current timestamping clock for the device.
297 */
298clockid_t iio_device_get_clock(const struct iio_dev *indio_dev)
299{
300 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
301
302 return iio_dev_opaque->clock_id;
303}
304EXPORT_SYMBOL(iio_device_get_clock);
305
306/**
307 * iio_get_time_ns() - utility function to get a time stamp for events etc
308 * @indio_dev: device
309 *
310 * Returns: Timestamp of the event in nanoseconds.
311 */
312s64 iio_get_time_ns(const struct iio_dev *indio_dev)
313{
314 struct timespec64 tp;
315
316 switch (iio_device_get_clock(indio_dev)) {
317 case CLOCK_REALTIME:
318 return ktime_get_real_ns();
319 case CLOCK_MONOTONIC:
320 return ktime_get_ns();
321 case CLOCK_MONOTONIC_RAW:
322 return ktime_get_raw_ns();
323 case CLOCK_REALTIME_COARSE:
324 return ktime_to_ns(ktime_get_coarse_real());
325 case CLOCK_MONOTONIC_COARSE:
326 ktime_get_coarse_ts64(&tp);
327 return timespec64_to_ns(&tp);
328 case CLOCK_BOOTTIME:
329 return ktime_get_boottime_ns();
330 case CLOCK_TAI:
331 return ktime_get_clocktai_ns();
332 default:
333 BUG();
334 }
335}
336EXPORT_SYMBOL(iio_get_time_ns);
337
338static int __init iio_init(void)
339{
340 int ret;
341
342 /* Register sysfs bus */
343 ret = bus_register(&iio_bus_type);
344 if (ret < 0) {
345 pr_err("could not register bus type\n");
346 goto error_nothing;
347 }
348
349 ret = alloc_chrdev_region(&iio_devt, 0, IIO_DEV_MAX, "iio");
350 if (ret < 0) {
351 pr_err("failed to allocate char dev region\n");
352 goto error_unregister_bus_type;
353 }
354
355 iio_debugfs_dentry = debugfs_create_dir("iio", NULL);
356
357 return 0;
358
359error_unregister_bus_type:
360 bus_unregister(&iio_bus_type);
361error_nothing:
362 return ret;
363}
364
365static void __exit iio_exit(void)
366{
367 if (iio_devt)
368 unregister_chrdev_region(iio_devt, IIO_DEV_MAX);
369 bus_unregister(&iio_bus_type);
370 debugfs_remove(iio_debugfs_dentry);
371}
372
373#if defined(CONFIG_DEBUG_FS)
374static ssize_t iio_debugfs_read_reg(struct file *file, char __user *userbuf,
375 size_t count, loff_t *ppos)
376{
377 struct iio_dev *indio_dev = file->private_data;
378 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
379 unsigned int val = 0;
380 int ret;
381
382 if (*ppos > 0)
383 return simple_read_from_buffer(userbuf, count, ppos,
384 iio_dev_opaque->read_buf,
385 iio_dev_opaque->read_buf_len);
386
387 ret = indio_dev->info->debugfs_reg_access(indio_dev,
388 iio_dev_opaque->cached_reg_addr,
389 0, &val);
390 if (ret) {
391 dev_err(indio_dev->dev.parent, "%s: read failed\n", __func__);
392 return ret;
393 }
394
395 iio_dev_opaque->read_buf_len = snprintf(iio_dev_opaque->read_buf,
396 sizeof(iio_dev_opaque->read_buf),
397 "0x%X\n", val);
398
399 return simple_read_from_buffer(userbuf, count, ppos,
400 iio_dev_opaque->read_buf,
401 iio_dev_opaque->read_buf_len);
402}
403
404static ssize_t iio_debugfs_write_reg(struct file *file,
405 const char __user *userbuf, size_t count, loff_t *ppos)
406{
407 struct iio_dev *indio_dev = file->private_data;
408 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
409 unsigned int reg, val;
410 char buf[80];
411 int ret;
412
413 count = min(count, sizeof(buf) - 1);
414 if (copy_from_user(buf, userbuf, count))
415 return -EFAULT;
416
417 buf[count] = 0;
418
419 ret = sscanf(buf, "%i %i", ®, &val);
420
421 switch (ret) {
422 case 1:
423 iio_dev_opaque->cached_reg_addr = reg;
424 break;
425 case 2:
426 iio_dev_opaque->cached_reg_addr = reg;
427 ret = indio_dev->info->debugfs_reg_access(indio_dev, reg,
428 val, NULL);
429 if (ret) {
430 dev_err(indio_dev->dev.parent, "%s: write failed\n",
431 __func__);
432 return ret;
433 }
434 break;
435 default:
436 return -EINVAL;
437 }
438
439 return count;
440}
441
442static const struct file_operations iio_debugfs_reg_fops = {
443 .open = simple_open,
444 .read = iio_debugfs_read_reg,
445 .write = iio_debugfs_write_reg,
446};
447
448static void iio_device_unregister_debugfs(struct iio_dev *indio_dev)
449{
450 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
451
452 debugfs_remove_recursive(iio_dev_opaque->debugfs_dentry);
453}
454
455static void iio_device_register_debugfs(struct iio_dev *indio_dev)
456{
457 struct iio_dev_opaque *iio_dev_opaque;
458
459 if (indio_dev->info->debugfs_reg_access == NULL)
460 return;
461
462 if (!iio_debugfs_dentry)
463 return;
464
465 iio_dev_opaque = to_iio_dev_opaque(indio_dev);
466
467 iio_dev_opaque->debugfs_dentry =
468 debugfs_create_dir(dev_name(&indio_dev->dev),
469 iio_debugfs_dentry);
470
471 debugfs_create_file("direct_reg_access", 0644,
472 iio_dev_opaque->debugfs_dentry, indio_dev,
473 &iio_debugfs_reg_fops);
474}
475#else
476static void iio_device_register_debugfs(struct iio_dev *indio_dev)
477{
478}
479
480static void iio_device_unregister_debugfs(struct iio_dev *indio_dev)
481{
482}
483#endif /* CONFIG_DEBUG_FS */
484
485static ssize_t iio_read_channel_ext_info(struct device *dev,
486 struct device_attribute *attr,
487 char *buf)
488{
489 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
490 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
491 const struct iio_chan_spec_ext_info *ext_info;
492
493 ext_info = &this_attr->c->ext_info[this_attr->address];
494
495 return ext_info->read(indio_dev, ext_info->private, this_attr->c, buf);
496}
497
498static ssize_t iio_write_channel_ext_info(struct device *dev,
499 struct device_attribute *attr,
500 const char *buf, size_t len)
501{
502 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
503 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
504 const struct iio_chan_spec_ext_info *ext_info;
505
506 ext_info = &this_attr->c->ext_info[this_attr->address];
507
508 return ext_info->write(indio_dev, ext_info->private,
509 this_attr->c, buf, len);
510}
511
512ssize_t iio_enum_available_read(struct iio_dev *indio_dev,
513 uintptr_t priv, const struct iio_chan_spec *chan, char *buf)
514{
515 const struct iio_enum *e = (const struct iio_enum *)priv;
516 unsigned int i;
517 size_t len = 0;
518
519 if (!e->num_items)
520 return 0;
521
522 for (i = 0; i < e->num_items; ++i) {
523 if (!e->items[i])
524 continue;
525 len += sysfs_emit_at(buf, len, "%s ", e->items[i]);
526 }
527
528 /* replace last space with a newline */
529 buf[len - 1] = '\n';
530
531 return len;
532}
533EXPORT_SYMBOL_GPL(iio_enum_available_read);
534
535ssize_t iio_enum_read(struct iio_dev *indio_dev,
536 uintptr_t priv, const struct iio_chan_spec *chan, char *buf)
537{
538 const struct iio_enum *e = (const struct iio_enum *)priv;
539 int i;
540
541 if (!e->get)
542 return -EINVAL;
543
544 i = e->get(indio_dev, chan);
545 if (i < 0)
546 return i;
547 if (i >= e->num_items || !e->items[i])
548 return -EINVAL;
549
550 return sysfs_emit(buf, "%s\n", e->items[i]);
551}
552EXPORT_SYMBOL_GPL(iio_enum_read);
553
554ssize_t iio_enum_write(struct iio_dev *indio_dev,
555 uintptr_t priv, const struct iio_chan_spec *chan, const char *buf,
556 size_t len)
557{
558 const struct iio_enum *e = (const struct iio_enum *)priv;
559 int ret;
560
561 if (!e->set)
562 return -EINVAL;
563
564 ret = __sysfs_match_string(e->items, e->num_items, buf);
565 if (ret < 0)
566 return ret;
567
568 ret = e->set(indio_dev, chan, ret);
569 return ret ? ret : len;
570}
571EXPORT_SYMBOL_GPL(iio_enum_write);
572
573static const struct iio_mount_matrix iio_mount_idmatrix = {
574 .rotation = {
575 "1", "0", "0",
576 "0", "1", "0",
577 "0", "0", "1"
578 }
579};
580
581static int iio_setup_mount_idmatrix(const struct device *dev,
582 struct iio_mount_matrix *matrix)
583{
584 *matrix = iio_mount_idmatrix;
585 dev_info(dev, "mounting matrix not found: using identity...\n");
586 return 0;
587}
588
589ssize_t iio_show_mount_matrix(struct iio_dev *indio_dev, uintptr_t priv,
590 const struct iio_chan_spec *chan, char *buf)
591{
592 const struct iio_mount_matrix *mtx;
593
594 mtx = ((iio_get_mount_matrix_t *)priv)(indio_dev, chan);
595 if (IS_ERR(mtx))
596 return PTR_ERR(mtx);
597
598 if (!mtx)
599 mtx = &iio_mount_idmatrix;
600
601 return sysfs_emit(buf, "%s, %s, %s; %s, %s, %s; %s, %s, %s\n",
602 mtx->rotation[0], mtx->rotation[1], mtx->rotation[2],
603 mtx->rotation[3], mtx->rotation[4], mtx->rotation[5],
604 mtx->rotation[6], mtx->rotation[7], mtx->rotation[8]);
605}
606EXPORT_SYMBOL_GPL(iio_show_mount_matrix);
607
608/**
609 * iio_read_mount_matrix() - retrieve iio device mounting matrix from
610 * device "mount-matrix" property
611 * @dev: device the mounting matrix property is assigned to
612 * @matrix: where to store retrieved matrix
613 *
614 * If device is assigned no mounting matrix property, a default 3x3 identity
615 * matrix will be filled in.
616 *
617 * Returns: 0 if success, or a negative error code on failure.
618 */
619int iio_read_mount_matrix(struct device *dev, struct iio_mount_matrix *matrix)
620{
621 size_t len = ARRAY_SIZE(iio_mount_idmatrix.rotation);
622 int err;
623
624 err = device_property_read_string_array(dev, "mount-matrix", matrix->rotation, len);
625 if (err == len)
626 return 0;
627
628 if (err >= 0)
629 /* Invalid number of matrix entries. */
630 return -EINVAL;
631
632 if (err != -EINVAL)
633 /* Invalid matrix declaration format. */
634 return err;
635
636 /* Matrix was not declared at all: fallback to identity. */
637 return iio_setup_mount_idmatrix(dev, matrix);
638}
639EXPORT_SYMBOL(iio_read_mount_matrix);
640
641static ssize_t __iio_format_value(char *buf, size_t offset, unsigned int type,
642 int size, const int *vals)
643{
644 int tmp0, tmp1;
645 s64 tmp2;
646 bool scale_db = false;
647
648 switch (type) {
649 case IIO_VAL_INT:
650 return sysfs_emit_at(buf, offset, "%d", vals[0]);
651 case IIO_VAL_INT_PLUS_MICRO_DB:
652 scale_db = true;
653 fallthrough;
654 case IIO_VAL_INT_PLUS_MICRO:
655 if (vals[1] < 0)
656 return sysfs_emit_at(buf, offset, "-%d.%06u%s",
657 abs(vals[0]), -vals[1],
658 scale_db ? " dB" : "");
659 else
660 return sysfs_emit_at(buf, offset, "%d.%06u%s", vals[0],
661 vals[1], scale_db ? " dB" : "");
662 case IIO_VAL_INT_PLUS_NANO:
663 if (vals[1] < 0)
664 return sysfs_emit_at(buf, offset, "-%d.%09u",
665 abs(vals[0]), -vals[1]);
666 else
667 return sysfs_emit_at(buf, offset, "%d.%09u", vals[0],
668 vals[1]);
669 case IIO_VAL_FRACTIONAL:
670 tmp2 = div_s64((s64)vals[0] * 1000000000LL, vals[1]);
671 tmp1 = vals[1];
672 tmp0 = (int)div_s64_rem(tmp2, 1000000000, &tmp1);
673 if ((tmp2 < 0) && (tmp0 == 0))
674 return sysfs_emit_at(buf, offset, "-0.%09u", abs(tmp1));
675 else
676 return sysfs_emit_at(buf, offset, "%d.%09u", tmp0,
677 abs(tmp1));
678 case IIO_VAL_FRACTIONAL_LOG2:
679 tmp2 = shift_right((s64)vals[0] * 1000000000LL, vals[1]);
680 tmp0 = (int)div_s64_rem(tmp2, 1000000000LL, &tmp1);
681 if (tmp0 == 0 && tmp2 < 0)
682 return sysfs_emit_at(buf, offset, "-0.%09u", abs(tmp1));
683 else
684 return sysfs_emit_at(buf, offset, "%d.%09u", tmp0,
685 abs(tmp1));
686 case IIO_VAL_INT_MULTIPLE:
687 {
688 int i;
689 int l = 0;
690
691 for (i = 0; i < size; ++i)
692 l += sysfs_emit_at(buf, offset + l, "%d ", vals[i]);
693 return l;
694 }
695 case IIO_VAL_CHAR:
696 return sysfs_emit_at(buf, offset, "%c", (char)vals[0]);
697 case IIO_VAL_INT_64:
698 tmp2 = (s64)((((u64)vals[1]) << 32) | (u32)vals[0]);
699 return sysfs_emit_at(buf, offset, "%lld", tmp2);
700 default:
701 return 0;
702 }
703}
704
705/**
706 * iio_format_value() - Formats a IIO value into its string representation
707 * @buf: The buffer to which the formatted value gets written
708 * which is assumed to be big enough (i.e. PAGE_SIZE).
709 * @type: One of the IIO_VAL_* constants. This decides how the val
710 * and val2 parameters are formatted.
711 * @size: Number of IIO value entries contained in vals
712 * @vals: Pointer to the values, exact meaning depends on the
713 * type parameter.
714 *
715 * Returns:
716 * 0 by default, a negative number on failure or the total number of characters
717 * written for a type that belongs to the IIO_VAL_* constant.
718 */
719ssize_t iio_format_value(char *buf, unsigned int type, int size, int *vals)
720{
721 ssize_t len;
722
723 len = __iio_format_value(buf, 0, type, size, vals);
724 if (len >= PAGE_SIZE - 1)
725 return -EFBIG;
726
727 return len + sysfs_emit_at(buf, len, "\n");
728}
729EXPORT_SYMBOL_GPL(iio_format_value);
730
731static ssize_t iio_read_channel_label(struct device *dev,
732 struct device_attribute *attr,
733 char *buf)
734{
735 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
736 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
737
738 if (indio_dev->info->read_label)
739 return indio_dev->info->read_label(indio_dev, this_attr->c, buf);
740
741 if (this_attr->c->extend_name)
742 return sysfs_emit(buf, "%s\n", this_attr->c->extend_name);
743
744 return -EINVAL;
745}
746
747static ssize_t iio_read_channel_info(struct device *dev,
748 struct device_attribute *attr,
749 char *buf)
750{
751 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
752 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
753 int vals[INDIO_MAX_RAW_ELEMENTS];
754 int ret;
755 int val_len = 2;
756
757 if (indio_dev->info->read_raw_multi)
758 ret = indio_dev->info->read_raw_multi(indio_dev, this_attr->c,
759 INDIO_MAX_RAW_ELEMENTS,
760 vals, &val_len,
761 this_attr->address);
762 else
763 ret = indio_dev->info->read_raw(indio_dev, this_attr->c,
764 &vals[0], &vals[1], this_attr->address);
765
766 if (ret < 0)
767 return ret;
768
769 return iio_format_value(buf, ret, val_len, vals);
770}
771
772static ssize_t iio_format_list(char *buf, const int *vals, int type, int length,
773 const char *prefix, const char *suffix)
774{
775 ssize_t len;
776 int stride;
777 int i;
778
779 switch (type) {
780 case IIO_VAL_INT:
781 stride = 1;
782 break;
783 default:
784 stride = 2;
785 break;
786 }
787
788 len = sysfs_emit(buf, prefix);
789
790 for (i = 0; i <= length - stride; i += stride) {
791 if (i != 0) {
792 len += sysfs_emit_at(buf, len, " ");
793 if (len >= PAGE_SIZE)
794 return -EFBIG;
795 }
796
797 len += __iio_format_value(buf, len, type, stride, &vals[i]);
798 if (len >= PAGE_SIZE)
799 return -EFBIG;
800 }
801
802 len += sysfs_emit_at(buf, len, "%s\n", suffix);
803
804 return len;
805}
806
807static ssize_t iio_format_avail_list(char *buf, const int *vals,
808 int type, int length)
809{
810
811 return iio_format_list(buf, vals, type, length, "", "");
812}
813
814static ssize_t iio_format_avail_range(char *buf, const int *vals, int type)
815{
816 int length;
817
818 /*
819 * length refers to the array size , not the number of elements.
820 * The purpose is to print the range [min , step ,max] so length should
821 * be 3 in case of int, and 6 for other types.
822 */
823 switch (type) {
824 case IIO_VAL_INT:
825 length = 3;
826 break;
827 default:
828 length = 6;
829 break;
830 }
831
832 return iio_format_list(buf, vals, type, length, "[", "]");
833}
834
835static ssize_t iio_read_channel_info_avail(struct device *dev,
836 struct device_attribute *attr,
837 char *buf)
838{
839 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
840 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
841 const int *vals;
842 int ret;
843 int length;
844 int type;
845
846 ret = indio_dev->info->read_avail(indio_dev, this_attr->c,
847 &vals, &type, &length,
848 this_attr->address);
849
850 if (ret < 0)
851 return ret;
852 switch (ret) {
853 case IIO_AVAIL_LIST:
854 return iio_format_avail_list(buf, vals, type, length);
855 case IIO_AVAIL_RANGE:
856 return iio_format_avail_range(buf, vals, type);
857 default:
858 return -EINVAL;
859 }
860}
861
862/**
863 * __iio_str_to_fixpoint() - Parse a fixed-point number from a string
864 * @str: The string to parse
865 * @fract_mult: Multiplier for the first decimal place, should be a power of 10
866 * @integer: The integer part of the number
867 * @fract: The fractional part of the number
868 * @scale_db: True if this should parse as dB
869 *
870 * Returns:
871 * 0 on success, or a negative error code if the string could not be parsed.
872 */
873static int __iio_str_to_fixpoint(const char *str, int fract_mult,
874 int *integer, int *fract, bool scale_db)
875{
876 int i = 0, f = 0;
877 bool integer_part = true, negative = false;
878
879 if (fract_mult == 0) {
880 *fract = 0;
881
882 return kstrtoint(str, 0, integer);
883 }
884
885 if (str[0] == '-') {
886 negative = true;
887 str++;
888 } else if (str[0] == '+') {
889 str++;
890 }
891
892 while (*str) {
893 if ('0' <= *str && *str <= '9') {
894 if (integer_part) {
895 i = i * 10 + *str - '0';
896 } else {
897 f += fract_mult * (*str - '0');
898 fract_mult /= 10;
899 }
900 } else if (*str == '\n') {
901 if (*(str + 1) == '\0')
902 break;
903 return -EINVAL;
904 } else if (!strncmp(str, " dB", sizeof(" dB") - 1) && scale_db) {
905 /* Ignore the dB suffix */
906 str += sizeof(" dB") - 1;
907 continue;
908 } else if (!strncmp(str, "dB", sizeof("dB") - 1) && scale_db) {
909 /* Ignore the dB suffix */
910 str += sizeof("dB") - 1;
911 continue;
912 } else if (*str == '.' && integer_part) {
913 integer_part = false;
914 } else {
915 return -EINVAL;
916 }
917 str++;
918 }
919
920 if (negative) {
921 if (i)
922 i = -i;
923 else
924 f = -f;
925 }
926
927 *integer = i;
928 *fract = f;
929
930 return 0;
931}
932
933/**
934 * iio_str_to_fixpoint() - Parse a fixed-point number from a string
935 * @str: The string to parse
936 * @fract_mult: Multiplier for the first decimal place, should be a power of 10
937 * @integer: The integer part of the number
938 * @fract: The fractional part of the number
939 *
940 * Returns:
941 * 0 on success, or a negative error code if the string could not be parsed.
942 */
943int iio_str_to_fixpoint(const char *str, int fract_mult,
944 int *integer, int *fract)
945{
946 return __iio_str_to_fixpoint(str, fract_mult, integer, fract, false);
947}
948EXPORT_SYMBOL_GPL(iio_str_to_fixpoint);
949
950static ssize_t iio_write_channel_info(struct device *dev,
951 struct device_attribute *attr,
952 const char *buf,
953 size_t len)
954{
955 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
956 struct iio_dev_attr *this_attr = to_iio_dev_attr(attr);
957 int ret, fract_mult = 100000;
958 int integer, fract = 0;
959 bool is_char = false;
960 bool scale_db = false;
961
962 /* Assumes decimal - precision based on number of digits */
963 if (!indio_dev->info->write_raw)
964 return -EINVAL;
965
966 if (indio_dev->info->write_raw_get_fmt)
967 switch (indio_dev->info->write_raw_get_fmt(indio_dev,
968 this_attr->c, this_attr->address)) {
969 case IIO_VAL_INT:
970 fract_mult = 0;
971 break;
972 case IIO_VAL_INT_PLUS_MICRO_DB:
973 scale_db = true;
974 fallthrough;
975 case IIO_VAL_INT_PLUS_MICRO:
976 fract_mult = 100000;
977 break;
978 case IIO_VAL_INT_PLUS_NANO:
979 fract_mult = 100000000;
980 break;
981 case IIO_VAL_CHAR:
982 is_char = true;
983 break;
984 default:
985 return -EINVAL;
986 }
987
988 if (is_char) {
989 char ch;
990
991 if (sscanf(buf, "%c", &ch) != 1)
992 return -EINVAL;
993 integer = ch;
994 } else {
995 ret = __iio_str_to_fixpoint(buf, fract_mult, &integer, &fract,
996 scale_db);
997 if (ret)
998 return ret;
999 }
1000
1001 ret = indio_dev->info->write_raw(indio_dev, this_attr->c,
1002 integer, fract, this_attr->address);
1003 if (ret)
1004 return ret;
1005
1006 return len;
1007}
1008
1009static
1010int __iio_device_attr_init(struct device_attribute *dev_attr,
1011 const char *postfix,
1012 struct iio_chan_spec const *chan,
1013 ssize_t (*readfunc)(struct device *dev,
1014 struct device_attribute *attr,
1015 char *buf),
1016 ssize_t (*writefunc)(struct device *dev,
1017 struct device_attribute *attr,
1018 const char *buf,
1019 size_t len),
1020 enum iio_shared_by shared_by)
1021{
1022 int ret = 0;
1023 char *name = NULL;
1024 char *full_postfix;
1025
1026 sysfs_attr_init(&dev_attr->attr);
1027
1028 /* Build up postfix of <extend_name>_<modifier>_postfix */
1029 if (chan->modified && (shared_by == IIO_SEPARATE)) {
1030 if (chan->extend_name)
1031 full_postfix = kasprintf(GFP_KERNEL, "%s_%s_%s",
1032 iio_modifier_names[chan->channel2],
1033 chan->extend_name,
1034 postfix);
1035 else
1036 full_postfix = kasprintf(GFP_KERNEL, "%s_%s",
1037 iio_modifier_names[chan->channel2],
1038 postfix);
1039 } else {
1040 if (chan->extend_name == NULL || shared_by != IIO_SEPARATE)
1041 full_postfix = kstrdup(postfix, GFP_KERNEL);
1042 else
1043 full_postfix = kasprintf(GFP_KERNEL,
1044 "%s_%s",
1045 chan->extend_name,
1046 postfix);
1047 }
1048 if (full_postfix == NULL)
1049 return -ENOMEM;
1050
1051 if (chan->differential) { /* Differential can not have modifier */
1052 switch (shared_by) {
1053 case IIO_SHARED_BY_ALL:
1054 name = kasprintf(GFP_KERNEL, "%s", full_postfix);
1055 break;
1056 case IIO_SHARED_BY_DIR:
1057 name = kasprintf(GFP_KERNEL, "%s_%s",
1058 iio_direction[chan->output],
1059 full_postfix);
1060 break;
1061 case IIO_SHARED_BY_TYPE:
1062 name = kasprintf(GFP_KERNEL, "%s_%s-%s_%s",
1063 iio_direction[chan->output],
1064 iio_chan_type_name_spec[chan->type],
1065 iio_chan_type_name_spec[chan->type],
1066 full_postfix);
1067 break;
1068 case IIO_SEPARATE:
1069 if (!chan->indexed) {
1070 WARN(1, "Differential channels must be indexed\n");
1071 ret = -EINVAL;
1072 goto error_free_full_postfix;
1073 }
1074 name = kasprintf(GFP_KERNEL,
1075 "%s_%s%d-%s%d_%s",
1076 iio_direction[chan->output],
1077 iio_chan_type_name_spec[chan->type],
1078 chan->channel,
1079 iio_chan_type_name_spec[chan->type],
1080 chan->channel2,
1081 full_postfix);
1082 break;
1083 }
1084 } else { /* Single ended */
1085 switch (shared_by) {
1086 case IIO_SHARED_BY_ALL:
1087 name = kasprintf(GFP_KERNEL, "%s", full_postfix);
1088 break;
1089 case IIO_SHARED_BY_DIR:
1090 name = kasprintf(GFP_KERNEL, "%s_%s",
1091 iio_direction[chan->output],
1092 full_postfix);
1093 break;
1094 case IIO_SHARED_BY_TYPE:
1095 name = kasprintf(GFP_KERNEL, "%s_%s_%s",
1096 iio_direction[chan->output],
1097 iio_chan_type_name_spec[chan->type],
1098 full_postfix);
1099 break;
1100
1101 case IIO_SEPARATE:
1102 if (chan->indexed)
1103 name = kasprintf(GFP_KERNEL, "%s_%s%d_%s",
1104 iio_direction[chan->output],
1105 iio_chan_type_name_spec[chan->type],
1106 chan->channel,
1107 full_postfix);
1108 else
1109 name = kasprintf(GFP_KERNEL, "%s_%s_%s",
1110 iio_direction[chan->output],
1111 iio_chan_type_name_spec[chan->type],
1112 full_postfix);
1113 break;
1114 }
1115 }
1116 if (name == NULL) {
1117 ret = -ENOMEM;
1118 goto error_free_full_postfix;
1119 }
1120 dev_attr->attr.name = name;
1121
1122 if (readfunc) {
1123 dev_attr->attr.mode |= 0444;
1124 dev_attr->show = readfunc;
1125 }
1126
1127 if (writefunc) {
1128 dev_attr->attr.mode |= 0200;
1129 dev_attr->store = writefunc;
1130 }
1131
1132error_free_full_postfix:
1133 kfree(full_postfix);
1134
1135 return ret;
1136}
1137
1138static void __iio_device_attr_deinit(struct device_attribute *dev_attr)
1139{
1140 kfree(dev_attr->attr.name);
1141}
1142
1143int __iio_add_chan_devattr(const char *postfix,
1144 struct iio_chan_spec const *chan,
1145 ssize_t (*readfunc)(struct device *dev,
1146 struct device_attribute *attr,
1147 char *buf),
1148 ssize_t (*writefunc)(struct device *dev,
1149 struct device_attribute *attr,
1150 const char *buf,
1151 size_t len),
1152 u64 mask,
1153 enum iio_shared_by shared_by,
1154 struct device *dev,
1155 struct iio_buffer *buffer,
1156 struct list_head *attr_list)
1157{
1158 int ret;
1159 struct iio_dev_attr *iio_attr, *t;
1160
1161 iio_attr = kzalloc(sizeof(*iio_attr), GFP_KERNEL);
1162 if (iio_attr == NULL)
1163 return -ENOMEM;
1164 ret = __iio_device_attr_init(&iio_attr->dev_attr,
1165 postfix, chan,
1166 readfunc, writefunc, shared_by);
1167 if (ret)
1168 goto error_iio_dev_attr_free;
1169 iio_attr->c = chan;
1170 iio_attr->address = mask;
1171 iio_attr->buffer = buffer;
1172 list_for_each_entry(t, attr_list, l)
1173 if (strcmp(t->dev_attr.attr.name,
1174 iio_attr->dev_attr.attr.name) == 0) {
1175 if (shared_by == IIO_SEPARATE)
1176 dev_err(dev, "tried to double register : %s\n",
1177 t->dev_attr.attr.name);
1178 ret = -EBUSY;
1179 goto error_device_attr_deinit;
1180 }
1181 list_add(&iio_attr->l, attr_list);
1182
1183 return 0;
1184
1185error_device_attr_deinit:
1186 __iio_device_attr_deinit(&iio_attr->dev_attr);
1187error_iio_dev_attr_free:
1188 kfree(iio_attr);
1189 return ret;
1190}
1191
1192static int iio_device_add_channel_label(struct iio_dev *indio_dev,
1193 struct iio_chan_spec const *chan)
1194{
1195 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1196 int ret;
1197
1198 if (!indio_dev->info->read_label && !chan->extend_name)
1199 return 0;
1200
1201 ret = __iio_add_chan_devattr("label",
1202 chan,
1203 &iio_read_channel_label,
1204 NULL,
1205 0,
1206 IIO_SEPARATE,
1207 &indio_dev->dev,
1208 NULL,
1209 &iio_dev_opaque->channel_attr_list);
1210 if (ret < 0)
1211 return ret;
1212
1213 return 1;
1214}
1215
1216static int iio_device_add_info_mask_type(struct iio_dev *indio_dev,
1217 struct iio_chan_spec const *chan,
1218 enum iio_shared_by shared_by,
1219 const long *infomask)
1220{
1221 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1222 int i, ret, attrcount = 0;
1223
1224 for_each_set_bit(i, infomask, sizeof(*infomask)*8) {
1225 if (i >= ARRAY_SIZE(iio_chan_info_postfix))
1226 return -EINVAL;
1227 ret = __iio_add_chan_devattr(iio_chan_info_postfix[i],
1228 chan,
1229 &iio_read_channel_info,
1230 &iio_write_channel_info,
1231 i,
1232 shared_by,
1233 &indio_dev->dev,
1234 NULL,
1235 &iio_dev_opaque->channel_attr_list);
1236 if ((ret == -EBUSY) && (shared_by != IIO_SEPARATE))
1237 continue;
1238 if (ret < 0)
1239 return ret;
1240 attrcount++;
1241 }
1242
1243 return attrcount;
1244}
1245
1246static int iio_device_add_info_mask_type_avail(struct iio_dev *indio_dev,
1247 struct iio_chan_spec const *chan,
1248 enum iio_shared_by shared_by,
1249 const long *infomask)
1250{
1251 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1252 int i, ret, attrcount = 0;
1253 char *avail_postfix;
1254
1255 for_each_set_bit(i, infomask, sizeof(*infomask) * 8) {
1256 if (i >= ARRAY_SIZE(iio_chan_info_postfix))
1257 return -EINVAL;
1258 avail_postfix = kasprintf(GFP_KERNEL,
1259 "%s_available",
1260 iio_chan_info_postfix[i]);
1261 if (!avail_postfix)
1262 return -ENOMEM;
1263
1264 ret = __iio_add_chan_devattr(avail_postfix,
1265 chan,
1266 &iio_read_channel_info_avail,
1267 NULL,
1268 i,
1269 shared_by,
1270 &indio_dev->dev,
1271 NULL,
1272 &iio_dev_opaque->channel_attr_list);
1273 kfree(avail_postfix);
1274 if ((ret == -EBUSY) && (shared_by != IIO_SEPARATE))
1275 continue;
1276 if (ret < 0)
1277 return ret;
1278 attrcount++;
1279 }
1280
1281 return attrcount;
1282}
1283
1284static int iio_device_add_channel_sysfs(struct iio_dev *indio_dev,
1285 struct iio_chan_spec const *chan)
1286{
1287 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1288 int ret, attrcount = 0;
1289 const struct iio_chan_spec_ext_info *ext_info;
1290
1291 if (chan->channel < 0)
1292 return 0;
1293 ret = iio_device_add_info_mask_type(indio_dev, chan,
1294 IIO_SEPARATE,
1295 &chan->info_mask_separate);
1296 if (ret < 0)
1297 return ret;
1298 attrcount += ret;
1299
1300 ret = iio_device_add_info_mask_type_avail(indio_dev, chan,
1301 IIO_SEPARATE,
1302 &chan->info_mask_separate_available);
1303 if (ret < 0)
1304 return ret;
1305 attrcount += ret;
1306
1307 ret = iio_device_add_info_mask_type(indio_dev, chan,
1308 IIO_SHARED_BY_TYPE,
1309 &chan->info_mask_shared_by_type);
1310 if (ret < 0)
1311 return ret;
1312 attrcount += ret;
1313
1314 ret = iio_device_add_info_mask_type_avail(indio_dev, chan,
1315 IIO_SHARED_BY_TYPE,
1316 &chan->info_mask_shared_by_type_available);
1317 if (ret < 0)
1318 return ret;
1319 attrcount += ret;
1320
1321 ret = iio_device_add_info_mask_type(indio_dev, chan,
1322 IIO_SHARED_BY_DIR,
1323 &chan->info_mask_shared_by_dir);
1324 if (ret < 0)
1325 return ret;
1326 attrcount += ret;
1327
1328 ret = iio_device_add_info_mask_type_avail(indio_dev, chan,
1329 IIO_SHARED_BY_DIR,
1330 &chan->info_mask_shared_by_dir_available);
1331 if (ret < 0)
1332 return ret;
1333 attrcount += ret;
1334
1335 ret = iio_device_add_info_mask_type(indio_dev, chan,
1336 IIO_SHARED_BY_ALL,
1337 &chan->info_mask_shared_by_all);
1338 if (ret < 0)
1339 return ret;
1340 attrcount += ret;
1341
1342 ret = iio_device_add_info_mask_type_avail(indio_dev, chan,
1343 IIO_SHARED_BY_ALL,
1344 &chan->info_mask_shared_by_all_available);
1345 if (ret < 0)
1346 return ret;
1347 attrcount += ret;
1348
1349 ret = iio_device_add_channel_label(indio_dev, chan);
1350 if (ret < 0)
1351 return ret;
1352 attrcount += ret;
1353
1354 if (chan->ext_info) {
1355 unsigned int i = 0;
1356
1357 for (ext_info = chan->ext_info; ext_info->name; ext_info++) {
1358 ret = __iio_add_chan_devattr(ext_info->name,
1359 chan,
1360 ext_info->read ?
1361 &iio_read_channel_ext_info : NULL,
1362 ext_info->write ?
1363 &iio_write_channel_ext_info : NULL,
1364 i,
1365 ext_info->shared,
1366 &indio_dev->dev,
1367 NULL,
1368 &iio_dev_opaque->channel_attr_list);
1369 i++;
1370 if (ret == -EBUSY && ext_info->shared)
1371 continue;
1372
1373 if (ret)
1374 return ret;
1375
1376 attrcount++;
1377 }
1378 }
1379
1380 return attrcount;
1381}
1382
1383/**
1384 * iio_free_chan_devattr_list() - Free a list of IIO device attributes
1385 * @attr_list: List of IIO device attributes
1386 *
1387 * This function frees the memory allocated for each of the IIO device
1388 * attributes in the list.
1389 */
1390void iio_free_chan_devattr_list(struct list_head *attr_list)
1391{
1392 struct iio_dev_attr *p, *n;
1393
1394 list_for_each_entry_safe(p, n, attr_list, l) {
1395 kfree_const(p->dev_attr.attr.name);
1396 list_del(&p->l);
1397 kfree(p);
1398 }
1399}
1400
1401static ssize_t name_show(struct device *dev, struct device_attribute *attr,
1402 char *buf)
1403{
1404 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
1405
1406 return sysfs_emit(buf, "%s\n", indio_dev->name);
1407}
1408
1409static DEVICE_ATTR_RO(name);
1410
1411static ssize_t label_show(struct device *dev, struct device_attribute *attr,
1412 char *buf)
1413{
1414 struct iio_dev *indio_dev = dev_to_iio_dev(dev);
1415
1416 return sysfs_emit(buf, "%s\n", indio_dev->label);
1417}
1418
1419static DEVICE_ATTR_RO(label);
1420
1421static const char * const clock_names[] = {
1422 [CLOCK_REALTIME] = "realtime",
1423 [CLOCK_MONOTONIC] = "monotonic",
1424 [CLOCK_PROCESS_CPUTIME_ID] = "process_cputime_id",
1425 [CLOCK_THREAD_CPUTIME_ID] = "thread_cputime_id",
1426 [CLOCK_MONOTONIC_RAW] = "monotonic_raw",
1427 [CLOCK_REALTIME_COARSE] = "realtime_coarse",
1428 [CLOCK_MONOTONIC_COARSE] = "monotonic_coarse",
1429 [CLOCK_BOOTTIME] = "boottime",
1430 [CLOCK_REALTIME_ALARM] = "realtime_alarm",
1431 [CLOCK_BOOTTIME_ALARM] = "boottime_alarm",
1432 [CLOCK_SGI_CYCLE] = "sgi_cycle",
1433 [CLOCK_TAI] = "tai",
1434};
1435
1436static ssize_t current_timestamp_clock_show(struct device *dev,
1437 struct device_attribute *attr,
1438 char *buf)
1439{
1440 const struct iio_dev *indio_dev = dev_to_iio_dev(dev);
1441 const clockid_t clk = iio_device_get_clock(indio_dev);
1442
1443 switch (clk) {
1444 case CLOCK_REALTIME:
1445 case CLOCK_MONOTONIC:
1446 case CLOCK_MONOTONIC_RAW:
1447 case CLOCK_REALTIME_COARSE:
1448 case CLOCK_MONOTONIC_COARSE:
1449 case CLOCK_BOOTTIME:
1450 case CLOCK_TAI:
1451 break;
1452 default:
1453 BUG();
1454 }
1455
1456 return sysfs_emit(buf, "%s\n", clock_names[clk]);
1457}
1458
1459static ssize_t current_timestamp_clock_store(struct device *dev,
1460 struct device_attribute *attr,
1461 const char *buf, size_t len)
1462{
1463 clockid_t clk;
1464 int ret;
1465
1466 ret = sysfs_match_string(clock_names, buf);
1467 if (ret < 0)
1468 return ret;
1469 clk = ret;
1470
1471 switch (clk) {
1472 case CLOCK_REALTIME:
1473 case CLOCK_MONOTONIC:
1474 case CLOCK_MONOTONIC_RAW:
1475 case CLOCK_REALTIME_COARSE:
1476 case CLOCK_MONOTONIC_COARSE:
1477 case CLOCK_BOOTTIME:
1478 case CLOCK_TAI:
1479 break;
1480 default:
1481 return -EINVAL;
1482 }
1483
1484 ret = iio_device_set_clock(dev_to_iio_dev(dev), clk);
1485 if (ret)
1486 return ret;
1487
1488 return len;
1489}
1490
1491int iio_device_register_sysfs_group(struct iio_dev *indio_dev,
1492 const struct attribute_group *group)
1493{
1494 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1495 const struct attribute_group **new, **old = iio_dev_opaque->groups;
1496 unsigned int cnt = iio_dev_opaque->groupcounter;
1497
1498 new = krealloc_array(old, cnt + 2, sizeof(*new), GFP_KERNEL);
1499 if (!new)
1500 return -ENOMEM;
1501
1502 new[iio_dev_opaque->groupcounter++] = group;
1503 new[iio_dev_opaque->groupcounter] = NULL;
1504
1505 iio_dev_opaque->groups = new;
1506
1507 return 0;
1508}
1509
1510static DEVICE_ATTR_RW(current_timestamp_clock);
1511
1512static int iio_device_register_sysfs(struct iio_dev *indio_dev)
1513{
1514 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1515 int i, ret = 0, attrcount, attrn, attrcount_orig = 0;
1516 struct iio_dev_attr *p;
1517 struct attribute **attr, *clk = NULL;
1518
1519 /* First count elements in any existing group */
1520 if (indio_dev->info->attrs) {
1521 attr = indio_dev->info->attrs->attrs;
1522 while (*attr++ != NULL)
1523 attrcount_orig++;
1524 }
1525 attrcount = attrcount_orig;
1526 /*
1527 * New channel registration method - relies on the fact a group does
1528 * not need to be initialized if its name is NULL.
1529 */
1530 if (indio_dev->channels)
1531 for (i = 0; i < indio_dev->num_channels; i++) {
1532 const struct iio_chan_spec *chan =
1533 &indio_dev->channels[i];
1534
1535 if (chan->type == IIO_TIMESTAMP)
1536 clk = &dev_attr_current_timestamp_clock.attr;
1537
1538 ret = iio_device_add_channel_sysfs(indio_dev, chan);
1539 if (ret < 0)
1540 goto error_clear_attrs;
1541 attrcount += ret;
1542 }
1543
1544 if (iio_dev_opaque->event_interface)
1545 clk = &dev_attr_current_timestamp_clock.attr;
1546
1547 if (indio_dev->name)
1548 attrcount++;
1549 if (indio_dev->label)
1550 attrcount++;
1551 if (clk)
1552 attrcount++;
1553
1554 iio_dev_opaque->chan_attr_group.attrs =
1555 kcalloc(attrcount + 1,
1556 sizeof(iio_dev_opaque->chan_attr_group.attrs[0]),
1557 GFP_KERNEL);
1558 if (iio_dev_opaque->chan_attr_group.attrs == NULL) {
1559 ret = -ENOMEM;
1560 goto error_clear_attrs;
1561 }
1562 /* Copy across original attributes, and point to original binary attributes */
1563 if (indio_dev->info->attrs) {
1564 memcpy(iio_dev_opaque->chan_attr_group.attrs,
1565 indio_dev->info->attrs->attrs,
1566 sizeof(iio_dev_opaque->chan_attr_group.attrs[0])
1567 *attrcount_orig);
1568 iio_dev_opaque->chan_attr_group.is_visible =
1569 indio_dev->info->attrs->is_visible;
1570 iio_dev_opaque->chan_attr_group.bin_attrs =
1571 indio_dev->info->attrs->bin_attrs;
1572 }
1573 attrn = attrcount_orig;
1574 /* Add all elements from the list. */
1575 list_for_each_entry(p, &iio_dev_opaque->channel_attr_list, l)
1576 iio_dev_opaque->chan_attr_group.attrs[attrn++] = &p->dev_attr.attr;
1577 if (indio_dev->name)
1578 iio_dev_opaque->chan_attr_group.attrs[attrn++] = &dev_attr_name.attr;
1579 if (indio_dev->label)
1580 iio_dev_opaque->chan_attr_group.attrs[attrn++] = &dev_attr_label.attr;
1581 if (clk)
1582 iio_dev_opaque->chan_attr_group.attrs[attrn++] = clk;
1583
1584 ret = iio_device_register_sysfs_group(indio_dev,
1585 &iio_dev_opaque->chan_attr_group);
1586 if (ret)
1587 goto error_free_chan_attrs;
1588
1589 return 0;
1590
1591error_free_chan_attrs:
1592 kfree(iio_dev_opaque->chan_attr_group.attrs);
1593 iio_dev_opaque->chan_attr_group.attrs = NULL;
1594error_clear_attrs:
1595 iio_free_chan_devattr_list(&iio_dev_opaque->channel_attr_list);
1596
1597 return ret;
1598}
1599
1600static void iio_device_unregister_sysfs(struct iio_dev *indio_dev)
1601{
1602 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1603
1604 iio_free_chan_devattr_list(&iio_dev_opaque->channel_attr_list);
1605 kfree(iio_dev_opaque->chan_attr_group.attrs);
1606 iio_dev_opaque->chan_attr_group.attrs = NULL;
1607 kfree(iio_dev_opaque->groups);
1608 iio_dev_opaque->groups = NULL;
1609}
1610
1611static void iio_dev_release(struct device *device)
1612{
1613 struct iio_dev *indio_dev = dev_to_iio_dev(device);
1614 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1615
1616 if (indio_dev->modes & INDIO_ALL_TRIGGERED_MODES)
1617 iio_device_unregister_trigger_consumer(indio_dev);
1618 iio_device_unregister_eventset(indio_dev);
1619 iio_device_unregister_sysfs(indio_dev);
1620
1621 iio_device_detach_buffers(indio_dev);
1622
1623 lockdep_unregister_key(&iio_dev_opaque->mlock_key);
1624
1625 ida_free(&iio_ida, iio_dev_opaque->id);
1626 kfree(iio_dev_opaque);
1627}
1628
1629const struct device_type iio_device_type = {
1630 .name = "iio_device",
1631 .release = iio_dev_release,
1632};
1633
1634/**
1635 * iio_device_alloc() - allocate an iio_dev from a driver
1636 * @parent: Parent device.
1637 * @sizeof_priv: Space to allocate for private structure.
1638 *
1639 * Returns:
1640 * Pointer to allocated iio_dev on success, NULL on failure.
1641 */
1642struct iio_dev *iio_device_alloc(struct device *parent, int sizeof_priv)
1643{
1644 struct iio_dev_opaque *iio_dev_opaque;
1645 struct iio_dev *indio_dev;
1646 size_t alloc_size;
1647
1648 alloc_size = sizeof(struct iio_dev_opaque);
1649 if (sizeof_priv) {
1650 alloc_size = ALIGN(alloc_size, IIO_DMA_MINALIGN);
1651 alloc_size += sizeof_priv;
1652 }
1653
1654 iio_dev_opaque = kzalloc(alloc_size, GFP_KERNEL);
1655 if (!iio_dev_opaque)
1656 return NULL;
1657
1658 indio_dev = &iio_dev_opaque->indio_dev;
1659 indio_dev->priv = (char *)iio_dev_opaque +
1660 ALIGN(sizeof(struct iio_dev_opaque), IIO_DMA_MINALIGN);
1661
1662 indio_dev->dev.parent = parent;
1663 indio_dev->dev.type = &iio_device_type;
1664 indio_dev->dev.bus = &iio_bus_type;
1665 device_initialize(&indio_dev->dev);
1666 mutex_init(&iio_dev_opaque->mlock);
1667 mutex_init(&iio_dev_opaque->info_exist_lock);
1668 INIT_LIST_HEAD(&iio_dev_opaque->channel_attr_list);
1669
1670 iio_dev_opaque->id = ida_alloc(&iio_ida, GFP_KERNEL);
1671 if (iio_dev_opaque->id < 0) {
1672 /* cannot use a dev_err as the name isn't available */
1673 pr_err("failed to get device id\n");
1674 kfree(iio_dev_opaque);
1675 return NULL;
1676 }
1677
1678 if (dev_set_name(&indio_dev->dev, "iio:device%d", iio_dev_opaque->id)) {
1679 ida_free(&iio_ida, iio_dev_opaque->id);
1680 kfree(iio_dev_opaque);
1681 return NULL;
1682 }
1683
1684 INIT_LIST_HEAD(&iio_dev_opaque->buffer_list);
1685 INIT_LIST_HEAD(&iio_dev_opaque->ioctl_handlers);
1686
1687 lockdep_register_key(&iio_dev_opaque->mlock_key);
1688 lockdep_set_class(&iio_dev_opaque->mlock, &iio_dev_opaque->mlock_key);
1689
1690 return indio_dev;
1691}
1692EXPORT_SYMBOL(iio_device_alloc);
1693
1694/**
1695 * iio_device_free() - free an iio_dev from a driver
1696 * @dev: the iio_dev associated with the device
1697 */
1698void iio_device_free(struct iio_dev *dev)
1699{
1700 if (dev)
1701 put_device(&dev->dev);
1702}
1703EXPORT_SYMBOL(iio_device_free);
1704
1705static void devm_iio_device_release(void *iio_dev)
1706{
1707 iio_device_free(iio_dev);
1708}
1709
1710/**
1711 * devm_iio_device_alloc - Resource-managed iio_device_alloc()
1712 * @parent: Device to allocate iio_dev for, and parent for this IIO device
1713 * @sizeof_priv: Space to allocate for private structure.
1714 *
1715 * Managed iio_device_alloc. iio_dev allocated with this function is
1716 * automatically freed on driver detach.
1717 *
1718 * Returns:
1719 * Pointer to allocated iio_dev on success, NULL on failure.
1720 */
1721struct iio_dev *devm_iio_device_alloc(struct device *parent, int sizeof_priv)
1722{
1723 struct iio_dev *iio_dev;
1724 int ret;
1725
1726 iio_dev = iio_device_alloc(parent, sizeof_priv);
1727 if (!iio_dev)
1728 return NULL;
1729
1730 ret = devm_add_action_or_reset(parent, devm_iio_device_release,
1731 iio_dev);
1732 if (ret)
1733 return NULL;
1734
1735 return iio_dev;
1736}
1737EXPORT_SYMBOL_GPL(devm_iio_device_alloc);
1738
1739/**
1740 * iio_chrdev_open() - chrdev file open for buffer access and ioctls
1741 * @inode: Inode structure for identifying the device in the file system
1742 * @filp: File structure for iio device used to keep and later access
1743 * private data
1744 *
1745 * Returns: 0 on success or -EBUSY if the device is already opened
1746 */
1747static int iio_chrdev_open(struct inode *inode, struct file *filp)
1748{
1749 struct iio_dev_opaque *iio_dev_opaque =
1750 container_of(inode->i_cdev, struct iio_dev_opaque, chrdev);
1751 struct iio_dev *indio_dev = &iio_dev_opaque->indio_dev;
1752 struct iio_dev_buffer_pair *ib;
1753
1754 if (test_and_set_bit(IIO_BUSY_BIT_POS, &iio_dev_opaque->flags))
1755 return -EBUSY;
1756
1757 iio_device_get(indio_dev);
1758
1759 ib = kmalloc(sizeof(*ib), GFP_KERNEL);
1760 if (!ib) {
1761 iio_device_put(indio_dev);
1762 clear_bit(IIO_BUSY_BIT_POS, &iio_dev_opaque->flags);
1763 return -ENOMEM;
1764 }
1765
1766 ib->indio_dev = indio_dev;
1767 ib->buffer = indio_dev->buffer;
1768
1769 filp->private_data = ib;
1770
1771 return 0;
1772}
1773
1774/**
1775 * iio_chrdev_release() - chrdev file close buffer access and ioctls
1776 * @inode: Inode structure pointer for the char device
1777 * @filp: File structure pointer for the char device
1778 *
1779 * Returns: 0 for successful release.
1780 */
1781static int iio_chrdev_release(struct inode *inode, struct file *filp)
1782{
1783 struct iio_dev_buffer_pair *ib = filp->private_data;
1784 struct iio_dev_opaque *iio_dev_opaque =
1785 container_of(inode->i_cdev, struct iio_dev_opaque, chrdev);
1786 struct iio_dev *indio_dev = &iio_dev_opaque->indio_dev;
1787
1788 kfree(ib);
1789 clear_bit(IIO_BUSY_BIT_POS, &iio_dev_opaque->flags);
1790 iio_device_put(indio_dev);
1791
1792 return 0;
1793}
1794
1795void iio_device_ioctl_handler_register(struct iio_dev *indio_dev,
1796 struct iio_ioctl_handler *h)
1797{
1798 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1799
1800 list_add_tail(&h->entry, &iio_dev_opaque->ioctl_handlers);
1801}
1802
1803void iio_device_ioctl_handler_unregister(struct iio_ioctl_handler *h)
1804{
1805 list_del(&h->entry);
1806}
1807
1808static long iio_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
1809{
1810 struct iio_dev_buffer_pair *ib = filp->private_data;
1811 struct iio_dev *indio_dev = ib->indio_dev;
1812 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1813 struct iio_ioctl_handler *h;
1814 int ret = -ENODEV;
1815
1816 mutex_lock(&iio_dev_opaque->info_exist_lock);
1817
1818 /*
1819 * The NULL check here is required to prevent crashing when a device
1820 * is being removed while userspace would still have open file handles
1821 * to try to access this device.
1822 */
1823 if (!indio_dev->info)
1824 goto out_unlock;
1825
1826 list_for_each_entry(h, &iio_dev_opaque->ioctl_handlers, entry) {
1827 ret = h->ioctl(indio_dev, filp, cmd, arg);
1828 if (ret != IIO_IOCTL_UNHANDLED)
1829 break;
1830 }
1831
1832 if (ret == IIO_IOCTL_UNHANDLED)
1833 ret = -ENODEV;
1834
1835out_unlock:
1836 mutex_unlock(&iio_dev_opaque->info_exist_lock);
1837
1838 return ret;
1839}
1840
1841static const struct file_operations iio_buffer_fileops = {
1842 .owner = THIS_MODULE,
1843 .llseek = noop_llseek,
1844 .read = iio_buffer_read_outer_addr,
1845 .write = iio_buffer_write_outer_addr,
1846 .poll = iio_buffer_poll_addr,
1847 .unlocked_ioctl = iio_ioctl,
1848 .compat_ioctl = compat_ptr_ioctl,
1849 .open = iio_chrdev_open,
1850 .release = iio_chrdev_release,
1851};
1852
1853static const struct file_operations iio_event_fileops = {
1854 .owner = THIS_MODULE,
1855 .llseek = noop_llseek,
1856 .unlocked_ioctl = iio_ioctl,
1857 .compat_ioctl = compat_ptr_ioctl,
1858 .open = iio_chrdev_open,
1859 .release = iio_chrdev_release,
1860};
1861
1862static int iio_check_unique_scan_index(struct iio_dev *indio_dev)
1863{
1864 int i, j;
1865 const struct iio_chan_spec *channels = indio_dev->channels;
1866
1867 if (!(indio_dev->modes & INDIO_ALL_BUFFER_MODES))
1868 return 0;
1869
1870 for (i = 0; i < indio_dev->num_channels - 1; i++) {
1871 if (channels[i].scan_index < 0)
1872 continue;
1873 for (j = i + 1; j < indio_dev->num_channels; j++)
1874 if (channels[i].scan_index == channels[j].scan_index) {
1875 dev_err(&indio_dev->dev,
1876 "Duplicate scan index %d\n",
1877 channels[i].scan_index);
1878 return -EINVAL;
1879 }
1880 }
1881
1882 return 0;
1883}
1884
1885static int iio_check_extended_name(const struct iio_dev *indio_dev)
1886{
1887 unsigned int i;
1888
1889 if (!indio_dev->info->read_label)
1890 return 0;
1891
1892 for (i = 0; i < indio_dev->num_channels; i++) {
1893 if (indio_dev->channels[i].extend_name) {
1894 dev_err(&indio_dev->dev,
1895 "Cannot use labels and extend_name at the same time\n");
1896 return -EINVAL;
1897 }
1898 }
1899
1900 return 0;
1901}
1902
1903static const struct iio_buffer_setup_ops noop_ring_setup_ops;
1904
1905static void iio_sanity_check_avail_scan_masks(struct iio_dev *indio_dev)
1906{
1907 unsigned int num_masks, masklength, longs_per_mask;
1908 const unsigned long *av_masks;
1909 int i;
1910
1911 av_masks = indio_dev->available_scan_masks;
1912 masklength = indio_dev->masklength;
1913 longs_per_mask = BITS_TO_LONGS(masklength);
1914
1915 /*
1916 * The code determining how many available_scan_masks is in the array
1917 * will be assuming the end of masks when first long with all bits
1918 * zeroed is encountered. This is incorrect for masks where mask
1919 * consists of more than one long, and where some of the available masks
1920 * has long worth of bits zeroed (but has subsequent bit(s) set). This
1921 * is a safety measure against bug where array of masks is terminated by
1922 * a single zero while mask width is greater than width of a long.
1923 */
1924 if (longs_per_mask > 1)
1925 dev_warn(indio_dev->dev.parent,
1926 "multi long available scan masks not fully supported\n");
1927
1928 if (bitmap_empty(av_masks, masklength))
1929 dev_warn(indio_dev->dev.parent, "empty scan mask\n");
1930
1931 for (num_masks = 0; *av_masks; num_masks++)
1932 av_masks += longs_per_mask;
1933
1934 if (num_masks < 2)
1935 return;
1936
1937 av_masks = indio_dev->available_scan_masks;
1938
1939 /*
1940 * Go through all the masks from first to one before the last, and see
1941 * that no mask found later from the available_scan_masks array is a
1942 * subset of mask found earlier. If this happens, then the mask found
1943 * later will never get used because scanning the array is stopped when
1944 * the first suitable mask is found. Drivers should order the array of
1945 * available masks in the order of preference (presumably the least
1946 * costy to access masks first).
1947 */
1948 for (i = 0; i < num_masks - 1; i++) {
1949 const unsigned long *mask1;
1950 int j;
1951
1952 mask1 = av_masks + i * longs_per_mask;
1953 for (j = i + 1; j < num_masks; j++) {
1954 const unsigned long *mask2;
1955
1956 mask2 = av_masks + j * longs_per_mask;
1957 if (bitmap_subset(mask2, mask1, masklength))
1958 dev_warn(indio_dev->dev.parent,
1959 "available_scan_mask %d subset of %d. Never used\n",
1960 j, i);
1961 }
1962 }
1963}
1964
1965int __iio_device_register(struct iio_dev *indio_dev, struct module *this_mod)
1966{
1967 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
1968 struct fwnode_handle *fwnode = NULL;
1969 int ret;
1970
1971 if (!indio_dev->info)
1972 return -EINVAL;
1973
1974 iio_dev_opaque->driver_module = this_mod;
1975
1976 /* If the calling driver did not initialize firmware node, do it here */
1977 if (dev_fwnode(&indio_dev->dev))
1978 fwnode = dev_fwnode(&indio_dev->dev);
1979 /* The default dummy IIO device has no parent */
1980 else if (indio_dev->dev.parent)
1981 fwnode = dev_fwnode(indio_dev->dev.parent);
1982 device_set_node(&indio_dev->dev, fwnode);
1983
1984 fwnode_property_read_string(fwnode, "label", &indio_dev->label);
1985
1986 ret = iio_check_unique_scan_index(indio_dev);
1987 if (ret < 0)
1988 return ret;
1989
1990 ret = iio_check_extended_name(indio_dev);
1991 if (ret < 0)
1992 return ret;
1993
1994 iio_device_register_debugfs(indio_dev);
1995
1996 ret = iio_buffers_alloc_sysfs_and_mask(indio_dev);
1997 if (ret) {
1998 dev_err(indio_dev->dev.parent,
1999 "Failed to create buffer sysfs interfaces\n");
2000 goto error_unreg_debugfs;
2001 }
2002
2003 if (indio_dev->available_scan_masks)
2004 iio_sanity_check_avail_scan_masks(indio_dev);
2005
2006 ret = iio_device_register_sysfs(indio_dev);
2007 if (ret) {
2008 dev_err(indio_dev->dev.parent,
2009 "Failed to register sysfs interfaces\n");
2010 goto error_buffer_free_sysfs;
2011 }
2012 ret = iio_device_register_eventset(indio_dev);
2013 if (ret) {
2014 dev_err(indio_dev->dev.parent,
2015 "Failed to register event set\n");
2016 goto error_free_sysfs;
2017 }
2018 if (indio_dev->modes & INDIO_ALL_TRIGGERED_MODES)
2019 iio_device_register_trigger_consumer(indio_dev);
2020
2021 if ((indio_dev->modes & INDIO_ALL_BUFFER_MODES) &&
2022 indio_dev->setup_ops == NULL)
2023 indio_dev->setup_ops = &noop_ring_setup_ops;
2024
2025 if (iio_dev_opaque->attached_buffers_cnt)
2026 cdev_init(&iio_dev_opaque->chrdev, &iio_buffer_fileops);
2027 else if (iio_dev_opaque->event_interface)
2028 cdev_init(&iio_dev_opaque->chrdev, &iio_event_fileops);
2029
2030 if (iio_dev_opaque->attached_buffers_cnt || iio_dev_opaque->event_interface) {
2031 indio_dev->dev.devt = MKDEV(MAJOR(iio_devt), iio_dev_opaque->id);
2032 iio_dev_opaque->chrdev.owner = this_mod;
2033 }
2034
2035 /* assign device groups now; they should be all registered now */
2036 indio_dev->dev.groups = iio_dev_opaque->groups;
2037
2038 ret = cdev_device_add(&iio_dev_opaque->chrdev, &indio_dev->dev);
2039 if (ret < 0)
2040 goto error_unreg_eventset;
2041
2042 return 0;
2043
2044error_unreg_eventset:
2045 iio_device_unregister_eventset(indio_dev);
2046error_free_sysfs:
2047 iio_device_unregister_sysfs(indio_dev);
2048error_buffer_free_sysfs:
2049 iio_buffers_free_sysfs_and_mask(indio_dev);
2050error_unreg_debugfs:
2051 iio_device_unregister_debugfs(indio_dev);
2052 return ret;
2053}
2054EXPORT_SYMBOL(__iio_device_register);
2055
2056/**
2057 * iio_device_unregister() - unregister a device from the IIO subsystem
2058 * @indio_dev: Device structure representing the device.
2059 */
2060void iio_device_unregister(struct iio_dev *indio_dev)
2061{
2062 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
2063
2064 cdev_device_del(&iio_dev_opaque->chrdev, &indio_dev->dev);
2065
2066 mutex_lock(&iio_dev_opaque->info_exist_lock);
2067
2068 iio_device_unregister_debugfs(indio_dev);
2069
2070 iio_disable_all_buffers(indio_dev);
2071
2072 indio_dev->info = NULL;
2073
2074 iio_device_wakeup_eventset(indio_dev);
2075 iio_buffer_wakeup_poll(indio_dev);
2076
2077 mutex_unlock(&iio_dev_opaque->info_exist_lock);
2078
2079 iio_buffers_free_sysfs_and_mask(indio_dev);
2080}
2081EXPORT_SYMBOL(iio_device_unregister);
2082
2083static void devm_iio_device_unreg(void *indio_dev)
2084{
2085 iio_device_unregister(indio_dev);
2086}
2087
2088int __devm_iio_device_register(struct device *dev, struct iio_dev *indio_dev,
2089 struct module *this_mod)
2090{
2091 int ret;
2092
2093 ret = __iio_device_register(indio_dev, this_mod);
2094 if (ret)
2095 return ret;
2096
2097 return devm_add_action_or_reset(dev, devm_iio_device_unreg, indio_dev);
2098}
2099EXPORT_SYMBOL_GPL(__devm_iio_device_register);
2100
2101/**
2102 * iio_device_claim_direct_mode - Keep device in direct mode
2103 * @indio_dev: the iio_dev associated with the device
2104 *
2105 * If the device is in direct mode it is guaranteed to stay
2106 * that way until iio_device_release_direct_mode() is called.
2107 *
2108 * Use with iio_device_release_direct_mode()
2109 *
2110 * Returns: 0 on success, -EBUSY on failure.
2111 */
2112int iio_device_claim_direct_mode(struct iio_dev *indio_dev)
2113{
2114 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
2115
2116 mutex_lock(&iio_dev_opaque->mlock);
2117
2118 if (iio_buffer_enabled(indio_dev)) {
2119 mutex_unlock(&iio_dev_opaque->mlock);
2120 return -EBUSY;
2121 }
2122 return 0;
2123}
2124EXPORT_SYMBOL_GPL(iio_device_claim_direct_mode);
2125
2126/**
2127 * iio_device_release_direct_mode - releases claim on direct mode
2128 * @indio_dev: the iio_dev associated with the device
2129 *
2130 * Release the claim. Device is no longer guaranteed to stay
2131 * in direct mode.
2132 *
2133 * Use with iio_device_claim_direct_mode()
2134 */
2135void iio_device_release_direct_mode(struct iio_dev *indio_dev)
2136{
2137 mutex_unlock(&to_iio_dev_opaque(indio_dev)->mlock);
2138}
2139EXPORT_SYMBOL_GPL(iio_device_release_direct_mode);
2140
2141/**
2142 * iio_device_claim_buffer_mode - Keep device in buffer mode
2143 * @indio_dev: the iio_dev associated with the device
2144 *
2145 * If the device is in buffer mode it is guaranteed to stay
2146 * that way until iio_device_release_buffer_mode() is called.
2147 *
2148 * Use with iio_device_release_buffer_mode().
2149 *
2150 * Returns: 0 on success, -EBUSY on failure.
2151 */
2152int iio_device_claim_buffer_mode(struct iio_dev *indio_dev)
2153{
2154 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
2155
2156 mutex_lock(&iio_dev_opaque->mlock);
2157
2158 if (iio_buffer_enabled(indio_dev))
2159 return 0;
2160
2161 mutex_unlock(&iio_dev_opaque->mlock);
2162 return -EBUSY;
2163}
2164EXPORT_SYMBOL_GPL(iio_device_claim_buffer_mode);
2165
2166/**
2167 * iio_device_release_buffer_mode - releases claim on buffer mode
2168 * @indio_dev: the iio_dev associated with the device
2169 *
2170 * Release the claim. Device is no longer guaranteed to stay
2171 * in buffer mode.
2172 *
2173 * Use with iio_device_claim_buffer_mode().
2174 */
2175void iio_device_release_buffer_mode(struct iio_dev *indio_dev)
2176{
2177 mutex_unlock(&to_iio_dev_opaque(indio_dev)->mlock);
2178}
2179EXPORT_SYMBOL_GPL(iio_device_release_buffer_mode);
2180
2181/**
2182 * iio_device_get_current_mode() - helper function providing read-only access to
2183 * the opaque @currentmode variable
2184 * @indio_dev: IIO device structure for device
2185 */
2186int iio_device_get_current_mode(struct iio_dev *indio_dev)
2187{
2188 struct iio_dev_opaque *iio_dev_opaque = to_iio_dev_opaque(indio_dev);
2189
2190 return iio_dev_opaque->currentmode;
2191}
2192EXPORT_SYMBOL_GPL(iio_device_get_current_mode);
2193
2194subsys_initcall(iio_init);
2195module_exit(iio_exit);
2196
2197MODULE_AUTHOR("Jonathan Cameron <jic23@kernel.org>");
2198MODULE_DESCRIPTION("Industrial I/O core");
2199MODULE_LICENSE("GPL");