[media] DocBook: fix incorrect or missing links
[linux-2.6/btrfs-unstable.git] / drivers / iio / inkern.c
blobb5afc2ff34fd5c5f58d0c83e3d1538a3d901acac
1 /* The industrial I/O core in kernel channel mapping
3 * Copyright (c) 2011 Jonathan Cameron
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 as published by
7 * the Free Software Foundation.
8 */
9 #include <linux/err.h>
10 #include <linux/export.h>
11 #include <linux/slab.h>
12 #include <linux/mutex.h>
14 #include <linux/iio/iio.h>
15 #include "iio_core.h"
16 #include <linux/iio/machine.h>
17 #include <linux/iio/driver.h>
18 #include <linux/iio/consumer.h>
20 struct iio_map_internal {
21 struct iio_dev *indio_dev;
22 struct iio_map *map;
23 struct list_head l;
26 static LIST_HEAD(iio_map_list);
27 static DEFINE_MUTEX(iio_map_list_lock);
29 int iio_map_array_register(struct iio_dev *indio_dev, struct iio_map *maps)
31 int i = 0, ret = 0;
32 struct iio_map_internal *mapi;
34 if (maps == NULL)
35 return 0;
37 mutex_lock(&iio_map_list_lock);
38 while (maps[i].consumer_dev_name != NULL) {
39 mapi = kzalloc(sizeof(*mapi), GFP_KERNEL);
40 if (mapi == NULL) {
41 ret = -ENOMEM;
42 goto error_ret;
44 mapi->map = &maps[i];
45 mapi->indio_dev = indio_dev;
46 list_add(&mapi->l, &iio_map_list);
47 i++;
49 error_ret:
50 mutex_unlock(&iio_map_list_lock);
52 return ret;
54 EXPORT_SYMBOL_GPL(iio_map_array_register);
57 /* Assumes the exact same array (e.g. memory locations)
58 * used at unregistration as used at registration rather than
59 * more complex checking of contents.
61 int iio_map_array_unregister(struct iio_dev *indio_dev,
62 struct iio_map *maps)
64 int i = 0, ret = 0;
65 bool found_it;
66 struct iio_map_internal *mapi;
68 if (maps == NULL)
69 return 0;
71 mutex_lock(&iio_map_list_lock);
72 while (maps[i].consumer_dev_name != NULL) {
73 found_it = false;
74 list_for_each_entry(mapi, &iio_map_list, l)
75 if (&maps[i] == mapi->map) {
76 list_del(&mapi->l);
77 kfree(mapi);
78 found_it = true;
79 break;
81 if (found_it == false) {
82 ret = -ENODEV;
83 goto error_ret;
85 i++;
87 error_ret:
88 mutex_unlock(&iio_map_list_lock);
90 return ret;
92 EXPORT_SYMBOL_GPL(iio_map_array_unregister);
94 static const struct iio_chan_spec
95 *iio_chan_spec_from_name(const struct iio_dev *indio_dev, const char *name)
97 int i;
98 const struct iio_chan_spec *chan = NULL;
100 for (i = 0; i < indio_dev->num_channels; i++)
101 if (indio_dev->channels[i].datasheet_name &&
102 strcmp(name, indio_dev->channels[i].datasheet_name) == 0) {
103 chan = &indio_dev->channels[i];
104 break;
106 return chan;
110 struct iio_channel *iio_channel_get(const char *name, const char *channel_name)
112 struct iio_map_internal *c_i = NULL, *c = NULL;
113 struct iio_channel *channel;
115 if (name == NULL && channel_name == NULL)
116 return ERR_PTR(-ENODEV);
118 /* first find matching entry the channel map */
119 mutex_lock(&iio_map_list_lock);
120 list_for_each_entry(c_i, &iio_map_list, l) {
121 if ((name && strcmp(name, c_i->map->consumer_dev_name) != 0) ||
122 (channel_name &&
123 strcmp(channel_name, c_i->map->consumer_channel) != 0))
124 continue;
125 c = c_i;
126 iio_device_get(c->indio_dev);
127 break;
129 mutex_unlock(&iio_map_list_lock);
130 if (c == NULL)
131 return ERR_PTR(-ENODEV);
133 channel = kmalloc(sizeof(*channel), GFP_KERNEL);
134 if (channel == NULL)
135 return ERR_PTR(-ENOMEM);
137 channel->indio_dev = c->indio_dev;
139 if (c->map->adc_channel_label)
140 channel->channel =
141 iio_chan_spec_from_name(channel->indio_dev,
142 c->map->adc_channel_label);
144 return channel;
146 EXPORT_SYMBOL_GPL(iio_channel_get);
148 void iio_channel_release(struct iio_channel *channel)
150 iio_device_put(channel->indio_dev);
151 kfree(channel);
153 EXPORT_SYMBOL_GPL(iio_channel_release);
155 struct iio_channel *iio_channel_get_all(const char *name)
157 struct iio_channel *chans;
158 struct iio_map_internal *c = NULL;
159 int nummaps = 0;
160 int mapind = 0;
161 int i, ret;
163 if (name == NULL)
164 return ERR_PTR(-EINVAL);
166 mutex_lock(&iio_map_list_lock);
167 /* first count the matching maps */
168 list_for_each_entry(c, &iio_map_list, l)
169 if (name && strcmp(name, c->map->consumer_dev_name) != 0)
170 continue;
171 else
172 nummaps++;
174 if (nummaps == 0) {
175 ret = -ENODEV;
176 goto error_ret;
179 /* NULL terminated array to save passing size */
180 chans = kzalloc(sizeof(*chans)*(nummaps + 1), GFP_KERNEL);
181 if (chans == NULL) {
182 ret = -ENOMEM;
183 goto error_ret;
186 /* for each map fill in the chans element */
187 list_for_each_entry(c, &iio_map_list, l) {
188 if (name && strcmp(name, c->map->consumer_dev_name) != 0)
189 continue;
190 chans[mapind].indio_dev = c->indio_dev;
191 chans[mapind].channel =
192 iio_chan_spec_from_name(chans[mapind].indio_dev,
193 c->map->adc_channel_label);
194 if (chans[mapind].channel == NULL) {
195 ret = -EINVAL;
196 goto error_free_chans;
198 iio_device_get(chans[mapind].indio_dev);
199 mapind++;
201 if (mapind == 0) {
202 ret = -ENODEV;
203 goto error_free_chans;
205 mutex_unlock(&iio_map_list_lock);
207 return chans;
209 error_free_chans:
210 for (i = 0; i < nummaps; i++)
211 iio_device_put(chans[i].indio_dev);
212 kfree(chans);
213 error_ret:
214 mutex_unlock(&iio_map_list_lock);
216 return ERR_PTR(ret);
218 EXPORT_SYMBOL_GPL(iio_channel_get_all);
220 void iio_channel_release_all(struct iio_channel *channels)
222 struct iio_channel *chan = &channels[0];
224 while (chan->indio_dev) {
225 iio_device_put(chan->indio_dev);
226 chan++;
228 kfree(channels);
230 EXPORT_SYMBOL_GPL(iio_channel_release_all);
232 int iio_read_channel_raw(struct iio_channel *chan, int *val)
234 int val2, ret;
236 mutex_lock(&chan->indio_dev->info_exist_lock);
237 if (chan->indio_dev->info == NULL) {
238 ret = -ENODEV;
239 goto err_unlock;
242 ret = chan->indio_dev->info->read_raw(chan->indio_dev, chan->channel,
243 val, &val2, 0);
244 err_unlock:
245 mutex_unlock(&chan->indio_dev->info_exist_lock);
247 return ret;
249 EXPORT_SYMBOL_GPL(iio_read_channel_raw);
251 int iio_read_channel_scale(struct iio_channel *chan, int *val, int *val2)
253 int ret;
255 mutex_lock(&chan->indio_dev->info_exist_lock);
256 if (chan->indio_dev->info == NULL) {
257 ret = -ENODEV;
258 goto err_unlock;
261 ret = chan->indio_dev->info->read_raw(chan->indio_dev,
262 chan->channel,
263 val, val2,
264 IIO_CHAN_INFO_SCALE);
265 err_unlock:
266 mutex_unlock(&chan->indio_dev->info_exist_lock);
268 return ret;
270 EXPORT_SYMBOL_GPL(iio_read_channel_scale);
272 int iio_get_channel_type(struct iio_channel *chan, enum iio_chan_type *type)
274 int ret = 0;
275 /* Need to verify underlying driver has not gone away */
277 mutex_lock(&chan->indio_dev->info_exist_lock);
278 if (chan->indio_dev->info == NULL) {
279 ret = -ENODEV;
280 goto err_unlock;
283 *type = chan->channel->type;
284 err_unlock:
285 mutex_unlock(&chan->indio_dev->info_exist_lock);
287 return ret;
289 EXPORT_SYMBOL_GPL(iio_get_channel_type);