Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Copyright (c) 2003 Evgeniy Polyakov <zbr@ioremap.net>
4 */
5
6#include <linux/slab.h>
7#include <linux/skbuff.h>
8#include <linux/netlink.h>
9#include <linux/connector.h>
10
11#include "w1_internal.h"
12#include "w1_netlink.h"
13
14#if defined(CONFIG_W1_CON) && (defined(CONFIG_CONNECTOR) || (defined(CONFIG_CONNECTOR_MODULE) && defined(CONFIG_W1_MODULE)))
15
16/* Bundle together everything required to process a request in one memory
17 * allocation.
18 */
19struct w1_cb_block {
20 atomic_t refcnt;
21 u32 portid; /* Sending process port ID */
22 /* maximum value for first_cn->len */
23 u16 maxlen;
24 /* pointers to building up the reply message */
25 struct cn_msg *first_cn; /* fixed once the structure is populated */
26 struct cn_msg *cn; /* advances as cn_msg is appeneded */
27 struct w1_netlink_msg *msg; /* advances as w1_netlink_msg is appened */
28 struct w1_netlink_cmd *cmd; /* advances as cmds are appened */
29 struct w1_netlink_msg *cur_msg; /* currently message being processed */
30 /* copy of the original request follows */
31 struct cn_msg request_cn;
32 /* followed by variable length:
33 * cn_msg, data (w1_netlink_msg and w1_netlink_cmd)
34 * one or more struct w1_cb_node
35 * reply first_cn, data (w1_netlink_msg and w1_netlink_cmd)
36 */
37};
38struct w1_cb_node {
39 struct w1_async_cmd async;
40 /* pointers within w1_cb_block and cn data */
41 struct w1_cb_block *block;
42 struct w1_netlink_msg *msg;
43 struct w1_slave *sl;
44 struct w1_master *dev;
45};
46
47/**
48 * w1_reply_len() - calculate current reply length, compare to maxlen
49 * @block: block to calculate
50 *
51 * Calculates the current message length including possible multiple
52 * cn_msg and data, excludes the first sizeof(struct cn_msg). Direclty
53 * compariable to maxlen and usable to send the message.
54 */
55static u16 w1_reply_len(struct w1_cb_block *block)
56{
57 if (!block->cn)
58 return 0;
59 return (u8 *)block->cn - (u8 *)block->first_cn + block->cn->len;
60}
61
62static void w1_unref_block(struct w1_cb_block *block)
63{
64 if (atomic_sub_return(1, &block->refcnt) == 0) {
65 u16 len = w1_reply_len(block);
66 if (len) {
67 cn_netlink_send_mult(block->first_cn, len,
68 block->portid, 0, GFP_KERNEL);
69 }
70 kfree(block);
71 }
72}
73
74/**
75 * w1_reply_make_space() - send message if needed to make space
76 * @block: block to make space on
77 * @space: how many bytes requested
78 *
79 * Verify there is enough room left for the caller to add "space" bytes to the
80 * message, if there isn't send the message and reset.
81 */
82static void w1_reply_make_space(struct w1_cb_block *block, u16 space)
83{
84 u16 len = w1_reply_len(block);
85 if (len + space >= block->maxlen) {
86 cn_netlink_send_mult(block->first_cn, len, block->portid, 0, GFP_KERNEL);
87 block->first_cn->len = 0;
88 block->cn = NULL;
89 block->msg = NULL;
90 block->cmd = NULL;
91 }
92}
93
94/* Early send when replies aren't bundled. */
95static void w1_netlink_check_send(struct w1_cb_block *block)
96{
97 if (!(block->request_cn.flags & W1_CN_BUNDLE) && block->cn)
98 w1_reply_make_space(block, block->maxlen);
99}
100
101/**
102 * w1_netlink_setup_msg() - prepare to write block->msg
103 * @block: block to operate on
104 * @ack: determines if cn can be reused
105 *
106 * block->cn will be setup with the correct ack, advancing if needed
107 * block->cn->len does not include space for block->msg
108 * block->msg advances but remains uninitialized
109 */
110static void w1_netlink_setup_msg(struct w1_cb_block *block, u32 ack)
111{
112 if (block->cn && block->cn->ack == ack) {
113 block->msg = (struct w1_netlink_msg *)(block->cn->data + block->cn->len);
114 } else {
115 /* advance or set to data */
116 if (block->cn)
117 block->cn = (struct cn_msg *)(block->cn->data +
118 block->cn->len);
119 else
120 block->cn = block->first_cn;
121
122 memcpy(block->cn, &block->request_cn, sizeof(*block->cn));
123 block->cn->len = 0;
124 block->cn->ack = ack;
125 block->msg = (struct w1_netlink_msg *)block->cn->data;
126 }
127}
128
129/* Append cmd to msg, include cmd->data as well. This is because
130 * any following data goes with the command and in the case of a read is
131 * the results.
132 */
133static void w1_netlink_queue_cmd(struct w1_cb_block *block,
134 struct w1_netlink_cmd *cmd)
135{
136 u32 space;
137 w1_reply_make_space(block, sizeof(struct cn_msg) +
138 sizeof(struct w1_netlink_msg) + sizeof(*cmd) + cmd->len);
139
140 /* There's a status message sent after each command, so no point
141 * in trying to bundle this cmd after an existing one, because
142 * there won't be one. Allocate and copy over a new cn_msg.
143 */
144 w1_netlink_setup_msg(block, block->request_cn.seq + 1);
145 memcpy(block->msg, block->cur_msg, sizeof(*block->msg));
146 block->cn->len += sizeof(*block->msg);
147 block->msg->len = 0;
148 block->cmd = (struct w1_netlink_cmd *)(block->msg->data);
149
150 space = sizeof(*cmd) + cmd->len;
151 if (block->cmd != cmd)
152 memcpy(block->cmd, cmd, space);
153 block->cn->len += space;
154 block->msg->len += space;
155}
156
157/* Append req_msg and req_cmd, no other commands and no data from req_cmd are
158 * copied.
159 */
160static void w1_netlink_queue_status(struct w1_cb_block *block,
161 struct w1_netlink_msg *req_msg, struct w1_netlink_cmd *req_cmd,
162 int error)
163{
164 u16 space = sizeof(struct cn_msg) + sizeof(*req_msg) + sizeof(*req_cmd);
165 w1_reply_make_space(block, space);
166 w1_netlink_setup_msg(block, block->request_cn.ack);
167
168 memcpy(block->msg, req_msg, sizeof(*req_msg));
169 block->cn->len += sizeof(*req_msg);
170 block->msg->len = 0;
171 block->msg->status = (u8)-error;
172 if (req_cmd) {
173 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)block->msg->data;
174 memcpy(cmd, req_cmd, sizeof(*cmd));
175 block->cn->len += sizeof(*cmd);
176 block->msg->len += sizeof(*cmd);
177 cmd->len = 0;
178 }
179 w1_netlink_check_send(block);
180}
181
182/**
183 * w1_netlink_send_error() - sends the error message now
184 * @cn: original cn_msg
185 * @msg: original w1_netlink_msg
186 * @portid: where to send it
187 * @error: error status
188 *
189 * Use when a block isn't available to queue the message to and cn, msg
190 * might not be contiguous.
191 */
192static void w1_netlink_send_error(struct cn_msg *cn, struct w1_netlink_msg *msg,
193 int portid, int error)
194{
195 struct {
196 struct cn_msg cn;
197 struct w1_netlink_msg msg;
198 } packet;
199 memcpy(&packet.cn, cn, sizeof(packet.cn));
200 memcpy(&packet.msg, msg, sizeof(packet.msg));
201 packet.cn.len = sizeof(packet.msg);
202 packet.msg.len = 0;
203 packet.msg.status = (u8)-error;
204 cn_netlink_send(&packet.cn, portid, 0, GFP_KERNEL);
205}
206
207/**
208 * w1_netlink_send() - sends w1 netlink notifications
209 * @dev: w1_master the even is associated with or for
210 * @msg: w1_netlink_msg message to be sent
211 *
212 * This are notifications generated from the kernel.
213 */
214void w1_netlink_send(struct w1_master *dev, struct w1_netlink_msg *msg)
215{
216 struct {
217 struct cn_msg cn;
218 struct w1_netlink_msg msg;
219 } packet;
220 memset(&packet, 0, sizeof(packet));
221
222 packet.cn.id.idx = CN_W1_IDX;
223 packet.cn.id.val = CN_W1_VAL;
224
225 packet.cn.seq = dev->seq++;
226 packet.cn.len = sizeof(*msg);
227
228 memcpy(&packet.msg, msg, sizeof(*msg));
229 packet.msg.len = 0;
230
231 cn_netlink_send(&packet.cn, 0, 0, GFP_KERNEL);
232}
233
234static void w1_send_slave(struct w1_master *dev, u64 rn)
235{
236 struct w1_cb_block *block = dev->priv;
237 struct w1_netlink_cmd *cache_cmd = block->cmd;
238 u64 *data;
239
240 w1_reply_make_space(block, sizeof(*data));
241
242 /* Add cmd back if the packet was sent */
243 if (!block->cmd) {
244 cache_cmd->len = 0;
245 w1_netlink_queue_cmd(block, cache_cmd);
246 }
247
248 data = (u64 *)(block->cmd->data + block->cmd->len);
249
250 *data = rn;
251 block->cn->len += sizeof(*data);
252 block->msg->len += sizeof(*data);
253 block->cmd->len += sizeof(*data);
254}
255
256static void w1_found_send_slave(struct w1_master *dev, u64 rn)
257{
258 /* update kernel slave list */
259 w1_slave_found(dev, rn);
260
261 w1_send_slave(dev, rn);
262}
263
264/* Get the current slave list, or search (with or without alarm) */
265static int w1_get_slaves(struct w1_master *dev, struct w1_netlink_cmd *req_cmd)
266{
267 struct w1_slave *sl;
268
269 req_cmd->len = 0;
270 w1_netlink_queue_cmd(dev->priv, req_cmd);
271
272 if (req_cmd->cmd == W1_CMD_LIST_SLAVES) {
273 u64 rn;
274 mutex_lock(&dev->list_mutex);
275 list_for_each_entry(sl, &dev->slist, w1_slave_entry) {
276 memcpy(&rn, &sl->reg_num, sizeof(rn));
277 w1_send_slave(dev, rn);
278 }
279 mutex_unlock(&dev->list_mutex);
280 } else {
281 w1_search_process_cb(dev, req_cmd->cmd == W1_CMD_ALARM_SEARCH ?
282 W1_ALARM_SEARCH : W1_SEARCH, w1_found_send_slave);
283 }
284
285 return 0;
286}
287
288static int w1_process_command_io(struct w1_master *dev,
289 struct w1_netlink_cmd *cmd)
290{
291 int err = 0;
292
293 switch (cmd->cmd) {
294 case W1_CMD_TOUCH:
295 w1_touch_block(dev, cmd->data, cmd->len);
296 w1_netlink_queue_cmd(dev->priv, cmd);
297 break;
298 case W1_CMD_READ:
299 w1_read_block(dev, cmd->data, cmd->len);
300 w1_netlink_queue_cmd(dev->priv, cmd);
301 break;
302 case W1_CMD_WRITE:
303 w1_write_block(dev, cmd->data, cmd->len);
304 break;
305 default:
306 err = -EINVAL;
307 break;
308 }
309
310 return err;
311}
312
313static int w1_process_command_addremove(struct w1_master *dev,
314 struct w1_netlink_cmd *cmd)
315{
316 struct w1_slave *sl;
317 int err = 0;
318 struct w1_reg_num *id;
319
320 if (cmd->len != sizeof(*id))
321 return -EINVAL;
322
323 id = (struct w1_reg_num *)cmd->data;
324
325 sl = w1_slave_search_device(dev, id);
326 switch (cmd->cmd) {
327 case W1_CMD_SLAVE_ADD:
328 if (sl)
329 err = -EINVAL;
330 else
331 err = w1_attach_slave_device(dev, id);
332 break;
333 case W1_CMD_SLAVE_REMOVE:
334 if (sl)
335 w1_slave_detach(sl);
336 else
337 err = -EINVAL;
338 break;
339 default:
340 err = -EINVAL;
341 break;
342 }
343
344 return err;
345}
346
347static int w1_process_command_master(struct w1_master *dev,
348 struct w1_netlink_cmd *req_cmd)
349{
350 int err = -EINVAL;
351
352 /* drop bus_mutex for search (does it's own locking), and add/remove
353 * which doesn't use the bus
354 */
355 switch (req_cmd->cmd) {
356 case W1_CMD_SEARCH:
357 case W1_CMD_ALARM_SEARCH:
358 case W1_CMD_LIST_SLAVES:
359 mutex_unlock(&dev->bus_mutex);
360 err = w1_get_slaves(dev, req_cmd);
361 mutex_lock(&dev->bus_mutex);
362 break;
363 case W1_CMD_READ:
364 case W1_CMD_WRITE:
365 case W1_CMD_TOUCH:
366 err = w1_process_command_io(dev, req_cmd);
367 break;
368 case W1_CMD_RESET:
369 err = w1_reset_bus(dev);
370 break;
371 case W1_CMD_SLAVE_ADD:
372 case W1_CMD_SLAVE_REMOVE:
373 mutex_unlock(&dev->bus_mutex);
374 mutex_lock(&dev->mutex);
375 err = w1_process_command_addremove(dev, req_cmd);
376 mutex_unlock(&dev->mutex);
377 mutex_lock(&dev->bus_mutex);
378 break;
379 default:
380 err = -EINVAL;
381 break;
382 }
383
384 return err;
385}
386
387static int w1_process_command_slave(struct w1_slave *sl,
388 struct w1_netlink_cmd *cmd)
389{
390 dev_dbg(&sl->master->dev, "%s: %02x.%012llx.%02x: cmd=%02x, len=%u.\n",
391 __func__, sl->reg_num.family, (unsigned long long)sl->reg_num.id,
392 sl->reg_num.crc, cmd->cmd, cmd->len);
393
394 return w1_process_command_io(sl->master, cmd);
395}
396
397static int w1_process_command_root(struct cn_msg *req_cn, u32 portid)
398{
399 struct w1_master *dev;
400 struct cn_msg *cn;
401 struct w1_netlink_msg *msg;
402 u32 *id;
403
404 cn = kmalloc(PAGE_SIZE, GFP_KERNEL);
405 if (!cn)
406 return -ENOMEM;
407
408 cn->id.idx = CN_W1_IDX;
409 cn->id.val = CN_W1_VAL;
410
411 cn->seq = req_cn->seq;
412 cn->ack = req_cn->seq + 1;
413 cn->len = sizeof(struct w1_netlink_msg);
414 msg = (struct w1_netlink_msg *)cn->data;
415
416 msg->type = W1_LIST_MASTERS;
417 msg->status = 0;
418 msg->len = 0;
419 id = (u32 *)msg->data;
420
421 mutex_lock(&w1_mlock);
422 list_for_each_entry(dev, &w1_masters, w1_master_entry) {
423 if (cn->len + sizeof(*id) > PAGE_SIZE - sizeof(struct cn_msg)) {
424 cn_netlink_send(cn, portid, 0, GFP_KERNEL);
425 cn->len = sizeof(struct w1_netlink_msg);
426 msg->len = 0;
427 id = (u32 *)msg->data;
428 }
429
430 *id = dev->id;
431 msg->len += sizeof(*id);
432 cn->len += sizeof(*id);
433 id++;
434 }
435 cn_netlink_send(cn, portid, 0, GFP_KERNEL);
436 mutex_unlock(&w1_mlock);
437
438 kfree(cn);
439 return 0;
440}
441
442static void w1_process_cb(struct w1_master *dev, struct w1_async_cmd *async_cmd)
443{
444 struct w1_cb_node *node = container_of(async_cmd, struct w1_cb_node,
445 async);
446 u16 mlen = node->msg->len;
447 u16 len;
448 int err = 0;
449 struct w1_slave *sl = node->sl;
450 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)node->msg->data;
451
452 mutex_lock(&dev->bus_mutex);
453 dev->priv = node->block;
454 if (sl && w1_reset_select_slave(sl))
455 err = -ENODEV;
456 node->block->cur_msg = node->msg;
457
458 while (mlen && !err) {
459 if (cmd->len + sizeof(struct w1_netlink_cmd) > mlen) {
460 err = -E2BIG;
461 break;
462 }
463
464 if (sl)
465 err = w1_process_command_slave(sl, cmd);
466 else
467 err = w1_process_command_master(dev, cmd);
468 w1_netlink_check_send(node->block);
469
470 w1_netlink_queue_status(node->block, node->msg, cmd, err);
471 err = 0;
472
473 len = sizeof(*cmd) + cmd->len;
474 cmd = (struct w1_netlink_cmd *)((u8 *)cmd + len);
475 mlen -= len;
476 }
477
478 if (!cmd || err)
479 w1_netlink_queue_status(node->block, node->msg, cmd, err);
480
481 /* ref taken in w1_search_slave or w1_search_master_id when building
482 * the block
483 */
484 if (sl)
485 w1_unref_slave(sl);
486 else
487 atomic_dec(&dev->refcnt);
488 dev->priv = NULL;
489 mutex_unlock(&dev->bus_mutex);
490
491 mutex_lock(&dev->list_mutex);
492 list_del(&async_cmd->async_entry);
493 mutex_unlock(&dev->list_mutex);
494
495 w1_unref_block(node->block);
496}
497
498static void w1_list_count_cmds(struct w1_netlink_msg *msg, int *cmd_count,
499 u16 *slave_len)
500{
501 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)msg->data;
502 u16 mlen = msg->len;
503 u16 len;
504 int slave_list = 0;
505 while (mlen) {
506 if (cmd->len + sizeof(struct w1_netlink_cmd) > mlen)
507 break;
508
509 switch (cmd->cmd) {
510 case W1_CMD_SEARCH:
511 case W1_CMD_ALARM_SEARCH:
512 case W1_CMD_LIST_SLAVES:
513 ++slave_list;
514 }
515 ++*cmd_count;
516 len = sizeof(*cmd) + cmd->len;
517 cmd = (struct w1_netlink_cmd *)((u8 *)cmd + len);
518 mlen -= len;
519 }
520
521 if (slave_list) {
522 struct w1_master *dev = w1_search_master_id(msg->id.mst.id);
523 if (dev) {
524 /* Bytes, and likely an overstimate, and if it isn't
525 * the results can still be split between packets.
526 */
527 *slave_len += sizeof(struct w1_reg_num) * slave_list *
528 (dev->slave_count + dev->max_slave_count);
529 /* search incremented it */
530 atomic_dec(&dev->refcnt);
531 }
532 }
533}
534
535static void w1_cn_callback(struct cn_msg *cn, struct netlink_skb_parms *nsp)
536{
537 struct w1_netlink_msg *msg = (struct w1_netlink_msg *)(cn + 1);
538 struct w1_slave *sl;
539 struct w1_master *dev;
540 u16 msg_len;
541 u16 slave_len = 0;
542 int err = 0;
543 struct w1_cb_block *block = NULL;
544 struct w1_cb_node *node = NULL;
545 int node_count = 0;
546 int cmd_count = 0;
547
548 /* If any unknown flag is set let the application know, that way
549 * applications can detect the absence of features in kernels that
550 * don't know about them. http://lwn.net/Articles/587527/
551 */
552 if (cn->flags & ~(W1_CN_BUNDLE)) {
553 w1_netlink_send_error(cn, msg, nsp->portid, -EINVAL);
554 return;
555 }
556
557 /* Count the number of master or slave commands there are to allocate
558 * space for one cb_node each.
559 */
560 msg_len = cn->len;
561 while (msg_len && !err) {
562 if (msg->len + sizeof(struct w1_netlink_msg) > msg_len) {
563 err = -E2BIG;
564 break;
565 }
566
567 /* count messages for nodes and allocate any additional space
568 * required for slave lists
569 */
570 if (msg->type == W1_MASTER_CMD || msg->type == W1_SLAVE_CMD) {
571 ++node_count;
572 w1_list_count_cmds(msg, &cmd_count, &slave_len);
573 }
574
575 msg_len -= sizeof(struct w1_netlink_msg) + msg->len;
576 msg = (struct w1_netlink_msg *)(((u8 *)msg) +
577 sizeof(struct w1_netlink_msg) + msg->len);
578 }
579 msg = (struct w1_netlink_msg *)(cn + 1);
580 if (node_count) {
581 int size;
582 int reply_size = sizeof(*cn) + cn->len + slave_len;
583 if (cn->flags & W1_CN_BUNDLE) {
584 /* bundling duplicats some of the messages */
585 reply_size += 2 * cmd_count * (sizeof(struct cn_msg) +
586 sizeof(struct w1_netlink_msg) +
587 sizeof(struct w1_netlink_cmd));
588 }
589 reply_size = min(CONNECTOR_MAX_MSG_SIZE, reply_size);
590
591 /* allocate space for the block, a copy of the original message,
592 * one node per cmd to point into the original message,
593 * space for replies which is the original message size plus
594 * space for any list slave data and status messages
595 * cn->len doesn't include itself which is part of the block
596 * */
597 size = /* block + original message */
598 sizeof(struct w1_cb_block) + sizeof(*cn) + cn->len +
599 /* space for nodes */
600 node_count * sizeof(struct w1_cb_node) +
601 /* replies */
602 sizeof(struct cn_msg) + reply_size;
603 block = kzalloc(size, GFP_KERNEL);
604 if (!block) {
605 /* if the system is already out of memory,
606 * (A) will this work, and (B) would it be better
607 * to not try?
608 */
609 w1_netlink_send_error(cn, msg, nsp->portid, -ENOMEM);
610 return;
611 }
612 atomic_set(&block->refcnt, 1);
613 block->portid = nsp->portid;
614 memcpy(&block->request_cn, cn, sizeof(*cn) + cn->len);
615 node = (struct w1_cb_node *)(block->request_cn.data + cn->len);
616
617 /* Sneeky, when not bundling, reply_size is the allocated space
618 * required for the reply, cn_msg isn't part of maxlen so
619 * it should be reply_size - sizeof(struct cn_msg), however
620 * when checking if there is enough space, w1_reply_make_space
621 * is called with the full message size including cn_msg,
622 * because it isn't known at that time if an additional cn_msg
623 * will need to be allocated. So an extra cn_msg is added
624 * above in "size".
625 */
626 block->maxlen = reply_size;
627 block->first_cn = (struct cn_msg *)(node + node_count);
628 memset(block->first_cn, 0, sizeof(*block->first_cn));
629 }
630
631 msg_len = cn->len;
632 while (msg_len && !err) {
633
634 dev = NULL;
635 sl = NULL;
636
637 if (msg->len + sizeof(struct w1_netlink_msg) > msg_len) {
638 err = -E2BIG;
639 break;
640 }
641
642 /* execute on this thread, no need to process later */
643 if (msg->type == W1_LIST_MASTERS) {
644 err = w1_process_command_root(cn, nsp->portid);
645 goto out_cont;
646 }
647
648 /* All following message types require additional data,
649 * check here before references are taken.
650 */
651 if (!msg->len) {
652 err = -EPROTO;
653 goto out_cont;
654 }
655
656 /* both search calls take references */
657 if (msg->type == W1_MASTER_CMD) {
658 dev = w1_search_master_id(msg->id.mst.id);
659 } else if (msg->type == W1_SLAVE_CMD) {
660 sl = w1_search_slave((struct w1_reg_num *)msg->id.id);
661 if (sl)
662 dev = sl->master;
663 } else {
664 pr_notice("%s: cn: %x.%x, wrong type: %u, len: %u.\n",
665 __func__, cn->id.idx, cn->id.val,
666 msg->type, msg->len);
667 err = -EPROTO;
668 goto out_cont;
669 }
670
671 if (!dev) {
672 err = -ENODEV;
673 goto out_cont;
674 }
675
676 err = 0;
677
678 atomic_inc(&block->refcnt);
679 node->async.cb = w1_process_cb;
680 node->block = block;
681 node->msg = (struct w1_netlink_msg *)((u8 *)&block->request_cn +
682 (size_t)((u8 *)msg - (u8 *)cn));
683 node->sl = sl;
684 node->dev = dev;
685
686 mutex_lock(&dev->list_mutex);
687 list_add_tail(&node->async.async_entry, &dev->async_list);
688 wake_up_process(dev->thread);
689 mutex_unlock(&dev->list_mutex);
690 ++node;
691
692out_cont:
693 /* Can't queue because that modifies block and another
694 * thread could be processing the messages by now and
695 * there isn't a lock, send directly.
696 */
697 if (err)
698 w1_netlink_send_error(cn, msg, nsp->portid, err);
699 msg_len -= sizeof(struct w1_netlink_msg) + msg->len;
700 msg = (struct w1_netlink_msg *)(((u8 *)msg) +
701 sizeof(struct w1_netlink_msg) + msg->len);
702
703 /*
704 * Let's allow requests for nonexisting devices.
705 */
706 if (err == -ENODEV)
707 err = 0;
708 }
709 if (block)
710 w1_unref_block(block);
711}
712
713int w1_init_netlink(void)
714{
715 struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL};
716
717 return cn_add_callback(&w1_id, "w1", &w1_cn_callback);
718}
719
720void w1_fini_netlink(void)
721{
722 struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL};
723
724 cn_del_callback(&w1_id);
725}
726#else
727void w1_netlink_send(struct w1_master *dev, struct w1_netlink_msg *cn)
728{
729}
730
731int w1_init_netlink(void)
732{
733 return 0;
734}
735
736void w1_fini_netlink(void)
737{
738}
739#endif
1/*
2 * w1_netlink.c
3 *
4 * Copyright (c) 2003 Evgeniy Polyakov <zbr@ioremap.net>
5 *
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License as published by
9 * the Free Software Foundation; either version 2 of the License, or
10 * (at your option) any later version.
11 *
12 * This program is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 * GNU General Public License for more details.
16 *
17 * You should have received a copy of the GNU General Public License
18 * along with this program; if not, write to the Free Software
19 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
20 */
21
22#include <linux/slab.h>
23#include <linux/skbuff.h>
24#include <linux/netlink.h>
25#include <linux/connector.h>
26
27#include "w1.h"
28#include "w1_log.h"
29#include "w1_netlink.h"
30
31#if defined(CONFIG_W1_CON) && (defined(CONFIG_CONNECTOR) || (defined(CONFIG_CONNECTOR_MODULE) && defined(CONFIG_W1_MODULE)))
32void w1_netlink_send(struct w1_master *dev, struct w1_netlink_msg *msg)
33{
34 char buf[sizeof(struct cn_msg) + sizeof(struct w1_netlink_msg)];
35 struct cn_msg *m = (struct cn_msg *)buf;
36 struct w1_netlink_msg *w = (struct w1_netlink_msg *)(m+1);
37
38 memset(buf, 0, sizeof(buf));
39
40 m->id.idx = CN_W1_IDX;
41 m->id.val = CN_W1_VAL;
42
43 m->seq = dev->seq++;
44 m->len = sizeof(struct w1_netlink_msg);
45
46 memcpy(w, msg, sizeof(struct w1_netlink_msg));
47
48 cn_netlink_send(m, 0, GFP_KERNEL);
49}
50
51static void w1_send_slave(struct w1_master *dev, u64 rn)
52{
53 struct cn_msg *msg = dev->priv;
54 struct w1_netlink_msg *hdr = (struct w1_netlink_msg *)(msg + 1);
55 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)(hdr + 1);
56 int avail;
57
58 /* update kernel slave list */
59 w1_slave_found(dev, rn);
60
61 avail = dev->priv_size - cmd->len;
62
63 if (avail > 8) {
64 u64 *data = (void *)(cmd + 1) + cmd->len;
65
66 *data = rn;
67 cmd->len += 8;
68 hdr->len += 8;
69 msg->len += 8;
70 return;
71 }
72
73 msg->ack++;
74 cn_netlink_send(msg, 0, GFP_KERNEL);
75
76 msg->len = sizeof(struct w1_netlink_msg) + sizeof(struct w1_netlink_cmd);
77 hdr->len = sizeof(struct w1_netlink_cmd);
78 cmd->len = 0;
79}
80
81static int w1_process_search_command(struct w1_master *dev, struct cn_msg *msg,
82 unsigned int avail)
83{
84 struct w1_netlink_msg *hdr = (struct w1_netlink_msg *)(msg + 1);
85 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)(hdr + 1);
86 int search_type = (cmd->cmd == W1_CMD_ALARM_SEARCH)?W1_ALARM_SEARCH:W1_SEARCH;
87
88 dev->priv = msg;
89 dev->priv_size = avail;
90
91 w1_search_process_cb(dev, search_type, w1_send_slave);
92
93 msg->ack = 0;
94 cn_netlink_send(msg, 0, GFP_KERNEL);
95
96 dev->priv = NULL;
97 dev->priv_size = 0;
98
99 return 0;
100}
101
102static int w1_send_read_reply(struct cn_msg *msg, struct w1_netlink_msg *hdr,
103 struct w1_netlink_cmd *cmd)
104{
105 void *data;
106 struct w1_netlink_msg *h;
107 struct w1_netlink_cmd *c;
108 struct cn_msg *cm;
109 int err;
110
111 data = kzalloc(sizeof(struct cn_msg) +
112 sizeof(struct w1_netlink_msg) +
113 sizeof(struct w1_netlink_cmd) +
114 cmd->len, GFP_KERNEL);
115 if (!data)
116 return -ENOMEM;
117
118 cm = (struct cn_msg *)(data);
119 h = (struct w1_netlink_msg *)(cm + 1);
120 c = (struct w1_netlink_cmd *)(h + 1);
121
122 memcpy(cm, msg, sizeof(struct cn_msg));
123 memcpy(h, hdr, sizeof(struct w1_netlink_msg));
124 memcpy(c, cmd, sizeof(struct w1_netlink_cmd));
125
126 cm->ack = msg->seq+1;
127 cm->len = sizeof(struct w1_netlink_msg) +
128 sizeof(struct w1_netlink_cmd) + cmd->len;
129
130 h->len = sizeof(struct w1_netlink_cmd) + cmd->len;
131
132 memcpy(c->data, cmd->data, c->len);
133
134 err = cn_netlink_send(cm, 0, GFP_KERNEL);
135
136 kfree(data);
137
138 return err;
139}
140
141static int w1_process_command_io(struct w1_master *dev, struct cn_msg *msg,
142 struct w1_netlink_msg *hdr, struct w1_netlink_cmd *cmd)
143{
144 int err = 0;
145
146 switch (cmd->cmd) {
147 case W1_CMD_TOUCH:
148 w1_touch_block(dev, cmd->data, cmd->len);
149 w1_send_read_reply(msg, hdr, cmd);
150 break;
151 case W1_CMD_READ:
152 w1_read_block(dev, cmd->data, cmd->len);
153 w1_send_read_reply(msg, hdr, cmd);
154 break;
155 case W1_CMD_WRITE:
156 w1_write_block(dev, cmd->data, cmd->len);
157 break;
158 default:
159 err = -EINVAL;
160 break;
161 }
162
163 return err;
164}
165
166static int w1_process_command_master(struct w1_master *dev, struct cn_msg *req_msg,
167 struct w1_netlink_msg *req_hdr, struct w1_netlink_cmd *req_cmd)
168{
169 int err = -EINVAL;
170 struct cn_msg *msg;
171 struct w1_netlink_msg *hdr;
172 struct w1_netlink_cmd *cmd;
173
174 msg = kzalloc(PAGE_SIZE, GFP_KERNEL);
175 if (!msg)
176 return -ENOMEM;
177
178 msg->id = req_msg->id;
179 msg->seq = req_msg->seq;
180 msg->ack = 0;
181 msg->len = sizeof(struct w1_netlink_msg) + sizeof(struct w1_netlink_cmd);
182
183 hdr = (struct w1_netlink_msg *)(msg + 1);
184 cmd = (struct w1_netlink_cmd *)(hdr + 1);
185
186 hdr->type = W1_MASTER_CMD;
187 hdr->id = req_hdr->id;
188 hdr->len = sizeof(struct w1_netlink_cmd);
189
190 cmd->cmd = req_cmd->cmd;
191 cmd->len = 0;
192
193 switch (cmd->cmd) {
194 case W1_CMD_SEARCH:
195 case W1_CMD_ALARM_SEARCH:
196 err = w1_process_search_command(dev, msg,
197 PAGE_SIZE - msg->len - sizeof(struct cn_msg));
198 break;
199 case W1_CMD_READ:
200 case W1_CMD_WRITE:
201 case W1_CMD_TOUCH:
202 err = w1_process_command_io(dev, req_msg, req_hdr, req_cmd);
203 break;
204 case W1_CMD_RESET:
205 err = w1_reset_bus(dev);
206 break;
207 default:
208 err = -EINVAL;
209 break;
210 }
211
212 kfree(msg);
213 return err;
214}
215
216static int w1_process_command_slave(struct w1_slave *sl, struct cn_msg *msg,
217 struct w1_netlink_msg *hdr, struct w1_netlink_cmd *cmd)
218{
219 dev_dbg(&sl->master->dev, "%s: %02x.%012llx.%02x: cmd=%02x, len=%u.\n",
220 __func__, sl->reg_num.family, (unsigned long long)sl->reg_num.id,
221 sl->reg_num.crc, cmd->cmd, cmd->len);
222
223 return w1_process_command_io(sl->master, msg, hdr, cmd);
224}
225
226static int w1_process_command_root(struct cn_msg *msg, struct w1_netlink_msg *mcmd)
227{
228 struct w1_master *m;
229 struct cn_msg *cn;
230 struct w1_netlink_msg *w;
231 u32 *id;
232
233 if (mcmd->type != W1_LIST_MASTERS) {
234 printk(KERN_NOTICE "%s: msg: %x.%x, wrong type: %u, len: %u.\n",
235 __func__, msg->id.idx, msg->id.val, mcmd->type, mcmd->len);
236 return -EPROTO;
237 }
238
239 cn = kmalloc(PAGE_SIZE, GFP_KERNEL);
240 if (!cn)
241 return -ENOMEM;
242
243 cn->id.idx = CN_W1_IDX;
244 cn->id.val = CN_W1_VAL;
245
246 cn->seq = msg->seq;
247 cn->ack = 1;
248 cn->len = sizeof(struct w1_netlink_msg);
249 w = (struct w1_netlink_msg *)(cn + 1);
250
251 w->type = W1_LIST_MASTERS;
252 w->status = 0;
253 w->len = 0;
254 id = (u32 *)(w + 1);
255
256 mutex_lock(&w1_mlock);
257 list_for_each_entry(m, &w1_masters, w1_master_entry) {
258 if (cn->len + sizeof(*id) > PAGE_SIZE - sizeof(struct cn_msg)) {
259 cn_netlink_send(cn, 0, GFP_KERNEL);
260 cn->ack++;
261 cn->len = sizeof(struct w1_netlink_msg);
262 w->len = 0;
263 id = (u32 *)(w + 1);
264 }
265
266 *id = m->id;
267 w->len += sizeof(*id);
268 cn->len += sizeof(*id);
269 id++;
270 }
271 cn->ack = 0;
272 cn_netlink_send(cn, 0, GFP_KERNEL);
273 mutex_unlock(&w1_mlock);
274
275 kfree(cn);
276 return 0;
277}
278
279static int w1_netlink_send_error(struct cn_msg *rcmsg, struct w1_netlink_msg *rmsg,
280 struct w1_netlink_cmd *rcmd, int error)
281{
282 struct cn_msg *cmsg;
283 struct w1_netlink_msg *msg;
284 struct w1_netlink_cmd *cmd;
285
286 cmsg = kzalloc(sizeof(*msg) + sizeof(*cmd) + sizeof(*cmsg), GFP_KERNEL);
287 if (!cmsg)
288 return -ENOMEM;
289
290 msg = (struct w1_netlink_msg *)(cmsg + 1);
291 cmd = (struct w1_netlink_cmd *)(msg + 1);
292
293 memcpy(cmsg, rcmsg, sizeof(*cmsg));
294 cmsg->len = sizeof(*msg);
295
296 memcpy(msg, rmsg, sizeof(*msg));
297 msg->len = 0;
298 msg->status = (short)-error;
299
300 if (rcmd) {
301 memcpy(cmd, rcmd, sizeof(*cmd));
302 cmd->len = 0;
303 msg->len += sizeof(*cmd);
304 cmsg->len += sizeof(*cmd);
305 }
306
307 error = cn_netlink_send(cmsg, 0, GFP_KERNEL);
308 kfree(cmsg);
309
310 return error;
311}
312
313static void w1_cn_callback(struct cn_msg *msg, struct netlink_skb_parms *nsp)
314{
315 struct w1_netlink_msg *m = (struct w1_netlink_msg *)(msg + 1);
316 struct w1_netlink_cmd *cmd;
317 struct w1_slave *sl;
318 struct w1_master *dev;
319 int err = 0;
320
321 while (msg->len && !err) {
322 struct w1_reg_num id;
323 u16 mlen = m->len;
324 u8 *cmd_data = m->data;
325
326 dev = NULL;
327 sl = NULL;
328 cmd = NULL;
329
330 memcpy(&id, m->id.id, sizeof(id));
331#if 0
332 printk("%s: %02x.%012llx.%02x: type=%02x, len=%u.\n",
333 __func__, id.family, (unsigned long long)id.id, id.crc, m->type, m->len);
334#endif
335 if (m->len + sizeof(struct w1_netlink_msg) > msg->len) {
336 err = -E2BIG;
337 break;
338 }
339
340 if (m->type == W1_MASTER_CMD) {
341 dev = w1_search_master_id(m->id.mst.id);
342 } else if (m->type == W1_SLAVE_CMD) {
343 sl = w1_search_slave(&id);
344 if (sl)
345 dev = sl->master;
346 } else {
347 err = w1_process_command_root(msg, m);
348 goto out_cont;
349 }
350
351 if (!dev) {
352 err = -ENODEV;
353 goto out_cont;
354 }
355
356 err = 0;
357 if (!mlen)
358 goto out_cont;
359
360 mutex_lock(&dev->mutex);
361
362 if (sl && w1_reset_select_slave(sl)) {
363 err = -ENODEV;
364 goto out_up;
365 }
366
367 while (mlen) {
368 cmd = (struct w1_netlink_cmd *)cmd_data;
369
370 if (cmd->len + sizeof(struct w1_netlink_cmd) > mlen) {
371 err = -E2BIG;
372 break;
373 }
374
375 if (sl)
376 err = w1_process_command_slave(sl, msg, m, cmd);
377 else
378 err = w1_process_command_master(dev, msg, m, cmd);
379
380 w1_netlink_send_error(msg, m, cmd, err);
381 err = 0;
382
383 cmd_data += cmd->len + sizeof(struct w1_netlink_cmd);
384 mlen -= cmd->len + sizeof(struct w1_netlink_cmd);
385 }
386out_up:
387 atomic_dec(&dev->refcnt);
388 if (sl)
389 atomic_dec(&sl->refcnt);
390 mutex_unlock(&dev->mutex);
391out_cont:
392 if (!cmd || err)
393 w1_netlink_send_error(msg, m, cmd, err);
394 msg->len -= sizeof(struct w1_netlink_msg) + m->len;
395 m = (struct w1_netlink_msg *)(((u8 *)m) + sizeof(struct w1_netlink_msg) + m->len);
396
397 /*
398 * Let's allow requests for nonexisting devices.
399 */
400 if (err == -ENODEV)
401 err = 0;
402 }
403}
404
405int w1_init_netlink(void)
406{
407 struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL};
408
409 return cn_add_callback(&w1_id, "w1", &w1_cn_callback);
410}
411
412void w1_fini_netlink(void)
413{
414 struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL};
415
416 cn_del_callback(&w1_id);
417}
418#else
419void w1_netlink_send(struct w1_master *dev, struct w1_netlink_msg *msg)
420{
421}
422
423int w1_init_netlink(void)
424{
425 return 0;
426}
427
428void w1_fini_netlink(void)
429{
430}
431#endif