2 * Copyright (c) 2003 Evgeniy Polyakov <zbr@ioremap.net>
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License as published by
6 * the Free Software Foundation; either version 2 of the License, or
7 * (at your option) any later version.
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
15 #include <linux/slab.h>
16 #include <linux/skbuff.h>
17 #include <linux/netlink.h>
18 #include <linux/connector.h>
20 #include "w1_internal.h"
21 #include "w1_netlink.h"
23 #if defined(CONFIG_W1_CON) && (defined(CONFIG_CONNECTOR) || (defined(CONFIG_CONNECTOR_MODULE) && defined(CONFIG_W1_MODULE)))
25 /* Bundle together everything required to process a request in one memory
30 u32 portid; /* Sending process port ID */
31 /* maximum value for first_cn->len */
33 /* pointers to building up the reply message */
34 struct cn_msg *first_cn; /* fixed once the structure is populated */
35 struct cn_msg *cn; /* advances as cn_msg is appeneded */
36 struct w1_netlink_msg *msg; /* advances as w1_netlink_msg is appened */
37 struct w1_netlink_cmd *cmd; /* advances as cmds are appened */
38 struct w1_netlink_msg *cur_msg; /* currently message being processed */
39 /* copy of the original request follows */
40 struct cn_msg request_cn;
41 /* followed by variable length:
42 * cn_msg, data (w1_netlink_msg and w1_netlink_cmd)
43 * one or more struct w1_cb_node
44 * reply first_cn, data (w1_netlink_msg and w1_netlink_cmd)
48 struct w1_async_cmd async;
49 /* pointers within w1_cb_block and cn data */
50 struct w1_cb_block *block;
51 struct w1_netlink_msg *msg;
53 struct w1_master *dev;
57 * w1_reply_len() - calculate current reply length, compare to maxlen
58 * @block: block to calculate
60 * Calculates the current message length including possible multiple
61 * cn_msg and data, excludes the first sizeof(struct cn_msg). Direclty
62 * compariable to maxlen and usable to send the message.
64 static u16 w1_reply_len(struct w1_cb_block *block)
68 return (u8 *)block->cn - (u8 *)block->first_cn + block->cn->len;
71 static void w1_unref_block(struct w1_cb_block *block)
73 if (atomic_sub_return(1, &block->refcnt) == 0) {
74 u16 len = w1_reply_len(block);
76 cn_netlink_send_mult(block->first_cn, len,
77 block->portid, 0, GFP_KERNEL);
84 * w1_reply_make_space() - send message if needed to make space
85 * @block: block to make space on
86 * @space: how many bytes requested
88 * Verify there is enough room left for the caller to add "space" bytes to the
89 * message, if there isn't send the message and reset.
91 static void w1_reply_make_space(struct w1_cb_block *block, u16 space)
93 u16 len = w1_reply_len(block);
94 if (len + space >= block->maxlen) {
95 cn_netlink_send_mult(block->first_cn, len, block->portid, 0, GFP_KERNEL);
96 block->first_cn->len = 0;
103 /* Early send when replies aren't bundled. */
104 static void w1_netlink_check_send(struct w1_cb_block *block)
106 if (!(block->request_cn.flags & W1_CN_BUNDLE) && block->cn)
107 w1_reply_make_space(block, block->maxlen);
111 * w1_netlink_setup_msg() - prepare to write block->msg
112 * @block: block to operate on
113 * @ack: determines if cn can be reused
115 * block->cn will be setup with the correct ack, advancing if needed
116 * block->cn->len does not include space for block->msg
117 * block->msg advances but remains uninitialized
119 static void w1_netlink_setup_msg(struct w1_cb_block *block, u32 ack)
121 if (block->cn && block->cn->ack == ack) {
122 block->msg = (struct w1_netlink_msg *)(block->cn->data + block->cn->len);
124 /* advance or set to data */
126 block->cn = (struct cn_msg *)(block->cn->data +
129 block->cn = block->first_cn;
131 memcpy(block->cn, &block->request_cn, sizeof(*block->cn));
133 block->cn->ack = ack;
134 block->msg = (struct w1_netlink_msg *)block->cn->data;
138 /* Append cmd to msg, include cmd->data as well. This is because
139 * any following data goes with the command and in the case of a read is
142 static void w1_netlink_queue_cmd(struct w1_cb_block *block,
143 struct w1_netlink_cmd *cmd)
146 w1_reply_make_space(block, sizeof(struct cn_msg) +
147 sizeof(struct w1_netlink_msg) + sizeof(*cmd) + cmd->len);
149 /* There's a status message sent after each command, so no point
150 * in trying to bundle this cmd after an existing one, because
151 * there won't be one. Allocate and copy over a new cn_msg.
153 w1_netlink_setup_msg(block, block->request_cn.seq + 1);
154 memcpy(block->msg, block->cur_msg, sizeof(*block->msg));
155 block->cn->len += sizeof(*block->msg);
157 block->cmd = (struct w1_netlink_cmd *)(block->msg->data);
159 space = sizeof(*cmd) + cmd->len;
160 if (block->cmd != cmd)
161 memcpy(block->cmd, cmd, space);
162 block->cn->len += space;
163 block->msg->len += space;
166 /* Append req_msg and req_cmd, no other commands and no data from req_cmd are
169 static void w1_netlink_queue_status(struct w1_cb_block *block,
170 struct w1_netlink_msg *req_msg, struct w1_netlink_cmd *req_cmd,
173 u16 space = sizeof(struct cn_msg) + sizeof(*req_msg) + sizeof(*req_cmd);
174 w1_reply_make_space(block, space);
175 w1_netlink_setup_msg(block, block->request_cn.ack);
177 memcpy(block->msg, req_msg, sizeof(*req_msg));
178 block->cn->len += sizeof(*req_msg);
180 block->msg->status = (u8)-error;
182 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)block->msg->data;
183 memcpy(cmd, req_cmd, sizeof(*cmd));
184 block->cn->len += sizeof(*cmd);
185 block->msg->len += sizeof(*cmd);
188 w1_netlink_check_send(block);
192 * w1_netlink_send_error() - sends the error message now
193 * @cn: original cn_msg
194 * @msg: original w1_netlink_msg
195 * @portid: where to send it
196 * @error: error status
198 * Use when a block isn't available to queue the message to and cn, msg
199 * might not be contiguous.
201 static void w1_netlink_send_error(struct cn_msg *cn, struct w1_netlink_msg *msg,
202 int portid, int error)
206 struct w1_netlink_msg msg;
208 memcpy(&packet.cn, cn, sizeof(packet.cn));
209 memcpy(&packet.msg, msg, sizeof(packet.msg));
210 packet.cn.len = sizeof(packet.msg);
212 packet.msg.status = (u8)-error;
213 cn_netlink_send(&packet.cn, portid, 0, GFP_KERNEL);
217 * w1_netlink_send() - sends w1 netlink notifications
218 * @dev: w1_master the even is associated with or for
219 * @msg: w1_netlink_msg message to be sent
221 * This are notifications generated from the kernel.
223 void w1_netlink_send(struct w1_master *dev, struct w1_netlink_msg *msg)
227 struct w1_netlink_msg msg;
229 memset(&packet, 0, sizeof(packet));
231 packet.cn.id.idx = CN_W1_IDX;
232 packet.cn.id.val = CN_W1_VAL;
234 packet.cn.seq = dev->seq++;
235 packet.cn.len = sizeof(*msg);
237 memcpy(&packet.msg, msg, sizeof(*msg));
240 cn_netlink_send(&packet.cn, 0, 0, GFP_KERNEL);
243 static void w1_send_slave(struct w1_master *dev, u64 rn)
245 struct w1_cb_block *block = dev->priv;
246 struct w1_netlink_cmd *cache_cmd = block->cmd;
249 w1_reply_make_space(block, sizeof(*data));
251 /* Add cmd back if the packet was sent */
254 w1_netlink_queue_cmd(block, cache_cmd);
257 data = (u64 *)(block->cmd->data + block->cmd->len);
260 block->cn->len += sizeof(*data);
261 block->msg->len += sizeof(*data);
262 block->cmd->len += sizeof(*data);
265 static void w1_found_send_slave(struct w1_master *dev, u64 rn)
267 /* update kernel slave list */
268 w1_slave_found(dev, rn);
270 w1_send_slave(dev, rn);
273 /* Get the current slave list, or search (with or without alarm) */
274 static int w1_get_slaves(struct w1_master *dev, struct w1_netlink_cmd *req_cmd)
279 w1_netlink_queue_cmd(dev->priv, req_cmd);
281 if (req_cmd->cmd == W1_CMD_LIST_SLAVES) {
283 mutex_lock(&dev->list_mutex);
284 list_for_each_entry(sl, &dev->slist, w1_slave_entry) {
285 memcpy(&rn, &sl->reg_num, sizeof(rn));
286 w1_send_slave(dev, rn);
288 mutex_unlock(&dev->list_mutex);
290 w1_search_process_cb(dev, req_cmd->cmd == W1_CMD_ALARM_SEARCH ?
291 W1_ALARM_SEARCH : W1_SEARCH, w1_found_send_slave);
297 static int w1_process_command_io(struct w1_master *dev,
298 struct w1_netlink_cmd *cmd)
304 w1_touch_block(dev, cmd->data, cmd->len);
305 w1_netlink_queue_cmd(dev->priv, cmd);
308 w1_read_block(dev, cmd->data, cmd->len);
309 w1_netlink_queue_cmd(dev->priv, cmd);
312 w1_write_block(dev, cmd->data, cmd->len);
322 static int w1_process_command_addremove(struct w1_master *dev,
323 struct w1_netlink_cmd *cmd)
327 struct w1_reg_num *id;
329 if (cmd->len != sizeof(*id))
332 id = (struct w1_reg_num *)cmd->data;
334 sl = w1_slave_search_device(dev, id);
336 case W1_CMD_SLAVE_ADD:
340 err = w1_attach_slave_device(dev, id);
342 case W1_CMD_SLAVE_REMOVE:
356 static int w1_process_command_master(struct w1_master *dev,
357 struct w1_netlink_cmd *req_cmd)
361 /* drop bus_mutex for search (does it's own locking), and add/remove
362 * which doesn't use the bus
364 switch (req_cmd->cmd) {
366 case W1_CMD_ALARM_SEARCH:
367 case W1_CMD_LIST_SLAVES:
368 mutex_unlock(&dev->bus_mutex);
369 err = w1_get_slaves(dev, req_cmd);
370 mutex_lock(&dev->bus_mutex);
375 err = w1_process_command_io(dev, req_cmd);
378 err = w1_reset_bus(dev);
380 case W1_CMD_SLAVE_ADD:
381 case W1_CMD_SLAVE_REMOVE:
382 mutex_unlock(&dev->bus_mutex);
383 mutex_lock(&dev->mutex);
384 err = w1_process_command_addremove(dev, req_cmd);
385 mutex_unlock(&dev->mutex);
386 mutex_lock(&dev->bus_mutex);
396 static int w1_process_command_slave(struct w1_slave *sl,
397 struct w1_netlink_cmd *cmd)
399 dev_dbg(&sl->master->dev, "%s: %02x.%012llx.%02x: cmd=%02x, len=%u.\n",
400 __func__, sl->reg_num.family, (unsigned long long)sl->reg_num.id,
401 sl->reg_num.crc, cmd->cmd, cmd->len);
403 return w1_process_command_io(sl->master, cmd);
406 static int w1_process_command_root(struct cn_msg *req_cn, u32 portid)
408 struct w1_master *dev;
410 struct w1_netlink_msg *msg;
413 cn = kmalloc(PAGE_SIZE, GFP_KERNEL);
417 cn->id.idx = CN_W1_IDX;
418 cn->id.val = CN_W1_VAL;
420 cn->seq = req_cn->seq;
421 cn->ack = req_cn->seq + 1;
422 cn->len = sizeof(struct w1_netlink_msg);
423 msg = (struct w1_netlink_msg *)cn->data;
425 msg->type = W1_LIST_MASTERS;
428 id = (u32 *)msg->data;
430 mutex_lock(&w1_mlock);
431 list_for_each_entry(dev, &w1_masters, w1_master_entry) {
432 if (cn->len + sizeof(*id) > PAGE_SIZE - sizeof(struct cn_msg)) {
433 cn_netlink_send(cn, portid, 0, GFP_KERNEL);
434 cn->len = sizeof(struct w1_netlink_msg);
436 id = (u32 *)msg->data;
440 msg->len += sizeof(*id);
441 cn->len += sizeof(*id);
444 cn_netlink_send(cn, portid, 0, GFP_KERNEL);
445 mutex_unlock(&w1_mlock);
451 static void w1_process_cb(struct w1_master *dev, struct w1_async_cmd *async_cmd)
453 struct w1_cb_node *node = container_of(async_cmd, struct w1_cb_node,
455 u16 mlen = node->msg->len;
458 struct w1_slave *sl = node->sl;
459 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)node->msg->data;
461 mutex_lock(&dev->bus_mutex);
462 dev->priv = node->block;
463 if (sl && w1_reset_select_slave(sl))
465 node->block->cur_msg = node->msg;
467 while (mlen && !err) {
468 if (cmd->len + sizeof(struct w1_netlink_cmd) > mlen) {
474 err = w1_process_command_slave(sl, cmd);
476 err = w1_process_command_master(dev, cmd);
477 w1_netlink_check_send(node->block);
479 w1_netlink_queue_status(node->block, node->msg, cmd, err);
482 len = sizeof(*cmd) + cmd->len;
483 cmd = (struct w1_netlink_cmd *)((u8 *)cmd + len);
488 w1_netlink_queue_status(node->block, node->msg, cmd, err);
490 /* ref taken in w1_search_slave or w1_search_master_id when building
496 atomic_dec(&dev->refcnt);
498 mutex_unlock(&dev->bus_mutex);
500 mutex_lock(&dev->list_mutex);
501 list_del(&async_cmd->async_entry);
502 mutex_unlock(&dev->list_mutex);
504 w1_unref_block(node->block);
507 static void w1_list_count_cmds(struct w1_netlink_msg *msg, int *cmd_count,
510 struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)msg->data;
515 if (cmd->len + sizeof(struct w1_netlink_cmd) > mlen)
520 case W1_CMD_ALARM_SEARCH:
521 case W1_CMD_LIST_SLAVES:
525 len = sizeof(*cmd) + cmd->len;
526 cmd = (struct w1_netlink_cmd *)((u8 *)cmd + len);
531 struct w1_master *dev = w1_search_master_id(msg->id.mst.id);
533 /* Bytes, and likely an overstimate, and if it isn't
534 * the results can still be split between packets.
536 *slave_len += sizeof(struct w1_reg_num) * slave_list *
537 (dev->slave_count + dev->max_slave_count);
538 /* search incremented it */
539 atomic_dec(&dev->refcnt);
544 static void w1_cn_callback(struct cn_msg *cn, struct netlink_skb_parms *nsp)
546 struct w1_netlink_msg *msg = (struct w1_netlink_msg *)(cn + 1);
548 struct w1_master *dev;
552 struct w1_cb_block *block = NULL;
553 struct w1_cb_node *node = NULL;
557 /* If any unknown flag is set let the application know, that way
558 * applications can detect the absence of features in kernels that
559 * don't know about them. http://lwn.net/Articles/587527/
561 if (cn->flags & ~(W1_CN_BUNDLE)) {
562 w1_netlink_send_error(cn, msg, nsp->portid, -EINVAL);
566 /* Count the number of master or slave commands there are to allocate
567 * space for one cb_node each.
570 while (msg_len && !err) {
571 if (msg->len + sizeof(struct w1_netlink_msg) > msg_len) {
576 /* count messages for nodes and allocate any additional space
577 * required for slave lists
579 if (msg->type == W1_MASTER_CMD || msg->type == W1_SLAVE_CMD) {
581 w1_list_count_cmds(msg, &cmd_count, &slave_len);
584 msg_len -= sizeof(struct w1_netlink_msg) + msg->len;
585 msg = (struct w1_netlink_msg *)(((u8 *)msg) +
586 sizeof(struct w1_netlink_msg) + msg->len);
588 msg = (struct w1_netlink_msg *)(cn + 1);
591 int reply_size = sizeof(*cn) + cn->len + slave_len;
592 if (cn->flags & W1_CN_BUNDLE) {
593 /* bundling duplicats some of the messages */
594 reply_size += 2 * cmd_count * (sizeof(struct cn_msg) +
595 sizeof(struct w1_netlink_msg) +
596 sizeof(struct w1_netlink_cmd));
598 reply_size = min(CONNECTOR_MAX_MSG_SIZE, reply_size);
600 /* allocate space for the block, a copy of the original message,
601 * one node per cmd to point into the original message,
602 * space for replies which is the original message size plus
603 * space for any list slave data and status messages
604 * cn->len doesn't include itself which is part of the block
606 size = /* block + original message */
607 sizeof(struct w1_cb_block) + sizeof(*cn) + cn->len +
608 /* space for nodes */
609 node_count * sizeof(struct w1_cb_node) +
611 sizeof(struct cn_msg) + reply_size;
612 block = kzalloc(size, GFP_KERNEL);
614 /* if the system is already out of memory,
615 * (A) will this work, and (B) would it be better
618 w1_netlink_send_error(cn, msg, nsp->portid, -ENOMEM);
621 atomic_set(&block->refcnt, 1);
622 block->portid = nsp->portid;
623 memcpy(&block->request_cn, cn, sizeof(*cn) + cn->len);
624 node = (struct w1_cb_node *)(block->request_cn.data + cn->len);
626 /* Sneeky, when not bundling, reply_size is the allocated space
627 * required for the reply, cn_msg isn't part of maxlen so
628 * it should be reply_size - sizeof(struct cn_msg), however
629 * when checking if there is enough space, w1_reply_make_space
630 * is called with the full message size including cn_msg,
631 * because it isn't known at that time if an additional cn_msg
632 * will need to be allocated. So an extra cn_msg is added
635 block->maxlen = reply_size;
636 block->first_cn = (struct cn_msg *)(node + node_count);
637 memset(block->first_cn, 0, sizeof(*block->first_cn));
641 while (msg_len && !err) {
646 if (msg->len + sizeof(struct w1_netlink_msg) > msg_len) {
651 /* execute on this thread, no need to process later */
652 if (msg->type == W1_LIST_MASTERS) {
653 err = w1_process_command_root(cn, nsp->portid);
657 /* All following message types require additional data,
658 * check here before references are taken.
665 /* both search calls take references */
666 if (msg->type == W1_MASTER_CMD) {
667 dev = w1_search_master_id(msg->id.mst.id);
668 } else if (msg->type == W1_SLAVE_CMD) {
669 sl = w1_search_slave((struct w1_reg_num *)msg->id.id);
673 pr_notice("%s: cn: %x.%x, wrong type: %u, len: %u.\n",
674 __func__, cn->id.idx, cn->id.val,
675 msg->type, msg->len);
687 atomic_inc(&block->refcnt);
688 node->async.cb = w1_process_cb;
690 node->msg = (struct w1_netlink_msg *)((u8 *)&block->request_cn +
691 (size_t)((u8 *)msg - (u8 *)cn));
695 mutex_lock(&dev->list_mutex);
696 list_add_tail(&node->async.async_entry, &dev->async_list);
697 wake_up_process(dev->thread);
698 mutex_unlock(&dev->list_mutex);
702 /* Can't queue because that modifies block and another
703 * thread could be processing the messages by now and
704 * there isn't a lock, send directly.
707 w1_netlink_send_error(cn, msg, nsp->portid, err);
708 msg_len -= sizeof(struct w1_netlink_msg) + msg->len;
709 msg = (struct w1_netlink_msg *)(((u8 *)msg) +
710 sizeof(struct w1_netlink_msg) + msg->len);
713 * Let's allow requests for nonexisting devices.
719 w1_unref_block(block);
722 int w1_init_netlink(void)
724 struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL};
726 return cn_add_callback(&w1_id, "w1", &w1_cn_callback);
729 void w1_fini_netlink(void)
731 struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL};
733 cn_del_callback(&w1_id);
736 void w1_netlink_send(struct w1_master *dev, struct w1_netlink_msg *cn)
740 int w1_init_netlink(void)
745 void w1_fini_netlink(void)