Linux Audio

Check our new training course

Loading...
v3.1
  1/*
  2 * Copyright (c) 2005 Voltaire Inc.  All rights reserved.
  3 * Copyright (c) 2002-2005, Network Appliance, Inc. All rights reserved.
  4 * Copyright (c) 1999-2005, Mellanox Technologies, Inc. All rights reserved.
  5 * Copyright (c) 2005 Intel Corporation.  All rights reserved.
  6 *
  7 * This software is available to you under a choice of one of two
  8 * licenses.  You may choose to be licensed under the terms of the GNU
  9 * General Public License (GPL) Version 2, available from the file
 10 * COPYING in the main directory of this source tree, or the
 11 * OpenIB.org BSD license below:
 12 *
 13 *     Redistribution and use in source and binary forms, with or
 14 *     without modification, are permitted provided that the following
 15 *     conditions are met:
 16 *
 17 *      - Redistributions of source code must retain the above
 18 *        copyright notice, this list of conditions and the following
 19 *        disclaimer.
 20 *
 21 *      - Redistributions in binary form must reproduce the above
 22 *        copyright notice, this list of conditions and the following
 23 *        disclaimer in the documentation and/or other materials
 24 *        provided with the distribution.
 25 *
 26 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 27 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 28 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 29 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
 30 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
 31 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
 32 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 33 * SOFTWARE.
 34 */
 35
 36#include <linux/mutex.h>
 37#include <linux/inetdevice.h>
 38#include <linux/slab.h>
 39#include <linux/workqueue.h>
 
 40#include <net/arp.h>
 41#include <net/neighbour.h>
 42#include <net/route.h>
 43#include <net/netevent.h>
 44#include <net/addrconf.h>
 45#include <net/ip6_route.h>
 46#include <rdma/ib_addr.h>
 47
 48MODULE_AUTHOR("Sean Hefty");
 49MODULE_DESCRIPTION("IB Address Translation");
 50MODULE_LICENSE("Dual BSD/GPL");
 51
 52struct addr_req {
 53	struct list_head list;
 54	struct sockaddr_storage src_addr;
 55	struct sockaddr_storage dst_addr;
 56	struct rdma_dev_addr *addr;
 57	struct rdma_addr_client *client;
 58	void *context;
 59	void (*callback)(int status, struct sockaddr *src_addr,
 60			 struct rdma_dev_addr *addr, void *context);
 61	unsigned long timeout;
 62	int status;
 63};
 64
 65static void process_req(struct work_struct *work);
 66
 67static DEFINE_MUTEX(lock);
 68static LIST_HEAD(req_list);
 69static DECLARE_DELAYED_WORK(work, process_req);
 70static struct workqueue_struct *addr_wq;
 71
 72void rdma_addr_register_client(struct rdma_addr_client *client)
 73{
 74	atomic_set(&client->refcount, 1);
 75	init_completion(&client->comp);
 76}
 77EXPORT_SYMBOL(rdma_addr_register_client);
 78
 79static inline void put_client(struct rdma_addr_client *client)
 80{
 81	if (atomic_dec_and_test(&client->refcount))
 82		complete(&client->comp);
 83}
 84
 85void rdma_addr_unregister_client(struct rdma_addr_client *client)
 86{
 87	put_client(client);
 88	wait_for_completion(&client->comp);
 89}
 90EXPORT_SYMBOL(rdma_addr_unregister_client);
 91
 92int rdma_copy_addr(struct rdma_dev_addr *dev_addr, struct net_device *dev,
 93		     const unsigned char *dst_dev_addr)
 94{
 95	dev_addr->dev_type = dev->type;
 96	memcpy(dev_addr->src_dev_addr, dev->dev_addr, MAX_ADDR_LEN);
 97	memcpy(dev_addr->broadcast, dev->broadcast, MAX_ADDR_LEN);
 98	if (dst_dev_addr)
 99		memcpy(dev_addr->dst_dev_addr, dst_dev_addr, MAX_ADDR_LEN);
100	dev_addr->bound_dev_if = dev->ifindex;
101	return 0;
102}
103EXPORT_SYMBOL(rdma_copy_addr);
104
105int rdma_translate_ip(struct sockaddr *addr, struct rdma_dev_addr *dev_addr)
106{
107	struct net_device *dev;
108	int ret = -EADDRNOTAVAIL;
109
110	if (dev_addr->bound_dev_if) {
111		dev = dev_get_by_index(&init_net, dev_addr->bound_dev_if);
112		if (!dev)
113			return -ENODEV;
114		ret = rdma_copy_addr(dev_addr, dev, NULL);
115		dev_put(dev);
116		return ret;
117	}
118
119	switch (addr->sa_family) {
120	case AF_INET:
121		dev = ip_dev_find(&init_net,
122			((struct sockaddr_in *) addr)->sin_addr.s_addr);
123
124		if (!dev)
125			return ret;
126
127		ret = rdma_copy_addr(dev_addr, dev, NULL);
128		dev_put(dev);
129		break;
130
131#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE)
132	case AF_INET6:
133		rcu_read_lock();
134		for_each_netdev_rcu(&init_net, dev) {
135			if (ipv6_chk_addr(&init_net,
136					  &((struct sockaddr_in6 *) addr)->sin6_addr,
137					  dev, 1)) {
138				ret = rdma_copy_addr(dev_addr, dev, NULL);
139				break;
140			}
141		}
142		rcu_read_unlock();
143		break;
144#endif
145	}
146	return ret;
147}
148EXPORT_SYMBOL(rdma_translate_ip);
149
150static void set_timeout(unsigned long time)
151{
152	unsigned long delay;
153
154	cancel_delayed_work(&work);
155
156	delay = time - jiffies;
157	if ((long)delay <= 0)
158		delay = 1;
159
160	queue_delayed_work(addr_wq, &work, delay);
161}
162
163static void queue_req(struct addr_req *req)
164{
165	struct addr_req *temp_req;
166
167	mutex_lock(&lock);
168	list_for_each_entry_reverse(temp_req, &req_list, list) {
169		if (time_after_eq(req->timeout, temp_req->timeout))
170			break;
171	}
172
173	list_add(&req->list, &temp_req->list);
174
175	if (req_list.next == &req->list)
176		set_timeout(req->timeout);
177	mutex_unlock(&lock);
178}
179
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
180static int addr4_resolve(struct sockaddr_in *src_in,
181			 struct sockaddr_in *dst_in,
182			 struct rdma_dev_addr *addr)
183{
184	__be32 src_ip = src_in->sin_addr.s_addr;
185	__be32 dst_ip = dst_in->sin_addr.s_addr;
186	struct rtable *rt;
187	struct neighbour *neigh;
188	struct flowi4 fl4;
189	int ret;
190
191	memset(&fl4, 0, sizeof(fl4));
192	fl4.daddr = dst_ip;
193	fl4.saddr = src_ip;
194	fl4.flowi4_oif = addr->bound_dev_if;
195	rt = ip_route_output_key(&init_net, &fl4);
196	if (IS_ERR(rt)) {
197		ret = PTR_ERR(rt);
198		goto out;
199	}
200	src_in->sin_family = AF_INET;
201	src_in->sin_addr.s_addr = fl4.saddr;
202
203	if (rt->dst.dev->flags & IFF_LOOPBACK) {
204		ret = rdma_translate_ip((struct sockaddr *) dst_in, addr);
205		if (!ret)
206			memcpy(addr->dst_dev_addr, addr->src_dev_addr, MAX_ADDR_LEN);
207		goto put;
208	}
209
210	/* If the device does ARP internally, return 'done' */
211	if (rt->dst.dev->flags & IFF_NOARP) {
212		ret = rdma_copy_addr(addr, rt->dst.dev, NULL);
213		goto put;
214	}
215
216	neigh = neigh_lookup(&arp_tbl, &rt->rt_gateway, rt->dst.dev);
217	if (!neigh || !(neigh->nud_state & NUD_VALID)) {
218		neigh_event_send(dst_get_neighbour(&rt->dst), NULL);
219		ret = -ENODATA;
220		if (neigh)
221			goto release;
222		goto put;
223	}
224
225	ret = rdma_copy_addr(addr, neigh->dev, neigh->ha);
226release:
227	neigh_release(neigh);
228put:
229	ip_rt_put(rt);
230out:
231	return ret;
232}
233
234#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE)
235static int addr6_resolve(struct sockaddr_in6 *src_in,
236			 struct sockaddr_in6 *dst_in,
237			 struct rdma_dev_addr *addr)
238{
239	struct flowi6 fl6;
240	struct neighbour *neigh;
241	struct dst_entry *dst;
242	int ret;
243
244	memset(&fl6, 0, sizeof fl6);
245	ipv6_addr_copy(&fl6.daddr, &dst_in->sin6_addr);
246	ipv6_addr_copy(&fl6.saddr, &src_in->sin6_addr);
247	fl6.flowi6_oif = addr->bound_dev_if;
248
249	dst = ip6_route_output(&init_net, NULL, &fl6);
250	if ((ret = dst->error))
251		goto put;
252
253	if (ipv6_addr_any(&fl6.saddr)) {
254		ret = ipv6_dev_get_saddr(&init_net, ip6_dst_idev(dst)->dev,
255					 &fl6.daddr, 0, &fl6.saddr);
256		if (ret)
257			goto put;
258
259		src_in->sin6_family = AF_INET6;
260		ipv6_addr_copy(&src_in->sin6_addr, &fl6.saddr);
261	}
262
263	if (dst->dev->flags & IFF_LOOPBACK) {
264		ret = rdma_translate_ip((struct sockaddr *) dst_in, addr);
265		if (!ret)
266			memcpy(addr->dst_dev_addr, addr->src_dev_addr, MAX_ADDR_LEN);
267		goto put;
268	}
269
270	/* If the device does ARP internally, return 'done' */
271	if (dst->dev->flags & IFF_NOARP) {
272		ret = rdma_copy_addr(addr, dst->dev, NULL);
273		goto put;
274	}
275
276	neigh = dst_get_neighbour(dst);
277	if (!neigh || !(neigh->nud_state & NUD_VALID)) {
278		if (neigh)
279			neigh_event_send(neigh, NULL);
280		ret = -ENODATA;
281		goto put;
282	}
283
284	ret = rdma_copy_addr(addr, dst->dev, neigh->ha);
285put:
286	dst_release(dst);
287	return ret;
288}
289#else
290static int addr6_resolve(struct sockaddr_in6 *src_in,
291			 struct sockaddr_in6 *dst_in,
292			 struct rdma_dev_addr *addr)
293{
294	return -EADDRNOTAVAIL;
295}
296#endif
297
298static int addr_resolve(struct sockaddr *src_in,
299			struct sockaddr *dst_in,
300			struct rdma_dev_addr *addr)
301{
302	if (src_in->sa_family == AF_INET) {
303		return addr4_resolve((struct sockaddr_in *) src_in,
304			(struct sockaddr_in *) dst_in, addr);
305	} else
306		return addr6_resolve((struct sockaddr_in6 *) src_in,
307			(struct sockaddr_in6 *) dst_in, addr);
308}
309
310static void process_req(struct work_struct *work)
311{
312	struct addr_req *req, *temp_req;
313	struct sockaddr *src_in, *dst_in;
314	struct list_head done_list;
315
316	INIT_LIST_HEAD(&done_list);
317
318	mutex_lock(&lock);
319	list_for_each_entry_safe(req, temp_req, &req_list, list) {
320		if (req->status == -ENODATA) {
321			src_in = (struct sockaddr *) &req->src_addr;
322			dst_in = (struct sockaddr *) &req->dst_addr;
323			req->status = addr_resolve(src_in, dst_in, req->addr);
324			if (req->status && time_after_eq(jiffies, req->timeout))
325				req->status = -ETIMEDOUT;
326			else if (req->status == -ENODATA)
327				continue;
328		}
329		list_move_tail(&req->list, &done_list);
330	}
331
332	if (!list_empty(&req_list)) {
333		req = list_entry(req_list.next, struct addr_req, list);
334		set_timeout(req->timeout);
335	}
336	mutex_unlock(&lock);
337
338	list_for_each_entry_safe(req, temp_req, &done_list, list) {
339		list_del(&req->list);
340		req->callback(req->status, (struct sockaddr *) &req->src_addr,
341			req->addr, req->context);
342		put_client(req->client);
343		kfree(req);
344	}
345}
346
347int rdma_resolve_ip(struct rdma_addr_client *client,
348		    struct sockaddr *src_addr, struct sockaddr *dst_addr,
349		    struct rdma_dev_addr *addr, int timeout_ms,
350		    void (*callback)(int status, struct sockaddr *src_addr,
351				     struct rdma_dev_addr *addr, void *context),
352		    void *context)
353{
354	struct sockaddr *src_in, *dst_in;
355	struct addr_req *req;
356	int ret = 0;
357
358	req = kzalloc(sizeof *req, GFP_KERNEL);
359	if (!req)
360		return -ENOMEM;
361
362	src_in = (struct sockaddr *) &req->src_addr;
363	dst_in = (struct sockaddr *) &req->dst_addr;
364
365	if (src_addr) {
366		if (src_addr->sa_family != dst_addr->sa_family) {
367			ret = -EINVAL;
368			goto err;
369		}
370
371		memcpy(src_in, src_addr, ip_addr_size(src_addr));
372	} else {
373		src_in->sa_family = dst_addr->sa_family;
374	}
375
376	memcpy(dst_in, dst_addr, ip_addr_size(dst_addr));
377	req->addr = addr;
378	req->callback = callback;
379	req->context = context;
380	req->client = client;
381	atomic_inc(&client->refcount);
382
383	req->status = addr_resolve(src_in, dst_in, addr);
384	switch (req->status) {
385	case 0:
386		req->timeout = jiffies;
387		queue_req(req);
388		break;
389	case -ENODATA:
390		req->timeout = msecs_to_jiffies(timeout_ms) + jiffies;
391		queue_req(req);
392		break;
393	default:
394		ret = req->status;
395		atomic_dec(&client->refcount);
396		goto err;
397	}
398	return ret;
399err:
400	kfree(req);
401	return ret;
402}
403EXPORT_SYMBOL(rdma_resolve_ip);
404
405void rdma_addr_cancel(struct rdma_dev_addr *addr)
406{
407	struct addr_req *req, *temp_req;
408
409	mutex_lock(&lock);
410	list_for_each_entry_safe(req, temp_req, &req_list, list) {
411		if (req->addr == addr) {
412			req->status = -ECANCELED;
413			req->timeout = jiffies;
414			list_move(&req->list, &req_list);
415			set_timeout(req->timeout);
416			break;
417		}
418	}
419	mutex_unlock(&lock);
420}
421EXPORT_SYMBOL(rdma_addr_cancel);
422
423static int netevent_callback(struct notifier_block *self, unsigned long event,
424	void *ctx)
425{
426	if (event == NETEVENT_NEIGH_UPDATE) {
427		struct neighbour *neigh = ctx;
428
429		if (neigh->nud_state & NUD_VALID) {
430			set_timeout(jiffies);
431		}
432	}
433	return 0;
434}
435
436static struct notifier_block nb = {
437	.notifier_call = netevent_callback
438};
439
440static int __init addr_init(void)
441{
442	addr_wq = create_singlethread_workqueue("ib_addr");
443	if (!addr_wq)
444		return -ENOMEM;
445
446	register_netevent_notifier(&nb);
447	return 0;
448}
449
450static void __exit addr_cleanup(void)
451{
452	unregister_netevent_notifier(&nb);
453	destroy_workqueue(addr_wq);
454}
455
456module_init(addr_init);
457module_exit(addr_cleanup);
v3.5.6
  1/*
  2 * Copyright (c) 2005 Voltaire Inc.  All rights reserved.
  3 * Copyright (c) 2002-2005, Network Appliance, Inc. All rights reserved.
  4 * Copyright (c) 1999-2005, Mellanox Technologies, Inc. All rights reserved.
  5 * Copyright (c) 2005 Intel Corporation.  All rights reserved.
  6 *
  7 * This software is available to you under a choice of one of two
  8 * licenses.  You may choose to be licensed under the terms of the GNU
  9 * General Public License (GPL) Version 2, available from the file
 10 * COPYING in the main directory of this source tree, or the
 11 * OpenIB.org BSD license below:
 12 *
 13 *     Redistribution and use in source and binary forms, with or
 14 *     without modification, are permitted provided that the following
 15 *     conditions are met:
 16 *
 17 *      - Redistributions of source code must retain the above
 18 *        copyright notice, this list of conditions and the following
 19 *        disclaimer.
 20 *
 21 *      - Redistributions in binary form must reproduce the above
 22 *        copyright notice, this list of conditions and the following
 23 *        disclaimer in the documentation and/or other materials
 24 *        provided with the distribution.
 25 *
 26 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 27 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 28 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 29 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
 30 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
 31 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
 32 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 33 * SOFTWARE.
 34 */
 35
 36#include <linux/mutex.h>
 37#include <linux/inetdevice.h>
 38#include <linux/slab.h>
 39#include <linux/workqueue.h>
 40#include <linux/module.h>
 41#include <net/arp.h>
 42#include <net/neighbour.h>
 43#include <net/route.h>
 44#include <net/netevent.h>
 45#include <net/addrconf.h>
 46#include <net/ip6_route.h>
 47#include <rdma/ib_addr.h>
 48
 49MODULE_AUTHOR("Sean Hefty");
 50MODULE_DESCRIPTION("IB Address Translation");
 51MODULE_LICENSE("Dual BSD/GPL");
 52
 53struct addr_req {
 54	struct list_head list;
 55	struct sockaddr_storage src_addr;
 56	struct sockaddr_storage dst_addr;
 57	struct rdma_dev_addr *addr;
 58	struct rdma_addr_client *client;
 59	void *context;
 60	void (*callback)(int status, struct sockaddr *src_addr,
 61			 struct rdma_dev_addr *addr, void *context);
 62	unsigned long timeout;
 63	int status;
 64};
 65
 66static void process_req(struct work_struct *work);
 67
 68static DEFINE_MUTEX(lock);
 69static LIST_HEAD(req_list);
 70static DECLARE_DELAYED_WORK(work, process_req);
 71static struct workqueue_struct *addr_wq;
 72
 73void rdma_addr_register_client(struct rdma_addr_client *client)
 74{
 75	atomic_set(&client->refcount, 1);
 76	init_completion(&client->comp);
 77}
 78EXPORT_SYMBOL(rdma_addr_register_client);
 79
 80static inline void put_client(struct rdma_addr_client *client)
 81{
 82	if (atomic_dec_and_test(&client->refcount))
 83		complete(&client->comp);
 84}
 85
 86void rdma_addr_unregister_client(struct rdma_addr_client *client)
 87{
 88	put_client(client);
 89	wait_for_completion(&client->comp);
 90}
 91EXPORT_SYMBOL(rdma_addr_unregister_client);
 92
 93int rdma_copy_addr(struct rdma_dev_addr *dev_addr, struct net_device *dev,
 94		     const unsigned char *dst_dev_addr)
 95{
 96	dev_addr->dev_type = dev->type;
 97	memcpy(dev_addr->src_dev_addr, dev->dev_addr, MAX_ADDR_LEN);
 98	memcpy(dev_addr->broadcast, dev->broadcast, MAX_ADDR_LEN);
 99	if (dst_dev_addr)
100		memcpy(dev_addr->dst_dev_addr, dst_dev_addr, MAX_ADDR_LEN);
101	dev_addr->bound_dev_if = dev->ifindex;
102	return 0;
103}
104EXPORT_SYMBOL(rdma_copy_addr);
105
106int rdma_translate_ip(struct sockaddr *addr, struct rdma_dev_addr *dev_addr)
107{
108	struct net_device *dev;
109	int ret = -EADDRNOTAVAIL;
110
111	if (dev_addr->bound_dev_if) {
112		dev = dev_get_by_index(&init_net, dev_addr->bound_dev_if);
113		if (!dev)
114			return -ENODEV;
115		ret = rdma_copy_addr(dev_addr, dev, NULL);
116		dev_put(dev);
117		return ret;
118	}
119
120	switch (addr->sa_family) {
121	case AF_INET:
122		dev = ip_dev_find(&init_net,
123			((struct sockaddr_in *) addr)->sin_addr.s_addr);
124
125		if (!dev)
126			return ret;
127
128		ret = rdma_copy_addr(dev_addr, dev, NULL);
129		dev_put(dev);
130		break;
131
132#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE)
133	case AF_INET6:
134		rcu_read_lock();
135		for_each_netdev_rcu(&init_net, dev) {
136			if (ipv6_chk_addr(&init_net,
137					  &((struct sockaddr_in6 *) addr)->sin6_addr,
138					  dev, 1)) {
139				ret = rdma_copy_addr(dev_addr, dev, NULL);
140				break;
141			}
142		}
143		rcu_read_unlock();
144		break;
145#endif
146	}
147	return ret;
148}
149EXPORT_SYMBOL(rdma_translate_ip);
150
151static void set_timeout(unsigned long time)
152{
153	unsigned long delay;
154
155	cancel_delayed_work(&work);
156
157	delay = time - jiffies;
158	if ((long)delay <= 0)
159		delay = 1;
160
161	queue_delayed_work(addr_wq, &work, delay);
162}
163
164static void queue_req(struct addr_req *req)
165{
166	struct addr_req *temp_req;
167
168	mutex_lock(&lock);
169	list_for_each_entry_reverse(temp_req, &req_list, list) {
170		if (time_after_eq(req->timeout, temp_req->timeout))
171			break;
172	}
173
174	list_add(&req->list, &temp_req->list);
175
176	if (req_list.next == &req->list)
177		set_timeout(req->timeout);
178	mutex_unlock(&lock);
179}
180
181static int dst_fetch_ha(struct dst_entry *dst, struct rdma_dev_addr *dev_addr, void *daddr)
182{
183	struct neighbour *n;
184	int ret;
185
186	n = dst_neigh_lookup(dst, daddr);
187
188	rcu_read_lock();
189	if (!n || !(n->nud_state & NUD_VALID)) {
190		if (n)
191			neigh_event_send(n, NULL);
192		ret = -ENODATA;
193	} else {
194		ret = rdma_copy_addr(dev_addr, dst->dev, n->ha);
195	}
196	rcu_read_unlock();
197
198	if (n)
199		neigh_release(n);
200
201	return ret;
202}
203
204static int addr4_resolve(struct sockaddr_in *src_in,
205			 struct sockaddr_in *dst_in,
206			 struct rdma_dev_addr *addr)
207{
208	__be32 src_ip = src_in->sin_addr.s_addr;
209	__be32 dst_ip = dst_in->sin_addr.s_addr;
210	struct rtable *rt;
 
211	struct flowi4 fl4;
212	int ret;
213
214	memset(&fl4, 0, sizeof(fl4));
215	fl4.daddr = dst_ip;
216	fl4.saddr = src_ip;
217	fl4.flowi4_oif = addr->bound_dev_if;
218	rt = ip_route_output_key(&init_net, &fl4);
219	if (IS_ERR(rt)) {
220		ret = PTR_ERR(rt);
221		goto out;
222	}
223	src_in->sin_family = AF_INET;
224	src_in->sin_addr.s_addr = fl4.saddr;
225
226	if (rt->dst.dev->flags & IFF_LOOPBACK) {
227		ret = rdma_translate_ip((struct sockaddr *) dst_in, addr);
228		if (!ret)
229			memcpy(addr->dst_dev_addr, addr->src_dev_addr, MAX_ADDR_LEN);
230		goto put;
231	}
232
233	/* If the device does ARP internally, return 'done' */
234	if (rt->dst.dev->flags & IFF_NOARP) {
235		ret = rdma_copy_addr(addr, rt->dst.dev, NULL);
236		goto put;
237	}
238
239	ret = dst_fetch_ha(&rt->dst, addr, &fl4.daddr);
 
 
 
 
 
 
 
 
 
 
 
240put:
241	ip_rt_put(rt);
242out:
243	return ret;
244}
245
246#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE)
247static int addr6_resolve(struct sockaddr_in6 *src_in,
248			 struct sockaddr_in6 *dst_in,
249			 struct rdma_dev_addr *addr)
250{
251	struct flowi6 fl6;
 
252	struct dst_entry *dst;
253	int ret;
254
255	memset(&fl6, 0, sizeof fl6);
256	fl6.daddr = dst_in->sin6_addr;
257	fl6.saddr = src_in->sin6_addr;
258	fl6.flowi6_oif = addr->bound_dev_if;
259
260	dst = ip6_route_output(&init_net, NULL, &fl6);
261	if ((ret = dst->error))
262		goto put;
263
264	if (ipv6_addr_any(&fl6.saddr)) {
265		ret = ipv6_dev_get_saddr(&init_net, ip6_dst_idev(dst)->dev,
266					 &fl6.daddr, 0, &fl6.saddr);
267		if (ret)
268			goto put;
269
270		src_in->sin6_family = AF_INET6;
271		src_in->sin6_addr = fl6.saddr;
272	}
273
274	if (dst->dev->flags & IFF_LOOPBACK) {
275		ret = rdma_translate_ip((struct sockaddr *) dst_in, addr);
276		if (!ret)
277			memcpy(addr->dst_dev_addr, addr->src_dev_addr, MAX_ADDR_LEN);
278		goto put;
279	}
280
281	/* If the device does ARP internally, return 'done' */
282	if (dst->dev->flags & IFF_NOARP) {
283		ret = rdma_copy_addr(addr, dst->dev, NULL);
284		goto put;
285	}
286
287	ret = dst_fetch_ha(dst, addr, &fl6.daddr);
 
 
 
 
 
 
 
 
288put:
289	dst_release(dst);
290	return ret;
291}
292#else
293static int addr6_resolve(struct sockaddr_in6 *src_in,
294			 struct sockaddr_in6 *dst_in,
295			 struct rdma_dev_addr *addr)
296{
297	return -EADDRNOTAVAIL;
298}
299#endif
300
301static int addr_resolve(struct sockaddr *src_in,
302			struct sockaddr *dst_in,
303			struct rdma_dev_addr *addr)
304{
305	if (src_in->sa_family == AF_INET) {
306		return addr4_resolve((struct sockaddr_in *) src_in,
307			(struct sockaddr_in *) dst_in, addr);
308	} else
309		return addr6_resolve((struct sockaddr_in6 *) src_in,
310			(struct sockaddr_in6 *) dst_in, addr);
311}
312
313static void process_req(struct work_struct *work)
314{
315	struct addr_req *req, *temp_req;
316	struct sockaddr *src_in, *dst_in;
317	struct list_head done_list;
318
319	INIT_LIST_HEAD(&done_list);
320
321	mutex_lock(&lock);
322	list_for_each_entry_safe(req, temp_req, &req_list, list) {
323		if (req->status == -ENODATA) {
324			src_in = (struct sockaddr *) &req->src_addr;
325			dst_in = (struct sockaddr *) &req->dst_addr;
326			req->status = addr_resolve(src_in, dst_in, req->addr);
327			if (req->status && time_after_eq(jiffies, req->timeout))
328				req->status = -ETIMEDOUT;
329			else if (req->status == -ENODATA)
330				continue;
331		}
332		list_move_tail(&req->list, &done_list);
333	}
334
335	if (!list_empty(&req_list)) {
336		req = list_entry(req_list.next, struct addr_req, list);
337		set_timeout(req->timeout);
338	}
339	mutex_unlock(&lock);
340
341	list_for_each_entry_safe(req, temp_req, &done_list, list) {
342		list_del(&req->list);
343		req->callback(req->status, (struct sockaddr *) &req->src_addr,
344			req->addr, req->context);
345		put_client(req->client);
346		kfree(req);
347	}
348}
349
350int rdma_resolve_ip(struct rdma_addr_client *client,
351		    struct sockaddr *src_addr, struct sockaddr *dst_addr,
352		    struct rdma_dev_addr *addr, int timeout_ms,
353		    void (*callback)(int status, struct sockaddr *src_addr,
354				     struct rdma_dev_addr *addr, void *context),
355		    void *context)
356{
357	struct sockaddr *src_in, *dst_in;
358	struct addr_req *req;
359	int ret = 0;
360
361	req = kzalloc(sizeof *req, GFP_KERNEL);
362	if (!req)
363		return -ENOMEM;
364
365	src_in = (struct sockaddr *) &req->src_addr;
366	dst_in = (struct sockaddr *) &req->dst_addr;
367
368	if (src_addr) {
369		if (src_addr->sa_family != dst_addr->sa_family) {
370			ret = -EINVAL;
371			goto err;
372		}
373
374		memcpy(src_in, src_addr, ip_addr_size(src_addr));
375	} else {
376		src_in->sa_family = dst_addr->sa_family;
377	}
378
379	memcpy(dst_in, dst_addr, ip_addr_size(dst_addr));
380	req->addr = addr;
381	req->callback = callback;
382	req->context = context;
383	req->client = client;
384	atomic_inc(&client->refcount);
385
386	req->status = addr_resolve(src_in, dst_in, addr);
387	switch (req->status) {
388	case 0:
389		req->timeout = jiffies;
390		queue_req(req);
391		break;
392	case -ENODATA:
393		req->timeout = msecs_to_jiffies(timeout_ms) + jiffies;
394		queue_req(req);
395		break;
396	default:
397		ret = req->status;
398		atomic_dec(&client->refcount);
399		goto err;
400	}
401	return ret;
402err:
403	kfree(req);
404	return ret;
405}
406EXPORT_SYMBOL(rdma_resolve_ip);
407
408void rdma_addr_cancel(struct rdma_dev_addr *addr)
409{
410	struct addr_req *req, *temp_req;
411
412	mutex_lock(&lock);
413	list_for_each_entry_safe(req, temp_req, &req_list, list) {
414		if (req->addr == addr) {
415			req->status = -ECANCELED;
416			req->timeout = jiffies;
417			list_move(&req->list, &req_list);
418			set_timeout(req->timeout);
419			break;
420		}
421	}
422	mutex_unlock(&lock);
423}
424EXPORT_SYMBOL(rdma_addr_cancel);
425
426static int netevent_callback(struct notifier_block *self, unsigned long event,
427	void *ctx)
428{
429	if (event == NETEVENT_NEIGH_UPDATE) {
430		struct neighbour *neigh = ctx;
431
432		if (neigh->nud_state & NUD_VALID) {
433			set_timeout(jiffies);
434		}
435	}
436	return 0;
437}
438
439static struct notifier_block nb = {
440	.notifier_call = netevent_callback
441};
442
443static int __init addr_init(void)
444{
445	addr_wq = create_singlethread_workqueue("ib_addr");
446	if (!addr_wq)
447		return -ENOMEM;
448
449	register_netevent_notifier(&nb);
450	return 0;
451}
452
453static void __exit addr_cleanup(void)
454{
455	unregister_netevent_notifier(&nb);
456	destroy_workqueue(addr_wq);
457}
458
459module_init(addr_init);
460module_exit(addr_cleanup);