• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /* -----------------------------------------------------------------------------
2  * Copyright (c) 2011 Ozmo Inc
3  * Released under the GNU General Public License Version 2 (GPLv2).
4  * -----------------------------------------------------------------------------
5  */
6 
7 #include <linux/module.h>
8 #include <linux/timer.h>
9 #include <linux/sched.h>
10 #include <linux/netdevice.h>
11 #include <linux/etherdevice.h>
12 #include <linux/errno.h>
13 #include <linux/ieee80211.h>
14 #include <linux/slab.h>
15 #include "ozdbg.h"
16 #include "ozprotocol.h"
17 #include "ozeltbuf.h"
18 #include "ozpd.h"
19 #include "ozproto.h"
20 #include "ozusbsvc.h"
21 
22 #include "ozappif.h"
23 #include <asm/unaligned.h>
24 #include <linux/uaccess.h>
25 #include <net/psnap.h>
26 
27 #define OZ_CF_CONN_SUCCESS	1
28 #define OZ_CF_CONN_FAILURE	2
29 
30 #define OZ_DO_STOP		1
31 #define OZ_DO_SLEEP		2
32 
33 struct oz_binding {
34 	struct packet_type ptype;
35 	char name[OZ_MAX_BINDING_LEN];
36 	struct list_head link;
37 };
38 
39 /*
40  * External variable
41  */
42 
43 DEFINE_SPINLOCK(g_polling_lock);
44 /*
45  * Static external variables.
46  */
47 static LIST_HEAD(g_pd_list);
48 static LIST_HEAD(g_binding);
49 static DEFINE_SPINLOCK(g_binding_lock);
50 static struct sk_buff_head g_rx_queue;
51 static u8 g_session_id;
52 static u16 g_apps = 0x1;
53 static int g_processing_rx;
54 
55 struct kmem_cache *oz_elt_info_cache;
56 struct kmem_cache *oz_tx_frame_cache;
57 
58 /*
59  * Context: softirq-serialized
60  */
oz_get_new_session_id(u8 exclude)61 static u8 oz_get_new_session_id(u8 exclude)
62 {
63 	if (++g_session_id == 0)
64 		g_session_id = 1;
65 	if (g_session_id == exclude) {
66 		if (++g_session_id == 0)
67 			g_session_id = 1;
68 	}
69 	return g_session_id;
70 }
71 
72 /*
73  * Context: softirq-serialized
74  */
oz_send_conn_rsp(struct oz_pd * pd,u8 status)75 static void oz_send_conn_rsp(struct oz_pd *pd, u8 status)
76 {
77 	struct sk_buff *skb;
78 	struct net_device *dev = pd->net_dev;
79 	struct oz_hdr *oz_hdr;
80 	struct oz_elt *elt;
81 	struct oz_elt_connect_rsp *body;
82 
83 	int sz = sizeof(struct oz_hdr) + sizeof(struct oz_elt) +
84 			sizeof(struct oz_elt_connect_rsp);
85 	skb = alloc_skb(sz + OZ_ALLOCATED_SPACE(dev), GFP_ATOMIC);
86 	if (skb == NULL)
87 		return;
88 	skb_reserve(skb, LL_RESERVED_SPACE(dev));
89 	skb_reset_network_header(skb);
90 	oz_hdr = (struct oz_hdr *)skb_put(skb, sz);
91 	elt = (struct oz_elt *)(oz_hdr+1);
92 	body = (struct oz_elt_connect_rsp *)(elt+1);
93 	skb->dev = dev;
94 	skb->protocol = htons(OZ_ETHERTYPE);
95 	/* Fill in device header */
96 	if (dev_hard_header(skb, dev, OZ_ETHERTYPE, pd->mac_addr,
97 			dev->dev_addr, skb->len) < 0) {
98 		kfree_skb(skb);
99 		return;
100 	}
101 	oz_hdr->control = (OZ_PROTOCOL_VERSION<<OZ_VERSION_SHIFT);
102 	oz_hdr->last_pkt_num = 0;
103 	put_unaligned(0, &oz_hdr->pkt_num);
104 	elt->type = OZ_ELT_CONNECT_RSP;
105 	elt->length = sizeof(struct oz_elt_connect_rsp);
106 	memset(body, 0, sizeof(struct oz_elt_connect_rsp));
107 	body->status = status;
108 	if (status == 0) {
109 		body->mode = pd->mode;
110 		body->session_id = pd->session_id;
111 		put_unaligned(cpu_to_le16(pd->total_apps), &body->apps);
112 	}
113 	oz_dbg(ON, "TX: OZ_ELT_CONNECT_RSP %d", status);
114 	dev_queue_xmit(skb);
115 }
116 
117 /*
118  * Context: softirq-serialized
119  */
pd_set_keepalive(struct oz_pd * pd,u8 kalive)120 static void pd_set_keepalive(struct oz_pd *pd, u8 kalive)
121 {
122 	unsigned long keep_alive = kalive & OZ_KALIVE_VALUE_MASK;
123 
124 	switch (kalive & OZ_KALIVE_TYPE_MASK) {
125 	case OZ_KALIVE_SPECIAL:
126 		pd->keep_alive = keep_alive * 1000*60*60*24*20;
127 		break;
128 	case OZ_KALIVE_SECS:
129 		pd->keep_alive = keep_alive*1000;
130 		break;
131 	case OZ_KALIVE_MINS:
132 		pd->keep_alive = keep_alive*1000*60;
133 		break;
134 	case OZ_KALIVE_HOURS:
135 		pd->keep_alive = keep_alive*1000*60*60;
136 		break;
137 	default:
138 		pd->keep_alive = 0;
139 	}
140 	oz_dbg(ON, "Keepalive = %lu mSec\n", pd->keep_alive);
141 }
142 
143 /*
144  * Context: softirq-serialized
145  */
pd_set_presleep(struct oz_pd * pd,u8 presleep,u8 start_timer)146 static void pd_set_presleep(struct oz_pd *pd, u8 presleep, u8 start_timer)
147 {
148 	if (presleep)
149 		pd->presleep = presleep*100;
150 	else
151 		pd->presleep = OZ_PRESLEEP_TOUT;
152 	if (start_timer) {
153 		spin_unlock(&g_polling_lock);
154 		oz_timer_add(pd, OZ_TIMER_TOUT, pd->presleep);
155 		spin_lock(&g_polling_lock);
156 	}
157 	oz_dbg(ON, "Presleep time = %lu mSec\n", pd->presleep);
158 }
159 
160 /*
161  * Context: softirq-serialized
162  */
oz_connect_req(struct oz_pd * cur_pd,struct oz_elt * elt,const u8 * pd_addr,struct net_device * net_dev)163 static struct oz_pd *oz_connect_req(struct oz_pd *cur_pd, struct oz_elt *elt,
164 			const u8 *pd_addr, struct net_device *net_dev)
165 {
166 	struct oz_pd *pd;
167 	struct oz_elt_connect_req *body =
168 			(struct oz_elt_connect_req *)(elt+1);
169 	u8 rsp_status = OZ_STATUS_SUCCESS;
170 	u8 stop_needed = 0;
171 	u16 new_apps = g_apps;
172 	struct net_device *old_net_dev = NULL;
173 	struct oz_pd *free_pd = NULL;
174 
175 	if (cur_pd) {
176 		pd = cur_pd;
177 		spin_lock_bh(&g_polling_lock);
178 	} else {
179 		struct oz_pd *pd2 = NULL;
180 		struct list_head *e;
181 
182 		pd = oz_pd_alloc(pd_addr);
183 		if (pd == NULL)
184 			return NULL;
185 		getnstimeofday(&pd->last_rx_timestamp);
186 		spin_lock_bh(&g_polling_lock);
187 		list_for_each(e, &g_pd_list) {
188 			pd2 = list_entry(e, struct oz_pd, link);
189 			if (ether_addr_equal(pd2->mac_addr, pd_addr)) {
190 				free_pd = pd;
191 				pd = pd2;
192 				break;
193 			}
194 		}
195 		if (pd != pd2)
196 			list_add_tail(&pd->link, &g_pd_list);
197 	}
198 	if (pd == NULL) {
199 		spin_unlock_bh(&g_polling_lock);
200 		return NULL;
201 	}
202 	if (pd->net_dev != net_dev) {
203 		old_net_dev = pd->net_dev;
204 		dev_hold(net_dev);
205 		pd->net_dev = net_dev;
206 	}
207 	oz_dbg(ON, "Host vendor: %d\n", body->host_vendor);
208 	pd->max_tx_size = OZ_MAX_TX_SIZE;
209 	pd->mode = body->mode;
210 	pd->pd_info = body->pd_info;
211 	if (pd->mode & OZ_F_ISOC_NO_ELTS) {
212 		pd->ms_per_isoc = body->ms_per_isoc;
213 		if (!pd->ms_per_isoc)
214 			pd->ms_per_isoc = 4;
215 
216 		switch (body->ms_isoc_latency & OZ_LATENCY_MASK) {
217 		case OZ_ONE_MS_LATENCY:
218 			pd->isoc_latency = (body->ms_isoc_latency &
219 					~OZ_LATENCY_MASK) / pd->ms_per_isoc;
220 			break;
221 		case OZ_TEN_MS_LATENCY:
222 			pd->isoc_latency = ((body->ms_isoc_latency &
223 				~OZ_LATENCY_MASK) * 10) / pd->ms_per_isoc;
224 			break;
225 		default:
226 			pd->isoc_latency = OZ_MAX_TX_QUEUE_ISOC;
227 		}
228 	}
229 	if (body->max_len_div16)
230 		pd->max_tx_size = ((u16)body->max_len_div16)<<4;
231 	oz_dbg(ON, "Max frame:%u Ms per isoc:%u\n",
232 	       pd->max_tx_size, pd->ms_per_isoc);
233 	pd->max_stream_buffering = 3*1024;
234 	pd->pulse_period = OZ_QUANTUM;
235 	pd_set_presleep(pd, body->presleep, 0);
236 	pd_set_keepalive(pd, body->keep_alive);
237 
238 	new_apps &= le16_to_cpu(get_unaligned(&body->apps));
239 	if ((new_apps & 0x1) && (body->session_id)) {
240 		if (pd->session_id) {
241 			if (pd->session_id != body->session_id) {
242 				rsp_status = OZ_STATUS_SESSION_MISMATCH;
243 				goto done;
244 			}
245 		} else {
246 			new_apps &= ~0x1;  /* Resume not permitted */
247 			pd->session_id =
248 				oz_get_new_session_id(body->session_id);
249 		}
250 	} else {
251 		if (pd->session_id && !body->session_id) {
252 			rsp_status = OZ_STATUS_SESSION_TEARDOWN;
253 			stop_needed = 1;
254 		} else {
255 			new_apps &= ~0x1;  /* Resume not permitted */
256 			pd->session_id =
257 				oz_get_new_session_id(body->session_id);
258 		}
259 	}
260 done:
261 	if (rsp_status == OZ_STATUS_SUCCESS) {
262 		u16 start_apps = new_apps & ~pd->total_apps & ~0x1;
263 		u16 stop_apps = pd->total_apps & ~new_apps & ~0x1;
264 		u16 resume_apps = new_apps & pd->paused_apps  & ~0x1;
265 
266 		spin_unlock_bh(&g_polling_lock);
267 		oz_pd_set_state(pd, OZ_PD_S_CONNECTED);
268 		oz_dbg(ON, "new_apps=0x%x total_apps=0x%x paused_apps=0x%x\n",
269 		       new_apps, pd->total_apps, pd->paused_apps);
270 		if (start_apps) {
271 			if (oz_services_start(pd, start_apps, 0))
272 				rsp_status = OZ_STATUS_TOO_MANY_PDS;
273 		}
274 		if (resume_apps)
275 			if (oz_services_start(pd, resume_apps, 1))
276 				rsp_status = OZ_STATUS_TOO_MANY_PDS;
277 		if (stop_apps)
278 			oz_services_stop(pd, stop_apps, 0);
279 		oz_pd_request_heartbeat(pd);
280 	} else {
281 		spin_unlock_bh(&g_polling_lock);
282 	}
283 	oz_send_conn_rsp(pd, rsp_status);
284 	if (rsp_status != OZ_STATUS_SUCCESS) {
285 		if (stop_needed)
286 			oz_pd_stop(pd);
287 		oz_pd_put(pd);
288 		pd = NULL;
289 	}
290 	if (old_net_dev)
291 		dev_put(old_net_dev);
292 	if (free_pd)
293 		oz_pd_destroy(free_pd);
294 	return pd;
295 }
296 
297 /*
298  * Context: softirq-serialized
299  */
oz_add_farewell(struct oz_pd * pd,u8 ep_num,u8 index,const u8 * report,u8 len)300 static void oz_add_farewell(struct oz_pd *pd, u8 ep_num, u8 index,
301 			const u8 *report, u8 len)
302 {
303 	struct oz_farewell *f;
304 	struct oz_farewell *f2;
305 	int found = 0;
306 
307 	f = kmalloc(sizeof(struct oz_farewell) + len, GFP_ATOMIC);
308 	if (!f)
309 		return;
310 	f->ep_num = ep_num;
311 	f->index = index;
312 	f->len = len;
313 	memcpy(f->report, report, len);
314 	oz_dbg(ON, "RX: Adding farewell report\n");
315 	spin_lock(&g_polling_lock);
316 	list_for_each_entry(f2, &pd->farewell_list, link) {
317 		if ((f2->ep_num == ep_num) && (f2->index == index)) {
318 			found = 1;
319 			list_del(&f2->link);
320 			break;
321 		}
322 	}
323 	list_add_tail(&f->link, &pd->farewell_list);
324 	spin_unlock(&g_polling_lock);
325 	if (found)
326 		kfree(f2);
327 }
328 
329 /*
330  * Context: softirq-serialized
331  */
oz_rx_frame(struct sk_buff * skb)332 static void oz_rx_frame(struct sk_buff *skb)
333 {
334 	u8 *mac_hdr;
335 	u8 *src_addr;
336 	struct oz_elt *elt;
337 	int length;
338 	struct oz_pd *pd = NULL;
339 	struct oz_hdr *oz_hdr = (struct oz_hdr *)skb_network_header(skb);
340 	struct timespec current_time;
341 	int dup = 0;
342 	u32 pkt_num;
343 
344 	oz_dbg(RX_FRAMES, "RX frame PN=0x%x LPN=0x%x control=0x%x\n",
345 	       oz_hdr->pkt_num, oz_hdr->last_pkt_num, oz_hdr->control);
346 	mac_hdr = skb_mac_header(skb);
347 	src_addr = &mac_hdr[ETH_ALEN];
348 	length = skb->len;
349 
350 	/* Check the version field */
351 	if (oz_get_prot_ver(oz_hdr->control) != OZ_PROTOCOL_VERSION) {
352 		oz_dbg(ON, "Incorrect protocol version: %d\n",
353 		       oz_get_prot_ver(oz_hdr->control));
354 		goto done;
355 	}
356 
357 	pkt_num = le32_to_cpu(get_unaligned(&oz_hdr->pkt_num));
358 
359 	pd = oz_pd_find(src_addr);
360 	if (pd) {
361 		if (!(pd->state & OZ_PD_S_CONNECTED))
362 			oz_pd_set_state(pd, OZ_PD_S_CONNECTED);
363 		getnstimeofday(&current_time);
364 		if ((current_time.tv_sec != pd->last_rx_timestamp.tv_sec) ||
365 			(pd->presleep < MSEC_PER_SEC))  {
366 			oz_timer_add(pd, OZ_TIMER_TOUT,	pd->presleep);
367 			pd->last_rx_timestamp = current_time;
368 		}
369 		if (pkt_num != pd->last_rx_pkt_num) {
370 			pd->last_rx_pkt_num = pkt_num;
371 		} else {
372 			dup = 1;
373 			oz_dbg(ON, "Duplicate frame\n");
374 		}
375 	}
376 
377 	if (pd && !dup && ((pd->mode & OZ_MODE_MASK) == OZ_MODE_TRIGGERED)) {
378 		oz_dbg(RX_FRAMES, "Received TRIGGER Frame\n");
379 		pd->last_sent_frame = &pd->tx_queue;
380 		if (oz_hdr->control & OZ_F_ACK) {
381 			/* Retire completed frames */
382 			oz_retire_tx_frames(pd, oz_hdr->last_pkt_num);
383 		}
384 		if ((oz_hdr->control & OZ_F_ACK_REQUESTED) &&
385 				(pd->state == OZ_PD_S_CONNECTED)) {
386 			int backlog = pd->nb_queued_frames;
387 
388 			pd->trigger_pkt_num = pkt_num;
389 			/* Send queued frames */
390 			oz_send_queued_frames(pd, backlog);
391 		}
392 	}
393 
394 	length -= sizeof(struct oz_hdr);
395 	elt = (struct oz_elt *)((u8 *)oz_hdr + sizeof(struct oz_hdr));
396 
397 	while (length >= sizeof(struct oz_elt)) {
398 		length -= sizeof(struct oz_elt) + elt->length;
399 		if (length < 0)
400 			break;
401 		switch (elt->type) {
402 		case OZ_ELT_CONNECT_REQ:
403 			oz_dbg(ON, "RX: OZ_ELT_CONNECT_REQ\n");
404 			pd = oz_connect_req(pd, elt, src_addr, skb->dev);
405 			break;
406 		case OZ_ELT_DISCONNECT:
407 			oz_dbg(ON, "RX: OZ_ELT_DISCONNECT\n");
408 			if (pd)
409 				oz_pd_sleep(pd);
410 			break;
411 		case OZ_ELT_UPDATE_PARAM_REQ: {
412 				struct oz_elt_update_param *body =
413 					(struct oz_elt_update_param *)(elt + 1);
414 				oz_dbg(ON, "RX: OZ_ELT_UPDATE_PARAM_REQ\n");
415 				if (pd && (pd->state & OZ_PD_S_CONNECTED)) {
416 					spin_lock(&g_polling_lock);
417 					pd_set_keepalive(pd, body->keepalive);
418 					pd_set_presleep(pd, body->presleep, 1);
419 					spin_unlock(&g_polling_lock);
420 				}
421 			}
422 			break;
423 		case OZ_ELT_FAREWELL_REQ: {
424 				struct oz_elt_farewell *body =
425 					(struct oz_elt_farewell *)(elt + 1);
426 				oz_dbg(ON, "RX: OZ_ELT_FAREWELL_REQ\n");
427 				oz_add_farewell(pd, body->ep_num,
428 					body->index, body->report,
429 					elt->length + 1 - sizeof(*body));
430 			}
431 			break;
432 		case OZ_ELT_APP_DATA:
433 			if (pd && (pd->state & OZ_PD_S_CONNECTED)) {
434 				struct oz_app_hdr *app_hdr =
435 					(struct oz_app_hdr *)(elt+1);
436 				if (dup)
437 					break;
438 				oz_handle_app_elt(pd, app_hdr->app_id, elt);
439 			}
440 			break;
441 		default:
442 			oz_dbg(ON, "RX: Unknown elt %02x\n", elt->type);
443 		}
444 		elt = oz_next_elt(elt);
445 	}
446 done:
447 	if (pd)
448 		oz_pd_put(pd);
449 	consume_skb(skb);
450 }
451 
452 /*
453  * Context: process
454  */
oz_protocol_term(void)455 void oz_protocol_term(void)
456 {
457 	struct oz_binding *b, *t;
458 
459 	/* Walk the list of bindings and remove each one.
460 	 */
461 	spin_lock_bh(&g_binding_lock);
462 	list_for_each_entry_safe(b, t, &g_binding, link) {
463 		list_del(&b->link);
464 		spin_unlock_bh(&g_binding_lock);
465 		dev_remove_pack(&b->ptype);
466 		if (b->ptype.dev)
467 			dev_put(b->ptype.dev);
468 		kfree(b);
469 		spin_lock_bh(&g_binding_lock);
470 	}
471 	spin_unlock_bh(&g_binding_lock);
472 	/* Walk the list of PDs and stop each one. This causes the PD to be
473 	 * removed from the list so we can just pull each one from the head
474 	 * of the list.
475 	 */
476 	spin_lock_bh(&g_polling_lock);
477 	while (!list_empty(&g_pd_list)) {
478 		struct oz_pd *pd =
479 			list_first_entry(&g_pd_list, struct oz_pd, link);
480 		oz_pd_get(pd);
481 		spin_unlock_bh(&g_polling_lock);
482 		oz_pd_stop(pd);
483 		oz_pd_put(pd);
484 		spin_lock_bh(&g_polling_lock);
485 	}
486 	spin_unlock_bh(&g_polling_lock);
487 	oz_dbg(ON, "Protocol stopped\n");
488 
489 	kmem_cache_destroy(oz_tx_frame_cache);
490 	kmem_cache_destroy(oz_elt_info_cache);
491 }
492 
493 /*
494  * Context: softirq
495  */
oz_pd_heartbeat_handler(unsigned long data)496 void oz_pd_heartbeat_handler(unsigned long data)
497 {
498 	struct oz_pd *pd = (struct oz_pd *)data;
499 	u16 apps = 0;
500 
501 	spin_lock_bh(&g_polling_lock);
502 	if (pd->state & OZ_PD_S_CONNECTED)
503 		apps = pd->total_apps;
504 	spin_unlock_bh(&g_polling_lock);
505 	if (apps)
506 		oz_pd_heartbeat(pd, apps);
507 	oz_pd_put(pd);
508 }
509 
510 /*
511  * Context: softirq
512  */
oz_pd_timeout_handler(unsigned long data)513 void oz_pd_timeout_handler(unsigned long data)
514 {
515 	int type;
516 	struct oz_pd *pd = (struct oz_pd *)data;
517 
518 	spin_lock_bh(&g_polling_lock);
519 	type = pd->timeout_type;
520 	spin_unlock_bh(&g_polling_lock);
521 	switch (type) {
522 	case OZ_TIMER_TOUT:
523 		oz_pd_sleep(pd);
524 		break;
525 	case OZ_TIMER_STOP:
526 		oz_pd_stop(pd);
527 		break;
528 	}
529 	oz_pd_put(pd);
530 }
531 
532 /*
533  * Context: Interrupt
534  */
oz_pd_heartbeat_event(struct hrtimer * timer)535 enum hrtimer_restart oz_pd_heartbeat_event(struct hrtimer *timer)
536 {
537 	struct oz_pd *pd;
538 
539 	pd = container_of(timer, struct oz_pd, heartbeat);
540 	hrtimer_forward_now(timer, ktime_set(pd->pulse_period /
541 	MSEC_PER_SEC, (pd->pulse_period % MSEC_PER_SEC) * NSEC_PER_MSEC));
542 	oz_pd_get(pd);
543 	tasklet_schedule(&pd->heartbeat_tasklet);
544 	return HRTIMER_RESTART;
545 }
546 
547 /*
548  * Context: Interrupt
549  */
oz_pd_timeout_event(struct hrtimer * timer)550 enum hrtimer_restart oz_pd_timeout_event(struct hrtimer *timer)
551 {
552 	struct oz_pd *pd;
553 
554 	pd = container_of(timer, struct oz_pd, timeout);
555 	oz_pd_get(pd);
556 	tasklet_schedule(&pd->timeout_tasklet);
557 	return HRTIMER_NORESTART;
558 }
559 
560 /*
561  * Context: softirq or process
562  */
oz_timer_add(struct oz_pd * pd,int type,unsigned long due_time)563 void oz_timer_add(struct oz_pd *pd, int type, unsigned long due_time)
564 {
565 	spin_lock_bh(&g_polling_lock);
566 	switch (type) {
567 	case OZ_TIMER_TOUT:
568 	case OZ_TIMER_STOP:
569 		if (hrtimer_active(&pd->timeout)) {
570 			hrtimer_set_expires(&pd->timeout, ktime_set(due_time /
571 			MSEC_PER_SEC, (due_time % MSEC_PER_SEC) *
572 							NSEC_PER_MSEC));
573 			hrtimer_start_expires(&pd->timeout, HRTIMER_MODE_REL);
574 		} else {
575 			hrtimer_start(&pd->timeout, ktime_set(due_time /
576 			MSEC_PER_SEC, (due_time % MSEC_PER_SEC) *
577 					NSEC_PER_MSEC), HRTIMER_MODE_REL);
578 		}
579 		pd->timeout_type = type;
580 		break;
581 	case OZ_TIMER_HEARTBEAT:
582 		if (!hrtimer_active(&pd->heartbeat))
583 			hrtimer_start(&pd->heartbeat, ktime_set(due_time /
584 			MSEC_PER_SEC, (due_time % MSEC_PER_SEC) *
585 					NSEC_PER_MSEC), HRTIMER_MODE_REL);
586 		break;
587 	}
588 	spin_unlock_bh(&g_polling_lock);
589 }
590 
591 /*
592  * Context: softirq or process
593  */
oz_pd_request_heartbeat(struct oz_pd * pd)594 void oz_pd_request_heartbeat(struct oz_pd *pd)
595 {
596 	oz_timer_add(pd, OZ_TIMER_HEARTBEAT, pd->pulse_period > 0 ?
597 					pd->pulse_period : OZ_QUANTUM);
598 }
599 
600 /*
601  * Context: softirq or process
602  */
oz_pd_find(const u8 * mac_addr)603 struct oz_pd *oz_pd_find(const u8 *mac_addr)
604 {
605 	struct oz_pd *pd;
606 
607 	spin_lock_bh(&g_polling_lock);
608 	list_for_each_entry(pd, &g_pd_list, link) {
609 		if (ether_addr_equal(pd->mac_addr, mac_addr)) {
610 			oz_pd_get(pd);
611 			spin_unlock_bh(&g_polling_lock);
612 			return pd;
613 		}
614 	}
615 	spin_unlock_bh(&g_polling_lock);
616 	return NULL;
617 }
618 
619 /*
620  * Context: process
621  */
oz_app_enable(int app_id,int enable)622 void oz_app_enable(int app_id, int enable)
623 {
624 	if (app_id < OZ_NB_APPS) {
625 		spin_lock_bh(&g_polling_lock);
626 		if (enable)
627 			g_apps |= (1<<app_id);
628 		else
629 			g_apps &= ~(1<<app_id);
630 		spin_unlock_bh(&g_polling_lock);
631 	}
632 }
633 
634 /*
635  * Context: softirq
636  */
oz_pkt_recv(struct sk_buff * skb,struct net_device * dev,struct packet_type * pt,struct net_device * orig_dev)637 static int oz_pkt_recv(struct sk_buff *skb, struct net_device *dev,
638 		struct packet_type *pt, struct net_device *orig_dev)
639 {
640 	skb = skb_share_check(skb, GFP_ATOMIC);
641 	if (skb == NULL)
642 		return 0;
643 	spin_lock_bh(&g_rx_queue.lock);
644 	if (g_processing_rx) {
645 		/* We already hold the lock so use __ variant.
646 		 */
647 		__skb_queue_head(&g_rx_queue, skb);
648 		spin_unlock_bh(&g_rx_queue.lock);
649 	} else {
650 		g_processing_rx = 1;
651 		do {
652 
653 			spin_unlock_bh(&g_rx_queue.lock);
654 			oz_rx_frame(skb);
655 			spin_lock_bh(&g_rx_queue.lock);
656 			if (skb_queue_empty(&g_rx_queue)) {
657 				g_processing_rx = 0;
658 				spin_unlock_bh(&g_rx_queue.lock);
659 				break;
660 			}
661 			/* We already hold the lock so use __ variant.
662 			 */
663 			skb = __skb_dequeue(&g_rx_queue);
664 		} while (1);
665 	}
666 	return 0;
667 }
668 
669 /*
670  * Context: process
671  */
oz_binding_add(const char * net_dev)672 void oz_binding_add(const char *net_dev)
673 {
674 	struct oz_binding *binding;
675 
676 	binding = kzalloc(sizeof(struct oz_binding), GFP_KERNEL);
677 	if (!binding)
678 		return;
679 
680 	binding->ptype.type = htons(OZ_ETHERTYPE);
681 	binding->ptype.func = oz_pkt_recv;
682 	if (net_dev && *net_dev) {
683 		memcpy(binding->name, net_dev, OZ_MAX_BINDING_LEN);
684 		oz_dbg(ON, "Adding binding: %s\n", net_dev);
685 		binding->ptype.dev = dev_get_by_name(&init_net, net_dev);
686 		if (binding->ptype.dev == NULL) {
687 			oz_dbg(ON, "Netdev %s not found\n", net_dev);
688 			kfree(binding);
689 			return;
690 		}
691 	}
692 	dev_add_pack(&binding->ptype);
693 	spin_lock_bh(&g_binding_lock);
694 	list_add_tail(&binding->link, &g_binding);
695 	spin_unlock_bh(&g_binding_lock);
696 }
697 
698 /*
699  * Context: process
700  */
pd_stop_all_for_device(struct net_device * net_dev)701 static void pd_stop_all_for_device(struct net_device *net_dev)
702 {
703 	LIST_HEAD(h);
704 	struct oz_pd *pd;
705 	struct oz_pd *n;
706 
707 	spin_lock_bh(&g_polling_lock);
708 	list_for_each_entry_safe(pd, n, &g_pd_list, link) {
709 		if (pd->net_dev == net_dev) {
710 			list_move(&pd->link, &h);
711 			oz_pd_get(pd);
712 		}
713 	}
714 	spin_unlock_bh(&g_polling_lock);
715 	while (!list_empty(&h)) {
716 		pd = list_first_entry(&h, struct oz_pd, link);
717 		oz_pd_stop(pd);
718 		oz_pd_put(pd);
719 	}
720 }
721 
722 /*
723  * Context: process
724  */
oz_binding_remove(const char * net_dev)725 void oz_binding_remove(const char *net_dev)
726 {
727 	struct oz_binding *binding;
728 	int found = 0;
729 
730 	oz_dbg(ON, "Removing binding: %s\n", net_dev);
731 	spin_lock_bh(&g_binding_lock);
732 	list_for_each_entry(binding, &g_binding, link) {
733 		if (strncmp(binding->name, net_dev, OZ_MAX_BINDING_LEN) == 0) {
734 			oz_dbg(ON, "Binding '%s' found\n", net_dev);
735 			found = 1;
736 			break;
737 		}
738 	}
739 	spin_unlock_bh(&g_binding_lock);
740 	if (found) {
741 		dev_remove_pack(&binding->ptype);
742 		if (binding->ptype.dev) {
743 			dev_put(binding->ptype.dev);
744 			pd_stop_all_for_device(binding->ptype.dev);
745 		}
746 		list_del(&binding->link);
747 		kfree(binding);
748 	}
749 }
750 
751 /*
752  * Context: process
753  */
oz_get_next_device_name(char * s,char * dname,int max_size)754 static char *oz_get_next_device_name(char *s, char *dname, int max_size)
755 {
756 	while (*s == ',')
757 		s++;
758 	while (*s && (*s != ',') && max_size > 1) {
759 		*dname++ = *s++;
760 		max_size--;
761 	}
762 	*dname = 0;
763 	return s;
764 }
765 
766 /*
767  * Context: process
768  */
oz_protocol_init(char * devs)769 int oz_protocol_init(char *devs)
770 {
771 	oz_elt_info_cache = KMEM_CACHE(oz_elt_info, 0);
772 	if (!oz_elt_info_cache)
773 		return -ENOMEM;
774 
775 	oz_tx_frame_cache = KMEM_CACHE(oz_tx_frame, 0);
776 	if (!oz_tx_frame_cache) {
777 		kmem_cache_destroy(oz_elt_info_cache);
778 		return -ENOMEM;
779 	}
780 
781 	skb_queue_head_init(&g_rx_queue);
782 	if (devs[0] == '*') {
783 		oz_binding_add(NULL);
784 	} else {
785 		char d[32];
786 
787 		while (*devs) {
788 			devs = oz_get_next_device_name(devs, d, sizeof(d));
789 			if (d[0])
790 				oz_binding_add(d);
791 		}
792 	}
793 	return 0;
794 }
795 
796 /*
797  * Context: process
798  */
oz_get_pd_list(struct oz_mac_addr * addr,int max_count)799 int oz_get_pd_list(struct oz_mac_addr *addr, int max_count)
800 {
801 	struct oz_pd *pd;
802 	int count = 0;
803 
804 	spin_lock_bh(&g_polling_lock);
805 	list_for_each_entry(pd, &g_pd_list, link) {
806 		if (count >= max_count)
807 			break;
808 		ether_addr_copy((u8 *)&addr[count++], pd->mac_addr);
809 	}
810 	spin_unlock_bh(&g_polling_lock);
811 	return count;
812 }
813 
814