• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * UWB reservation management.
3  *
4  * Copyright (C) 2008 Cambridge Silicon Radio Ltd.
5  *
6  * This program is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU General Public License version
8  * 2 as published by the Free Software Foundation.
9  *
10  * This program is distributed in the hope that it will be useful,
11  * but WITHOUT ANY WARRANTY; without even the implied warranty of
12  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
13  * GNU General Public License for more details.
14  *
15  * You should have received a copy of the GNU General Public License
16  * along with this program.  If not, see <http://www.gnu.org/licenses/>.
17  */
18 #include <linux/kernel.h>
19 #include <linux/uwb.h>
20 #include <linux/slab.h>
21 #include <linux/random.h>
22 #include <linux/export.h>
23 
24 #include "uwb-internal.h"
25 
26 static void uwb_rsv_timer(unsigned long arg);
27 
28 static const char *rsv_states[] = {
29 	[UWB_RSV_STATE_NONE]                 = "none            ",
30 	[UWB_RSV_STATE_O_INITIATED]          = "o initiated     ",
31 	[UWB_RSV_STATE_O_PENDING]            = "o pending       ",
32 	[UWB_RSV_STATE_O_MODIFIED]           = "o modified      ",
33 	[UWB_RSV_STATE_O_ESTABLISHED]        = "o established   ",
34 	[UWB_RSV_STATE_O_TO_BE_MOVED]        = "o to be moved   ",
35 	[UWB_RSV_STATE_O_MOVE_EXPANDING]     = "o move expanding",
36 	[UWB_RSV_STATE_O_MOVE_COMBINING]     = "o move combining",
37 	[UWB_RSV_STATE_O_MOVE_REDUCING]      = "o move reducing ",
38 	[UWB_RSV_STATE_T_ACCEPTED]           = "t accepted      ",
39 	[UWB_RSV_STATE_T_CONFLICT]           = "t conflict      ",
40 	[UWB_RSV_STATE_T_PENDING]            = "t pending       ",
41 	[UWB_RSV_STATE_T_DENIED]             = "t denied        ",
42 	[UWB_RSV_STATE_T_RESIZED]            = "t resized       ",
43 	[UWB_RSV_STATE_T_EXPANDING_ACCEPTED] = "t expanding acc ",
44 	[UWB_RSV_STATE_T_EXPANDING_CONFLICT] = "t expanding conf",
45 	[UWB_RSV_STATE_T_EXPANDING_PENDING]  = "t expanding pend",
46 	[UWB_RSV_STATE_T_EXPANDING_DENIED]   = "t expanding den ",
47 };
48 
49 static const char *rsv_types[] = {
50 	[UWB_DRP_TYPE_ALIEN_BP] = "alien-bp",
51 	[UWB_DRP_TYPE_HARD]     = "hard",
52 	[UWB_DRP_TYPE_SOFT]     = "soft",
53 	[UWB_DRP_TYPE_PRIVATE]  = "private",
54 	[UWB_DRP_TYPE_PCA]      = "pca",
55 };
56 
uwb_rsv_has_two_drp_ies(struct uwb_rsv * rsv)57 bool uwb_rsv_has_two_drp_ies(struct uwb_rsv *rsv)
58 {
59 	static const bool has_two_drp_ies[] = {
60 		[UWB_RSV_STATE_O_INITIATED]               = false,
61 		[UWB_RSV_STATE_O_PENDING]                 = false,
62 		[UWB_RSV_STATE_O_MODIFIED]                = false,
63 		[UWB_RSV_STATE_O_ESTABLISHED]             = false,
64 		[UWB_RSV_STATE_O_TO_BE_MOVED]             = false,
65 		[UWB_RSV_STATE_O_MOVE_COMBINING]          = false,
66 		[UWB_RSV_STATE_O_MOVE_REDUCING]           = false,
67 		[UWB_RSV_STATE_O_MOVE_EXPANDING]          = true,
68 		[UWB_RSV_STATE_T_ACCEPTED]                = false,
69 		[UWB_RSV_STATE_T_CONFLICT]                = false,
70 		[UWB_RSV_STATE_T_PENDING]                 = false,
71 		[UWB_RSV_STATE_T_DENIED]                  = false,
72 		[UWB_RSV_STATE_T_RESIZED]                 = false,
73 		[UWB_RSV_STATE_T_EXPANDING_ACCEPTED]      = true,
74 		[UWB_RSV_STATE_T_EXPANDING_CONFLICT]      = true,
75 		[UWB_RSV_STATE_T_EXPANDING_PENDING]       = true,
76 		[UWB_RSV_STATE_T_EXPANDING_DENIED]        = true,
77 	};
78 
79 	return has_two_drp_ies[rsv->state];
80 }
81 
82 /**
83  * uwb_rsv_state_str - return a string for a reservation state
84  * @state: the reservation state.
85  */
uwb_rsv_state_str(enum uwb_rsv_state state)86 const char *uwb_rsv_state_str(enum uwb_rsv_state state)
87 {
88 	if (state < UWB_RSV_STATE_NONE || state >= UWB_RSV_STATE_LAST)
89 		return "unknown";
90 	return rsv_states[state];
91 }
92 EXPORT_SYMBOL_GPL(uwb_rsv_state_str);
93 
94 /**
95  * uwb_rsv_type_str - return a string for a reservation type
96  * @type: the reservation type
97  */
uwb_rsv_type_str(enum uwb_drp_type type)98 const char *uwb_rsv_type_str(enum uwb_drp_type type)
99 {
100 	if (type < UWB_DRP_TYPE_ALIEN_BP || type > UWB_DRP_TYPE_PCA)
101 		return "invalid";
102 	return rsv_types[type];
103 }
104 EXPORT_SYMBOL_GPL(uwb_rsv_type_str);
105 
uwb_rsv_dump(char * text,struct uwb_rsv * rsv)106 void uwb_rsv_dump(char *text, struct uwb_rsv *rsv)
107 {
108 	struct device *dev = &rsv->rc->uwb_dev.dev;
109 	struct uwb_dev_addr devaddr;
110 	char owner[UWB_ADDR_STRSIZE], target[UWB_ADDR_STRSIZE];
111 
112 	uwb_dev_addr_print(owner, sizeof(owner), &rsv->owner->dev_addr);
113 	if (rsv->target.type == UWB_RSV_TARGET_DEV)
114 		devaddr = rsv->target.dev->dev_addr;
115 	else
116 		devaddr = rsv->target.devaddr;
117 	uwb_dev_addr_print(target, sizeof(target), &devaddr);
118 
119 	dev_dbg(dev, "rsv %s %s -> %s: %s\n",
120 		text, owner, target, uwb_rsv_state_str(rsv->state));
121 }
122 
uwb_rsv_release(struct kref * kref)123 static void uwb_rsv_release(struct kref *kref)
124 {
125 	struct uwb_rsv *rsv = container_of(kref, struct uwb_rsv, kref);
126 
127 	kfree(rsv);
128 }
129 
uwb_rsv_get(struct uwb_rsv * rsv)130 void uwb_rsv_get(struct uwb_rsv *rsv)
131 {
132 	kref_get(&rsv->kref);
133 }
134 
uwb_rsv_put(struct uwb_rsv * rsv)135 void uwb_rsv_put(struct uwb_rsv *rsv)
136 {
137 	kref_put(&rsv->kref, uwb_rsv_release);
138 }
139 
140 /*
141  * Get a free stream index for a reservation.
142  *
143  * If the target is a DevAddr (e.g., a WUSB cluster reservation) then
144  * the stream is allocated from a pool of per-RC stream indexes,
145  * otherwise a unique stream index for the target is selected.
146  */
uwb_rsv_get_stream(struct uwb_rsv * rsv)147 static int uwb_rsv_get_stream(struct uwb_rsv *rsv)
148 {
149 	struct uwb_rc *rc = rsv->rc;
150 	struct device *dev = &rc->uwb_dev.dev;
151 	unsigned long *streams_bm;
152 	int stream;
153 
154 	switch (rsv->target.type) {
155 	case UWB_RSV_TARGET_DEV:
156 		streams_bm = rsv->target.dev->streams;
157 		break;
158 	case UWB_RSV_TARGET_DEVADDR:
159 		streams_bm = rc->uwb_dev.streams;
160 		break;
161 	default:
162 		return -EINVAL;
163 	}
164 
165 	stream = find_first_zero_bit(streams_bm, UWB_NUM_STREAMS);
166 	if (stream >= UWB_NUM_STREAMS)
167 		return -EBUSY;
168 
169 	rsv->stream = stream;
170 	set_bit(stream, streams_bm);
171 
172 	dev_dbg(dev, "get stream %d\n", rsv->stream);
173 
174 	return 0;
175 }
176 
uwb_rsv_put_stream(struct uwb_rsv * rsv)177 static void uwb_rsv_put_stream(struct uwb_rsv *rsv)
178 {
179 	struct uwb_rc *rc = rsv->rc;
180 	struct device *dev = &rc->uwb_dev.dev;
181 	unsigned long *streams_bm;
182 
183 	switch (rsv->target.type) {
184 	case UWB_RSV_TARGET_DEV:
185 		streams_bm = rsv->target.dev->streams;
186 		break;
187 	case UWB_RSV_TARGET_DEVADDR:
188 		streams_bm = rc->uwb_dev.streams;
189 		break;
190 	default:
191 		return;
192 	}
193 
194 	clear_bit(rsv->stream, streams_bm);
195 
196 	dev_dbg(dev, "put stream %d\n", rsv->stream);
197 }
198 
uwb_rsv_backoff_win_timer(unsigned long arg)199 void uwb_rsv_backoff_win_timer(unsigned long arg)
200 {
201 	struct uwb_drp_backoff_win *bow = (struct uwb_drp_backoff_win *)arg;
202 	struct uwb_rc *rc = container_of(bow, struct uwb_rc, bow);
203 	struct device *dev = &rc->uwb_dev.dev;
204 
205 	bow->can_reserve_extra_mases = true;
206 	if (bow->total_expired <= 4) {
207 		bow->total_expired++;
208 	} else {
209 		/* after 4 backoff window has expired we can exit from
210 		 * the backoff procedure */
211 		bow->total_expired = 0;
212 		bow->window = UWB_DRP_BACKOFF_WIN_MIN >> 1;
213 	}
214 	dev_dbg(dev, "backoff_win_timer total_expired=%d, n=%d\n: ", bow->total_expired, bow->n);
215 
216 	/* try to relocate all the "to be moved" relocations */
217 	uwb_rsv_handle_drp_avail_change(rc);
218 }
219 
uwb_rsv_backoff_win_increment(struct uwb_rc * rc)220 void uwb_rsv_backoff_win_increment(struct uwb_rc *rc)
221 {
222 	struct uwb_drp_backoff_win *bow = &rc->bow;
223 	struct device *dev = &rc->uwb_dev.dev;
224 	unsigned timeout_us;
225 
226 	dev_dbg(dev, "backoff_win_increment: window=%d\n", bow->window);
227 
228 	bow->can_reserve_extra_mases = false;
229 
230 	if((bow->window << 1) == UWB_DRP_BACKOFF_WIN_MAX)
231 		return;
232 
233 	bow->window <<= 1;
234 	bow->n = random32() & (bow->window - 1);
235 	dev_dbg(dev, "new_window=%d, n=%d\n: ", bow->window, bow->n);
236 
237 	/* reset the timer associated variables */
238 	timeout_us = bow->n * UWB_SUPERFRAME_LENGTH_US;
239 	bow->total_expired = 0;
240 	mod_timer(&bow->timer, jiffies + usecs_to_jiffies(timeout_us));
241 }
242 
uwb_rsv_stroke_timer(struct uwb_rsv * rsv)243 static void uwb_rsv_stroke_timer(struct uwb_rsv *rsv)
244 {
245 	int sframes = UWB_MAX_LOST_BEACONS;
246 
247 	/*
248 	 * Multicast reservations can become established within 1
249 	 * super frame and should not be terminated if no response is
250 	 * received.
251 	 */
252 	if (rsv->is_multicast) {
253 		if (rsv->state == UWB_RSV_STATE_O_INITIATED
254 		    || rsv->state == UWB_RSV_STATE_O_MOVE_EXPANDING
255 		    || rsv->state == UWB_RSV_STATE_O_MOVE_COMBINING
256 		    || rsv->state == UWB_RSV_STATE_O_MOVE_REDUCING)
257 			sframes = 1;
258 		if (rsv->state == UWB_RSV_STATE_O_ESTABLISHED)
259 			sframes = 0;
260 
261 	}
262 
263 	if (sframes > 0) {
264 		/*
265 		 * Add an additional 2 superframes to account for the
266 		 * time to send the SET DRP IE command.
267 		 */
268 		unsigned timeout_us = (sframes + 2) * UWB_SUPERFRAME_LENGTH_US;
269 		mod_timer(&rsv->timer, jiffies + usecs_to_jiffies(timeout_us));
270 	} else
271 		del_timer(&rsv->timer);
272 }
273 
274 /*
275  * Update a reservations state, and schedule an update of the
276  * transmitted DRP IEs.
277  */
uwb_rsv_state_update(struct uwb_rsv * rsv,enum uwb_rsv_state new_state)278 static void uwb_rsv_state_update(struct uwb_rsv *rsv,
279 				 enum uwb_rsv_state new_state)
280 {
281 	rsv->state = new_state;
282 	rsv->ie_valid = false;
283 
284 	uwb_rsv_dump("SU", rsv);
285 
286 	uwb_rsv_stroke_timer(rsv);
287 	uwb_rsv_sched_update(rsv->rc);
288 }
289 
uwb_rsv_callback(struct uwb_rsv * rsv)290 static void uwb_rsv_callback(struct uwb_rsv *rsv)
291 {
292 	if (rsv->callback)
293 		rsv->callback(rsv);
294 }
295 
uwb_rsv_set_state(struct uwb_rsv * rsv,enum uwb_rsv_state new_state)296 void uwb_rsv_set_state(struct uwb_rsv *rsv, enum uwb_rsv_state new_state)
297 {
298 	struct uwb_rsv_move *mv = &rsv->mv;
299 
300 	if (rsv->state == new_state) {
301 		switch (rsv->state) {
302 		case UWB_RSV_STATE_O_ESTABLISHED:
303 		case UWB_RSV_STATE_O_MOVE_EXPANDING:
304 		case UWB_RSV_STATE_O_MOVE_COMBINING:
305 		case UWB_RSV_STATE_O_MOVE_REDUCING:
306 		case UWB_RSV_STATE_T_ACCEPTED:
307 		case UWB_RSV_STATE_T_EXPANDING_ACCEPTED:
308 		case UWB_RSV_STATE_T_RESIZED:
309 		case UWB_RSV_STATE_NONE:
310 			uwb_rsv_stroke_timer(rsv);
311 			break;
312 		default:
313 			/* Expecting a state transition so leave timer
314 			   as-is. */
315 			break;
316 		}
317 		return;
318 	}
319 
320 	uwb_rsv_dump("SC", rsv);
321 
322 	switch (new_state) {
323 	case UWB_RSV_STATE_NONE:
324 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_NONE);
325 		uwb_rsv_callback(rsv);
326 		break;
327 	case UWB_RSV_STATE_O_INITIATED:
328 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_INITIATED);
329 		break;
330 	case UWB_RSV_STATE_O_PENDING:
331 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_PENDING);
332 		break;
333 	case UWB_RSV_STATE_O_MODIFIED:
334 		/* in the companion there are the MASes to drop */
335 		bitmap_andnot(rsv->mas.bm, rsv->mas.bm, mv->companion_mas.bm, UWB_NUM_MAS);
336 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MODIFIED);
337 		break;
338 	case UWB_RSV_STATE_O_ESTABLISHED:
339 		if (rsv->state == UWB_RSV_STATE_O_MODIFIED
340 		    || rsv->state == UWB_RSV_STATE_O_MOVE_REDUCING) {
341 			uwb_drp_avail_release(rsv->rc, &mv->companion_mas);
342 			rsv->needs_release_companion_mas = false;
343 		}
344 		uwb_drp_avail_reserve(rsv->rc, &rsv->mas);
345 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_ESTABLISHED);
346 		uwb_rsv_callback(rsv);
347 		break;
348 	case UWB_RSV_STATE_O_MOVE_EXPANDING:
349 		rsv->needs_release_companion_mas = true;
350 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MOVE_EXPANDING);
351 		break;
352 	case UWB_RSV_STATE_O_MOVE_COMBINING:
353 		rsv->needs_release_companion_mas = false;
354 		uwb_drp_avail_reserve(rsv->rc, &mv->companion_mas);
355 		bitmap_or(rsv->mas.bm, rsv->mas.bm, mv->companion_mas.bm, UWB_NUM_MAS);
356 		rsv->mas.safe   += mv->companion_mas.safe;
357 		rsv->mas.unsafe += mv->companion_mas.unsafe;
358 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MOVE_COMBINING);
359 		break;
360 	case UWB_RSV_STATE_O_MOVE_REDUCING:
361 		bitmap_andnot(mv->companion_mas.bm, rsv->mas.bm, mv->final_mas.bm, UWB_NUM_MAS);
362 		rsv->needs_release_companion_mas = true;
363 		rsv->mas.safe   = mv->final_mas.safe;
364 		rsv->mas.unsafe = mv->final_mas.unsafe;
365 		bitmap_copy(rsv->mas.bm, mv->final_mas.bm, UWB_NUM_MAS);
366 		bitmap_copy(rsv->mas.unsafe_bm, mv->final_mas.unsafe_bm, UWB_NUM_MAS);
367 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_O_MOVE_REDUCING);
368 		break;
369 	case UWB_RSV_STATE_T_ACCEPTED:
370 	case UWB_RSV_STATE_T_RESIZED:
371 		rsv->needs_release_companion_mas = false;
372 		uwb_drp_avail_reserve(rsv->rc, &rsv->mas);
373 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_ACCEPTED);
374 		uwb_rsv_callback(rsv);
375 		break;
376 	case UWB_RSV_STATE_T_DENIED:
377 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_DENIED);
378 		break;
379 	case UWB_RSV_STATE_T_CONFLICT:
380 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_CONFLICT);
381 		break;
382 	case UWB_RSV_STATE_T_PENDING:
383 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_PENDING);
384 		break;
385 	case UWB_RSV_STATE_T_EXPANDING_ACCEPTED:
386 		rsv->needs_release_companion_mas = true;
387 		uwb_drp_avail_reserve(rsv->rc, &mv->companion_mas);
388 		uwb_rsv_state_update(rsv, UWB_RSV_STATE_T_EXPANDING_ACCEPTED);
389 		break;
390 	default:
391 		dev_err(&rsv->rc->uwb_dev.dev, "unhandled state: %s (%d)\n",
392 			uwb_rsv_state_str(new_state), new_state);
393 	}
394 }
395 
uwb_rsv_handle_timeout_work(struct work_struct * work)396 static void uwb_rsv_handle_timeout_work(struct work_struct *work)
397 {
398 	struct uwb_rsv *rsv = container_of(work, struct uwb_rsv,
399 					   handle_timeout_work);
400 	struct uwb_rc *rc = rsv->rc;
401 
402 	mutex_lock(&rc->rsvs_mutex);
403 
404 	uwb_rsv_dump("TO", rsv);
405 
406 	switch (rsv->state) {
407 	case UWB_RSV_STATE_O_INITIATED:
408 		if (rsv->is_multicast) {
409 			uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_ESTABLISHED);
410 			goto unlock;
411 		}
412 		break;
413 	case UWB_RSV_STATE_O_MOVE_EXPANDING:
414 		if (rsv->is_multicast) {
415 			uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_MOVE_COMBINING);
416 			goto unlock;
417 		}
418 		break;
419 	case UWB_RSV_STATE_O_MOVE_COMBINING:
420 		if (rsv->is_multicast) {
421 			uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_MOVE_REDUCING);
422 			goto unlock;
423 		}
424 		break;
425 	case UWB_RSV_STATE_O_MOVE_REDUCING:
426 		if (rsv->is_multicast) {
427 			uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_ESTABLISHED);
428 			goto unlock;
429 		}
430 		break;
431 	case UWB_RSV_STATE_O_ESTABLISHED:
432 		if (rsv->is_multicast)
433 			goto unlock;
434 		break;
435 	case UWB_RSV_STATE_T_EXPANDING_ACCEPTED:
436 		/*
437 		 * The time out could be for the main or of the
438 		 * companion DRP, assume it's for the companion and
439 		 * drop that first.  A further time out is required to
440 		 * drop the main.
441 		 */
442 		uwb_rsv_set_state(rsv, UWB_RSV_STATE_T_ACCEPTED);
443 		uwb_drp_avail_release(rsv->rc, &rsv->mv.companion_mas);
444 		goto unlock;
445 	default:
446 		break;
447 	}
448 
449 	uwb_rsv_remove(rsv);
450 
451 unlock:
452 	mutex_unlock(&rc->rsvs_mutex);
453 }
454 
uwb_rsv_alloc(struct uwb_rc * rc)455 static struct uwb_rsv *uwb_rsv_alloc(struct uwb_rc *rc)
456 {
457 	struct uwb_rsv *rsv;
458 
459 	rsv = kzalloc(sizeof(struct uwb_rsv), GFP_KERNEL);
460 	if (!rsv)
461 		return NULL;
462 
463 	INIT_LIST_HEAD(&rsv->rc_node);
464 	INIT_LIST_HEAD(&rsv->pal_node);
465 	kref_init(&rsv->kref);
466 	init_timer(&rsv->timer);
467 	rsv->timer.function = uwb_rsv_timer;
468 	rsv->timer.data     = (unsigned long)rsv;
469 
470 	rsv->rc = rc;
471 	INIT_WORK(&rsv->handle_timeout_work, uwb_rsv_handle_timeout_work);
472 
473 	return rsv;
474 }
475 
476 /**
477  * uwb_rsv_create - allocate and initialize a UWB reservation structure
478  * @rc: the radio controller
479  * @cb: callback to use when the reservation completes or terminates
480  * @pal_priv: data private to the PAL to be passed in the callback
481  *
482  * The callback is called when the state of the reservation changes from:
483  *
484  *   - pending to accepted
485  *   - pending to denined
486  *   - accepted to terminated
487  *   - pending to terminated
488  */
uwb_rsv_create(struct uwb_rc * rc,uwb_rsv_cb_f cb,void * pal_priv)489 struct uwb_rsv *uwb_rsv_create(struct uwb_rc *rc, uwb_rsv_cb_f cb, void *pal_priv)
490 {
491 	struct uwb_rsv *rsv;
492 
493 	rsv = uwb_rsv_alloc(rc);
494 	if (!rsv)
495 		return NULL;
496 
497 	rsv->callback = cb;
498 	rsv->pal_priv = pal_priv;
499 
500 	return rsv;
501 }
502 EXPORT_SYMBOL_GPL(uwb_rsv_create);
503 
uwb_rsv_remove(struct uwb_rsv * rsv)504 void uwb_rsv_remove(struct uwb_rsv *rsv)
505 {
506 	uwb_rsv_dump("RM", rsv);
507 
508 	if (rsv->state != UWB_RSV_STATE_NONE)
509 		uwb_rsv_set_state(rsv, UWB_RSV_STATE_NONE);
510 
511 	if (rsv->needs_release_companion_mas)
512 		uwb_drp_avail_release(rsv->rc, &rsv->mv.companion_mas);
513 	uwb_drp_avail_release(rsv->rc, &rsv->mas);
514 
515 	if (uwb_rsv_is_owner(rsv))
516 		uwb_rsv_put_stream(rsv);
517 
518 	uwb_dev_put(rsv->owner);
519 	if (rsv->target.type == UWB_RSV_TARGET_DEV)
520 		uwb_dev_put(rsv->target.dev);
521 
522 	list_del_init(&rsv->rc_node);
523 	uwb_rsv_put(rsv);
524 }
525 
526 /**
527  * uwb_rsv_destroy - free a UWB reservation structure
528  * @rsv: the reservation to free
529  *
530  * The reservation must already be terminated.
531  */
uwb_rsv_destroy(struct uwb_rsv * rsv)532 void uwb_rsv_destroy(struct uwb_rsv *rsv)
533 {
534 	uwb_rsv_put(rsv);
535 }
536 EXPORT_SYMBOL_GPL(uwb_rsv_destroy);
537 
538 /**
539  * usb_rsv_establish - start a reservation establishment
540  * @rsv: the reservation
541  *
542  * The PAL should fill in @rsv's owner, target, type, max_mas,
543  * min_mas, max_interval and is_multicast fields.  If the target is a
544  * uwb_dev it must be referenced.
545  *
546  * The reservation's callback will be called when the reservation is
547  * accepted, denied or times out.
548  */
uwb_rsv_establish(struct uwb_rsv * rsv)549 int uwb_rsv_establish(struct uwb_rsv *rsv)
550 {
551 	struct uwb_rc *rc = rsv->rc;
552 	struct uwb_mas_bm available;
553 	int ret;
554 
555 	mutex_lock(&rc->rsvs_mutex);
556 	ret = uwb_rsv_get_stream(rsv);
557 	if (ret)
558 		goto out;
559 
560 	rsv->tiebreaker = random32() & 1;
561 	/* get available mas bitmap */
562 	uwb_drp_available(rc, &available);
563 
564 	ret = uwb_rsv_find_best_allocation(rsv, &available, &rsv->mas);
565 	if (ret == UWB_RSV_ALLOC_NOT_FOUND) {
566 		ret = -EBUSY;
567 		uwb_rsv_put_stream(rsv);
568 		goto out;
569 	}
570 
571 	ret = uwb_drp_avail_reserve_pending(rc, &rsv->mas);
572 	if (ret != 0) {
573 		uwb_rsv_put_stream(rsv);
574 		goto out;
575 	}
576 
577 	uwb_rsv_get(rsv);
578 	list_add_tail(&rsv->rc_node, &rc->reservations);
579 	rsv->owner = &rc->uwb_dev;
580 	uwb_dev_get(rsv->owner);
581 	uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_INITIATED);
582 out:
583 	mutex_unlock(&rc->rsvs_mutex);
584 	return ret;
585 }
586 EXPORT_SYMBOL_GPL(uwb_rsv_establish);
587 
588 /**
589  * uwb_rsv_modify - modify an already established reservation
590  * @rsv: the reservation to modify
591  * @max_mas: new maximum MAS to reserve
592  * @min_mas: new minimum MAS to reserve
593  * @max_interval: new max_interval to use
594  *
595  * FIXME: implement this once there are PALs that use it.
596  */
uwb_rsv_modify(struct uwb_rsv * rsv,int max_mas,int min_mas,int max_interval)597 int uwb_rsv_modify(struct uwb_rsv *rsv, int max_mas, int min_mas, int max_interval)
598 {
599 	return -ENOSYS;
600 }
601 EXPORT_SYMBOL_GPL(uwb_rsv_modify);
602 
603 /*
604  * move an already established reservation (rc->rsvs_mutex must to be
605  * taken when tis function is called)
606  */
uwb_rsv_try_move(struct uwb_rsv * rsv,struct uwb_mas_bm * available)607 int uwb_rsv_try_move(struct uwb_rsv *rsv, struct uwb_mas_bm *available)
608 {
609 	struct uwb_rc *rc = rsv->rc;
610 	struct uwb_drp_backoff_win *bow = &rc->bow;
611 	struct device *dev = &rc->uwb_dev.dev;
612 	struct uwb_rsv_move *mv;
613 	int ret = 0;
614 
615 	if (bow->can_reserve_extra_mases == false)
616 		return -EBUSY;
617 
618 	mv = &rsv->mv;
619 
620 	if (uwb_rsv_find_best_allocation(rsv, available, &mv->final_mas) == UWB_RSV_ALLOC_FOUND) {
621 
622 		if (!bitmap_equal(rsv->mas.bm, mv->final_mas.bm, UWB_NUM_MAS)) {
623 			/* We want to move the reservation */
624 			bitmap_andnot(mv->companion_mas.bm, mv->final_mas.bm, rsv->mas.bm, UWB_NUM_MAS);
625 			uwb_drp_avail_reserve_pending(rc, &mv->companion_mas);
626 			uwb_rsv_set_state(rsv, UWB_RSV_STATE_O_MOVE_EXPANDING);
627 		}
628 	} else {
629 		dev_dbg(dev, "new allocation not found\n");
630 	}
631 
632 	return ret;
633 }
634 
635 /* It will try to move every reservation in state O_ESTABLISHED giving
636  * to the MAS allocator algorithm an availability that is the real one
637  * plus the allocation already established from the reservation. */
uwb_rsv_handle_drp_avail_change(struct uwb_rc * rc)638 void uwb_rsv_handle_drp_avail_change(struct uwb_rc *rc)
639 {
640 	struct uwb_drp_backoff_win *bow = &rc->bow;
641 	struct uwb_rsv *rsv;
642 	struct uwb_mas_bm mas;
643 
644 	if (bow->can_reserve_extra_mases == false)
645 		return;
646 
647 	list_for_each_entry(rsv, &rc->reservations, rc_node) {
648 		if (rsv->state == UWB_RSV_STATE_O_ESTABLISHED ||
649 		    rsv->state == UWB_RSV_STATE_O_TO_BE_MOVED) {
650 			uwb_drp_available(rc, &mas);
651 			bitmap_or(mas.bm, mas.bm, rsv->mas.bm, UWB_NUM_MAS);
652 			uwb_rsv_try_move(rsv, &mas);
653 		}
654 	}
655 
656 }
657 
658 /**
659  * uwb_rsv_terminate - terminate an established reservation
660  * @rsv: the reservation to terminate
661  *
662  * A reservation is terminated by removing the DRP IE from the beacon,
663  * the other end will consider the reservation to be terminated when
664  * it does not see the DRP IE for at least mMaxLostBeacons.
665  *
666  * If applicable, the reference to the target uwb_dev will be released.
667  */
uwb_rsv_terminate(struct uwb_rsv * rsv)668 void uwb_rsv_terminate(struct uwb_rsv *rsv)
669 {
670 	struct uwb_rc *rc = rsv->rc;
671 
672 	mutex_lock(&rc->rsvs_mutex);
673 
674 	if (rsv->state != UWB_RSV_STATE_NONE)
675 		uwb_rsv_set_state(rsv, UWB_RSV_STATE_NONE);
676 
677 	mutex_unlock(&rc->rsvs_mutex);
678 }
679 EXPORT_SYMBOL_GPL(uwb_rsv_terminate);
680 
681 /**
682  * uwb_rsv_accept - accept a new reservation from a peer
683  * @rsv:      the reservation
684  * @cb:       call back for reservation changes
685  * @pal_priv: data to be passed in the above call back
686  *
687  * Reservation requests from peers are denied unless a PAL accepts it
688  * by calling this function.
689  *
690  * The PAL call uwb_rsv_destroy() for all accepted reservations before
691  * calling uwb_pal_unregister().
692  */
uwb_rsv_accept(struct uwb_rsv * rsv,uwb_rsv_cb_f cb,void * pal_priv)693 void uwb_rsv_accept(struct uwb_rsv *rsv, uwb_rsv_cb_f cb, void *pal_priv)
694 {
695 	uwb_rsv_get(rsv);
696 
697 	rsv->callback = cb;
698 	rsv->pal_priv = pal_priv;
699 	rsv->state    = UWB_RSV_STATE_T_ACCEPTED;
700 }
701 EXPORT_SYMBOL_GPL(uwb_rsv_accept);
702 
703 /*
704  * Is a received DRP IE for this reservation?
705  */
uwb_rsv_match(struct uwb_rsv * rsv,struct uwb_dev * src,struct uwb_ie_drp * drp_ie)706 static bool uwb_rsv_match(struct uwb_rsv *rsv, struct uwb_dev *src,
707 			  struct uwb_ie_drp *drp_ie)
708 {
709 	struct uwb_dev_addr *rsv_src;
710 	int stream;
711 
712 	stream = uwb_ie_drp_stream_index(drp_ie);
713 
714 	if (rsv->stream != stream)
715 		return false;
716 
717 	switch (rsv->target.type) {
718 	case UWB_RSV_TARGET_DEVADDR:
719 		return rsv->stream == stream;
720 	case UWB_RSV_TARGET_DEV:
721 		if (uwb_ie_drp_owner(drp_ie))
722 			rsv_src = &rsv->owner->dev_addr;
723 		else
724 			rsv_src = &rsv->target.dev->dev_addr;
725 		return uwb_dev_addr_cmp(&src->dev_addr, rsv_src) == 0;
726 	}
727 	return false;
728 }
729 
uwb_rsv_new_target(struct uwb_rc * rc,struct uwb_dev * src,struct uwb_ie_drp * drp_ie)730 static struct uwb_rsv *uwb_rsv_new_target(struct uwb_rc *rc,
731 					  struct uwb_dev *src,
732 					  struct uwb_ie_drp *drp_ie)
733 {
734 	struct uwb_rsv *rsv;
735 	struct uwb_pal *pal;
736 	enum uwb_rsv_state state;
737 
738 	rsv = uwb_rsv_alloc(rc);
739 	if (!rsv)
740 		return NULL;
741 
742 	rsv->rc          = rc;
743 	rsv->owner       = src;
744 	uwb_dev_get(rsv->owner);
745 	rsv->target.type = UWB_RSV_TARGET_DEV;
746 	rsv->target.dev  = &rc->uwb_dev;
747 	uwb_dev_get(&rc->uwb_dev);
748 	rsv->type        = uwb_ie_drp_type(drp_ie);
749 	rsv->stream      = uwb_ie_drp_stream_index(drp_ie);
750 	uwb_drp_ie_to_bm(&rsv->mas, drp_ie);
751 
752 	/*
753 	 * See if any PALs are interested in this reservation. If not,
754 	 * deny the request.
755 	 */
756 	rsv->state = UWB_RSV_STATE_T_DENIED;
757 	mutex_lock(&rc->uwb_dev.mutex);
758 	list_for_each_entry(pal, &rc->pals, node) {
759 		if (pal->new_rsv)
760 			pal->new_rsv(pal, rsv);
761 		if (rsv->state == UWB_RSV_STATE_T_ACCEPTED)
762 			break;
763 	}
764 	mutex_unlock(&rc->uwb_dev.mutex);
765 
766 	list_add_tail(&rsv->rc_node, &rc->reservations);
767 	state = rsv->state;
768 	rsv->state = UWB_RSV_STATE_NONE;
769 
770 	/* FIXME: do something sensible here */
771 	if (state == UWB_RSV_STATE_T_ACCEPTED
772 	    && uwb_drp_avail_reserve_pending(rc, &rsv->mas) == -EBUSY) {
773 		/* FIXME: do something sensible here */
774 	} else {
775 		uwb_rsv_set_state(rsv, state);
776 	}
777 
778 	return rsv;
779 }
780 
781 /**
782  * uwb_rsv_get_usable_mas - get the bitmap of the usable MAS of a reservations
783  * @rsv: the reservation.
784  * @mas: returns the available MAS.
785  *
786  * The usable MAS of a reservation may be less than the negotiated MAS
787  * if alien BPs are present.
788  */
uwb_rsv_get_usable_mas(struct uwb_rsv * rsv,struct uwb_mas_bm * mas)789 void uwb_rsv_get_usable_mas(struct uwb_rsv *rsv, struct uwb_mas_bm *mas)
790 {
791 	bitmap_zero(mas->bm, UWB_NUM_MAS);
792 	bitmap_andnot(mas->bm, rsv->mas.bm, rsv->rc->cnflt_alien_bitmap.bm, UWB_NUM_MAS);
793 }
794 EXPORT_SYMBOL_GPL(uwb_rsv_get_usable_mas);
795 
796 /**
797  * uwb_rsv_find - find a reservation for a received DRP IE.
798  * @rc: the radio controller
799  * @src: source of the DRP IE
800  * @drp_ie: the DRP IE
801  *
802  * If the reservation cannot be found and the DRP IE is from a peer
803  * attempting to establish a new reservation, create a new reservation
804  * and add it to the list.
805  */
uwb_rsv_find(struct uwb_rc * rc,struct uwb_dev * src,struct uwb_ie_drp * drp_ie)806 struct uwb_rsv *uwb_rsv_find(struct uwb_rc *rc, struct uwb_dev *src,
807 			     struct uwb_ie_drp *drp_ie)
808 {
809 	struct uwb_rsv *rsv;
810 
811 	list_for_each_entry(rsv, &rc->reservations, rc_node) {
812 		if (uwb_rsv_match(rsv, src, drp_ie))
813 			return rsv;
814 	}
815 
816 	if (uwb_ie_drp_owner(drp_ie))
817 		return uwb_rsv_new_target(rc, src, drp_ie);
818 
819 	return NULL;
820 }
821 
822 /*
823  * Go through all the reservations and check for timeouts and (if
824  * necessary) update their DRP IEs.
825  *
826  * FIXME: look at building the SET_DRP_IE command here rather than
827  * having to rescan the list in uwb_rc_send_all_drp_ie().
828  */
uwb_rsv_update_all(struct uwb_rc * rc)829 static bool uwb_rsv_update_all(struct uwb_rc *rc)
830 {
831 	struct uwb_rsv *rsv, *t;
832 	bool ie_updated = false;
833 
834 	list_for_each_entry_safe(rsv, t, &rc->reservations, rc_node) {
835 		if (!rsv->ie_valid) {
836 			uwb_drp_ie_update(rsv);
837 			ie_updated = true;
838 		}
839 	}
840 
841 	return ie_updated;
842 }
843 
uwb_rsv_queue_update(struct uwb_rc * rc)844 void uwb_rsv_queue_update(struct uwb_rc *rc)
845 {
846 	unsigned long delay_us = UWB_MAS_LENGTH_US * UWB_MAS_PER_ZONE;
847 
848 	queue_delayed_work(rc->rsv_workq, &rc->rsv_update_work, usecs_to_jiffies(delay_us));
849 }
850 
851 /**
852  * uwb_rsv_sched_update - schedule an update of the DRP IEs
853  * @rc: the radio controller.
854  *
855  * To improve performance and ensure correctness with [ECMA-368] the
856  * number of SET-DRP-IE commands that are done are limited.
857  *
858  * DRP IEs update come from two sources: DRP events from the hardware
859  * which all occur at the beginning of the superframe ('syncronous'
860  * events) and reservation establishment/termination requests from
861  * PALs or timers ('asynchronous' events).
862  *
863  * A delayed work ensures that all the synchronous events result in
864  * one SET-DRP-IE command.
865  *
866  * Additional logic (the set_drp_ie_pending and rsv_updated_postponed
867  * flags) will prevent an asynchrous event starting a SET-DRP-IE
868  * command if one is currently awaiting a response.
869  *
870  * FIXME: this does leave a window where an asynchrous event can delay
871  * the SET-DRP-IE for a synchronous event by one superframe.
872  */
uwb_rsv_sched_update(struct uwb_rc * rc)873 void uwb_rsv_sched_update(struct uwb_rc *rc)
874 {
875 	spin_lock_bh(&rc->rsvs_lock);
876 	if (!delayed_work_pending(&rc->rsv_update_work)) {
877 		if (rc->set_drp_ie_pending > 0) {
878 			rc->set_drp_ie_pending++;
879 			goto unlock;
880 		}
881 		uwb_rsv_queue_update(rc);
882 	}
883 unlock:
884 	spin_unlock_bh(&rc->rsvs_lock);
885 }
886 
887 /*
888  * Update DRP IEs and, if necessary, the DRP Availability IE and send
889  * the updated IEs to the radio controller.
890  */
uwb_rsv_update_work(struct work_struct * work)891 static void uwb_rsv_update_work(struct work_struct *work)
892 {
893 	struct uwb_rc *rc = container_of(work, struct uwb_rc,
894 					 rsv_update_work.work);
895 	bool ie_updated;
896 
897 	mutex_lock(&rc->rsvs_mutex);
898 
899 	ie_updated = uwb_rsv_update_all(rc);
900 
901 	if (!rc->drp_avail.ie_valid) {
902 		uwb_drp_avail_ie_update(rc);
903 		ie_updated = true;
904 	}
905 
906 	if (ie_updated && (rc->set_drp_ie_pending == 0))
907 		uwb_rc_send_all_drp_ie(rc);
908 
909 	mutex_unlock(&rc->rsvs_mutex);
910 }
911 
uwb_rsv_alien_bp_work(struct work_struct * work)912 static void uwb_rsv_alien_bp_work(struct work_struct *work)
913 {
914 	struct uwb_rc *rc = container_of(work, struct uwb_rc,
915 					 rsv_alien_bp_work.work);
916 	struct uwb_rsv *rsv;
917 
918 	mutex_lock(&rc->rsvs_mutex);
919 
920 	list_for_each_entry(rsv, &rc->reservations, rc_node) {
921 		if (rsv->type != UWB_DRP_TYPE_ALIEN_BP) {
922 			rsv->callback(rsv);
923 		}
924 	}
925 
926 	mutex_unlock(&rc->rsvs_mutex);
927 }
928 
uwb_rsv_timer(unsigned long arg)929 static void uwb_rsv_timer(unsigned long arg)
930 {
931 	struct uwb_rsv *rsv = (struct uwb_rsv *)arg;
932 
933 	queue_work(rsv->rc->rsv_workq, &rsv->handle_timeout_work);
934 }
935 
936 /**
937  * uwb_rsv_remove_all - remove all reservations
938  * @rc: the radio controller
939  *
940  * A DRP IE update is not done.
941  */
uwb_rsv_remove_all(struct uwb_rc * rc)942 void uwb_rsv_remove_all(struct uwb_rc *rc)
943 {
944 	struct uwb_rsv *rsv, *t;
945 
946 	mutex_lock(&rc->rsvs_mutex);
947 	list_for_each_entry_safe(rsv, t, &rc->reservations, rc_node) {
948 		if (rsv->state != UWB_RSV_STATE_NONE)
949 			uwb_rsv_set_state(rsv, UWB_RSV_STATE_NONE);
950 		del_timer_sync(&rsv->timer);
951 	}
952 	/* Cancel any postponed update. */
953 	rc->set_drp_ie_pending = 0;
954 	mutex_unlock(&rc->rsvs_mutex);
955 
956 	cancel_delayed_work_sync(&rc->rsv_update_work);
957 	flush_workqueue(rc->rsv_workq);
958 
959 	mutex_lock(&rc->rsvs_mutex);
960 	list_for_each_entry_safe(rsv, t, &rc->reservations, rc_node) {
961 		uwb_rsv_remove(rsv);
962 	}
963 	mutex_unlock(&rc->rsvs_mutex);
964 }
965 
uwb_rsv_init(struct uwb_rc * rc)966 void uwb_rsv_init(struct uwb_rc *rc)
967 {
968 	INIT_LIST_HEAD(&rc->reservations);
969 	INIT_LIST_HEAD(&rc->cnflt_alien_list);
970 	mutex_init(&rc->rsvs_mutex);
971 	spin_lock_init(&rc->rsvs_lock);
972 	INIT_DELAYED_WORK(&rc->rsv_update_work, uwb_rsv_update_work);
973 	INIT_DELAYED_WORK(&rc->rsv_alien_bp_work, uwb_rsv_alien_bp_work);
974 	rc->bow.can_reserve_extra_mases = true;
975 	rc->bow.total_expired = 0;
976 	rc->bow.window = UWB_DRP_BACKOFF_WIN_MIN >> 1;
977 	init_timer(&rc->bow.timer);
978 	rc->bow.timer.function = uwb_rsv_backoff_win_timer;
979 	rc->bow.timer.data     = (unsigned long)&rc->bow;
980 
981 	bitmap_complement(rc->uwb_dev.streams, rc->uwb_dev.streams, UWB_NUM_STREAMS);
982 }
983 
uwb_rsv_setup(struct uwb_rc * rc)984 int uwb_rsv_setup(struct uwb_rc *rc)
985 {
986 	char name[16];
987 
988 	snprintf(name, sizeof(name), "%s_rsvd", dev_name(&rc->uwb_dev.dev));
989 	rc->rsv_workq = create_singlethread_workqueue(name);
990 	if (rc->rsv_workq == NULL)
991 		return -ENOMEM;
992 
993 	return 0;
994 }
995 
uwb_rsv_cleanup(struct uwb_rc * rc)996 void uwb_rsv_cleanup(struct uwb_rc *rc)
997 {
998 	uwb_rsv_remove_all(rc);
999 	destroy_workqueue(rc->rsv_workq);
1000 }
1001