• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /* FS-Cache object state machine handler
2  *
3  * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
4  * Written by David Howells (dhowells@redhat.com)
5  *
6  * This program is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU General Public License
8  * as published by the Free Software Foundation; either version
9  * 2 of the License, or (at your option) any later version.
10  *
11  * See Documentation/filesystems/caching/object.txt for a description of the
12  * object state machine and the in-kernel representations.
13  */
14 
15 #define FSCACHE_DEBUG_LEVEL COOKIE
16 #include <linux/module.h>
17 #include <linux/slab.h>
18 #include <linux/prefetch.h>
19 #include "internal.h"
20 
21 static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *, int);
22 static const struct fscache_state *fscache_kill_dependents(struct fscache_object *, int);
23 static const struct fscache_state *fscache_drop_object(struct fscache_object *, int);
24 static const struct fscache_state *fscache_initialise_object(struct fscache_object *, int);
25 static const struct fscache_state *fscache_invalidate_object(struct fscache_object *, int);
26 static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *, int);
27 static const struct fscache_state *fscache_kill_object(struct fscache_object *, int);
28 static const struct fscache_state *fscache_lookup_failure(struct fscache_object *, int);
29 static const struct fscache_state *fscache_look_up_object(struct fscache_object *, int);
30 static const struct fscache_state *fscache_object_available(struct fscache_object *, int);
31 static const struct fscache_state *fscache_parent_ready(struct fscache_object *, int);
32 static const struct fscache_state *fscache_update_object(struct fscache_object *, int);
33 static const struct fscache_state *fscache_object_dead(struct fscache_object *, int);
34 
35 #define __STATE_NAME(n) fscache_osm_##n
36 #define STATE(n) (&__STATE_NAME(n))
37 
38 /*
39  * Define a work state.  Work states are execution states.  No event processing
40  * is performed by them.  The function attached to a work state returns a
41  * pointer indicating the next state to which the state machine should
42  * transition.  Returning NO_TRANSIT repeats the current state, but goes back
43  * to the scheduler first.
44  */
45 #define WORK_STATE(n, sn, f) \
46 	const struct fscache_state __STATE_NAME(n) = {			\
47 		.name = #n,						\
48 		.short_name = sn,					\
49 		.work = f						\
50 	}
51 
52 /*
53  * Returns from work states.
54  */
55 #define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); })
56 
57 #define NO_TRANSIT ((struct fscache_state *)NULL)
58 
59 /*
60  * Define a wait state.  Wait states are event processing states.  No execution
61  * is performed by them.  Wait states are just tables of "if event X occurs,
62  * clear it and transition to state Y".  The dispatcher returns to the
63  * scheduler if none of the events in which the wait state has an interest are
64  * currently pending.
65  */
66 #define WAIT_STATE(n, sn, ...) \
67 	const struct fscache_state __STATE_NAME(n) = {			\
68 		.name = #n,						\
69 		.short_name = sn,					\
70 		.work = NULL,						\
71 		.transitions = { __VA_ARGS__, { 0, NULL } }		\
72 	}
73 
74 #define TRANSIT_TO(state, emask) \
75 	{ .events = (emask), .transit_to = STATE(state) }
76 
77 /*
78  * The object state machine.
79  */
80 static WORK_STATE(INIT_OBJECT,		"INIT", fscache_initialise_object);
81 static WORK_STATE(PARENT_READY,		"PRDY", fscache_parent_ready);
82 static WORK_STATE(ABORT_INIT,		"ABRT", fscache_abort_initialisation);
83 static WORK_STATE(LOOK_UP_OBJECT,	"LOOK", fscache_look_up_object);
84 static WORK_STATE(CREATE_OBJECT,	"CRTO", fscache_look_up_object);
85 static WORK_STATE(OBJECT_AVAILABLE,	"AVBL", fscache_object_available);
86 static WORK_STATE(JUMPSTART_DEPS,	"JUMP", fscache_jumpstart_dependents);
87 
88 static WORK_STATE(INVALIDATE_OBJECT,	"INVL", fscache_invalidate_object);
89 static WORK_STATE(UPDATE_OBJECT,	"UPDT", fscache_update_object);
90 
91 static WORK_STATE(LOOKUP_FAILURE,	"LCFL", fscache_lookup_failure);
92 static WORK_STATE(KILL_OBJECT,		"KILL", fscache_kill_object);
93 static WORK_STATE(KILL_DEPENDENTS,	"KDEP", fscache_kill_dependents);
94 static WORK_STATE(DROP_OBJECT,		"DROP", fscache_drop_object);
95 static WORK_STATE(OBJECT_DEAD,		"DEAD", fscache_object_dead);
96 
97 static WAIT_STATE(WAIT_FOR_INIT,	"?INI",
98 		  TRANSIT_TO(INIT_OBJECT,	1 << FSCACHE_OBJECT_EV_NEW_CHILD));
99 
100 static WAIT_STATE(WAIT_FOR_PARENT,	"?PRN",
101 		  TRANSIT_TO(PARENT_READY,	1 << FSCACHE_OBJECT_EV_PARENT_READY));
102 
103 static WAIT_STATE(WAIT_FOR_CMD,		"?CMD",
104 		  TRANSIT_TO(INVALIDATE_OBJECT,	1 << FSCACHE_OBJECT_EV_INVALIDATE),
105 		  TRANSIT_TO(UPDATE_OBJECT,	1 << FSCACHE_OBJECT_EV_UPDATE),
106 		  TRANSIT_TO(JUMPSTART_DEPS,	1 << FSCACHE_OBJECT_EV_NEW_CHILD));
107 
108 static WAIT_STATE(WAIT_FOR_CLEARANCE,	"?CLR",
109 		  TRANSIT_TO(KILL_OBJECT,	1 << FSCACHE_OBJECT_EV_CLEARED));
110 
111 /*
112  * Out-of-band event transition tables.  These are for handling unexpected
113  * events, such as an I/O error.  If an OOB event occurs, the state machine
114  * clears and disables the event and forces a transition to the nominated work
115  * state (acurrently executing work states will complete first).
116  *
117  * In such a situation, object->state remembers the state the machine should
118  * have been in/gone to and returning NO_TRANSIT returns to that.
119  */
120 static const struct fscache_transition fscache_osm_init_oob[] = {
121 	   TRANSIT_TO(ABORT_INIT,
122 		      (1 << FSCACHE_OBJECT_EV_ERROR) |
123 		      (1 << FSCACHE_OBJECT_EV_KILL)),
124 	   { 0, NULL }
125 };
126 
127 static const struct fscache_transition fscache_osm_lookup_oob[] = {
128 	   TRANSIT_TO(LOOKUP_FAILURE,
129 		      (1 << FSCACHE_OBJECT_EV_ERROR) |
130 		      (1 << FSCACHE_OBJECT_EV_KILL)),
131 	   { 0, NULL }
132 };
133 
134 static const struct fscache_transition fscache_osm_run_oob[] = {
135 	   TRANSIT_TO(KILL_OBJECT,
136 		      (1 << FSCACHE_OBJECT_EV_ERROR) |
137 		      (1 << FSCACHE_OBJECT_EV_KILL)),
138 	   { 0, NULL }
139 };
140 
141 static int  fscache_get_object(struct fscache_object *);
142 static void fscache_put_object(struct fscache_object *);
143 static bool fscache_enqueue_dependents(struct fscache_object *, int);
144 static void fscache_dequeue_object(struct fscache_object *);
145 
146 /*
147  * we need to notify the parent when an op completes that we had outstanding
148  * upon it
149  */
fscache_done_parent_op(struct fscache_object * object)150 static inline void fscache_done_parent_op(struct fscache_object *object)
151 {
152 	struct fscache_object *parent = object->parent;
153 
154 	_enter("OBJ%x {OBJ%x,%x}",
155 	       object->debug_id, parent->debug_id, parent->n_ops);
156 
157 	spin_lock_nested(&parent->lock, 1);
158 	parent->n_obj_ops--;
159 	parent->n_ops--;
160 	if (parent->n_ops == 0)
161 		fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
162 	spin_unlock(&parent->lock);
163 }
164 
165 /*
166  * Object state machine dispatcher.
167  */
fscache_object_sm_dispatcher(struct fscache_object * object)168 static void fscache_object_sm_dispatcher(struct fscache_object *object)
169 {
170 	const struct fscache_transition *t;
171 	const struct fscache_state *state, *new_state;
172 	unsigned long events, event_mask;
173 	int event = -1;
174 
175 	ASSERT(object != NULL);
176 
177 	_enter("{OBJ%x,%s,%lx}",
178 	       object->debug_id, object->state->name, object->events);
179 
180 	event_mask = object->event_mask;
181 restart:
182 	object->event_mask = 0; /* Mask normal event handling */
183 	state = object->state;
184 restart_masked:
185 	events = object->events;
186 
187 	/* Handle any out-of-band events (typically an error) */
188 	if (events & object->oob_event_mask) {
189 		_debug("{OBJ%x} oob %lx",
190 		       object->debug_id, events & object->oob_event_mask);
191 		for (t = object->oob_table; t->events; t++) {
192 			if (events & t->events) {
193 				state = t->transit_to;
194 				ASSERT(state->work != NULL);
195 				event = fls(events & t->events) - 1;
196 				__clear_bit(event, &object->oob_event_mask);
197 				clear_bit(event, &object->events);
198 				goto execute_work_state;
199 			}
200 		}
201 	}
202 
203 	/* Wait states are just transition tables */
204 	if (!state->work) {
205 		if (events & event_mask) {
206 			for (t = state->transitions; t->events; t++) {
207 				if (events & t->events) {
208 					new_state = t->transit_to;
209 					event = fls(events & t->events) - 1;
210 					clear_bit(event, &object->events);
211 					_debug("{OBJ%x} ev %d: %s -> %s",
212 					       object->debug_id, event,
213 					       state->name, new_state->name);
214 					object->state = state = new_state;
215 					goto execute_work_state;
216 				}
217 			}
218 
219 			/* The event mask didn't include all the tabled bits */
220 			BUG();
221 		}
222 		/* Randomly woke up */
223 		goto unmask_events;
224 	}
225 
226 execute_work_state:
227 	_debug("{OBJ%x} exec %s", object->debug_id, state->name);
228 
229 	new_state = state->work(object, event);
230 	event = -1;
231 	if (new_state == NO_TRANSIT) {
232 		_debug("{OBJ%x} %s notrans", object->debug_id, state->name);
233 		if (unlikely(state == STATE(OBJECT_DEAD))) {
234 			_leave(" [dead]");
235 			return;
236 		}
237 		fscache_enqueue_object(object);
238 		event_mask = object->oob_event_mask;
239 		goto unmask_events;
240 	}
241 
242 	_debug("{OBJ%x} %s -> %s",
243 	       object->debug_id, state->name, new_state->name);
244 	object->state = state = new_state;
245 
246 	if (state->work) {
247 		if (unlikely(state == STATE(OBJECT_DEAD))) {
248 			_leave(" [dead]");
249 			return;
250 		}
251 		goto restart_masked;
252 	}
253 
254 	/* Transited to wait state */
255 	event_mask = object->oob_event_mask;
256 	for (t = state->transitions; t->events; t++)
257 		event_mask |= t->events;
258 
259 unmask_events:
260 	object->event_mask = event_mask;
261 	smp_mb();
262 	events = object->events;
263 	if (events & event_mask)
264 		goto restart;
265 	_leave(" [msk %lx]", event_mask);
266 }
267 
268 /*
269  * execute an object
270  */
fscache_object_work_func(struct work_struct * work)271 static void fscache_object_work_func(struct work_struct *work)
272 {
273 	struct fscache_object *object =
274 		container_of(work, struct fscache_object, work);
275 	unsigned long start;
276 
277 	_enter("{OBJ%x}", object->debug_id);
278 
279 	start = jiffies;
280 	fscache_object_sm_dispatcher(object);
281 	fscache_hist(fscache_objs_histogram, start);
282 	fscache_put_object(object);
283 }
284 
285 /**
286  * fscache_object_init - Initialise a cache object description
287  * @object: Object description
288  * @cookie: Cookie object will be attached to
289  * @cache: Cache in which backing object will be found
290  *
291  * Initialise a cache object description to its basic values.
292  *
293  * See Documentation/filesystems/caching/backend-api.txt for a complete
294  * description.
295  */
fscache_object_init(struct fscache_object * object,struct fscache_cookie * cookie,struct fscache_cache * cache)296 void fscache_object_init(struct fscache_object *object,
297 			 struct fscache_cookie *cookie,
298 			 struct fscache_cache *cache)
299 {
300 	const struct fscache_transition *t;
301 
302 	atomic_inc(&cache->object_count);
303 
304 	object->state = STATE(WAIT_FOR_INIT);
305 	object->oob_table = fscache_osm_init_oob;
306 	object->flags = 1 << FSCACHE_OBJECT_IS_LIVE;
307 	spin_lock_init(&object->lock);
308 	INIT_LIST_HEAD(&object->cache_link);
309 	INIT_HLIST_NODE(&object->cookie_link);
310 	INIT_WORK(&object->work, fscache_object_work_func);
311 	INIT_LIST_HEAD(&object->dependents);
312 	INIT_LIST_HEAD(&object->dep_link);
313 	INIT_LIST_HEAD(&object->pending_ops);
314 	object->n_children = 0;
315 	object->n_ops = object->n_in_progress = object->n_exclusive = 0;
316 	object->events = 0;
317 	object->store_limit = 0;
318 	object->store_limit_l = 0;
319 	object->cache = cache;
320 	object->cookie = cookie;
321 	object->parent = NULL;
322 #ifdef CONFIG_FSCACHE_OBJECT_LIST
323 	RB_CLEAR_NODE(&object->objlist_link);
324 #endif
325 
326 	object->oob_event_mask = 0;
327 	for (t = object->oob_table; t->events; t++)
328 		object->oob_event_mask |= t->events;
329 	object->event_mask = object->oob_event_mask;
330 	for (t = object->state->transitions; t->events; t++)
331 		object->event_mask |= t->events;
332 }
333 EXPORT_SYMBOL(fscache_object_init);
334 
335 /*
336  * Mark the object as no longer being live, making sure that we synchronise
337  * against op submission.
338  */
fscache_mark_object_dead(struct fscache_object * object)339 static inline void fscache_mark_object_dead(struct fscache_object *object)
340 {
341 	spin_lock(&object->lock);
342 	clear_bit(FSCACHE_OBJECT_IS_LIVE, &object->flags);
343 	spin_unlock(&object->lock);
344 }
345 
346 /*
347  * Abort object initialisation before we start it.
348  */
fscache_abort_initialisation(struct fscache_object * object,int event)349 static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *object,
350 								int event)
351 {
352 	_enter("{OBJ%x},%d", object->debug_id, event);
353 
354 	object->oob_event_mask = 0;
355 	fscache_dequeue_object(object);
356 	return transit_to(KILL_OBJECT);
357 }
358 
359 /*
360  * initialise an object
361  * - check the specified object's parent to see if we can make use of it
362  *   immediately to do a creation
363  * - we may need to start the process of creating a parent and we need to wait
364  *   for the parent's lookup and creation to complete if it's not there yet
365  */
fscache_initialise_object(struct fscache_object * object,int event)366 static const struct fscache_state *fscache_initialise_object(struct fscache_object *object,
367 							     int event)
368 {
369 	struct fscache_object *parent;
370 	bool success;
371 
372 	_enter("{OBJ%x},%d", object->debug_id, event);
373 
374 	ASSERT(list_empty(&object->dep_link));
375 
376 	parent = object->parent;
377 	if (!parent) {
378 		_leave(" [no parent]");
379 		return transit_to(DROP_OBJECT);
380 	}
381 
382 	_debug("parent: %s of:%lx", parent->state->name, parent->flags);
383 
384 	if (fscache_object_is_dying(parent)) {
385 		_leave(" [bad parent]");
386 		return transit_to(DROP_OBJECT);
387 	}
388 
389 	if (fscache_object_is_available(parent)) {
390 		_leave(" [ready]");
391 		return transit_to(PARENT_READY);
392 	}
393 
394 	_debug("wait");
395 
396 	spin_lock(&parent->lock);
397 	fscache_stat(&fscache_n_cop_grab_object);
398 	success = false;
399 	if (fscache_object_is_live(parent) &&
400 	    object->cache->ops->grab_object(object)) {
401 		list_add(&object->dep_link, &parent->dependents);
402 		success = true;
403 	}
404 	fscache_stat_d(&fscache_n_cop_grab_object);
405 	spin_unlock(&parent->lock);
406 	if (!success) {
407 		_leave(" [grab failed]");
408 		return transit_to(DROP_OBJECT);
409 	}
410 
411 	/* fscache_acquire_non_index_cookie() uses this
412 	 * to wake the chain up */
413 	fscache_raise_event(parent, FSCACHE_OBJECT_EV_NEW_CHILD);
414 	_leave(" [wait]");
415 	return transit_to(WAIT_FOR_PARENT);
416 }
417 
418 /*
419  * Once the parent object is ready, we should kick off our lookup op.
420  */
fscache_parent_ready(struct fscache_object * object,int event)421 static const struct fscache_state *fscache_parent_ready(struct fscache_object *object,
422 							int event)
423 {
424 	struct fscache_object *parent = object->parent;
425 
426 	_enter("{OBJ%x},%d", object->debug_id, event);
427 
428 	ASSERT(parent != NULL);
429 
430 	spin_lock(&parent->lock);
431 	parent->n_ops++;
432 	parent->n_obj_ops++;
433 	object->lookup_jif = jiffies;
434 	spin_unlock(&parent->lock);
435 
436 	_leave("");
437 	return transit_to(LOOK_UP_OBJECT);
438 }
439 
440 /*
441  * look an object up in the cache from which it was allocated
442  * - we hold an "access lock" on the parent object, so the parent object cannot
443  *   be withdrawn by either party till we've finished
444  */
fscache_look_up_object(struct fscache_object * object,int event)445 static const struct fscache_state *fscache_look_up_object(struct fscache_object *object,
446 							  int event)
447 {
448 	struct fscache_cookie *cookie = object->cookie;
449 	struct fscache_object *parent = object->parent;
450 	int ret;
451 
452 	_enter("{OBJ%x},%d", object->debug_id, event);
453 
454 	object->oob_table = fscache_osm_lookup_oob;
455 
456 	ASSERT(parent != NULL);
457 	ASSERTCMP(parent->n_ops, >, 0);
458 	ASSERTCMP(parent->n_obj_ops, >, 0);
459 
460 	/* make sure the parent is still available */
461 	ASSERT(fscache_object_is_available(parent));
462 
463 	if (fscache_object_is_dying(parent) ||
464 	    test_bit(FSCACHE_IOERROR, &object->cache->flags) ||
465 	    !fscache_use_cookie(object)) {
466 		_leave(" [unavailable]");
467 		return transit_to(LOOKUP_FAILURE);
468 	}
469 
470 	_debug("LOOKUP \"%s\" in \"%s\"",
471 	       cookie->def->name, object->cache->tag->name);
472 
473 	fscache_stat(&fscache_n_object_lookups);
474 	fscache_stat(&fscache_n_cop_lookup_object);
475 	ret = object->cache->ops->lookup_object(object);
476 	fscache_stat_d(&fscache_n_cop_lookup_object);
477 
478 	fscache_unuse_cookie(object);
479 
480 	if (ret == -ETIMEDOUT) {
481 		/* probably stuck behind another object, so move this one to
482 		 * the back of the queue */
483 		fscache_stat(&fscache_n_object_lookups_timed_out);
484 		_leave(" [timeout]");
485 		return NO_TRANSIT;
486 	}
487 
488 	if (ret < 0) {
489 		_leave(" [error]");
490 		return transit_to(LOOKUP_FAILURE);
491 	}
492 
493 	_leave(" [ok]");
494 	return transit_to(OBJECT_AVAILABLE);
495 }
496 
497 /**
498  * fscache_object_lookup_negative - Note negative cookie lookup
499  * @object: Object pointing to cookie to mark
500  *
501  * Note negative lookup, permitting those waiting to read data from an already
502  * existing backing object to continue as there's no data for them to read.
503  */
fscache_object_lookup_negative(struct fscache_object * object)504 void fscache_object_lookup_negative(struct fscache_object *object)
505 {
506 	struct fscache_cookie *cookie = object->cookie;
507 
508 	_enter("{OBJ%x,%s}", object->debug_id, object->state->name);
509 
510 	if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
511 		fscache_stat(&fscache_n_object_lookups_negative);
512 
513 		/* Allow write requests to begin stacking up and read requests to begin
514 		 * returning ENODATA.
515 		 */
516 		set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
517 		clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
518 
519 		_debug("wake up lookup %p", &cookie->flags);
520 		clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
521 		wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
522 	}
523 	_leave("");
524 }
525 EXPORT_SYMBOL(fscache_object_lookup_negative);
526 
527 /**
528  * fscache_obtained_object - Note successful object lookup or creation
529  * @object: Object pointing to cookie to mark
530  *
531  * Note successful lookup and/or creation, permitting those waiting to write
532  * data to a backing object to continue.
533  *
534  * Note that after calling this, an object's cookie may be relinquished by the
535  * netfs, and so must be accessed with object lock held.
536  */
fscache_obtained_object(struct fscache_object * object)537 void fscache_obtained_object(struct fscache_object *object)
538 {
539 	struct fscache_cookie *cookie = object->cookie;
540 
541 	_enter("{OBJ%x,%s}", object->debug_id, object->state->name);
542 
543 	/* if we were still looking up, then we must have a positive lookup
544 	 * result, in which case there may be data available */
545 	if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
546 		fscache_stat(&fscache_n_object_lookups_positive);
547 
548 		/* We do (presumably) have data */
549 		clear_bit_unlock(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
550 		clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
551 
552 		/* Allow write requests to begin stacking up and read requests
553 		 * to begin shovelling data.
554 		 */
555 		clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
556 		wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
557 	} else {
558 		fscache_stat(&fscache_n_object_created);
559 	}
560 
561 	set_bit(FSCACHE_OBJECT_IS_AVAILABLE, &object->flags);
562 	_leave("");
563 }
564 EXPORT_SYMBOL(fscache_obtained_object);
565 
566 /*
567  * handle an object that has just become available
568  */
fscache_object_available(struct fscache_object * object,int event)569 static const struct fscache_state *fscache_object_available(struct fscache_object *object,
570 							    int event)
571 {
572 	_enter("{OBJ%x},%d", object->debug_id, event);
573 
574 	object->oob_table = fscache_osm_run_oob;
575 
576 	spin_lock(&object->lock);
577 
578 	fscache_done_parent_op(object);
579 	if (object->n_in_progress == 0) {
580 		if (object->n_ops > 0) {
581 			ASSERTCMP(object->n_ops, >=, object->n_obj_ops);
582 			fscache_start_operations(object);
583 		} else {
584 			ASSERT(list_empty(&object->pending_ops));
585 		}
586 	}
587 	spin_unlock(&object->lock);
588 
589 	fscache_stat(&fscache_n_cop_lookup_complete);
590 	object->cache->ops->lookup_complete(object);
591 	fscache_stat_d(&fscache_n_cop_lookup_complete);
592 
593 	fscache_hist(fscache_obj_instantiate_histogram, object->lookup_jif);
594 	fscache_stat(&fscache_n_object_avail);
595 
596 	_leave("");
597 	return transit_to(JUMPSTART_DEPS);
598 }
599 
600 /*
601  * Wake up this object's dependent objects now that we've become available.
602  */
fscache_jumpstart_dependents(struct fscache_object * object,int event)603 static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *object,
604 								int event)
605 {
606 	_enter("{OBJ%x},%d", object->debug_id, event);
607 
608 	if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_PARENT_READY))
609 		return NO_TRANSIT; /* Not finished; requeue */
610 	return transit_to(WAIT_FOR_CMD);
611 }
612 
613 /*
614  * Handle lookup or creation failute.
615  */
fscache_lookup_failure(struct fscache_object * object,int event)616 static const struct fscache_state *fscache_lookup_failure(struct fscache_object *object,
617 							  int event)
618 {
619 	struct fscache_cookie *cookie;
620 
621 	_enter("{OBJ%x},%d", object->debug_id, event);
622 
623 	object->oob_event_mask = 0;
624 
625 	fscache_stat(&fscache_n_cop_lookup_complete);
626 	object->cache->ops->lookup_complete(object);
627 	fscache_stat_d(&fscache_n_cop_lookup_complete);
628 
629 	set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags);
630 
631 	cookie = object->cookie;
632 	set_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
633 	if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags))
634 		wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
635 
636 	fscache_done_parent_op(object);
637 	return transit_to(KILL_OBJECT);
638 }
639 
640 /*
641  * Wait for completion of all active operations on this object and the death of
642  * all child objects of this object.
643  */
fscache_kill_object(struct fscache_object * object,int event)644 static const struct fscache_state *fscache_kill_object(struct fscache_object *object,
645 						       int event)
646 {
647 	_enter("{OBJ%x,%d,%d},%d",
648 	       object->debug_id, object->n_ops, object->n_children, event);
649 
650 	fscache_mark_object_dead(object);
651 	object->oob_event_mask = 0;
652 
653 	if (test_bit(FSCACHE_OBJECT_RETIRED, &object->flags)) {
654 		/* Reject any new read/write ops and abort any that are pending. */
655 		clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
656 		fscache_cancel_all_ops(object);
657 	}
658 
659 	if (list_empty(&object->dependents) &&
660 	    object->n_ops == 0 &&
661 	    object->n_children == 0)
662 		return transit_to(DROP_OBJECT);
663 
664 	if (object->n_in_progress == 0) {
665 		spin_lock(&object->lock);
666 		if (object->n_ops > 0 && object->n_in_progress == 0)
667 			fscache_start_operations(object);
668 		spin_unlock(&object->lock);
669 	}
670 
671 	if (!list_empty(&object->dependents))
672 		return transit_to(KILL_DEPENDENTS);
673 
674 	return transit_to(WAIT_FOR_CLEARANCE);
675 }
676 
677 /*
678  * Kill dependent objects.
679  */
fscache_kill_dependents(struct fscache_object * object,int event)680 static const struct fscache_state *fscache_kill_dependents(struct fscache_object *object,
681 							   int event)
682 {
683 	_enter("{OBJ%x},%d", object->debug_id, event);
684 
685 	if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_KILL))
686 		return NO_TRANSIT; /* Not finished */
687 	return transit_to(WAIT_FOR_CLEARANCE);
688 }
689 
690 /*
691  * Drop an object's attachments
692  */
fscache_drop_object(struct fscache_object * object,int event)693 static const struct fscache_state *fscache_drop_object(struct fscache_object *object,
694 						       int event)
695 {
696 	struct fscache_object *parent = object->parent;
697 	struct fscache_cookie *cookie = object->cookie;
698 	struct fscache_cache *cache = object->cache;
699 	bool awaken = false;
700 
701 	_enter("{OBJ%x,%d},%d", object->debug_id, object->n_children, event);
702 
703 	ASSERT(cookie != NULL);
704 	ASSERT(!hlist_unhashed(&object->cookie_link));
705 
706 	/* Make sure the cookie no longer points here and that the netfs isn't
707 	 * waiting for us.
708 	 */
709 	spin_lock(&cookie->lock);
710 	hlist_del_init(&object->cookie_link);
711 	if (hlist_empty(&cookie->backing_objects) &&
712 	    test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
713 		awaken = true;
714 	spin_unlock(&cookie->lock);
715 
716 	if (awaken)
717 		wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
718 	if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags))
719 		wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
720 
721 
722 	/* Prevent a race with our last child, which has to signal EV_CLEARED
723 	 * before dropping our spinlock.
724 	 */
725 	spin_lock(&object->lock);
726 	spin_unlock(&object->lock);
727 
728 	/* Discard from the cache's collection of objects */
729 	spin_lock(&cache->object_list_lock);
730 	list_del_init(&object->cache_link);
731 	spin_unlock(&cache->object_list_lock);
732 
733 	fscache_stat(&fscache_n_cop_drop_object);
734 	cache->ops->drop_object(object);
735 	fscache_stat_d(&fscache_n_cop_drop_object);
736 
737 	/* The parent object wants to know when all it dependents have gone */
738 	if (parent) {
739 		_debug("release parent OBJ%x {%d}",
740 		       parent->debug_id, parent->n_children);
741 
742 		spin_lock(&parent->lock);
743 		parent->n_children--;
744 		if (parent->n_children == 0)
745 			fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
746 		spin_unlock(&parent->lock);
747 		object->parent = NULL;
748 	}
749 
750 	/* this just shifts the object release to the work processor */
751 	fscache_put_object(object);
752 	fscache_stat(&fscache_n_object_dead);
753 
754 	_leave("");
755 	return transit_to(OBJECT_DEAD);
756 }
757 
758 /*
759  * get a ref on an object
760  */
fscache_get_object(struct fscache_object * object)761 static int fscache_get_object(struct fscache_object *object)
762 {
763 	int ret;
764 
765 	fscache_stat(&fscache_n_cop_grab_object);
766 	ret = object->cache->ops->grab_object(object) ? 0 : -EAGAIN;
767 	fscache_stat_d(&fscache_n_cop_grab_object);
768 	return ret;
769 }
770 
771 /*
772  * Discard a ref on an object
773  */
fscache_put_object(struct fscache_object * object)774 static void fscache_put_object(struct fscache_object *object)
775 {
776 	fscache_stat(&fscache_n_cop_put_object);
777 	object->cache->ops->put_object(object);
778 	fscache_stat_d(&fscache_n_cop_put_object);
779 }
780 
781 /**
782  * fscache_object_destroy - Note that a cache object is about to be destroyed
783  * @object: The object to be destroyed
784  *
785  * Note the imminent destruction and deallocation of a cache object record.
786  */
fscache_object_destroy(struct fscache_object * object)787 void fscache_object_destroy(struct fscache_object *object)
788 {
789 	fscache_objlist_remove(object);
790 
791 	/* We can get rid of the cookie now */
792 	fscache_cookie_put(object->cookie);
793 	object->cookie = NULL;
794 }
795 EXPORT_SYMBOL(fscache_object_destroy);
796 
797 /*
798  * enqueue an object for metadata-type processing
799  */
fscache_enqueue_object(struct fscache_object * object)800 void fscache_enqueue_object(struct fscache_object *object)
801 {
802 	_enter("{OBJ%x}", object->debug_id);
803 
804 	if (fscache_get_object(object) >= 0) {
805 		wait_queue_head_t *cong_wq =
806 			&get_cpu_var(fscache_object_cong_wait);
807 
808 		if (queue_work(fscache_object_wq, &object->work)) {
809 			if (fscache_object_congested())
810 				wake_up(cong_wq);
811 		} else
812 			fscache_put_object(object);
813 
814 		put_cpu_var(fscache_object_cong_wait);
815 	}
816 }
817 
818 /**
819  * fscache_object_sleep_till_congested - Sleep until object wq is congested
820  * @timeoutp: Scheduler sleep timeout
821  *
822  * Allow an object handler to sleep until the object workqueue is congested.
823  *
824  * The caller must set up a wake up event before calling this and must have set
825  * the appropriate sleep mode (such as TASK_UNINTERRUPTIBLE) and tested its own
826  * condition before calling this function as no test is made here.
827  *
828  * %true is returned if the object wq is congested, %false otherwise.
829  */
fscache_object_sleep_till_congested(signed long * timeoutp)830 bool fscache_object_sleep_till_congested(signed long *timeoutp)
831 {
832 	wait_queue_head_t *cong_wq = this_cpu_ptr(&fscache_object_cong_wait);
833 	DEFINE_WAIT(wait);
834 
835 	if (fscache_object_congested())
836 		return true;
837 
838 	add_wait_queue_exclusive(cong_wq, &wait);
839 	if (!fscache_object_congested())
840 		*timeoutp = schedule_timeout(*timeoutp);
841 	finish_wait(cong_wq, &wait);
842 
843 	return fscache_object_congested();
844 }
845 EXPORT_SYMBOL_GPL(fscache_object_sleep_till_congested);
846 
847 /*
848  * Enqueue the dependents of an object for metadata-type processing.
849  *
850  * If we don't manage to finish the list before the scheduler wants to run
851  * again then return false immediately.  We return true if the list was
852  * cleared.
853  */
fscache_enqueue_dependents(struct fscache_object * object,int event)854 static bool fscache_enqueue_dependents(struct fscache_object *object, int event)
855 {
856 	struct fscache_object *dep;
857 	bool ret = true;
858 
859 	_enter("{OBJ%x}", object->debug_id);
860 
861 	if (list_empty(&object->dependents))
862 		return true;
863 
864 	spin_lock(&object->lock);
865 
866 	while (!list_empty(&object->dependents)) {
867 		dep = list_entry(object->dependents.next,
868 				 struct fscache_object, dep_link);
869 		list_del_init(&dep->dep_link);
870 
871 		fscache_raise_event(dep, event);
872 		fscache_put_object(dep);
873 
874 		if (!list_empty(&object->dependents) && need_resched()) {
875 			ret = false;
876 			break;
877 		}
878 	}
879 
880 	spin_unlock(&object->lock);
881 	return ret;
882 }
883 
884 /*
885  * remove an object from whatever queue it's waiting on
886  */
fscache_dequeue_object(struct fscache_object * object)887 static void fscache_dequeue_object(struct fscache_object *object)
888 {
889 	_enter("{OBJ%x}", object->debug_id);
890 
891 	if (!list_empty(&object->dep_link)) {
892 		spin_lock(&object->parent->lock);
893 		list_del_init(&object->dep_link);
894 		spin_unlock(&object->parent->lock);
895 	}
896 
897 	_leave("");
898 }
899 
900 /**
901  * fscache_check_aux - Ask the netfs whether an object on disk is still valid
902  * @object: The object to ask about
903  * @data: The auxiliary data for the object
904  * @datalen: The size of the auxiliary data
905  *
906  * This function consults the netfs about the coherency state of an object.
907  * The caller must be holding a ref on cookie->n_active (held by
908  * fscache_look_up_object() on behalf of the cache backend during object lookup
909  * and creation).
910  */
fscache_check_aux(struct fscache_object * object,const void * data,uint16_t datalen)911 enum fscache_checkaux fscache_check_aux(struct fscache_object *object,
912 					const void *data, uint16_t datalen)
913 {
914 	enum fscache_checkaux result;
915 
916 	if (!object->cookie->def->check_aux) {
917 		fscache_stat(&fscache_n_checkaux_none);
918 		return FSCACHE_CHECKAUX_OKAY;
919 	}
920 
921 	result = object->cookie->def->check_aux(object->cookie->netfs_data,
922 						data, datalen);
923 	switch (result) {
924 		/* entry okay as is */
925 	case FSCACHE_CHECKAUX_OKAY:
926 		fscache_stat(&fscache_n_checkaux_okay);
927 		break;
928 
929 		/* entry requires update */
930 	case FSCACHE_CHECKAUX_NEEDS_UPDATE:
931 		fscache_stat(&fscache_n_checkaux_update);
932 		break;
933 
934 		/* entry requires deletion */
935 	case FSCACHE_CHECKAUX_OBSOLETE:
936 		fscache_stat(&fscache_n_checkaux_obsolete);
937 		break;
938 
939 	default:
940 		BUG();
941 	}
942 
943 	return result;
944 }
945 EXPORT_SYMBOL(fscache_check_aux);
946 
947 /*
948  * Asynchronously invalidate an object.
949  */
_fscache_invalidate_object(struct fscache_object * object,int event)950 static const struct fscache_state *_fscache_invalidate_object(struct fscache_object *object,
951 							      int event)
952 {
953 	struct fscache_operation *op;
954 	struct fscache_cookie *cookie = object->cookie;
955 
956 	_enter("{OBJ%x},%d", object->debug_id, event);
957 
958 	/* We're going to need the cookie.  If the cookie is not available then
959 	 * retire the object instead.
960 	 */
961 	if (!fscache_use_cookie(object)) {
962 		ASSERT(object->cookie->stores.rnode == NULL);
963 		set_bit(FSCACHE_OBJECT_RETIRED, &object->flags);
964 		_leave(" [no cookie]");
965 		return transit_to(KILL_OBJECT);
966 	}
967 
968 	/* Reject any new read/write ops and abort any that are pending. */
969 	fscache_invalidate_writes(cookie);
970 	clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
971 	fscache_cancel_all_ops(object);
972 
973 	/* Now we have to wait for in-progress reads and writes */
974 	op = kzalloc(sizeof(*op), GFP_KERNEL);
975 	if (!op)
976 		goto nomem;
977 
978 	fscache_operation_init(op, object->cache->ops->invalidate_object,
979 			       NULL, NULL);
980 	op->flags = FSCACHE_OP_ASYNC |
981 		(1 << FSCACHE_OP_EXCLUSIVE) |
982 		(1 << FSCACHE_OP_UNUSE_COOKIE);
983 
984 	spin_lock(&cookie->lock);
985 	if (fscache_submit_exclusive_op(object, op) < 0)
986 		goto submit_op_failed;
987 	spin_unlock(&cookie->lock);
988 	fscache_put_operation(op);
989 
990 	/* Once we've completed the invalidation, we know there will be no data
991 	 * stored in the cache and thus we can reinstate the data-check-skip
992 	 * optimisation.
993 	 */
994 	set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
995 
996 	/* We can allow read and write requests to come in once again.  They'll
997 	 * queue up behind our exclusive invalidation operation.
998 	 */
999 	if (test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
1000 		wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
1001 	_leave(" [ok]");
1002 	return transit_to(UPDATE_OBJECT);
1003 
1004 nomem:
1005 	fscache_mark_object_dead(object);
1006 	fscache_unuse_cookie(object);
1007 	_leave(" [ENOMEM]");
1008 	return transit_to(KILL_OBJECT);
1009 
1010 submit_op_failed:
1011 	fscache_mark_object_dead(object);
1012 	spin_unlock(&cookie->lock);
1013 	fscache_unuse_cookie(object);
1014 	kfree(op);
1015 	_leave(" [EIO]");
1016 	return transit_to(KILL_OBJECT);
1017 }
1018 
fscache_invalidate_object(struct fscache_object * object,int event)1019 static const struct fscache_state *fscache_invalidate_object(struct fscache_object *object,
1020 							     int event)
1021 {
1022 	const struct fscache_state *s;
1023 
1024 	fscache_stat(&fscache_n_invalidates_run);
1025 	fscache_stat(&fscache_n_cop_invalidate_object);
1026 	s = _fscache_invalidate_object(object, event);
1027 	fscache_stat_d(&fscache_n_cop_invalidate_object);
1028 	return s;
1029 }
1030 
1031 /*
1032  * Asynchronously update an object.
1033  */
fscache_update_object(struct fscache_object * object,int event)1034 static const struct fscache_state *fscache_update_object(struct fscache_object *object,
1035 							 int event)
1036 {
1037 	_enter("{OBJ%x},%d", object->debug_id, event);
1038 
1039 	fscache_stat(&fscache_n_updates_run);
1040 	fscache_stat(&fscache_n_cop_update_object);
1041 	object->cache->ops->update_object(object);
1042 	fscache_stat_d(&fscache_n_cop_update_object);
1043 
1044 	_leave("");
1045 	return transit_to(WAIT_FOR_CMD);
1046 }
1047 
1048 /**
1049  * fscache_object_retrying_stale - Note retrying stale object
1050  * @object: The object that will be retried
1051  *
1052  * Note that an object lookup found an on-disk object that was adjudged to be
1053  * stale and has been deleted.  The lookup will be retried.
1054  */
fscache_object_retrying_stale(struct fscache_object * object)1055 void fscache_object_retrying_stale(struct fscache_object *object)
1056 {
1057 	fscache_stat(&fscache_n_cache_no_space_reject);
1058 }
1059 EXPORT_SYMBOL(fscache_object_retrying_stale);
1060 
1061 /**
1062  * fscache_object_mark_killed - Note that an object was killed
1063  * @object: The object that was culled
1064  * @why: The reason the object was killed.
1065  *
1066  * Note that an object was killed.  Returns true if the object was
1067  * already marked killed, false if it wasn't.
1068  */
fscache_object_mark_killed(struct fscache_object * object,enum fscache_why_object_killed why)1069 void fscache_object_mark_killed(struct fscache_object *object,
1070 				enum fscache_why_object_killed why)
1071 {
1072 	if (test_and_set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags)) {
1073 		pr_err("Error: Object already killed by cache [%s]\n",
1074 		       object->cache->identifier);
1075 		return;
1076 	}
1077 
1078 	switch (why) {
1079 	case FSCACHE_OBJECT_NO_SPACE:
1080 		fscache_stat(&fscache_n_cache_no_space_reject);
1081 		break;
1082 	case FSCACHE_OBJECT_IS_STALE:
1083 		fscache_stat(&fscache_n_cache_stale_objects);
1084 		break;
1085 	case FSCACHE_OBJECT_WAS_RETIRED:
1086 		fscache_stat(&fscache_n_cache_retired_objects);
1087 		break;
1088 	case FSCACHE_OBJECT_WAS_CULLED:
1089 		fscache_stat(&fscache_n_cache_culled_objects);
1090 		break;
1091 	}
1092 }
1093 EXPORT_SYMBOL(fscache_object_mark_killed);
1094 
1095 /*
1096  * The object is dead.  We can get here if an object gets queued by an event
1097  * that would lead to its death (such as EV_KILL) when the dispatcher is
1098  * already running (and so can be requeued) but hasn't yet cleared the event
1099  * mask.
1100  */
fscache_object_dead(struct fscache_object * object,int event)1101 static const struct fscache_state *fscache_object_dead(struct fscache_object *object,
1102 						       int event)
1103 {
1104 	if (!test_and_set_bit(FSCACHE_OBJECT_RUN_AFTER_DEAD,
1105 			      &object->flags))
1106 		return NO_TRANSIT;
1107 
1108 	WARN(true, "FS-Cache object redispatched after death");
1109 	return NO_TRANSIT;
1110 }
1111