• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1  /* SPDX-License-Identifier: GPL-2.0+ */
2  /*
3   * Sleepable Read-Copy Update mechanism for mutual exclusion,
4   *	tiny variant.
5   *
6   * Copyright (C) IBM Corporation, 2017
7   *
8   * Author: Paul McKenney <paulmck@linux.ibm.com>
9   */
10  
11  #ifndef _LINUX_SRCU_TINY_H
12  #define _LINUX_SRCU_TINY_H
13  
14  #include <linux/swait.h>
15  
16  struct srcu_struct {
17  	short srcu_lock_nesting[2];	/* srcu_read_lock() nesting depth. */
18  	unsigned short srcu_idx;	/* Current reader array element in bit 0x2. */
19  	unsigned short srcu_idx_max;	/* Furthest future srcu_idx request. */
20  	u8 srcu_gp_running;		/* GP workqueue running? */
21  	u8 srcu_gp_waiting;		/* GP waiting for readers? */
22  	struct swait_queue_head srcu_wq;
23  					/* Last srcu_read_unlock() wakes GP. */
24  	struct rcu_head *srcu_cb_head;	/* Pending callbacks: Head. */
25  	struct rcu_head **srcu_cb_tail;	/* Pending callbacks: Tail. */
26  	struct work_struct srcu_work;	/* For driving grace periods. */
27  #ifdef CONFIG_DEBUG_LOCK_ALLOC
28  	struct lockdep_map dep_map;
29  #endif /* #ifdef CONFIG_DEBUG_LOCK_ALLOC */
30  };
31  
32  void srcu_drive_gp(struct work_struct *wp);
33  
34  #define __SRCU_STRUCT_INIT(name, __ignored)				\
35  {									\
36  	.srcu_wq = __SWAIT_QUEUE_HEAD_INITIALIZER(name.srcu_wq),	\
37  	.srcu_cb_tail = &name.srcu_cb_head,				\
38  	.srcu_work = __WORK_INITIALIZER(name.srcu_work, srcu_drive_gp),	\
39  	__SRCU_DEP_MAP_INIT(name)					\
40  }
41  
42  /*
43   * This odd _STATIC_ arrangement is needed for API compatibility with
44   * Tree SRCU, which needs some per-CPU data.
45   */
46  #define DEFINE_SRCU(name) \
47  	struct srcu_struct name = __SRCU_STRUCT_INIT(name, name)
48  #define DEFINE_STATIC_SRCU(name) \
49  	static struct srcu_struct name = __SRCU_STRUCT_INIT(name, name)
50  
51  void synchronize_srcu(struct srcu_struct *ssp);
52  
53  /*
54   * Counts the new reader in the appropriate per-CPU element of the
55   * srcu_struct.  Can be invoked from irq/bh handlers, but the matching
56   * __srcu_read_unlock() must be in the same handler instance.  Returns an
57   * index that must be passed to the matching srcu_read_unlock().
58   */
__srcu_read_lock(struct srcu_struct * ssp)59  static inline int __srcu_read_lock(struct srcu_struct *ssp)
60  {
61  	int idx;
62  
63  	idx = ((READ_ONCE(ssp->srcu_idx) + 1) & 0x2) >> 1;
64  	WRITE_ONCE(ssp->srcu_lock_nesting[idx], READ_ONCE(ssp->srcu_lock_nesting[idx]) + 1);
65  	return idx;
66  }
67  
synchronize_srcu_expedited(struct srcu_struct * ssp)68  static inline void synchronize_srcu_expedited(struct srcu_struct *ssp)
69  {
70  	synchronize_srcu(ssp);
71  }
72  
srcu_barrier(struct srcu_struct * ssp)73  static inline void srcu_barrier(struct srcu_struct *ssp)
74  {
75  	synchronize_srcu(ssp);
76  }
77  
78  /* Defined here to avoid size increase for non-torture kernels. */
srcu_torture_stats_print(struct srcu_struct * ssp,char * tt,char * tf)79  static inline void srcu_torture_stats_print(struct srcu_struct *ssp,
80  					    char *tt, char *tf)
81  {
82  	int idx;
83  
84  	idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1;
85  	pr_alert("%s%s Tiny SRCU per-CPU(idx=%d): (%hd,%hd)\n",
86  		 tt, tf, idx,
87  		 data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])),
88  		 data_race(READ_ONCE(ssp->srcu_lock_nesting[idx])));
89  }
90  
91  #endif
92