1 /*
2 *
3 * Copyright 2016 gRPC authors.
4 *
5 * Licensed under the Apache License, Version 2.0 (the "License");
6 * you may not use this file except in compliance with the License.
7 * You may obtain a copy of the License at
8 *
9 * http://www.apache.org/licenses/LICENSE-2.0
10 *
11 * Unless required by applicable law or agreed to in writing, software
12 * distributed under the License is distributed on an "AS IS" BASIS,
13 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14 * See the License for the specific language governing permissions and
15 * limitations under the License.
16 *
17 */
18
19 #include <grpc/support/port_platform.h>
20
21 #include "src/core/lib/iomgr/combiner.h"
22
23 #include <assert.h>
24 #include <inttypes.h>
25 #include <string.h>
26
27 #include <grpc/support/alloc.h>
28 #include <grpc/support/log.h>
29
30 #include "src/core/lib/debug/stats.h"
31 #include "src/core/lib/iomgr/executor.h"
32 #include "src/core/lib/profiling/timers.h"
33
34 grpc_core::DebugOnlyTraceFlag grpc_combiner_trace(false, "combiner");
35
36 #define GRPC_COMBINER_TRACE(fn) \
37 do { \
38 if (grpc_combiner_trace.enabled()) { \
39 fn; \
40 } \
41 } while (0)
42
43 #define STATE_UNORPHANED 1
44 #define STATE_ELEM_COUNT_LOW_BIT 2
45
46 struct grpc_combiner {
47 grpc_combiner* next_combiner_on_this_exec_ctx;
48 grpc_closure_scheduler scheduler;
49 grpc_closure_scheduler finally_scheduler;
50 gpr_mpscq queue;
51 // either:
52 // a pointer to the initiating exec ctx if that is the only exec_ctx that has
53 // ever queued to this combiner, or NULL. If this is non-null, it's not
54 // dereferencable (since the initiating exec_ctx may have gone out of scope)
55 gpr_atm initiating_exec_ctx_or_null;
56 // state is:
57 // lower bit - zero if orphaned (STATE_UNORPHANED)
58 // other bits - number of items queued on the lock (STATE_ELEM_COUNT_LOW_BIT)
59 gpr_atm state;
60 bool time_to_execute_final_list;
61 grpc_closure_list final_list;
62 grpc_closure offload;
63 gpr_refcount refs;
64 };
65
66 static void combiner_run(grpc_closure* closure, grpc_error* error);
67 static void combiner_exec(grpc_closure* closure, grpc_error* error);
68 static void combiner_finally_exec(grpc_closure* closure, grpc_error* error);
69
70 static const grpc_closure_scheduler_vtable scheduler = {
71 combiner_run, combiner_exec, "combiner:immediately"};
72 static const grpc_closure_scheduler_vtable finally_scheduler = {
73 combiner_finally_exec, combiner_finally_exec, "combiner:finally"};
74
75 static void offload(void* arg, grpc_error* error);
76
grpc_combiner_create(void)77 grpc_combiner* grpc_combiner_create(void) {
78 grpc_combiner* lock = static_cast<grpc_combiner*>(gpr_zalloc(sizeof(*lock)));
79 gpr_ref_init(&lock->refs, 1);
80 lock->scheduler.vtable = &scheduler;
81 lock->finally_scheduler.vtable = &finally_scheduler;
82 gpr_atm_no_barrier_store(&lock->state, STATE_UNORPHANED);
83 gpr_mpscq_init(&lock->queue);
84 grpc_closure_list_init(&lock->final_list);
85 GRPC_CLOSURE_INIT(&lock->offload, offload, lock,
86 grpc_executor_scheduler(GRPC_EXECUTOR_SHORT));
87 GRPC_COMBINER_TRACE(gpr_log(GPR_INFO, "C:%p create", lock));
88 return lock;
89 }
90
really_destroy(grpc_combiner * lock)91 static void really_destroy(grpc_combiner* lock) {
92 GRPC_COMBINER_TRACE(gpr_log(GPR_INFO, "C:%p really_destroy", lock));
93 GPR_ASSERT(gpr_atm_no_barrier_load(&lock->state) == 0);
94 gpr_mpscq_destroy(&lock->queue);
95 gpr_free(lock);
96 }
97
start_destroy(grpc_combiner * lock)98 static void start_destroy(grpc_combiner* lock) {
99 gpr_atm old_state = gpr_atm_full_fetch_add(&lock->state, -STATE_UNORPHANED);
100 GRPC_COMBINER_TRACE(gpr_log(
101 GPR_INFO, "C:%p really_destroy old_state=%" PRIdPTR, lock, old_state));
102 if (old_state == 1) {
103 really_destroy(lock);
104 }
105 }
106
107 #ifndef NDEBUG
108 #define GRPC_COMBINER_DEBUG_SPAM(op, delta) \
109 if (grpc_combiner_trace.enabled()) { \
110 gpr_log(file, line, GPR_LOG_SEVERITY_DEBUG, \
111 "C:%p %s %" PRIdPTR " --> %" PRIdPTR " %s", lock, (op), \
112 gpr_atm_no_barrier_load(&lock->refs.count), \
113 gpr_atm_no_barrier_load(&lock->refs.count) + (delta), reason); \
114 }
115 #else
116 #define GRPC_COMBINER_DEBUG_SPAM(op, delta)
117 #endif
118
grpc_combiner_unref(grpc_combiner * lock GRPC_COMBINER_DEBUG_ARGS)119 void grpc_combiner_unref(grpc_combiner* lock GRPC_COMBINER_DEBUG_ARGS) {
120 GRPC_COMBINER_DEBUG_SPAM("UNREF", -1);
121 if (gpr_unref(&lock->refs)) {
122 start_destroy(lock);
123 }
124 }
125
grpc_combiner_ref(grpc_combiner * lock GRPC_COMBINER_DEBUG_ARGS)126 grpc_combiner* grpc_combiner_ref(grpc_combiner* lock GRPC_COMBINER_DEBUG_ARGS) {
127 GRPC_COMBINER_DEBUG_SPAM(" REF", 1);
128 gpr_ref(&lock->refs);
129 return lock;
130 }
131
push_last_on_exec_ctx(grpc_combiner * lock)132 static void push_last_on_exec_ctx(grpc_combiner* lock) {
133 lock->next_combiner_on_this_exec_ctx = nullptr;
134 if (grpc_core::ExecCtx::Get()->combiner_data()->active_combiner == nullptr) {
135 grpc_core::ExecCtx::Get()->combiner_data()->active_combiner =
136 grpc_core::ExecCtx::Get()->combiner_data()->last_combiner = lock;
137 } else {
138 grpc_core::ExecCtx::Get()
139 ->combiner_data()
140 ->last_combiner->next_combiner_on_this_exec_ctx = lock;
141 grpc_core::ExecCtx::Get()->combiner_data()->last_combiner = lock;
142 }
143 }
144
push_first_on_exec_ctx(grpc_combiner * lock)145 static void push_first_on_exec_ctx(grpc_combiner* lock) {
146 lock->next_combiner_on_this_exec_ctx =
147 grpc_core::ExecCtx::Get()->combiner_data()->active_combiner;
148 grpc_core::ExecCtx::Get()->combiner_data()->active_combiner = lock;
149 if (lock->next_combiner_on_this_exec_ctx == nullptr) {
150 grpc_core::ExecCtx::Get()->combiner_data()->last_combiner = lock;
151 }
152 }
153
154 #define COMBINER_FROM_CLOSURE_SCHEDULER(closure, scheduler_name) \
155 ((grpc_combiner*)(((char*)((closure)->scheduler)) - \
156 offsetof(grpc_combiner, scheduler_name)))
157
combiner_exec(grpc_closure * cl,grpc_error * error)158 static void combiner_exec(grpc_closure* cl, grpc_error* error) {
159 GPR_TIMER_SCOPE("combiner.execute", 0);
160 GRPC_STATS_INC_COMBINER_LOCKS_SCHEDULED_ITEMS();
161 grpc_combiner* lock = COMBINER_FROM_CLOSURE_SCHEDULER(cl, scheduler);
162 gpr_atm last = gpr_atm_full_fetch_add(&lock->state, STATE_ELEM_COUNT_LOW_BIT);
163 GRPC_COMBINER_TRACE(gpr_log(GPR_INFO,
164 "C:%p grpc_combiner_execute c=%p last=%" PRIdPTR,
165 lock, cl, last));
166 if (last == 1) {
167 GRPC_STATS_INC_COMBINER_LOCKS_INITIATED();
168 GPR_TIMER_MARK("combiner.initiated", 0);
169 gpr_atm_no_barrier_store(&lock->initiating_exec_ctx_or_null,
170 (gpr_atm)grpc_core::ExecCtx::Get());
171 // first element on this list: add it to the list of combiner locks
172 // executing within this exec_ctx
173 push_last_on_exec_ctx(lock);
174 } else {
175 // there may be a race with setting here: if that happens, we may delay
176 // offload for one or two actions, and that's fine
177 gpr_atm initiator =
178 gpr_atm_no_barrier_load(&lock->initiating_exec_ctx_or_null);
179 if (initiator != 0 && initiator != (gpr_atm)grpc_core::ExecCtx::Get()) {
180 gpr_atm_no_barrier_store(&lock->initiating_exec_ctx_or_null, 0);
181 }
182 }
183 GPR_ASSERT(last & STATE_UNORPHANED); // ensure lock has not been destroyed
184 assert(cl->cb);
185 cl->error_data.error = error;
186 gpr_mpscq_push(&lock->queue, &cl->next_data.atm_next);
187 }
188
move_next()189 static void move_next() {
190 grpc_core::ExecCtx::Get()->combiner_data()->active_combiner =
191 grpc_core::ExecCtx::Get()
192 ->combiner_data()
193 ->active_combiner->next_combiner_on_this_exec_ctx;
194 if (grpc_core::ExecCtx::Get()->combiner_data()->active_combiner == nullptr) {
195 grpc_core::ExecCtx::Get()->combiner_data()->last_combiner = nullptr;
196 }
197 }
198
offload(void * arg,grpc_error * error)199 static void offload(void* arg, grpc_error* error) {
200 grpc_combiner* lock = static_cast<grpc_combiner*>(arg);
201 push_last_on_exec_ctx(lock);
202 }
203
queue_offload(grpc_combiner * lock)204 static void queue_offload(grpc_combiner* lock) {
205 GRPC_STATS_INC_COMBINER_LOCKS_OFFLOADED();
206 move_next();
207 GRPC_COMBINER_TRACE(gpr_log(GPR_INFO, "C:%p queue_offload", lock));
208 GRPC_CLOSURE_SCHED(&lock->offload, GRPC_ERROR_NONE);
209 }
210
grpc_combiner_continue_exec_ctx()211 bool grpc_combiner_continue_exec_ctx() {
212 GPR_TIMER_SCOPE("combiner.continue_exec_ctx", 0);
213 grpc_combiner* lock =
214 grpc_core::ExecCtx::Get()->combiner_data()->active_combiner;
215 if (lock == nullptr) {
216 return false;
217 }
218
219 bool contended =
220 gpr_atm_no_barrier_load(&lock->initiating_exec_ctx_or_null) == 0;
221
222 GRPC_COMBINER_TRACE(gpr_log(GPR_INFO,
223 "C:%p grpc_combiner_continue_exec_ctx "
224 "contended=%d "
225 "exec_ctx_ready_to_finish=%d "
226 "time_to_execute_final_list=%d",
227 lock, contended,
228 grpc_core::ExecCtx::Get()->IsReadyToFinish(),
229 lock->time_to_execute_final_list));
230
231 if (contended && grpc_core::ExecCtx::Get()->IsReadyToFinish() &&
232 grpc_executor_is_threaded()) {
233 GPR_TIMER_MARK("offload_from_finished_exec_ctx", 0);
234 // this execution context wants to move on: schedule remaining work to be
235 // picked up on the executor
236 queue_offload(lock);
237 return true;
238 }
239
240 if (!lock->time_to_execute_final_list ||
241 // peek to see if something new has shown up, and execute that with
242 // priority
243 (gpr_atm_acq_load(&lock->state) >> 1) > 1) {
244 gpr_mpscq_node* n = gpr_mpscq_pop(&lock->queue);
245 GRPC_COMBINER_TRACE(
246 gpr_log(GPR_INFO, "C:%p maybe_finish_one n=%p", lock, n));
247 if (n == nullptr) {
248 // queue is in an inconsistent state: use this as a cue that we should
249 // go off and do something else for a while (and come back later)
250 GPR_TIMER_MARK("delay_busy", 0);
251 queue_offload(lock);
252 return true;
253 }
254 GPR_TIMER_SCOPE("combiner.exec1", 0);
255 grpc_closure* cl = reinterpret_cast<grpc_closure*>(n);
256 grpc_error* cl_err = cl->error_data.error;
257 #ifndef NDEBUG
258 cl->scheduled = false;
259 #endif
260 cl->cb(cl->cb_arg, cl_err);
261 GRPC_ERROR_UNREF(cl_err);
262 } else {
263 grpc_closure* c = lock->final_list.head;
264 GPR_ASSERT(c != nullptr);
265 grpc_closure_list_init(&lock->final_list);
266 int loops = 0;
267 while (c != nullptr) {
268 GPR_TIMER_SCOPE("combiner.exec_1final", 0);
269 GRPC_COMBINER_TRACE(
270 gpr_log(GPR_INFO, "C:%p execute_final[%d] c=%p", lock, loops, c));
271 grpc_closure* next = c->next_data.next;
272 grpc_error* error = c->error_data.error;
273 #ifndef NDEBUG
274 c->scheduled = false;
275 #endif
276 c->cb(c->cb_arg, error);
277 GRPC_ERROR_UNREF(error);
278 c = next;
279 }
280 }
281
282 GPR_TIMER_MARK("unref", 0);
283 move_next();
284 lock->time_to_execute_final_list = false;
285 gpr_atm old_state =
286 gpr_atm_full_fetch_add(&lock->state, -STATE_ELEM_COUNT_LOW_BIT);
287 GRPC_COMBINER_TRACE(
288 gpr_log(GPR_INFO, "C:%p finish old_state=%" PRIdPTR, lock, old_state));
289 // Define a macro to ease readability of the following switch statement.
290 #define OLD_STATE_WAS(orphaned, elem_count) \
291 (((orphaned) ? 0 : STATE_UNORPHANED) | \
292 ((elem_count)*STATE_ELEM_COUNT_LOW_BIT))
293 // Depending on what the previous state was, we need to perform different
294 // actions.
295 switch (old_state) {
296 default:
297 // we have multiple queued work items: just continue executing them
298 break;
299 case OLD_STATE_WAS(false, 2):
300 case OLD_STATE_WAS(true, 2):
301 // we're down to one queued item: if it's the final list we should do that
302 if (!grpc_closure_list_empty(lock->final_list)) {
303 lock->time_to_execute_final_list = true;
304 }
305 break;
306 case OLD_STATE_WAS(false, 1):
307 // had one count, one unorphaned --> unlocked unorphaned
308 return true;
309 case OLD_STATE_WAS(true, 1):
310 // and one count, one orphaned --> unlocked and orphaned
311 really_destroy(lock);
312 return true;
313 case OLD_STATE_WAS(false, 0):
314 case OLD_STATE_WAS(true, 0):
315 // these values are illegal - representing an already unlocked or
316 // deleted lock
317 GPR_UNREACHABLE_CODE(return true);
318 }
319 push_first_on_exec_ctx(lock);
320 return true;
321 }
322
323 static void enqueue_finally(void* closure, grpc_error* error);
324
combiner_finally_exec(grpc_closure * closure,grpc_error * error)325 static void combiner_finally_exec(grpc_closure* closure, grpc_error* error) {
326 GPR_TIMER_SCOPE("combiner.execute_finally", 0);
327 GRPC_STATS_INC_COMBINER_LOCKS_SCHEDULED_FINAL_ITEMS();
328 grpc_combiner* lock =
329 COMBINER_FROM_CLOSURE_SCHEDULER(closure, finally_scheduler);
330 GRPC_COMBINER_TRACE(gpr_log(
331 GPR_INFO, "C:%p grpc_combiner_execute_finally c=%p; ac=%p", lock, closure,
332 grpc_core::ExecCtx::Get()->combiner_data()->active_combiner));
333 if (grpc_core::ExecCtx::Get()->combiner_data()->active_combiner != lock) {
334 GPR_TIMER_MARK("slowpath", 0);
335 GRPC_CLOSURE_SCHED(GRPC_CLOSURE_CREATE(enqueue_finally, closure,
336 grpc_combiner_scheduler(lock)),
337 error);
338 return;
339 }
340
341 if (grpc_closure_list_empty(lock->final_list)) {
342 gpr_atm_full_fetch_add(&lock->state, STATE_ELEM_COUNT_LOW_BIT);
343 }
344 grpc_closure_list_append(&lock->final_list, closure, error);
345 }
346
combiner_run(grpc_closure * closure,grpc_error * error)347 static void combiner_run(grpc_closure* closure, grpc_error* error) {
348 grpc_combiner* lock = COMBINER_FROM_CLOSURE_SCHEDULER(closure, scheduler);
349 #ifndef NDEBUG
350 closure->scheduled = false;
351 GRPC_COMBINER_TRACE(gpr_log(
352 GPR_DEBUG,
353 "Combiner:%p grpc_combiner_run closure:%p created [%s:%d] run [%s:%d]",
354 lock, closure, closure->file_created, closure->line_created,
355 closure->file_initiated, closure->line_initiated));
356 #endif
357 GPR_ASSERT(grpc_core::ExecCtx::Get()->combiner_data()->active_combiner ==
358 lock);
359 closure->cb(closure->cb_arg, error);
360 GRPC_ERROR_UNREF(error);
361 }
362
enqueue_finally(void * closure,grpc_error * error)363 static void enqueue_finally(void* closure, grpc_error* error) {
364 combiner_finally_exec(static_cast<grpc_closure*>(closure),
365 GRPC_ERROR_REF(error));
366 }
367
grpc_combiner_scheduler(grpc_combiner * combiner)368 grpc_closure_scheduler* grpc_combiner_scheduler(grpc_combiner* combiner) {
369 return &combiner->scheduler;
370 }
371
grpc_combiner_finally_scheduler(grpc_combiner * combiner)372 grpc_closure_scheduler* grpc_combiner_finally_scheduler(
373 grpc_combiner* combiner) {
374 return &combiner->finally_scheduler;
375 }
376