• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 //
2 //
3 // Copyright 2015 gRPC authors.
4 //
5 // Licensed under the Apache License, Version 2.0 (the "License");
6 // you may not use this file except in compliance with the License.
7 // You may obtain a copy of the License at
8 //
9 //     http://www.apache.org/licenses/LICENSE-2.0
10 //
11 // Unless required by applicable law or agreed to in writing, software
12 // distributed under the License is distributed on an "AS IS" BASIS,
13 // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14 // See the License for the specific language governing permissions and
15 // limitations under the License.
16 //
17 //
18 
19 // Test of gpr synchronization support.
20 
21 #include <grpc/support/alloc.h>
22 #include <grpc/support/sync.h>
23 #include <grpc/support/time.h>
24 #include <stdint.h>
25 #include <stdio.h>
26 
27 #include <memory>
28 
29 #include "gtest/gtest.h"
30 #include "src/core/util/thd.h"
31 #include "test/core/test_util/test_config.h"
32 
33 // ==================Example use of interface===================
34 
35 // A producer-consumer queue of up to N integers,
36 // illustrating the use of the calls in this interface.
37 
38 #define N 4
39 
40 typedef struct queue {
41   gpr_cv non_empty;  // Signalled when length becomes non-zero.
42   gpr_cv non_full;   // Signalled when length becomes non-N.
43   gpr_mu mu;         // Protects all fields below.
44                      // (That is, except during initialization or
45                      // destruction, the fields below should be accessed
46                      // only by a thread that holds mu.)
47   int head;          // Index of head of queue 0..N-1.
48   int length;        // Number of valid elements in queue 0..N.
49   int elem[N];       // elem[head .. head+length-1] are queue elements.
50 } queue;
51 
52 // Initialize *q.
queue_init(queue * q)53 void queue_init(queue* q) {
54   gpr_mu_init(&q->mu);
55   gpr_cv_init(&q->non_empty);
56   gpr_cv_init(&q->non_full);
57   q->head = 0;
58   q->length = 0;
59 }
60 
61 // Free storage associated with *q.
queue_destroy(queue * q)62 void queue_destroy(queue* q) {
63   gpr_mu_destroy(&q->mu);
64   gpr_cv_destroy(&q->non_empty);
65   gpr_cv_destroy(&q->non_full);
66 }
67 
68 // Wait until there is room in *q, then append x to *q.
queue_append(queue * q,int x)69 void queue_append(queue* q, int x) {
70   gpr_mu_lock(&q->mu);
71   // To wait for a predicate without a deadline, loop on the negation of the
72   // predicate, and use gpr_cv_wait(..., gpr_inf_future(GPR_CLOCK_REALTIME))
73   // inside the loop
74   // to release the lock, wait, and reacquire on each iteration.  Code that
75   // makes the condition true should use gpr_cv_broadcast() on the
76   // corresponding condition variable.  The predicate must be on state
77   // protected by the lock.
78   while (q->length == N) {
79     gpr_cv_wait(&q->non_full, &q->mu, gpr_inf_future(GPR_CLOCK_MONOTONIC));
80   }
81   if (q->length == 0) {  // Wake threads blocked in queue_remove().
82     // It's normal to use gpr_cv_broadcast() or gpr_signal() while
83     // holding the lock.
84     gpr_cv_broadcast(&q->non_empty);
85   }
86   q->elem[(q->head + q->length) % N] = x;
87   q->length++;
88   gpr_mu_unlock(&q->mu);
89 }
90 
91 // If it can be done without blocking, append x to *q and return non-zero.
92 // Otherwise return 0.
queue_try_append(queue * q,int x)93 int queue_try_append(queue* q, int x) {
94   int result = 0;
95   if (gpr_mu_trylock(&q->mu)) {
96     if (q->length != N) {
97       if (q->length == 0) {  // Wake threads blocked in queue_remove().
98         gpr_cv_broadcast(&q->non_empty);
99       }
100       q->elem[(q->head + q->length) % N] = x;
101       q->length++;
102       result = 1;
103     }
104     gpr_mu_unlock(&q->mu);
105   }
106   return result;
107 }
108 
109 // Wait until the *q is non-empty or deadline abs_deadline passes.  If the
110 // queue is non-empty, remove its head entry, place it in *head, and return
111 // non-zero.  Otherwise return 0.
queue_remove(queue * q,int * head,gpr_timespec abs_deadline)112 int queue_remove(queue* q, int* head, gpr_timespec abs_deadline) {
113   int result = 0;
114   gpr_mu_lock(&q->mu);
115   // To wait for a predicate with a deadline, loop on the negation of the
116   // predicate or until gpr_cv_wait() returns true.  Code that makes
117   // the condition true should use gpr_cv_broadcast() on the corresponding
118   // condition variable.  The predicate must be on state protected by the
119   // lock.
120   while (q->length == 0 && !gpr_cv_wait(&q->non_empty, &q->mu, abs_deadline)) {
121   }
122   if (q->length != 0) {  // Queue is non-empty.
123     result = 1;
124     if (q->length == N) {  // Wake threads blocked in queue_append().
125       gpr_cv_broadcast(&q->non_full);
126     }
127     *head = q->elem[q->head];
128     q->head = (q->head + 1) % N;
129     q->length--;
130   }  // else deadline exceeded
131   gpr_mu_unlock(&q->mu);
132   return result;
133 }
134 
135 // -------------------------------------------------
136 // Tests for gpr_mu and gpr_cv, and the queue example.
137 struct test {
138   int nthreads;  // number of threads
139   grpc_core::Thread* threads;
140 
141   int64_t iterations;  // number of iterations per thread
142   int64_t counter;
143   int thread_count;  // used to allocate thread ids
144   int done;          // threads not yet completed
145   int incr_step;     // how much to increment/decrement refcount each time
146 
147   gpr_mu mu;  // protects iterations, counter, thread_count, done
148 
149   gpr_cv cv;  // signalling depends on test
150 
151   gpr_cv done_cv;  // signalled when done == 0
152 
153   queue q;
154 
155   gpr_stats_counter stats_counter;
156 
157   gpr_refcount refcount;
158   gpr_refcount thread_refcount;
159   gpr_event event;
160 };
161 
162 // Return pointer to a new struct test.
test_new(int nthreads,int64_t iterations,int incr_step)163 static struct test* test_new(int nthreads, int64_t iterations, int incr_step) {
164   struct test* m = static_cast<struct test*>(gpr_malloc(sizeof(*m)));
165   m->nthreads = nthreads;
166   m->threads = static_cast<grpc_core::Thread*>(
167       gpr_malloc(sizeof(*m->threads) * nthreads));
168   m->iterations = iterations;
169   m->counter = 0;
170   m->thread_count = 0;
171   m->done = nthreads;
172   m->incr_step = incr_step;
173   gpr_mu_init(&m->mu);
174   gpr_cv_init(&m->cv);
175   gpr_cv_init(&m->done_cv);
176   queue_init(&m->q);
177   gpr_stats_init(&m->stats_counter, 0);
178   gpr_ref_init(&m->refcount, 0);
179   gpr_ref_init(&m->thread_refcount, nthreads);
180   gpr_event_init(&m->event);
181   return m;
182 }
183 
184 // Return pointer to a new struct test.
test_destroy(struct test * m)185 static void test_destroy(struct test* m) {
186   gpr_mu_destroy(&m->mu);
187   gpr_cv_destroy(&m->cv);
188   gpr_cv_destroy(&m->done_cv);
189   queue_destroy(&m->q);
190   gpr_free(m->threads);
191   gpr_free(m);
192 }
193 
194 // Create m->nthreads threads, each running (*body)(m)
test_create_threads(struct test * m,void (* body)(void * arg))195 static void test_create_threads(struct test* m, void (*body)(void* arg)) {
196   int i;
197   for (i = 0; i != m->nthreads; i++) {
198     m->threads[i] = grpc_core::Thread("grpc_create_threads", body, m);
199     m->threads[i].Start();
200   }
201 }
202 
203 // Wait until all threads report done.
test_wait(struct test * m)204 static void test_wait(struct test* m) {
205   gpr_mu_lock(&m->mu);
206   while (m->done != 0) {
207     gpr_cv_wait(&m->done_cv, &m->mu, gpr_inf_future(GPR_CLOCK_MONOTONIC));
208   }
209   gpr_mu_unlock(&m->mu);
210   for (int i = 0; i != m->nthreads; i++) {
211     m->threads[i].Join();
212   }
213 }
214 
215 // Get an integer thread id in the raneg 0..nthreads-1
thread_id(struct test * m)216 static int thread_id(struct test* m) {
217   int id;
218   gpr_mu_lock(&m->mu);
219   id = m->thread_count++;
220   gpr_mu_unlock(&m->mu);
221   return id;
222 }
223 
224 // Indicate that a thread is done, by decrementing m->done
225 // and signalling done_cv if m->done==0.
mark_thread_done(struct test * m)226 static void mark_thread_done(struct test* m) {
227   gpr_mu_lock(&m->mu);
228   ASSERT_NE(m->done, 0);
229   m->done--;
230   if (m->done == 0) {
231     gpr_cv_signal(&m->done_cv);
232   }
233   gpr_mu_unlock(&m->mu);
234 }
235 
236 // Test several threads running (*body)(struct test *m) for increasing settings
237 // of m->iterations, until about timeout_s to 2*timeout_s seconds have elapsed.
238 // If extra!=NULL, run (*extra)(m) in an additional thread.
239 // incr_step controls by how much m->refcount should be incremented/decremented
240 // (if at all) each time in the tests.
241 //
test(const char * name,void (* body)(void * m),void (* extra)(void * m),int timeout_s,int incr_step)242 static void test(const char* name, void (*body)(void* m),
243                  void (*extra)(void* m), int timeout_s, int incr_step) {
244   int64_t iterations = 8;
245   struct test* m;
246   gpr_timespec start = gpr_now(GPR_CLOCK_REALTIME);
247   gpr_timespec time_taken;
248   gpr_timespec deadline = gpr_time_add(
249       start, gpr_time_from_micros(static_cast<int64_t>(timeout_s) * 1000000,
250                                   GPR_TIMESPAN));
251   fprintf(stderr, "%s:", name);
252   fflush(stderr);
253   while (gpr_time_cmp(gpr_now(GPR_CLOCK_REALTIME), deadline) < 0) {
254     fprintf(stderr, " %ld", static_cast<long>(iterations));
255     fflush(stderr);
256     m = test_new(10, iterations, incr_step);
257     grpc_core::Thread extra_thd;
258     if (extra != nullptr) {
259       extra_thd = grpc_core::Thread(name, extra, m);
260       extra_thd.Start();
261       m->done++;  // one more thread to wait for
262     }
263     test_create_threads(m, body);
264     test_wait(m);
265     if (extra != nullptr) {
266       extra_thd.Join();
267     }
268     if (m->counter != m->nthreads * m->iterations * m->incr_step) {
269       fprintf(stderr, "counter %ld  threads %d  iterations %ld\n",
270               static_cast<long>(m->counter), m->nthreads,
271               static_cast<long>(m->iterations));
272       fflush(stderr);
273       ASSERT_TRUE(0);
274     }
275     test_destroy(m);
276     iterations <<= 1;
277   }
278   time_taken = gpr_time_sub(gpr_now(GPR_CLOCK_REALTIME), start);
279   fprintf(stderr, " done %lld.%09d s\n",
280           static_cast<long long>(time_taken.tv_sec),
281           static_cast<int>(time_taken.tv_nsec));
282   fflush(stderr);
283 }
284 
285 // Increment m->counter on each iteration; then mark thread as done.
inc(void * v)286 static void inc(void* v /*=m*/) {
287   struct test* m = static_cast<struct test*>(v);
288   int64_t i;
289   for (i = 0; i != m->iterations; i++) {
290     gpr_mu_lock(&m->mu);
291     m->counter++;
292     gpr_mu_unlock(&m->mu);
293   }
294   mark_thread_done(m);
295 }
296 
297 // Increment m->counter under lock acquired with trylock, m->iterations times;
298 // then mark thread as done.
inctry(void * v)299 static void inctry(void* v /*=m*/) {
300   struct test* m = static_cast<struct test*>(v);
301   int64_t i;
302   for (i = 0; i != m->iterations;) {
303     if (gpr_mu_trylock(&m->mu)) {
304       m->counter++;
305       gpr_mu_unlock(&m->mu);
306       i++;
307     }
308   }
309   mark_thread_done(m);
310 }
311 
312 // Increment counter only when (m->counter%m->nthreads)==m->thread_id; then mark
313 // thread as done.
inc_by_turns(void * v)314 static void inc_by_turns(void* v /*=m*/) {
315   struct test* m = static_cast<struct test*>(v);
316   int64_t i;
317   int id = thread_id(m);
318   for (i = 0; i != m->iterations; i++) {
319     gpr_mu_lock(&m->mu);
320     while ((m->counter % m->nthreads) != id) {
321       gpr_cv_wait(&m->cv, &m->mu, gpr_inf_future(GPR_CLOCK_MONOTONIC));
322     }
323     m->counter++;
324     gpr_cv_broadcast(&m->cv);
325     gpr_mu_unlock(&m->mu);
326   }
327   mark_thread_done(m);
328 }
329 
330 // Wait a millisecond and increment counter on each iteration;
331 // then mark thread as done.
inc_with_1ms_delay(void * v)332 static void inc_with_1ms_delay(void* v /*=m*/) {
333   struct test* m = static_cast<struct test*>(v);
334   int64_t i;
335   for (i = 0; i != m->iterations; i++) {
336     gpr_timespec deadline;
337     gpr_mu_lock(&m->mu);
338     deadline = gpr_time_add(gpr_now(GPR_CLOCK_MONOTONIC),
339                             gpr_time_from_micros(1000, GPR_TIMESPAN));
340     while (!gpr_cv_wait(&m->cv, &m->mu, deadline)) {
341     }
342     m->counter++;
343     gpr_mu_unlock(&m->mu);
344   }
345   mark_thread_done(m);
346 }
347 
348 // Wait a millisecond and increment counter on each iteration, using an event
349 // for timing; then mark thread as done.
inc_with_1ms_delay_event(void * v)350 static void inc_with_1ms_delay_event(void* v /*=m*/) {
351   struct test* m = static_cast<struct test*>(v);
352   int64_t i;
353   for (i = 0; i != m->iterations; i++) {
354     gpr_timespec deadline;
355     deadline = gpr_time_add(gpr_now(GPR_CLOCK_REALTIME),
356                             gpr_time_from_micros(1000, GPR_TIMESPAN));
357     ASSERT_EQ(gpr_event_wait(&m->event, deadline), nullptr);
358     gpr_mu_lock(&m->mu);
359     m->counter++;
360     gpr_mu_unlock(&m->mu);
361   }
362   mark_thread_done(m);
363 }
364 
365 // Produce m->iterations elements on queue m->q, then mark thread as done.
366 // Even threads use queue_append(), and odd threads use queue_try_append()
367 // until it succeeds.
many_producers(void * v)368 static void many_producers(void* v /*=m*/) {
369   struct test* m = static_cast<struct test*>(v);
370   int64_t i;
371   int x = thread_id(m);
372   if ((x & 1) == 0) {
373     for (i = 0; i != m->iterations; i++) {
374       queue_append(&m->q, 1);
375     }
376   } else {
377     for (i = 0; i != m->iterations; i++) {
378       while (!queue_try_append(&m->q, 1)) {
379       }
380     }
381   }
382   mark_thread_done(m);
383 }
384 
385 // Consume elements from m->q until m->nthreads*m->iterations are seen,
386 // wait an extra second to confirm that no more elements are arriving,
387 // then mark thread as done.
consumer(void * v)388 static void consumer(void* v /*=m*/) {
389   struct test* m = static_cast<struct test*>(v);
390   int64_t n = m->iterations * m->nthreads;
391   int64_t i;
392   int value;
393   for (i = 0; i != n; i++) {
394     queue_remove(&m->q, &value, gpr_inf_future(GPR_CLOCK_MONOTONIC));
395   }
396   gpr_mu_lock(&m->mu);
397   m->counter = n;
398   gpr_mu_unlock(&m->mu);
399   ASSERT_TRUE(
400       !queue_remove(&m->q, &value,
401                     gpr_time_add(gpr_now(GPR_CLOCK_MONOTONIC),
402                                  gpr_time_from_micros(1000000, GPR_TIMESPAN))));
403   mark_thread_done(m);
404 }
405 
406 // Increment m->stats_counter m->iterations times, transfer counter value to
407 // m->counter, then mark thread as done.
statsinc(void * v)408 static void statsinc(void* v /*=m*/) {
409   struct test* m = static_cast<struct test*>(v);
410   int64_t i;
411   for (i = 0; i != m->iterations; i++) {
412     gpr_stats_inc(&m->stats_counter, 1);
413   }
414   gpr_mu_lock(&m->mu);
415   m->counter = gpr_stats_read(&m->stats_counter);
416   gpr_mu_unlock(&m->mu);
417   mark_thread_done(m);
418 }
419 
420 // Increment m->refcount by m->incr_step for m->iterations times. Decrement
421 // m->thread_refcount once, and if it reaches zero, set m->event to (void*)1;
422 // then mark thread as done.
refinc(void * v)423 static void refinc(void* v /*=m*/) {
424   struct test* m = static_cast<struct test*>(v);
425   int64_t i;
426   for (i = 0; i != m->iterations; i++) {
427     if (m->incr_step == 1) {
428       gpr_ref(&m->refcount);
429     } else {
430       gpr_refn(&m->refcount, m->incr_step);
431     }
432   }
433   if (gpr_unref(&m->thread_refcount)) {
434     gpr_event_set(&m->event, reinterpret_cast<void*>(1));
435   }
436   mark_thread_done(m);
437 }
438 
439 // Wait until m->event is set to (void *)1, then decrement m->refcount by 1
440 // (m->nthreads * m->iterations * m->incr_step) times, and ensure that the last
441 // decrement caused the counter to reach zero, then mark thread as done.
refcheck(void * v)442 static void refcheck(void* v /*=m*/) {
443   struct test* m = static_cast<struct test*>(v);
444   int64_t n = m->iterations * m->nthreads * m->incr_step;
445   int64_t i;
446   ASSERT_EQ(gpr_event_wait(&m->event, gpr_inf_future(GPR_CLOCK_REALTIME)),
447             (void*)1);
448   ASSERT_EQ(gpr_event_get(&m->event), (void*)1);
449   for (i = 1; i != n; i++) {
450     ASSERT_FALSE(gpr_unref(&m->refcount));
451     m->counter++;
452   }
453   ASSERT_TRUE(gpr_unref(&m->refcount));
454   m->counter++;
455   mark_thread_done(m);
456 }
457 
458 // -------------------------------------------------
459 
TEST(SyncTest,MainTest)460 TEST(SyncTest, MainTest) {
461   test("mutex", &inc, nullptr, 1, 1);
462   test("mutex try", &inctry, nullptr, 1, 1);
463   test("cv", &inc_by_turns, nullptr, 1, 1);
464   test("timedcv", &inc_with_1ms_delay, nullptr, 1, 1);
465   test("queue", &many_producers, &consumer, 10, 1);
466   test("stats_counter", &statsinc, nullptr, 1, 1);
467   test("refcount by 1", &refinc, &refcheck, 1, 1);
468   test("refcount by 3", &refinc, &refcheck, 1, 3);  // incr_step of 3 is an
469                                                     // arbitrary choice. Any
470                                                     // number > 1 is okay here
471   test("timedevent", &inc_with_1ms_delay_event, nullptr, 1, 1);
472 }
473 
main(int argc,char ** argv)474 int main(int argc, char** argv) {
475   grpc::testing::TestEnvironment env(&argc, argv);
476   ::testing::InitGoogleTest(&argc, argv);
477   return RUN_ALL_TESTS();
478 }
479