• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /* Copyright Joyent, Inc. and other Node contributors. All rights reserved.
2  *
3  * Permission is hereby granted, free of charge, to any person obtaining a copy
4  * of this software and associated documentation files (the "Software"), to
5  * deal in the Software without restriction, including without limitation the
6  * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
7  * sell copies of the Software, and to permit persons to whom the Software is
8  * furnished to do so, subject to the following conditions:
9  *
10  * The above copyright notice and this permission notice shall be included in
11  * all copies or substantial portions of the Software.
12  *
13  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
18  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
19  * IN THE SOFTWARE.
20  */
21 
22 #include "uv.h"
23 #include "uv/tree.h"
24 #include "internal.h"
25 #include "heap-inl.h"
26 #include <stdlib.h>
27 #include <string.h>
28 #include <unistd.h>
29 
uv_loop_init(uv_loop_t * loop)30 int uv_loop_init(uv_loop_t* loop) {
31   uv__loop_internal_fields_t* lfields;
32   void* saved_data;
33   int err;
34 
35 
36   saved_data = loop->data;
37   memset(loop, 0, sizeof(*loop));
38   loop->data = saved_data;
39 
40   lfields = (uv__loop_internal_fields_t*) uv__calloc(1, sizeof(*lfields));
41   if (lfields == NULL)
42     return UV_ENOMEM;
43   loop->internal_fields = lfields;
44 
45   err = uv_mutex_init(&lfields->loop_metrics.lock);
46   if (err)
47     goto fail_metrics_mutex_init;
48 
49   heap_init((struct heap*) &loop->timer_heap);
50   QUEUE_INIT(&loop->wq);
51 #ifdef USE_FFRT
52   uv__loop_internal_fields_t* lfields_qos = uv__get_internal_fields(loop);
53   QUEUE_INIT(&(lfields_qos->wq_sub[uv_qos_background]));
54   QUEUE_INIT(&(lfields_qos->wq_sub[uv_qos_utility]));
55   QUEUE_INIT(&(lfields_qos->wq_sub[uv_qos_default]));
56   QUEUE_INIT(&(lfields_qos->wq_sub[uv_qos_user_initiated]));
57 #endif
58   QUEUE_INIT(&loop->idle_handles);
59   QUEUE_INIT(&loop->async_handles);
60   QUEUE_INIT(&loop->check_handles);
61   QUEUE_INIT(&loop->prepare_handles);
62   QUEUE_INIT(&loop->handle_queue);
63 
64   loop->active_handles = 0;
65   loop->active_reqs.count = 0;
66   loop->nfds = 0;
67   loop->watchers = NULL;
68   loop->nwatchers = 0;
69   QUEUE_INIT(&loop->pending_queue);
70   QUEUE_INIT(&loop->watcher_queue);
71 
72   loop->closing_handles = NULL;
73   uv__update_time(loop);
74   loop->async_io_watcher.fd = -1;
75   loop->async_wfd = -1;
76   loop->signal_pipefd[0] = -1;
77   loop->signal_pipefd[1] = -1;
78   loop->backend_fd = -1;
79   loop->emfile_fd = -1;
80 
81   loop->timer_counter = 0;
82   loop->stop_flag = 0;
83 
84   err = uv__platform_loop_init(loop);
85   if (err)
86     goto fail_platform_init;
87 
88   uv__signal_global_once_init();
89   err = uv_signal_init(loop, &loop->child_watcher);
90   if (err)
91     goto fail_signal_init;
92 
93   uv__handle_unref(&loop->child_watcher);
94   loop->child_watcher.flags |= UV_HANDLE_INTERNAL;
95   QUEUE_INIT(&loop->process_handles);
96 
97   err = uv_rwlock_init(&loop->cloexec_lock);
98   if (err)
99     goto fail_rwlock_init;
100 
101   err = uv_mutex_init(&loop->wq_mutex);
102   if (err)
103     goto fail_mutex_init;
104 
105   err = uv_async_init(loop, &loop->wq_async, uv__work_done);
106   if (err)
107     goto fail_async_init;
108 
109   uv__handle_unref(&loop->wq_async);
110   loop->wq_async.flags |= UV_HANDLE_INTERNAL;
111 
112   loop->magic = UV_LOOP_MAGIC;
113   return 0;
114 
115 fail_async_init:
116   uv_mutex_destroy(&loop->wq_mutex);
117 
118 fail_mutex_init:
119   uv_rwlock_destroy(&loop->cloexec_lock);
120 
121 fail_rwlock_init:
122   uv__signal_loop_cleanup(loop);
123 
124 fail_signal_init:
125   uv__platform_loop_delete(loop);
126 
127 fail_platform_init:
128   uv_mutex_destroy(&lfields->loop_metrics.lock);
129 
130 fail_metrics_mutex_init:
131   uv__free(lfields);
132   loop->internal_fields = NULL;
133 
134   uv__free(loop->watchers);
135   loop->nwatchers = 0;
136   return err;
137 }
138 
139 
uv_loop_fork(uv_loop_t * loop)140 int uv_loop_fork(uv_loop_t* loop) {
141   int err;
142   unsigned int i;
143   uv__io_t* w;
144 
145   err = uv__io_fork(loop);
146   if (err)
147     return err;
148 
149   err = uv__async_fork(loop);
150   if (err)
151     return err;
152 
153   err = uv__signal_loop_fork(loop);
154   if (err)
155     return err;
156 
157   /* Rearm all the watchers that aren't re-queued by the above. */
158   for (i = 0; i < loop->nwatchers; i++) {
159     w = loop->watchers[i];
160     if (w == NULL)
161       continue;
162 
163     if (w->pevents != 0 && QUEUE_EMPTY(&w->watcher_queue)) {
164       w->events = 0; /* Force re-registration in uv__io_poll. */
165       QUEUE_INSERT_TAIL(&loop->watcher_queue, &w->watcher_queue);
166     }
167   }
168 
169   return 0;
170 }
171 
172 
uv__loop_close(uv_loop_t * loop)173 void uv__loop_close(uv_loop_t* loop) {
174   uv__loop_internal_fields_t* lfields;
175 
176   uv__signal_loop_cleanup(loop);
177   uv__platform_loop_delete(loop);
178   uv__async_stop(loop);
179 
180   if (loop->emfile_fd != -1) {
181     uv__close(loop->emfile_fd);
182     loop->emfile_fd = -1;
183   }
184 
185   if (loop->backend_fd != -1) {
186     uv__close(loop->backend_fd);
187     loop->backend_fd = -1;
188   }
189 
190   uv_mutex_lock(&loop->wq_mutex);
191 #ifndef USE_FFRT
192   assert(QUEUE_EMPTY(&loop->wq) && "thread pool work queue not empty!");
193 #else
194   uv__loop_internal_fields_t* lfields_qos = uv__get_internal_fields(loop);
195   assert(QUEUE_EMPTY(&(lfields_qos->wq_sub[uv_qos_background])) && "thread pool work queue qos_background not empty!");
196   assert(QUEUE_EMPTY(&(lfields_qos->wq_sub[uv_qos_utility])) && "thread pool work queue qos_utility not empty!");
197   assert(QUEUE_EMPTY(&(lfields_qos->wq_sub[uv_qos_default])) && "thread pool work queue qos_default not empty!");
198   assert(QUEUE_EMPTY(&(lfields_qos->wq_sub[uv_qos_user_initiated])) && "thread pool work queue qos_user_initiated not empty!");
199 #endif
200   assert(!uv__has_active_reqs(loop));
201   uv_mutex_unlock(&loop->wq_mutex);
202   uv_mutex_destroy(&loop->wq_mutex);
203 
204   /*
205    * Note that all thread pool stuff is finished at this point and
206    * it is safe to just destroy rw lock
207    */
208   uv_rwlock_destroy(&loop->cloexec_lock);
209 
210 #if 0
211   assert(QUEUE_EMPTY(&loop->pending_queue));
212   assert(QUEUE_EMPTY(&loop->watcher_queue));
213   assert(loop->nfds == 0);
214 #endif
215 
216   uv__free(loop->watchers);
217   loop->watchers = NULL;
218   loop->nwatchers = 0;
219 
220   lfields = uv__get_internal_fields(loop);
221   uv_mutex_destroy(&lfields->loop_metrics.lock);
222   uv__free(lfields);
223   loop->internal_fields = NULL;
224   loop->magic = ~UV_LOOP_MAGIC;
225 }
226 
227 
uv__loop_configure(uv_loop_t * loop,uv_loop_option option,va_list ap)228 int uv__loop_configure(uv_loop_t* loop, uv_loop_option option, va_list ap) {
229   uv__loop_internal_fields_t* lfields;
230 
231   lfields = uv__get_internal_fields(loop);
232   if (option == UV_METRICS_IDLE_TIME) {
233     lfields->flags |= UV_METRICS_IDLE_TIME;
234     return 0;
235   }
236 
237   if (option != UV_LOOP_BLOCK_SIGNAL)
238     return UV_ENOSYS;
239 
240   if (va_arg(ap, int) != SIGPROF)
241     return UV_EINVAL;
242 
243   loop->flags |= UV_LOOP_BLOCK_SIGPROF;
244   return 0;
245 }
246