• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * QEMU System Emulator
3  *
4  * Copyright (c) 2003-2008 Fabrice Bellard
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a copy
7  * of this software and associated documentation files (the "Software"), to deal
8  * in the Software without restriction, including without limitation the rights
9  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10  * copies of the Software, and to permit persons to whom the Software is
11  * furnished to do so, subject to the following conditions:
12  *
13  * The above copyright notice and this permission notice shall be included in
14  * all copies or substantial portions of the Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22  * THE SOFTWARE.
23  */
24 
25 #include "qemu-common.h"
26 #include "block/aio.h"
27 
28 /*
29  * An AsyncContext protects the callbacks of AIO requests and Bottom Halves
30  * against interfering with each other. A typical example is qcow2 that accepts
31  * asynchronous requests, but relies for manipulation of its metadata on
32  * synchronous bdrv_read/write that doesn't trigger any callbacks.
33  *
34  * However, these functions are often emulated using AIO which means that AIO
35  * callbacks must be run - but at the same time we must not run callbacks of
36  * other requests as they might start to modify metadata and corrupt the
37  * internal state of the caller of bdrv_read/write.
38  *
39  * To achieve the desired semantics we switch into a new AsyncContext.
40  * Callbacks must only be run if they belong to the current AsyncContext.
41  * Otherwise they need to be queued until their own context is active again.
42  * This is how you can make qemu_aio_wait() wait only for your own callbacks.
43  *
44  * The AsyncContexts form a stack. When you leave a AsyncContexts, you always
45  * return to the old ("parent") context.
46  */
47 struct AsyncContext {
48     /* Consecutive number of the AsyncContext (position in the stack) */
49     int id;
50 
51     /* Anchor of the list of Bottom Halves belonging to the context */
52     struct QEMUBH *first_bh;
53 
54     /* Link to parent context */
55     struct AsyncContext *parent;
56 };
57 
58 /* The currently active AsyncContext */
59 static struct AsyncContext *async_context = &(struct AsyncContext) { 0 };
60 
61 /*
62  * Enter a new AsyncContext. Already scheduled Bottom Halves and AIO callbacks
63  * won't be called until this context is left again.
64  */
async_context_push(void)65 void async_context_push(void)
66 {
67     struct AsyncContext *new = g_malloc0(sizeof(*new));
68     new->parent = async_context;
69     new->id = async_context->id + 1;
70     async_context = new;
71 }
72 
73 /* Run queued AIO completions and destroy Bottom Half */
bh_run_aio_completions(void * opaque)74 static void bh_run_aio_completions(void *opaque)
75 {
76     QEMUBH **bh = opaque;
77     qemu_bh_delete(*bh);
78     g_free(bh);
79     qemu_aio_process_queue();
80 }
81 /*
82  * Leave the currently active AsyncContext. All Bottom Halves belonging to the
83  * old context are executed before changing the context.
84  */
async_context_pop(void)85 void async_context_pop(void)
86 {
87     struct AsyncContext *old = async_context;
88     QEMUBH **bh;
89 
90     /* Flush the bottom halves, we don't want to lose them */
91     while (qemu_bh_poll());
92 
93     /* Switch back to the parent context */
94     async_context = async_context->parent;
95     g_free(old);
96 
97     if (async_context == NULL) {
98         abort();
99     }
100 
101     /* Schedule BH to run any queued AIO completions as soon as possible */
102     bh = g_malloc(sizeof(*bh));
103     *bh = qemu_bh_new(bh_run_aio_completions, bh);
104     qemu_bh_schedule(*bh);
105 }
106 
107 /*
108  * Returns the ID of the currently active AsyncContext
109  */
get_async_context_id(void)110 int get_async_context_id(void)
111 {
112     return async_context->id;
113 }
114 
115 /***********************************************************/
116 /* bottom halves (can be seen as timers which expire ASAP) */
117 
118 struct QEMUBH {
119     QEMUBHFunc *cb;
120     void *opaque;
121     int scheduled;
122     int idle;
123     int deleted;
124     QEMUBH *next;
125 };
126 
qemu_bh_new(QEMUBHFunc * cb,void * opaque)127 QEMUBH *qemu_bh_new(QEMUBHFunc *cb, void *opaque)
128 {
129     QEMUBH *bh;
130     bh = g_malloc0(sizeof(QEMUBH));
131     bh->cb = cb;
132     bh->opaque = opaque;
133     bh->next = async_context->first_bh;
134     async_context->first_bh = bh;
135     return bh;
136 }
137 
qemu_bh_poll(void)138 int qemu_bh_poll(void)
139 {
140     QEMUBH *bh, **bhp;
141     int ret;
142 
143     ret = 0;
144     for (bh = async_context->first_bh; bh; bh = bh->next) {
145         if (!bh->deleted && bh->scheduled) {
146             bh->scheduled = 0;
147             if (!bh->idle)
148                 ret = 1;
149             bh->idle = 0;
150             bh->cb(bh->opaque);
151         }
152     }
153 
154     /* remove deleted bhs */
155     bhp = &async_context->first_bh;
156     while (*bhp) {
157         bh = *bhp;
158         if (bh->deleted) {
159             *bhp = bh->next;
160             g_free(bh);
161         } else
162             bhp = &bh->next;
163     }
164 
165     return ret;
166 }
167 
qemu_bh_schedule_idle(QEMUBH * bh)168 void qemu_bh_schedule_idle(QEMUBH *bh)
169 {
170     if (bh->scheduled)
171         return;
172     bh->scheduled = 1;
173     bh->idle = 1;
174 }
175 
qemu_bh_schedule(QEMUBH * bh)176 void qemu_bh_schedule(QEMUBH *bh)
177 {
178     if (bh->scheduled)
179         return;
180     bh->scheduled = 1;
181     bh->idle = 0;
182     /* stop the currently executing CPU to execute the BH ASAP */
183     qemu_notify_event();
184 }
185 
qemu_bh_cancel(QEMUBH * bh)186 void qemu_bh_cancel(QEMUBH *bh)
187 {
188     bh->scheduled = 0;
189 }
190 
qemu_bh_delete(QEMUBH * bh)191 void qemu_bh_delete(QEMUBH *bh)
192 {
193     bh->scheduled = 0;
194     bh->deleted = 1;
195 }
196 
qemu_bh_update_timeout(int * timeout)197 void qemu_bh_update_timeout(int *timeout)
198 {
199     QEMUBH *bh;
200 
201     for (bh = async_context->first_bh; bh; bh = bh->next) {
202         if (!bh->deleted && bh->scheduled) {
203             if (bh->idle) {
204                 /* idle bottom halves will be polled at least
205                  * every 10ms */
206                 *timeout = MIN(10, *timeout);
207             } else {
208                 /* non-idle bottom halves will be executed
209                  * immediately */
210                 *timeout = 0;
211                 break;
212             }
213         }
214     }
215 }
216 
217