async.c 8.3 KB
Newer Older
K
Kevin Wolf 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
/*
 * QEMU System Emulator
 *
 * Copyright (c) 2003-2008 Fabrice Bellard
 *
 * Permission is hereby granted, free of charge, to any person obtaining a copy
 * of this software and associated documentation files (the "Software"), to deal
 * in the Software without restriction, including without limitation the rights
 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 * copies of the Software, and to permit persons to whom the Software is
 * furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice shall be included in
 * all copies or substantial portions of the Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
 * THE SOFTWARE.
 */

#include "qemu-common.h"
26
#include "block/aio.h"
27
#include "block/thread-pool.h"
28
#include "qemu/main-loop.h"
P
Paolo Bonzini 已提交
29
#include "qemu/atomic.h"
30

K
Kevin Wolf 已提交
31 32 33 34
/***********************************************************/
/* bottom halves (can be seen as timers which expire ASAP) */

struct QEMUBH {
P
Paolo Bonzini 已提交
35
    AioContext *ctx;
K
Kevin Wolf 已提交
36 37 38
    QEMUBHFunc *cb;
    void *opaque;
    QEMUBH *next;
39 40 41
    bool scheduled;
    bool idle;
    bool deleted;
K
Kevin Wolf 已提交
42 43
};

44
QEMUBH *aio_bh_new(AioContext *ctx, QEMUBHFunc *cb, void *opaque)
K
Kevin Wolf 已提交
45 46
{
    QEMUBH *bh;
47 48 49 50 51 52
    bh = g_new(QEMUBH, 1);
    *bh = (QEMUBH){
        .ctx = ctx,
        .cb = cb,
        .opaque = opaque,
    };
53
    qemu_mutex_lock(&ctx->bh_lock);
54
    bh->next = ctx->first_bh;
55 56
    /* Make sure that the members are ready before putting bh into list */
    smp_wmb();
57
    ctx->first_bh = bh;
58
    qemu_mutex_unlock(&ctx->bh_lock);
K
Kevin Wolf 已提交
59 60 61
    return bh;
}

62
/* Multiple occurrences of aio_bh_poll cannot be called concurrently */
63
int aio_bh_poll(AioContext *ctx)
K
Kevin Wolf 已提交
64
{
65
    QEMUBH *bh, **bhp, *next;
K
Kevin Wolf 已提交
66
    int ret;
67

68
    ctx->walking_bh++;
K
Kevin Wolf 已提交
69 70

    ret = 0;
71
    for (bh = ctx->first_bh; bh; bh = next) {
72 73
        /* Make sure that fetching bh happens before accessing its members */
        smp_read_barrier_depends();
74
        next = bh->next;
K
Kevin Wolf 已提交
75 76
        if (!bh->deleted && bh->scheduled) {
            bh->scheduled = 0;
77 78 79 80
            /* Paired with write barrier in bh schedule to ensure reading for
             * idle & callbacks coming after bh's scheduling.
             */
            smp_rmb();
K
Kevin Wolf 已提交
81 82 83 84 85 86 87
            if (!bh->idle)
                ret = 1;
            bh->idle = 0;
            bh->cb(bh->opaque);
        }
    }

88
    ctx->walking_bh--;
89

K
Kevin Wolf 已提交
90
    /* remove deleted bhs */
91
    if (!ctx->walking_bh) {
92
        qemu_mutex_lock(&ctx->bh_lock);
93
        bhp = &ctx->first_bh;
94 95 96 97 98 99 100 101 102
        while (*bhp) {
            bh = *bhp;
            if (bh->deleted) {
                *bhp = bh->next;
                g_free(bh);
            } else {
                bhp = &bh->next;
            }
        }
103
        qemu_mutex_unlock(&ctx->bh_lock);
K
Kevin Wolf 已提交
104 105 106 107 108 109 110 111 112 113
    }

    return ret;
}

void qemu_bh_schedule_idle(QEMUBH *bh)
{
    if (bh->scheduled)
        return;
    bh->idle = 1;
114 115 116 117 118
    /* Make sure that idle & any writes needed by the callback are done
     * before the locations are read in the aio_bh_poll.
     */
    smp_wmb();
    bh->scheduled = 1;
K
Kevin Wolf 已提交
119 120 121 122
}

void qemu_bh_schedule(QEMUBH *bh)
{
123 124
    AioContext *ctx;

K
Kevin Wolf 已提交
125 126
    if (bh->scheduled)
        return;
127
    ctx = bh->ctx;
K
Kevin Wolf 已提交
128
    bh->idle = 0;
129 130 131 132 133
    /* Make sure that:
     * 1. idle & any writes needed by the callback are done before the
     *    locations are read in the aio_bh_poll.
     * 2. ctx is loaded before scheduled is set and the callback has a chance
     *    to execute.
134
     */
135
    smp_mb();
136
    bh->scheduled = 1;
137
    aio_notify(ctx);
K
Kevin Wolf 已提交
138 139
}

140 141 142

/* This func is async.
 */
K
Kevin Wolf 已提交
143 144 145 146 147
void qemu_bh_cancel(QEMUBH *bh)
{
    bh->scheduled = 0;
}

148 149 150
/* This func is async.The bottom half will do the delete action at the finial
 * end.
 */
K
Kevin Wolf 已提交
151 152 153 154 155 156
void qemu_bh_delete(QEMUBH *bh)
{
    bh->scheduled = 0;
    bh->deleted = 1;
}

157 158
int64_t
aio_compute_timeout(AioContext *ctx)
K
Kevin Wolf 已提交
159
{
160 161
    int64_t deadline;
    int timeout = -1;
K
Kevin Wolf 已提交
162 163
    QEMUBH *bh;

164
    for (bh = ctx->first_bh; bh; bh = bh->next) {
K
Kevin Wolf 已提交
165 166 167 168
        if (!bh->deleted && bh->scheduled) {
            if (bh->idle) {
                /* idle bottom halves will be polled at least
                 * every 10ms */
169
                timeout = 10000000;
K
Kevin Wolf 已提交
170 171 172
            } else {
                /* non-idle bottom halves will be executed
                 * immediately */
173
                return 0;
K
Kevin Wolf 已提交
174 175 176
            }
        }
    }
P
Paolo Bonzini 已提交
177

178
    deadline = timerlistgroup_deadline_ns(&ctx->tlg);
179
    if (deadline == 0) {
180
        return 0;
181
    } else {
182
        return qemu_soonest_timeout(timeout, deadline);
183
    }
184
}
185

186 187 188 189 190 191 192
static gboolean
aio_ctx_prepare(GSource *source, gint    *timeout)
{
    AioContext *ctx = (AioContext *) source;

    /* We assume there is no timeout already supplied */
    *timeout = qemu_timeout_ns_to_ms(aio_compute_timeout(ctx));
193 194 195 196 197

    if (aio_prepare(ctx)) {
        *timeout = 0;
    }

198
    return *timeout == 0;
P
Paolo Bonzini 已提交
199 200 201 202 203 204 205 206 207 208 209 210 211
}

static gboolean
aio_ctx_check(GSource *source)
{
    AioContext *ctx = (AioContext *) source;
    QEMUBH *bh;

    for (bh = ctx->first_bh; bh; bh = bh->next) {
        if (!bh->deleted && bh->scheduled) {
            return true;
	}
    }
212
    return aio_pending(ctx) || (timerlistgroup_deadline_ns(&ctx->tlg) == 0);
P
Paolo Bonzini 已提交
213 214 215 216 217 218 219 220 221 222
}

static gboolean
aio_ctx_dispatch(GSource     *source,
                 GSourceFunc  callback,
                 gpointer     user_data)
{
    AioContext *ctx = (AioContext *) source;

    assert(callback == NULL);
223
    aio_dispatch(ctx);
P
Paolo Bonzini 已提交
224 225 226
    return true;
}

P
Paolo Bonzini 已提交
227 228 229 230 231
static void
aio_ctx_finalize(GSource     *source)
{
    AioContext *ctx = (AioContext *) source;

232
    thread_pool_free(ctx->thread_pool);
S
Stefan Hajnoczi 已提交
233
    aio_set_event_notifier(ctx, &ctx->notifier, NULL);
P
Paolo Bonzini 已提交
234
    event_notifier_cleanup(&ctx->notifier);
235
    rfifolock_destroy(&ctx->lock);
236
    qemu_mutex_destroy(&ctx->bh_lock);
237
    g_array_free(ctx->pollfds, TRUE);
238
    timerlistgroup_deinit(&ctx->tlg);
P
Paolo Bonzini 已提交
239 240
}

P
Paolo Bonzini 已提交
241 242 243 244
static GSourceFuncs aio_source_funcs = {
    aio_ctx_prepare,
    aio_ctx_check,
    aio_ctx_dispatch,
P
Paolo Bonzini 已提交
245
    aio_ctx_finalize
P
Paolo Bonzini 已提交
246 247 248 249 250 251 252
};

GSource *aio_get_g_source(AioContext *ctx)
{
    g_source_ref(&ctx->source);
    return &ctx->source;
}
253

254 255 256 257 258 259 260 261
ThreadPool *aio_get_thread_pool(AioContext *ctx)
{
    if (!ctx->thread_pool) {
        ctx->thread_pool = thread_pool_new(ctx);
    }
    return ctx->thread_pool;
}

P
Paolo Bonzini 已提交
262 263 264 265 266 267 268 269 270 271 272 273
void aio_set_dispatching(AioContext *ctx, bool dispatching)
{
    ctx->dispatching = dispatching;
    if (!dispatching) {
        /* Write ctx->dispatching before reading e.g. bh->scheduled.
         * Optimization: this is only needed when we're entering the "unsafe"
         * phase where other threads must call event_notifier_set.
         */
        smp_mb();
    }
}

P
Paolo Bonzini 已提交
274 275
void aio_notify(AioContext *ctx)
{
P
Paolo Bonzini 已提交
276 277 278 279 280
    /* Write e.g. bh->scheduled before reading ctx->dispatching.  */
    smp_mb();
    if (!ctx->dispatching) {
        event_notifier_set(&ctx->notifier);
    }
P
Paolo Bonzini 已提交
281 282
}

283 284 285 286 287
static void aio_timerlist_notify(void *opaque)
{
    aio_notify(opaque);
}

288 289 290 291 292 293
static void aio_rfifolock_cb(void *opaque)
{
    /* Kick owner thread in case they are blocked in aio_poll() */
    aio_notify(opaque);
}

294
AioContext *aio_context_new(Error **errp)
295
{
296
    int ret;
P
Paolo Bonzini 已提交
297 298
    AioContext *ctx;
    ctx = (AioContext *) g_source_new(&aio_source_funcs, sizeof(AioContext));
299 300 301 302 303 304
    ret = event_notifier_init(&ctx->notifier, false);
    if (ret < 0) {
        g_source_destroy(&ctx->source);
        error_setg_errno(errp, -ret, "Failed to initialize event notifier");
        return NULL;
    }
305
    g_source_set_can_recurse(&ctx->source, true);
306 307 308
    aio_set_event_notifier(ctx, &ctx->notifier,
                           (EventNotifierHandler *)
                           event_notifier_test_and_clear);
309
    ctx->pollfds = g_array_new(FALSE, FALSE, sizeof(GPollFD));
310
    ctx->thread_pool = NULL;
311
    qemu_mutex_init(&ctx->bh_lock);
312
    rfifolock_init(&ctx->lock, aio_rfifolock_cb, ctx);
313
    timerlistgroup_init(&ctx->tlg, aio_timerlist_notify, ctx);
P
Paolo Bonzini 已提交
314 315

    return ctx;
P
Paolo Bonzini 已提交
316 317 318 319 320 321 322 323 324 325
}

void aio_context_ref(AioContext *ctx)
{
    g_source_ref(&ctx->source);
}

void aio_context_unref(AioContext *ctx)
{
    g_source_unref(&ctx->source);
326
}
327 328 329 330 331 332 333 334 335 336

void aio_context_acquire(AioContext *ctx)
{
    rfifolock_lock(&ctx->lock);
}

void aio_context_release(AioContext *ctx)
{
    rfifolock_unlock(&ctx->lock);
}