From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from eggs.gnu.org ([2001:4830:134:3::10]:46251) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1aK63K-0005X4-Un for qemu-devel@nongnu.org; Fri, 15 Jan 2016 10:12:43 -0500 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1aK63G-0006FW-Tt for qemu-devel@nongnu.org; Fri, 15 Jan 2016 10:12:38 -0500 Received: from mx1.redhat.com ([209.132.183.28]:54668) by eggs.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1aK63G-0006FR-Ol for qemu-devel@nongnu.org; Fri, 15 Jan 2016 10:12:34 -0500 Received: from int-mx09.intmail.prod.int.phx2.redhat.com (int-mx09.intmail.prod.int.phx2.redhat.com [10.5.11.22]) by mx1.redhat.com (Postfix) with ESMTPS id 7E2098F875 for ; Fri, 15 Jan 2016 15:12:34 +0000 (UTC) From: Paolo Bonzini Date: Fri, 15 Jan 2016 16:12:11 +0100 Message-Id: <1452870739-28484-9-git-send-email-pbonzini@redhat.com> In-Reply-To: <1452870739-28484-1-git-send-email-pbonzini@redhat.com> References: <1452870739-28484-1-git-send-email-pbonzini@redhat.com> Subject: [Qemu-devel] [PATCH 08/16] aio: rename bh_lock to list_lock List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: qemu-devel@nongnu.org Cc: stefanha@redhat.com This will be used for AioHandlers too. There is going to be little or no contention, so it is better to reuse the same lock. Signed-off-by: Paolo Bonzini --- async.c | 16 ++++++++-------- include/block/aio.h | 2 +- 2 files changed, 9 insertions(+), 9 deletions(-) diff --git a/async.c b/async.c index 6eee89d..b3d9006 100644 --- a/async.c +++ b/async.c @@ -50,12 +50,12 @@ QEMUBH *aio_bh_new(AioContext *ctx, QEMUBHFunc *cb, void *opaque) .cb = cb, .opaque = opaque, }; - qemu_mutex_lock(&ctx->bh_lock); + qemu_mutex_lock(&ctx->list_lock); bh->next = ctx->first_bh; /* Make sure that the members are ready before putting bh into list */ smp_wmb(); ctx->first_bh = bh; - qemu_mutex_unlock(&ctx->bh_lock); + qemu_mutex_unlock(&ctx->list_lock); return bh; } @@ -97,7 +97,7 @@ int aio_bh_poll(AioContext *ctx) /* remove deleted bhs */ if (!ctx->walking_bh) { - qemu_mutex_lock(&ctx->bh_lock); + qemu_mutex_lock(&ctx->list_lock); bhp = &ctx->first_bh; while (*bhp) { bh = *bhp; @@ -108,7 +108,7 @@ int aio_bh_poll(AioContext *ctx) bhp = &bh->next; } } - qemu_mutex_unlock(&ctx->bh_lock); + qemu_mutex_unlock(&ctx->list_lock); } return ret; @@ -239,7 +239,7 @@ aio_ctx_finalize(GSource *source) thread_pool_free(ctx->thread_pool); - qemu_mutex_lock(&ctx->bh_lock); + qemu_mutex_lock(&ctx->list_lock); while (ctx->first_bh) { QEMUBH *next = ctx->first_bh->next; @@ -249,12 +249,12 @@ aio_ctx_finalize(GSource *source) g_free(ctx->first_bh); ctx->first_bh = next; } - qemu_mutex_unlock(&ctx->bh_lock); + qemu_mutex_unlock(&ctx->list_lock); aio_set_event_notifier(ctx, &ctx->notifier, false, NULL); event_notifier_cleanup(&ctx->notifier); qemu_rec_mutex_destroy(&ctx->lock); - qemu_mutex_destroy(&ctx->bh_lock); + qemu_mutex_destroy(&ctx->list_lock); timerlistgroup_deinit(&ctx->tlg); } @@ -548,7 +548,7 @@ AioContext *aio_context_new(Error **errp) (EventNotifierHandler *) event_notifier_dummy_cb); ctx->thread_pool = NULL; - qemu_mutex_init(&ctx->bh_lock); + qemu_mutex_init(&ctx->list_lock); qemu_rec_mutex_init(&ctx->lock); timerlistgroup_init(&ctx->tlg, aio_timerlist_notify, ctx); diff --git a/include/block/aio.h b/include/block/aio.h index 3f055d2..322a10e 100644 --- a/include/block/aio.h +++ b/include/block/aio.h @@ -88,7 +88,7 @@ struct AioContext { uint32_t notify_me; /* lock to protect between bh's adders and deleter */ - QemuMutex bh_lock; + QemuMutex list_lock; /* Anchor of the list of Bottom Halves belonging to the context */ struct QEMUBH *first_bh; -- 2.5.0