block: explicitly acquire aiocontext in timers that need it
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> Reviewed-by: Fam Zheng <famz@redhat.com> Reviewed-by: Daniel P. Berrange <berrange@redhat.com> Message-id: 20170213135235.12274-13-pbonzini@redhat.com Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
This commit is contained in:
		
							parent
							
								
									0836c72f70
								
							
						
					
					
						commit
						2f47da5f7f
					
				| 
						 | 
				
			
			@ -424,9 +424,11 @@ static void curl_multi_timeout_do(void *arg)
 | 
			
		|||
        return;
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    aio_context_acquire(s->aio_context);
 | 
			
		||||
    curl_multi_socket_action(s->multi, CURL_SOCKET_TIMEOUT, 0, &running);
 | 
			
		||||
 | 
			
		||||
    curl_multi_check_completion(s);
 | 
			
		||||
    aio_context_release(s->aio_context);
 | 
			
		||||
#else
 | 
			
		||||
    abort();
 | 
			
		||||
#endif
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -2080,6 +2080,11 @@ void bdrv_aio_cancel(BlockAIOCB *acb)
 | 
			
		|||
        if (acb->aiocb_info->get_aio_context) {
 | 
			
		||||
            aio_poll(acb->aiocb_info->get_aio_context(acb), true);
 | 
			
		||||
        } else if (acb->bs) {
 | 
			
		||||
            /* qemu_aio_ref and qemu_aio_unref are not thread-safe, so
 | 
			
		||||
             * assert that we're not using an I/O thread.  Thread-safe
 | 
			
		||||
             * code should use bdrv_aio_cancel_async exclusively.
 | 
			
		||||
             */
 | 
			
		||||
            assert(bdrv_get_aio_context(acb->bs) == qemu_get_aio_context());
 | 
			
		||||
            aio_poll(bdrv_get_aio_context(acb->bs), true);
 | 
			
		||||
        } else {
 | 
			
		||||
            abort();
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -174,7 +174,7 @@ static void iscsi_retry_timer_expired(void *opaque)
 | 
			
		|||
    struct IscsiTask *iTask = opaque;
 | 
			
		||||
    iTask->complete = 1;
 | 
			
		||||
    if (iTask->co) {
 | 
			
		||||
        qemu_coroutine_enter(iTask->co);
 | 
			
		||||
        aio_co_wake(iTask->co);
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -1392,16 +1392,20 @@ static void iscsi_nop_timed_event(void *opaque)
 | 
			
		|||
{
 | 
			
		||||
    IscsiLun *iscsilun = opaque;
 | 
			
		||||
 | 
			
		||||
    aio_context_acquire(iscsilun->aio_context);
 | 
			
		||||
    if (iscsi_get_nops_in_flight(iscsilun->iscsi) >= MAX_NOP_FAILURES) {
 | 
			
		||||
        error_report("iSCSI: NOP timeout. Reconnecting...");
 | 
			
		||||
        iscsilun->request_timed_out = true;
 | 
			
		||||
    } else if (iscsi_nop_out_async(iscsilun->iscsi, NULL, NULL, 0, NULL) != 0) {
 | 
			
		||||
        error_report("iSCSI: failed to sent NOP-Out. Disabling NOP messages.");
 | 
			
		||||
        return;
 | 
			
		||||
        goto out;
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    timer_mod(iscsilun->nop_timer, qemu_clock_get_ms(QEMU_CLOCK_REALTIME) + NOP_INTERVAL);
 | 
			
		||||
    iscsi_set_events(iscsilun);
 | 
			
		||||
 | 
			
		||||
out:
 | 
			
		||||
    aio_context_release(iscsilun->aio_context);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static void iscsi_readcapacity_sync(IscsiLun *iscsilun, Error **errp)
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -141,7 +141,11 @@ static void null_bh_cb(void *opaque)
 | 
			
		|||
static void null_timer_cb(void *opaque)
 | 
			
		||||
{
 | 
			
		||||
    NullAIOCB *acb = opaque;
 | 
			
		||||
    AioContext *ctx = bdrv_get_aio_context(acb->common.bs);
 | 
			
		||||
 | 
			
		||||
    aio_context_acquire(ctx);
 | 
			
		||||
    acb->common.cb(acb->common.opaque, 0);
 | 
			
		||||
    aio_context_release(ctx);
 | 
			
		||||
    timer_deinit(&acb->timer);
 | 
			
		||||
    qemu_aio_unref(acb);
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
							
								
								
									
										12
									
								
								block/qed.c
								
								
								
								
							
							
						
						
									
										12
									
								
								block/qed.c
								
								
								
								
							| 
						 | 
				
			
			@ -345,10 +345,22 @@ static void qed_need_check_timer_cb(void *opaque)
 | 
			
		|||
 | 
			
		||||
    trace_qed_need_check_timer_cb(s);
 | 
			
		||||
 | 
			
		||||
    qed_acquire(s);
 | 
			
		||||
    qed_plug_allocating_write_reqs(s);
 | 
			
		||||
 | 
			
		||||
    /* Ensure writes are on disk before clearing flag */
 | 
			
		||||
    bdrv_aio_flush(s->bs->file->bs, qed_clear_need_check, s);
 | 
			
		||||
    qed_release(s);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
void qed_acquire(BDRVQEDState *s)
 | 
			
		||||
{
 | 
			
		||||
    aio_context_acquire(bdrv_get_aio_context(s->bs));
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
void qed_release(BDRVQEDState *s)
 | 
			
		||||
{
 | 
			
		||||
    aio_context_release(bdrv_get_aio_context(s->bs));
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static void qed_start_need_check_timer(BDRVQEDState *s)
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -198,6 +198,9 @@ enum {
 | 
			
		|||
 */
 | 
			
		||||
typedef void QEDFindClusterFunc(void *opaque, int ret, uint64_t offset, size_t len);
 | 
			
		||||
 | 
			
		||||
void qed_acquire(BDRVQEDState *s);
 | 
			
		||||
void qed_release(BDRVQEDState *s);
 | 
			
		||||
 | 
			
		||||
/**
 | 
			
		||||
 * Generic callback for chaining async callbacks
 | 
			
		||||
 */
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -416,7 +416,9 @@ static void timer_cb(BlockBackend *blk, bool is_write)
 | 
			
		|||
    qemu_mutex_unlock(&tg->lock);
 | 
			
		||||
 | 
			
		||||
    /* Run the request that was waiting for this timer */
 | 
			
		||||
    aio_context_acquire(blk_get_aio_context(blk));
 | 
			
		||||
    empty_queue = !qemu_co_enter_next(&blkp->throttled_reqs[is_write]);
 | 
			
		||||
    aio_context_release(blk_get_aio_context(blk));
 | 
			
		||||
 | 
			
		||||
    /* If the request queue was empty then we have to take care of
 | 
			
		||||
     * scheduling the next one */
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -454,9 +454,7 @@ bool aio_dispatch(AioContext *ctx, bool dispatch_fds)
 | 
			
		|||
    }
 | 
			
		||||
 | 
			
		||||
    /* Run our timers */
 | 
			
		||||
    aio_context_acquire(ctx);
 | 
			
		||||
    progress |= timerlistgroup_run_timers(&ctx->tlg);
 | 
			
		||||
    aio_context_release(ctx);
 | 
			
		||||
 | 
			
		||||
    return progress;
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -403,9 +403,7 @@ bool aio_poll(AioContext *ctx, bool blocking)
 | 
			
		|||
        progress |= aio_dispatch_handlers(ctx, event);
 | 
			
		||||
    } while (count > 0);
 | 
			
		||||
 | 
			
		||||
    aio_context_acquire(ctx);
 | 
			
		||||
    progress |= timerlistgroup_run_timers(&ctx->tlg);
 | 
			
		||||
    aio_context_release(ctx);
 | 
			
		||||
    return progress;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -25,7 +25,7 @@ static void co_sleep_cb(void *opaque)
 | 
			
		|||
{
 | 
			
		||||
    CoSleepCB *sleep_cb = opaque;
 | 
			
		||||
 | 
			
		||||
    qemu_coroutine_enter(sleep_cb->co);
 | 
			
		||||
    aio_co_wake(sleep_cb->co);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
void coroutine_fn co_aio_sleep_ns(AioContext *ctx, QEMUClockType type,
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in New Issue