@@ -1376,14 +1376,17 @@ BlockAIOCB *blk_abort_aio_request(BlockBackend *blk,
void *opaque, int ret)
{
struct BlockBackendAIOCB *acb;
+ AioContext *ctx;
blk_inc_in_flight(blk);
acb = blk_aio_get(&block_backend_aiocb_info, blk, cb, opaque);
acb->blk = blk;
acb->ret = ret;
-
- replay_bh_schedule_oneshot_event(blk_get_aio_context(blk),
- error_callback_bh, acb);
+ ctx = blk_get_aio_context(blk);
+ if (!replay_bh_schedule_oneshot_event(ctx, error_callback_bh, acb)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, error_callback_bh, acb);
+ }
return &acb->common;
}
@@ -1447,8 +1450,12 @@ static BlockAIOCB *blk_aio_prwv(BlockBackend *blk, int64_t offset, int bytes,
acb->has_returned = true;
if (acb->rwco.ret != NOT_DONE) {
- replay_bh_schedule_oneshot_event(blk_get_aio_context(blk),
- blk_aio_complete_bh, acb);
+ AioContext *ctx = blk_get_aio_context(blk);
+
+ if (!replay_bh_schedule_oneshot_event(ctx, blk_aio_complete_bh, acb)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, blk_aio_complete_bh, acb);
+ }
}
return &acb->common;
@@ -348,6 +348,7 @@ static void coroutine_fn bdrv_co_yield_to_drain(BlockDriverState *bs,
int *drained_end_counter)
{
BdrvCoDrainData data;
+ AioContext *ctx;
/* Calling bdrv_drain() from a BH ensures the current coroutine yields and
* other coroutines run if they were queued by aio_co_enter(). */
@@ -368,8 +369,11 @@ static void coroutine_fn bdrv_co_yield_to_drain(BlockDriverState *bs,
if (bs) {
bdrv_inc_in_flight(bs);
}
- replay_bh_schedule_oneshot_event(bdrv_get_aio_context(bs),
- bdrv_co_drain_bh_cb, &data);
+ ctx = bdrv_get_aio_context(bs);
+ if (!replay_bh_schedule_oneshot_event(ctx, bdrv_co_drain_bh_cb, &data)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, bdrv_co_drain_bh_cb, &data);
+ }
qemu_coroutine_yield();
/* If we are resumed from some other event (such as an aio completion or a
@@ -600,12 +604,12 @@ void bdrv_drain_all_begin(void)
return;
}
- /*
- * bdrv queue is managed by record/replay,
- * waiting for finishing the I/O requests may
- * be infinite
- */
if (replay_events_enabled()) {
+ /*
+ * bdrv queue is managed by record/replay,
+ * waiting for finishing the I/O requests may
+ * be infinite
+ */
return;
}
@@ -638,12 +642,12 @@ void bdrv_drain_all_end(void)
BlockDriverState *bs = NULL;
int drained_end_counter = 0;
- /*
- * bdrv queue is managed by record/replay,
- * waiting for finishing the I/O requests may
- * be endless
- */
if (replay_events_enabled()) {
+ /*
+ * bdrv queue is managed by record/replay,
+ * waiting for finishing the I/O requests may
+ * be endless
+ */
return;
}
@@ -2122,12 +2126,12 @@ int bdrv_flush_all(void)
BlockDriverState *bs = NULL;
int result = 0;
- /*
- * bdrv queue is managed by record/replay,
- * creating new flush request for stopping
- * the VM may break the determinism
- */
if (replay_events_enabled()) {
+ /*
+ * bdrv queue is managed by record/replay,
+ * creating new flush request for stopping
+ * the VM may break the determinism
+ */
return result;
}
@@ -283,8 +283,13 @@ iscsi_co_generic_cb(struct iscsi_context *iscsi, int status,
}
if (iTask->co) {
- replay_bh_schedule_oneshot_event(iTask->iscsilun->aio_context,
- iscsi_co_generic_bh_cb, iTask);
+ AioContext *ctx = iTask->iscsilun->aio_context;
+
+ if (!replay_bh_schedule_oneshot_event(ctx,
+ iscsi_co_generic_bh_cb, iTask)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, iscsi_co_generic_bh_cb, iTask);
+ }
} else {
iTask->complete = 1;
}
@@ -7,7 +7,6 @@ block_ss.add(files(
'backup-top.c',
'blkdebug.c',
'blklogwrites.c',
- 'blkreplay.c',
'blkverify.c',
'block-backend.c',
'block-copy.c',
@@ -42,6 +41,8 @@ block_ss.add(files(
'write-threshold.c',
), zstd, zlib)
+block_ss.add(when: 'CONFIG_TCG', if_true: files('blkreplay.c'))
+
block_ss.add(when: 'CONFIG_QCOW1', if_true: files('qcow.c'))
block_ss.add(when: 'CONFIG_VDI', if_true: files('vdi.c'))
block_ss.add(when: 'CONFIG_CLOOP', if_true: files('cloop.c'))
@@ -246,7 +246,9 @@ static void
nfs_co_generic_cb(int ret, struct nfs_context *nfs, void *data,
void *private_data)
{
+ AioContext *ctx;
NFSRPC *task = private_data;
+
task->ret = ret;
assert(!task->st);
if (task->ret > 0 && task->iov) {
@@ -259,8 +261,12 @@ nfs_co_generic_cb(int ret, struct nfs_context *nfs, void *data,
if (task->ret < 0) {
error_report("NFS Error: %s", nfs_get_error(nfs));
}
- replay_bh_schedule_oneshot_event(task->client->aio_context,
- nfs_co_generic_bh_cb, task);
+
+ ctx = task->client->aio_context;
+ if (!replay_bh_schedule_oneshot_event(ctx, nfs_co_generic_bh_cb, task)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, nfs_co_generic_bh_cb, task);
+ }
}
static int coroutine_fn nfs_co_preadv(BlockDriverState *bs, uint64_t offset,
@@ -180,8 +180,12 @@ static inline BlockAIOCB *null_aio_common(BlockDriverState *bs,
timer_mod_ns(&acb->timer,
qemu_clock_get_ns(QEMU_CLOCK_REALTIME) + s->latency_ns);
} else {
- replay_bh_schedule_oneshot_event(bdrv_get_aio_context(bs),
- null_bh_cb, acb);
+ AioContext *ctx = bdrv_get_aio_context(bs);
+
+ if (!replay_bh_schedule_oneshot_event(ctx, null_bh_cb, acb)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, null_bh_cb, acb);
+ }
}
return &acb->common;
}
@@ -313,8 +313,12 @@ static void nvme_put_free_req_locked(NVMeQueuePair *q, NVMeRequest *req)
static void nvme_wake_free_req_locked(NVMeQueuePair *q)
{
if (!qemu_co_queue_empty(&q->free_req_queue)) {
- replay_bh_schedule_oneshot_event(q->s->aio_context,
- nvme_free_req_queue_cb, q);
+ AioContext *ctx = q->s->aio_context;
+
+ if (!replay_bh_schedule_oneshot_event(ctx, nvme_free_req_queue_cb, q)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, nvme_free_req_queue_cb, q);
+ }
}
}
@@ -1068,7 +1072,10 @@ static void nvme_rw_cb(void *opaque, int ret)
/* The rw coroutine hasn't yielded, don't try to enter. */
return;
}
- replay_bh_schedule_oneshot_event(data->ctx, nvme_rw_cb_bh, data);
+ if (!replay_bh_schedule_oneshot_event(data->ctx, nvme_rw_cb_bh, data)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(data->ctx, nvme_rw_cb_bh, data);
+ }
}
static coroutine_fn int nvme_co_prw_aligned(BlockDriverState *bs,
@@ -857,13 +857,17 @@ static void rbd_finish_bh(void *opaque)
*/
static void rbd_finish_aiocb(rbd_completion_t c, RADOSCB *rcb)
{
+ AioContext *ctx;
RBDAIOCB *acb = rcb->acb;
rcb->ret = rbd_aio_get_return_value(c);
rbd_aio_release(c);
- replay_bh_schedule_oneshot_event(bdrv_get_aio_context(acb->common.bs),
- rbd_finish_bh, rcb);
+ ctx = bdrv_get_aio_context(acb->common.bs);
+ if (!replay_bh_schedule_oneshot_event(ctx, rbd_finish_bh, rcb)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, rbd_finish_bh, rcb);
+ }
}
static int rbd_aio_discard_wrapper(rbd_image_t image,
@@ -493,7 +493,10 @@ static void ide_issue_trim_cb(void *opaque, int ret)
done:
iocb->aiocb = NULL;
if (iocb->bh) {
- replay_bh_schedule_event(iocb->bh);
+ if (!replay_bh_schedule_event(iocb->bh)) {
+ /* regular case without replay */
+ qemu_bh_schedule(iocb->bh);
+ }
}
}
@@ -2277,6 +2280,7 @@ void ide_ctrl_write(void *opaque, uint32_t addr, uint32_t val)
IDEBus *bus = opaque;
IDEState *s;
int i;
+ AioContext *ctx;
trace_ide_ctrl_write(addr, val, bus);
@@ -2289,8 +2293,12 @@ void ide_ctrl_write(void *opaque, uint32_t addr, uint32_t val)
s = &bus->ifs[i];
s->status |= BUSY_STAT;
}
- replay_bh_schedule_oneshot_event(qemu_get_aio_context(),
- ide_bus_perform_srst, bus);
+
+ ctx = qemu_get_aio_context();
+ if (!replay_bh_schedule_oneshot_event(ctx, ide_bus_perform_srst, bus)) {
+ /* regular case without replay */
+ aio_bh_schedule_oneshot(ctx, ide_bus_perform_srst, bus);
+ }
}
bus->cmd = val;
@@ -173,14 +173,14 @@ bool replay_events_enabled(void);
/* Flushes events queue */
void replay_flush_events(void);
/*! Adds bottom half event to the queue */
-void replay_bh_schedule_event(QEMUBH *bh);
+bool replay_bh_schedule_event(QEMUBH *bh);
/* Adds oneshot bottom half event to the queue */
-void replay_bh_schedule_oneshot_event(AioContext *ctx,
+bool replay_bh_schedule_oneshot_event(AioContext *ctx,
QEMUBHFunc *cb, void *opaque);
/*! Adds input event to the queue */
-void replay_input_event(QemuConsole *src, InputEvent *evt);
+bool replay_input_event(QemuConsole *src, InputEvent *evt);
/*! Adds input sync event to the queue */
-void replay_input_sync_event(void);
+bool replay_input_sync_event(void);
/*! Adds block layer event to the queue */
void replay_block_event(QEMUBH *bh, uint64_t id);
/*! Returns ID for the next block event */
@@ -63,6 +63,7 @@
#include "migration/colo.h"
#include "qemu/bitmap.h"
#include "net/announce.h"
+#include "sysemu/tcg.h"
const unsigned int postcopy_ram_discard_version = 0;
@@ -2674,10 +2675,12 @@ int save_snapshot(const char *name, Error **errp)
return ret;
}
- if (!replay_can_snapshot()) {
- error_setg(errp, "Record/replay does not allow making snapshot "
- "right now. Try once more later.");
- return ret;
+ if (tcg_enabled()) {
+ if (!replay_can_snapshot()) {
+ error_setg(errp, "Record/replay does not allow making snapshot "
+ "right now. Try once more later.");
+ return ret;
+ }
}
if (!bdrv_all_can_snapshot(&bs)) {
@@ -7,7 +7,6 @@ softmmu_ss.add(files(
'eth.c',
'filter-buffer.c',
'filter-mirror.c',
- 'filter-replay.c',
'filter-rewriter.c',
'filter.c',
'hub.c',
@@ -17,6 +16,8 @@ softmmu_ss.add(files(
'util.c',
))
+softmmu_ss.add(when: 'CONFIG_TCG', if_true: files('filter-replay.c'))
+
softmmu_ss.add(when: 'CONFIG_L2TPV3', if_true: files('l2tpv3.c'))
softmmu_ss.add(when: slirp, if_true: files('slirp.c'))
softmmu_ss.add(when: ['CONFIG_VDE', vde], if_true: files('vde.c'))
@@ -1,4 +1,4 @@
-softmmu_ss.add(files(
+softmmu_ss.add(when: 'CONFIG_TCG', if_true: files(
'replay.c',
'replay-internal.c',
'replay-events.c',
@@ -128,24 +128,34 @@ void replay_add_event(ReplayAsyncEventKind event_kind,
QTAILQ_INSERT_TAIL(&events_list, event, events);
}
-void replay_bh_schedule_event(QEMUBH *bh)
+/*
+ * return true if replay has eaten the event,
+ * false if normal event handling should be done.
+ */
+bool replay_bh_schedule_event(QEMUBH *bh)
{
if (events_enabled) {
uint64_t id = replay_get_current_icount();
replay_add_event(REPLAY_ASYNC_EVENT_BH, bh, NULL, id);
+ return true;
} else {
- qemu_bh_schedule(bh);
+ return false;
}
}
-void replay_bh_schedule_oneshot_event(AioContext *ctx,
- QEMUBHFunc *cb, void *opaque)
+/*
+ * return true if replay has eaten the event,
+ * false if normal event handling should be done.
+ */
+bool replay_bh_schedule_oneshot_event(AioContext *ctx,
+ QEMUBHFunc *cb, void *opaque)
{
if (events_enabled) {
uint64_t id = replay_get_current_icount();
replay_add_event(REPLAY_ASYNC_EVENT_BH_ONESHOT, cb, opaque, id);
+ return true;
} else {
- aio_bh_schedule_oneshot(ctx, cb, opaque);
+ return false;
}
}
@@ -117,24 +117,36 @@ InputEvent *replay_read_input_event(void)
return QAPI_CLONE(InputEvent, &evt);
}
-void replay_input_event(QemuConsole *src, InputEvent *evt)
+/*
+ * return true if replay has eaten the event,
+ * false if normal event handling should be done.
+ */
+bool replay_input_event(QemuConsole *src, InputEvent *evt)
{
if (replay_mode == REPLAY_MODE_PLAY) {
/* Nothing */
+ return true;
} else if (replay_mode == REPLAY_MODE_RECORD) {
replay_add_input_event(QAPI_CLONE(InputEvent, evt));
+ return true;
} else {
- qemu_input_event_send_impl(src, evt);
+ return false;
}
}
-void replay_input_sync_event(void)
+/*
+ * return true if replay has eaten the event,
+ * false if normal event handling should be done.
+ */
+bool replay_input_sync_event(void)
{
if (replay_mode == REPLAY_MODE_PLAY) {
/* Nothing */
+ return true;
} else if (replay_mode == REPLAY_MODE_RECORD) {
replay_add_input_sync_event();
+ return true;
} else {
- qemu_input_event_sync_impl();
+ return false;
}
}
@@ -32,7 +32,6 @@ stub_ss.add(files('qtest.c'))
stub_ss.add(files('ram-block.c'))
stub_ss.add(files('ramfb.c'))
stub_ss.add(files('replay.c'))
-stub_ss.add(files('replay-user.c'))
stub_ss.add(files('runstate-check.c'))
stub_ss.add(files('set-fd-handler.c'))
stub_ss.add(files('sysbus.c'))
deleted file mode 100644
@@ -1,9 +0,0 @@
-#include "qemu/osdep.h"
-#include "sysemu/replay.h"
-#include "sysemu/sysemu.h"
-
-void replay_bh_schedule_oneshot_event(AioContext *ctx,
- QEMUBHFunc *cb, void *opaque)
-{
- aio_bh_schedule_oneshot(ctx, cb, opaque);
-}
@@ -103,3 +103,109 @@ bool replay_reverse_continue(void)
{
return false;
}
+
+/*
+ * the following event-related stubs need to return false,
+ * so that normal events processing can happen when the replay framework
+ * is not available (!CONFIG_TCG)
+ */
+bool replay_input_event(QemuConsole *src, InputEvent *evt)
+{
+ return false;
+}
+bool replay_input_sync_event(void)
+{
+ return false;
+}
+bool replay_bh_schedule_event(QEMUBH *bh)
+{
+ return false;
+}
+bool replay_bh_schedule_oneshot_event(AioContext *ctx,
+ QEMUBHFunc *cb, void *opaque)
+{
+ return false;
+}
+
+void replay_add_blocker(Error *reason)
+{
+}
+void replay_audio_in(size_t *recorded, void *samples, size_t *wpos, size_t size)
+{
+}
+void replay_audio_out(size_t *played)
+{
+}
+void replay_breakpoint(void)
+{
+}
+bool replay_can_snapshot(void)
+{
+ return false;
+}
+void replay_configure(struct QemuOpts *opts)
+{
+}
+void replay_flush_events(void)
+{
+}
+void replay_gdb_attached(void)
+{
+}
+bool replay_running_debug(void)
+{
+ return false;
+}
+void replay_shutdown_request(ShutdownCause cause)
+{
+}
+void replay_start(void)
+{
+}
+void replay_vmstate_init(void)
+{
+}
+
+#include "monitor/monitor.h"
+#include "monitor/hmp.h"
+#include "qapi/qapi-commands-replay.h"
+#include "qapi/error.h"
+#include "qemu/error-report.h"
+
+void hmp_info_replay(Monitor *mon, const QDict *qdict)
+{
+ error_report("replay support not available");
+}
+void hmp_replay_break(Monitor *mon, const QDict *qdict)
+{
+ error_report("replay support not available");
+}
+void hmp_replay_delete_break(Monitor *mon, const QDict *qdict)
+{
+ error_report("replay support not available");
+}
+void hmp_replay_seek(Monitor *mon, const QDict *qdict)
+{
+ error_report("replay support not available");
+}
+ReplayInfo *qmp_query_replay(Error **errp)
+{
+ error_set(errp, ERROR_CLASS_COMMAND_NOT_FOUND,
+ "replay support not available");
+ return NULL;
+}
+void qmp_replay_break(int64_t icount, Error **errp)
+{
+ error_set(errp, ERROR_CLASS_COMMAND_NOT_FOUND,
+ "replay support not available");
+}
+void qmp_replay_delete_break(Error **errp)
+{
+ error_set(errp, ERROR_CLASS_COMMAND_NOT_FOUND,
+ "replay support not available");
+}
+void qmp_replay_seek(int64_t icount, Error **errp)
+{
+ error_set(errp, ERROR_CLASS_COMMAND_NOT_FOUND,
+ "replay support not available");
+}
@@ -122,8 +122,3 @@ void qemu_bh_delete(QEMUBH *bh)
{
g_free(bh);
}
-
-void replay_bh_schedule_event(QEMUBH *bh)
-{
- bh->cb(bh->opaque);
-}
@@ -31,6 +31,9 @@ static int query_error_class(const char *cmd)
#ifndef CONFIG_SPICE
{ "query-spice", ERROR_CLASS_COMMAND_NOT_FOUND },
#endif
+#ifndef CONFIG_TCG
+ { "query-replay", ERROR_CLASS_COMMAND_NOT_FOUND },
+#endif
#ifndef CONFIG_VNC
{ "query-vnc", ERROR_CLASS_GENERIC_ERROR },
{ "query-vnc-servers", ERROR_CLASS_GENERIC_ERROR },
@@ -375,7 +375,10 @@ void qemu_input_event_send(QemuConsole *src, InputEvent *evt)
return;
}
- replay_input_event(src, evt);
+ if (!replay_input_event(src, evt)) {
+ /* regular case without replay */
+ qemu_input_event_send_impl(src, evt);
+ }
}
void qemu_input_event_sync_impl(void)
@@ -401,7 +404,10 @@ void qemu_input_event_sync(void)
return;
}
- replay_input_sync_event();
+ if (!replay_input_sync_event()) {
+ /* regular case without replay */
+ qemu_input_event_sync_impl();
+ }
}
static InputEvent *qemu_input_event_new_key(KeyValue *key, bool down)
this fixes non-TCG builds broken recently by replay reverse debugging. stub the needed functions in stub/, including errors for hmp and qmp. change hooks for the case when replay code is disabled (!CONFIG_TCG), as we need to avoid sinking all the events in replay in this case. Surprisingly, only _one_ qtest was affected by this, ide-test.c, which resulted in a buzz as the bh events were never delivered, and the bh never executed. Many other subsystems _should_ have been affected. This fixes the immediate issue, however a better way to group replay functionality to TCG-only code could be developed in the long term. Signed-off-by: Claudio Fontana <cfontana@suse.de> --- block/block-backend.c | 17 ++++-- block/io.c | 38 +++++++------ block/iscsi.c | 9 +++- block/meson.build | 3 +- block/nfs.c | 10 +++- block/null.c | 8 ++- block/nvme.c | 13 +++-- block/rbd.c | 8 ++- hw/ide/core.c | 14 +++-- include/sysemu/replay.h | 8 +-- migration/savevm.c | 11 ++-- net/meson.build | 3 +- replay/meson.build | 2 +- replay/replay-events.c | 20 +++++-- replay/replay-input.c | 20 +++++-- stubs/meson.build | 1 - stubs/replay-user.c | 9 ---- stubs/replay.c | 106 +++++++++++++++++++++++++++++++++++++ tests/ptimer-test-stubs.c | 5 -- tests/qtest/qmp-cmd-test.c | 3 ++ ui/input.c | 10 +++- 21 files changed, 245 insertions(+), 73 deletions(-) delete mode 100644 stubs/replay-user.c