From: "Michael R. Hines" <mrhines@linux.vnet.ibm.com>
To: Chegu Vinod <chegu_vinod@hp.com>
Cc: aliguori@us.ibm.com, quintela@redhat.com, qemu-devel@nongnu.org,
owasserm@redhat.com, abali@us.ibm.com, mrhines@us.ibm.com,
gokul@us.ibm.com, pbonzini@redhat.com, knoel@redhat.com
Subject: Re: [Qemu-devel] [PATCH v9 14/14] rdma: add pin-all accounting timestamp to QMP statistics
Date: Tue, 18 Jun 2013 16:17:43 -0400 [thread overview]
Message-ID: <51C0C067.9080603@linux.vnet.ibm.com> (raw)
In-Reply-To: <51BE05CC.8040706@hp.com>
On 06/16/2013 02:37 PM, Chegu Vinod wrote:
> On 6/14/2013 1:35 PM, mrhines@linux.vnet.ibm.com wrote:
>> From: "Michael R. Hines" <mrhines@us.ibm.com>
>>
>> For very large virtual machines, pinning can take a long time.
>> While this does not affect the migration's *actual* time itself,
>> it is still important for the user to know what's going on and
>> to know what component of the total time is actual taken up by
>> pinning.
>>
>> For example, using a 14GB virtual machine, pinning can take as
>> long as 5 seconds, for which the user would not otherwise know
>> what was happening.
>>
>> Reviewed-by: Paolo Bonzini <pbonzini@redhat.com>
>> Signed-off-by: Michael R. Hines <mrhines@us.ibm.com>
>> ---
>> hmp.c | 4 +++
>> include/migration/migration.h | 1 +
>> migration-rdma.c | 55
>> +++++++++++++++++++++++++++--------------
>> migration.c | 13 +++++++++-
>> qapi-schema.json | 3 ++-
>> 5 files changed, 56 insertions(+), 20 deletions(-)
>>
>> diff --git a/hmp.c b/hmp.c
>> index 148a3fb..90c55f2 100644
>> --- a/hmp.c
>> +++ b/hmp.c
>> @@ -164,6 +164,10 @@ void hmp_info_migrate(Monitor *mon, const QDict
>> *qdict)
>> monitor_printf(mon, "downtime: %" PRIu64 "
>> milliseconds\n",
>> info->downtime);
>> }
>> + if (info->has_pin_all_time) {
>> + monitor_printf(mon, "pin-all: %" PRIu64 " milliseconds\n",
>> + info->pin_all_time);
>> + }
>> }
>> if (info->has_ram) {
>> diff --git a/include/migration/migration.h
>> b/include/migration/migration.h
>> index b49e68b..d2ca75b 100644
>> --- a/include/migration/migration.h
>> +++ b/include/migration/migration.h
>> @@ -49,6 +49,7 @@ struct MigrationState
>> bool enabled_capabilities[MIGRATION_CAPABILITY_MAX];
>> int64_t xbzrle_cache_size;
>> double mbps;
>> + int64_t pin_all_time;
>> };
>> void process_incoming_migration(QEMUFile *f);
>> diff --git a/migration-rdma.c b/migration-rdma.c
>> index 853de18..e407dce 100644
>> --- a/migration-rdma.c
>> +++ b/migration-rdma.c
>> @@ -699,11 +699,11 @@ static int qemu_rdma_alloc_qp(RDMAContext *rdma)
>> return 0;
>> }
>> -static int qemu_rdma_reg_whole_ram_blocks(RDMAContext *rdma,
>> - RDMALocalBlocks *rdma_local_ram_blocks)
>> +static int qemu_rdma_reg_whole_ram_blocks(RDMAContext *rdma)
>> {
>> int i;
>> - uint64_t start = qemu_get_clock_ms(rt_clock);
>> + int64_t start = qemu_get_clock_ms(host_clock);
>> + RDMALocalBlocks *rdma_local_ram_blocks = &rdma->local_ram_blocks;
>> (void)start;
>> for (i = 0; i < rdma_local_ram_blocks->num_blocks; i++) {
>> @@ -721,7 +721,8 @@ static int
>> qemu_rdma_reg_whole_ram_blocks(RDMAContext *rdma,
>> rdma->total_registrations++;
>> }
>> - DPRINTF("lock time: %" PRIu64 "\n",
>> qemu_get_clock_ms(rt_clock) - start);
>> + DPRINTF("local lock time: %" PRId64 "\n",
>> + qemu_get_clock_ms(host_clock) - start);
>> if (i >= rdma_local_ram_blocks->num_blocks) {
>> return 0;
>> @@ -1262,7 +1263,8 @@ static void qemu_rdma_move_header(RDMAContext
>> *rdma, int idx,
>> */
>> static int qemu_rdma_exchange_send(RDMAContext *rdma,
>> RDMAControlHeader *head,
>> uint8_t *data, RDMAControlHeader
>> *resp,
>> - int *resp_idx)
>> + int *resp_idx,
>> + int (*callback)(RDMAContext *rdma))
>> {
>> int ret = 0;
>> int idx = 0;
>> @@ -1315,6 +1317,14 @@ static int qemu_rdma_exchange_send(RDMAContext
>> *rdma, RDMAControlHeader *head,
>> * If we're expecting a response, block and wait for it.
>> */
>> if (resp) {
>> + if (callback) {
>> + DPRINTF("Issuing callback before receiving response...\n");
>> + ret = callback(rdma);
>> + if (ret < 0) {
>> + return ret;
>> + }
>> + }
>> +
>> DDPRINTF("Waiting for response %s\n",
>> control_desc[resp->type]);
>> ret = qemu_rdma_exchange_get_response(rdma, resp,
>> resp->type, idx + 1);
>> @@ -1464,7 +1474,7 @@ static int qemu_rdma_write_one(QEMUFile *f,
>> RDMAContext *rdma,
>> chunk, sge.length, current_index, offset);
>> ret = qemu_rdma_exchange_send(rdma, &head,
>> - (uint8_t *) &comp, NULL, NULL);
>> + (uint8_t *) &comp, NULL, NULL, NULL);
>> if (ret < 0) {
>> return -EIO;
>> @@ -1487,7 +1497,7 @@ static int qemu_rdma_write_one(QEMUFile *f,
>> RDMAContext *rdma,
>> chunk, sge.length, current_index, offset);
>> ret = qemu_rdma_exchange_send(rdma, &head, (uint8_t
>> *) ®,
>> - &resp, ®_result_idx);
>> + &resp, ®_result_idx, NULL);
>> if (ret < 0) {
>> return ret;
>> }
>> @@ -2126,7 +2136,7 @@ static int qemu_rdma_put_buffer(void *opaque,
>> const uint8_t *buf,
>> head.len = r->len;
>> head.type = RDMA_CONTROL_QEMU_FILE;
>> - ret = qemu_rdma_exchange_send(rdma, &head, data, NULL, NULL);
>> + ret = qemu_rdma_exchange_send(rdma, &head, data, NULL, NULL,
>> NULL);
>> if (ret < 0) {
>> rdma->error_state = ret;
>> @@ -2482,7 +2492,7 @@ static int
>> qemu_rdma_registration_handle(QEMUFile *f, void *opaque,
>> DPRINTF("Initial setup info requested.\n");
>> if (rdma->pin_all) {
>> - ret = qemu_rdma_reg_whole_ram_blocks(rdma,
>> &rdma->local_ram_blocks);
>> + ret = qemu_rdma_reg_whole_ram_blocks(rdma);
>> if (ret) {
>> fprintf(stderr, "rdma migration: error dest "
>> "registering ram blocks!\n");
>> @@ -2614,10 +2624,22 @@ static int
>> qemu_rdma_registration_stop(QEMUFile *f, void *opaque,
>> }
>> if (flags == RAM_CONTROL_SETUP) {
>> + int64_t start = qemu_get_clock_ms(host_clock);
>> + MigrationState *s = migrate_get_current();
>> +
>> head.type = RDMA_CONTROL_RAM_BLOCKS_REQUEST;
>> DPRINTF("Sending registration setup for ram blocks...\n");
>> - ret = qemu_rdma_exchange_send(rdma, &head, NULL, &resp,
>> ®_result_idx);
>> + /*
>> + * Make sure that we parallelize the pinning on both size.
>> + * For very large guests, doing this serially takes a really
>> + * long time, so we have to 'interleave' the pinning locally
>> + * by performing it before we receive the response from the
>> + * destination that the pinning has completed.
>> + */
>> + ret = qemu_rdma_exchange_send(rdma, &head, NULL, &resp,
>> + ®_result_idx, rdma->pin_all ?
>> + qemu_rdma_reg_whole_ram_blocks : NULL);
>> if (ret < 0) {
>> ERROR(errp, "receiving remote info!\n");
>> return ret;
>> @@ -2634,12 +2656,7 @@ static int
>> qemu_rdma_registration_stop(QEMUFile *f, void *opaque,
>> }
>> if (rdma->pin_all) {
>> - ret = qemu_rdma_reg_whole_ram_blocks(rdma,
>> &rdma->local_ram_blocks);
>> - if (ret) {
>> - fprintf(stderr, "rdma migration: error source "
>> - "registering ram blocks!\n");
>> - return ret;
>> - }
>> + s->pin_all_time = qemu_get_clock_ms(host_clock) - start;
>> } else {
>> int x = 0;
>> for (x = 0; x < rdma->local_ram_blocks.num_blocks; x++) {
>> @@ -2653,7 +2670,7 @@ static int qemu_rdma_registration_stop(QEMUFile
>> *f, void *opaque,
>> DDDPRINTF("Sending registration finish %" PRIu64 "...\n", flags);
>> head.type = RDMA_CONTROL_REGISTER_FINISHED;
>> - ret = qemu_rdma_exchange_send(rdma, &head, NULL, NULL, NULL);
>> + ret = qemu_rdma_exchange_send(rdma, &head, NULL, NULL, NULL, NULL);
>> if (ret < 0) {
>> goto err;
>> @@ -2801,8 +2818,10 @@ void rdma_start_outgoing_migration(void *opaque,
>> DPRINTF("qemu_rdma_source_connect success\n");
>> + if (rdma->pin_all) {
>> + s->pin_all_time = 0;
>> + }
>> s->file = qemu_fopen_rdma(rdma, "wb");
>> - s->total_time = qemu_get_clock_ms(rt_clock);
>> migrate_fd_connect(s);
>> return;
>> err:
>> diff --git a/migration.c b/migration.c
>> index 62c6b85..17531e9 100644
>> --- a/migration.c
>> +++ b/migration.c
>> @@ -67,6 +67,7 @@ MigrationState *migrate_get_current(void)
>> .bandwidth_limit = MAX_THROTTLE,
>> .xbzrle_cache_size = DEFAULT_MIGRATE_CACHE_SIZE,
>> .mbps = -1,
>> + .pin_all_time = -1,
>> };
>> return ¤t_migration;
>> @@ -189,7 +190,7 @@ MigrationInfo *qmp_query_migrate(Error **errp)
>> break;
>> case MIG_STATE_ACTIVE:
>> info->has_status = true;
>> - info->status = g_strdup("active");
>> + info->status = g_strdup(s->pin_all_time != 0 ? "active" :
>> "pinning");
>
> Pl. check the logic for setting the migration status correctly. In the
> current logic where you rely on "s->pin_all_time" if we did non-rdma
> migration the migration status will remain in "pinning" (instead of
> "active") till the migration completes....
>
> Thx
> Vinod
I just sent out a V10 which I have also validated with a separate patch
to libvirt. (I'm maintaining a sister patch for RDMA to libvirt on
github as well, in case you need to use libvirt too).
- Michael
next prev parent reply other threads:[~2013-06-18 20:18 UTC|newest]
Thread overview: 27+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-06-14 20:35 [Qemu-devel] [PATCH v9 00/14] rdma: migration support mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 01/14] rdma: add documentation mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 02/14] rdma: introduce qemu_update_position() mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 03/14] rdma: export yield_until_fd_readable() mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 04/14] rdma: export throughput w/ MigrationStats QMP mrhines
2013-06-15 9:09 ` Eric Blake
2013-06-16 4:04 ` Michael R. Hines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 05/14] rdma: introduce qemu_file_mode_is_not_valid() mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 06/14] rdma: export qemu_fflush() mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 07/14] rdma: introduce ram_handle_compressed() mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 08/14] rdma: introduce qemu_ram_foreach_block() mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 09/14] rdma: new QEMUFileOps hooks mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 10/14] rdma: introduce capability x-rdma-pin-all mrhines
2013-06-15 9:05 ` Eric Blake
2013-06-16 4:04 ` Michael R. Hines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 11/14] rdma: core logic mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 12/14] rdma: send pc.ram mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 13/14] rdma: fix mlock() freezes and accounting mrhines
2013-06-14 20:35 ` [Qemu-devel] [PATCH v9 14/14] rdma: add pin-all accounting timestamp to QMP statistics mrhines
2013-06-15 9:06 ` Eric Blake
2013-06-15 17:32 ` Paolo Bonzini
2013-06-16 4:06 ` Michael R. Hines
2013-06-16 4:05 ` Michael R. Hines
2013-06-16 18:37 ` Chegu Vinod
2013-06-18 20:17 ` Michael R. Hines [this message]
2013-06-15 22:51 ` [Qemu-devel] [PATCH v9 00/14] rdma: migration support Chegu Vinod
2013-06-16 4:07 ` Michael R. Hines
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=51C0C067.9080603@linux.vnet.ibm.com \
--to=mrhines@linux.vnet.ibm.com \
--cc=abali@us.ibm.com \
--cc=aliguori@us.ibm.com \
--cc=chegu_vinod@hp.com \
--cc=gokul@us.ibm.com \
--cc=knoel@redhat.com \
--cc=mrhines@us.ibm.com \
--cc=owasserm@redhat.com \
--cc=pbonzini@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=quintela@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).