From: Peter Xu <peterx@redhat.com>
To: Fabiano Rosas <farosas@suse.de>
Cc: qemu-devel@nongnu.org, Thomas Huth <thuth@redhat.com>
Subject: Re: [PATCH 6/7] migration: Fix arrays of pointers in JSON writer
Date: Wed, 8 Jan 2025 11:14:01 -0500 [thread overview]
Message-ID: <Z36kScJti9LrWVU7@x1n> (raw)
In-Reply-To: <87cygx4dlt.fsf@suse.de>
On Wed, Jan 08, 2025 at 10:52:30AM -0300, Fabiano Rosas wrote:
> Peter Xu <peterx@redhat.com> writes:
>
> > On Tue, Jan 07, 2025 at 04:50:24PM -0300, Fabiano Rosas wrote:
> >> Currently, if an array of pointers contains a NULL pointer, that
> >> pointer will be encoded as '0' in the stream. Since the JSON writer
> >> doesn't define a "pointer" type, that '0' will now be an uint64, which
> >> is different from the original type being pointed to, e.g. struct.
> >>
> >> That mixed-type array shouldn't be compressed, otherwise data is lost
> >> as the code currently makes the whole array have the type of the first
> >> element.
> >>
> >> While we could disable the array compression when a NULL pointer is
> >> found, the JSON part of the stream still makes part of downtime, so we
> >> should avoid writing unecessary bytes to it.
> >>
> >> Keep the array compression in place, but break the array into several
> >> type-contiguous pieces if NULL and non-NULL pointers are mixed.
> >
> > Could I request for a sample JSON dump for an example array in the commit
> > log? This whole solution looks working but is tricky. A sample could help
> > people understand (e.g. showing the same "name" being dumped multiple
> > times..).
>
> {"name": "s390_css", "instance_id": 0, "vmsd_name": "s390_css",
> "version": 1, "fields": [
> ...,
> {"name": "css", "array_len": 254, "type": "uint8", "size": 1},
> {"name": "css", "type": "struct", "struct": {
> "vmsd_name": "s390_css_img", "version": 1, "fields": [{"name":
> "chpids", "array_len": 256, "type": "struct", "struct": {"vmsd_name":
> "s390_chp_info", "version": 1, "fields": [{"name": "in_use", "type":
> "uint8", "size": 1}, {"name": "type", "type": "uint8", "size": 1},
> {"name": "is_virtual", "type": "uint8", "size": 1}]}, "size": 3}]},
> "size": 768},
> {"name": "css", "type": "uint8", "size": 1},
> ...
> ]}
Yes something like this would work, thanks. We could even omit most of the
struct details but only show the important ones:
{"name": "s390_css", "instance_id": 0, "vmsd_name": "s390_css",
"version": 1, "fields": [
...,
{"name": "css", "array_len": 254, "type": "uint8", "size": 1},
{"name": "css", "type": "struct", "struct": {"vmsd_name": "s390_css_img", ... }, "size": 768},
{"name": "css", "type": "uint8", "size": 1},
...
]}
>
> >
> > Side note: I tried to dump a very basic VM's JSON out to disk, it scares me
> > on the size:
> >
> > $ ls -lhS JSON.out
> > -rw-r--r--. 1 peterx peterx 106K Jan 7 17:18 JSON.out
> >
> > That's a simplest VM with all default stuff, mostly nothing complex.. I may
> > really need to measure how the JSON debug strings affect migration function
> > or perf at some point..
> >
>
> Agreed.
>
> >>
> >> Signed-off-by: Fabiano Rosas <farosas@suse.de>
> >> ---
> >> migration/vmstate.c | 33 ++++++++++++++++++++++++++++++++-
> >> scripts/analyze-migration.py | 9 ++++++++-
> >> 2 files changed, 40 insertions(+), 2 deletions(-)
> >>
> >> diff --git a/migration/vmstate.c b/migration/vmstate.c
> >> index 52704c822c..a79ccf3875 100644
> >> --- a/migration/vmstate.c
> >> +++ b/migration/vmstate.c
> >> @@ -425,15 +425,19 @@ int vmstate_save_state_v(QEMUFile *f, const VMStateDescription *vmsd,
> >> int size = vmstate_size(opaque, field);
> >> uint64_t old_offset, written_bytes;
> >> JSONWriter *vmdesc_loop = vmdesc;
> >> + bool is_prev_null = false;
> >>
> >> trace_vmstate_save_state_loop(vmsd->name, field->name, n_elems);
> >> if (field->flags & VMS_POINTER) {
> >> first_elem = *(void **)first_elem;
> >> assert(first_elem || !n_elems || !size);
> >> }
> >> +
> >> for (i = 0; i < n_elems; i++) {
> >> void *curr_elem = first_elem + size * i;
> >> const VMStateField *inner_field;
> >> + bool is_null;
> >> + int max_elems = n_elems - i;
> >>
> >> old_offset = qemu_file_transferred(f);
> >> if (field->flags & VMS_ARRAY_OF_POINTER) {
> >> @@ -448,12 +452,39 @@ int vmstate_save_state_v(QEMUFile *f, const VMStateDescription *vmsd,
> >> * not follow.
> >> */
> >> inner_field = vmsd_create_fake_nullptr_field(field);
> >> + is_null = true;
> >> } else {
> >> inner_field = field;
> >> + is_null = false;
> >> + }
> >> +
> >> + /*
> >> + * Due to the fake nullptr handling above, if there's mixed
> >> + * null/non-null data, it doesn't make sense to emit a
> >> + * compressed array representation spanning the entire array
> >> + * because the field types will be different (e.g. struct
> >> + * vs. uint64_t). Search ahead for the next null/non-null
> >> + * element and start a new compressed array if found.
> >> + */
> >> + if (field->flags & VMS_ARRAY_OF_POINTER &&
> >> + is_null != is_prev_null) {
> >> +
> >> + is_prev_null = is_null;
> >> + vmdesc_loop = vmdesc;
> >> +
> >> + for (int j = i + 1; j < n_elems; j++) {
> >> + void *elem = *(void **)(first_elem + size * j);
> >> + bool elem_is_null = !elem && size;
> >> +
> >> + if (is_null != elem_is_null) {
> >> + max_elems = j - i;
> >> + break;
> >> + }
> >> + }
> >> }
> >>
> >> vmsd_desc_field_start(vmsd, vmdesc_loop, inner_field,
> >> - i, n_elems);
> >> + i, max_elems);
> >>
> >> if (inner_field->flags & VMS_STRUCT) {
> >> ret = vmstate_save_state(f, inner_field->vmsd,
> >> diff --git a/scripts/analyze-migration.py b/scripts/analyze-migration.py
> >> index 4836920ddc..9138e91a11 100755
> >> --- a/scripts/analyze-migration.py
> >> +++ b/scripts/analyze-migration.py
> >> @@ -497,7 +497,14 @@ def read(self):
> >> raise Exception("internal index of data field unmatched (%d/%d)" % (len(a), int(field['index'])))
> >> a.append(field['data'])
>
> There's actually a bug here, the code above does:
>
> if len(a) != int(field['index']):
> raise Exception()
>
> Which only works with this patch because the compressed array happens to
> come first.
I think it will work no matter how it's ordered after your patch? IOW I'd
hope it'll keep working if the 1st is a nullptr:
{"name": "css", "type": "uint8", "size": 1},
{"name": "css", "type": "struct", "struct": {"vmsd_name": "s390_css_img", ... }, "size": 768},
{"name": "css", "array_len": 254, "type": "uint8", "size": 1},
Because IIUC the python script will parse each of the lines above into a
VMSD field.
>
> >> else:
> >> - self.data[field['name']] = field['data']
> >> + # There could be multiple entries for the same field
> >> + # name, e.g. when a compressed array was broken in
> >> + # more than one piece.
> >> + if (field['name'] in self.data and
> >> + type(self.data[field['name']]) == list):
> >> + self.data[field['name']].append(field['data'])
> >> + else:
> >> + self.data[field['name']] = field['data']
> >
> > Do we realy need these script changes? I thought VMSDFieldStruct always
> > breaks array_len field into "index" based anyway?
> >
> > new_fields = []
> > for field in self.desc['struct']['fields']:
> > if not 'array_len' in field:
> > new_fields.append(field)
> > continue
> > array_len = field.pop('array_len')
> > field['index'] = 0
> > new_fields.append(field)
> > for i in range(1, array_len):
> > c = field.copy()
> > c['index'] = i
> > new_fields.append(c)
> >
> > self.desc['struct']['fields'] = new_fields
>
> This code is about decompressing the array, it doesn't handle multiple
> entries with the same name. See the JSON I posted up there.
>
> This makes the single:
>
> {"name": "css", "array_len": 254, "type": "uint8", "size": 1},
>
> become multiple:
>
> {"name": "css", "index": 0, "type": "uint8", "size": 1},
> {"name": "css", "index": 1, "type": "uint8", "size": 1},
> ...
> {"name": "css", "index": 253, "type": "uint8", "size": 1},
Correct.
I think that means for each of the break-down entries there'll be an
"index" if it's an array. What you changed above is the case where "index"
is not available, which is processing the non-array entry. Why does that
need change? What happens if you run this without the python part you
changed in this patch?
--
Peter Xu
next prev parent reply other threads:[~2025-01-08 16:15 UTC|newest]
Thread overview: 18+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-01-07 19:50 [PATCH 0/7] migration: Fix s390 regressions + migration script Fabiano Rosas
2025-01-07 19:50 ` [PATCH 1/7] migration: Add more error handling to analyze-migration.py Fabiano Rosas
2025-01-07 19:50 ` [PATCH 2/7] migration: Remove unused argument in vmsd_desc_field_end Fabiano Rosas
2025-01-07 21:14 ` Peter Xu
2025-01-07 19:50 ` [PATCH 3/7] migration: Document the effect of vmstate_info_nullptr Fabiano Rosas
2025-01-07 21:24 ` Peter Xu
2025-01-08 13:31 ` Fabiano Rosas
2025-01-08 13:48 ` Peter Xu
2025-01-08 14:37 ` Fabiano Rosas
2025-01-07 19:50 ` [PATCH 4/7] migration: Fix parsing of s390 stream Fabiano Rosas
2025-01-07 19:50 ` [PATCH 5/7] migration: Dump correct JSON format for nullptr replacement Fabiano Rosas
2025-01-07 19:50 ` [PATCH 6/7] migration: Fix arrays of pointers in JSON writer Fabiano Rosas
2025-01-07 23:25 ` Peter Xu
2025-01-08 13:52 ` Fabiano Rosas
2025-01-08 16:14 ` Peter Xu [this message]
2025-01-08 17:15 ` Fabiano Rosas
2025-01-08 17:56 ` Peter Xu
2025-01-07 19:50 ` [PATCH 7/7] s390x: Fix CSS migration Fabiano Rosas
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=Z36kScJti9LrWVU7@x1n \
--to=peterx@redhat.com \
--cc=farosas@suse.de \
--cc=qemu-devel@nongnu.org \
--cc=thuth@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).