From: Andrew Cooper <andrew.cooper3@citrix.com>
To: Don Slutz <dslutz@verizon.com>
Cc: Keir Fraser <keir@xen.org>,
Ian Campbell <ian.campbell@citrix.com>,
Stefano Stabellini <stefano.stabellini@eu.citrix.com>,
Ian Jackson <ian.jackson@eu.citrix.com>,
Jan Beulich <JBeulich@suse.com>,
xen-devel <xen-devel@lists.xenproject.org>
Subject: Re: [BUGFIX][PATCH 3/4] hvm_save_one: return correct data.
Date: Sun, 15 Dec 2013 18:11:17 +0000 [thread overview]
Message-ID: <52ADF0C5.5060408@citrix.com> (raw)
In-Reply-To: <52ADEA05.7090907@terremark.com>
[-- Attachment #1.1: Type: text/plain, Size: 8927 bytes --]
On 15/12/2013 17:42, Don Slutz wrote:
> On 12/15/13 12:22, Andrew Cooper wrote:
>> On 15/12/2013 17:19, Don Slutz wrote:
>>> On 12/15/13 11:51, Andrew Cooper wrote:
>>>> On 15/12/2013 00:29, Don Slutz wrote:
>>>>>
>>>>> I think I have corrected all coding errors (please check again).
>>>>> And done all requested changes. I did add the reviewed by (not
>>>>> sure if I should since this changes a large part of the patch, but
>>>>> they are all what Jan said).
>>>>>
>>>>> I have unit tested it and it appears to work the same as the
>>>>> previous version (as expected).
>>>>>
>>>>> Here is the new version, also attached.
>>>>>
>>>>> From e0e8f5246ba492b153884cea93bfe753f1b0782e Mon Sep 17 00:00:00
>>>>> 2001
>>>>> From: Don Slutz <dslutz@verizon.com>
>>>>> Date: Tue, 12 Nov 2013 08:22:53 -0500
>>>>> Subject: [PATCH v2 3/4] hvm_save_one: return correct data.
>>>>>
>>>>> It is possible that hvm_sr_handlers[typecode].save does not use all
>>>>> the provided room. In that case, using:
>>>>>
>>>>> instance * hvm_sr_handlers[typecode].size
>>>>>
>>>>> does not select the correct instance. Add code to search for the
>>>>> correct instance.
>>>>>
>>>>> Signed-off-by: Don Slutz <dslutz@verizon.com>
>>>>> Reviewed-by: Jan Beulich <jbeulich@suse.com>
>>>>
>>>> but this fairs no better at selecting the correct subset in the
>>>> case that less data than hvm_sr_handlers[typecode].size is written
>>>> by hvm_sr_handlers[typecode].save.
>>>>
>>> True, but the inverse is the case here; .save writes 'n' 'size'
>>> blocks. Form the loop above:
>>>
>>> if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU )
>>> for_each_vcpu(d, v)
>>> sz += hvm_sr_handlers[typecode].size;
>>> else
>>> sz = hvm_sr_handlers[typecode].size;
>>>
>>> so sz is in multiples of 'size'. Normally sz == ctxt.cur. With
>>> some offline vcpus it write fewer 'size' blocks.
>>>> It always increments by 'size' bytes, and will only copy the data
>>>> back if the bytes under desc->instance happen to match the instance
>>>> we are looking for.
>>>>
>>> The only time it does not find one is for an offline vcpu. Try out
>>> the unit test code in patch #1 on an unchanged xen. It should not
>>> display anything. Then offline a cpu in a domU (echo 0 >
>>> /sys/devices/system/cpu/cpu1/online). And with 3 vcpus, it will
>>> report an error.
>>>
>>> -Don Slutz
>>
>> Ah - so there are actually two problems. I see now the one you are
>> trying to solve, and would agree that your code does solve it.
>>
>> However, some of the save handlers are themselves variable length,
>> and will write records shorter than hvm_sr_handlers[typecode].size if
>> they can get away with doing so. In this case, the new logic still
>> wont get the correct instance.
>>
> Not sure which one(s) you are referring to.
>
> From the full dump:
>
> xen-hvmctx 1| grep -i entry
> Entry 0: type 1 instance 0, length 24
> Entry 1: type 2 instance 0, length 1024
> Entry 2: type 2 instance 2, length 1024
> Entry 3: type 2 instance 3, length 1024
> Entry 4: type 2 instance 4, length 1024
> Entry 5: type 2 instance 5, length 1024
> Entry 6: type 2 instance 6, length 1024
> Entry 7: type 2 instance 7, length 1024
> Entry 8: type 3 instance 0, length 8
> Entry 9: type 3 instance 1, length 8
> Entry 10: type 4 instance 0, length 400
> Entry 11: type 5 instance 0, length 24
> Entry 12: type 5 instance 1, length 24
> Entry 13: type 5 instance 2, length 24
> Entry 14: type 5 instance 3, length 24
> Entry 15: type 5 instance 4, length 24
> Entry 16: type 5 instance 5, length 24
> Entry 17: type 5 instance 6, length 24
> Entry 18: type 5 instance 7, length 24
> Entry 19: type 6 instance 0, length 1024
> Entry 20: type 6 instance 1, length 1024
> Entry 21: type 6 instance 2, length 1024
> Entry 22: type 6 instance 3, length 1024
> Entry 23: type 6 instance 4, length 1024
> Entry 24: type 6 instance 5, length 1024
> Entry 25: type 6 instance 6, length 1024
> Entry 26: type 6 instance 7, length 1024
> Entry 27: type 7 instance 0, length 16
> Entry 28: type 8 instance 0, length 8
> Entry 29: type 9 instance 0, length 8
> Entry 30: type 10 instance 0, length 56
> Entry 31: type 11 instance 0, length 16
> Entry 32: type 12 instance 0, length 1048
> Entry 33: type 13 instance 0, length 8
> Entry 34: type 14 instance 0, length 240
> Entry 35: type 14 instance 1, length 240
> Entry 36: type 14 instance 2, length 240
> Entry 37: type 14 instance 3, length 240
> Entry 38: type 14 instance 4, length 240
> Entry 39: type 14 instance 5, length 240
> Entry 40: type 14 instance 6, length 240
> Entry 41: type 14 instance 7, length 240
> Entry 42: type 16 instance 0, length 856
> Entry 43: type 16 instance 1, length 856
> Entry 44: type 16 instance 2, length 856
> Entry 45: type 16 instance 3, length 856
> Entry 46: type 16 instance 4, length 856
> Entry 47: type 16 instance 5, length 856
> Entry 48: type 16 instance 6, length 856
> Entry 49: type 16 instance 7, length 856
> Entry 50: type 18 instance 0, length 24
> Entry 51: type 18 instance 1, length 24
> Entry 52: type 18 instance 2, length 24
> Entry 53: type 18 instance 3, length 24
> Entry 54: type 18 instance 4, length 24
> Entry 55: type 18 instance 5, length 24
> Entry 56: type 18 instance 6, length 24
> Entry 57: type 18 instance 7, length 24
> Entry 58: type 19 instance 0, length 8
> Entry 59: type 19 instance 1, length 8
> Entry 60: type 19 instance 2, length 8
> Entry 61: type 19 instance 3, length 8
> Entry 62: type 19 instance 4, length 8
> Entry 63: type 19 instance 5, length 8
> Entry 64: type 19 instance 6, length 8
> Entry 65: type 19 instance 7, length 8
> Entry 66: type 0 instance 0, length 0
>
> All typecode's appear to save the same amount per instance.
>
> Most use hvm_save_entry:
>
> ...
> _hvm_write_entry((_h), (_src), HVM_SAVE_LENGTH(_x)); \
>
> and
>
> /* Syntactic sugar around that function: specify the max number
> of
> * saves, and this calculates the size of buffer needed */
> #define HVM_REGISTER_SAVE_RESTORE(_x, _save, _load, _num,
> _k) \
> static int __init
> __hvm_register_##_x##_save_and_restore(void) \
> {
> \
>
> hvm_register_savevm(HVM_SAVE_CODE(_x), \
>
> #_x, \
>
> &_save, \
>
> &_load, \
> (_num) *
> (HVM_SAVE_LENGTH(_x) \
> + sizeof (struct
> hvm_save_descriptor)), \
>
> _k); \
> return
> 0; \
> }
> \
>
> I do not find any that call on _hvm_write_entry directly.
>
> The only special one I found: CPU_XSAVE_CODE
>
> Still "writes" a full sized entry:
>
> if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id,
> HVM_CPU_XSAVE_SIZE) )
> return 1;
> ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
> h->cur += HVM_CPU_XSAVE_SIZE;
> memset(ctxt, 0, HVM_CPU_XSAVE_SIZE);
>
> It then modifies the zeros conditionaly.
>
> if ( v->fpu_initialised )
> memcpy(&ctxt->save_area,
> v->arch.xsave_area, xsave_cntxt_size);
>
> #define HVM_CPU_XSAVE_SIZE (3 * sizeof(uint64_t) + xsave_cntxt_size)
>
> is part of this.
>
> /* We need variable length data chunk for xsave area, hence
> customized
> * declaration other than
> HVM_REGISTER_SAVE_RESTORE.
>
> */
> static int __init __hvm_register_CPU_XSAVE_save_and_restore(void)
> {
> hvm_register_savevm(CPU_XSAVE_CODE,
> "CPU_XSAVE",
> hvm_save_cpu_xsave_states,
> hvm_load_cpu_xsave_states,
> HVM_CPU_XSAVE_SIZE + sizeof (struct
> hvm_save_descriptor),
> HVMSR_PER_VCPU);
> return 0;
> }
> __initcall(__hvm_register_CPU_XSAVE_save_and_restore);
>
> is the final part of this one. So I do not find any code that does
> what you are wondering about.
>
> -Don
>
HVM_CPU_XSAVE_SIZE() changes depending on which xsave features have ever
been enabled by a vcpu (size is proportional to the contents of
v->arch.xcr0_accum). It is not guaranteed to be the same for each vcpu
in a domain, (although almost certainly will be the same for any
recognisable OS)
Jan's new generic MSR save record will also write less than the maximum
if it can.
~Andrew
[-- Attachment #1.2: Type: text/html, Size: 17726 bytes --]
[-- Attachment #2: Type: text/plain, Size: 126 bytes --]
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
http://lists.xen.org/xen-devel
next prev parent reply other threads:[~2013-12-15 18:11 UTC|newest]
Thread overview: 34+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-12-12 0:56 [BUGFIX][PATCH 0/4] hvm_save_one: return correct data Don Slutz
2013-12-12 0:56 ` [PATCH 1/4] tools/test: Add check-hvmctx Don Slutz
2013-12-12 0:56 ` [PATCH 2/4] Add tools/tests/offline_module Don Slutz
2013-12-12 10:01 ` Ian Campbell
2013-12-12 11:09 ` David Vrabel
2013-12-12 14:24 ` Don Slutz
2013-12-12 14:32 ` Don Slutz
2013-12-12 0:56 ` [BUGFIX][PATCH 3/4] hvm_save_one: return correct data Don Slutz
2013-12-13 14:20 ` Jan Beulich
2013-12-15 0:29 ` Don Slutz
2013-12-15 16:51 ` Andrew Cooper
2013-12-15 17:19 ` Don Slutz
2013-12-15 17:22 ` Andrew Cooper
2013-12-15 17:42 ` Don Slutz
2013-12-15 18:11 ` Andrew Cooper [this message]
2013-12-15 18:41 ` Don Slutz
2013-12-15 19:06 ` Andrew Cooper
2013-12-15 19:23 ` Don Slutz
2013-12-16 8:17 ` Jan Beulich
2013-12-16 17:51 ` Don Slutz
2013-12-16 18:33 ` Andrew Cooper
2013-12-22 19:40 ` Don Slutz
2013-12-22 21:13 ` Andrew Cooper
2014-01-07 15:55 ` Keir Fraser
2013-12-17 8:20 ` Jan Beulich
2013-12-17 10:40 ` Andrew Cooper
2013-12-20 0:32 ` Don Slutz
2013-12-20 13:31 ` George Dunlap
2013-12-22 19:44 ` Don Slutz
2013-12-17 15:58 ` Don Slutz
2013-12-12 0:56 ` [BUGFIX][PATCH 4/4] hvm_save_one: allow the 2nd instance to be fetched for PIC Don Slutz
2013-12-13 14:38 ` Jan Beulich
2013-12-15 1:38 ` Don Slutz
2013-12-16 8:22 ` Jan Beulich
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=52ADF0C5.5060408@citrix.com \
--to=andrew.cooper3@citrix.com \
--cc=JBeulich@suse.com \
--cc=dslutz@verizon.com \
--cc=ian.campbell@citrix.com \
--cc=ian.jackson@eu.citrix.com \
--cc=keir@xen.org \
--cc=stefano.stabellini@eu.citrix.com \
--cc=xen-devel@lists.xenproject.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).