From: Don Slutz <dslutz@verizon.com>
To: Andrew Cooper <andrew.cooper3@citrix.com>
Cc: Keir Fraser <keir@xen.org>,
Ian Campbell <ian.campbell@citrix.com>,
Stefano Stabellini <stefano.stabellini@eu.citrix.com>,
Ian Jackson <ian.jackson@eu.citrix.com>,
Don Slutz <dslutz@verizon.com>, Jan Beulich <JBeulich@suse.com>,
xen-devel <xen-devel@lists.xenproject.org>
Subject: Re: [BUGFIX][PATCH 3/4] hvm_save_one: return correct data.
Date: Sun, 15 Dec 2013 13:41:59 -0500 [thread overview]
Message-ID: <52ADF7F7.9060007@terremark.com> (raw)
In-Reply-To: <52ADF0C5.5060408@citrix.com>
[-- Attachment #1.1: Type: text/plain, Size: 9401 bytes --]
On 12/15/13 13:11, Andrew Cooper wrote:
> On 15/12/2013 17:42, Don Slutz wrote:
>> On 12/15/13 12:22, Andrew Cooper wrote:
>>> On 15/12/2013 17:19, Don Slutz wrote:
>>>> On 12/15/13 11:51, Andrew Cooper wrote:
>>>>> On 15/12/2013 00:29, Don Slutz wrote:
>>>>>>
>>>>>> I think I have corrected all coding errors (please check again).
>>>>>> And done all requested changes. I did add the reviewed by (not
>>>>>> sure if I should since this changes a large part of the patch,
>>>>>> but they are all what Jan said).
>>>>>>
>>>>>> I have unit tested it and it appears to work the same as the
>>>>>> previous version (as expected).
>>>>>>
>>>>>> Here is the new version, also attached.
>>>>>>
>>>>>> From e0e8f5246ba492b153884cea93bfe753f1b0782e Mon Sep 17 00:00:00
>>>>>> 2001
>>>>>> From: Don Slutz <dslutz@verizon.com>
>>>>>> Date: Tue, 12 Nov 2013 08:22:53 -0500
>>>>>> Subject: [PATCH v2 3/4] hvm_save_one: return correct data.
>>>>>>
>>>>>> It is possible that hvm_sr_handlers[typecode].save does not use all
>>>>>> the provided room. In that case, using:
>>>>>>
>>>>>> instance * hvm_sr_handlers[typecode].size
>>>>>>
>>>>>> does not select the correct instance. Add code to search for the
>>>>>> correct instance.
>>>>>>
>>>>>> Signed-off-by: Don Slutz <dslutz@verizon.com>
>>>>>> Reviewed-by: Jan Beulich <jbeulich@suse.com>
>>>>>
>>>>> but this fairs no better at selecting the correct subset in the
>>>>> case that less data than hvm_sr_handlers[typecode].size is written
>>>>> by hvm_sr_handlers[typecode].save.
>>>>>
>>>> True, but the inverse is the case here; .save writes 'n' 'size'
>>>> blocks. Form the loop above:
>>>>
>>>> if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU )
>>>> for_each_vcpu(d, v)
>>>> sz += hvm_sr_handlers[typecode].size;
>>>> else
>>>> sz = hvm_sr_handlers[typecode].size;
>>>>
>>>> so sz is in multiples of 'size'. Normally sz == ctxt.cur. With
>>>> some offline vcpus it write fewer 'size' blocks.
>>>>> It always increments by 'size' bytes, and will only copy the data
>>>>> back if the bytes under desc->instance happen to match the
>>>>> instance we are looking for.
>>>>>
>>>> The only time it does not find one is for an offline vcpu. Try out
>>>> the unit test code in patch #1 on an unchanged xen. It should not
>>>> display anything. Then offline a cpu in a domU (echo 0 >
>>>> /sys/devices/system/cpu/cpu1/online). And with 3 vcpus, it will
>>>> report an error.
>>>>
>>>> -Don Slutz
>>>
>>> Ah - so there are actually two problems. I see now the one you are
>>> trying to solve, and would agree that your code does solve it.
>>>
>>> However, some of the save handlers are themselves variable length,
>>> and will write records shorter than hvm_sr_handlers[typecode].size
>>> if they can get away with doing so. In this case, the new logic
>>> still wont get the correct instance.
>>>
>> Not sure which one(s) you are referring to.
>>
>> From the full dump:
>>
>> xen-hvmctx 1| grep -i entry
>> Entry 0: type 1 instance 0, length 24
>> Entry 1: type 2 instance 0, length 1024
>> Entry 2: type 2 instance 2, length 1024
>> Entry 3: type 2 instance 3, length 1024
>> Entry 4: type 2 instance 4, length 1024
>> Entry 5: type 2 instance 5, length 1024
>> Entry 6: type 2 instance 6, length 1024
>> Entry 7: type 2 instance 7, length 1024
>> Entry 8: type 3 instance 0, length 8
>> Entry 9: type 3 instance 1, length 8
>> Entry 10: type 4 instance 0, length 400
>> Entry 11: type 5 instance 0, length 24
>> Entry 12: type 5 instance 1, length 24
>> Entry 13: type 5 instance 2, length 24
>> Entry 14: type 5 instance 3, length 24
>> Entry 15: type 5 instance 4, length 24
>> Entry 16: type 5 instance 5, length 24
>> Entry 17: type 5 instance 6, length 24
>> Entry 18: type 5 instance 7, length 24
>> Entry 19: type 6 instance 0, length 1024
>> Entry 20: type 6 instance 1, length 1024
>> Entry 21: type 6 instance 2, length 1024
>> Entry 22: type 6 instance 3, length 1024
>> Entry 23: type 6 instance 4, length 1024
>> Entry 24: type 6 instance 5, length 1024
>> Entry 25: type 6 instance 6, length 1024
>> Entry 26: type 6 instance 7, length 1024
>> Entry 27: type 7 instance 0, length 16
>> Entry 28: type 8 instance 0, length 8
>> Entry 29: type 9 instance 0, length 8
>> Entry 30: type 10 instance 0, length 56
>> Entry 31: type 11 instance 0, length 16
>> Entry 32: type 12 instance 0, length 1048
>> Entry 33: type 13 instance 0, length 8
>> Entry 34: type 14 instance 0, length 240
>> Entry 35: type 14 instance 1, length 240
>> Entry 36: type 14 instance 2, length 240
>> Entry 37: type 14 instance 3, length 240
>> Entry 38: type 14 instance 4, length 240
>> Entry 39: type 14 instance 5, length 240
>> Entry 40: type 14 instance 6, length 240
>> Entry 41: type 14 instance 7, length 240
>> Entry 42: type 16 instance 0, length 856
>> Entry 43: type 16 instance 1, length 856
>> Entry 44: type 16 instance 2, length 856
>> Entry 45: type 16 instance 3, length 856
>> Entry 46: type 16 instance 4, length 856
>> Entry 47: type 16 instance 5, length 856
>> Entry 48: type 16 instance 6, length 856
>> Entry 49: type 16 instance 7, length 856
>> Entry 50: type 18 instance 0, length 24
>> Entry 51: type 18 instance 1, length 24
>> Entry 52: type 18 instance 2, length 24
>> Entry 53: type 18 instance 3, length 24
>> Entry 54: type 18 instance 4, length 24
>> Entry 55: type 18 instance 5, length 24
>> Entry 56: type 18 instance 6, length 24
>> Entry 57: type 18 instance 7, length 24
>> Entry 58: type 19 instance 0, length 8
>> Entry 59: type 19 instance 1, length 8
>> Entry 60: type 19 instance 2, length 8
>> Entry 61: type 19 instance 3, length 8
>> Entry 62: type 19 instance 4, length 8
>> Entry 63: type 19 instance 5, length 8
>> Entry 64: type 19 instance 6, length 8
>> Entry 65: type 19 instance 7, length 8
>> Entry 66: type 0 instance 0, length 0
>>
>> All typecode's appear to save the same amount per instance.
>>
>> Most use hvm_save_entry:
>>
>> ...
>> _hvm_write_entry((_h), (_src), HVM_SAVE_LENGTH(_x)); \
>>
>> and
>>
>> /* Syntactic sugar around that function: specify the max number of
>> * saves, and this calculates the size of buffer needed */
>> #define HVM_REGISTER_SAVE_RESTORE(_x, _save, _load, _num,
>> _k) \
>> static int __init
>> __hvm_register_##_x##_save_and_restore(void) \
>> { \
>> hvm_register_savevm(HVM_SAVE_CODE(_x), \
>> #_x, \
>> &_save, \
>> &_load, \
>> (_num) *
>> (HVM_SAVE_LENGTH(_x) \
>> + sizeof (struct
>> hvm_save_descriptor)), \
>> _k); \
>> return
>> 0; \
>> } \
>>
>> I do not find any that call on _hvm_write_entry directly.
>>
>> The only special one I found: CPU_XSAVE_CODE
>>
>> Still "writes" a full sized entry:
>>
>> if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id,
>> HVM_CPU_XSAVE_SIZE) )
>> return 1;
>> ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
>> h->cur += HVM_CPU_XSAVE_SIZE;
>> memset(ctxt, 0, HVM_CPU_XSAVE_SIZE);
>>
>> It then modifies the zeros conditionaly.
>>
>> if ( v->fpu_initialised )
>> memcpy(&ctxt->save_area,
>> v->arch.xsave_area, xsave_cntxt_size);
>>
>> #define HVM_CPU_XSAVE_SIZE (3 * sizeof(uint64_t) + xsave_cntxt_size)
>>
>> is part of this.
>>
>> /* We need variable length data chunk for xsave area, hence customized
>> * declaration other than HVM_REGISTER_SAVE_RESTORE.
>> */
>> static int __init __hvm_register_CPU_XSAVE_save_and_restore(void)
>> {
>> hvm_register_savevm(CPU_XSAVE_CODE,
>> "CPU_XSAVE",
>> hvm_save_cpu_xsave_states,
>> hvm_load_cpu_xsave_states,
>> HVM_CPU_XSAVE_SIZE + sizeof (struct
>> hvm_save_descriptor),
>> HVMSR_PER_VCPU);
>> return 0;
>> }
>> __initcall(__hvm_register_CPU_XSAVE_save_and_restore);
>>
>> is the final part of this one. So I do not find any code that does
>> what you are wondering about.
>>
>> -Don
>>
>
> HVM_CPU_XSAVE_SIZE() changes depending on which xsave features have
> ever been enabled by a vcpu (size is proportional to the contents of
> v->arch.xcr0_accum). It is not guaranteed to be the same for each
> vcpu in a domain, (although almost certainly will be the same for any
> recognisable OS)
>
Ah, I see.
Well, hvm_save_one, hvm_save_size, and hvm_save all expect that
hvm_sr_handlers[typecode].size has the max size. I do not see that
being true for XSAVE.
> Jan's new generic MSR save record will also write less than the
> maximum if it can.
>
This looks to be Jan's patch:
http://lists.xen.org/archives/html/xen-devel/2013-12/msg02061.html
Does look to set hvm_sr_handlers[typecode].size to the max size.
And it looks like the code I did in patch #4 would actually fix this
issue. Since it now uses the length stored in the save descriptor to
find each instance.
Jan has some questions about patch #4; so what to do about it is still
pending.
Clearly I can merge #3 and #4 into 1 patch.
-Don Slutz
> ~Andrew
[-- Attachment #1.2: Type: text/html, Size: 19622 bytes --]
[-- Attachment #2: Type: text/plain, Size: 126 bytes --]
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xen.org
http://lists.xen.org/xen-devel
next prev parent reply other threads:[~2013-12-15 18:42 UTC|newest]
Thread overview: 34+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-12-12 0:56 [BUGFIX][PATCH 0/4] hvm_save_one: return correct data Don Slutz
2013-12-12 0:56 ` [PATCH 1/4] tools/test: Add check-hvmctx Don Slutz
2013-12-12 0:56 ` [PATCH 2/4] Add tools/tests/offline_module Don Slutz
2013-12-12 10:01 ` Ian Campbell
2013-12-12 11:09 ` David Vrabel
2013-12-12 14:24 ` Don Slutz
2013-12-12 14:32 ` Don Slutz
2013-12-12 0:56 ` [BUGFIX][PATCH 3/4] hvm_save_one: return correct data Don Slutz
2013-12-13 14:20 ` Jan Beulich
2013-12-15 0:29 ` Don Slutz
2013-12-15 16:51 ` Andrew Cooper
2013-12-15 17:19 ` Don Slutz
2013-12-15 17:22 ` Andrew Cooper
2013-12-15 17:42 ` Don Slutz
2013-12-15 18:11 ` Andrew Cooper
2013-12-15 18:41 ` Don Slutz [this message]
2013-12-15 19:06 ` Andrew Cooper
2013-12-15 19:23 ` Don Slutz
2013-12-16 8:17 ` Jan Beulich
2013-12-16 17:51 ` Don Slutz
2013-12-16 18:33 ` Andrew Cooper
2013-12-22 19:40 ` Don Slutz
2013-12-22 21:13 ` Andrew Cooper
2014-01-07 15:55 ` Keir Fraser
2013-12-17 8:20 ` Jan Beulich
2013-12-17 10:40 ` Andrew Cooper
2013-12-20 0:32 ` Don Slutz
2013-12-20 13:31 ` George Dunlap
2013-12-22 19:44 ` Don Slutz
2013-12-17 15:58 ` Don Slutz
2013-12-12 0:56 ` [BUGFIX][PATCH 4/4] hvm_save_one: allow the 2nd instance to be fetched for PIC Don Slutz
2013-12-13 14:38 ` Jan Beulich
2013-12-15 1:38 ` Don Slutz
2013-12-16 8:22 ` Jan Beulich
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=52ADF7F7.9060007@terremark.com \
--to=dslutz@verizon.com \
--cc=JBeulich@suse.com \
--cc=andrew.cooper3@citrix.com \
--cc=ian.campbell@citrix.com \
--cc=ian.jackson@eu.citrix.com \
--cc=keir@xen.org \
--cc=stefano.stabellini@eu.citrix.com \
--cc=xen-devel@lists.xenproject.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).