From: David Hildenbrand <david@redhat.com>
To: Christian Borntraeger <borntraeger@de.ibm.com>,
Cornelia Huck <cohuck@redhat.com>
Cc: Collin Walling <walling@linux.ibm.com>,
mst@redhat.com, qemu-devel@nongnu.org, pasic@linux.ibm.com,
qemu-s390x@nongnu.org, pbonzini@redhat.com, rth@twiddle.net
Subject: Re: [Qemu-devel] [qemu-s390x] [PATCH v4] s390: diagnose 318 info reset and migration support
Date: Tue, 14 May 2019 11:27:32 +0200 [thread overview]
Message-ID: <078b7bcd-de03-743d-a150-456be0b09362@redhat.com> (raw)
In-Reply-To: <27bd36e1-a102-e793-6a61-3e7acb6f1255@de.ibm.com>
On 14.05.19 11:25, Christian Borntraeger wrote:
>
>
> On 14.05.19 11:23, Christian Borntraeger wrote:
>>
>>
>> On 14.05.19 11:20, David Hildenbrand wrote:
>>> On 14.05.19 11:10, Christian Borntraeger wrote:
>>>>
>>>>
>>>> On 14.05.19 10:59, David Hildenbrand wrote:
>>>>> On 14.05.19 10:49, Cornelia Huck wrote:
>>>>>> On Tue, 14 May 2019 10:37:32 +0200
>>>>>> Christian Borntraeger <borntraeger@de.ibm.com> wrote:
>>>>>>
>>>>>>> On 14.05.19 09:28, David Hildenbrand wrote:
>>>>>>>>>>> But that can be tested using the runability information if I am not wrong.
>>>>>>>>>>
>>>>>>>>>> You mean the cpu level information, right?
>>>>>>>>
>>>>>>>> Yes, query-cpu-definition includes for each model runability information
>>>>>>>> via "unavailable-features" (valid under the started QEMU machine).
>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>> and others that we have today.
>>>>>>>>>>>>
>>>>>>>>>>>> So yes, I think this would be acceptable.
>>>>>>>>>>>
>>>>>>>>>>> I guess it is acceptable yes. I doubt anybody uses that many CPUs in
>>>>>>>>>>> production either way. But you never know.
>>>>>>>>>>
>>>>>>>>>> I think that using that many cpus is a more uncommon setup, but I still
>>>>>>>>>> think that having to wait for actual failure
>>>>>>>>>
>>>>>>>>> That can happen all the time today. You can easily say z14 in the xml when
>>>>>>>>> on a zEC12. Only at startup you get the error. The question is really:
>>>>>>>>
>>>>>>>> "-smp 248 -cpu host" will no longer work, while e.g. "-smp 248 -cpu z12"
>>>>>>>> will work. Actually, even "-smp 248" will no longer work on affected
>>>>>>>> machines.
>>>>>>>>
>>>>>>>> That is why wonder if it is better to disable the feature and print a
>>>>>>>> warning. Similar to CMMA, where want want to tolerate when CMMA is not
>>>>>>>> possible in the current environment (huge pages).
>>>>>>>>
>>>>>>>> "Diag318 will not be enabled because it is not compatible with more than
>>>>>>>> 240 CPUs".
>>>>>>>>
>>>>>>>> However, I still think that implementing support for more than one SCLP
>>>>>>>> response page is the best solution. Guests will need adaptions for > 240
>>>>>>>> CPUs with Diag318, but who cares? Existing setups will continue to work.
>>>>>>>>
>>>>>>>> Implementing that SCLP thingy will avoid any warnings and any errors. It
>>>>>>>> just works from the QEMU perspective.
>>>>>>>>
>>>>>>>> Is implementing this realistic?
>>>>>>>
>>>>>>> Yes it is but it will take time. I will try to get this rolling. To make
>>>>>>> progress on the diag318 thing, can we error on startup now and simply
>>>>>>> remove that check when when have implemented a larger sccb? If we would
>>>>>>> now do all kinds of "change the max number games" would be harder to "fix".
>>>>>>
>>>>>> So, the idea right now is:
>>>>>>
>>>>>> - fail to start if you try to specify a diag318 device and more than
>>>>>> 240 cpus (do we need a knob to turn off the device?)
>>>>>> - in the future, support more than one SCLP response page
>>>>>>
>>>>>> I'm getting a bit lost in the discussion; but the above sounds
>>>>>> reasonable to me.
>>>>>>
>>>>>
>>>>> We can
>>>>>
>>>>> 1. Fail to start with #cpus > 240 when diag318=on
>>>>> 2. Remove the error once we support more than one SCLP response page
>>>>>
>>>>> Or
>>>>>
>>>>> 1. Allow to start with #cpus > 240 when diag318=on, but indicate only
>>>>> 240 CPUs via SCLP
>>>>> 2. Print a warning
>>>>> 3. Remove the restriction and the warning once we support more than one
>>>>> SCLP response page
>>>>>
>>>>> While I prefer the second approach (similar to defining zPCI devices
>>>>> without zpci=on), I could also live with the first approach.
>>>>
>>>> I prefer approach 1.
>>>>
>>>
>>> Isn't approach #2 what we discussed (limiting sclp, but of course to 247
>>> CPUs), but with an additional warning? I'm confused.
>>
>> Different numbering interpretion. I was talking about 1 = "Allow to start with #cpus > 240 when diag318=on, but indicate only
>> 240 CPUs via SCLP"
>
> So yes, variant 2 when I use your numbering. The only question is: do we need
> a warning? It probably does not hurt.
After all, we are talking about 1 VCPU that the guest can only use by
indirect probing ... I leave that up to Collin :)
--
Thanks,
David / dhildenb
next prev parent reply other threads:[~2019-05-14 9:29 UTC|newest]
Thread overview: 39+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-05-01 22:31 [Qemu-devel] [PATCH v4] s390: diagnose 318 info reset and migration support Collin Walling
2019-05-01 22:31 ` Collin Walling
2019-05-09 9:58 ` Christian Borntraeger
2019-05-09 10:05 ` David Hildenbrand
2019-05-09 20:50 ` Collin Walling
2019-05-13 5:56 ` [Qemu-devel] [qemu-s390x] " Thomas Huth
2019-05-13 7:46 ` [Qemu-devel] " David Hildenbrand
2019-05-13 8:03 ` David Hildenbrand
2019-05-13 9:34 ` Christian Borntraeger
2019-05-13 9:40 ` David Hildenbrand
2019-05-13 9:51 ` Christian Borntraeger
2019-05-13 9:57 ` David Hildenbrand
2019-05-13 10:55 ` Christian Borntraeger
2019-05-13 11:34 ` David Hildenbrand
2019-05-13 11:46 ` Cornelia Huck
2019-05-14 7:09 ` Christian Borntraeger
2019-05-14 7:28 ` David Hildenbrand
2019-05-14 8:37 ` Christian Borntraeger
2019-05-14 8:49 ` Cornelia Huck
2019-05-14 8:53 ` Christian Borntraeger
2019-05-14 8:59 ` David Hildenbrand
2019-05-14 9:07 ` Christian Borntraeger
2019-05-14 9:12 ` Cornelia Huck
2019-05-14 9:10 ` Christian Borntraeger
2019-05-14 9:20 ` David Hildenbrand
2019-05-14 9:23 ` Christian Borntraeger
2019-05-14 9:25 ` [Qemu-devel] [qemu-s390x] " Christian Borntraeger
2019-05-14 9:27 ` David Hildenbrand [this message]
2019-05-14 9:30 ` Cornelia Huck
2019-05-16 13:35 ` Collin Walling
2019-05-16 14:10 ` Christian Borntraeger
2019-05-14 8:50 ` [Qemu-devel] " David Hildenbrand
2019-05-14 8:56 ` Christian Borntraeger
2019-05-14 9:00 ` Cornelia Huck
2019-05-14 9:03 ` David Hildenbrand
2019-05-14 9:05 ` David Hildenbrand
2019-05-14 9:00 ` David Hildenbrand
2019-05-14 9:04 ` Christian Borntraeger
2019-05-16 12:42 ` Collin Walling
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=078b7bcd-de03-743d-a150-456be0b09362@redhat.com \
--to=david@redhat.com \
--cc=borntraeger@de.ibm.com \
--cc=cohuck@redhat.com \
--cc=mst@redhat.com \
--cc=pasic@linux.ibm.com \
--cc=pbonzini@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=qemu-s390x@nongnu.org \
--cc=rth@twiddle.net \
--cc=walling@linux.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).