xen-devel.lists.xenproject.org archive mirror
 help / color / mirror / Atom feed
* [xen-unstable test] 118078: regressions - FAIL
@ 2018-01-16  8:43 osstest service owner
  2018-01-16  8:57 ` Jan Beulich
  0 siblings, 1 reply; 7+ messages in thread
From: osstest service owner @ 2018-01-16  8:43 UTC (permalink / raw)
  To: xen-devel, osstest-admin

flight 118078 xen-unstable real [real]
http://logs.test-lab.xenproject.org/osstest/logs/118078/

Regressions :-(

Tests which did not succeed and are blocking,
including tests which could not be run:
 build-arm64-pvops             6 kernel-build             fail REGR. vs. 118003
 test-amd64-amd64-xl-qemut-win7-amd64 10 windows-install  fail REGR. vs. 118003

Tests which did not succeed, but are not blocking:
 test-arm64-arm64-xl           1 build-check(1)               blocked  n/a
 test-arm64-arm64-xl-credit2   1 build-check(1)               blocked  n/a
 test-arm64-arm64-examine      1 build-check(1)               blocked  n/a
 test-arm64-arm64-xl-xsm       1 build-check(1)               blocked  n/a
 test-arm64-arm64-libvirt-xsm  1 build-check(1)               blocked  n/a
 test-armhf-armhf-libvirt     14 saverestore-support-check    fail  like 118003
 test-armhf-armhf-libvirt-xsm 14 saverestore-support-check    fail  like 118003
 test-amd64-amd64-xl-qemuu-win7-amd64 17 guest-stop            fail like 118003
 test-amd64-i386-xl-qemuu-win7-amd64 17 guest-stop             fail like 118003
 test-amd64-amd64-xl-qemuu-ws16-amd64 17 guest-stop            fail like 118003
 test-amd64-i386-xl-qemut-win7-amd64 17 guest-stop             fail like 118003
 test-armhf-armhf-libvirt-raw 13 saverestore-support-check    fail  like 118003
 test-amd64-i386-xl-qemuu-ws16-amd64 17 guest-stop             fail like 118003
 test-amd64-amd64-xl-qemut-ws16-amd64 17 guest-stop            fail like 118003
 test-amd64-amd64-xl-pvhv2-intel 12 guest-start                 fail never pass
 test-amd64-amd64-xl-pvhv2-amd 12 guest-start                  fail  never pass
 test-amd64-amd64-libvirt     13 migrate-support-check        fail   never pass
 test-amd64-amd64-libvirt-xsm 13 migrate-support-check        fail   never pass
 test-amd64-i386-libvirt      13 migrate-support-check        fail   never pass
 test-amd64-i386-libvirt-xsm  13 migrate-support-check        fail   never pass
 test-amd64-amd64-libvirt-qemuu-debianhvm-amd64-xsm 11 migrate-support-check fail never pass
 test-amd64-i386-libvirt-qemuu-debianhvm-amd64-xsm 11 migrate-support-check fail never pass
 test-amd64-amd64-qemuu-nested-amd 17 debian-hvm-install/l1/l2  fail never pass
 test-amd64-amd64-libvirt-vhd 12 migrate-support-check        fail   never pass
 test-armhf-armhf-xl          13 migrate-support-check        fail   never pass
 test-armhf-armhf-xl          14 saverestore-support-check    fail   never pass
 test-armhf-armhf-xl-xsm      13 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-xsm      14 saverestore-support-check    fail   never pass
 test-armhf-armhf-libvirt     13 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-cubietruck 13 migrate-support-check        fail never pass
 test-armhf-armhf-libvirt-xsm 13 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-cubietruck 14 saverestore-support-check    fail never pass
 test-armhf-armhf-xl-multivcpu 13 migrate-support-check        fail  never pass
 test-armhf-armhf-xl-multivcpu 14 saverestore-support-check    fail  never pass
 test-armhf-armhf-xl-rtds     13 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-rtds     14 saverestore-support-check    fail   never pass
 test-armhf-armhf-libvirt-raw 12 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-vhd      12 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-vhd      13 saverestore-support-check    fail   never pass
 test-armhf-armhf-xl-credit2  13 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-credit2  14 saverestore-support-check    fail   never pass
 test-armhf-armhf-xl-arndale  13 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-arndale  14 saverestore-support-check    fail   never pass
 test-amd64-i386-xl-qemut-ws16-amd64 17 guest-stop              fail never pass
 test-amd64-i386-xl-qemuu-win10-i386 10 windows-install         fail never pass
 test-amd64-amd64-xl-qemuu-win10-i386 10 windows-install        fail never pass
 test-amd64-amd64-xl-qemut-win10-i386 10 windows-install        fail never pass
 test-amd64-i386-xl-qemut-win10-i386 10 windows-install         fail never pass

version targeted for testing:
 xen                  bd61fe94bee0556bc2f64999a4a8315b93f90f21
baseline version:
 xen                  2d70b54e055635ff60526b6949156504b6194b7c

Last test of basis   118003  2018-01-15 04:26:28 Z    1 days
Testing same since   118078  2018-01-15 19:14:52 Z    0 days    1 attempts

------------------------------------------------------------
People who touched revisions under test:
  Andrew Cooper <andrew.cooper3@citrix.com>

jobs:
 build-amd64-xsm                                              pass    
 build-arm64-xsm                                              pass    
 build-armhf-xsm                                              pass    
 build-i386-xsm                                               pass    
 build-amd64-xtf                                              pass    
 build-amd64                                                  pass    
 build-arm64                                                  pass    
 build-armhf                                                  pass    
 build-i386                                                   pass    
 build-amd64-libvirt                                          pass    
 build-arm64-libvirt                                          pass    
 build-armhf-libvirt                                          pass    
 build-i386-libvirt                                           pass    
 build-amd64-prev                                             pass    
 build-i386-prev                                              pass    
 build-amd64-pvops                                            pass    
 build-arm64-pvops                                            fail    
 build-armhf-pvops                                            pass    
 build-i386-pvops                                             pass    
 build-amd64-rumprun                                          pass    
 build-i386-rumprun                                           pass    
 test-xtf-amd64-amd64-1                                       pass    
 test-xtf-amd64-amd64-2                                       pass    
 test-xtf-amd64-amd64-3                                       pass    
 test-xtf-amd64-amd64-4                                       pass    
 test-xtf-amd64-amd64-5                                       pass    
 test-amd64-amd64-xl                                          pass    
 test-arm64-arm64-xl                                          blocked 
 test-armhf-armhf-xl                                          pass    
 test-amd64-i386-xl                                           pass    
 test-amd64-amd64-xl-qemut-debianhvm-amd64-xsm                pass    
 test-amd64-i386-xl-qemut-debianhvm-amd64-xsm                 pass    
 test-amd64-amd64-libvirt-qemuu-debianhvm-amd64-xsm           pass    
 test-amd64-i386-libvirt-qemuu-debianhvm-amd64-xsm            pass    
 test-amd64-amd64-xl-qemuu-debianhvm-amd64-xsm                pass    
 test-amd64-i386-xl-qemuu-debianhvm-amd64-xsm                 pass    
 test-amd64-amd64-xl-qemut-stubdom-debianhvm-amd64-xsm        pass    
 test-amd64-i386-xl-qemut-stubdom-debianhvm-amd64-xsm         pass    
 test-amd64-amd64-libvirt-xsm                                 pass    
 test-arm64-arm64-libvirt-xsm                                 blocked 
 test-armhf-armhf-libvirt-xsm                                 pass    
 test-amd64-i386-libvirt-xsm                                  pass    
 test-amd64-amd64-xl-xsm                                      pass    
 test-arm64-arm64-xl-xsm                                      blocked 
 test-armhf-armhf-xl-xsm                                      pass    
 test-amd64-i386-xl-xsm                                       pass    
 test-amd64-amd64-qemuu-nested-amd                            fail    
 test-amd64-amd64-xl-pvhv2-amd                                fail    
 test-amd64-i386-qemut-rhel6hvm-amd                           pass    
 test-amd64-i386-qemuu-rhel6hvm-amd                           pass    
 test-amd64-amd64-xl-qemut-debianhvm-amd64                    pass    
 test-amd64-i386-xl-qemut-debianhvm-amd64                     pass    
 test-amd64-amd64-xl-qemuu-debianhvm-amd64                    pass    
 test-amd64-i386-xl-qemuu-debianhvm-amd64                     pass    
 test-amd64-i386-freebsd10-amd64                              pass    
 test-amd64-amd64-xl-qemuu-ovmf-amd64                         pass    
 test-amd64-i386-xl-qemuu-ovmf-amd64                          pass    
 test-amd64-amd64-rumprun-amd64                               pass    
 test-amd64-amd64-xl-qemut-win7-amd64                         fail    
 test-amd64-i386-xl-qemut-win7-amd64                          fail    
 test-amd64-amd64-xl-qemuu-win7-amd64                         fail    
 test-amd64-i386-xl-qemuu-win7-amd64                          fail    
 test-amd64-amd64-xl-qemut-ws16-amd64                         fail    
 test-amd64-i386-xl-qemut-ws16-amd64                          fail    
 test-amd64-amd64-xl-qemuu-ws16-amd64                         fail    
 test-amd64-i386-xl-qemuu-ws16-amd64                          fail    
 test-armhf-armhf-xl-arndale                                  pass    
 test-amd64-amd64-xl-credit2                                  pass    
 test-arm64-arm64-xl-credit2                                  blocked 
 test-armhf-armhf-xl-credit2                                  pass    
 test-armhf-armhf-xl-cubietruck                               pass    
 test-amd64-amd64-examine                                     pass    
 test-arm64-arm64-examine                                     blocked 
 test-armhf-armhf-examine                                     pass    
 test-amd64-i386-examine                                      pass    
 test-amd64-i386-freebsd10-i386                               pass    
 test-amd64-i386-rumprun-i386                                 pass    
 test-amd64-amd64-xl-qemut-win10-i386                         fail    
 test-amd64-i386-xl-qemut-win10-i386                          fail    
 test-amd64-amd64-xl-qemuu-win10-i386                         fail    
 test-amd64-i386-xl-qemuu-win10-i386                          fail    
 test-amd64-amd64-qemuu-nested-intel                          pass    
 test-amd64-amd64-xl-pvhv2-intel                              fail    
 test-amd64-i386-qemut-rhel6hvm-intel                         pass    
 test-amd64-i386-qemuu-rhel6hvm-intel                         pass    
 test-amd64-amd64-libvirt                                     pass    
 test-armhf-armhf-libvirt                                     pass    
 test-amd64-i386-libvirt                                      pass    
 test-amd64-amd64-livepatch                                   pass    
 test-amd64-i386-livepatch                                    pass    
 test-amd64-amd64-migrupgrade                                 pass    
 test-amd64-i386-migrupgrade                                  pass    
 test-amd64-amd64-xl-multivcpu                                pass    
 test-armhf-armhf-xl-multivcpu                                pass    
 test-amd64-amd64-pair                                        pass    
 test-amd64-i386-pair                                         pass    
 test-amd64-amd64-libvirt-pair                                pass    
 test-amd64-i386-libvirt-pair                                 pass    
 test-amd64-amd64-amd64-pvgrub                                pass    
 test-amd64-amd64-i386-pvgrub                                 pass    
 test-amd64-amd64-pygrub                                      pass    
 test-amd64-amd64-xl-qcow2                                    pass    
 test-armhf-armhf-libvirt-raw                                 pass    
 test-amd64-i386-xl-raw                                       pass    
 test-amd64-amd64-xl-rtds                                     pass    
 test-armhf-armhf-xl-rtds                                     pass    
 test-amd64-amd64-libvirt-vhd                                 pass    
 test-armhf-armhf-xl-vhd                                      pass    


------------------------------------------------------------
sg-report-flight on osstest.test-lab.xenproject.org
logs: /home/logs/logs
images: /home/logs/images

Logs, config files, etc. are available at
    http://logs.test-lab.xenproject.org/osstest/logs

Explanation of these reports, and of osstest in general, is at
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README.email;hb=master
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README;hb=master

Test harness code can be found at
    http://xenbits.xen.org/gitweb?p=osstest.git;a=summary


Not pushing.

------------------------------------------------------------
commit bd61fe94bee0556bc2f64999a4a8315b93f90f21
Author: Andrew Cooper <andrew.cooper3@citrix.com>
Date:   Fri Sep 1 12:15:39 2017 +0100

    x86/mm: Always set _PAGE_ACCESSED on L4e updates
    
    Signed-off-by: Andrew Cooper <andrew.cooper3@citrix.com>
    Reviewed-by: Jan Beulich <jbeulich@suse.com>
(qemu changes not included)

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [xen-unstable test] 118078: regressions - FAIL
  2018-01-16  8:43 [xen-unstable test] 118078: regressions - FAIL osstest service owner
@ 2018-01-16  8:57 ` Jan Beulich
  2018-01-16  9:26   ` Paul Durrant
  0 siblings, 1 reply; 7+ messages in thread
From: Jan Beulich @ 2018-01-16  8:57 UTC (permalink / raw)
  To: Paul Durrant; +Cc: xen-devel, osstest-admin

>>> On 16.01.18 at 09:43, <osstest-admin@xenproject.org> wrote:
> flight 118078 xen-unstable real [real]
> http://logs.test-lab.xenproject.org/osstest/logs/118078/ 
> 
> Regressions :-(
> 
> Tests which did not succeed and are blocking,
> including tests which could not be run:
>  build-arm64-pvops             6 kernel-build             fail REGR. vs. 118003
>  test-amd64-amd64-xl-qemut-win7-amd64 10 windows-install  fail REGR. vs. 118003

Paul,

is this last one something you could look into?

(XEN) d4: VIRIDIAN GUEST_OS_ID: vendor: 1 os: 4 major: 6 minor: 1 sp: 0 build: 1db0
(XEN) d4: VIRIDIAN HYPERCALL: enabled: 1 pfn: 3ffff
(XEN) d4v0: VIRIDIAN VP_ASSIST_PAGE: enabled: 1 pfn: 3fffe
(XEN) domain_crash called from viridian.c:452
(XEN) Domain 4 (vcpu#0) crashed on cpu#1:
(XEN) ----[ Xen-4.11-unstable  x86_64  debug=y   Not tainted ]----
(XEN) CPU:    1
(XEN) RIP:    0010:[<fffff8000265d479>]
(XEN) RFLAGS: 0000000000000286   CONTEXT: hvm guest (d4v0)
(XEN) rax: 0000000000000000   rbx: fffff800027f7e80   rcx: 0000000000000001
(XEN) rdx: 0000000000000000   rsi: fffffa800129d040   rdi: fffff80002805c40
(XEN) rbp: 0000000000000080   rsp: fffff880009b0d80   r8:  0000000000000000
(XEN) r9:  fffff800027f7e80   r10: fffffa800129d040   r11: fffff800027f7e90
(XEN) r12: fffff800008129a0   r13: fffff800028b9be0   r14: fffffa8001239b30
(XEN) r15: fffff80000b96080   cr0: 0000000080050031   cr4: 00000000000006b8
(XEN) cr3: 0000000000187000   cr2: 0000000000000000
(XEN) fsb: 0000000000000000   gsb: fffff800027f7d00   gss: fffff800027f7d00
(XEN) ds: 002b   es: 002b   fs: 0053   gs: 002b   ss: 0018   cs: 0010

I.e. the domain_crash() in viridian_start_apic_assist().

Jan


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [xen-unstable test] 118078: regressions - FAIL
  2018-01-16  8:57 ` Jan Beulich
@ 2018-01-16  9:26   ` Paul Durrant
  2018-01-16 17:30     ` Paul Durrant
  0 siblings, 1 reply; 7+ messages in thread
From: Paul Durrant @ 2018-01-16  9:26 UTC (permalink / raw)
  To: 'Jan Beulich'; +Cc: xen-devel, osstest-admin@xenproject.org

> -----Original Message-----
> From: Jan Beulich [mailto:JBeulich@suse.com]
> Sent: 16 January 2018 08:58
> To: Paul Durrant <Paul.Durrant@citrix.com>
> Cc: xen-devel <xen-devel@lists.xenproject.org>; osstest-
> admin@xenproject.org
> Subject: Re: [Xen-devel] [xen-unstable test] 118078: regressions - FAIL
> 
> >>> On 16.01.18 at 09:43, <osstest-admin@xenproject.org> wrote:
> > flight 118078 xen-unstable real [real]
> > http://logs.test-lab.xenproject.org/osstest/logs/118078/
> >
> > Regressions :-(
> >
> > Tests which did not succeed and are blocking,
> > including tests which could not be run:
> >  build-arm64-pvops             6 kernel-build             fail REGR. vs. 118003
> >  test-amd64-amd64-xl-qemut-win7-amd64 10 windows-install  fail REGR. vs.
> 118003
> 
> Paul,
> 
> is this last one something you could look into?
> 
> (XEN) d4: VIRIDIAN GUEST_OS_ID: vendor: 1 os: 4 major: 6 minor: 1 sp: 0
> build: 1db0
> (XEN) d4: VIRIDIAN HYPERCALL: enabled: 1 pfn: 3ffff
> (XEN) d4v0: VIRIDIAN VP_ASSIST_PAGE: enabled: 1 pfn: 3fffe
> (XEN) domain_crash called from viridian.c:452
> (XEN) Domain 4 (vcpu#0) crashed on cpu#1:
> (XEN) ----[ Xen-4.11-unstable  x86_64  debug=y   Not tainted ]----
> (XEN) CPU:    1
> (XEN) RIP:    0010:[<fffff8000265d479>]
> (XEN) RFLAGS: 0000000000000286   CONTEXT: hvm guest (d4v0)
> (XEN) rax: 0000000000000000   rbx: fffff800027f7e80   rcx: 0000000000000001
> (XEN) rdx: 0000000000000000   rsi: fffffa800129d040   rdi: fffff80002805c40
> (XEN) rbp: 0000000000000080   rsp: fffff880009b0d80   r8:  0000000000000000
> (XEN) r9:  fffff800027f7e80   r10: fffffa800129d040   r11: fffff800027f7e90
> (XEN) r12: fffff800008129a0   r13: fffff800028b9be0   r14: fffffa8001239b30
> (XEN) r15: fffff80000b96080   cr0: 0000000080050031   cr4: 00000000000006b8
> (XEN) cr3: 0000000000187000   cr2: 0000000000000000
> (XEN) fsb: 0000000000000000   gsb: fffff800027f7d00   gss: fffff800027f7d00
> (XEN) ds: 002b   es: 002b   fs: 0053   gs: 002b   ss: 0018   cs: 0010
> 
> I.e. the domain_crash() in viridian_start_apic_assist().
> 

Yes, I'll have a look at that.

  Paul

> Jan


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [xen-unstable test] 118078: regressions - FAIL
  2018-01-16  9:26   ` Paul Durrant
@ 2018-01-16 17:30     ` Paul Durrant
  2018-01-17  8:51       ` Jan Beulich
  0 siblings, 1 reply; 7+ messages in thread
From: Paul Durrant @ 2018-01-16 17:30 UTC (permalink / raw)
  To: Paul Durrant, 'Jan Beulich'
  Cc: xen-devel, osstest-admin@xenproject.org

> -----Original Message-----
> From: Xen-devel [mailto:xen-devel-bounces@lists.xenproject.org] On Behalf
> Of Paul Durrant
> Sent: 16 January 2018 09:27
> To: 'Jan Beulich' <JBeulich@suse.com>
> Cc: xen-devel <xen-devel@lists.xenproject.org>; osstest-
> admin@xenproject.org
> Subject: Re: [Xen-devel] [xen-unstable test] 118078: regressions - FAIL
> 
> > -----Original Message-----
> > From: Jan Beulich [mailto:JBeulich@suse.com]
> > Sent: 16 January 2018 08:58
> > To: Paul Durrant <Paul.Durrant@citrix.com>
> > Cc: xen-devel <xen-devel@lists.xenproject.org>; osstest-
> > admin@xenproject.org
> > Subject: Re: [Xen-devel] [xen-unstable test] 118078: regressions - FAIL
> >
> > >>> On 16.01.18 at 09:43, <osstest-admin@xenproject.org> wrote:
> > > flight 118078 xen-unstable real [real]
> > > http://logs.test-lab.xenproject.org/osstest/logs/118078/
> > >
> > > Regressions :-(
> > >
> > > Tests which did not succeed and are blocking,
> > > including tests which could not be run:
> > >  build-arm64-pvops             6 kernel-build             fail REGR. vs. 118003
> > >  test-amd64-amd64-xl-qemut-win7-amd64 10 windows-install  fail REGR.
> vs.
> > 118003
> >
> > Paul,
> >
> > is this last one something you could look into?
> >
> > (XEN) d4: VIRIDIAN GUEST_OS_ID: vendor: 1 os: 4 major: 6 minor: 1 sp: 0
> > build: 1db0
> > (XEN) d4: VIRIDIAN HYPERCALL: enabled: 1 pfn: 3ffff
> > (XEN) d4v0: VIRIDIAN VP_ASSIST_PAGE: enabled: 1 pfn: 3fffe
> > (XEN) domain_crash called from viridian.c:452
> > (XEN) Domain 4 (vcpu#0) crashed on cpu#1:
> > (XEN) ----[ Xen-4.11-unstable  x86_64  debug=y   Not tainted ]----
> > (XEN) CPU:    1
> > (XEN) RIP:    0010:[<fffff8000265d479>]
> > (XEN) RFLAGS: 0000000000000286   CONTEXT: hvm guest (d4v0)
> > (XEN) rax: 0000000000000000   rbx: fffff800027f7e80   rcx: 0000000000000001
> > (XEN) rdx: 0000000000000000   rsi: fffffa800129d040   rdi: fffff80002805c40
> > (XEN) rbp: 0000000000000080   rsp: fffff880009b0d80   r8:
> 0000000000000000
> > (XEN) r9:  fffff800027f7e80   r10: fffffa800129d040   r11: fffff800027f7e90
> > (XEN) r12: fffff800008129a0   r13: fffff800028b9be0   r14: fffffa8001239b30
> > (XEN) r15: fffff80000b96080   cr0: 0000000080050031   cr4:
> 00000000000006b8
> > (XEN) cr3: 0000000000187000   cr2: 0000000000000000
> > (XEN) fsb: 0000000000000000   gsb: fffff800027f7d00   gss: fffff800027f7d00
> > (XEN) ds: 002b   es: 002b   fs: 0053   gs: 002b   ss: 0018   cs: 0010
> >
> > I.e. the domain_crash() in viridian_start_apic_assist().
> >
> 
> Yes, I'll have a look at that.

No real clue about this as yet. It is odd that the guest has only set up one of the APIC assist pages and yet has taken an interrupt...

Jan 16 01:46:05.691223 (XEN) Dumping guest's current state at key_handler...
Jan 16 01:46:05.691265 (XEN) Size of VMCB = 4096, paddr = 000000020f7f7000, vaddr = ffff83020f7f7000
Jan 16 01:46:05.699269 (XEN) cr_intercepts = 0xfef3fef3 dr_intercepts = 0xffffffff exception_intercepts = 0x60082
Jan 16 01:46:05.707128 (XEN) general1_intercepts = 0xbdc4000f general2_intercepts = 0x2e7f
Jan 16 01:46:05.715222 (XEN) iopm_base_pa = 0xdfd71000 msrpm_base_pa = 0x20f7f4000 tsc_offset = 0xfffffc36684278c9
Jan 16 01:46:05.723116 (XEN) tlb_control = 0 vintr = 0x1020001 interrupt_shadow = 0
Jan 16 01:46:05.723153 (XEN) eventinj 000000008000002f, valid? 1, ec? 0, type 0, vector 0x2f
Jan 16 01:46:05.731141 (XEN) exitcode = 0x64 exitintinfo = 0
Jan 16 01:46:05.739123 (XEN) exitinfo1 = 0 exitinfo2 = 0
Jan 16 01:46:05.739157 (XEN) np_enable = 0x1 guest_asid = 0x4b49
Jan 16 01:46:05.739187 (XEN) virtual vmload/vmsave = 0, virt_ext = 0

I'd expect it to have interrupts disabled at this point. Seemingly doesn't repro on Intel h/w (although I was testing with Win7 SP1 rather than RTM) so I'll try to find some AMD h/w and try again.

  Paul
 
>   Paul
> 
> > Jan
> 
> 
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@lists.xenproject.org
> https://lists.xenproject.org/mailman/listinfo/xen-devel
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [xen-unstable test] 118078: regressions - FAIL
  2018-01-16 17:30     ` Paul Durrant
@ 2018-01-17  8:51       ` Jan Beulich
  2018-01-17  9:37         ` Paul Durrant
  0 siblings, 1 reply; 7+ messages in thread
From: Jan Beulich @ 2018-01-17  8:51 UTC (permalink / raw)
  To: Paul Durrant; +Cc: xen-devel, osstest-admin@xenproject.org

>>> On 16.01.18 at 18:30, <Paul.Durrant@citrix.com> wrote:
>> From: Xen-devel [mailto:xen-devel-bounces@lists.xenproject.org] On Behalf
>> Of Paul Durrant
>> Sent: 16 January 2018 09:27
>> > From: Jan Beulich [mailto:JBeulich@suse.com]
>> > Sent: 16 January 2018 08:58
>> > >>> On 16.01.18 at 09:43, <osstest-admin@xenproject.org> wrote:
>> > > flight 118078 xen-unstable real [real]
>> > > http://logs.test-lab.xenproject.org/osstest/logs/118078/ 
>> > >
>> > > Regressions :-(
>> > >
>> > > Tests which did not succeed and are blocking,
>> > > including tests which could not be run:
>> > >  build-arm64-pvops             6 kernel-build             fail REGR. vs. 118003
>> > >  test-amd64-amd64-xl-qemut-win7-amd64 10 windows-install  fail REGR.
>> vs.
>> > 118003
>> >
>> > Paul,
>> >
>> > is this last one something you could look into?
>> >
>> > (XEN) d4: VIRIDIAN GUEST_OS_ID: vendor: 1 os: 4 major: 6 minor: 1 sp: 0
>> > build: 1db0
>> > (XEN) d4: VIRIDIAN HYPERCALL: enabled: 1 pfn: 3ffff
>> > (XEN) d4v0: VIRIDIAN VP_ASSIST_PAGE: enabled: 1 pfn: 3fffe
>> > (XEN) domain_crash called from viridian.c:452
>> > (XEN) Domain 4 (vcpu#0) crashed on cpu#1:
>> > (XEN) ----[ Xen-4.11-unstable  x86_64  debug=y   Not tainted ]----
>> > (XEN) CPU:    1
>> > (XEN) RIP:    0010:[<fffff8000265d479>]
>> > (XEN) RFLAGS: 0000000000000286   CONTEXT: hvm guest (d4v0)
>> > (XEN) rax: 0000000000000000   rbx: fffff800027f7e80   rcx: 0000000000000001
>> > (XEN) rdx: 0000000000000000   rsi: fffffa800129d040   rdi: fffff80002805c40
>> > (XEN) rbp: 0000000000000080   rsp: fffff880009b0d80   r8:
>> 0000000000000000
>> > (XEN) r9:  fffff800027f7e80   r10: fffffa800129d040   r11: fffff800027f7e90
>> > (XEN) r12: fffff800008129a0   r13: fffff800028b9be0   r14: fffffa8001239b30
>> > (XEN) r15: fffff80000b96080   cr0: 0000000080050031   cr4:
>> 00000000000006b8
>> > (XEN) cr3: 0000000000187000   cr2: 0000000000000000
>> > (XEN) fsb: 0000000000000000   gsb: fffff800027f7d00   gss: fffff800027f7d00
>> > (XEN) ds: 002b   es: 002b   fs: 0053   gs: 002b   ss: 0018   cs: 0010
>> >
>> > I.e. the domain_crash() in viridian_start_apic_assist().
>> >
>> 
>> Yes, I'll have a look at that.
> 
> No real clue about this as yet. It is odd that the guest has only set up one 
> of the APIC assist pages and yet has taken an interrupt...
> 
> Jan 16 01:46:05.691223 (XEN) Dumping guest's current state at key_handler...
> Jan 16 01:46:05.691265 (XEN) Size of VMCB = 4096, paddr = 000000020f7f7000, 
> vaddr = ffff83020f7f7000
> Jan 16 01:46:05.699269 (XEN) cr_intercepts = 0xfef3fef3 dr_intercepts = 
> 0xffffffff exception_intercepts = 0x60082
> Jan 16 01:46:05.707128 (XEN) general1_intercepts = 0xbdc4000f 
> general2_intercepts = 0x2e7f
> Jan 16 01:46:05.715222 (XEN) iopm_base_pa = 0xdfd71000 msrpm_base_pa = 
> 0x20f7f4000 tsc_offset = 0xfffffc36684278c9
> Jan 16 01:46:05.723116 (XEN) tlb_control = 0 vintr = 0x1020001 
> interrupt_shadow = 0
> Jan 16 01:46:05.723153 (XEN) eventinj 000000008000002f, valid? 1, ec? 0, 
> type 0, vector 0x2f
> Jan 16 01:46:05.731141 (XEN) exitcode = 0x64 exitintinfo = 0
> Jan 16 01:46:05.739123 (XEN) exitinfo1 = 0 exitinfo2 = 0
> Jan 16 01:46:05.739157 (XEN) np_enable = 0x1 guest_asid = 0x4b49
> Jan 16 01:46:05.739187 (XEN) virtual vmload/vmsave = 0, virt_ext = 0
> 
> I'd expect it to have interrupts disabled at this point. Seemingly doesn't 
> repro on Intel h/w (although I was testing with Win7 SP1 rather than RTM) so 
> I'll try to find some AMD h/w and try again.

Well, it looks to be a random problem in the first place, or else we
would have known about the issue much earlier I think. I.e. I'm
not sure I see what you take the "AMD only" from here.

As to interrupt state - isn't it quite normal for an OS to bring up
the BSP first, enable interrupts on it, and then bring up APs?
That's how we do it in Xen.

Jan


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [xen-unstable test] 118078: regressions - FAIL
  2018-01-17  8:51       ` Jan Beulich
@ 2018-01-17  9:37         ` Paul Durrant
  2018-01-17 10:17           ` Paul Durrant
  0 siblings, 1 reply; 7+ messages in thread
From: Paul Durrant @ 2018-01-17  9:37 UTC (permalink / raw)
  To: 'Jan Beulich'; +Cc: xen-devel, osstest-admin@xenproject.org

> -----Original Message-----
> From: Jan Beulich [mailto:JBeulich@suse.com]
> Sent: 17 January 2018 08:52
> To: Paul Durrant <Paul.Durrant@citrix.com>
> Cc: xen-devel <xen-devel@lists.xenproject.org>; osstest-
> admin@xenproject.org
> Subject: RE: [Xen-devel] [xen-unstable test] 118078: regressions - FAIL
> 
> >>> On 16.01.18 at 18:30, <Paul.Durrant@citrix.com> wrote:
> >> From: Xen-devel [mailto:xen-devel-bounces@lists.xenproject.org] On
> Behalf
> >> Of Paul Durrant
> >> Sent: 16 January 2018 09:27
> >> > From: Jan Beulich [mailto:JBeulich@suse.com]
> >> > Sent: 16 January 2018 08:58
> >> > >>> On 16.01.18 at 09:43, <osstest-admin@xenproject.org> wrote:
> >> > > flight 118078 xen-unstable real [real]
> >> > > http://logs.test-lab.xenproject.org/osstest/logs/118078/
> >> > >
> >> > > Regressions :-(
> >> > >
> >> > > Tests which did not succeed and are blocking,
> >> > > including tests which could not be run:
> >> > >  build-arm64-pvops             6 kernel-build             fail REGR. vs. 118003
> >> > >  test-amd64-amd64-xl-qemut-win7-amd64 10 windows-install  fail
> REGR.
> >> vs.
> >> > 118003
> >> >
> >> > Paul,
> >> >
> >> > is this last one something you could look into?
> >> >
> >> > (XEN) d4: VIRIDIAN GUEST_OS_ID: vendor: 1 os: 4 major: 6 minor: 1 sp: 0
> >> > build: 1db0
> >> > (XEN) d4: VIRIDIAN HYPERCALL: enabled: 1 pfn: 3ffff
> >> > (XEN) d4v0: VIRIDIAN VP_ASSIST_PAGE: enabled: 1 pfn: 3fffe
> >> > (XEN) domain_crash called from viridian.c:452
> >> > (XEN) Domain 4 (vcpu#0) crashed on cpu#1:
> >> > (XEN) ----[ Xen-4.11-unstable  x86_64  debug=y   Not tainted ]----
> >> > (XEN) CPU:    1
> >> > (XEN) RIP:    0010:[<fffff8000265d479>]
> >> > (XEN) RFLAGS: 0000000000000286   CONTEXT: hvm guest (d4v0)
> >> > (XEN) rax: 0000000000000000   rbx: fffff800027f7e80   rcx:
> 0000000000000001
> >> > (XEN) rdx: 0000000000000000   rsi: fffffa800129d040   rdi:
> fffff80002805c40
> >> > (XEN) rbp: 0000000000000080   rsp: fffff880009b0d80   r8:
> >> 0000000000000000
> >> > (XEN) r9:  fffff800027f7e80   r10: fffffa800129d040   r11: fffff800027f7e90
> >> > (XEN) r12: fffff800008129a0   r13: fffff800028b9be0   r14:
> fffffa8001239b30
> >> > (XEN) r15: fffff80000b96080   cr0: 0000000080050031   cr4:
> >> 00000000000006b8
> >> > (XEN) cr3: 0000000000187000   cr2: 0000000000000000
> >> > (XEN) fsb: 0000000000000000   gsb: fffff800027f7d00   gss:
> fffff800027f7d00
> >> > (XEN) ds: 002b   es: 002b   fs: 0053   gs: 002b   ss: 0018   cs: 0010
> >> >
> >> > I.e. the domain_crash() in viridian_start_apic_assist().
> >> >
> >>
> >> Yes, I'll have a look at that.
> >
> > No real clue about this as yet. It is odd that the guest has only set up one
> > of the APIC assist pages and yet has taken an interrupt...
> >
> > Jan 16 01:46:05.691223 (XEN) Dumping guest's current state at
> key_handler...
> > Jan 16 01:46:05.691265 (XEN) Size of VMCB = 4096, paddr =
> 000000020f7f7000,
> > vaddr = ffff83020f7f7000
> > Jan 16 01:46:05.699269 (XEN) cr_intercepts = 0xfef3fef3 dr_intercepts =
> > 0xffffffff exception_intercepts = 0x60082
> > Jan 16 01:46:05.707128 (XEN) general1_intercepts = 0xbdc4000f
> > general2_intercepts = 0x2e7f
> > Jan 16 01:46:05.715222 (XEN) iopm_base_pa = 0xdfd71000 msrpm_base_pa
> =
> > 0x20f7f4000 tsc_offset = 0xfffffc36684278c9
> > Jan 16 01:46:05.723116 (XEN) tlb_control = 0 vintr = 0x1020001
> > interrupt_shadow = 0
> > Jan 16 01:46:05.723153 (XEN) eventinj 000000008000002f, valid? 1, ec? 0,
> > type 0, vector 0x2f
> > Jan 16 01:46:05.731141 (XEN) exitcode = 0x64 exitintinfo = 0
> > Jan 16 01:46:05.739123 (XEN) exitinfo1 = 0 exitinfo2 = 0
> > Jan 16 01:46:05.739157 (XEN) np_enable = 0x1 guest_asid = 0x4b49
> > Jan 16 01:46:05.739187 (XEN) virtual vmload/vmsave = 0, virt_ext = 0
> >
> > I'd expect it to have interrupts disabled at this point. Seemingly doesn't
> > repro on Intel h/w (although I was testing with Win7 SP1 rather than RTM)
> so
> > I'll try to find some AMD h/w and try again.
> 
> Well, it looks to be a random problem in the first place, or else we
> would have known about the issue much earlier I think. I.e. I'm
> not sure I see what you take the "AMD only" from here.
> 

Well, I don't see any repro on Intel and the VMCB (rather than VMCS) dump identifies this case as using AMD h/w. I agree that it is random though, so it could indeed be purely coincidental.

> As to interrupt state - isn't it quite normal for an OS to bring up
> the BSP first, enable interrupts on it, and then bring up APs?
> That's how we do it in Xen.

What I meant was that I'd expect the guest to have interrupts disabled whilst poking the MSR to enable APIC assist on that CPU, since enabling APIC assist is clearly going to modify the way in which interrupts are handled. If that's not the case though then I guess that is probably the cause of the issue; I never really considered protecting interrupt handling against APIC assist being enabled on the same CPU.

  Paul

> 
> Jan


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [xen-unstable test] 118078: regressions - FAIL
  2018-01-17  9:37         ` Paul Durrant
@ 2018-01-17 10:17           ` Paul Durrant
  0 siblings, 0 replies; 7+ messages in thread
From: Paul Durrant @ 2018-01-17 10:17 UTC (permalink / raw)
  To: Paul Durrant, 'Jan Beulich'
  Cc: xen-devel, osstest-admin@xenproject.org

> -----Original Message-----
[snip]
> What I meant was that I'd expect the guest to have interrupts disabled whilst
> poking the MSR to enable APIC assist on that CPU, since enabling APIC assist
> is clearly going to modify the way in which interrupts are handled. If that's
> not the case though then I guess that is probably the cause of the issue; I
> never really considered protecting interrupt handling against APIC assist
> being enabled on the same CPU.
> 

I think I've spotted it...

If the guest is indeed setting up APIC assist but not actually using it then I think we get into this situation...

- On return to guest vlapic_has_pending_irq() finds a bit set in the IRR, then vlapic_ack_pending_irq() calls viridian_start_apic_assist() which latches the vector, sets the bit in the ISR and clears it from the IRR.
- The guest then processes the interrupt but EOIs it normally, therefore clearing the bit in the ISR.
- On next return to guest vlapic_has_pending_irq() calls viridian_complete_apic_assist(), which discovers the assist bit still set in the page and therefore leaves the latched vector in place, but finds another bit set in the IRR.
- vlapic_ack_pending_irq() is then called, the ISR is clear, so another call is made to viridian_start_apic_assist() and this then calls domain_crash() because it finds the previously latched vector.

I think the correct solution to this is to call viridian_abort_apic_assist() in vlapic_EOI_set() when the vector is cleared from the ISR.

  Paul
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2018-01-17 10:18 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2018-01-16  8:43 [xen-unstable test] 118078: regressions - FAIL osstest service owner
2018-01-16  8:57 ` Jan Beulich
2018-01-16  9:26   ` Paul Durrant
2018-01-16 17:30     ` Paul Durrant
2018-01-17  8:51       ` Jan Beulich
2018-01-17  9:37         ` Paul Durrant
2018-01-17 10:17           ` Paul Durrant

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).