From: Artyom Tarasenko <atar4qemu@googlemail.com>
To: Blue Swirl <blauwirbel@gmail.com>
Cc: qemu-devel <qemu-devel@nongnu.org>
Subject: [Qemu-devel] Re: target-sparc/TODO
Date: Fri, 21 Aug 2009 23:01:39 +0200 [thread overview]
Message-ID: <fb8d4f70908211401k5954e674g8b38b038b3769b4e@mail.gmail.com> (raw)
In-Reply-To: <f43fc5580908211245v74a52b17w77a2ce4aa030ffe9@mail.gmail.com>
2009/8/21 Blue Swirl <blauwirbel@gmail.com>:
> On Fri, Aug 21, 2009 at 3:40 PM, Artyom
> Tarasenko<atar4qemu@googlemail.com> wrote:
>> 2009/8/21 Artyom Tarasenko <atar4qemu@googlemail.com>:
>>> 2009/8/20 Blue Swirl <blauwirbel@gmail.com>:
>>>> On Thu, Aug 20, 2009 at 12:44 PM, Artyom
>>>> Tarasenko<atar4qemu@googlemail.com> wrote:
>>>>>>> Particularly I'm interested if
>>>>>>>
>>>>>>> jmp %l1, %g4, %g0
>>>>>>>
>>>>>>> may behave other than on a real hw.
>>>>>>
>>>>>> No, if rd is %g0, the current PC will not be written anywhere (not by
>>>>>> real HW either).
>>>>>
>>>>> The reason I asked is the two following pieces of code work
>>>>> differently on a real and emulated SS-5. On a real one spacel! does an
>>>>> asi write, and spacel@ does an asi read, and under qemu spacel! seems
>>>>> to do nothing, and spacel@ returns its second parameter multiplied by
>>>>> 4. Both of them don't even try to call an [unimplemented] asi
>>>>> operation, I've runned the tests with mmu and asi debug turned on.
>>>>>
>>>>> Real SS-5:
>>>>>
>>>>> ok 0 0 spacel@ .
>>>>> Data Access Error
>>>>> ok 0 20 spacel@ .
>>>>> 0
>>>>> ok 12345678 0 20 spacel!
>>>>> ok 0 20 spacel@ .
>>>>> 12345678
>>>>> ok
>>>>>
>>>>>
>>>>> qemu SS-5:
>>>>>
>>>>> ok 0 0 spacel@ .
>>>>> 0
>>>>> ok 0 20 spacel@ .
>>>>> 80
>>>>> ok 12345678 0 20 spacel!
>>>>> ok 0 20 spacel@ .
>>>>> 80
>>>>> ok
>>>>>
>>>>> I don't know sparc asm good enogh, but qemu behavior seems to be
>>>>> logical: in the first case I see no store op, and there are shifts
>>>>> which would multiply by 4:
>>>>>
>>>>> ok see spacel!
>>>>> code spacel!
>>>>> ffd26e0c ld [%g7], %l2
>>>>> ffd26e10 add %g7, 4, %g7
>>>>> ffd26e14 ld [%g7], %l0
>>>>> ffd26e18 add %g7, 4, %g7
>>>>> ffd26e1c sll %g4, 2, %g4
>>>>> ffd26e20 call ffd26e24
>>>>> ffd26e24 add %g0, 14, %l1
>>>>>
>>>>> ok ffd26e24 dis
>>>>> ffd26e24 add %g0, 14, %l1
>>>>> ffd26e28 add %o7, %l1, %l1
>>>>> ffd26e2c jmp %l1, %g4, %g0
>>>>> ffd26e30 ba ffd26f68
>>>>> ok
>>>>>
>>>>> ok see spacel@
>>>>> code spacel@
>>>>> ffd26830 ld [%g7], %l0
>>>>> ffd26834 add %g7, 4, %g7
>>>>> ffd26838 sll %g4, 2, %g4
>>>>> ffd2683c call ffd26840
>>>>> ffd26840 add %g0, 14, %l1
>>>>>
>>>>> ok ffd26840 dis
>>>>> ffd26840 add %g0, 14, %l1
>>>>> ffd26844 add %o7, %l1, %l1
>>>>> ffd26848 jmp %l1, %g4, %g0
>>>>> ffd2684c ba ffd26984
>>>>>
>>>>>
>>>>> The code is identical on a real and emulated SS.
>>>>>
>>>>> It must be the jump, which jumps differently on a real hw and under
>>>>> qemu. Do you see from the code where the jump would jump to, or maybe
>>>>> you have a suggestion how to check where the jump jumps to on the real
>>>>> hw?
>>>>
>>>> The target of the call instruction is also a delay slot instruction
>>>> for the call itself. Maybe this case is not handled correctly?
>>>
>>> Good idea! Don't know how to test it though.
>>>
>>> And what about "ba" in the delay slot of "jmp"? Is the correct
>>> behavior described somewhere? Would jump just be ignored? Whould it
>>> execute one instruction on jump destination and then branch? Would
>>> branch be ignored?
>>
>> Page 55 of The SPARC v8 Architecture Manual
>> (http://www.sparc.org/standards/V8.pdf) describes this case
>> explicitly:
>> cpu should execute one instruction on the jump target and then branch.
>> Is it what qemu currently does?
>
> I may be blind, I don't see the description of this case in that page.
I wasn't referring the call case, but jmp+ba case (two last ops in the
listing above). This DCTI is described on pages marked 55-56 (pages
54-54 in a pdf reader). That's the first case in the table 5-12.
> Both QEMU and real (Sparc64) hardware exit with return value of 3, so
> the inc is re-executed. If I add a nop in the call delay slot, the
> return value is 2.
Can you make a similar test, but with ba in the jmp's delay slot?
next prev parent reply other threads:[~2009-08-21 21:02 UTC|newest]
Thread overview: 18+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-08-17 10:52 [Qemu-devel] target-sparc/TODO Artyom Tarasenko
2009-08-17 17:35 ` [Qemu-devel] target-sparc/TODO Blue Swirl
2009-08-19 10:17 ` Artyom Tarasenko
2009-08-19 16:43 ` Blue Swirl
2009-08-20 9:44 ` Artyom Tarasenko
2009-08-20 19:15 ` Blue Swirl
2009-08-21 9:58 ` Artyom Tarasenko
2009-08-21 12:40 ` Artyom Tarasenko
2009-08-21 19:45 ` Blue Swirl
2009-08-21 21:01 ` Artyom Tarasenko [this message]
2009-08-21 21:10 ` Igor Kovalenko
2009-08-21 21:17 ` Artyom Tarasenko
2009-08-22 6:51 ` Blue Swirl
2009-08-22 12:40 ` Artyom Tarasenko
2009-08-22 13:30 ` Robert Reif
2009-08-22 17:25 ` Artyom Tarasenko
2009-08-22 18:46 ` Robert Reif
-- strict thread matches above, loose matches on Subject: below --
2010-08-20 19:59 [Qemu-devel] target-sparc/TODO Artyom Tarasenko
2010-08-20 20:19 ` [Qemu-devel] target-sparc/TODO Blue Swirl
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=fb8d4f70908211401k5954e674g8b38b038b3769b4e@mail.gmail.com \
--to=atar4qemu@googlemail.com \
--cc=blauwirbel@gmail.com \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).