From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from eggs.gnu.org ([2001:4830:134:3::10]:58760) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1ZoBZq-0006Wy-GN for qemu-devel@nongnu.org; Mon, 19 Oct 2015 10:38:23 -0400 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1ZoBZl-0001NT-Au for qemu-devel@nongnu.org; Mon, 19 Oct 2015 10:38:18 -0400 Received: from smtp.aimale.com ([166.78.138.199]:48243) by eggs.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1ZoBZl-0001HO-39 for qemu-devel@nongnu.org; Mon, 19 Oct 2015 10:38:13 -0400 References: <1444952643-5033-1-git-send-email-valerio@aimale.com> <87h9lrkz56.fsf@blackfin.pond.sub.org> <56210A17.6080401@aimale.com> <87io63xpke.fsf@blackfin.pond.sub.org> From: Valerio Aimale Message-ID: <56250035.40805@aimale.com> Date: Mon, 19 Oct 2015 08:37:41 -0600 MIME-Version: 1.0 In-Reply-To: <87io63xpke.fsf@blackfin.pond.sub.org> Content-Type: text/plain; charset=windows-1252; format=flowed Content-Transfer-Encoding: 7bit Subject: Re: [Qemu-devel] QEMU patch to allow VM introspection via libvmi List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: Markus Armbruster Cc: qemu-devel@nongnu.org, ehabkost@redhat.com, lcapitulino@redhat.com On 10/19/15 1:52 AM, Markus Armbruster wrote: > Valerio Aimale writes: > >> On 10/16/15 2:15 AM, Markus Armbruster wrote: >>> valerio@aimale.com writes: >>> >>>> All- >>>> >>>> I've produced a patch for the current QEMU HEAD, for libvmi to >>>> introspect QEMU/KVM VMs. >>>> >>>> Libvmi has patches for the old qeum-kvm fork, inside its source tree: >>>> https://github.com/libvmi/libvmi/tree/master/tools/qemu-kvm-patch >>>> >>>> This patch adds a hmp and a qmp command, "pmemaccess". When the >>>> commands is invoked with a string arguments (a filename), it will open >>>> a UNIX socket and spawn a listening thread. >>>> >>>> The client writes binary commands to the socket, in the form of a c >>>> structure: >>>> >>>> struct request { >>>> uint8_t type; // 0 quit, 1 read, 2 write, ... rest reserved >>>> uint64_t address; // address to read from OR write to >>>> uint64_t length; // number of bytes to read OR write >>>> }; >>>> >>>> The client receives as a response, either (length+1) bytes, if it is a >>>> read operation, or 1 byte ifit is a write operation. >>>> >>>> The last bytes of a read operation response indicates success (1 >>>> success, 0 failure). The single byte returned for a write operation >>>> indicates same (1 success, 0 failure). >>> So, if you ask to read 1 MiB, and it fails, you get back 1 MiB of >>> garbage followed by the "it failed" byte? >> Markus, that appear to be the case. However, I did not write the >> communication protocol between libvmi and qemu. I'm assuming that the >> person that wrote the protocol, did not want to bother with over >> complicating things. >> >> https://github.com/libvmi/libvmi/blob/master/libvmi/driver/kvm/kvm.c >> >> I'm thinking he assumed reads would be small in size and the price of >> reading garbage was less than the price of writing a more complicated >> protocol. I can see his point, confronted with the same problem, I >> might have done the same. > All right, the interface is designed for *small* memory blocks then. > > Makes me wonder why he needs a separate binary protocol on a separate > socket. Small blocks could be done just fine in QMP. The problem is speed. if one's analyzing the memory space of a running process (physical and paged), libvmi will make a large number of small and mid-sized reads. If one uses xp, or pmemsave, the overhead is quite significant. xp has overhead due to encoding, and pmemsave has overhead due to file open/write (server), file open/read/close/unlink (client). Others have gone through the problem before me. It appears that pmemsave and xp are significantly slower than reading memory using a socket via pmemaccess. The following data is not mine, but it shows the time, in milliseconds, required to resolve the content of a paged memory address via socket (pmemaccess) , pmemsave and xp http://cl.ly/image/322a3s0h1V05 Again, I did not produce those data points, they come from an old libvmi thread. I think it might be conceivable that there could be a QMP command that returns the content of an arbitrarily size memory region as a base64 or a base85 json string. It would still have both time- (due to encoding/decoding) and space- (base64 has 33% and ase85 would be 7%) overhead, + json encoding/decoding overhead. It might still be the case that socket would outperform such a command as well, speed-vise. I don't think it would be any faster than xp. There's also a similar patch, floating around the internet, the uses shared memory, instead of sockets, as inter-process communication between libvmi and QEMU. I've never used that. > >>>> The socket API was written by the libvmi author and it works the with >>>> current libvmi version. The libvmi client-side implementation is at: >>>> >>>> https://github.com/libvmi/libvmi/blob/master/libvmi/driver/kvm/kvm.c >>>> >>>> As many use kvm VM's for introspection, malware and security analysis, >>>> it might be worth thinking about making the pmemaccess a permanent >>>> hmp/qmp command, as opposed to having to produce a patch at each QEMU >>>> point release. >>> Related existing commands: memsave, pmemsave, dump-guest-memory. >>> >>> Can you explain why these won't do for your use case? >> For people who do security analysis there are two use cases, static >> and dynamic analysis. With memsave, pmemsave and dum-guest-memory one >> can do static analysis. I.e. snapshotting a VM and see what was >> happening at that point in time. >> Dynamic analysis require to be able to 'introspect' a VM while it's running. >> >> If you take a snapshot of two people exchanging a glass of water, and >> you happen to take it at the very moment both persons have their hands >> on the glass, it's hard to tell who passed the glass to whom. If you >> have a movie of the same scene, it's obvious who's the giver and who's >> the receiver. Same use case. > I understand the need for introspecting a running guest. What exactly > makes the existing commands unsuitable for that? Speed. See discussion above. > >> More to the point, there's a host of C and python frameworks to >> dynamically analyze VMs: volatility, rekal, "drakvuf", etc. They all >> build on top of libvmi. I did not want to reinvent the wheel. > Fair enough. > > Front page http://libvmi.com/ claims "Works with Xen, KVM, Qemu, and Raw > memory files." What exactly is missing for KVM? When they say they support kvm, what they really mean they support the (retired, I understand) qemu-kvm fork via a patch that is provided in the libvmi source tree. I think the most recent qem-kvm supported is 1.6.0 https://github.com/libvmi/libvmi/tree/master/tools/qemu-kvm-patch I wanted to bring support to the head revision of QEMU, to bring libvmi level with more modern QEMU. Maybe the solution is simply to put this patch in the libvmi source tree, which I've already asked to do via pull request, leaving QEMU alone. However, the patch has to be updated at every QEMU point release. I wanted to avoid that, if at all possible. > >> Mind you, 99.9% of people that do dynamic VM analysis use xen. They >> contend that xen has better introspection support. In my case, I did >> not want to bother with dedicating a full server to be a xen domain >> 0. I just wanted to do a quick test by standing up a QEMU/kvm VM, in >> an otherwise purposed server. > I'm not at all against better introspection support in QEMU. I'm just > trying to understand the problem you're trying to solve with your > patches. What all users of libvmi would love to have is super high speed access to VM physical memory as part of the QEMU source tree, and not supported via a patch. Implemented as the QEMU owners see fit, as long as it is blazing fast and easy accessed via client library or inter-process communication. My gut feeling is that it has to bypass QMP protocol/encoding/file access/json to be fast, but, it is just a gut feeling - worth nothing. > >>>> Also, the pmemsave commands QAPI should be changed to be usable with >>>> 64bit VM's >>>> >>>> in qapi-schema.json >>>> >>>> from >>>> >>>> --- >>>> { 'command': 'pmemsave', >>>> 'data': {'val': 'int', 'size': 'int', 'filename': 'str'} } >>>> --- >>>> >>>> to >>>> >>>> --- >>>> { 'command': 'pmemsave', >>>> 'data': {'val': 'int64', 'size': 'int64', 'filename': 'str'} } >>>> --- >>> In the QAPI schema, 'int' is actually an alias for 'int64'. Yes, that's >>> confusing. >> I think it's confusing for the HMP parser too. If you have a VM with >> 8Gb of RAM and want to snapshot the whole physical memory, via HMP >> over telnet this is what happens: >> >> $ telnet localhost 1234 >> Trying 127.0.0.1... >> Connected to localhost. >> Escape character is '^]'. >> QEMU 2.4.0.1 monitor - type 'help' for more information >> (qemu) help pmemsave >> pmemsave addr size file -- save to disk physical memory dump starting >> at 'addr' of size 'size' >> (qemu) pmemsave 0 8589934591 "/tmp/memorydump" >> 'pmemsave' has failed: integer is for 32-bit values >> Try "help pmemsave" for more information >> (qemu) quit > Your change to pmemsave's definition in qapi-schema.json is effectively a > no-op. > > Your example shows *HMP* command pmemsave. The definition of an HMP > command is *independent* of the QMP command. The implementation *uses* > the QMP command. > > QMP pmemsave is defined in qapi-schema.json as > > { 'command': 'pmemsave', > 'data': {'val': 'int', 'size': 'int', 'filename': 'str'} } > > Its implementation is in cpus.c: > > void qmp_pmemsave(int64_t addr, int64_t size, const char *filename, > Error **errp) > > Note the int64_t size. > > HMP pmemsave is defined in hmp-commands.hx as > > { > .name = "pmemsave", > .args_type = "val:l,size:i,filename:s", > .params = "addr size file", > .help = "save to disk physical memory dump starting at 'addr' of size 'size'", > .mhandler.cmd = hmp_pmemsave, > }, > > Its implementation is in hmp.c: > > void hmp_pmemsave(Monitor *mon, const QDict *qdict) > { > uint32_t size = qdict_get_int(qdict, "size"); > const char *filename = qdict_get_str(qdict, "filename"); > uint64_t addr = qdict_get_int(qdict, "val"); > Error *err = NULL; > > qmp_pmemsave(addr, size, filename, &err); > hmp_handle_error(mon, &err); > } > > Note uint32_t size. > > Arguably, the QMP size argument should use 'size' (an alias for > 'uint64'), and the HMP args_type should use 'size:o'. Understand all that. Indeed, I've re-implemented 'pmemaccess' the same way pmemsave is implemented. There is a single function, and two points of entrance, one for HMP and one for QMP. I think pmemacess mimics pmemsave closely. However, if one wants to simply dump a memory region, via HMP for human easy of use/debug/testing purposes, one cannot dump memory regions that resides higher than 2^32-1 >> With the changes I suggested, the command succeeds >> >> $ telnet localhost 1234 >> Trying 127.0.0.1... >> Connected to localhost. >> Escape character is '^]'. >> QEMU 2.4.0.1 monitor - type 'help' for more information >> (qemu) help pmemsave >> pmemsave addr size file -- save to disk physical memory dump starting >> at 'addr' of size 'size' >> (qemu) pmemsave 0 8589934591 "/tmp/memorydump" >> (qemu) quit >> >> However I just noticed that the dump is just about 4GB in size, so >> there might be more changes needed to snapshot all physical memory of >> a 64 but VM. I did not investigate any further. >> >> ls -l /tmp/memorydump >> -rw-rw-r-- 1 libvirt-qemu kvm 4294967295 Oct 16 08:04 /tmp/memorydump >> >>>> hmp-commands.hx and qmp-commands.hx should be edited accordingly. I >>>> did not make the above pmemsave changes part of my patch. >>>> >>>> Let me know if you have any questions, >>>> >>>> Valerio