From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from eggs.gnu.org ([2001:4830:134:3::10]:59517) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1XVzVC-00029m-LU for qemu-devel@nongnu.org; Mon, 22 Sep 2014 05:01:51 -0400 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1XVzV8-000640-BV for qemu-devel@nongnu.org; Mon, 22 Sep 2014 05:01:46 -0400 Received: from [59.151.112.132] (port=36797 helo=heian.cn.fujitsu.com) by eggs.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1XVzV8-00060G-0O for qemu-devel@nongnu.org; Mon, 22 Sep 2014 05:01:42 -0400 Message-ID: <541FE5D0.9050705@cn.fujitsu.com> Date: Mon, 22 Sep 2014 17:03:12 +0800 From: Tang Chen MIME-Version: 1.0 References: <1410863956-11056-1-git-send-email-zhang.zhanghailiang@huawei.com> <1410863956-11056-2-git-send-email-zhang.zhanghailiang@huawei.com> <20140917083220.GA17032@G08FNSTD100614.fnst.cn.fujitsu.com> <20140919142642.7d7fbd3d@nial.usersys.redhat.com> In-Reply-To: <20140919142642.7d7fbd3d@nial.usersys.redhat.com> Content-Type: text/plain; charset="ISO-8859-1"; format=flowed Content-Transfer-Encoding: 7bit Subject: Re: [Qemu-devel] [PATCH 1/2] pc-dimm: No numa option shouldn't break hotplug memory feature List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: Igor Mammedov , Hu Tao Cc: zhanghailiang , mst@redhat.com, luonengjun@huawei.com, qemu-devel@nongnu.org, peter.huangpeng@huawei.com, pbonzini@redhat.com, gaowanlong@cn.fujitsu.com Hi Igor, On 09/19/2014 08:26 PM, Igor Mammedov wrote: > On Wed, 17 Sep 2014 16:32:20 +0800 > Hu Tao wrote: > >> On Tue, Sep 16, 2014 at 06:39:15PM +0800, zhanghailiang wrote: >>> If we do not configure numa option, memory hotplug should work as well. >>> It should not depend on numa option. >>> >>> Steps to reproduce: >>> (1) Start VM: qemu-kvm -m 1024,slots=4,maxmem=8G >>> (2) Hotplug memory >>> It will fail and reports: >>> "'DIMM property node has value 0' which exceeds the number of numa nodes: 0" >>> >> I rememberd Tang Chen had a patch for this bug, this is what Andrey suggested: >> >> I thnk that there should be no >> cases when dimm is plugged (and check from patch is fired up) without >> actually populated NUMA, because not every OS will workaround this by >> faking the node. > This doesn't take in to account that dimm device by itself has nothing to do > with numa (numa is just optional property of its representation in ACPI land > and nothing else). > > In case initial memory is converted to dimm devices, qemu can be > started without numa option and it still must work. > > So I'm in favor of this path. I just did some tests. Even if I modify qemu code and make hotpluggable bit in SRAT 0, memory hotplug will still work on Linux guest, which means Linux kernel doesn't check SRAT info after system is up when doing memory hotplug. I did the following modification in hw/i386/acpi-build.c - ram_addr_t hotplugabble_address_space_size = - object_property_get_int(OBJECT(pcms), PC_MACHINE_MEMHP_REGION_SIZE, - NULL); + ram_addr_t hotplugabble_address_space_size = 0; And when the guest is up, no memory should be hotpluggable, I think. But I hot-added memory successfully. IMHO, I think memory hotplug should based on ACPI on Linux. And SRAT tells system which memory ranges are hotpluggable, and we should follow it. So I think Linux kernel has some problem in this issue. I'd like to fix it like this: 1. Send patches to make Linux kernel to check SRAT info when doing memory hotplug. (Now, SRAT is only checked at boot time.) 2. In qemu, when users gave a memory hotplug option without NUMA options, we create node0 and node1, and make node1 hotpluggable. This is because when using MOVABLE_NODE, node0 in which the kernel resides in should not be hotpluggable. Of course, make part of memory in node0 hotpluggable is OK, but on a real machine, no one will do this, I think. So I suggest above idea. Thanks. :) > >> https://lists.nongnu.org/archive/html/qemu-devel/2014-08/msg04587.html >> >> Have you tested this patch with Windows guest? >> >> Regards, >> Hu > > . >