From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id CFE0BC07E9D for ; Thu, 29 Sep 2022 07:47:20 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235252AbiI2HrT (ORCPT ); Thu, 29 Sep 2022 03:47:19 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:40960 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234940AbiI2HrS (ORCPT ); Thu, 29 Sep 2022 03:47:18 -0400 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id EA60CEEEB4 for ; Thu, 29 Sep 2022 00:47:16 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1664437636; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=PsFKA0JcnshzZDoVR/7R357TCzayyacAiy4vW0aVEaY=; b=YduBBko3FML/DaZT+jos0HbhalfuSUQu+jT1EpYznMDU+fdr0vYTmE0wNsBbtPRAUgNAPI UCTfjZz2VDPpr+2wxh6jrPTYfPzPNhnt/TQEEwUk9tTfM29+SQwAolZmT2dlrn2KTmXc2+ Y/8v3QO4BSlU8r6ZlR9RagJeTb8aoPo= Received: from mail-wr1-f69.google.com (mail-wr1-f69.google.com [209.85.221.69]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_128_GCM_SHA256) id us-mta-44-Q7aLSrYeO32awCH49MemAA-1; Thu, 29 Sep 2022 03:47:14 -0400 X-MC-Unique: Q7aLSrYeO32awCH49MemAA-1 Received: by mail-wr1-f69.google.com with SMTP id u20-20020adfc654000000b0022cc05e9119so188737wrg.16 for ; Thu, 29 Sep 2022 00:47:14 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:x-gm-message-state:from:to:cc:subject:date; bh=PsFKA0JcnshzZDoVR/7R357TCzayyacAiy4vW0aVEaY=; b=tpJe1hLHd3rJEQTijiYca3xXny1Tgiv1OMF1ylAVZSuACukv0CT4Ud+niz7O0/5hug S9pPLlyOuHY+kQDTeCZS/331lVVU7ADyPY8YZQqwc08F0egk5PXVO3lYmXkbfHkhv3SQ lRT2kgpP1VscjXFvFChWFIHK/uP3h/qwmI4ZjV7nDQZ77S/cWS5FGNPmhvhkm3gGe+1i GrR5fVkg/d3Jabpo6bTx/7kXv9kf06BT5C/LADEc4yYYuEHX3ZvE26eiGIbMgKih0E5A DN9SVd/1psbOn/dMQsFsLAZiZsyDEwjxt+MGH7U5bU57vdlUeIaNrKIuC33YqGR6Iqlt hBYg== X-Gm-Message-State: ACrzQf3ophJde+ySb056IpeGUFeeNbU/jEv5Te5h0qhK3Ezj7unWpLVo 3QSx0AiMqHpdtAgeIw+wGiFHHJVN+bBRhH9pEVJLvlUYyjRMJxBQYgmMj3DxUMelb16486O1goa 0DpEEAXRjsyEV4cpv X-Received: by 2002:a5d:4284:0:b0:22a:291e:fa8f with SMTP id k4-20020a5d4284000000b0022a291efa8fmr1154735wrq.553.1664437633412; Thu, 29 Sep 2022 00:47:13 -0700 (PDT) X-Google-Smtp-Source: AMsMyM4+U+FYi8gr6euyVSrck8fFbYRHG3C6n/KBq6oIquAU+3k+fHj1pTJKlSFhr1Cm9nI7F+MMDw== X-Received: by 2002:a5d:4284:0:b0:22a:291e:fa8f with SMTP id k4-20020a5d4284000000b0022a291efa8fmr1154726wrq.553.1664437633210; Thu, 29 Sep 2022 00:47:13 -0700 (PDT) Received: from redhat.com ([2.55.17.78]) by smtp.gmail.com with ESMTPSA id 3-20020a05600c230300b003b4727d199asm3639886wmo.15.2022.09.29.00.47.10 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 29 Sep 2022 00:47:12 -0700 (PDT) Date: Thu, 29 Sep 2022 03:47:08 -0400 From: "Michael S. Tsirkin" To: Stefano Garzarella Cc: Junichi Uekawa , Stefan Hajnoczi , Jason Wang , Eric Dumazet , davem@davemloft.net, netdev@vger.kernel.org, Jakub Kicinski , virtualization@lists.linux-foundation.org, kvm@vger.kernel.org, Paolo Abeni , linux-kernel@vger.kernel.org, Bobby Eshleman Subject: Re: [PATCH] vhost/vsock: Use kvmalloc/kvfree for larger packets. Message-ID: <20220929034552-mutt-send-email-mst@kernel.org> References: <20220928064538.667678-1-uekawa@chromium.org> <20220928082823.wyxplop5wtpuurwo@sgarzare-redhat> <20220928052738-mutt-send-email-mst@kernel.org> <20220928151135.pvrlsylg6j3hzh74@sgarzare-redhat> <20220928160116-mutt-send-email-mst@kernel.org> <20220929074010.37mksjmwr3l4wlwt@sgarzare-redhat> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20220929074010.37mksjmwr3l4wlwt@sgarzare-redhat> Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org On Thu, Sep 29, 2022 at 09:40:10AM +0200, Stefano Garzarella wrote: > On Wed, Sep 28, 2022 at 04:02:12PM -0400, Michael S. Tsirkin wrote: > > On Wed, Sep 28, 2022 at 05:11:35PM +0200, Stefano Garzarella wrote: > > > On Wed, Sep 28, 2022 at 05:31:58AM -0400, Michael S. Tsirkin wrote: > > > > On Wed, Sep 28, 2022 at 10:28:23AM +0200, Stefano Garzarella wrote: > > > > > On Wed, Sep 28, 2022 at 03:45:38PM +0900, Junichi Uekawa wrote: > > > > > > When copying a large file over sftp over vsock, data size is usually 32kB, > > > > > > and kmalloc seems to fail to try to allocate 32 32kB regions. > > > > > > > > > > > > Call Trace: > > > > > > [] dump_stack+0x97/0xdb > > > > > > [] warn_alloc_failed+0x10f/0x138 > > > > > > [] ? __alloc_pages_direct_compact+0x38/0xc8 > > > > > > [] __alloc_pages_nodemask+0x84c/0x90d > > > > > > [] alloc_kmem_pages+0x17/0x19 > > > > > > [] kmalloc_order_trace+0x2b/0xdb > > > > > > [] __kmalloc+0x177/0x1f7 > > > > > > [] ? copy_from_iter+0x8d/0x31d > > > > > > [] vhost_vsock_handle_tx_kick+0x1fa/0x301 [vhost_vsock] > > > > > > [] vhost_worker+0xf7/0x157 [vhost] > > > > > > [] kthread+0xfd/0x105 > > > > > > [] ? vhost_dev_set_owner+0x22e/0x22e [vhost] > > > > > > [] ? flush_kthread_worker+0xf3/0xf3 > > > > > > [] ret_from_fork+0x4e/0x80 > > > > > > [] ? flush_kthread_worker+0xf3/0xf3 > > > > > > > > > > > > Work around by doing kvmalloc instead. > > > > > > > > > > > > Signed-off-by: Junichi Uekawa > > > > > > > > My worry here is that this in more of a work around. > > > > It would be better to not allocate memory so aggressively: > > > > if we are so short on memory we should probably process > > > > packets one at a time. Is that very hard to implement? > > > > > > Currently the "virtio_vsock_pkt" is allocated in the "handle_kick" callback > > > of TX virtqueue. Then the packet is multiplexed on the right socket queue, > > > then the user space can de-queue it whenever they want. > > > > > > So maybe we can stop processing the virtqueue if we are short on memory, but > > > when can we restart the TX virtqueue processing? > > > > Assuming you added at least one buffer, the time to restart would be > > after that buffer has been used. > > Yes, but we still might not have as many continuous pages to allocate, so I > would use kvmalloc the same. you would do something like if (is_vmalloc_addr()) stop adding buffers. > I agree that we should do better, I hope that moving to sk_buff will allow > us to better manage allocation. Maybe after we merge that part we should > spend some time to solve these problems. > > Thanks, > Stefano