From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id AD0FEC433EF for ; Thu, 17 Mar 2022 15:42:43 +0000 (UTC) Received: from localhost ([::1]:40358 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1nUsGw-000118-LJ for qemu-devel@archiver.kernel.org; Thu, 17 Mar 2022 11:42:42 -0400 Received: from eggs.gnu.org ([209.51.188.92]:52360) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1nUsEe-000778-LR for qemu-devel@nongnu.org; Thu, 17 Mar 2022 11:40:20 -0400 Received: from [2607:f8b0:4864:20::1033] (port=54268 helo=mail-pj1-x1033.google.com) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_128_GCM_SHA256:128) (Exim 4.90_1) (envelope-from ) id 1nUsEa-000538-LM for qemu-devel@nongnu.org; Thu, 17 Mar 2022 11:40:19 -0400 Received: by mail-pj1-x1033.google.com with SMTP id bx5so5185043pjb.3 for ; Thu, 17 Mar 2022 08:40:13 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=message-id:date:mime-version:user-agent:subject:content-language:to :cc:references:from:in-reply-to:content-transfer-encoding; bh=TGFZ/3KqPBPXJdG4gwrJAExeOzs1pExm3NzAeNruZ88=; b=BAg7vcJTP4HuXWBFJdDxxQlo/MBDK6m7zTgtYkdMmcUHhmNlP5CURpHg6f2minDF6K ljC/9wghvG6woW3CdTpvXz5kopEc8sd9oYUnElQt8V1kEec56U4vMIVMOETru2mPNk5V CmdST1fpieFKJTa0VXoup25q4FkHN4nmlkl/lnuCxIhYIimROJ3qa8Qn0rlbqTReo4Yt iVbFbTvnHXdtgjfvgITT8gTtFZdg/GTqWW0AUEMAe3Tr9auNKPYbipNSmkDMS+is0CBc 8keoTcHWg7w4uLtKgzQPqrrkUF74FJ5204wtDNj4Wnu8M94R7dlM1q4GcK+pOcBJcwhS yUSA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:message-id:date:mime-version:user-agent:subject :content-language:to:cc:references:from:in-reply-to :content-transfer-encoding; bh=TGFZ/3KqPBPXJdG4gwrJAExeOzs1pExm3NzAeNruZ88=; b=cnucmPter9Bp4dBi62IQ398y44XoT3HEmhHnY8CbAzDL9GDByKKQThu54bhu/m0vc3 OucIPEQ52xyTBcNiKRd8NnwdKJoB0IXRbuPv2zZWgMj152paQH0kKxq/3LJsXh3WfMYh yvc8XzKjvDC2OC12K9VXUl6g43C5VE2b0zqfZ79/o6Dc8kdmu9GtMWxr4CEtlwnAHEkZ VfqxiCvM+SRDNnq6g5bXD+KxFHu2b51mMJWYTCvuj7cOy1qkAcL4TbAyvsMl59wGg/H9 uqz58zyA7Q/CRTWq676IJji1T+nSm9vQdLnPcaMaGtVaTKAdoMeLjfJzeOHBHbofAQUA 4Auw== X-Gm-Message-State: AOAM531rsSPWlEfeod4I013uj6yu1iRbxidGVKWQm0DrGTPiXgFPlTBV kHOn5Gq03EHSxZI9n1fqQDo= X-Google-Smtp-Source: ABdhPJxo06RneRXsJWNlSwzeBHGwZCUrIyCkGjEf5sm4IGfww2bjpTm7bzr5lXjVga7ovJEYEgh2YQ== X-Received: by 2002:a17:90b:3910:b0:1bf:2972:ccea with SMTP id ob16-20020a17090b391000b001bf2972cceamr16400790pjb.221.1647531612143; Thu, 17 Mar 2022 08:40:12 -0700 (PDT) Received: from [192.168.66.3] (p912131-ipoe.ipoe.ocn.ne.jp. [153.243.13.130]) by smtp.gmail.com with ESMTPSA id g1-20020a17090adac100b001c67cedd84esm2684229pjx.42.2022.03.17.08.40.08 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Thu, 17 Mar 2022 08:40:11 -0700 (PDT) Message-ID: <07854f4f-985a-47b3-0a22-f49e9663f451@gmail.com> Date: Fri, 18 Mar 2022 00:40:06 +0900 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux aarch64; rv:91.0) Gecko/20100101 Thunderbird/91.6.2 Subject: Re: [PATCH v21 3/7] net/vmnet: implement shared mode (vmnet-shared) Content-Language: en-US To: Vladislav Yaroshchuk , qemu-devel@nongnu.org References: <20220317102833.20775-1-Vladislav.Yaroshchuk@jetbrains.com> <20220317102833.20775-4-Vladislav.Yaroshchuk@jetbrains.com> From: Akihiko Odaki In-Reply-To: <20220317102833.20775-4-Vladislav.Yaroshchuk@jetbrains.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Host-Lookup-Failed: Reverse DNS lookup failed for 2607:f8b0:4864:20::1033 (failed) Received-SPF: pass client-ip=2607:f8b0:4864:20::1033; envelope-from=akihiko.odaki@gmail.com; helo=mail-pj1-x1033.google.com X-Spam_score_int: -6 X-Spam_score: -0.7 X-Spam_bar: / X-Spam_report: (-0.7 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, FREEMAIL_FROM=0.001, NICE_REPLY_A=-0.001, PDS_HP_HELO_NORDNS=0.659, RCVD_IN_DNSWL_NONE=-0.0001, RDNS_NONE=0.793, SPF_HELO_NONE=0.001, SPF_PASS=-0.001, T_SCC_BODY_TEXT_LINE=-0.01 autolearn=no autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: peter.maydell@linaro.org, alex.bennee@linaro.org, jasowang@redhat.com, phillip.ennen@gmail.com, armbru@redhat.com, dirty@apple.com, f4bug@amsat.org, r.bolshakov@yadro.com, agraf@csgraf.de, phillip@axleos.com, roman@roolebo.dev, hsp.cat7@gmail.com, hello@adns.io, qemu_oss@crudebyte.com, eblake@redhat.com, kraxel@redhat.com Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" On 2022/03/17 19:28, Vladislav Yaroshchuk wrote: > Interaction with vmnet.framework in different modes > differs only on configuration stage, so we can create > common `send`, `receive`, etc. procedures and reuse them. > > Signed-off-by: Phillip Tennen > Signed-off-by: Vladislav Yaroshchuk > --- > net/vmnet-common.m | 358 +++++++++++++++++++++++++++++++++++++++++++++ > net/vmnet-shared.c | 90 +++++++++++- > net/vmnet_int.h | 40 ++++- > 3 files changed, 483 insertions(+), 5 deletions(-) > > diff --git a/net/vmnet-common.m b/net/vmnet-common.m > index 06326efb1c..2cb60b9ddd 100644 > --- a/net/vmnet-common.m > +++ b/net/vmnet-common.m > @@ -10,6 +10,8 @@ > */ > > #include "qemu/osdep.h" > +#include "qemu/main-loop.h" > +#include "qemu/log.h" > #include "qapi/qapi-types-net.h" > #include "vmnet_int.h" > #include "clients.h" > @@ -17,4 +19,360 @@ > #include "qapi/error.h" > > #include > +#include > > + > +static void vmnet_send_completed(NetClientState *nc, ssize_t len); > + > + > +const char *vmnet_status_map_str(vmnet_return_t status) > +{ > + switch (status) { > + case VMNET_SUCCESS: > + return "success"; > + case VMNET_FAILURE: > + return "general failure (possibly not enough privileges)"; > + case VMNET_MEM_FAILURE: > + return "memory allocation failure"; > + case VMNET_INVALID_ARGUMENT: > + return "invalid argument specified"; > + case VMNET_SETUP_INCOMPLETE: > + return "interface setup is not complete"; > + case VMNET_INVALID_ACCESS: > + return "invalid access, permission denied"; > + case VMNET_PACKET_TOO_BIG: > + return "packet size is larger than MTU"; > + case VMNET_BUFFER_EXHAUSTED: > + return "buffers exhausted in kernel"; > + case VMNET_TOO_MANY_PACKETS: > + return "packet count exceeds limit"; > +#if defined(MAC_OS_VERSION_11_0) && \ > + MAC_OS_X_VERSION_MIN_REQUIRED >= MAC_OS_VERSION_11_0 > + case VMNET_SHARING_SERVICE_BUSY: > + return "conflict, sharing service is in use"; > +#endif > + default: > + return "unknown vmnet error"; > + } > +} > + > + > +/** > + * Write packets from QEMU to vmnet interface. > + * > + * vmnet.framework supports iov, but writing more than > + * one iov into vmnet interface fails with > + * 'VMNET_INVALID_ARGUMENT'. Collecting provided iovs into > + * one and passing it to vmnet works fine. That's the > + * reason why receive_iov() left unimplemented. But it still > + * works with good performance having .receive() only. > + */ > +ssize_t vmnet_receive_common(NetClientState *nc, > + const uint8_t *buf, > + size_t size) > +{ > + VmnetState *s = DO_UPCAST(VmnetState, nc, nc); > + struct vmpktdesc packet; > + struct iovec iov; > + int pkt_cnt; > + vmnet_return_t if_status; > + > + if (size > s->max_packet_size) { > + warn_report("vmnet: packet is too big, %zu > %" PRIu64, > + packet.vm_pkt_size, > + s->max_packet_size); > + return -1; > + } > + > + iov.iov_base = (char *) buf; > + iov.iov_len = size; > + > + packet.vm_pkt_iovcnt = 1; > + packet.vm_flags = 0; > + packet.vm_pkt_size = size; > + packet.vm_pkt_iov = &iov; > + pkt_cnt = 1; > + > + if_status = vmnet_write(s->vmnet_if, &packet, &pkt_cnt); > + if (if_status != VMNET_SUCCESS) { > + error_report("vmnet: write error: %s\n", > + vmnet_status_map_str(if_status)); > + return -1; > + } > + > + if (pkt_cnt) { > + return size; > + } > + return 0; > +} > + > + > +/** > + * Read packets from vmnet interface and write them > + * to temporary buffers in VmnetState. > + * > + * Returns read packets number (may be 0) on success, > + * -1 on error > + */ > +static int vmnet_read_packets(VmnetState *s) > +{ > + assert(s->packets_send_current_pos == s->packets_send_end_pos); > + > + struct vmpktdesc *packets = s->packets_buf; > + vmnet_return_t status; > + int i; > + > + /* Read as many packets as present */ > + s->packets_send_current_pos = 0; > + s->packets_send_end_pos = VMNET_PACKETS_LIMIT; > + for (i = 0; i < s->packets_send_end_pos; ++i) { > + packets[i].vm_pkt_size = s->max_packet_size; > + packets[i].vm_pkt_iovcnt = 1; > + packets[i].vm_flags = 0; > + } > + > + status = vmnet_read(s->vmnet_if, packets, &s->packets_send_end_pos); > + if (status != VMNET_SUCCESS) { > + error_printf("vmnet: read failed: %s\n", > + vmnet_status_map_str(status)); > + s->packets_send_current_pos = 0; > + s->packets_send_end_pos = 0; > + return -1; > + } > + return s->packets_send_end_pos; > +} > + > + > +/** > + * Write packets from temporary buffers in VmnetState > + * to QEMU. > + */ > +static void vmnet_write_packets_to_qemu(VmnetState *s) > +{ > + while (s->packets_send_current_pos < s->packets_send_end_pos) { > + ssize_t size = qemu_send_packet_async(&s->nc, > + s->iov_buf[s->packets_send_current_pos].iov_base, > + s->packets_buf[s->packets_send_current_pos].vm_pkt_size, > + vmnet_send_completed); > + > + if (size == 0) { > + /* QEMU is not ready to consume more packets - > + * stop and wait for completion callback call */ > + return; > + } > + ++s->packets_send_current_pos; > + } > +} > + > + > +/** > + * Bottom half callback that transfers packets from vmnet interface > + * to QEMU. > + * > + * The process of transferring packets is three-staged: > + * 1. Handle vmnet event; > + * 2. Read packets from vmnet interface into temporary buffer; > + * 3. Write packets from temporary buffer to QEMU. > + * > + * QEMU may suspend this process on the last stage, returning 0 from > + * qemu_send_packet_async function. If this happens, we should > + * respectfully wait until it is ready to consume more packets, > + * write left ones in temporary buffer and only after this > + * continue reading more packets from vmnet interface. > + * > + * Packets to be transferred are stored into packets_buf, > + * in the window [packets_send_current_pos..packets_send_end_pos) > + * including current_pos, excluding end_pos. > + * > + * Thus, if QEMU is not ready, buffer is not read and > + * packets_send_current_pos < packets_send_end_pos. > + */ > +static void vmnet_send_bh(void *opaque) > +{ > + NetClientState *nc = (NetClientState *) opaque; > + VmnetState *s = DO_UPCAST(VmnetState, nc, nc); > + > + /* > + * Do nothing if QEMU is not ready - wait > + * for completion callback invocation > + */ > + if (s->packets_send_current_pos < s->packets_send_end_pos) { > + return; > + } > + > + /* Read packets from vmnet interface */ > + if (vmnet_read_packets(s) > 0) { > + /* Send them to QEMU */ > + vmnet_write_packets_to_qemu(s); > + } > +} > + > + > +/** > + * Completion callback to be invoked by QEMU when it becomes > + * ready to consume more packets. > + */ > +static void vmnet_send_completed(NetClientState *nc, ssize_t len) > +{ > + VmnetState *s = DO_UPCAST(VmnetState, nc, nc); > + > + /* Callback is invoked eq queued packet is sent */ > + ++s->packets_send_current_pos; > + > + /* Complete sending packets left in VmnetState buffers */ > + vmnet_write_packets_to_qemu(s); > + > + /* And read new ones from vmnet if VmnetState buffer is ready */ > + if (s->packets_send_current_pos < s->packets_send_end_pos) { > + qemu_bh_schedule(s->send_bh); > + } > +} > + > + > +static void vmnet_bufs_init(VmnetState *s) > +{ > + struct vmpktdesc *packets = s->packets_buf; > + struct iovec *iov = s->iov_buf; > + int i; > + > + for (i = 0; i < VMNET_PACKETS_LIMIT; ++i) { > + iov[i].iov_len = s->max_packet_size; > + iov[i].iov_base = g_malloc0(iov[i].iov_len); > + packets[i].vm_pkt_iov = iov + i; > + } > +} > + > + > +int vmnet_if_create(NetClientState *nc, > + xpc_object_t if_desc, > + Error **errp) > +{ > + VmnetState *s = DO_UPCAST(VmnetState, nc, nc); > + dispatch_semaphore_t if_created_sem = dispatch_semaphore_create(0); > + __block vmnet_return_t if_status; > + > + s->if_queue = dispatch_queue_create( > + "org.qemu.vmnet.if_queue", > + DISPATCH_QUEUE_SERIAL > + ); > + > + xpc_dictionary_set_bool( > + if_desc, > + vmnet_allocate_mac_address_key, > + false > + ); > + > +#ifdef DEBUG > + qemu_log("vmnet.start.interface_desc:\n"); > + xpc_dictionary_apply(if_desc, > + ^bool(const char *k, xpc_object_t v) { > + char *desc = xpc_copy_description(v); > + qemu_log(" %s=%s\n", k, desc); > + free(desc); > + return true; > + }); > +#endif /* DEBUG */ > + > + s->vmnet_if = vmnet_start_interface( > + if_desc, > + s->if_queue, > + ^(vmnet_return_t status, xpc_object_t interface_param) { > + if_status = status; > + if (status != VMNET_SUCCESS || !interface_param) { > + dispatch_semaphore_signal(if_created_sem); > + return; > + } > + > +#ifdef DEBUG > + qemu_log("vmnet.start.interface_param:\n"); > + xpc_dictionary_apply(interface_param, > + ^bool(const char *k, xpc_object_t v) { > + char *desc = xpc_copy_description(v); > + qemu_log(" %s=%s\n", k, desc); > + free(desc); > + return true; > + }); > +#endif /* DEBUG */ > + > + s->mtu = xpc_dictionary_get_uint64( > + interface_param, > + vmnet_mtu_key); > + s->max_packet_size = xpc_dictionary_get_uint64( > + interface_param, > + vmnet_max_packet_size_key); > + > + dispatch_semaphore_signal(if_created_sem); > + }); > + > + if (s->vmnet_if == NULL) { > + dispatch_release(s->if_queue); > + dispatch_release(if_created_sem); > + error_setg(errp, > + "unable to create interface with requested params"); > + return -1; > + } > + > + dispatch_semaphore_wait(if_created_sem, DISPATCH_TIME_FOREVER); > + dispatch_release(if_created_sem); > + > + if (if_status != VMNET_SUCCESS) { > + dispatch_release(s->if_queue); > + error_setg(errp, > + "cannot create vmnet interface: %s", > + vmnet_status_map_str(if_status)); > + return -1; > + } > + > + s->send_bh = aio_bh_new(qemu_get_aio_context(), vmnet_send_bh, nc); > + vmnet_bufs_init(s); > + > + s->packets_send_current_pos = 0; > + s->packets_send_end_pos = 0; > + > + vmnet_interface_set_event_callback( > + s->vmnet_if, > + VMNET_INTERFACE_PACKETS_AVAILABLE, > + s->if_queue, > + ^(interface_event_t event_id, xpc_object_t event) { > + assert(event_id == VMNET_INTERFACE_PACKETS_AVAILABLE); > + /* > + * This function is being called from a non qemu thread, so > + * we only schedule a BH, and do the rest of the io completion > + * handling from vmnet_send_bh() which runs in a qemu context. > + */ > + qemu_bh_schedule(s->send_bh); > + }); > + > + return 0; > +} > + > + > +void vmnet_cleanup_common(NetClientState *nc) > +{ > + VmnetState *s = DO_UPCAST(VmnetState, nc, nc); > + dispatch_semaphore_t if_stopped_sem; > + > + if (s->vmnet_if == NULL) { > + return; > + } > + > + if_stopped_sem = dispatch_semaphore_create(0); > + vmnet_stop_interface( > + s->vmnet_if, > + s->if_queue, > + ^(vmnet_return_t status) { > + assert(status == VMNET_SUCCESS); > + dispatch_semaphore_signal(if_stopped_sem); > + }); > + dispatch_semaphore_wait(if_stopped_sem, DISPATCH_TIME_FOREVER); > + > + qemu_purge_queued_packets(nc); > + > + qemu_bh_delete(s->send_bh); > + dispatch_release(if_stopped_sem); > + dispatch_release(s->if_queue); > + > + for (int i = 0; i < VMNET_PACKETS_LIMIT; ++i) { > + g_free(s->iov_buf[i].iov_base); > + } > +} > diff --git a/net/vmnet-shared.c b/net/vmnet-shared.c > index 6dfb133a18..2ce2323bdb 100644 > --- a/net/vmnet-shared.c > +++ b/net/vmnet-shared.c > @@ -10,16 +10,98 @@ > > #include "qemu/osdep.h" > #include "qapi/qapi-types-net.h" > +#include "qapi/error.h" > #include "vmnet_int.h" > #include "clients.h" > -#include "qemu/error-report.h" > -#include "qapi/error.h" > > #include > > + > +static bool validate_options(const Netdev *netdev, Error **errp) > +{ > + const NetdevVmnetSharedOptions *options = &(netdev->u.vmnet_shared); > + > +#if !defined(MAC_OS_VERSION_11_0) || \ > + MAC_OS_X_VERSION_MIN_REQUIRED < MAC_OS_VERSION_11_0 > + if (options->has_isolated) { > + error_setg(errp, > + "vmnet-shared.isolated feature is " > + "unavailable: outdated vmnet.framework API"); > + return false; > + } > +#endif > + > + if ((options->has_start_address || > + options->has_end_address || > + options->has_subnet_mask) && > + !(options->has_start_address && > + options->has_end_address && > + options->has_subnet_mask)) { > + error_setg(errp, > + "'start-address', 'end-address', 'subnet-mask' " > + "should be provided together" > + ); > + return false; > + } > + > + return true; > +} > + > +static xpc_object_t build_if_desc(const Netdev *netdev) > +{ > + const NetdevVmnetSharedOptions *options = &(netdev->u.vmnet_shared); > + xpc_object_t if_desc = xpc_dictionary_create(NULL, NULL, 0); > + > + xpc_dictionary_set_uint64( > + if_desc, > + vmnet_operation_mode_key, > + VMNET_SHARED_MODE > + ); > + > + if (options->has_nat66_prefix) { > + xpc_dictionary_set_string(if_desc, > + vmnet_nat66_prefix_key, > + options->nat66_prefix); > + } > + > + if (options->has_start_address) { > + xpc_dictionary_set_string(if_desc, > + vmnet_start_address_key, > + options->start_address); > + xpc_dictionary_set_string(if_desc, > + vmnet_end_address_key, > + options->end_address); > + xpc_dictionary_set_string(if_desc, > + vmnet_subnet_mask_key, > + options->subnet_mask); > + } > + > +#if defined(MAC_OS_VERSION_11_0) && \ > + MAC_OS_X_VERSION_MIN_REQUIRED >= MAC_OS_VERSION_11_0 > + xpc_dictionary_set_bool( > + if_desc, > + vmnet_enable_isolation_key, > + options->isolated > + ); > +#endif > + > + return if_desc; > +} > + > +static NetClientInfo net_vmnet_shared_info = { > + .type = NET_CLIENT_DRIVER_VMNET_SHARED, > + .size = sizeof(VmnetState), > + .receive = vmnet_receive_common, > + .cleanup = vmnet_cleanup_common, > +}; > + > int net_init_vmnet_shared(const Netdev *netdev, const char *name, > NetClientState *peer, Error **errp) > { > - error_setg(errp, "vmnet-shared is not implemented yet"); > - return -1; > + NetClientState *nc = qemu_new_net_client(&net_vmnet_shared_info, > + peer, "vmnet-shared", name); > + if (!validate_options(netdev, errp)) { > + return -1; > + } > + return vmnet_if_create(nc, build_if_desc(netdev), errp); if_desc apparently leaks. I'm sorry requesting another change again, but please fix it. Regards, Akihiko Odaki > } > diff --git a/net/vmnet_int.h b/net/vmnet_int.h > index c383038a1d..adf6e8c20d 100644 > --- a/net/vmnet_int.h > +++ b/net/vmnet_int.h > @@ -15,11 +15,49 @@ > #include "clients.h" > > #include > +#include > + > +/** > + * From vmnet.framework documentation > + * > + * Each read/write call allows up to 200 packets to be > + * read or written for a maximum of 256KB. > + * > + * Each packet written should be a complete > + * ethernet frame. > + * > + * https://developer.apple.com/documentation/vmnet > + */ > +#define VMNET_PACKETS_LIMIT 200 > > typedef struct VmnetState { > - NetClientState nc; > + NetClientState nc; > + interface_ref vmnet_if; > + > + uint64_t mtu; > + uint64_t max_packet_size; > > + dispatch_queue_t if_queue; > + > + QEMUBH *send_bh; > + > + struct vmpktdesc packets_buf[VMNET_PACKETS_LIMIT]; > + int packets_send_current_pos; > + int packets_send_end_pos; > + > + struct iovec iov_buf[VMNET_PACKETS_LIMIT]; > } VmnetState; > > +const char *vmnet_status_map_str(vmnet_return_t status); > + > +int vmnet_if_create(NetClientState *nc, > + xpc_object_t if_desc, > + Error **errp); > + > +ssize_t vmnet_receive_common(NetClientState *nc, > + const uint8_t *buf, > + size_t size); > + > +void vmnet_cleanup_common(NetClientState *nc); > > #endif /* VMNET_INT_H */