From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-11.0 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 222CBC433E1 for ; Wed, 24 Mar 2021 12:20:28 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id E36F461A1F for ; Wed, 24 Mar 2021 12:20:27 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233243AbhCXMT5 (ORCPT ); Wed, 24 Mar 2021 08:19:57 -0400 Received: from us-smtp-delivery-124.mimecast.com ([170.10.133.124]:38031 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232459AbhCXMTo (ORCPT ); Wed, 24 Mar 2021 08:19:44 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1616588384; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding; bh=szHor9C63JSP474Bo4aHGwNm4F6F4NN93pYHW51drAQ=; b=FflHsFZTFWUmTC7TVxxT2Z/JJL5CQ/woaascizg5aBjERzuM2nBj4cX/Td48+1BHsxuvbU OQm4bLMVB2Wa8ZPj2tEJCk5P+TD9H0FrkPSowftm2NWIPmykaCuCA5QMKcFDHo/fblZmBc u5kp6ZA/8Oq7lIbIwDkG0k4uX6GDsy0= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-120-C7lG3O_sPhm9Ujps0gVS6Q-1; Wed, 24 Mar 2021 08:19:42 -0400 X-MC-Unique: C7lG3O_sPhm9Ujps0gVS6Q-1 Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.phx2.redhat.com [10.5.11.22]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 45A1A81624; Wed, 24 Mar 2021 12:19:41 +0000 (UTC) Received: from localhost (ovpn-13-127.pek2.redhat.com [10.72.13.127]) by smtp.corp.redhat.com (Postfix) with ESMTP id D932D1001281; Wed, 24 Mar 2021 12:19:32 +0000 (UTC) From: Ming Lei To: Jens Axboe Cc: linux-block@vger.kernel.org, Jeffle Xu , Mike Snitzer , dm-devel@redhat.com, Ming Lei Subject: [PATCH V3 00/13] block: support bio based io polling Date: Wed, 24 Mar 2021 20:19:14 +0800 Message-Id: <20210324121927.362525-1-ming.lei@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Scanned-By: MIMEDefang 2.84 on 10.5.11.22 Precedence: bulk List-ID: X-Mailing-List: linux-block@vger.kernel.org Hi, Add per-task io poll context for holding HIPRI blk-mq/underlying bios queued from bio based driver's io submission context, and reuse one bio padding field for storing 'cookie' returned from submit_bio() for these bios. Also explicitly end these bios in poll context by adding two new bio flags. In this way, we needn't to poll all underlying hw queues any more, which is implemented in Jeffle's patches. And we can just poll hw queues in which there is HIPRI IO queued. Usually io submission and io poll share same context, so the added io poll context data is just like one stack variable, and the cost for saving bios is cheap. Any comments are welcome. V3: - fix cookie returned for bio based driver, as suggested by Jeffle Xu - draining pending bios when submission context is exiting - patch style and comment fix, as suggested by Mike - allow poll context data to be NULL by always polling on submission queue - remove RFC, and reviewed-by V2: - address queue depth scalability issue reported by Jeffle via bio group list. Reuse .bi_end_io for linking bios which share same .bi_end_io, and support 32 such groups in submit queue. With this way, the scalability issue caused by kfifio is solved. Before really ending bio, .bi_end_io is recovered from the group head. Jeffle Xu (4): block/mq: extract one helper function polling hw queue block: add queue_to_disk() to get gendisk from request_queue block: add poll_capable method to support bio-based IO polling dm: support IO polling for bio-based dm device Ming Lei (9): block: add helper of blk_queue_poll block: add one helper to free io_context block: add helper of blk_create_io_context block: create io poll context for submission and poll task block: add req flag of REQ_POLL_CTX block: add new field into 'struct bvec_iter' block: prepare for supporting bio_list via other link block: use per-task poll context to implement bio based io polling blk-mq: limit hw queues to be polled in each blk_poll() block/bio.c | 5 + block/blk-core.c | 251 ++++++++++++++++++++++++++-- block/blk-ioc.c | 14 +- block/blk-mq.c | 300 +++++++++++++++++++++++++++++++++- block/blk-sysfs.c | 14 +- block/blk.h | 65 ++++++++ drivers/md/dm-table.c | 24 +++ drivers/md/dm.c | 14 ++ drivers/nvme/host/core.c | 2 +- include/linux/bio.h | 132 +++++++-------- include/linux/blk_types.h | 22 ++- include/linux/blkdev.h | 4 + include/linux/bvec.h | 8 + include/linux/device-mapper.h | 1 + include/linux/iocontext.h | 2 + include/trace/events/kyber.h | 6 +- 16 files changed, 770 insertions(+), 94 deletions(-) -- 2.29.2