From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-19.0 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 476E2C433ED for ; Thu, 22 Apr 2021 12:02:33 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 1125F6145C for ; Thu, 22 Apr 2021 12:02:33 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236298AbhDVMDC (ORCPT ); Thu, 22 Apr 2021 08:03:02 -0400 Received: from mail.kernel.org ([198.145.29.99]:56592 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S236214AbhDVMC5 (ORCPT ); Thu, 22 Apr 2021 08:02:57 -0400 Received: by mail.kernel.org (Postfix) with ESMTPSA id 0E2AB6145C; Thu, 22 Apr 2021 12:02:20 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1619092942; bh=RiX3HZWIh+Y7OR8HTcVGUXmwRu2q2Kly3DM0/c5TVQ4=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=FfsX73ZHsHLaHTxIK7g/GsBCQlnqkg/s387ovSX/lSPSd+DkTNWGp4E5+5dRLshyJ vSP0W5Xn3IR9EhXw7AMubWZtw927wCaHt9fCRl6dUio4ZlIy+bC/t7SQE5J0/GIfQC WOGKKwjLUt0nje0f+zLv8zfH+7PfLrVOtW6kBnspjEzfCJZ+Rg8YtZPgAhOhdWX2hc PC6nF28LqXtrkKGzE72GweIq6JyOSzefVmlj2Nf4ncSCNnHum1gU1oqCtNzKXVDxOo SIcGTLLW02VVhCFqi4cnhvsqKmn5M0/AR4fQ2SNSKqGNXUI4fqklzFRr090mPl/aX+ IhwD20BD7MXYQ== From: Frederic Weisbecker To: Peter Zijlstra , Thomas Gleixner Cc: LKML , Marcelo Tosatti , "Rafael J . Wysocki" , Yunfeng Ye , Frederic Weisbecker Subject: [PATCH 8/8] tick/nohz: Kick only _queued_ task whose tick dependency is updated Date: Thu, 22 Apr 2021 14:01:58 +0200 Message-Id: <20210422120158.33629-9-frederic@kernel.org> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20210422120158.33629-1-frederic@kernel.org> References: <20210422120158.33629-1-frederic@kernel.org> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Marcelo Tosatti When the tick dependency of a task is updated, we want it to aknowledge the new state and restart the tick if needed. If the task is not running, we don't need to kick it because it will observe the new dependency upon scheduling in. But if the task is running, we may need to send an IPI to it so that it gets notified. Unfortunately we don't have the means to check if a task is running in a race free way. Checking p->on_cpu in a synchronized way against p->tick_dep_mask would imply adding a full barrier between prepare_task_switch() and tick_nohz_task_switch(), which we want to avoid in this fast-path. Therefore we blindly fire an IPI to the task's CPU. Meanwhile we can check if the task is queued on the CPU rq because p->on_rq is always set to TASK_ON_RQ_QUEUED _before_ schedule() and its full barrier that precedes tick_nohz_task_switch(). And if the task is queued on a nohz_full CPU, it also has fair chances to be running as the isolation constraints prescribe running single tasks on full dynticks CPUs. So use this as a trick to check if we can spare an IPI toward a non-running task. NOTE: For the ordering to be correct, it is assumed that we never deactivate a task while it is running, the only exception being the task deactivating itself while scheduling out. Suggested-by: Peter Zijlstra Signed-off-by: Marcelo Tosatti Cc: Yunfeng Ye Cc: Thomas Gleixner Cc: Rafael J. Wysocki Signed-off-by: Frederic Weisbecker --- include/linux/sched.h | 2 ++ kernel/sched/core.c | 5 +++++ kernel/time/tick-sched.c | 19 +++++++++++++++++-- 3 files changed, 24 insertions(+), 2 deletions(-) diff --git a/include/linux/sched.h b/include/linux/sched.h index ef00bb22164c..64dd6f698f3a 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -1999,6 +1999,8 @@ static inline void set_task_cpu(struct task_struct *p, unsigned int cpu) #endif /* CONFIG_SMP */ +extern bool sched_task_on_rq(struct task_struct *p); + /* * In order to reduce various lock holder preemption latencies provide an * interface to see if a vCPU is currently running or not. diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 98191218d891..08526227d200 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -1580,6 +1580,11 @@ static inline void uclamp_post_fork(struct task_struct *p) { } static inline void init_uclamp(void) { } #endif /* CONFIG_UCLAMP_TASK */ +bool sched_task_on_rq(struct task_struct *p) +{ + return task_on_rq_queued(p); +} + static inline void enqueue_task(struct rq *rq, struct task_struct *p, int flags) { if (!(flags & ENQUEUE_NOCLOCK)) diff --git a/kernel/time/tick-sched.c b/kernel/time/tick-sched.c index ad5c3905196a..faba7881048f 100644 --- a/kernel/time/tick-sched.c +++ b/kernel/time/tick-sched.c @@ -324,8 +324,6 @@ void tick_nohz_full_kick_cpu(int cpu) static void tick_nohz_kick_task(struct task_struct *tsk) { - int cpu = task_cpu(tsk); - /* * If the task concurrently migrates to another cpu, * we guarantee it sees the new tick dependency upon @@ -340,6 +338,23 @@ static void tick_nohz_kick_task(struct task_struct *tsk) * tick_nohz_task_switch() smp_mb() (atomic_fetch_or()) * LOAD p->tick_dep_mask LOAD p->cpu */ + int cpu = task_cpu(tsk); + + /* + * If the task is not running, run_posix_cpu_timers + * has nothing to elapsed, can spare IPI in that + * case. + * + * activate_task() STORE p->tick_dep_mask + * STORE p->on_rq + * __schedule() (switch to task 'p') smp_mb() (atomic_fetch_or()) + * LOCK rq->lock LOAD p->on_rq + * smp_mb__after_spin_lock() + * tick_nohz_task_switch() + * LOAD p->tick_dep_mask + */ + if (!sched_task_on_rq(tsk)) + return; preempt_disable(); if (cpu_online(cpu)) -- 2.25.1