From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754067Ab3LPNo4 (ORCPT ); Mon, 16 Dec 2013 08:44:56 -0500 Received: from mail-ea0-f169.google.com ([209.85.215.169]:42905 "EHLO mail-ea0-f169.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753781Ab3LPNoz (ORCPT ); Mon, 16 Dec 2013 08:44:55 -0500 Date: Mon, 16 Dec 2013 14:44:49 +0100 From: Ingo Molnar To: Mel Gorman Cc: Linus Torvalds , Alex Shi , Thomas Gleixner , Andrew Morton , Fengguang Wu , H Peter Anvin , Linux-X86 , Linux-MM , LKML , Peter Zijlstra Subject: Re: [PATCH 0/4] Fix ebizzy performance regression due to X86 TLB range flush v2 Message-ID: <20131216134449.GA3034@gmail.com> References: <1386964870-6690-1-git-send-email-mgorman@suse.de> <20131215155539.GM11295@suse.de> <20131216102439.GA21624@gmail.com> <20131216125923.GS11295@suse.de> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20131216125923.GS11295@suse.de> User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org * Mel Gorman wrote: > > Whatever we did right in v3.4 we want to do in v3.13 as well - or > > at least understand it. > > Also agreed. I started a bisection before answering this mail. It > would be cooler and potentially faster to figure it out from direct > analysis but bisection is reliable and less guesswork. Trying to guess can potentially last a _lot_ longer than a generic, no-assumptions bisection ... The symptoms could point to anything: scheduler, locking details, some stupid little change in a wakeup sequence somewhere, etc. It might even be a non-deterministic effect of some timing change causing the workload 'just' to avoid a common point of preemption and not scheduling as much - and become more unfair and thus certain threads lasting longer to finish. Does the benchmark execute a fixed amount of transactions per thread? That might artificially increase the numeric regression: with more threads it 'magnifies' any unfairness effects because slower threads will become slower, faster threads will become faster, as the thread count increases. [ That in itself is somewhat artificial, because real workloads tend to balance between threads dynamically and don't insist on keeping the fastest threads idle near the end of a run. It does not invalidate the complaint about the unfairness itself, obviously. ] Ingo