public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Paul Gortmaker <p_gortmaker@yahoo.com>
To: Linus Torvalds <torvalds@transmeta.com>
Cc: linux-kernel@vger.kernel.org
Subject: [PATCH] Speedup SMP kernel on UP box
Date: Sat, 16 Mar 2002 06:51:35 -0500	[thread overview]
Message-ID: <3C9331C7.5BDDE3BA@yahoo.com> (raw)

[From the Futexes thread...]

Linus wrote:

> UP vs SMP - a UP system still wants to do locking, but it doesn't need the
> lock prefix. And that lock prefix makes a _huge_ difference
> performance-wise.

> The point being that the difference between a "decl" and a "lock ;  decl"
> is about 1:12 or so in performance.

> On SMP, it's a memory barrier. This is why a "lock ; decl" is more
> expensive than a "decl" - it's the implied memory ordering constraints (on
> other architectures they are explicit). On an intel CPU, this basically
> means that the pipeline is drained, so a locked instruction takes roughly
> 12 cycles on a PPro core (AMD's K7 core seems to be rather more graceful
> about this one). I haven't timed a P4 lately, I think it's worse.

I had a think on this from the perspective of increasing UP performance
of SMP kernels, and came up with the following rather interesting (IMHO)
patch.  Executive summary is that when a SMP kernel finds itself on a UP
box, it modifies itself (ooohh!) by going along and essentially doing a
sed '_text,_etext s/lock/nop/'    :)

Details:  Address of each relevant lock opcode is stored (similar to
the way an exception table is) in a separate section which is marked
init, and so it is zero cost.  After we determine smp_num_cpus==1,
a bit of code (also __init) uses the table to replace "lock" with "nop".
On a very basic system (IDE disk, net) I get about 3300 locks replaced.
For modules, insmod could detect the case of a SMP kernel on a UP box
and do the lock -> nop change prior to starting the module.

Patch is against 2.4.19pre2 - not that it is necessarily 2.4 material,
but rather that I just wanted a kernel that worked to start with.  :)

Paul.

--- ../linux/arch/i386/kernel/smpboot.c	Sun Mar  3 05:04:19 2002
+++ arch/i386/kernel/smpboot.c	Sat Mar 16 06:37:03 2002
@@ -30,6 +30,7 @@
  *		Tigran Aivazian	:	fixed "0.00 in /proc/uptime on SMP" bug.
  *	Maciej W. Rozycki	:	Bits for genuine 82489DX APICs
  *		Martin J. Bligh	: 	Added support for multi-quad systems
+ *		Paul Gortmaker	:	Self-modifying code; speedup UP boxes
  */
 
 #include <linux/config.h>
@@ -964,6 +965,44 @@
 }
 
 /*
+ * Self modifying kernel - oh my...  The lock.init section contains
+ * the address of all the "lock" opcodes (think like an exception table)
+ * and this just replaces all the "lock" with "nop".  Number of cycles
+ * saved depends on how processor handles "lock", but is substantial.
+ */
+
+extern const unsigned long __lock_tbl_start[];
+extern const unsigned long __lock_tbl_end[];
+extern char _text, _etext;
+
+#define IN_TEXT(x) (((char*)(x) >= &_text) && ((char*)(x) < &_etext))
+
+void __init modify_kernel_for_UP(void)
+{
+	unsigned long i, count = 0;
+
+	printk("Dynamically optimizing SMP kernel code for UP operation...\n");
+
+	for (i = 0; __lock_tbl_start[i] != __lock_tbl_end[0]; i++) {
+
+		u8 *here = (u8*)__lock_tbl_start[i];
+
+		if (!IN_TEXT(here))
+			continue;	/* skip _init, _exit, etc. */
+
+		if (*here != 0xf0) {
+			printk("0x%p is not 0xf0 (lock opcode)\n", here);
+			return;
+		}
+
+		*here = 0x90;	/* nop */
+		count++;
+	}
+
+	printk("OK, made %ld modifications to kernel code.\n", count);
+}
+
+/*
  * Cycle through the processors sending APIC IPIs to boot each.
  */
 
@@ -1220,5 +1259,8 @@
 		synchronize_tsc_bp();
 
 smp_done:
+	if (smp_num_cpus == 1)
+		modify_kernel_for_UP();
+
 	zap_low_mappings();
 }
--- ../linux/arch/i386/vmlinux.lds	Thu Feb 28 09:36:30 2002
+++ arch/i386/vmlinux.lds	Fri Mar 15 06:04:15 2002
@@ -49,6 +49,10 @@
   __initcall_start = .;
   .initcall.init : { *(.initcall.init) }
   __initcall_end = .;
+  . = ALIGN(4);
+  __lock_tbl_start = .;
+  .lock.init : { *(.lock.init) }
+  __lock_tbl_end = .;
   . = ALIGN(4096);
   __init_end = .;
 
diff -u ../linux/include/asm-i386/atomic.h include/asm-i386/atomic.h
--- ../linux/include/asm-i386/atomic.h	Mon Feb 18 05:22:14 2002
+++ include/asm-i386/atomic.h	Fri Mar 15 04:07:46 2002
@@ -9,9 +9,15 @@
  */
 
 #ifdef CONFIG_SMP
-#define LOCK "lock ; "
+#define LOCK "\n1:\tlock ; "
+#define LOCK_ADDR	"\n" \
+			".section .lock.init,\"a\"\n\t" \
+			".align 4\n\t" \
+			".long 1b\n" \
+			".previous\n"
 #else
 #define LOCK ""
+#define LOCK_ADDR ""
 #endif
 
 /*
@@ -54,6 +60,7 @@
 {
 	__asm__ __volatile__(
 		LOCK "addl %1,%0"
+		LOCK_ADDR
 		:"=m" (v->counter)
 		:"ir" (i), "m" (v->counter));
 }
@@ -70,6 +77,7 @@
 {
 	__asm__ __volatile__(
 		LOCK "subl %1,%0"
+		LOCK_ADDR
 		:"=m" (v->counter)
 		:"ir" (i), "m" (v->counter));
 }
@@ -90,6 +98,7 @@
 
 	__asm__ __volatile__(
 		LOCK "subl %2,%0; sete %1"
+		LOCK_ADDR
 		:"=m" (v->counter), "=qm" (c)
 		:"ir" (i), "m" (v->counter) : "memory");
 	return c;
@@ -106,6 +115,7 @@
 {
 	__asm__ __volatile__(
 		LOCK "incl %0"
+		LOCK_ADDR
 		:"=m" (v->counter)
 		:"m" (v->counter));
 }
@@ -121,6 +131,7 @@
 {
 	__asm__ __volatile__(
 		LOCK "decl %0"
+		LOCK_ADDR
 		:"=m" (v->counter)
 		:"m" (v->counter));
 }
@@ -140,6 +151,7 @@
 
 	__asm__ __volatile__(
 		LOCK "decl %0; sete %1"
+		LOCK_ADDR
 		:"=m" (v->counter), "=qm" (c)
 		:"m" (v->counter) : "memory");
 	return c != 0;
@@ -160,6 +172,7 @@
 
 	__asm__ __volatile__(
 		LOCK "incl %0; sete %1"
+		LOCK_ADDR
 		:"=m" (v->counter), "=qm" (c)
 		:"m" (v->counter) : "memory");
 	return c != 0;
@@ -181,6 +194,7 @@
 
 	__asm__ __volatile__(
 		LOCK "addl %2,%0; sets %1"
+		LOCK_ADDR
 		:"=m" (v->counter), "=qm" (c)
 		:"ir" (i), "m" (v->counter) : "memory");
 	return c;
@@ -189,10 +203,12 @@
 /* These are x86-specific, used by some header files */
 #define atomic_clear_mask(mask, addr) \
 __asm__ __volatile__(LOCK "andl %0,%1" \
+LOCK_ADDR \
 : : "r" (~(mask)),"m" (*addr) : "memory")
 
 #define atomic_set_mask(mask, addr) \
 __asm__ __volatile__(LOCK "orl %0,%1" \
+LOCK_ADDR \
 : : "r" (mask),"m" (*addr) : "memory")
 
 /* Atomic operations are already serializing on x86 */
diff -u ../linux/include/asm-i386/bitops.h include/asm-i386/bitops.h
--- ../linux/include/asm-i386/bitops.h	Mon Feb 18 05:22:14 2002
+++ include/asm-i386/bitops.h	Sat Mar 16 05:17:08 2002
@@ -16,9 +16,15 @@
  */
 
 #ifdef CONFIG_SMP
-#define LOCK_PREFIX "lock ; "
+#define LOCK_PREFIX "\n1:\tlock ; "
+#define LOCK_SUFFIX	"\n" \
+			".section .lock.init,\"a\"\n\t" \
+			".align 4\n\t" \
+			".long 1b\n" \
+			".previous"
 #else
 #define LOCK_PREFIX ""
+#define LOCK_SUFFIX ""
 #endif
 
 #define ADDR (*(volatile long *) addr)
@@ -37,6 +43,7 @@
 {
 	__asm__ __volatile__( LOCK_PREFIX
 		"btsl %1,%0"
+		LOCK_SUFFIX
 		:"=m" (ADDR)
 		:"Ir" (nr));
 }
@@ -72,6 +79,7 @@
 {
 	__asm__ __volatile__( LOCK_PREFIX
 		"btrl %1,%0"
+		LOCK_SUFFIX
 		:"=m" (ADDR)
 		:"Ir" (nr));
 }
@@ -108,6 +116,7 @@
 {
 	__asm__ __volatile__( LOCK_PREFIX
 		"btcl %1,%0"
+		LOCK_SUFFIX
 		:"=m" (ADDR)
 		:"Ir" (nr));
 }
@@ -126,6 +135,7 @@
 
 	__asm__ __volatile__( LOCK_PREFIX
 		"btsl %2,%1\n\tsbbl %0,%0"
+		LOCK_SUFFIX
 		:"=r" (oldbit),"=m" (ADDR)
 		:"Ir" (nr) : "memory");
 	return oldbit;
@@ -165,6 +175,7 @@
 
 	__asm__ __volatile__( LOCK_PREFIX
 		"btrl %2,%1\n\tsbbl %0,%0"
+		LOCK_SUFFIX
 		:"=r" (oldbit),"=m" (ADDR)
 		:"Ir" (nr) : "memory");
 	return oldbit;
@@ -216,6 +227,7 @@
 
 	__asm__ __volatile__( LOCK_PREFIX
 		"btcl %2,%1\n\tsbbl %0,%0"
+		LOCK_SUFFIX
 		:"=r" (oldbit),"=m" (ADDR)
 		:"Ir" (nr) : "memory");
 	return oldbit;
diff -u ../linux/include/asm-i386/rwlock.h include/asm-i386/rwlock.h
--- ../linux/include/asm-i386/rwlock.h	Sun Mar  3 05:05:10 2002
+++ include/asm-i386/rwlock.h	Fri Mar 15 03:50:13 2002
@@ -22,6 +22,7 @@
 
 #define __build_read_lock_ptr(rw, helper)   \
 	asm volatile(LOCK "subl $1,(%0)\n\t" \
+		     LOCK_ADDR "\t" \
 		     "js 2f\n" \
 		     "1:\n" \
 		     LOCK_SECTION_START("") \
@@ -32,6 +33,7 @@
 
 #define __build_read_lock_const(rw, helper)   \
 	asm volatile(LOCK "subl $1,%0\n\t" \
+		     LOCK_ADDR "\t" \
 		     "js 2f\n" \
 		     "1:\n" \
 		     LOCK_SECTION_START("") \
@@ -52,6 +54,7 @@
 
 #define __build_write_lock_ptr(rw, helper) \
 	asm volatile(LOCK "subl $" RW_LOCK_BIAS_STR ",(%0)\n\t" \
+		     LOCK_ADDR "\t" \
 		     "jnz 2f\n" \
 		     "1:\n" \
 		     LOCK_SECTION_START("") \
@@ -62,6 +65,7 @@
 
 #define __build_write_lock_const(rw, helper) \
 	asm volatile(LOCK "subl $" RW_LOCK_BIAS_STR ",(%0)\n\t" \
+		     LOCK_ADDR "\t" \
 		     "jnz 2f\n" \
 		     "1:\n" \
 		     LOCK_SECTION_START("") \
diff -u ../linux/include/asm-i386/semaphore.h include/asm-i386/semaphore.h
--- ../linux/include/asm-i386/semaphore.h	Sun Mar  3 05:05:10 2002
+++ include/asm-i386/semaphore.h	Fri Mar 15 03:51:20 2002
@@ -120,6 +120,7 @@
 	__asm__ __volatile__(
 		"# atomic down operation\n\t"
 		LOCK "decl %0\n\t"     /* --sem->count */
+		LOCK_ADDR "\t"
 		"js 2f\n"
 		"1:\n"
 		LOCK_SECTION_START("")
@@ -146,6 +147,7 @@
 	__asm__ __volatile__(
 		"# atomic interruptible down operation\n\t"
 		LOCK "decl %1\n\t"     /* --sem->count */
+		LOCK_ADDR "\t"
 		"js 2f\n\t"
 		"xorl %0,%0\n"
 		"1:\n"
@@ -174,6 +176,7 @@
 	__asm__ __volatile__(
 		"# atomic interruptible down operation\n\t"
 		LOCK "decl %1\n\t"     /* --sem->count */
+		LOCK_ADDR "\t"
 		"js 2f\n\t"
 		"xorl %0,%0\n"
 		"1:\n"
@@ -201,6 +204,7 @@
 	__asm__ __volatile__(
 		"# atomic up operation\n\t"
 		LOCK "incl %0\n\t"     /* ++sem->count */
+		LOCK_ADDR "\t"
 		"jle 2f\n"
 		"1:\n"
 		LOCK_SECTION_START("")
diff -u ../linux/include/asm-i386/spinlock.h include/asm-i386/spinlock.h
--- ../linux/include/asm-i386/spinlock.h	Sun Mar  3 05:05:10 2002
+++ include/asm-i386/spinlock.h	Fri Mar 15 04:05:42 2002
@@ -54,7 +54,11 @@
 
 #define spin_lock_string \
 	"\n1:\t" \
-	"lock ; decb %0\n\t" \
+	"lock ; decb %0\n" \
+	".section .lock.init,\"a\"\n\t" \
+	".align 4\n\t" \
+	".long 1b\n" \
+	".previous\n\t" \
 	"js 2f\n" \
 	LOCK_SECTION_START("") \
 	"2:\t" \
@@ -197,8 +201,21 @@
 	__build_write_lock(rw, "__write_lock_failed");
 }
 
-#define read_unlock(rw)		asm volatile("lock ; incl %0" :"=m" ((rw)->lock) : : "memory")
-#define write_unlock(rw)	asm volatile("lock ; addl $" RW_LOCK_BIAS_STR ",%0":"=m" ((rw)->lock) : : "memory")
+#define read_unlock(rw)		asm volatile( \
+	"\n1:\tlock ; incl %0\n" \
+	".section .lock.init,\"a\"\n\t" \
+	".align 4\n\t" \
+	".long 1b\n" \
+	".previous" \
+	:"=m" ((rw)->lock) : : "memory")
+
+#define write_unlock(rw)	asm volatile( \
+	"\n1:\tlock ; addl $" RW_LOCK_BIAS_STR ",%0\n" \
+	".section .lock.init,\"a\"\n\t" \
+	".align 4\n\t" \
+	".long 1b\n" \
+	".previous" \
+	:"=m" ((rw)->lock) : : "memory")
 
 static inline int write_trylock(rwlock_t *lock)
 {



             reply	other threads:[~2002-03-16 12:22 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2002-03-16 11:51 Paul Gortmaker [this message]
2002-03-16 12:42 ` [PATCH] Speedup SMP kernel on UP box Keith Owens
     [not found] <0C01A29FBAE24448A792F5C68F5EA47D238DE0@nasdaq.ms.ensim.com>
2002-03-16 23:37 ` Paul Menage
2002-03-16 23:54   ` Daniel Jacobowitz

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=3C9331C7.5BDDE3BA@yahoo.com \
    --to=p_gortmaker@yahoo.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=torvalds@transmeta.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox