* [Linux-ia64] O(1) MQ scheduler J9 patch for IA64
@ 2002-02-01 14:45 Erich Focht
2002-02-02 0:21 ` Keith Owens
` (2 more replies)
0 siblings, 3 replies; 4+ messages in thread
From: Erich Focht @ 2002-02-01 14:45 UTC (permalink / raw)
To: linux-ia64
[-- Attachment #1: Type: TEXT/PLAIN, Size: 776 bytes --]
Hi,
please find attached a patch which aplies to
linux-2.4.17 + kdb_common + ia64_patch + kdb_ia64 + sched-O1-2.4.17-J9.patch
I had to add a function smp_call_function_nowait which sends an IPI and
returns immediately in order to get rid of a IA64-specific race condition
when the mmu_context wraps around. Now it makes a very stable and fast
impression, interactive work is even possible while doing a "hackbench
55" and ping flooding the machine... Tested on 2 CPU BigSur, 4 CPU LION,
16 CPU AzusA.
The patch should aply to the K0 version of the scheduler, too. As this
isn't out yet for 2.4.17 I couldn't test it.
Regards,
Erich
---
Erich Focht <efocht@ess.nec.de>
NEC European Supercomputer Systems, European HPC Technology Center
[-- Attachment #2: Type: TEXT/PLAIN, Size: 21053 bytes --]
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kdb/kdba_bt.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/kdb/kdba_bt.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/kdb/kdba_bt.c Thu Jan 31 16:42:51 2002
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kdb/kdba_bt.c Thu Jan 31 18:40:40 2002
@@ -197,7 +197,7 @@
}
#ifdef CONFIG_SMP
else if (task_has_cpu(p)) {
- sw = kdb_sw[p->processor];
+ sw = kdb_sw[p->cpu];
}
#endif
else {
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kernel/entry.S 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/entry.S
--- 2.4.17-IA64-kdb-J9/arch/ia64/kernel/entry.S Thu Jan 31 16:42:26 2002
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/entry.S Thu Jan 31 18:40:40 2002
@@ -161,7 +161,8 @@
mov r8=r13 // return pointer to previously running task
mov r13=in0 // set "current" pointer
;;
-(p6) ssm psr.i // renable psr.i AFTER the ic bit is serialized
+//(p6) ssm psr.i // interrupt delivery should not be enabled
+ // with the new O(1) MQ scheduler
DO_LOAD_SWITCH_STACK
#ifdef CONFIG_SMP
@@ -170,7 +171,8 @@
br.ret.sptk.many rp // boogie on out in new context
.map:
- rsm psr.i | psr.ic
+ //rsm psr.i | psr.ic
+ rsm psr.ic
movl r25=PAGE_KERNEL
;;
srlz.d
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kernel/irq_ia64.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/irq_ia64.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/kernel/irq_ia64.c Fri Nov 9 23:26:17 2001
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/irq_ia64.c Thu Jan 31 18:40:40 2002
@@ -148,6 +148,14 @@
flags: SA_INTERRUPT,
name: "IPI"
};
+
+extern void smp_task_migration_interrupt(int irq, void *dev_id, struct pt_regs *regs);
+
+static struct irqaction task_migration_irqaction = {
+ handler: smp_task_migration_interrupt,
+ flags: SA_INTERRUPT,
+ name: "Task migration"
+};
#endif
void
@@ -172,6 +180,7 @@
register_percpu_irq(IA64_SPURIOUS_INT_VECTOR, NULL);
#ifdef CONFIG_SMP
register_percpu_irq(IA64_IPI_VECTOR, &ipi_irqaction);
+ register_percpu_irq(IA64_TASK_MIGRATION, &task_migration_irqaction);
#endif
platform_irq_init();
}
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kernel/process.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/process.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/kernel/process.c Thu Jan 31 16:42:26 2002
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/process.c Thu Jan 31 18:40:40 2002
@@ -125,9 +125,6 @@
cpu_idle (void *unused)
{
/* endless idle loop with no priority at all */
- init_idle();
- current->nice = 20;
- current->counter = -100;
while (1) {
@@ -136,11 +133,10 @@
min_xtp();
#endif
- while (!current->need_resched) {
+ if (!current->need_resched) {
#ifdef CONFIG_IA64_SGI_SN
snidle();
#endif
- continue;
}
#ifdef CONFIG_IA64_SGI_SN
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kernel/setup.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/setup.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/kernel/setup.c Thu Jan 31 16:42:26 2002
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/setup.c Thu Jan 31 18:40:40 2002
@@ -375,10 +375,10 @@
{
#ifdef CONFIG_SMP
# define lpj c->loops_per_jiffy
-# define cpu c->processor
+# define cpum c->processor
#else
# define lpj loops_per_jiffy
-# define cpu 0
+# define cpum 0
#endif
char family[32], features[128], *cp;
struct cpuinfo_ia64 *c = v;
@@ -417,7 +417,7 @@
"cpu MHz : %lu.%06lu\n"
"itc MHz : %lu.%06lu\n"
"BogoMIPS : %lu.%02lu\n\n",
- cpu, c->vendor, family, c->model, c->revision, c->archrev,
+ cpum, c->vendor, family, c->model, c->revision, c->archrev,
features, c->ppn, c->number,
c->proc_freq / 1000000, c->proc_freq % 1000000,
c->itc_freq / 1000000, c->itc_freq % 1000000,
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kernel/smp.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/smp.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/kernel/smp.c Thu Jan 31 16:42:51 2002
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/smp.c Fri Feb 1 14:36:45 2002
@@ -17,6 +17,8 @@
* scheme.
* 10/13/00 Goutham Rao <goutham.rao@intel.com> Updated smp_call_function and
* smp_call_function_single to resend IPI on timeouts
+ * 01/31/02 Erich Focht <efocht@ess.nec.de> smp_call_function_nowait for avoiding
+ * race conditions with O(1) scheduler.
*/
#define __KERNEL_SYSCALLS__
@@ -71,10 +73,16 @@
};
static volatile struct call_data_struct *call_data;
+static struct call_data_struct nowait_data = { NULL,
+ NULL,
+ 0L,
+ ATOMIC_INIT(0),
+ ATOMIC_INIT(0) };
#define IPI_CALL_FUNC 0
#define IPI_CPU_STOP 1
#define IPI_KDB_INTERRUPT 2
+#define IPI_CALL_FUNC_NOWAIT 3
static void
stop_this_cpu (void)
@@ -146,6 +154,26 @@
break;
#endif
+ case IPI_CALL_FUNC_NOWAIT:
+ {
+ struct call_data_struct *data;
+ void (*func)(void *info);
+ void *info;
+
+ /* release the 'pointer lock' */
+ data = (struct call_data_struct *) &nowait_data;
+ func = data->func;
+ info = data->info;
+
+ mb();
+ atomic_dec(&data->started);
+
+ /* At this point the structure may be gone unless wait is true. */
+ (*func)(info);
+
+ }
+ break;
+
default:
printk(KERN_CRIT "Unknown IPI on CPU %d: %lu\n", this_cpu, which);
break;
@@ -196,12 +224,55 @@
}
void
+smp_send_reschedule_all(void)
+{
+ send_IPI_all(IA64_IPI_RESCHEDULE);
+}
+
+static spinlock_t migration_lock = SPIN_LOCK_UNLOCKED;
+static task_t *new_task;
+
+
+/*
+ * This function sends a 'task migration' IPI to another CPU.
+ * Must be called from syscall contexts, with interrupts *enabled*.
+ */
+void smp_migrate_task(int cpu, task_t *p)
+{
+ /*
+ * The target CPU will unlock the migration spinlock:
+ */
+ spin_lock(&migration_lock);
+ new_task = p;
+ platform_send_ipi(cpu, IA64_TASK_MIGRATION, IA64_IPI_DM_INT, 0);
+}
+
+/*
+ * Task migration callback.
+ */
+asmlinkage void smp_task_migration_interrupt(int irq, void *dev_id, struct pt_regs *regs)
+{
+ task_t *p;
+
+ p = new_task;
+ spin_unlock(&migration_lock);
+ sched_task_migrated(p);
+}
+
+void
smp_flush_tlb_all (void)
{
smp_call_function ((void (*)(void *))__flush_tlb_all,0,1,1);
__flush_tlb_all();
}
+void
+smp_flush_tlb_all_nowait (void)
+{
+ smp_call_function_nowait ((void (*)(void *))__flush_tlb_all,0);
+ __flush_tlb_all();
+}
+
/*
* Run a function on another CPU
* <func> The function to run. This must be fast and non-blocking.
@@ -305,6 +376,42 @@
spin_unlock_bh(&call_lock);
return 0;
}
+
+/*
+ * [SUMMARY] Run a function on all other CPUs.
+ * <func> The function to run. This must be fast and non-blocking.
+ * <info> An arbitrary pointer to pass to the function.
+ * [RETURNS] 0
+ *
+ * Returns immediately. Data is stored in nowait_data where it has chances
+ * to live long enough...
+ * Used by smp_flush_tlb_all_nowait() and avoids a race condition with the
+ * O(1) MQ scheduler (wrap_mmu_context while another CPU waits for the
+ * runqueue lock of the processor doing context wrap). [EF]
+ */
+int
+smp_call_function_nowait (void (*func) (void *info), void *info)
+{
+ int cpus = smp_num_cpus-1;
+
+ if (!cpus)
+ return 0;
+
+ if (atomic_read(&nowait_data.started) > 0) {
+ printk("smp_call_function_nowait: called before previous call executed on all CPUs!\n");
+ }
+
+ nowait_data.func = func;
+ nowait_data.info = info;
+ atomic_set(&nowait_data.started, cpus);
+ nowait_data.wait = wait;
+
+ spin_lock_bh(&call_lock);
+ mb(); /* ensure store to call_data precedes setting of IPI_CALL_FUNC_NOWAIT */
+ send_IPI_allbutself(IPI_CALL_FUNC_NOWAIT);
+ spin_unlock_bh(&call_lock);
+ return 0;
+}
void
smp_do_timer (struct pt_regs *regs)
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kernel/smpboot.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/smpboot.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/kernel/smpboot.c Thu Jan 31 16:42:26 2002
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/smpboot.c Fri Feb 1 14:27:50 2002
@@ -356,6 +356,7 @@
local_irq_enable();
calibrate_delay();
local_cpu_data->loops_per_jiffy = loops_per_jiffy;
+ ia64_disable_timer();
/*
* Allow the master to continue.
*/
@@ -379,7 +380,8 @@
Dprintk("CPU %d is set to go.\n", smp_processor_id());
while (!atomic_read(&smp_commenced))
;
-
+ /* reenable timer interrupts */
+ ia64_cpu_local_tick();
Dprintk("CPU %d is starting idle.\n", smp_processor_id());
return cpu_idle();
}
@@ -416,11 +418,10 @@
if (!idle)
panic("No idle process for CPU %d", cpu);
- task_set_cpu(idle, cpu); /* we schedule the first task manually */
+ init_idle(idle, cpu);
ia64_cpu_to_sapicid[cpu] = sapicid;
- del_from_runqueue(idle);
unhash_process(idle);
init_tasks[cpu] = idle;
@@ -481,8 +482,7 @@
printk("Boot processor id 0x%x/0x%x\n", 0, boot_cpu_id);
global_irq_holder = 0;
- current->processor = 0;
- init_idle();
+ current->cpu = 0;
/*
* If SMP should be disabled, then really disable it!
@@ -569,3 +569,9 @@
smp_num_cpus = 1;
}
}
+
+/* Number of ticks we consider an idle tasks still cache-hot.
+ * For Itanium: with 1GB/s bandwidth we need 4ms to fill up 4MB L3 cache...
+ * So let's try 10 ticks.
+ */
+unsigned long cache_decay_ticks=10;
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/kernel/time.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/time.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/kernel/time.c Fri Nov 9 23:26:17 2001
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/kernel/time.c Thu Jan 31 18:41:56 2002
@@ -209,7 +209,7 @@
/*
* Encapsulate access to the itm structure for SMP.
*/
-void __init
+void
ia64_cpu_local_tick (void)
{
int cpu = smp_processor_id();
@@ -298,3 +298,9 @@
efi_gettimeofday((struct timeval *) &xtime);
ia64_init_itm();
}
+
+void __init ia64_disable_timer(void)
+{
+ ia64_set_itv(IA64_TIMER_VECTOR | IA64_TIMER_MASK);
+}
+
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/mm/fault.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/mm/fault.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/mm/fault.c Fri Nov 9 23:26:17 2001
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/mm/fault.c Thu Jan 31 18:40:40 2002
@@ -194,8 +194,7 @@
out_of_memory:
up_read(&mm->mmap_sem);
if (current->pid == 1) {
- current->policy |= SCHED_YIELD;
- schedule();
+ yield();
down_read(&mm->mmap_sem);
goto survive;
}
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/mm/tlb.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/mm/tlb.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/mm/tlb.c Fri Nov 9 23:26:17 2001
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/mm/tlb.c Fri Feb 1 14:27:08 2002
@@ -76,7 +76,8 @@
ia64_ctx.limit = tsk_context;
}
read_unlock(&tasklist_lock);
- flush_tlb_all();
+ //flush_tlb_all(); /* potential race condition with O(1) scheduler [EF] */
+ smp_flush_tlb_all_nowait();
}
static inline void
diff -urN 2.4.17-IA64-kdb-J9/arch/ia64/tools/print_offsets.c 2.4.17-IA64-kdb-J9ia64/arch/ia64/tools/print_offsets.c
--- 2.4.17-IA64-kdb-J9/arch/ia64/tools/print_offsets.c Fri Nov 9 23:26:17 2001
+++ 2.4.17-IA64-kdb-J9ia64/arch/ia64/tools/print_offsets.c Thu Jan 31 18:40:40 2002
@@ -54,7 +54,7 @@
{ "IA64_TASK_PTRACE_OFFSET", offsetof (struct task_struct, ptrace) },
{ "IA64_TASK_SIGPENDING_OFFSET", offsetof (struct task_struct, sigpending) },
{ "IA64_TASK_NEED_RESCHED_OFFSET", offsetof (struct task_struct, need_resched) },
- { "IA64_TASK_PROCESSOR_OFFSET", offsetof (struct task_struct, processor) },
+ { "IA64_TASK_PROCESSOR_OFFSET", offsetof (struct task_struct, cpu) },
{ "IA64_TASK_THREAD_OFFSET", offsetof (struct task_struct, thread) },
{ "IA64_TASK_THREAD_KSP_OFFSET", offsetof (struct task_struct, thread.ksp) },
#ifdef CONFIG_PERFMON
diff -urN 2.4.17-IA64-kdb-J9/include/asm-ia64/bitops.h 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/bitops.h
--- 2.4.17-IA64-kdb-J9/include/asm-ia64/bitops.h Thu Jan 31 16:42:28 2002
+++ 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/bitops.h Fri Feb 1 14:39:33 2002
@@ -280,6 +280,20 @@
return result;
}
+/**
+ * __ffs - find first bit in a 64 bit long.
+ *
+ * Undefined if no bit exists, so code should check against 0 first.
+ */
+static inline unsigned long
+__ffs (unsigned long x)
+{
+ unsigned long result;
+
+ __asm__ ("popcnt %0=%1" : "=r" (result) : "r" (~x & (x - 1)));
+ return result;
+}
+
#ifdef __KERNEL__
/*
@@ -357,6 +371,8 @@
tmp = *p;
found_first:
tmp |= ~0UL << size;
+ if (tmp == ~0UL) /* Are any bits zero? */
+ return result + size; /* Nope. */
found_middle:
return result + ffz(tmp);
}
@@ -366,8 +382,52 @@
*/
#define find_first_zero_bit(addr, size) find_next_zero_bit((addr), (size), 0)
+/*
+ * Find next bit in a bitmap reasonably efficiently..
+ */
+static inline int
+find_next_bit (void *addr, unsigned long size, unsigned long offset)
+{
+ unsigned long *p = ((unsigned long *) addr) + (offset >> 6);
+ unsigned long result = offset & ~63UL;
+ unsigned long tmp;
+
+ if (offset >= size)
+ return size;
+ size -= result;
+ offset &= 63UL;
+ if (offset) {
+ tmp = *(p++);
+ tmp &= ~0UL << offset;
+ if (size < 64)
+ goto found_first;
+ if (tmp)
+ goto found_middle;
+ size -= 64;
+ result += 64;
+ }
+ while (size & ~63UL) {
+ if ((tmp = *(p++)))
+ goto found_middle;
+ result += 64;
+ size -= 64;
+ }
+ if (!size)
+ return result;
+ tmp = *p;
+found_first:
+ tmp &= ~0UL >> (64-size);
+ if (tmp == 0UL) /* Are any bits set? */
+ return result + size; /* Nope. */
+found_middle:
+ return result + __ffs(tmp);
+}
+
+#define find_first_bit(addr, size) find_next_bit((addr), (size), 0)
+
#ifdef __KERNEL__
+#define __clear_bit(nr, addr) clear_bit(nr, addr)
#define ext2_set_bit test_and_set_bit
#define ext2_clear_bit test_and_clear_bit
#define ext2_test_bit test_bit
diff -urN 2.4.17-IA64-kdb-J9/include/asm-ia64/hw_irq.h 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/hw_irq.h
--- 2.4.17-IA64-kdb-J9/include/asm-ia64/hw_irq.h Tue Jul 31 19:30:09 2001
+++ 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/hw_irq.h Fri Feb 1 14:39:33 2002
@@ -49,6 +49,7 @@
#define IA64_PERFMON_VECTOR 0xee /* performanc monitor interrupt vector */
#define IA64_TIMER_VECTOR 0xef /* use highest-prio group 15 interrupt for timer */
#define IA64_MCA_WAKEUP_VECTOR 0xf0 /* MCA wakeup (must be >MCA_RENDEZ_VECTOR) */
+#define IA64_TASK_MIGRATION 0xfb /* task migration interrupt vector */
#define IA64_IPI_RESCHEDULE 0xfd /* SMP reschedule */
#define IA64_IPI_VECTOR 0xfe /* inter-processor interrupt vector */
@@ -65,6 +66,9 @@
IA64_IPI_DM_EXTINT = 0x7, /* pend an 8259-compatible interrupt. */
};
+/* bit for masking and discarding timer interrupts on IA64 */
+#define IA64_TIMER_MASK (1<<16)
+
extern __u8 isa_irq_to_vector_map[16];
#define isa_irq_to_vector(x) isa_irq_to_vector_map[(x)]
diff -urN 2.4.17-IA64-kdb-J9/include/asm-ia64/mmu_context.h 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/mmu_context.h
--- 2.4.17-IA64-kdb-J9/include/asm-ia64/mmu_context.h Fri Nov 9 23:26:17 2001
+++ 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/mmu_context.h Fri Feb 1 14:39:33 2002
@@ -118,6 +118,22 @@
reload_context(next);
}
+/*
+ * Needed for the O(1) MQ scheduler.
+ */
+#if MAX_RT_PRIO != 128 || MAX_PRIO >= 192
+# error update this function. */
+#endif
+
+static inline int sched_find_first_bit(unsigned long *b)
+{
+ if (unlikely(b[0]))
+ return __ffs(b[0]);
+ if (unlikely(b[1]))
+ return 64 + __ffs(b[1]);
+ return __ffs(b[2]) + 128;
+}
+
#define switch_mm(prev_mm,next_mm,next_task,cpu) activate_mm(prev_mm, next_mm)
# endif /* ! __ASSEMBLY__ */
diff -urN 2.4.17-IA64-kdb-J9/include/asm-ia64/smp.h 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/smp.h
--- 2.4.17-IA64-kdb-J9/include/asm-ia64/smp.h Fri Nov 9 23:26:17 2001
+++ 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/smp.h Fri Feb 1 14:39:33 2002
@@ -27,7 +27,7 @@
#define SMP_IRQ_REDIRECTION (1 << 0)
#define SMP_IPI_REDIRECTION (1 << 1)
-#define smp_processor_id() (current->processor)
+#define smp_processor_id() (current->cpu)
extern struct smp_boot_data {
int cpu_count;
@@ -48,6 +48,9 @@
extern unsigned long ap_wakeup_vector;
+extern void smp_send_reschedule(int cpu);
+extern void smp_send_reschedule_all(void);
+
/*
* Function to map hard smp processor id to logical id. Slow, so
* don't use this in performance-critical code.
@@ -110,12 +113,6 @@
#define NO_PROC_ID 0xffffffff /* no processor magic marker */
-/*
- * Extra overhead to move a task from one cpu to another (due to TLB and cache misses).
- * Expressed in "negative nice value" units (larger number means higher priority/penalty).
- */
-#define PROC_CHANGE_PENALTY 20
-
extern void __init init_smp_config (void);
extern void smp_do_timer (struct pt_regs *regs);
diff -urN 2.4.17-IA64-kdb-J9/include/asm-ia64/spinlock.h 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/spinlock.h
--- 2.4.17-IA64-kdb-J9/include/asm-ia64/spinlock.h Thu Jan 31 16:42:28 2002
+++ 2.4.17-IA64-kdb-J9ia64/include/asm-ia64/spinlock.h Fri Feb 1 14:39:33 2002
@@ -84,7 +84,7 @@
"mov r29 = 1\n" \
";;\n" \
"1:\n" \
- "ld4.bias r2 = [%0]\n" \
+ "ld4 r2 = [%0]\n" \
";;\n" \
"cmp4.eq p0,p7 = r0,r2\n" \
"(p7) br.cond.spnt.few 1b \n" \
diff -urN 2.4.17-IA64-kdb-J9/include/linux/smp.h 2.4.17-IA64-kdb-J9ia64/include/linux/smp.h
--- 2.4.17-IA64-kdb-J9/include/linux/smp.h Thu Jan 31 18:39:37 2002
+++ 2.4.17-IA64-kdb-J9ia64/include/linux/smp.h Fri Feb 1 14:39:33 2002
@@ -24,12 +24,6 @@
extern void smp_send_stop(void);
/*
- * sends a 'reschedule' event to another CPU:
- */
-extern void FASTCALL(smp_send_reschedule(int cpu));
-
-
-/*
* Boot processor call to load the other CPU's
*/
extern void smp_boot_cpus(void);
diff -urN 2.4.17-IA64-kdb-J9/kdb/kdbmain.c 2.4.17-IA64-kdb-J9ia64/kdb/kdbmain.c
--- 2.4.17-IA64-kdb-J9/kdb/kdbmain.c Thu Jan 31 16:42:04 2002
+++ 2.4.17-IA64-kdb-J9ia64/kdb/kdbmain.c Thu Jan 31 18:40:40 2002
@@ -2360,7 +2360,7 @@
for_each_task(p) {
kdb_printf("0x%p %08d %08d %1.1d %3.3d %s 0x%p%c%s\n",
(void *)p, p->pid, p->p_pptr->pid,
- task_has_cpu(p), p->processor,
+ task_has_cpu(p), p->cpu,
(p->state == 0)?"run ":(p->state>0)?"stop":"unrn",
(void *)(&p->thread),
(p == current) ? '*': ' ',
diff -urN 2.4.17-IA64-kdb-J9/kernel/printk.c 2.4.17-IA64-kdb-J9ia64/kernel/printk.c
--- 2.4.17-IA64-kdb-J9/kernel/printk.c Thu Jan 31 18:39:37 2002
+++ 2.4.17-IA64-kdb-J9ia64/kernel/printk.c Thu Jan 31 18:40:40 2002
@@ -25,6 +25,8 @@
#include <linux/init.h>
#include <linux/module.h>
#include <linux/interrupt.h> /* For in_interrupt() */
+#include <linux/config.h>
+#include <linux/delay.h>
#include <asm/uaccess.h>
diff -urN 2.4.17-IA64-kdb-J9/kernel/sched.c 2.4.17-IA64-kdb-J9ia64/kernel/sched.c
--- 2.4.17-IA64-kdb-J9/kernel/sched.c Thu Jan 31 18:39:37 2002
+++ 2.4.17-IA64-kdb-J9ia64/kernel/sched.c Fri Feb 1 16:06:40 2002
@@ -93,6 +93,15 @@
p->array = array;
}
+/* needed on IA64, arch/ia64/kernel/head.S relies on it (EF) */
+struct task_struct * init_tasks[NR_CPUS] __initdata = {&init_task, };
+
+/* needed in kdb (EF) */
+int task_has_cpu(task_t *p)
+{
+ return (p == task_rq(p)->curr);
+}
+
/*
* A task is 'heavily interactive' if it either has reached the
* bottom 25% of the SCHED_OTHER priority range, or if it is below
@@ -859,16 +868,16 @@
p->cpus_allowed = new_mask;
/*
- * Can the task run on the current CPU? If not then
+ * Can the task run on its current CPU? If not then
* migrate the process off to a proper CPU.
*/
- if (new_mask & (1UL << smp_processor_id()))
+ if (new_mask & (1UL << p->cpu))
return;
#if CONFIG_SMP
- current->state = TASK_UNINTERRUPTIBLE;
- smp_migrate_task(__ffs(new_mask), current);
-
- schedule();
+ p->state = TASK_UNINTERRUPTIBLE;
+ smp_migrate_task(__ffs(new_mask), p);
+ if (p == current)
+ schedule();
#endif
}
diff -urN 2.4.17-IA64-kdb-J9/kernel/timer.c 2.4.17-IA64-kdb-J9ia64/kernel/timer.c
--- 2.4.17-IA64-kdb-J9/kernel/timer.c Thu Jan 31 18:39:37 2002
+++ 2.4.17-IA64-kdb-J9ia64/kernel/timer.c Thu Jan 31 18:40:40 2002
@@ -585,17 +585,16 @@
update_one_process(p, user_tick, system, cpu);
if (p->pid) {
- if (--p->counter <= 0) {
- p->counter = 0;
- p->need_resched = 1;
- }
- if (p->nice > 0)
+ if (p->__nice > 0)
kstat.per_cpu_nice[cpu] += user_tick;
else
kstat.per_cpu_user[cpu] += user_tick;
kstat.per_cpu_system[cpu] += system;
- } else if (really_local_bh_count() || really_local_irq_count() > 1)
- kstat.per_cpu_system[cpu] += system;
+ } else {
+ if (really_local_bh_count() || really_local_irq_count() > 1)
+ kstat.per_cpu_system[cpu] += system;
+ }
+ scheduler_tick(p);
}
/*
^ permalink raw reply [flat|nested] 4+ messages in thread* Re: [Linux-ia64] O(1) MQ scheduler J9 patch for IA64
2002-02-01 14:45 [Linux-ia64] O(1) MQ scheduler J9 patch for IA64 Erich Focht
@ 2002-02-02 0:21 ` Keith Owens
2002-02-04 17:21 ` Erich Focht
2002-02-04 19:22 ` Ingo Molnar
2 siblings, 0 replies; 4+ messages in thread
From: Keith Owens @ 2002-02-02 0:21 UTC (permalink / raw)
To: linux-ia64
On Fri, 1 Feb 2002 15:45:32 +0100 (MET),
Erich Focht <efocht@ess.nec.de> wrote:
>please find attached a patch which aplies to
>
>linux-2.4.17 + kdb_common + ia64_patch + kdb_ia64 + sched-O1-2.4.17-J9.patch
>
>I had to add a function smp_call_function_nowait which sends an IPI and
>returns immediately in order to get rid of a IA64-specific race condition
>when the mmu_context wraps around. Now it makes a very stable and fast
>impression, interactive work is even possible while doing a "hackbench
>55" and ping flooding the machine... Tested on 2 CPU BigSur, 4 CPU LION,
>16 CPU AzusA.
Why is there a global change of current->processor to current->cpu?
That has knock on effects because #define cpu and references to that
macro also have to be changed. I don't see any benefit.
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [Linux-ia64] O(1) MQ scheduler J9 patch for IA64
2002-02-01 14:45 [Linux-ia64] O(1) MQ scheduler J9 patch for IA64 Erich Focht
2002-02-02 0:21 ` Keith Owens
@ 2002-02-04 17:21 ` Erich Focht
2002-02-04 19:22 ` Ingo Molnar
2 siblings, 0 replies; 4+ messages in thread
From: Erich Focht @ 2002-02-04 17:21 UTC (permalink / raw)
To: linux-ia64
On Sat, 2 Feb 2002, Keith Owens wrote:
> Why is there a global change of current->processor to current->cpu?
> That has knock on effects because #define cpu and references to that
> macro also have to be changed. I don't see any benefit.
Keith,
I can only guess that Ingo wanted to see the problematic spots when he
compiles. I think there was a discussion on linux-kernel ML but I didn't
follow it, so maybe Ingo can tell us...
Regards,
Erich
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [Linux-ia64] O(1) MQ scheduler J9 patch for IA64
2002-02-01 14:45 [Linux-ia64] O(1) MQ scheduler J9 patch for IA64 Erich Focht
2002-02-02 0:21 ` Keith Owens
2002-02-04 17:21 ` Erich Focht
@ 2002-02-04 19:22 ` Ingo Molnar
2 siblings, 0 replies; 4+ messages in thread
From: Ingo Molnar @ 2002-02-04 19:22 UTC (permalink / raw)
To: linux-ia64
On Mon, 4 Feb 2002, Erich Focht wrote:
> > Why is there a global change of current->processor to current->cpu?
> > That has knock on effects because #define cpu and references to that
> > macro also have to be changed. I don't see any benefit.
>
> Keith,
> I can only guess that Ingo wanted to see the problematic spots when he
> compiles. I think there was a discussion on linux-kernel ML but I didn't
> follow it, so maybe Ingo can tell us...
the current->processor => current->cpu indeed was such a helper, to find
places that rely on p->cpu. (it's mostly code that has scheduling
relevance.)
renaming ->nice to ->__nice had another purpose, eg. in the -K2 patch
there is no 'nice' field anymore, ->static_prio is used.
Ingo
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2002-02-04 19:22 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2002-02-01 14:45 [Linux-ia64] O(1) MQ scheduler J9 patch for IA64 Erich Focht
2002-02-02 0:21 ` Keith Owens
2002-02-04 17:21 ` Erich Focht
2002-02-04 19:22 ` Ingo Molnar
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox