--- include/linux/init_task.h 938a1587ab9e35bb8d24cf843d4e7424e3030a4c +++ include/linux/init_task.h f9469678db8c3609c2f07ddb885ec4f6afa7812c @@ -126,7 +126,8 @@ .cpu_timers = INIT_CPU_TIMERS(tsk.cpu_timers), \ .fs_excl = ATOMIC_INIT(0), \ .posix_timer_list = NULL, \ - INIT_RT_MUTEXES(tsk) \ + INIT_RT_MUTEXES(tsk), \ + .delayed_drop = LIST_HEAD_INIT(tsk.children) \ } ============================================================ --- include/linux/sched.h 0ed8993484be9c13728f4ebdaa51fc0f0c229018 +++ include/linux/sched.h c65ebaa452498f611280baafd8ee6282ea0746f2 @@ -1082,6 +1091,9 @@ * cache last used pipe for splice */ struct pipe_inode_info *splice_pipe; + + /* --billh */ + struct list_head delayed_drop; // should investigate how do_fork() handles this as well }; static inline pid_t process_group(struct task_struct *tsk) ============================================================ --- kernel/exit.c f4cc2f8e48a262bd26b6fc5d1253a8482c8c2d04 +++ kernel/exit.c 9af3ab7b1d0174e3b8fa7aacec97c8e85c559e68 @@ -177,7 +177,7 @@ spin_unlock(&p->proc_lock); proc_pid_flush(proc_dentry); release_thread(p); - call_rcu(&p->rcu, delayed_put_task_struct); + put_task_struct(p); p = leader; if (unlikely(zap_leader)) ============================================================ --- kernel/fork.c 506dabd42d242f78e0321594c7723481e0cd87dc +++ kernel/fork.c ee8452887ab91be38ffd366bb57be2cf46b2f08b @@ -75,7 +75,10 @@ */ static DEFINE_PER_CPU(struct task_struct *, desched_task); -static DEFINE_PER_CPU(struct list_head, delayed_drop_list); +static DEFINE_PER_CPU(struct list_head, delayed_mmdrop_list); +#ifdef CONFIG_PREEMPT_RT +static DEFINE_PER_CPU(struct list_head, delayed_free_task_list); //--bilh +#endif int nr_processes(void) { @@ -120,6 +123,8 @@ } EXPORT_SYMBOL(free_task); +void fastcall free_task_delayed(struct task_struct *task); + void __put_task_struct(struct task_struct *tsk) { WARN_ON(!(tsk->exit_state & (EXIT_DEAD | EXIT_ZOMBIE))); @@ -132,9 +137,13 @@ put_group_info(tsk->group_info); if (!profile_handoff_task(tsk)) - free_task(tsk); +#ifdef CONFIG_PREEMPT_RT + free_task_delayed(tsk); +#else + free_task(tsk); // --billh +#endif } - + void __init fork_init(unsigned long mempages) { int i; @@ -167,8 +176,12 @@ init_task.signal->rlim[RLIMIT_SIGPENDING] = init_task.signal->rlim[RLIMIT_NPROC]; - for (i = 0; i < NR_CPUS; i++) - INIT_LIST_HEAD(&per_cpu(delayed_drop_list, i)); + for (i = 0; i < NR_CPUS; i++) { + INIT_LIST_HEAD(&per_cpu(delayed_mmdrop_list, i)); +#ifdef CONFIG_PREEMPT_RT + INIT_LIST_HEAD(&per_cpu(delayed_free_task_list, i)); //--billh +#endif + } } static struct task_struct *dup_task_struct(struct task_struct *orig) @@ -1067,6 +1080,9 @@ #endif rt_mutex_init_task(p); +#ifdef CONFIG_PREEMPT_RT + INIT_LIST_HEAD(&p->delayed_drop); //--billh +#endif #ifdef CONFIG_DEBUG_MUTEXES p->blocked_on = NULL; /* not blocked yet */ @@ -1693,24 +1709,73 @@ return err; } +static void wake_cpu_desched_task(void) +{ + struct task_struct *desched_task; + + desched_task = __get_cpu_var(desched_task); + if (desched_task) + wake_up_process(desched_task); +} + +#ifdef CONFIG_PREEMPT_RT +static int free_task_complete(void) +{ + struct list_head *head; + int ret = 0; + + head = &get_cpu_var(delayed_free_task_list); + while (!list_empty(head)) { + struct task_struct *task = list_entry(head->next, + struct task_struct, delayed_drop); + list_del(&task->delayed_drop); + put_cpu_var(delayed_free_task_list); + + free_task(task); + ret = 1; + + head = &get_cpu_var(delayed_free_task_list); + } + put_cpu_var(delayed_free_task_list); + + return ret; +} + +/* + * We dont want to do complex work from the scheduler, thus + * we delay the work to a per-CPU worker thread: + */ +void fastcall free_task_delayed(struct task_struct *task) +{ + struct list_head *head; + + head = &get_cpu_var(delayed_free_task_list); + list_add_tail(&task->delayed_drop, head); + + wake_cpu_desched_task(); + + put_cpu_var(delayed_mmdrop_list); +} +#endif + static int mmdrop_complete(void) { struct list_head *head; int ret = 0; - head = &get_cpu_var(delayed_drop_list); + head = &get_cpu_var(delayed_mmdrop_list); while (!list_empty(head)) { struct mm_struct *mm = list_entry(head->next, struct mm_struct, delayed_drop); list_del(&mm->delayed_drop); - put_cpu_var(delayed_drop_list); + put_cpu_var(delayed_mmdrop_list); __mmdrop(mm); ret = 1; - head = &get_cpu_var(delayed_drop_list); + head = &get_cpu_var(delayed_mmdrop_list); } - put_cpu_var(delayed_drop_list); + put_cpu_var(delayed_mmdrop_list); return ret; } @@ -1721,15 +1786,14 @@ */ void fastcall __mmdrop_delayed(struct mm_struct *mm) { - struct task_struct *desched_task; struct list_head *head; - head = &get_cpu_var(delayed_drop_list); + head = &get_cpu_var(delayed_mmdrop_list); list_add_tail(&mm->delayed_drop, head); - desched_task = __get_cpu_var(desched_task); - if (desched_task) - wake_up_process(desched_task); - put_cpu_var(delayed_drop_list); + + wake_cpu_desched_task(); + + put_cpu_var(delayed_mmdrop_list); } static int desched_thread(void * __bind_cpu) @@ -1743,6 +1807,9 @@ if (mmdrop_complete()) continue; + if (free_task_complete()) + continue; + schedule(); /* This must be called from time to time on ia64, and is a no-op on other archs. @@ -1767,7 +1834,10 @@ case CPU_UP_PREPARE: BUG_ON(per_cpu(desched_task, hotcpu)); - INIT_LIST_HEAD(&per_cpu(delayed_drop_list, hotcpu)); + INIT_LIST_HEAD(&per_cpu(delayed_mmdrop_list, hotcpu)); +#ifdef CONFIG_PREEMPT_RT + INIT_LIST_HEAD(&per_cpu(delayed_free_task_list, hotcpu)); // --billh +#endif p = kthread_create(desched_thread, hcpu, "desched/%d", hotcpu); if (IS_ERR(p)) { printk("desched_thread for %i failed\n", hotcpu);