X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=mm%2Foom_kill.c;h=ac1055f2ba9ce6055f2bd367e8658cd02f0317c4;hb=c7942184966421370e59fff55a0c4aabba7edb58;hp=3de04b6173c51b474f5ff4110f08d79150d439c9;hpb=48ed9e9ed158dedf557fbe4b9e8b09f109e2a79a;p=linux-2.6.git diff --git a/mm/oom_kill.c b/mm/oom_kill.c index 3de04b617..ac1055f2b 100644 --- a/mm/oom_kill.c +++ b/mm/oom_kill.c @@ -6,8 +6,8 @@ * for goading me into coding this file... * * The routines in this file are used to kill a process when - * we're seriously out of memory. This gets called from kswapd() - * in linux/mm/vmscan.c when we really run out of memory. + * we're seriously out of memory. This gets called from __alloc_pages() + * in mm/page_alloc.c when we really run out of memory. * * Since we won't call these routines often (on a well-configured * machine) this file will double as a 'coding guide' and a signpost @@ -20,12 +20,15 @@ #include #include #include +#include +int sysctl_panic_on_oom; /* #define DEBUG */ /** - * oom_badness - calculate a numeric value for how bad this task has been + * badness - calculate a numeric value for how bad this task has been * @p: task struct of which task we should calculate + * @uptime: current uptime in seconds * * The formula used is relatively simple and documented inline in the * function. The main rationale is that we want to select a good task @@ -41,28 +44,64 @@ * of least surprise ... (be careful when you change it) */ -static int badness(struct task_struct *p) +unsigned long badness(struct task_struct *p, unsigned long uptime) { - int points, cpu_time, run_time, s; + unsigned long points, cpu_time, run_time, s; + struct mm_struct *mm; + struct task_struct *child; - if (!p->mm) + task_lock(p); + mm = p->mm; + if (!mm) { + task_unlock(p); return 0; + } + + /* + * swapoff can easily use up all memory, so kill those first. + */ + if (p->flags & PF_SWAPOFF) + return ULONG_MAX; - if (p->flags & PF_MEMDIE) - return 0; /* * The memory size of the process is the basis for the badness. */ - points = p->mm->total_vm; - /* add vserver badness ;) */ + points = mm->total_vm; + + /* + * After this unlock we can no longer dereference local variable `mm' + */ + task_unlock(p); + + /* FIXME: add vserver badness ;) */ + + /* + * Processes which fork a lot of child processes are likely + * a good choice. We add half the vmsize of the children if they + * have an own mm. This prevents forking servers to flood the + * machine with an endless amount of children. In case a single + * child is eating the vast majority of memory, adding only half + * to the parents will make the child our kill candidate of choice. + */ + list_for_each_entry(child, &p->children, sibling) { + task_lock(child); + if (child->mm != mm && child->mm) + points += child->mm->total_vm/2 + 1; + task_unlock(child); + } /* - * CPU time is in seconds and run time is in minutes. There is no - * particular reason for this other than that it turned out to work - * very well in practice. + * CPU time is in tens of seconds and run time is in thousands + * of seconds. There is no particular reason for this other than + * that it turned out to work very well in practice. */ - cpu_time = (p->utime + p->stime) >> (SHIFT_HZ + 3); - run_time = (get_jiffies_64() - p->start_time) >> (SHIFT_HZ + 10); + cpu_time = (cputime_to_jiffies(p->utime) + cputime_to_jiffies(p->stime)) + >> (SHIFT_HZ + 3); + + if (uptime >= p->start_time.tv_sec) + run_time = (uptime - p->start_time.tv_sec) >> 10; + else + run_time = 0; s = int_sqrt(cpu_time); if (s) @@ -94,6 +133,25 @@ static int badness(struct task_struct *p) */ if (cap_t(p->cap_effective) & CAP_TO_MASK(CAP_SYS_RAWIO)) points /= 4; + + /* + * If p's nodes don't overlap ours, it may still help to kill p + * because p may have allocated or otherwise mapped memory on + * this node before. However it will be less likely. + */ + if (!cpuset_excl_nodes_overlap(p)) + points /= 8; + + /* + * Adjust the score by oomkilladj. + */ + if (p->oomkilladj) { + if (p->oomkilladj > 0) + points <<= p->oomkilladj; + else + points >>= -(p->oomkilladj); + } + #ifdef DEBUG printk(KERN_DEBUG "OOMkill: task %d (%s) got %d points\n", p->pid, p->comm, points); @@ -101,29 +159,98 @@ static int badness(struct task_struct *p) return points; } +#if defined(CONFIG_OOM_PANIC) && defined(CONFIG_OOM_KILLER) +#warning Only define OOM_PANIC or OOM_KILLER; not both +#endif + +#ifdef CONFIG_OOM_KILLER +/* + * Types of limitations to the nodes from which allocations may occur + */ +#define CONSTRAINT_NONE 1 +#define CONSTRAINT_MEMORY_POLICY 2 +#define CONSTRAINT_CPUSET 3 + +/* + * Determine the type of allocation constraint. + */ +static inline int constrained_alloc(struct zonelist *zonelist, gfp_t gfp_mask) +{ +#ifdef CONFIG_NUMA + struct zone **z; + nodemask_t nodes = node_online_map; + + for (z = zonelist->zones; *z; z++) + if (cpuset_zone_allowed(*z, gfp_mask)) + node_clear((*z)->zone_pgdat->node_id, + nodes); + else + return CONSTRAINT_CPUSET; + + if (!nodes_empty(nodes)) + return CONSTRAINT_MEMORY_POLICY; +#endif + + return CONSTRAINT_NONE; +} + /* * Simple selection loop. We chose the process with the highest * number of 'points'. We expect the caller will lock the tasklist. * * (not docbooked, we don't want this one cluttering up the manual) */ -static struct task_struct * select_bad_process(void) +static struct task_struct *select_bad_process(unsigned long *ppoints) { - int maxpoints = 0; struct task_struct *g, *p; struct task_struct *chosen = NULL; - - do_each_thread(g, p) - if (p->pid) { - int points = badness(p); - if (points > maxpoints) { + struct timespec uptime; + *ppoints = 0; + + do_posix_clock_monotonic_gettime(&uptime); + do_each_thread(g, p) { + unsigned long points; + int releasing; + + /* skip kernel threads */ + if (!p->mm) + continue; + + /* skip the init task with pid == 1 */ + if (p->pid == 1) + continue; + /* + * This is in the process of releasing memory so wait for it + * to finish before killing some other task by mistake. + * + * However, if p is the current task, we allow the 'kill' to + * go ahead if it is exiting: this will simply set TIF_MEMDIE, + * which will allow it to gain access to memory reserves in + * the process of exiting and releasing its resources. + * Otherwise we could get an OOM deadlock. + */ + releasing = test_tsk_thread_flag(p, TIF_MEMDIE) || + p->flags & PF_EXITING; + if (releasing) { + /* PF_DEAD tasks have already released their mm */ + if (p->flags & PF_DEAD) + continue; + if (p->flags & PF_EXITING && p == current) { chosen = p; - maxpoints = points; + *ppoints = ULONG_MAX; + break; } - if (p->flags & PF_SWAPOFF) - return p; + return ERR_PTR(-1UL); } - while_each_thread(g, p); + if (p->oomkilladj == OOM_DISABLE) + continue; + + points = badness(p, uptime.tv_sec); + if (points > *ppoints || !chosen) { + chosen = p; + *ppoints = points; + } + } while_each_thread(g, p); return chosen; } @@ -132,8 +259,14 @@ static struct task_struct * select_bad_process(void) * CAP_SYS_RAW_IO set, send SIGTERM instead (but it's unlikely that * we select a process with CAP_SYS_RAW_IO set). */ -static void __oom_kill_task(task_t *p) +static void __oom_kill_task(struct task_struct *p, const char *message) { + if (p->pid == 1) { + WARN_ON(1); + printk(KERN_WARNING "tried to kill init!\n"); + return; + } + task_lock(p); if (!p->mm || p->mm == &init_mm) { WARN_ON(1); @@ -142,7 +275,9 @@ static void __oom_kill_task(task_t *p) return; } task_unlock(p); - printk(KERN_ERR "Out of Memory: Killed process %d (%s).\n", p->pid, p->comm); + if (message) + printk(KERN_ERR "%s: Killed process %d (%s).\n", + message, p->pid, p->comm); /* * We give our sacrificial lamb high priority and access to @@ -150,86 +285,75 @@ static void __oom_kill_task(task_t *p) * exit() and clear out its resources quickly... */ p->time_slice = HZ; - p->flags |= PF_MEMALLOC | PF_MEMDIE; + set_tsk_thread_flag(p, TIF_MEMDIE); - /* This process has hardware access, be more careful. */ - if (cap_t(p->cap_effective) & CAP_TO_MASK(CAP_SYS_RAWIO)) { - force_sig(SIGTERM, p); - } else { - force_sig(SIGKILL, p); - } + force_sig(SIGKILL, p); } -static struct mm_struct *oom_kill_task(task_t *p) +static int oom_kill_task(struct task_struct *p, const char *message) { - struct mm_struct *mm = get_task_mm(p); - if (!mm || mm == &init_mm) - return NULL; - __oom_kill_task(p); - return mm; -} + struct mm_struct *mm; + struct task_struct *g, *q; + mm = p->mm; -/** - * oom_kill - kill the "best" process when we run out of memory - * - * If we run out of memory, we have the choice between either - * killing a random task (bad), letting the system crash (worse) - * OR try to be smart about which process to kill. Note that we - * don't have to be perfect here, we just have to be good. - */ -static void oom_kill(void) -{ - struct mm_struct *mm; - struct task_struct *g, *p, *q; - - read_lock(&tasklist_lock); -retry: - p = select_bad_process(); + /* WARNING: mm may not be dereferenced since we did not obtain its + * value from get_task_mm(p). This is OK since all we need to do is + * compare mm to q->mm below. + * + * Furthermore, even if mm contains a non-NULL value, p->mm may + * change to NULL at any time since we do not hold task_lock(p). + * However, this is of no concern to us. + */ - /* Found nothing?!?! Either we hang forever, or we panic. */ - if (!p) { - show_free_areas(); - panic("Out of memory and no killable processes...\n"); - } + if (mm == NULL || mm == &init_mm) + return 1; - mm = oom_kill_task(p); - if (!mm) - goto retry; + __oom_kill_task(p, message); /* * kill all processes that share the ->mm (i.e. all threads), * but are in a different thread group */ do_each_thread(g, q) if (q->mm == mm && q->tgid != p->tgid) - __oom_kill_task(q); + __oom_kill_task(q, message); while_each_thread(g, q); - if (!p->mm) - printk(KERN_INFO "Fixed up OOM kill of mm-less task\n"); - read_unlock(&tasklist_lock); - mmput(mm); - /* - * Make kswapd go out of the way, so "p" has a good chance of - * killing itself before someone else gets the chance to ask - * for more memory. - */ - yield(); - return; + return 0; } -/** - * out_of_memory - is the system out of memory? - */ -void out_of_memory(int gfp_mask) +static int oom_kill_process(struct task_struct *p, unsigned long points, + const char *message) { + struct task_struct *c; + struct list_head *tsk; + /* - * oom_lock protects out_of_memory()'s static variables. - * It's a global lock; this is not performance-critical. + * If the task is already exiting, don't alarm the sysadmin or kill + * its children or threads, just set TIF_MEMDIE so it can die quickly */ + if (p->flags & PF_EXITING) { + __oom_kill_task(p, NULL); + return 0; + } + + /* Try to kill a child first */ + list_for_each(tsk, &p->children) { + c = list_entry(tsk, struct task_struct, sibling); + if (c->mm == p->mm) + continue; + if (!oom_kill_task(c, message)) + return 0; + } + return oom_kill_task(p, message); +} + +int should_oom_kill(void) +{ static spinlock_t oom_lock = SPIN_LOCK_UNLOCKED; static unsigned long first, last, count, lastkill; unsigned long now, since; + int ret = 0; spin_lock(&oom_lock); now = jiffies; @@ -253,7 +377,7 @@ void out_of_memory(int gfp_mask) /* * If we have gotten only a few failures, - * we're not really oom. + * we're not really oom. */ if (++count < 10) goto out_unlock; @@ -271,24 +395,132 @@ void out_of_memory(int gfp_mask) * Ok, really out of memory. Kill something. */ lastkill = now; + ret = 1; - printk("oom-killer: gfp_mask=0x%x\n", gfp_mask); - show_free_areas(); +reset: +/* + * We dropped the lock above, so check to be sure the variable + * first only ever increases to prevent false OOM's. + */ + if (time_after(now, first)) + first = now; + count = 0; - /* oom_kill() sleeps */ +out_unlock: spin_unlock(&oom_lock); - oom_kill(); + return ret; +} + +/** + * out_of_memory - kill the "best" process when we run out of memory + * + * If we run out of memory, we have the choice between either + * killing a random task (bad), letting the system crash (worse) + * OR try to be smart about which process to kill. Note that we + * don't have to be perfect here, we just have to be good. + */ +void out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, int order) +{ + struct task_struct *p; + unsigned long points = 0; + + if (printk_ratelimit()) { + printk(KERN_WARNING "%s invoked oom-killer: " + "gfp_mask=0x%x, order=%d, oomkilladj=%d\n", + current->comm, gfp_mask, order, current->oomkilladj); + dump_stack(); + show_mem(); + } + + if (!should_oom_kill()) + return; + + cpuset_lock(); + read_lock(&tasklist_lock); + + /* + * Check if there were limitations on the allocation (only relevant for + * NUMA) that may require different handling. + */ + switch (constrained_alloc(zonelist, gfp_mask)) { + case CONSTRAINT_MEMORY_POLICY: + oom_kill_process(current, points, + "No available memory (MPOL_BIND)"); + break; + + case CONSTRAINT_CPUSET: + oom_kill_process(current, points, + "No available memory in cpuset"); + break; + + case CONSTRAINT_NONE: + if (sysctl_panic_on_oom) + panic("out of memory. panic_on_oom is selected\n"); +retry: + /* + * Rambo mode: Shoot down a process and hope it solves whatever + * issues we may have. + */ + p = select_bad_process(&points); + + if (PTR_ERR(p) == -1UL) + goto out; + + /* Found nothing?!?! Either we hang forever, or we panic. */ + if (!p) { + read_unlock(&tasklist_lock); + cpuset_unlock(); + panic("Out of memory and no killable processes...\n"); + } + + if (oom_kill_process(p, points, "Out of memory")) + goto retry; + + break; + } + +out: + read_unlock(&tasklist_lock); + cpuset_unlock(); + + /* + * Give "p" a good chance of killing itself before we + * retry to allocate memory unless "p" is current + */ + if (!test_thread_flag(TIF_MEMDIE)) + schedule_timeout_uninterruptible(1); +} +#endif /* CONFIG_OOM_KILLER */ + +#ifdef CONFIG_OOM_PANIC +/** + * out_of_memory - panic if the system out of memory? + */ +void out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, int order) +{ + /* + * oom_lock protects out_of_memory()'s static variables. + * It's a global lock; this is not performance-critical. + */ + static spinlock_t oom_lock = SPIN_LOCK_UNLOCKED; + static unsigned long count; + spin_lock(&oom_lock); -reset: /* - * We dropped the lock above, so check to be sure the variable - * first only ever increases to prevent false OOM's. + * If we have gotten only a few failures, + * we're not really oom. */ - if (time_after(now, first)) - first = now; - count = 0; + if (++count >= 10) { + /* + * Ok, really out of memory. Panic. + */ -out_unlock: + printk("oom-killer: gfp_mask=0x%x\n", gfp_mask); + show_free_areas(); + + panic("Out Of Memory"); + } spin_unlock(&oom_lock); } +#endif /* CONFIG_OOM_PANIC */