| /* | 
 |  * mm/thrash.c | 
 |  * | 
 |  * Copyright (C) 2004, Red Hat, Inc. | 
 |  * Copyright (C) 2004, Rik van Riel <riel@redhat.com> | 
 |  * Released under the GPL, see the file COPYING for details. | 
 |  * | 
 |  * Simple token based thrashing protection, using the algorithm | 
 |  * described in: http://www.cse.ohio-state.edu/hpcs/WWW/HTML/publications/abs05-1.html | 
 |  * | 
 |  * Sep 2006, Ashwin Chaugule <ashwin.chaugule@celunite.com> | 
 |  * Improved algorithm to pass token: | 
 |  * Each task has a priority which is incremented if it contended | 
 |  * for the token in an interval less than its previous attempt. | 
 |  * If the token is acquired, that task's priority is boosted to prevent | 
 |  * the token from bouncing around too often and to let the task make | 
 |  * some progress in its execution. | 
 |  */ | 
 |  | 
 | #include <linux/jiffies.h> | 
 | #include <linux/mm.h> | 
 | #include <linux/sched.h> | 
 | #include <linux/swap.h> | 
 | #include <linux/memcontrol.h> | 
 |  | 
 | #include <trace/events/vmscan.h> | 
 |  | 
 | #define TOKEN_AGING_INTERVAL	(0xFF) | 
 |  | 
 | static DEFINE_SPINLOCK(swap_token_lock); | 
 | struct mm_struct *swap_token_mm; | 
 | static struct mem_cgroup *swap_token_memcg; | 
 |  | 
 | #ifdef CONFIG_CGROUP_MEM_RES_CTLR | 
 | static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm) | 
 | { | 
 | 	struct mem_cgroup *memcg; | 
 |  | 
 | 	memcg = try_get_mem_cgroup_from_mm(mm); | 
 | 	if (memcg) | 
 | 		css_put(mem_cgroup_css(memcg)); | 
 |  | 
 | 	return memcg; | 
 | } | 
 | #else | 
 | static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm) | 
 | { | 
 | 	return NULL; | 
 | } | 
 | #endif | 
 |  | 
 | void grab_swap_token(struct mm_struct *mm) | 
 | { | 
 | 	int current_interval; | 
 | 	unsigned int old_prio = mm->token_priority; | 
 | 	static unsigned int global_faults; | 
 | 	static unsigned int last_aging; | 
 |  | 
 | 	global_faults++; | 
 |  | 
 | 	current_interval = global_faults - mm->faultstamp; | 
 |  | 
 | 	if (!spin_trylock(&swap_token_lock)) | 
 | 		return; | 
 |  | 
 | 	/* First come first served */ | 
 | 	if (!swap_token_mm) | 
 | 		goto replace_token; | 
 |  | 
 | 	/* | 
 | 	 * Usually, we don't need priority aging because long interval faults | 
 | 	 * makes priority decrease quickly. But there is one exception. If the | 
 | 	 * token owner task is sleeping, it never make long interval faults. | 
 | 	 * Thus, we need a priority aging mechanism instead. The requirements | 
 | 	 * of priority aging are | 
 | 	 *  1) An aging interval is reasonable enough long. Too short aging | 
 | 	 *     interval makes quick swap token lost and decrease performance. | 
 | 	 *  2) The swap token owner task have to get priority aging even if | 
 | 	 *     it's under sleep. | 
 | 	 */ | 
 | 	if ((global_faults - last_aging) > TOKEN_AGING_INTERVAL) { | 
 | 		swap_token_mm->token_priority /= 2; | 
 | 		last_aging = global_faults; | 
 | 	} | 
 |  | 
 | 	if (mm == swap_token_mm) { | 
 | 		mm->token_priority += 2; | 
 | 		goto update_priority; | 
 | 	} | 
 |  | 
 | 	if (current_interval < mm->last_interval) | 
 | 		mm->token_priority++; | 
 | 	else { | 
 | 		if (likely(mm->token_priority > 0)) | 
 | 			mm->token_priority--; | 
 | 	} | 
 |  | 
 | 	/* Check if we deserve the token */ | 
 | 	if (mm->token_priority > swap_token_mm->token_priority) | 
 | 		goto replace_token; | 
 |  | 
 | update_priority: | 
 | 	trace_update_swap_token_priority(mm, old_prio, swap_token_mm); | 
 |  | 
 | out: | 
 | 	mm->faultstamp = global_faults; | 
 | 	mm->last_interval = current_interval; | 
 | 	spin_unlock(&swap_token_lock); | 
 | 	return; | 
 |  | 
 | replace_token: | 
 | 	mm->token_priority += 2; | 
 | 	trace_replace_swap_token(swap_token_mm, mm); | 
 | 	swap_token_mm = mm; | 
 | 	swap_token_memcg = swap_token_memcg_from_mm(mm); | 
 | 	last_aging = global_faults; | 
 | 	goto out; | 
 | } | 
 |  | 
 | /* Called on process exit. */ | 
 | void __put_swap_token(struct mm_struct *mm) | 
 | { | 
 | 	spin_lock(&swap_token_lock); | 
 | 	if (likely(mm == swap_token_mm)) { | 
 | 		trace_put_swap_token(swap_token_mm); | 
 | 		swap_token_mm = NULL; | 
 | 		swap_token_memcg = NULL; | 
 | 	} | 
 | 	spin_unlock(&swap_token_lock); | 
 | } | 
 |  | 
 | static bool match_memcg(struct mem_cgroup *a, struct mem_cgroup *b) | 
 | { | 
 | 	if (!a) | 
 | 		return true; | 
 | 	if (!b) | 
 | 		return true; | 
 | 	if (a == b) | 
 | 		return true; | 
 | 	return false; | 
 | } | 
 |  | 
 | void disable_swap_token(struct mem_cgroup *memcg) | 
 | { | 
 | 	/* memcg reclaim don't disable unrelated mm token. */ | 
 | 	if (match_memcg(memcg, swap_token_memcg)) { | 
 | 		spin_lock(&swap_token_lock); | 
 | 		if (match_memcg(memcg, swap_token_memcg)) { | 
 | 			trace_disable_swap_token(swap_token_mm); | 
 | 			swap_token_mm = NULL; | 
 | 			swap_token_memcg = NULL; | 
 | 		} | 
 | 		spin_unlock(&swap_token_lock); | 
 | 	} | 
 | } |