On large memory systems, the VM can spend way too much time scanningthrough pages that it cannot (or should not) evict from memory. Notonly does it use up CPU time, but it also provokes lock contentionand can leave large systems under memory presure in a catatonic state.

Against 2.6.26-rc2-mm1

This patch series improves VM scalability by:

1) putting filesystem backed, swap backed and non-reclaimable pages onto their own LRUs, so the system only scans the pages that it can/should evict from memory

2) switching to SEQ replacement for the anonymous LRUs, so the number of pages that need to be scanned when the system starts swapping is bound to a reasonable number

3) keeping non-reclaimable pages off the LRU completely, so the VM does not waste CPU time scanning them. Currently only ramfs and SHM_LOCKED pages are kept on the noreclaim list, mlock()ed VMAs will be added later

- pull the memcontrol lru arrayification earlier into the patch series- use a pagevec array similar to the lru array- clean up the code in various places- improved pageout balancing and reduced pageout cpu use

- fix compilation on PPC and without memcontrol- make page_is_pagecache more readable- replace get_scan_ratio with correct version

- merge memcontroller split LRU code into the main split LRU patch, since it is not functionally different (it was split up only to help people who had seen the last version of the patch series review it)- drop the page_file_cache debugging patch, since it never triggered- reintroduce code to not scan anon list if swap is full- add code to scan anon list if page cache is very small already- use lumpy reclaim more aggressively for smaller order > 1 allocations-- All Rights Reversed