<feed xmlns='http://www.w3.org/2005/Atom'>
<title>linux/mm, branch v2.6.22.13</title>
<subtitle>Linux kernel source tree</subtitle>
<id>https://git.amat.us/linux/atom/mm?h=v2.6.22.13</id>
<link rel='self' href='https://git.amat.us/linux/atom/mm?h=v2.6.22.13'/>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/'/>
<updated>2007-11-05T17:56:08Z</updated>
<entry>
<title>Revert "x86_64: allocate sparsemem memmap above 4G"</title>
<updated>2007-11-05T17:56:08Z</updated>
<author>
<name>Linus Torvalds</name>
<email>torvalds@linux-foundation.org</email>
</author>
<published>2007-10-29T18:36:04Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=572b445e0403ec0b7b6e3cc3a98588dbe660139a'/>
<id>urn:sha1:572b445e0403ec0b7b6e3cc3a98588dbe660139a</id>
<content type='text'>
patch 6a22c57b8d2a62dea7280a6b2ac807a539ef0716 in mainline.

This reverts commit 2e1c49db4c640b35df13889b86b9d62215ade4b6.

First off, testing in Fedora has shown it to cause boot failures,
bisected down by Martin Ebourne, and reported by Dave Jobes.  So the
commit will likely be reverted in the 2.6.23 stable kernels.

Secondly, in the 2.6.24 model, x86-64 has now grown support for
SPARSEMEM_VMEMMAP, which disables the relevant code anyway, so while the
bug is not visible any more, it's become invisible due to the code just
being irrelevant and no longer enabled on the only architecture that
this ever affected.

backported to 2.6.22 by Chuck Ebbert

Reported-by: Dave Jones &lt;davej@redhat.com&gt;
Tested-by: Martin Ebourne &lt;fedora@ebourne.me.uk&gt;
Cc: Zou Nan hai &lt;nanhai.zou@intel.com&gt;
Cc: Suresh Siddha &lt;suresh.b.siddha@intel.com&gt;
Cc: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Acked-by: Andy Whitcroft &lt;apw@shadowen.org&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
Cc: Chuck Ebbert &lt;cebbert@redhat.com&gt;
Signed-off-by: Greg Kroah-Hartman &lt;gregkh@suse.de&gt;


</content>
</entry>
<entry>
<title>do not limit locked memory when RLIMIT_MEMLOCK is RLIM_INFINITY</title>
<updated>2007-08-09T21:27:41Z</updated>
<author>
<name>Herbert van den Bergh</name>
<email>Herbert.van.den.Bergh@oracle.com</email>
</author>
<published>2007-07-16T06:38:25Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=ed173ab97b4faeeb2809a81c2fa0e6e4f979d5bb'/>
<id>urn:sha1:ed173ab97b4faeeb2809a81c2fa0e6e4f979d5bb</id>
<content type='text'>
Fix a bug in mm/mlock.c on 32-bit architectures that prevents a user from
locking more than 4GB of shared memory, or allocating more than 4GB of
shared memory in hugepages, when rlim[RLIMIT_MEMLOCK] is set to
RLIM_INFINITY.

Signed-off-by: Herbert van den Bergh &lt;herbert.van.den.bergh@oracle.com&gt;
Acked-by: Chris Mason &lt;chris.mason@oracle.com&gt;
Signed-off-by: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Signed-off-by: Greg Kroah-Hartman &lt;gregkh@suse.de&gt;

</content>
</entry>
<entry>
<title>hugetlb: fix race in alloc_fresh_huge_page()</title>
<updated>2007-08-09T21:27:40Z</updated>
<author>
<name>Joe Jin</name>
<email>joe.jin@oracle.com</email>
</author>
<published>2007-07-16T06:38:12Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=88656e13bea01b02f02a1dfcd3ba75a83e708648'/>
<id>urn:sha1:88656e13bea01b02f02a1dfcd3ba75a83e708648</id>
<content type='text'>
That static `nid' index needs locking.  Without it we can end up calling
alloc_pages_node() with an illegal node ID and the kernel crashes.

Acked-by: Gurudas Pai &lt;gurudas.pai@oracle.com&gt;
Signed-off-by: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Signed-off-by: Greg Kroah-Hartman &lt;gregkh@suse.de&gt;

</content>
</entry>
<entry>
<title>readahead: MIN_RA_PAGES/MAX_RA_PAGES macros</title>
<updated>2007-08-09T21:27:39Z</updated>
<author>
<name>Fengguang Wu</name>
<email>wfg@mail.ustc.edu.cn</email>
</author>
<published>2007-07-19T08:47:58Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=608fe2dd0715bc40d43e9b0c0fb7343e4fd7d719'/>
<id>urn:sha1:608fe2dd0715bc40d43e9b0c0fb7343e4fd7d719</id>
<content type='text'>
Define two convenient macros for read-ahead:
	- MAX_RA_PAGES: rounded down counterpart of VM_MAX_READAHEAD
	- MIN_RA_PAGES: rounded _up_ counterpart of VM_MIN_READAHEAD

Note that the rounded up MIN_RA_PAGES will work flawlessly with _large_
page sizes like 64k.

Signed-off-by: Fengguang Wu &lt;wfg@mail.ustc.edu.cn&gt;
Cc: Steven Pratt &lt;slpratt@austin.ibm.com&gt;
Cc: Ram Pai &lt;linuxram@us.ibm.com&gt;
Cc: Rusty Russell &lt;rusty@rustcorp.com.au&gt;
Signed-off-by: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Signed-off-by: Greg Kroah-Hartman &lt;gregkh@suse.de&gt;

</content>
</entry>
<entry>
<title>mm: double mark_page_accessed() in read_cache_page_async()</title>
<updated>2007-07-08T17:13:21Z</updated>
<author>
<name>Peter Zijlstra</name>
<email>peterz@infradead.org</email>
</author>
<published>2007-07-08T08:13:06Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=4e99325b462ba18075768582621af74a6b79d2a5'/>
<id>urn:sha1:4e99325b462ba18075768582621af74a6b79d2a5</id>
<content type='text'>
Fix a post-2.6.21 regression.

read_cache_page_async() has two invocations of mark_page_accessed() which will
launch pages right onto the active list.

Remove the first one, keeping the latter one.  This avoids marking unwanted
pages active (in the retry loop).

Signed-off-by: Peter Zijlstra &lt;a.p.zijlstra@chello.nl&gt;
Acked-by: Nick Piggin &lt;nickpiggin@yahoo.com.au&gt;
Signed-off-by: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
</content>
</entry>
<entry>
<title>slub: remove useless EXPORT_SYMBOL</title>
<updated>2007-07-06T18:45:11Z</updated>
<author>
<name>Christoph Lameter</name>
<email>clameter@sgi.com</email>
</author>
<published>2007-07-06T18:17:32Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=d23cf676d0e9d1fdc066d2dcb7f8bc0e2d4c75bd'/>
<id>urn:sha1:d23cf676d0e9d1fdc066d2dcb7f8bc0e2d4c75bd</id>
<content type='text'>
kmem_cache_open is static. EXPORT_SYMBOL was leftover from some earlier
time period where kmem_cache_open was usable outside of slub.

(Fixes powerpc build error)

Signed-off-by: Chrsitoph Lameter &lt;clameter@sgi.com&gt;
Cc: Johannes Berg &lt;johannes@sipsolutions.net&gt;
Cc: Benjamin Herrenschmidt &lt;benh@kernel.crashing.org&gt;
Cc: Paul Mackerras &lt;paulus@samba.org&gt;
Signed-off-by: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
</content>
</entry>
<entry>
<title>mm: fixup /proc/vmstat output</title>
<updated>2007-07-06T17:26:50Z</updated>
<author>
<name>Peter Zijlstra</name>
<email>peterz@infradead.org</email>
</author>
<published>2007-07-06T11:35:34Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=23c1fb52961bc24bd3a8078eefc49eed533b2b38'/>
<id>urn:sha1:23c1fb52961bc24bd3a8078eefc49eed533b2b38</id>
<content type='text'>
Line up the vmstat_text with zone_stat_item

enum zone_stat_item {
	/* First 128 byte cacheline (assuming 64 bit words) */
	NR_FREE_PAGES,
	NR_INACTIVE,
	NR_ACTIVE,

We current have nr_active and nr_inactive reversed.

[ "OK with patch, though using initializers canbe handy to prevent such
   things in future:

	static const char * const vmstat_text[] = {
		[NR_FREE_PAGES] = "nr_free_pages",
		..."
							 - Alexey ]

Signed-off-by: Peter Zijlstra &lt;a.p.zijlstra@chello.nl&gt;
Acked-by: Alexey Dobriyan &lt;adobriyan@gmail.com&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
</content>
</entry>
<entry>
<title>Fix slab redzone alignment</title>
<updated>2007-07-05T22:54:13Z</updated>
<author>
<name>David Woodhouse</name>
<email>dwmw2@infradead.org</email>
</author>
<published>2007-07-05T01:26:44Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=87a927c715789853cc8331d76039a2fd657a832a'/>
<id>urn:sha1:87a927c715789853cc8331d76039a2fd657a832a</id>
<content type='text'>
Commit b46b8f19c9cd435ecac4d9d12b39d78c137ecd66 fixed a couple of bugs
by switching the redzone to 64 bits. Unfortunately, it neglected to
ensure that the _second_ redzone, after the slab object, is aligned
correctly. This caused illegal instruction faults on sparc32, which for
some reason not entirely clear to me are not trapped and fixed up.

Two things need to be done to fix this:
  - increase the object size, rounding up to alignof(long long) so
    that the second redzone can be aligned correctly.
  - If SLAB_STORE_USER is set but alignof(long long)==8, allow a
    full 64 bits of space for the user word at the end of the buffer,
    even though we may not _use_ the whole 64 bits.

This patch should be a no-op on any 64-bit architecture or any 32-bit
architecture where alignof(long long) == 4. Of the others, it's tested
on ppc32 by myself and a very similar patch was tested on sparc32 by
Mark Fortescue, who reported the new problem.

Also, fix the conditions for FORCED_DEBUG, which hadn't been adjusted to
the new sizes. Again noticed by Mark.

Signed-off-by: David Woodhouse &lt;dwmw2@infradead.org&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
</content>
</entry>
<entry>
<title>SLUB: Make lockdep happy by not calling add_partial with interrupts enabled during bootstrap</title>
<updated>2007-07-03T20:56:13Z</updated>
<author>
<name>Christoph Lameter</name>
<email>clameter@sgi.com</email>
</author>
<published>2007-07-03T16:31:04Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=dbc55faa64c12f4c9fab6e2bd131d771bc026ed1'/>
<id>urn:sha1:dbc55faa64c12f4c9fab6e2bd131d771bc026ed1</id>
<content type='text'>
If we move the local_irq_enable() to the end of the function then
add_partial() in early_kmem_cache_node_alloc() will be called
with interrupts disabled like during regular operations.

This makes lockdep happy.

Signed-off-by: Christoph Lameter &lt;clameter@sgi.com&gt;
Tested-by: Andre Noll &lt;maan@systemlinux.org&gt;
Acked-by: Ingo Molnar &lt;mingo@elte.hu&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
</content>
</entry>
<entry>
<title>SLAB: remove WARN_ON_ONCE for zero sized objects for 2.6.22 release</title>
<updated>2007-07-01T19:29:43Z</updated>
<author>
<name>Christoph Lameter</name>
<email>clameter@sgi.com</email>
</author>
<published>2007-07-01T19:06:35Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=17022220dd70378b2f17fa9b0248f15067d88c58'/>
<id>urn:sha1:17022220dd70378b2f17fa9b0248f15067d88c58</id>
<content type='text'>
We agreed to remove the WARN_ON_ONCE before 2.6.22 is released.

Signed-off-by: Christoph Lameter &lt;clameter@sgi.com&gt;
Signed-off-by: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
</content>
</entry>
</feed>
