<feed xmlns='http://www.w3.org/2005/Atom'>
<title>linux/fs/afs, branch v3.0.32</title>
<subtitle>Linux kernel source tree</subtitle>
<id>https://git.amat.us/linux/atom/fs/afs?h=v3.0.32</id>
<link rel='self' href='https://git.amat.us/linux/atom/fs/afs?h=v3.0.32'/>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/'/>
<updated>2012-03-23T18:20:51Z</updated>
<entry>
<title>afs: Remote abort can cause BUG in rxrpc code</title>
<updated>2012-03-23T18:20:51Z</updated>
<author>
<name>Anton Blanchard</name>
<email>anton@samba.org</email>
</author>
<published>2012-03-16T10:28:19Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=086b5909dfe7c79b68ed76d798a30dfc7c329852'/>
<id>urn:sha1:086b5909dfe7c79b68ed76d798a30dfc7c329852</id>
<content type='text'>
commit c0173863528a8c9212c53e080d63a1aaae5ef4f4 upstream.

When writing files to afs I sometimes hit a BUG:

kernel BUG at fs/afs/rxrpc.c:179!

With a backtrace of:

	afs_free_call
	afs_make_call
	afs_fs_store_data
	afs_vnode_store_data
	afs_write_back_from_locked_page
	afs_writepages_region
	afs_writepages

The cause is:

	ASSERT(skb_queue_empty(&amp;call-&gt;rx_queue));

Looking at a tcpdump of the session the abort happens because we
are exceeding our disk quota:

	rx abort fs reply store-data error diskquota exceeded (32)

So the abort error is valid. We hit the BUG because we haven't
freed all the resources for the call.

By freeing any skbs in call-&gt;rx_queue before calling afs_free_call
we avoid hitting leaking memory and avoid hitting the BUG.

Signed-off-by: Anton Blanchard &lt;anton@samba.org&gt;
Signed-off-by: David Howells &lt;dhowells@redhat.com&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
Signed-off-by: Greg Kroah-Hartman &lt;gregkh@linuxfoundation.org&gt;

</content>
</entry>
<entry>
<title>afs: Read of file returns EBADMSG</title>
<updated>2012-03-23T18:20:51Z</updated>
<author>
<name>Anton Blanchard</name>
<email>anton@samba.org</email>
</author>
<published>2012-03-16T10:28:07Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=4376d0751532595d76d8dbe9a7ddb6d92f08ff82'/>
<id>urn:sha1:4376d0751532595d76d8dbe9a7ddb6d92f08ff82</id>
<content type='text'>
commit 2c724fb92732c0b2a5629eb8af74e82eb62ac947 upstream.

A read of a large file on an afs mount failed:

# cat junk.file &gt; /dev/null
cat: junk.file: Bad message

Looking at the trace, call-&gt;offset wrapped since it is only an
unsigned short. In afs_extract_data:

        _enter("{%u},{%zu},%d,,%zu", call-&gt;offset, len, last, count);
...

        if (call-&gt;offset &lt; count) {
                if (last) {
                        _leave(" = -EBADMSG [%d &lt; %zu]", call-&gt;offset, count);
                        return -EBADMSG;
                }

Which matches the trace:

[cat   ] ==&gt; afs_extract_data({65132},{524},1,,65536)
[cat   ] &lt;== afs_extract_data() = -EBADMSG [0 &lt; 65536]

call-&gt;offset went from 65132 to 0. Fix this by making call-&gt;offset an
unsigned int.

Signed-off-by: Anton Blanchard &lt;anton@samba.org&gt;
Signed-off-by: David Howells &lt;dhowells@redhat.com&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
Signed-off-by: Greg Kroah-Hartman &lt;gregkh@linuxfoundation.org&gt;

</content>
</entry>
<entry>
<title>AFS: Use i_generation not i_version for the vnode uniquifier</title>
<updated>2011-06-16T15:44:48Z</updated>
<author>
<name>David Howells</name>
<email>dhowells@redhat.com</email>
</author>
<published>2011-06-13T23:45:44Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=d6e43f751f252c68ca69fa6d18665d88d69ef8b7'/>
<id>urn:sha1:d6e43f751f252c68ca69fa6d18665d88d69ef8b7</id>
<content type='text'>
Store the AFS vnode uniquifier in the i_generation field, not the i_version
field of the inode struct.  i_version can then be given the AFS data version
number.

Signed-off-by: David Howells &lt;dhowells@redhat.com&gt;
Signed-off-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
</content>
</entry>
<entry>
<title>AFS: Set s_id in the superblock to the volume name</title>
<updated>2011-06-16T15:44:47Z</updated>
<author>
<name>David Howells</name>
<email>dhowells@redhat.com</email>
</author>
<published>2011-06-13T23:38:44Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=2e41ae225f742ded5b7d9847cd8bd605f27daba8'/>
<id>urn:sha1:2e41ae225f742ded5b7d9847cd8bd605f27daba8</id>
<content type='text'>
Set s_id in the superblock to the name of the AFS volume that this superblock
corresponds to.

Signed-off-by: David Howells &lt;dhowells@redhat.com&gt;
Signed-off-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
</content>
</entry>
<entry>
<title>afs: afs_fill_page reads too much, or wrong data</title>
<updated>2011-06-16T15:44:46Z</updated>
<author>
<name>Anton Blanchard</name>
<email>anton@samba.org</email>
</author>
<published>2011-06-13T21:31:12Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=5e7f23373bf9a853e9256e81e86724cdd0a33c29'/>
<id>urn:sha1:5e7f23373bf9a853e9256e81e86724cdd0a33c29</id>
<content type='text'>
afs_fill_page should read the page that is about to be written but
the current implementation has a number of issues. If we aren't
extending the file we always read PAGE_CACHE_SIZE at offset 0. If we
are extending the file we try to read the entire file.

Change afs_fill_page to read PAGE_CACHE_SIZE at the right offset,
clamped to i_size.

While here, avoid calling afs_fill_page when we are doing a
PAGE_CACHE_SIZE write.

Signed-off-by: Anton Blanchard &lt;anton@samba.org&gt;
Signed-off-by: David Howells &lt;dhowells@redhat.com&gt;
Signed-off-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
</content>
</entry>
<entry>
<title>afs: fix sget() races, close leak on umount</title>
<updated>2011-06-12T21:45:36Z</updated>
<author>
<name>Al Viro</name>
<email>viro@zeniv.linux.org.uk</email>
</author>
<published>2011-06-12T20:01:21Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=dde194a64bb5c3fd05d965775dc92e8a4920a53a'/>
<id>urn:sha1:dde194a64bb5c3fd05d965775dc92e8a4920a53a</id>
<content type='text'>
* set -&gt;s_fs_info in set() callback passed to sget()
* allocate the thing and set it up enough for afs_test_super() before
making it visible
* have it freed in -&gt;kill_sb() (current tree simply leaks it)
* have -&gt;put_super() leave -&gt;s_fs_info-&gt;volume alone; it's too early for
dropping it; do that from -&gt;kill_sb() after having called kill_anon_super().

Signed-off-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
</content>
</entry>
<entry>
<title>afs: remove unnecessary dentry_unhash on rmdir, dir rename</title>
<updated>2011-05-28T05:02:53Z</updated>
<author>
<name>Sage Weil</name>
<email>sage@newdream.net</email>
</author>
<published>2011-05-27T20:42:13Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=705da2de95ce79199a52f21d82dd9f36b89afd12'/>
<id>urn:sha1:705da2de95ce79199a52f21d82dd9f36b89afd12</id>
<content type='text'>
afs has no problems with references to unlinked directories.

CC: David Howells &lt;dhowells@redhat.com&gt;
CC: linux-afs@lists.infradead.org
Signed-off-by: Sage Weil &lt;sage@newdream.net&gt;
Signed-off-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
</content>
</entry>
<entry>
<title>vfs: push dentry_unhash on rename_dir into file systems</title>
<updated>2011-05-26T11:26:48Z</updated>
<author>
<name>Sage Weil</name>
<email>sage@newdream.net</email>
</author>
<published>2011-05-24T20:06:07Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=e4eaac06bcccb2a70bca6a2de9871882dce2aa14'/>
<id>urn:sha1:e4eaac06bcccb2a70bca6a2de9871882dce2aa14</id>
<content type='text'>
Only a few file systems need this.  Start by pushing it down into each
rename method (except gfs2 and xfs) so that it can be dealt with on a
per-fs basis.

Acked-by: Christoph Hellwig &lt;hch@lst.de&gt;
Signed-off-by: Sage Weil &lt;sage@newdream.net&gt;
Signed-off-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
</content>
</entry>
<entry>
<title>vfs: push dentry_unhash on rmdir into file systems</title>
<updated>2011-05-26T11:26:47Z</updated>
<author>
<name>Sage Weil</name>
<email>sage@newdream.net</email>
</author>
<published>2011-05-24T20:06:06Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=79bf7c732b5ff75b96022ed9d29181afd3d2509c'/>
<id>urn:sha1:79bf7c732b5ff75b96022ed9d29181afd3d2509c</id>
<content type='text'>
Only a few file systems need this.  Start by pushing it down into each
fs rmdir method (except gfs2 and xfs) so it can be dealt with on a per-fs
basis.

This does not change behavior for any in-tree file systems.

Acked-by: Christoph Hellwig &lt;hch@lst.de&gt;
Signed-off-by: Sage Weil &lt;sage@newdream.net&gt;
Signed-off-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
</content>
</entry>
<entry>
<title>Fix common misspellings</title>
<updated>2011-03-31T14:26:23Z</updated>
<author>
<name>Lucas De Marchi</name>
<email>lucas.demarchi@profusion.mobi</email>
</author>
<published>2011-03-31T01:57:33Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=25985edcedea6396277003854657b5f3cb31a628'/>
<id>urn:sha1:25985edcedea6396277003854657b5f3cb31a628</id>
<content type='text'>
Fixes generated by 'codespell' and manually reviewed.

Signed-off-by: Lucas De Marchi &lt;lucas.demarchi@profusion.mobi&gt;
</content>
</entry>
</feed>
