<feed xmlns='http://www.w3.org/2005/Atom'>
<title>linux/include/net/sock.h, branch v3.4.33</title>
<subtitle>Linux kernel source tree</subtitle>
<id>https://git.amat.us/linux/atom/include/net/sock.h?h=v3.4.33</id>
<link rel='self' href='https://git.amat.us/linux/atom/include/net/sock.h?h=v3.4.33'/>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/'/>
<updated>2012-10-02T17:29:34Z</updated>
<entry>
<title>tcp: Apply device TSO segment limit earlier</title>
<updated>2012-10-02T17:29:34Z</updated>
<author>
<name>Ben Hutchings</name>
<email>bhutchings@solarflare.com</email>
</author>
<published>2012-07-30T16:11:42Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=4658b24b2dd0e4c6215db2203743fa999765e8a0'/>
<id>urn:sha1:4658b24b2dd0e4c6215db2203743fa999765e8a0</id>
<content type='text'>
[ Upstream commit 1485348d2424e1131ea42efc033cbd9366462b01 ]

Cache the device gso_max_segs in sock::sk_gso_max_segs and use it to
limit the size of TSO skbs.  This avoids the need to fall back to
software GSO for local TCP senders.

Signed-off-by: Ben Hutchings &lt;bhutchings@solarflare.com&gt;
Signed-off-by: David S. Miller &lt;davem@davemloft.net&gt;
Signed-off-by: Greg Kroah-Hartman &lt;gregkh@linuxfoundation.org&gt;
</content>
</entry>
<entry>
<title>net: fix sk_sockets_allocated_read_positive</title>
<updated>2012-04-30T17:37:59Z</updated>
<author>
<name>Eric Dumazet</name>
<email>edumazet@google.com</email>
</author>
<published>2012-04-28T23:21:56Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=518fbf9cdf17875d808596afd77fc115a6f942ca'/>
<id>urn:sha1:518fbf9cdf17875d808596afd77fc115a6f942ca</id>
<content type='text'>
Denys Fedoryshchenko reported frequent crashes on a proxy server and kindly
provided a lockdep report that explains it all :

  [  762.903868]
  [  762.903880] =================================
  [  762.903890] [ INFO: inconsistent lock state ]
  [  762.903903] 3.3.4-build-0061 #8 Not tainted
  [  762.904133] ---------------------------------
  [  762.904344] inconsistent {IN-SOFTIRQ-W} -&gt; {SOFTIRQ-ON-W} usage.
  [  762.904542] squid/1603 [HC0[0]:SC0[0]:HE1:SE1] takes:
  [  762.904542]  (key#3){+.?...}, at: [&lt;c0232cc4&gt;]
__percpu_counter_sum+0xd/0x58
  [  762.904542] {IN-SOFTIRQ-W} state was registered at:
  [  762.904542]   [&lt;c0158b84&gt;] __lock_acquire+0x284/0xc26
  [  762.904542]   [&lt;c01598e8&gt;] lock_acquire+0x71/0x85
  [  762.904542]   [&lt;c0349765&gt;] _raw_spin_lock+0x33/0x40
  [  762.904542]   [&lt;c0232c93&gt;] __percpu_counter_add+0x58/0x7c
  [  762.904542]   [&lt;c02cfde1&gt;] sk_clone_lock+0x1e5/0x200
  [  762.904542]   [&lt;c0303ee4&gt;] inet_csk_clone_lock+0xe/0x78
  [  762.904542]   [&lt;c0315778&gt;] tcp_create_openreq_child+0x1b/0x404
  [  762.904542]   [&lt;c031339c&gt;] tcp_v4_syn_recv_sock+0x32/0x1c1
  [  762.904542]   [&lt;c031615a&gt;] tcp_check_req+0x1fd/0x2d7
  [  762.904542]   [&lt;c0313f77&gt;] tcp_v4_do_rcv+0xab/0x194
  [  762.904542]   [&lt;c03153bb&gt;] tcp_v4_rcv+0x3b3/0x5cc
  [  762.904542]   [&lt;c02fc0c4&gt;] ip_local_deliver_finish+0x13a/0x1e9
  [  762.904542]   [&lt;c02fc539&gt;] NF_HOOK.clone.11+0x46/0x4d
  [  762.904542]   [&lt;c02fc652&gt;] ip_local_deliver+0x41/0x45
  [  762.904542]   [&lt;c02fc4d1&gt;] ip_rcv_finish+0x31a/0x33c
  [  762.904542]   [&lt;c02fc539&gt;] NF_HOOK.clone.11+0x46/0x4d
  [  762.904542]   [&lt;c02fc857&gt;] ip_rcv+0x201/0x23e
  [  762.904542]   [&lt;c02daa3a&gt;] __netif_receive_skb+0x319/0x368
  [  762.904542]   [&lt;c02dac07&gt;] netif_receive_skb+0x4e/0x7d
  [  762.904542]   [&lt;c02dacf6&gt;] napi_skb_finish+0x1e/0x34
  [  762.904542]   [&lt;c02db122&gt;] napi_gro_receive+0x20/0x24
  [  762.904542]   [&lt;f85d1743&gt;] e1000_receive_skb+0x3f/0x45 [e1000e]
  [  762.904542]   [&lt;f85d3464&gt;] e1000_clean_rx_irq+0x1f9/0x284 [e1000e]
  [  762.904542]   [&lt;f85d3926&gt;] e1000_clean+0x62/0x1f4 [e1000e]
  [  762.904542]   [&lt;c02db228&gt;] net_rx_action+0x90/0x160
  [  762.904542]   [&lt;c012a445&gt;] __do_softirq+0x7b/0x118
  [  762.904542] irq event stamp: 156915469
  [  762.904542] hardirqs last  enabled at (156915469): [&lt;c019b4f4&gt;]
__slab_alloc.clone.58.clone.63+0xc4/0x2de
  [  762.904542] hardirqs last disabled at (156915468): [&lt;c019b452&gt;]
__slab_alloc.clone.58.clone.63+0x22/0x2de
  [  762.904542] softirqs last  enabled at (156915466): [&lt;c02ce677&gt;]
lock_sock_nested+0x64/0x6c
  [  762.904542] softirqs last disabled at (156915464): [&lt;c0349914&gt;]
_raw_spin_lock_bh+0xe/0x45
  [  762.904542]
  [  762.904542] other info that might help us debug this:
  [  762.904542]  Possible unsafe locking scenario:
  [  762.904542]
  [  762.904542]        CPU0
  [  762.904542]        ----
  [  762.904542]   lock(key#3);
  [  762.904542]   &lt;Interrupt&gt;
  [  762.904542]     lock(key#3);
  [  762.904542]
  [  762.904542]  *** DEADLOCK ***
  [  762.904542]
  [  762.904542] 1 lock held by squid/1603:
  [  762.904542]  #0:  (sk_lock-AF_INET){+.+.+.}, at: [&lt;c03055c0&gt;]
lock_sock+0xa/0xc
  [  762.904542]
  [  762.904542] stack backtrace:
  [  762.904542] Pid: 1603, comm: squid Not tainted 3.3.4-build-0061 #8
  [  762.904542] Call Trace:
  [  762.904542]  [&lt;c0347b73&gt;] ? printk+0x18/0x1d
  [  762.904542]  [&lt;c015873a&gt;] valid_state+0x1f6/0x201
  [  762.904542]  [&lt;c0158816&gt;] mark_lock+0xd1/0x1bb
  [  762.904542]  [&lt;c015876b&gt;] ? mark_lock+0x26/0x1bb
  [  762.904542]  [&lt;c015805d&gt;] ? check_usage_forwards+0x77/0x77
  [  762.904542]  [&lt;c0158bf8&gt;] __lock_acquire+0x2f8/0xc26
  [  762.904542]  [&lt;c0159b8e&gt;] ? mark_held_locks+0x5d/0x7b
  [  762.904542]  [&lt;c0159cf6&gt;] ? trace_hardirqs_on+0xb/0xd
  [  762.904542]  [&lt;c0158dd4&gt;] ? __lock_acquire+0x4d4/0xc26
  [  762.904542]  [&lt;c01598e8&gt;] lock_acquire+0x71/0x85
  [  762.904542]  [&lt;c0232cc4&gt;] ? __percpu_counter_sum+0xd/0x58
  [  762.904542]  [&lt;c0349765&gt;] _raw_spin_lock+0x33/0x40
  [  762.904542]  [&lt;c0232cc4&gt;] ? __percpu_counter_sum+0xd/0x58
  [  762.904542]  [&lt;c0232cc4&gt;] __percpu_counter_sum+0xd/0x58
  [  762.904542]  [&lt;c02cebc4&gt;] __sk_mem_schedule+0xdd/0x1c7
  [  762.904542]  [&lt;c02d178d&gt;] ? __alloc_skb+0x76/0x100
  [  762.904542]  [&lt;c0305e8e&gt;] sk_wmem_schedule+0x21/0x2d
  [  762.904542]  [&lt;c0306370&gt;] sk_stream_alloc_skb+0x42/0xaa
  [  762.904542]  [&lt;c0306567&gt;] tcp_sendmsg+0x18f/0x68b
  [  762.904542]  [&lt;c031f3dc&gt;] ? ip_fast_csum+0x30/0x30
  [  762.904542]  [&lt;c0320193&gt;] inet_sendmsg+0x53/0x5a
  [  762.904542]  [&lt;c02cb633&gt;] sock_aio_write+0xd2/0xda
  [  762.904542]  [&lt;c015876b&gt;] ? mark_lock+0x26/0x1bb
  [  762.904542]  [&lt;c01a1017&gt;] do_sync_write+0x9f/0xd9
  [  762.904542]  [&lt;c01a2111&gt;] ? file_free_rcu+0x2f/0x2f
  [  762.904542]  [&lt;c01a17a1&gt;] vfs_write+0x8f/0xab
  [  762.904542]  [&lt;c01a284d&gt;] ? fget_light+0x75/0x7c
  [  762.904542]  [&lt;c01a1900&gt;] sys_write+0x3d/0x5e
  [  762.904542]  [&lt;c0349ec9&gt;] syscall_call+0x7/0xb
  [  762.904542]  [&lt;c0340000&gt;] ? rp_sidt+0x41/0x83

Bug is that sk_sockets_allocated_read_positive() calls
percpu_counter_sum_positive() without BH being disabled.

This bug was added in commit 180d8cd942ce33
(foundations of per-cgroup memory pressure controlling.), since previous
code was using percpu_counter_read_positive() which is IRQ safe.

In __sk_mem_schedule() we dont need the precise count of allocated
sockets and can revert to previous behavior.

Reported-by: Denys Fedoryshchenko &lt;denys@visp.net.lb&gt;
Sined-off-by: Eric Dumazet &lt;edumazet@google.com&gt;
Cc: Glauber Costa &lt;glommer@parallels.com&gt;
Acked-by: Neal Cardwell &lt;ncardwell@google.com&gt;

Signed-off-by: David S. Miller &lt;davem@davemloft.net&gt;
</content>
</entry>
<entry>
<title>net/sock.h: fix sk_peek_off kernel-doc warning</title>
<updated>2012-04-18T02:32:00Z</updated>
<author>
<name>Randy Dunlap</name>
<email>rdunlap@xenotime.net</email>
</author>
<published>2012-04-17T14:03:53Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=d3d4f0a025e621b82da08a76df4036d4267739dd'/>
<id>urn:sha1:d3d4f0a025e621b82da08a76df4036d4267739dd</id>
<content type='text'>
Fix kernel-doc warning in net/sock.h:

Warning(include/net/sock.h:377): No description found for parameter 'sk_peek_off'

Signed-off-by: Randy Dunlap &lt;rdunlap@xenotime.net&gt;
Signed-off-by: David S. Miller &lt;davem@davemloft.net&gt;
</content>
</entry>
<entry>
<title>poll: add poll_requested_events() and poll_does_not_wait() functions</title>
<updated>2012-03-23T23:58:38Z</updated>
<author>
<name>Hans Verkuil</name>
<email>hans.verkuil@cisco.com</email>
</author>
<published>2012-03-23T22:02:27Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=626cf236608505d376e4799adb4f7eb00a8594af'/>
<id>urn:sha1:626cf236608505d376e4799adb4f7eb00a8594af</id>
<content type='text'>
In some cases the poll() implementation in a driver has to do different
things depending on the events the caller wants to poll for.  An example
is when a driver needs to start a DMA engine if the caller polls for
POLLIN, but doesn't want to do that if POLLIN is not requested but instead
only POLLOUT or POLLPRI is requested.  This is something that can happen
in the video4linux subsystem among others.

Unfortunately, the current epoll/poll/select implementation doesn't
provide that information reliably.  The poll_table_struct does have it: it
has a key field with the event mask.  But once a poll() call matches one
or more bits of that mask any following poll() calls are passed a NULL
poll_table pointer.

Also, the eventpoll implementation always left the key field at ~0 instead
of using the requested events mask.

This was changed in eventpoll.c so the key field now contains the actual
events that should be polled for as set by the caller.

The solution to the NULL poll_table pointer is to set the qproc field to
NULL in poll_table once poll() matches the events, not the poll_table
pointer itself.  That way drivers can obtain the mask through a new
poll_requested_events inline.

The poll_table_struct can still be NULL since some kernel code calls it
internally (netfs_state_poll() in ./drivers/staging/pohmelfs/netfs.h).  In
that case poll_requested_events() returns ~0 (i.e.  all events).

Very rarely drivers might want to know whether poll_wait will actually
wait.  If another earlier file descriptor in the set already matched the
events the caller wanted to wait for, then the kernel will return from the
select() call without waiting.  This might be useful information in order
to avoid doing expensive work.

A new helper function poll_does_not_wait() is added that drivers can use
to detect this situation.  This is now used in sock_poll_wait() in
include/net/sock.h.  This was the only place in the kernel that needed
this information.

Drivers should no longer access any of the poll_table internals, but use
the poll_requested_events() and poll_does_not_wait() access functions
instead.  In order to enforce that the poll_table fields are now prepended
with an underscore and a comment was added warning against using them
directly.

This required a change in unix_dgram_poll() in unix/af_unix.c which used
the key field to get the requested events.  It's been replaced by a call
to poll_requested_events().

For qproc it was especially important to change its name since the
behavior of that field changes with this patch since this function pointer
can now be NULL when that wasn't possible in the past.

Any driver accessing the qproc or key fields directly will now fail to compile.

Some notes regarding the correctness of this patch: the driver's poll()
function is called with a 'struct poll_table_struct *wait' argument.  This
pointer may or may not be NULL, drivers can never rely on it being one or
the other as that depends on whether or not an earlier file descriptor in
the select()'s fdset matched the requested events.

There are only three things a driver can do with the wait argument:

1) obtain the key field:

	events = wait ? wait-&gt;key : ~0;

   This will still work although it should be replaced with the new
   poll_requested_events() function (which does exactly the same).
   This will now even work better, since wait is no longer set to NULL
   unnecessarily.

2) use the qproc callback. This could be deadly since qproc can now be
   NULL. Renaming qproc should prevent this from happening. There are no
   kernel drivers that actually access this callback directly, BTW.

3) test whether wait == NULL to determine whether poll would return without
   waiting. This is no longer sufficient as the correct test is now
   wait == NULL || wait-&gt;_qproc == NULL.

   However, the worst that can happen here is a slight performance hit in
   the case where wait != NULL and wait-&gt;_qproc == NULL. In that case the
   driver will assume that poll_wait() will actually add the fd to the set
   of waiting file descriptors. Of course, poll_wait() will not do that
   since it tests for wait-&gt;_qproc. This will not break anything, though.

   There is only one place in the whole kernel where this happens
   (sock_poll_wait() in include/net/sock.h) and that code will be replaced
   by a call to poll_does_not_wait() in the next patch.

   Note that even if wait-&gt;_qproc != NULL drivers cannot rely on poll_wait()
   actually waiting. The next file descriptor from the set might match the
   event mask and thus any possible waits will never happen.

Signed-off-by: Hans Verkuil &lt;hans.verkuil@cisco.com&gt;
Reviewed-by: Jonathan Corbet &lt;corbet@lwn.net&gt;
Reviewed-by: Al Viro &lt;viro@zeniv.linux.org.uk&gt;
Cc: Davide Libenzi &lt;davidel@xmailserver.org&gt;
Signed-off-by: Hans de Goede &lt;hdegoede@redhat.com&gt;
Cc: Mauro Carvalho Chehab &lt;mchehab@infradead.org&gt;
Cc: David Miller &lt;davem@davemloft.net&gt;
Cc: Eric Dumazet &lt;eric.dumazet@gmail.com&gt;
Signed-off-by: Andrew Morton &lt;akpm@linux-foundation.org&gt;
Signed-off-by: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
</content>
</entry>
<entry>
<title>Merge branch 'next' of git://git.kernel.org/pub/scm/linux/kernel/git/jmorris/linux-security</title>
<updated>2012-03-21T20:25:04Z</updated>
<author>
<name>Linus Torvalds</name>
<email>torvalds@linux-foundation.org</email>
</author>
<published>2012-03-21T20:25:04Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=3556485f1595e3964ba539e39ea682acbb835cee'/>
<id>urn:sha1:3556485f1595e3964ba539e39ea682acbb835cee</id>
<content type='text'>
Pull security subsystem updates for 3.4 from James Morris:
 "The main addition here is the new Yama security module from Kees Cook,
  which was discussed at the Linux Security Summit last year.  Its
  purpose is to collect miscellaneous DAC security enhancements in one
  place.  This also marks a departure in policy for LSM modules, which
  were previously limited to being standalone access control systems.
  Chromium OS is using Yama, and I believe there are plans for Ubuntu,
  at least.

  This patchset also includes maintenance updates for AppArmor, TOMOYO
  and others."

Fix trivial conflict in &lt;net/sock.h&gt; due to the jumo_label-&gt;static_key
rename.

* 'next' of git://git.kernel.org/pub/scm/linux/kernel/git/jmorris/linux-security: (38 commits)
  AppArmor: Fix location of const qualifier on generated string tables
  TOMOYO: Return error if fails to delete a domain
  AppArmor: add const qualifiers to string arrays
  AppArmor: Add ability to load extended policy
  TOMOYO: Return appropriate value to poll().
  AppArmor: Move path failure information into aa_get_name and rename
  AppArmor: Update dfa matching routines.
  AppArmor: Minor cleanup of d_namespace_path to consolidate error handling
  AppArmor: Retrieve the dentry_path for error reporting when path lookup fails
  AppArmor: Add const qualifiers to generated string tables
  AppArmor: Fix oops in policy unpack auditing
  AppArmor: Fix error returned when a path lookup is disconnected
  KEYS: testing wrong bit for KEY_FLAG_REVOKED
  TOMOYO: Fix mount flags checking order.
  security: fix ima kconfig warning
  AppArmor: Fix the error case for chroot relative path name lookup
  AppArmor: fix mapping of META_READ to audit and quiet flags
  AppArmor: Fix underflow in xindex calculation
  AppArmor: Fix dropping of allowed operations that are force audited
  AppArmor: Add mising end of structure test to caps unpacking
  ...
</content>
</entry>
<entry>
<title>Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-next</title>
<updated>2012-03-21T04:04:47Z</updated>
<author>
<name>Linus Torvalds</name>
<email>torvalds@linux-foundation.org</email>
</author>
<published>2012-03-21T04:04:47Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=3b59bf081622b6446db77ad06c93fe23677bc533'/>
<id>urn:sha1:3b59bf081622b6446db77ad06c93fe23677bc533</id>
<content type='text'>
Pull networking merge from David Miller:
 "1) Move ixgbe driver over to purely page based buffering on receive.
     From Alexander Duyck.

  2) Add receive packet steering support to e1000e, from Bruce Allan.

  3) Convert TCP MD5 support over to RCU, from Eric Dumazet.

  4) Reduce cpu usage in handling out-of-order TCP packets on modern
     systems, also from Eric Dumazet.

  5) Support the IP{,V6}_UNICAST_IF socket options, making the wine
     folks happy, from Erich Hoover.

  6) Support VLAN trunking from guests in hyperv driver, from Haiyang
     Zhang.

  7) Support byte-queue-limtis in r8169, from Igor Maravic.

  8) Outline code intended for IP_RECVTOS in IP_PKTOPTIONS existed but
     was never properly implemented, Jiri Benc fixed that.

  9) 64-bit statistics support in r8169 and 8139too, from Junchang Wang.

  10) Support kernel side dump filtering by ctmark in netfilter
      ctnetlink, from Pablo Neira Ayuso.

  11) Support byte-queue-limits in gianfar driver, from Paul Gortmaker.

  12) Add new peek socket options to assist with socket migration, from
      Pavel Emelyanov.

  13) Add sch_plug packet scheduler whose queue is controlled by
      userland daemons using explicit freeze and release commands.  From
      Shriram Rajagopalan.

  14) Fix FCOE checksum offload handling on transmit, from Yi Zou."

* git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-next: (1846 commits)
  Fix pppol2tp getsockname()
  Remove printk from rds_sendmsg
  ipv6: fix incorrent ipv6 ipsec packet fragment
  cpsw: Hook up default ndo_change_mtu.
  net: qmi_wwan: fix build error due to cdc-wdm dependecy
  netdev: driver: ethernet: Add TI CPSW driver
  netdev: driver: ethernet: add cpsw address lookup engine support
  phy: add am79c874 PHY support
  mlx4_core: fix race on comm channel
  bonding: send igmp report for its master
  fs_enet: Add MPC5125 FEC support and PHY interface selection
  net: bpf_jit: fix BPF_S_LDX_B_MSH compilation
  net: update the usage of CHECKSUM_UNNECESSARY
  fcoe: use CHECKSUM_UNNECESSARY instead of CHECKSUM_PARTIAL on tx
  net: do not do gso for CHECKSUM_UNNECESSARY in netif_needs_gso
  ixgbe: Fix issues with SR-IOV loopback when flow control is disabled
  net/hyperv: Fix the code handling tx busy
  ixgbe: fix namespace issues when FCoE/DCB is not enabled
  rtlwifi: Remove unused ETH_ADDR_LEN defines
  igbvf: Use ETH_ALEN
  ...

Fix up fairly trivial conflicts in drivers/isdn/gigaset/interface.c and
drivers/net/usb/{Kconfig,qmi_wwan.c} as per David.
</content>
</entry>
<entry>
<title>Merge branch 'for-3.4' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/cgroup</title>
<updated>2012-03-21T01:11:21Z</updated>
<author>
<name>Linus Torvalds</name>
<email>torvalds@linux-foundation.org</email>
</author>
<published>2012-03-21T01:11:21Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=0d9cabdccedb79ee5f27b77ff51f29a9e7d23275'/>
<id>urn:sha1:0d9cabdccedb79ee5f27b77ff51f29a9e7d23275</id>
<content type='text'>
Pull cgroup changes from Tejun Heo:
 "Out of the 8 commits, one fixes a long-standing locking issue around
  tasklist walking and others are cleanups."

* 'for-3.4' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/cgroup:
  cgroup: Walk task list under tasklist_lock in cgroup_enable_task_cg_list
  cgroup: Remove wrong comment on cgroup_enable_task_cg_list()
  cgroup: remove cgroup_subsys argument from callbacks
  cgroup: remove extra calls to find_existing_css_set
  cgroup: replace tasklist_lock with rcu_read_lock
  cgroup: simplify double-check locking in cgroup_attach_proc
  cgroup: move struct cgroup_pidlist out from the header file
  cgroup: remove cgroup_attach_task_current_cg()
</content>
</entry>
<entry>
<title>net: Add framework to allow sending packets with customized CRC.</title>
<updated>2012-02-24T09:37:35Z</updated>
<author>
<name>Ben Greear</name>
<email>greearb@candelatech.com</email>
</author>
<published>2012-02-11T15:39:30Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=3bdc0eba0b8b47797f4a76e377dd8360f317450f'/>
<id>urn:sha1:3bdc0eba0b8b47797f4a76e377dd8360f317450f</id>
<content type='text'>
This is useful for testing RX handling of frames with bad
CRCs.

Requires driver support to actually put the packet on the
wire properly.

Signed-off-by: Ben Greear &lt;greearb@candelatech.com&gt;
Tested-by: Aaron Brown &lt;aaron.f.brown@intel.com&gt;
Signed-off-by: Jeff Kirsher &lt;jeffrey.t.kirsher@intel.com&gt;
</content>
</entry>
<entry>
<title>static keys: Introduce 'struct static_key', static_key_true()/false() and static_key_slow_[inc|dec]()</title>
<updated>2012-02-24T09:05:59Z</updated>
<author>
<name>Ingo Molnar</name>
<email>mingo@elte.hu</email>
</author>
<published>2012-02-24T07:31:31Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=c5905afb0ee6550b42c49213da1c22d67316c194'/>
<id>urn:sha1:c5905afb0ee6550b42c49213da1c22d67316c194</id>
<content type='text'>
So here's a boot tested patch on top of Jason's series that does
all the cleanups I talked about and turns jump labels into a
more intuitive to use facility. It should also address the
various misconceptions and confusions that surround jump labels.

Typical usage scenarios:

        #include &lt;linux/static_key.h&gt;

        struct static_key key = STATIC_KEY_INIT_TRUE;

        if (static_key_false(&amp;key))
                do unlikely code
        else
                do likely code

Or:

        if (static_key_true(&amp;key))
                do likely code
        else
                do unlikely code

The static key is modified via:

        static_key_slow_inc(&amp;key);
        ...
        static_key_slow_dec(&amp;key);

The 'slow' prefix makes it abundantly clear that this is an
expensive operation.

I've updated all in-kernel code to use this everywhere. Note
that I (intentionally) have not pushed through the rename
blindly through to the lowest levels: the actual jump-label
patching arch facility should be named like that, so we want to
decouple jump labels from the static-key facility a bit.

On non-jump-label enabled architectures static keys default to
likely()/unlikely() branches.

Signed-off-by: Ingo Molnar &lt;mingo@elte.hu&gt;
Acked-by: Jason Baron &lt;jbaron@redhat.com&gt;
Acked-by: Steven Rostedt &lt;rostedt@goodmis.org&gt;
Cc: a.p.zijlstra@chello.nl
Cc: mathieu.desnoyers@efficios.com
Cc: davem@davemloft.net
Cc: ddaney.cavm@gmail.com
Cc: Linus Torvalds &lt;torvalds@linux-foundation.org&gt;
Link: http://lkml.kernel.org/r/20120222085809.GA26397@elte.hu
Signed-off-by: Ingo Molnar &lt;mingo@elte.hu&gt;
</content>
</entry>
<entry>
<title>sock: Introduce the SO_PEEK_OFF sock option</title>
<updated>2012-02-21T20:03:48Z</updated>
<author>
<name>Pavel Emelyanov</name>
<email>xemul@parallels.com</email>
</author>
<published>2012-02-21T07:31:34Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=ef64a54f6e558155b4f149bb10666b9e914b6c54'/>
<id>urn:sha1:ef64a54f6e558155b4f149bb10666b9e914b6c54</id>
<content type='text'>
This one specifies where to start MSG_PEEK-ing queue data from. When
set to negative value means that MSG_PEEK works as ususally -- peeks
from the head of the queue always.

When some bytes are peeked from queue and the peeking offset is non
negative it is moved forward so that the next peek will return next
portion of data.

When non-peeking recvmsg occurs and the peeking offset is non negative
is is moved backward so that the next peek will still peek the proper
data (i.e. the one that would have been picked if there were no non
peeking recv in between).

The offset is set using per-proto opteration to let the protocol handle
the locking issues and to check whether the peeking offset feature is
supported by the protocol the socket belongs to.

Signed-off-by: Pavel Emelyanov &lt;xemul@parallels.com&gt;
Signed-off-by: David S. Miller &lt;davem@davemloft.net&gt;
</content>
</entry>
</feed>
