<feed xmlns='http://www.w3.org/2005/Atom'>
<title>linux/include/crypto, branch v2.6.34-rc2</title>
<subtitle>Linux kernel source tree</subtitle>
<id>https://git.amat.us/linux/atom/include/crypto?h=v2.6.34-rc2</id>
<link rel='self' href='https://git.amat.us/linux/atom/include/crypto?h=v2.6.34-rc2'/>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/'/>
<updated>2010-01-17T10:55:31Z</updated>
<entry>
<title>crypto: md5 - Add export support</title>
<updated>2010-01-17T10:55:31Z</updated>
<author>
<name>Max Vozeler</name>
<email>max@hinterhof.net</email>
</author>
<published>2010-01-17T10:55:31Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=7d6f75eb21b84cdc5dfb09789974f02b42a89058'/>
<id>urn:sha1:7d6f75eb21b84cdc5dfb09789974f02b42a89058</id>
<content type='text'>
This patch adds export/import support to md5. The exported type is
defined by struct md5_state.

This is modeled after the equivalent change to sha1_generic.

Signed-off-by: Max Vozeler &lt;max@hinterhof.net&gt;
Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>crypto: pcrypt - Add pcrypt crypto parallelization wrapper</title>
<updated>2010-01-07T04:57:19Z</updated>
<author>
<name>Steffen Klassert</name>
<email>steffen.klassert@secunet.com</email>
</author>
<published>2010-01-07T04:57:19Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=5068c7a883d1694f0fb3631b664827644153cd08'/>
<id>urn:sha1:5068c7a883d1694f0fb3631b664827644153cd08</id>
<content type='text'>
This patch adds a parallel crypto template that takes a crypto
algorithm and converts it to process the crypto transforms in
parallel. For the moment only aead algorithms are supported.

Signed-off-by: Steffen Klassert &lt;steffen.klassert@secunet.com&gt;
Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>crypto: hash - Remove legacy hash/digest code</title>
<updated>2009-10-19T03:53:37Z</updated>
<author>
<name>Benjamin Gilbert</name>
<email>bgilbert@cs.cmu.edu</email>
</author>
<published>2009-10-19T03:53:37Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=2141b6309b1fce535329c195cb5e5274a4c84ebc'/>
<id>urn:sha1:2141b6309b1fce535329c195cb5e5274a4c84ebc</id>
<content type='text'>
6941c3a0 disabled compilation of the legacy digest code but didn't
actually remove it.  Rectify this.  Also, remove the crypto_hash_type
extern declaration from algapi.h now that the struct is gone.

Signed-off-by: Benjamin Gilbert &lt;bgilbert@cs.cmu.edu&gt;
Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>crypto: ghash - Add PCLMULQDQ accelerated implementation</title>
<updated>2009-10-19T02:53:06Z</updated>
<author>
<name>Huang Ying</name>
<email>ying.huang@intel.com</email>
</author>
<published>2009-10-19T02:53:06Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=0e1227d356e9b2fe0500d6cc7084f752040a1e0e'/>
<id>urn:sha1:0e1227d356e9b2fe0500d6cc7084f752040a1e0e</id>
<content type='text'>
PCLMULQDQ is used to accelerate the most time-consuming part of GHASH,
carry-less multiplication. More information about PCLMULQDQ can be
found at:

http://software.intel.com/en-us/articles/carry-less-multiplication-and-its-usage-for-computing-the-gcm-mode/

Because PCLMULQDQ changes XMM state, its usage must be enclosed with
kernel_fpu_begin/end, which can be used only in process context, the
acceleration is implemented as crypto_ahash. That is, request in soft
IRQ context will be defered to the cryptd kernel thread.

Signed-off-by: Huang Ying &lt;ying.huang@intel.com&gt;
Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>Merge git://git.kernel.org/pub/scm/linux/kernel/git/herbert/crypto-2.6</title>
<updated>2009-09-11T16:38:37Z</updated>
<author>
<name>Linus Torvalds</name>
<email>torvalds@linux-foundation.org</email>
</author>
<published>2009-09-11T16:38:37Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=332a3392188e0ad966543c87b8da2b9d246f301d'/>
<id>urn:sha1:332a3392188e0ad966543c87b8da2b9d246f301d</id>
<content type='text'>
* git://git.kernel.org/pub/scm/linux/kernel/git/herbert/crypto-2.6: (102 commits)
  crypto: sha-s390 - Fix warnings in import function
  crypto: vmac - New hash algorithm for intel_txt support
  crypto: api - Do not displace newly registered algorithms
  crypto: ansi_cprng - Fix module initialization
  crypto: xcbc - Fix alignment calculation of xcbc_tfm_ctx
  crypto: fips - Depend on ansi_cprng
  crypto: blkcipher - Do not use eseqiv on stream ciphers
  crypto: ctr - Use chainiv on raw counter mode
  Revert crypto: fips - Select CPRNG
  crypto: rng - Fix typo
  crypto: talitos - add support for 36 bit addressing
  crypto: talitos - align locks on cache lines
  crypto: talitos - simplify hmac data size calculation
  crypto: mv_cesa - Add support for Orion5X crypto engine
  crypto: cryptd - Add support to access underlaying shash
  crypto: gcm - Use GHASH digest algorithm
  crypto: ghash - Add GHASH digest algorithm for GCM
  crypto: authenc - Convert to ahash
  crypto: api - Fix aligned ctx helper
  crypto: hmac - Prehash ipad/opad
  ...
</content>
</entry>
<entry>
<title>crypto: vmac - New hash algorithm for intel_txt support</title>
<updated>2009-09-02T10:05:22Z</updated>
<author>
<name>Shane Wang</name>
<email>shane.wang@intel.com</email>
</author>
<published>2009-09-02T10:05:22Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=f1939f7c56456d22a559d2c75156e91912a2e97e'/>
<id>urn:sha1:f1939f7c56456d22a559d2c75156e91912a2e97e</id>
<content type='text'>
This patch adds VMAC (a fast MAC) support into crypto framework.

Signed-off-by: Shane Wang &lt;shane.wang@intel.com&gt;
Signed-off-by: Joseph Cihula &lt;joseph.cihula@intel.com&gt;
Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>crypto: skcipher - Fix skcipher_dequeue_givcrypt NULL test</title>
<updated>2009-08-29T10:44:04Z</updated>
<author>
<name>Herbert Xu</name>
<email>herbert@gondor.apana.org.au</email>
</author>
<published>2009-08-29T10:44:04Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=0c7d400fafaeab6014504a6a6249f01bac7f7db4'/>
<id>urn:sha1:0c7d400fafaeab6014504a6a6249f01bac7f7db4</id>
<content type='text'>
As struct skcipher_givcrypt_request includes struct crypto_request
at a non-zero offset, testing for NULL after converting the pointer
returned by crypto_dequeue_request does not work.  This can result
in IPsec crashes when the queue is depleted.

This patch fixes it by doing the pointer conversion only when the
return value is non-NULL.  In particular, we create a new function
__crypto_dequeue_request that does the pointer conversion.

Reported-by: Brad Bosch &lt;bradbosch@comcast.net&gt;
Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>crypto: cryptd - Add support to access underlaying shash</title>
<updated>2009-08-06T05:35:20Z</updated>
<author>
<name>Huang Ying</name>
<email>ying.huang@intel.com</email>
</author>
<published>2009-08-06T05:35:20Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=ace1366369841c9c3a9788f79baa4d73f1c53107'/>
<id>urn:sha1:ace1366369841c9c3a9788f79baa4d73f1c53107</id>
<content type='text'>
cryptd_alloc_ahash() will allocate a cryptd-ed ahash for specified
algorithm name. The new allocated one is guaranteed to be cryptd-ed
ahash, so the shash underlying can be gotten via cryptd_ahash_child().

Signed-off-by: Huang Ying &lt;ying.huang@intel.com&gt;
Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>crypto: api - Fix aligned ctx helper</title>
<updated>2009-07-24T07:26:15Z</updated>
<author>
<name>Herbert Xu</name>
<email>herbert@gondor.apana.org.au</email>
</author>
<published>2009-07-24T07:26:15Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=ab300465676b0c0559af62d57ec9a902f5680b03'/>
<id>urn:sha1:ab300465676b0c0559af62d57ec9a902f5680b03</id>
<content type='text'>
The aligned ctx helper was using a bogus alignment value thas was
one off the correct value.  Fortunately the current users do not
require anything beyond the natural alignment of the platform so
this hasn't caused a problem.

This patch fixes that and also removes the unnecessary minimum
check since if the alignment is less than the natural alignment
then the subsequent ALIGN operation should be a noop.

Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
<entry>
<title>crypto: sha512_generic - Use 64-bit counters</title>
<updated>2009-07-22T06:38:13Z</updated>
<author>
<name>Herbert Xu</name>
<email>herbert@gondor.apana.org.au</email>
</author>
<published>2009-07-22T04:22:43Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/linux/commit/?id=13887ed6888dad1608eb9530ebd83b6ba29db577'/>
<id>urn:sha1:13887ed6888dad1608eb9530ebd83b6ba29db577</id>
<content type='text'>
This patch replaces the 32-bit counters in sha512_generic with
64-bit counters.  It also switches the bit count to the simpler
byte count.

Signed-off-by: Herbert Xu &lt;herbert@gondor.apana.org.au&gt;
</content>
</entry>
</feed>
