<feed xmlns='http://www.w3.org/2005/Atom'>
<title>llvm/test/CodeGen/CellSPU, branch master</title>
<subtitle>http://llvm.org</subtitle>
<id>https://git.amat.us/llvm/atom/test/CodeGen/CellSPU?h=master</id>
<link rel='self' href='https://git.amat.us/llvm/atom/test/CodeGen/CellSPU?h=master'/>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/'/>
<updated>2012-11-14T22:09:20Z</updated>
<entry>
<title>Remove the CellSPU port.</title>
<updated>2012-11-14T22:09:20Z</updated>
<author>
<name>Eric Christopher</name>
<email>echristo@gmail.com</email>
</author>
<published>2012-11-14T22:09:20Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=06b423452c85f5a78a1b0555b767cf27b36c0752'/>
<id>urn:sha1:06b423452c85f5a78a1b0555b767cf27b36c0752</id>
<content type='text'>
Approved by Chris Lattner.


git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@167984 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>Fix broken tests.</title>
<updated>2012-10-02T15:49:34Z</updated>
<author>
<name>Benjamin Kramer</name>
<email>benny.kra@googlemail.com</email>
</author>
<published>2012-10-02T15:49:34Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=fba80d9e97685f4aff959a6ba9fb076536bf285e'/>
<id>urn:sha1:fba80d9e97685f4aff959a6ba9fb076536bf285e</id>
<content type='text'>
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@165019 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>Add test triples to fix win32 failures.  Revert workaround from r161292.</title>
<updated>2012-08-08T20:31:37Z</updated>
<author>
<name>Bob Wilson</name>
<email>bob.wilson@apple.com</email>
</author>
<published>2012-08-08T20:31:37Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=5f91a99427c0be5899835fc6407c6e55bfcd606d'/>
<id>urn:sha1:5f91a99427c0be5899835fc6407c6e55bfcd606d</id>
<content type='text'>
I don't have a win32 system to test, so hopefully I got them all fixed here.

git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@161519 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>Refactor and check "onlyReadsMemory" before optimizing builtins.</title>
<updated>2012-08-03T23:29:17Z</updated>
<author>
<name>Bob Wilson</name>
<email>bob.wilson@apple.com</email>
</author>
<published>2012-08-03T23:29:17Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=53624a2df557b4a24f2ee98cfce1a69bf83243af'/>
<id>urn:sha1:53624a2df557b4a24f2ee98cfce1a69bf83243af</id>
<content type='text'>
This patch is mostly just refactoring a bunch of copy-and-pasted code, but
it also adds a check that the call instructions are readnone or readonly.
That check was already present for sin, cos, sqrt, log2, and exp2 calls, but
it was missing for the rest of the builtins being handled in this code.

git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@161282 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>Implement r160312 as target indepedenet dag combine.</title>
<updated>2012-07-17T08:31:11Z</updated>
<author>
<name>Evan Cheng</name>
<email>evan.cheng@apple.com</email>
</author>
<published>2012-07-17T08:31:11Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=f5c0539092996771824893309f311378e719e32e'/>
<id>urn:sha1:f5c0539092996771824893309f311378e719e32e</id>
<content type='text'>
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@160354 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>Convert all tests using TCL-style quoting to use shell-style quoting.</title>
<updated>2012-07-02T12:47:22Z</updated>
<author>
<name>Chandler Carruth</name>
<email>chandlerc@gmail.com</email>
</author>
<published>2012-07-02T12:47:22Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=4177e6fff50552908bab510f1e896fa974a6f155'/>
<id>urn:sha1:4177e6fff50552908bab510f1e896fa974a6f155</id>
<content type='text'>
This was done through the aid of a terrible Perl creation. I will not
paste any of the horrors here. Suffice to say, it require multiple
staged rounds of replacements, state carried between, and a few
nested-construct-parsing hacks that I'm not proud of. It happens, by
luck, to be able to deal with all the TCL-quoting patterns in evidence
in the LLVM test suite.

If anyone is maintaining large out-of-tree test trees, feel free to poke
me and I'll send you the steps I used to convert things, as well as
answer any painful questions etc. IRC works best for this type of thing
I find.

Once converted, switch the LLVM lit config to use ShTests the same as
Clang. In addition to being able to delete large amounts of Python code
from 'lit', this will also simplify the entire test suite and some of
lit's architecture.

Finally, the test suite runs 33% faster on Linux now. ;]
For my 16-hardware-thread (2x 4-core xeon e5520): 36s -&gt; 24s

git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@159525 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>Rewrite three tests that had truly egregious abuses of 'grep' in them to</title>
<updated>2012-07-02T12:20:14Z</updated>
<author>
<name>Chandler Carruth</name>
<email>chandlerc@gmail.com</email>
</author>
<published>2012-07-02T12:20:14Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=38adf0bdaadfb042d336764a2a922768a22c044a'/>
<id>urn:sha1:38adf0bdaadfb042d336764a2a922768a22c044a</id>
<content type='text'>
use FileCheck.

Aside from removing a dependence on TCL-style quoting, this also makes
the tests ... significantly more robust. =] It would be really, *really*
great of the maintainer(s) of the CellSPU backend went through and
systematically rewrite these tests to use FileCheck. There are a lot
more that have nearly this bad of abuses.

Another step along the path to a TclTest-free testsuite.

git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@159523 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>Flip the new block-placement pass to be on by default.</title>
<updated>2012-04-16T13:49:17Z</updated>
<author>
<name>Chandler Carruth</name>
<email>chandlerc@gmail.com</email>
</author>
<published>2012-04-16T13:49:17Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=9e67db4af13abb967cae5858502207a43d26bf84'/>
<id>urn:sha1:9e67db4af13abb967cae5858502207a43d26bf84</id>
<content type='text'>
This is mostly to test the waters. I'd like to get results from FNT
build bots and other bots running on non-x86 platforms.

This feature has been pretty heavily tested over the last few months by
me, and it fixes several of the execution time regressions caused by the
inlining work by preventing inlining decisions from radically impacting
block layout.

I've seen very large improvements in yacr2 and ackermann benchmarks,
along with the expected noise across all of the benchmark suite whenever
code layout changes. I've analyzed all of the regressions and fixed
them, or found them to be impossible to fix. See my email to llvmdev for
more details.

I'd like for this to be in 3.1 as it complements the inliner changes,
but if any failures are showing up or anyone has concerns, it is just
a flag flip and so can be easily turned off.

I'm switching it on tonight to try and get at least one run through
various folks' performance suites in case SPEC or something else has
serious issues with it. I'll watch bots and revert if anything shows up.

git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@154816 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>1. Remove the part of r153848 which optimizes shuffle-of-shuffle into a new</title>
<updated>2012-04-07T21:19:08Z</updated>
<author>
<name>Nadav Rotem</name>
<email>nadav.rotem@intel.com</email>
</author>
<published>2012-04-07T21:19:08Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=d16c8d0d336638225378466bc17c9db156401817'/>
<id>urn:sha1:d16c8d0d336638225378466bc17c9db156401817</id>
<content type='text'>
   shuffle node because it could introduce new shuffle nodes that were not
   supported efficiently by the target.

2. Add a more restrictive shuffle-of-shuffle optimization for cases where the
   second shuffle reverses the transformation of the first shuffle.



git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@154266 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
<entry>
<title>This commit contains a few changes that had to go in together.</title>
<updated>2012-04-01T19:31:22Z</updated>
<author>
<name>Nadav Rotem</name>
<email>nadav.rotem@intel.com</email>
</author>
<published>2012-04-01T19:31:22Z</published>
<link rel='alternate' type='text/html' href='https://git.amat.us/llvm/commit/?id=4ac9081c7110dbf099d682fa51c091741e763491'/>
<id>urn:sha1:4ac9081c7110dbf099d682fa51c091741e763491</id>
<content type='text'>
1. Simplify xor/and/or (bitcast(A), bitcast(B)) -&gt; bitcast(op (A,B))
   (and also scalar_to_vector).

2. Xor/and/or are indifferent to the swizzle operation (shuffle of one src).
   Simplify xor/and/or (shuff(A), shuff(B)) -&gt; shuff(op (A, B))

3. Optimize swizzles of shuffles:  shuff(shuff(x, y), undef) -&gt; shuff(x, y).

4. Fix an X86ISelLowering optimization which was very bitcast-sensitive.

Code which was previously compiled to this:

movd    (%rsi), %xmm0
movdqa  .LCPI0_0(%rip), %xmm2
pshufb  %xmm2, %xmm0
movd    (%rdi), %xmm1
pshufb  %xmm2, %xmm1
pxor    %xmm0, %xmm1
pshufb  .LCPI0_1(%rip), %xmm1
movd    %xmm1, (%rdi)
ret

Now compiles to this:

movl    (%rsi), %eax
xorl    %eax, (%rdi)
ret




git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@153848 91177308-0d34-0410-b5e6-96231b3b80d8
</content>
</entry>
</feed>
