diff options
author | Craig Topper <craig.topper@gmail.com> | 2011-12-20 09:55:26 +0000 |
---|---|---|
committer | Craig Topper <craig.topper@gmail.com> | 2011-12-20 09:55:26 +0000 |
commit | 4c07c5dfebd270b2f0660e86f056eeafdb26a4fb (patch) | |
tree | 3c21d6c3d721e66804b3ae1ea2818735b4b740f2 /lib/Headers/avx2intrin.h | |
parent | 3ff53b3587862439a1a84a7b0bc9a7dbee0cf4eb (diff) |
Add AVX2 intrinsics for pavg, pblend, and pcmp instructions. Also remove unneeded builtins for SSE pcmp. Change SSE pcmpeqq and pcmpgtq to not use builtins and just use vector == and >.
git-svn-id: https://llvm.org/svn/llvm-project/cfe/trunk@146969 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'lib/Headers/avx2intrin.h')
-rw-r--r-- | lib/Headers/avx2intrin.h | 72 |
1 files changed, 72 insertions, 0 deletions
diff --git a/lib/Headers/avx2intrin.h b/lib/Headers/avx2intrin.h index df0450d7d1..813c602527 100644 --- a/lib/Headers/avx2intrin.h +++ b/lib/Headers/avx2intrin.h @@ -136,6 +136,78 @@ _mm256_andnot_si256(__m256i a, __m256i b) } static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_avg_epu8(__m256i a, __m256i b) +{ + return (__m256i)__builtin_ia32_pavgb256((__v32qi)a, (__v32qi)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_avg_epu16(__m256i a, __m256i b) +{ + return (__m256i)__builtin_ia32_pavgw256((__v16hi)a, (__v16hi)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_blendv_epi8(__m256i __V1, __m256i __V2, __m256i __M) +{ + return (__m256i)__builtin_ia32_pblendvb256((__v32qi)__V1, (__v32qi)__V2, + (__v32qi)__M); +} + +#define _mm256_blend_epi16(V1, V2, M) __extension__ ({ \ + __m256i __V1 = (V1); \ + __m256i __V2 = (V2); \ + (__m256i)__builtin_ia32_pblendw256((__v16hi)__V1, (__v16hi)__V2, M); }) + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpeq_epi8(__m256i a, __m256i b) +{ + return (__m256i)((__v32qi)a == (__v32qi)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpeq_epi16(__m256i a, __m256i b) +{ + return (__m256i)((__v16hi)a == (__v16hi)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpeq_epi32(__m256i a, __m256i b) +{ + return (__m256i)((__v8si)a == (__v8si)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpeq_epi64(__m256i a, __m256i b) +{ + return (__m256i)((__v4di)a == (__v4di)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpgt_epi8(__m256i a, __m256i b) +{ + return (__m256i)((__v32qi)a > (__v32qi)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpgt_epi16(__m256i a, __m256i b) +{ + return (__m256i)((__v16hi)a > (__v16hi)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpgt_epi32(__m256i a, __m256i b) +{ + return (__m256i)((__v8si)a > (__v8si)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) +_mm256_cmpgt_epi64(__m256i a, __m256i b) +{ + return (__m256i)((__v4di)a > (__v4di)b); +} + +static __inline__ __m256i __attribute__((__always_inline__, __nodebug__)) _mm256_or_si256(__m256i a, __m256i b) { return a | b; |