48 return _mm512_set_ps(b,a,b,a,b,a,b,a,b,a,b,a,b,a,b,a);
52 return _mm512_set1_ps(a);
56 return _mm512_set_pd(b,a,b,a,b,a,b,a);
60 return _mm512_set1_pd(a);
64 return _mm512_set1_epi32(a);
79 _mm512_store_si512((__m512i *)I,a);
87 _mm512_stream_ps(a,b);
92 _mm512_stream_pd(a,b);
113 return _mm512_set_ps( a[15],a[14],a[13],a[12],a[11],a[10],a[9],a[8],
114 a[7],a[6],a[5],a[4],a[3],a[2],a[1],a[0]);
118 return _mm512_set_pd(a[7],a[6],a[5],a[4],a[3],a[2],a[1],a[0]);
122 return _mm512_set_epi32( a[15],a[14],a[13],a[12],a[11],a[10],a[9],a[8],
123 a[7],a[6],a[5],a[4],a[3],a[2],a[1],a[0]);
128template <
typename Out_type,
typename In_type>
133 printf(
"Error, using wrong Reduce function\n");
146 return _mm512_add_ps(a,b);
150 return _mm512_add_pd(a,b);
154 return _mm512_add_epi32(a,b);
161 return _mm512_sub_ps(a,b);
165 return _mm512_sub_pd(a,b);
169 return _mm512_sub_epi32(a,b);
188 ymm0 = _mm512_moveldup_ps(a);
189 return _mm512_mul_ps(ymm0,b);
193 ymm0 = _mm512_shuffle_pd(a,a,0x00);
194 return _mm512_mul_pd(ymm0,b);
199 __m512 ymm0 = _mm512_moveldup_ps(a);
200 return _mm512_fmadd_ps( ymm0, b, c);
203 __m512d ymm0 = _mm512_shuffle_pd( a, a, 0x00 );
204 return _mm512_fmadd_pd( ymm0, b, c);
212 __m512 a_real = _mm512_moveldup_ps( a );
213 __m512 a_imag = _mm512_movehdup_ps( a );
214 a_imag = _mm512_mul_ps( a_imag, _mm512_permute_ps( b, 0xB1 ) );
215 return _mm512_fmaddsub_ps( a_real, b, a_imag );
219 __m512d a_real = _mm512_shuffle_pd( a, a, 0x00 );
220 __m512d a_imag = _mm512_shuffle_pd( a, a, 0xFF );
221 a_imag = _mm512_mul_pd( a_imag, _mm512_permute_pd( b, 0x55 ) );
222 return _mm512_fmaddsub_pd( a_real, b, a_imag );
228 inline void mac(__m512 &a, __m512 b, __m512 c){
229 a= _mm512_fmadd_ps( b, c, a);
231 inline void mac(__m512d &a, __m512d b, __m512d c){
232 a= _mm512_fmadd_pd( b, c, a);
236 return _mm512_mul_ps(a,b);
240 return _mm512_mul_pd(a,b);
244 return _mm512_mullo_epi32(a,b);
251 return _mm512_div_ps(a,b);
255 return _mm512_div_pd(a,b);
263 return _mm512_mask_sub_ps(in,0xaaaa,_mm512_setzero_ps(),in);
267 return _mm512_mask_sub_pd(in, 0xaa,_mm512_setzero_pd(), in);
278 return _mm512_mask_sub_ps(tmp,0xaaaa,_mm512_setzero_ps(),tmp);
284 __m512d tmp = _mm512_shuffle_pd(in,in,0x55);
285 return _mm512_mask_sub_pd(tmp,0xaa,_mm512_setzero_pd(),tmp);
293 return _mm512_mask_sub_ps(tmp,0x5555,_mm512_setzero_ps(),tmp);
297 __m512d tmp = _mm512_shuffle_pd(in,in,0x55);
298 return _mm512_mask_sub_pd(tmp,0x55,_mm512_setzero_pd(),tmp);
326 return _mm512_shuffle_pd(in,in,0x55);
335 static inline __m512i
StoH (__m512 a,__m512 b) {
338 __m256i ha = _mm512_cvtps_ph(a,0);
339 __m256i hb = _mm512_cvtps_ph(b,0);
340 h =(__m512i) _mm512_castps256_ps512((__m256)ha);
341 h =(__m512i) _mm512_insertf64x4((__m512d)h,(__m256d)hb,1);
348 static inline void HtoS (__m512i h,__m512 &sa,__m512 &sb) {
350 sa = _mm512_cvtph_ps((__m256i)_mm512_extractf64x4_pd((__m512d)h,0));
351 sb = _mm512_cvtph_ps((__m256i)_mm512_extractf64x4_pd((__m512d)h,1));
357 static inline __m512
DtoS (__m512d a,__m512d b) {
358 __m256 sa = _mm512_cvtpd_ps(a);
359 __m256 sb = _mm512_cvtpd_ps(b);
360 __m512 s = _mm512_castps256_ps512(sa);
361 s =(__m512) _mm512_insertf64x4((__m512d)s,(__m256d)sb,1);
365 static inline void StoD (__m512 s,__m512d &a,__m512d &b) {
366 a = _mm512_cvtps_pd((__m256)_mm512_extractf64x4_pd((__m512d)s,0));
367 b = _mm512_cvtps_pd((__m256)_mm512_extractf64x4_pd((__m512d)s,1));
370 static inline __m512i
DtoH (__m512d a,__m512d b,__m512d c,__m512d d) {
377 static inline void HtoD (__m512i h,__m512d &a,__m512d &b,__m512d &c,__m512d &d) {
389 static inline void Exchange0(__m512 &out1,__m512 &out2,__m512 in1,__m512 in2){
393 static inline void Exchange1(__m512 &out1,__m512 &out2,__m512 in1,__m512 in2){
399 static inline void Exchange2(__m512 &out1,__m512 &out2,__m512 in1,__m512 in2){
403 static inline void Exchange3(__m512 &out1,__m512 &out2,__m512 in1,__m512 in2){
410 static inline void Exchange0(__m512d &out1,__m512d &out2,__m512d in1,__m512d in2){
414 static inline void Exchange1(__m512d &out1,__m512d &out2,__m512d in1,__m512d in2){
420 static inline void Exchange2(__m512d &out1,__m512d &out2,__m512d in1,__m512d in2){
421 out1 = _mm512_shuffle_pd(in1,in2,0x00);
422 out2 = _mm512_shuffle_pd(in1,in2,0xFF);
424 static inline void Exchange3(__m512d &out1,__m512d &out2,__m512d in1,__m512d in2){
433 static inline __m512
rotate(__m512 in,
int n){
455 static inline __m512d
rotate(__m512d in,
int n){
469 template<
int n>
static inline __m512
tRotate(__m512 in){
470 return (__m512)_mm512_alignr_epi32((__m512i)in,(__m512i)in,n);
473 template<
int n>
static inline __m512d
tRotate(__m512d in){
474 return (__m512d)_mm512_alignr_epi64((__m512i)in,(__m512i)in,n);
486 return Grid::ComplexF(_mm512_mask_reduce_add_ps(0x5555, in),_mm512_mask_reduce_add_ps(0xAAAA, in));
491 return _mm512_reduce_add_ps(in);
497 return Grid::ComplexD(_mm512_mask_reduce_add_pd(0x55, in),_mm512_mask_reduce_add_pd(0xAA, in));
503 return _mm512_reduce_add_pd(in);
509 return _mm512_reduce_add_epi32(in);
524 for(
int i=0;i<size;i+=64){
525 _mm_prefetch(ptr+i+4096,_MM_HINT_T1);
526 _mm_prefetch(ptr+i+512,_MM_HINT_T0);
530 _mm_prefetch(ptr,_MM_HINT_T0);
538template <
typename S,
typename T>
using ReduceSIMD = Optimization::Reduce<S,T>;
541typedef Optimization::Sum
SumSIMD;
542typedef Optimization::Sub
SubSIMD;
544typedef Optimization::Div
DivSIMD;
Optimization::Vstream VstreamSIMD
Optimization::TimesMinusI TimesMinusISIMD
Optimization::MultComplex MultComplexSIMD
Optimization::TimesI TimesISIMD
Optimization::Reduce< S, T > ReduceSIMD
Optimization::Mult MultSIMD
Optimization::MaddRealPart MaddRealPartSIMD
Optimization::vecd SIMD_Dtype
Optimization::veci SIMD_Itype
Optimization::Vstore VstoreSIMD
Optimization::Conj ConjSIMD
Optimization::vecf SIMD_Ftype
Optimization::Vsplat VsplatSIMD
Optimization::Sum SumSIMD
Optimization::Sub SubSIMD
Optimization::Div DivSIMD
Optimization::MultRealPart MultRealPartSIMD
Optimization::Vset VsetSIMD
Optimization::vech SIMD_Htype
void prefetch_HINT_T0(const char *ptr)
void v_prefetch0(int size, const char *ptr)
Lattice< vobj > real(const Lattice< vobj > &lhs)
Lattice< vobj > imag(const Lattice< vobj > &lhs)
#define NAMESPACE_BEGIN(A)
#define _MM_SELECT_FOUR_FOUR(A, B, C, D)
static INTERNAL_PRECISION F
__m512 operator()(__m512 in)
__m512d operator()(__m512d in)
__m512 operator()(__m512 a, __m512 b)
__m512d operator()(__m512d a, __m512d b)
static void Exchange2(__m512d &out1, __m512d &out2, __m512d in1, __m512d in2)
static void Exchange2(__m512 &out1, __m512 &out2, __m512 in1, __m512 in2)
static void Exchange0(__m512d &out1, __m512d &out2, __m512d in1, __m512d in2)
static void Exchange1(__m512 &out1, __m512 &out2, __m512 in1, __m512 in2)
static void Exchange3(__m512d &out1, __m512d &out2, __m512d in1, __m512d in2)
static void Exchange3(__m512 &out1, __m512 &out2, __m512 in1, __m512 in2)
static void Exchange0(__m512 &out1, __m512 &out2, __m512 in1, __m512 in2)
static void Exchange1(__m512d &out1, __m512d &out2, __m512d in1, __m512d in2)
__m512d operator()(__m512d a, __m512d b, __m512d c)
__m512 operator()(__m512 a, __m512 b, __m512 c)
__m512 operator()(__m512 a, __m512 b)
__m512d operator()(__m512d a, __m512d b)
__m512 operator()(__m512 a, __m512 b)
__m512d operator()(__m512d a, __m512d b)
__m512 operator()(__m512 a, __m512 b)
void mac(__m512d &a, __m512d b, __m512d c)
void mac(__m512 &a, __m512 b, __m512 c)
__m512d operator()(__m512d a, __m512d b)
__m512i operator()(__m512i a, __m512i b)
static __m512d Permute0(__m512d in)
static __m512d Permute1(__m512d in)
static __m512 Permute2(__m512 in)
static __m512d Permute2(__m512d in)
static __m512 Permute3(__m512 in)
static __m512d Permute3(__m512d in)
static __m512 Permute1(__m512 in)
static __m512 Permute0(__m512 in)
static vech StoH(const vecf &sa, const vecf &sb)
static void HtoD(__m512i h, __m512d &a, __m512d &b, __m512d &c, __m512d &d)
static __m512i DtoH(__m512d a, __m512d b, __m512d c, __m512d d)
static void HtoS(__m512i h, __m512 &sa, __m512 &sb)
static void StoD(vecf s, vecd &a, vecd &b)
static vecf DtoS(vecd a, vecd b)
static __m512i StoH(__m512 a, __m512 b)
static __m512 DtoS(__m512d a, __m512d b)
static void HtoS(vech h, vecf &sa, vecf &sb)
static void StoD(__m512 s, __m512d &a, __m512d &b)
Out_type operator()(In_type in)
static vec< T > tRotate(vec< T > in)
static __m512 rotate(__m512 in, int n)
static __m512 tRotate(__m512 in)
static __m512d rotate(__m512d in, int n)
static __m512d tRotate(__m512d in)
__m512 operator()(__m512 a, __m512 b)
__m512i operator()(__m512i a, __m512i b)
__m512d operator()(__m512d a, __m512d b)
__m512 operator()(__m512 a, __m512 b)
__m512i operator()(__m512i a, __m512i b)
__m512d operator()(__m512d a, __m512d b)
__m512d operator()(__m512d in)
__m512 operator()(__m512 in)
__m512 operator()(__m512 in)
__m512d operator()(__m512d in)
__m512 operator()(float *a)
__m512d operator()(Grid::ComplexD *a)
__m512i operator()(Integer *a)
__m512d operator()(double *a)
__m512 operator()(Grid::ComplexF *a)
__m512d operator()(double a, double b)
__m512d operator()(double a)
__m512 operator()(float a)
__m512i operator()(Integer a)
__m512 operator()(float a, float b)
void operator()(__m512 a, float *F)
void operator()(__m512d a, double *D)
void operator()(__m512i a, Integer *I)
void operator()(float *a, __m512 b)
void operator()(double *a, __m512d b)