Grid 0.7.0
Grid_avx.h
Go to the documentation of this file.
1/*************************************************************************************
2
3 Grid physics library, www.github.com/paboyle/Grid
4
5 Source file: ./lib/simd/Grid_avx.h
6
7 Copyright (C) 2015
8
9Author: Azusa Yamaguchi <ayamaguc@staffmail.ed.ac.uk>
10Author: Guido Cossu <cossu@iroiro-pc.kek.jp>
11Author: Peter Boyle <paboyle@ph.ed.ac.uk>
12Author: neo <cossu@post.kek.jp>
13Author: paboyle <paboyle@ph.ed.ac.uk>
14
15 This program is free software; you can redistribute it and/or modify
16 it under the terms of the GNU General Public License as published by
17 the Free Software Foundation; either version 2 of the License, or
18 (at your option) any later version.
19
20 This program is distributed in the hope that it will be useful,
21 but WITHOUT ANY WARRANTY; without even the implied warranty of
22 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
23 GNU General Public License for more details.
24
25 You should have received a copy of the GNU General Public License along
26 with this program; if not, write to the Free Software Foundation, Inc.,
27 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
28
29 See the full license in the file "LICENSE" in the top level distribution directory
30*************************************************************************************/
31/* END LEGAL */
32#include <immintrin.h>
33#ifdef AVXFMA4
34#include <x86intrin.h>
35#endif
36// _mm256_set_m128i(hi,lo); // not defined in all versions of immintrin.h
37#ifndef _mm256_set_m128i
38#define _mm256_set_m128i(hi,lo) _mm256_insertf128_si256(_mm256_castsi128_si256(lo),(hi),1)
39#endif
40
42NAMESPACE_BEGIN(Optimization);
43
44template<class vtype>
45union uconv {
46 __m256 f;
47 vtype v;
48};
49
50union u256f {
51 __m256 v;
52 float f[8];
53};
54
55union u256d {
56 __m256d v;
57 double f[4];
58};
59
60struct Vsplat{
61 // Complex float
62 inline __m256 operator()(float a, float b) {
63 return _mm256_set_ps(b,a,b,a,b,a,b,a);
64 }
65 // Real float
66 inline __m256 operator()(float a){
67 return _mm256_set_ps(a,a,a,a,a,a,a,a);
68 }
69 //Complex double
70 inline __m256d operator()(double a, double b){
71 return _mm256_set_pd(b,a,b,a);
72 }
73 //Real double
74 inline __m256d operator()(double a){
75 return _mm256_set_pd(a,a,a,a);
76 }
77 //Integer
78 inline __m256i operator()(Integer a){
79 return _mm256_set1_epi32(a);
80 }
81};
82
83struct Vstore{
84 //Float
85 inline void operator()(__m256 a, float* F){
86 _mm256_store_ps(F,a);
87 }
88 //Double
89 inline void operator()(__m256d a, double* D){
90 _mm256_store_pd(D,a);
91 }
92 //Integer
93 inline void operator()(__m256i a, Integer* I){
94 _mm256_store_si256((__m256i*)I,a);
95 }
96
97};
98
99struct Vstream{
100 //Float
101 inline void operator()(float * a, __m256 b){
102 _mm256_stream_ps(a,b);
103 }
104 //Double
105 inline void operator()(double * a, __m256d b){
106 _mm256_stream_pd(a,b);
107 }
108
109
110};
111
112struct Vset{
113 // Complex float
114 inline __m256 operator()(Grid::ComplexF *a){
115 return _mm256_set_ps(a[3].imag(),a[3].real(),a[2].imag(),a[2].real(),a[1].imag(),a[1].real(),a[0].imag(),a[0].real());
116 }
117 // Complex double
118 inline __m256d operator()(Grid::ComplexD *a){
119 return _mm256_set_pd(a[1].imag(),a[1].real(),a[0].imag(),a[0].real());
120 }
121 // Real float
122 inline __m256 operator()(float *a){
123 return _mm256_set_ps(a[7],a[6],a[5],a[4],a[3],a[2],a[1],a[0]);
124 }
125 // Real double
126 inline __m256d operator()(double *a){
127 return _mm256_set_pd(a[3],a[2],a[1],a[0]);
128 }
129 // Integer
130 inline __m256i operator()(Integer *a){
131 return _mm256_set_epi32(a[7],a[6],a[5],a[4],a[3],a[2],a[1],a[0]);
132 }
133
134};
135
136template <typename Out_type, typename In_type>
137struct Reduce{
138 // Need templated class to overload output type
139 // General form must generate error if compiled
140 inline Out_type operator()(In_type in){
141 printf("Error, using wrong Reduce function\n");
142 exit(1);
143 return 0;
144 }
145};
146
148// Arithmetic operations
150struct Sum{
151 //Complex/Real float
152 inline __m256 operator()(__m256 a, __m256 b){
153 return _mm256_add_ps(a,b);
154 }
155 //Complex/Real double
156 inline __m256d operator()(__m256d a, __m256d b){
157 return _mm256_add_pd(a,b);
158 }
159 //Integer
160 inline __m256i operator()(__m256i a, __m256i b){
161#if defined (AVX1) || defined (AVXFMA) || defined (AVXFMA4)
162 __m128i a0,a1;
163 __m128i b0,b1;
164 a0 = _mm256_extractf128_si256(a,0);
165 b0 = _mm256_extractf128_si256(b,0);
166 a1 = _mm256_extractf128_si256(a,1);
167 b1 = _mm256_extractf128_si256(b,1);
168 a0 = _mm_add_epi32(a0,b0);
169 a1 = _mm_add_epi32(a1,b1);
170 return _mm256_set_m128i(a1,a0);
171#endif
172#if defined (AVX2)
173 return _mm256_add_epi32(a,b);
174#endif
175 }
176};
177
178struct Sub{
179 //Complex/Real float
180 inline __m256 operator()(__m256 a, __m256 b){
181 return _mm256_sub_ps(a,b);
182 }
183 //Complex/Real double
184 inline __m256d operator()(__m256d a, __m256d b){
185 return _mm256_sub_pd(a,b);
186 }
187 //Integer
188 inline __m256i operator()(__m256i a, __m256i b){
189#if defined (AVX1) || defined (AVXFMA) || defined (AVXFMA4)
190 __m128i a0,a1;
191 __m128i b0,b1;
192 a0 = _mm256_extractf128_si256(a,0);
193 b0 = _mm256_extractf128_si256(b,0);
194 a1 = _mm256_extractf128_si256(a,1);
195 b1 = _mm256_extractf128_si256(b,1);
196 a0 = _mm_sub_epi32(a0,b0);
197 a1 = _mm_sub_epi32(a1,b1);
198 return _mm256_set_m128i(a1,a0);
199#endif
200#if defined (AVX2)
201 return _mm256_sub_epi32(a,b);
202#endif
203
204 }
205};
206
207struct MultRealPart{
208 inline __m256 operator()(__m256 a, __m256 b){
209 __m256 ymm0;
210 ymm0 = _mm256_shuffle_ps(a,a,_MM_SELECT_FOUR_FOUR(2,2,0,0)); // ymm0 <- ar ar,
211 return _mm256_mul_ps(ymm0,b); // ymm0 <- ar bi, ar br
212 }
213 inline __m256d operator()(__m256d a, __m256d b){
214 __m256d ymm0;
215 ymm0 = _mm256_shuffle_pd(a,a,0x0); // ymm0 <- ar ar, ar,ar b'00,00
216 return _mm256_mul_pd(ymm0,b); // ymm0 <- ar bi, ar br
217 }
218};
219struct MaddRealPart{
220 inline __m256 operator()(__m256 a, __m256 b, __m256 c){
221 __m256 ymm0 = _mm256_moveldup_ps(a); // ymm0 <- ar ar,
222 return _mm256_add_ps(_mm256_mul_ps( ymm0, b),c);
223 }
224 inline __m256d operator()(__m256d a, __m256d b, __m256d c){
225 __m256d ymm0 = _mm256_shuffle_pd( a, a, 0x0 );
226 return _mm256_add_pd(_mm256_mul_pd( ymm0, b),c);
227 }
228};
229
230struct MultComplex{
231 // Complex float
232 inline __m256 operator()(__m256 a, __m256 b){
233#if defined (AVX1)
234 __m256 ymm0,ymm1,ymm2;
235 ymm0 = _mm256_shuffle_ps(a,a,_MM_SELECT_FOUR_FOUR(2,2,0,0)); // ymm0 <- ar ar,
236 ymm0 = _mm256_mul_ps(ymm0,b); // ymm0 <- ar bi, ar br
237 // FIXME AVX2 could MAC
238 ymm1 = _mm256_shuffle_ps(b,b,_MM_SELECT_FOUR_FOUR(2,3,0,1)); // ymm1 <- br,bi
239 ymm2 = _mm256_shuffle_ps(a,a,_MM_SELECT_FOUR_FOUR(3,3,1,1)); // ymm2 <- ai,ai
240 ymm1 = _mm256_mul_ps(ymm1,ymm2); // ymm1 <- br ai, ai bi
241 return _mm256_addsub_ps(ymm0,ymm1);
242#endif
243#if defined (AVXFMA4)
244 __m256 a_real = _mm256_shuffle_ps(a,a,_MM_SELECT_FOUR_FOUR(2,2,0,0)); // ar ar,
245 __m256 a_imag = _mm256_shuffle_ps(a,a,_MM_SELECT_FOUR_FOUR(3,3,1,1)); // ai ai
246 __m256 tmp = _mm256_shuffle_ps( b,b, _MM_SELECT_FOUR_FOUR(2,3,0,1));
247 a_imag = _mm256_mul_ps( a_imag,tmp ); // (Ai, Ai) * (Bi, Br) = Ai Bi, Ai Br
248 return _mm256_maddsub_ps( a_real, b, a_imag ); // Ar Br , Ar Bi +- Ai Bi = ArBr-AiBi , ArBi+AiBr
249#endif
250#if defined (AVX2) || defined (AVXFMA)
251 __m256 a_real = _mm256_moveldup_ps( a ); // Ar Ar
252 __m256 a_imag = _mm256_movehdup_ps( a ); // Ai Ai
253 a_imag = _mm256_mul_ps( a_imag, _mm256_shuffle_ps( b,b, _MM_SELECT_FOUR_FOUR(2,3,0,1) )); // (Ai, Ai) * (Bi, Br) = Ai Bi, Ai Br
254 return _mm256_fmaddsub_ps( a_real, b, a_imag ); // Ar Br , Ar Bi +- Ai Bi = ArBr-AiBi , ArBi+AiBr
255#endif
256 }
257 // Complex double
258 inline __m256d operator()(__m256d a, __m256d b) {
259 // Multiplication of (ak+ibk)*(ck+idk)
260 // a + i b can be stored as a data structure
261 // From intel optimisation reference guide
262 /*
263 movsldup xmm0, Src1; load real parts into the destination,
264 ; a1, a1, a0, a0
265 movaps xmm1, src2; load the 2nd pair of complex values, ; i.e. d1, c1, d0, c0
266 mulps xmm0, xmm1; temporary results, a1d1, a1c1, a0d0, ; a0c0
267 shufps xmm1, xmm1, b1; reorder the real and imaginary ; parts, c1, d1, c0, d0
268 movshdup xmm2, Src1; load the imaginary parts into the ; destination, b1, b1, b0, b0
269 mulps xmm2, xmm1; temporary results, b1c1, b1d1, b0c0, ; b0d0
270 addsubps xmm0, xmm2; b1c1+a1d1, a1c1 -b1d1, b0c0+a0d
271 VSHUFPD (VEX.256 encoded version)
272 IF IMM0[0] = 0
273 THEN DEST[63:0]=SRC1[63:0] ELSE DEST[63:0]=SRC1[127:64] FI;
274 IF IMM0[1] = 0
275 THEN DEST[127:64]=SRC2[63:0] ELSE DEST[127:64]=SRC2[127:64] FI;
276 IF IMM0[2] = 0
277 THEN DEST[191:128]=SRC1[191:128] ELSE DEST[191:128]=SRC1[255:192] FI;
278 IF IMM0[3] = 0
279 THEN DEST[255:192]=SRC2[191:128] ELSE DEST[255:192]=SRC2[255:192] FI; // Ox5 r<->i ; 0xC unchanged
280 */
281#if defined (AVX1)
282 __m256d ymm0,ymm1,ymm2;
283 ymm0 = _mm256_shuffle_pd(a,a,0x0); // ymm0 <- ar ar, ar,ar b'00,00
284 ymm0 = _mm256_mul_pd(ymm0,b); // ymm0 <- ar bi, ar br
285 ymm1 = _mm256_shuffle_pd(b,b,0x5); // ymm1 <- br,bi b'01,01
286 ymm2 = _mm256_shuffle_pd(a,a,0xF); // ymm2 <- ai,ai b'11,11
287 ymm1 = _mm256_mul_pd(ymm1,ymm2); // ymm1 <- br ai, ai bi
288 return _mm256_addsub_pd(ymm0,ymm1);
289#endif
290#if defined (AVXFMA4)
291 __m256d a_real = _mm256_shuffle_pd(a,a,0x0);//arar
292 __m256d a_imag = _mm256_shuffle_pd(a,a,0xF);//aiai
293 a_imag = _mm256_mul_pd( a_imag, _mm256_permute_pd( b, 0x5 ) ); // (Ai, Ai) * (Bi, Br) = Ai Bi, Ai Br
294 return _mm256_maddsub_pd( a_real, b, a_imag ); // Ar Br , Ar Bi +- Ai Bi = ArBr-AiBi , ArBi+AiBr
295#endif
296#if defined (AVX2) || defined (AVXFMA)
297 __m256d a_real = _mm256_movedup_pd( a ); // Ar Ar
298 __m256d a_imag = _mm256_shuffle_pd(a,a,0xF);//aiai
299 a_imag = _mm256_mul_pd( a_imag, _mm256_permute_pd( b, 0x5 ) ); // (Ai, Ai) * (Bi, Br) = Ai Bi, Ai Br
300 return _mm256_fmaddsub_pd( a_real, b, a_imag ); // Ar Br , Ar Bi +- Ai Bi = ArBr-AiBi , ArBi+AiBr
301#endif
302 }
303
304
305};
306
307#if 0
308struct ComplexDot {
309
310 inline void Prep(__m256 ari,__m256 &air) {
311 cdotRIperm(ari,air);
312 }
313 inline void Mul(__m256 ari,__m256 air,__m256 b,__m256 &riir,__m256 &iirr) {
314 riir=air*b;
315 iirr=arr*b;
316 };
317 inline void Madd(__m256 ari,__m256 air,__m256 b,__m256 &riir,__m256 &iirr) {
318 mac(riir,air,b);
319 mac(iirr,ari,b);
320 }
321 inline void End(__m256 ari,__m256 &air) {
322 // cdotRI
323 }
324
325};
326#endif
327
328struct Mult{
329
330 inline void mac(__m256 &a, __m256 b, __m256 c){
331#if defined (AVX1)
332 a= _mm256_add_ps(_mm256_mul_ps(b,c),a);
333#endif
334#if defined (AVXFMA4)
335 a= _mm256_macc_ps(b,c,a);
336#endif
337#if defined (AVX2) || defined (AVXFMA)
338 a= _mm256_fmadd_ps( b, c, a);
339#endif
340 }
341
342 inline void mac(__m256d &a, __m256d b, __m256d c){
343#if defined (AVX1)
344 a= _mm256_add_pd(_mm256_mul_pd(b,c),a);
345#endif
346#if defined (AVXFMA4)
347 a= _mm256_macc_pd(b,c,a);
348#endif
349#if defined (AVX2) || defined (AVXFMA)
350 a= _mm256_fmadd_pd( b, c, a);
351#endif
352 }
353
354 // Real float
355 inline __m256 operator()(__m256 a, __m256 b){
356 return _mm256_mul_ps(a,b);
357 }
358 // Real double
359 inline __m256d operator()(__m256d a, __m256d b){
360 return _mm256_mul_pd(a,b);
361 }
362 // Integer
363 inline __m256i operator()(__m256i a, __m256i b){
364#if defined (AVX1) || defined (AVXFMA)
365 __m128i a0,a1;
366 __m128i b0,b1;
367 a0 = _mm256_extractf128_si256(a,0);
368 b0 = _mm256_extractf128_si256(b,0);
369 a1 = _mm256_extractf128_si256(a,1);
370 b1 = _mm256_extractf128_si256(b,1);
371 a0 = _mm_mullo_epi32(a0,b0);
372 a1 = _mm_mullo_epi32(a1,b1);
373 return _mm256_set_m128i(a1,a0);
374#endif
375#if defined (AVX2)
376 return _mm256_mullo_epi32(a,b);
377#endif
378
379 }
380};
381
382struct Div {
383 // Real float
384 inline __m256 operator()(__m256 a, __m256 b) {
385 return _mm256_div_ps(a, b);
386 }
387 // Real double
388 inline __m256d operator()(__m256d a, __m256d b){
389 return _mm256_div_pd(a,b);
390 }
391};
392
393
394struct Conj{
395 // Complex single
396 inline __m256 operator()(__m256 in){
397 return _mm256_xor_ps(_mm256_addsub_ps(_mm256_setzero_ps(),in), _mm256_set1_ps(-0.f));
398 }
399 // Complex double
400 inline __m256d operator()(__m256d in){
401 return _mm256_xor_pd(_mm256_addsub_pd(_mm256_setzero_pd(),in), _mm256_set1_pd(-0.f));
402 }
403 // do not define for integer input
404};
405
406struct TimesMinusI{
407 //Complex single
408 inline __m256 operator()(__m256 in){
409 __m256 tmp =_mm256_addsub_ps(_mm256_setzero_ps(),in); // r,-i
410 return _mm256_shuffle_ps(tmp,tmp,_MM_SELECT_FOUR_FOUR(2,3,0,1)); //-i,r
411 }
412 //Complex double
413 inline __m256d operator()(__m256d in){
414 __m256d tmp = _mm256_addsub_pd(_mm256_setzero_pd(),in); // r,-i
415 return _mm256_shuffle_pd(tmp,tmp,0x5);
416 }
417};
418
419struct TimesI{
420 //Complex single
421 inline __m256 operator()(__m256 in){
422 __m256 tmp =_mm256_shuffle_ps(in,in,_MM_SELECT_FOUR_FOUR(2,3,0,1)); // i,r
423 return _mm256_addsub_ps(_mm256_setzero_ps(),tmp); // i,-r
424 }
425 //Complex double
426 inline __m256d operator()(__m256d in){
427 __m256d tmp = _mm256_shuffle_pd(in,in,0x5);
428 return _mm256_addsub_pd(_mm256_setzero_pd(),tmp); // i,-r
429 }
430};
431
433// Some Template specialization
435
436struct Permute{
437
438 static inline __m256 Permute0(__m256 in){
439 return _mm256_permute2f128_ps(in,in,0x01); //ABCD EFGH -> EFGH ABCD
440 };
441 static inline __m256 Permute1(__m256 in){
442 return _mm256_shuffle_ps(in,in,_MM_SELECT_FOUR_FOUR(1,0,3,2)); //ABCD EFGH -> CDAB GHEF
443 };
444 static inline __m256 Permute2(__m256 in){
445 return _mm256_shuffle_ps(in,in,_MM_SELECT_FOUR_FOUR(2,3,0,1)); //ABCD EFGH -> BADC FEHG
446 };
447 static inline __m256 Permute3(__m256 in){
448 return in;
449 };
450
451 static inline __m256d Permute0(__m256d in){
452 return _mm256_permute2f128_pd(in,in,0x01); //AB CD -> CD AB
453 };
454 static inline __m256d Permute1(__m256d in){ //AB CD -> BA DC
455 return _mm256_shuffle_pd(in,in,0x5);
456 };
457 static inline __m256d Permute2(__m256d in){
458 return in;
459 };
460 static inline __m256d Permute3(__m256d in){
461 return in;
462 };
463};
464#define USE_FP16
465struct PrecisionChange {
466 static inline __m256i StoH (__m256 a,__m256 b) {
467 __m256i h;
468#ifdef USE_FP16
469 __m128i ha = _mm256_cvtps_ph(a,0);
470 __m128i hb = _mm256_cvtps_ph(b,0);
471 h =(__m256i) _mm256_castps128_ps256((__m128)ha);
472 h =(__m256i) _mm256_insertf128_ps((__m256)h,(__m128)hb,1);
473#else
474 assert(0);
475#endif
476 return h;
477 }
478 static inline void HtoS (__m256i h,__m256 &sa,__m256 &sb) {
479#ifdef USE_FP16
480 sa = _mm256_cvtph_ps((__m128i)_mm256_extractf128_ps((__m256)h,0));
481 sb = _mm256_cvtph_ps((__m128i)_mm256_extractf128_ps((__m256)h,1));
482#else
483 assert(0);
484#endif
485 }
486 static inline __m256 DtoS (__m256d a,__m256d b) {
487 __m128 sa = _mm256_cvtpd_ps(a);
488 __m128 sb = _mm256_cvtpd_ps(b);
489 __m256 s = _mm256_castps128_ps256(sa);
490 s = _mm256_insertf128_ps(s,sb,1);
491 return s;
492 }
493 static inline void StoD (__m256 s,__m256d &a,__m256d &b) {
494 a = _mm256_cvtps_pd(_mm256_extractf128_ps(s,0));
495 b = _mm256_cvtps_pd(_mm256_extractf128_ps(s,1));
496 }
497 static inline __m256i DtoH (__m256d a,__m256d b,__m256d c,__m256d d) {
498 __m256 sa,sb;
499 sa = DtoS(a,b);
500 sb = DtoS(c,d);
501 return StoH(sa,sb);
502 }
503 static inline void HtoD (__m256i h,__m256d &a,__m256d &b,__m256d &c,__m256d &d) {
504 __m256 sa,sb;
505 HtoS(h,sa,sb);
506 StoD(sa,a,b);
507 StoD(sb,c,d);
508 }
509};
510struct Exchange{
511 // 3210 ordering
512 static inline void Exchange0(__m256 &out1,__m256 &out2,__m256 in1,__m256 in2){
513 //Invertible
514 //AB CD -> AC BD
515 //AC BD -> AB CD
516 out1= _mm256_permute2f128_ps(in1,in2,0x20);
517 out2= _mm256_permute2f128_ps(in1,in2,0x31);
518 };
519 static inline void Exchange1(__m256 &out1,__m256 &out2,__m256 in1,__m256 in2){
520 //Invertible
521 // ABCD EFGH ->ABEF CDGH
522 // ABEF CDGH ->ABCD EFGH
523 out1= _mm256_shuffle_ps(in1,in2,_MM_SELECT_FOUR_FOUR(1,0,1,0));
524 out2= _mm256_shuffle_ps(in1,in2,_MM_SELECT_FOUR_FOUR(3,2,3,2));
525 };
526 static inline void Exchange2(__m256 &out1,__m256 &out2,__m256 in1,__m256 in2){
527 // Invertible ?
528 // ABCD EFGH -> ACEG BDFH
529 // ACEG BDFH -> AEBF CGDH
530 // out1= _mm256_shuffle_ps(in1,in2,_MM_SELECT_FOUR_FOUR(2,0,2,0));
531 // out2= _mm256_shuffle_ps(in1,in2,_MM_SELECT_FOUR_FOUR(3,1,3,1));
532 // Bollocks; need
533 // AECG BFDH -> ABCD EFGH
534 out1= _mm256_shuffle_ps(in1,in2,_MM_SELECT_FOUR_FOUR(2,0,2,0)); /*ACEG*/
535 out2= _mm256_shuffle_ps(in1,in2,_MM_SELECT_FOUR_FOUR(3,1,3,1)); /*BDFH*/
536 out1= _mm256_shuffle_ps(out1,out1,_MM_SELECT_FOUR_FOUR(3,1,2,0)); /*AECG*/
537 out2= _mm256_shuffle_ps(out2,out2,_MM_SELECT_FOUR_FOUR(3,1,2,0)); /*AECG*/
538 };
539 static inline void Exchange3(__m256 &out1,__m256 &out2,__m256 in1,__m256 in2){
540 assert(0);
541 return;
542 };
543
544 static inline void Exchange0(__m256d &out1,__m256d &out2,__m256d in1,__m256d in2){
545 out1= _mm256_permute2f128_pd(in1,in2,0x20);
546 out2= _mm256_permute2f128_pd(in1,in2,0x31);
547 return;
548 };
549 static inline void Exchange1(__m256d &out1,__m256d &out2,__m256d in1,__m256d in2){
550 out1= _mm256_shuffle_pd(in1,in2,0x0);
551 out2= _mm256_shuffle_pd(in1,in2,0xF);
552 };
553 static inline void Exchange2(__m256d &out1,__m256d &out2,__m256d in1,__m256d in2){
554 assert(0);
555 return;
556 };
557 static inline void Exchange3(__m256d &out1,__m256d &out2,__m256d in1,__m256d in2){
558 assert(0);
559 return;
560 };
561};
562
563
564#if defined (AVX2)
565#define _mm256_alignr_epi32_grid(ret,a,b,n) ret=(__m256) _mm256_alignr_epi8((__m256i)a,(__m256i)b,(n*4)%16)
566#define _mm256_alignr_epi64_grid(ret,a,b,n) ret=(__m256d) _mm256_alignr_epi8((__m256i)a,(__m256i)b,(n*8)%16)
567#endif
568
569#if defined (AVX1) || defined (AVXFMA)
570#define _mm256_alignr_epi32_grid(ret,a,b,n) { \
571 __m128 aa, bb; \
572 \
573 aa = _mm256_extractf128_ps(a,1); \
574 bb = _mm256_extractf128_ps(b,1); \
575 aa = (__m128)_mm_alignr_epi8((__m128i)aa,(__m128i)bb,(n*4)%16); \
576 ret = _mm256_insertf128_ps(ret,aa,1); \
577 \
578 aa = _mm256_extractf128_ps(a,0); \
579 bb = _mm256_extractf128_ps(b,0); \
580 aa = (__m128)_mm_alignr_epi8((__m128i)aa,(__m128i)bb,(n*4)%16); \
581 ret = _mm256_insertf128_ps(ret,aa,0); \
582 }
583
584#define _mm256_alignr_epi64_grid(ret,a,b,n) { \
585 __m128d aa, bb; \
586 \
587 aa = _mm256_extractf128_pd(a,1); \
588 bb = _mm256_extractf128_pd(b,1); \
589 aa = (__m128d)_mm_alignr_epi8((__m128i)aa,(__m128i)bb,(n*8)%16); \
590 ret = _mm256_insertf128_pd(ret,aa,1); \
591 \
592 aa = _mm256_extractf128_pd(a,0); \
593 bb = _mm256_extractf128_pd(b,0); \
594 aa = (__m128d)_mm_alignr_epi8((__m128i)aa,(__m128i)bb,(n*8)%16); \
595 ret = _mm256_insertf128_pd(ret,aa,0); \
596 }
597
598#endif
599
600struct Rotate{
601
602 static inline __m256 rotate(__m256 in,int n){
603 switch(n){
604 case 0: return tRotate<0>(in);break;
605 case 1: return tRotate<1>(in);break;
606 case 2: return tRotate<2>(in);break;
607 case 3: return tRotate<3>(in);break;
608 case 4: return tRotate<4>(in);break;
609 case 5: return tRotate<5>(in);break;
610 case 6: return tRotate<6>(in);break;
611 case 7: return tRotate<7>(in);break;
612 default: assert(0);
613 }
614 }
615 static inline __m256d rotate(__m256d in,int n){
616 switch(n){
617 case 0: return tRotate<0>(in);break;
618 case 1: return tRotate<1>(in);break;
619 case 2: return tRotate<2>(in);break;
620 case 3: return tRotate<3>(in);break;
621 default: assert(0);
622 }
623 }
624
625
626 template<int n>
627 static inline __m256 tRotate(__m256 in){
628 __m256 tmp = Permute::Permute0(in);
629 __m256 ret;
630 if ( n > 3 ) {
631 _mm256_alignr_epi32_grid(ret,in,tmp,n);
632 } else {
633 _mm256_alignr_epi32_grid(ret,tmp,in,n);
634 }
635 return ret;
636 }
637
638 template<int n>
639 static inline __m256d tRotate(__m256d in){
640 __m256d tmp = Permute::Permute0(in);
641 __m256d ret;
642 if ( n > 1 ) {
643 _mm256_alignr_epi64_grid(ret,in,tmp,n);
644 } else {
645 _mm256_alignr_epi64_grid(ret,tmp,in,n);
646 }
647 return ret;
648 };
649
650};
651
652//Complex float Reduce
653template<>
654inline Grid::ComplexF Reduce<Grid::ComplexF, __m256>::operator()(__m256 in){
655 __m256 v1,v2;
656 v1=Optimization::Permute::Permute0(in); // avx 256; quad complex single
657 v1= _mm256_add_ps(v1,in);
658 v2=Optimization::Permute::Permute1(v1);
659 v1 = _mm256_add_ps(v1,v2);
660 u256f conv; conv.v = v1;
661 return Grid::ComplexF(conv.f[0],conv.f[1]);
662}
663
664//Real float Reduce
665template<>
666inline Grid::RealF Reduce<Grid::RealF, __m256>::operator()(__m256 in){
667 __m256 v1,v2;
668 v1 = Optimization::Permute::Permute0(in); // avx 256; octo-double
669 v1 = _mm256_add_ps(v1,in);
670 v2 = Optimization::Permute::Permute1(v1);
671 v1 = _mm256_add_ps(v1,v2);
672 v2 = Optimization::Permute::Permute2(v1);
673 v1 = _mm256_add_ps(v1,v2);
674 u256f conv; conv.v=v1;
675 return conv.f[0];
676}
677
678
679//Complex double Reduce
680template<>
681inline Grid::ComplexD Reduce<Grid::ComplexD, __m256d>::operator()(__m256d in){
682 __m256d v1;
683 v1 = Optimization::Permute::Permute0(in); // sse 128; paired complex single
684 v1 = _mm256_add_pd(v1,in);
685 u256d conv; conv.v = v1;
686 return Grid::ComplexD(conv.f[0],conv.f[1]);
687}
688
689//Real double Reduce
690template<>
691inline Grid::RealD Reduce<Grid::RealD, __m256d>::operator()(__m256d in){
692 __m256d v1,v2;
693 v1 = Optimization::Permute::Permute0(in); // avx 256; quad double
694 v1 = _mm256_add_pd(v1,in);
695 v2 = Optimization::Permute::Permute1(v1);
696 v1 = _mm256_add_pd(v1,v2);
697 u256d conv; conv.v = v1;
698 return conv.f[0];
699}
700
701//Integer Reduce
702template<>
704 __m128i ret;
705#if defined (AVX2)
706 // AVX2 horizontal adds within upper and lower halves of register; use
707 // SSE to add upper and lower halves for result.
708 __m256i v1, v2;
709 __m128i u1, u2;
710 v1 = _mm256_hadd_epi32(in, in);
711 v2 = _mm256_hadd_epi32(v1, v1);
712 u1 = _mm256_castsi256_si128(v2); // upper half
713 u2 = _mm256_extracti128_si256(v2, 1); // lower half
714 ret = _mm_add_epi32(u1, u2);
715#else
716 // No AVX horizontal add; extract upper and lower halves of register & use
717 // SSE intrinsics.
718 __m128i u1, u2, u3;
719 u1 = _mm256_extractf128_si256(in, 0); // upper half
720 u2 = _mm256_extractf128_si256(in, 1); // lower half
721 u3 = _mm_add_epi32(u1, u2);
722 u1 = _mm_hadd_epi32(u3, u3);
723 ret = _mm_hadd_epi32(u1, u1);
724#endif
725 return _mm_cvtsi128_si32(ret);
726}
727
728NAMESPACE_END(Optimization);
729
731// Here assign types
732
733typedef __m256i SIMD_Htype; // Single precision type
734typedef __m256 SIMD_Ftype; // Single precision type
735typedef __m256d SIMD_Dtype; // Double precision type
736typedef __m256i SIMD_Itype; // Integer type
737
738// prefecthing
739inline void v_prefetch0(int size, const char *ptr){
740 for(int i=0;i<size;i+=64){ // Define L1 linesize above
741 _mm_prefetch(ptr+i+4096,_MM_HINT_T1);
742 _mm_prefetch(ptr+i+512,_MM_HINT_T0);
743 }
744}
745inline void prefetch_HINT_T0(const char *ptr){
746 _mm_prefetch(ptr, _MM_HINT_T0);
747}
748
749// Function name aliases
750typedef Optimization::Vsplat VsplatSIMD;
751typedef Optimization::Vstore VstoreSIMD;
752typedef Optimization::Vset VsetSIMD;
753typedef Optimization::Vstream VstreamSIMD;
754
755template <typename S, typename T> using ReduceSIMD = Optimization::Reduce<S, T>;
756
757// Arithmetic operations
758typedef Optimization::Sum SumSIMD;
759typedef Optimization::Sub SubSIMD;
760typedef Optimization::Div DivSIMD;
761typedef Optimization::Mult MultSIMD;
762typedef Optimization::MultComplex MultComplexSIMD;
763typedef Optimization::MultRealPart MultRealPartSIMD;
764typedef Optimization::MaddRealPart MaddRealPartSIMD;
765typedef Optimization::Conj ConjSIMD;
766typedef Optimization::TimesMinusI TimesMinusISIMD;
767typedef Optimization::TimesI TimesISIMD;
768
Optimization::Vstream VstreamSIMD
Optimization::TimesMinusI TimesMinusISIMD
Optimization::MultComplex MultComplexSIMD
Optimization::TimesI TimesISIMD
Optimization::Reduce< S, T > ReduceSIMD
Optimization::Mult MultSIMD
Optimization::MaddRealPart MaddRealPartSIMD
Optimization::vecd SIMD_Dtype
Optimization::veci SIMD_Itype
Optimization::Vstore VstoreSIMD
Optimization::Conj ConjSIMD
Optimization::vecf SIMD_Ftype
Optimization::Vsplat VsplatSIMD
Optimization::Sum SumSIMD
Optimization::Sub SubSIMD
Optimization::Div DivSIMD
Optimization::MultRealPart MultRealPartSIMD
Optimization::Vset VsetSIMD
Optimization::vech SIMD_Htype
#define _mm256_set_m128i(hi, lo)
Definition Grid_avx.h:38
void prefetch_HINT_T0(const char *ptr)
Definition Grid_avx.h:745
void v_prefetch0(int size, const char *ptr)
Definition Grid_avx.h:739
void mac(Lattice< obj1 > &ret, const Lattice< obj2 > &lhs, const Lattice< obj3 > &rhs)
Lattice< vobj > real(const Lattice< vobj > &lhs)
Lattice< vobj > imag(const Lattice< vobj > &lhs)
#define NAMESPACE_BEGIN(A)
Definition Namespace.h:35
#define NAMESPACE_END(A)
Definition Namespace.h:36
uint32_t Integer
Definition Simd.h:58
#define _MM_SELECT_FOUR_FOUR(A, B, C, D)
Definition Simd.h:48
static INTERNAL_PRECISION F
Definition Zolotarev.cc:230
__m256 operator()(__m256 in)
Definition Grid_avx.h:396
__m256d operator()(__m256d in)
Definition Grid_avx.h:400
__m256d operator()(__m256d a, __m256d b)
Definition Grid_avx.h:388
__m256 operator()(__m256 a, __m256 b)
Definition Grid_avx.h:384
static void Exchange2(__m256 &out1, __m256 &out2, __m256 in1, __m256 in2)
Definition Grid_avx.h:526
static void Exchange1(__m256d &out1, __m256d &out2, __m256d in1, __m256d in2)
Definition Grid_avx.h:549
static void Exchange3(__m256 &out1, __m256 &out2, __m256 in1, __m256 in2)
Definition Grid_avx.h:539
static void Exchange1(__m256 &out1, __m256 &out2, __m256 in1, __m256 in2)
Definition Grid_avx.h:519
static void Exchange3(__m256d &out1, __m256d &out2, __m256d in1, __m256d in2)
Definition Grid_avx.h:557
static void Exchange0(__m256d &out1, __m256d &out2, __m256d in1, __m256d in2)
Definition Grid_avx.h:544
static void Exchange2(__m256d &out1, __m256d &out2, __m256d in1, __m256d in2)
Definition Grid_avx.h:553
static void Exchange0(__m256 &out1, __m256 &out2, __m256 in1, __m256 in2)
Definition Grid_avx.h:512
__m256d operator()(__m256d a, __m256d b, __m256d c)
Definition Grid_avx.h:224
__m256 operator()(__m256 a, __m256 b, __m256 c)
Definition Grid_avx.h:220
__m256d operator()(__m256d a, __m256d b)
Definition Grid_avx.h:258
__m256 operator()(__m256 a, __m256 b)
Definition Grid_avx.h:232
__m256 operator()(__m256 a, __m256 b)
Definition Grid_avx.h:208
__m256d operator()(__m256d a, __m256d b)
Definition Grid_avx.h:213
__m256d operator()(__m256d a, __m256d b)
Definition Grid_avx.h:359
void mac(__m256d &a, __m256d b, __m256d c)
Definition Grid_avx.h:342
__m256i operator()(__m256i a, __m256i b)
Definition Grid_avx.h:363
__m256 operator()(__m256 a, __m256 b)
Definition Grid_avx.h:355
void mac(__m256 &a, __m256 b, __m256 c)
Definition Grid_avx.h:330
static vec< T > Permute0(vec< T > in)
static __m256 Permute0(__m256 in)
Definition Grid_avx.h:438
static __m256d Permute2(__m256d in)
Definition Grid_avx.h:457
static __m256 Permute2(__m256 in)
Definition Grid_avx.h:444
static __m256 Permute3(__m256 in)
Definition Grid_avx.h:447
static __m256d Permute0(__m256d in)
Definition Grid_avx.h:451
static __m256d Permute3(__m256d in)
Definition Grid_avx.h:460
static __m256d Permute1(__m256d in)
Definition Grid_avx.h:454
static __m256 Permute1(__m256 in)
Definition Grid_avx.h:441
static void HtoD(__m256i h, __m256d &a, __m256d &b, __m256d &c, __m256d &d)
Definition Grid_avx.h:503
static vech StoH(const vecf &sa, const vecf &sb)
static __m256 DtoS(__m256d a, __m256d b)
Definition Grid_avx.h:486
static void HtoS(__m256i h, __m256 &sa, __m256 &sb)
Definition Grid_avx.h:478
static __m256i StoH(__m256 a, __m256 b)
Definition Grid_avx.h:466
static void StoD(vecf s, vecd &a, vecd &b)
static vecf DtoS(vecd a, vecd b)
static __m256i DtoH(__m256d a, __m256d b, __m256d c, __m256d d)
Definition Grid_avx.h:497
static void StoD(__m256 s, __m256d &a, __m256d &b)
Definition Grid_avx.h:493
static void HtoS(vech h, vecf &sa, vecf &sb)
Out_type operator()(In_type in)
Definition Grid_avx.h:140
static vec< T > tRotate(vec< T > in)
static __m256 tRotate(__m256 in)
Definition Grid_avx.h:627
static __m256d rotate(__m256d in, int n)
Definition Grid_avx.h:615
static __m256 rotate(__m256 in, int n)
Definition Grid_avx.h:602
static __m256d tRotate(__m256d in)
Definition Grid_avx.h:639
__m256i operator()(__m256i a, __m256i b)
Definition Grid_avx.h:188
__m256 operator()(__m256 a, __m256 b)
Definition Grid_avx.h:180
__m256d operator()(__m256d a, __m256d b)
Definition Grid_avx.h:184
__m256d operator()(__m256d a, __m256d b)
Definition Grid_avx.h:156
__m256 operator()(__m256 a, __m256 b)
Definition Grid_avx.h:152
__m256i operator()(__m256i a, __m256i b)
Definition Grid_avx.h:160
__m256 operator()(__m256 in)
Definition Grid_avx.h:421
__m256d operator()(__m256d in)
Definition Grid_avx.h:426
__m256d operator()(__m256d in)
Definition Grid_avx.h:413
__m256 operator()(__m256 in)
Definition Grid_avx.h:408
__m256d operator()(Grid::ComplexD *a)
Definition Grid_avx.h:118
__m256 operator()(Grid::ComplexF *a)
Definition Grid_avx.h:114
__m256 operator()(float *a)
Definition Grid_avx.h:122
__m256i operator()(Integer *a)
Definition Grid_avx.h:130
__m256d operator()(double *a)
Definition Grid_avx.h:126
__m256d operator()(double a)
Definition Grid_avx.h:74
__m256 operator()(float a)
Definition Grid_avx.h:66
__m256d operator()(double a, double b)
Definition Grid_avx.h:70
__m256i operator()(Integer a)
Definition Grid_avx.h:78
__m256 operator()(float a, float b)
Definition Grid_avx.h:62
void operator()(__m256d a, double *D)
Definition Grid_avx.h:89
void operator()(__m256 a, float *F)
Definition Grid_avx.h:85
void operator()(__m256i a, Integer *I)
Definition Grid_avx.h:93
void operator()(float *a, __m256 b)
Definition Grid_avx.h:101
void operator()(double *a, __m256d b)
Definition Grid_avx.h:105
__m256d v
Definition Grid_avx.h:56
double f[4]
Definition Grid_avx.h:57
float f[8]
Definition Grid_avx.h:52
__m256 v
Definition Grid_avx.h:51
__m256 f
Definition Grid_avx.h:46
vtype v
Definition Grid_avx.h:47