2 * Note: this file was generated by the Gromacs avx_256_double kernel generator.
4 * This source code is part of
8 * Copyright (c) 2001-2012, The GROMACS Development Team
10 * Gromacs is a library for molecular simulation and trajectory analysis,
11 * written by Erik Lindahl, David van der Spoel, Berk Hess, and others - for
12 * a full list of developers and information, check out http://www.gromacs.org
14 * This program is free software; you can redistribute it and/or modify it under
15 * the terms of the GNU Lesser General Public License as published by the Free
16 * Software Foundation; either version 2 of the License, or (at your option) any
19 * To help fund GROMACS development, we humbly ask that you cite
20 * the papers people have written on it - you can find them on the website.
28 #include "../nb_kernel.h"
29 #include "types/simple.h"
33 #include "gmx_math_x86_avx_256_double.h"
34 #include "kernelutil_x86_avx_256_double.h"
37 * Gromacs nonbonded kernel: nb_kernel_ElecCSTab_VdwNone_GeomW3W3_VF_avx_256_double
38 * Electrostatics interaction: CubicSplineTable
39 * VdW interaction: None
40 * Geometry: Water3-Water3
41 * Calculate force/pot: PotentialAndForce
44 nb_kernel_ElecCSTab_VdwNone_GeomW3W3_VF_avx_256_double
45 (t_nblist * gmx_restrict nlist,
46 rvec * gmx_restrict xx,
47 rvec * gmx_restrict ff,
48 t_forcerec * gmx_restrict fr,
49 t_mdatoms * gmx_restrict mdatoms,
50 nb_kernel_data_t * gmx_restrict kernel_data,
51 t_nrnb * gmx_restrict nrnb)
53 /* Suffixes 0,1,2,3 refer to particle indices for waters in the inner or outer loop, or
54 * just 0 for non-waters.
55 * Suffixes A,B,C,D refer to j loop unrolling done with AVX, e.g. for the four different
56 * jnr indices corresponding to data put in the four positions in the SIMD register.
58 int i_shift_offset,i_coord_offset,outeriter,inneriter;
59 int j_index_start,j_index_end,jidx,nri,inr,ggid,iidx;
60 int jnrA,jnrB,jnrC,jnrD;
61 int jnrlistA,jnrlistB,jnrlistC,jnrlistD;
62 int jnrlistE,jnrlistF,jnrlistG,jnrlistH;
63 int j_coord_offsetA,j_coord_offsetB,j_coord_offsetC,j_coord_offsetD;
64 int *iinr,*jindex,*jjnr,*shiftidx,*gid;
66 real *shiftvec,*fshift,*x,*f;
67 real *fjptrA,*fjptrB,*fjptrC,*fjptrD;
69 __m256d tx,ty,tz,fscal,rcutoff,rcutoff2,jidxall;
70 real * vdwioffsetptr0;
71 __m256d ix0,iy0,iz0,fix0,fiy0,fiz0,iq0,isai0;
72 real * vdwioffsetptr1;
73 __m256d ix1,iy1,iz1,fix1,fiy1,fiz1,iq1,isai1;
74 real * vdwioffsetptr2;
75 __m256d ix2,iy2,iz2,fix2,fiy2,fiz2,iq2,isai2;
76 int vdwjidx0A,vdwjidx0B,vdwjidx0C,vdwjidx0D;
77 __m256d jx0,jy0,jz0,fjx0,fjy0,fjz0,jq0,isaj0;
78 int vdwjidx1A,vdwjidx1B,vdwjidx1C,vdwjidx1D;
79 __m256d jx1,jy1,jz1,fjx1,fjy1,fjz1,jq1,isaj1;
80 int vdwjidx2A,vdwjidx2B,vdwjidx2C,vdwjidx2D;
81 __m256d jx2,jy2,jz2,fjx2,fjy2,fjz2,jq2,isaj2;
82 __m256d dx00,dy00,dz00,rsq00,rinv00,rinvsq00,r00,qq00,c6_00,c12_00;
83 __m256d dx01,dy01,dz01,rsq01,rinv01,rinvsq01,r01,qq01,c6_01,c12_01;
84 __m256d dx02,dy02,dz02,rsq02,rinv02,rinvsq02,r02,qq02,c6_02,c12_02;
85 __m256d dx10,dy10,dz10,rsq10,rinv10,rinvsq10,r10,qq10,c6_10,c12_10;
86 __m256d dx11,dy11,dz11,rsq11,rinv11,rinvsq11,r11,qq11,c6_11,c12_11;
87 __m256d dx12,dy12,dz12,rsq12,rinv12,rinvsq12,r12,qq12,c6_12,c12_12;
88 __m256d dx20,dy20,dz20,rsq20,rinv20,rinvsq20,r20,qq20,c6_20,c12_20;
89 __m256d dx21,dy21,dz21,rsq21,rinv21,rinvsq21,r21,qq21,c6_21,c12_21;
90 __m256d dx22,dy22,dz22,rsq22,rinv22,rinvsq22,r22,qq22,c6_22,c12_22;
91 __m256d velec,felec,velecsum,facel,crf,krf,krf2;
94 __m128i ifour = _mm_set1_epi32(4);
95 __m256d rt,vfeps,vftabscale,Y,F,G,H,Heps,Fp,VV,FF;
97 __m256d dummy_mask,cutoff_mask;
98 __m128 tmpmask0,tmpmask1;
99 __m256d signbit = _mm256_castsi256_pd( _mm256_set1_epi32(0x80000000) );
100 __m256d one = _mm256_set1_pd(1.0);
101 __m256d two = _mm256_set1_pd(2.0);
107 jindex = nlist->jindex;
109 shiftidx = nlist->shift;
111 shiftvec = fr->shift_vec[0];
112 fshift = fr->fshift[0];
113 facel = _mm256_set1_pd(fr->epsfac);
114 charge = mdatoms->chargeA;
116 vftab = kernel_data->table_elec->data;
117 vftabscale = _mm256_set1_pd(kernel_data->table_elec->scale);
119 /* Setup water-specific parameters */
120 inr = nlist->iinr[0];
121 iq0 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+0]));
122 iq1 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+1]));
123 iq2 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+2]));
125 jq0 = _mm256_set1_pd(charge[inr+0]);
126 jq1 = _mm256_set1_pd(charge[inr+1]);
127 jq2 = _mm256_set1_pd(charge[inr+2]);
128 qq00 = _mm256_mul_pd(iq0,jq0);
129 qq01 = _mm256_mul_pd(iq0,jq1);
130 qq02 = _mm256_mul_pd(iq0,jq2);
131 qq10 = _mm256_mul_pd(iq1,jq0);
132 qq11 = _mm256_mul_pd(iq1,jq1);
133 qq12 = _mm256_mul_pd(iq1,jq2);
134 qq20 = _mm256_mul_pd(iq2,jq0);
135 qq21 = _mm256_mul_pd(iq2,jq1);
136 qq22 = _mm256_mul_pd(iq2,jq2);
138 /* Avoid stupid compiler warnings */
139 jnrA = jnrB = jnrC = jnrD = 0;
148 for(iidx=0;iidx<4*DIM;iidx++)
153 /* Start outer loop over neighborlists */
154 for(iidx=0; iidx<nri; iidx++)
156 /* Load shift vector for this list */
157 i_shift_offset = DIM*shiftidx[iidx];
159 /* Load limits for loop over neighbors */
160 j_index_start = jindex[iidx];
161 j_index_end = jindex[iidx+1];
163 /* Get outer coordinate index */
165 i_coord_offset = DIM*inr;
167 /* Load i particle coords and add shift vector */
168 gmx_mm256_load_shift_and_3rvec_broadcast_pd(shiftvec+i_shift_offset,x+i_coord_offset,
169 &ix0,&iy0,&iz0,&ix1,&iy1,&iz1,&ix2,&iy2,&iz2);
171 fix0 = _mm256_setzero_pd();
172 fiy0 = _mm256_setzero_pd();
173 fiz0 = _mm256_setzero_pd();
174 fix1 = _mm256_setzero_pd();
175 fiy1 = _mm256_setzero_pd();
176 fiz1 = _mm256_setzero_pd();
177 fix2 = _mm256_setzero_pd();
178 fiy2 = _mm256_setzero_pd();
179 fiz2 = _mm256_setzero_pd();
181 /* Reset potential sums */
182 velecsum = _mm256_setzero_pd();
184 /* Start inner kernel loop */
185 for(jidx=j_index_start; jidx<j_index_end && jjnr[jidx+3]>=0; jidx+=4)
188 /* Get j neighbor index, and coordinate index */
193 j_coord_offsetA = DIM*jnrA;
194 j_coord_offsetB = DIM*jnrB;
195 j_coord_offsetC = DIM*jnrC;
196 j_coord_offsetD = DIM*jnrD;
198 /* load j atom coordinates */
199 gmx_mm256_load_3rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
200 x+j_coord_offsetC,x+j_coord_offsetD,
201 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,&jy2,&jz2);
203 /* Calculate displacement vector */
204 dx00 = _mm256_sub_pd(ix0,jx0);
205 dy00 = _mm256_sub_pd(iy0,jy0);
206 dz00 = _mm256_sub_pd(iz0,jz0);
207 dx01 = _mm256_sub_pd(ix0,jx1);
208 dy01 = _mm256_sub_pd(iy0,jy1);
209 dz01 = _mm256_sub_pd(iz0,jz1);
210 dx02 = _mm256_sub_pd(ix0,jx2);
211 dy02 = _mm256_sub_pd(iy0,jy2);
212 dz02 = _mm256_sub_pd(iz0,jz2);
213 dx10 = _mm256_sub_pd(ix1,jx0);
214 dy10 = _mm256_sub_pd(iy1,jy0);
215 dz10 = _mm256_sub_pd(iz1,jz0);
216 dx11 = _mm256_sub_pd(ix1,jx1);
217 dy11 = _mm256_sub_pd(iy1,jy1);
218 dz11 = _mm256_sub_pd(iz1,jz1);
219 dx12 = _mm256_sub_pd(ix1,jx2);
220 dy12 = _mm256_sub_pd(iy1,jy2);
221 dz12 = _mm256_sub_pd(iz1,jz2);
222 dx20 = _mm256_sub_pd(ix2,jx0);
223 dy20 = _mm256_sub_pd(iy2,jy0);
224 dz20 = _mm256_sub_pd(iz2,jz0);
225 dx21 = _mm256_sub_pd(ix2,jx1);
226 dy21 = _mm256_sub_pd(iy2,jy1);
227 dz21 = _mm256_sub_pd(iz2,jz1);
228 dx22 = _mm256_sub_pd(ix2,jx2);
229 dy22 = _mm256_sub_pd(iy2,jy2);
230 dz22 = _mm256_sub_pd(iz2,jz2);
232 /* Calculate squared distance and things based on it */
233 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
234 rsq01 = gmx_mm256_calc_rsq_pd(dx01,dy01,dz01);
235 rsq02 = gmx_mm256_calc_rsq_pd(dx02,dy02,dz02);
236 rsq10 = gmx_mm256_calc_rsq_pd(dx10,dy10,dz10);
237 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
238 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
239 rsq20 = gmx_mm256_calc_rsq_pd(dx20,dy20,dz20);
240 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
241 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
243 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
244 rinv01 = gmx_mm256_invsqrt_pd(rsq01);
245 rinv02 = gmx_mm256_invsqrt_pd(rsq02);
246 rinv10 = gmx_mm256_invsqrt_pd(rsq10);
247 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
248 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
249 rinv20 = gmx_mm256_invsqrt_pd(rsq20);
250 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
251 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
253 fjx0 = _mm256_setzero_pd();
254 fjy0 = _mm256_setzero_pd();
255 fjz0 = _mm256_setzero_pd();
256 fjx1 = _mm256_setzero_pd();
257 fjy1 = _mm256_setzero_pd();
258 fjz1 = _mm256_setzero_pd();
259 fjx2 = _mm256_setzero_pd();
260 fjy2 = _mm256_setzero_pd();
261 fjz2 = _mm256_setzero_pd();
263 /**************************
264 * CALCULATE INTERACTIONS *
265 **************************/
267 r00 = _mm256_mul_pd(rsq00,rinv00);
269 /* Calculate table index by multiplying r with table scale and truncate to integer */
270 rt = _mm256_mul_pd(r00,vftabscale);
271 vfitab = _mm256_cvttpd_epi32(rt);
272 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
273 vfitab = _mm_slli_epi32(vfitab,2);
275 /* CUBIC SPLINE TABLE ELECTROSTATICS */
276 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
277 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
278 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
279 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
280 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
281 Heps = _mm256_mul_pd(vfeps,H);
282 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
283 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
284 velec = _mm256_mul_pd(qq00,VV);
285 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
286 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq00,FF),_mm256_mul_pd(vftabscale,rinv00)));
288 /* Update potential sum for this i atom from the interaction with this j atom. */
289 velecsum = _mm256_add_pd(velecsum,velec);
293 /* Calculate temporary vectorial force */
294 tx = _mm256_mul_pd(fscal,dx00);
295 ty = _mm256_mul_pd(fscal,dy00);
296 tz = _mm256_mul_pd(fscal,dz00);
298 /* Update vectorial force */
299 fix0 = _mm256_add_pd(fix0,tx);
300 fiy0 = _mm256_add_pd(fiy0,ty);
301 fiz0 = _mm256_add_pd(fiz0,tz);
303 fjx0 = _mm256_add_pd(fjx0,tx);
304 fjy0 = _mm256_add_pd(fjy0,ty);
305 fjz0 = _mm256_add_pd(fjz0,tz);
307 /**************************
308 * CALCULATE INTERACTIONS *
309 **************************/
311 r01 = _mm256_mul_pd(rsq01,rinv01);
313 /* Calculate table index by multiplying r with table scale and truncate to integer */
314 rt = _mm256_mul_pd(r01,vftabscale);
315 vfitab = _mm256_cvttpd_epi32(rt);
316 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
317 vfitab = _mm_slli_epi32(vfitab,2);
319 /* CUBIC SPLINE TABLE ELECTROSTATICS */
320 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
321 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
322 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
323 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
324 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
325 Heps = _mm256_mul_pd(vfeps,H);
326 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
327 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
328 velec = _mm256_mul_pd(qq01,VV);
329 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
330 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq01,FF),_mm256_mul_pd(vftabscale,rinv01)));
332 /* Update potential sum for this i atom from the interaction with this j atom. */
333 velecsum = _mm256_add_pd(velecsum,velec);
337 /* Calculate temporary vectorial force */
338 tx = _mm256_mul_pd(fscal,dx01);
339 ty = _mm256_mul_pd(fscal,dy01);
340 tz = _mm256_mul_pd(fscal,dz01);
342 /* Update vectorial force */
343 fix0 = _mm256_add_pd(fix0,tx);
344 fiy0 = _mm256_add_pd(fiy0,ty);
345 fiz0 = _mm256_add_pd(fiz0,tz);
347 fjx1 = _mm256_add_pd(fjx1,tx);
348 fjy1 = _mm256_add_pd(fjy1,ty);
349 fjz1 = _mm256_add_pd(fjz1,tz);
351 /**************************
352 * CALCULATE INTERACTIONS *
353 **************************/
355 r02 = _mm256_mul_pd(rsq02,rinv02);
357 /* Calculate table index by multiplying r with table scale and truncate to integer */
358 rt = _mm256_mul_pd(r02,vftabscale);
359 vfitab = _mm256_cvttpd_epi32(rt);
360 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
361 vfitab = _mm_slli_epi32(vfitab,2);
363 /* CUBIC SPLINE TABLE ELECTROSTATICS */
364 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
365 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
366 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
367 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
368 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
369 Heps = _mm256_mul_pd(vfeps,H);
370 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
371 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
372 velec = _mm256_mul_pd(qq02,VV);
373 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
374 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq02,FF),_mm256_mul_pd(vftabscale,rinv02)));
376 /* Update potential sum for this i atom from the interaction with this j atom. */
377 velecsum = _mm256_add_pd(velecsum,velec);
381 /* Calculate temporary vectorial force */
382 tx = _mm256_mul_pd(fscal,dx02);
383 ty = _mm256_mul_pd(fscal,dy02);
384 tz = _mm256_mul_pd(fscal,dz02);
386 /* Update vectorial force */
387 fix0 = _mm256_add_pd(fix0,tx);
388 fiy0 = _mm256_add_pd(fiy0,ty);
389 fiz0 = _mm256_add_pd(fiz0,tz);
391 fjx2 = _mm256_add_pd(fjx2,tx);
392 fjy2 = _mm256_add_pd(fjy2,ty);
393 fjz2 = _mm256_add_pd(fjz2,tz);
395 /**************************
396 * CALCULATE INTERACTIONS *
397 **************************/
399 r10 = _mm256_mul_pd(rsq10,rinv10);
401 /* Calculate table index by multiplying r with table scale and truncate to integer */
402 rt = _mm256_mul_pd(r10,vftabscale);
403 vfitab = _mm256_cvttpd_epi32(rt);
404 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
405 vfitab = _mm_slli_epi32(vfitab,2);
407 /* CUBIC SPLINE TABLE ELECTROSTATICS */
408 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
409 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
410 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
411 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
412 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
413 Heps = _mm256_mul_pd(vfeps,H);
414 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
415 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
416 velec = _mm256_mul_pd(qq10,VV);
417 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
418 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq10,FF),_mm256_mul_pd(vftabscale,rinv10)));
420 /* Update potential sum for this i atom from the interaction with this j atom. */
421 velecsum = _mm256_add_pd(velecsum,velec);
425 /* Calculate temporary vectorial force */
426 tx = _mm256_mul_pd(fscal,dx10);
427 ty = _mm256_mul_pd(fscal,dy10);
428 tz = _mm256_mul_pd(fscal,dz10);
430 /* Update vectorial force */
431 fix1 = _mm256_add_pd(fix1,tx);
432 fiy1 = _mm256_add_pd(fiy1,ty);
433 fiz1 = _mm256_add_pd(fiz1,tz);
435 fjx0 = _mm256_add_pd(fjx0,tx);
436 fjy0 = _mm256_add_pd(fjy0,ty);
437 fjz0 = _mm256_add_pd(fjz0,tz);
439 /**************************
440 * CALCULATE INTERACTIONS *
441 **************************/
443 r11 = _mm256_mul_pd(rsq11,rinv11);
445 /* Calculate table index by multiplying r with table scale and truncate to integer */
446 rt = _mm256_mul_pd(r11,vftabscale);
447 vfitab = _mm256_cvttpd_epi32(rt);
448 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
449 vfitab = _mm_slli_epi32(vfitab,2);
451 /* CUBIC SPLINE TABLE ELECTROSTATICS */
452 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
453 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
454 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
455 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
456 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
457 Heps = _mm256_mul_pd(vfeps,H);
458 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
459 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
460 velec = _mm256_mul_pd(qq11,VV);
461 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
462 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq11,FF),_mm256_mul_pd(vftabscale,rinv11)));
464 /* Update potential sum for this i atom from the interaction with this j atom. */
465 velecsum = _mm256_add_pd(velecsum,velec);
469 /* Calculate temporary vectorial force */
470 tx = _mm256_mul_pd(fscal,dx11);
471 ty = _mm256_mul_pd(fscal,dy11);
472 tz = _mm256_mul_pd(fscal,dz11);
474 /* Update vectorial force */
475 fix1 = _mm256_add_pd(fix1,tx);
476 fiy1 = _mm256_add_pd(fiy1,ty);
477 fiz1 = _mm256_add_pd(fiz1,tz);
479 fjx1 = _mm256_add_pd(fjx1,tx);
480 fjy1 = _mm256_add_pd(fjy1,ty);
481 fjz1 = _mm256_add_pd(fjz1,tz);
483 /**************************
484 * CALCULATE INTERACTIONS *
485 **************************/
487 r12 = _mm256_mul_pd(rsq12,rinv12);
489 /* Calculate table index by multiplying r with table scale and truncate to integer */
490 rt = _mm256_mul_pd(r12,vftabscale);
491 vfitab = _mm256_cvttpd_epi32(rt);
492 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
493 vfitab = _mm_slli_epi32(vfitab,2);
495 /* CUBIC SPLINE TABLE ELECTROSTATICS */
496 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
497 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
498 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
499 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
500 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
501 Heps = _mm256_mul_pd(vfeps,H);
502 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
503 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
504 velec = _mm256_mul_pd(qq12,VV);
505 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
506 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq12,FF),_mm256_mul_pd(vftabscale,rinv12)));
508 /* Update potential sum for this i atom from the interaction with this j atom. */
509 velecsum = _mm256_add_pd(velecsum,velec);
513 /* Calculate temporary vectorial force */
514 tx = _mm256_mul_pd(fscal,dx12);
515 ty = _mm256_mul_pd(fscal,dy12);
516 tz = _mm256_mul_pd(fscal,dz12);
518 /* Update vectorial force */
519 fix1 = _mm256_add_pd(fix1,tx);
520 fiy1 = _mm256_add_pd(fiy1,ty);
521 fiz1 = _mm256_add_pd(fiz1,tz);
523 fjx2 = _mm256_add_pd(fjx2,tx);
524 fjy2 = _mm256_add_pd(fjy2,ty);
525 fjz2 = _mm256_add_pd(fjz2,tz);
527 /**************************
528 * CALCULATE INTERACTIONS *
529 **************************/
531 r20 = _mm256_mul_pd(rsq20,rinv20);
533 /* Calculate table index by multiplying r with table scale and truncate to integer */
534 rt = _mm256_mul_pd(r20,vftabscale);
535 vfitab = _mm256_cvttpd_epi32(rt);
536 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
537 vfitab = _mm_slli_epi32(vfitab,2);
539 /* CUBIC SPLINE TABLE ELECTROSTATICS */
540 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
541 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
542 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
543 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
544 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
545 Heps = _mm256_mul_pd(vfeps,H);
546 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
547 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
548 velec = _mm256_mul_pd(qq20,VV);
549 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
550 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq20,FF),_mm256_mul_pd(vftabscale,rinv20)));
552 /* Update potential sum for this i atom from the interaction with this j atom. */
553 velecsum = _mm256_add_pd(velecsum,velec);
557 /* Calculate temporary vectorial force */
558 tx = _mm256_mul_pd(fscal,dx20);
559 ty = _mm256_mul_pd(fscal,dy20);
560 tz = _mm256_mul_pd(fscal,dz20);
562 /* Update vectorial force */
563 fix2 = _mm256_add_pd(fix2,tx);
564 fiy2 = _mm256_add_pd(fiy2,ty);
565 fiz2 = _mm256_add_pd(fiz2,tz);
567 fjx0 = _mm256_add_pd(fjx0,tx);
568 fjy0 = _mm256_add_pd(fjy0,ty);
569 fjz0 = _mm256_add_pd(fjz0,tz);
571 /**************************
572 * CALCULATE INTERACTIONS *
573 **************************/
575 r21 = _mm256_mul_pd(rsq21,rinv21);
577 /* Calculate table index by multiplying r with table scale and truncate to integer */
578 rt = _mm256_mul_pd(r21,vftabscale);
579 vfitab = _mm256_cvttpd_epi32(rt);
580 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
581 vfitab = _mm_slli_epi32(vfitab,2);
583 /* CUBIC SPLINE TABLE ELECTROSTATICS */
584 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
585 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
586 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
587 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
588 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
589 Heps = _mm256_mul_pd(vfeps,H);
590 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
591 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
592 velec = _mm256_mul_pd(qq21,VV);
593 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
594 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq21,FF),_mm256_mul_pd(vftabscale,rinv21)));
596 /* Update potential sum for this i atom from the interaction with this j atom. */
597 velecsum = _mm256_add_pd(velecsum,velec);
601 /* Calculate temporary vectorial force */
602 tx = _mm256_mul_pd(fscal,dx21);
603 ty = _mm256_mul_pd(fscal,dy21);
604 tz = _mm256_mul_pd(fscal,dz21);
606 /* Update vectorial force */
607 fix2 = _mm256_add_pd(fix2,tx);
608 fiy2 = _mm256_add_pd(fiy2,ty);
609 fiz2 = _mm256_add_pd(fiz2,tz);
611 fjx1 = _mm256_add_pd(fjx1,tx);
612 fjy1 = _mm256_add_pd(fjy1,ty);
613 fjz1 = _mm256_add_pd(fjz1,tz);
615 /**************************
616 * CALCULATE INTERACTIONS *
617 **************************/
619 r22 = _mm256_mul_pd(rsq22,rinv22);
621 /* Calculate table index by multiplying r with table scale and truncate to integer */
622 rt = _mm256_mul_pd(r22,vftabscale);
623 vfitab = _mm256_cvttpd_epi32(rt);
624 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
625 vfitab = _mm_slli_epi32(vfitab,2);
627 /* CUBIC SPLINE TABLE ELECTROSTATICS */
628 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
629 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
630 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
631 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
632 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
633 Heps = _mm256_mul_pd(vfeps,H);
634 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
635 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
636 velec = _mm256_mul_pd(qq22,VV);
637 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
638 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq22,FF),_mm256_mul_pd(vftabscale,rinv22)));
640 /* Update potential sum for this i atom from the interaction with this j atom. */
641 velecsum = _mm256_add_pd(velecsum,velec);
645 /* Calculate temporary vectorial force */
646 tx = _mm256_mul_pd(fscal,dx22);
647 ty = _mm256_mul_pd(fscal,dy22);
648 tz = _mm256_mul_pd(fscal,dz22);
650 /* Update vectorial force */
651 fix2 = _mm256_add_pd(fix2,tx);
652 fiy2 = _mm256_add_pd(fiy2,ty);
653 fiz2 = _mm256_add_pd(fiz2,tz);
655 fjx2 = _mm256_add_pd(fjx2,tx);
656 fjy2 = _mm256_add_pd(fjy2,ty);
657 fjz2 = _mm256_add_pd(fjz2,tz);
659 fjptrA = f+j_coord_offsetA;
660 fjptrB = f+j_coord_offsetB;
661 fjptrC = f+j_coord_offsetC;
662 fjptrD = f+j_coord_offsetD;
664 gmx_mm256_decrement_3rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
665 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,fjx2,fjy2,fjz2);
667 /* Inner loop uses 387 flops */
673 /* Get j neighbor index, and coordinate index */
674 jnrlistA = jjnr[jidx];
675 jnrlistB = jjnr[jidx+1];
676 jnrlistC = jjnr[jidx+2];
677 jnrlistD = jjnr[jidx+3];
678 /* Sign of each element will be negative for non-real atoms.
679 * This mask will be 0xFFFFFFFF for dummy entries and 0x0 for real ones,
680 * so use it as val = _mm_andnot_pd(mask,val) to clear dummy entries.
682 tmpmask0 = gmx_mm_castsi128_pd(_mm_cmplt_epi32(_mm_loadu_si128((const __m128i *)(jjnr+jidx)),_mm_setzero_si128()));
684 tmpmask1 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(3,3,2,2));
685 tmpmask0 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(1,1,0,0));
686 dummy_mask = _mm256_castps_pd(gmx_mm256_set_m128(tmpmask1,tmpmask0));
688 jnrA = (jnrlistA>=0) ? jnrlistA : 0;
689 jnrB = (jnrlistB>=0) ? jnrlistB : 0;
690 jnrC = (jnrlistC>=0) ? jnrlistC : 0;
691 jnrD = (jnrlistD>=0) ? jnrlistD : 0;
692 j_coord_offsetA = DIM*jnrA;
693 j_coord_offsetB = DIM*jnrB;
694 j_coord_offsetC = DIM*jnrC;
695 j_coord_offsetD = DIM*jnrD;
697 /* load j atom coordinates */
698 gmx_mm256_load_3rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
699 x+j_coord_offsetC,x+j_coord_offsetD,
700 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,&jy2,&jz2);
702 /* Calculate displacement vector */
703 dx00 = _mm256_sub_pd(ix0,jx0);
704 dy00 = _mm256_sub_pd(iy0,jy0);
705 dz00 = _mm256_sub_pd(iz0,jz0);
706 dx01 = _mm256_sub_pd(ix0,jx1);
707 dy01 = _mm256_sub_pd(iy0,jy1);
708 dz01 = _mm256_sub_pd(iz0,jz1);
709 dx02 = _mm256_sub_pd(ix0,jx2);
710 dy02 = _mm256_sub_pd(iy0,jy2);
711 dz02 = _mm256_sub_pd(iz0,jz2);
712 dx10 = _mm256_sub_pd(ix1,jx0);
713 dy10 = _mm256_sub_pd(iy1,jy0);
714 dz10 = _mm256_sub_pd(iz1,jz0);
715 dx11 = _mm256_sub_pd(ix1,jx1);
716 dy11 = _mm256_sub_pd(iy1,jy1);
717 dz11 = _mm256_sub_pd(iz1,jz1);
718 dx12 = _mm256_sub_pd(ix1,jx2);
719 dy12 = _mm256_sub_pd(iy1,jy2);
720 dz12 = _mm256_sub_pd(iz1,jz2);
721 dx20 = _mm256_sub_pd(ix2,jx0);
722 dy20 = _mm256_sub_pd(iy2,jy0);
723 dz20 = _mm256_sub_pd(iz2,jz0);
724 dx21 = _mm256_sub_pd(ix2,jx1);
725 dy21 = _mm256_sub_pd(iy2,jy1);
726 dz21 = _mm256_sub_pd(iz2,jz1);
727 dx22 = _mm256_sub_pd(ix2,jx2);
728 dy22 = _mm256_sub_pd(iy2,jy2);
729 dz22 = _mm256_sub_pd(iz2,jz2);
731 /* Calculate squared distance and things based on it */
732 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
733 rsq01 = gmx_mm256_calc_rsq_pd(dx01,dy01,dz01);
734 rsq02 = gmx_mm256_calc_rsq_pd(dx02,dy02,dz02);
735 rsq10 = gmx_mm256_calc_rsq_pd(dx10,dy10,dz10);
736 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
737 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
738 rsq20 = gmx_mm256_calc_rsq_pd(dx20,dy20,dz20);
739 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
740 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
742 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
743 rinv01 = gmx_mm256_invsqrt_pd(rsq01);
744 rinv02 = gmx_mm256_invsqrt_pd(rsq02);
745 rinv10 = gmx_mm256_invsqrt_pd(rsq10);
746 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
747 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
748 rinv20 = gmx_mm256_invsqrt_pd(rsq20);
749 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
750 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
752 fjx0 = _mm256_setzero_pd();
753 fjy0 = _mm256_setzero_pd();
754 fjz0 = _mm256_setzero_pd();
755 fjx1 = _mm256_setzero_pd();
756 fjy1 = _mm256_setzero_pd();
757 fjz1 = _mm256_setzero_pd();
758 fjx2 = _mm256_setzero_pd();
759 fjy2 = _mm256_setzero_pd();
760 fjz2 = _mm256_setzero_pd();
762 /**************************
763 * CALCULATE INTERACTIONS *
764 **************************/
766 r00 = _mm256_mul_pd(rsq00,rinv00);
767 r00 = _mm256_andnot_pd(dummy_mask,r00);
769 /* Calculate table index by multiplying r with table scale and truncate to integer */
770 rt = _mm256_mul_pd(r00,vftabscale);
771 vfitab = _mm256_cvttpd_epi32(rt);
772 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
773 vfitab = _mm_slli_epi32(vfitab,2);
775 /* CUBIC SPLINE TABLE ELECTROSTATICS */
776 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
777 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
778 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
779 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
780 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
781 Heps = _mm256_mul_pd(vfeps,H);
782 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
783 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
784 velec = _mm256_mul_pd(qq00,VV);
785 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
786 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq00,FF),_mm256_mul_pd(vftabscale,rinv00)));
788 /* Update potential sum for this i atom from the interaction with this j atom. */
789 velec = _mm256_andnot_pd(dummy_mask,velec);
790 velecsum = _mm256_add_pd(velecsum,velec);
794 fscal = _mm256_andnot_pd(dummy_mask,fscal);
796 /* Calculate temporary vectorial force */
797 tx = _mm256_mul_pd(fscal,dx00);
798 ty = _mm256_mul_pd(fscal,dy00);
799 tz = _mm256_mul_pd(fscal,dz00);
801 /* Update vectorial force */
802 fix0 = _mm256_add_pd(fix0,tx);
803 fiy0 = _mm256_add_pd(fiy0,ty);
804 fiz0 = _mm256_add_pd(fiz0,tz);
806 fjx0 = _mm256_add_pd(fjx0,tx);
807 fjy0 = _mm256_add_pd(fjy0,ty);
808 fjz0 = _mm256_add_pd(fjz0,tz);
810 /**************************
811 * CALCULATE INTERACTIONS *
812 **************************/
814 r01 = _mm256_mul_pd(rsq01,rinv01);
815 r01 = _mm256_andnot_pd(dummy_mask,r01);
817 /* Calculate table index by multiplying r with table scale and truncate to integer */
818 rt = _mm256_mul_pd(r01,vftabscale);
819 vfitab = _mm256_cvttpd_epi32(rt);
820 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
821 vfitab = _mm_slli_epi32(vfitab,2);
823 /* CUBIC SPLINE TABLE ELECTROSTATICS */
824 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
825 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
826 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
827 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
828 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
829 Heps = _mm256_mul_pd(vfeps,H);
830 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
831 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
832 velec = _mm256_mul_pd(qq01,VV);
833 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
834 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq01,FF),_mm256_mul_pd(vftabscale,rinv01)));
836 /* Update potential sum for this i atom from the interaction with this j atom. */
837 velec = _mm256_andnot_pd(dummy_mask,velec);
838 velecsum = _mm256_add_pd(velecsum,velec);
842 fscal = _mm256_andnot_pd(dummy_mask,fscal);
844 /* Calculate temporary vectorial force */
845 tx = _mm256_mul_pd(fscal,dx01);
846 ty = _mm256_mul_pd(fscal,dy01);
847 tz = _mm256_mul_pd(fscal,dz01);
849 /* Update vectorial force */
850 fix0 = _mm256_add_pd(fix0,tx);
851 fiy0 = _mm256_add_pd(fiy0,ty);
852 fiz0 = _mm256_add_pd(fiz0,tz);
854 fjx1 = _mm256_add_pd(fjx1,tx);
855 fjy1 = _mm256_add_pd(fjy1,ty);
856 fjz1 = _mm256_add_pd(fjz1,tz);
858 /**************************
859 * CALCULATE INTERACTIONS *
860 **************************/
862 r02 = _mm256_mul_pd(rsq02,rinv02);
863 r02 = _mm256_andnot_pd(dummy_mask,r02);
865 /* Calculate table index by multiplying r with table scale and truncate to integer */
866 rt = _mm256_mul_pd(r02,vftabscale);
867 vfitab = _mm256_cvttpd_epi32(rt);
868 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
869 vfitab = _mm_slli_epi32(vfitab,2);
871 /* CUBIC SPLINE TABLE ELECTROSTATICS */
872 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
873 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
874 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
875 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
876 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
877 Heps = _mm256_mul_pd(vfeps,H);
878 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
879 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
880 velec = _mm256_mul_pd(qq02,VV);
881 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
882 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq02,FF),_mm256_mul_pd(vftabscale,rinv02)));
884 /* Update potential sum for this i atom from the interaction with this j atom. */
885 velec = _mm256_andnot_pd(dummy_mask,velec);
886 velecsum = _mm256_add_pd(velecsum,velec);
890 fscal = _mm256_andnot_pd(dummy_mask,fscal);
892 /* Calculate temporary vectorial force */
893 tx = _mm256_mul_pd(fscal,dx02);
894 ty = _mm256_mul_pd(fscal,dy02);
895 tz = _mm256_mul_pd(fscal,dz02);
897 /* Update vectorial force */
898 fix0 = _mm256_add_pd(fix0,tx);
899 fiy0 = _mm256_add_pd(fiy0,ty);
900 fiz0 = _mm256_add_pd(fiz0,tz);
902 fjx2 = _mm256_add_pd(fjx2,tx);
903 fjy2 = _mm256_add_pd(fjy2,ty);
904 fjz2 = _mm256_add_pd(fjz2,tz);
906 /**************************
907 * CALCULATE INTERACTIONS *
908 **************************/
910 r10 = _mm256_mul_pd(rsq10,rinv10);
911 r10 = _mm256_andnot_pd(dummy_mask,r10);
913 /* Calculate table index by multiplying r with table scale and truncate to integer */
914 rt = _mm256_mul_pd(r10,vftabscale);
915 vfitab = _mm256_cvttpd_epi32(rt);
916 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
917 vfitab = _mm_slli_epi32(vfitab,2);
919 /* CUBIC SPLINE TABLE ELECTROSTATICS */
920 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
921 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
922 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
923 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
924 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
925 Heps = _mm256_mul_pd(vfeps,H);
926 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
927 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
928 velec = _mm256_mul_pd(qq10,VV);
929 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
930 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq10,FF),_mm256_mul_pd(vftabscale,rinv10)));
932 /* Update potential sum for this i atom from the interaction with this j atom. */
933 velec = _mm256_andnot_pd(dummy_mask,velec);
934 velecsum = _mm256_add_pd(velecsum,velec);
938 fscal = _mm256_andnot_pd(dummy_mask,fscal);
940 /* Calculate temporary vectorial force */
941 tx = _mm256_mul_pd(fscal,dx10);
942 ty = _mm256_mul_pd(fscal,dy10);
943 tz = _mm256_mul_pd(fscal,dz10);
945 /* Update vectorial force */
946 fix1 = _mm256_add_pd(fix1,tx);
947 fiy1 = _mm256_add_pd(fiy1,ty);
948 fiz1 = _mm256_add_pd(fiz1,tz);
950 fjx0 = _mm256_add_pd(fjx0,tx);
951 fjy0 = _mm256_add_pd(fjy0,ty);
952 fjz0 = _mm256_add_pd(fjz0,tz);
954 /**************************
955 * CALCULATE INTERACTIONS *
956 **************************/
958 r11 = _mm256_mul_pd(rsq11,rinv11);
959 r11 = _mm256_andnot_pd(dummy_mask,r11);
961 /* Calculate table index by multiplying r with table scale and truncate to integer */
962 rt = _mm256_mul_pd(r11,vftabscale);
963 vfitab = _mm256_cvttpd_epi32(rt);
964 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
965 vfitab = _mm_slli_epi32(vfitab,2);
967 /* CUBIC SPLINE TABLE ELECTROSTATICS */
968 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
969 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
970 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
971 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
972 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
973 Heps = _mm256_mul_pd(vfeps,H);
974 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
975 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
976 velec = _mm256_mul_pd(qq11,VV);
977 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
978 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq11,FF),_mm256_mul_pd(vftabscale,rinv11)));
980 /* Update potential sum for this i atom from the interaction with this j atom. */
981 velec = _mm256_andnot_pd(dummy_mask,velec);
982 velecsum = _mm256_add_pd(velecsum,velec);
986 fscal = _mm256_andnot_pd(dummy_mask,fscal);
988 /* Calculate temporary vectorial force */
989 tx = _mm256_mul_pd(fscal,dx11);
990 ty = _mm256_mul_pd(fscal,dy11);
991 tz = _mm256_mul_pd(fscal,dz11);
993 /* Update vectorial force */
994 fix1 = _mm256_add_pd(fix1,tx);
995 fiy1 = _mm256_add_pd(fiy1,ty);
996 fiz1 = _mm256_add_pd(fiz1,tz);
998 fjx1 = _mm256_add_pd(fjx1,tx);
999 fjy1 = _mm256_add_pd(fjy1,ty);
1000 fjz1 = _mm256_add_pd(fjz1,tz);
1002 /**************************
1003 * CALCULATE INTERACTIONS *
1004 **************************/
1006 r12 = _mm256_mul_pd(rsq12,rinv12);
1007 r12 = _mm256_andnot_pd(dummy_mask,r12);
1009 /* Calculate table index by multiplying r with table scale and truncate to integer */
1010 rt = _mm256_mul_pd(r12,vftabscale);
1011 vfitab = _mm256_cvttpd_epi32(rt);
1012 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1013 vfitab = _mm_slli_epi32(vfitab,2);
1015 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1016 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1017 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1018 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1019 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1020 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1021 Heps = _mm256_mul_pd(vfeps,H);
1022 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1023 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
1024 velec = _mm256_mul_pd(qq12,VV);
1025 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1026 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq12,FF),_mm256_mul_pd(vftabscale,rinv12)));
1028 /* Update potential sum for this i atom from the interaction with this j atom. */
1029 velec = _mm256_andnot_pd(dummy_mask,velec);
1030 velecsum = _mm256_add_pd(velecsum,velec);
1034 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1036 /* Calculate temporary vectorial force */
1037 tx = _mm256_mul_pd(fscal,dx12);
1038 ty = _mm256_mul_pd(fscal,dy12);
1039 tz = _mm256_mul_pd(fscal,dz12);
1041 /* Update vectorial force */
1042 fix1 = _mm256_add_pd(fix1,tx);
1043 fiy1 = _mm256_add_pd(fiy1,ty);
1044 fiz1 = _mm256_add_pd(fiz1,tz);
1046 fjx2 = _mm256_add_pd(fjx2,tx);
1047 fjy2 = _mm256_add_pd(fjy2,ty);
1048 fjz2 = _mm256_add_pd(fjz2,tz);
1050 /**************************
1051 * CALCULATE INTERACTIONS *
1052 **************************/
1054 r20 = _mm256_mul_pd(rsq20,rinv20);
1055 r20 = _mm256_andnot_pd(dummy_mask,r20);
1057 /* Calculate table index by multiplying r with table scale and truncate to integer */
1058 rt = _mm256_mul_pd(r20,vftabscale);
1059 vfitab = _mm256_cvttpd_epi32(rt);
1060 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1061 vfitab = _mm_slli_epi32(vfitab,2);
1063 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1064 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1065 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1066 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1067 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1068 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1069 Heps = _mm256_mul_pd(vfeps,H);
1070 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1071 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
1072 velec = _mm256_mul_pd(qq20,VV);
1073 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1074 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq20,FF),_mm256_mul_pd(vftabscale,rinv20)));
1076 /* Update potential sum for this i atom from the interaction with this j atom. */
1077 velec = _mm256_andnot_pd(dummy_mask,velec);
1078 velecsum = _mm256_add_pd(velecsum,velec);
1082 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1084 /* Calculate temporary vectorial force */
1085 tx = _mm256_mul_pd(fscal,dx20);
1086 ty = _mm256_mul_pd(fscal,dy20);
1087 tz = _mm256_mul_pd(fscal,dz20);
1089 /* Update vectorial force */
1090 fix2 = _mm256_add_pd(fix2,tx);
1091 fiy2 = _mm256_add_pd(fiy2,ty);
1092 fiz2 = _mm256_add_pd(fiz2,tz);
1094 fjx0 = _mm256_add_pd(fjx0,tx);
1095 fjy0 = _mm256_add_pd(fjy0,ty);
1096 fjz0 = _mm256_add_pd(fjz0,tz);
1098 /**************************
1099 * CALCULATE INTERACTIONS *
1100 **************************/
1102 r21 = _mm256_mul_pd(rsq21,rinv21);
1103 r21 = _mm256_andnot_pd(dummy_mask,r21);
1105 /* Calculate table index by multiplying r with table scale and truncate to integer */
1106 rt = _mm256_mul_pd(r21,vftabscale);
1107 vfitab = _mm256_cvttpd_epi32(rt);
1108 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1109 vfitab = _mm_slli_epi32(vfitab,2);
1111 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1112 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1113 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1114 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1115 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1116 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1117 Heps = _mm256_mul_pd(vfeps,H);
1118 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1119 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
1120 velec = _mm256_mul_pd(qq21,VV);
1121 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1122 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq21,FF),_mm256_mul_pd(vftabscale,rinv21)));
1124 /* Update potential sum for this i atom from the interaction with this j atom. */
1125 velec = _mm256_andnot_pd(dummy_mask,velec);
1126 velecsum = _mm256_add_pd(velecsum,velec);
1130 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1132 /* Calculate temporary vectorial force */
1133 tx = _mm256_mul_pd(fscal,dx21);
1134 ty = _mm256_mul_pd(fscal,dy21);
1135 tz = _mm256_mul_pd(fscal,dz21);
1137 /* Update vectorial force */
1138 fix2 = _mm256_add_pd(fix2,tx);
1139 fiy2 = _mm256_add_pd(fiy2,ty);
1140 fiz2 = _mm256_add_pd(fiz2,tz);
1142 fjx1 = _mm256_add_pd(fjx1,tx);
1143 fjy1 = _mm256_add_pd(fjy1,ty);
1144 fjz1 = _mm256_add_pd(fjz1,tz);
1146 /**************************
1147 * CALCULATE INTERACTIONS *
1148 **************************/
1150 r22 = _mm256_mul_pd(rsq22,rinv22);
1151 r22 = _mm256_andnot_pd(dummy_mask,r22);
1153 /* Calculate table index by multiplying r with table scale and truncate to integer */
1154 rt = _mm256_mul_pd(r22,vftabscale);
1155 vfitab = _mm256_cvttpd_epi32(rt);
1156 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1157 vfitab = _mm_slli_epi32(vfitab,2);
1159 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1160 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1161 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1162 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1163 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1164 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1165 Heps = _mm256_mul_pd(vfeps,H);
1166 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1167 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
1168 velec = _mm256_mul_pd(qq22,VV);
1169 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1170 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq22,FF),_mm256_mul_pd(vftabscale,rinv22)));
1172 /* Update potential sum for this i atom from the interaction with this j atom. */
1173 velec = _mm256_andnot_pd(dummy_mask,velec);
1174 velecsum = _mm256_add_pd(velecsum,velec);
1178 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1180 /* Calculate temporary vectorial force */
1181 tx = _mm256_mul_pd(fscal,dx22);
1182 ty = _mm256_mul_pd(fscal,dy22);
1183 tz = _mm256_mul_pd(fscal,dz22);
1185 /* Update vectorial force */
1186 fix2 = _mm256_add_pd(fix2,tx);
1187 fiy2 = _mm256_add_pd(fiy2,ty);
1188 fiz2 = _mm256_add_pd(fiz2,tz);
1190 fjx2 = _mm256_add_pd(fjx2,tx);
1191 fjy2 = _mm256_add_pd(fjy2,ty);
1192 fjz2 = _mm256_add_pd(fjz2,tz);
1194 fjptrA = (jnrlistA>=0) ? f+j_coord_offsetA : scratch;
1195 fjptrB = (jnrlistB>=0) ? f+j_coord_offsetB : scratch;
1196 fjptrC = (jnrlistC>=0) ? f+j_coord_offsetC : scratch;
1197 fjptrD = (jnrlistD>=0) ? f+j_coord_offsetD : scratch;
1199 gmx_mm256_decrement_3rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
1200 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,fjx2,fjy2,fjz2);
1202 /* Inner loop uses 396 flops */
1205 /* End of innermost loop */
1207 gmx_mm256_update_iforce_3atom_swizzle_pd(fix0,fiy0,fiz0,fix1,fiy1,fiz1,fix2,fiy2,fiz2,
1208 f+i_coord_offset,fshift+i_shift_offset);
1211 /* Update potential energies */
1212 gmx_mm256_update_1pot_pd(velecsum,kernel_data->energygrp_elec+ggid);
1214 /* Increment number of inner iterations */
1215 inneriter += j_index_end - j_index_start;
1217 /* Outer loop uses 19 flops */
1220 /* Increment number of outer iterations */
1223 /* Update outer/inner flops */
1225 inc_nrnb(nrnb,eNR_NBKERNEL_ELEC_W3W3_VF,outeriter*19 + inneriter*396);
1228 * Gromacs nonbonded kernel: nb_kernel_ElecCSTab_VdwNone_GeomW3W3_F_avx_256_double
1229 * Electrostatics interaction: CubicSplineTable
1230 * VdW interaction: None
1231 * Geometry: Water3-Water3
1232 * Calculate force/pot: Force
1235 nb_kernel_ElecCSTab_VdwNone_GeomW3W3_F_avx_256_double
1236 (t_nblist * gmx_restrict nlist,
1237 rvec * gmx_restrict xx,
1238 rvec * gmx_restrict ff,
1239 t_forcerec * gmx_restrict fr,
1240 t_mdatoms * gmx_restrict mdatoms,
1241 nb_kernel_data_t * gmx_restrict kernel_data,
1242 t_nrnb * gmx_restrict nrnb)
1244 /* Suffixes 0,1,2,3 refer to particle indices for waters in the inner or outer loop, or
1245 * just 0 for non-waters.
1246 * Suffixes A,B,C,D refer to j loop unrolling done with AVX, e.g. for the four different
1247 * jnr indices corresponding to data put in the four positions in the SIMD register.
1249 int i_shift_offset,i_coord_offset,outeriter,inneriter;
1250 int j_index_start,j_index_end,jidx,nri,inr,ggid,iidx;
1251 int jnrA,jnrB,jnrC,jnrD;
1252 int jnrlistA,jnrlistB,jnrlistC,jnrlistD;
1253 int jnrlistE,jnrlistF,jnrlistG,jnrlistH;
1254 int j_coord_offsetA,j_coord_offsetB,j_coord_offsetC,j_coord_offsetD;
1255 int *iinr,*jindex,*jjnr,*shiftidx,*gid;
1256 real rcutoff_scalar;
1257 real *shiftvec,*fshift,*x,*f;
1258 real *fjptrA,*fjptrB,*fjptrC,*fjptrD;
1259 real scratch[4*DIM];
1260 __m256d tx,ty,tz,fscal,rcutoff,rcutoff2,jidxall;
1261 real * vdwioffsetptr0;
1262 __m256d ix0,iy0,iz0,fix0,fiy0,fiz0,iq0,isai0;
1263 real * vdwioffsetptr1;
1264 __m256d ix1,iy1,iz1,fix1,fiy1,fiz1,iq1,isai1;
1265 real * vdwioffsetptr2;
1266 __m256d ix2,iy2,iz2,fix2,fiy2,fiz2,iq2,isai2;
1267 int vdwjidx0A,vdwjidx0B,vdwjidx0C,vdwjidx0D;
1268 __m256d jx0,jy0,jz0,fjx0,fjy0,fjz0,jq0,isaj0;
1269 int vdwjidx1A,vdwjidx1B,vdwjidx1C,vdwjidx1D;
1270 __m256d jx1,jy1,jz1,fjx1,fjy1,fjz1,jq1,isaj1;
1271 int vdwjidx2A,vdwjidx2B,vdwjidx2C,vdwjidx2D;
1272 __m256d jx2,jy2,jz2,fjx2,fjy2,fjz2,jq2,isaj2;
1273 __m256d dx00,dy00,dz00,rsq00,rinv00,rinvsq00,r00,qq00,c6_00,c12_00;
1274 __m256d dx01,dy01,dz01,rsq01,rinv01,rinvsq01,r01,qq01,c6_01,c12_01;
1275 __m256d dx02,dy02,dz02,rsq02,rinv02,rinvsq02,r02,qq02,c6_02,c12_02;
1276 __m256d dx10,dy10,dz10,rsq10,rinv10,rinvsq10,r10,qq10,c6_10,c12_10;
1277 __m256d dx11,dy11,dz11,rsq11,rinv11,rinvsq11,r11,qq11,c6_11,c12_11;
1278 __m256d dx12,dy12,dz12,rsq12,rinv12,rinvsq12,r12,qq12,c6_12,c12_12;
1279 __m256d dx20,dy20,dz20,rsq20,rinv20,rinvsq20,r20,qq20,c6_20,c12_20;
1280 __m256d dx21,dy21,dz21,rsq21,rinv21,rinvsq21,r21,qq21,c6_21,c12_21;
1281 __m256d dx22,dy22,dz22,rsq22,rinv22,rinvsq22,r22,qq22,c6_22,c12_22;
1282 __m256d velec,felec,velecsum,facel,crf,krf,krf2;
1285 __m128i ifour = _mm_set1_epi32(4);
1286 __m256d rt,vfeps,vftabscale,Y,F,G,H,Heps,Fp,VV,FF;
1288 __m256d dummy_mask,cutoff_mask;
1289 __m128 tmpmask0,tmpmask1;
1290 __m256d signbit = _mm256_castsi256_pd( _mm256_set1_epi32(0x80000000) );
1291 __m256d one = _mm256_set1_pd(1.0);
1292 __m256d two = _mm256_set1_pd(2.0);
1298 jindex = nlist->jindex;
1300 shiftidx = nlist->shift;
1302 shiftvec = fr->shift_vec[0];
1303 fshift = fr->fshift[0];
1304 facel = _mm256_set1_pd(fr->epsfac);
1305 charge = mdatoms->chargeA;
1307 vftab = kernel_data->table_elec->data;
1308 vftabscale = _mm256_set1_pd(kernel_data->table_elec->scale);
1310 /* Setup water-specific parameters */
1311 inr = nlist->iinr[0];
1312 iq0 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+0]));
1313 iq1 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+1]));
1314 iq2 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+2]));
1316 jq0 = _mm256_set1_pd(charge[inr+0]);
1317 jq1 = _mm256_set1_pd(charge[inr+1]);
1318 jq2 = _mm256_set1_pd(charge[inr+2]);
1319 qq00 = _mm256_mul_pd(iq0,jq0);
1320 qq01 = _mm256_mul_pd(iq0,jq1);
1321 qq02 = _mm256_mul_pd(iq0,jq2);
1322 qq10 = _mm256_mul_pd(iq1,jq0);
1323 qq11 = _mm256_mul_pd(iq1,jq1);
1324 qq12 = _mm256_mul_pd(iq1,jq2);
1325 qq20 = _mm256_mul_pd(iq2,jq0);
1326 qq21 = _mm256_mul_pd(iq2,jq1);
1327 qq22 = _mm256_mul_pd(iq2,jq2);
1329 /* Avoid stupid compiler warnings */
1330 jnrA = jnrB = jnrC = jnrD = 0;
1331 j_coord_offsetA = 0;
1332 j_coord_offsetB = 0;
1333 j_coord_offsetC = 0;
1334 j_coord_offsetD = 0;
1339 for(iidx=0;iidx<4*DIM;iidx++)
1341 scratch[iidx] = 0.0;
1344 /* Start outer loop over neighborlists */
1345 for(iidx=0; iidx<nri; iidx++)
1347 /* Load shift vector for this list */
1348 i_shift_offset = DIM*shiftidx[iidx];
1350 /* Load limits for loop over neighbors */
1351 j_index_start = jindex[iidx];
1352 j_index_end = jindex[iidx+1];
1354 /* Get outer coordinate index */
1356 i_coord_offset = DIM*inr;
1358 /* Load i particle coords and add shift vector */
1359 gmx_mm256_load_shift_and_3rvec_broadcast_pd(shiftvec+i_shift_offset,x+i_coord_offset,
1360 &ix0,&iy0,&iz0,&ix1,&iy1,&iz1,&ix2,&iy2,&iz2);
1362 fix0 = _mm256_setzero_pd();
1363 fiy0 = _mm256_setzero_pd();
1364 fiz0 = _mm256_setzero_pd();
1365 fix1 = _mm256_setzero_pd();
1366 fiy1 = _mm256_setzero_pd();
1367 fiz1 = _mm256_setzero_pd();
1368 fix2 = _mm256_setzero_pd();
1369 fiy2 = _mm256_setzero_pd();
1370 fiz2 = _mm256_setzero_pd();
1372 /* Start inner kernel loop */
1373 for(jidx=j_index_start; jidx<j_index_end && jjnr[jidx+3]>=0; jidx+=4)
1376 /* Get j neighbor index, and coordinate index */
1378 jnrB = jjnr[jidx+1];
1379 jnrC = jjnr[jidx+2];
1380 jnrD = jjnr[jidx+3];
1381 j_coord_offsetA = DIM*jnrA;
1382 j_coord_offsetB = DIM*jnrB;
1383 j_coord_offsetC = DIM*jnrC;
1384 j_coord_offsetD = DIM*jnrD;
1386 /* load j atom coordinates */
1387 gmx_mm256_load_3rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
1388 x+j_coord_offsetC,x+j_coord_offsetD,
1389 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,&jy2,&jz2);
1391 /* Calculate displacement vector */
1392 dx00 = _mm256_sub_pd(ix0,jx0);
1393 dy00 = _mm256_sub_pd(iy0,jy0);
1394 dz00 = _mm256_sub_pd(iz0,jz0);
1395 dx01 = _mm256_sub_pd(ix0,jx1);
1396 dy01 = _mm256_sub_pd(iy0,jy1);
1397 dz01 = _mm256_sub_pd(iz0,jz1);
1398 dx02 = _mm256_sub_pd(ix0,jx2);
1399 dy02 = _mm256_sub_pd(iy0,jy2);
1400 dz02 = _mm256_sub_pd(iz0,jz2);
1401 dx10 = _mm256_sub_pd(ix1,jx0);
1402 dy10 = _mm256_sub_pd(iy1,jy0);
1403 dz10 = _mm256_sub_pd(iz1,jz0);
1404 dx11 = _mm256_sub_pd(ix1,jx1);
1405 dy11 = _mm256_sub_pd(iy1,jy1);
1406 dz11 = _mm256_sub_pd(iz1,jz1);
1407 dx12 = _mm256_sub_pd(ix1,jx2);
1408 dy12 = _mm256_sub_pd(iy1,jy2);
1409 dz12 = _mm256_sub_pd(iz1,jz2);
1410 dx20 = _mm256_sub_pd(ix2,jx0);
1411 dy20 = _mm256_sub_pd(iy2,jy0);
1412 dz20 = _mm256_sub_pd(iz2,jz0);
1413 dx21 = _mm256_sub_pd(ix2,jx1);
1414 dy21 = _mm256_sub_pd(iy2,jy1);
1415 dz21 = _mm256_sub_pd(iz2,jz1);
1416 dx22 = _mm256_sub_pd(ix2,jx2);
1417 dy22 = _mm256_sub_pd(iy2,jy2);
1418 dz22 = _mm256_sub_pd(iz2,jz2);
1420 /* Calculate squared distance and things based on it */
1421 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
1422 rsq01 = gmx_mm256_calc_rsq_pd(dx01,dy01,dz01);
1423 rsq02 = gmx_mm256_calc_rsq_pd(dx02,dy02,dz02);
1424 rsq10 = gmx_mm256_calc_rsq_pd(dx10,dy10,dz10);
1425 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
1426 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
1427 rsq20 = gmx_mm256_calc_rsq_pd(dx20,dy20,dz20);
1428 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
1429 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
1431 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
1432 rinv01 = gmx_mm256_invsqrt_pd(rsq01);
1433 rinv02 = gmx_mm256_invsqrt_pd(rsq02);
1434 rinv10 = gmx_mm256_invsqrt_pd(rsq10);
1435 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
1436 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
1437 rinv20 = gmx_mm256_invsqrt_pd(rsq20);
1438 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
1439 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
1441 fjx0 = _mm256_setzero_pd();
1442 fjy0 = _mm256_setzero_pd();
1443 fjz0 = _mm256_setzero_pd();
1444 fjx1 = _mm256_setzero_pd();
1445 fjy1 = _mm256_setzero_pd();
1446 fjz1 = _mm256_setzero_pd();
1447 fjx2 = _mm256_setzero_pd();
1448 fjy2 = _mm256_setzero_pd();
1449 fjz2 = _mm256_setzero_pd();
1451 /**************************
1452 * CALCULATE INTERACTIONS *
1453 **************************/
1455 r00 = _mm256_mul_pd(rsq00,rinv00);
1457 /* Calculate table index by multiplying r with table scale and truncate to integer */
1458 rt = _mm256_mul_pd(r00,vftabscale);
1459 vfitab = _mm256_cvttpd_epi32(rt);
1460 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1461 vfitab = _mm_slli_epi32(vfitab,2);
1463 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1464 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1465 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1466 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1467 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1468 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1469 Heps = _mm256_mul_pd(vfeps,H);
1470 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1471 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1472 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq00,FF),_mm256_mul_pd(vftabscale,rinv00)));
1476 /* Calculate temporary vectorial force */
1477 tx = _mm256_mul_pd(fscal,dx00);
1478 ty = _mm256_mul_pd(fscal,dy00);
1479 tz = _mm256_mul_pd(fscal,dz00);
1481 /* Update vectorial force */
1482 fix0 = _mm256_add_pd(fix0,tx);
1483 fiy0 = _mm256_add_pd(fiy0,ty);
1484 fiz0 = _mm256_add_pd(fiz0,tz);
1486 fjx0 = _mm256_add_pd(fjx0,tx);
1487 fjy0 = _mm256_add_pd(fjy0,ty);
1488 fjz0 = _mm256_add_pd(fjz0,tz);
1490 /**************************
1491 * CALCULATE INTERACTIONS *
1492 **************************/
1494 r01 = _mm256_mul_pd(rsq01,rinv01);
1496 /* Calculate table index by multiplying r with table scale and truncate to integer */
1497 rt = _mm256_mul_pd(r01,vftabscale);
1498 vfitab = _mm256_cvttpd_epi32(rt);
1499 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1500 vfitab = _mm_slli_epi32(vfitab,2);
1502 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1503 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1504 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1505 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1506 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1507 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1508 Heps = _mm256_mul_pd(vfeps,H);
1509 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1510 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1511 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq01,FF),_mm256_mul_pd(vftabscale,rinv01)));
1515 /* Calculate temporary vectorial force */
1516 tx = _mm256_mul_pd(fscal,dx01);
1517 ty = _mm256_mul_pd(fscal,dy01);
1518 tz = _mm256_mul_pd(fscal,dz01);
1520 /* Update vectorial force */
1521 fix0 = _mm256_add_pd(fix0,tx);
1522 fiy0 = _mm256_add_pd(fiy0,ty);
1523 fiz0 = _mm256_add_pd(fiz0,tz);
1525 fjx1 = _mm256_add_pd(fjx1,tx);
1526 fjy1 = _mm256_add_pd(fjy1,ty);
1527 fjz1 = _mm256_add_pd(fjz1,tz);
1529 /**************************
1530 * CALCULATE INTERACTIONS *
1531 **************************/
1533 r02 = _mm256_mul_pd(rsq02,rinv02);
1535 /* Calculate table index by multiplying r with table scale and truncate to integer */
1536 rt = _mm256_mul_pd(r02,vftabscale);
1537 vfitab = _mm256_cvttpd_epi32(rt);
1538 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1539 vfitab = _mm_slli_epi32(vfitab,2);
1541 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1542 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1543 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1544 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1545 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1546 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1547 Heps = _mm256_mul_pd(vfeps,H);
1548 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1549 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1550 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq02,FF),_mm256_mul_pd(vftabscale,rinv02)));
1554 /* Calculate temporary vectorial force */
1555 tx = _mm256_mul_pd(fscal,dx02);
1556 ty = _mm256_mul_pd(fscal,dy02);
1557 tz = _mm256_mul_pd(fscal,dz02);
1559 /* Update vectorial force */
1560 fix0 = _mm256_add_pd(fix0,tx);
1561 fiy0 = _mm256_add_pd(fiy0,ty);
1562 fiz0 = _mm256_add_pd(fiz0,tz);
1564 fjx2 = _mm256_add_pd(fjx2,tx);
1565 fjy2 = _mm256_add_pd(fjy2,ty);
1566 fjz2 = _mm256_add_pd(fjz2,tz);
1568 /**************************
1569 * CALCULATE INTERACTIONS *
1570 **************************/
1572 r10 = _mm256_mul_pd(rsq10,rinv10);
1574 /* Calculate table index by multiplying r with table scale and truncate to integer */
1575 rt = _mm256_mul_pd(r10,vftabscale);
1576 vfitab = _mm256_cvttpd_epi32(rt);
1577 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1578 vfitab = _mm_slli_epi32(vfitab,2);
1580 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1581 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1582 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1583 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1584 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1585 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1586 Heps = _mm256_mul_pd(vfeps,H);
1587 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1588 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1589 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq10,FF),_mm256_mul_pd(vftabscale,rinv10)));
1593 /* Calculate temporary vectorial force */
1594 tx = _mm256_mul_pd(fscal,dx10);
1595 ty = _mm256_mul_pd(fscal,dy10);
1596 tz = _mm256_mul_pd(fscal,dz10);
1598 /* Update vectorial force */
1599 fix1 = _mm256_add_pd(fix1,tx);
1600 fiy1 = _mm256_add_pd(fiy1,ty);
1601 fiz1 = _mm256_add_pd(fiz1,tz);
1603 fjx0 = _mm256_add_pd(fjx0,tx);
1604 fjy0 = _mm256_add_pd(fjy0,ty);
1605 fjz0 = _mm256_add_pd(fjz0,tz);
1607 /**************************
1608 * CALCULATE INTERACTIONS *
1609 **************************/
1611 r11 = _mm256_mul_pd(rsq11,rinv11);
1613 /* Calculate table index by multiplying r with table scale and truncate to integer */
1614 rt = _mm256_mul_pd(r11,vftabscale);
1615 vfitab = _mm256_cvttpd_epi32(rt);
1616 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1617 vfitab = _mm_slli_epi32(vfitab,2);
1619 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1620 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1621 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1622 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1623 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1624 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1625 Heps = _mm256_mul_pd(vfeps,H);
1626 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1627 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1628 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq11,FF),_mm256_mul_pd(vftabscale,rinv11)));
1632 /* Calculate temporary vectorial force */
1633 tx = _mm256_mul_pd(fscal,dx11);
1634 ty = _mm256_mul_pd(fscal,dy11);
1635 tz = _mm256_mul_pd(fscal,dz11);
1637 /* Update vectorial force */
1638 fix1 = _mm256_add_pd(fix1,tx);
1639 fiy1 = _mm256_add_pd(fiy1,ty);
1640 fiz1 = _mm256_add_pd(fiz1,tz);
1642 fjx1 = _mm256_add_pd(fjx1,tx);
1643 fjy1 = _mm256_add_pd(fjy1,ty);
1644 fjz1 = _mm256_add_pd(fjz1,tz);
1646 /**************************
1647 * CALCULATE INTERACTIONS *
1648 **************************/
1650 r12 = _mm256_mul_pd(rsq12,rinv12);
1652 /* Calculate table index by multiplying r with table scale and truncate to integer */
1653 rt = _mm256_mul_pd(r12,vftabscale);
1654 vfitab = _mm256_cvttpd_epi32(rt);
1655 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1656 vfitab = _mm_slli_epi32(vfitab,2);
1658 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1659 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1660 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1661 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1662 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1663 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1664 Heps = _mm256_mul_pd(vfeps,H);
1665 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1666 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1667 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq12,FF),_mm256_mul_pd(vftabscale,rinv12)));
1671 /* Calculate temporary vectorial force */
1672 tx = _mm256_mul_pd(fscal,dx12);
1673 ty = _mm256_mul_pd(fscal,dy12);
1674 tz = _mm256_mul_pd(fscal,dz12);
1676 /* Update vectorial force */
1677 fix1 = _mm256_add_pd(fix1,tx);
1678 fiy1 = _mm256_add_pd(fiy1,ty);
1679 fiz1 = _mm256_add_pd(fiz1,tz);
1681 fjx2 = _mm256_add_pd(fjx2,tx);
1682 fjy2 = _mm256_add_pd(fjy2,ty);
1683 fjz2 = _mm256_add_pd(fjz2,tz);
1685 /**************************
1686 * CALCULATE INTERACTIONS *
1687 **************************/
1689 r20 = _mm256_mul_pd(rsq20,rinv20);
1691 /* Calculate table index by multiplying r with table scale and truncate to integer */
1692 rt = _mm256_mul_pd(r20,vftabscale);
1693 vfitab = _mm256_cvttpd_epi32(rt);
1694 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1695 vfitab = _mm_slli_epi32(vfitab,2);
1697 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1698 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1699 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1700 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1701 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1702 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1703 Heps = _mm256_mul_pd(vfeps,H);
1704 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1705 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1706 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq20,FF),_mm256_mul_pd(vftabscale,rinv20)));
1710 /* Calculate temporary vectorial force */
1711 tx = _mm256_mul_pd(fscal,dx20);
1712 ty = _mm256_mul_pd(fscal,dy20);
1713 tz = _mm256_mul_pd(fscal,dz20);
1715 /* Update vectorial force */
1716 fix2 = _mm256_add_pd(fix2,tx);
1717 fiy2 = _mm256_add_pd(fiy2,ty);
1718 fiz2 = _mm256_add_pd(fiz2,tz);
1720 fjx0 = _mm256_add_pd(fjx0,tx);
1721 fjy0 = _mm256_add_pd(fjy0,ty);
1722 fjz0 = _mm256_add_pd(fjz0,tz);
1724 /**************************
1725 * CALCULATE INTERACTIONS *
1726 **************************/
1728 r21 = _mm256_mul_pd(rsq21,rinv21);
1730 /* Calculate table index by multiplying r with table scale and truncate to integer */
1731 rt = _mm256_mul_pd(r21,vftabscale);
1732 vfitab = _mm256_cvttpd_epi32(rt);
1733 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1734 vfitab = _mm_slli_epi32(vfitab,2);
1736 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1737 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1738 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1739 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1740 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1741 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1742 Heps = _mm256_mul_pd(vfeps,H);
1743 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1744 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1745 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq21,FF),_mm256_mul_pd(vftabscale,rinv21)));
1749 /* Calculate temporary vectorial force */
1750 tx = _mm256_mul_pd(fscal,dx21);
1751 ty = _mm256_mul_pd(fscal,dy21);
1752 tz = _mm256_mul_pd(fscal,dz21);
1754 /* Update vectorial force */
1755 fix2 = _mm256_add_pd(fix2,tx);
1756 fiy2 = _mm256_add_pd(fiy2,ty);
1757 fiz2 = _mm256_add_pd(fiz2,tz);
1759 fjx1 = _mm256_add_pd(fjx1,tx);
1760 fjy1 = _mm256_add_pd(fjy1,ty);
1761 fjz1 = _mm256_add_pd(fjz1,tz);
1763 /**************************
1764 * CALCULATE INTERACTIONS *
1765 **************************/
1767 r22 = _mm256_mul_pd(rsq22,rinv22);
1769 /* Calculate table index by multiplying r with table scale and truncate to integer */
1770 rt = _mm256_mul_pd(r22,vftabscale);
1771 vfitab = _mm256_cvttpd_epi32(rt);
1772 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1773 vfitab = _mm_slli_epi32(vfitab,2);
1775 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1776 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1777 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1778 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1779 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1780 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1781 Heps = _mm256_mul_pd(vfeps,H);
1782 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1783 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1784 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq22,FF),_mm256_mul_pd(vftabscale,rinv22)));
1788 /* Calculate temporary vectorial force */
1789 tx = _mm256_mul_pd(fscal,dx22);
1790 ty = _mm256_mul_pd(fscal,dy22);
1791 tz = _mm256_mul_pd(fscal,dz22);
1793 /* Update vectorial force */
1794 fix2 = _mm256_add_pd(fix2,tx);
1795 fiy2 = _mm256_add_pd(fiy2,ty);
1796 fiz2 = _mm256_add_pd(fiz2,tz);
1798 fjx2 = _mm256_add_pd(fjx2,tx);
1799 fjy2 = _mm256_add_pd(fjy2,ty);
1800 fjz2 = _mm256_add_pd(fjz2,tz);
1802 fjptrA = f+j_coord_offsetA;
1803 fjptrB = f+j_coord_offsetB;
1804 fjptrC = f+j_coord_offsetC;
1805 fjptrD = f+j_coord_offsetD;
1807 gmx_mm256_decrement_3rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
1808 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,fjx2,fjy2,fjz2);
1810 /* Inner loop uses 351 flops */
1813 if(jidx<j_index_end)
1816 /* Get j neighbor index, and coordinate index */
1817 jnrlistA = jjnr[jidx];
1818 jnrlistB = jjnr[jidx+1];
1819 jnrlistC = jjnr[jidx+2];
1820 jnrlistD = jjnr[jidx+3];
1821 /* Sign of each element will be negative for non-real atoms.
1822 * This mask will be 0xFFFFFFFF for dummy entries and 0x0 for real ones,
1823 * so use it as val = _mm_andnot_pd(mask,val) to clear dummy entries.
1825 tmpmask0 = gmx_mm_castsi128_pd(_mm_cmplt_epi32(_mm_loadu_si128((const __m128i *)(jjnr+jidx)),_mm_setzero_si128()));
1827 tmpmask1 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(3,3,2,2));
1828 tmpmask0 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(1,1,0,0));
1829 dummy_mask = _mm256_castps_pd(gmx_mm256_set_m128(tmpmask1,tmpmask0));
1831 jnrA = (jnrlistA>=0) ? jnrlistA : 0;
1832 jnrB = (jnrlistB>=0) ? jnrlistB : 0;
1833 jnrC = (jnrlistC>=0) ? jnrlistC : 0;
1834 jnrD = (jnrlistD>=0) ? jnrlistD : 0;
1835 j_coord_offsetA = DIM*jnrA;
1836 j_coord_offsetB = DIM*jnrB;
1837 j_coord_offsetC = DIM*jnrC;
1838 j_coord_offsetD = DIM*jnrD;
1840 /* load j atom coordinates */
1841 gmx_mm256_load_3rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
1842 x+j_coord_offsetC,x+j_coord_offsetD,
1843 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,&jy2,&jz2);
1845 /* Calculate displacement vector */
1846 dx00 = _mm256_sub_pd(ix0,jx0);
1847 dy00 = _mm256_sub_pd(iy0,jy0);
1848 dz00 = _mm256_sub_pd(iz0,jz0);
1849 dx01 = _mm256_sub_pd(ix0,jx1);
1850 dy01 = _mm256_sub_pd(iy0,jy1);
1851 dz01 = _mm256_sub_pd(iz0,jz1);
1852 dx02 = _mm256_sub_pd(ix0,jx2);
1853 dy02 = _mm256_sub_pd(iy0,jy2);
1854 dz02 = _mm256_sub_pd(iz0,jz2);
1855 dx10 = _mm256_sub_pd(ix1,jx0);
1856 dy10 = _mm256_sub_pd(iy1,jy0);
1857 dz10 = _mm256_sub_pd(iz1,jz0);
1858 dx11 = _mm256_sub_pd(ix1,jx1);
1859 dy11 = _mm256_sub_pd(iy1,jy1);
1860 dz11 = _mm256_sub_pd(iz1,jz1);
1861 dx12 = _mm256_sub_pd(ix1,jx2);
1862 dy12 = _mm256_sub_pd(iy1,jy2);
1863 dz12 = _mm256_sub_pd(iz1,jz2);
1864 dx20 = _mm256_sub_pd(ix2,jx0);
1865 dy20 = _mm256_sub_pd(iy2,jy0);
1866 dz20 = _mm256_sub_pd(iz2,jz0);
1867 dx21 = _mm256_sub_pd(ix2,jx1);
1868 dy21 = _mm256_sub_pd(iy2,jy1);
1869 dz21 = _mm256_sub_pd(iz2,jz1);
1870 dx22 = _mm256_sub_pd(ix2,jx2);
1871 dy22 = _mm256_sub_pd(iy2,jy2);
1872 dz22 = _mm256_sub_pd(iz2,jz2);
1874 /* Calculate squared distance and things based on it */
1875 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
1876 rsq01 = gmx_mm256_calc_rsq_pd(dx01,dy01,dz01);
1877 rsq02 = gmx_mm256_calc_rsq_pd(dx02,dy02,dz02);
1878 rsq10 = gmx_mm256_calc_rsq_pd(dx10,dy10,dz10);
1879 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
1880 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
1881 rsq20 = gmx_mm256_calc_rsq_pd(dx20,dy20,dz20);
1882 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
1883 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
1885 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
1886 rinv01 = gmx_mm256_invsqrt_pd(rsq01);
1887 rinv02 = gmx_mm256_invsqrt_pd(rsq02);
1888 rinv10 = gmx_mm256_invsqrt_pd(rsq10);
1889 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
1890 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
1891 rinv20 = gmx_mm256_invsqrt_pd(rsq20);
1892 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
1893 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
1895 fjx0 = _mm256_setzero_pd();
1896 fjy0 = _mm256_setzero_pd();
1897 fjz0 = _mm256_setzero_pd();
1898 fjx1 = _mm256_setzero_pd();
1899 fjy1 = _mm256_setzero_pd();
1900 fjz1 = _mm256_setzero_pd();
1901 fjx2 = _mm256_setzero_pd();
1902 fjy2 = _mm256_setzero_pd();
1903 fjz2 = _mm256_setzero_pd();
1905 /**************************
1906 * CALCULATE INTERACTIONS *
1907 **************************/
1909 r00 = _mm256_mul_pd(rsq00,rinv00);
1910 r00 = _mm256_andnot_pd(dummy_mask,r00);
1912 /* Calculate table index by multiplying r with table scale and truncate to integer */
1913 rt = _mm256_mul_pd(r00,vftabscale);
1914 vfitab = _mm256_cvttpd_epi32(rt);
1915 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1916 vfitab = _mm_slli_epi32(vfitab,2);
1918 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1919 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1920 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1921 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1922 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1923 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1924 Heps = _mm256_mul_pd(vfeps,H);
1925 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1926 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1927 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq00,FF),_mm256_mul_pd(vftabscale,rinv00)));
1931 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1933 /* Calculate temporary vectorial force */
1934 tx = _mm256_mul_pd(fscal,dx00);
1935 ty = _mm256_mul_pd(fscal,dy00);
1936 tz = _mm256_mul_pd(fscal,dz00);
1938 /* Update vectorial force */
1939 fix0 = _mm256_add_pd(fix0,tx);
1940 fiy0 = _mm256_add_pd(fiy0,ty);
1941 fiz0 = _mm256_add_pd(fiz0,tz);
1943 fjx0 = _mm256_add_pd(fjx0,tx);
1944 fjy0 = _mm256_add_pd(fjy0,ty);
1945 fjz0 = _mm256_add_pd(fjz0,tz);
1947 /**************************
1948 * CALCULATE INTERACTIONS *
1949 **************************/
1951 r01 = _mm256_mul_pd(rsq01,rinv01);
1952 r01 = _mm256_andnot_pd(dummy_mask,r01);
1954 /* Calculate table index by multiplying r with table scale and truncate to integer */
1955 rt = _mm256_mul_pd(r01,vftabscale);
1956 vfitab = _mm256_cvttpd_epi32(rt);
1957 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1958 vfitab = _mm_slli_epi32(vfitab,2);
1960 /* CUBIC SPLINE TABLE ELECTROSTATICS */
1961 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1962 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1963 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1964 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1965 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1966 Heps = _mm256_mul_pd(vfeps,H);
1967 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1968 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1969 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq01,FF),_mm256_mul_pd(vftabscale,rinv01)));
1973 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1975 /* Calculate temporary vectorial force */
1976 tx = _mm256_mul_pd(fscal,dx01);
1977 ty = _mm256_mul_pd(fscal,dy01);
1978 tz = _mm256_mul_pd(fscal,dz01);
1980 /* Update vectorial force */
1981 fix0 = _mm256_add_pd(fix0,tx);
1982 fiy0 = _mm256_add_pd(fiy0,ty);
1983 fiz0 = _mm256_add_pd(fiz0,tz);
1985 fjx1 = _mm256_add_pd(fjx1,tx);
1986 fjy1 = _mm256_add_pd(fjy1,ty);
1987 fjz1 = _mm256_add_pd(fjz1,tz);
1989 /**************************
1990 * CALCULATE INTERACTIONS *
1991 **************************/
1993 r02 = _mm256_mul_pd(rsq02,rinv02);
1994 r02 = _mm256_andnot_pd(dummy_mask,r02);
1996 /* Calculate table index by multiplying r with table scale and truncate to integer */
1997 rt = _mm256_mul_pd(r02,vftabscale);
1998 vfitab = _mm256_cvttpd_epi32(rt);
1999 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2000 vfitab = _mm_slli_epi32(vfitab,2);
2002 /* CUBIC SPLINE TABLE ELECTROSTATICS */
2003 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2004 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2005 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2006 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2007 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2008 Heps = _mm256_mul_pd(vfeps,H);
2009 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2010 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2011 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq02,FF),_mm256_mul_pd(vftabscale,rinv02)));
2015 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2017 /* Calculate temporary vectorial force */
2018 tx = _mm256_mul_pd(fscal,dx02);
2019 ty = _mm256_mul_pd(fscal,dy02);
2020 tz = _mm256_mul_pd(fscal,dz02);
2022 /* Update vectorial force */
2023 fix0 = _mm256_add_pd(fix0,tx);
2024 fiy0 = _mm256_add_pd(fiy0,ty);
2025 fiz0 = _mm256_add_pd(fiz0,tz);
2027 fjx2 = _mm256_add_pd(fjx2,tx);
2028 fjy2 = _mm256_add_pd(fjy2,ty);
2029 fjz2 = _mm256_add_pd(fjz2,tz);
2031 /**************************
2032 * CALCULATE INTERACTIONS *
2033 **************************/
2035 r10 = _mm256_mul_pd(rsq10,rinv10);
2036 r10 = _mm256_andnot_pd(dummy_mask,r10);
2038 /* Calculate table index by multiplying r with table scale and truncate to integer */
2039 rt = _mm256_mul_pd(r10,vftabscale);
2040 vfitab = _mm256_cvttpd_epi32(rt);
2041 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2042 vfitab = _mm_slli_epi32(vfitab,2);
2044 /* CUBIC SPLINE TABLE ELECTROSTATICS */
2045 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2046 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2047 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2048 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2049 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2050 Heps = _mm256_mul_pd(vfeps,H);
2051 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2052 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2053 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq10,FF),_mm256_mul_pd(vftabscale,rinv10)));
2057 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2059 /* Calculate temporary vectorial force */
2060 tx = _mm256_mul_pd(fscal,dx10);
2061 ty = _mm256_mul_pd(fscal,dy10);
2062 tz = _mm256_mul_pd(fscal,dz10);
2064 /* Update vectorial force */
2065 fix1 = _mm256_add_pd(fix1,tx);
2066 fiy1 = _mm256_add_pd(fiy1,ty);
2067 fiz1 = _mm256_add_pd(fiz1,tz);
2069 fjx0 = _mm256_add_pd(fjx0,tx);
2070 fjy0 = _mm256_add_pd(fjy0,ty);
2071 fjz0 = _mm256_add_pd(fjz0,tz);
2073 /**************************
2074 * CALCULATE INTERACTIONS *
2075 **************************/
2077 r11 = _mm256_mul_pd(rsq11,rinv11);
2078 r11 = _mm256_andnot_pd(dummy_mask,r11);
2080 /* Calculate table index by multiplying r with table scale and truncate to integer */
2081 rt = _mm256_mul_pd(r11,vftabscale);
2082 vfitab = _mm256_cvttpd_epi32(rt);
2083 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2084 vfitab = _mm_slli_epi32(vfitab,2);
2086 /* CUBIC SPLINE TABLE ELECTROSTATICS */
2087 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2088 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2089 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2090 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2091 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2092 Heps = _mm256_mul_pd(vfeps,H);
2093 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2094 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2095 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq11,FF),_mm256_mul_pd(vftabscale,rinv11)));
2099 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2101 /* Calculate temporary vectorial force */
2102 tx = _mm256_mul_pd(fscal,dx11);
2103 ty = _mm256_mul_pd(fscal,dy11);
2104 tz = _mm256_mul_pd(fscal,dz11);
2106 /* Update vectorial force */
2107 fix1 = _mm256_add_pd(fix1,tx);
2108 fiy1 = _mm256_add_pd(fiy1,ty);
2109 fiz1 = _mm256_add_pd(fiz1,tz);
2111 fjx1 = _mm256_add_pd(fjx1,tx);
2112 fjy1 = _mm256_add_pd(fjy1,ty);
2113 fjz1 = _mm256_add_pd(fjz1,tz);
2115 /**************************
2116 * CALCULATE INTERACTIONS *
2117 **************************/
2119 r12 = _mm256_mul_pd(rsq12,rinv12);
2120 r12 = _mm256_andnot_pd(dummy_mask,r12);
2122 /* Calculate table index by multiplying r with table scale and truncate to integer */
2123 rt = _mm256_mul_pd(r12,vftabscale);
2124 vfitab = _mm256_cvttpd_epi32(rt);
2125 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2126 vfitab = _mm_slli_epi32(vfitab,2);
2128 /* CUBIC SPLINE TABLE ELECTROSTATICS */
2129 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2130 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2131 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2132 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2133 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2134 Heps = _mm256_mul_pd(vfeps,H);
2135 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2136 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2137 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq12,FF),_mm256_mul_pd(vftabscale,rinv12)));
2141 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2143 /* Calculate temporary vectorial force */
2144 tx = _mm256_mul_pd(fscal,dx12);
2145 ty = _mm256_mul_pd(fscal,dy12);
2146 tz = _mm256_mul_pd(fscal,dz12);
2148 /* Update vectorial force */
2149 fix1 = _mm256_add_pd(fix1,tx);
2150 fiy1 = _mm256_add_pd(fiy1,ty);
2151 fiz1 = _mm256_add_pd(fiz1,tz);
2153 fjx2 = _mm256_add_pd(fjx2,tx);
2154 fjy2 = _mm256_add_pd(fjy2,ty);
2155 fjz2 = _mm256_add_pd(fjz2,tz);
2157 /**************************
2158 * CALCULATE INTERACTIONS *
2159 **************************/
2161 r20 = _mm256_mul_pd(rsq20,rinv20);
2162 r20 = _mm256_andnot_pd(dummy_mask,r20);
2164 /* Calculate table index by multiplying r with table scale and truncate to integer */
2165 rt = _mm256_mul_pd(r20,vftabscale);
2166 vfitab = _mm256_cvttpd_epi32(rt);
2167 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2168 vfitab = _mm_slli_epi32(vfitab,2);
2170 /* CUBIC SPLINE TABLE ELECTROSTATICS */
2171 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2172 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2173 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2174 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2175 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2176 Heps = _mm256_mul_pd(vfeps,H);
2177 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2178 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2179 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq20,FF),_mm256_mul_pd(vftabscale,rinv20)));
2183 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2185 /* Calculate temporary vectorial force */
2186 tx = _mm256_mul_pd(fscal,dx20);
2187 ty = _mm256_mul_pd(fscal,dy20);
2188 tz = _mm256_mul_pd(fscal,dz20);
2190 /* Update vectorial force */
2191 fix2 = _mm256_add_pd(fix2,tx);
2192 fiy2 = _mm256_add_pd(fiy2,ty);
2193 fiz2 = _mm256_add_pd(fiz2,tz);
2195 fjx0 = _mm256_add_pd(fjx0,tx);
2196 fjy0 = _mm256_add_pd(fjy0,ty);
2197 fjz0 = _mm256_add_pd(fjz0,tz);
2199 /**************************
2200 * CALCULATE INTERACTIONS *
2201 **************************/
2203 r21 = _mm256_mul_pd(rsq21,rinv21);
2204 r21 = _mm256_andnot_pd(dummy_mask,r21);
2206 /* Calculate table index by multiplying r with table scale and truncate to integer */
2207 rt = _mm256_mul_pd(r21,vftabscale);
2208 vfitab = _mm256_cvttpd_epi32(rt);
2209 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2210 vfitab = _mm_slli_epi32(vfitab,2);
2212 /* CUBIC SPLINE TABLE ELECTROSTATICS */
2213 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2214 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2215 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2216 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2217 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2218 Heps = _mm256_mul_pd(vfeps,H);
2219 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2220 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2221 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq21,FF),_mm256_mul_pd(vftabscale,rinv21)));
2225 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2227 /* Calculate temporary vectorial force */
2228 tx = _mm256_mul_pd(fscal,dx21);
2229 ty = _mm256_mul_pd(fscal,dy21);
2230 tz = _mm256_mul_pd(fscal,dz21);
2232 /* Update vectorial force */
2233 fix2 = _mm256_add_pd(fix2,tx);
2234 fiy2 = _mm256_add_pd(fiy2,ty);
2235 fiz2 = _mm256_add_pd(fiz2,tz);
2237 fjx1 = _mm256_add_pd(fjx1,tx);
2238 fjy1 = _mm256_add_pd(fjy1,ty);
2239 fjz1 = _mm256_add_pd(fjz1,tz);
2241 /**************************
2242 * CALCULATE INTERACTIONS *
2243 **************************/
2245 r22 = _mm256_mul_pd(rsq22,rinv22);
2246 r22 = _mm256_andnot_pd(dummy_mask,r22);
2248 /* Calculate table index by multiplying r with table scale and truncate to integer */
2249 rt = _mm256_mul_pd(r22,vftabscale);
2250 vfitab = _mm256_cvttpd_epi32(rt);
2251 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2252 vfitab = _mm_slli_epi32(vfitab,2);
2254 /* CUBIC SPLINE TABLE ELECTROSTATICS */
2255 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2256 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2257 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2258 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2259 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2260 Heps = _mm256_mul_pd(vfeps,H);
2261 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2262 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2263 felec = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_mul_pd(qq22,FF),_mm256_mul_pd(vftabscale,rinv22)));
2267 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2269 /* Calculate temporary vectorial force */
2270 tx = _mm256_mul_pd(fscal,dx22);
2271 ty = _mm256_mul_pd(fscal,dy22);
2272 tz = _mm256_mul_pd(fscal,dz22);
2274 /* Update vectorial force */
2275 fix2 = _mm256_add_pd(fix2,tx);
2276 fiy2 = _mm256_add_pd(fiy2,ty);
2277 fiz2 = _mm256_add_pd(fiz2,tz);
2279 fjx2 = _mm256_add_pd(fjx2,tx);
2280 fjy2 = _mm256_add_pd(fjy2,ty);
2281 fjz2 = _mm256_add_pd(fjz2,tz);
2283 fjptrA = (jnrlistA>=0) ? f+j_coord_offsetA : scratch;
2284 fjptrB = (jnrlistB>=0) ? f+j_coord_offsetB : scratch;
2285 fjptrC = (jnrlistC>=0) ? f+j_coord_offsetC : scratch;
2286 fjptrD = (jnrlistD>=0) ? f+j_coord_offsetD : scratch;
2288 gmx_mm256_decrement_3rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
2289 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,fjx2,fjy2,fjz2);
2291 /* Inner loop uses 360 flops */
2294 /* End of innermost loop */
2296 gmx_mm256_update_iforce_3atom_swizzle_pd(fix0,fiy0,fiz0,fix1,fiy1,fiz1,fix2,fiy2,fiz2,
2297 f+i_coord_offset,fshift+i_shift_offset);
2299 /* Increment number of inner iterations */
2300 inneriter += j_index_end - j_index_start;
2302 /* Outer loop uses 18 flops */
2305 /* Increment number of outer iterations */
2308 /* Update outer/inner flops */
2310 inc_nrnb(nrnb,eNR_NBKERNEL_ELEC_W3W3_F,outeriter*18 + inneriter*360);