2 * This file is part of the GROMACS molecular simulation package.
4 * Copyright (c) 2012,2013,2014, by the GROMACS development team, led by
5 * Mark Abraham, David van der Spoel, Berk Hess, and Erik Lindahl,
6 * and including many others, as listed in the AUTHORS file in the
7 * top-level source directory and at http://www.gromacs.org.
9 * GROMACS is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU Lesser General Public License
11 * as published by the Free Software Foundation; either version 2.1
12 * of the License, or (at your option) any later version.
14 * GROMACS is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17 * Lesser General Public License for more details.
19 * You should have received a copy of the GNU Lesser General Public
20 * License along with GROMACS; if not, see
21 * http://www.gnu.org/licenses, or write to the Free Software Foundation,
22 * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
24 * If you want to redistribute modifications to GROMACS, please
25 * consider that scientific software is very special. Version
26 * control is crucial - bugs must be traceable. We will be happy to
27 * consider code for inclusion in the official distribution, but
28 * derived work must not be called official GROMACS. Details are found
29 * in the README & COPYING files - if they are missing, get the
30 * official version at http://www.gromacs.org.
32 * To help us fund GROMACS development, we humbly ask that you cite
33 * the research papers on the package. Check out http://www.gromacs.org.
36 * Note: this file was generated by the GROMACS avx_256_double kernel generator.
42 #include "../nb_kernel.h"
43 #include "types/simple.h"
44 #include "gromacs/math/vec.h"
47 #include "gromacs/simd/math_x86_avx_256_double.h"
48 #include "kernelutil_x86_avx_256_double.h"
51 * Gromacs nonbonded kernel: nb_kernel_ElecRFCut_VdwCSTab_GeomW4W4_VF_avx_256_double
52 * Electrostatics interaction: ReactionField
53 * VdW interaction: CubicSplineTable
54 * Geometry: Water4-Water4
55 * Calculate force/pot: PotentialAndForce
58 nb_kernel_ElecRFCut_VdwCSTab_GeomW4W4_VF_avx_256_double
59 (t_nblist * gmx_restrict nlist,
60 rvec * gmx_restrict xx,
61 rvec * gmx_restrict ff,
62 t_forcerec * gmx_restrict fr,
63 t_mdatoms * gmx_restrict mdatoms,
64 nb_kernel_data_t gmx_unused * gmx_restrict kernel_data,
65 t_nrnb * gmx_restrict nrnb)
67 /* Suffixes 0,1,2,3 refer to particle indices for waters in the inner or outer loop, or
68 * just 0 for non-waters.
69 * Suffixes A,B,C,D refer to j loop unrolling done with AVX, e.g. for the four different
70 * jnr indices corresponding to data put in the four positions in the SIMD register.
72 int i_shift_offset,i_coord_offset,outeriter,inneriter;
73 int j_index_start,j_index_end,jidx,nri,inr,ggid,iidx;
74 int jnrA,jnrB,jnrC,jnrD;
75 int jnrlistA,jnrlistB,jnrlistC,jnrlistD;
76 int jnrlistE,jnrlistF,jnrlistG,jnrlistH;
77 int j_coord_offsetA,j_coord_offsetB,j_coord_offsetC,j_coord_offsetD;
78 int *iinr,*jindex,*jjnr,*shiftidx,*gid;
80 real *shiftvec,*fshift,*x,*f;
81 real *fjptrA,*fjptrB,*fjptrC,*fjptrD;
83 __m256d tx,ty,tz,fscal,rcutoff,rcutoff2,jidxall;
84 real * vdwioffsetptr0;
85 __m256d ix0,iy0,iz0,fix0,fiy0,fiz0,iq0,isai0;
86 real * vdwioffsetptr1;
87 __m256d ix1,iy1,iz1,fix1,fiy1,fiz1,iq1,isai1;
88 real * vdwioffsetptr2;
89 __m256d ix2,iy2,iz2,fix2,fiy2,fiz2,iq2,isai2;
90 real * vdwioffsetptr3;
91 __m256d ix3,iy3,iz3,fix3,fiy3,fiz3,iq3,isai3;
92 int vdwjidx0A,vdwjidx0B,vdwjidx0C,vdwjidx0D;
93 __m256d jx0,jy0,jz0,fjx0,fjy0,fjz0,jq0,isaj0;
94 int vdwjidx1A,vdwjidx1B,vdwjidx1C,vdwjidx1D;
95 __m256d jx1,jy1,jz1,fjx1,fjy1,fjz1,jq1,isaj1;
96 int vdwjidx2A,vdwjidx2B,vdwjidx2C,vdwjidx2D;
97 __m256d jx2,jy2,jz2,fjx2,fjy2,fjz2,jq2,isaj2;
98 int vdwjidx3A,vdwjidx3B,vdwjidx3C,vdwjidx3D;
99 __m256d jx3,jy3,jz3,fjx3,fjy3,fjz3,jq3,isaj3;
100 __m256d dx00,dy00,dz00,rsq00,rinv00,rinvsq00,r00,qq00,c6_00,c12_00;
101 __m256d dx11,dy11,dz11,rsq11,rinv11,rinvsq11,r11,qq11,c6_11,c12_11;
102 __m256d dx12,dy12,dz12,rsq12,rinv12,rinvsq12,r12,qq12,c6_12,c12_12;
103 __m256d dx13,dy13,dz13,rsq13,rinv13,rinvsq13,r13,qq13,c6_13,c12_13;
104 __m256d dx21,dy21,dz21,rsq21,rinv21,rinvsq21,r21,qq21,c6_21,c12_21;
105 __m256d dx22,dy22,dz22,rsq22,rinv22,rinvsq22,r22,qq22,c6_22,c12_22;
106 __m256d dx23,dy23,dz23,rsq23,rinv23,rinvsq23,r23,qq23,c6_23,c12_23;
107 __m256d dx31,dy31,dz31,rsq31,rinv31,rinvsq31,r31,qq31,c6_31,c12_31;
108 __m256d dx32,dy32,dz32,rsq32,rinv32,rinvsq32,r32,qq32,c6_32,c12_32;
109 __m256d dx33,dy33,dz33,rsq33,rinv33,rinvsq33,r33,qq33,c6_33,c12_33;
110 __m256d velec,felec,velecsum,facel,crf,krf,krf2;
113 __m256d rinvsix,rvdw,vvdw,vvdw6,vvdw12,fvdw,fvdw6,fvdw12,vvdwsum,sh_vdw_invrcut6;
116 __m256d one_sixth = _mm256_set1_pd(1.0/6.0);
117 __m256d one_twelfth = _mm256_set1_pd(1.0/12.0);
119 __m128i ifour = _mm_set1_epi32(4);
120 __m256d rt,vfeps,vftabscale,Y,F,G,H,Heps,Fp,VV,FF;
122 __m256d dummy_mask,cutoff_mask;
123 __m128 tmpmask0,tmpmask1;
124 __m256d signbit = _mm256_castsi256_pd( _mm256_set1_epi32(0x80000000) );
125 __m256d one = _mm256_set1_pd(1.0);
126 __m256d two = _mm256_set1_pd(2.0);
132 jindex = nlist->jindex;
134 shiftidx = nlist->shift;
136 shiftvec = fr->shift_vec[0];
137 fshift = fr->fshift[0];
138 facel = _mm256_set1_pd(fr->epsfac);
139 charge = mdatoms->chargeA;
140 krf = _mm256_set1_pd(fr->ic->k_rf);
141 krf2 = _mm256_set1_pd(fr->ic->k_rf*2.0);
142 crf = _mm256_set1_pd(fr->ic->c_rf);
143 nvdwtype = fr->ntype;
145 vdwtype = mdatoms->typeA;
147 vftab = kernel_data->table_vdw->data;
148 vftabscale = _mm256_set1_pd(kernel_data->table_vdw->scale);
150 /* Setup water-specific parameters */
151 inr = nlist->iinr[0];
152 iq1 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+1]));
153 iq2 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+2]));
154 iq3 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+3]));
155 vdwioffsetptr0 = vdwparam+2*nvdwtype*vdwtype[inr+0];
157 jq1 = _mm256_set1_pd(charge[inr+1]);
158 jq2 = _mm256_set1_pd(charge[inr+2]);
159 jq3 = _mm256_set1_pd(charge[inr+3]);
160 vdwjidx0A = 2*vdwtype[inr+0];
161 c6_00 = _mm256_set1_pd(vdwioffsetptr0[vdwjidx0A]);
162 c12_00 = _mm256_set1_pd(vdwioffsetptr0[vdwjidx0A+1]);
163 qq11 = _mm256_mul_pd(iq1,jq1);
164 qq12 = _mm256_mul_pd(iq1,jq2);
165 qq13 = _mm256_mul_pd(iq1,jq3);
166 qq21 = _mm256_mul_pd(iq2,jq1);
167 qq22 = _mm256_mul_pd(iq2,jq2);
168 qq23 = _mm256_mul_pd(iq2,jq3);
169 qq31 = _mm256_mul_pd(iq3,jq1);
170 qq32 = _mm256_mul_pd(iq3,jq2);
171 qq33 = _mm256_mul_pd(iq3,jq3);
173 /* When we use explicit cutoffs the value must be identical for elec and VdW, so use elec as an arbitrary choice */
174 rcutoff_scalar = fr->rcoulomb;
175 rcutoff = _mm256_set1_pd(rcutoff_scalar);
176 rcutoff2 = _mm256_mul_pd(rcutoff,rcutoff);
178 /* Avoid stupid compiler warnings */
179 jnrA = jnrB = jnrC = jnrD = 0;
188 for(iidx=0;iidx<4*DIM;iidx++)
193 /* Start outer loop over neighborlists */
194 for(iidx=0; iidx<nri; iidx++)
196 /* Load shift vector for this list */
197 i_shift_offset = DIM*shiftidx[iidx];
199 /* Load limits for loop over neighbors */
200 j_index_start = jindex[iidx];
201 j_index_end = jindex[iidx+1];
203 /* Get outer coordinate index */
205 i_coord_offset = DIM*inr;
207 /* Load i particle coords and add shift vector */
208 gmx_mm256_load_shift_and_4rvec_broadcast_pd(shiftvec+i_shift_offset,x+i_coord_offset,
209 &ix0,&iy0,&iz0,&ix1,&iy1,&iz1,&ix2,&iy2,&iz2,&ix3,&iy3,&iz3);
211 fix0 = _mm256_setzero_pd();
212 fiy0 = _mm256_setzero_pd();
213 fiz0 = _mm256_setzero_pd();
214 fix1 = _mm256_setzero_pd();
215 fiy1 = _mm256_setzero_pd();
216 fiz1 = _mm256_setzero_pd();
217 fix2 = _mm256_setzero_pd();
218 fiy2 = _mm256_setzero_pd();
219 fiz2 = _mm256_setzero_pd();
220 fix3 = _mm256_setzero_pd();
221 fiy3 = _mm256_setzero_pd();
222 fiz3 = _mm256_setzero_pd();
224 /* Reset potential sums */
225 velecsum = _mm256_setzero_pd();
226 vvdwsum = _mm256_setzero_pd();
228 /* Start inner kernel loop */
229 for(jidx=j_index_start; jidx<j_index_end && jjnr[jidx+3]>=0; jidx+=4)
232 /* Get j neighbor index, and coordinate index */
237 j_coord_offsetA = DIM*jnrA;
238 j_coord_offsetB = DIM*jnrB;
239 j_coord_offsetC = DIM*jnrC;
240 j_coord_offsetD = DIM*jnrD;
242 /* load j atom coordinates */
243 gmx_mm256_load_4rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
244 x+j_coord_offsetC,x+j_coord_offsetD,
245 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,
246 &jy2,&jz2,&jx3,&jy3,&jz3);
248 /* Calculate displacement vector */
249 dx00 = _mm256_sub_pd(ix0,jx0);
250 dy00 = _mm256_sub_pd(iy0,jy0);
251 dz00 = _mm256_sub_pd(iz0,jz0);
252 dx11 = _mm256_sub_pd(ix1,jx1);
253 dy11 = _mm256_sub_pd(iy1,jy1);
254 dz11 = _mm256_sub_pd(iz1,jz1);
255 dx12 = _mm256_sub_pd(ix1,jx2);
256 dy12 = _mm256_sub_pd(iy1,jy2);
257 dz12 = _mm256_sub_pd(iz1,jz2);
258 dx13 = _mm256_sub_pd(ix1,jx3);
259 dy13 = _mm256_sub_pd(iy1,jy3);
260 dz13 = _mm256_sub_pd(iz1,jz3);
261 dx21 = _mm256_sub_pd(ix2,jx1);
262 dy21 = _mm256_sub_pd(iy2,jy1);
263 dz21 = _mm256_sub_pd(iz2,jz1);
264 dx22 = _mm256_sub_pd(ix2,jx2);
265 dy22 = _mm256_sub_pd(iy2,jy2);
266 dz22 = _mm256_sub_pd(iz2,jz2);
267 dx23 = _mm256_sub_pd(ix2,jx3);
268 dy23 = _mm256_sub_pd(iy2,jy3);
269 dz23 = _mm256_sub_pd(iz2,jz3);
270 dx31 = _mm256_sub_pd(ix3,jx1);
271 dy31 = _mm256_sub_pd(iy3,jy1);
272 dz31 = _mm256_sub_pd(iz3,jz1);
273 dx32 = _mm256_sub_pd(ix3,jx2);
274 dy32 = _mm256_sub_pd(iy3,jy2);
275 dz32 = _mm256_sub_pd(iz3,jz2);
276 dx33 = _mm256_sub_pd(ix3,jx3);
277 dy33 = _mm256_sub_pd(iy3,jy3);
278 dz33 = _mm256_sub_pd(iz3,jz3);
280 /* Calculate squared distance and things based on it */
281 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
282 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
283 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
284 rsq13 = gmx_mm256_calc_rsq_pd(dx13,dy13,dz13);
285 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
286 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
287 rsq23 = gmx_mm256_calc_rsq_pd(dx23,dy23,dz23);
288 rsq31 = gmx_mm256_calc_rsq_pd(dx31,dy31,dz31);
289 rsq32 = gmx_mm256_calc_rsq_pd(dx32,dy32,dz32);
290 rsq33 = gmx_mm256_calc_rsq_pd(dx33,dy33,dz33);
292 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
293 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
294 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
295 rinv13 = gmx_mm256_invsqrt_pd(rsq13);
296 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
297 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
298 rinv23 = gmx_mm256_invsqrt_pd(rsq23);
299 rinv31 = gmx_mm256_invsqrt_pd(rsq31);
300 rinv32 = gmx_mm256_invsqrt_pd(rsq32);
301 rinv33 = gmx_mm256_invsqrt_pd(rsq33);
303 rinvsq11 = _mm256_mul_pd(rinv11,rinv11);
304 rinvsq12 = _mm256_mul_pd(rinv12,rinv12);
305 rinvsq13 = _mm256_mul_pd(rinv13,rinv13);
306 rinvsq21 = _mm256_mul_pd(rinv21,rinv21);
307 rinvsq22 = _mm256_mul_pd(rinv22,rinv22);
308 rinvsq23 = _mm256_mul_pd(rinv23,rinv23);
309 rinvsq31 = _mm256_mul_pd(rinv31,rinv31);
310 rinvsq32 = _mm256_mul_pd(rinv32,rinv32);
311 rinvsq33 = _mm256_mul_pd(rinv33,rinv33);
313 fjx0 = _mm256_setzero_pd();
314 fjy0 = _mm256_setzero_pd();
315 fjz0 = _mm256_setzero_pd();
316 fjx1 = _mm256_setzero_pd();
317 fjy1 = _mm256_setzero_pd();
318 fjz1 = _mm256_setzero_pd();
319 fjx2 = _mm256_setzero_pd();
320 fjy2 = _mm256_setzero_pd();
321 fjz2 = _mm256_setzero_pd();
322 fjx3 = _mm256_setzero_pd();
323 fjy3 = _mm256_setzero_pd();
324 fjz3 = _mm256_setzero_pd();
326 /**************************
327 * CALCULATE INTERACTIONS *
328 **************************/
330 if (gmx_mm256_any_lt(rsq00,rcutoff2))
333 r00 = _mm256_mul_pd(rsq00,rinv00);
335 /* Calculate table index by multiplying r with table scale and truncate to integer */
336 rt = _mm256_mul_pd(r00,vftabscale);
337 vfitab = _mm256_cvttpd_epi32(rt);
338 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
339 vfitab = _mm_slli_epi32(vfitab,3);
341 /* CUBIC SPLINE TABLE DISPERSION */
342 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
343 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
344 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
345 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
346 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
347 Heps = _mm256_mul_pd(vfeps,H);
348 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
349 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
350 vvdw6 = _mm256_mul_pd(c6_00,VV);
351 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
352 fvdw6 = _mm256_mul_pd(c6_00,FF);
354 /* CUBIC SPLINE TABLE REPULSION */
355 vfitab = _mm_add_epi32(vfitab,ifour);
356 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
357 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
358 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
359 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
360 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
361 Heps = _mm256_mul_pd(vfeps,H);
362 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
363 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
364 vvdw12 = _mm256_mul_pd(c12_00,VV);
365 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
366 fvdw12 = _mm256_mul_pd(c12_00,FF);
367 vvdw = _mm256_add_pd(vvdw12,vvdw6);
368 fvdw = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_add_pd(fvdw6,fvdw12),_mm256_mul_pd(vftabscale,rinv00)));
370 cutoff_mask = _mm256_cmp_pd(rsq00,rcutoff2,_CMP_LT_OQ);
372 /* Update potential sum for this i atom from the interaction with this j atom. */
373 vvdw = _mm256_and_pd(vvdw,cutoff_mask);
374 vvdwsum = _mm256_add_pd(vvdwsum,vvdw);
378 fscal = _mm256_and_pd(fscal,cutoff_mask);
380 /* Calculate temporary vectorial force */
381 tx = _mm256_mul_pd(fscal,dx00);
382 ty = _mm256_mul_pd(fscal,dy00);
383 tz = _mm256_mul_pd(fscal,dz00);
385 /* Update vectorial force */
386 fix0 = _mm256_add_pd(fix0,tx);
387 fiy0 = _mm256_add_pd(fiy0,ty);
388 fiz0 = _mm256_add_pd(fiz0,tz);
390 fjx0 = _mm256_add_pd(fjx0,tx);
391 fjy0 = _mm256_add_pd(fjy0,ty);
392 fjz0 = _mm256_add_pd(fjz0,tz);
396 /**************************
397 * CALCULATE INTERACTIONS *
398 **************************/
400 if (gmx_mm256_any_lt(rsq11,rcutoff2))
403 /* REACTION-FIELD ELECTROSTATICS */
404 velec = _mm256_mul_pd(qq11,_mm256_sub_pd(_mm256_add_pd(rinv11,_mm256_mul_pd(krf,rsq11)),crf));
405 felec = _mm256_mul_pd(qq11,_mm256_sub_pd(_mm256_mul_pd(rinv11,rinvsq11),krf2));
407 cutoff_mask = _mm256_cmp_pd(rsq11,rcutoff2,_CMP_LT_OQ);
409 /* Update potential sum for this i atom from the interaction with this j atom. */
410 velec = _mm256_and_pd(velec,cutoff_mask);
411 velecsum = _mm256_add_pd(velecsum,velec);
415 fscal = _mm256_and_pd(fscal,cutoff_mask);
417 /* Calculate temporary vectorial force */
418 tx = _mm256_mul_pd(fscal,dx11);
419 ty = _mm256_mul_pd(fscal,dy11);
420 tz = _mm256_mul_pd(fscal,dz11);
422 /* Update vectorial force */
423 fix1 = _mm256_add_pd(fix1,tx);
424 fiy1 = _mm256_add_pd(fiy1,ty);
425 fiz1 = _mm256_add_pd(fiz1,tz);
427 fjx1 = _mm256_add_pd(fjx1,tx);
428 fjy1 = _mm256_add_pd(fjy1,ty);
429 fjz1 = _mm256_add_pd(fjz1,tz);
433 /**************************
434 * CALCULATE INTERACTIONS *
435 **************************/
437 if (gmx_mm256_any_lt(rsq12,rcutoff2))
440 /* REACTION-FIELD ELECTROSTATICS */
441 velec = _mm256_mul_pd(qq12,_mm256_sub_pd(_mm256_add_pd(rinv12,_mm256_mul_pd(krf,rsq12)),crf));
442 felec = _mm256_mul_pd(qq12,_mm256_sub_pd(_mm256_mul_pd(rinv12,rinvsq12),krf2));
444 cutoff_mask = _mm256_cmp_pd(rsq12,rcutoff2,_CMP_LT_OQ);
446 /* Update potential sum for this i atom from the interaction with this j atom. */
447 velec = _mm256_and_pd(velec,cutoff_mask);
448 velecsum = _mm256_add_pd(velecsum,velec);
452 fscal = _mm256_and_pd(fscal,cutoff_mask);
454 /* Calculate temporary vectorial force */
455 tx = _mm256_mul_pd(fscal,dx12);
456 ty = _mm256_mul_pd(fscal,dy12);
457 tz = _mm256_mul_pd(fscal,dz12);
459 /* Update vectorial force */
460 fix1 = _mm256_add_pd(fix1,tx);
461 fiy1 = _mm256_add_pd(fiy1,ty);
462 fiz1 = _mm256_add_pd(fiz1,tz);
464 fjx2 = _mm256_add_pd(fjx2,tx);
465 fjy2 = _mm256_add_pd(fjy2,ty);
466 fjz2 = _mm256_add_pd(fjz2,tz);
470 /**************************
471 * CALCULATE INTERACTIONS *
472 **************************/
474 if (gmx_mm256_any_lt(rsq13,rcutoff2))
477 /* REACTION-FIELD ELECTROSTATICS */
478 velec = _mm256_mul_pd(qq13,_mm256_sub_pd(_mm256_add_pd(rinv13,_mm256_mul_pd(krf,rsq13)),crf));
479 felec = _mm256_mul_pd(qq13,_mm256_sub_pd(_mm256_mul_pd(rinv13,rinvsq13),krf2));
481 cutoff_mask = _mm256_cmp_pd(rsq13,rcutoff2,_CMP_LT_OQ);
483 /* Update potential sum for this i atom from the interaction with this j atom. */
484 velec = _mm256_and_pd(velec,cutoff_mask);
485 velecsum = _mm256_add_pd(velecsum,velec);
489 fscal = _mm256_and_pd(fscal,cutoff_mask);
491 /* Calculate temporary vectorial force */
492 tx = _mm256_mul_pd(fscal,dx13);
493 ty = _mm256_mul_pd(fscal,dy13);
494 tz = _mm256_mul_pd(fscal,dz13);
496 /* Update vectorial force */
497 fix1 = _mm256_add_pd(fix1,tx);
498 fiy1 = _mm256_add_pd(fiy1,ty);
499 fiz1 = _mm256_add_pd(fiz1,tz);
501 fjx3 = _mm256_add_pd(fjx3,tx);
502 fjy3 = _mm256_add_pd(fjy3,ty);
503 fjz3 = _mm256_add_pd(fjz3,tz);
507 /**************************
508 * CALCULATE INTERACTIONS *
509 **************************/
511 if (gmx_mm256_any_lt(rsq21,rcutoff2))
514 /* REACTION-FIELD ELECTROSTATICS */
515 velec = _mm256_mul_pd(qq21,_mm256_sub_pd(_mm256_add_pd(rinv21,_mm256_mul_pd(krf,rsq21)),crf));
516 felec = _mm256_mul_pd(qq21,_mm256_sub_pd(_mm256_mul_pd(rinv21,rinvsq21),krf2));
518 cutoff_mask = _mm256_cmp_pd(rsq21,rcutoff2,_CMP_LT_OQ);
520 /* Update potential sum for this i atom from the interaction with this j atom. */
521 velec = _mm256_and_pd(velec,cutoff_mask);
522 velecsum = _mm256_add_pd(velecsum,velec);
526 fscal = _mm256_and_pd(fscal,cutoff_mask);
528 /* Calculate temporary vectorial force */
529 tx = _mm256_mul_pd(fscal,dx21);
530 ty = _mm256_mul_pd(fscal,dy21);
531 tz = _mm256_mul_pd(fscal,dz21);
533 /* Update vectorial force */
534 fix2 = _mm256_add_pd(fix2,tx);
535 fiy2 = _mm256_add_pd(fiy2,ty);
536 fiz2 = _mm256_add_pd(fiz2,tz);
538 fjx1 = _mm256_add_pd(fjx1,tx);
539 fjy1 = _mm256_add_pd(fjy1,ty);
540 fjz1 = _mm256_add_pd(fjz1,tz);
544 /**************************
545 * CALCULATE INTERACTIONS *
546 **************************/
548 if (gmx_mm256_any_lt(rsq22,rcutoff2))
551 /* REACTION-FIELD ELECTROSTATICS */
552 velec = _mm256_mul_pd(qq22,_mm256_sub_pd(_mm256_add_pd(rinv22,_mm256_mul_pd(krf,rsq22)),crf));
553 felec = _mm256_mul_pd(qq22,_mm256_sub_pd(_mm256_mul_pd(rinv22,rinvsq22),krf2));
555 cutoff_mask = _mm256_cmp_pd(rsq22,rcutoff2,_CMP_LT_OQ);
557 /* Update potential sum for this i atom from the interaction with this j atom. */
558 velec = _mm256_and_pd(velec,cutoff_mask);
559 velecsum = _mm256_add_pd(velecsum,velec);
563 fscal = _mm256_and_pd(fscal,cutoff_mask);
565 /* Calculate temporary vectorial force */
566 tx = _mm256_mul_pd(fscal,dx22);
567 ty = _mm256_mul_pd(fscal,dy22);
568 tz = _mm256_mul_pd(fscal,dz22);
570 /* Update vectorial force */
571 fix2 = _mm256_add_pd(fix2,tx);
572 fiy2 = _mm256_add_pd(fiy2,ty);
573 fiz2 = _mm256_add_pd(fiz2,tz);
575 fjx2 = _mm256_add_pd(fjx2,tx);
576 fjy2 = _mm256_add_pd(fjy2,ty);
577 fjz2 = _mm256_add_pd(fjz2,tz);
581 /**************************
582 * CALCULATE INTERACTIONS *
583 **************************/
585 if (gmx_mm256_any_lt(rsq23,rcutoff2))
588 /* REACTION-FIELD ELECTROSTATICS */
589 velec = _mm256_mul_pd(qq23,_mm256_sub_pd(_mm256_add_pd(rinv23,_mm256_mul_pd(krf,rsq23)),crf));
590 felec = _mm256_mul_pd(qq23,_mm256_sub_pd(_mm256_mul_pd(rinv23,rinvsq23),krf2));
592 cutoff_mask = _mm256_cmp_pd(rsq23,rcutoff2,_CMP_LT_OQ);
594 /* Update potential sum for this i atom from the interaction with this j atom. */
595 velec = _mm256_and_pd(velec,cutoff_mask);
596 velecsum = _mm256_add_pd(velecsum,velec);
600 fscal = _mm256_and_pd(fscal,cutoff_mask);
602 /* Calculate temporary vectorial force */
603 tx = _mm256_mul_pd(fscal,dx23);
604 ty = _mm256_mul_pd(fscal,dy23);
605 tz = _mm256_mul_pd(fscal,dz23);
607 /* Update vectorial force */
608 fix2 = _mm256_add_pd(fix2,tx);
609 fiy2 = _mm256_add_pd(fiy2,ty);
610 fiz2 = _mm256_add_pd(fiz2,tz);
612 fjx3 = _mm256_add_pd(fjx3,tx);
613 fjy3 = _mm256_add_pd(fjy3,ty);
614 fjz3 = _mm256_add_pd(fjz3,tz);
618 /**************************
619 * CALCULATE INTERACTIONS *
620 **************************/
622 if (gmx_mm256_any_lt(rsq31,rcutoff2))
625 /* REACTION-FIELD ELECTROSTATICS */
626 velec = _mm256_mul_pd(qq31,_mm256_sub_pd(_mm256_add_pd(rinv31,_mm256_mul_pd(krf,rsq31)),crf));
627 felec = _mm256_mul_pd(qq31,_mm256_sub_pd(_mm256_mul_pd(rinv31,rinvsq31),krf2));
629 cutoff_mask = _mm256_cmp_pd(rsq31,rcutoff2,_CMP_LT_OQ);
631 /* Update potential sum for this i atom from the interaction with this j atom. */
632 velec = _mm256_and_pd(velec,cutoff_mask);
633 velecsum = _mm256_add_pd(velecsum,velec);
637 fscal = _mm256_and_pd(fscal,cutoff_mask);
639 /* Calculate temporary vectorial force */
640 tx = _mm256_mul_pd(fscal,dx31);
641 ty = _mm256_mul_pd(fscal,dy31);
642 tz = _mm256_mul_pd(fscal,dz31);
644 /* Update vectorial force */
645 fix3 = _mm256_add_pd(fix3,tx);
646 fiy3 = _mm256_add_pd(fiy3,ty);
647 fiz3 = _mm256_add_pd(fiz3,tz);
649 fjx1 = _mm256_add_pd(fjx1,tx);
650 fjy1 = _mm256_add_pd(fjy1,ty);
651 fjz1 = _mm256_add_pd(fjz1,tz);
655 /**************************
656 * CALCULATE INTERACTIONS *
657 **************************/
659 if (gmx_mm256_any_lt(rsq32,rcutoff2))
662 /* REACTION-FIELD ELECTROSTATICS */
663 velec = _mm256_mul_pd(qq32,_mm256_sub_pd(_mm256_add_pd(rinv32,_mm256_mul_pd(krf,rsq32)),crf));
664 felec = _mm256_mul_pd(qq32,_mm256_sub_pd(_mm256_mul_pd(rinv32,rinvsq32),krf2));
666 cutoff_mask = _mm256_cmp_pd(rsq32,rcutoff2,_CMP_LT_OQ);
668 /* Update potential sum for this i atom from the interaction with this j atom. */
669 velec = _mm256_and_pd(velec,cutoff_mask);
670 velecsum = _mm256_add_pd(velecsum,velec);
674 fscal = _mm256_and_pd(fscal,cutoff_mask);
676 /* Calculate temporary vectorial force */
677 tx = _mm256_mul_pd(fscal,dx32);
678 ty = _mm256_mul_pd(fscal,dy32);
679 tz = _mm256_mul_pd(fscal,dz32);
681 /* Update vectorial force */
682 fix3 = _mm256_add_pd(fix3,tx);
683 fiy3 = _mm256_add_pd(fiy3,ty);
684 fiz3 = _mm256_add_pd(fiz3,tz);
686 fjx2 = _mm256_add_pd(fjx2,tx);
687 fjy2 = _mm256_add_pd(fjy2,ty);
688 fjz2 = _mm256_add_pd(fjz2,tz);
692 /**************************
693 * CALCULATE INTERACTIONS *
694 **************************/
696 if (gmx_mm256_any_lt(rsq33,rcutoff2))
699 /* REACTION-FIELD ELECTROSTATICS */
700 velec = _mm256_mul_pd(qq33,_mm256_sub_pd(_mm256_add_pd(rinv33,_mm256_mul_pd(krf,rsq33)),crf));
701 felec = _mm256_mul_pd(qq33,_mm256_sub_pd(_mm256_mul_pd(rinv33,rinvsq33),krf2));
703 cutoff_mask = _mm256_cmp_pd(rsq33,rcutoff2,_CMP_LT_OQ);
705 /* Update potential sum for this i atom from the interaction with this j atom. */
706 velec = _mm256_and_pd(velec,cutoff_mask);
707 velecsum = _mm256_add_pd(velecsum,velec);
711 fscal = _mm256_and_pd(fscal,cutoff_mask);
713 /* Calculate temporary vectorial force */
714 tx = _mm256_mul_pd(fscal,dx33);
715 ty = _mm256_mul_pd(fscal,dy33);
716 tz = _mm256_mul_pd(fscal,dz33);
718 /* Update vectorial force */
719 fix3 = _mm256_add_pd(fix3,tx);
720 fiy3 = _mm256_add_pd(fiy3,ty);
721 fiz3 = _mm256_add_pd(fiz3,tz);
723 fjx3 = _mm256_add_pd(fjx3,tx);
724 fjy3 = _mm256_add_pd(fjy3,ty);
725 fjz3 = _mm256_add_pd(fjz3,tz);
729 fjptrA = f+j_coord_offsetA;
730 fjptrB = f+j_coord_offsetB;
731 fjptrC = f+j_coord_offsetC;
732 fjptrD = f+j_coord_offsetD;
734 gmx_mm256_decrement_4rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
735 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,
736 fjx2,fjy2,fjz2,fjx3,fjy3,fjz3);
738 /* Inner loop uses 387 flops */
744 /* Get j neighbor index, and coordinate index */
745 jnrlistA = jjnr[jidx];
746 jnrlistB = jjnr[jidx+1];
747 jnrlistC = jjnr[jidx+2];
748 jnrlistD = jjnr[jidx+3];
749 /* Sign of each element will be negative for non-real atoms.
750 * This mask will be 0xFFFFFFFF for dummy entries and 0x0 for real ones,
751 * so use it as val = _mm_andnot_pd(mask,val) to clear dummy entries.
753 tmpmask0 = gmx_mm_castsi128_ps(_mm_cmplt_epi32(_mm_loadu_si128((const __m128i *)(jjnr+jidx)),_mm_setzero_si128()));
755 tmpmask1 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(3,3,2,2));
756 tmpmask0 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(1,1,0,0));
757 dummy_mask = _mm256_castps_pd(gmx_mm256_set_m128(tmpmask1,tmpmask0));
759 jnrA = (jnrlistA>=0) ? jnrlistA : 0;
760 jnrB = (jnrlistB>=0) ? jnrlistB : 0;
761 jnrC = (jnrlistC>=0) ? jnrlistC : 0;
762 jnrD = (jnrlistD>=0) ? jnrlistD : 0;
763 j_coord_offsetA = DIM*jnrA;
764 j_coord_offsetB = DIM*jnrB;
765 j_coord_offsetC = DIM*jnrC;
766 j_coord_offsetD = DIM*jnrD;
768 /* load j atom coordinates */
769 gmx_mm256_load_4rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
770 x+j_coord_offsetC,x+j_coord_offsetD,
771 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,
772 &jy2,&jz2,&jx3,&jy3,&jz3);
774 /* Calculate displacement vector */
775 dx00 = _mm256_sub_pd(ix0,jx0);
776 dy00 = _mm256_sub_pd(iy0,jy0);
777 dz00 = _mm256_sub_pd(iz0,jz0);
778 dx11 = _mm256_sub_pd(ix1,jx1);
779 dy11 = _mm256_sub_pd(iy1,jy1);
780 dz11 = _mm256_sub_pd(iz1,jz1);
781 dx12 = _mm256_sub_pd(ix1,jx2);
782 dy12 = _mm256_sub_pd(iy1,jy2);
783 dz12 = _mm256_sub_pd(iz1,jz2);
784 dx13 = _mm256_sub_pd(ix1,jx3);
785 dy13 = _mm256_sub_pd(iy1,jy3);
786 dz13 = _mm256_sub_pd(iz1,jz3);
787 dx21 = _mm256_sub_pd(ix2,jx1);
788 dy21 = _mm256_sub_pd(iy2,jy1);
789 dz21 = _mm256_sub_pd(iz2,jz1);
790 dx22 = _mm256_sub_pd(ix2,jx2);
791 dy22 = _mm256_sub_pd(iy2,jy2);
792 dz22 = _mm256_sub_pd(iz2,jz2);
793 dx23 = _mm256_sub_pd(ix2,jx3);
794 dy23 = _mm256_sub_pd(iy2,jy3);
795 dz23 = _mm256_sub_pd(iz2,jz3);
796 dx31 = _mm256_sub_pd(ix3,jx1);
797 dy31 = _mm256_sub_pd(iy3,jy1);
798 dz31 = _mm256_sub_pd(iz3,jz1);
799 dx32 = _mm256_sub_pd(ix3,jx2);
800 dy32 = _mm256_sub_pd(iy3,jy2);
801 dz32 = _mm256_sub_pd(iz3,jz2);
802 dx33 = _mm256_sub_pd(ix3,jx3);
803 dy33 = _mm256_sub_pd(iy3,jy3);
804 dz33 = _mm256_sub_pd(iz3,jz3);
806 /* Calculate squared distance and things based on it */
807 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
808 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
809 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
810 rsq13 = gmx_mm256_calc_rsq_pd(dx13,dy13,dz13);
811 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
812 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
813 rsq23 = gmx_mm256_calc_rsq_pd(dx23,dy23,dz23);
814 rsq31 = gmx_mm256_calc_rsq_pd(dx31,dy31,dz31);
815 rsq32 = gmx_mm256_calc_rsq_pd(dx32,dy32,dz32);
816 rsq33 = gmx_mm256_calc_rsq_pd(dx33,dy33,dz33);
818 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
819 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
820 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
821 rinv13 = gmx_mm256_invsqrt_pd(rsq13);
822 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
823 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
824 rinv23 = gmx_mm256_invsqrt_pd(rsq23);
825 rinv31 = gmx_mm256_invsqrt_pd(rsq31);
826 rinv32 = gmx_mm256_invsqrt_pd(rsq32);
827 rinv33 = gmx_mm256_invsqrt_pd(rsq33);
829 rinvsq11 = _mm256_mul_pd(rinv11,rinv11);
830 rinvsq12 = _mm256_mul_pd(rinv12,rinv12);
831 rinvsq13 = _mm256_mul_pd(rinv13,rinv13);
832 rinvsq21 = _mm256_mul_pd(rinv21,rinv21);
833 rinvsq22 = _mm256_mul_pd(rinv22,rinv22);
834 rinvsq23 = _mm256_mul_pd(rinv23,rinv23);
835 rinvsq31 = _mm256_mul_pd(rinv31,rinv31);
836 rinvsq32 = _mm256_mul_pd(rinv32,rinv32);
837 rinvsq33 = _mm256_mul_pd(rinv33,rinv33);
839 fjx0 = _mm256_setzero_pd();
840 fjy0 = _mm256_setzero_pd();
841 fjz0 = _mm256_setzero_pd();
842 fjx1 = _mm256_setzero_pd();
843 fjy1 = _mm256_setzero_pd();
844 fjz1 = _mm256_setzero_pd();
845 fjx2 = _mm256_setzero_pd();
846 fjy2 = _mm256_setzero_pd();
847 fjz2 = _mm256_setzero_pd();
848 fjx3 = _mm256_setzero_pd();
849 fjy3 = _mm256_setzero_pd();
850 fjz3 = _mm256_setzero_pd();
852 /**************************
853 * CALCULATE INTERACTIONS *
854 **************************/
856 if (gmx_mm256_any_lt(rsq00,rcutoff2))
859 r00 = _mm256_mul_pd(rsq00,rinv00);
860 r00 = _mm256_andnot_pd(dummy_mask,r00);
862 /* Calculate table index by multiplying r with table scale and truncate to integer */
863 rt = _mm256_mul_pd(r00,vftabscale);
864 vfitab = _mm256_cvttpd_epi32(rt);
865 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
866 vfitab = _mm_slli_epi32(vfitab,3);
868 /* CUBIC SPLINE TABLE DISPERSION */
869 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
870 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
871 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
872 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
873 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
874 Heps = _mm256_mul_pd(vfeps,H);
875 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
876 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
877 vvdw6 = _mm256_mul_pd(c6_00,VV);
878 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
879 fvdw6 = _mm256_mul_pd(c6_00,FF);
881 /* CUBIC SPLINE TABLE REPULSION */
882 vfitab = _mm_add_epi32(vfitab,ifour);
883 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
884 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
885 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
886 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
887 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
888 Heps = _mm256_mul_pd(vfeps,H);
889 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
890 VV = _mm256_add_pd(Y,_mm256_mul_pd(vfeps,Fp));
891 vvdw12 = _mm256_mul_pd(c12_00,VV);
892 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
893 fvdw12 = _mm256_mul_pd(c12_00,FF);
894 vvdw = _mm256_add_pd(vvdw12,vvdw6);
895 fvdw = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_add_pd(fvdw6,fvdw12),_mm256_mul_pd(vftabscale,rinv00)));
897 cutoff_mask = _mm256_cmp_pd(rsq00,rcutoff2,_CMP_LT_OQ);
899 /* Update potential sum for this i atom from the interaction with this j atom. */
900 vvdw = _mm256_and_pd(vvdw,cutoff_mask);
901 vvdw = _mm256_andnot_pd(dummy_mask,vvdw);
902 vvdwsum = _mm256_add_pd(vvdwsum,vvdw);
906 fscal = _mm256_and_pd(fscal,cutoff_mask);
908 fscal = _mm256_andnot_pd(dummy_mask,fscal);
910 /* Calculate temporary vectorial force */
911 tx = _mm256_mul_pd(fscal,dx00);
912 ty = _mm256_mul_pd(fscal,dy00);
913 tz = _mm256_mul_pd(fscal,dz00);
915 /* Update vectorial force */
916 fix0 = _mm256_add_pd(fix0,tx);
917 fiy0 = _mm256_add_pd(fiy0,ty);
918 fiz0 = _mm256_add_pd(fiz0,tz);
920 fjx0 = _mm256_add_pd(fjx0,tx);
921 fjy0 = _mm256_add_pd(fjy0,ty);
922 fjz0 = _mm256_add_pd(fjz0,tz);
926 /**************************
927 * CALCULATE INTERACTIONS *
928 **************************/
930 if (gmx_mm256_any_lt(rsq11,rcutoff2))
933 /* REACTION-FIELD ELECTROSTATICS */
934 velec = _mm256_mul_pd(qq11,_mm256_sub_pd(_mm256_add_pd(rinv11,_mm256_mul_pd(krf,rsq11)),crf));
935 felec = _mm256_mul_pd(qq11,_mm256_sub_pd(_mm256_mul_pd(rinv11,rinvsq11),krf2));
937 cutoff_mask = _mm256_cmp_pd(rsq11,rcutoff2,_CMP_LT_OQ);
939 /* Update potential sum for this i atom from the interaction with this j atom. */
940 velec = _mm256_and_pd(velec,cutoff_mask);
941 velec = _mm256_andnot_pd(dummy_mask,velec);
942 velecsum = _mm256_add_pd(velecsum,velec);
946 fscal = _mm256_and_pd(fscal,cutoff_mask);
948 fscal = _mm256_andnot_pd(dummy_mask,fscal);
950 /* Calculate temporary vectorial force */
951 tx = _mm256_mul_pd(fscal,dx11);
952 ty = _mm256_mul_pd(fscal,dy11);
953 tz = _mm256_mul_pd(fscal,dz11);
955 /* Update vectorial force */
956 fix1 = _mm256_add_pd(fix1,tx);
957 fiy1 = _mm256_add_pd(fiy1,ty);
958 fiz1 = _mm256_add_pd(fiz1,tz);
960 fjx1 = _mm256_add_pd(fjx1,tx);
961 fjy1 = _mm256_add_pd(fjy1,ty);
962 fjz1 = _mm256_add_pd(fjz1,tz);
966 /**************************
967 * CALCULATE INTERACTIONS *
968 **************************/
970 if (gmx_mm256_any_lt(rsq12,rcutoff2))
973 /* REACTION-FIELD ELECTROSTATICS */
974 velec = _mm256_mul_pd(qq12,_mm256_sub_pd(_mm256_add_pd(rinv12,_mm256_mul_pd(krf,rsq12)),crf));
975 felec = _mm256_mul_pd(qq12,_mm256_sub_pd(_mm256_mul_pd(rinv12,rinvsq12),krf2));
977 cutoff_mask = _mm256_cmp_pd(rsq12,rcutoff2,_CMP_LT_OQ);
979 /* Update potential sum for this i atom from the interaction with this j atom. */
980 velec = _mm256_and_pd(velec,cutoff_mask);
981 velec = _mm256_andnot_pd(dummy_mask,velec);
982 velecsum = _mm256_add_pd(velecsum,velec);
986 fscal = _mm256_and_pd(fscal,cutoff_mask);
988 fscal = _mm256_andnot_pd(dummy_mask,fscal);
990 /* Calculate temporary vectorial force */
991 tx = _mm256_mul_pd(fscal,dx12);
992 ty = _mm256_mul_pd(fscal,dy12);
993 tz = _mm256_mul_pd(fscal,dz12);
995 /* Update vectorial force */
996 fix1 = _mm256_add_pd(fix1,tx);
997 fiy1 = _mm256_add_pd(fiy1,ty);
998 fiz1 = _mm256_add_pd(fiz1,tz);
1000 fjx2 = _mm256_add_pd(fjx2,tx);
1001 fjy2 = _mm256_add_pd(fjy2,ty);
1002 fjz2 = _mm256_add_pd(fjz2,tz);
1006 /**************************
1007 * CALCULATE INTERACTIONS *
1008 **************************/
1010 if (gmx_mm256_any_lt(rsq13,rcutoff2))
1013 /* REACTION-FIELD ELECTROSTATICS */
1014 velec = _mm256_mul_pd(qq13,_mm256_sub_pd(_mm256_add_pd(rinv13,_mm256_mul_pd(krf,rsq13)),crf));
1015 felec = _mm256_mul_pd(qq13,_mm256_sub_pd(_mm256_mul_pd(rinv13,rinvsq13),krf2));
1017 cutoff_mask = _mm256_cmp_pd(rsq13,rcutoff2,_CMP_LT_OQ);
1019 /* Update potential sum for this i atom from the interaction with this j atom. */
1020 velec = _mm256_and_pd(velec,cutoff_mask);
1021 velec = _mm256_andnot_pd(dummy_mask,velec);
1022 velecsum = _mm256_add_pd(velecsum,velec);
1026 fscal = _mm256_and_pd(fscal,cutoff_mask);
1028 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1030 /* Calculate temporary vectorial force */
1031 tx = _mm256_mul_pd(fscal,dx13);
1032 ty = _mm256_mul_pd(fscal,dy13);
1033 tz = _mm256_mul_pd(fscal,dz13);
1035 /* Update vectorial force */
1036 fix1 = _mm256_add_pd(fix1,tx);
1037 fiy1 = _mm256_add_pd(fiy1,ty);
1038 fiz1 = _mm256_add_pd(fiz1,tz);
1040 fjx3 = _mm256_add_pd(fjx3,tx);
1041 fjy3 = _mm256_add_pd(fjy3,ty);
1042 fjz3 = _mm256_add_pd(fjz3,tz);
1046 /**************************
1047 * CALCULATE INTERACTIONS *
1048 **************************/
1050 if (gmx_mm256_any_lt(rsq21,rcutoff2))
1053 /* REACTION-FIELD ELECTROSTATICS */
1054 velec = _mm256_mul_pd(qq21,_mm256_sub_pd(_mm256_add_pd(rinv21,_mm256_mul_pd(krf,rsq21)),crf));
1055 felec = _mm256_mul_pd(qq21,_mm256_sub_pd(_mm256_mul_pd(rinv21,rinvsq21),krf2));
1057 cutoff_mask = _mm256_cmp_pd(rsq21,rcutoff2,_CMP_LT_OQ);
1059 /* Update potential sum for this i atom from the interaction with this j atom. */
1060 velec = _mm256_and_pd(velec,cutoff_mask);
1061 velec = _mm256_andnot_pd(dummy_mask,velec);
1062 velecsum = _mm256_add_pd(velecsum,velec);
1066 fscal = _mm256_and_pd(fscal,cutoff_mask);
1068 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1070 /* Calculate temporary vectorial force */
1071 tx = _mm256_mul_pd(fscal,dx21);
1072 ty = _mm256_mul_pd(fscal,dy21);
1073 tz = _mm256_mul_pd(fscal,dz21);
1075 /* Update vectorial force */
1076 fix2 = _mm256_add_pd(fix2,tx);
1077 fiy2 = _mm256_add_pd(fiy2,ty);
1078 fiz2 = _mm256_add_pd(fiz2,tz);
1080 fjx1 = _mm256_add_pd(fjx1,tx);
1081 fjy1 = _mm256_add_pd(fjy1,ty);
1082 fjz1 = _mm256_add_pd(fjz1,tz);
1086 /**************************
1087 * CALCULATE INTERACTIONS *
1088 **************************/
1090 if (gmx_mm256_any_lt(rsq22,rcutoff2))
1093 /* REACTION-FIELD ELECTROSTATICS */
1094 velec = _mm256_mul_pd(qq22,_mm256_sub_pd(_mm256_add_pd(rinv22,_mm256_mul_pd(krf,rsq22)),crf));
1095 felec = _mm256_mul_pd(qq22,_mm256_sub_pd(_mm256_mul_pd(rinv22,rinvsq22),krf2));
1097 cutoff_mask = _mm256_cmp_pd(rsq22,rcutoff2,_CMP_LT_OQ);
1099 /* Update potential sum for this i atom from the interaction with this j atom. */
1100 velec = _mm256_and_pd(velec,cutoff_mask);
1101 velec = _mm256_andnot_pd(dummy_mask,velec);
1102 velecsum = _mm256_add_pd(velecsum,velec);
1106 fscal = _mm256_and_pd(fscal,cutoff_mask);
1108 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1110 /* Calculate temporary vectorial force */
1111 tx = _mm256_mul_pd(fscal,dx22);
1112 ty = _mm256_mul_pd(fscal,dy22);
1113 tz = _mm256_mul_pd(fscal,dz22);
1115 /* Update vectorial force */
1116 fix2 = _mm256_add_pd(fix2,tx);
1117 fiy2 = _mm256_add_pd(fiy2,ty);
1118 fiz2 = _mm256_add_pd(fiz2,tz);
1120 fjx2 = _mm256_add_pd(fjx2,tx);
1121 fjy2 = _mm256_add_pd(fjy2,ty);
1122 fjz2 = _mm256_add_pd(fjz2,tz);
1126 /**************************
1127 * CALCULATE INTERACTIONS *
1128 **************************/
1130 if (gmx_mm256_any_lt(rsq23,rcutoff2))
1133 /* REACTION-FIELD ELECTROSTATICS */
1134 velec = _mm256_mul_pd(qq23,_mm256_sub_pd(_mm256_add_pd(rinv23,_mm256_mul_pd(krf,rsq23)),crf));
1135 felec = _mm256_mul_pd(qq23,_mm256_sub_pd(_mm256_mul_pd(rinv23,rinvsq23),krf2));
1137 cutoff_mask = _mm256_cmp_pd(rsq23,rcutoff2,_CMP_LT_OQ);
1139 /* Update potential sum for this i atom from the interaction with this j atom. */
1140 velec = _mm256_and_pd(velec,cutoff_mask);
1141 velec = _mm256_andnot_pd(dummy_mask,velec);
1142 velecsum = _mm256_add_pd(velecsum,velec);
1146 fscal = _mm256_and_pd(fscal,cutoff_mask);
1148 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1150 /* Calculate temporary vectorial force */
1151 tx = _mm256_mul_pd(fscal,dx23);
1152 ty = _mm256_mul_pd(fscal,dy23);
1153 tz = _mm256_mul_pd(fscal,dz23);
1155 /* Update vectorial force */
1156 fix2 = _mm256_add_pd(fix2,tx);
1157 fiy2 = _mm256_add_pd(fiy2,ty);
1158 fiz2 = _mm256_add_pd(fiz2,tz);
1160 fjx3 = _mm256_add_pd(fjx3,tx);
1161 fjy3 = _mm256_add_pd(fjy3,ty);
1162 fjz3 = _mm256_add_pd(fjz3,tz);
1166 /**************************
1167 * CALCULATE INTERACTIONS *
1168 **************************/
1170 if (gmx_mm256_any_lt(rsq31,rcutoff2))
1173 /* REACTION-FIELD ELECTROSTATICS */
1174 velec = _mm256_mul_pd(qq31,_mm256_sub_pd(_mm256_add_pd(rinv31,_mm256_mul_pd(krf,rsq31)),crf));
1175 felec = _mm256_mul_pd(qq31,_mm256_sub_pd(_mm256_mul_pd(rinv31,rinvsq31),krf2));
1177 cutoff_mask = _mm256_cmp_pd(rsq31,rcutoff2,_CMP_LT_OQ);
1179 /* Update potential sum for this i atom from the interaction with this j atom. */
1180 velec = _mm256_and_pd(velec,cutoff_mask);
1181 velec = _mm256_andnot_pd(dummy_mask,velec);
1182 velecsum = _mm256_add_pd(velecsum,velec);
1186 fscal = _mm256_and_pd(fscal,cutoff_mask);
1188 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1190 /* Calculate temporary vectorial force */
1191 tx = _mm256_mul_pd(fscal,dx31);
1192 ty = _mm256_mul_pd(fscal,dy31);
1193 tz = _mm256_mul_pd(fscal,dz31);
1195 /* Update vectorial force */
1196 fix3 = _mm256_add_pd(fix3,tx);
1197 fiy3 = _mm256_add_pd(fiy3,ty);
1198 fiz3 = _mm256_add_pd(fiz3,tz);
1200 fjx1 = _mm256_add_pd(fjx1,tx);
1201 fjy1 = _mm256_add_pd(fjy1,ty);
1202 fjz1 = _mm256_add_pd(fjz1,tz);
1206 /**************************
1207 * CALCULATE INTERACTIONS *
1208 **************************/
1210 if (gmx_mm256_any_lt(rsq32,rcutoff2))
1213 /* REACTION-FIELD ELECTROSTATICS */
1214 velec = _mm256_mul_pd(qq32,_mm256_sub_pd(_mm256_add_pd(rinv32,_mm256_mul_pd(krf,rsq32)),crf));
1215 felec = _mm256_mul_pd(qq32,_mm256_sub_pd(_mm256_mul_pd(rinv32,rinvsq32),krf2));
1217 cutoff_mask = _mm256_cmp_pd(rsq32,rcutoff2,_CMP_LT_OQ);
1219 /* Update potential sum for this i atom from the interaction with this j atom. */
1220 velec = _mm256_and_pd(velec,cutoff_mask);
1221 velec = _mm256_andnot_pd(dummy_mask,velec);
1222 velecsum = _mm256_add_pd(velecsum,velec);
1226 fscal = _mm256_and_pd(fscal,cutoff_mask);
1228 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1230 /* Calculate temporary vectorial force */
1231 tx = _mm256_mul_pd(fscal,dx32);
1232 ty = _mm256_mul_pd(fscal,dy32);
1233 tz = _mm256_mul_pd(fscal,dz32);
1235 /* Update vectorial force */
1236 fix3 = _mm256_add_pd(fix3,tx);
1237 fiy3 = _mm256_add_pd(fiy3,ty);
1238 fiz3 = _mm256_add_pd(fiz3,tz);
1240 fjx2 = _mm256_add_pd(fjx2,tx);
1241 fjy2 = _mm256_add_pd(fjy2,ty);
1242 fjz2 = _mm256_add_pd(fjz2,tz);
1246 /**************************
1247 * CALCULATE INTERACTIONS *
1248 **************************/
1250 if (gmx_mm256_any_lt(rsq33,rcutoff2))
1253 /* REACTION-FIELD ELECTROSTATICS */
1254 velec = _mm256_mul_pd(qq33,_mm256_sub_pd(_mm256_add_pd(rinv33,_mm256_mul_pd(krf,rsq33)),crf));
1255 felec = _mm256_mul_pd(qq33,_mm256_sub_pd(_mm256_mul_pd(rinv33,rinvsq33),krf2));
1257 cutoff_mask = _mm256_cmp_pd(rsq33,rcutoff2,_CMP_LT_OQ);
1259 /* Update potential sum for this i atom from the interaction with this j atom. */
1260 velec = _mm256_and_pd(velec,cutoff_mask);
1261 velec = _mm256_andnot_pd(dummy_mask,velec);
1262 velecsum = _mm256_add_pd(velecsum,velec);
1266 fscal = _mm256_and_pd(fscal,cutoff_mask);
1268 fscal = _mm256_andnot_pd(dummy_mask,fscal);
1270 /* Calculate temporary vectorial force */
1271 tx = _mm256_mul_pd(fscal,dx33);
1272 ty = _mm256_mul_pd(fscal,dy33);
1273 tz = _mm256_mul_pd(fscal,dz33);
1275 /* Update vectorial force */
1276 fix3 = _mm256_add_pd(fix3,tx);
1277 fiy3 = _mm256_add_pd(fiy3,ty);
1278 fiz3 = _mm256_add_pd(fiz3,tz);
1280 fjx3 = _mm256_add_pd(fjx3,tx);
1281 fjy3 = _mm256_add_pd(fjy3,ty);
1282 fjz3 = _mm256_add_pd(fjz3,tz);
1286 fjptrA = (jnrlistA>=0) ? f+j_coord_offsetA : scratch;
1287 fjptrB = (jnrlistB>=0) ? f+j_coord_offsetB : scratch;
1288 fjptrC = (jnrlistC>=0) ? f+j_coord_offsetC : scratch;
1289 fjptrD = (jnrlistD>=0) ? f+j_coord_offsetD : scratch;
1291 gmx_mm256_decrement_4rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
1292 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,
1293 fjx2,fjy2,fjz2,fjx3,fjy3,fjz3);
1295 /* Inner loop uses 388 flops */
1298 /* End of innermost loop */
1300 gmx_mm256_update_iforce_4atom_swizzle_pd(fix0,fiy0,fiz0,fix1,fiy1,fiz1,fix2,fiy2,fiz2,fix3,fiy3,fiz3,
1301 f+i_coord_offset,fshift+i_shift_offset);
1304 /* Update potential energies */
1305 gmx_mm256_update_1pot_pd(velecsum,kernel_data->energygrp_elec+ggid);
1306 gmx_mm256_update_1pot_pd(vvdwsum,kernel_data->energygrp_vdw+ggid);
1308 /* Increment number of inner iterations */
1309 inneriter += j_index_end - j_index_start;
1311 /* Outer loop uses 26 flops */
1314 /* Increment number of outer iterations */
1317 /* Update outer/inner flops */
1319 inc_nrnb(nrnb,eNR_NBKERNEL_ELEC_VDW_W4W4_VF,outeriter*26 + inneriter*388);
1322 * Gromacs nonbonded kernel: nb_kernel_ElecRFCut_VdwCSTab_GeomW4W4_F_avx_256_double
1323 * Electrostatics interaction: ReactionField
1324 * VdW interaction: CubicSplineTable
1325 * Geometry: Water4-Water4
1326 * Calculate force/pot: Force
1329 nb_kernel_ElecRFCut_VdwCSTab_GeomW4W4_F_avx_256_double
1330 (t_nblist * gmx_restrict nlist,
1331 rvec * gmx_restrict xx,
1332 rvec * gmx_restrict ff,
1333 t_forcerec * gmx_restrict fr,
1334 t_mdatoms * gmx_restrict mdatoms,
1335 nb_kernel_data_t gmx_unused * gmx_restrict kernel_data,
1336 t_nrnb * gmx_restrict nrnb)
1338 /* Suffixes 0,1,2,3 refer to particle indices for waters in the inner or outer loop, or
1339 * just 0 for non-waters.
1340 * Suffixes A,B,C,D refer to j loop unrolling done with AVX, e.g. for the four different
1341 * jnr indices corresponding to data put in the four positions in the SIMD register.
1343 int i_shift_offset,i_coord_offset,outeriter,inneriter;
1344 int j_index_start,j_index_end,jidx,nri,inr,ggid,iidx;
1345 int jnrA,jnrB,jnrC,jnrD;
1346 int jnrlistA,jnrlistB,jnrlistC,jnrlistD;
1347 int jnrlistE,jnrlistF,jnrlistG,jnrlistH;
1348 int j_coord_offsetA,j_coord_offsetB,j_coord_offsetC,j_coord_offsetD;
1349 int *iinr,*jindex,*jjnr,*shiftidx,*gid;
1350 real rcutoff_scalar;
1351 real *shiftvec,*fshift,*x,*f;
1352 real *fjptrA,*fjptrB,*fjptrC,*fjptrD;
1353 real scratch[4*DIM];
1354 __m256d tx,ty,tz,fscal,rcutoff,rcutoff2,jidxall;
1355 real * vdwioffsetptr0;
1356 __m256d ix0,iy0,iz0,fix0,fiy0,fiz0,iq0,isai0;
1357 real * vdwioffsetptr1;
1358 __m256d ix1,iy1,iz1,fix1,fiy1,fiz1,iq1,isai1;
1359 real * vdwioffsetptr2;
1360 __m256d ix2,iy2,iz2,fix2,fiy2,fiz2,iq2,isai2;
1361 real * vdwioffsetptr3;
1362 __m256d ix3,iy3,iz3,fix3,fiy3,fiz3,iq3,isai3;
1363 int vdwjidx0A,vdwjidx0B,vdwjidx0C,vdwjidx0D;
1364 __m256d jx0,jy0,jz0,fjx0,fjy0,fjz0,jq0,isaj0;
1365 int vdwjidx1A,vdwjidx1B,vdwjidx1C,vdwjidx1D;
1366 __m256d jx1,jy1,jz1,fjx1,fjy1,fjz1,jq1,isaj1;
1367 int vdwjidx2A,vdwjidx2B,vdwjidx2C,vdwjidx2D;
1368 __m256d jx2,jy2,jz2,fjx2,fjy2,fjz2,jq2,isaj2;
1369 int vdwjidx3A,vdwjidx3B,vdwjidx3C,vdwjidx3D;
1370 __m256d jx3,jy3,jz3,fjx3,fjy3,fjz3,jq3,isaj3;
1371 __m256d dx00,dy00,dz00,rsq00,rinv00,rinvsq00,r00,qq00,c6_00,c12_00;
1372 __m256d dx11,dy11,dz11,rsq11,rinv11,rinvsq11,r11,qq11,c6_11,c12_11;
1373 __m256d dx12,dy12,dz12,rsq12,rinv12,rinvsq12,r12,qq12,c6_12,c12_12;
1374 __m256d dx13,dy13,dz13,rsq13,rinv13,rinvsq13,r13,qq13,c6_13,c12_13;
1375 __m256d dx21,dy21,dz21,rsq21,rinv21,rinvsq21,r21,qq21,c6_21,c12_21;
1376 __m256d dx22,dy22,dz22,rsq22,rinv22,rinvsq22,r22,qq22,c6_22,c12_22;
1377 __m256d dx23,dy23,dz23,rsq23,rinv23,rinvsq23,r23,qq23,c6_23,c12_23;
1378 __m256d dx31,dy31,dz31,rsq31,rinv31,rinvsq31,r31,qq31,c6_31,c12_31;
1379 __m256d dx32,dy32,dz32,rsq32,rinv32,rinvsq32,r32,qq32,c6_32,c12_32;
1380 __m256d dx33,dy33,dz33,rsq33,rinv33,rinvsq33,r33,qq33,c6_33,c12_33;
1381 __m256d velec,felec,velecsum,facel,crf,krf,krf2;
1384 __m256d rinvsix,rvdw,vvdw,vvdw6,vvdw12,fvdw,fvdw6,fvdw12,vvdwsum,sh_vdw_invrcut6;
1387 __m256d one_sixth = _mm256_set1_pd(1.0/6.0);
1388 __m256d one_twelfth = _mm256_set1_pd(1.0/12.0);
1390 __m128i ifour = _mm_set1_epi32(4);
1391 __m256d rt,vfeps,vftabscale,Y,F,G,H,Heps,Fp,VV,FF;
1393 __m256d dummy_mask,cutoff_mask;
1394 __m128 tmpmask0,tmpmask1;
1395 __m256d signbit = _mm256_castsi256_pd( _mm256_set1_epi32(0x80000000) );
1396 __m256d one = _mm256_set1_pd(1.0);
1397 __m256d two = _mm256_set1_pd(2.0);
1403 jindex = nlist->jindex;
1405 shiftidx = nlist->shift;
1407 shiftvec = fr->shift_vec[0];
1408 fshift = fr->fshift[0];
1409 facel = _mm256_set1_pd(fr->epsfac);
1410 charge = mdatoms->chargeA;
1411 krf = _mm256_set1_pd(fr->ic->k_rf);
1412 krf2 = _mm256_set1_pd(fr->ic->k_rf*2.0);
1413 crf = _mm256_set1_pd(fr->ic->c_rf);
1414 nvdwtype = fr->ntype;
1415 vdwparam = fr->nbfp;
1416 vdwtype = mdatoms->typeA;
1418 vftab = kernel_data->table_vdw->data;
1419 vftabscale = _mm256_set1_pd(kernel_data->table_vdw->scale);
1421 /* Setup water-specific parameters */
1422 inr = nlist->iinr[0];
1423 iq1 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+1]));
1424 iq2 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+2]));
1425 iq3 = _mm256_mul_pd(facel,_mm256_set1_pd(charge[inr+3]));
1426 vdwioffsetptr0 = vdwparam+2*nvdwtype*vdwtype[inr+0];
1428 jq1 = _mm256_set1_pd(charge[inr+1]);
1429 jq2 = _mm256_set1_pd(charge[inr+2]);
1430 jq3 = _mm256_set1_pd(charge[inr+3]);
1431 vdwjidx0A = 2*vdwtype[inr+0];
1432 c6_00 = _mm256_set1_pd(vdwioffsetptr0[vdwjidx0A]);
1433 c12_00 = _mm256_set1_pd(vdwioffsetptr0[vdwjidx0A+1]);
1434 qq11 = _mm256_mul_pd(iq1,jq1);
1435 qq12 = _mm256_mul_pd(iq1,jq2);
1436 qq13 = _mm256_mul_pd(iq1,jq3);
1437 qq21 = _mm256_mul_pd(iq2,jq1);
1438 qq22 = _mm256_mul_pd(iq2,jq2);
1439 qq23 = _mm256_mul_pd(iq2,jq3);
1440 qq31 = _mm256_mul_pd(iq3,jq1);
1441 qq32 = _mm256_mul_pd(iq3,jq2);
1442 qq33 = _mm256_mul_pd(iq3,jq3);
1444 /* When we use explicit cutoffs the value must be identical for elec and VdW, so use elec as an arbitrary choice */
1445 rcutoff_scalar = fr->rcoulomb;
1446 rcutoff = _mm256_set1_pd(rcutoff_scalar);
1447 rcutoff2 = _mm256_mul_pd(rcutoff,rcutoff);
1449 /* Avoid stupid compiler warnings */
1450 jnrA = jnrB = jnrC = jnrD = 0;
1451 j_coord_offsetA = 0;
1452 j_coord_offsetB = 0;
1453 j_coord_offsetC = 0;
1454 j_coord_offsetD = 0;
1459 for(iidx=0;iidx<4*DIM;iidx++)
1461 scratch[iidx] = 0.0;
1464 /* Start outer loop over neighborlists */
1465 for(iidx=0; iidx<nri; iidx++)
1467 /* Load shift vector for this list */
1468 i_shift_offset = DIM*shiftidx[iidx];
1470 /* Load limits for loop over neighbors */
1471 j_index_start = jindex[iidx];
1472 j_index_end = jindex[iidx+1];
1474 /* Get outer coordinate index */
1476 i_coord_offset = DIM*inr;
1478 /* Load i particle coords and add shift vector */
1479 gmx_mm256_load_shift_and_4rvec_broadcast_pd(shiftvec+i_shift_offset,x+i_coord_offset,
1480 &ix0,&iy0,&iz0,&ix1,&iy1,&iz1,&ix2,&iy2,&iz2,&ix3,&iy3,&iz3);
1482 fix0 = _mm256_setzero_pd();
1483 fiy0 = _mm256_setzero_pd();
1484 fiz0 = _mm256_setzero_pd();
1485 fix1 = _mm256_setzero_pd();
1486 fiy1 = _mm256_setzero_pd();
1487 fiz1 = _mm256_setzero_pd();
1488 fix2 = _mm256_setzero_pd();
1489 fiy2 = _mm256_setzero_pd();
1490 fiz2 = _mm256_setzero_pd();
1491 fix3 = _mm256_setzero_pd();
1492 fiy3 = _mm256_setzero_pd();
1493 fiz3 = _mm256_setzero_pd();
1495 /* Start inner kernel loop */
1496 for(jidx=j_index_start; jidx<j_index_end && jjnr[jidx+3]>=0; jidx+=4)
1499 /* Get j neighbor index, and coordinate index */
1501 jnrB = jjnr[jidx+1];
1502 jnrC = jjnr[jidx+2];
1503 jnrD = jjnr[jidx+3];
1504 j_coord_offsetA = DIM*jnrA;
1505 j_coord_offsetB = DIM*jnrB;
1506 j_coord_offsetC = DIM*jnrC;
1507 j_coord_offsetD = DIM*jnrD;
1509 /* load j atom coordinates */
1510 gmx_mm256_load_4rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
1511 x+j_coord_offsetC,x+j_coord_offsetD,
1512 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,
1513 &jy2,&jz2,&jx3,&jy3,&jz3);
1515 /* Calculate displacement vector */
1516 dx00 = _mm256_sub_pd(ix0,jx0);
1517 dy00 = _mm256_sub_pd(iy0,jy0);
1518 dz00 = _mm256_sub_pd(iz0,jz0);
1519 dx11 = _mm256_sub_pd(ix1,jx1);
1520 dy11 = _mm256_sub_pd(iy1,jy1);
1521 dz11 = _mm256_sub_pd(iz1,jz1);
1522 dx12 = _mm256_sub_pd(ix1,jx2);
1523 dy12 = _mm256_sub_pd(iy1,jy2);
1524 dz12 = _mm256_sub_pd(iz1,jz2);
1525 dx13 = _mm256_sub_pd(ix1,jx3);
1526 dy13 = _mm256_sub_pd(iy1,jy3);
1527 dz13 = _mm256_sub_pd(iz1,jz3);
1528 dx21 = _mm256_sub_pd(ix2,jx1);
1529 dy21 = _mm256_sub_pd(iy2,jy1);
1530 dz21 = _mm256_sub_pd(iz2,jz1);
1531 dx22 = _mm256_sub_pd(ix2,jx2);
1532 dy22 = _mm256_sub_pd(iy2,jy2);
1533 dz22 = _mm256_sub_pd(iz2,jz2);
1534 dx23 = _mm256_sub_pd(ix2,jx3);
1535 dy23 = _mm256_sub_pd(iy2,jy3);
1536 dz23 = _mm256_sub_pd(iz2,jz3);
1537 dx31 = _mm256_sub_pd(ix3,jx1);
1538 dy31 = _mm256_sub_pd(iy3,jy1);
1539 dz31 = _mm256_sub_pd(iz3,jz1);
1540 dx32 = _mm256_sub_pd(ix3,jx2);
1541 dy32 = _mm256_sub_pd(iy3,jy2);
1542 dz32 = _mm256_sub_pd(iz3,jz2);
1543 dx33 = _mm256_sub_pd(ix3,jx3);
1544 dy33 = _mm256_sub_pd(iy3,jy3);
1545 dz33 = _mm256_sub_pd(iz3,jz3);
1547 /* Calculate squared distance and things based on it */
1548 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
1549 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
1550 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
1551 rsq13 = gmx_mm256_calc_rsq_pd(dx13,dy13,dz13);
1552 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
1553 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
1554 rsq23 = gmx_mm256_calc_rsq_pd(dx23,dy23,dz23);
1555 rsq31 = gmx_mm256_calc_rsq_pd(dx31,dy31,dz31);
1556 rsq32 = gmx_mm256_calc_rsq_pd(dx32,dy32,dz32);
1557 rsq33 = gmx_mm256_calc_rsq_pd(dx33,dy33,dz33);
1559 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
1560 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
1561 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
1562 rinv13 = gmx_mm256_invsqrt_pd(rsq13);
1563 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
1564 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
1565 rinv23 = gmx_mm256_invsqrt_pd(rsq23);
1566 rinv31 = gmx_mm256_invsqrt_pd(rsq31);
1567 rinv32 = gmx_mm256_invsqrt_pd(rsq32);
1568 rinv33 = gmx_mm256_invsqrt_pd(rsq33);
1570 rinvsq11 = _mm256_mul_pd(rinv11,rinv11);
1571 rinvsq12 = _mm256_mul_pd(rinv12,rinv12);
1572 rinvsq13 = _mm256_mul_pd(rinv13,rinv13);
1573 rinvsq21 = _mm256_mul_pd(rinv21,rinv21);
1574 rinvsq22 = _mm256_mul_pd(rinv22,rinv22);
1575 rinvsq23 = _mm256_mul_pd(rinv23,rinv23);
1576 rinvsq31 = _mm256_mul_pd(rinv31,rinv31);
1577 rinvsq32 = _mm256_mul_pd(rinv32,rinv32);
1578 rinvsq33 = _mm256_mul_pd(rinv33,rinv33);
1580 fjx0 = _mm256_setzero_pd();
1581 fjy0 = _mm256_setzero_pd();
1582 fjz0 = _mm256_setzero_pd();
1583 fjx1 = _mm256_setzero_pd();
1584 fjy1 = _mm256_setzero_pd();
1585 fjz1 = _mm256_setzero_pd();
1586 fjx2 = _mm256_setzero_pd();
1587 fjy2 = _mm256_setzero_pd();
1588 fjz2 = _mm256_setzero_pd();
1589 fjx3 = _mm256_setzero_pd();
1590 fjy3 = _mm256_setzero_pd();
1591 fjz3 = _mm256_setzero_pd();
1593 /**************************
1594 * CALCULATE INTERACTIONS *
1595 **************************/
1597 if (gmx_mm256_any_lt(rsq00,rcutoff2))
1600 r00 = _mm256_mul_pd(rsq00,rinv00);
1602 /* Calculate table index by multiplying r with table scale and truncate to integer */
1603 rt = _mm256_mul_pd(r00,vftabscale);
1604 vfitab = _mm256_cvttpd_epi32(rt);
1605 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
1606 vfitab = _mm_slli_epi32(vfitab,3);
1608 /* CUBIC SPLINE TABLE DISPERSION */
1609 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1610 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1611 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1612 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1613 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1614 Heps = _mm256_mul_pd(vfeps,H);
1615 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1616 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1617 fvdw6 = _mm256_mul_pd(c6_00,FF);
1619 /* CUBIC SPLINE TABLE REPULSION */
1620 vfitab = _mm_add_epi32(vfitab,ifour);
1621 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
1622 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
1623 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
1624 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
1625 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
1626 Heps = _mm256_mul_pd(vfeps,H);
1627 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
1628 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
1629 fvdw12 = _mm256_mul_pd(c12_00,FF);
1630 fvdw = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_add_pd(fvdw6,fvdw12),_mm256_mul_pd(vftabscale,rinv00)));
1632 cutoff_mask = _mm256_cmp_pd(rsq00,rcutoff2,_CMP_LT_OQ);
1636 fscal = _mm256_and_pd(fscal,cutoff_mask);
1638 /* Calculate temporary vectorial force */
1639 tx = _mm256_mul_pd(fscal,dx00);
1640 ty = _mm256_mul_pd(fscal,dy00);
1641 tz = _mm256_mul_pd(fscal,dz00);
1643 /* Update vectorial force */
1644 fix0 = _mm256_add_pd(fix0,tx);
1645 fiy0 = _mm256_add_pd(fiy0,ty);
1646 fiz0 = _mm256_add_pd(fiz0,tz);
1648 fjx0 = _mm256_add_pd(fjx0,tx);
1649 fjy0 = _mm256_add_pd(fjy0,ty);
1650 fjz0 = _mm256_add_pd(fjz0,tz);
1654 /**************************
1655 * CALCULATE INTERACTIONS *
1656 **************************/
1658 if (gmx_mm256_any_lt(rsq11,rcutoff2))
1661 /* REACTION-FIELD ELECTROSTATICS */
1662 felec = _mm256_mul_pd(qq11,_mm256_sub_pd(_mm256_mul_pd(rinv11,rinvsq11),krf2));
1664 cutoff_mask = _mm256_cmp_pd(rsq11,rcutoff2,_CMP_LT_OQ);
1668 fscal = _mm256_and_pd(fscal,cutoff_mask);
1670 /* Calculate temporary vectorial force */
1671 tx = _mm256_mul_pd(fscal,dx11);
1672 ty = _mm256_mul_pd(fscal,dy11);
1673 tz = _mm256_mul_pd(fscal,dz11);
1675 /* Update vectorial force */
1676 fix1 = _mm256_add_pd(fix1,tx);
1677 fiy1 = _mm256_add_pd(fiy1,ty);
1678 fiz1 = _mm256_add_pd(fiz1,tz);
1680 fjx1 = _mm256_add_pd(fjx1,tx);
1681 fjy1 = _mm256_add_pd(fjy1,ty);
1682 fjz1 = _mm256_add_pd(fjz1,tz);
1686 /**************************
1687 * CALCULATE INTERACTIONS *
1688 **************************/
1690 if (gmx_mm256_any_lt(rsq12,rcutoff2))
1693 /* REACTION-FIELD ELECTROSTATICS */
1694 felec = _mm256_mul_pd(qq12,_mm256_sub_pd(_mm256_mul_pd(rinv12,rinvsq12),krf2));
1696 cutoff_mask = _mm256_cmp_pd(rsq12,rcutoff2,_CMP_LT_OQ);
1700 fscal = _mm256_and_pd(fscal,cutoff_mask);
1702 /* Calculate temporary vectorial force */
1703 tx = _mm256_mul_pd(fscal,dx12);
1704 ty = _mm256_mul_pd(fscal,dy12);
1705 tz = _mm256_mul_pd(fscal,dz12);
1707 /* Update vectorial force */
1708 fix1 = _mm256_add_pd(fix1,tx);
1709 fiy1 = _mm256_add_pd(fiy1,ty);
1710 fiz1 = _mm256_add_pd(fiz1,tz);
1712 fjx2 = _mm256_add_pd(fjx2,tx);
1713 fjy2 = _mm256_add_pd(fjy2,ty);
1714 fjz2 = _mm256_add_pd(fjz2,tz);
1718 /**************************
1719 * CALCULATE INTERACTIONS *
1720 **************************/
1722 if (gmx_mm256_any_lt(rsq13,rcutoff2))
1725 /* REACTION-FIELD ELECTROSTATICS */
1726 felec = _mm256_mul_pd(qq13,_mm256_sub_pd(_mm256_mul_pd(rinv13,rinvsq13),krf2));
1728 cutoff_mask = _mm256_cmp_pd(rsq13,rcutoff2,_CMP_LT_OQ);
1732 fscal = _mm256_and_pd(fscal,cutoff_mask);
1734 /* Calculate temporary vectorial force */
1735 tx = _mm256_mul_pd(fscal,dx13);
1736 ty = _mm256_mul_pd(fscal,dy13);
1737 tz = _mm256_mul_pd(fscal,dz13);
1739 /* Update vectorial force */
1740 fix1 = _mm256_add_pd(fix1,tx);
1741 fiy1 = _mm256_add_pd(fiy1,ty);
1742 fiz1 = _mm256_add_pd(fiz1,tz);
1744 fjx3 = _mm256_add_pd(fjx3,tx);
1745 fjy3 = _mm256_add_pd(fjy3,ty);
1746 fjz3 = _mm256_add_pd(fjz3,tz);
1750 /**************************
1751 * CALCULATE INTERACTIONS *
1752 **************************/
1754 if (gmx_mm256_any_lt(rsq21,rcutoff2))
1757 /* REACTION-FIELD ELECTROSTATICS */
1758 felec = _mm256_mul_pd(qq21,_mm256_sub_pd(_mm256_mul_pd(rinv21,rinvsq21),krf2));
1760 cutoff_mask = _mm256_cmp_pd(rsq21,rcutoff2,_CMP_LT_OQ);
1764 fscal = _mm256_and_pd(fscal,cutoff_mask);
1766 /* Calculate temporary vectorial force */
1767 tx = _mm256_mul_pd(fscal,dx21);
1768 ty = _mm256_mul_pd(fscal,dy21);
1769 tz = _mm256_mul_pd(fscal,dz21);
1771 /* Update vectorial force */
1772 fix2 = _mm256_add_pd(fix2,tx);
1773 fiy2 = _mm256_add_pd(fiy2,ty);
1774 fiz2 = _mm256_add_pd(fiz2,tz);
1776 fjx1 = _mm256_add_pd(fjx1,tx);
1777 fjy1 = _mm256_add_pd(fjy1,ty);
1778 fjz1 = _mm256_add_pd(fjz1,tz);
1782 /**************************
1783 * CALCULATE INTERACTIONS *
1784 **************************/
1786 if (gmx_mm256_any_lt(rsq22,rcutoff2))
1789 /* REACTION-FIELD ELECTROSTATICS */
1790 felec = _mm256_mul_pd(qq22,_mm256_sub_pd(_mm256_mul_pd(rinv22,rinvsq22),krf2));
1792 cutoff_mask = _mm256_cmp_pd(rsq22,rcutoff2,_CMP_LT_OQ);
1796 fscal = _mm256_and_pd(fscal,cutoff_mask);
1798 /* Calculate temporary vectorial force */
1799 tx = _mm256_mul_pd(fscal,dx22);
1800 ty = _mm256_mul_pd(fscal,dy22);
1801 tz = _mm256_mul_pd(fscal,dz22);
1803 /* Update vectorial force */
1804 fix2 = _mm256_add_pd(fix2,tx);
1805 fiy2 = _mm256_add_pd(fiy2,ty);
1806 fiz2 = _mm256_add_pd(fiz2,tz);
1808 fjx2 = _mm256_add_pd(fjx2,tx);
1809 fjy2 = _mm256_add_pd(fjy2,ty);
1810 fjz2 = _mm256_add_pd(fjz2,tz);
1814 /**************************
1815 * CALCULATE INTERACTIONS *
1816 **************************/
1818 if (gmx_mm256_any_lt(rsq23,rcutoff2))
1821 /* REACTION-FIELD ELECTROSTATICS */
1822 felec = _mm256_mul_pd(qq23,_mm256_sub_pd(_mm256_mul_pd(rinv23,rinvsq23),krf2));
1824 cutoff_mask = _mm256_cmp_pd(rsq23,rcutoff2,_CMP_LT_OQ);
1828 fscal = _mm256_and_pd(fscal,cutoff_mask);
1830 /* Calculate temporary vectorial force */
1831 tx = _mm256_mul_pd(fscal,dx23);
1832 ty = _mm256_mul_pd(fscal,dy23);
1833 tz = _mm256_mul_pd(fscal,dz23);
1835 /* Update vectorial force */
1836 fix2 = _mm256_add_pd(fix2,tx);
1837 fiy2 = _mm256_add_pd(fiy2,ty);
1838 fiz2 = _mm256_add_pd(fiz2,tz);
1840 fjx3 = _mm256_add_pd(fjx3,tx);
1841 fjy3 = _mm256_add_pd(fjy3,ty);
1842 fjz3 = _mm256_add_pd(fjz3,tz);
1846 /**************************
1847 * CALCULATE INTERACTIONS *
1848 **************************/
1850 if (gmx_mm256_any_lt(rsq31,rcutoff2))
1853 /* REACTION-FIELD ELECTROSTATICS */
1854 felec = _mm256_mul_pd(qq31,_mm256_sub_pd(_mm256_mul_pd(rinv31,rinvsq31),krf2));
1856 cutoff_mask = _mm256_cmp_pd(rsq31,rcutoff2,_CMP_LT_OQ);
1860 fscal = _mm256_and_pd(fscal,cutoff_mask);
1862 /* Calculate temporary vectorial force */
1863 tx = _mm256_mul_pd(fscal,dx31);
1864 ty = _mm256_mul_pd(fscal,dy31);
1865 tz = _mm256_mul_pd(fscal,dz31);
1867 /* Update vectorial force */
1868 fix3 = _mm256_add_pd(fix3,tx);
1869 fiy3 = _mm256_add_pd(fiy3,ty);
1870 fiz3 = _mm256_add_pd(fiz3,tz);
1872 fjx1 = _mm256_add_pd(fjx1,tx);
1873 fjy1 = _mm256_add_pd(fjy1,ty);
1874 fjz1 = _mm256_add_pd(fjz1,tz);
1878 /**************************
1879 * CALCULATE INTERACTIONS *
1880 **************************/
1882 if (gmx_mm256_any_lt(rsq32,rcutoff2))
1885 /* REACTION-FIELD ELECTROSTATICS */
1886 felec = _mm256_mul_pd(qq32,_mm256_sub_pd(_mm256_mul_pd(rinv32,rinvsq32),krf2));
1888 cutoff_mask = _mm256_cmp_pd(rsq32,rcutoff2,_CMP_LT_OQ);
1892 fscal = _mm256_and_pd(fscal,cutoff_mask);
1894 /* Calculate temporary vectorial force */
1895 tx = _mm256_mul_pd(fscal,dx32);
1896 ty = _mm256_mul_pd(fscal,dy32);
1897 tz = _mm256_mul_pd(fscal,dz32);
1899 /* Update vectorial force */
1900 fix3 = _mm256_add_pd(fix3,tx);
1901 fiy3 = _mm256_add_pd(fiy3,ty);
1902 fiz3 = _mm256_add_pd(fiz3,tz);
1904 fjx2 = _mm256_add_pd(fjx2,tx);
1905 fjy2 = _mm256_add_pd(fjy2,ty);
1906 fjz2 = _mm256_add_pd(fjz2,tz);
1910 /**************************
1911 * CALCULATE INTERACTIONS *
1912 **************************/
1914 if (gmx_mm256_any_lt(rsq33,rcutoff2))
1917 /* REACTION-FIELD ELECTROSTATICS */
1918 felec = _mm256_mul_pd(qq33,_mm256_sub_pd(_mm256_mul_pd(rinv33,rinvsq33),krf2));
1920 cutoff_mask = _mm256_cmp_pd(rsq33,rcutoff2,_CMP_LT_OQ);
1924 fscal = _mm256_and_pd(fscal,cutoff_mask);
1926 /* Calculate temporary vectorial force */
1927 tx = _mm256_mul_pd(fscal,dx33);
1928 ty = _mm256_mul_pd(fscal,dy33);
1929 tz = _mm256_mul_pd(fscal,dz33);
1931 /* Update vectorial force */
1932 fix3 = _mm256_add_pd(fix3,tx);
1933 fiy3 = _mm256_add_pd(fiy3,ty);
1934 fiz3 = _mm256_add_pd(fiz3,tz);
1936 fjx3 = _mm256_add_pd(fjx3,tx);
1937 fjy3 = _mm256_add_pd(fjy3,ty);
1938 fjz3 = _mm256_add_pd(fjz3,tz);
1942 fjptrA = f+j_coord_offsetA;
1943 fjptrB = f+j_coord_offsetB;
1944 fjptrC = f+j_coord_offsetC;
1945 fjptrD = f+j_coord_offsetD;
1947 gmx_mm256_decrement_4rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
1948 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,
1949 fjx2,fjy2,fjz2,fjx3,fjy3,fjz3);
1951 /* Inner loop uses 324 flops */
1954 if(jidx<j_index_end)
1957 /* Get j neighbor index, and coordinate index */
1958 jnrlistA = jjnr[jidx];
1959 jnrlistB = jjnr[jidx+1];
1960 jnrlistC = jjnr[jidx+2];
1961 jnrlistD = jjnr[jidx+3];
1962 /* Sign of each element will be negative for non-real atoms.
1963 * This mask will be 0xFFFFFFFF for dummy entries and 0x0 for real ones,
1964 * so use it as val = _mm_andnot_pd(mask,val) to clear dummy entries.
1966 tmpmask0 = gmx_mm_castsi128_ps(_mm_cmplt_epi32(_mm_loadu_si128((const __m128i *)(jjnr+jidx)),_mm_setzero_si128()));
1968 tmpmask1 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(3,3,2,2));
1969 tmpmask0 = _mm_permute_ps(tmpmask0,_GMX_MM_PERMUTE(1,1,0,0));
1970 dummy_mask = _mm256_castps_pd(gmx_mm256_set_m128(tmpmask1,tmpmask0));
1972 jnrA = (jnrlistA>=0) ? jnrlistA : 0;
1973 jnrB = (jnrlistB>=0) ? jnrlistB : 0;
1974 jnrC = (jnrlistC>=0) ? jnrlistC : 0;
1975 jnrD = (jnrlistD>=0) ? jnrlistD : 0;
1976 j_coord_offsetA = DIM*jnrA;
1977 j_coord_offsetB = DIM*jnrB;
1978 j_coord_offsetC = DIM*jnrC;
1979 j_coord_offsetD = DIM*jnrD;
1981 /* load j atom coordinates */
1982 gmx_mm256_load_4rvec_4ptr_swizzle_pd(x+j_coord_offsetA,x+j_coord_offsetB,
1983 x+j_coord_offsetC,x+j_coord_offsetD,
1984 &jx0,&jy0,&jz0,&jx1,&jy1,&jz1,&jx2,
1985 &jy2,&jz2,&jx3,&jy3,&jz3);
1987 /* Calculate displacement vector */
1988 dx00 = _mm256_sub_pd(ix0,jx0);
1989 dy00 = _mm256_sub_pd(iy0,jy0);
1990 dz00 = _mm256_sub_pd(iz0,jz0);
1991 dx11 = _mm256_sub_pd(ix1,jx1);
1992 dy11 = _mm256_sub_pd(iy1,jy1);
1993 dz11 = _mm256_sub_pd(iz1,jz1);
1994 dx12 = _mm256_sub_pd(ix1,jx2);
1995 dy12 = _mm256_sub_pd(iy1,jy2);
1996 dz12 = _mm256_sub_pd(iz1,jz2);
1997 dx13 = _mm256_sub_pd(ix1,jx3);
1998 dy13 = _mm256_sub_pd(iy1,jy3);
1999 dz13 = _mm256_sub_pd(iz1,jz3);
2000 dx21 = _mm256_sub_pd(ix2,jx1);
2001 dy21 = _mm256_sub_pd(iy2,jy1);
2002 dz21 = _mm256_sub_pd(iz2,jz1);
2003 dx22 = _mm256_sub_pd(ix2,jx2);
2004 dy22 = _mm256_sub_pd(iy2,jy2);
2005 dz22 = _mm256_sub_pd(iz2,jz2);
2006 dx23 = _mm256_sub_pd(ix2,jx3);
2007 dy23 = _mm256_sub_pd(iy2,jy3);
2008 dz23 = _mm256_sub_pd(iz2,jz3);
2009 dx31 = _mm256_sub_pd(ix3,jx1);
2010 dy31 = _mm256_sub_pd(iy3,jy1);
2011 dz31 = _mm256_sub_pd(iz3,jz1);
2012 dx32 = _mm256_sub_pd(ix3,jx2);
2013 dy32 = _mm256_sub_pd(iy3,jy2);
2014 dz32 = _mm256_sub_pd(iz3,jz2);
2015 dx33 = _mm256_sub_pd(ix3,jx3);
2016 dy33 = _mm256_sub_pd(iy3,jy3);
2017 dz33 = _mm256_sub_pd(iz3,jz3);
2019 /* Calculate squared distance and things based on it */
2020 rsq00 = gmx_mm256_calc_rsq_pd(dx00,dy00,dz00);
2021 rsq11 = gmx_mm256_calc_rsq_pd(dx11,dy11,dz11);
2022 rsq12 = gmx_mm256_calc_rsq_pd(dx12,dy12,dz12);
2023 rsq13 = gmx_mm256_calc_rsq_pd(dx13,dy13,dz13);
2024 rsq21 = gmx_mm256_calc_rsq_pd(dx21,dy21,dz21);
2025 rsq22 = gmx_mm256_calc_rsq_pd(dx22,dy22,dz22);
2026 rsq23 = gmx_mm256_calc_rsq_pd(dx23,dy23,dz23);
2027 rsq31 = gmx_mm256_calc_rsq_pd(dx31,dy31,dz31);
2028 rsq32 = gmx_mm256_calc_rsq_pd(dx32,dy32,dz32);
2029 rsq33 = gmx_mm256_calc_rsq_pd(dx33,dy33,dz33);
2031 rinv00 = gmx_mm256_invsqrt_pd(rsq00);
2032 rinv11 = gmx_mm256_invsqrt_pd(rsq11);
2033 rinv12 = gmx_mm256_invsqrt_pd(rsq12);
2034 rinv13 = gmx_mm256_invsqrt_pd(rsq13);
2035 rinv21 = gmx_mm256_invsqrt_pd(rsq21);
2036 rinv22 = gmx_mm256_invsqrt_pd(rsq22);
2037 rinv23 = gmx_mm256_invsqrt_pd(rsq23);
2038 rinv31 = gmx_mm256_invsqrt_pd(rsq31);
2039 rinv32 = gmx_mm256_invsqrt_pd(rsq32);
2040 rinv33 = gmx_mm256_invsqrt_pd(rsq33);
2042 rinvsq11 = _mm256_mul_pd(rinv11,rinv11);
2043 rinvsq12 = _mm256_mul_pd(rinv12,rinv12);
2044 rinvsq13 = _mm256_mul_pd(rinv13,rinv13);
2045 rinvsq21 = _mm256_mul_pd(rinv21,rinv21);
2046 rinvsq22 = _mm256_mul_pd(rinv22,rinv22);
2047 rinvsq23 = _mm256_mul_pd(rinv23,rinv23);
2048 rinvsq31 = _mm256_mul_pd(rinv31,rinv31);
2049 rinvsq32 = _mm256_mul_pd(rinv32,rinv32);
2050 rinvsq33 = _mm256_mul_pd(rinv33,rinv33);
2052 fjx0 = _mm256_setzero_pd();
2053 fjy0 = _mm256_setzero_pd();
2054 fjz0 = _mm256_setzero_pd();
2055 fjx1 = _mm256_setzero_pd();
2056 fjy1 = _mm256_setzero_pd();
2057 fjz1 = _mm256_setzero_pd();
2058 fjx2 = _mm256_setzero_pd();
2059 fjy2 = _mm256_setzero_pd();
2060 fjz2 = _mm256_setzero_pd();
2061 fjx3 = _mm256_setzero_pd();
2062 fjy3 = _mm256_setzero_pd();
2063 fjz3 = _mm256_setzero_pd();
2065 /**************************
2066 * CALCULATE INTERACTIONS *
2067 **************************/
2069 if (gmx_mm256_any_lt(rsq00,rcutoff2))
2072 r00 = _mm256_mul_pd(rsq00,rinv00);
2073 r00 = _mm256_andnot_pd(dummy_mask,r00);
2075 /* Calculate table index by multiplying r with table scale and truncate to integer */
2076 rt = _mm256_mul_pd(r00,vftabscale);
2077 vfitab = _mm256_cvttpd_epi32(rt);
2078 vfeps = _mm256_sub_pd(rt,_mm256_round_pd(rt, _MM_FROUND_FLOOR));
2079 vfitab = _mm_slli_epi32(vfitab,3);
2081 /* CUBIC SPLINE TABLE DISPERSION */
2082 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2083 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2084 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2085 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2086 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2087 Heps = _mm256_mul_pd(vfeps,H);
2088 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2089 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2090 fvdw6 = _mm256_mul_pd(c6_00,FF);
2092 /* CUBIC SPLINE TABLE REPULSION */
2093 vfitab = _mm_add_epi32(vfitab,ifour);
2094 Y = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,0) );
2095 F = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,1) );
2096 G = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,2) );
2097 H = _mm256_load_pd( vftab + _mm_extract_epi32(vfitab,3) );
2098 GMX_MM256_FULLTRANSPOSE4_PD(Y,F,G,H);
2099 Heps = _mm256_mul_pd(vfeps,H);
2100 Fp = _mm256_add_pd(F,_mm256_mul_pd(vfeps,_mm256_add_pd(G,Heps)));
2101 FF = _mm256_add_pd(Fp,_mm256_mul_pd(vfeps,_mm256_add_pd(G,_mm256_add_pd(Heps,Heps))));
2102 fvdw12 = _mm256_mul_pd(c12_00,FF);
2103 fvdw = _mm256_xor_pd(signbit,_mm256_mul_pd(_mm256_add_pd(fvdw6,fvdw12),_mm256_mul_pd(vftabscale,rinv00)));
2105 cutoff_mask = _mm256_cmp_pd(rsq00,rcutoff2,_CMP_LT_OQ);
2109 fscal = _mm256_and_pd(fscal,cutoff_mask);
2111 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2113 /* Calculate temporary vectorial force */
2114 tx = _mm256_mul_pd(fscal,dx00);
2115 ty = _mm256_mul_pd(fscal,dy00);
2116 tz = _mm256_mul_pd(fscal,dz00);
2118 /* Update vectorial force */
2119 fix0 = _mm256_add_pd(fix0,tx);
2120 fiy0 = _mm256_add_pd(fiy0,ty);
2121 fiz0 = _mm256_add_pd(fiz0,tz);
2123 fjx0 = _mm256_add_pd(fjx0,tx);
2124 fjy0 = _mm256_add_pd(fjy0,ty);
2125 fjz0 = _mm256_add_pd(fjz0,tz);
2129 /**************************
2130 * CALCULATE INTERACTIONS *
2131 **************************/
2133 if (gmx_mm256_any_lt(rsq11,rcutoff2))
2136 /* REACTION-FIELD ELECTROSTATICS */
2137 felec = _mm256_mul_pd(qq11,_mm256_sub_pd(_mm256_mul_pd(rinv11,rinvsq11),krf2));
2139 cutoff_mask = _mm256_cmp_pd(rsq11,rcutoff2,_CMP_LT_OQ);
2143 fscal = _mm256_and_pd(fscal,cutoff_mask);
2145 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2147 /* Calculate temporary vectorial force */
2148 tx = _mm256_mul_pd(fscal,dx11);
2149 ty = _mm256_mul_pd(fscal,dy11);
2150 tz = _mm256_mul_pd(fscal,dz11);
2152 /* Update vectorial force */
2153 fix1 = _mm256_add_pd(fix1,tx);
2154 fiy1 = _mm256_add_pd(fiy1,ty);
2155 fiz1 = _mm256_add_pd(fiz1,tz);
2157 fjx1 = _mm256_add_pd(fjx1,tx);
2158 fjy1 = _mm256_add_pd(fjy1,ty);
2159 fjz1 = _mm256_add_pd(fjz1,tz);
2163 /**************************
2164 * CALCULATE INTERACTIONS *
2165 **************************/
2167 if (gmx_mm256_any_lt(rsq12,rcutoff2))
2170 /* REACTION-FIELD ELECTROSTATICS */
2171 felec = _mm256_mul_pd(qq12,_mm256_sub_pd(_mm256_mul_pd(rinv12,rinvsq12),krf2));
2173 cutoff_mask = _mm256_cmp_pd(rsq12,rcutoff2,_CMP_LT_OQ);
2177 fscal = _mm256_and_pd(fscal,cutoff_mask);
2179 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2181 /* Calculate temporary vectorial force */
2182 tx = _mm256_mul_pd(fscal,dx12);
2183 ty = _mm256_mul_pd(fscal,dy12);
2184 tz = _mm256_mul_pd(fscal,dz12);
2186 /* Update vectorial force */
2187 fix1 = _mm256_add_pd(fix1,tx);
2188 fiy1 = _mm256_add_pd(fiy1,ty);
2189 fiz1 = _mm256_add_pd(fiz1,tz);
2191 fjx2 = _mm256_add_pd(fjx2,tx);
2192 fjy2 = _mm256_add_pd(fjy2,ty);
2193 fjz2 = _mm256_add_pd(fjz2,tz);
2197 /**************************
2198 * CALCULATE INTERACTIONS *
2199 **************************/
2201 if (gmx_mm256_any_lt(rsq13,rcutoff2))
2204 /* REACTION-FIELD ELECTROSTATICS */
2205 felec = _mm256_mul_pd(qq13,_mm256_sub_pd(_mm256_mul_pd(rinv13,rinvsq13),krf2));
2207 cutoff_mask = _mm256_cmp_pd(rsq13,rcutoff2,_CMP_LT_OQ);
2211 fscal = _mm256_and_pd(fscal,cutoff_mask);
2213 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2215 /* Calculate temporary vectorial force */
2216 tx = _mm256_mul_pd(fscal,dx13);
2217 ty = _mm256_mul_pd(fscal,dy13);
2218 tz = _mm256_mul_pd(fscal,dz13);
2220 /* Update vectorial force */
2221 fix1 = _mm256_add_pd(fix1,tx);
2222 fiy1 = _mm256_add_pd(fiy1,ty);
2223 fiz1 = _mm256_add_pd(fiz1,tz);
2225 fjx3 = _mm256_add_pd(fjx3,tx);
2226 fjy3 = _mm256_add_pd(fjy3,ty);
2227 fjz3 = _mm256_add_pd(fjz3,tz);
2231 /**************************
2232 * CALCULATE INTERACTIONS *
2233 **************************/
2235 if (gmx_mm256_any_lt(rsq21,rcutoff2))
2238 /* REACTION-FIELD ELECTROSTATICS */
2239 felec = _mm256_mul_pd(qq21,_mm256_sub_pd(_mm256_mul_pd(rinv21,rinvsq21),krf2));
2241 cutoff_mask = _mm256_cmp_pd(rsq21,rcutoff2,_CMP_LT_OQ);
2245 fscal = _mm256_and_pd(fscal,cutoff_mask);
2247 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2249 /* Calculate temporary vectorial force */
2250 tx = _mm256_mul_pd(fscal,dx21);
2251 ty = _mm256_mul_pd(fscal,dy21);
2252 tz = _mm256_mul_pd(fscal,dz21);
2254 /* Update vectorial force */
2255 fix2 = _mm256_add_pd(fix2,tx);
2256 fiy2 = _mm256_add_pd(fiy2,ty);
2257 fiz2 = _mm256_add_pd(fiz2,tz);
2259 fjx1 = _mm256_add_pd(fjx1,tx);
2260 fjy1 = _mm256_add_pd(fjy1,ty);
2261 fjz1 = _mm256_add_pd(fjz1,tz);
2265 /**************************
2266 * CALCULATE INTERACTIONS *
2267 **************************/
2269 if (gmx_mm256_any_lt(rsq22,rcutoff2))
2272 /* REACTION-FIELD ELECTROSTATICS */
2273 felec = _mm256_mul_pd(qq22,_mm256_sub_pd(_mm256_mul_pd(rinv22,rinvsq22),krf2));
2275 cutoff_mask = _mm256_cmp_pd(rsq22,rcutoff2,_CMP_LT_OQ);
2279 fscal = _mm256_and_pd(fscal,cutoff_mask);
2281 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2283 /* Calculate temporary vectorial force */
2284 tx = _mm256_mul_pd(fscal,dx22);
2285 ty = _mm256_mul_pd(fscal,dy22);
2286 tz = _mm256_mul_pd(fscal,dz22);
2288 /* Update vectorial force */
2289 fix2 = _mm256_add_pd(fix2,tx);
2290 fiy2 = _mm256_add_pd(fiy2,ty);
2291 fiz2 = _mm256_add_pd(fiz2,tz);
2293 fjx2 = _mm256_add_pd(fjx2,tx);
2294 fjy2 = _mm256_add_pd(fjy2,ty);
2295 fjz2 = _mm256_add_pd(fjz2,tz);
2299 /**************************
2300 * CALCULATE INTERACTIONS *
2301 **************************/
2303 if (gmx_mm256_any_lt(rsq23,rcutoff2))
2306 /* REACTION-FIELD ELECTROSTATICS */
2307 felec = _mm256_mul_pd(qq23,_mm256_sub_pd(_mm256_mul_pd(rinv23,rinvsq23),krf2));
2309 cutoff_mask = _mm256_cmp_pd(rsq23,rcutoff2,_CMP_LT_OQ);
2313 fscal = _mm256_and_pd(fscal,cutoff_mask);
2315 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2317 /* Calculate temporary vectorial force */
2318 tx = _mm256_mul_pd(fscal,dx23);
2319 ty = _mm256_mul_pd(fscal,dy23);
2320 tz = _mm256_mul_pd(fscal,dz23);
2322 /* Update vectorial force */
2323 fix2 = _mm256_add_pd(fix2,tx);
2324 fiy2 = _mm256_add_pd(fiy2,ty);
2325 fiz2 = _mm256_add_pd(fiz2,tz);
2327 fjx3 = _mm256_add_pd(fjx3,tx);
2328 fjy3 = _mm256_add_pd(fjy3,ty);
2329 fjz3 = _mm256_add_pd(fjz3,tz);
2333 /**************************
2334 * CALCULATE INTERACTIONS *
2335 **************************/
2337 if (gmx_mm256_any_lt(rsq31,rcutoff2))
2340 /* REACTION-FIELD ELECTROSTATICS */
2341 felec = _mm256_mul_pd(qq31,_mm256_sub_pd(_mm256_mul_pd(rinv31,rinvsq31),krf2));
2343 cutoff_mask = _mm256_cmp_pd(rsq31,rcutoff2,_CMP_LT_OQ);
2347 fscal = _mm256_and_pd(fscal,cutoff_mask);
2349 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2351 /* Calculate temporary vectorial force */
2352 tx = _mm256_mul_pd(fscal,dx31);
2353 ty = _mm256_mul_pd(fscal,dy31);
2354 tz = _mm256_mul_pd(fscal,dz31);
2356 /* Update vectorial force */
2357 fix3 = _mm256_add_pd(fix3,tx);
2358 fiy3 = _mm256_add_pd(fiy3,ty);
2359 fiz3 = _mm256_add_pd(fiz3,tz);
2361 fjx1 = _mm256_add_pd(fjx1,tx);
2362 fjy1 = _mm256_add_pd(fjy1,ty);
2363 fjz1 = _mm256_add_pd(fjz1,tz);
2367 /**************************
2368 * CALCULATE INTERACTIONS *
2369 **************************/
2371 if (gmx_mm256_any_lt(rsq32,rcutoff2))
2374 /* REACTION-FIELD ELECTROSTATICS */
2375 felec = _mm256_mul_pd(qq32,_mm256_sub_pd(_mm256_mul_pd(rinv32,rinvsq32),krf2));
2377 cutoff_mask = _mm256_cmp_pd(rsq32,rcutoff2,_CMP_LT_OQ);
2381 fscal = _mm256_and_pd(fscal,cutoff_mask);
2383 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2385 /* Calculate temporary vectorial force */
2386 tx = _mm256_mul_pd(fscal,dx32);
2387 ty = _mm256_mul_pd(fscal,dy32);
2388 tz = _mm256_mul_pd(fscal,dz32);
2390 /* Update vectorial force */
2391 fix3 = _mm256_add_pd(fix3,tx);
2392 fiy3 = _mm256_add_pd(fiy3,ty);
2393 fiz3 = _mm256_add_pd(fiz3,tz);
2395 fjx2 = _mm256_add_pd(fjx2,tx);
2396 fjy2 = _mm256_add_pd(fjy2,ty);
2397 fjz2 = _mm256_add_pd(fjz2,tz);
2401 /**************************
2402 * CALCULATE INTERACTIONS *
2403 **************************/
2405 if (gmx_mm256_any_lt(rsq33,rcutoff2))
2408 /* REACTION-FIELD ELECTROSTATICS */
2409 felec = _mm256_mul_pd(qq33,_mm256_sub_pd(_mm256_mul_pd(rinv33,rinvsq33),krf2));
2411 cutoff_mask = _mm256_cmp_pd(rsq33,rcutoff2,_CMP_LT_OQ);
2415 fscal = _mm256_and_pd(fscal,cutoff_mask);
2417 fscal = _mm256_andnot_pd(dummy_mask,fscal);
2419 /* Calculate temporary vectorial force */
2420 tx = _mm256_mul_pd(fscal,dx33);
2421 ty = _mm256_mul_pd(fscal,dy33);
2422 tz = _mm256_mul_pd(fscal,dz33);
2424 /* Update vectorial force */
2425 fix3 = _mm256_add_pd(fix3,tx);
2426 fiy3 = _mm256_add_pd(fiy3,ty);
2427 fiz3 = _mm256_add_pd(fiz3,tz);
2429 fjx3 = _mm256_add_pd(fjx3,tx);
2430 fjy3 = _mm256_add_pd(fjy3,ty);
2431 fjz3 = _mm256_add_pd(fjz3,tz);
2435 fjptrA = (jnrlistA>=0) ? f+j_coord_offsetA : scratch;
2436 fjptrB = (jnrlistB>=0) ? f+j_coord_offsetB : scratch;
2437 fjptrC = (jnrlistC>=0) ? f+j_coord_offsetC : scratch;
2438 fjptrD = (jnrlistD>=0) ? f+j_coord_offsetD : scratch;
2440 gmx_mm256_decrement_4rvec_4ptr_swizzle_pd(fjptrA,fjptrB,fjptrC,fjptrD,
2441 fjx0,fjy0,fjz0,fjx1,fjy1,fjz1,
2442 fjx2,fjy2,fjz2,fjx3,fjy3,fjz3);
2444 /* Inner loop uses 325 flops */
2447 /* End of innermost loop */
2449 gmx_mm256_update_iforce_4atom_swizzle_pd(fix0,fiy0,fiz0,fix1,fiy1,fiz1,fix2,fiy2,fiz2,fix3,fiy3,fiz3,
2450 f+i_coord_offset,fshift+i_shift_offset);
2452 /* Increment number of inner iterations */
2453 inneriter += j_index_end - j_index_start;
2455 /* Outer loop uses 24 flops */
2458 /* Increment number of outer iterations */
2461 /* Update outer/inner flops */
2463 inc_nrnb(nrnb,eNR_NBKERNEL_ELEC_VDW_W4W4_F,outeriter*24 + inneriter*325);