config.stream = stream_;
auto kernelPtr = leapfrog_kernel;
- const float3 *gm_x = d_x;
- float3 *gm_xp = d_xp;
- float3 *gm_v = d_v;
- const float3 *gm_f = d_f;
- const float *gm_inverseMasses = d_inverseMasses_;
const auto kernelArgs = prepareGpuKernelArguments(kernelPtr, config,
&numAtoms_,
- &gm_x, &gm_xp,
- &gm_v,
- &gm_f,
- &gm_inverseMasses, &dt);
+ &d_x, &d_xp,
+ &d_v,
+ &d_f,
+ &d_inverseMasses_, &dt);
launchGpuKernel(kernelPtr, config, nullptr, "leapfrog_kernel", kernelArgs);
return;
}
config.stream = stream_;
- // This is to satisfy prepareGpuKernelArguments(...)
- // It there a better way?
- const float3 * gm_x = d_x;
- float3 * gm_xp = d_xp;
- float3 * gm_v = d_v;
-
const auto kernelArgs = prepareGpuKernelArguments(kernelPtr, config,
&kernelParams_,
- &gm_x, &gm_xp,
- &gm_v, &invdt);
+ &d_x, &d_xp,
+ &d_v, &invdt);
launchGpuKernel(kernelPtr, config, nullptr,
"lincs_kernel<updateVelocities, computeVirial>", kernelArgs);
}
config.stream = stream_;
- const int3 *gm_atomIds = d_atomIds_;
- const float3 *gm_x = d_x;
- float3 *gm_xp = d_xp;
- float3 *gm_v = d_v;
- float *gm_virialScaled = d_virialScaled_;
-
const auto kernelArgs = prepareGpuKernelArguments(kernelPtr, config,
&numSettles_,
- &gm_atomIds,
+ &d_atomIds_,
&settleParameters_,
- &gm_x,
- &gm_xp,
+ &d_x,
+ &d_xp,
&pbcAiuc_,
&invdt,
- &gm_v,
- &gm_virialScaled);
+ &d_v,
+ &d_virialScaled_);
launchGpuKernel(kernelPtr, config, nullptr,
"settle_kernel<updateVelocities, computeVirial>", kernelArgs);