{
// Create streams to manage pipelining
ppCommManagers_.reserve(ppRanks.size());
- for (auto& ppRank : ppRanks)
+ for (const auto& ppRank : ppRanks)
{
ppCommManagers_.emplace_back(PpCommManager{
ppRank,
#if GMX_MPI
// Receive event from PP task
- // NOLINTNEXTLINE(bugprone-sizeof-expression)
MPI_Irecv(&ppCommManagers_[ppRank].sync,
- sizeof(GpuEventSynchronizer*),
+ sizeof(GpuEventSynchronizer*), // NOLINT(bugprone-sizeof-expression)
MPI_BYTE,
ppRank,
0,
*/
#if GMX_GPU_CUDA
- pmeGpu->kernelParams->usePipeline = false;
+ pmeGpu->kernelParams->usePipeline = char(false);
pmeGpu->kernelParams->pipelineAtomStart = 0;
pmeGpu->kernelParams->pipelineAtomEnd = 0;
pmeGpu->maxGridWidthX = deviceContext.deviceInfo().prop.maxGridSize[0];
pme_gpu_start_timing(pmeGpu, timingId);
auto* timingEvent = pme_gpu_fetch_timing_event(pmeGpu, timingId);
- kernelParamsPtr->usePipeline = computeSplines && spreadCharges && useGpuDirectComm
- && (pmeCoordinateReceiverGpu->ppCommNumSenderRanks() > 1)
- && !writeGlobalOrSaveSplines;
- if (kernelParamsPtr->usePipeline)
+ kernelParamsPtr->usePipeline = char(computeSplines && spreadCharges && useGpuDirectComm
+ && (pmeCoordinateReceiverGpu->ppCommNumSenderRanks() > 1)
+ && !writeGlobalOrSaveSplines);
+ if (kernelParamsPtr->usePipeline != 0)
{
int numStagesInPipeline = pmeCoordinateReceiverGpu->ppCommNumSenderRanks();