| 233 |  | } | 
| 234 |  | } | 
| 235 |  |  | 
| 236 | < | #endif | 
| 237 | < |  | 
| 238 | < | // allocate memory for the parallel objects | 
| 239 | < | atypesLocal.resize(nLocal_); | 
| 240 | < |  | 
| 241 | < | for (int i = 0; i < nLocal_; i++) | 
| 242 | < | atypesLocal[i] = ff_->getAtomType(idents[i]); | 
| 243 | < |  | 
| 244 | < | groupList_.clear(); | 
| 245 | < | groupList_.resize(nGroups_); | 
| 246 | < | for (int i = 0; i < nGroups_; i++) { | 
| 247 | < | int gid = cgLocalToGlobal[i]; | 
| 248 | < | for (int j = 0; j < nLocal_; j++) { | 
| 249 | < | int aid = AtomLocalToGlobal[j]; | 
| 250 | < | if (globalGroupMembership[aid] == gid) { | 
| 251 | < | groupList_[i].push_back(j); | 
| 252 | < | } | 
| 253 | < | } | 
| 254 | < | } | 
| 255 | < |  | 
| 236 | > | #else | 
| 237 |  | excludesForAtom.clear(); | 
| 238 |  | excludesForAtom.resize(nLocal_); | 
| 239 |  | toposForAtom.clear(); | 
| 247 |  | for (int j = 0; j < nLocal_; j++) { | 
| 248 |  | int jglob = AtomLocalToGlobal[j]; | 
| 249 |  |  | 
| 250 | < | if (excludes->hasPair(iglob, jglob)) | 
| 250 | > | if (excludes->hasPair(iglob, jglob)) | 
| 251 |  | excludesForAtom[i].push_back(j); | 
| 252 |  |  | 
| 253 | + |  | 
| 254 |  | if (oneTwo->hasPair(iglob, jglob)) { | 
| 255 |  | toposForAtom[i].push_back(j); | 
| 256 |  | topoDist[i].push_back(1); | 
| 267 |  | } | 
| 268 |  | } | 
| 269 |  | } | 
| 270 | < |  | 
| 270 | > | #endif | 
| 271 | > |  | 
| 272 | > | // allocate memory for the parallel objects | 
| 273 | > | atypesLocal.resize(nLocal_); | 
| 274 | > |  | 
| 275 | > | for (int i = 0; i < nLocal_; i++) | 
| 276 | > | atypesLocal[i] = ff_->getAtomType(idents[i]); | 
| 277 | > |  | 
| 278 | > | groupList_.clear(); | 
| 279 | > | groupList_.resize(nGroups_); | 
| 280 | > | for (int i = 0; i < nGroups_; i++) { | 
| 281 | > | int gid = cgLocalToGlobal[i]; | 
| 282 | > | for (int j = 0; j < nLocal_; j++) { | 
| 283 | > | int aid = AtomLocalToGlobal[j]; | 
| 284 | > | if (globalGroupMembership[aid] == gid) { | 
| 285 | > | groupList_[i].push_back(j); | 
| 286 | > | } | 
| 287 | > | } | 
| 288 | > | } | 
| 289 | > |  | 
| 290 | > |  | 
| 291 |  | createGtypeCutoffMap(); | 
| 292 |  |  | 
| 293 |  | } | 
| 685 |  | } | 
| 686 |  |  | 
| 687 |  | AtomPlanRealColumn->scatter(atomColData.skippedCharge, skch_tmp); | 
| 688 | < | for (int i = 0; i < ns; i++) | 
| 688 | > | for (int i = 0; i < ns; i++) | 
| 689 |  | snap_->atomData.skippedCharge[i] += skch_tmp[i]; | 
| 690 | + |  | 
| 691 |  | } | 
| 692 |  |  | 
| 693 |  | nLocal_ = snap_->getNumberOfAtoms(); | 
| 717 |  | pairwisePot[ii] = ploc2; | 
| 718 |  | } | 
| 719 |  |  | 
| 720 | + | for (int ii = 0; ii < N_INTERACTION_FAMILIES; ii++) { | 
| 721 | + | RealType ploc1 = embeddingPot[ii]; | 
| 722 | + | RealType ploc2 = 0.0; | 
| 723 | + | MPI::COMM_WORLD.Allreduce(&ploc1, &ploc2, 1, MPI::REALTYPE, MPI::SUM); | 
| 724 | + | embeddingPot[ii] = ploc2; | 
| 725 | + | } | 
| 726 | + |  | 
| 727 |  | #endif | 
| 728 |  |  | 
| 729 |  | } | 
| 865 |  | * field) must still be handled for these pairs. | 
| 866 |  | */ | 
| 867 |  | bool ForceMatrixDecomposition::excludeAtomPair(int atom1, int atom2) { | 
| 868 | < | int unique_id_2; | 
| 869 | < | #ifdef IS_MPI | 
| 870 | < | // in MPI, we have to look up the unique IDs for the row atom. | 
| 861 | < | unique_id_2 = AtomColToGlobal[atom2]; | 
| 862 | < | #else | 
| 863 | < | // in the normal loop, the atom numbers are unique | 
| 864 | < | unique_id_2 = atom2; | 
| 865 | < | #endif | 
| 868 | > |  | 
| 869 | > | // excludesForAtom was constructed to use row/column indices in the MPI | 
| 870 | > | // version, and to use local IDs in the non-MPI version: | 
| 871 |  |  | 
| 872 |  | for (vector<int>::iterator i = excludesForAtom[atom1].begin(); | 
| 873 |  | i != excludesForAtom[atom1].end(); ++i) { | 
| 874 | < | if ( (*i) == unique_id_2 ) return true; | 
| 874 | > | if ( (*i) == atom2 )  return true; | 
| 875 |  | } | 
| 876 |  |  | 
| 877 |  | return false; |