| 1 |  | /* | 
| 2 | < | * Copyright (c) 2005 The University of Notre Dame. All Rights Reserved. | 
| 2 | > | * copyright (c) 2005 The University of Notre Dame. All Rights Reserved. | 
| 3 |  | * | 
| 4 |  | * The University of Notre Dame grants you ("Licensee") a | 
| 5 |  | * non-exclusive, royalty free, license to use, modify and | 
| 100 |  | #ifdef IS_MPI | 
| 101 |  | int streamSize; | 
| 102 |  | const int masterNode = 0; | 
| 103 | < | int commStatus; | 
| 103 | > |  | 
| 104 |  | if (worldRank == masterNode) { | 
| 105 | < | commStatus = MPI_Bcast(&mdFileVersion, 1, MPI_INT, masterNode, MPI_COMM_WORLD); | 
| 105 | > | MPI::COMM_WORLD.Bcast(&mdFileVersion, 1, MPI::INT, masterNode); | 
| 106 |  | #endif | 
| 107 |  | SimplePreprocessor preprocessor; | 
| 108 | < | preprocessor.preprocess(rawMetaDataStream, filename, startOfMetaDataBlock, ppStream); | 
| 108 | > | preprocessor.preprocess(rawMetaDataStream, filename, startOfMetaDataBlock, | 
| 109 | > | ppStream); | 
| 110 |  |  | 
| 111 |  | #ifdef IS_MPI | 
| 112 |  | //brocasting the stream size | 
| 113 |  | streamSize = ppStream.str().size() +1; | 
| 114 | < | commStatus = MPI_Bcast(&streamSize, 1, MPI_LONG, masterNode, MPI_COMM_WORLD); | 
| 115 | < |  | 
| 116 | < | commStatus = MPI_Bcast(static_cast<void*>(const_cast<char*>(ppStream.str().c_str())), streamSize, MPI_CHAR, masterNode, MPI_COMM_WORLD); | 
| 116 | < |  | 
| 114 | > | MPI::COMM_WORLD.Bcast(&streamSize, 1, MPI::LONG, masterNode); | 
| 115 | > | MPI::COMM_WORLD.Bcast(static_cast<void*>(const_cast<char*>(ppStream.str().c_str())), | 
| 116 | > | streamSize, MPI::CHAR, masterNode); | 
| 117 |  |  | 
| 118 |  | } else { | 
| 119 |  |  | 
| 120 | < | commStatus = MPI_Bcast(&mdFileVersion, 1, MPI_INT, masterNode, MPI_COMM_WORLD); | 
| 120 | > | MPI::COMM_WORLD.Bcast(&mdFileVersion, 1, MPI::INT, masterNode); | 
| 121 |  |  | 
| 122 |  | //get stream size | 
| 123 | < | commStatus = MPI_Bcast(&streamSize, 1, MPI_LONG, masterNode, MPI_COMM_WORLD); | 
| 123 | > | MPI::COMM_WORLD.Bcast(&streamSize, 1, MPI::LONG, masterNode); | 
| 124 |  |  | 
| 125 |  | char* buf = new char[streamSize]; | 
| 126 |  | assert(buf); | 
| 127 |  |  | 
| 128 |  | //receive file content | 
| 129 | < | commStatus = MPI_Bcast(buf, streamSize, MPI_CHAR, masterNode, MPI_COMM_WORLD); | 
| 129 | > | MPI::COMM_WORLD.Bcast(buf, streamSize, MPI::CHAR, masterNode); | 
| 130 |  |  | 
| 131 |  | ppStream.str(buf); | 
| 132 |  | delete [] buf; | 
| 255 |  | std::string mdRawData; | 
| 256 |  | int metaDataBlockStart = -1; | 
| 257 |  | int metaDataBlockEnd = -1; | 
| 258 | < | int i; | 
| 258 | > | int i, j; | 
| 259 |  | streamoff mdOffset(0); | 
| 260 |  | int mdFileVersion; | 
| 261 |  |  | 
| 262 | + | // Create a string for embedding the version information in the MetaData | 
| 263 | + | std::string version; | 
| 264 | + | version.assign("## Last run using OpenMD Version: "); | 
| 265 | + | version.append(OPENMD_VERSION_MAJOR); | 
| 266 | + | version.append("."); | 
| 267 | + | version.append(OPENMD_VERSION_MINOR); | 
| 268 |  |  | 
| 269 | + | std::string svnrev; | 
| 270 | + | //convert a macro from compiler to a string in c++ | 
| 271 | + | STR_DEFINE(svnrev, SVN_REV ); | 
| 272 | + | version.append(" Revision: "); | 
| 273 | + | // If there's no SVN revision, just call this the RELEASE revision. | 
| 274 | + | if (!svnrev.empty()) { | 
| 275 | + | version.append(svnrev); | 
| 276 | + | } else { | 
| 277 | + | version.append("RELEASE"); | 
| 278 | + | } | 
| 279 | + |  | 
| 280 |  | #ifdef IS_MPI | 
| 281 |  | const int masterNode = 0; | 
| 282 |  | if (worldRank == masterNode) { | 
| 371 |  |  | 
| 372 |  | mdRawData.clear(); | 
| 373 |  |  | 
| 374 | + | bool foundVersion = false; | 
| 375 | + |  | 
| 376 |  | for (int i = 0; i < metaDataBlockEnd - metaDataBlockStart - 1; ++i) { | 
| 377 |  | mdFile_.getline(buffer, bufferSize); | 
| 378 | < | mdRawData += buffer; | 
| 378 | > | std::string line = trimLeftCopy(buffer); | 
| 379 | > | j = CaseInsensitiveFind(line, "## Last run using OpenMD Version"); | 
| 380 | > | if (static_cast<size_t>(j) != string::npos) { | 
| 381 | > | foundVersion = true; | 
| 382 | > | mdRawData += version; | 
| 383 | > | } else { | 
| 384 | > | mdRawData += buffer; | 
| 385 | > | } | 
| 386 |  | mdRawData += "\n"; | 
| 387 |  | } | 
| 388 | < |  | 
| 388 | > |  | 
| 389 | > | if (!foundVersion) mdRawData += version + "\n"; | 
| 390 | > |  | 
| 391 |  | mdFile_.close(); | 
| 392 |  |  | 
| 393 |  | #ifdef IS_MPI | 
| 527 |  | int nTarget; | 
| 528 |  | int done; | 
| 529 |  | int i; | 
| 502 | – | int j; | 
| 530 |  | int loops; | 
| 531 |  | int which_proc; | 
| 532 |  | int nProcessors; | 
| 534 |  | int nGlobalMols = info->getNGlobalMolecules(); | 
| 535 |  | std::vector<int> molToProcMap(nGlobalMols, -1); // default to an error condition: | 
| 536 |  |  | 
| 537 | < | MPI_Comm_size(MPI_COMM_WORLD, &nProcessors); | 
| 537 | > | nProcessors = MPI::COMM_WORLD.Get_size(); | 
| 538 |  |  | 
| 539 |  | if (nProcessors > nGlobalMols) { | 
| 540 |  | sprintf(painCave.errMsg, | 
| 572 |  | nTarget = (int)(precast + 0.5); | 
| 573 |  |  | 
| 574 |  | for(i = 0; i < nGlobalMols; i++) { | 
| 575 | + |  | 
| 576 |  | done = 0; | 
| 577 |  | loops = 0; | 
| 578 |  |  | 
| 597 |  | // and be done with it. | 
| 598 |  |  | 
| 599 |  | if (loops > 100) { | 
| 600 | + |  | 
| 601 |  | sprintf(painCave.errMsg, | 
| 602 | < | "I've tried 100 times to assign molecule %d to a " | 
| 603 | < | " processor, but can't find a good spot.\n" | 
| 604 | < | "I'm assigning it at random to processor %d.\n", | 
| 602 | > | "There have been 100 attempts to assign molecule %d to an\n" | 
| 603 | > | "\tunderworked processor, but there's no good place to\n" | 
| 604 | > | "\tleave it.  OpenMD is assigning it at random to processor %d.\n", | 
| 605 |  | i, which_proc); | 
| 606 | < |  | 
| 606 | > |  | 
| 607 |  | painCave.isFatal = 0; | 
| 608 | + | painCave.severity = OPENMD_INFO; | 
| 609 |  | simError(); | 
| 610 |  |  | 
| 611 |  | molToProcMap[i] = which_proc; | 
| 650 |  | } | 
| 651 |  |  | 
| 652 |  | delete myRandom; | 
| 653 | < |  | 
| 653 | > |  | 
| 654 |  | // Spray out this nonsense to all other processors: | 
| 655 | < |  | 
| 626 | < | MPI_Bcast(&molToProcMap[0], nGlobalMols, MPI_INT, 0, MPI_COMM_WORLD); | 
| 655 | > | MPI::COMM_WORLD.Bcast(&molToProcMap[0], nGlobalMols, MPI::INT, 0); | 
| 656 |  | } else { | 
| 657 |  |  | 
| 658 |  | // Listen to your marching orders from processor 0: | 
| 659 | < |  | 
| 660 | < | MPI_Bcast(&molToProcMap[0], nGlobalMols, MPI_INT, 0, MPI_COMM_WORLD); | 
| 659 | > | MPI::COMM_WORLD.Bcast(&molToProcMap[0], nGlobalMols, MPI::INT, 0); | 
| 660 | > |  | 
| 661 |  | } | 
| 662 |  |  | 
| 663 |  | info->setMolToProcMap(molToProcMap); | 
| 815 |  | int beginRigidBodyIndex; | 
| 816 |  | int beginCutoffGroupIndex; | 
| 817 |  | int nGlobalAtoms = info->getNGlobalAtoms(); | 
| 818 | + | int nGlobalRigidBodies = info->getNGlobalRigidBodies(); | 
| 819 |  |  | 
| 820 |  | beginAtomIndex = 0; | 
| 821 |  | //rigidbody's index begins right after atom's | 
| 882 |  | // This would be prettier if we could use MPI_IN_PLACE like the MPI-2 | 
| 883 |  | // docs said we could. | 
| 884 |  | std::vector<int> tmpGroupMembership(info->getNGlobalAtoms(), 0); | 
| 885 | < | MPI_Allreduce(&globalGroupMembership[0], &tmpGroupMembership[0], nGlobalAtoms, | 
| 886 | < | MPI_INT, MPI_SUM, MPI_COMM_WORLD); | 
| 885 | > | MPI::COMM_WORLD.Allreduce(&globalGroupMembership[0], | 
| 886 | > | &tmpGroupMembership[0], nGlobalAtoms, | 
| 887 | > | MPI::INT, MPI::SUM); | 
| 888 |  | info->setGlobalGroupMembership(tmpGroupMembership); | 
| 889 |  | #else | 
| 890 |  | info->setGlobalGroupMembership(globalGroupMembership); | 
| 891 |  | #endif | 
| 892 |  |  | 
| 893 |  | //fill molMembership | 
| 894 | < | std::vector<int> globalMolMembership(info->getNGlobalAtoms(), 0); | 
| 894 | > | std::vector<int> globalMolMembership(info->getNGlobalAtoms() + | 
| 895 | > | info->getNGlobalRigidBodies(), 0); | 
| 896 |  |  | 
| 897 | < | for(mol = info->beginMolecule(mi); mol != NULL; mol = info->nextMolecule(mi)) { | 
| 897 | > | for(mol = info->beginMolecule(mi); mol != NULL; | 
| 898 | > | mol = info->nextMolecule(mi)) { | 
| 899 |  | for(atom = mol->beginAtom(ai); atom != NULL; atom = mol->nextAtom(ai)) { | 
| 900 |  | globalMolMembership[atom->getGlobalIndex()] = mol->getGlobalIndex(); | 
| 901 |  | } | 
| 902 | + | for (rb = mol->beginRigidBody(ri); rb != NULL; | 
| 903 | + | rb = mol->nextRigidBody(ri)) { | 
| 904 | + | globalMolMembership[rb->getGlobalIndex()] = mol->getGlobalIndex(); | 
| 905 | + | } | 
| 906 |  | } | 
| 907 |  |  | 
| 908 |  | #ifdef IS_MPI | 
| 909 | < | std::vector<int> tmpMolMembership(info->getNGlobalAtoms(), 0); | 
| 909 | > | std::vector<int> tmpMolMembership(info->getNGlobalAtoms() + | 
| 910 | > | info->getNGlobalRigidBodies(), 0); | 
| 911 | > | MPI::COMM_WORLD.Allreduce(&globalMolMembership[0], &tmpMolMembership[0], | 
| 912 | > | nGlobalAtoms + nGlobalRigidBodies, | 
| 913 | > | MPI::INT, MPI::SUM); | 
| 914 |  |  | 
| 874 | – | MPI_Allreduce(&globalMolMembership[0], &tmpMolMembership[0], nGlobalAtoms, | 
| 875 | – | MPI_INT, MPI_SUM, MPI_COMM_WORLD); | 
| 876 | – |  | 
| 915 |  | info->setGlobalMolMembership(tmpMolMembership); | 
| 916 |  | #else | 
| 917 |  | info->setGlobalMolMembership(globalMolMembership); | 
| 921 |  | // here the molecules are listed by their global indices. | 
| 922 |  |  | 
| 923 |  | std::vector<int> nIOPerMol(info->getNGlobalMolecules(), 0); | 
| 924 | < | for (mol = info->beginMolecule(mi); mol != NULL; mol = info->nextMolecule(mi)) { | 
| 924 | > | for (mol = info->beginMolecule(mi); mol != NULL; | 
| 925 | > | mol = info->nextMolecule(mi)) { | 
| 926 |  | nIOPerMol[mol->getGlobalIndex()] = mol->getNIntegrableObjects(); | 
| 927 |  | } | 
| 928 |  |  | 
| 929 |  | #ifdef IS_MPI | 
| 930 |  | std::vector<int> numIntegrableObjectsPerMol(info->getNGlobalMolecules(), 0); | 
| 931 | < | MPI_Allreduce(&nIOPerMol[0], &numIntegrableObjectsPerMol[0], | 
| 932 | < | info->getNGlobalMolecules(), MPI_INT, MPI_SUM, MPI_COMM_WORLD); | 
| 931 | > | MPI::COMM_WORLD.Allreduce(&nIOPerMol[0], &numIntegrableObjectsPerMol[0], | 
| 932 | > | info->getNGlobalMolecules(), MPI::INT, MPI::SUM); | 
| 933 |  | #else | 
| 934 |  | std::vector<int> numIntegrableObjectsPerMol = nIOPerMol; | 
| 935 |  | #endif | 
| 943 |  | } | 
| 944 |  |  | 
| 945 |  | std::vector<StuntDouble*> IOIndexToIntegrableObject(info->getNGlobalIntegrableObjects(), (StuntDouble*)NULL); | 
| 946 | < | for (mol = info->beginMolecule(mi); mol != NULL; mol = info->nextMolecule(mi)) { | 
| 946 | > | for (mol = info->beginMolecule(mi); mol != NULL; | 
| 947 | > | mol = info->nextMolecule(mi)) { | 
| 948 |  | int myGlobalIndex = mol->getGlobalIndex(); | 
| 949 |  | int globalIO = startingIOIndexForMol[myGlobalIndex]; | 
| 950 |  | for (StuntDouble* sd = mol->beginIntegrableObject(ioi); sd != NULL; | 
| 960 |  | } | 
| 961 |  |  | 
| 962 |  | void SimCreator::loadCoordinates(SimInfo* info, const std::string& mdFileName) { | 
| 923 | – | Globals* simParams; | 
| 963 |  |  | 
| 925 | – | simParams = info->getSimParams(); | 
| 926 | – |  | 
| 964 |  | DumpReader reader(info, mdFileName); | 
| 965 |  | int nframes = reader.getNFrames(); | 
| 966 |  |  |