6 |
|
#include <mpi.h> |
7 |
|
#include <mpi++.h> |
8 |
|
#include "mpiSimulation.hpp" |
9 |
< |
#define TAKE_THIS_TAG 1 |
9 |
> |
#define TAKE_THIS_TAG_CHAR 1 |
10 |
> |
#define TAKE_THIS_TAG_INT 2 |
11 |
> |
|
12 |
> |
namespace dWrite{ |
13 |
> |
void nodeZeroError( void ); |
14 |
> |
void anonymousNodeDie( void ); |
15 |
> |
} |
16 |
> |
|
17 |
> |
using namespace dWrite; |
18 |
|
#endif //is_mpi |
19 |
|
|
20 |
|
#include "ReadWrite.hpp" |
128 |
|
|
129 |
|
#else // is_mpi |
130 |
|
|
131 |
+ |
// first thing first, suspend fatalities. |
132 |
+ |
painCave.isEventLoop = 1; |
133 |
+ |
|
134 |
+ |
int myStatus; // 1 = wakeup & success; 0 = error; -1 = AllDone |
135 |
+ |
int haveError; |
136 |
+ |
|
137 |
|
MPI::Status istatus; |
138 |
|
int *AtomToProcMap = mpiSim->getAtomToProcMap(); |
139 |
|
|
152 |
|
|
153 |
|
which_node = AtomToProcMap[i]; |
154 |
|
|
155 |
< |
if (which_node == mpiSim->getMyNode()) { |
142 |
< |
sprintf( tempBuffer, |
143 |
< |
"%s\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t", |
144 |
< |
atoms[i]->getType(), |
145 |
< |
atoms[i]->getX(), |
146 |
< |
atoms[i]->getY(), |
147 |
< |
atoms[i]->getZ(), |
148 |
< |
atoms[i]->get_vx(), |
149 |
< |
atoms[i]->get_vy(), |
150 |
< |
atoms[i]->get_vz()); |
151 |
< |
strcpy( writeLine, tempBuffer ); |
155 |
> |
if (which_node == 0 ) { |
156 |
|
|
157 |
< |
if( atoms[i]->isDirectional() ){ |
158 |
< |
|
159 |
< |
dAtom = (DirectionalAtom *)atoms[i]; |
160 |
< |
dAtom->getQ( q ); |
157 |
< |
|
158 |
< |
sprintf( tempBuffer, |
159 |
< |
"%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\n", |
160 |
< |
q[0], |
161 |
< |
q[1], |
162 |
< |
q[2], |
163 |
< |
q[3], |
164 |
< |
dAtom->getJx(), |
165 |
< |
dAtom->getJy(), |
166 |
< |
dAtom->getJz()); |
167 |
< |
strcat( writeLine, tempBuffer ); |
168 |
< |
} |
169 |
< |
else |
170 |
< |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
171 |
< |
|
172 |
< |
} else { |
173 |
< |
|
174 |
< |
std::cerr << "node 0: sending node " << which_node << " request for atom " << i << "\n"; |
175 |
< |
MPI::COMM_WORLD.Send(&i, 1, MPI_INT, which_node, TAKE_THIS_TAG); |
176 |
< |
std::cerr << "node 0: sent!\n"; |
177 |
< |
MPI::COMM_WORLD.Recv(writeLine, BUFFERSIZE, MPI_CHAR, which_node, |
178 |
< |
TAKE_THIS_TAG, istatus); |
179 |
< |
std::cerr << "node 0: got this line: " << writeLine; |
180 |
< |
} |
181 |
< |
|
182 |
< |
outFile << writeLine; |
183 |
< |
outFile.flush(); |
184 |
< |
} |
185 |
< |
|
186 |
< |
// kill everyone off: |
187 |
< |
game_over = -1; |
188 |
< |
for (j = 0; j < mpiSim->getNumberProcessors(); j++) { |
189 |
< |
MPI::COMM_WORLD.Send(&game_over, 1, MPI_INT, j, TAKE_THIS_TAG); |
190 |
< |
} |
191 |
< |
|
192 |
< |
} else { |
193 |
< |
|
194 |
< |
done = 0; |
195 |
< |
while (!done) { |
196 |
< |
std::cerr << "node: " << mpiSim->getMyNode() << " Waiting for receive \n"; |
197 |
< |
MPI::COMM_WORLD.Recv(&which_atom, 1, MPI_INT, 0, |
198 |
< |
TAKE_THIS_TAG, istatus); |
199 |
< |
std::cerr << "node: " << mpiSim->getMyNode() << " got request for atom " << which_atom << "\n"; |
200 |
< |
if (which_atom == -1) { |
201 |
< |
done=1; |
202 |
< |
continue; |
203 |
< |
} else { |
204 |
< |
local_index=-1; |
205 |
< |
for (j=0; j < mpiSim->getMyNlocal(); j++) { |
157 |
> |
haveError = 0; |
158 |
> |
which_atom = i; |
159 |
> |
local_index=-1; |
160 |
> |
for (j=0; (j<mpiSim->getMyNlocal()) && (local_index < 0); j++) { |
161 |
|
if (atoms[j]->getGlobalIndex() == which_atom) local_index = j; |
162 |
|
} |
163 |
|
if (local_index != -1) { |
188 |
|
dAtom->getJy(), |
189 |
|
dAtom->getJz()); |
190 |
|
strcat( writeLine, tempBuffer ); |
191 |
< |
} |
192 |
< |
else |
193 |
< |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
194 |
< |
std::cerr << "node: " << mpiSim->getMyNode() << " sending this line" << writeLine; |
195 |
< |
MPI::COMM_WORLD.Send(writeLine, BUFFERSIZE, MPI_CHAR, 0, |
196 |
< |
TAKE_THIS_TAG); |
197 |
< |
} else { |
198 |
< |
strcpy( writeLine, "ATOM NOT FOUND ON THIS PROCESSOR"); |
199 |
< |
MPI::COMM_WORLD.Send(writeLine, BUFFERSIZE, MPI_CHAR, 0, |
200 |
< |
TAKE_THIS_TAG); |
201 |
< |
} |
191 |
> |
|
192 |
> |
} |
193 |
> |
else |
194 |
> |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
195 |
> |
} |
196 |
> |
else { |
197 |
> |
sprintf(painCave.errMsg, |
198 |
> |
"Atom %d not found on processor %d\n", |
199 |
> |
i, worldRank ); |
200 |
> |
haveError= 1; |
201 |
> |
simError(); |
202 |
> |
} |
203 |
> |
|
204 |
> |
if(haveError) nodeZeroError(); |
205 |
> |
|
206 |
> |
} |
207 |
> |
else { |
208 |
> |
myStatus = 1; |
209 |
> |
MPI::COMM_WORLD.Send(&myStatus, 1, MPI_INT, which_node, |
210 |
> |
TAKE_THIS_TAG_INT); |
211 |
> |
MPI::COMM_WORLD.Send(&i, 1, MPI_INT, which_node, TAKE_THIS_TAG_INT); |
212 |
> |
MPI::COMM_WORLD.Recv(writeLine, BUFFERSIZE, MPI_CHAR, which_node, |
213 |
> |
TAKE_THIS_TAG_CHAR, istatus); |
214 |
> |
MPI::COMM_WORLD.Recv(&myStatus, 1, MPI_INT, which_node, |
215 |
> |
TAKE_THIS_TAG_INT, istatus); |
216 |
> |
|
217 |
> |
if(!myStatus) nodeZeroError(); |
218 |
> |
|
219 |
|
} |
220 |
+ |
|
221 |
+ |
outFile << writeLine; |
222 |
+ |
outFile.flush(); |
223 |
|
} |
224 |
+ |
|
225 |
+ |
// kill everyone off: |
226 |
+ |
myStatus = -1; |
227 |
+ |
for (j = 0; j < mpiSim->getNumberProcessors(); j++) { |
228 |
+ |
MPI::COMM_WORLD.Send(&myStatus, 1, MPI_INT, j, |
229 |
+ |
TAKE_THIS_TAG_INT); |
230 |
+ |
} |
231 |
+ |
|
232 |
+ |
} else { |
233 |
+ |
|
234 |
+ |
done = 0; |
235 |
+ |
while (!done) { |
236 |
+ |
|
237 |
+ |
MPI::COMM_WORLD.Recv(&myStatus, 1, MPI_INT, 0, |
238 |
+ |
TAKE_THIS_TAG_INT, istatus); |
239 |
+ |
|
240 |
+ |
if(!myStatus) anonymousNodeDie(); |
241 |
+ |
|
242 |
+ |
if(myStatus < 0) break; |
243 |
+ |
|
244 |
+ |
MPI::COMM_WORLD.Recv(&which_atom, 1, MPI_INT, 0, |
245 |
+ |
TAKE_THIS_TAG_INT, istatus); |
246 |
+ |
|
247 |
+ |
myStatus = 1; |
248 |
+ |
local_index=-1; |
249 |
+ |
for (j=0; (j<mpiSim->getMyNlocal()) && (local_index < 0); j++) { |
250 |
+ |
if (atoms[j]->getGlobalIndex() == which_atom) local_index = j; |
251 |
+ |
} |
252 |
+ |
if (local_index != -1) { |
253 |
+ |
//format the line |
254 |
+ |
sprintf( tempBuffer, |
255 |
+ |
"%s\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t", |
256 |
+ |
atoms[local_index]->getType(), |
257 |
+ |
atoms[local_index]->getX(), |
258 |
+ |
atoms[local_index]->getY(), |
259 |
+ |
atoms[local_index]->getZ(), |
260 |
+ |
atoms[local_index]->get_vx(), |
261 |
+ |
atoms[local_index]->get_vy(), |
262 |
+ |
atoms[local_index]->get_vz()); // check here. |
263 |
+ |
strcpy( writeLine, tempBuffer ); |
264 |
+ |
|
265 |
+ |
if( atoms[local_index]->isDirectional() ){ |
266 |
+ |
|
267 |
+ |
dAtom = (DirectionalAtom *)atoms[local_index]; |
268 |
+ |
dAtom->getQ( q ); |
269 |
+ |
|
270 |
+ |
sprintf( tempBuffer, |
271 |
+ |
"%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\n", |
272 |
+ |
q[0], |
273 |
+ |
q[1], |
274 |
+ |
q[2], |
275 |
+ |
q[3], |
276 |
+ |
dAtom->getJx(), |
277 |
+ |
dAtom->getJy(), |
278 |
+ |
dAtom->getJz()); |
279 |
+ |
strcat( writeLine, tempBuffer ); |
280 |
+ |
} |
281 |
+ |
else{ |
282 |
+ |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
283 |
+ |
} |
284 |
+ |
} |
285 |
+ |
else { |
286 |
+ |
sprintf(painCave.errMsg, |
287 |
+ |
"Atom %d not found on processor %d\n", |
288 |
+ |
which_atom, worldRank ); |
289 |
+ |
myStatus = 0; |
290 |
+ |
simError(); |
291 |
+ |
|
292 |
+ |
strcpy( writeLine, "Hello, I'm an error.\n"); |
293 |
+ |
} |
294 |
+ |
|
295 |
+ |
MPI::COMM_WORLD.Send(writeLine, BUFFERSIZE, MPI_CHAR, 0, |
296 |
+ |
TAKE_THIS_TAG_CHAR); |
297 |
+ |
MPI::COMM_WORLD.Send( &myStatus, 1, MPI_INT, 0, |
298 |
+ |
TAKE_THIS_TAG_INT); |
299 |
+ |
} |
300 |
|
} |
301 |
|
outFile.flush(); |
302 |
|
sprintf( checkPointMsg, |
303 |
|
"Sucessfully took a dump.\n"); |
304 |
|
MPIcheckPoint(); |
305 |
+ |
|
306 |
+ |
// last thing last, enable fatalities. |
307 |
+ |
painCave.isEventLoop = 0; |
308 |
+ |
|
309 |
|
#endif // is_mpi |
310 |
|
} |
311 |
|
|
398 |
|
|
399 |
|
#else // is_mpi |
400 |
|
|
401 |
+ |
// first thing first, suspend fatalities. |
402 |
+ |
painCave.isEventLoop = 1; |
403 |
+ |
|
404 |
+ |
int myStatus; // 1 = wakeup & success; 0 = error; -1 = AllDone |
405 |
+ |
int haveError; |
406 |
+ |
|
407 |
|
MPI::Status istatus; |
408 |
|
int *AtomToProcMap = mpiSim->getAtomToProcMap(); |
409 |
|
|
410 |
|
// write out header and node 0's coordinates |
411 |
|
|
412 |
+ |
haveError = 0; |
413 |
|
if( worldRank == 0 ){ |
414 |
|
finalOut << mpiSim->getTotAtoms() << "\n"; |
415 |
|
|
423 |
|
which_node = AtomToProcMap[i]; |
424 |
|
|
425 |
|
if (which_node == mpiSim->getMyNode()) { |
426 |
< |
|
427 |
< |
sprintf( tempBuffer, |
428 |
< |
"%s\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t", |
429 |
< |
atoms[i]->getType(), |
430 |
< |
atoms[i]->getX(), |
369 |
< |
atoms[i]->getY(), |
370 |
< |
atoms[i]->getZ(), |
371 |
< |
atoms[i]->get_vx(), |
372 |
< |
atoms[i]->get_vy(), |
373 |
< |
atoms[i]->get_vz()); |
374 |
< |
strcpy( writeLine, tempBuffer ); |
375 |
< |
|
376 |
< |
if( atoms[i]->isDirectional() ){ |
377 |
< |
|
378 |
< |
dAtom = (DirectionalAtom *)atoms[i]; |
379 |
< |
dAtom->getQ( q ); |
380 |
< |
|
381 |
< |
sprintf( tempBuffer, |
382 |
< |
"%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\n", |
383 |
< |
q[0], |
384 |
< |
q[1], |
385 |
< |
q[2], |
386 |
< |
q[3], |
387 |
< |
dAtom->getJx(), |
388 |
< |
dAtom->getJy(), |
389 |
< |
dAtom->getJz()); |
390 |
< |
strcat( writeLine, tempBuffer ); |
426 |
> |
|
427 |
> |
which_atom = i; |
428 |
> |
local_index=-1; |
429 |
> |
for (j=0; (j<mpiSim->getMyNlocal()) && (local_index < 0); j++) { |
430 |
> |
if (atoms[j]->getGlobalIndex() == which_atom) local_index = j; |
431 |
|
} |
432 |
< |
else |
433 |
< |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
432 |
> |
if (local_index != -1) { |
433 |
> |
sprintf( tempBuffer, |
434 |
> |
"%s\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t", |
435 |
> |
atoms[local_index]->getType(), |
436 |
> |
atoms[local_index]->getX(), |
437 |
> |
atoms[local_index]->getY(), |
438 |
> |
atoms[local_index]->getZ(), |
439 |
> |
atoms[local_index]->get_vx(), |
440 |
> |
atoms[local_index]->get_vy(), |
441 |
> |
atoms[local_index]->get_vz()); |
442 |
> |
strcpy( writeLine, tempBuffer ); |
443 |
> |
|
444 |
> |
if( atoms[local_index]->isDirectional() ){ |
445 |
> |
|
446 |
> |
dAtom = (DirectionalAtom *)atoms[local_index]; |
447 |
> |
dAtom->getQ( q ); |
448 |
> |
|
449 |
> |
sprintf( tempBuffer, |
450 |
> |
"%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\n", |
451 |
> |
q[0], |
452 |
> |
q[1], |
453 |
> |
q[2], |
454 |
> |
q[3], |
455 |
> |
dAtom->getJx(), |
456 |
> |
dAtom->getJy(), |
457 |
> |
dAtom->getJz()); |
458 |
> |
strcat( writeLine, tempBuffer ); |
459 |
> |
} |
460 |
> |
else |
461 |
> |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
462 |
> |
} |
463 |
> |
else { |
464 |
> |
sprintf(painCave.errMsg, |
465 |
> |
"Atom %d not found on processor %d\n", |
466 |
> |
i, worldRank ); |
467 |
> |
haveError= 1; |
468 |
> |
simError(); |
469 |
> |
} |
470 |
> |
|
471 |
> |
if(haveError) nodeZeroError(); |
472 |
> |
|
473 |
> |
} |
474 |
> |
else { |
475 |
|
|
476 |
< |
} else { |
477 |
< |
|
478 |
< |
MPI::COMM_WORLD.Send(&i, 1, MPI_INT, which_node, TAKE_THIS_TAG); |
476 |
> |
myStatus = 1; |
477 |
> |
MPI::COMM_WORLD.Send(&myStatus, 1, MPI_INT, which_node, |
478 |
> |
TAKE_THIS_TAG_INT); |
479 |
> |
MPI::COMM_WORLD.Send(&i, 1, MPI_INT, which_node, TAKE_THIS_TAG_INT); |
480 |
|
MPI::COMM_WORLD.Recv(writeLine, BUFFERSIZE, MPI_CHAR, which_node, |
481 |
< |
TAKE_THIS_TAG, istatus); |
481 |
> |
TAKE_THIS_TAG_CHAR, istatus); |
482 |
> |
MPI::COMM_WORLD.Recv(&myStatus, 1, MPI_INT, which_node, |
483 |
> |
TAKE_THIS_TAG_INT, istatus); |
484 |
> |
|
485 |
> |
if(!myStatus) nodeZeroError(); |
486 |
|
} |
487 |
|
|
488 |
|
finalOut << writeLine; |
489 |
|
} |
490 |
|
|
491 |
|
// kill everyone off: |
492 |
< |
game_over = -1; |
493 |
< |
for (j = 0; j < mpiSim->getNumberProcessors(); j++) { |
494 |
< |
MPI::COMM_WORLD.Send(&game_over, 1, MPI_INT, j, TAKE_THIS_TAG); |
492 |
> |
myStatus = -1; |
493 |
> |
for (j = 0; j < mpiSim->getNumberProcessors(); j++) { |
494 |
> |
MPI::COMM_WORLD.Send(&myStatus, 1, MPI_INT, j, |
495 |
> |
TAKE_THIS_TAG_INT); |
496 |
|
} |
497 |
|
|
498 |
|
} else { |
499 |
|
|
500 |
|
done = 0; |
501 |
|
while (!done) { |
502 |
+ |
|
503 |
+ |
MPI::COMM_WORLD.Recv(&myStatus, 1, MPI_INT, 0, |
504 |
+ |
TAKE_THIS_TAG_INT, istatus); |
505 |
+ |
|
506 |
+ |
if(!myStatus) anonymousNodeDie(); |
507 |
+ |
|
508 |
+ |
if(myStatus < 0) break; |
509 |
+ |
|
510 |
|
MPI::COMM_WORLD.Recv(&which_atom, 1, MPI_INT, 0, |
511 |
< |
TAKE_THIS_TAG, istatus); |
511 |
> |
TAKE_THIS_TAG_INT, istatus); |
512 |
> |
|
513 |
> |
myStatus = 1; |
514 |
> |
local_index=-1; |
515 |
> |
for (j=0; j < mpiSim->getMyNlocal(); j++) { |
516 |
> |
if (atoms[j]->getGlobalIndex() == which_atom) local_index = j; |
517 |
> |
} |
518 |
> |
if (local_index != -1) { |
519 |
|
|
520 |
< |
if (which_atom == -1) { |
521 |
< |
done=1; |
522 |
< |
continue; |
523 |
< |
} else { |
524 |
< |
|
525 |
< |
local_index=-1; |
526 |
< |
for (j=0; j < mpiSim->getMyNlocal(); j++) { |
527 |
< |
if (atoms[j]->getGlobalIndex() == which_atom) local_index = j; |
528 |
< |
} |
529 |
< |
if (local_index != -1) { |
530 |
< |
|
531 |
< |
//format the line |
532 |
< |
sprintf( tempBuffer, |
431 |
< |
"%s\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t", |
432 |
< |
atoms[local_index]->getType(), |
433 |
< |
atoms[local_index]->getX(), |
434 |
< |
atoms[local_index]->getY(), |
435 |
< |
atoms[local_index]->getZ(), |
436 |
< |
atoms[local_index]->get_vx(), |
437 |
< |
atoms[local_index]->get_vy(), |
438 |
< |
atoms[local_index]->get_vz()); // check here. |
439 |
< |
strcpy( writeLine, tempBuffer ); |
520 |
> |
//format the line |
521 |
> |
sprintf( tempBuffer, |
522 |
> |
"%s\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t", |
523 |
> |
atoms[local_index]->getType(), |
524 |
> |
atoms[local_index]->getX(), |
525 |
> |
atoms[local_index]->getY(), |
526 |
> |
atoms[local_index]->getZ(), |
527 |
> |
atoms[local_index]->get_vx(), |
528 |
> |
atoms[local_index]->get_vy(), |
529 |
> |
atoms[local_index]->get_vz()); // check here. |
530 |
> |
strcpy( writeLine, tempBuffer ); |
531 |
> |
|
532 |
> |
if( atoms[local_index]->isDirectional() ){ |
533 |
|
|
534 |
< |
if( atoms[local_index]->isDirectional() ){ |
535 |
< |
|
536 |
< |
dAtom = (DirectionalAtom *)atoms[local_index]; |
537 |
< |
dAtom->getQ( q ); |
538 |
< |
|
539 |
< |
sprintf( tempBuffer, |
540 |
< |
"%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\n", |
541 |
< |
q[0], |
542 |
< |
q[1], |
543 |
< |
q[2], |
544 |
< |
q[3], |
545 |
< |
dAtom->getJx(), |
546 |
< |
dAtom->getJy(), |
547 |
< |
dAtom->getJz()); |
548 |
< |
strcat( writeLine, tempBuffer ); |
549 |
< |
} |
550 |
< |
else |
551 |
< |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
552 |
< |
|
553 |
< |
MPI::COMM_WORLD.Send(writeLine, BUFFERSIZE, MPI_CHAR, 0, |
554 |
< |
TAKE_THIS_TAG); |
555 |
< |
} else { |
556 |
< |
strcpy( writeLine, "ATOM NOT FOUND ON THIS PROCESSOR"); |
557 |
< |
MPI::COMM_WORLD.Send(writeLine, BUFFERSIZE, MPI_CHAR, 0, |
558 |
< |
TAKE_THIS_TAG); |
559 |
< |
} |
534 |
> |
dAtom = (DirectionalAtom *)atoms[local_index]; |
535 |
> |
dAtom->getQ( q ); |
536 |
> |
|
537 |
> |
sprintf( tempBuffer, |
538 |
> |
"%lf\t%lf\t%lf\t%lf\t%lf\t%lf\t%lf\n", |
539 |
> |
q[0], |
540 |
> |
q[1], |
541 |
> |
q[2], |
542 |
> |
q[3], |
543 |
> |
dAtom->getJx(), |
544 |
> |
dAtom->getJy(), |
545 |
> |
dAtom->getJz()); |
546 |
> |
strcat( writeLine, tempBuffer ); |
547 |
> |
} |
548 |
> |
else{ |
549 |
> |
strcat( writeLine, "0.0\t0.0\t0.0\t0.0\t0.0\t0.0\t0.0\n" ); |
550 |
> |
} |
551 |
> |
} |
552 |
> |
else { |
553 |
> |
sprintf(painCave.errMsg, |
554 |
> |
"Atom %d not found on processor %d\n", |
555 |
> |
which_atom, worldRank ); |
556 |
> |
myStatus = 0; |
557 |
> |
simError(); |
558 |
> |
|
559 |
> |
strcpy( writeLine, "Hello, I'm an error.\n"); |
560 |
|
} |
561 |
+ |
|
562 |
+ |
MPI::COMM_WORLD.Send(writeLine, BUFFERSIZE, MPI_CHAR, 0, |
563 |
+ |
TAKE_THIS_TAG_CHAR); |
564 |
+ |
MPI::COMM_WORLD.Send( &myStatus, 1, MPI_INT, 0, |
565 |
+ |
TAKE_THIS_TAG_INT); |
566 |
|
} |
567 |
|
} |
568 |
|
finalOut.flush(); |
569 |
|
sprintf( checkPointMsg, |
570 |
|
"Sucessfully took a dump.\n"); |
571 |
|
MPIcheckPoint(); |
572 |
< |
|
572 |
> |
|
573 |
|
if( worldRank == 0 ) finalOut.close(); |
574 |
|
#endif // is_mpi |
575 |
|
} |
576 |
+ |
|
577 |
+ |
|
578 |
+ |
|
579 |
+ |
#ifdef IS_MPI |
580 |
+ |
|
581 |
+ |
// a couple of functions to let us escape the write loop |
582 |
+ |
|
583 |
+ |
void dWrite::nodeZeroError( void ){ |
584 |
+ |
int j, myStatus; |
585 |
+ |
|
586 |
+ |
myStatus = 0; |
587 |
+ |
for (j = 0; j < mpiSim->getNumberProcessors(); j++) { |
588 |
+ |
MPI::COMM_WORLD.Send( &myStatus, 1, MPI_INT, j, |
589 |
+ |
TAKE_THIS_TAG_INT); |
590 |
+ |
} |
591 |
+ |
|
592 |
+ |
|
593 |
+ |
MPI_Finalize(); |
594 |
+ |
exit (0); |
595 |
+ |
|
596 |
+ |
} |
597 |
+ |
|
598 |
+ |
void dWrite::anonymousNodeDie( void ){ |
599 |
+ |
|
600 |
+ |
MPI_Finalize(); |
601 |
+ |
exit (0); |
602 |
+ |
} |
603 |
+ |
|
604 |
+ |
#endif //is_mpi |