Actual source code: ex4f.F90
1: !
2: ! This introductory example illustrates running PETSc on a subset
3: ! of processes
4: !
5: ! -----------------------------------------------------------------------
7: program main
8: #include <petsc/finclude/petscsys.h>
9: use petscmpi ! or mpi or mpi_f08
10: use petscsys
11: implicit none
12: PetscErrorCode ierr
13: PetscMPIInt rank, size,grank,zero,two
14: PetscReal globalrank
16: ! We must call MPI_Init() first, making us, not PETSc, responsible for MPI
18: PetscCallMPIA(MPI_Init(ierr))
19: #if defined(PETSC_HAVE_ELEMENTAL)
20: PetscCallA(PetscElementalInitializePackage(ierr))
21: #endif
22: ! We can now change the communicator universe for PETSc
24: zero = 0
25: two = 2
26: PetscCallMPIA(MPI_Comm_rank(MPI_COMM_WORLD,rank,ierr))
27: PetscCallMPIA(MPI_Comm_split(MPI_COMM_WORLD,mod(rank,two),zero,PETSC_COMM_WORLD,ierr))
29: ! Every PETSc routine should begin with the PetscInitialize()
30: ! routine.
31: PetscCallA(PetscInitializeNoArguments(ierr))
33: ! The following MPI calls return the number of processes being used
34: ! and the rank of this process in the group.
36: PetscCallMPIA(MPI_Comm_size(PETSC_COMM_WORLD,size,ierr))
37: PetscCallMPIA(MPI_Comm_rank(PETSC_COMM_WORLD,rank,ierr))
39: ! Here we would like to print only one message that represents all
40: ! the processes in the group. Sleep so that IO from different ranks
41: ! don't get mixed up. Note this is not an ideal solution
42: PetscCallMPIA(MPI_Comm_rank(MPI_COMM_WORLD,grank,ierr))
43: globalrank = grank
44: PetscCallA(PetscSleep(globalrank,ierr))
45: if (rank .eq. 0) write(6,100) size,rank
46: 100 format('No of Procs = ',i4,' rank = ',i4)
48: ! Always call PetscFinalize() before exiting a program. This
49: ! routine - finalizes the PETSc libraries as well as MPI - provides
50: ! summary and diagnostic information if certain runtime options are
51: ! chosen (e.g., -log_view). See PetscFinalize() manpage for more
52: ! information.
54: PetscCallA(PetscFinalize(ierr))
55: PetscCallMPIA(MPI_Comm_free(PETSC_COMM_WORLD,ierr))
56: #if defined(PETSC_HAVE_ELEMENTAL)
57: PetscCallA(PetscElementalFinalizePackage(ierr))
58: #endif
60: ! Since we initialized MPI, we must call MPI_Finalize()
62: PetscCallMPIA(MPI_Finalize(ierr))
63: end
65: !/*TEST
66: !
67: ! test:
68: ! nsize: 5
69: ! filter: sort -b
70: ! filter_output: sort -b
71: ! requires: !cuda !saws
72: !
73: !TEST*/