Actual source code: ex8f.F90

  1: !
  2: ! Description: Demonstrates using a local ordering to set values into a parallel vector
  3: !

  5:   program main
  6: #include <petsc/finclude/petscvec.h>
  7:   use petscvec

  9:   implicit none

 11:   PetscErrorCode ierr
 12:   PetscMPIInt    rank
 13:   PetscInt    ::   i,ng,rstart,rend,M
 14:   PetscInt, pointer, dimension(:) :: gindices
 15:   PetscScalar, parameter :: sone = 1.0
 16:   Vec   ::         x
 17:   ISLocalToGlobalMapping :: ltog
 18:   PetscInt,parameter :: one = 1

 20:   PetscCallA(PetscInitialize(ierr))
 21:   PetscCallMPIA(MPI_Comm_rank(PETSC_COMM_WORLD,rank,ierr))

 23: !
 24: !     Create a parallel vector.
 25: !      - In this case, we specify the size of each processor's local
 26: !        portion, and PETSc computes the global size.  Alternatively,
 27: !        PETSc could determine the vector's distribution if we specify
 28: !        just the global size.
 29: !
 30:   PetscCallA(VecCreate(PETSC_COMM_WORLD,x,ierr))
 31:   PetscCallA(VecSetSizes(x,rank+one,PETSC_DECIDE,ierr))
 32:   PetscCallA(VecSetFromOptions(x,ierr))

 34:   PetscCallA(VecSet(x,sone,ierr))

 36: !
 37: !     Set the local to global ordering for the vector. Each processor
 38: !     generates a list of the global indices for each local index. Note that
 39: !     the local indices are just whatever is convenient for a particular application.
 40: !     In this case we treat the vector as lying on a one dimensional grid and
 41: !     have one ghost point on each end of the blocks owned by each processor.
 42: !

 44:   PetscCallA(VecGetSize(x,M,ierr))
 45:   PetscCallA(VecGetOwnershipRange(x,rstart,rend,ierr))
 46:   ng = rend - rstart + 2
 47:   allocate(gindices(0:ng-1))
 48:   gindices(0) = rstart -1

 50:   do i=0,ng-2
 51:    gindices(i+1) = gindices(i) + 1
 52:   end do

 54: ! map the first and last point as periodic

 56:   if (gindices(0) == -1) gindices(0) = M - 1

 58:   if (gindices(ng-1) == M) gindices(ng-1) = 0

 60:   PetscCallA(ISLocalToGlobalMappingCreate(PETSC_COMM_SELF,one,ng,gindices,PETSC_COPY_VALUES,ltog,ierr))
 61:   PetscCallA(VecSetLocalToGlobalMapping(x,ltog,ierr))
 62:   PetscCallA(ISLocalToGlobalMappingDestroy(ltog,ierr))
 63:   deallocate(gindices)

 65:      ! Set the vector elements.
 66:      ! - In this case set the values using the local ordering
 67:      ! - Each processor can contribute any vector entries,
 68:      !   regardless of which processor "owns" them; any nonlocal
 69:      !   contributions will be transferred to the appropriate processor
 70:      !   during the assembly process.
 71:      ! - In this example, the flag ADD_VALUES indicates that all
 72:      !   contributions will be added together.

 74:   do i=0,ng-1
 75:    PetscCallA(VecSetValuesLocal(x,one,i,sone,ADD_VALUES,ierr))
 76:   end do

 78:   !
 79:   ! Assemble vector, using the 2-step process:
 80:   ! VecAssemblyBegin(), VecAssemblyEnd()
 81:   ! Computations can be done while messages are in transition
 82:   ! by placing code between these two statements.
 83:   !
 84:   PetscCallA(VecAssemblyBegin(x,ierr))
 85:   PetscCallA(VecAssemblyEnd(x,ierr))
 86:   !
 87:   ! View the vector; then destroy it.
 88:   !
 89:   PetscCallA(VecView(x,PETSC_VIEWER_STDOUT_WORLD,ierr))
 90:   PetscCallA(VecDestroy(x,ierr))
 91:   PetscCallA(PetscFinalize(ierr))

 93: end program

 95: !/*TEST
 96: !
 97: !     test:
 98: !       nsize: 4
 99: !       output_file: output/ex8_1.out
100: !
101: !TEST*/