Actual source code: gasm.c

  1: /*
  2:   This file defines an "generalized" additive Schwarz preconditioner for any Mat implementation.
  3:   In this version, each MPI process may intersect multiple subdomains and any subdomain may
  4:   intersect multiple MPI processes.  Intersections of subdomains with MPI processes are called *local
  5:   subdomains*.

  7:        N    - total number of distinct global subdomains  (set explicitly in PCGASMSetTotalSubdomains() or implicitly PCGASMSetSubdomains() and then calculated in PCSetUp_GASM())
  8:        n    - actual number of local subdomains on this process (set in `PCGASMSetSubdomains()` or calculated in `PCGASMSetTotalSubdomains()`)
  9:        nmax - maximum number of local subdomains per process (calculated in PCSetUp_GASM())
 10: */
 11: #include <petsc/private/pcimpl.h>
 12: #include <petscdm.h>

 14: typedef struct {
 15:   PetscInt   N, n, nmax;
 16:   PetscInt   overlap;                /* overlap requested by user */
 17:   PCGASMType type;                   /* use reduced interpolation, restriction or both */
 18:   PetscBool  type_set;               /* if user set this value (so won't change it for symmetric problems) */
 19:   PetscBool  same_subdomain_solvers; /* flag indicating whether all local solvers are same */
 20:   PetscBool  sort_indices;           /* flag to sort subdomain indices */
 21:   PetscBool  user_subdomains;        /* whether the user set explicit subdomain index sets -- keep them on PCReset() */
 22:   PetscBool  dm_subdomains;          /* whether DM is allowed to define subdomains */
 23:   PetscBool  hierarchicalpartitioning;
 24:   IS        *ois;           /* index sets that define the outer (conceptually, overlapping) subdomains */
 25:   IS        *iis;           /* index sets that define the inner (conceptually, nonoverlapping) subdomains */
 26:   KSP       *ksp;           /* linear solvers for each subdomain */
 27:   Mat       *pmat;          /* subdomain block matrices */
 28:   Vec        gx, gy;        /* Merged work vectors */
 29:   Vec       *x, *y;         /* Split work vectors; storage aliases pieces of storage of the above merged vectors. */
 30:   VecScatter gorestriction; /* merged restriction to disjoint union of outer subdomains */
 31:   VecScatter girestriction; /* merged restriction to disjoint union of inner subdomains */
 32:   VecScatter pctoouter;
 33:   IS         permutationIS;
 34:   Mat        permutationP;
 35:   Mat        pcmat;
 36:   Vec        pcx, pcy;
 37: } PC_GASM;

 39: static PetscErrorCode PCGASMComputeGlobalSubdomainNumbering_Private(PC pc, PetscInt **numbering, PetscInt **permutation)
 40: {
 41:   PC_GASM *osm = (PC_GASM *)pc->data;
 42:   PetscInt i;

 44:   PetscFunctionBegin;
 45:   /* Determine the number of globally-distinct subdomains and compute a global numbering for them. */
 46:   PetscCall(PetscMalloc2(osm->n, numbering, osm->n, permutation));
 47:   PetscCall(PetscObjectsListGetGlobalNumbering(PetscObjectComm((PetscObject)pc), osm->n, (PetscObject *)osm->iis, NULL, *numbering));
 48:   for (i = 0; i < osm->n; ++i) (*permutation)[i] = i;
 49:   PetscCall(PetscSortIntWithPermutation(osm->n, *numbering, *permutation));
 50:   PetscFunctionReturn(PETSC_SUCCESS);
 51: }

 53: static PetscErrorCode PCGASMSubdomainView_Private(PC pc, PetscInt i, PetscViewer viewer)
 54: {
 55:   PC_GASM        *osm = (PC_GASM *)pc->data;
 56:   PetscInt        j, nidx;
 57:   const PetscInt *idx;
 58:   PetscViewer     sviewer;
 59:   char           *cidx;

 61:   PetscFunctionBegin;
 62:   PetscCheck(i >= -1 && i < osm->n, PetscObjectComm((PetscObject)viewer), PETSC_ERR_ARG_WRONG, "Invalid subdomain %" PetscInt_FMT ": must nonnegative and less than %" PetscInt_FMT, i, osm->n);

 64:   /* Inner subdomains. */
 65:   /*
 66:    No more than 15 characters per index plus a space.
 67:    PetscViewerStringSPrintf requires a string of size at least 2, so use (nidx+1) instead of nidx,
 68:    in case nidx == 0. That will take care of the space for the trailing '\0' as well.
 69:    For nidx == 0, the whole string 16 '\0'.
 70:    */
 71:   PetscCall(PetscViewerASCIIPrintf(viewer, "Inner subdomain:\n"));
 72:   PetscCall(PetscViewerFlush(viewer));
 73:   PetscCall(PetscViewerASCIIPushSynchronized(viewer));
 74:   if (i > -1) {
 75:     PetscCall(ISGetLocalSize(osm->iis[i], &nidx));
 76:     PetscCall(PetscMalloc1(16 * (nidx + 1) + 1, &cidx));
 77:     PetscCall(PetscViewerStringOpen(PETSC_COMM_SELF, cidx, 16 * (nidx + 1) + 1, &sviewer));
 78:     PetscCall(ISGetIndices(osm->iis[i], &idx));
 79:     for (j = 0; j < nidx; ++j) PetscCall(PetscViewerStringSPrintf(sviewer, "%" PetscInt_FMT " ", idx[j]));
 80:     PetscCall(ISRestoreIndices(osm->iis[i], &idx));
 81:     PetscCall(PetscViewerDestroy(&sviewer));
 82:     PetscCall(PetscViewerASCIISynchronizedPrintf(viewer, "%s", cidx));
 83:     PetscCall(PetscFree(cidx));
 84:   }
 85:   PetscCall(PetscViewerFlush(viewer));
 86:   PetscCall(PetscViewerASCIIPopSynchronized(viewer));
 87:   PetscCall(PetscViewerASCIIPrintf(viewer, "\n"));
 88:   PetscCall(PetscViewerFlush(viewer));

 90:   /* Outer subdomains. */
 91:   /*
 92:    No more than 15 characters per index plus a space.
 93:    PetscViewerStringSPrintf requires a string of size at least 2, so use (nidx+1) instead of nidx,
 94:    in case nidx == 0. That will take care of the space for the trailing '\0' as well.
 95:    For nidx == 0, the whole string 16 '\0'.
 96:    */
 97:   PetscCall(PetscViewerASCIIPrintf(viewer, "Outer subdomain:\n"));
 98:   PetscCall(PetscViewerFlush(viewer));
 99:   PetscCall(PetscViewerASCIIPushSynchronized(viewer));
100:   if (i > -1) {
101:     PetscCall(ISGetLocalSize(osm->ois[i], &nidx));
102:     PetscCall(PetscMalloc1(16 * (nidx + 1) + 1, &cidx));
103:     PetscCall(PetscViewerStringOpen(PETSC_COMM_SELF, cidx, 16 * (nidx + 1) + 1, &sviewer));
104:     PetscCall(ISGetIndices(osm->ois[i], &idx));
105:     for (j = 0; j < nidx; ++j) PetscCall(PetscViewerStringSPrintf(sviewer, "%" PetscInt_FMT " ", idx[j]));
106:     PetscCall(PetscViewerDestroy(&sviewer));
107:     PetscCall(ISRestoreIndices(osm->ois[i], &idx));
108:     PetscCall(PetscViewerASCIISynchronizedPrintf(viewer, "%s", cidx));
109:     PetscCall(PetscFree(cidx));
110:   }
111:   PetscCall(PetscViewerFlush(viewer));
112:   PetscCall(PetscViewerASCIIPopSynchronized(viewer));
113:   PetscCall(PetscViewerASCIIPrintf(viewer, "\n"));
114:   PetscCall(PetscViewerFlush(viewer));
115:   PetscFunctionReturn(PETSC_SUCCESS);
116: }

118: static PetscErrorCode PCGASMPrintSubdomains(PC pc)
119: {
120:   PC_GASM    *osm = (PC_GASM *)pc->data;
121:   const char *prefix;
122:   char        fname[PETSC_MAX_PATH_LEN + 1];
123:   PetscInt    l, d, count;
124:   PetscBool   found;
125:   PetscViewer viewer;
126:   PetscInt   *numbering, *permutation; /* global numbering of locally-supported subdomains and the permutation from the local ordering */

128:   PetscFunctionBegin;
129:   PetscCall(PCGetOptionsPrefix(pc, &prefix));
130:   PetscCall(PetscOptionsHasName(NULL, prefix, "-pc_gasm_print_subdomains", &found));
131:   if (!found) PetscFunctionReturn(PETSC_SUCCESS);
132:   PetscCall(PetscOptionsGetString(NULL, prefix, "-pc_gasm_print_subdomains", fname, sizeof(fname), &found));
133:   if (!found) PetscCall(PetscStrncpy(fname, "stdout", sizeof(fname)));
134:   PetscCall(PetscViewerASCIIOpen(PetscObjectComm((PetscObject)pc), fname, &viewer));
135:   /*
136:    Make sure the viewer has a name. Otherwise this may cause a deadlock or other weird errors when creating a subcomm viewer:
137:    the subcomm viewer will attempt to inherit the viewer's name, which, if not set, will be constructed collectively on the comm.
138:   */
139:   PetscCall(PetscObjectName((PetscObject)viewer));
140:   l = 0;
141:   PetscCall(PCGASMComputeGlobalSubdomainNumbering_Private(pc, &numbering, &permutation));
142:   for (count = 0; count < osm->N; ++count) {
143:     /* Now let subdomains go one at a time in the global numbering order and print their subdomain/solver info. */
144:     if (l < osm->n) {
145:       d = permutation[l]; /* d is the local number of the l-th smallest (in the global ordering) among the locally supported subdomains */
146:       if (numbering[d] == count) l++;
147:       else d = -1;
148:     } else d = -1;
149:     PetscCall(PCGASMSubdomainView_Private(pc, d, viewer));
150:   }
151:   PetscCall(PetscFree2(numbering, permutation));
152:   PetscCall(PetscViewerDestroy(&viewer));
153:   PetscFunctionReturn(PETSC_SUCCESS);
154: }

156: static PetscErrorCode PCView_GASM(PC pc, PetscViewer viewer)
157: {
158:   PC_GASM    *osm = (PC_GASM *)pc->data;
159:   const char *prefix;
160:   PetscMPIInt rank, size;
161:   PetscInt    bsz;
162:   PetscBool   iascii, view_subdomains = PETSC_FALSE;
163:   PetscViewer sviewer;
164:   PetscInt    count, l;
165:   char        overlap[256]     = "user-defined overlap";
166:   char        gsubdomains[256] = "unknown total number of subdomains";
167:   char        msubdomains[256] = "unknown max number of local subdomains";
168:   PetscInt   *numbering, *permutation; /* global numbering of locally-supported subdomains and the permutation from the local ordering */

170:   PetscFunctionBegin;
171:   PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)pc), &size));
172:   PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)pc), &rank));

174:   if (osm->overlap >= 0) PetscCall(PetscSNPrintf(overlap, sizeof(overlap), "requested amount of overlap = %" PetscInt_FMT, osm->overlap));
175:   if (osm->N != PETSC_DETERMINE) PetscCall(PetscSNPrintf(gsubdomains, sizeof(gsubdomains), "total number of subdomains = %" PetscInt_FMT, osm->N));
176:   if (osm->nmax != PETSC_DETERMINE) PetscCall(PetscSNPrintf(msubdomains, sizeof(msubdomains), "max number of local subdomains = %" PetscInt_FMT, osm->nmax));

178:   PetscCall(PCGetOptionsPrefix(pc, &prefix));
179:   PetscCall(PetscOptionsGetBool(NULL, prefix, "-pc_gasm_view_subdomains", &view_subdomains, NULL));

181:   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERASCII, &iascii));
182:   if (iascii) {
183:     /*
184:      Make sure the viewer has a name. Otherwise this may cause a deadlock when creating a subcomm viewer:
185:      the subcomm viewer will attempt to inherit the viewer's name, which, if not set, will be constructed
186:      collectively on the comm.
187:      */
188:     PetscCall(PetscObjectName((PetscObject)viewer));
189:     PetscCall(PetscViewerASCIIPrintf(viewer, "  Restriction/interpolation type: %s\n", PCGASMTypes[osm->type]));
190:     PetscCall(PetscViewerASCIIPrintf(viewer, "  %s\n", overlap));
191:     PetscCall(PetscViewerASCIIPrintf(viewer, "  %s\n", gsubdomains));
192:     PetscCall(PetscViewerASCIIPrintf(viewer, "  %s\n", msubdomains));
193:     PetscCall(PetscViewerASCIIPushSynchronized(viewer));
194:     PetscCall(PetscViewerASCIISynchronizedPrintf(viewer, "  [%d|%d] number of locally-supported subdomains = %" PetscInt_FMT "\n", rank, size, osm->n));
195:     PetscCall(PetscViewerFlush(viewer));
196:     PetscCall(PetscViewerASCIIPopSynchronized(viewer));
197:     /* Cannot take advantage of osm->same_subdomain_solvers without a global numbering of subdomains. */
198:     PetscCall(PetscViewerASCIIPrintf(viewer, "  Subdomain solver info is as follows:\n"));
199:     PetscCall(PetscViewerASCIIPushTab(viewer));
200:     PetscCall(PetscViewerASCIIPrintf(viewer, "  - - - - - - - - - - - - - - - - - -\n"));
201:     /* Now let subdomains go one at a time in the global numbering order and print their subdomain/solver info. */
202:     PetscCall(PCGASMComputeGlobalSubdomainNumbering_Private(pc, &numbering, &permutation));
203:     l = 0;
204:     for (count = 0; count < osm->N; ++count) {
205:       PetscMPIInt srank, ssize;
206:       if (l < osm->n) {
207:         PetscInt d = permutation[l]; /* d is the local number of the l-th smallest (in the global ordering) among the locally supported subdomains */
208:         if (numbering[d] == count) {
209:           PetscCallMPI(MPI_Comm_size(((PetscObject)osm->ois[d])->comm, &ssize));
210:           PetscCallMPI(MPI_Comm_rank(((PetscObject)osm->ois[d])->comm, &srank));
211:           PetscCall(PetscViewerGetSubViewer(viewer, ((PetscObject)osm->ois[d])->comm, &sviewer));
212:           PetscCall(ISGetLocalSize(osm->ois[d], &bsz));
213:           PetscCall(PetscViewerASCIISynchronizedPrintf(sviewer, "  [%d|%d] (subcomm [%d|%d]) local subdomain number %" PetscInt_FMT ", local size = %" PetscInt_FMT "\n", rank, size, srank, ssize, d, bsz));
214:           PetscCall(PetscViewerFlush(sviewer));
215:           PetscCall(PetscViewerASCIIPushTab(sviewer));
216:           if (view_subdomains) PetscCall(PCGASMSubdomainView_Private(pc, d, sviewer));
217:           if (!pc->setupcalled) {
218:             PetscCall(PetscViewerASCIISynchronizedPrintf(sviewer, "  Solver not set up yet: PCSetUp() not yet called\n"));
219:           } else {
220:             PetscCall(KSPView(osm->ksp[d], sviewer));
221:           }
222:           PetscCall(PetscViewerASCIIPopTab(sviewer));
223:           PetscCall(PetscViewerASCIIPrintf(sviewer, "  - - - - - - - - - - - - - - - - - -\n"));
224:           PetscCall(PetscViewerFlush(sviewer));
225:           PetscCall(PetscViewerRestoreSubViewer(viewer, ((PetscObject)osm->ois[d])->comm, &sviewer));
226:           ++l;
227:         } else {
228:           PetscCall(PetscViewerGetSubViewer(viewer, PETSC_COMM_SELF, &sviewer));
229:           PetscCall(PetscViewerRestoreSubViewer(viewer, PETSC_COMM_SELF, &sviewer));
230:         }
231:       } else {
232:         PetscCall(PetscViewerGetSubViewer(viewer, PETSC_COMM_SELF, &sviewer));
233:         PetscCall(PetscViewerRestoreSubViewer(viewer, PETSC_COMM_SELF, &sviewer));
234:       }
235:     }
236:     PetscCall(PetscFree2(numbering, permutation));
237:     PetscCall(PetscViewerASCIIPopTab(viewer));
238:     PetscCall(PetscViewerFlush(viewer));
239:     /* this line is needed to match the extra PetscViewerASCIIPushSynchronized() in PetscViewerGetSubViewer() */
240:     PetscCall(PetscViewerASCIIPopSynchronized(viewer));
241:   }
242:   PetscFunctionReturn(PETSC_SUCCESS);
243: }

245: PETSC_INTERN PetscErrorCode PCGASMCreateLocalSubdomains(Mat A, PetscInt nloc, IS *iis[]);

247: static PetscErrorCode PCGASMSetHierarchicalPartitioning(PC pc)
248: {
249:   PC_GASM        *osm = (PC_GASM *)pc->data;
250:   MatPartitioning part;
251:   MPI_Comm        comm;
252:   PetscMPIInt     size;
253:   PetscInt        nlocalsubdomains, fromrows_localsize;
254:   IS              partitioning, fromrows, isn;
255:   Vec             outervec;

257:   PetscFunctionBegin;
258:   PetscCall(PetscObjectGetComm((PetscObject)pc, &comm));
259:   PetscCallMPI(MPI_Comm_size(comm, &size));
260:   /* we do not need a hierarchical partitioning when
261:     * the total number of subdomains is consistent with
262:     * the number of MPI tasks.
263:     * For the following cases, we do not need to use HP
264:     * */
265:   if (osm->N == PETSC_DETERMINE || osm->N >= size || osm->N == 1) PetscFunctionReturn(PETSC_SUCCESS);
266:   PetscCheck(size % osm->N == 0, PETSC_COMM_WORLD, PETSC_ERR_ARG_INCOMP, "have to specify the total number of subdomains %" PetscInt_FMT " to be a factor of the number of ranks %d ", osm->N, size);
267:   nlocalsubdomains = size / osm->N;
268:   osm->n           = 1;
269:   PetscCall(MatPartitioningCreate(comm, &part));
270:   PetscCall(MatPartitioningSetAdjacency(part, pc->pmat));
271:   PetscCall(MatPartitioningSetType(part, MATPARTITIONINGHIERARCH));
272:   PetscCall(MatPartitioningHierarchicalSetNcoarseparts(part, osm->N));
273:   PetscCall(MatPartitioningHierarchicalSetNfineparts(part, nlocalsubdomains));
274:   PetscCall(MatPartitioningSetFromOptions(part));
275:   /* get new rank owner number of each vertex */
276:   PetscCall(MatPartitioningApply(part, &partitioning));
277:   PetscCall(ISBuildTwoSided(partitioning, NULL, &fromrows));
278:   PetscCall(ISPartitioningToNumbering(partitioning, &isn));
279:   PetscCall(ISDestroy(&isn));
280:   PetscCall(ISGetLocalSize(fromrows, &fromrows_localsize));
281:   PetscCall(MatPartitioningDestroy(&part));
282:   PetscCall(MatCreateVecs(pc->pmat, &outervec, NULL));
283:   PetscCall(VecCreateMPI(comm, fromrows_localsize, PETSC_DETERMINE, &osm->pcx));
284:   PetscCall(VecDuplicate(osm->pcx, &osm->pcy));
285:   PetscCall(VecScatterCreate(osm->pcx, NULL, outervec, fromrows, &osm->pctoouter));
286:   PetscCall(MatCreateSubMatrix(pc->pmat, fromrows, fromrows, MAT_INITIAL_MATRIX, &osm->permutationP));
287:   PetscCall(PetscObjectReference((PetscObject)fromrows));
288:   osm->permutationIS = fromrows;
289:   osm->pcmat         = pc->pmat;
290:   PetscCall(PetscObjectReference((PetscObject)osm->permutationP));
291:   pc->pmat = osm->permutationP;
292:   PetscCall(VecDestroy(&outervec));
293:   PetscCall(ISDestroy(&fromrows));
294:   PetscCall(ISDestroy(&partitioning));
295:   osm->n = PETSC_DETERMINE;
296:   PetscFunctionReturn(PETSC_SUCCESS);
297: }

299: static PetscErrorCode PCSetUp_GASM(PC pc)
300: {
301:   PC_GASM        *osm = (PC_GASM *)pc->data;
302:   PetscInt        i, nInnerIndices, nTotalInnerIndices;
303:   PetscMPIInt     rank, size;
304:   MatReuse        scall = MAT_REUSE_MATRIX;
305:   KSP             ksp;
306:   PC              subpc;
307:   const char     *prefix, *pprefix;
308:   Vec             x, y;
309:   PetscInt        oni;   /* Number of indices in the i-th local outer subdomain.               */
310:   const PetscInt *oidxi; /* Indices from the i-th subdomain local outer subdomain.             */
311:   PetscInt        on;    /* Number of indices in the disjoint union of local outer subdomains. */
312:   PetscInt       *oidx;  /* Indices in the disjoint union of local outer subdomains. */
313:   IS              gois;  /* Disjoint union the global indices of outer subdomains.             */
314:   IS              goid;  /* Identity IS of the size of the disjoint union of outer subdomains. */
315:   PetscScalar    *gxarray, *gyarray;
316:   PetscInt        gostart; /* Start of locally-owned indices in the vectors -- osm->gx,osm->gy -- over the disjoint union of outer subdomains. */
317:   PetscInt        num_subdomains  = 0;
318:   DM             *subdomain_dm    = NULL;
319:   char          **subdomain_names = NULL;
320:   PetscInt       *numbering;

322:   PetscFunctionBegin;
323:   PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)pc), &size));
324:   PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)pc), &rank));
325:   if (!pc->setupcalled) {
326:     /* use a hierarchical partitioning */
327:     if (osm->hierarchicalpartitioning) PetscCall(PCGASMSetHierarchicalPartitioning(pc));
328:     if (osm->n == PETSC_DETERMINE) {
329:       if (osm->N != PETSC_DETERMINE) {
330:         /* No local subdomains given, but the desired number of total subdomains is known, so construct them accordingly. */
331:         PetscCall(PCGASMCreateSubdomains(pc->pmat, osm->N, &osm->n, &osm->iis));
332:       } else if (osm->dm_subdomains && pc->dm) {
333:         /* try pc->dm next, if allowed */
334:         PetscInt d;
335:         IS      *inner_subdomain_is, *outer_subdomain_is;
336:         PetscCall(DMCreateDomainDecomposition(pc->dm, &num_subdomains, &subdomain_names, &inner_subdomain_is, &outer_subdomain_is, &subdomain_dm));
337:         if (num_subdomains) PetscCall(PCGASMSetSubdomains(pc, num_subdomains, inner_subdomain_is, outer_subdomain_is));
338:         for (d = 0; d < num_subdomains; ++d) {
339:           if (inner_subdomain_is) PetscCall(ISDestroy(&inner_subdomain_is[d]));
340:           if (outer_subdomain_is) PetscCall(ISDestroy(&outer_subdomain_is[d]));
341:         }
342:         PetscCall(PetscFree(inner_subdomain_is));
343:         PetscCall(PetscFree(outer_subdomain_is));
344:       } else {
345:         /* still no subdomains; use one per rank */
346:         osm->nmax = osm->n = 1;
347:         PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)pc), &size));
348:         osm->N = size;
349:         PetscCall(PCGASMCreateLocalSubdomains(pc->pmat, osm->n, &osm->iis));
350:       }
351:     }
352:     if (!osm->iis) {
353:       /*
354:        osm->n was set in PCGASMSetSubdomains(), but the actual subdomains have not been supplied.
355:        We create the requisite number of local inner subdomains and then expand them into
356:        out subdomains, if necessary.
357:        */
358:       PetscCall(PCGASMCreateLocalSubdomains(pc->pmat, osm->n, &osm->iis));
359:     }
360:     if (!osm->ois) {
361:       /*
362:             Initially make outer subdomains the same as inner subdomains. If nonzero additional overlap
363:             has been requested, copy the inner subdomains over so they can be modified.
364:       */
365:       PetscCall(PetscMalloc1(osm->n, &osm->ois));
366:       for (i = 0; i < osm->n; ++i) {
367:         if (osm->overlap > 0 && osm->N > 1) { /* With positive overlap, osm->iis[i] will be modified */
368:           PetscCall(ISDuplicate(osm->iis[i], (osm->ois) + i));
369:           PetscCall(ISCopy(osm->iis[i], osm->ois[i]));
370:         } else {
371:           PetscCall(PetscObjectReference((PetscObject)osm->iis[i]));
372:           osm->ois[i] = osm->iis[i];
373:         }
374:       }
375:       if (osm->overlap > 0 && osm->N > 1) {
376:         /* Extend the "overlapping" regions by a number of steps */
377:         PetscCall(MatIncreaseOverlapSplit(pc->pmat, osm->n, osm->ois, osm->overlap));
378:       }
379:     }

381:     /* Now the subdomains are defined.  Determine their global and max local numbers, if necessary. */
382:     if (osm->nmax == PETSC_DETERMINE) {
383:       PetscMPIInt inwork, outwork;
384:       /* determine global number of subdomains and the max number of local subdomains */
385:       inwork = osm->n;
386:       PetscCall(MPIU_Allreduce(&inwork, &outwork, 1, MPI_INT, MPI_MAX, PetscObjectComm((PetscObject)pc)));
387:       osm->nmax = outwork;
388:     }
389:     if (osm->N == PETSC_DETERMINE) {
390:       /* Determine the number of globally-distinct subdomains and compute a global numbering for them. */
391:       PetscCall(PetscObjectsListGetGlobalNumbering(PetscObjectComm((PetscObject)pc), osm->n, (PetscObject *)osm->ois, &osm->N, NULL));
392:     }

394:     if (osm->sort_indices) {
395:       for (i = 0; i < osm->n; i++) {
396:         PetscCall(ISSort(osm->ois[i]));
397:         PetscCall(ISSort(osm->iis[i]));
398:       }
399:     }
400:     PetscCall(PCGetOptionsPrefix(pc, &prefix));
401:     PetscCall(PCGASMPrintSubdomains(pc));

403:     /*
404:        Merge the ISs, create merged vectors and restrictions.
405:      */
406:     /* Merge outer subdomain ISs and construct a restriction onto the disjoint union of local outer subdomains. */
407:     on = 0;
408:     for (i = 0; i < osm->n; i++) {
409:       PetscCall(ISGetLocalSize(osm->ois[i], &oni));
410:       on += oni;
411:     }
412:     PetscCall(PetscMalloc1(on, &oidx));
413:     on = 0;
414:     /* Merge local indices together */
415:     for (i = 0; i < osm->n; i++) {
416:       PetscCall(ISGetLocalSize(osm->ois[i], &oni));
417:       PetscCall(ISGetIndices(osm->ois[i], &oidxi));
418:       PetscCall(PetscArraycpy(oidx + on, oidxi, oni));
419:       PetscCall(ISRestoreIndices(osm->ois[i], &oidxi));
420:       on += oni;
421:     }
422:     PetscCall(ISCreateGeneral(((PetscObject)(pc))->comm, on, oidx, PETSC_OWN_POINTER, &gois));
423:     nTotalInnerIndices = 0;
424:     for (i = 0; i < osm->n; i++) {
425:       PetscCall(ISGetLocalSize(osm->iis[i], &nInnerIndices));
426:       nTotalInnerIndices += nInnerIndices;
427:     }
428:     PetscCall(VecCreateMPI(((PetscObject)(pc))->comm, nTotalInnerIndices, PETSC_DETERMINE, &x));
429:     PetscCall(VecDuplicate(x, &y));

431:     PetscCall(VecCreateMPI(PetscObjectComm((PetscObject)pc), on, PETSC_DECIDE, &osm->gx));
432:     PetscCall(VecDuplicate(osm->gx, &osm->gy));
433:     PetscCall(VecGetOwnershipRange(osm->gx, &gostart, NULL));
434:     PetscCall(ISCreateStride(PetscObjectComm((PetscObject)pc), on, gostart, 1, &goid));
435:     /* gois might indices not on local */
436:     PetscCall(VecScatterCreate(x, gois, osm->gx, goid, &osm->gorestriction));
437:     PetscCall(PetscMalloc1(osm->n, &numbering));
438:     PetscCall(PetscObjectsListGetGlobalNumbering(PetscObjectComm((PetscObject)pc), osm->n, (PetscObject *)osm->ois, NULL, numbering));
439:     PetscCall(VecDestroy(&x));
440:     PetscCall(ISDestroy(&gois));

442:     /* Merge inner subdomain ISs and construct a restriction onto the disjoint union of local inner subdomains. */
443:     {
444:       PetscInt        ini;   /* Number of indices the i-th a local inner subdomain. */
445:       PetscInt        in;    /* Number of indices in the disjoint union of local inner subdomains. */
446:       PetscInt       *iidx;  /* Global indices in the merged local inner subdomain. */
447:       PetscInt       *ioidx; /* Global indices of the disjoint union of inner subdomains within the disjoint union of outer subdomains. */
448:       IS              giis;  /* IS for the disjoint union of inner subdomains. */
449:       IS              giois; /* IS for the disjoint union of inner subdomains within the disjoint union of outer subdomains. */
450:       PetscScalar    *array;
451:       const PetscInt *indices;
452:       PetscInt        k;
453:       on = 0;
454:       for (i = 0; i < osm->n; i++) {
455:         PetscCall(ISGetLocalSize(osm->ois[i], &oni));
456:         on += oni;
457:       }
458:       PetscCall(PetscMalloc1(on, &iidx));
459:       PetscCall(PetscMalloc1(on, &ioidx));
460:       PetscCall(VecGetArray(y, &array));
461:       /* set communicator id to determine where overlap is */
462:       in = 0;
463:       for (i = 0; i < osm->n; i++) {
464:         PetscCall(ISGetLocalSize(osm->iis[i], &ini));
465:         for (k = 0; k < ini; ++k) array[in + k] = numbering[i];
466:         in += ini;
467:       }
468:       PetscCall(VecRestoreArray(y, &array));
469:       PetscCall(VecScatterBegin(osm->gorestriction, y, osm->gy, INSERT_VALUES, SCATTER_FORWARD));
470:       PetscCall(VecScatterEnd(osm->gorestriction, y, osm->gy, INSERT_VALUES, SCATTER_FORWARD));
471:       PetscCall(VecGetOwnershipRange(osm->gy, &gostart, NULL));
472:       PetscCall(VecGetArray(osm->gy, &array));
473:       on = 0;
474:       in = 0;
475:       for (i = 0; i < osm->n; i++) {
476:         PetscCall(ISGetLocalSize(osm->ois[i], &oni));
477:         PetscCall(ISGetIndices(osm->ois[i], &indices));
478:         for (k = 0; k < oni; k++) {
479:           /*  skip overlapping indices to get inner domain */
480:           if (PetscRealPart(array[on + k]) != numbering[i]) continue;
481:           iidx[in]    = indices[k];
482:           ioidx[in++] = gostart + on + k;
483:         }
484:         PetscCall(ISRestoreIndices(osm->ois[i], &indices));
485:         on += oni;
486:       }
487:       PetscCall(VecRestoreArray(osm->gy, &array));
488:       PetscCall(ISCreateGeneral(PetscObjectComm((PetscObject)pc), in, iidx, PETSC_OWN_POINTER, &giis));
489:       PetscCall(ISCreateGeneral(PetscObjectComm((PetscObject)pc), in, ioidx, PETSC_OWN_POINTER, &giois));
490:       PetscCall(VecScatterCreate(y, giis, osm->gy, giois, &osm->girestriction));
491:       PetscCall(VecDestroy(&y));
492:       PetscCall(ISDestroy(&giis));
493:       PetscCall(ISDestroy(&giois));
494:     }
495:     PetscCall(ISDestroy(&goid));
496:     PetscCall(PetscFree(numbering));

498:     /* Create the subdomain work vectors. */
499:     PetscCall(PetscMalloc1(osm->n, &osm->x));
500:     PetscCall(PetscMalloc1(osm->n, &osm->y));
501:     PetscCall(VecGetArray(osm->gx, &gxarray));
502:     PetscCall(VecGetArray(osm->gy, &gyarray));
503:     for (i = 0, on = 0; i < osm->n; ++i, on += oni) {
504:       PetscInt oNi;
505:       PetscCall(ISGetLocalSize(osm->ois[i], &oni));
506:       /* on a sub communicator */
507:       PetscCall(ISGetSize(osm->ois[i], &oNi));
508:       PetscCall(VecCreateMPIWithArray(((PetscObject)osm->ois[i])->comm, 1, oni, oNi, gxarray + on, &osm->x[i]));
509:       PetscCall(VecCreateMPIWithArray(((PetscObject)osm->ois[i])->comm, 1, oni, oNi, gyarray + on, &osm->y[i]));
510:     }
511:     PetscCall(VecRestoreArray(osm->gx, &gxarray));
512:     PetscCall(VecRestoreArray(osm->gy, &gyarray));
513:     /* Create the subdomain solvers */
514:     PetscCall(PetscMalloc1(osm->n, &osm->ksp));
515:     for (i = 0; i < osm->n; i++) {
516:       char subprefix[PETSC_MAX_PATH_LEN + 1];
517:       PetscCall(KSPCreate(((PetscObject)osm->ois[i])->comm, &ksp));
518:       PetscCall(KSPSetNestLevel(ksp, pc->kspnestlevel));
519:       PetscCall(KSPSetErrorIfNotConverged(ksp, pc->erroriffailure));
520:       PetscCall(PetscObjectIncrementTabLevel((PetscObject)ksp, (PetscObject)pc, 1));
521:       PetscCall(KSPSetType(ksp, KSPPREONLY));
522:       PetscCall(KSPGetPC(ksp, &subpc)); /* Why do we need this here? */
523:       if (subdomain_dm) {
524:         PetscCall(KSPSetDM(ksp, subdomain_dm[i]));
525:         PetscCall(DMDestroy(subdomain_dm + i));
526:       }
527:       PetscCall(PCGetOptionsPrefix(pc, &prefix));
528:       PetscCall(KSPSetOptionsPrefix(ksp, prefix));
529:       if (subdomain_names && subdomain_names[i]) {
530:         PetscCall(PetscSNPrintf(subprefix, PETSC_MAX_PATH_LEN, "sub_%s_", subdomain_names[i]));
531:         PetscCall(KSPAppendOptionsPrefix(ksp, subprefix));
532:         PetscCall(PetscFree(subdomain_names[i]));
533:       }
534:       PetscCall(KSPAppendOptionsPrefix(ksp, "sub_"));
535:       osm->ksp[i] = ksp;
536:     }
537:     PetscCall(PetscFree(subdomain_dm));
538:     PetscCall(PetscFree(subdomain_names));
539:     scall = MAT_INITIAL_MATRIX;
540:   } else { /* if (pc->setupcalled) */
541:     /*
542:        Destroy the submatrices from the previous iteration
543:     */
544:     if (pc->flag == DIFFERENT_NONZERO_PATTERN) {
545:       PetscCall(MatDestroyMatrices(osm->n, &osm->pmat));
546:       scall = MAT_INITIAL_MATRIX;
547:     }
548:     if (osm->permutationIS) {
549:       PetscCall(MatCreateSubMatrix(pc->pmat, osm->permutationIS, osm->permutationIS, scall, &osm->permutationP));
550:       PetscCall(PetscObjectReference((PetscObject)osm->permutationP));
551:       osm->pcmat = pc->pmat;
552:       pc->pmat   = osm->permutationP;
553:     }
554:   }

556:   /*
557:      Extract the submatrices.
558:   */
559:   if (size > 1) {
560:     PetscCall(MatCreateSubMatricesMPI(pc->pmat, osm->n, osm->ois, osm->ois, scall, &osm->pmat));
561:   } else {
562:     PetscCall(MatCreateSubMatrices(pc->pmat, osm->n, osm->ois, osm->ois, scall, &osm->pmat));
563:   }
564:   if (scall == MAT_INITIAL_MATRIX) {
565:     PetscCall(PetscObjectGetOptionsPrefix((PetscObject)pc->pmat, &pprefix));
566:     for (i = 0; i < osm->n; i++) PetscCall(PetscObjectSetOptionsPrefix((PetscObject)osm->pmat[i], pprefix));
567:   }

569:   /* Return control to the user so that the submatrices can be modified (e.g., to apply
570:      different boundary conditions for the submatrices than for the global problem) */
571:   PetscCall(PCModifySubMatrices(pc, osm->n, osm->ois, osm->ois, osm->pmat, pc->modifysubmatricesP));

573:   /*
574:      Loop over submatrices putting them into local ksps
575:   */
576:   for (i = 0; i < osm->n; i++) {
577:     PetscCall(KSPSetOperators(osm->ksp[i], osm->pmat[i], osm->pmat[i]));
578:     PetscCall(KSPGetOptionsPrefix(osm->ksp[i], &prefix));
579:     PetscCall(MatSetOptionsPrefix(osm->pmat[i], prefix));
580:     if (!pc->setupcalled) PetscCall(KSPSetFromOptions(osm->ksp[i]));
581:   }
582:   if (osm->pcmat) {
583:     PetscCall(MatDestroy(&pc->pmat));
584:     pc->pmat   = osm->pcmat;
585:     osm->pcmat = NULL;
586:   }
587:   PetscFunctionReturn(PETSC_SUCCESS);
588: }

590: static PetscErrorCode PCSetUpOnBlocks_GASM(PC pc)
591: {
592:   PC_GASM *osm = (PC_GASM *)pc->data;
593:   PetscInt i;

595:   PetscFunctionBegin;
596:   for (i = 0; i < osm->n; i++) PetscCall(KSPSetUp(osm->ksp[i]));
597:   PetscFunctionReturn(PETSC_SUCCESS);
598: }

600: static PetscErrorCode PCApply_GASM(PC pc, Vec xin, Vec yout)
601: {
602:   PC_GASM    *osm = (PC_GASM *)pc->data;
603:   PetscInt    i;
604:   Vec         x, y;
605:   ScatterMode forward = SCATTER_FORWARD, reverse = SCATTER_REVERSE;

607:   PetscFunctionBegin;
608:   if (osm->pctoouter) {
609:     PetscCall(VecScatterBegin(osm->pctoouter, xin, osm->pcx, INSERT_VALUES, SCATTER_REVERSE));
610:     PetscCall(VecScatterEnd(osm->pctoouter, xin, osm->pcx, INSERT_VALUES, SCATTER_REVERSE));
611:     x = osm->pcx;
612:     y = osm->pcy;
613:   } else {
614:     x = xin;
615:     y = yout;
616:   }
617:   /*
618:      support for limiting the restriction or interpolation only to the inner
619:      subdomain values (leaving the other values 0).
620:   */
621:   if (!(osm->type & PC_GASM_RESTRICT)) {
622:     /* have to zero the work RHS since scatter may leave some slots empty */
623:     PetscCall(VecZeroEntries(osm->gx));
624:     PetscCall(VecScatterBegin(osm->girestriction, x, osm->gx, INSERT_VALUES, forward));
625:   } else {
626:     PetscCall(VecScatterBegin(osm->gorestriction, x, osm->gx, INSERT_VALUES, forward));
627:   }
628:   PetscCall(VecZeroEntries(osm->gy));
629:   if (!(osm->type & PC_GASM_RESTRICT)) {
630:     PetscCall(VecScatterEnd(osm->girestriction, x, osm->gx, INSERT_VALUES, forward));
631:   } else {
632:     PetscCall(VecScatterEnd(osm->gorestriction, x, osm->gx, INSERT_VALUES, forward));
633:   }
634:   /* do the subdomain solves */
635:   for (i = 0; i < osm->n; ++i) {
636:     PetscCall(KSPSolve(osm->ksp[i], osm->x[i], osm->y[i]));
637:     PetscCall(KSPCheckSolve(osm->ksp[i], pc, osm->y[i]));
638:   }
639:   /* do we need to zero y? */
640:   PetscCall(VecZeroEntries(y));
641:   if (!(osm->type & PC_GASM_INTERPOLATE)) {
642:     PetscCall(VecScatterBegin(osm->girestriction, osm->gy, y, ADD_VALUES, reverse));
643:     PetscCall(VecScatterEnd(osm->girestriction, osm->gy, y, ADD_VALUES, reverse));
644:   } else {
645:     PetscCall(VecScatterBegin(osm->gorestriction, osm->gy, y, ADD_VALUES, reverse));
646:     PetscCall(VecScatterEnd(osm->gorestriction, osm->gy, y, ADD_VALUES, reverse));
647:   }
648:   if (osm->pctoouter) {
649:     PetscCall(VecScatterBegin(osm->pctoouter, y, yout, INSERT_VALUES, SCATTER_FORWARD));
650:     PetscCall(VecScatterEnd(osm->pctoouter, y, yout, INSERT_VALUES, SCATTER_FORWARD));
651:   }
652:   PetscFunctionReturn(PETSC_SUCCESS);
653: }

655: static PetscErrorCode PCMatApply_GASM(PC pc, Mat Xin, Mat Yout)
656: {
657:   PC_GASM    *osm = (PC_GASM *)pc->data;
658:   Mat         X, Y, O = NULL, Z, W;
659:   Vec         x, y;
660:   PetscInt    i, m, M, N;
661:   ScatterMode forward = SCATTER_FORWARD, reverse = SCATTER_REVERSE;

663:   PetscFunctionBegin;
664:   PetscCheck(osm->n == 1, PetscObjectComm((PetscObject)pc), PETSC_ERR_SUP, "Not yet implemented");
665:   PetscCall(MatGetSize(Xin, NULL, &N));
666:   if (osm->pctoouter) {
667:     PetscCall(VecGetLocalSize(osm->pcx, &m));
668:     PetscCall(VecGetSize(osm->pcx, &M));
669:     PetscCall(MatCreateDense(PetscObjectComm((PetscObject)osm->ois[0]), m, PETSC_DECIDE, M, N, NULL, &O));
670:     for (i = 0; i < N; ++i) {
671:       PetscCall(MatDenseGetColumnVecRead(Xin, i, &x));
672:       PetscCall(MatDenseGetColumnVecWrite(O, i, &y));
673:       PetscCall(VecScatterBegin(osm->pctoouter, x, y, INSERT_VALUES, SCATTER_REVERSE));
674:       PetscCall(VecScatterEnd(osm->pctoouter, x, y, INSERT_VALUES, SCATTER_REVERSE));
675:       PetscCall(MatDenseRestoreColumnVecWrite(O, i, &y));
676:       PetscCall(MatDenseRestoreColumnVecRead(Xin, i, &x));
677:     }
678:     X = Y = O;
679:   } else {
680:     X = Xin;
681:     Y = Yout;
682:   }
683:   /*
684:      support for limiting the restriction or interpolation only to the inner
685:      subdomain values (leaving the other values 0).
686:   */
687:   PetscCall(VecGetLocalSize(osm->x[0], &m));
688:   PetscCall(VecGetSize(osm->x[0], &M));
689:   PetscCall(MatCreateDense(PetscObjectComm((PetscObject)osm->ois[0]), m, PETSC_DECIDE, M, N, NULL, &Z));
690:   for (i = 0; i < N; ++i) {
691:     PetscCall(MatDenseGetColumnVecRead(X, i, &x));
692:     PetscCall(MatDenseGetColumnVecWrite(Z, i, &y));
693:     if (!(osm->type & PC_GASM_RESTRICT)) {
694:       /* have to zero the work RHS since scatter may leave some slots empty */
695:       PetscCall(VecZeroEntries(y));
696:       PetscCall(VecScatterBegin(osm->girestriction, x, y, INSERT_VALUES, forward));
697:       PetscCall(VecScatterEnd(osm->girestriction, x, y, INSERT_VALUES, forward));
698:     } else {
699:       PetscCall(VecScatterBegin(osm->gorestriction, x, y, INSERT_VALUES, forward));
700:       PetscCall(VecScatterEnd(osm->gorestriction, x, y, INSERT_VALUES, forward));
701:     }
702:     PetscCall(MatDenseRestoreColumnVecWrite(Z, i, &y));
703:     PetscCall(MatDenseRestoreColumnVecRead(X, i, &x));
704:   }
705:   PetscCall(MatCreateDense(PetscObjectComm((PetscObject)osm->ois[0]), m, PETSC_DECIDE, M, N, NULL, &W));
706:   PetscCall(MatSetOption(Z, MAT_NO_OFF_PROC_ENTRIES, PETSC_TRUE));
707:   PetscCall(MatAssemblyBegin(Z, MAT_FINAL_ASSEMBLY));
708:   PetscCall(MatAssemblyEnd(Z, MAT_FINAL_ASSEMBLY));
709:   /* do the subdomain solve */
710:   PetscCall(KSPMatSolve(osm->ksp[0], Z, W));
711:   PetscCall(KSPCheckSolve(osm->ksp[0], pc, NULL));
712:   PetscCall(MatDestroy(&Z));
713:   /* do we need to zero y? */
714:   PetscCall(MatZeroEntries(Y));
715:   for (i = 0; i < N; ++i) {
716:     PetscCall(MatDenseGetColumnVecWrite(Y, i, &y));
717:     PetscCall(MatDenseGetColumnVecRead(W, i, &x));
718:     if (!(osm->type & PC_GASM_INTERPOLATE)) {
719:       PetscCall(VecScatterBegin(osm->girestriction, x, y, ADD_VALUES, reverse));
720:       PetscCall(VecScatterEnd(osm->girestriction, x, y, ADD_VALUES, reverse));
721:     } else {
722:       PetscCall(VecScatterBegin(osm->gorestriction, x, y, ADD_VALUES, reverse));
723:       PetscCall(VecScatterEnd(osm->gorestriction, x, y, ADD_VALUES, reverse));
724:     }
725:     PetscCall(MatDenseRestoreColumnVecRead(W, i, &x));
726:     if (osm->pctoouter) {
727:       PetscCall(MatDenseGetColumnVecWrite(Yout, i, &x));
728:       PetscCall(VecScatterBegin(osm->pctoouter, y, x, INSERT_VALUES, SCATTER_FORWARD));
729:       PetscCall(VecScatterEnd(osm->pctoouter, y, x, INSERT_VALUES, SCATTER_FORWARD));
730:       PetscCall(MatDenseRestoreColumnVecRead(Yout, i, &x));
731:     }
732:     PetscCall(MatDenseRestoreColumnVecWrite(Y, i, &y));
733:   }
734:   PetscCall(MatDestroy(&W));
735:   PetscCall(MatDestroy(&O));
736:   PetscFunctionReturn(PETSC_SUCCESS);
737: }

739: static PetscErrorCode PCApplyTranspose_GASM(PC pc, Vec xin, Vec yout)
740: {
741:   PC_GASM    *osm = (PC_GASM *)pc->data;
742:   PetscInt    i;
743:   Vec         x, y;
744:   ScatterMode forward = SCATTER_FORWARD, reverse = SCATTER_REVERSE;

746:   PetscFunctionBegin;
747:   if (osm->pctoouter) {
748:     PetscCall(VecScatterBegin(osm->pctoouter, xin, osm->pcx, INSERT_VALUES, SCATTER_REVERSE));
749:     PetscCall(VecScatterEnd(osm->pctoouter, xin, osm->pcx, INSERT_VALUES, SCATTER_REVERSE));
750:     x = osm->pcx;
751:     y = osm->pcy;
752:   } else {
753:     x = xin;
754:     y = yout;
755:   }
756:   /*
757:      Support for limiting the restriction or interpolation to only local
758:      subdomain values (leaving the other values 0).

760:      Note: these are reversed from the PCApply_GASM() because we are applying the
761:      transpose of the three terms
762:   */
763:   if (!(osm->type & PC_GASM_INTERPOLATE)) {
764:     /* have to zero the work RHS since scatter may leave some slots empty */
765:     PetscCall(VecZeroEntries(osm->gx));
766:     PetscCall(VecScatterBegin(osm->girestriction, x, osm->gx, INSERT_VALUES, forward));
767:   } else {
768:     PetscCall(VecScatterBegin(osm->gorestriction, x, osm->gx, INSERT_VALUES, forward));
769:   }
770:   PetscCall(VecZeroEntries(osm->gy));
771:   if (!(osm->type & PC_GASM_INTERPOLATE)) {
772:     PetscCall(VecScatterEnd(osm->girestriction, x, osm->gx, INSERT_VALUES, forward));
773:   } else {
774:     PetscCall(VecScatterEnd(osm->gorestriction, x, osm->gx, INSERT_VALUES, forward));
775:   }
776:   /* do the local solves */
777:   for (i = 0; i < osm->n; ++i) { /* Note that the solves are local, so we can go to osm->n, rather than osm->nmax. */
778:     PetscCall(KSPSolveTranspose(osm->ksp[i], osm->x[i], osm->y[i]));
779:     PetscCall(KSPCheckSolve(osm->ksp[i], pc, osm->y[i]));
780:   }
781:   PetscCall(VecZeroEntries(y));
782:   if (!(osm->type & PC_GASM_RESTRICT)) {
783:     PetscCall(VecScatterBegin(osm->girestriction, osm->gy, y, ADD_VALUES, reverse));
784:     PetscCall(VecScatterEnd(osm->girestriction, osm->gy, y, ADD_VALUES, reverse));
785:   } else {
786:     PetscCall(VecScatterBegin(osm->gorestriction, osm->gy, y, ADD_VALUES, reverse));
787:     PetscCall(VecScatterEnd(osm->gorestriction, osm->gy, y, ADD_VALUES, reverse));
788:   }
789:   if (osm->pctoouter) {
790:     PetscCall(VecScatterBegin(osm->pctoouter, y, yout, INSERT_VALUES, SCATTER_FORWARD));
791:     PetscCall(VecScatterEnd(osm->pctoouter, y, yout, INSERT_VALUES, SCATTER_FORWARD));
792:   }
793:   PetscFunctionReturn(PETSC_SUCCESS);
794: }

796: static PetscErrorCode PCReset_GASM(PC pc)
797: {
798:   PC_GASM *osm = (PC_GASM *)pc->data;
799:   PetscInt i;

801:   PetscFunctionBegin;
802:   if (osm->ksp) {
803:     for (i = 0; i < osm->n; i++) PetscCall(KSPReset(osm->ksp[i]));
804:   }
805:   if (osm->pmat) {
806:     if (osm->n > 0) {
807:       PetscMPIInt size;
808:       PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)pc), &size));
809:       if (size > 1) {
810:         /* osm->pmat is created by MatCreateSubMatricesMPI(), cannot use MatDestroySubMatrices() */
811:         PetscCall(MatDestroyMatrices(osm->n, &osm->pmat));
812:       } else {
813:         PetscCall(MatDestroySubMatrices(osm->n, &osm->pmat));
814:       }
815:     }
816:   }
817:   if (osm->x) {
818:     for (i = 0; i < osm->n; i++) {
819:       PetscCall(VecDestroy(&osm->x[i]));
820:       PetscCall(VecDestroy(&osm->y[i]));
821:     }
822:   }
823:   PetscCall(VecDestroy(&osm->gx));
824:   PetscCall(VecDestroy(&osm->gy));

826:   PetscCall(VecScatterDestroy(&osm->gorestriction));
827:   PetscCall(VecScatterDestroy(&osm->girestriction));
828:   if (!osm->user_subdomains) {
829:     PetscCall(PCGASMDestroySubdomains(osm->n, &osm->ois, &osm->iis));
830:     osm->N    = PETSC_DETERMINE;
831:     osm->nmax = PETSC_DETERMINE;
832:   }
833:   if (osm->pctoouter) PetscCall(VecScatterDestroy(&osm->pctoouter));
834:   if (osm->permutationIS) PetscCall(ISDestroy(&osm->permutationIS));
835:   if (osm->pcx) PetscCall(VecDestroy(&osm->pcx));
836:   if (osm->pcy) PetscCall(VecDestroy(&osm->pcy));
837:   if (osm->permutationP) PetscCall(MatDestroy(&osm->permutationP));
838:   if (osm->pcmat) PetscCall(MatDestroy(&osm->pcmat));
839:   PetscFunctionReturn(PETSC_SUCCESS);
840: }

842: static PetscErrorCode PCDestroy_GASM(PC pc)
843: {
844:   PC_GASM *osm = (PC_GASM *)pc->data;
845:   PetscInt i;

847:   PetscFunctionBegin;
848:   PetscCall(PCReset_GASM(pc));
849:   /* PCReset will not destroy subdomains, if user_subdomains is true. */
850:   PetscCall(PCGASMDestroySubdomains(osm->n, &osm->ois, &osm->iis));
851:   if (osm->ksp) {
852:     for (i = 0; i < osm->n; i++) PetscCall(KSPDestroy(&osm->ksp[i]));
853:     PetscCall(PetscFree(osm->ksp));
854:   }
855:   PetscCall(PetscFree(osm->x));
856:   PetscCall(PetscFree(osm->y));
857:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetSubdomains_C", NULL));
858:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetOverlap_C", NULL));
859:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetType_C", NULL));
860:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetSortIndices_C", NULL));
861:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMGetSubKSP_C", NULL));
862:   PetscCall(PetscFree(pc->data));
863:   PetscFunctionReturn(PETSC_SUCCESS);
864: }

866: static PetscErrorCode PCSetFromOptions_GASM(PC pc, PetscOptionItems *PetscOptionsObject)
867: {
868:   PC_GASM   *osm = (PC_GASM *)pc->data;
869:   PetscInt   blocks, ovl;
870:   PetscBool  flg;
871:   PCGASMType gasmtype;

873:   PetscFunctionBegin;
874:   PetscOptionsHeadBegin(PetscOptionsObject, "Generalized additive Schwarz options");
875:   PetscCall(PetscOptionsBool("-pc_gasm_use_dm_subdomains", "If subdomains aren't set, use DMCreateDomainDecomposition() to define subdomains.", "PCGASMSetUseDMSubdomains", osm->dm_subdomains, &osm->dm_subdomains, &flg));
876:   PetscCall(PetscOptionsInt("-pc_gasm_total_subdomains", "Total number of subdomains across communicator", "PCGASMSetTotalSubdomains", osm->N, &blocks, &flg));
877:   if (flg) PetscCall(PCGASMSetTotalSubdomains(pc, blocks));
878:   PetscCall(PetscOptionsInt("-pc_gasm_overlap", "Number of overlapping degrees of freedom", "PCGASMSetOverlap", osm->overlap, &ovl, &flg));
879:   if (flg) {
880:     PetscCall(PCGASMSetOverlap(pc, ovl));
881:     osm->dm_subdomains = PETSC_FALSE;
882:   }
883:   flg = PETSC_FALSE;
884:   PetscCall(PetscOptionsEnum("-pc_gasm_type", "Type of restriction/extension", "PCGASMSetType", PCGASMTypes, (PetscEnum)osm->type, (PetscEnum *)&gasmtype, &flg));
885:   if (flg) PetscCall(PCGASMSetType(pc, gasmtype));
886:   PetscCall(PetscOptionsBool("-pc_gasm_use_hierachical_partitioning", "use hierarchical partitioning", NULL, osm->hierarchicalpartitioning, &osm->hierarchicalpartitioning, &flg));
887:   PetscOptionsHeadEnd();
888:   PetscFunctionReturn(PETSC_SUCCESS);
889: }

891: /*@
892:   PCGASMSetTotalSubdomains - sets the total number of subdomains to use across the communicator for `PCGASM`

894:   Logically Collective

896:   Input Parameters:
897: + pc - the preconditioner
898: - N  - total number of subdomains

900:   Level: beginner

902: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetSubdomains()`, `PCGASMSetOverlap()`
903:           `PCGASMCreateSubdomains2D()`
904: @*/
905: PetscErrorCode PCGASMSetTotalSubdomains(PC pc, PetscInt N)
906: {
907:   PC_GASM    *osm = (PC_GASM *)pc->data;
908:   PetscMPIInt size, rank;

910:   PetscFunctionBegin;
911:   PetscCheck(N >= 1, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Total number of subdomains must be 1 or more, got N = %" PetscInt_FMT, N);
912:   PetscCheck(!pc->setupcalled, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_WRONGSTATE, "PCGASMSetTotalSubdomains() should be called before calling PCSetUp().");

914:   PetscCall(PCGASMDestroySubdomains(osm->n, &osm->iis, &osm->ois));
915:   osm->ois = osm->iis = NULL;

917:   PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)pc), &size));
918:   PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)pc), &rank));
919:   osm->N             = N;
920:   osm->n             = PETSC_DETERMINE;
921:   osm->nmax          = PETSC_DETERMINE;
922:   osm->dm_subdomains = PETSC_FALSE;
923:   PetscFunctionReturn(PETSC_SUCCESS);
924: }

926: static PetscErrorCode PCGASMSetSubdomains_GASM(PC pc, PetscInt n, IS iis[], IS ois[])
927: {
928:   PC_GASM *osm = (PC_GASM *)pc->data;
929:   PetscInt i;

931:   PetscFunctionBegin;
932:   PetscCheck(n >= 1, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Each MPI rank must have 1 or more subdomains, got n = %" PetscInt_FMT, n);
933:   PetscCheck(!pc->setupcalled, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_WRONGSTATE, "PCGASMSetSubdomains() should be called before calling PCSetUp().");

935:   PetscCall(PCGASMDestroySubdomains(osm->n, &osm->iis, &osm->ois));
936:   osm->iis = osm->ois = NULL;
937:   osm->n              = n;
938:   osm->N              = PETSC_DETERMINE;
939:   osm->nmax           = PETSC_DETERMINE;
940:   if (ois) {
941:     PetscCall(PetscMalloc1(n, &osm->ois));
942:     for (i = 0; i < n; i++) {
943:       PetscCall(PetscObjectReference((PetscObject)ois[i]));
944:       osm->ois[i] = ois[i];
945:     }
946:     /*
947:        Since the user set the outer subdomains, even if nontrivial overlap was requested via PCGASMSetOverlap(),
948:        it will be ignored.  To avoid confusion later on (e.g., when viewing the PC), the overlap size is set to -1.
949:     */
950:     osm->overlap = -1;
951:     /* inner subdomains must be provided  */
952:     PetscCheck(iis, PETSC_COMM_SELF, PETSC_ERR_ARG_NULL, "inner indices have to be provided ");
953:   } /* end if */
954:   if (iis) {
955:     PetscCall(PetscMalloc1(n, &osm->iis));
956:     for (i = 0; i < n; i++) {
957:       PetscCall(PetscObjectReference((PetscObject)iis[i]));
958:       osm->iis[i] = iis[i];
959:     }
960:     if (!ois) {
961:       osm->ois = NULL;
962:       /* if user does not provide outer indices, we will create the corresponding outer indices using  osm->overlap =1 in PCSetUp_GASM */
963:     }
964:   }
965:   if (PetscDefined(USE_DEBUG)) {
966:     PetscInt        j, rstart, rend, *covered, lsize;
967:     const PetscInt *indices;
968:     /* check if the inner indices cover and only cover the local portion of the preconditioning matrix */
969:     PetscCall(MatGetOwnershipRange(pc->pmat, &rstart, &rend));
970:     PetscCall(PetscCalloc1(rend - rstart, &covered));
971:     /* check if the current MPI process owns indices from others */
972:     for (i = 0; i < n; i++) {
973:       PetscCall(ISGetIndices(osm->iis[i], &indices));
974:       PetscCall(ISGetLocalSize(osm->iis[i], &lsize));
975:       for (j = 0; j < lsize; j++) {
976:         PetscCheck(indices[j] >= rstart && indices[j] < rend, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "inner subdomains can not own an index %" PetscInt_FMT " from other ranks", indices[j]);
977:         PetscCheck(covered[indices[j] - rstart] != 1, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "inner subdomains can not have an overlapping index %" PetscInt_FMT " ", indices[j]);
978:         covered[indices[j] - rstart] = 1;
979:       }
980:       PetscCall(ISRestoreIndices(osm->iis[i], &indices));
981:     }
982:     /* check if we miss any indices */
983:     for (i = rstart; i < rend; i++) PetscCheck(covered[i - rstart], PETSC_COMM_SELF, PETSC_ERR_ARG_NULL, "local entity %" PetscInt_FMT " was not covered by inner subdomains", i);
984:     PetscCall(PetscFree(covered));
985:   }
986:   if (iis) osm->user_subdomains = PETSC_TRUE;
987:   PetscFunctionReturn(PETSC_SUCCESS);
988: }

990: static PetscErrorCode PCGASMSetOverlap_GASM(PC pc, PetscInt ovl)
991: {
992:   PC_GASM *osm = (PC_GASM *)pc->data;

994:   PetscFunctionBegin;
995:   PetscCheck(ovl >= 0, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_OUTOFRANGE, "Negative overlap value requested");
996:   PetscCheck(!pc->setupcalled || ovl == osm->overlap, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_WRONGSTATE, "PCGASMSetOverlap() should be called before PCSetUp().");
997:   if (!pc->setupcalled) osm->overlap = ovl;
998:   PetscFunctionReturn(PETSC_SUCCESS);
999: }

1001: static PetscErrorCode PCGASMSetType_GASM(PC pc, PCGASMType type)
1002: {
1003:   PC_GASM *osm = (PC_GASM *)pc->data;

1005:   PetscFunctionBegin;
1006:   osm->type     = type;
1007:   osm->type_set = PETSC_TRUE;
1008:   PetscFunctionReturn(PETSC_SUCCESS);
1009: }

1011: static PetscErrorCode PCGASMSetSortIndices_GASM(PC pc, PetscBool doSort)
1012: {
1013:   PC_GASM *osm = (PC_GASM *)pc->data;

1015:   PetscFunctionBegin;
1016:   osm->sort_indices = doSort;
1017:   PetscFunctionReturn(PETSC_SUCCESS);
1018: }

1020: /*
1021:    FIXME: This routine might need to be modified now that multiple processes per subdomain are allowed.
1022:         In particular, it would upset the global subdomain number calculation.
1023: */
1024: static PetscErrorCode PCGASMGetSubKSP_GASM(PC pc, PetscInt *n, PetscInt *first, KSP **ksp)
1025: {
1026:   PC_GASM *osm = (PC_GASM *)pc->data;

1028:   PetscFunctionBegin;
1029:   PetscCheck(osm->n >= 1, PetscObjectComm((PetscObject)pc), PETSC_ERR_ORDER, "Need to call PCSetUp() on PC (or KSPSetUp() on the outer KSP object) before calling here");

1031:   if (n) *n = osm->n;
1032:   if (first) {
1033:     PetscCallMPI(MPI_Scan(&osm->n, first, 1, MPIU_INT, MPI_SUM, PetscObjectComm((PetscObject)pc)));
1034:     *first -= osm->n;
1035:   }
1036:   if (ksp) {
1037:     /* Assume that local solves are now different; not necessarily
1038:        true, though!  This flag is used only for PCView_GASM() */
1039:     *ksp                        = osm->ksp;
1040:     osm->same_subdomain_solvers = PETSC_FALSE;
1041:   }
1042:   PetscFunctionReturn(PETSC_SUCCESS);
1043: } /* PCGASMGetSubKSP_GASM() */

1045: /*@
1046:   PCGASMSetSubdomains - Sets the subdomains for this MPI process
1047:   for the additive Schwarz preconditioner with multiple MPI processes per subdomain, `PCGASM`

1049:   Collective

1051:   Input Parameters:
1052: + pc  - the preconditioner object
1053: . n   - the number of subdomains for this MPI process
1054: . iis - the index sets that define the inner subdomains (or `NULL` for PETSc to determine subdomains), the `iis` array is
1055:         copied so may be freed after this call.
1056: - ois - the index sets that define the outer subdomains (or `NULL` to use the same as `iis`, or to construct by expanding `iis` by
1057:         the requested overlap), the `ois` array is copied so may be freed after this call.

1059:   Level: advanced

1061:   Notes:
1062:   The `IS` indices use the parallel, global numbering of the vector entries.

1064:   Inner subdomains are those where the correction is applied.

1066:   Outer subdomains are those where the residual necessary to obtain the
1067:   corrections is obtained (see `PCGASMType` for the use of inner/outer subdomains).

1069:   Both inner and outer subdomains can extend over several MPI processes.
1070:   This process' portion of a subdomain is known as a local subdomain.

1072:   Inner subdomains can not overlap with each other, do not have any entities from remote processes,
1073:   and  have to cover the entire local subdomain owned by the current process. The index sets on each
1074:   process should be ordered such that the ith local subdomain is connected to the ith remote subdomain
1075:   on another MPI process.

1077:   By default the `PGASM` preconditioner uses 1 (local) subdomain per MPI process.

1079:   The `iis` and `ois` arrays may be freed after this call using `PCGASMDestroySubdomains()`

1081: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetOverlap()`, `PCGASMGetSubKSP()`, `PCGASMDestroySubdomains()`,
1082:           `PCGASMCreateSubdomains2D()`, `PCGASMGetSubdomains()`
1083: @*/
1084: PetscErrorCode PCGASMSetSubdomains(PC pc, PetscInt n, IS iis[], IS ois[])
1085: {
1086:   PC_GASM *osm = (PC_GASM *)pc->data;

1088:   PetscFunctionBegin;
1090:   PetscTryMethod(pc, "PCGASMSetSubdomains_C", (PC, PetscInt, IS[], IS[]), (pc, n, iis, ois));
1091:   osm->dm_subdomains = PETSC_FALSE;
1092:   PetscFunctionReturn(PETSC_SUCCESS);
1093: }

1095: /*@
1096:   PCGASMSetOverlap - Sets the overlap between a pair of subdomains for the
1097:   additive Schwarz preconditioner `PCGASM`.  Either all or no MPI processes in the
1098:   pc communicator must call this routine.

1100:   Logically Collective

1102:   Input Parameters:
1103: + pc  - the preconditioner context
1104: - ovl - the amount of overlap between subdomains (ovl >= 0, default value = 0)

1106:   Options Database Key:
1107: . -pc_gasm_overlap <overlap> - Sets overlap

1109:   Level: intermediate

1111:   Notes:
1112:   By default the `PCGASM` preconditioner uses 1 subdomain per process.  To use
1113:   multiple subdomain per perocessor or "straddling" subdomains that intersect
1114:   multiple processes use `PCGASMSetSubdomains()` (or option `-pc_gasm_total_subdomains` <n>).

1116:   The overlap defaults to 0, so if one desires that no additional
1117:   overlap be computed beyond what may have been set with a call to
1118:   `PCGASMSetSubdomains()`, then `ovl` must be set to be 0.  In particular, if one does
1119:   not explicitly set the subdomains in application code, then all overlap would be computed
1120:   internally by PETSc, and using an overlap of 0 would result in an `PCGASM`
1121:   variant that is equivalent to the block Jacobi preconditioner.

1123:   One can define initial index sets with any overlap via
1124:   `PCGASMSetSubdomains()`; the routine `PCGASMSetOverlap()` merely allows
1125:   PETSc to extend that overlap further, if desired.

1127: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetSubdomains()`, `PCGASMGetSubKSP()`,
1128:           `PCGASMCreateSubdomains2D()`, `PCGASMGetSubdomains()`
1129: @*/
1130: PetscErrorCode PCGASMSetOverlap(PC pc, PetscInt ovl)
1131: {
1132:   PC_GASM *osm = (PC_GASM *)pc->data;

1134:   PetscFunctionBegin;
1137:   PetscTryMethod(pc, "PCGASMSetOverlap_C", (PC, PetscInt), (pc, ovl));
1138:   osm->dm_subdomains = PETSC_FALSE;
1139:   PetscFunctionReturn(PETSC_SUCCESS);
1140: }

1142: /*@
1143:   PCGASMSetType - Sets the type of restriction and interpolation used
1144:   for local problems in the `PCGASM` additive Schwarz method.

1146:   Logically Collective

1148:   Input Parameters:
1149: + pc   - the preconditioner context
1150: - type - variant of `PCGASM`, one of
1151: .vb
1152:       `PC_GASM_BASIC`       - full interpolation and restriction
1153:       `PC_GASM_RESTRICT`    - full restriction, local MPI process interpolation
1154:       `PC_GASM_INTERPOLATE` - full interpolation, local MPI process restriction
1155:       `PC_GASM_NONE`        - local MPI process restriction and interpolation
1156: .ve

1158:   Options Database Key:
1159: . -pc_gasm_type [basic,restrict,interpolate,none] - Sets `PCGASM` type

1161:   Level: intermediate

1163: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetSubdomains()`, `PCGASMGetSubKSP()`,
1164:           `PCGASMCreateSubdomains2D()`, `PCASM`, `PCASMSetType()`
1165: @*/
1166: PetscErrorCode PCGASMSetType(PC pc, PCGASMType type)
1167: {
1168:   PetscFunctionBegin;
1171:   PetscTryMethod(pc, "PCGASMSetType_C", (PC, PCGASMType), (pc, type));
1172:   PetscFunctionReturn(PETSC_SUCCESS);
1173: }

1175: /*@
1176:   PCGASMSetSortIndices - Determines whether subdomain indices are sorted.

1178:   Logically Collective

1180:   Input Parameters:
1181: + pc     - the preconditioner context
1182: - doSort - sort the subdomain indices

1184:   Level: intermediate

1186: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetSubdomains()`, `PCGASMGetSubKSP()`,
1187:           `PCGASMCreateSubdomains2D()`
1188: @*/
1189: PetscErrorCode PCGASMSetSortIndices(PC pc, PetscBool doSort)
1190: {
1191:   PetscFunctionBegin;
1194:   PetscTryMethod(pc, "PCGASMSetSortIndices_C", (PC, PetscBool), (pc, doSort));
1195:   PetscFunctionReturn(PETSC_SUCCESS);
1196: }

1198: /*@C
1199:   PCGASMGetSubKSP - Gets the local `KSP` contexts for all subdomains on this MPI process.

1201:   Collective iff first_local is requested

1203:   Input Parameter:
1204: . pc - the preconditioner context

1206:   Output Parameters:
1207: + n_local     - the number of blocks on this MPI process or `NULL`
1208: . first_local - the global number of the first block on this process or `NULL`, all processes must request or all must pass `NULL`
1209: - ksp         - the array of `KSP` contexts

1211:   Level: advanced

1213:   Note:
1214:   After `PCGASMGetSubKSP()` the array of `KSP`es is not to be freed

1216:   Currently for some matrix implementations only 1 block per MPI process
1217:   is supported.

1219:   You must call `KSPSetUp()` before calling `PCGASMGetSubKSP()`.

1221: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetSubdomains()`, `PCGASMSetOverlap()`,
1222:           `PCGASMCreateSubdomains2D()`,
1223: @*/
1224: PetscErrorCode PCGASMGetSubKSP(PC pc, PetscInt *n_local, PetscInt *first_local, KSP *ksp[])
1225: {
1226:   PetscFunctionBegin;
1228:   PetscUseMethod(pc, "PCGASMGetSubKSP_C", (PC, PetscInt *, PetscInt *, KSP **), (pc, n_local, first_local, ksp));
1229:   PetscFunctionReturn(PETSC_SUCCESS);
1230: }

1232: /*MC
1233:    PCGASM - Use the (restricted) additive Schwarz method, each block is (approximately) solved with
1234:            its own `KSP` object on a subset of MPI processes

1236:    Options Database Keys:
1237: +  -pc_gasm_total_subdomains <n>                   - Sets total number of local subdomains to be distributed among the MPI processes
1238: .  -pc_gasm_view_subdomains                        - activates the printing of subdomain indices in `PCView()`, -ksp_view or -snes_view
1239: .  -pc_gasm_print_subdomains                       - activates the printing of subdomain indices in `PCSetUp()`
1240: .  -pc_gasm_overlap <ovl>                          - Sets overlap by which to (automatically) extend local subdomains
1241: -  -pc_gasm_type [basic,restrict,interpolate,none] - Sets `PCGASMType`

1243:    Level: beginner

1245:    Notes:
1246:    To set options on the solvers for each block append `-sub_` to all the `KSP`, and `PC`
1247:    options database keys. For example, `-sub_pc_type ilu -sub_pc_factor_levels 1 -sub_ksp_type preonly`

1249:    To set the options on the solvers separate for each block call `PCGASMGetSubKSP()`
1250:    and set the options directly on the resulting `KSP` object (you can access its `PC`
1251:    with `KSPGetPC()`)

1253:    See {cite}`dryja1987additive` and {cite}`1sbg` for details on additive Schwarz algorithms

1255: .seealso: [](ch_ksp), `PCCreate()`, `PCSetType()`, `PCType`, `PC`, `PCASM`, `PCGASMType`, `PCGASMSetType()`,
1256:           `PCBJACOBI`, `PCGASMGetSubKSP()`, `PCGASMSetSubdomains()`,
1257:           `PCSetModifySubMatrices()`, `PCGASMSetOverlap()`, `PCGASMSetType()`
1258: M*/

1260: PETSC_EXTERN PetscErrorCode PCCreate_GASM(PC pc)
1261: {
1262:   PC_GASM *osm;

1264:   PetscFunctionBegin;
1265:   PetscCall(PetscNew(&osm));

1267:   osm->N                        = PETSC_DETERMINE;
1268:   osm->n                        = PETSC_DECIDE;
1269:   osm->nmax                     = PETSC_DETERMINE;
1270:   osm->overlap                  = 0;
1271:   osm->ksp                      = NULL;
1272:   osm->gorestriction            = NULL;
1273:   osm->girestriction            = NULL;
1274:   osm->pctoouter                = NULL;
1275:   osm->gx                       = NULL;
1276:   osm->gy                       = NULL;
1277:   osm->x                        = NULL;
1278:   osm->y                        = NULL;
1279:   osm->pcx                      = NULL;
1280:   osm->pcy                      = NULL;
1281:   osm->permutationIS            = NULL;
1282:   osm->permutationP             = NULL;
1283:   osm->pcmat                    = NULL;
1284:   osm->ois                      = NULL;
1285:   osm->iis                      = NULL;
1286:   osm->pmat                     = NULL;
1287:   osm->type                     = PC_GASM_RESTRICT;
1288:   osm->same_subdomain_solvers   = PETSC_TRUE;
1289:   osm->sort_indices             = PETSC_TRUE;
1290:   osm->dm_subdomains            = PETSC_FALSE;
1291:   osm->hierarchicalpartitioning = PETSC_FALSE;

1293:   pc->data                 = (void *)osm;
1294:   pc->ops->apply           = PCApply_GASM;
1295:   pc->ops->matapply        = PCMatApply_GASM;
1296:   pc->ops->applytranspose  = PCApplyTranspose_GASM;
1297:   pc->ops->setup           = PCSetUp_GASM;
1298:   pc->ops->reset           = PCReset_GASM;
1299:   pc->ops->destroy         = PCDestroy_GASM;
1300:   pc->ops->setfromoptions  = PCSetFromOptions_GASM;
1301:   pc->ops->setuponblocks   = PCSetUpOnBlocks_GASM;
1302:   pc->ops->view            = PCView_GASM;
1303:   pc->ops->applyrichardson = NULL;

1305:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetSubdomains_C", PCGASMSetSubdomains_GASM));
1306:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetOverlap_C", PCGASMSetOverlap_GASM));
1307:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetType_C", PCGASMSetType_GASM));
1308:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMSetSortIndices_C", PCGASMSetSortIndices_GASM));
1309:   PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCGASMGetSubKSP_C", PCGASMGetSubKSP_GASM));
1310:   PetscFunctionReturn(PETSC_SUCCESS);
1311: }

1313: PetscErrorCode PCGASMCreateLocalSubdomains(Mat A, PetscInt nloc, IS *iis[])
1314: {
1315:   MatPartitioning mpart;
1316:   const char     *prefix;
1317:   PetscInt        i, j, rstart, rend, bs;
1318:   PetscBool       hasop, isbaij = PETSC_FALSE, foundpart = PETSC_FALSE;
1319:   Mat             Ad = NULL, adj;
1320:   IS              ispart, isnumb, *is;

1322:   PetscFunctionBegin;
1323:   PetscCheck(nloc >= 1, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "number of local subdomains must > 0, got nloc = %" PetscInt_FMT, nloc);

1325:   /* Get prefix, row distribution, and block size */
1326:   PetscCall(MatGetOptionsPrefix(A, &prefix));
1327:   PetscCall(MatGetOwnershipRange(A, &rstart, &rend));
1328:   PetscCall(MatGetBlockSize(A, &bs));
1329:   PetscCheck(rstart / bs * bs == rstart && rend / bs * bs == rend, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "bad row distribution [%" PetscInt_FMT ",%" PetscInt_FMT ") for matrix block size %" PetscInt_FMT, rstart, rend, bs);

1331:   /* Get diagonal block from matrix if possible */
1332:   PetscCall(MatHasOperation(A, MATOP_GET_DIAGONAL_BLOCK, &hasop));
1333:   if (hasop) PetscCall(MatGetDiagonalBlock(A, &Ad));
1334:   if (Ad) {
1335:     PetscCall(PetscObjectBaseTypeCompare((PetscObject)Ad, MATSEQBAIJ, &isbaij));
1336:     if (!isbaij) PetscCall(PetscObjectBaseTypeCompare((PetscObject)Ad, MATSEQSBAIJ, &isbaij));
1337:   }
1338:   if (Ad && nloc > 1) {
1339:     PetscBool match, done;
1340:     /* Try to setup a good matrix partitioning if available */
1341:     PetscCall(MatPartitioningCreate(PETSC_COMM_SELF, &mpart));
1342:     PetscCall(PetscObjectSetOptionsPrefix((PetscObject)mpart, prefix));
1343:     PetscCall(MatPartitioningSetFromOptions(mpart));
1344:     PetscCall(PetscObjectTypeCompare((PetscObject)mpart, MATPARTITIONINGCURRENT, &match));
1345:     if (!match) PetscCall(PetscObjectTypeCompare((PetscObject)mpart, MATPARTITIONINGSQUARE, &match));
1346:     if (!match) { /* assume a "good" partitioner is available */
1347:       PetscInt        na;
1348:       const PetscInt *ia, *ja;
1349:       PetscCall(MatGetRowIJ(Ad, 0, PETSC_TRUE, isbaij, &na, &ia, &ja, &done));
1350:       if (done) {
1351:         /* Build adjacency matrix by hand. Unfortunately a call to
1352:            MatConvert(Ad,MATMPIADJ,MAT_INITIAL_MATRIX,&adj) will
1353:            remove the block-aij structure and we cannot expect
1354:            MatPartitioning to split vertices as we need */
1355:         PetscInt        i, j, len, nnz, cnt, *iia = NULL, *jja = NULL;
1356:         const PetscInt *row;
1357:         nnz = 0;
1358:         for (i = 0; i < na; i++) { /* count number of nonzeros */
1359:           len = ia[i + 1] - ia[i];
1360:           row = ja + ia[i];
1361:           for (j = 0; j < len; j++) {
1362:             if (row[j] == i) { /* don't count diagonal */
1363:               len--;
1364:               break;
1365:             }
1366:           }
1367:           nnz += len;
1368:         }
1369:         PetscCall(PetscMalloc1(na + 1, &iia));
1370:         PetscCall(PetscMalloc1(nnz, &jja));
1371:         nnz    = 0;
1372:         iia[0] = 0;
1373:         for (i = 0; i < na; i++) { /* fill adjacency */
1374:           cnt = 0;
1375:           len = ia[i + 1] - ia[i];
1376:           row = ja + ia[i];
1377:           for (j = 0; j < len; j++) {
1378:             if (row[j] != i) jja[nnz + cnt++] = row[j]; /* if not diagonal */
1379:           }
1380:           nnz += cnt;
1381:           iia[i + 1] = nnz;
1382:         }
1383:         /* Partitioning of the adjacency matrix */
1384:         PetscCall(MatCreateMPIAdj(PETSC_COMM_SELF, na, na, iia, jja, NULL, &adj));
1385:         PetscCall(MatPartitioningSetAdjacency(mpart, adj));
1386:         PetscCall(MatPartitioningSetNParts(mpart, nloc));
1387:         PetscCall(MatPartitioningApply(mpart, &ispart));
1388:         PetscCall(ISPartitioningToNumbering(ispart, &isnumb));
1389:         PetscCall(MatDestroy(&adj));
1390:         foundpart = PETSC_TRUE;
1391:       }
1392:       PetscCall(MatRestoreRowIJ(Ad, 0, PETSC_TRUE, isbaij, &na, &ia, &ja, &done));
1393:     }
1394:     PetscCall(MatPartitioningDestroy(&mpart));
1395:   }
1396:   PetscCall(PetscMalloc1(nloc, &is));
1397:   if (!foundpart) {
1398:     /* Partitioning by contiguous chunks of rows */

1400:     PetscInt mbs   = (rend - rstart) / bs;
1401:     PetscInt start = rstart;
1402:     for (i = 0; i < nloc; i++) {
1403:       PetscInt count = (mbs / nloc + ((mbs % nloc) > i)) * bs;
1404:       PetscCall(ISCreateStride(PETSC_COMM_SELF, count, start, 1, &is[i]));
1405:       start += count;
1406:     }

1408:   } else {
1409:     /* Partitioning by adjacency of diagonal block  */

1411:     const PetscInt *numbering;
1412:     PetscInt       *count, nidx, *indices, *newidx, start = 0;
1413:     /* Get node count in each partition */
1414:     PetscCall(PetscMalloc1(nloc, &count));
1415:     PetscCall(ISPartitioningCount(ispart, nloc, count));
1416:     if (isbaij && bs > 1) { /* adjust for the block-aij case */
1417:       for (i = 0; i < nloc; i++) count[i] *= bs;
1418:     }
1419:     /* Build indices from node numbering */
1420:     PetscCall(ISGetLocalSize(isnumb, &nidx));
1421:     PetscCall(PetscMalloc1(nidx, &indices));
1422:     for (i = 0; i < nidx; i++) indices[i] = i; /* needs to be initialized */
1423:     PetscCall(ISGetIndices(isnumb, &numbering));
1424:     PetscCall(PetscSortIntWithPermutation(nidx, numbering, indices));
1425:     PetscCall(ISRestoreIndices(isnumb, &numbering));
1426:     if (isbaij && bs > 1) { /* adjust for the block-aij case */
1427:       PetscCall(PetscMalloc1(nidx * bs, &newidx));
1428:       for (i = 0; i < nidx; i++) {
1429:         for (j = 0; j < bs; j++) newidx[i * bs + j] = indices[i] * bs + j;
1430:       }
1431:       PetscCall(PetscFree(indices));
1432:       nidx *= bs;
1433:       indices = newidx;
1434:     }
1435:     /* Shift to get global indices */
1436:     for (i = 0; i < nidx; i++) indices[i] += rstart;

1438:     /* Build the index sets for each block */
1439:     for (i = 0; i < nloc; i++) {
1440:       PetscCall(ISCreateGeneral(PETSC_COMM_SELF, count[i], &indices[start], PETSC_COPY_VALUES, &is[i]));
1441:       PetscCall(ISSort(is[i]));
1442:       start += count[i];
1443:     }

1445:     PetscCall(PetscFree(count));
1446:     PetscCall(PetscFree(indices));
1447:     PetscCall(ISDestroy(&isnumb));
1448:     PetscCall(ISDestroy(&ispart));
1449:   }
1450:   *iis = is;
1451:   PetscFunctionReturn(PETSC_SUCCESS);
1452: }

1454: PETSC_INTERN PetscErrorCode PCGASMCreateStraddlingSubdomains(Mat A, PetscInt N, PetscInt *n, IS *iis[])
1455: {
1456:   PetscFunctionBegin;
1457:   PetscCall(MatSubdomainsCreateCoalesce(A, N, n, iis));
1458:   PetscFunctionReturn(PETSC_SUCCESS);
1459: }

1461: /*@C
1462:   PCGASMCreateSubdomains - Creates `n` index sets defining `n` nonoverlapping subdomains on this MPI process for the `PCGASM` additive
1463:   Schwarz preconditioner for a any problem based on its matrix.

1465:   Collective

1467:   Input Parameters:
1468: + A - The global matrix operator
1469: - N - the number of global subdomains requested

1471:   Output Parameters:
1472: + n   - the number of subdomains created on this MPI process
1473: - iis - the array of index sets defining the local inner subdomains (on which the correction is applied)

1475:   Level: advanced

1477:   Notes:
1478:   When `N` >= A's communicator size, each subdomain is local -- contained within a single MPI process.
1479:   When `N` < size, the subdomains are 'straddling' (process boundaries) and are no longer local.
1480:   The resulting subdomains can be use in `PCGASMSetSubdomains`(pc,n,iss,`NULL`).  The overlapping
1481:   outer subdomains will be automatically generated from these according to the requested amount of
1482:   overlap; this is currently supported only with local subdomains.

1484:   Use `PCGASMDestroySubdomains()` to free the array and the list of index sets.

1486: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetSubdomains()`, `PCGASMDestroySubdomains()`
1487: @*/
1488: PetscErrorCode PCGASMCreateSubdomains(Mat A, PetscInt N, PetscInt *n, IS *iis[])
1489: {
1490:   PetscMPIInt size;

1492:   PetscFunctionBegin;
1494:   PetscAssertPointer(iis, 4);

1496:   PetscCheck(N >= 1, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "Number of subdomains must be > 0, N = %" PetscInt_FMT, N);
1497:   PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)A), &size));
1498:   if (N >= size) {
1499:     *n = N / size + (N % size);
1500:     PetscCall(PCGASMCreateLocalSubdomains(A, *n, iis));
1501:   } else {
1502:     PetscCall(PCGASMCreateStraddlingSubdomains(A, N, n, iis));
1503:   }
1504:   PetscFunctionReturn(PETSC_SUCCESS);
1505: }

1507: /*@C
1508:   PCGASMDestroySubdomains - Destroys the index sets created with
1509:   `PCGASMCreateSubdomains()` or `PCGASMCreateSubdomains2D()`. Should be
1510:   called after setting subdomains with `PCGASMSetSubdomains()`.

1512:   Collective

1514:   Input Parameters:
1515: + n   - the number of index sets
1516: . iis - the array of inner subdomains
1517: - ois - the array of outer subdomains, can be `NULL`

1519:   Level: intermediate

1521:   Note:
1522:   This is a convenience subroutine that walks each list,
1523:   destroys each `IS` on the list, and then frees the list. At the end the
1524:   list pointers are set to `NULL`.

1526:   Fortran Note:
1527:   The arrays are not freed, only the `IS` within the arrays are destroyed

1529: .seealso: [](ch_ksp), `PCGASM`, `PCGASMCreateSubdomains()`, `PCGASMSetSubdomains()`
1530: @*/
1531: PetscErrorCode PCGASMDestroySubdomains(PetscInt n, IS *iis[], IS *ois[])
1532: {
1533:   PetscInt i;

1535:   PetscFunctionBegin;
1536:   if (n <= 0) PetscFunctionReturn(PETSC_SUCCESS);
1537:   if (ois) {
1538:     PetscAssertPointer(ois, 3);
1539:     if (*ois) {
1540:       PetscAssertPointer(*ois, 3);
1541:       for (i = 0; i < n; i++) PetscCall(ISDestroy(&(*ois)[i]));
1542:       PetscCall(PetscFree(*ois));
1543:     }
1544:   }
1545:   if (iis) {
1546:     PetscAssertPointer(iis, 2);
1547:     if (*iis) {
1548:       PetscAssertPointer(*iis, 2);
1549:       for (i = 0; i < n; i++) PetscCall(ISDestroy(&(*iis)[i]));
1550:       PetscCall(PetscFree(*iis));
1551:     }
1552:   }
1553:   PetscFunctionReturn(PETSC_SUCCESS);
1554: }

1556: #define PCGASMLocalSubdomainBounds2D(M, N, xleft, ylow, xright, yhigh, first, last, xleft_loc, ylow_loc, xright_loc, yhigh_loc, n) \
1557:   do { \
1558:     PetscInt first_row = first / M, last_row = last / M + 1; \
1559:     /*                                                                                                    \
1560:      Compute ylow_loc and yhigh_loc so that (ylow_loc,xleft) and (yhigh_loc,xright) are the corners       \
1561:      of the bounding box of the intersection of the subdomain with the local ownership range (local       \
1562:      subdomain).                                                                                          \
1563:      Also compute xleft_loc and xright_loc as the lower and upper bounds on the first and last rows       \
1564:      of the intersection.                                                                                 \
1565:     */ \
1566:     /* ylow_loc is the grid row containing the first element of the local sumbdomain */ \
1567:     *ylow_loc = PetscMax(first_row, ylow); \
1568:     /* xleft_loc is the offset of first element of the local subdomain within its grid row (might actually be outside the local subdomain) */ \
1569:     *xleft_loc = *ylow_loc == first_row ? PetscMax(first % M, xleft) : xleft; \
1570:     /* yhigh_loc is the grid row above the last local subdomain element */ \
1571:     *yhigh_loc = PetscMin(last_row, yhigh); \
1572:     /* xright is the offset of the end of the  local subdomain within its grid row (might actually be outside the local subdomain) */ \
1573:     *xright_loc = *yhigh_loc == last_row ? PetscMin(xright, last % M) : xright; \
1574:     /* Now compute the size of the local subdomain n. */ \
1575:     *n = 0; \
1576:     if (*ylow_loc < *yhigh_loc) { \
1577:       PetscInt width = xright - xleft; \
1578:       *n += width * (*yhigh_loc - *ylow_loc - 1); \
1579:       *n += PetscMin(PetscMax(*xright_loc - xleft, 0), width); \
1580:       *n -= PetscMin(PetscMax(*xleft_loc - xleft, 0), width); \
1581:     } \
1582:   } while (0)

1584: /*@C
1585:   PCGASMCreateSubdomains2D - Creates the index sets for the `PCGASM` overlapping Schwarz
1586:   preconditioner for a two-dimensional problem on a regular grid.

1588:   Collective

1590:   Input Parameters:
1591: + pc       - the preconditioner context
1592: . M        - the global number of grid points in the x direction
1593: . N        - the global number of grid points in the y direction
1594: . Mdomains - the global number of subdomains in the x direction
1595: . Ndomains - the global number of subdomains in the y direction
1596: . dof      - degrees of freedom per node
1597: - overlap  - overlap in mesh lines

1599:   Output Parameters:
1600: + nsub - the number of local subdomains created
1601: . iis  - array of index sets defining inner (nonoverlapping) subdomains
1602: - ois  - array of index sets defining outer (overlapping, if overlap > 0) subdomains

1604:   Level: advanced

1606:   Note:
1607:   Use `PCGASMDestroySubdomains()` to free the index sets and the arrays

1609:   Fortran Notes:
1610:   The `IS` must be declared as an array of length long enough to hold `Nsub` entries

1612: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetSubdomains()`, `PCGASMGetSubKSP()`, `PCGASMSetOverlap()`, `PCASMCreateSubdomains2D()`,
1613:           `PCGASMDestroySubdomains()`
1614: @*/
1615: PetscErrorCode PCGASMCreateSubdomains2D(PC pc, PetscInt M, PetscInt N, PetscInt Mdomains, PetscInt Ndomains, PetscInt dof, PetscInt overlap, PetscInt *nsub, IS *iis[], IS *ois[])
1616: {
1617:   PetscMPIInt size, rank;
1618:   PetscInt    i, j;
1619:   PetscInt    maxheight, maxwidth;
1620:   PetscInt    xstart, xleft, xright, xleft_loc, xright_loc;
1621:   PetscInt    ystart, ylow, yhigh, ylow_loc, yhigh_loc;
1622:   PetscInt    x[2][2], y[2][2], n[2];
1623:   PetscInt    first, last;
1624:   PetscInt    nidx, *idx;
1625:   PetscInt    ii, jj, s, q, d;
1626:   PetscInt    k, kk;
1627:   PetscMPIInt color;
1628:   MPI_Comm    comm, subcomm;
1629:   IS        **xis = NULL, **is = ois, **is_local = iis;

1631:   PetscFunctionBegin;
1632:   PetscCall(PetscObjectGetComm((PetscObject)pc, &comm));
1633:   PetscCallMPI(MPI_Comm_size(comm, &size));
1634:   PetscCallMPI(MPI_Comm_rank(comm, &rank));
1635:   PetscCall(MatGetOwnershipRange(pc->pmat, &first, &last));
1636:   PetscCheck((first % dof) == 0 && (last % dof) == 0, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE,
1637:              "Matrix row partitioning unsuitable for domain decomposition: local row range (%" PetscInt_FMT ",%" PetscInt_FMT ") "
1638:              "does not respect the number of degrees of freedom per grid point %" PetscInt_FMT,
1639:              first, last, dof);

1641:   /* Determine the number of domains with nonzero intersections with the local ownership range. */
1642:   s      = 0;
1643:   ystart = 0;
1644:   for (j = 0; j < Ndomains; ++j) {
1645:     maxheight = N / Ndomains + ((N % Ndomains) > j); /* Maximal height of subdomain */
1646:     PetscCheck(maxheight >= 2, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Too many %" PetscInt_FMT " subdomains in the vertical direction for mesh height %" PetscInt_FMT, Ndomains, N);
1647:     /* Vertical domain limits with an overlap. */
1648:     ylow   = PetscMax(ystart - overlap, 0);
1649:     yhigh  = PetscMin(ystart + maxheight + overlap, N);
1650:     xstart = 0;
1651:     for (i = 0; i < Mdomains; ++i) {
1652:       maxwidth = M / Mdomains + ((M % Mdomains) > i); /* Maximal width of subdomain */
1653:       PetscCheck(maxwidth >= 2, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Too many %" PetscInt_FMT " subdomains in the horizontal direction for mesh width %" PetscInt_FMT, Mdomains, M);
1654:       /* Horizontal domain limits with an overlap. */
1655:       xleft  = PetscMax(xstart - overlap, 0);
1656:       xright = PetscMin(xstart + maxwidth + overlap, M);
1657:       /*
1658:          Determine whether this subdomain intersects this rank's ownership range of pc->pmat.
1659:       */
1660:       PCGASMLocalSubdomainBounds2D(M, N, xleft, ylow, xright, yhigh, first, last, (&xleft_loc), (&ylow_loc), (&xright_loc), (&yhigh_loc), (&nidx));
1661:       if (nidx) ++s;
1662:       xstart += maxwidth;
1663:     } /* for (i = 0; i < Mdomains; ++i) */
1664:     ystart += maxheight;
1665:   } /* for (j = 0; j < Ndomains; ++j) */

1667:   /* Now we can allocate the necessary number of ISs. */
1668:   *nsub = s;
1669:   PetscCall(PetscMalloc1(*nsub, is));
1670:   PetscCall(PetscMalloc1(*nsub, is_local));
1671:   s      = 0;
1672:   ystart = 0;
1673:   for (j = 0; j < Ndomains; ++j) {
1674:     maxheight = N / Ndomains + ((N % Ndomains) > j); /* Maximal height of subdomain */
1675:     PetscCheck(maxheight >= 2, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Too many %" PetscInt_FMT " subdomains in the vertical direction for mesh height %" PetscInt_FMT, Ndomains, N);
1676:     /* Vertical domain limits with an overlap. */
1677:     y[0][0] = PetscMax(ystart - overlap, 0);
1678:     y[0][1] = PetscMin(ystart + maxheight + overlap, N);
1679:     /* Vertical domain limits without an overlap. */
1680:     y[1][0] = ystart;
1681:     y[1][1] = PetscMin(ystart + maxheight, N);
1682:     xstart  = 0;
1683:     for (i = 0; i < Mdomains; ++i) {
1684:       maxwidth = M / Mdomains + ((M % Mdomains) > i); /* Maximal width of subdomain */
1685:       PetscCheck(maxwidth >= 2, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Too many %" PetscInt_FMT " subdomains in the horizontal direction for mesh width %" PetscInt_FMT, Mdomains, M);
1686:       /* Horizontal domain limits with an overlap. */
1687:       x[0][0] = PetscMax(xstart - overlap, 0);
1688:       x[0][1] = PetscMin(xstart + maxwidth + overlap, M);
1689:       /* Horizontal domain limits without an overlap. */
1690:       x[1][0] = xstart;
1691:       x[1][1] = PetscMin(xstart + maxwidth, M);
1692:       /*
1693:          Determine whether this domain intersects this rank's ownership range of pc->pmat.
1694:          Do this twice: first for the domains with overlaps, and once without.
1695:          During the first pass create the subcommunicators, and use them on the second pass as well.
1696:       */
1697:       for (q = 0; q < 2; ++q) {
1698:         PetscBool split = PETSC_FALSE;
1699:         /*
1700:           domain limits, (xleft, xright) and (ylow, yheigh) are adjusted
1701:           according to whether the domain with an overlap or without is considered.
1702:         */
1703:         xleft  = x[q][0];
1704:         xright = x[q][1];
1705:         ylow   = y[q][0];
1706:         yhigh  = y[q][1];
1707:         PCGASMLocalSubdomainBounds2D(M, N, xleft, ylow, xright, yhigh, first, last, (&xleft_loc), (&ylow_loc), (&xright_loc), (&yhigh_loc), (&nidx));
1708:         nidx *= dof;
1709:         n[q] = nidx;
1710:         /*
1711:          Based on the counted number of indices in the local domain *with an overlap*,
1712:          construct a subcommunicator of all the MPI ranks supporting this domain.
1713:          Observe that a domain with an overlap might have nontrivial local support,
1714:          while the domain without an overlap might not.  Hence, the decision to participate
1715:          in the subcommunicator must be based on the domain with an overlap.
1716:          */
1717:         if (q == 0) {
1718:           if (nidx) color = 1;
1719:           else color = MPI_UNDEFINED;
1720:           PetscCallMPI(MPI_Comm_split(comm, color, rank, &subcomm));
1721:           split = PETSC_TRUE;
1722:         }
1723:         /*
1724:          Proceed only if the number of local indices *with an overlap* is nonzero.
1725:          */
1726:         if (n[0]) {
1727:           if (q == 0) xis = is;
1728:           if (q == 1) {
1729:             /*
1730:              The IS for the no-overlap subdomain shares a communicator with the overlapping domain.
1731:              Moreover, if the overlap is zero, the two ISs are identical.
1732:              */
1733:             if (overlap == 0) {
1734:               (*is_local)[s] = (*is)[s];
1735:               PetscCall(PetscObjectReference((PetscObject)(*is)[s]));
1736:               continue;
1737:             } else {
1738:               xis     = is_local;
1739:               subcomm = ((PetscObject)(*is)[s])->comm;
1740:             }
1741:           } /* if (q == 1) */
1742:           idx = NULL;
1743:           PetscCall(PetscMalloc1(nidx, &idx));
1744:           if (nidx) {
1745:             k = 0;
1746:             for (jj = ylow_loc; jj < yhigh_loc; ++jj) {
1747:               PetscInt x0 = (jj == ylow_loc) ? xleft_loc : xleft;
1748:               PetscInt x1 = (jj == yhigh_loc - 1) ? xright_loc : xright;
1749:               kk          = dof * (M * jj + x0);
1750:               for (ii = x0; ii < x1; ++ii) {
1751:                 for (d = 0; d < dof; ++d) idx[k++] = kk++;
1752:               }
1753:             }
1754:           }
1755:           PetscCall(ISCreateGeneral(subcomm, nidx, idx, PETSC_OWN_POINTER, (*xis) + s));
1756:           if (split) PetscCallMPI(MPI_Comm_free(&subcomm));
1757:         } /* if (n[0]) */
1758:       } /* for (q = 0; q < 2; ++q) */
1759:       if (n[0]) ++s;
1760:       xstart += maxwidth;
1761:     } /* for (i = 0; i < Mdomains; ++i) */
1762:     ystart += maxheight;
1763:   } /* for (j = 0; j < Ndomains; ++j) */
1764:   PetscFunctionReturn(PETSC_SUCCESS);
1765: }

1767: /*@C
1768:   PCGASMGetSubdomains - Gets the subdomains supported on this MPI process
1769:   for the `PCGASM` additive Schwarz preconditioner.

1771:   Not Collective

1773:   Input Parameter:
1774: . pc - the preconditioner context

1776:   Output Parameters:
1777: + n   - the number of subdomains for this MPI process (default value = 1)
1778: . iis - the index sets that define the inner subdomains (without overlap) supported on this process (can be `NULL`)
1779: - ois - the index sets that define the outer subdomains (with overlap) supported on this process (can be `NULL`)

1781:   Level: advanced

1783:   Notes:
1784:   The user is responsible for destroying the `IS`s and freeing the returned arrays, this can be done with
1785:   `PCGASMDestroySubdomains()`

1787:   The `IS` numbering is in the parallel, global numbering of the vector.

1789: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetOverlap()`, `PCGASMGetSubKSP()`, `PCGASMCreateSubdomains2D()`,
1790:           `PCGASMSetSubdomains()`, `PCGASMGetSubmatrices()`, `PCGASMDestroySubdomains()`
1791: @*/
1792: PetscErrorCode PCGASMGetSubdomains(PC pc, PetscInt *n, IS *iis[], IS *ois[])
1793: {
1794:   PC_GASM  *osm;
1795:   PetscBool match;
1796:   PetscInt  i;

1798:   PetscFunctionBegin;
1800:   PetscCall(PetscObjectTypeCompare((PetscObject)pc, PCGASM, &match));
1801:   PetscCheck(match, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_WRONG, "Incorrect object type: expected %s, got %s instead", PCGASM, ((PetscObject)pc)->type_name);
1802:   osm = (PC_GASM *)pc->data;
1803:   if (n) *n = osm->n;
1804:   if (iis) PetscCall(PetscMalloc1(osm->n, iis));
1805:   if (ois) PetscCall(PetscMalloc1(osm->n, ois));
1806:   if (iis || ois) {
1807:     for (i = 0; i < osm->n; ++i) {
1808:       if (iis) (*iis)[i] = osm->iis[i];
1809:       if (ois) (*ois)[i] = osm->ois[i];
1810:     }
1811:   }
1812:   PetscFunctionReturn(PETSC_SUCCESS);
1813: }

1815: /*@C
1816:   PCGASMGetSubmatrices - Gets the local submatrices (for this MPI process
1817:   only) for the `PCGASM` additive Schwarz preconditioner.

1819:   Not Collective

1821:   Input Parameter:
1822: . pc - the preconditioner context

1824:   Output Parameters:
1825: + n   - the number of matrices for this MPI process (default value = 1)
1826: - mat - the matrices

1828:   Level: advanced

1830:   Note:
1831:   Matrices returned by this routine have the same communicators as the index sets (`IS`)
1832:   used to define subdomains in `PCGASMSetSubdomains()`

1834: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetOverlap()`, `PCGASMGetSubKSP()`,
1835:           `PCGASMCreateSubdomains2D()`, `PCGASMSetSubdomains()`, `PCGASMGetSubdomains()`
1836: @*/
1837: PetscErrorCode PCGASMGetSubmatrices(PC pc, PetscInt *n, Mat *mat[])
1838: {
1839:   PC_GASM  *osm;
1840:   PetscBool match;

1842:   PetscFunctionBegin;
1844:   PetscAssertPointer(n, 2);
1845:   if (mat) PetscAssertPointer(mat, 3);
1846:   PetscCheck(pc->setupcalled, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_WRONGSTATE, "Must call after KSPSetUp() or PCSetUp().");
1847:   PetscCall(PetscObjectTypeCompare((PetscObject)pc, PCGASM, &match));
1848:   PetscCheck(match, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_WRONG, "Expected %s, got %s instead", PCGASM, ((PetscObject)pc)->type_name);
1849:   osm = (PC_GASM *)pc->data;
1850:   if (n) *n = osm->n;
1851:   if (mat) *mat = osm->pmat;
1852:   PetscFunctionReturn(PETSC_SUCCESS);
1853: }

1855: /*@
1856:   PCGASMSetUseDMSubdomains - Indicates whether to use `DMCreateDomainDecomposition()` to define the subdomains, whenever possible for `PCGASM`

1858:   Logically Collective

1860:   Input Parameters:
1861: + pc  - the preconditioner
1862: - flg - boolean indicating whether to use subdomains defined by the `DM`

1864:   Options Database Key:
1865: + -pc_gasm_dm_subdomains    - configure subdomains
1866: . -pc_gasm_overlap          - set overlap
1867: - -pc_gasm_total_subdomains - set number of subdomains

1869:   Level: intermediate

1871:   Note:
1872:   `PCGASMSetSubdomains()`, `PCGASMSetTotalSubdomains()` or `PCGASMSetOverlap()` take precedence over `PCGASMSetUseDMSubdomains()`,
1873:   so setting `PCGASMSetSubdomains()` with nontrivial subdomain ISs or any of `PCGASMSetTotalSubdomains()` and `PCGASMSetOverlap()`
1874:   automatically turns the latter off.

1876: .seealso: [](ch_ksp), `PCGASM`, `PCGASMGetUseDMSubdomains()`, `PCGASMSetSubdomains()`, `PCGASMSetOverlap()`
1877:           `PCGASMCreateSubdomains2D()`
1878: @*/
1879: PetscErrorCode PCGASMSetUseDMSubdomains(PC pc, PetscBool flg)
1880: {
1881:   PC_GASM  *osm = (PC_GASM *)pc->data;
1882:   PetscBool match;

1884:   PetscFunctionBegin;
1887:   PetscCheck(!pc->setupcalled, ((PetscObject)pc)->comm, PETSC_ERR_ARG_WRONGSTATE, "Not for a setup PC.");
1888:   PetscCall(PetscObjectTypeCompare((PetscObject)pc, PCGASM, &match));
1889:   if (match) {
1890:     if (!osm->user_subdomains && osm->N == PETSC_DETERMINE && osm->overlap < 0) osm->dm_subdomains = flg;
1891:   }
1892:   PetscFunctionReturn(PETSC_SUCCESS);
1893: }

1895: /*@
1896:   PCGASMGetUseDMSubdomains - Returns flag indicating whether to use `DMCreateDomainDecomposition()` to define the subdomains, whenever possible with `PCGASM`

1898:   Not Collective

1900:   Input Parameter:
1901: . pc - the preconditioner

1903:   Output Parameter:
1904: . flg - boolean indicating whether to use subdomains defined by the `DM`

1906:   Level: intermediate

1908: .seealso: [](ch_ksp), `PCGASM`, `PCGASMSetUseDMSubdomains()`, `PCGASMSetOverlap()`
1909:           `PCGASMCreateSubdomains2D()`
1910: @*/
1911: PetscErrorCode PCGASMGetUseDMSubdomains(PC pc, PetscBool *flg)
1912: {
1913:   PC_GASM  *osm = (PC_GASM *)pc->data;
1914:   PetscBool match;

1916:   PetscFunctionBegin;
1918:   PetscAssertPointer(flg, 2);
1919:   PetscCall(PetscObjectTypeCompare((PetscObject)pc, PCGASM, &match));
1920:   if (match) {
1921:     if (flg) *flg = osm->dm_subdomains;
1922:   }
1923:   PetscFunctionReturn(PETSC_SUCCESS);
1924: }