/usr/lib/petscdir/3.1/include/private/vecimpl.h is in libpetsc3.1-dev 3.1.dfsg-11ubuntu1.
This file is owned by root:root, with mode 0o644.
The actual contents of the file can be viewed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 | /*
This private file should not be included in users' code.
Defines the fields shared by all vector implementations.
*/
#ifndef __VECIMPL_H
#define __VECIMPL_H
#include "petscvec.h"
PETSC_EXTERN_CXX_BEGIN
/*S
PetscLayout - defines layout of vectors and matrices across processes (which rows are owned by which processes)
Level: developer
.seealso: PetscLayoutCreate(), PetscLayoutDestroy()
S*/
typedef struct _p_PetscLayout* PetscLayout;
struct _p_PetscLayout{
MPI_Comm comm;
PetscInt n,N; /* local, global vector size */
PetscInt rstart,rend; /* local start, local end + 1 */
PetscInt *range; /* the offset of each processor */
PetscInt bs; /* number of elements in each block (generally for multi-component problems) Do NOT multiply above numbers by bs */
PetscInt refcnt; /* MPI Vecs obtained with VecDuplicate() and from MatGetVecs() reuse map of input object */
};
EXTERN PetscErrorCode PetscLayoutCreate(MPI_Comm,PetscLayout*);
EXTERN PetscErrorCode PetscLayoutSetUp(PetscLayout);
EXTERN PetscErrorCode PetscLayoutDestroy(PetscLayout);
EXTERN PetscErrorCode PetscLayoutCopy(PetscLayout,PetscLayout*);
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutSetLocalSize(PetscLayout,PetscInt);
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutGetLocalSize(PetscLayout,PetscInt *);
PetscPolymorphicFunction(PetscLayoutGetLocalSize,(PetscLayout m),(m,&s),PetscInt,s)
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutSetSize(PetscLayout,PetscInt);
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutGetSize(PetscLayout,PetscInt *);
PetscPolymorphicFunction(PetscLayoutGetSize,(PetscLayout m),(m,&s),PetscInt,s)
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutSetBlockSize(PetscLayout,PetscInt);
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutGetBlockSize(PetscLayout,PetscInt*);
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutGetRange(PetscLayout,PetscInt *,PetscInt *);
EXTERN PetscErrorCode PETSCVEC_DLLEXPORT PetscLayoutGetRanges(PetscLayout,const PetscInt *[]);
/* ----------------------------------------------------------------------------*/
typedef struct _VecOps *VecOps;
struct _VecOps {
PetscErrorCode (*duplicate)(Vec,Vec*); /* get single vector */
PetscErrorCode (*duplicatevecs)(Vec,PetscInt,Vec**); /* get array of vectors */
PetscErrorCode (*destroyvecs)(Vec[],PetscInt); /* free array of vectors */
PetscErrorCode (*dot)(Vec,Vec,PetscScalar*); /* z = x^H * y */
PetscErrorCode (*mdot)(Vec,PetscInt,const Vec[],PetscScalar*); /* z[j] = x dot y[j] */
PetscErrorCode (*norm)(Vec,NormType,PetscReal*); /* z = sqrt(x^H * x) */
PetscErrorCode (*tdot)(Vec,Vec,PetscScalar*); /* x'*y */
PetscErrorCode (*mtdot)(Vec,PetscInt,const Vec[],PetscScalar*);/* z[j] = x dot y[j] */
PetscErrorCode (*scale)(Vec,PetscScalar); /* x = alpha * x */
PetscErrorCode (*copy)(Vec,Vec); /* y = x */
PetscErrorCode (*set)(Vec,PetscScalar); /* y = alpha */
PetscErrorCode (*swap)(Vec,Vec); /* exchange x and y */
PetscErrorCode (*axpy)(Vec,PetscScalar,Vec); /* y = y + alpha * x */
PetscErrorCode (*axpby)(Vec,PetscScalar,PetscScalar,Vec); /* y = alpha * x + beta * y*/
PetscErrorCode (*maxpy)(Vec,PetscInt,const PetscScalar*,Vec*); /* y = y + alpha[j] x[j] */
PetscErrorCode (*aypx)(Vec,PetscScalar,Vec); /* y = x + alpha * y */
PetscErrorCode (*waxpy)(Vec,PetscScalar,Vec,Vec); /* w = y + alpha * x */
PetscErrorCode (*axpbypcz)(Vec,PetscScalar,PetscScalar,PetscScalar,Vec,Vec); /* z = alpha * x + beta *y + gamma *z*/
PetscErrorCode (*pointwisemult)(Vec,Vec,Vec); /* w = x .* y */
PetscErrorCode (*pointwisedivide)(Vec,Vec,Vec); /* w = x ./ y */
PetscErrorCode (*setvalues)(Vec,PetscInt,const PetscInt[],const PetscScalar[],InsertMode);
PetscErrorCode (*assemblybegin)(Vec); /* start global assembly */
PetscErrorCode (*assemblyend)(Vec); /* end global assembly */
PetscErrorCode (*getarray)(Vec,PetscScalar**); /* get data array */
PetscErrorCode (*getsize)(Vec,PetscInt*);
PetscErrorCode (*getlocalsize)(Vec,PetscInt*);
PetscErrorCode (*restorearray)(Vec,PetscScalar**); /* restore data array */
PetscErrorCode (*max)(Vec,PetscInt*,PetscReal*); /* z = max(x); idx=index of max(x) */
PetscErrorCode (*min)(Vec,PetscInt*,PetscReal*); /* z = min(x); idx=index of min(x) */
PetscErrorCode (*setrandom)(Vec,PetscRandom); /* set y[j] = random numbers */
PetscErrorCode (*setoption)(Vec,VecOption,PetscTruth);
PetscErrorCode (*setvaluesblocked)(Vec,PetscInt,const PetscInt[],const PetscScalar[],InsertMode);
PetscErrorCode (*destroy)(Vec);
PetscErrorCode (*view)(Vec,PetscViewer);
PetscErrorCode (*placearray)(Vec,const PetscScalar*); /* place data array */
PetscErrorCode (*replacearray)(Vec,const PetscScalar*); /* replace data array */
PetscErrorCode (*dot_local)(Vec,Vec,PetscScalar*);
PetscErrorCode (*tdot_local)(Vec,Vec,PetscScalar*);
PetscErrorCode (*norm_local)(Vec,NormType,PetscReal*);
PetscErrorCode (*mdot_local)(Vec,PetscInt,const Vec[],PetscScalar*);
PetscErrorCode (*mtdot_local)(Vec,PetscInt,const Vec[],PetscScalar*);
PetscErrorCode (*loadintovector)(PetscViewer,Vec);
PetscErrorCode (*loadintovectornative)(PetscViewer,Vec);
PetscErrorCode (*reciprocal)(Vec);
PetscErrorCode (*viewnative)(Vec,PetscViewer);
PetscErrorCode (*conjugate)(Vec);
PetscErrorCode (*setlocaltoglobalmapping)(Vec,ISLocalToGlobalMapping);
PetscErrorCode (*setvalueslocal)(Vec,PetscInt,const PetscInt *,const PetscScalar *,InsertMode);
PetscErrorCode (*resetarray)(Vec); /* vector points to its original array, i.e. undoes any VecPlaceArray() */
PetscErrorCode (*setfromoptions)(Vec);
PetscErrorCode (*maxpointwisedivide)(Vec,Vec,PetscReal*); /* m = max abs(x ./ y) */
PetscErrorCode (*load)(PetscViewer,const VecType,Vec*);
PetscErrorCode (*pointwisemax)(Vec,Vec,Vec);
PetscErrorCode (*pointwisemaxabs)(Vec,Vec,Vec);
PetscErrorCode (*pointwisemin)(Vec,Vec,Vec);
PetscErrorCode (*getvalues)(Vec,PetscInt,const PetscInt[],PetscScalar[]);
PetscErrorCode (*sqrt)(Vec);
PetscErrorCode (*abs)(Vec);
PetscErrorCode (*exp)(Vec);
PetscErrorCode (*log)(Vec);
PetscErrorCode (*shift)(Vec);
PetscErrorCode (*create)(Vec);
PetscErrorCode (*dotnorm2)(Vec,Vec,PetscScalar*,PetscScalar*);
};
/*
The stash is used to temporarily store inserted vec values that
belong to another processor. During the assembly phase the stashed
values are moved to the correct processor and
*/
typedef struct {
PetscInt nmax; /* maximum stash size */
PetscInt umax; /* max stash size user wants */
PetscInt oldnmax; /* the nmax value used previously */
PetscInt n; /* stash size */
PetscInt bs; /* block size of the stash */
PetscInt reallocs; /* preserve the no of mallocs invoked */
PetscInt *idx; /* global row numbers in stash */
PetscScalar *array; /* array to hold stashed values */
/* The following variables are used for communication */
MPI_Comm comm;
PetscMPIInt size,rank;
PetscMPIInt tag1,tag2;
MPI_Request *send_waits; /* array of send requests */
MPI_Request *recv_waits; /* array of receive requests */
MPI_Status *send_status; /* array of send status */
PetscInt nsends,nrecvs; /* numbers of sends and receives */
PetscScalar *svalues,*rvalues; /* sending and receiving data */
PetscInt *sindices,*rindices;
PetscInt rmax; /* maximum message length */
PetscInt *nprocs; /* tmp data used both during scatterbegin and end */
PetscInt nprocessed; /* number of messages already processed */
PetscTruth donotstash;
PetscTruth ignorenegidx; /* ignore negative indices passed into VecSetValues/VetGetValues */
InsertMode insertmode;
PetscInt *bowners;
} VecStash;
struct _p_Vec {
PETSCHEADER(struct _VecOps);
PetscLayout map;
void *data; /* implementation-specific data */
ISLocalToGlobalMapping mapping; /* mapping used in VecSetValuesLocal() */
ISLocalToGlobalMapping bmapping; /* mapping used in VecSetValuesBlockedLocal() */
PetscTruth array_gotten;
VecStash stash,bstash; /* used for storing off-proc values during assembly */
PetscTruth petscnative; /* means the ->data starts with VECHEADER and can use VecGetArrayFast()*/
};
#define VecGetArray(x,a) ((x)->petscnative ? (*(a) = *((PetscScalar **)(x)->data),0) : VecGetArray_Private((x),(a)))
#define VecRestoreArray(x,a) ((x)->petscnative ? PetscObjectStateIncrease((PetscObject)x) : VecRestoreArray_Private((x),(a)))
/*
Common header shared by array based vectors,
currently Vec_Seq and Vec_MPI
*/
#define VECHEADER \
PetscScalar *array; \
PetscScalar *array_allocated; /* if the array was allocated by PETSc this is its pointer */ \
PetscScalar *unplacedarray; /* if one called VecPlaceArray(), this is where it stashed the original */
/* Default obtain and release vectors; can be used by any implementation */
EXTERN PetscErrorCode VecDuplicateVecs_Default(Vec,PetscInt,Vec *[]);
EXTERN PetscErrorCode VecDestroyVecs_Default(Vec [],PetscInt);
EXTERN PetscErrorCode VecLoadIntoVector_Default(PetscViewer,Vec);
extern PetscInt NormIds[7]; /* map from NormType to IDs used to cache/retreive values of norms */
/* --------------------------------------------------------------------*/
/* */
/* Defines the data structures used in the Vec Scatter operations */
typedef enum { VEC_SCATTER_SEQ_GENERAL,VEC_SCATTER_SEQ_STRIDE,
VEC_SCATTER_MPI_GENERAL,VEC_SCATTER_MPI_TOALL,
VEC_SCATTER_MPI_TOONE} VecScatterType;
/*
These scatters are for the purely local case.
*/
typedef struct {
VecScatterType type;
PetscInt n; /* number of components to scatter */
PetscInt *vslots; /* locations of components */
/*
The next three fields are used in parallel scatters, they contain
optimization in the special case that the "to" vector and the "from"
vector are the same, so one only needs copy components that truly
copies instead of just y[idx[i]] = y[jdx[i]] where idx[i] == jdx[i].
*/
PetscTruth nonmatching_computed;
PetscInt n_nonmatching; /* number of "from"s != "to"s */
PetscInt *slots_nonmatching; /* locations of "from"s != "to"s */
PetscTruth is_copy;
PetscInt copy_start; /* local scatter is a copy starting at copy_start */
PetscInt copy_length;
} VecScatter_Seq_General;
typedef struct {
VecScatterType type;
PetscInt n;
PetscInt first;
PetscInt step;
} VecScatter_Seq_Stride;
/*
This scatter is for a global vector copied (completely) to each processor (or all to one)
*/
typedef struct {
VecScatterType type;
PetscMPIInt *count; /* elements of vector on each processor */
PetscMPIInt *displx;
PetscScalar *work1;
PetscScalar *work2;
} VecScatter_MPI_ToAll;
/*
This is the general parallel scatter
*/
typedef struct {
VecScatterType type;
PetscInt n; /* number of processors to send/receive */
PetscInt *starts; /* starting point in indices and values for each proc*/
PetscInt *indices; /* list of all components sent or received */
PetscMPIInt *procs; /* processors we are communicating with in scatter */
MPI_Request *requests,*rev_requests;
PetscScalar *values; /* buffer for all sends or receives */
VecScatter_Seq_General local; /* any part that happens to be local */
MPI_Status *sstatus,*rstatus;
PetscTruth use_readyreceiver;
PetscInt bs;
PetscTruth sendfirst;
PetscTruth contiq;
/* for MPI_Alltoallv() approach */
PetscTruth use_alltoallv;
PetscMPIInt *counts,*displs;
/* for MPI_Alltoallw() approach */
PetscTruth use_alltoallw;
#if defined(PETSC_HAVE_MPI_ALLTOALLW)
PetscMPIInt *wcounts,*wdispls;
MPI_Datatype *types;
#endif
PetscTruth use_window;
#if defined(PETSC_HAVE_MPI_WIN_CREATE)
MPI_Win window;
PetscInt *winstarts; /* displacements in the processes I am putting to */
#endif
} VecScatter_MPI_General;
struct _p_VecScatter {
PETSCHEADER(int);
PetscInt to_n,from_n;
PetscTruth inuse; /* prevents corruption from mixing two scatters */
PetscTruth beginandendtogether; /* indicates that the scatter begin and end function are called together, VecScatterEnd()
is then treated as a nop */
PetscTruth packtogether; /* packs all the messages before sending, same with receive */
PetscTruth reproduce; /* always receive the ghost points in the same order of processes */
PetscErrorCode (*begin)(VecScatter,Vec,Vec,InsertMode,ScatterMode);
PetscErrorCode (*end)(VecScatter,Vec,Vec,InsertMode,ScatterMode);
PetscErrorCode (*copy)(VecScatter,VecScatter);
PetscErrorCode (*destroy)(VecScatter);
PetscErrorCode (*view)(VecScatter,PetscViewer);
void *fromdata,*todata;
};
EXTERN PetscErrorCode VecStashCreate_Private(MPI_Comm,PetscInt,VecStash*);
EXTERN PetscErrorCode VecStashDestroy_Private(VecStash*);
EXTERN PetscErrorCode VecStashExpand_Private(VecStash*,PetscInt);
EXTERN PetscErrorCode VecStashScatterEnd_Private(VecStash*);
EXTERN PetscErrorCode VecStashSetInitialSize_Private(VecStash*,PetscInt);
EXTERN PetscErrorCode VecStashGetInfo_Private(VecStash*,PetscInt*,PetscInt*);
EXTERN PetscErrorCode VecStashScatterBegin_Private(VecStash*,PetscInt*);
EXTERN PetscErrorCode VecStashScatterGetMesg_Private(VecStash*,PetscMPIInt*,PetscInt**,PetscScalar**,PetscInt*);
/*
VecStashValue_Private - inserts a single value into the stash.
Input Parameters:
stash - the stash
idx - the global of the inserted value
values - the value inserted
*/
PETSC_STATIC_INLINE PetscErrorCode VecStashValue_Private(VecStash *stash,PetscInt row,PetscScalar value)
{
PetscErrorCode ierr;
/* Check and see if we have sufficient memory */
if (((stash)->n + 1) > (stash)->nmax) {
ierr = VecStashExpand_Private(stash,1);CHKERRQ(ierr);
}
(stash)->idx[(stash)->n] = row;
(stash)->array[(stash)->n] = value;
(stash)->n++;
return 0;
}
/*
VecStashValuesBlocked_Private - inserts 1 block of values into the stash.
Input Parameters:
stash - the stash
idx - the global block index
values - the values inserted
*/
PETSC_STATIC_INLINE PetscErrorCode VecStashValuesBlocked_Private(VecStash *stash,PetscInt row,PetscScalar *values)
{
PetscInt jj,stash_bs=(stash)->bs;
PetscScalar *array;
PetscErrorCode ierr;
if (((stash)->n+1) > (stash)->nmax) {
ierr = VecStashExpand_Private(stash,1);CHKERRQ(ierr);
}
array = (stash)->array + stash_bs*(stash)->n;
(stash)->idx[(stash)->n] = row;
for (jj=0; jj<stash_bs; jj++) { array[jj] = values[jj];}
(stash)->n++;
return 0;
}
EXTERN PetscErrorCode VecReciprocal_Default(Vec);
extern PetscLogEvent VEC_View, VEC_Max, VEC_Min, VEC_DotBarrier, VEC_Dot, VEC_MDotBarrier, VEC_MDot, VEC_TDot, VEC_MTDot;
extern PetscLogEvent VEC_Norm, VEC_Normalize, VEC_Scale, VEC_Copy, VEC_Set, VEC_AXPY, VEC_AYPX, VEC_WAXPY, VEC_MAXPY;
extern PetscLogEvent VEC_AssemblyEnd, VEC_PointwiseMult, VEC_SetValues, VEC_Load, VEC_ScatterBarrier, VEC_ScatterBegin, VEC_ScatterEnd;
extern PetscLogEvent VEC_SetRandom, VEC_ReduceArithmetic, VEC_ReduceBarrier, VEC_ReduceCommunication;
extern PetscLogEvent VEC_Swap, VEC_AssemblyBegin, VEC_NormBarrier, VEC_DotNormBarrier, VEC_DotNorm, VEC_AXPBYPCZ, VEC_Ops;
#if defined(PETSC_HAVE_MATLAB_ENGINE)
EXTERN_C_BEGIN
EXTERN PetscErrorCode VecMatlabEnginePut_Default(PetscObject,void*);
EXTERN PetscErrorCode VecMatlabEngineGet_Default(PetscObject,void*);
EXTERN_C_END
#endif
PETSC_EXTERN_CXX_END
#endif
|