, including all inherited members.
AlocateMemoryByPattern(IConectMatrix *bskl) | SparseGridMtxLU | |
BackSubstU(double *x, long fixed_blocks) | SparseGridMtxLU | |
block_order | SparseGridMtx | |
block_size | SparseGridMtx | |
block_storage | SparseGridMtx | |
BlockArith | SparseGridMtx | |
blocks | SparseGridMtx | [protected] |
Blocks() | SparseGridMtx | [inline] |
BlockSize() | SparseGridMtx | [inline] |
clock_start | TraceableMatrix | [protected] |
Columns | SparseGridMtx | |
Columns_data | SparseGridMtxLU | |
columns_data_length | SparseGridMtx | |
ComputeBlocks() | SparseGridMtx | |
CS() | TraceableMatrix | |
Diagonal_data | SparseGridMtxLU | |
ElementAt(int i, int j) | SparseGridMtxLU | [virtual] |
eMT | TraceableMatrix | |
Factorize() | SparseGridMtxLU | [virtual] |
ForwardSubstL(double *x, long fixed_blocks) | SparseGridMtxLU | |
GetValue(long bi, long bj, long si, long sj, long &aux_bi_idx, long &aux_bj_idx) | SparseGridMtxLU | [inline] |
GetWaste() | SparseGridMtx | |
LoadMatrixNumbers(SparseMatrixF &sm) | SparseGridMtxLU | [virtual] |
LoadZeros() | SparseGridMtxLU | [virtual] |
MC_() | TraceableMatrix | |
MT | TraceableMatrix | |
MultiplyByVector(const LargeVectorAttach &x, LargeVectorAttach &y) | SparseGridMtxLU | [virtual] |
N() const | SparseGridMtx | [inline, virtual] |
n | SparseGridMtx | [protected] |
n_blocks | SparseGridMtx | [protected] |
N_blocks() | SparseGridMtx | [inline] |
No_Multiplications() | SparseGridMtx | [inline, virtual] |
no_multiplications | SparseGridMtx | [protected] |
node_order | SparseGridMtx | |
noDummyDOFs | SparseGridMtx | [protected] |
nonzeros | SparseGridMtx | [protected] |
Nonzeros() const | SparseGridMtx | [inline, virtual] |
Rows_data | SparseGridMtxLU | |
SchurComplementFactorization(int fixed_blocks) | SparseGridMtxLU | [virtual] |
Solve(double *b, double *x) | SparseGridMtxLU | [virtual] |
SolveA11(double *x, long fixed_blocks) | SparseGridMtxLU | [virtual] |
SolveLU(double *x, long fixed_blocks=0) | SparseGridMtxLU | |
SolveLV(const LargeVector &b, LargeVector &x) | SparseGridMtxLU | [virtual] |
SparseGridMtx(SparseMatrixF &sm, BYTE block_size, Ordering *block_order, MathTracer *eMT) | SparseGridMtx | |
SparseGridMtx(SparseMatrixF &sm, BYTE block_size, Ordering *block_order, Ordering *node_order, MathTracer *eMT) | SparseGridMtx | |
SparseGridMtxLU(SparseMatrixF &sm, BYTE block_size, Ordering *block_order, MathTracer *eMT, BOOL load_data=1) | SparseGridMtxLU | |
SparseGridMtxLU(SparseMatrixF &sm, BYTE block_size, Ordering *block_order, Ordering *node_order, MathTracer *eMT, BOOL load_data=1) | SparseGridMtxLU | |
Sub_A11inv_A12(double *x, long fixed_blocks) | SparseGridMtxLU | [virtual] |
Sub_A21_A11inv(double *x, long fixed_blocks) | SparseGridMtxLU | [virtual] |
SubMultL21(double *px, double *py, long fixed_blocks) | SparseGridMtxLU | |
SubMultU12(double *px, double *py, long fixed_blocks) | SparseGridMtxLU | |
temporary_measure_start | TraceableMatrix | [protected] |
TraceableMatrix() | TraceableMatrix | |
Write(char *cmd) | TraceableMatrix | |
WriteCondensedMatrixA22(double *a, Ordering *mcn, IntArrayList *lncn) | SparseGridMtxLU | [virtual] |
Writeln(char *cmd) | TraceableMatrix | |
WriteStatistics(long no_init_blocks, long no_nonzeros) | SparseGridMtx | [virtual] |
~IMatrix() | IMatrix | [inline, virtual] |
~SparseGridMtx() | SparseGridMtx | [virtual] |
~SparseGridMtxLU() | SparseGridMtxLU | [virtual] |
~TraceableMatrix() | TraceableMatrix | [inline, virtual] |