Preconditioning multiphysics problems is a complicated issue... but I'll try to shed some light. For further reading there is a fairly lengthy explanation of all of this in the MOOSE Workshop Manual which you can snag here: http://mooseframework.org/documentation/

Note that in MOOSE we separate out the ideas of the "Preconditioning Matrix" (which is what we call a Preconditioner in MOOSE) and the "Preconditioning Process" which is the approximate solve done by something like PETSc to invert the Preconditioning Matrix... for specifying that you utilize PETSc options (look at the Workshop Manual)

The first rule of Preconditioning in MOOSE is to do the simplest thing possible that allows you to solve your problem while you're developing your application. Only after your application is mostly "complete" (meaning you have all of the relevant physics in there and you are basically getting the answers you want) should you start delving into the more complicated preconditioning options.

Because of this we have a hierarchy of Preconditioners spanning from simple to more complicated in this order:

  • Block Diagonal Preconditioning (default)
  • Single Matrix Preconditioner (SMP)
  • Finite Difference Preconditioner (FDP) - only use for debugging
  • Physics Based Preconditioner (PBP)
  • Split Based Preconditioner (Split)

As noted above, the default Preconditioner is a "Block Diagonal Preconditioner". This will basically call your computeQpJacobian() statements and fill in the diagonal block for each variable in your problem. This basically creates a "decoupled" preconditioner that works well on a variety of problems. You can then use PETSc options (as described in the Workshop Manual) to apply PETSc preconditioning processes (like ILU or HYPRE-boomeramg) to that matrix.

Single Matrix Preconditioner (SMP) will fill the block diagonals by default and then allow you to specify some off-diagonal blocks to also fill in. This is accomplished by specifying off_diag_row and off_diag_column. Let's say you have two variables in your system: u and v. And v shows up strongly in the u equation (like a v^2 is in the u PDE somewhere). It might be important to tell the preconditioner about that relationship so you could do it like so:

    type = SMP
    off_diag_row = 'u'
    off_diag_column = 'v'

Note that off_diag_row and off_diag_column are matched "up and down" as in the first thing in off_diag_row goes with the first thing in off_diag_column... the second entries in both go together... the third entries go together... etc.

This will cause the SMP to call your computeQpOffDiagJacobian() statements for your "u" Kernels and ask them to compute the derivative of your PDE terms with respect to "v".... filling in the u-v block of your matrix.

You can then use PETSc options to apply a PETSc preconditioning process to that matrix (see the Workshop Manual for more information).

Finite Difference Preconditioner (FDP) takes the same options as SMP but does not call your computeQpOffDiagJacobian() statements.... instead it does a numerical finite differencing to fill in your preconditioning matrix. This only works in serial and is just there to help you track down problems.

PBP and Split are both "advanced" preconditioners. Each of them allows you to dial up a different PETSc preconditioning process to apply to each equation separately and then try to combine them back together. This allows you to, for instance, use ILU to target a convection dominated PDE in your system while utilizing an Algebraic Multigrid (AMG) preconditioner on your more elliptic looking PDEs.

PBP takes the same options as SMP and FDP for specifying off-diagonal blocks but also takes a solve_order option and a preconditioner option that are matched together. solve_order is the order you want to solve each equation while preconditioner is the type of preconditioning process (like ILU or AMG) that you want to use for that solve. solve is highlighted here because it's really a approximate solve. PBP basically gives you the ability to dial up a "block Gauss-Seidel" solve for your preconditioner.

For examples of all of these you can look in the examples directory at example 11 and 12. You can also look in moose/test/tests/preconditioners for additional examples.

Current rating: 4.5


slope game 2 weeks, 6 days ago

You can then use PETSc options to apply a PETSc preconditioning process to that matrix!!!

Link | Reply
Currently unrated

Bulk SMS 2 weeks, 5 days ago

The SMS is quick and fast enough to reach anywhere from over 80% of the world. Try the SMS service that allows users to send and receive hundreds of SMS numbers from numbers around the world. SMS services, however, are much better and have more jobs than email or phone that require a lot of work and investment. The purchase is cheaper and loved by all.

Link | Reply
Currently unrated

New Comment


required (not published)


R&D 100

Recent Blog Posts

Using our new clang-format Style

8 months, 2 weeks ago

Steady State Algorithm Change

1 year, 5 months ago

Making an App inherit from an App

1 year, 6 months ago

Changing a published API (Deprecation)

1 year, 6 months ago

DiscreteMaterial Objects

1 year, 8 months ago

Recent Posts






RSS / Atom