-
Notifications
You must be signed in to change notification settings - Fork 17
Packages
Packages are Parthenon's way of introducing modularity in the code, separating different physics and features so that they don't interfere with one another or render the code unreadable. Each package has its own C++ namespace for functions, as well as grid variables ("fields"), and a space for parameters/options/globals (a parthenon::Params
object). Packages operate by defining functions, which they then assign as "callbacks," to be invoked at particular points during each step (or at various other times, e.g. before each output file is produced).
Every package should have an Initialize
function in its namespace, which exists to declare the parameters, fields, and callbacks which a package will use. The function should return a KHARMAPackage
object, which is added to a master list of packages accessible from throughout the code.
The KHARMA::ProcessPackages
function is responsible for loading packages by calling each Initialize
. Remember to add your package to this function in order to load it!
ProcessPackages
uses a task list in order to load packages in the correct order. The basic packages which are always loaded are Globals
, Driver
, GRMHD
, Inverter
, Flux
, Reductions
, and Boundaries
. A package for magnetic field is loaded if the problem includes magnetic fields, and the Implicit
package is loaded if any variables are marked with the "Implicit"
flag -- usually for Extended MHD problems (see ImEx driver). Additional packages are loaded as needed, usually because either they are explicitly enabled in the parameters (e.g., Wind
), or they are necessary to calculate some variable (e.g., Current
, CoordinateOutput
).
The Params
class acts quite a bit like a Dictionary in Python, and one is allocated for each package. Members of any type can be added with pkg->Add("name", object)
-- this includes not only the usual int
, Real
, etc. but also structs, objects, or even Kokkos View
s containing any array-type information which should not be block-attached (though this last use is discouraged in KHARMA if it can be avoided).
Unless marked private
, variables can be accessed by nearly and function in KHARMA via the Mesh or MeshBlock pointer, usually called pmb
. This is a feature, not a bug -- it allows any function to access parameters from a categorized global repository, and acts as an alternative to dangerous global variables or cumbersome extra function parameters. For example, one should access the fluid adiabatic index gamma
via the GRMHD
package:
auto gam = pmb->packages.Get("GRMHD")->Param<Real>("gamma");
and this is done all across the code.
In additional to global constants that are available through the params
object, packages can also include variables that are defined over the entire mesh (also called fields). These can be scalar, vectors or even tensor objects, attached to each cell center, face, or edge of each zone, or free of the grid structure entirely. Each variable carries metadata flags, which contains all the information about how a variable should be allocated and handled (size, datatype, whether the variable should be synchronized at MPI communications, and so on).
Metadata m_real = Metadata({Metadata::Real, Metadata::Cell, Metadata::Derived, Metadata::OneCopy});
pkg->AddField("solve_norm", m_real);
Here we have defined an object of Parthenon's Metadata
class called m_real
by calling Metadata
's constructor. We've provided a MetadataFlag
vector that specifies four flags, Metadata::Real
tells it that the field defined by this metadata will consist of Reals, Metadata::Cell
clarifies that array of Reals will reside at the cell centers, Metadata::Derived
explains that the data in this array can be computed on-the-fly from a more fundamental state of variables, and finally Metadata::OneCopy
tells Parthenon that any field defined with this metadata will be shared between stages (recall that we have a multistage driver). The Parthenon docs have a page that lists all the metadata flags available.
The constructor takes an optional second argument, a std::vector<int>
defining a shape for the value at each location the field is defined. Thus a 3-vector might pass std::vector<int>{3}
, or a metric tensor std::vector<int>{4, 4}
. A separate flag Metadata::Vector
indicates that the field represents a spatial vector -- that is, that the second component should be inverted at reflecting boundaries. The second line in the code excerpt above defines a field solve_norm
in the implicit package with m_real
as its metadata. "solve_norm"
is intended for output so it is named succinctly -- generally, fields that will only be used within one package should be preceded by the package name, for example "Flux.vr"
for the velocity at right of faces, used within the Flux package.
User flags may be added with Metadata::AddUserFlag("FlagName")
and retrieved with Metadata::GetUserFlag("FlagName")
. Flags are useful for categorizing variables, to pack only those needed for each kernel.
Each package can additionally register functions to be called at certain points during each step, e.g. to add a source term, update fluxes before they are applied, etc, etc. These callbacks are intended to be the primary way most packages operate, rather than adding code to the different Driver functions explicitly.
The callbacks available are:
// PHYSICS
// Recovery of primitive variables from conserved.
// These can be host-side functions because they are not called from GetFlux()
// rather, they are called on zone center values once per step only.
std::function<void(MeshBlockData<Real>*, IndexDomain, bool)> BlockUtoP = nullptr;
std::function<void(MeshData<Real>*, IndexDomain, bool)> MeshUtoP = nullptr;
// Allow applying UtoP only/separately for boundary domains after sync/prolong/restrict ops
// All packages with independent variables should register this for AMR
std::function<void(MeshBlockData<Real>*, IndexDomain, bool)> BoundaryUtoP = nullptr;
// On domain boundaries, however, we sometimes need to respect the primitive variables.
// Currently only the GRMHD primitives (rho, u, uvec) do this
std::function<void(MeshBlockData<Real>*, IndexDomain, bool)> DomainBoundaryPtoU = nullptr;
// Going the other way, however, is handled by Flux::{Block,Mesh}PtoU.
// All PtoU implementations are device-side (called prim_to_flux),
// so we do not need something like
//std::function<void(MeshBlockData<Real>*, IndexDomain, bool)> BlockPtoU = nullptr;
// Source term to add to the conserved variables during each step
std::function<void(MeshData<Real>*, MeshData<Real>*, IndexDomain)> AddSource = nullptr;
// Source term to apply to primitive variables, needed for some problems in order
// to control dissipation (Hubble, turbulence).
// Must be applied over entire domain!
std::function<void(MeshBlockData<Real>*)> BlockApplyPrimSource = nullptr;
// Apply any fixes after the initial fluxes are calculated
std::function<void(MeshData<Real>*)> FixFlux = nullptr;
// Apply any floors or limiters specific to the package (that is, on the package's variables)
std::function<void(MeshBlockData<Real>*, IndexDomain)> BlockApplyFloors = nullptr;
std::function<void(MeshData<Real>*, IndexDomain)> MeshApplyFloors = nullptr;
// CONVENIENCE
// Anything to be done before each step begins -- currently just updating global "in_loop"
std::function<void(Mesh*, ParameterInput*, const SimTime&)> PreStepWork = nullptr;
// Anything to be done after every step is fully complete -- usually reductions or preservation of variables
// Note that most diagnostics should go in "PostStepDiagnosticsMesh" instead
std::function<void(Mesh*, ParameterInput*, const SimTime&)> PostStepWork = nullptr;
// Anything to be done just before any outputs (dump files, restarts, history files) are made
// Usually for filling output-only variables
// TODO Add MeshUserWorkBeforeOutput to Parthenon
std::function<void(MeshBlock*, ParameterInput*)> BlockUserWorkBeforeOutput = nullptr;
// Anything at the very end of simulation. Cleanup, summaries, outputs if you're brave
std::function<void(Mesh*, ParameterInput*, const SimTime&)> PostExecute = nullptr;
KHARMA's packages can also register any of the callbacks used by generic Parthenon StateDescriptor
(a.k.a. package) objects. The full StateDescriptor
documentation is here.
Note KHARMA has three different kinds of callbacks for recovering primitive variables: {Block,Mesh}UtoP
, BoundaryUtoP
and DomainBoundaryPtoU
. These are for flexibility when handling boundary synchronizations in your package: UtoP
is called over the whole domain once per step, and provides most of the primitive variable values for reconstruction to cell faces and calculation of resulting fluxes in the next step. If your package maintains separate primitive and conserved variables at all, you likely need at least this. The other two calls are only used when synchronizing conserved variables between meshblocks/MPI ranks, as is default for the kharma
step (but not imex
! You can ignore them if using only the imex
step). BoundaryUtoP
is called over ghost zones (i.e. with domain == IndexDomain::inner_x1
etc etc) during each boundary synchronization, after conserved variables are synchronized. That is, by default each mesh block will call this function six times for its six faces, regardless of whether they border other meshblocks or the edge of the domain. DomainBoundaryPtoU
, in contrast, will be called only on domain boundaries (i.e., the reflecting boundary at the pole or outflow boundaries at the domain edges). It overrides BoundaryUtoP
, if both are registered for the same package -- indeed, you should only need it for overriding BoundaryUtoP
, not for any other purpose. DomainBoundaryPtoU
exists so that even when synchronizing conserved variables, the outflow or reflecting boundaries can be applied to the primitive variables instead -- by default KHARMA does this for the GRMHD fluid variables, as is common practice. (Note that if you want to use this, your primitive variables will have to be marked Metadata::FillGhost
to be accessible!)
As noted in the listing, there are no callbacks for PtoU
, because it is integrated into the kernels computing fluxes of conserved variables between zones. Instead, one should write a device-side function prim_to_flux
and add a call to your function in Flux::prim_to_flux
in flux/flux_functions.hpp
.
KHARMA's structure is maybe best summarized with the dependency structure of its many packages:
Most packages are fairly small, so we detail the layers of this structure together. A few large packages make large changes/additions to the algorithm and deserve individual documentation, covered in the next section.
The core level consists of mostly convenience features, parameters, and common datatypes -- anything common to the whole code, which could be used in any other package. Among these only Globals
, Reductions
, and Boundaries
are formal packages.
The Package
core module/namespace defines the standard interface for packages in KHARMA, which adds features to the Parthenon definition for convenience. KHARMA provides several new callbacks for its packages, structured around common points in a GRMHD update step; e.g., for adding source terms, operator-split physics, or conversions between primitive and conserved variables.
Other core modules house common data structures and code for easily calculating index ranges, as well as code for performing reductions among MPI ranks, used for diagnostics and summary statistics by several different packages. The core modules also include a Globals
package, which provides more universal access to a couple of useful variables, such as the current simulation time and state (initializing vs. stepping).
Finally, the core includes a differential-geometry-aware coordinates object, GRCoordinates
, which takes the place of Parthenon's built-in coordinates. A new GRCoordinates
object is created automatically with each MeshBlock
, providing data for locations within that block. In addition to cell locations in the code's coordinates, as provided by Parthenon, GRCoordinates
introduces the concept of a coordinate transformation between the code's coordinates (in which zones are evenly spaced) and a system of "embedding" coordinates representing the basic coordinates for the spacetime (e.g. Spherical Kerr-Schild). The GRCoordinates
object also carries caches of expensive-to-calculate quantities such as the metric and connection coefficients.
The next level of abstraction implements reconstruction schemes, Riemann solvers (in the Flux
package), and semi-implicit stepping (in the Implicit
package). Together, these are intended to form a framework for solving any hyperbolic system in curved spacetime via finite volumes, allowing for the same code to be shared between different schemes with or without magnetic fields, viscous effects, radiative feedback, etc.
This separation is imperfect still, as the Riemann solver requires determining sound speeds and states of conserved variables from primitive variables, both of which are theory-dependent but must live in the same kernels as the Riemann solve itself, to maintain performance. However, the distinction is enough that KHARMA supports a few theories (hydrodynamics, magnetohydrodynamics, and extended MHD) in a relatively natural way with a minimum of code duplication.
The Flux
package also handles correcting any fluxes which will lead to floor hits or primitive recovery failures (i.e., first-order flux corrections).
The implicit solver is written to be general, allowing any variables marked with the tag Metadata::Implicit
, to be evolved semi-implicitly. Any time-dependent source terms acting on the new variable must still be added manually, however. The implicit solver is described in detail below.
The GRMHD
package handles updating the primitive hydrodynamic fluid variables associated with a HARM scheme -- density, internal energy density, and velocity. It does not evolve the magnetic field components, however. Thus, while it optionally uses a magnetic field to compute conserved variables like stress-energy tensor (and the signal speeds, etc), it is agnostic to the magnetic field transport scheme, and supports several different transports implemented in different packages.
When evolving the GRMHD variables explicitly, the Inverter
package handles the numerical inversion of conserved to primitive fluid quantities (when evolved implicitly, the variable inversion is done automatically by the nonlinear solver). As they share substantial structure and code, this package implements both of the available primitive variable recovery schemes, which can be selected at runtime.
The Floors
package, implementing various density and internal energy floors and injection frames, as well as a number of variable limits, also lives alongside/within the GRMHD algorithm.
There are a few additional packages which enhance the main GRMHD package. Generally, these implement any additional physics beyond evolving the core GRHD variables: electron temperatures, dissipative effects, and so on.
These packages also include the magnetic field transports, solutions for evolving the magnetic field while preserving low divergence. KHARMA supports several different magnetic field transport schemes, but modern simulations generally use the face-centered constrained transport scheme unless another package is required for backward compatibility, testing, or speed.
Time-stepping is orchestrated by the Driver
, which constructs a list of tasks constituting a single simulation sub-step. KHARMA implements three different time-steps, described in The Driver.
Fluid initialization code can also take advantage of any package: different problems can require or make optional whatever fields make sense in context. Where there is a common, standard initialization which makes sense (i.e. a nominal "cold" value for electron temperatures which will be heated to equilibrium by the fluid) this is applied if the package is loaded but the field is not otherwise initialized. Initialization routines which require fields not enabled by the user generate errors.
Since different magnetic field transports preserve different representations of the magnetic field divergence, initialization of the magnetic field is dependent on what field transport will be used, and thus is handled separately from problem initialization. Where problems require a specific magnetic field configuration, they can set parameters for this later initialization, so as not to have to implement two magnetic field initializations per problem.
While it contains quite a few different modules, KHARMA is not (yet) a terribly complicated code overall -- the figure lays out almost all of the code modules and internal dependencies. Freed from implementing the extensive bookkeeping of mesh refinement or specialized syntax of device code, KHARMA consists almost exclusively of physics-related code, which makes up 3 of the 4 main layers.
The implicit solver is essentially a multidimensional Newton-Raphson scheme. In its current form it finds the root for the residual of the system of balance laws,
As is the case for the multidimensional Newton-Raphson scheme, the solver computes a Jacobian
NOTE: This is a local solve ie., it minimizes the residual and computes the update to the primitives on a zone-by-zone basis.
The details of the scheme are present in section 3 of this paper.
The Implicit
package defines a bunch of constants that are relevant to the implicit solve. These are,
-
min_nonlinear_iter
(default:1
) - The minimum number of iterations that must be performed by the solver. -
max_nonlinear_iter
(default:3
) - The maximum number of iterations that must be performed by the solver. -
rootfind_tol
(default:1.e-12
) - The tolerance that must be met during the solve to consider it a success. -
jacobian_delta
(default:4.e-8
) - The small change considered for each primitive -$\Delta P$ - to compute the Jacobian. -
use_qr
(default:true
) - A boolean that dictates whether the QR decomposition should be considered while solving the system of linear equations. Iffalse
, the solver uses LU decomposition. While QR has certain stability benefits, we've found that it can lead to potentially inaccurate results when performing a >= 8x8 solve on GPUs. -
linesearch
(default:true
) - A boolean that decides if backtracking linesearch must be performed.
The following arguments are relevant only if linesearch has been enabled.
-
max_linesearch_iter
(default:3
) - The maximum number of iterations that must be performed by the linesearch. -
linesearch_eps
(default:1.e-4
) - A parameter that determines if the rate of decrease of$\boldsymbol{R}$ is small enough to exit linesearch. -
linesearch_lambda
(default:1.0
) - The initial value of the step length.
Additionally, the Implicit
package defines to two scalar fields,
-
solve_norm
- The L2 norm of the residual. -
solve_fail
- It is possible that the solver may not converge to a physically viable solution in zones where the loss landscape is difficult to navigate. In these scenarios, we manually backtrack by setting the step length to 0.1, recording the failure by settingsolve_fail
to1
, and performing the linesearch. In the event the manual backtrack is not sufficient, we setsolve_fail
to2
and exit the solver (for that particular zone). Later, we consider the average of the primitives of the neighboring zones that did converge.
The package contains a single task Step
that performs the solve over the implicit variables. This task is called by the ImexDriver
after the explicit variables are updated. Since we are only concerned with evolving the implicit variables, we need to reorder the variables in relevant fluid states. We achieve this with the package's get_ordered_names
function. It does this by leveraging metadata flags. The function creates a vector of MetadataFlag
by picking variables marked with the ImplicitFlag
first and then those marked with ExplicitFlag
. We then pack the variables in this order in all fluid states via Parthenon's PackVariables
function, and finally generate a VarMap
to refer to the right variable index. VarMaps are explained in detail here.
The syntax for the rest of the task is much like KHARMA's GetFlux
. We first estimate the scratch memory required and launch a Kokkos kernel (aka. parallel dispatch). Inside the outer for loop (across blocks and along X3 and X2) we allocate temporaries using Parthenon's ScratchPad and initialize them using the corresponding MeshBlockPack
. In the inner for loop (along X1) we define Kokkos subviews - slices over the scratchpads - that are lightweight and faster. The remainder of the computational body carries out the necessary operations needed to update the primitives,
- Compute and save implicit source terms. Currently, we only include source terms for the EMHD variables, but one can just as easily include function calls for other implicit sources.
- Compute the Jacobian.
- Solve the system of equations (refer this) for the update
$\boldsymbol{\delta P}$ . - Perform linesearch to estimate the right step length.
- Update the primitives, calculate the residual and its L2 norm.
Additionally, the package has two device side functions calc_residual
and calc_jacobian
that are defined in the package's header file and as the name implies compute the residual and jacobian respectively.
DISCLAIMER: This package is in a developmental phase.
The EMHD package includes anisotropic heat conduction and momentum transport. The theory is detailed in this work and in the theory subsection we present just the governing equations with variable definitions for brevity.
The evolution equations for the extended magnetohydrodynamics model are,
where the first three equations are similar to those for ideal magnetohydrodynamics (see here), and the latter two dictate the evolution of the new non-ideal variables - conductive heat flux scalar
This is done to increase the stability of the numerical implementation of these equations in regions of low density.
Note that the stress-energy tensor
The magnetic field variables are updated explicitly prior to the implicit solver. Therefore, a full non-ideal run contains 7 variables to be updated implicitly and it utilizes the newly updated B-field.
We now elucidate how the various source terms are computed ie., explicitly or implicitly. The source terms for the internal energy and velocity field are evaluated explicitly in Flux::AddSource
prior to the solve. Since the evolution equations for the rest-mass density and magnetic fields contain no source terms we are done calculating all the source terms for the ideal MHD sector. For the two remaining equations corresponding to the non-ideal equations we selectively evaluate some explicitly and some implicitly. Let's see the source terms for
What is worth noting is that
The terms in blue are treated implicitly and the rest are computed explicitly prior to the solve. The first term on the RHS has a factor of EMHD::implicit_sources
). The remainder of the implicit terms contain a time derivative and NEED to be calculated within the solver as the fluid state is iteratively updated (EMHD::time_derivative_sources
). The terms in black are computationally expensive and do not warrant an implicit update; they are therefore evaluated prior to the solve via EMHD::AddSource
and passed as arguments to Implicit::Step
.
Many of the constants in the EMHD
package have to do with the closure scheme used to compute the transfer coefficients. We express this as,
where conduction_alpha
and viscosity_alpha
respectively.
-
conduction_alpha
(default:1.
) - The closure parameter for thermal diffusivity. -
viscosity_alpha
(default:1.
) - The closure parameter for kinematic viscosity. -
tau
(default:1.
) - Relaxation timescale. Must be provided for test problems. It is computed for the torus problems based on instability conditions. -
closure_type
(default:torus
) - The closure scheme to be used. This is a misnomer of sorts since ultimately we use the scheme mentioned above but a few additional operations may need to be performed either to compute$\tau_{R}$ or if$\kappa$ or$\eta$ are provided instead (for eg. in the conducting atmosphere and viscous Bondi problem). -
kappa
(default:1.
) - Thermal conductivity. -
eta
(default:1.
) - Dynamic viscosity. -
higher_order_terms
(default:false
) - Should second order terms be considered? -
feedback
(default:true
) - Should the EMHD sector feedback on to the ideal MHD variables? We generally want this enabled to get the physics right but there may be cases where one might want to disable it (for eg., the viscous Bondi problem or for debugging purposes).
With the introduction of two new variables, the EMHD
package has four new scalar fields,
-
prims.q
-$\tilde{q}$ -
prims.dP
-$\Delta\tilde{P}$ -
cons.q
-$\sqrt{-g}\tilde{q}u^{t}$ -
cons.dP
-$\sqrt{-g}\Delta\tilde{P}u^{t}$
The EMHD::AddSource
task is called by the ImEx driver prior to an implicit update, it computes all the explicit source terms associated with the evolution equations of the non-ideal variables. These source terms contain temperature and velocity gradients (cf. the decomposition of source terms as implicit/explicit above). This means we need to reconstruct these variables at the zone faces to obtain a second order slope value to zone centers. To facilitate this, the EMHD package has a bunch of auxiliary functions present in emhd_utils.hpp
.
KHARMA does not cache scalar fields such as set_parameters
function in the header file computes the transport coefficients based on the closure scheme, while the convert_prims_to_q_dP
function calculates calc_tensor
that is called when source terms or fluxes need to be estimated.