Model Exploration
ABMs are too heterogeneous to be managed and explored in a traditional way, because each model has its agents, its ecosystem and variadic number of parameters. You need to write code able to generate code: metaprogramming, and in Rust is possible using macros.
Rust language has a powerful macro system, based on two categories: declarative and procedural. Procedural macros need to be written in a specific library, and cannot be directly included in krABMaga, so we chose declarative ones, which are also the most used.
Our framework provides several algorithms to explore parameter space:
Because the number of simulations to run can easily explode based on the parameters, the amount of values generated, the algorithm and its settings, we provide parallel, distributed (using MPI) and "on Cloud" (using AWS API) explorations to run multiple simulations simultaneously. At the moment, they are available for Parameter Sweeping and Genetic Algorithm.
Model Exploration with MPI is based on rsmpi
, an MPI binding to Rust. Of course to use it you need a MPI distribution installed on your machine(s).
rsmpi
officially supports:
- OpenMPI 4.0.3 on Ubuntu-20.04, 4.1.2 on macOS.
- MPICH 3.3.2 on Ubuntu 20.04.
- MS-MPI (Windows) 10.1.1 on Windows 2022.
To use exploration with MPI you have to enable distributed-mpi
feature.
Parameter Sweeping§
Runs a model many times, varying the model’s settings based on a discrete range of values and recording the results.
For each parameter, the user has to call a macro to generate the set of values to test on different simulation configurations: with macro, user can pass types as a parameter, so the same code can be adapted for any parameter type.
Generated parameters, the user can call the macro that provides the exploration: the macro creates an ad hoc dataframe to store all configurations and output of each run, define all configurations (all possible combinations) and finally run the simulation, thanks to another macro able to schedule steps of simulations.
The only restriction is defining input and output parameters inside your State, and inout parameter names need to match with generated ones. For example, if you have and input parameter called X of type u32 inside State, you have to generate a subset of u32 values and store in a Vec called X. By default, the exploration is sequential.
// Sequential scenario
// Computing mode for sequential isn't mandatory. See below.
let result = explore!(
STEP, REPS, State, // Simulation Step, Repetitions for each configuration, name of your State struct
input{
par1: u32 // Parameters generated
par2: f64
},
output [ output: f64],
ExploreMode::Matched
);
// Parallel scenario
let result = explore!(
STEP, REPS, State, // Simulation Step, Repetitions for each configuration, name of your State struct
input{
par1: u32 // Parameters generated
par2: f64
},
output [ output: f64],
ExploreMode::Matched,
ComputingMode::Parallel // Distributed or Cloud as other options
);
There are two Explore Mode options:
- Exaustive, the standard mode;
- Matched, if you have to test only specific combinations;
Evolutionary Search§
krABMaga provides a macro to optimize paramaters of your model using an evolutioary searching strategy. A genetic algorithm is an heuristic search that is inspired by Charles Darwin’s theory of natural evolution. Starting from an initial population, at each generation this algorithm selects the fittest individuals to create offspring of next generation. An individual is a string called chromosome that, in this case, represents a combination of parameters. A single parameter is called gene.
To use this algorithm we need to define 5 functions:
- Init population;
- Fitness function to evaluate an individual, so that you can compare if one is better than one other;
- Selection to select which part of population can survive and/or can be used to create the offspring;
- Crossover to combine genes of a couple of individuals;
- Mutation to apply some little changes to offspring genes with a low random probability;
By default, the search is sequential.
// Sequential scenario
// Computing mode for sequential isn't mandatory. See below.
let result = evolutionary_search!(
init_population,
fitness,
selection, //with macro u can pass function you define
mutation,
crossover,
cmp, // function to compare two fitness
State,
DESIRED_FITNESS,
MAX_GENERATION,
STEP,
REPETITIONS, // optional
);
// Parallel scenario
let result = evolutionary_search!(
init_population,
fitness,
selection, //with macro u can pass function you define
mutation,
crossover,
cmp, // function to compare two fitness
State,
DESIRED_FITNESS,
MAX_GENERATION,
STEP,
REPETITIONS, // optional
ComputingMode::Parallel // Distributed or Cloud as other options
);
To use model exploration using Genetic Algorithms you have to pass several functions to krABMaga macro. In this way modelist can use specific functions based on his model or needs.
More details are available in the example SIR
Bayesian Optimization§
Bayesian optimization is a global optimization strategy. It is very useful to evaluate black-boxes, continuous functions that don't assume any functional forms, and functions with an high computational cost. Both are features of a simulation.
This strategy is based on two functions:
- Surrogate Function: Bayesian approximation of the objective function that can be sampled efficiently;
- Acquisition Function: Technique by which the posterior probability is used to select the next sample from the search space;
let (x, y) = bayesian_opt!(
init_population, // initial points to setup algorithm
costly_function, // setup and exectution of a simulation, returning a cost
acquisition_function,
get_points, // generate point from parameter space as base of the iteration
check_domain, // check the point returned from an iteration of the algorithm
ITERATIONS,
);
To use model exploration based on Bayesian Optimization you have to pass several functions to krABMaga macro. In this way modelist can use specific functions based on his model or needs.
bayesian_opt!
returns a couple composed of the point of the optimal solution found and the value of function cost in that point.
This macro needs bayesian
feature enabled.
More details are available in the following examples:
Random Search§
Random Search is a family of algorithms based on the concept of: starting from a point X of paramater space, sample a set of points based on current position and move to the point Y if f(Y) is the best of sampling and better than f(X), where f is the cost function. This operation is iterated until you got your goal or reached tha maximum number of iterations.