Methods
This chapter introduces the utility modules provided by D²MTOLab, including batch experiments, data analysis, performance metrics, and algorithm components. These modules provide standardized testing workflows and rich algorithm building tools.
Batch Experiments
from Methods.batch_experiment import BatchExperiment
The batch experiment module provides a complete framework for running multiple optimization algorithms on multiple benchmark problems, supporting parallel processing, automatic logging, and configuration management.
Module Features
The BatchExperiment class offers:
Flexible Configuration: Support for adding multiple test problems and algorithms with their parameter configurations
Parallel Computing: Utilize multi-core CPU for parallel execution to significantly improve efficiency
Complete Experiment Recording: Automatically record execution time, status, and error information
Configuration Persistence: Save experiment configurations as YAML files for reproducibility
Time Statistics: Generate CSV files with detailed timing information
Optional Folder Cleanup: Support for cleaning old data before experiments
Progress Visualization: Real-time display of experiment progress and completion status
Class Initialization
Initialize the BatchExperiment class:
batch_exp = BatchExperiment(
base_path='./Data', # Data storage path
clear_folder=False # Whether to clear folder
)
Parameters:
base_path: Storage path for experiment data, default:./Dataclear_folder: IfTrue, clear all contents in the target folder before initialization
Adding Problems
Use the add_problem method to add optimization problems:
from Problems.MTSO.cec17_mtso import CEC17MTSO
cec17mtso = CEC17MTSO()
# Add problems to batch experiment
batch_exp.add_problem(problem_creator=cec17mtso.P1, problem_name='P1')
batch_exp.add_problem(problem_creator=cec17mtso.P2, problem_name='P2')
Parameters:
problem_creator: Problem creation function that generates problem instancesproblem_name: Problem name for result file naming**problem_params: Optional parameters passed to the problem creator (e.g., maximum number of fitness evaluations)
Adding Algorithms
Use the add_algorithm method to add optimization algorithms:
from Algorithms.STSO.GA import GA
from Algorithms.STSO.DE import DE
from Algorithms.STSO.PSO import PSO
# Add algorithms with parameters
batch_exp.add_algorithm(algorithm_class=GA, algorithm_name='GA',
n=100, max_nfes=10000)
batch_exp.add_algorithm(algorithm_class=DE, algorithm_name='DE',
n=100, max_nfes=10000)
batch_exp.add_algorithm(algorithm_class=PSO, algorithm_name='PSO',
n=100, max_nfes=10000)
Parameters:
algorithm_class: Algorithm class (e.g.,GA,DE,PSO)algorithm_name: Algorithm name for subfolder and file naming**algorithm_params: Algorithm parameters (problem,save_path, andnameare set automatically)
Running Experiments
Execute the batch experiment using the run method:
batch_exp.run(n_runs=30, verbose=True, max_workers=8)
Parameters:
n_runs: Number of independent runs for each algorithm on each problemverbose: Whether to print detailed progress information, default:Truemax_workers: Maximum number of parallel worker processes, default: CPU core count
Example Output:
Clearing existing data folder: ./Data
Configuration saved to: ./Data/experiment_config.yaml
============================================================
Starting Batch Experiment (Parallel Mode)!
============================================================
Number of problems: 2
Number of algorithms: 3
Number of independent runs: 30
Total experiments: 180
Max workers: 8
Progress: 18/180 (10.0%)
Progress: 36/180 (20.0%)
...
Total time: 1200.00 seconds (20.00 minutes)
Parallel speedup: 10.76x
Timing summary saved to: ./Data/time_summary_20251203_143022.csv
============================================================
All Experiments Completed!
============================================================
Configuration Management
Experiment configurations are automatically saved as YAML files (experiment_config.yaml) when running, including:
Creation time and base path
Detailed problem configurations
Algorithm parameters
Run settings (number of runs, workers, etc.)
Loading from Configuration:
# Load experiment from saved configuration
batch_exp = BatchExperiment.from_config('./Data/experiment_config.yaml')
batch_exp.run() # Use settings from config file
# Override settings
batch_exp = BatchExperiment.from_config('./Data/experiment_config.yaml')
batch_exp.run(n_runs=50, max_workers=16)
Output Structure
Batch experiments generate three types of files:
Configuration File:
experiment_config.yamlAlgorithm Results: Organized in subfolders
Data/ ├── GA/ │ ├── GA_P1_1.pkl │ ├── GA_P1_2.pkl │ └── ... ├── DE/ │ └── ... └── PSO/ └── ...Timing Statistics:
time_summary_[timestamp].csvAlgorithm
Problem
Run
Filename
Time(s)
Status
Error
GA
P1
1
GA_P1_1
1.2345
Success
GA
P1
2
GA_P1_2
1.2198
Success
PSO
P2
5
PSO_P2_5
0.0000
Failed
Division by zero
Data Analysis
from Methods.data_analysis import DataAnalyzer
The data analysis module provides comprehensive analysis and visualization for optimization results, including metric calculation, statistical comparison tables, convergence curves, runtime analysis, Pareto front visualization, etc.
Module Features
The DataAnalyzer class offers:
Automatic Data Scanning: Automatically identify algorithms, problems, and run counts
Multiple Performance Metrics: Support for objective values (SO), IGD, and HV (MO)
Statistical Analysis: Mean, median, max, min statistics with Wilcoxon rank-sum test
Table Generation: Excel or LaTeX format tables with significance annotations
Convergence Curves: Plot algorithm convergence on each task with log-scale support
Runtime Analysis: Generate runtime comparison bar charts
Pareto Front Visualization: Support 2D, 3D, and high-dimensional non-dominated solutions
Flexible Configuration: Customizable color schemes, marker styles, and statistics
Complete Pipeline: One-step analysis or step-by-step execution
Class Initialization
Initialize the DataAnalyzer with configuration options:
analyzer = DataAnalyzer(
data_path='./Data', # Data directory path
settings=None, # Problem settings (for complex metrics)
algorithm_order=None, # Algorithm display order
save_path='./Results', # Results save path
table_format='excel', # Table format: 'excel' or 'latex'
figure_format='pdf', # Figure format: 'pdf', 'png', 'svg'
statistic_type='mean', # Statistic: 'mean', 'median', 'max', 'min'
significance_level=0.05, # Significance level for tests
rank_sum_test=True, # Whether to perform rank-sum test
log_scale=False, # Whether to use log scale
show_pf=True, # Whether to show true Pareto front
show_nd=True, # Whether to show only non-dominated
best_so_far=True, # Whether to use best-so-far values
clear_results=True # Whether to clear results folder
)
Metric Configuration
For problems requiring complex metrics (e.g., multiobjective optimization), provide a settings configuration dictionary:
SETTINGS = {
'metric': 'IGD', # Performance metric: 'IGD' or 'HV'
'ref_path': './MOReference', # Reference file path
'n_ref': 10000, # Number of reference points
# Problem P1 reference definitions
'P1': {
'T1': 'P1_T1_ref.npy', # Method 1: File path
'T2': 'P1_T2_ref.csv', # Supports .npy and .csv
},
# Problem P2 reference definitions
'P2': {
'T1': lambda n, m: generate_pf(n, m), # Method 2: Callable function
'T2': [[1.0, 0.0], [0.0, 1.0]], # Method 3: Direct array
},
}
# Use settings to create analyzer
analyzer = DataAnalyzer(data_path='./Data', settings=SETTINGS)
Reference definitions support three methods:
File Path: String filename or full path, supports
.npyand.csvCallable Function: Accepts
(n_points, n_objectives)parameters, returns reference arrayArray Data: Directly provide list, tuple, or NumPy array
Complete Analysis Pipeline
One-Step Analysis:
from Methods.data_analysis import DataAnalyzer
# Create analyzer instance (settings optional for SO)
analyzer = DataAnalyzer()
# Execute complete analysis pipeline
results = analyzer.run()
Step-by-Step Execution:
# Create analyzer
analyzer = DataAnalyzer(
data_path='./Data',
settings=SETTINGS,
algorithm_order=['NSGA-II', 'MOEA/D', 'MyAlgo'],
clear_results=False
)
# Step 1: Scan data directory
scan_result = analyzer.scan_data()
# Step 2: Calculate metrics
metric_results = analyzer.calculate_metrics()
# Step 3: Selective generation
analyzer.generate_tables() # Statistical tables
analyzer.generate_convergence_plots() # Convergence curves
analyzer.generate_runtime_plots() # Runtime plots
analyzer.generate_nd_solution_plots() # Pareto front plots
Accessing Raw Results
Access raw data through the returned MetricResults object:
# Run analysis
results = analyzer.run()
# Access metric values (per generation)
algo1_p1_run1_task0 = results.metric_values['GA']['P1'][1][0]
print(f"Convergence length: {len(algo1_p1_run1_task0)}")
# Access best values
best_vals = results.best_values['GA']['P1'][1]
print(f"Best values per task: {best_vals}")
# Access objective values (Pareto solutions)
pareto_solutions = results.objective_values['GA']['P1'][1][0]
print(f"Solution shape: {pareto_solutions.shape}")
# Access runtime
runtime_seconds = results.runtime['GA']['P1'][1]
print(f"Runtime: {runtime_seconds:.2f}s")
# Access max function evaluations
max_nfes_list = results.max_nfes['GA']['P1']
print(f"Max NFEs per task: {max_nfes_list}")
# Access metric name
print(f"Metric used: {results.metric_name}")
Output Structure
Complete analysis generates the following output files:
./Results/
├── results_table_mean.xlsx # Statistical table (Excel)
├── results_table_mean.tex # Statistical table (LaTeX)
├── P1.pdf # Convergence curve for P1
├── P2-Task1.pdf # Convergence curve for P2 Task1
├── P2-Task2.pdf # Convergence curve for P2 Task2
├── runtime_comparison.pdf # Runtime comparison
└── ND_Solutions/ # Non-dominated solutions
├── P1-GA.pdf
├── P1-DE.pdf
├── P2-Task1-GA.pdf
└── ...
Reference Data Loading
The reference data loading system provides a flexible interface for loading Pareto fronts, reference points, or other reference data required for performance metric calculation and visualization.
Supported Reference Types:
The system supports three types of reference definitions:
Callable Functions: Dynamically generate reference data based on problem parameters
File Paths: Load pre-computed reference data from .npy or .csv files
Array Data: Directly use numpy arrays, lists, or tuples as reference data
Core Interface:
from Methods.data_utils import DataUtils
reference = DataUtils.load_reference(
settings=SETTINGS,
problem='DTLZ1',
task_identifier='T1', # or task index: 0
M=3, # Number of objectives (required)
D=10, # Number of variables (optional)
C=0 # Number of constraints (optional)
)
Parameters:
settings: Dictionary containing problem configurationsproblem: Problem name (e.g., “DTLZ1”, “DTLZ2”)task_identifier: Task name (str “T1”) or index (int 0)M: Number of objectives (required)D: Number of decision variables (optional)C: Number of constraints (optional, default: 0)
Returns: NumPy array with shape (n_points, M), or None if unavailable
Example 1: Callable Reference Function
Most common for benchmark problems:
from Methods.Algo_Methods.uniform_point import uniform_point
# Define reference generation function
def DTLZ1_PF(N, M):
W, _ = uniform_point(N, M)
return W / 2
# Configure in settings
SETTINGS = {
'metric': 'IGD',
'n_ref': 2000,
'DTLZ1': {
'T1': DTLZ1_PF, # Function reference
'T2': DTLZ1_PF,
}
}
# Load reference (automatically calls DTLZ1_PF(2000, 3))
reference = DataUtils.load_reference(SETTINGS, 'DTLZ1', 'T1', M=3)
Function Signatures:
Reference functions can have different signatures based on requirements:
# Signature 1: Basic (N, M)
def basic_ref(N, M):
return generate_reference(N, M)
# Signature 2: With dimension (N, M, D)
def dimension_ref(N, M, D):
scale = np.sqrt(D)
return generate_reference(N, M) * scale
# Signature 3: Full parameters (N, M, D, C)
def full_ref(N, M, D, C):
ref = generate_reference(N, M)
if C > 0:
# Apply constraint-based filtering
pass
return ref
The system automatically detects the function signature and passes appropriate parameters.
Example 2: File-Based Reference
Load pre-computed reference from files:
SETTINGS = {
'ref_path': './MOReference',
'MyProblem': {
'T1': 'myproblem_t1_pf.npy', # Relative path
'T2': '/abs/path/to/reference.csv', # Absolute path
}
}
reference = DataUtils.load_reference(SETTINGS, 'MyProblem', 'T1', M=3)
Supported file formats: .npy (NumPy binary) and .csv (comma-separated)
Automatic File Search:
If the specified file is not found, the system searches for:
{ref_path}/{problem}_{task}_ref.npy{ref_path}/{problem}_{task}_ref.csv
Example 3: Direct Array Reference
Provide reference data directly:
# Predefined reference points
predefined_pf = np.array([[0.0, 1.0], [0.5, 0.5], [1.0, 0.0]])
SETTINGS = {
'SimpleProblem': {
'T1': predefined_pf, # NumPy array
'T2': [[0, 1], [1, 0]], # List
'T3': ([0, 1], [1, 0]) # Tuple
}
}
reference = DataUtils.load_reference(SETTINGS, 'SimpleProblem', 'T1', M=2)
Example 4: Shared Reference for All Tasks
Use the same reference for all tasks:
SETTINGS = {
'n_ref': 10000,
'DTLZ2': {
'all_tasks': DTLZ2_PF # Applied to all tasks
}
}
# All tasks automatically use the same reference
ref_t1 = DataUtils.load_reference(SETTINGS, 'DTLZ2', 'T1', M=3)
ref_t2 = DataUtils.load_reference(SETTINGS, 'DTLZ2', 'T2', M=3)
Integration with DataAnalyzer
The reference loading is automatically handled by DataAnalyzer when settings are provided:
# Define references in settings
SETTINGS = {
'metric': 'IGD',
'n_ref': 5000,
'DTLZ1': {'T1': DTLZ1_PF, 'T2': DTLZ1_PF},
'DTLZ2': {'all_tasks': DTLZ2_PF}
}
# DataAnalyzer automatically uses references for:
# - Metric calculation (IGD, HV, etc.)
# - Pareto front visualization
analyzer = DataAnalyzer(data_path='./Data', settings=SETTINGS)
results = analyzer.run()
Best Practices:
Organize reference files systematically:
MOReference/ ├── DTLZ1_T1_ref.npy ├── DTLZ1_T2_ref.npy ├── DTLZ2_T1_ref.csv └── CustomProblem/ ├── T1_ref.npy └── T2_ref.npySet appropriate n_ref for metrics and visualization:
When calculating multiobjective metrics (e.g., IGD), it is recommended to set
n_refto 1000 (preferably not exceeding 2000). Using too many reference points can result in very large PDF files when visualizing Pareto fronts with the true PF overlay.SETTINGS = { 'metric': 'IGD', 'n_ref': 1000, # Recommended: balance accuracy and file size 'DTLZ1': {'T1': DTLZ1_PF} }
Always provide M parameter (number of objectives)
Provide D and C if your reference function requires them
Use meaningful function signatures:
(N, M)for simple problems(N, M, D)when dimension matters(N, M, D, C)for constrained problems
Error Handling:
The system provides informative warnings:
# Problem not found
reference = DataUtils.load_reference(SETTINGS, 'NonexistentProblem', 'T1', M=3)
# Warning: Problem 'NonexistentProblem' not found in settings
# Returns: None
# File not found
# Warning: File not found: './MOReference/missing_file.npy'
# Returns: None
# Missing parameter D (when needed)
# Warning: D not provided for Problem_T1, using 0
Test Data Analysis
from Methods.test_data_analysis import TestDataAnalyzer
The TestDataAnalyzer is a lightweight version of DataAnalyzer for quickly analyzing single test runs. It directly reads files with _test.pkl suffix without statistical tests or multi-run aggregation, suitable for algorithm development and debugging.
Module Features
Simplified Data Structure: Read test files directly without algorithm-classified subfolders
Fast Analysis: Skip statistical tests and multi-run aggregation
Complete Visualization: Convergence curves, runtime comparison, and Pareto fronts
Table Generation: LaTeX format result tables and convergence summaries
Flexible Configuration: Same configuration options as
DataAnalyzer
Class Initialization
analyzer = TestDataAnalyzer(
data_path='./TestData', # Test data directory
settings=None, # Problem settings (for MO)
algorithm_order=None, # Algorithm display order
save_path='./TestResults', # Results save path
figure_format='pdf', # Figure format
log_scale=False, # Log scale
show_pf=True, # Show true Pareto front
show_nd=True, # Show only non-dominated
best_so_far=True, # Use best-so-far values
clear_results=True, # Clear results folder
file_suffix='_test.pkl' # Test file suffix
)
Basic Usage
from Methods.test_data_analysis import TestDataAnalyzer
# Create analyzer (settings optional for SO)
analyzer = TestDataAnalyzer(data_path='./TestData',
save_path='./TestResults')
# Execute complete analysis
results = analyzer.run()
Output Structure
./TestResults/
├── test_results_table.tex # Results comparison table
├── convergence_summary_table.tex # Convergence summary table
├── Task1_convergence.pdf # Task1 convergence
├── Task2_convergence.pdf # Task2 convergence (if any)
├── runtime_comparison.pdf # Runtime comparison
└── ND_Solutions/ # Non-dominated solutions
├── Task1-GA.pdf
├── Task1-DE.pdf
└── ...
Comparison with DataAnalyzer
Feature |
TestDataAnalyzer |
DataAnalyzer |
|---|---|---|
Data Source |
Single test files ( |
Multiple repeated experiments |
File Structure |
Direct test files in directory |
Subfolders per algorithm |
Statistical Analysis |
No statistical tests |
Wilcoxon rank-sum test |
Table Format |
LaTeX only |
Excel and LaTeX |
Use Case |
Development and quick validation |
Formal experiment analysis |
Problem Definition (MTOP)
from Methods.mtop import MTOP
The MTOP (Multitask Optimization Problem) class provides a unified interface for defining single-task and multitask optimization problems with support for objectives, constraints, and variable bounds.
Module Features
The MTOP class offers:
Flexible Task Definition: Add single or multiple tasks with different dimensions and objectives
Constraint Support: Define constraint functions for constrained optimization
Automatic Vectorization: Handle both vectorized and non-vectorized objective functions
Unified Evaluation Mode: Optionally pad outputs to consistent dimensions across tasks
Cross-Platform Compatibility: Pickle-compatible function wrappers for parallel execution
Selective Evaluation: Evaluate specific objectives or constraints as needed
Class Initialization
mtop = MTOP(
unified_eval_mode=False, # Pad outputs to max dimensions
fill_value=0.0 # Fill value for padding
)
Adding Tasks
Single Task with Default Bounds [0, 1]:
def sphere(x):
return np.sum(x**2, axis=1)
mtop = MTOP()
idx = mtop.add_task(sphere, dim=3)
Single Task with Custom Bounds:
# Array bounds
idx = mtop.add_task(sphere, dim=3, lower_bound=[-5, -5, -5], upper_bound=[5, 5, 5])
# Scalar bounds (broadcast to all dimensions)
idx = mtop.add_task(sphere, dim=5, lower_bound=-5, upper_bound=5)
Multiple Tasks at Once:
def f1(x): return np.sum(x**2, axis=1)
def f2(x): return np.sum((x-1)**2, axis=1)
indices = mtop.add_task(
objective_func=(f1, f2),
dim=(3, 4),
lower_bound=([-1]*3, [-2]*4),
upper_bound=([1]*3, [2]*4)
)
Task with Constraints:
def constraint(x):
return x[:, 0] - 0.5 # g(x) <= 0
idx = mtop.add_task(sphere, dim=3, constraint_func=constraint)
Multiobjective Task:
def multi_obj(x):
f1 = np.sum(x**2, axis=1)
f2 = np.sum((x-1)**2, axis=1)
return np.column_stack([f1, f2])
idx = mtop.add_task(multi_obj, dim=3)
Evaluating Tasks
# Evaluate a single task
X = np.random.rand(10, 3)
objs, cons = mtop.evaluate_task(0, X)
# Selective evaluation (evaluate only specific objectives)
objs, cons = mtop.evaluate_task(0, X, eval_objectives=[0, 2])
# Skip constraint evaluation
objs, cons = mtop.evaluate_task(0, X, eval_constraints=False)
# Evaluate multiple tasks
X_list = [np.random.rand(10, 3), np.random.rand(10, 4)]
objs_list, cons_list = mtop.evaluate_tasks([0, 1], X_list)
Querying Task Information
# Get number of tasks
n_tasks = mtop.n_tasks
# Get dimensions for all tasks
dims = mtop.dims
# Get number of objectives/constraints for a task
n_obj = mtop.get_n_objectives(0)
n_con = mtop.get_n_constraints(0)
# Get detailed task info
info = mtop.get_task_info(0)
# Returns: {'dimension', 'n_objectives', 'n_constraints', 'lower_bounds', 'upper_bounds', ...}
# Print MTOP summary
print(mtop)
Animation Generator
from Methods.animation_generator import AnimationGenerator, create_optimization_animation
The animation generator module provides comprehensive visualization tools for optimization processes, supporting both single-objective and multiobjective optimization with multiple comparison modes.
Module Features
The animation generator offers:
Multiple Visualization Types: Decision space evolution, convergence curves (SO), and Pareto front evolution (MO)
Flexible Comparison Modes: Support for individual animations or merged comparisons across algorithms
NFEs-based Tracking: Display convergence in terms of function evaluations for better comparability
Batch Processing: Automatically scan and generate animations for all result files
Customizable Display: Configure algorithm order, animation quality, frame rate, and format
Multi-format Output: Support for GIF and MP4 formats
Task-specific Configuration: Different NFEs settings for different optimization tasks
Visualization Components
For Single-Objective Optimization:
Decision Space (Left): Parallel coordinate plot showing decision variable evolution
Convergence Curve (Right): Best objective value vs. NFEs (Number of Function Evaluations)
For Multiobjective Optimization:
Decision Space (Left): Parallel coordinate plot showing decision variable evolution
Objective Space (Right): Pareto front evolution
2D: Scatter plot (f1 vs. f2)
3D: 3D scatter plot with rotation view
High-dimensional: Parallel coordinate plot with normalized objectives
Quick Start
Using AnimationGenerator Class:
from Methods.animation_generator import AnimationGenerator
# Create generator and run
generator = AnimationGenerator(data_path='./Data', save_path='./Results')
generator.run()
Single File Animation (Convenience Function):
from Methods.animation_generator import create_optimization_animation
# Generate animation for a single result file
create_optimization_animation(
pkl_path='./Data/GA/GA_P1_1.pkl',
max_nfes=10000,
format='gif'
)
Batch Generation:
# Automatically scan and generate animations for all .pkl files
create_optimization_animation(
data_path='./Data',
save_path='./Animations',
max_nfes=10000,
fps=10,
dpi=100
)
Comparison Modes
The animation generator supports four merge modes for algorithm comparison:
Mode 0: Individual Animations (No Merge)
Generate separate animation for each algorithm:
create_optimization_animation(
data_path='./Data',
save_path='./Animations',
merge=0, # Default: individual animations
max_nfes=10000
)
Output structure:
Animations/
├── GA_P1_1_animation.gif
├── DE_P1_1_animation.gif
└── PSO_P1_1_animation.gif
Mode 1: Full Merge
All algorithms in the same plots (side-by-side decision and objective spaces):
create_optimization_animation(
pkl_path=['./Data/GA/GA_P1_1.pkl',
'./Data/DE/DE_P1_1.pkl',
'./Data/PSO/PSO_P1_1.pkl'],
merge=1,
title='Algorithm Comparison',
algorithm_order=['GA', 'DE', 'PSO'],
max_nfes=10000
)
Layout: [Merged Decision Space | Merged Objective Space]
Mode 2: Decision Separated, Objective Merged
Separate decision space for each algorithm, merged objective space:
create_optimization_animation(
pkl_path=['./Data/GA/GA_P1_1.pkl',
'./Data/DE/DE_P1_1.pkl',
'./Data/PSO/PSO_P1_1.pkl'],
merge=2,
title='Comparison',
algorithm_order=['GA', 'DE', 'PSO'],
max_nfes=10000
)
Layout: [GA Decision | DE Decision | PSO Decision | Merged Objective]
Mode 3: All Separated
Both decision and objective spaces separated for each algorithm:
create_optimization_animation(
pkl_path=['./Data/GA/GA_P1_1.pkl',
'./Data/DE/DE_P1_1.pkl'],
merge=3,
algorithm_order=['GA', 'DE'],
max_nfes=10000
)
Layout: [GA Dec | DE Dec | GA Obj | DE Obj]
Class Initialization
Instantiate the AnimationGenerator class for batch processing:
from Methods.animation_generator import AnimationGenerator
generator = AnimationGenerator(
data_path='./Data', # Directory containing .pkl files
save_path='./Results', # Output directory
algorithm_order=['GA', 'DE'], # Optional: specify display order
title='My Comparison', # Optional: custom title
merge=0, # Merge mode (0-3)
max_nfes=10000, # Max function evaluations
fps=10, # Frames per second
dpi=100, # Resolution
interval=100, # Frame interval (ms)
format='gif', # Output format
log_scale=False, # Log scale for SO
file_suffix='.pkl' # File pattern
)
# Execute the pipeline
results = generator.run()
Parameters:
pkl_path: Path to .pkl file(s), string for single file or list for merge modeoutput_path: Output file path (optional, auto-generated if None)fps: Frames per second (default: 10)dpi: Resolution, affects file size and quality (default: 100)merge: Comparison mode (0-3, default: 0)title: Custom title for the animation (optional)algorithm_order: List of algorithm names specifying display order (merge mode only)max_nfes: Maximum NFEs, scalar or list for multitask problems (default: 100)
NFEs Configuration
The max_nfes parameter controls the x-axis scale for convergence curves in single-objective optimization.
Scalar NFEs (Same for All Tasks):
# All tasks use the same NFEs
create_optimization_animation(
pkl_path='results.pkl',
max_nfes=10000 # All tasks: 10000 NFEs
)
List NFEs (Different per Task):
# Multitask problem with different NFEs per task
create_optimization_animation(
pkl_path='multi_task_results.pkl',
max_nfes=[5000, 10000, 15000] # Task 1: 5000, Task 2: 10000, Task 3: 15000
)
Automatic Compatibility:
The system automatically handles single-task and multitask scenarios:
# Single-task optimization
create_optimization_animation('single_task.pkl', max_nfes=1000)
# Multitask optimization
create_optimization_animation('multi_task.pkl', max_nfes=[1000, 2000])
Algorithm Order
Control the display order of algorithms in merge modes:
# Specify custom order
create_optimization_animation(
pkl_path=['BO-LCB-BCKT.pkl', 'BO.pkl', 'MTBO.pkl', 'RAMTEA.pkl'],
merge=2,
algorithm_order=['BO', 'MTBO', 'RAMTEA', 'BO-LCB-BCKT'],
max_nfes=10000
)
Behavior:
Algorithms are reordered according to
algorithm_orderMissing algorithms in the list are excluded with a warning
Extra files not in
algorithm_orderare ignoredIf
algorithm_order=None, uses the original order frompkl_path
Output Formats
GIF Format (Default):
# Explicit GIF
create_optimization_animation('results.pkl', format='gif')
# Or specify output file
create_optimization_animation('results.pkl', output_path='animation.gif')
MP4 Format (Requires FFmpeg):
# Explicit MP4
create_optimization_animation('results.pkl', format='mp4')
# Or specify output file
create_optimization_animation('results.pkl', output_path='animation.mp4')
Note: MP4 requires FFmpeg installation:
pip install ffmpeg-python
If FFmpeg is unavailable, the system automatically falls back to GIF format.
Quality Settings
Adjust animation quality through fps and dpi parameters:
# High quality, larger file
create_optimization_animation(
'results.pkl',
fps=20, # Smoother animation
dpi=150, # Higher resolution
format='mp4'
)
# Fast generation, smaller file
create_optimization_animation(
'results.pkl',
fps=8, # Fewer frames
dpi=70, # Lower resolution
format='gif'
)
Recommended Settings:
Preview/Draft:
fps=8, dpi=70Standard:
fps=10, dpi=100(default)Publication:
fps=15, dpi=150
Batch Processing
Automatic Scanning:
# Scan ./TestData and save to ./TestResults
results = create_optimization_animation(
data_path='./TestData',
save_path='./TestResults',
max_nfes=10000,
fps=10,
dpi=100
)
# Check results
print(f"Success: {results['success']}")
print(f"Failed: {results['failed']}")
Batch with Merge Mode:
# Automatically scan and merge all files
create_optimization_animation(
data_path='./Data',
save_path='./Animations',
merge=1,
title='All Algorithms Comparison',
algorithm_order=['BO', 'MTBO', 'RAMTEA', 'BO-LCB-BCKT'],
max_nfes=[5000, 10000], # Two tasks
format='mp4'
)
Custom File Pattern:
# Only process specific files
create_optimization_animation(
data_path='./Data',
pattern='GA_*.pkl', # Only GA results
save_path='./Animations',
max_nfes=10000
)
Complete Example
Single-Objective Optimization:
from Methods.optimization_animator import create_optimization_animation
# Individual animations
create_optimization_animation(
data_path='./Data',
save_path='./Animations/Individual',
merge=0,
max_nfes=10000,
fps=10,
dpi=100,
format='gif'
)
# Merged comparison
create_optimization_animation(
pkl_path=['./Data/GA/GA_P1_1.pkl',
'./Data/DE/DE_P1_1.pkl',
'./Data/PSO/PSO_P1_1.pkl'],
output_path='./Animations/comparison.mp4',
merge=2,
title='SO Algorithm Comparison',
algorithm_order=['GA', 'DE', 'PSO'],
max_nfes=10000,
fps=15,
dpi=120,
format='mp4'
)
Multiobjective Multitask Optimization:
# Multitask with different NFEs
create_optimization_animation(
pkl_path=['./Data/NSGAII/NSGAII_DTLZ_1.pkl',
'./Data/MOEAD/MOEAD_DTLZ_1.pkl',
'./Data/MyAlgo/MyAlgo_DTLZ_1.pkl'],
output_path='./Animations/MO_comparison.gif',
merge=3,
title='Multiobjective Comparison',
algorithm_order=['NSGA-II', 'MOEA/D', 'MyAlgo'],
max_nfes=[5000, 8000, 10000], # Different NFEs for 3 tasks
fps=12,
dpi=100
)
Command Line Usage
The animation generator can be used from the command line:
# Single file
python -m Methods.optimization_animator results.pkl output.gif 10 100
# Auto-scan mode (no arguments)
python -m Methods.optimization_animator
Arguments:
pkl_file: Path to .pkl fileoutput_file: Output path (optional)fps: Frames per second (optional, default: 10)dpi: Resolution (optional, default: 100)
Output Structure
Individual Mode (merge=0):
Animations/
├── GA_P1_1_animation.gif
├── GA_P1_2_animation.gif
├── DE_P1_1_animation.gif
└── PSO_P1_1_animation.gif
Merge Mode (merge>0):
Animations/
└── test_animation.gif # Default name if title not specified
Custom Title:
Animations/
└── Algorithm_Comparison_animation.gif
Console Output Example
======================================================================
Optimization Animation Generator
======================================================================
Data path: ./TestData
Save path: ./TestResults
Found 4 result files
Animation params: FPS=10, DPI=100, Interval=100ms, Format=GIF
Max NFEs: [5000, 10000]
Mode: MERGE (Decision Separated)
Algorithm Order: ['BO', 'MTBO', 'RAMTEA', 'BO-LCB-BCKT']
======================================================================
Creating merged comparison animation...
Original algorithms: ['BO-LCB-BCKT', 'BO', 'MTBO', 'RAMTEA']
Ordered algorithms: ['BO', 'MTBO', 'RAMTEA', 'BO-LCB-BCKT']
Generating animation... (this may take a while)
Animation saved to: ./TestResults/test_animation.gif
✓ Success
======================================================================
Processing Complete!
Merged animation: Success
======================================================================
Animations saved to: ./TestResults
======================================================================
Best Practices
Use MP4 for publication quality:
MP4 files are typically smaller and higher quality than GIF for the same content.
Adjust frame rate based on convergence speed:
Slow convergence (>1000 generations):
fps=8-10Medium convergence (100-1000 generations):
fps=10-15Fast convergence (<100 generations):
fps=15-20
Balance DPI and file size:
For presentations:
dpi=100-120For papers:
dpi=120-150For web sharing:
dpi=70-100
Use merge mode 2 for comparing many algorithms:
Mode 2 allows clear visualization of individual decision spaces while comparing objectives together.
Specify max_nfes consistently:
Ensure
max_nfesmatches your actual experimental setup for accurate NFEs display.Use algorithm_order for clarity:
Order algorithms logically (e.g., baseline first, variants after) for easier comparison.
Integration with Batch Experiments
Combine with BatchExperiment for complete workflow:
from Methods.batch_experiment import BatchExperiment
from Methods.optimization_animator import create_optimization_animation
# Step 1: Run batch experiments
batch_exp = BatchExperiment(base_path='./Data')
batch_exp.add_problem(problem_creator=problem.P1, problem_name='P1')
batch_exp.add_algorithm(algorithm_class=GA, algorithm_name='GA', n=100, max_nfes=10000)
batch_exp.add_algorithm(algorithm_class=DE, algorithm_name='DE', n=100, max_nfes=10000)
batch_exp.run(n_runs=30)
# Step 2: Generate animations for first run of each algorithm
create_optimization_animation(
pkl_path=['./Data/GA/GA_P1_1.pkl',
'./Data/DE/DE_P1_1.pkl'],
merge=2,
title='GA vs DE on P1',
algorithm_order=['GA', 'DE'],
max_nfes=10000,
format='mp4'
)
Troubleshooting
Issue: “FFMpeg not installed, falling back to GIF”
Solution: Install FFmpeg:
pip install ffmpeg-python
Issue: Animation file is too large
Solutions:
Reduce
dpi(e.g., from 100 to 70)Reduce
fps(e.g., from 15 to 8)Use MP4 instead of GIF
Reduce number of frames by using fewer generations in data
Issue: “Incompatible data: file X has Y tasks, expected Z”
Solution: Ensure all .pkl files have the same number of tasks when using merge mode.
Issue: “Algorithm names not found in pkl_paths”
Solution: Check that algorithm names in algorithm_order match the file stems (filenames without .pkl).
Issue: Animation generation is slow
Solutions:
Reduce
dpifor faster processingUse fewer data points (subsample generations)
Process files in smaller batches
Use fewer algorithms in merge mode
Performance Metrics
from Methods.metrics import IGD, HV, GD, IGDp, FR, CV, DeltaP, Spacing, Spread
The performance metrics module provides comprehensive implementations of optimization algorithm evaluation metrics with a unified interface design.
Module Features
The metric module follows these design principles:
Unified Interface: All metric classes follow the same interface specification
Direction Indicator: Each metric has a
signattribute (-1for minimization,1for maximization)Callable Support: Metric instances support functional calling (
__call__method)
Available Metrics
Metric |
Sign |
Description |
|---|---|---|
|
-1 |
Inverted Generational Distance. Measures convergence and diversity by averaging distances from Pareto front points to nearest obtained solutions. |
|
-1 |
Generational Distance. Measures convergence by averaging distances from obtained solutions to the nearest Pareto front points. |
|
-1 |
Inverted Generational Distance Plus. A modified IGD that only penalizes dominated portions, making it Pareto-compliant. |
|
+1 |
Hypervolume. Measures the volume of objective space dominated by the obtained solutions. Supports 2D/3D exact calculation and Monte Carlo for higher dimensions. |
|
-1 |
Averaged Hausdorff Distance. Maximum of GD and IGD, measuring both convergence and diversity. |
|
-1 |
Spacing metric. Measures the standard deviation of nearest neighbor distances, indicating solution uniformity. |
|
-1 |
Spread metric. Measures distribution uniformity relative to extreme points of the Pareto front. |
|
+1 |
Feasible Rate. Proportion of feasible solutions in the population (for constrained optimization). |
|
-1 |
Constraint Violation. Sum of constraint violations for the best solution (for constrained optimization). |
Metric Interface
All metric classes follow this template:
class MetricTemplate:
"""Performance metric template"""
def __init__(self):
"""Initialize metric"""
self.name = "MetricName" # Metric name
self.sign = -1 or 1 # Direction: -1 minimize, 1 maximize
def calculate(self, *args, **kwargs) -> float:
"""Calculate metric value"""
# Implementation...
pass
def __call__(self, *args, **kwargs) -> float:
"""Support instance as function call"""
return self.calculate(*args, **kwargs)
Usage Examples
Multiobjective Metrics (IGD, GD, HV, etc.):
from Methods.metrics import IGD, HV, GD, IGDp, DeltaP, Spacing, Spread
import numpy as np
# Obtained solutions and true Pareto front
objs = np.random.rand(100, 2) # 100 solutions, 2 objectives
pf = np.random.rand(1000, 2) # True Pareto front
# IGD (requires Pareto front)
igd = IGD()
igd_value = igd(objs, pf)
# GD (requires Pareto front)
gd = GD()
gd_value = gd(objs, pf)
# IGD+ (Pareto-compliant version)
igdp = IGDp()
igdp_value = igdp(objs, pf)
# HV (requires Pareto front or reference point)
hv = HV()
hv_value = hv(objs, pf=pf) # With Pareto front for normalization
hv_value = hv(objs, reference=np.array([2.0, 2.0])) # With reference point
# Averaged Hausdorff Distance
deltap = DeltaP()
deltap_value = deltap(objs, pf)
# Spacing (only requires obtained solutions)
spacing = Spacing()
spacing_value = spacing(objs)
# Spread (requires Pareto front)
spread = Spread()
spread_value = spread(objs, pf)
Constrained Optimization Metrics (FR, CV):
from Methods.metrics import FR, CV
import numpy as np
# Constraint values (n_solutions x n_constraints)
# Constraint satisfied when cons <= 0
cons = np.array([
[-0.1, -0.2], # Feasible (all <= 0)
[0.5, -0.1], # Infeasible
[-0.3, 0.2], # Infeasible
[-0.1, -0.5], # Feasible
])
# Feasible Rate (proportion of feasible solutions)
fr = FR()
fr_value = fr(cons) # Returns 0.5 (2 out of 4 feasible)
# Constraint Violation (minimum CV in population)
cv = CV()
cv_value = cv(cons) # Returns 0.0 (best solution has CV=0)
Algorithm Components
Algorithm Utilities
from Methods.Algo_Methods.algo_utils import *
The algorithm utilities module provides a complete toolkit for building optimization algorithms, including population initialization, evaluation, selection, mutation, crossover, and auxiliary functions.
Function |
Description |
|---|---|
|
Initialize multitask decision variable matrices with Random or LHS sampling |
|
Batch evaluate multiple tasks with selective objective/constraint evaluation |
|
Evaluate a single specified task |
|
Simulated Binary Crossover (SBX) for two parent vectors |
|
Polynomial mutation on decision vectors |
|
Generate offspring using GA operators (SBX + mutation) |
|
Generate offspring using DE/rand/1/bin strategy |
|
Tournament selection with multi-criteria lexicographic ordering |
|
Single-objective elite selection considering constraint violation |
|
Fast non-dominated sorting algorithm |
|
Calculate crowding distance for diversity preservation |
|
Initialize population history storage structure |
|
Append current generation data to history |
|
Extract best solutions, build Results object, and save to file |
|
Trim history exceeding max function evaluations |
|
Transfer data between unified and real spaces |
|
Data normalization (min-max or z-score) |
|
Inverse normalization to restore original scale |
|
Vertically stack multiple population arrays |
|
Synchronously select rows from multiple arrays by index |
|
Convert single parameter to multitask parameter list |
|
Extract and print algorithm metadata |
Bayesian Optimization Utilities
from Methods.Algo_Methods.bo_utils import *
The BO utilities module provides core Bayesian optimization functionalities based on BoTorch and GPyTorch, including single-task and multitask Gaussian process modeling.
Function |
Description |
|---|---|
|
Build and train single-task Gaussian process model |
|
Predict using trained single-task GP model |
|
Get next sampling point via single-task BO (LogEI acquisition) |
|
Build multitask Gaussian process model |
|
Predict for specified task using multitask GP |
|
Extract task correlation matrix from multitask GP |
|
Get next sampling point via multitask BO |
Similarity Evaluation
from Methods.Algo_Methods.sim_evaluation import *
The similarity evaluation module computes inter-task similarity for knowledge transfer decisions.
Function |
Description |
|---|---|
|
Calculate similarity matrix between tasks using Pearson correlation |
Uniform Point Generation
from Methods.Algo_Methods.uniform_point import *
The uniform point generation module provides various methods for generating uniformly distributed points for multiobjective optimization and decision space sampling.
Function |
Description |
|---|---|
|
Unified interface for point generation (NBI/ILD/MUD/grid/Latin) |
|
Normal-Boundary Intersection for reference points on unit simplex |
|
Incremental Lattice Design for adaptive reference points |
|
Mixture Uniform Design using good lattice points |
|
Grid sampling in unit hypercube |
|
Latin Hypercube Sampling for decision space exploration |
|
Generate good lattice points for MUD method |
|
Calculate Centered Discrepancy (CD2) for uniformity evaluation |
See Also
Problems - Problem definition guide
Algorithms - Algorithm implementation guide
API Reference - Complete API documentation