Auto-Calibrating SABR Parameters with Bayesian Optimization
Auto-Calibrating SABR Parameters with Bayesian Optimization
The SABR model (Stochastic Alpha Beta Rho) is widely used for modeling interest-rate options and FX volatility. It has four parameters (alpha, beta, rho, nu) that must be calibrated to market option prices daily. Traditional calibration (solving an optimization problem) is expensive computationally. Bayesian optimization with machine learning surrogates enables fast, accurate calibration.
SABR Model Essentials
SABR models forward rates as a process with stochastic volatility:
dF = α × σ × F^β dW
dσ = ν × σ × dZ
Parameters:
- α: initial volatility level
- β: backbone CEV exponent (controls how volatility varies with level)
- ρ: correlation between rate and volatility processes
- ν: volatility of volatility
SABR has a closed-form approximation for option prices, enabling fast valuation. However, extracting parameters from market prices requires solving an inverse problem.
Traditional Calibration Issues
Standard approach: minimize squared difference between market and model prices:
Minimize: Σ (Market_IV(K) - Model_IV(K; α,β,ρ,ν))²
This optimization is computationally expensive (requires hundreds of SABR valuations), and non-convex (multiple local optima). Daily recalibration across many tenor points is slow.
Bayesian Optimization Approach
Rather than optimizing directly, maintain a surrogate model (Gaussian process) that approximates the objective function (squared pricing error). The surrogate is fast to evaluate. Using expected improvement criterion, select promising parameter combinations to evaluate.
Workflow:
- Initialize with 5-10 random SABR calibrations
- Fit GP surrogate to these calibration points
- Use expected improvement to select next promising parameters to test
- Run SABR calibration at those parameters
- Update surrogate with new point
- Repeat until convergence (typically 20-30 iterations)
This requires far fewer objective function evaluations than standard optimization.
Neural Network Surrogates
Instead of Gaussian processes, use neural networks as surrogates. Given (market vol surface), predict (optimal SABR parameters). This reverses the direction: normally we solve prices → parameters; here we directly predict parameters from prices.
Training data: collect (market vol surface, optimal SABR parameters) pairs from historical markets. Train network to map volumes to parameters. At inference time, given today's market, instantly predict parameters.
Advantages:
- Single forward pass provides parameters (no iterative optimization)
- Can capture regime-specific optimal parameters
- Easily parallelizable
Handling Multiple Tenors and Strikes
In practice, calibrate to a grid: multiple tenors (1 month, 3 months, ..., 5 years) and multiple strikes (ATM, 25 delta, 10 delta, etc.). The calibration surface is high-dimensional.
Joint calibration across the full grid is challenging. Approaches:
- Separate calibrations per tenor with consistency constraints
- Global optimization that fits all points simultaneously with regularization
- Neural network learning joint parameter structure
Parameter Stability and Constraints
SABR parameters have valid ranges. β ∈ [0, 1]; ρ ∈ [-1, 1]; α, ν > 0. Optimization must enforce these bounds.
Additionally, parameters should be stable: today's parameters should not be drastically different from yesterday's. Regularizing toward previous calibration helps.
Validation and Error Analysis
Calibration quality is evaluated by:
- RMS pricing error: how well do calibrated SABR prices match market prices?
- Greeks accuracy: do hedging Greeks (delta, gamma, vega) match market expectations?
- Extrapolation stability: do parameters change smoothly as new data arrives?
Extensions to Multi-Curve Framework
Modern interest-rate markets use multi-curve frameworks (different curves for different tenors and indices). SABR calibration extends to this setting: calibrate separate processes for each curve, with cross-curve correlations.
Practical Production Systems
In production:
- Calibrations run hourly or more frequently
- Results cached and used for pricing/hedging throughout the day
- Automated checks for parameter sanity and stability
- Manual overrides available if automated calibration produces suspect results
Conclusion
Bayesian optimization and neural network surrogates accelerate SABR calibration, enabling frequent recalibration with minimal computational cost. This is a practical example of machine learning improving real-world derivatives operations.