Learning-Augmented Adaptive MPC — Closing the Model-Reality Gap for Higher Performance
Public reference use case (by others): This page summarizes publicly available reference implementations and papers. Not client results of Dr. Rafal Noga.
Why This Matters (Executive Summary)
- The problem: First-principles models rarely capture every real-world effect — tire slip at the limit, unknown payloads, changing environments, or structural wear all create model mismatch that degrades MPC performance and can compromise safety.
- The solution class: Learning-augmented MPC adds a data-driven component (Gaussian process, neural network, or adaptive law) to a nominal MPC, correcting residual model error online or from prior data.
- Measurable outcomes: Published experiments report up to 82% tracking error reduction, 10% lap-time gains at the racing limit, and 3x speed increases for obstacle avoidance — all while preserving real-time feasibility and constraint satisfaction.
- Why it matters for operations: You can start with a solid nominal MPC and selectively add learning where model mismatch is measurable and limits performance — no full model rewrite required.
The Design Pattern Explained
Learning-augmented MPC keeps a physics-based nominal model as the backbone and adds a learned correction term that captures what the nominal model misses. The correction can take several forms:
- Gaussian Process MPC (GP-MPC): A GP learns the residual dynamics from operational data and provides calibrated uncertainty estimates. The MPC can then tighten constraints in high-uncertainty regions, maintaining safety while exploiting improved predictions where confidence is high. 1
- Neural MPC: A neural network (potentially with thousands of parameters) replaces or augments the dynamics model. The key engineering challenge is ensuring that the network evaluation fits within a real-time optimization loop — recent advances show this is feasible at 50 Hz on embedded hardware. 2
- Adaptive MPC (MRAC + MPC): Model Reference Adaptive Control updates parameters online without requiring offline retraining, making the controller robust to changing dynamics such as different loads or contact conditions. 3
- Sampling-based / GPU-accelerated MPC: When the learned model or environment is too complex for gradient-based solvers, massively parallel sampling on GPUs enables real-time randomized MPC with thousands of rollouts per control step. 4
The architecture follows a consistent pattern: estimation (state + model uncertainty) feeds a planning / optimization layer (MPC with learned dynamics), which outputs constrained control actions, with a safety fallback to the nominal controller if learning confidence drops.
Applications & Reference Implementations
Application 1: GP-MPC for Autonomous Racing — ETH Zurich
A full-size autonomous race car (AMZ, ETH Zurich) used Gaussian process regression to learn residual tire dynamics that a nominal single-track model could not capture at the performance limit. The GP-augmented contouring MPC operated at lateral accelerations up to 2 g and speeds of 15 m/s, with a dictionary-based data management approach for continual online updates. The experiments reported approximately 10% lap-time reduction compared to the nominal MPC baseline. This demonstrates that even a well-tuned physics model benefits from data-driven correction when operating near dynamic limits. 1
Application 2: Real-Time Neural MPC on a Quadrotor — University of Zurich
Researchers integrated a large neural-network dynamics model (with over 4,000x the parametric capacity of prior neural-MPC implementations) into an MPC pipeline running at 50 Hz on an embedded platform. On an agile quadrotor performing aggressive maneuvers, the neural MPC achieved up to 82% lower positional tracking error compared to MPC without neural dynamics. The work proves that modern neural architectures can be deployed inside real-time optimization loops without violating timing constraints, opening the door for industrial embedded applications. 2
Application 3: Learning-Based NMPC for Vision-Based Mobile Robots — University of Toronto
A learning-based nonlinear MPC was evaluated on two mobile robot platforms (50 kg and 160 kg) for vision-based path tracking over distances of 1.8 km and 500 m at speeds up to 1.6 m/s. The learned model compensated for terrain-dependent model mismatch that degraded pure vision-based tracking. Long-range field trials validated that the approach maintained robust performance across changing outdoor conditions — a critical requirement for logistics, agriculture, and inspection robots. 5
Application 4: Adaptive Interaction Control (MRAC + MPC) for Door Opening — ETH Zurich
A mobile manipulator used MRAC combined with MPC to open doors with varying dynamics (light and heavy doors). The adaptive layer updated interaction parameters online, reducing angular RMSE from 6.7 degrees (baseline) to 1.4 degrees (MRAC+MPC) on a light door and from 3.2 degrees to 1.6 degrees on a heavy door, with force profiles maintained at 10-15 N and 20-25 N respectively. This pattern transfers directly to any industrial task where a robot must interact with objects whose dynamics change between cycles — palletizing, machine tending, or assembly with variable fixtures. 3
Application 5: GPU Randomized MPC for Dynamic Obstacle Avoidance
A randomized MPC implementation on a 1/10 scale RC car compared CPU versus GPU execution for dynamic obstacle avoidance. The CPU version achieved only 30 Hz and collided at 5.1 km/h, while the GPU version ran at 200 Hz with 1,000 parallel trajectory samples and achieved smooth avoidance up to 11.5 km/h — a 3x speed improvement with a 30-step prediction horizon covering approximately 3 meters of look-ahead. This approach is directly relevant to AGV and AMR applications in warehouses or factory floors where obstacles are unpredictable. 4
Application 6: Incremental MPC with Time-Delay Estimation — TUM/DLR
An incremental (model-free) MPC used time-delay estimation to construct an implicit dynamics model without requiring explicit plant identification. Validated on a real 3-DoF manipulator with Maxon motor instrumentation, this approach avoids the cost and fragility of system identification while maintaining MPC-quality constraint handling. It is particularly attractive for legacy equipment where detailed models are unavailable. 6
What This Means for Your Operations
Learning-augmented MPC is most valuable when your system’s physics are well-understood in principle but site-specific conditions vary — product mix changes, equipment wear, environmental drift, or payload variability. The practical starting point is always a functioning nominal MPC; learning is added surgically where model mismatch is measurable and limits performance.
Common readiness indicators:
- You already run MPC or advanced control but see performance degrade with changing conditions.
- You have operational data (logs, sensors) that captures the conditions where performance drops.
- Your control hardware has compute headroom (or can be upgraded) for the learning component.
How We Deliver This (Engagement Model)
- Phase 0: NDA + data request — Collect operational logs, model documentation, and performance baselines. Identify where model mismatch is the bottleneck.
- Phase 1: Fixed-scope discovery (concept + feasibility) — Quantify model mismatch from data. Select the learning approach (GP, neural, adaptive) based on data availability, real-time budget, and safety requirements. Deliver a concept document with architecture, risk assessment, and validation plan.
- Phase 2: Implementation + validation + commissioning — Build the learning-augmented controller. Validate on representative scenarios. Commission with safe fallback to the nominal controller during ramp-up.
- Phase 3: Monitoring + training + scaling — Deploy monitoring for model confidence and constraint activity. Train operators on when the learning component is active versus fallback. Scale to additional operating points or sister plants.
Typical KPIs to Track
- Tracking error reduction (position, contour, force) versus nominal MPC baseline
- Model prediction error (residual) before and after learning
- Constraint violation rate and safety margin utilization
- Real-time solve time and worst-case computation budget
- Operator intervention frequency and manual override rate
Risks & Prerequisites
- Learning needs data: The learned component is only as good as the data it was trained on. Ensure sufficient coverage of the operating envelope.
- Safety under uncertainty: Learned models introduce epistemic uncertainty. Responsible approaches either constrain predictions to high-confidence regions or fall back to a nominal safe controller.
- Compute budget: Neural MPC and GPU-based approaches require hardware with sufficient compute. Verify real-time feasibility early.
- Maintenance: Learned models may need periodic retraining as the plant changes. Plan for model lifecycle management.
FAQ
Can I add learning to my existing MPC without replacing it? Yes — the most common pattern is to keep the nominal MPC intact and add a residual correction term. If the learned component fails or confidence drops, the system falls back to the nominal controller.
How much data do I need? It depends on the approach. GP-MPC can work with dozens to hundreds of data points per operating region. Neural MPC typically needs more data but can generalize better across conditions. Adaptive approaches (MRAC) update online from a few cycles.
Is this safe for production? Safety depends on the fallback architecture. All responsible implementations in the references include constraint tightening under uncertainty and/or a safe nominal fallback mode. The learning component should be treated as advisory until validated.
What compute hardware is required? GP-MPC runs on standard industrial PCs. Neural MPC at high rates may need GPU or optimized inference. GPU-RMPC explicitly requires a GPU. The right choice depends on your control rate and model complexity.
Book a 30-Minute Discovery Call
Ready to explore whether this pattern fits your system?
Dr. Rafal Noga — Independent APC/MPC Consultant
Fixed-scope discovery · NDA-first · DACH on-site available
Public References
Footnotes
-
Kabzan et al., “Learning-Based Model Predictive Control for Autonomous Racing” (ETH Research Collection). https://www.research-collection.ethz.ch/bitstreams/7d0faa11-1667-481c-a497-ca7ef4611521/download ↩ ↩2
-
Salzmann et al., “Real-time Neural MPC: Deep Learning Model Predictive Control for Quadrotors and Agile Robotic Platforms” (RAL, 2023). https://rpg.ifi.uzh.ch/docs/RAL2023_Salzmann.pdf ↩ ↩2
-
Batzianoulis et al., “Adaptive Interaction Control for Robotic Door Opening” (arXiv, 2021). https://arxiv.org/pdf/2106.04202 ↩ ↩2
-
Muraleedharan et al., “Randomized Model Predictive Control for Dynamic Obstacle Avoidance and Autonomous Racing” (IEEE T-IV, 2022). https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9911860 ↩ ↩2
-
Ostafew et al., “Learning-Based Nonlinear Model Predictive Control to Improve Vision-Based Mobile Robot Path Tracking” (ICRA, 2014). https://asrl.utias.utoronto.ca/wp-content/papercite-data/pdf/ostafew_icra14.pdf ↩
-
“Incremental MPC Exploiting Time-Delay Estimation” (TUM/DLR). https://mediatum.ub.tum.de/doc/1732774/1732774.pdf ↩
Related Use Cases
Interested in applying these methods?
Let's discuss how APC/MPC can deliver similar results for your operations.
Get in Touch