Zero-Shot Transferable Solution Method for Parametric Optimal Control Problems

Xinjian Li, Kelvin Kan, Deepanshu Verma, Krishna Kumar, Stanley Osher, Ján Drgoňa · 2025

[arxiv]

Zero-Shot Transferable Solution Method for Parametric Optimal Control Problems

Authors: Xinjian Li, Kelvin Kan, Deepanshu Verma, Krishna Kumar, Stanley Osher, Ján Drgoňa Year: 2025 Tags: optimal-control, transfer-learning, function-encoders, imitation-learning, zero-shot-generalization, neural-basis-functions

TL;DR

Learns a fixed set of neural network basis functions offline via imitation learning, then adapts to new parametric optimal control tasks zero-shot by estimating task-specific linear coefficients — either via least-squares projection onto a few trajectory samples or via a pre-trained operator network. This offline–online decomposition avoids re-solving the optimization problem for each new task while maintaining near-optimal performance across diverse dynamics and cost structures.

First pass — the five C's

Category. Research prototype — algorithmic framework with numerical validation.

Context. Parametric optimal control / transfer learning subfield. Builds directly on Ingebrand et al. (2025) function encoder (FE) theory for Hilbert-space transfer; extends Onken et al. (2022) neural network approach for high-dimensional OC applied to quadcopter; uses Li et al. (2024) stochastic OC neural network as a benchmark problem setting; draws on Ingebrand et al. (2025) basis-to-basis operator learning for the operator network variant.

Correctness. Load-bearing assumptions: (1) the true policy family lies approximately in a finite-dimensional Hilbert subspace spanned by learnable neural basis functions; (2) training tasks sufficiently cover the task distribution so basis functions generalize; (3) fixed dynamics across tasks — method does not address dynamics variability; (4) availability of solver-generated trajectory data for offline training. Assumption (1) is supported by Theorem 1 (universal approximation in Hilbert spaces), but the required number of bases for a given accuracy is not bounded empirically or theoretically in this paper.

Contributions. - Offline–online decomposition for parametric OC: basis functions learned once; online adaptation is a lightweight least-squares solve or a single forward pass through an operator network. - Two zero-shot inference modes (LS projection from data; operator network from task specification) with a principled trade-off analysis. - Statistical convergence guarantee (Theorem 2) bounding LS coefficient error as O(M^{−1/2}) in the number of trajectory samples. - Empirical validation spanning 2D linear, 12D nonlinear (quadcopter), and 4D nonlinear (bicycle) problems with varying terminal and running costs including obstacle avoidance.

Clarity. Generally well-structured with clear algorithm boxes and an explicit offline–online framing; notation is consistent, though the operator network training (Algorithm 2) receives less exposition than the LS path, and reproducibility details (exact basis count for bicycle, hyperparameter sensitivity) are scattered or absent.

Second pass — content

Main thrust: Learn p neural basis functions offline via imitation on N training tasks; at test time, express the new task's policy as a linear combination of those bases with coefficients found by least squares over ≤M new trajectory samples (or by evaluating a pre-trained operator network), achieving <4% objective error with no retraining.

Supporting evidence: - 2D path planning (linear dynamics, terminal cost variability): LS achieves <4% error in objective functional across seen targets, interpolation, and extrapolation; operator method shows up to ~9% error on extrapolation. - 12D quadcopter (nonlinear dynamics, terminal cost variability, 27 unseen test tasks): LS incurs 0.4% objective error on new targets; operator method incurs ~7.4% error on new targets (true objective 274.3089 vs. predicted 294.5627). - 4D bicycle, single obstacle (nonlinear dynamics, running cost variability, 36 unseen configurations): predicted terminal state deviation 0.0012 vs. ground truth 0.00005; predicted obstacle cost 0.0463 vs. ground truth 0.0385. - 4D bicycle, double obstacle (544/32 train/test split, 576 total configurations): predicted terminal state deviation 0.0046 vs. ground truth 0.00004; obstacle cost prediction 0.5750 vs. ground truth 0.5286 (~8.8% overestimate). - All experiments use 100 basis functions and 4-layer MLP with hidden size 256; 2D trained for 20K steps, quadcopter for 100K steps.

Figures & tables: Figure 1 (pipeline schematic) conveys the offline–online split clearly. Figures 2–4 show trajectory visualizations; axes are labeled with spatial coordinates. Figures 5–6 deliberately show worst-performing cases, which is commendable, though no error bars or confidence intervals appear anywhere. Tables I–II report mean objective losses but without variance, standard deviation, or statistical tests. Table III reports control cost, obstacle cost, and terminal deviation aggregated across new scenarios without distributional spread. No wall-clock timing comparisons against baselines appear in any figure or table.

Follow-up references: - Ingebrand, Thorpe & Topcu (2025) arXiv:2501.18373 — foundational function encoder theory this method directly extends. - Onken et al. (2022) IEEE TCST — neural network OC for high-dimensional problems; provides the quadcopter benchmark used here. - Verma et al. (2025) Foundations of Data Science — neural network approaches for parameterized OC; closest direct competitor cited. - Ingebrand et al. (2025) CMAME — basis-to-basis operator learning, motivates the operator network variant.

Third pass — critique

Implicit assumptions: - Fixed dynamics: the method is framed as handling varying objectives only; varying dynamics would invalidate the shared basis functions, but this is never discussed. - The task distribution is known and covered during offline training; out-of-distribution generalization beyond the convex hull is noted as less reliable (consistent with Figure 2c extrapolation results) but not characterized quantitatively. - The policy function space is a separable Hilbert space that can be adequately spanned by p=100 bases; no ablation on p is presented to show sensitivity or saturation. - Solver-generated open-loop trajectories are assumed to be near-optimal ground truth; any suboptimality in the data propagates directly into the learned bases. - Sub-Gaussian noise and full-rank Gram matrix are required by Theorem 2 but never verified empirically.

Missing context or citations: - No direct quantitative comparison to Verma et al. (2025) (cited as [45]), the closest stated competitor for parameterized OC, despite being cited in the related work. - No comparison to meta-learning or fine-tuning approaches (e.g., MAML-style) that also target fast adaptation to new tasks. - No comparison to explicit MPC / multi-parametric programming [2],[6] on the low-dimensional cases where those methods are tractable. - Online computation time per query is never reported; the claim of "real-time deployment" suitability is not numerically supported. - No discussion of constraint-satisfying control (input/state constraints beyond box bounds on U), which is critical in real robotics.

Possible experimental / analytical issues: - No error bars or statistical significance on any reported number; each table reports a single mean across tasks without variance, making it impossible to assess reliability. - Terminal state deviation for bicycle model: predicted 0.0012/0.0046 vs. ground truth ~0.00005 represents 24×/92× overshoot relative to the solver, yet this is presented without concern — its practical consequence (does the bicycle actually reach the target?) is not discussed. - Training and test task sets are generated from the same parameterized distributions (grids, Gaussian parameters); true out-of-distribution generalization to qualitatively different cost structures is not tested. - The "worst-case" framing in Figures 5–6 lacks context: worst out of how many? The overall failure rate (e.g., fraction of cases exceeding some threshold) is not reported. - Quadcopter operator method error (~7.4%) is substantially larger than LS (0.4%) but no analysis explains why or when the operator method would be preferred for this system. - Code is promised "upon publication" but not yet available, limiting reproducibility at time of preprint.

Ideas for future work: - Ablate number of basis functions p and number of training tasks N to characterize the accuracy–compute trade-off and guide practitioners on minimum data requirements. - Report online inference wall-clock times and compare to CasADi/SciPy re-solve times to substantiate real-time deployment claims quantitatively. - Extend to varying dynamics (e.g., parametric mass, friction) by conditioning basis functions on dynamics parameters, testing whether the linear combination structure still suffices. - Incorporate safety/constraint satisfaction (e.g., control barrier functions or projected LS) and test whether the zero-shot coefficient estimates preserve constraint feasibility under distribution shift.

Methods

  • function encoder (FE)
  • imitation learning
  • least-squares projection
  • operator network
  • multi-layer perceptron (MLP)
  • direct transcription
  • Tikhonov regularization
  • offline-online decomposition

Datasets

  • 2D trajectory planning (synthetic)
  • 12D quadcopter path planning (synthetic)
  • bicycle model single-obstacle (synthetic)
  • bicycle model double-obstacle (synthetic)

Claims

  • A function encoder framework can learn reusable neural basis functions that span the control policy space, enabling zero-shot adaptation to new optimal control tasks without retraining.
  • The offline-online decomposition confines intensive computation to the offline phase, making online adaptation lightweight and suitable for real-time deployment.
  • The proposed method achieves near-optimal performance with less than 4% error in objective value across 2D, 12D, and nonlinear benchmark problems.
  • Task-specific coefficients can be inferred either via least-squares projection from limited trajectory data or via a data-free operator network mapping from problem specifications.
  • The approach generalizes to unseen task parameters including interpolation and extrapolation of target locations and obstacle configurations.