Accepted Paper
Our Frontier4LCD workshop has received 134 outstanding submissions, and we are thrilled to announce that 100 high-quality papers have been accepted for presentation. Below is the list of accepted papers.
| Number | Title | 
|---|---|
| 1 | AbODE: Ab initio antibody design using conjoined ODEs | 
| 2 | Distributional Distance Classifiers for Goal-Conditioned Reinforcement Learning | 
| 3 | LEAD: Min-Max Optimization from a Physical Perspective | 
| 4 | Balancing exploration and exploitation in Partially Observed Linear Contextual Bandits via Thompson Sampling | 
| 5 | Stochastic Linear Bandits with Unknown Safety Constraints and Local Feedback | 
| 6 | Visual Dexterity: In-hand Dexterous Manipulation from Depth | 
| 7 | Regret Bounds for Risk-sensitive Reinforcement Learning with Lipschitz Dynamic Risk Measures | 
| 8 | On learning history-based policies for controlling Markov decision processes | 
| 9 | Improved sampling via learned diffusions | 
| 10 | Fast Approximation of the Generalized Sliced-Wasserstein Distance | 
| 11 | Synthetic Experience Replay | 
| 12 | A neural RDE approach for continuous-time non-Markovian stochastic control problems | 
| 13 | Importance Weighted Actor-Critic for Optimal Conservative Offline Reinforcement Learning | 
| 14 | Toward Understanding Latent Model Learning in MuZero: A Case Study in Linear Quadratic Gaussian Control | 
| 15 | Gradient-free training of neural ODEs for system identification and control using ensemble Kalman inversion | 
| 16 | Preventing Reward Hacking with Occupancy Measure Regularization | 
| 17 | Exponential weight averaging as damped harmonic motion | 
| 18 | Bridging RL Theory and Practice with the Effective Horizon | 
| 19 | Accelerated Policy Gradient: On the Nesterov Momentum for Reinforcement Learning | 
| 20 | Neural Optimal Transport with Lagrangian Costs | 
| 21 | Taylor TD-learning | 
| 22 | Coupled Gradient Flows for Strategic Non-Local Distribution Shift | 
| 23 | Kernel Mirror Prox and RKHS Gradient Flow for Mixed Functional Nash Equilibrium | 
| 24 | What is the Solution for State-Adversarial Multi-Agent Reinforcement Learning? | 
| 25 | On the Generalization Capacities of Neural Controlled Differential Equations | 
| 26 | A Best Arm Identification Approach for Stochastic Rising Bandits | 
| 27 | Maximum State Entropy Exploration using Predecessor and Successor Representations | 
| 28 | Guide Your Agent with Adaptive Multimodal Rewards | 
| 29 | Breaking the Curse of Multiagents in a Large State Space: RL in Markov Games with Independent Linear Function Approximation | 
| 30 | Unbalanced Optimal Transport meets Sliced-Wasserstein | 
| 31 | Randomly Coupled Oscillators for Time Series Processing | 
| 32 | Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware | 
| 33 | Boosting Off-policy RL with Policy Representation and Policy-extended Value Function Approximator | 
| 34 | Statistics estimation in neural network training: a recursive identification approach | 
| 35 | Embedding Surfaces by Optimizing Neural Networks with Prescribed Riemannian Metric and Beyond | 
| 36 | Delphic Offline Reinforcement Learning under Nonidentifiable Hidden Confounding | 
| 37 | A Flexible Diffusion Model | 
| 38 | Simulation-Free Schrödinger Bridges via Score and Flow Matching | 
| 39 | On the Imitation of Non-Markovian Demonstrations: From Low-Level Stability to High-Level Planning | 
| 40 | Fixed-Budget Hypothesis Best Arm Identification: On the Information Loss in Experimental Design | 
| 41 | Variational Principle and Variational Integrators for Neural Symplectic Forms | 
| 42 | A Policy-Decoupled Method for High-Quality Data Augmentation in Offline Reinforcement Learning | 
| 43 | When is Agnostic Reinforcement Learning Statistically Tractable? | 
| 44 | Algorithms for Optimal Adaptation ofDiffusion Models to Reward Functions | 
| 45 | On Convergence of Approximate Schr\"{o}dinger Bridge with Bounded Cost | 
| 46 | In-Context Decision-Making from Supervised Pretraining | 
| 47 | Unbalanced Diffusion Schrödinger Bridge | 
| 48 | Learning from Sparse Offline Datasets via Conservative Density Estimation | 
| 49 | Parameterized projected Bellman operator | 
| 50 | Randomized methods for computing optimal transport without regularization and their convergence analysis | 
| 51 | Bridging Physics-Informed Neural Networks with Reinforcement Learning: Hamilton-Jacobi-Bellman Proximal Policy Optimization (HJBPPO) | 
| 52 | On a Connection between Differential Games, Optimal Control, and Energy-based Models for Multi-Agent Interactions | 
| 53 | Dynamic Feature-based Newsvendor | 
| 54 | Game Theoretic Neural ODE Optimizer | 
| 55 | Diffusion Model-Augmented Behavioral Cloning | 
| 56 | Vector Quantile Regression on Manifolds | 
| 57 | PAC-Bayesian Bounds for Learning LTI-ss systems with Input from Empirical Loss | 
| 58 | Learning to Optimize with Recurrent Hierarchical Transformers | 
| 59 | Sample Complexity of Hierarchical Decompositions in Markov Decision Processes | 
| 60 | Fairness In a Non-Stationary Environment From an Optimal Control Perspective | 
| 61 | Modular Hierarchical Reinforcement Learning for Robotics: Improving Scalability and Generalizability | 
| 62 | Taylorformer: Probabalistic Modelling for Random Processes including Time Series | 
| 63 | Stability of Multi-Agent Learning: Convergence in Network Games with Many Players | 
| 64 | Improving and Generalizing Flow-Based Generative Models with Minibatch Optimal Transport | 
| 65 | Deep Equilibrium Based Neural Operators for Steady-State PDEs | 
| 66 | Informed POMDP: Leveraging Additional Information in Model-Based RL | 
| 67 | IQL-TD-MPC: Implicit Q-Learning for Hierarchical Model Predictive Control | 
| 68 | Modeling Accurate Long Rollouts with Temporal Neural PDE Solvers | 
| 69 | Sub-linear Regret in Adaptive Model Predictive Control | 
| 70 | Analyzing the Sample Complexity of Model-Free Opponent Shaping | 
| 71 | Continuous Vector Quantile Regression | 
| 72 | Trajectory Generation, Control, and Safety with Denoising Diffusion Probabilistic Models | 
| 73 | Structured State Space Models for In-Context Reinforcement Learning | 
| 74 | Latent Space Editing in Transformer-Based Flow Matching | 
| 75 | Transport, VI, and Diffusions | 
| 76 | Fit Like You Sample: Sample-Efficient Generalized Score Matching from Fast Mixing Markov Chains | 
| 77 | Policy Gradient Algorithms Implicitly Optimize by Continuation | 
| 78 | Improving Offline-to-Online Reinforcement Learning with Q-Ensembles | 
| 79 | Offline Goal-Conditioned RL with Latent States as Actions | 
| 80 | On the effectiveness of neural priors in modeling dynamical systems | 
| 81 | Action and Trajectory Planning for Urban Autonomous Driving with Hierarchical Reinforcement Learning | 
| 82 | Physics-informed Localized Learning for Advection-Diffusion-Reaction Systems | 
| 83 | Factor Learning Portfolio Optimization Informed by Continuous-Time Finance Models | 
| 84 | Equivalence Class Learning for GENERIC Systems | 
| 85 | Model-based Policy Optimization under Approximate Bayesian Inference | 
| 86 | Nonlinear Wasserstein Distributionally Robust Optimal Control | 
| 87 | Online Control with Adversarial Disturbance for Continuous-time Linear Systems | 
| 88 | Leveraging Factored Action Spaces for Off-Policy Evaluation | 
| 89 | Efficient RL with Impaired Observability: Learning to Act with Delayed and Missing State Observations | 
| 90 | Limited Information Opponent Modeling | 
| 91 | Tendiffpure: Tensorizing Diffusion Models for Purification | 
| 92 | Look Beneath the Surface: Exploiting Fundamental Symmetry for Sample-Efficient Offline RL | 
| 93 | Optimization or Architecture: What Matters in Non-Linear Filtering? | 
| 94 | Variational quantum dynamics of two-dimensional rotor models | 
| 95 | Parallel Sampling of Diffusion Models | 
| 96 | Actor-Critic Methods using Physics-Informed Neural Networks: Control of a 1D PDE Model for Fluid-Cooled Battery Packs | 
| 97 | Undo Maps: A Tool for Adapting Policies to Perceptual Distortions | 
| 98 | Learning with Learning Awareness using Meta-Values | 
| 99 | Aligned Diffusion Schrödinger Bridges | 
| 100 | On First-Order Meta-Reinforcement Learning with Moreau Envelopes |