Accepted Paper
Our Frontier4LCD workshop has received 134 outstanding submissions, and we are thrilled to announce that 100 high-quality papers have been accepted for presentation. Below is the list of accepted papers.
| Number | Title |
|---|---|
| 1 | AbODE: Ab initio antibody design using conjoined ODEs |
| 2 | Distributional Distance Classifiers for Goal-Conditioned Reinforcement Learning |
| 3 | LEAD: Min-Max Optimization from a Physical Perspective |
| 4 | Balancing exploration and exploitation in Partially Observed Linear Contextual Bandits via Thompson Sampling |
| 5 | Stochastic Linear Bandits with Unknown Safety Constraints and Local Feedback |
| 6 | Visual Dexterity: In-hand Dexterous Manipulation from Depth |
| 7 | Regret Bounds for Risk-sensitive Reinforcement Learning with Lipschitz Dynamic Risk Measures |
| 8 | On learning history-based policies for controlling Markov decision processes |
| 9 | Improved sampling via learned diffusions |
| 10 | Fast Approximation of the Generalized Sliced-Wasserstein Distance |
| 11 | Synthetic Experience Replay |
| 12 | A neural RDE approach for continuous-time non-Markovian stochastic control problems |
| 13 | Importance Weighted Actor-Critic for Optimal Conservative Offline Reinforcement Learning |
| 14 | Toward Understanding Latent Model Learning in MuZero: A Case Study in Linear Quadratic Gaussian Control |
| 15 | Gradient-free training of neural ODEs for system identification and control using ensemble Kalman inversion |
| 16 | Preventing Reward Hacking with Occupancy Measure Regularization |
| 17 | Exponential weight averaging as damped harmonic motion |
| 18 | Bridging RL Theory and Practice with the Effective Horizon |
| 19 | Accelerated Policy Gradient: On the Nesterov Momentum for Reinforcement Learning |
| 20 | Neural Optimal Transport with Lagrangian Costs |
| 21 | Taylor TD-learning |
| 22 | Coupled Gradient Flows for Strategic Non-Local Distribution Shift |
| 23 | Kernel Mirror Prox and RKHS Gradient Flow for Mixed Functional Nash Equilibrium |
| 24 | What is the Solution for State-Adversarial Multi-Agent Reinforcement Learning? |
| 25 | On the Generalization Capacities of Neural Controlled Differential Equations |
| 26 | A Best Arm Identification Approach for Stochastic Rising Bandits |
| 27 | Maximum State Entropy Exploration using Predecessor and Successor Representations |
| 28 | Guide Your Agent with Adaptive Multimodal Rewards |
| 29 | Breaking the Curse of Multiagents in a Large State Space: RL in Markov Games with Independent Linear Function Approximation |
| 30 | Unbalanced Optimal Transport meets Sliced-Wasserstein |
| 31 | Randomly Coupled Oscillators for Time Series Processing |
| 32 | Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware |
| 33 | Boosting Off-policy RL with Policy Representation and Policy-extended Value Function Approximator |
| 34 | Statistics estimation in neural network training: a recursive identification approach |
| 35 | Embedding Surfaces by Optimizing Neural Networks with Prescribed Riemannian Metric and Beyond |
| 36 | Delphic Offline Reinforcement Learning under Nonidentifiable Hidden Confounding |
| 37 | A Flexible Diffusion Model |
| 38 | Simulation-Free Schrödinger Bridges via Score and Flow Matching |
| 39 | On the Imitation of Non-Markovian Demonstrations: From Low-Level Stability to High-Level Planning |
| 40 | Fixed-Budget Hypothesis Best Arm Identification: On the Information Loss in Experimental Design |
| 41 | Variational Principle and Variational Integrators for Neural Symplectic Forms |
| 42 | A Policy-Decoupled Method for High-Quality Data Augmentation in Offline Reinforcement Learning |
| 43 | When is Agnostic Reinforcement Learning Statistically Tractable? |
| 44 | Algorithms for Optimal Adaptation ofDiffusion Models to Reward Functions |
| 45 | On Convergence of Approximate Schr\"{o}dinger Bridge with Bounded Cost |
| 46 | In-Context Decision-Making from Supervised Pretraining |
| 47 | Unbalanced Diffusion Schrödinger Bridge |
| 48 | Learning from Sparse Offline Datasets via Conservative Density Estimation |
| 49 | Parameterized projected Bellman operator |
| 50 | Randomized methods for computing optimal transport without regularization and their convergence analysis |
| 51 | Bridging Physics-Informed Neural Networks with Reinforcement Learning: Hamilton-Jacobi-Bellman Proximal Policy Optimization (HJBPPO) |
| 52 | On a Connection between Differential Games, Optimal Control, and Energy-based Models for Multi-Agent Interactions |
| 53 | Dynamic Feature-based Newsvendor |
| 54 | Game Theoretic Neural ODE Optimizer |
| 55 | Diffusion Model-Augmented Behavioral Cloning |
| 56 | Vector Quantile Regression on Manifolds |
| 57 | PAC-Bayesian Bounds for Learning LTI-ss systems with Input from Empirical Loss |
| 58 | Learning to Optimize with Recurrent Hierarchical Transformers |
| 59 | Sample Complexity of Hierarchical Decompositions in Markov Decision Processes |
| 60 | Fairness In a Non-Stationary Environment From an Optimal Control Perspective |
| 61 | Modular Hierarchical Reinforcement Learning for Robotics: Improving Scalability and Generalizability |
| 62 | Taylorformer: Probabalistic Modelling for Random Processes including Time Series |
| 63 | Stability of Multi-Agent Learning: Convergence in Network Games with Many Players |
| 64 | Improving and Generalizing Flow-Based Generative Models with Minibatch Optimal Transport |
| 65 | Deep Equilibrium Based Neural Operators for Steady-State PDEs |
| 66 | Informed POMDP: Leveraging Additional Information in Model-Based RL |
| 67 | IQL-TD-MPC: Implicit Q-Learning for Hierarchical Model Predictive Control |
| 68 | Modeling Accurate Long Rollouts with Temporal Neural PDE Solvers |
| 69 | Sub-linear Regret in Adaptive Model Predictive Control |
| 70 | Analyzing the Sample Complexity of Model-Free Opponent Shaping |
| 71 | Continuous Vector Quantile Regression |
| 72 | Trajectory Generation, Control, and Safety with Denoising Diffusion Probabilistic Models |
| 73 | Structured State Space Models for In-Context Reinforcement Learning |
| 74 | Latent Space Editing in Transformer-Based Flow Matching |
| 75 | Transport, VI, and Diffusions |
| 76 | Fit Like You Sample: Sample-Efficient Generalized Score Matching from Fast Mixing Markov Chains |
| 77 | Policy Gradient Algorithms Implicitly Optimize by Continuation |
| 78 | Improving Offline-to-Online Reinforcement Learning with Q-Ensembles |
| 79 | Offline Goal-Conditioned RL with Latent States as Actions |
| 80 | On the effectiveness of neural priors in modeling dynamical systems |
| 81 | Action and Trajectory Planning for Urban Autonomous Driving with Hierarchical Reinforcement Learning |
| 82 | Physics-informed Localized Learning for Advection-Diffusion-Reaction Systems |
| 83 | Factor Learning Portfolio Optimization Informed by Continuous-Time Finance Models |
| 84 | Equivalence Class Learning for GENERIC Systems |
| 85 | Model-based Policy Optimization under Approximate Bayesian Inference |
| 86 | Nonlinear Wasserstein Distributionally Robust Optimal Control |
| 87 | Online Control with Adversarial Disturbance for Continuous-time Linear Systems |
| 88 | Leveraging Factored Action Spaces for Off-Policy Evaluation |
| 89 | Efficient RL with Impaired Observability: Learning to Act with Delayed and Missing State Observations |
| 90 | Limited Information Opponent Modeling |
| 91 | Tendiffpure: Tensorizing Diffusion Models for Purification |
| 92 | Look Beneath the Surface: Exploiting Fundamental Symmetry for Sample-Efficient Offline RL |
| 93 | Optimization or Architecture: What Matters in Non-Linear Filtering? |
| 94 | Variational quantum dynamics of two-dimensional rotor models |
| 95 | Parallel Sampling of Diffusion Models |
| 96 | Actor-Critic Methods using Physics-Informed Neural Networks: Control of a 1D PDE Model for Fluid-Cooled Battery Packs |
| 97 | Undo Maps: A Tool for Adapting Policies to Perceptual Distortions |
| 98 | Learning with Learning Awareness using Meta-Values |
| 99 | Aligned Diffusion Schrödinger Bridges |
| 100 | On First-Order Meta-Reinforcement Learning with Moreau Envelopes |