Moritz Reuss

I am a fourth-year PhD student in the Intuitive Robots Lab (IRL) at the Karlsruhe Institute of Technology (KIT), Germany. My research focuses on developing machine learning methods to teach robots new behaviors from uncurated, multimodal human demonstrations, supervised by Rudolf Lioutikov. I am grateful for the Apple PhD Fellowship 2025 for supporting my research. I was an intern at Apple DMLI, working on robotics with Peide Huang and Jian Zhang. Previously, I obtained my Master's degree in Mechanical Engineering at KIT, where I wrote my thesis at Bosch Research, supervised by Gerhard Neumann. During my studies I interned at Audi AG, IPG Automotive, and the Research Center for Informatics (FZI).

Email  /  CV  /  Google Scholar  /  GitHub  /  LinkedIn  /  Blog

Portrait of Moritz Reuss
Research

My primary research goal is to build intelligent embodied agents that assist people in their everyday lives and communicate intuitively. I focus on language-conditioned multitask imitation learning from robot play data. My work explores efficient Vision-Language-Action (VLA) policies and novel policy representations that can learn from uncurated, multimodal human demonstrations without rewards. I have worked on score/flow-based diffusion policies and developed generalist VLA models that achieve strong performance with minimal computational requirements. Representative papers are highlighted.

  • Efficient VLAs: Compact flow/diffusion-based vision-language-action models (e.g., FLOWER, MoDE) and novel efficient action tokenization methods (e.g., BEAST) that achieve strong performance with minimal computational requirements.
  • Expressive Policy Representations: Diffusion and flow-based policies that capture multimodal action distributions and handle diverse goal specifications including language, images, and waypoints (e.g., BESO, MDT), with benchmarks for evaluating behavior diversity (e.g., D3IL).
  • Learning from Uncurated Data: Zero-shot dataset labeling with vision-language foundation models (e.g., NILS) to enable scalable learning from robot play data without manual annotation, and methods for handling multimodal observations (RGB + 3D inputs).
Research Papers
FLOWER VLA overview diagram
FLOWER: Democratizing Generalist Robot Policies with Efficient Vision-Language-Action Flow Policies
Moritz Reuss, Hongyi Zhou, Marcel Rühle, Ömer Erdinç Yağmurlu, Fabian Otto, Rudolf Lioutikov

CoRL, 2025
Project Page / Code / arXiv

We systematically analyze VLA design decisions for small and efficient VLAs. Our findings lead us to introduce FLOWER, a 950M-parameter Vision-Language-Action (VLA) policy that achieves state-of-the-art performance across 190 tasks in 10 benchmarks while requiring only ~1% of the pretraining compute of models like OpenVLA. FLOWER introduces intermediate-modality fusion and action-specific Global-AdaLN conditioning to achieve strong performance with improved efficiency. Our approach democratizes VLA development by making high-performance robotic foundation models accessible with commodity hardware, requiring significantly less GPU memory to run.

BEAST B-spline action tokenizer overview
BEAST: An Efficient Action Tokenizer with B-Splines
Hongyi Zhou, Weiran Liao, Xi Huang, Yucheng Tang, Fabian Otto, Xiaogang Jia, Xinkai Jiang, Simon Hilber, Ge Li, Qian Wang, Ömer Erdinç Yağmurlu, Nils Blank, Moritz Reuss, Rudolf Lioutikov

NeurIPS, 2025
Project Page / Code / arXiv

We introduce BEAST, a B-spline–based action tokenizer that efficiently represents continuous robot actions for generalist policies while maintaining smooth trajectories essential for robot control. BEAST enables efficient action representation and improved performance in VLA models by leveraging the mathematical properties of B-splines for smooth, continuous control. It flexibly supports continuous tokens and discrete tokenization. Experiments across various benchmarks verify good compression with strong performance and smooth behavior without additional temporal aggregation.

PointMapPolicy: structured point cloud grid for diffusion policy
PointMapPolicy: Structured Point Cloud Processing for Multi-Modal Imitation Learning
Xiaogang Jia, Qian Wang, Anrui Wang, Han A. Wang, Balázs Gyenes, Emiliyan Gospodinov, Xinkai Jiang, Ge Li, Hongyi Zhou, Weiran Liao, Xi Huang, Maximilian Beck, Moritz Reuss, Rudolf Lioutikov, Gerhard Neumann

NeurIPS, 2025
Project Page / Paper

We present PointMapPolicy, a multi-modal imitation learning method that conditions diffusion policies on structured grids of points without downsampling. Our approach fuses point maps with RGB using an xLSTM backbone, enabling direct application of computer vision techniques to 3D data while preserving fine-grained geometric details. On RoboCasa and CALVIN, plus real-robot evaluations, we achieve state-of-the-art performance across diverse manipulation tasks.

MoDE: mixture-of-denoising experts diffusion transformer policy
Efficient Diffusion Transformer Policies with Mixture of Expert Denoisers for Multitask Learning
Moritz Reuss*, Jyothish Pari*, Pulkit Agrawal, Rudolf Lioutikov

ICLR, 2025
Project Page / Code / arXiv

We propose Mixture-of-Denoising Experts (MoDE), a generalist policy for guided behavior generation that outperforms dense transformer-based diffusion policies in accuracy, parameter count, and efficiency. Our routing strategy conditions expert selection on the current noise level of the diffusion process. On four imitation learning benchmarks (including CALVIN and LIBERO), MoDE consistently exceeds dense transformer baselines. We pretrain MoDE on a subset of OXE for just 3 days on 6 GPUs and surpass OpenVLA and Octo on SIMPLER. MoDE achieves higher average performance with ~90% fewer FLOPs, ~20% faster inference, and ~40% fewer parameters compared to dense transformer diffusion policies.

Zero-shot relabeling pipeline with foundation models (LUPUS)
Scaling Robot Policy Learning via Zero-Shot Labeling with Foundation Models
Nils Blank, Moritz Reuss, Marcel Rühle, Ömer Erdinç Yağmurlu, Fabian Wenzel, Oier Mees, Rudolf Lioutikov

Conference on Robot Learning (CoRL), 2024 — Oral @ 2nd Workshop on Mobile Manipulation and Embodied Intelligence, ICRA 2024
Paper

We introduce a method to automatically label uncurated, long-horizon robot teleoperation data at scale in a zero-shot manner without human intervention. We combine pre-trained vision-language foundation models to detect objects, propose tasks, segment tasks in large unlabeled interaction datasets, and train language-conditioned policies on the relabeled data. Initial experiments show that our method enables policies that match those trained with oracle human annotations.

MDT: multimodal diffusion transformer on CALVIN
Multimodal Diffusion Transformer: Learning Versatile Behavior from Multimodal Goals
Moritz Reuss, Ömer Erdinç Yağmurlu, Fabian Wenzel, Rudolf Lioutikov

Robotics: Science and Systems (RSS), 2024 — Oral @ Workshop on Language and Robot Learning (LangRob), CoRL 2023
Project Page / Code / arXiv

We present a diffusion policy for learning from uncurated, reward-free offline data with sparse language labels. Multimodal Diffusion Transformer (MDT) learns complex, long-horizon behaviors and sets a new state of the art on CALVIN. MDT leverages pre-trained vision and language foundation models and aligns multimodal goal specifications in the transformer encoder’s latent space, using two self-supervised auxiliary objectives to better follow goals specified in language and images.

D3IL benchmark environments for imitation learning
Towards Diverse Behaviors: A Benchmark for Imitation Learning with Human Demonstrations
Xiaogang Jia, Denis Blessing, Xinkai Jiang, Moritz Reuss, Atalay Donat, Rudolf Lioutikov, Gerhard Neumann

ICLR, 2024
OpenReview

D3IL introduces simulation benchmark environments and datasets tailored for imitation learning, designed to evaluate a model’s ability to learn and replicate diverse, multimodal human behaviors. Environments encompass multiple sub-tasks and object manipulations, providing rich diversity often lacking in other datasets. We also propose practical metrics to quantify behavioral diversity and benchmark state-of-the-art methods.

BESO: goal-conditioned diffusion policy
Goal-Conditioned Imitation Learning using Score-based Diffusion Policies
Moritz Reuss, Maximilian Li, Xiaogang Jia, Rudolf Lioutikov

Best Paper Award @ Workshop on Learning from Diverse, Offline Data (L-DOD), ICRA 2023; RSS, 2023
Project Page / Code / arXiv

We present BESO, a policy representation for goal-conditioned imitation learning using score-based diffusion models. BESO effectively learns goal-directed, multimodal behavior from uncurated, reward-free offline data and achieves state of the art with as few as three denoising steps.

Information Maximizing Curriculum: obstacle avoidance result
Information Maximizing Curriculum: A Curriculum-Based Approach for Learning Versatile Skills
Denis Blessing, Onur Celik, Xiaogang Jia, Moritz Reuss, Maximilian Xiling, Rudolf Lioutikov, Gerhard Neumann

NeurIPS, 2023
arXiv

We introduce the Information Maximizing Curriculum to address mode-averaging in imitation learning by enabling specialization on representable data. A mixture-of-experts policy focuses on different data subsets, with a maximum-entropy objective for full dataset coverage.

Hybrid inverse dynamics model on Franka Emika Panda
End-to-End Learning of Hybrid Inverse Dynamics Models for Precise and Compliant Impedance Control
Moritz Reuss, Niels van Duijkeren, Robert Krug, Philipp Becker, Vaisakh Shaj, Gerhard Neumann

Robotics: Science and Systems (RSS), 2022
arXiv

We present a hybrid model combining a differentiable rigid-body model with a recurrent LSTM to accurately model the inverse dynamics of a robot manipulator. A differentiable formulation of barycentric parameters enables end-to-end training jointly with the residual network while preserving physical consistency. On a Franka Emika Panda, the model enables precise and compliant motion tracking.


The website is based on the source code.  |  Blog