Maximilian Böther

I am pursuing a Ph.D. in Computer Science at ETH Zurich's Systems Group and the Efficient Architectures and Systems Lab (EASL), supervised by Ana Klimovic and Gustavo Alonso. I work on the intersection of systems and data-centric AI. My interests span large-scale LLM/VLM training, data management for machine learning, and machine learning pipelines and deployments.
I have worked on Mixtera (Github), a lightweight data plane for LLM/VLM training, and on Modyn (Github), a platform for training models on datasets that grow over time. I also am involved in Apertus, Switzerland’s national LLM.
I published at venues such as SIGMOD, VLDB, MLSys, and ICLR, and interned at Google and Apple. I obtained B.Sc. and M.Sc. degrees in IT-Systems Engineering from Hasso Plattner Institute, Potsdam, Germany in 2020 and 2022. Please find my CV here.
news
Sep 1, 2025 | We just released Apertus v1, Switzerland’s national LLM. Happy to have contributed to the pretraining data! The models can be found on huggingface and the technical report (v0.1) on Github. |
---|---|
Jun 27, 2025 | I attended SIGMOD’25 in Berlin and presented Modyn at the conference! |
May 19, 2025 | I started as an ML Intern at Apple in Seattle, working on reinforcement learning infrastructure for diffusion models! |
May 15, 2025 | I presented our distributed data selection paper at MLSys 2025, and gave a talk at DatologyAI on Mixtera. |
Mar 19, 2025 | I have received the ML and Systems Rising Star Award 2025. Thank you so much! |
Mar 4, 2025 | I presented Mixtera and Modyn at BTW’25. Thank you for the great discussions! |
Mar 1, 2025 | We just released a preprint on Mixtera, our data plane for foundation model training. If you are training LLMs or VLMs, and are looking for infrastructure for data loading and mixing, please feel free to reach out! |
Feb 10, 2025 | Our paper on distributed submodular subset selection–a result from my Google internship–has been accepted to MLSys’25. See you soon in Santa Clara! |
Oct 31, 2024 | Our paper on Modyn has been accepted to SIGMOD’25 in Berlin! |
Oct 4, 2024 | We organized the Systems for Cost-Efficient AI Track at the AI+X Summit in Zurich. |
Sep 30, 2024 | Our vision paper on Mixtera, our lightweight data lake for LLM training, has been accepted to HotInfra’24 at SOSP. See you in Austin, TX! |
Aug 2, 2024 | Happy to have attended the Dagstuhl Seminar 24311: Resource-Efficient Machine Learning. |
Jun 17, 2024 | I will talk about Modyn at the Data-centric Machine Learning (DML) workshop at ICLR’24. See you in Vienna! |
Feb 26, 2024 | We just released a preprint of our paper on scaling out practical subset selection using submodular functions. This paper is a result of my internship at Google. |
Jun 17, 2023 | Our paper on analyzing vectorized hash tables across CPU architectures just got accepted at VLDB’23 in Vancouver! |
Jun 5, 2023 | I joined Google for a summer research internship in Sunnyvale, California, USA! I am working on scaling out submodular data subset selection. |
Apr 11, 2023 | Our work-in-progress workshop paper on Modyn, our research platform for model training on dynamic datasets, has been accepted at EuroMLSys’23 in Rome! |
Nov 1, 2022 | I joined the ETH Zurich Systems Group and the Efficient Architectures and Systems Lab (EASL) to do a Ph.D. in Machine Learning Systems, supervised by Professor Ana Klimovic. Looking forward to the new adventures in Switzerland! |
Oct 15, 2022 | Our paper on efficiently computing directed minimum spanning trees (arboresence) has been accepted for publication at ALENEX 2023. Check out the final version here. |
Jun 6, 2022 | Our Law Smells paper, which applies concepts of software engineering to the law, has been published in AI&Law. Check out the final version here. |
Jan 24, 2022 | Our paper on deep learning for combinatorial optimization just got accepted at ICLR! Check out the final version here. |
Dec 27, 2021 | This website just went online! |