The 5th Workshop on Machine Learning and Systems (EuroMLSys)

co-located with EuroSys '25

March 31st 2025, Rotterdam, The Netherlands


The recent wave of research focusing on machine intelligence (machine learning and artificial intelligence) and its applications has been fuelled by both hardware improvements and deep learning frameworks that simplify the design and training of neural models. Advances in AI also accelerate research towards Reinforcement Learning (RL), where dynamic control mechanisms are designed to tackle complex tasks. Further, machine learning based optimisation, such as Bayesian Optimisation, is gaining traction in the computer systems community where optimisation needs to scale with complex and large parameter spaces; areas of interest range from hyperparameter tuning to system configuration tuning,

The EuroMLSys workshop will provide a platform for discussing emerging trends in building frameworks, programming models, optimisation algorithms, and software engineering to support AI/ML applications. At the same time, using ML for building such frameworks or optimisation tools will be discussed. Recent emergence of LLM is remarked by their substantial computational requirements and optimisation in every possible part of the system will be important. EuroMLSys aims to bridge the gap between AI research and practice, through a technical program of fresh ideas on software infrastructure, tools, design principles, and theory/algorithms, from a systems perspective. We will also explore potential applications that will take advantages of ML.

News

  • The list of accepted papers is out!

Key dates

  • Paper submission deadline: February 7, 2025 (23:59 AoE) February 11, 2025 (23:59 AoE)
  • Acceptance notification: February 21, 2025 February 24, 2025
  • Final paper due: March 7, 2025 March 11, 2025
  • Registration due: March 17, 2025
  • Workshop: March 31, 2025 (full-day workshop)

Past Editions

Call for Papers

A growing area of interest in machine intelligence is at the intersection of AI/ML and systems design. At the same time, applications of ML are growing in complexity and so is the volume of data they produce/consume. For computer systems to scale, new learning approaches and advanced optimisation techniques are needed. We also need to understand better the current AI/ML frameworks, in terms of their functionality, limitations, and target applications. This will clarify potential desired functions and future architectures. Novel machine learning methods to optimise and accelerate software and hardware systems must also be developed.

EuroMLSys is an interdisciplinary workshop that brings together researchers in computer architecture, systems and machine learning, along with practitioners who are active in these emerging areas.

Topics of interest include, but are not limited to, the following:

  • Scheduling algorithms for data processing clusters
  • Custom hardware for machine learning
  • Programming languages for machine learning
  • Benchmarking systems (for machine learning algorithms)
  • Synthetic input data generation for training
  • Systems for training and serving machine learning models at scale
  • Graph neural networks
  • Neural network compression and pruning in systems
  • Systems for incremental learning algorithms
  • Large scale distributed learning algorithms in practice
  • Database systems for large scale learning
  • Model understanding tools (debugging, visualisation, etc.)
  • Systems for model-free and model-based Reinforcement Learning
  • Optimisation in end-to-end deep learning
  • System optimisation using Bayesian Optimisation
  • Acceleration of model building (e.g., imitation learning in RL)
  • Use of probabilistic models in ML/AI application
  • Learning models for inferring network attacks, device/service fingerprinting, congestion, etc.
  • Techniques to collect and analyze network data in a privacy-preserving manner
  • Learning models to capture network events and control actions
  • Machine learning in networking (e.g., use of Deep RL in networking)
  • Analysis of distributed ML algorithms
  • Semantics for distributed ML languages
  • Probabilistic modelling for distributed ML algorithms
  • Synchronisation and state control of distributed ML algorithms
  • ML Compiler Optimisation
  • Optimisation in Large Language Model (LLM)
  • Novel approaches to identify and mitigate bias in ML systems
  • Enhancing transparency and interpretability for fair AI
  • ML systems promoting equity, fairness, and diversity
  • Examining the societal and ecological impacts of ML systems

Accepted papers will be published in the ACM Digital Library (you can opt out from this).

Accepted Papers

  • Machine Learning-based Deep Packet Inspection at Line Rate for RDMA on FPGAsMaximilian Jakob Heer, Benjamin Ramhorst, Gustavo Alonso (ETH Zurich)

  • Practical Federated Learning without a ServerAkash Dhasade, Anne-Marie Kermarrec (EPFL); Erick Lavoie (University of Basel); Johan Pouwelse (Delft University of Technology); Rishi Sharma, Martijn de Vos (EPFL)

  • Leveraging Approximate Caching for Faster Retrieval-Augmented GenerationShai Aviram Bergman, Zhang Ji (Huawei); Anne-Marie Kermarrec, Diana Petrescu, Rafael Pires, Mathis Randl, Martijn de Vos (EPFL)

  • Efficient Federated Search for Retrieval-Augmented GenerationRachid Guerraoui, Anne-Marie Kermarrec, Diana Petrescu, Rafael Pires, Mathis Randl, Martijn de Vos (EPFL)

  • Verifying Semantic Equivalence of Large Models with Equality SaturationKahfi S. Zulkifli (University of Virginia); Wenbo Qian (Northeastern University); Shaowei Zhu, Yuan Zhou, Zhen Zhang (Amazon Web Services); Chang Lou (University of Virginia)

  • NeuraLUT-Assemble: Hardware-aware Assembling of Sub-Neural Networks for Efficient LUT InferenceMarta Andronic (Imperial College London); George A. Constantinides (Imperial College London, UK)

  • Systems Opportunities for LLM Fine-Tuning using Reinforcement LearningPedro F. Silvestre, Peter Pietzuch (Imperial College London)

  • Decentralized Adaptive Ranking using TransformersMarcel Gregoriadis, Quinten Stokkink, Johan Pouwelse (Delft University of Technology)

  • Performance Aware LLM Load Balancer for Mixed WorkloadsKunal Jain (Microsoft); Anjaly Parayil (Microsoft Research); Ankur Mallick, Esha Choukse (Microsoft); Xiaoting Qin, Jue Zhang (Microsoft Research); Íñigo Goiri, Rujia Wang, Chetan Bansal (Microsoft); Victor Rühle (Microsoft Research); Anoop Kulkarni, Steve Kofsky (Microsoft); Saravan Rajmohan (Microsoft 365)

  • Exploiting Unstructured Sparsity in Fully Homomorphic Encrypted DNNsAidan Ferguson, Perry Gibson, Lara D'Agata (University of Glasgow); Parker McLeod, Ferhat Yaman, Amitabh Das, Ian Colbert (AMD); José Cano (University of Glasgow)

  • Decoupling Structural and Quantitative Knowledge in ReLU-based Deep Neural NetworksJosé Duato (Qsimov Quantum Computing S.L.); Jose I. Mestre, Manuel F. Dolz (Universitat Jaume I); Enrique S. Quintana-Orti (Universitat Politècnica de València); José Cano (University of Glasgow)

  • RMAI: Rethinking Memory for AI (Inference)Amir Noohi (University of Edinburgh); Mostafa Derispour (Isfahan University of Technology); Antonio Barbalace (The University of Edinburgh)

  • Understanding Oversubscribed Memory Management for Deep Learning TrainingMao Lin, Hyeran Jeon (University of California, Merced)

  • Priority-Aware Preemptive Scheduling for Mixed-Priority Workloads in MoE InferenceMohammad Siavashi (KTH Royal Institute of Technology); Faezeh Keshmiri Dindarloo (Unaffiliated); Dejan Kostic, Marco Chiesa (KTH Royal Institute of Technology)

  • Diagnosing and Resolving Cloud Platform Instability with Multi-modal RAG LLMsYifan Wang, Kenneth P. Birman (Cornell University)

  • FlexInfer: Breaking Memory Constraint via Flexible and Efficient Offloading for On-Device LLM InferenceHongchao Du, Shangyu Wu (City University of Hong Kong); Arina Kharlamova (Mohamed bin Zayed University of Artificial Intelligence); Nan Guan (City University of Hong Kong); Chun Jason Xue (Mohamed bin Zayed University of Artificial Intelligence)

  • Deferred prefill for throughput maximization in LLM inferenceMoonmoon Mohanty, Gautham Bolar, Preetam Patil (Indian Institute of Science Bangalore); UmaMaheswari Devi, Felix George, Pratibha Moogi (IBM Research - India); Parimal Parag (Indian Institute of Science Bangalore)

  • AMPLE: Event-Driven Accelerator for Mixed-Precision Inference of Graph Neural NetworksPedro Gimenes (Imperial College London, UK); Aaron Zhao (Imperial College London); George A. Constantinides (Imperial College London, UK)

  • Client Availability in Federated Learning: It Matters!Dhruv Garg, Debopam Sanyal (Georgia Institute of Technology); Myungjin Lee (Cisco Systems); Alexey Tumanov, Ada Gavrilovska (Georgia Institute of Technology)
  • Accepted Poster Papers

  • Global-QSGD: Allreduce-Compatible Quantization for Distributed Learning with Theoretical GuaranteesJihao Xin, Marco Canini, Peter Richtárik (KAUST); Samuel Horváth (MBZUAI)

  • Hybrid Task Scheduling for Optimized Neural Network Inference on Skin Lesions in Resource-Constrained SystemsDiogen Babuc, Teodor-Florin Fortiş (West University of Timişoara)

  • Cross-Domain DRL Agents for Efficient Job Placement in the Cloud-Edge ContinuumTheodoros Aslanidis (University College Dublin); Sokol Kosta (Department of Electronic Systems, Aalborg University Copenhagen); Spyros Lalis (University of Thessaly); Dimitris Chatzopoulos (University College Dublin)

  • Towards a Unified Framework for Split LearningBoris Radovič (KAUST & University of Ljubljana); Marco Canini (KAUST); Samuel Horváth (MBZUAI); Veljko Pejović (University of Ljubljana); Praneeth Vepakomma (MBZUAI & MIT)

  • Manage the Workloads not the Cluster: Designing a Control Plane for Large-Scale AI ClustersRuiqi Lai, Siyu Cao, Leqi Li (NTU Singapore); Luo Mai (University of Edinburgh); Dmitrii Ustiugov (NTU Singapore)

  • Harnessing Increased Client Participation with Cohort-Parallel Federated LearningAkash Dhasade, Anne-Marie Kermarrec (EPFL); Tuan-Ahn Nguyen (Independent Researcher); Rafael Pires, Martijn de Vos (EPFL)

  • Accelerating MoE Model Inference with Expert ShardingOana Balmau (McGill); Anne-Marie Kermarrec, Rafael Pires, André Loureiro Espírito Santo, Martijn de Vos, Milos Vujasinovic (EPFL)

  • TAGC: Optimizing Gradient Communication in Distributed Transformer TrainingIgor Polyakov (VK, ITMO University); Alexey Dukhanov (ITMO University); Egor Spirin (VK Lab)

  • β-GNN: A Robust Ensemble Approach Against Graph Structure PerturbationHaci Ismail Aslan (Technical University of Berlin); Philipp Wiesner, Ping Xiong, Odej Kao (Technische Universität Berlin)

  • May the Memory Be With You: Efficient and Infinitely Updatable State for Large Language ModelsExcel Chukwu, Laurent Bindschaedler (Max Planck Institute for Software Systems)

  • Towards Asynchronous Peer-to-Peer Federated Learning for Heterogeneous SystemsChristos Sad (Aristotle University of Thessaloniki); George Retsinas, Dimitrios Soudris (National Technical University of Athens); Kostas Siozios (Aristotle University of Thessaloniki); Dimosthenis Masouros (National Technical University of Athens)

  • Beyond Test-Time Compute Strategies: Advocating Energy-per-Token in LLM InferencePatrick Wilhelm, Thorsten Wittkopp, Odej Kao (Technische Universität Berlin)

  • Utilizing Large Language Models for Ablation Studies in Machine Learning and Deep LearningSina Sheikholeslami, Hamid Ghasemirahni, Amir H. Payberah, Tianze Wang (KTH Royal Institute of Technology); Jim Dowling (Hopsworks AB); Vladimir Vlassov (KTH Royal Institute of Techonology, Sweden)

  • Rethinking Observability for AI workloads on Multi-tenant public cloudsTheophilus A. Benson (Carnegie Mellon University)

  • OptimusNIC: Offloading Optimizer State to SmartNICs for Efficient Large-Scale AI TrainingAchref Rebai, Marco Canini (KAUST)

  • Analysis of Information Propagation in Ethereum Network Using Combined Graph Attention Network and Reinforcement Learning to Optimize Network Efficiency and ScalabilityStefan Behfar, Richard Mortier, Jon Crowcroft (University of Cambridge)

Submission

Papers must be submitted electronically as PDF files, formatted for 8.5x11-inch paper. Submissions will be up to 6 pages long, including figures, and tables, with 10-point font, in a two-column format. Bibliographic references are not included in the 6-page limit. Submitted papers must use the official SIGPLAN Latex / MS Word templates.

Submissions will be single-blind.

Submit your paper at: https://euromlsys25.hotcrp.com/paper/new

Sponsors


Committees

Workshop and TPC Chairs

Technical Program Committee

  • Aaron Zhao, Imperial College London
  • Andy Twigg, Google
  • Ahmed Sayed, Queen Mary University of London
  • Alexandros Koliousis, Northeastern University London and Institute for Experiential AI
  • Amitabha Roy, Google
  • Chi Zhang, Brandeis University
  • Christos Bouganis, Imperial College London
  • Chunwei Xia , University of Leeds
  • Daniel Goodman, Oracle
  • Daniel Mendoza, Stanford University
  • Dawei Li, Amazon
  • Deepak George Thomas, Iowa State University
  • Dimitris Chatzopoulos, University College Dublin
  • Fiodar Kazhamiaka,Stanford University
  • Guilherme H. Apostolo, Vrije Universiteit Amsterdam
  • Guoliang He, University of Cambridge
  • Jenny Huang, Nvidia
  • Joana Tirana, University College Dublin
  • Jon Crowcroft, University of Cambridge
  • Jose Cano, University of Glasgow
  • Luo Mai, University of Edinburgh
  • Mark Zhao, Stanford University
  • Mengying Zhou, Fudan University
  • Nikolas Ioannou, Google
  • Paul Patras, University of Edinburgh
  • Peter Pietzuch, Imperial College London
  • Peter Triantafillou, University of Warwick
  • Pinar Tözün, IT University of Copenhagen
  • Pouya Hamadanian, MIT
  • Sam Ainsworth, University of Edinburgh
  • Sami Alabed, Deepmind
  • Sandra Siby, NYU Abu Dhabi
  • Shivaram Venkataraman, University of Wisconsin-Madison
  • Taiyi Wang, University of Cambridge
  • Thaleia Dimitra Doudali, IMDEA
  • Valentin Radu, University of Sheffield
  • Veljko Pejovic, University of Ljubljana
  • Xupeng Miao, Peking University
  • Yaniv Ben-Itzhak, Broadcom
  • Youhe Jiang, University of Cambridge
  • Yuchen Zhao, University of York
  • Zak Singh, University of Cambridge
  • Zheng Wang, University of Leeds
  • Zhihao Jia, CMU

Web Chair

  • Alexis Duque, Net AI

Contact

For any question(s) related to EuroMLSys 2025, please contact the TPC Chairs Eiko Yoneki and Amir H. Payberah.

Follow us on Twitter: @euromlsys