Current projects

Research Projects

Projects supported by industry partners


Semantics and Implementation of ACID properties in modern hardware [Gustavo Alonso]


SoftHier: Exploring Explicitly Managed Memory Hierarchy Architectures for Scalable Acceleration [Torsten Hoefler/Luca Benini]

Enlarged view: SoftHier project

The objective of SoftHier is to drastically reduce hardware inefficiency by replac-ing caches with software-managed, efficient SPMs and DMAs, while maintaining program-ming effort low for the application developer. This goal will be achieved thanks to: (i) An increased level of automation in the programming tool-flow; (ii) The use of domain-specific languages and abstractions; (iii) Hardware enhancements in the memory hierarchy that facilitate automatic instantiation of accelerated data-motion primitives from abstract, streamlined software APIs.

Key achievements:

  • Defined baseline template architecture for SoftHier, explored GEMM dataflow mapping, architecture design space, PPA
  • Developed a High-Level Simulation Model with Open-source release
  • PPA estimation for SoftHier Architecture

QuantSparse3D: A Stacked-Memory Multi-Chiplet inference Engine for Sparse, Quantized LLMs [Luca Benini]


LLM Agent Firewall [Srdjan Capkun]

ifsc projects page figure

Large Language Model (LLM) agents offer increasingly rich functionalities and capabilities to act on users' inputs and data. However, it remains unclear how to safely restrict LLM agents to the desired data and functionalities, and what are their implications on the privacy and integrity of the systems that they operate in.

In this project, we investigate how LLM agents need to be restricted in terms of the access to system data and resources, actions that they are allowed to take, and which access control models and mechanisms should be deployed to achieve access control and isolation properties. It is not yet clear whether existing solutions can be leveraged in an agentic setting and what are the tradeoffs of different solutions. For example, in a corporate setting where different roles have access to different projects, a single assistant trained on all the company's data provides efficient training but insufficient access control due to threats such as memorization and membership inference attacks. Instead, training one model for each permission level can quickly lead to training an exponential number of models.

We aim to analyze the combinations of novel LLM environments that will restrict the agents, novel system access control policies and LLM training, all of which combined will limit LLM agents. This is a challenging problem, as solutions must (i) guarantee information isolation between different groups while (ii) allowing computationally reasonable training, finetuning, and updating, and (iii) not harming functionality and providing high-quality results.  


Graph Computations and LLMs: A Synergy [Torsten Hoefler]


Unifying High-Performance Automated Differentiation for Machine Learning and Scientific Computing [Torsten Hoefler]


Processing-in-Memory Architectures for Data-Intensive Applications [Onur Mutlu]

Research Grants

Internal research projects supported by an EFCL grant







Student projects

Smaller projects based on pre-PhD research supported by an EFCL grant


JavaScript has been disabled in your browser