Energy Efficient Machine Learning and Cognitive Computing

11th Edition

Co-located with ASPLOS 2026 in Pittsburgh, PA

event Sunday, March 22, 2026 meeting_room Room: Ft. Pitt

emoji_events EMC2 Competition: AI Infrastructure Demos, sponsored by Runara.aiLearn more →

local_bar EMC2 Social: Join us after the workshop! 6:15 – 8:30 PM at The Lobby Bar, The Landing Hotel at Rivers Casino. Learn more →

emoji_events EMC2 Competition: AI Infrastructure Demos

Sponsored by Runara.ai

Overview

The EMC2 Competition invites researchers, practitioners, and students to present production-ready AI infrastructure systems demonstrating measurable improvements in efficiency, observability, and operational robustness. The competition is designed to surface systems that operate under real-world constraints- including scale, reliability, cost, and sustainability.

The primary goal is to move beyond toy benchmarks, isolated micro-optimizations, and paper-only abstractions. We want to see working demos, toolkits, and integrated systems that can be realistically deployed in modern AI infrastructure environments. Submissions should emphasize live execution, actionable telemetry, and end-to-end impact rather than synthetic evaluations.

Competition Tracks

Track 1: Efficient Inference

Systems improving cost, performance, and energy efficiency of inference workloads:

  • Runtime inference optimization
  • Scheduling, batching, parallelism
  • Hardware-aware execution
  • Cost/energy-aware pipelines
Track 2: Infrastructure Observability

Visibility, monitoring, and diagnostics across the AI stack:

  • Live telemetry and visualization
  • Utilization and bottleneck analysis
  • Cross-layer observability
  • Actionable operator insights

Hardware and Model Flexibility

Participants are free to choose any hardware platform and any inference model, including GPUs or custom accelerators, cloud-based, on-premise, or edge environments, and open-source or proprietary models. There are no restrictions on vendors, architectures, or model families, provided the submission demonstrates real execution, live metrics, and measurable impact.

Submission Requirements

  • Team size: Up to 2 participants
  • Format: Single-page proposal (PDF) describing the problem, system architecture, hardware/model used, live metrics captured, and how efficiency or observability improvements are demonstrated
  • Originality: The idea does not need to be novel, but implementation must be original. Prior work may be extended with clear attribution.
  • Demo requirement (mandatory): Submissions must include a working demo that tracks live metrics, demonstrates real execution, and clearly shows improvements in cost, performance, or sustainability. Simulation-only or slide-only submissions will not be considered.

Submission deadline: March 15

Selection & Presentation

Accepted teams will display their demos during the workshop lunch break and engage directly with judges during in-person evaluations. Based on judging scores, the top 3 teams from each track will be selected for oral presentations.

Final Presentations: 4:00 - 5:00 PM, 10 minutes per team, Live demo encouraged

At the conclusion of the session, one winning team per track will be selected.

Awards

Each winning team receives:

$500 cash prize + Certificate of Recognition

Winners eligible for Summer Internship opportunities at Runara.ai

Evaluation Criteria

Since this is an open-scope, applied competition, proposals will be evaluated based on their real-world application potential instead of rigid benchmark metrics. Judging prioritizes: production readiness and robustness, clarity and usefulness of metrics, real measurable end-to-end impact, clean system design, and practical relevance to modern AI infrastructure.

For any questions please contact raj@runara.ai

schedule 16:00 - 17:00 event Sunday, March 22, 2026 room Ft. Pitt

forum Architecting an energy-first stack for the AI age

As the cumulative demand for AI computing continues to grow, new bottlenecks in technology deployment are emerging. These range from compute capacity to memory supply to cluster reliability to simply finding enough power to supply the infrastructure. Since these aren’t easily solved with brute force addition of more chips, potential solutions encompass new model architectures, new computing architectures, high performance cooling designs, robust multi-device software stacks, and even orbital datacenters. In this panel, we will aim to identify the most pressing requirements for the computing stack of coming years. We will discuss both the promising technologies and the hypes to define a blueprint of the computing architecture where energy is the currency of success.

Moderator

Panelists

Esha Choukse
Esha Choukse link
Microsoft
Zain Asgar
Zain Asgar link
Gimlet AI
Gennady Pekhimenko
Gennady Pekhimenko link
University of Toronto

description Accepted Papers

Paper Session #1 (1:30 PM – 2:15 PM)
  • piPE-SA: Enabling Deeply Pipelined Processing Elements in Systolic Arrays

    Jiayi Wang, Chenyi Wang and Ang Li

  • CacheFlex: Explicitly Control What You Need in Your Cache

    Jingqun Zhang, Weihang Li, Maohua Nie, Yung-Jen Cheng, Jiayi Wang, Shwet Chitnis and Ang Li

  • IEI: A Composite Infrastructure Efficiency Index for GPU Inference

    Raj Parihar

Paper Session #2 (2:45 PM – 3:30 PM)
  • Fast NF4 Dequantization Kernels for Large Language Model Inference

    Xiangbo Qi, Chaoyi Jiang and Murali Annavar

  • I-Fuse: Profile-Guided Speculative Load Micro-op Fusion for Data Center Applications

    Deepanjali Mishra, Tanvir Ahmed Khan, Gilles Pokam, Heiner Litz and Akshitha Sriraman

  • Store or Recompute? Characterizing the Carbon Tradeoff of KV Cache Retention in LLM Inference

    Abnash Bassi, Jaylen Wang, Fiodar Kazhamiaka, Daniel S. Berger and Akshitha Sriraman

local_bar EMC2 Social

Sponsored by d-Matrix

Join us after the workshop for an informal social! This is a great opportunity to connect with fellow researchers, speakers, and organizers in a relaxed setting.

description Workshop Objective

In the Eleventh edition of EMC2 workshop, we plan to facilitate conversation about the sustainability of large-scale AI computing systems being developed to meet the ever-increasing demands of generative AI. This involves discussions spanning multiple interrelated areas. First, we continue to serve as the leading forums for discussing the energy-efficiency aspect of GenAI workloads which directly impact the overall viability and economic value of AI technology. Second, we reassess the scaling laws of AI with the prevalence of agentic, multi-modal, and reasoning-based models in conjunction with novel techniques such as a highly sparse expert architecture and disaggregated computation. Finally, we discuss sustainable and high-performance computing paradigms towards efficient datacenters and hybrid computing models that can cater to the exponential growth in model sizes, application areas, anduser base. This would allow us to explore ideas to build the hardware, software, systems, and scaling infrastructure, as well as model architectures that make AI technology even more prevalent and accessible.

chat Call for Papers

The goal of this Workshop is to provide a forum for researchers and industry experts who are exploring novel ideas, tools and techniques to improve the energy efficiency of MLLMs as it is practised today and would evolve in the next decade. We envision that only through close collaboration between industry and the academia we will be able to address the difficult challenges and opportunities of reducing the carbon footprint of AI and its uses. We have tailored our program to best serve the participants in a fully digital setting. Our forum facilitates active exchange of ideas through:

  • Keynotes, invited talks and discussion panels by leading researchers from industry and academia
  • Peer-reviewed papers on latest solutions including works-in-progress to seek directed feedback from experts
  • Independent publication of proceedings through IEEE CPS

We invite full-length papers describing original, cutting-edge, and even work-in-progress research projects about efficient machine learning. Suggested topics for papers include, but are not limited to the ones listed below:

format_list_bulleted Topics for the Workshop

  • Neural network architectures for resource constrained applications.
  • Efficient hardware designs to implement neural networks including sparsity, locality, and systolic designs.
  • Power and performance efficient memory architectures suited for neural networks.
  • Network reduction techniques – approximation, quantization, reduced precision, pruning, distillation, and reconfiguration.
  • Exploring interplay of precision, performance, power, and energy through benchmarks, workloads, and characterization.
  • Performance potential, limit studies, bottleneck analysis, profiling, and synthesis of workloads.
  • Explorations and architctures aimed to promote sustainable computing.
  • Simulation and emulation techniques, frameworks, tools, and platforms for machine learning.
  • Optimizations to improve performance of training techniques including on-device and large-scale learning.
  • Load balancing and efficient task distribution, communication and computation overlapping for optimal performance.
  • Verification, validation, determinism, robustness, bias, safety, and privacy challenges in AI systems.
  • Efficient deployment strategies for edge and distributed environments.
  • Model compression and optimization techniques that preserve reasoning and problem-solving capabilities.
  • Architectures and frameworks for multi-agent systems and retrieval-augmented generation (RAG) pipelines.
  • Systems-level approaches for scaling future foundation models (e.g., Llama 4, GPT-5 and beyond).
  • We will follow that same formatting guidelines and duplicate submission policies as ASPLOS.

08:00 - 08:15
Welcome

Welcome and Opening Remarks

Sushant Kondguli, Meta

08:15 - 09:00
Keynote

Future of Energy-Efficient Cognitive Computing: A Six-Word Story on Sentience, Systems, and Sustainability

Parthasarathy Ranganathan, Google link

We are at a pivotal inflection point in the design of computing systems. On one hand, demand for computing is accelerating at phenomenal rates, fueled by the AI revolution and increasingly deep processing on massive data volumes. On the other hand, Moore’s Law is slowing down. This widening supply-demand gap is forcing us to revisit traditional assumptions around systems design. In this talk, we will discuss Google’s experience designing and deploying large-scale AI systems optimized for efficiency, reliability, and velocity. Building on these lessons, we will identify key challenges and opportunities for future innovation, highlighting how the next generation of energy-efficient cognitive computing will be AI-driven, vertically integrated, and uncomfortably exciting!

09:00 - 09:30
Invited Talk

Architecting Responsible AI: Efficient and Carbon-Aware Caching for AI and Its Security Implications

Sihang Liu, University of Waterloo link

While generative AI offers transformative potential, its rapid scaling has introduced significant challenges in sustainability and responsible deployment. This talk explores the role of caching in AI systems, including both performance and its broader implications for sustainability and security. First, I will present our recent work on improving the efficiency of video generation systems through caching techniques. I will then discuss our context caching system that explores the trade-off between operational carbon savings and the increased embodied carbon associated with storage. Finally, I will highlight our recent findings on novel security vulnerabilities arising from caching mechanisms used in image generation services.

09:30 - 10:00
Invited Talk

Energy Consumption in AI Datacenters: Can We Address this Challenge?

Josep Torrellas, UIUC link

The global electricity demand from data centers, AI, and cryptocurrencies in expected to be around 800TW/h by the end of 2026. Currently, companies like Oracle are building 500MW AI campuses. Training a single model takes 100+ MW over many months, with clusters of about 100K GPUs, and the projected AI cluster size is expected to increase several times in the next few years. Clearly, power will be the primary limiter to the growth of AI. Given this scenario, what can we, researchers, do? In this talk, I will discuss the problem and suggest some techniques that we can use to try to mitigate it. Some of the ideas are being developed in the context of the ACE Center for Evolvable Computing, a center funded by SRC and DARPA for efficient distributed computing.

10:00 - 10:30
Break

Break

10:30 - 11:00
Invited Talk

Memory-Centric Computing: Solving Computing's Memory Problem

Onur Mutlu, ETH Zurich link

Computing is bottlenecked by data. Large amounts of application data overwhelm the storage capability, communication capability, and computation capability of the modern machines we design today. As a result, many key applications’ performance, efficiency, and scalability are bottlenecked by data movement. In this talk, we describe three major shortcomings of modern computers in terms of 1) dealing with data, 2) taking advantage of vast amounts of data, and 3) exploiting different semantic properties of application data. We argue that an intelligent computing architecture should be designed to handle data well. We posit that handling data well requires designing architectures based on three key principles: 1) data-centric, 2) data-driven, 3) data-aware. We give examples of how to exploit these principles to design a much more efficient and higher performance computing system. We especially discuss recent research that aims to fundamentally reduce memory latency and energy, and practically enable computation close to data, with at least two promising directions: 1) processing using memory, which exploits the fundamental operational properties of memory chips to perform massively-parallel computation in memory, with low-cost changes, 2) processing near memory, which integrates sophisticated additional processing capability in memory chips, the logic layer of 3D-stacked technologies, or memory controllers to enable near-memory computation with high memory bandwidth and low memory latency. We show both types of architectures can enable order(s) of magnitude improvements in performance and energy consumption of many important workloads, such as artificial intelligence, machine learning, graph analytics, database systems, video processing, climate modeling, genome analysis. We discuss how to enable adoption of such fundamentally more intelligent architectures, which are key to efficiency, performance, and sustainability. We conclude with some research opportunities in and guiding principles for future computing architecture and system designs.

An accompanying overview of modern memory-centric computing ideas & systems can be found at A Modern Primer on Processing in Memory (updated February 2025).

A shorter invited paper from IMW 2025 is at Memory-Centric Computing: Solving Computing’s Memory Problem (May 2025).

11:00 - 11:30
Invited Talk

Efficient and Scalable Agentic AI With Heterogeneous Systems

Zain Asgar, Gimlet AI link

AI agents are dynamic, multi-stage systems composed of diverse operations with very different resource characteristics, ranging from compute- and memory-intensive model execution to bandwidth- and IO-bound data processing. Yet most deployments today run on homogeneous infrastructure, despite inference itself being inherently heterogeneous across its phases. In this talk, we present a system for orchestrating agent workloads across a mix of CPUs, GPUs, and accelerators spanning vendors and hardware tiers. The system automatically decomposes agentic workloads into fine-grained execution graphs, compiles them into hardware-specific fragments, and dynamically places and stitches them across distributed infrastructure to meet latency and performance goals. By mapping each stage of execution to the hardware best suited for it, we show how we can achieve step-function performance gains and significant improvements in cost efficiency. We highlight results showing how emerging hardware like SRAM-centric chips, combined with existing GPU infrastructure, can outperform traditional homogeneous deployments.

11:30 - 12:00
Invited Talk

Private Neural Recommendation with Homomorphic Encryption and Orion

Brandon Reagen, NYU link

AI is great. It powers many of our favorite services and drives industries. However, today’s solutions pose a tradeoff between utility and privacy where receiving the highest quality service often requires disclosing private information. In this talk I will show how things don’t have to be this way. Fully Homomorphic Encryption (FHE) is a cryptographic method that enables computation directly on encrypted data, never disclosing sensitive inputs while still enabling access to high-quality services. I will then cover the challenges of using this technology, which include both extreme performance overhead and programming difficulty, and how our Orion framework addresses them. Finally, I will highlight our most recent work that demonstrates how FHE can be applied to neural recommendation. Time permitting, a demonstration will be given.

12:00 - 13:15
Lunch Break

Lunch Break

13:15 - 13:30
Invited Talk

Sponsor Talk

Max Sbabo, d-Matrix

13:30 - 14:15
Paper Session #1

piPE-SA: Enabling Deeply Pipelined Processing Elements in Systolic Arrays

Jiayi Wang, Chenyi Wang and Ang Li

CacheFlex: Explicitly Control What You Need in Your Cache

Jingqun Zhang, Weihang Li, Maohua Nie, Yung-Jen Cheng, Jiayi Wang, Shwet Chitnis and Ang Li

IEI: A Composite Infrastructure Efficiency Index for GPU Inference

Raj Parihar

14:15 - 14:45
Invited Talk

Making the Accuracy-Efficiency Trade-Off in Agentic Systems

Esha Choukse, Microsoft link

For decades, systems researchers have balanced competing objectives: latency versus throughput, performance versus power, and speed versus cost. Traditionally, accuracy was a fixed requirement—non-negotiable and external to the systems equation. But as modern computing increasingly hosts AI-driven workloads, accuracy itself has become a tunable system variable. In this talk, I’ll argue that the next frontier in systems design lies in treating accuracy as a first-class performance knob, to be traded for efficiency in a principled way. Drawing from our recent work, I will show how software–hardware co-design can explicitly expose, quantify, and manage the accuracy-efficiency tradeoff.

14:45 - 15:30
Paper Session #2

Fast NF4 Dequantization Kernels for Large Language Model Inference

Xiangbo Qi, Chaoyi Jiang and Murali Annavar

I-Fuse: Profile-Guided Speculative Load Micro-op Fusion for Data Center Applications

Deepanjali Mishra, Tanvir Ahmed Khan, Gilles Pokam, Heiner Litz and Akshitha Sriraman

Store or Recompute? Characterizing the Carbon Tradeoff of KV Cache Retention in LLM Inference

Abnash Bassi, Jaylen Wang, Fiodar Kazhamiaka, Daniel S. Berger and Akshitha Sriraman

15:30 - 16:00
Break

Break

16:00 - 17:00
Panel

Architecting an energy-first stack for the AI age

As the cumulative demand for AI computing continues to grow, new bottlenecks in technology deployment are emerging. These range from compute capacity to memory supply to cluster reliability to simply finding enough power to supply the infrastructure. Since these aren’t easily solved with brute force addition of more chips, potential solutions encompass new model architectures, new computing architectures, high performance cooling designs, robust multi-device software stacks, and even orbital datacenters. In this panel, we will aim to identify the most pressing requirements for the computing stack of coming years. We will discuss both the promising technologies and the hypes to define a blueprint of the computing architecture where energy is the currency of success.

Moderator: Satyam Srivastava, d-Matrix

Panelists:
  • Nishil Talati, UIUC
  • Esha Choukse, Microsoft
  • Josep Torrellas, UIUC
  • Zain Asgar, Gimlet AI
  • Gennady Pekhimenko, University of Toronto
17:00 - 17:30
Invited Talk

Agile and evolvable software construction in the era of rapidly evolving hardware accelerator designs

Charith Mendis, UIUC link

Modern AI workloads have become exceedingly abundant and important in the current computing landscape. As a result, there have been numerous software and hardware innovations aimed at accelerating these workloads. However, we observe a subtle disconnect between the software and hardware communities. Most software innovations target well-established hardware platforms such as CPUs (e.g., x86, ARM) and GPUs (e.g., NVidia GPUs), while hardware innovations produce plenty of other tensor accelerator designs (e.g., Gemmini, Feather, Trainium) each year.

We asked the question, why aren’t the software community using these accelerators or even evaluating on them? The simple yet undeniable reason is the lack of standardized software tooling compared to CPUs and GPUs. For an architecture to be used, properly designed compiler backends, correctness, and performance testing tools should be abundant (e.g., CUDA ecosystem).

In this talk, I will describe how we bridge this gap by automatically generating the necessary software tools for a large class of accelerators through the Accelerator Compiler Toolkit (ACT) ecosystem. Central to ACT is an ISA definition language, TAIDL, that for the first time standardizes the hardware-software interfaces for a large class of accelerators. Departing from the traditional approach of manually constructing test oracles, performance models, or retargetable compiler backends, we instead introduce agile and evolvable methodologies to automatically generate such necessary tooling using both formal methods and machine learning techniques for any TAIDL-defined accelerator interface. I will show how such automation enables rapid software prototyping, making rapidly evolving accelerator designs usable by the software community.

17:30 - 18:00
Invited Talk

Making Long-Context LLM Inference Practical: Leveraging Token Importance for Efficient Inference

Urmish Thakker, Ex-SambaNova Systems link

As LLM context windows scale to hundreds of thousands or even millions of tokens, inference efficiency is increasingly limited by KV-cache memory growth and the high cost of prompt prefilling. Prior approaches attempt to mitigate these costs through heuristic cache eviction or prompt compression techniques, but often rely on architecture-specific assumptions or require retraining to recover accuracy. In this talk, I show that LLMs implicitly encode token importance signals that reveal which parts of context matter most for downstream computation. By exploiting these signals, it is possible to selectively retain or process only the most relevant tokens, significantly improving both decoding throughput and prefill throughput while preserving model quality in long-context inference.

18:00 - 18:05
Close

Closing Remarks

Satyam Srivastava, d-Matrix

18:15 - 20:30
Social

Social