Energy Efficient Machine Learning and Cognitive Computing
11th Edition
Co-located with ASPLOS 2026 in Pittsburgh, PA
emoji_events EMC2 Competition: AI Infrastructure Demos
Sponsored by Runara.ai
Overview
The EMC2 Competition invites researchers, practitioners, and students to present production-ready AI infrastructure systems demonstrating measurable improvements in efficiency, observability, and operational robustness. The competition is designed to surface systems that operate under real-world constraints- including scale, reliability, cost, and sustainability.
The primary goal is to move beyond toy benchmarks, isolated micro-optimizations, and paper-only abstractions. We want to see working demos, toolkits, and integrated systems that can be realistically deployed in modern AI infrastructure environments. Submissions should emphasize live execution, actionable telemetry, and end-to-end impact rather than synthetic evaluations.
Competition Tracks
Systems improving cost, performance, and energy efficiency of inference workloads:
- Runtime inference optimization
- Scheduling, batching, parallelism
- Hardware-aware execution
- Cost/energy-aware pipelines
Visibility, monitoring, and diagnostics across the AI stack:
- Live telemetry and visualization
- Utilization and bottleneck analysis
- Cross-layer observability
- Actionable operator insights
Hardware and Model Flexibility
Participants are free to choose any hardware platform and any inference model, including GPUs or custom accelerators, cloud-based, on-premise, or edge environments, and open-source or proprietary models. There are no restrictions on vendors, architectures, or model families, provided the submission demonstrates real execution, live metrics, and measurable impact.
Submission Requirements
- Team size: Up to 2 participants
- Format: Single-page proposal (PDF) describing the problem, system architecture, hardware/model used, live metrics captured, and how efficiency or observability improvements are demonstrated
- Originality: The idea does not need to be novel, but implementation must be original. Prior work may be extended with clear attribution.
- Demo requirement (mandatory): Submissions must include a working demo that tracks live metrics, demonstrates real execution, and clearly shows improvements in cost, performance, or sustainability. Simulation-only or slide-only submissions will not be considered.
Submission deadline: March 15
Selection & Presentation
Accepted teams will display their demos during the workshop lunch break and engage directly with judges during in-person evaluations. Based on judging scores, the top 3 teams from each track will be selected for oral presentations.
Final Presentations: 4:00 - 5:00 PM, 10 minutes per team, Live demo encouraged
At the conclusion of the session, one winning team per track will be selected.
Awards
Each winning team receives:
$500 cash prize + Certificate of Recognition
Winners eligible for Summer Internship opportunities at Runara.ai
Evaluation Criteria
Since this is an open-scope, applied competition, proposals will be evaluated based on their real-world application potential instead of rigid benchmark metrics. Judging prioritizes: production readiness and robustness, clarity and usefulness of metrics, real measurable end-to-end impact, clean system design, and practical relevance to modern AI infrastructure.
For any questions please contact raj@runara.ai
forum Architecting an energy-first stack for the AI age
As the cumulative demand for AI computing continues to grow, new bottlenecks in technology deployment are emerging. These range from compute capacity to memory supply to cluster reliability to simply finding enough power to supply the infrastructure. Since these aren’t easily solved with brute force addition of more chips, potential solutions encompass new model architectures, new computing architectures, high performance cooling designs, robust multi-device software stacks, and even orbital datacenters. In this panel, we will aim to identify the most pressing requirements for the computing stack of coming years. We will discuss both the promising technologies and the hypes to define a blueprint of the computing architecture where energy is the currency of success.
Moderator
Panelists
description Accepted Papers
Paper Session #1 (1:30 PM – 2:15 PM)
-
piPE-SA: Enabling Deeply Pipelined Processing Elements in Systolic Arrays
Jiayi Wang, Chenyi Wang and Ang Li
-
CacheFlex: Explicitly Control What You Need in Your Cache
Jingqun Zhang, Weihang Li, Maohua Nie, Yung-Jen Cheng, Jiayi Wang, Shwet Chitnis and Ang Li
-
IEI: A Composite Infrastructure Efficiency Index for GPU Inference
Raj Parihar
Paper Session #2 (2:45 PM – 3:30 PM)
-
Fast NF4 Dequantization Kernels for Large Language Model Inference
Xiangbo Qi, Chaoyi Jiang and Murali Annavar
-
I-Fuse: Profile-Guided Speculative Load Micro-op Fusion for Data Center Applications
Deepanjali Mishra, Tanvir Ahmed Khan, Gilles Pokam, Heiner Litz and Akshitha Sriraman
-
Store or Recompute? Characterizing the Carbon Tradeoff of KV Cache Retention in LLM Inference
Abnash Bassi, Jaylen Wang, Fiodar Kazhamiaka, Daniel S. Berger and Akshitha Sriraman
description Workshop Objective
In the Eleventh edition of EMC2 workshop, we plan to facilitate conversation about the sustainability of large-scale AI computing systems being developed to meet the ever-increasing demands of generative AI. This involves discussions spanning multiple interrelated areas. First, we continue to serve as the leading forums for discussing the energy-efficiency aspect of GenAI workloads which directly impact the overall viability and economic value of AI technology. Second, we reassess the scaling laws of AI with the prevalence of agentic, multi-modal, and reasoning-based models in conjunction with novel techniques such as a highly sparse expert architecture and disaggregated computation. Finally, we discuss sustainable and high-performance computing paradigms towards efficient datacenters and hybrid computing models that can cater to the exponential growth in model sizes, application areas, anduser base. This would allow us to explore ideas to build the hardware, software, systems, and scaling infrastructure, as well as model architectures that make AI technology even more prevalent and accessible.
chat Call for Papers
The goal of this Workshop is to provide a forum for researchers and industry experts who are exploring novel ideas, tools and techniques to improve the energy efficiency of MLLMs as it is practised today and would evolve in the next decade. We envision that only through close collaboration between industry and the academia we will be able to address the difficult challenges and opportunities of reducing the carbon footprint of AI and its uses. We have tailored our program to best serve the participants in a fully digital setting. Our forum facilitates active exchange of ideas through:
- Keynotes, invited talks and discussion panels by leading researchers from industry and academia
- Peer-reviewed papers on latest solutions including works-in-progress to seek directed feedback from experts
- Independent publication of proceedings through IEEE CPS
We invite full-length papers describing original, cutting-edge, and even work-in-progress research projects about efficient machine learning. Suggested topics for papers include, but are not limited to the ones listed below:
format_list_bulleted Topics for the Workshop
- Neural network architectures for resource constrained applications.
- Efficient hardware designs to implement neural networks including sparsity, locality, and systolic designs.
- Power and performance efficient memory architectures suited for neural networks.
- Network reduction techniques – approximation, quantization, reduced precision, pruning, distillation, and reconfiguration.
- Exploring interplay of precision, performance, power, and energy through benchmarks, workloads, and characterization.
- Performance potential, limit studies, bottleneck analysis, profiling, and synthesis of workloads.
- Explorations and architctures aimed to promote sustainable computing.
- Simulation and emulation techniques, frameworks, tools, and platforms for machine learning.
- Optimizations to improve performance of training techniques including on-device and large-scale learning.
- Load balancing and efficient task distribution, communication and computation overlapping for optimal performance.
- Verification, validation, determinism, robustness, bias, safety, and privacy challenges in AI systems.
- Efficient deployment strategies for edge and distributed environments.
- Model compression and optimization techniques that preserve reasoning and problem-solving capabilities.
- Architectures and frameworks for multi-agent systems and retrieval-augmented generation (RAG) pipelines.
- Systems-level approaches for scaling future foundation models (e.g., Llama 4, GPT-5 and beyond).
We will follow that same formatting guidelines and duplicate submission policies as ASPLOS.
08:00 - 08:15
Welcome
Welcome and Opening Remarks
Sushant Kondguli, Meta
08:15 - 09:00
Keynote
Future of Energy-Efficient Cognitive Computing: A Six-Word Story on Sentience, Systems, and Sustainability
Parthasarathy Ranganathan, Google link
We are at a pivotal inflection point in the design of computing systems. On one hand, demand for computing is accelerating at phenomenal rates, fueled by the AI revolution and increasingly deep processing on massive data volumes. On the other hand, Moore’s Law is slowing down. This widening supply-demand gap is forcing us to revisit traditional assumptions around systems design. In this talk, we will discuss Google’s experience designing and deploying large-scale AI systems optimized for efficiency, reliability, and velocity. Building on these lessons, we will identify key challenges and opportunities for future innovation, highlighting how the next generation of energy-efficient cognitive computing will be AI-driven, vertically integrated, and uncomfortably exciting!
09:00 - 09:30
Invited Talk
Architecting Responsible AI: Efficient and Carbon-Aware Caching for AI and Its Security Implications
Sihang Liu, University of Waterloo link
While generative AI offers transformative potential, its rapid scaling has introduced significant challenges in sustainability and responsible deployment. This talk explores the role of caching in AI systems, including both performance and its broader implications for sustainability and security. First, I will present our recent work on improving the efficiency of video generation systems through caching techniques. I will then discuss our context caching system that explores the trade-off between operational carbon savings and the increased embodied carbon associated with storage. Finally, I will highlight our recent findings on novel security vulnerabilities arising from caching mechanisms used in image generation services.
09:30 - 10:00
Invited Talk
Energy Consumption in AI Datacenters: Can We Address this Challenge?
Josep Torrellas, UIUC link
The global electricity demand from data centers, AI, and cryptocurrencies in expected to be around 800TW/h by the end of 2026. Currently, companies like Oracle are building 500MW AI campuses. Training a single model takes 100+ MW over many months, with clusters of about 100K GPUs, and the projected AI cluster size is expected to increase several times in the next few years. Clearly, power will be the primary limiter to the growth of AI. Given this scenario, what can we, researchers, do? In this talk, I will discuss the problem and suggest some techniques that we can use to try to mitigate it. Some of the ideas are being developed in the context of the ACE Center for Evolvable Computing, a center funded by SRC and DARPA for efficient distributed computing.
10:00 - 10:30
Break
Break
10:30 - 11:00
Invited Talk
Memory-Centric Computing: Solving Computing's Memory Problem
Onur Mutlu, ETH Zurich link
Computing is bottlenecked by data. Large amounts of application data overwhelm the storage capability, communication capability, and computation capability of the modern machines we design today. As a result, many key applications’ performance, efficiency, and scalability are bottlenecked by data movement. In this talk, we describe three major shortcomings of modern computers in terms of 1) dealing with data, 2) taking advantage of vast amounts of data, and 3) exploiting different semantic properties of application data. We argue that an intelligent computing architecture should be designed to handle data well. We posit that handling data well requires designing architectures based on three key principles: 1) data-centric, 2) data-driven, 3) data-aware. We give examples of how to exploit these principles to design a much more efficient and higher performance computing system. We especially discuss recent research that aims to fundamentally reduce memory latency and energy, and practically enable computation close to data, with at least two promising directions: 1) processing using memory, which exploits the fundamental operational properties of memory chips to perform massively-parallel computation in memory, with low-cost changes, 2) processing near memory, which integrates sophisticated additional processing capability in memory chips, the logic layer of 3D-stacked technologies, or memory controllers to enable near-memory computation with high memory bandwidth and low memory latency. We show both types of architectures can enable order(s) of magnitude improvements in performance and energy consumption of many important workloads, such as artificial intelligence, machine learning, graph analytics, database systems, video processing, climate modeling, genome analysis. We discuss how to enable adoption of such fundamentally more intelligent architectures, which are key to efficiency, performance, and sustainability. We conclude with some research opportunities in and guiding principles for future computing architecture and system designs.
An accompanying overview of modern memory-centric computing ideas & systems can be found at A Modern Primer on Processing in Memory (updated February 2025).
A shorter invited paper from IMW 2025 is at Memory-Centric Computing: Solving Computing’s Memory Problem (May 2025).
11:00 - 11:30
Invited Talk
Efficient and Scalable Agentic AI With Heterogeneous Systems
Zain Asgar, Gimlet AI link
AI agents are dynamic, multi-stage systems composed of diverse operations with very different resource characteristics, ranging from compute- and memory-intensive model execution to bandwidth- and IO-bound data processing. Yet most deployments today run on homogeneous infrastructure, despite inference itself being inherently heterogeneous across its phases. In this talk, we present a system for orchestrating agent workloads across a mix of CPUs, GPUs, and accelerators spanning vendors and hardware tiers. The system automatically decomposes agentic workloads into fine-grained execution graphs, compiles them into hardware-specific fragments, and dynamically places and stitches them across distributed infrastructure to meet latency and performance goals. By mapping each stage of execution to the hardware best suited for it, we show how we can achieve step-function performance gains and significant improvements in cost efficiency. We highlight results showing how emerging hardware like SRAM-centric chips, combined with existing GPU infrastructure, can outperform traditional homogeneous deployments.
11:30 - 12:00
Invited Talk
Private Neural Recommendation with Homomorphic Encryption and Orion
Brandon Reagen, NYU link
AI is great. It powers many of our favorite services and drives industries. However, today’s solutions pose a tradeoff between utility and privacy where receiving the highest quality service often requires disclosing private information. In this talk I will show how things don’t have to be this way. Fully Homomorphic Encryption (FHE) is a cryptographic method that enables computation directly on encrypted data, never disclosing sensitive inputs while still enabling access to high-quality services. I will then cover the challenges of using this technology, which include both extreme performance overhead and programming difficulty, and how our Orion framework addresses them. Finally, I will highlight our most recent work that demonstrates how FHE can be applied to neural recommendation. Time permitting, a demonstration will be given.
12:00 - 13:15
Lunch Break
Lunch Break
13:15 - 13:30
Invited Talk
Sponsor Talk
Max Sbabo, d-Matrix
13:30 - 14:15
Paper Session #1
piPE-SA: Enabling Deeply Pipelined Processing Elements in Systolic Arrays
Jiayi Wang, Chenyi Wang and Ang Li
CacheFlex: Explicitly Control What You Need in Your Cache
Jingqun Zhang, Weihang Li, Maohua Nie, Yung-Jen Cheng, Jiayi Wang, Shwet Chitnis and Ang Li
IEI: A Composite Infrastructure Efficiency Index for GPU Inference
Raj Parihar
14:15 - 14:45
Invited Talk
Making the Accuracy-Efficiency Trade-Off in Agentic Systems
Esha Choukse, Microsoft link
For decades, systems researchers have balanced competing objectives: latency versus throughput, performance versus power, and speed versus cost. Traditionally, accuracy was a fixed requirement—non-negotiable and external to the systems equation. But as modern computing increasingly hosts AI-driven workloads, accuracy itself has become a tunable system variable. In this talk, I’ll argue that the next frontier in systems design lies in treating accuracy as a first-class performance knob, to be traded for efficiency in a principled way. Drawing from our recent work, I will show how software–hardware co-design can explicitly expose, quantify, and manage the accuracy-efficiency tradeoff.
14:45 - 15:30
Paper Session #2
Fast NF4 Dequantization Kernels for Large Language Model Inference
Xiangbo Qi, Chaoyi Jiang and Murali Annavar
I-Fuse: Profile-Guided Speculative Load Micro-op Fusion for Data Center Applications
Deepanjali Mishra, Tanvir Ahmed Khan, Gilles Pokam, Heiner Litz and Akshitha Sriraman
Store or Recompute? Characterizing the Carbon Tradeoff of KV Cache Retention in LLM Inference
Abnash Bassi, Jaylen Wang, Fiodar Kazhamiaka, Daniel S. Berger and Akshitha Sriraman
15:30 - 16:00
Break
Break
16:00 - 17:00
Panel
Architecting an energy-first stack for the AI age
As the cumulative demand for AI computing continues to grow, new bottlenecks in technology deployment are emerging. These range from compute capacity to memory supply to cluster reliability to simply finding enough power to supply the infrastructure. Since these aren’t easily solved with brute force addition of more chips, potential solutions encompass new model architectures, new computing architectures, high performance cooling designs, robust multi-device software stacks, and even orbital datacenters. In this panel, we will aim to identify the most pressing requirements for the computing stack of coming years. We will discuss both the promising technologies and the hypes to define a blueprint of the computing architecture where energy is the currency of success.
Moderator: Satyam Srivastava, d-Matrix
Panelists:
- Nishil Talati, UIUC
- Esha Choukse, Microsoft
- Josep Torrellas, UIUC
- Zain Asgar, Gimlet AI
- Gennady Pekhimenko, University of Toronto
17:00 - 17:30
Invited Talk
Agile and evolvable software construction in the era of rapidly evolving hardware accelerator designs
Charith Mendis, UIUC link
Modern AI workloads have become exceedingly abundant and important in the current computing landscape. As a result, there have been numerous software and hardware innovations aimed at accelerating these workloads. However, we observe a subtle disconnect between the software and hardware communities. Most software innovations target well-established hardware platforms such as CPUs (e.g., x86, ARM) and GPUs (e.g., NVidia GPUs), while hardware innovations produce plenty of other tensor accelerator designs (e.g., Gemmini, Feather, Trainium) each year.
We asked the question, why aren’t the software community using these accelerators or even evaluating on them? The simple yet undeniable reason is the lack of standardized software tooling compared to CPUs and GPUs. For an architecture to be used, properly designed compiler backends, correctness, and performance testing tools should be abundant (e.g., CUDA ecosystem).
In this talk, I will describe how we bridge this gap by automatically generating the necessary software tools for a large class of accelerators through the Accelerator Compiler Toolkit (ACT) ecosystem. Central to ACT is an ISA definition language, TAIDL, that for the first time standardizes the hardware-software interfaces for a large class of accelerators. Departing from the traditional approach of manually constructing test oracles, performance models, or retargetable compiler backends, we instead introduce agile and evolvable methodologies to automatically generate such necessary tooling using both formal methods and machine learning techniques for any TAIDL-defined accelerator interface. I will show how such automation enables rapid software prototyping, making rapidly evolving accelerator designs usable by the software community.
17:30 - 18:00
Invited Talk
Making Long-Context LLM Inference Practical: Leveraging Token Importance for Efficient Inference
Urmish Thakker, Ex-SambaNova Systems link
As LLM context windows scale to hundreds of thousands or even millions of tokens, inference efficiency is increasingly limited by KV-cache memory growth and the high cost of prompt prefilling. Prior approaches attempt to mitigate these costs through heuristic cache eviction or prompt compression techniques, but often rely on architecture-specific assumptions or require retraining to recover accuracy. In this talk, I show that LLMs implicitly encode token importance signals that reveal which parts of context matter most for downstream computation. By exploiting these signals, it is possible to selectively retain or process only the most relevant tokens, significantly improving both decoding throughput and prefill throughput while preserving model quality in long-context inference.