Picture for Shivaram Venkataraman

Shivaram Venkataraman

AI Benchmark Democratization and Carpentry

Add code
Dec 12, 2025
Viaarxiv icon

An MLCommons Scientific Benchmarks Ontology

Add code
Nov 06, 2025
Viaarxiv icon

Tesserae: Scalable Placement Policies for Deep Learning Workloads

Add code
Aug 07, 2025
Viaarxiv icon

PGT-I: Scaling Spatiotemporal GNNs with Memory-Efficient Distributed Training

Add code
Jul 15, 2025
Viaarxiv icon

Armada: Memory-Efficient Distributed Training of Large-Scale Graph Neural Networks

Add code
Feb 25, 2025
Viaarxiv icon

LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models

Add code
Feb 04, 2025
Figure 1 for LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models
Figure 2 for LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models
Figure 3 for LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models
Figure 4 for LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models
Viaarxiv icon

Scaling Inference-Efficient Language Models

Add code
Jan 30, 2025
Figure 1 for Scaling Inference-Efficient Language Models
Figure 2 for Scaling Inference-Efficient Language Models
Figure 3 for Scaling Inference-Efficient Language Models
Figure 4 for Scaling Inference-Efficient Language Models
Viaarxiv icon

Incremental IVF Index Maintenance for Streaming Vector Search

Add code
Nov 01, 2024
Figure 1 for Incremental IVF Index Maintenance for Streaming Vector Search
Figure 2 for Incremental IVF Index Maintenance for Streaming Vector Search
Figure 3 for Incremental IVF Index Maintenance for Streaming Vector Search
Figure 4 for Incremental IVF Index Maintenance for Streaming Vector Search
Viaarxiv icon

GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval

Add code
Jun 25, 2024
Figure 1 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Figure 2 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Figure 3 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Figure 4 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Viaarxiv icon

CHAI: Clustered Head Attention for Efficient LLM Inference

Add code
Mar 12, 2024
Figure 1 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 2 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 3 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 4 for CHAI: Clustered Head Attention for Efficient LLM Inference
Viaarxiv icon