Picture for Mohamed S. Abdelfattah

Mohamed S. Abdelfattah

NITRO: LLM Inference on Intel Laptop NPUs

Add code
Dec 15, 2024
Viaarxiv icon

Attamba: Attending To Multi-Token States

Add code
Nov 26, 2024
Viaarxiv icon

BitMoD: Bit-serial Mixture-of-Datatype LLM Acceleration

Add code
Nov 18, 2024
Figure 1 for BitMoD: Bit-serial Mixture-of-Datatype LLM Acceleration
Figure 2 for BitMoD: Bit-serial Mixture-of-Datatype LLM Acceleration
Figure 3 for BitMoD: Bit-serial Mixture-of-Datatype LLM Acceleration
Figure 4 for BitMoD: Bit-serial Mixture-of-Datatype LLM Acceleration
Viaarxiv icon

BBS: Bi-directional Bit-level Sparsity for Deep Learning Acceleration

Add code
Sep 08, 2024
Figure 1 for BBS: Bi-directional Bit-level Sparsity for Deep Learning Acceleration
Figure 2 for BBS: Bi-directional Bit-level Sparsity for Deep Learning Acceleration
Figure 3 for BBS: Bi-directional Bit-level Sparsity for Deep Learning Acceleration
Figure 4 for BBS: Bi-directional Bit-level Sparsity for Deep Learning Acceleration
Viaarxiv icon

Learning from Students: Applying t-Distributions to Explore Accurate and Efficient Formats for LLMs

Add code
May 06, 2024
Viaarxiv icon

Encodings for Prediction-based Neural Architecture Search

Add code
Mar 04, 2024
Viaarxiv icon

On Latency Predictors for Neural Architecture Search

Add code
Mar 04, 2024
Figure 1 for On Latency Predictors for Neural Architecture Search
Figure 2 for On Latency Predictors for Neural Architecture Search
Figure 3 for On Latency Predictors for Neural Architecture Search
Figure 4 for On Latency Predictors for Neural Architecture Search
Viaarxiv icon

Fast Inference Through The Reuse Of Attention Maps In Diffusion Models

Add code
Dec 13, 2023
Viaarxiv icon

FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search

Add code
Aug 07, 2023
Figure 1 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Figure 2 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Figure 3 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Figure 4 for FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search
Viaarxiv icon

DiviML: A Module-based Heuristic for Mapping Neural Networks onto Heterogeneous Platforms

Add code
Aug 02, 2023
Figure 1 for DiviML: A Module-based Heuristic for Mapping Neural Networks onto Heterogeneous Platforms
Figure 2 for DiviML: A Module-based Heuristic for Mapping Neural Networks onto Heterogeneous Platforms
Figure 3 for DiviML: A Module-based Heuristic for Mapping Neural Networks onto Heterogeneous Platforms
Figure 4 for DiviML: A Module-based Heuristic for Mapping Neural Networks onto Heterogeneous Platforms
Viaarxiv icon