Picture for Mahyar Najibi

Mahyar Najibi

From Dense to Dynamic: Token-Difficulty Driven MoEfication of Pre-Trained LLMs

Add code
Feb 17, 2025
Viaarxiv icon

M2R2: Mixture of Multi-Rate Residuals for Efficient Transformer Inference

Add code
Feb 04, 2025
Viaarxiv icon

Efficient Vision-Language Models by Summarizing Visual Tokens into Compact Registers

Add code
Oct 17, 2024
Viaarxiv icon

CtrlSynth: Controllable Image Text Synthesis for Data-Efficient Multimodal Learning

Add code
Oct 15, 2024
Viaarxiv icon

SeedLM: Compressing LLM Weights into Seeds of Pseudo-Random Generators

Add code
Oct 14, 2024
Figure 1 for SeedLM: Compressing LLM Weights into Seeds of Pseudo-Random Generators
Figure 2 for SeedLM: Compressing LLM Weights into Seeds of Pseudo-Random Generators
Figure 3 for SeedLM: Compressing LLM Weights into Seeds of Pseudo-Random Generators
Figure 4 for SeedLM: Compressing LLM Weights into Seeds of Pseudo-Random Generators
Viaarxiv icon

Apple Intelligence Foundation Language Models

Add code
Jul 29, 2024
Figure 1 for Apple Intelligence Foundation Language Models
Figure 2 for Apple Intelligence Foundation Language Models
Figure 3 for Apple Intelligence Foundation Language Models
Figure 4 for Apple Intelligence Foundation Language Models
Viaarxiv icon

LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inference

Add code
Jul 19, 2024
Figure 1 for LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inference
Figure 2 for LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inference
Figure 3 for LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inference
Figure 4 for LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inference
Viaarxiv icon

OpenELM: An Efficient Language Model Family with Open Training and Inference Framework

Add code
May 02, 2024
Viaarxiv icon

CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data

Add code
Apr 24, 2024
Figure 1 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Figure 2 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Figure 3 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Figure 4 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Viaarxiv icon

Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation

Add code
Apr 10, 2024
Figure 1 for Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation
Figure 2 for Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation
Figure 3 for Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation
Figure 4 for Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation
Viaarxiv icon