Picture for Armen Aghajanyan

Armen Aghajanyan

MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts

Add code
Jul 31, 2024
Viaarxiv icon

Small Molecule Optimization with Large Language Models

Add code
Jul 26, 2024
Viaarxiv icon

Text Quality-Based Pruning for Efficient Training of Language Models

Add code
Apr 26, 2024
Viaarxiv icon

DOMINO: A Dual-System for Multi-step Visual Language Reasoning

Add code
Oct 04, 2023
Viaarxiv icon

Jointly Training Large Autoregressive Multimodal Models

Add code
Sep 28, 2023
Viaarxiv icon

Scaling Autoregressive Multi-Modal Models: Pretraining and Instruction Tuning

Add code
Sep 05, 2023
Viaarxiv icon

D4: Improving LLM Pretraining via Document De-Duplication and Diversification

Add code
Aug 23, 2023
Figure 1 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 2 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 3 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Figure 4 for D4: Improving LLM Pretraining via Document De-Duplication and Diversification
Viaarxiv icon

MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers

Add code
May 19, 2023
Viaarxiv icon

Scaling Laws for Generative Mixed-Modal Language Models

Add code
Jan 10, 2023
Viaarxiv icon

BARTSmiles: Generative Masked Language Models for Molecular Representations

Add code
Nov 29, 2022
Viaarxiv icon