Picture for Ali Modarressi

Ali Modarressi

NoLiMa: Long-Context Evaluation Beyond Literal Matching

Add code
Feb 07, 2025
Viaarxiv icon

MEXA: Multilingual Evaluation of English-Centric LLMs via Cross-Lingual Alignment

Add code
Oct 08, 2024
Viaarxiv icon

Consistent Document-Level Relation Extraction via Counterfactuals

Add code
Jul 09, 2024
Viaarxiv icon

MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory

Add code
Apr 17, 2024
Figure 1 for MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory
Figure 2 for MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory
Figure 3 for MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory
Figure 4 for MemLLM: Finetuning LLMs to Use An Explicit Read-Write Memory
Viaarxiv icon

DecompX: Explaining Transformers Decisions by Propagating Token Decomposition

Add code
Jun 05, 2023
Figure 1 for DecompX: Explaining Transformers Decisions by Propagating Token Decomposition
Figure 2 for DecompX: Explaining Transformers Decisions by Propagating Token Decomposition
Figure 3 for DecompX: Explaining Transformers Decisions by Propagating Token Decomposition
Figure 4 for DecompX: Explaining Transformers Decisions by Propagating Token Decomposition
Viaarxiv icon

RET-LLM: Towards a General Read-Write Memory for Large Language Models

Add code
May 23, 2023
Viaarxiv icon

Guide the Learner: Controlling Product of Experts Debiasing Method Based on Token Attribution Similarities

Add code
Feb 06, 2023
Viaarxiv icon

BERT on a Data Diet: Finding Important Examples by Gradient-Based Pruning

Add code
Nov 10, 2022
Viaarxiv icon

GlobEnc: Quantifying Global Token Attribution by Incorporating the Whole Encoder Layer in Transformers

Add code
May 06, 2022
Figure 1 for GlobEnc: Quantifying Global Token Attribution by Incorporating the Whole Encoder Layer in Transformers
Figure 2 for GlobEnc: Quantifying Global Token Attribution by Incorporating the Whole Encoder Layer in Transformers
Figure 3 for GlobEnc: Quantifying Global Token Attribution by Incorporating the Whole Encoder Layer in Transformers
Figure 4 for GlobEnc: Quantifying Global Token Attribution by Incorporating the Whole Encoder Layer in Transformers
Viaarxiv icon

AdapLeR: Speeding up Inference by Adaptive Length Reduction

Add code
Mar 16, 2022
Figure 1 for AdapLeR: Speeding up Inference by Adaptive Length Reduction
Figure 2 for AdapLeR: Speeding up Inference by Adaptive Length Reduction
Figure 3 for AdapLeR: Speeding up Inference by Adaptive Length Reduction
Figure 4 for AdapLeR: Speeding up Inference by Adaptive Length Reduction
Viaarxiv icon