Picture for Goro Kobayashi

Goro Kobayashi

Contrastive Learning-based Sentence Encoders Implicitly Weight Informative Words

Add code
Oct 24, 2023
Viaarxiv icon

Assessing Step-by-Step Reasoning against Lexical Negation: A Case Study on Syllogism

Add code
Oct 23, 2023
Viaarxiv icon

Transformer Language Models Handle Word Frequency in Prediction Head

Add code
May 29, 2023
Viaarxiv icon

Feed-Forward Blocks Control Contextualization in Masked Language Models

Add code
Feb 01, 2023
Viaarxiv icon

Incorporating Residual and Normalization Layers into Analysis of Masked Language Models

Add code
Sep 15, 2021
Figure 1 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Figure 2 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Figure 3 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Figure 4 for Incorporating Residual and Normalization Layers into Analysis of Masked Language Models
Viaarxiv icon

Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms

Add code
Apr 21, 2020
Figure 1 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Figure 2 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Figure 3 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Figure 4 for Attention Module is Not Only a Weight: Analyzing Transformers with Vector Norms
Viaarxiv icon