Picture for Ye Lin

Ye Lin

Green Multigrid Network

Add code
Jul 04, 2024
Viaarxiv icon

Understanding Parameter Sharing in Transformers

Add code
Jun 15, 2023
Figure 1 for Understanding Parameter Sharing in Transformers
Figure 2 for Understanding Parameter Sharing in Transformers
Figure 3 for Understanding Parameter Sharing in Transformers
Figure 4 for Understanding Parameter Sharing in Transformers
Viaarxiv icon

MobileNMT: Enabling Translation in 15MB and 30ms

Add code
Jun 07, 2023
Figure 1 for MobileNMT: Enabling Translation in 15MB and 30ms
Figure 2 for MobileNMT: Enabling Translation in 15MB and 30ms
Figure 3 for MobileNMT: Enabling Translation in 15MB and 30ms
Figure 4 for MobileNMT: Enabling Translation in 15MB and 30ms
Viaarxiv icon

Multi-Path Transformer is Better: A Case Study on Neural Machine Translation

Add code
May 10, 2023
Figure 1 for Multi-Path Transformer is Better: A Case Study on Neural Machine Translation
Figure 2 for Multi-Path Transformer is Better: A Case Study on Neural Machine Translation
Figure 3 for Multi-Path Transformer is Better: A Case Study on Neural Machine Translation
Figure 4 for Multi-Path Transformer is Better: A Case Study on Neural Machine Translation
Viaarxiv icon

The NiuTrans System for WNGT 2020 Efficiency Task

Add code
Sep 16, 2021
Figure 1 for The NiuTrans System for WNGT 2020 Efficiency Task
Figure 2 for The NiuTrans System for WNGT 2020 Efficiency Task
Figure 3 for The NiuTrans System for WNGT 2020 Efficiency Task
Figure 4 for The NiuTrans System for WNGT 2020 Efficiency Task
Viaarxiv icon

The NiuTrans System for the WMT21 Efficiency Task

Add code
Sep 16, 2021
Figure 1 for The NiuTrans System for the WMT21 Efficiency Task
Figure 2 for The NiuTrans System for the WMT21 Efficiency Task
Figure 3 for The NiuTrans System for the WMT21 Efficiency Task
Figure 4 for The NiuTrans System for the WMT21 Efficiency Task
Viaarxiv icon

Bag of Tricks for Optimizing Transformer Efficiency

Add code
Sep 09, 2021
Figure 1 for Bag of Tricks for Optimizing Transformer Efficiency
Figure 2 for Bag of Tricks for Optimizing Transformer Efficiency
Figure 3 for Bag of Tricks for Optimizing Transformer Efficiency
Figure 4 for Bag of Tricks for Optimizing Transformer Efficiency
Viaarxiv icon

An Efficient Transformer Decoder with Compressed Sub-layers

Add code
Jan 03, 2021
Figure 1 for An Efficient Transformer Decoder with Compressed Sub-layers
Figure 2 for An Efficient Transformer Decoder with Compressed Sub-layers
Figure 3 for An Efficient Transformer Decoder with Compressed Sub-layers
Figure 4 for An Efficient Transformer Decoder with Compressed Sub-layers
Viaarxiv icon

A Simple and Effective Approach to Robust Unsupervised Bilingual Dictionary Induction

Add code
Nov 30, 2020
Figure 1 for A Simple and Effective Approach to Robust Unsupervised Bilingual Dictionary Induction
Figure 2 for A Simple and Effective Approach to Robust Unsupervised Bilingual Dictionary Induction
Figure 3 for A Simple and Effective Approach to Robust Unsupervised Bilingual Dictionary Induction
Figure 4 for A Simple and Effective Approach to Robust Unsupervised Bilingual Dictionary Induction
Viaarxiv icon

Weight Distillation: Transferring the Knowledge in Neural Network Parameters

Add code
Sep 19, 2020
Figure 1 for Weight Distillation: Transferring the Knowledge in Neural Network Parameters
Figure 2 for Weight Distillation: Transferring the Knowledge in Neural Network Parameters
Figure 3 for Weight Distillation: Transferring the Knowledge in Neural Network Parameters
Figure 4 for Weight Distillation: Transferring the Knowledge in Neural Network Parameters
Viaarxiv icon