Picture for Sheng Lin

Sheng Lin

Towards Zero Memory Footprint Spiking Neural Network Training

Add code
Aug 16, 2023
Viaarxiv icon

FAIVConf: Face enhancement for AI-based Video Conference with Low Bit-rate

Add code
Jul 08, 2022
Figure 1 for FAIVConf: Face enhancement for AI-based Video Conference with Low Bit-rate
Figure 2 for FAIVConf: Face enhancement for AI-based Video Conference with Low Bit-rate
Figure 3 for FAIVConf: Face enhancement for AI-based Video Conference with Low Bit-rate
Figure 4 for FAIVConf: Face enhancement for AI-based Video Conference with Low Bit-rate
Viaarxiv icon

A Secure and Efficient Federated Learning Framework for NLP

Add code
Jan 28, 2022
Figure 1 for A Secure and Efficient Federated Learning Framework for NLP
Figure 2 for A Secure and Efficient Federated Learning Framework for NLP
Figure 3 for A Secure and Efficient Federated Learning Framework for NLP
Figure 4 for A Secure and Efficient Federated Learning Framework for NLP
Viaarxiv icon

CAP-RAM: A Charge-Domain In-Memory Computing 6T-SRAM for Accurate and Precision-Programmable CNN Inference

Add code
Jul 06, 2021
Figure 1 for CAP-RAM: A Charge-Domain In-Memory Computing 6T-SRAM for Accurate and Precision-Programmable CNN Inference
Figure 2 for CAP-RAM: A Charge-Domain In-Memory Computing 6T-SRAM for Accurate and Precision-Programmable CNN Inference
Figure 3 for CAP-RAM: A Charge-Domain In-Memory Computing 6T-SRAM for Accurate and Precision-Programmable CNN Inference
Figure 4 for CAP-RAM: A Charge-Domain In-Memory Computing 6T-SRAM for Accurate and Precision-Programmable CNN Inference
Viaarxiv icon

FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator

Add code
Jun 16, 2021
Figure 1 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Figure 2 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Figure 3 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Figure 4 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Viaarxiv icon

Efficient Micro-Structured Weight Unification and Pruning for Neural Network Compression

Add code
Jun 16, 2021
Figure 1 for Efficient Micro-Structured Weight Unification and Pruning for Neural Network Compression
Figure 2 for Efficient Micro-Structured Weight Unification and Pruning for Neural Network Compression
Figure 3 for Efficient Micro-Structured Weight Unification and Pruning for Neural Network Compression
Figure 4 for Efficient Micro-Structured Weight Unification and Pruning for Neural Network Compression
Viaarxiv icon

ESMFL: Efficient and Secure Models for Federated Learning

Add code
Sep 03, 2020
Figure 1 for ESMFL: Efficient and Secure Models for Federated Learning
Figure 2 for ESMFL: Efficient and Secure Models for Federated Learning
Figure 3 for ESMFL: Efficient and Secure Models for Federated Learning
Figure 4 for ESMFL: Efficient and Secure Models for Federated Learning
Viaarxiv icon

An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices

Add code
Feb 22, 2020
Figure 1 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Figure 2 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Figure 3 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Figure 4 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Viaarxiv icon

RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition

Add code
Feb 19, 2020
Figure 1 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Figure 2 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Figure 3 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Figure 4 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Viaarxiv icon

PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning

Add code
Jan 22, 2020
Figure 1 for PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning
Figure 2 for PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning
Figure 3 for PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning
Figure 4 for PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning
Viaarxiv icon