Picture for Yiren Zhao

Yiren Zhao

Hardware and Software Platform Inference

Add code
Nov 07, 2024
Viaarxiv icon

Absorb & Escape: Overcoming Single Model Limitations in Generating Genomic Sequences

Add code
Oct 28, 2024
Viaarxiv icon

Scaling Laws for Mixed quantization in Large Language Models

Add code
Oct 09, 2024
Figure 1 for Scaling Laws for Mixed quantization in Large Language Models
Figure 2 for Scaling Laws for Mixed quantization in Large Language Models
Figure 3 for Scaling Laws for Mixed quantization in Large Language Models
Figure 4 for Scaling Laws for Mixed quantization in Large Language Models
Viaarxiv icon

QERA: an Analytical Framework for Quantization Error Reconstruction

Add code
Oct 08, 2024
Figure 1 for QERA: an Analytical Framework for Quantization Error Reconstruction
Figure 2 for QERA: an Analytical Framework for Quantization Error Reconstruction
Figure 3 for QERA: an Analytical Framework for Quantization Error Reconstruction
Figure 4 for QERA: an Analytical Framework for Quantization Error Reconstruction
Viaarxiv icon

GV-Rep: A Large-Scale Dataset for Genetic Variant Representation Learning

Add code
Jul 24, 2024
Viaarxiv icon

Unlocking the Global Synergies in Low-Rank Adapters

Add code
Jun 21, 2024
Viaarxiv icon

Optimised Grouped-Query Attention Mechanism for Transformers

Add code
Jun 21, 2024
Viaarxiv icon

HASS: Hardware-Aware Sparsity Search for Dataflow DNN Accelerator

Add code
Jun 05, 2024
Figure 1 for HASS: Hardware-Aware Sparsity Search for Dataflow DNN Accelerator
Figure 2 for HASS: Hardware-Aware Sparsity Search for Dataflow DNN Accelerator
Figure 3 for HASS: Hardware-Aware Sparsity Search for Dataflow DNN Accelerator
Figure 4 for HASS: Hardware-Aware Sparsity Search for Dataflow DNN Accelerator
Viaarxiv icon

$Δ$-DiT: A Training-Free Acceleration Method Tailored for Diffusion Transformers

Add code
Jun 03, 2024
Figure 1 for $Δ$-DiT: A Training-Free Acceleration Method Tailored for Diffusion Transformers
Figure 2 for $Δ$-DiT: A Training-Free Acceleration Method Tailored for Diffusion Transformers
Figure 3 for $Δ$-DiT: A Training-Free Acceleration Method Tailored for Diffusion Transformers
Figure 4 for $Δ$-DiT: A Training-Free Acceleration Method Tailored for Diffusion Transformers
Viaarxiv icon

Locking Machine Learning Models into Hardware

Add code
May 31, 2024
Viaarxiv icon