Picture for Yuanxing Zhang

Yuanxing Zhang

MIO: A Foundation Model on Multimodal Tokens

Add code
Sep 26, 2024
Figure 1 for MIO: A Foundation Model on Multimodal Tokens
Figure 2 for MIO: A Foundation Model on Multimodal Tokens
Figure 3 for MIO: A Foundation Model on Multimodal Tokens
Figure 4 for MIO: A Foundation Model on Multimodal Tokens
Viaarxiv icon

DDK: Distilling Domain Knowledge for Efficient Large Language Models

Add code
Jul 23, 2024
Viaarxiv icon

R2C2-Coder: Enhancing and Benchmarking Real-world Repository-level Code Completion Abilities of Code Large Language Models

Add code
Jun 04, 2024
Viaarxiv icon

D-CPT Law: Domain-specific Continual Pre-Training Scaling Law for Large Language Models

Add code
Jun 03, 2024
Viaarxiv icon

ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models

Add code
Feb 23, 2024
Figure 1 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Figure 2 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Figure 3 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Figure 4 for ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models
Viaarxiv icon

Adaptive Modulus RF Beamforming for Enhanced Self-Interference Suppression in Full-Duplex Massive MIMO Systems

Add code
Feb 13, 2024
Viaarxiv icon

E^2-LLM: Efficient and Extreme Length Extension of Large Language Models

Add code
Jan 18, 2024
Figure 1 for E^2-LLM: Efficient and Extreme Length Extension of Large Language Models
Figure 2 for E^2-LLM: Efficient and Extreme Length Extension of Large Language Models
Figure 3 for E^2-LLM: Efficient and Extreme Length Extension of Large Language Models
Figure 4 for E^2-LLM: Efficient and Extreme Length Extension of Large Language Models
Viaarxiv icon

GBA: A Tuning-free Approach to Switch between Synchronous and Asynchronous Training for Recommendation Model

Add code
May 23, 2022
Figure 1 for GBA: A Tuning-free Approach to Switch between Synchronous and Asynchronous Training for Recommendation Model
Figure 2 for GBA: A Tuning-free Approach to Switch between Synchronous and Asynchronous Training for Recommendation Model
Figure 3 for GBA: A Tuning-free Approach to Switch between Synchronous and Asynchronous Training for Recommendation Model
Figure 4 for GBA: A Tuning-free Approach to Switch between Synchronous and Asynchronous Training for Recommendation Model
Viaarxiv icon

PICASSO: Unleashing the Potential of GPU-centric Training for Wide-and-deep Recommender Systems

Add code
Apr 17, 2022
Figure 1 for PICASSO: Unleashing the Potential of GPU-centric Training for Wide-and-deep Recommender Systems
Figure 2 for PICASSO: Unleashing the Potential of GPU-centric Training for Wide-and-deep Recommender Systems
Figure 3 for PICASSO: Unleashing the Potential of GPU-centric Training for Wide-and-deep Recommender Systems
Figure 4 for PICASSO: Unleashing the Potential of GPU-centric Training for Wide-and-deep Recommender Systems
Viaarxiv icon

TSSRGCN: Temporal Spectral Spatial Retrieval Graph Convolutional Network for Traffic Flow Forecasting

Add code
Nov 30, 2020
Figure 1 for TSSRGCN: Temporal Spectral Spatial Retrieval Graph Convolutional Network for Traffic Flow Forecasting
Figure 2 for TSSRGCN: Temporal Spectral Spatial Retrieval Graph Convolutional Network for Traffic Flow Forecasting
Figure 3 for TSSRGCN: Temporal Spectral Spatial Retrieval Graph Convolutional Network for Traffic Flow Forecasting
Viaarxiv icon