Picture for Athanasios Mouchtaris

Athanasios Mouchtaris

Alexa Machine Learning, Amazon, USA

Wanda++: Pruning Large Language Models via Regional Gradients

Add code
Mar 06, 2025
Viaarxiv icon

QuZO: Quantized Zeroth-Order Fine-Tuning for Large Language Models

Add code
Feb 17, 2025
Viaarxiv icon

MaZO: Masked Zeroth-Order Optimization for Multi-Task Fine-Tuning of Large Language Models

Add code
Feb 17, 2025
Viaarxiv icon

AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning

Add code
Jun 26, 2024
Figure 1 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Figure 2 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Figure 3 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Figure 4 for AdaZeta: Adaptive Zeroth-Order Tensor-Train Adaption for Memory-Efficient Large Language Models Fine-Tuning
Viaarxiv icon

Accelerator-Aware Training for Transducer-Based Speech Recognition

Add code
May 12, 2023
Figure 1 for Accelerator-Aware Training for Transducer-Based Speech Recognition
Figure 2 for Accelerator-Aware Training for Transducer-Based Speech Recognition
Figure 3 for Accelerator-Aware Training for Transducer-Based Speech Recognition
Figure 4 for Accelerator-Aware Training for Transducer-Based Speech Recognition
Viaarxiv icon

Robust Acoustic and Semantic Contextual Biasing in Neural Transducers for Speech Recognition

Add code
May 09, 2023
Viaarxiv icon

Lookahead When It Matters: Adaptive Non-causal Transformers for Streaming Neural Transducers

Add code
May 09, 2023
Viaarxiv icon

Dual-Attention Neural Transducers for Efficient Wake Word Spotting in Speech Recognition

Add code
Apr 05, 2023
Figure 1 for Dual-Attention Neural Transducers for Efficient Wake Word Spotting in Speech Recognition
Figure 2 for Dual-Attention Neural Transducers for Efficient Wake Word Spotting in Speech Recognition
Figure 3 for Dual-Attention Neural Transducers for Efficient Wake Word Spotting in Speech Recognition
Figure 4 for Dual-Attention Neural Transducers for Efficient Wake Word Spotting in Speech Recognition
Viaarxiv icon

Leveraging Redundancy in Multiple Audio Signals for Far-Field Speech Recognition

Add code
Mar 01, 2023
Viaarxiv icon

Sub-8-bit quantization for on-device speech recognition: a regularization-free approach

Add code
Oct 17, 2022
Figure 1 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 2 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 3 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 4 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Viaarxiv icon