Picture for Yaman Umuroglu

Yaman Umuroglu

A2Q+: Improving Accumulator-Aware Weight Quantization

Add code
Jan 19, 2024
Viaarxiv icon

Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark

Add code
Jun 23, 2022
Figure 1 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Figure 2 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Figure 3 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Figure 4 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Viaarxiv icon

QONNX: Representing Arbitrary-Precision Quantized Neural Networks

Add code
Jun 17, 2022
Figure 1 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Figure 2 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Figure 3 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Figure 4 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Viaarxiv icon

EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators

Add code
Feb 04, 2022
Figure 1 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Figure 2 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Figure 3 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Figure 4 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Viaarxiv icon

Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference

Add code
Feb 22, 2021
Figure 1 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Figure 2 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Figure 3 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Figure 4 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Viaarxiv icon

LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications

Add code
Apr 06, 2020
Figure 1 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Figure 2 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Figure 3 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Figure 4 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Viaarxiv icon

Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic

Add code
Jun 26, 2018
Figure 1 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Figure 2 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Figure 3 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Figure 4 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Viaarxiv icon

Streamlined Deployment for Quantized Neural Networks

Add code
May 30, 2018
Figure 1 for Streamlined Deployment for Quantized Neural Networks
Figure 2 for Streamlined Deployment for Quantized Neural Networks
Figure 3 for Streamlined Deployment for Quantized Neural Networks
Figure 4 for Streamlined Deployment for Quantized Neural Networks
Viaarxiv icon

Scaling Binarized Neural Networks on Reconfigurable Logic

Add code
Jan 27, 2017
Figure 1 for Scaling Binarized Neural Networks on Reconfigurable Logic
Figure 2 for Scaling Binarized Neural Networks on Reconfigurable Logic
Figure 3 for Scaling Binarized Neural Networks on Reconfigurable Logic
Figure 4 for Scaling Binarized Neural Networks on Reconfigurable Logic
Viaarxiv icon

FINN: A Framework for Fast, Scalable Binarized Neural Network Inference

Add code
Dec 01, 2016
Figure 1 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Figure 2 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Figure 3 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Figure 4 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Viaarxiv icon