Help us make Indico better by taking this survey! Aidez-nous à améliorer Indico en répondant à ce sondage !

15–18 Oct 2024
Purdue University
America/Indiana/Indianapolis timezone

EnsembleLUT: Scaling up LUT-based Neural Networks with Ensemble Learning

16 Oct 2024, 15:30
5m
Steward Center 306 (Third floor) (Purdue University)

Steward Center 306 (Third floor)

Purdue University

128 Memorial Mall Dr, West Lafayette, IN 47907
Lightning 5 min talk + poster Lighting talks

Speaker

Olivia Weng

Description

Applications like high-energy physics and cybersecurity require extremely high throughput and low latency neural network (NN) inference. Lookup-table-based NNs address these constraints by implementing NNs purely as lookup tables (LUTs), achieving inference latency on the order of nanoseconds. Since LUTs are a fundamental FPGA building block, LUT-based NNs map to FPGAs easily. LogicNets (and its successors) form one such class of LUT-based NNs that target FPGAs, mapping neurons directly to LUTs to meet the low latency constraints with minimal resources. However, it is difficult to implement larger, more performant LUT-based NNs like LogicNets because LUT usage increases exponentially with respect to neuron fan-in (i.e., number of synapses $\times$ synapse bitwidth). A large LUT-based NN quickly runs out of LUTs on an FPGA, which is unideal. Our work EnsembleLUT addresses this issue by creating ensembles of smaller LUT-based NNs that scale linearly with respect to the number of models, achieving higher accuracy within the resource constraints of an FPGA. We demonstrate that EnsembleLUT improves the scalability of LUT-based NNs on various scientific machine learning benchmarks such as jet substructure classification and high-granularity endcap calorimeter data compression found at the LHC CMS experiment, reaching higher accuracy with fewer resources than the largest LogicNets.

Primary authors

Olivia Weng Marta Andronic (Imperial College London) Danial Zuberi (UC San Diego) Jiaqing Chen (Arizona State University) Caleb Geniesse (Lawrence Berkeley National Laboratory) George Constantinides (Imperial College London) Javier Duarte (UCSD) Nhan Tran (Fermi National Accelerator Lab. (US)) Nicholas Fraser (AMD) Ryan Kastner

Presentation materials