This episode explores how a 2018 paper brings neural network inference into the Level-1 trigger at the Large Hadron Collider, where event decisions must be made under sub-microsecond latency constraints. It explains why FPGAs are a natural fit for this setting, emphasizing batch-one, deterministic inference and the hardware realities that make model size, timing, memory use, and routing just as important as accuracy. The discussion centers on a compact dense network for jet substructure classification, using 16 engineered features to distinguish quark, gluon, W, Z, and top jets while preserving rare physics signals. It also highlights the paper’s broader argument: tools like High-Level Synthesis and hls4ml can let physicists deploy hardware-aware ML workflows directly, making real-time AI a practical part of scientific instrumentation rather than just a benchmark exercise. Sources: 1. Fast inference of deep neural networks in FPGAs for particle physics — Javier Duarte, Song Han, Philip Harris, Sergo Jindariani, Edward Kreinar, Benjamin Kreis, Jennifer Ngadiuba, Maurizio Pierini, Ryan Rivera, Nhan Tran, Zhenbin Wu, 2018 http://arxiv.org/abs/1804.06913 2. A Survey on Performance Optimization of High-Level Synthesis Tools — Lan Huang, Da-Lin Li, Kang-Ping Wang, Teng Gao, Adriano Tavares, 2020 https://scholar.google.com/scholar?q=A+Survey+on+Performance+Optimization+of+High-Level+Synthesis+Tools 3. FINN-R: An End-to-End Deep-Learning Framework for Fast Exploration of Quantized Neural Networks — Michaela Blott, Thomas B. Preusser, Nicholas J. Fraser, Giulio Gambardella, Kenneth O'Brien, Yaman Umuroglu, Miriam Leeser, Kees Vissers, 2018 https://scholar.google.com/scholar?q=FINN-R:+An+End-to-End+Deep-Learning+Framework+for+Fast+Exploration+of+Quantized+Neural+Networks 4. Fast inference of deep neural networks in FPGAs for particle physics — Javier Duarte, Song Han, Philip Harris, Sergo Jindariani, Edward Kreinar, Jennifer Ngadiuba, Maurizio Pierini, Nhan Tran, Zhenbin Wu, et al., 2018 https://scholar.google.com/scholar?q=Fast+inference+of+deep+neural+networks+in+FPGAs+for+particle+physics 5. Fast convolutional neural networks on FPGAs with hls4ml — Thea Aarrestad, Vladimir Loncar, Nicolo Ghielmetti, Maurizio Pierini, Sioni Summers, Jennifer Ngadiuba, Javier Duarte, Philip Harris, et al., 2021 https://scholar.google.com/scholar?q=Fast+convolutional+neural+networks+on+FPGAs+with+hls4ml 6. FINN: A Framework for Fast, Scalable Binarized Neural Network Inference — Yaman Umuroglu, Nicholas J. Fraser, Giulio Gambardella, Michaela Blott, Philip H. W. Leong, Magnus Jahre, Kees Vissers, 2017 https://scholar.google.com/scholar?q=FINN:+A+Framework+for+Fast,+Scalable+Binarized+Neural+Network+Inference 7. Serving DNNs in Real Time at Datacenter Scale with Project Brainwave — Eric Chung, Jeremy Fowers, Kalin Ovtcharov, Michael Papamichael, Adrian Caulfield, Todd Massengill, Ming Liu, Doug Burger, et al., 2018 https://scholar.google.com/scholar?q=Serving+DNNs+in+Real+Time+at+Datacenter+Scale+with+Project+Brainwave 8. Automatic heterogeneous quantization of deep neural networks for low-latency inference on the edge for particle detectors — Claudionor N. Coelho Jr., Aki Kuusela, Shan Li, Hao Zhuang, Jennifer Ngadiuba, Thea K. Aarrestad, Vladimir Loncar, Maurizio Pierini, et al., 2021 https://scholar.google.com/scholar?q=Automatic+heterogeneous+quantization+of+deep+neural+networks+for+low-latency+inference+on+the+edge+for+particle+detectors 9. Learning both Weights and Connections for Efficient Neural Network — Song Han, Jeff Pool, John Tran, William J. Dally, 2015 https://scholar.google.com/scholar?q=Learning+both+Weights+and+Connections+for+Efficient+Neural+Network 10. Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding — Song Han, Huizi Mao, William J. Dally, 2016 https://scholar.google.com/scholar?q=Deep+Compression:+Compressing+Deep+Neural+Networks+with+Pruning,+Trained+Quantization+and+Huffman+Coding 11. Jet Substructure at the Large Hadron Collider: A Review of Recent Advances in Theory and Machine Learning — Andrew J. Larkoski, Ian Moult, Benjamin Nachman, 2020 https://scholar.google.com/scholar?q=Jet+Substructure+at+the+Large+Hadron+Collider:+A+Review+of+Recent+Advances+in+Theory+and+Machine+Learning 12. Deep-learning Top Taggers or The End of QCD? — Gregor Kasieczka, Tilman Plehn, Michael Russell, Torben Schell, 2017 https://scholar.google.com/scholar?q=Deep-learning+Top+Taggers+or+The+End+of+QCD? 13. From High-Level Deep Neural Models to FPGAs — Hardik Sharma, Jongse Park, Divya Mahajan, Emmanuel Amaro, Joon Kyung Kim, Chenkai Shao, Asit Mishra, Hadi Esmaeilzadeh, 2016 https://scholar.google.com/scholar?q=From+High-Level+Deep+Neural+Models+to+FPGAs 14. Distance-Weighted Graph Neural Networks on FPGAs for Real-Time Particle Reconstruction in High Energy Physics — Yutaro Iiyama, Gianluca Cerminara, Abhijay Gupta, Jan Kieseler, Vladimir Loncar, Maurizio Pierini, Shah Rukh Qasim and collaborators, 2021 https://scholar.google.com/scholar?q=Distance-Weighted+Graph+Neural+Networks+on+FPGAs+for+Real-Time+Particle+Reconstruction+in+High+Energy+Physics 15. Low latency transformer inference on FPGAs for physics applications with hls4ml — not confirmed from snippet; likely hls4ml/particle-physics collaboration, recent https://scholar.google.com/scholar?q=Low+latency+transformer+inference+on+FPGAs+for+physics+applications+with+hls4ml 16. Optimizing transformer models for low-latency inference: techniques, architectures, and code implementations — not confirmed from snippet, recent https://scholar.google.com/scholar?q=Optimizing+transformer+models+for+low-latency+inference:+techniques,+architectures,+and+code+implementations 17. Low-bit mixed-precision quantization and acceleration of CNN for FPGA deployment — not confirmed from snippet, recent https://scholar.google.com/scholar?q=Low-bit+mixed-precision+quantization+and+acceleration+of+CNN+for+FPGA+deployment 18. MPQA: Mixed-Precision Quantization Accelerator for CNN Inference — not confirmed from snippet, recent https://scholar.google.com/scholar?q=MPQA:+Mixed-Precision+Quantization+Accelerator+for+CNN+Inference 19. Fine-grained structured sparse computing for FPGA-based AI inference — not confirmed from snippet, recent https://scholar.google.com/scholar?q=Fine-grained+structured+sparse+computing+for+FPGA-based+AI+inference 20. Efficient CNN inference acceleration on FPGAs: a pattern pruning-driven approach — not confirmed from snippet, recent https://scholar.google.com/scholar?q=Efficient+CNN+inference+acceleration+on+FPGAs:+a+pattern+pruning-driven+approach 21. Online Learning Extreme Learning Machine with Low-Complexity Predictive Plasticity Rule and FPGA Implementation — not confirmed from snippet, recent https://scholar.google.com/scholar?q=Online+Learning+Extreme+Learning+Machine+with+Low-Complexity+Predictive+Plasticity+Rule+and+FPGA+Implementation 22. An FPGA architecture for online learning using the Tsetlin machine — not confirmed from snippet, recent https://scholar.google.com/scholar?q=An+FPGA+architecture+for+online+learning+using+the+Tsetlin+machine 23. AI Post Transformers: FPGA Neural Network Accelerators for Space — Hal Turing & Dr. Ada Shannon, 2026 https://podcast.do-not-panic.com/episodes/2026-04-26-fpga-neural-network-accelerators-for-spa-3087ae.mp3 Interactive Visualization: Fast FPGA Inference for LHC Triggers