

Forward and Inverse Tasks for Inference Rate Estimation in Binary Neural Networks
https://doi.org/10.17586/0021-3454-2025-68-7-567-575
Abstract
Deploying Binary Neural Networks (BNNs) presents significant challenges, particularly in selecting suitable hardware to achieve desired performance levels and accurately estimating computational costs. To address these issues, a novel metric named "XNOROP" was recently introduced, offering a simplified approach for estimating the computational cost of BNNs and introducing a method for compressing binary filters. This paper leverages " XNOROP" to define and solve two critical tasks in BNNs. The first task, referred to as the "Forward Task", involves estimating the inference rate of a given model M when deployed on a specific target device T. The second task, known as the "Inverse Task", outlines a systematic procedure to identify a set of target devices T capable of achieving a required inference rate M when deploying the model . Additionally, we extend the foundational formula of "XNOROP" and introduce “LXNOROP” which incorporates considerations for memory access time, enhancing its applicability for real-world deployment scenarios.
About the Authors
A. ShakkoufRussian Federation
Ali Shakkouf — Post-Graduate Student; Faculty of Control Systems and Robotics
St. Petersburg
V. S. Gromov
Russian Federation
Vladislav S. Gromov — PhD; Faculty of Control Systems and Robotics; Associate Professor
St. Petersburg
References
1. Tan S., Zhang Z., Cai Y., Ergu D., Wu L., Hu B., Yu P., & Zhao Y. arXiv preprint, arXiv:2408.00496, 2024.
2. Lin X.V., Shrivastava A., Luo L., Iyer S., Lewis M., Gosh G., Zettlemoyer L., & Aghajanyan A. arXiv preprint, arXiv:2407.21770, 2024.
3. Cui Z., Yao J., Zeng L., Yang J., Liu W., & Wang X. arXiv preprint, arXiv: 2407.18054, 2024.
4. Calculate Computational Efficiency of Deep Learning Models with FLOPs and MACs, KDnuggets, June 19, 2023, https://www.kdnuggets.com/2023/06/calculate-computational-efficiency-deep-learning-models-flops-macs.html.
5. Dongarra J. Performance of various computers using standard linear equations software in a Fortran environment, Argonne National Laboratory Report ANL-80-70, Argonne, Illinois, 1979.
6. Courbariaux M., Bengio Y., & David J.-P. Binary Connect: Training Deep Neural Networks with binary weights during propagations, 2016, https://arxiv.org/abs/1511.00363.
7. Shakkouf A. and Sergeevich G.V. 36th Conference of Open Innovations Association (FRUCT), Lappeenranta, Finland, 2024, pp. 721–728, DOI: 10.23919/FRUCT64283.2024.10749876.
8. Wang L., Zhan J., Gao W., Yang K., Jiang Z., Ren R., He X., & Luo C. BOPS, Not FLOPS! A New Metric and Roofline Performance Model For Datacenter Computing, 2019, https://arxiv.org/abs/1801.09212.
9. Moosmann J., Bonazzi P., Li Y., Bian S., Mayer P., Benini L., & Magno M. Ultra-Efficient On-Device Object Detection on AI-Integrated Smart Glasses with TinyissimoYOLO, 2023, https://arxiv.org/abs/2311.01057.
10. Geiger et al. Journal of Open-Source Software, 2020, no. 5(45), pp. 1746, https://doi.org/10.21105/joss.01746.
Review
For citations:
Shakkouf A., Gromov V.S. Forward and Inverse Tasks for Inference Rate Estimation in Binary Neural Networks. Journal of Instrument Engineering. 2025;68(7):567-575. (In Russ.) https://doi.org/10.17586/0021-3454-2025-68-7-567-575