Preview

Journal of Instrument Engineering

Advanced search
Open Access Open Access  Restricted Access Subscription Access

Forward and Inverse Tasks for Inference Rate Estimation in Binary Neural Networks

https://doi.org/10.17586/0021-3454-2025-68-7-567-575

Abstract

Deploying Binary Neural Networks (BNNs) presents significant challenges, particularly in selecting suitable hardware to achieve desired performance levels and accurately estimating computational costs. To address these issues, a novel metric named "XNOROP" was recently introduced, offering a simplified approach for estimating the computational cost of BNNs and introducing a method for compressing binary filters. This paper leverages " XNOROP" to define and solve two critical tasks in BNNs. The first task, referred to as the "Forward Task", involves estimating the inference rate of a given model M when deployed on a specific target device T. The second task, known as the "Inverse Task", outlines a systematic procedure to identify a set of target devices T capable of achieving a required inference rate M when deploying the model . Additionally, we extend the foundational formula of "XNOROP" and introduce “LXNOROP” which incorporates considerations for memory access time, enhancing its applicability for real-world deployment scenarios.

About the Authors

A. Shakkouf
ITMO University
Russian Federation

Ali Shakkouf — Post-Graduate Student; Faculty of Control Systems and Robotics

St. Petersburg



V. S. Gromov
ITMO University
Russian Federation

Vladislav S. Gromov — PhD; Faculty of Control Systems and Robotics; Associate Professor

St. Petersburg



References

1. Tan S., Zhang Z., Cai Y., Ergu D., Wu L., Hu B., Yu P., & Zhao Y. arXiv preprint, arXiv:2408.00496, 2024.

2. Lin X.V., Shrivastava A., Luo L., Iyer S., Lewis M., Gosh G., Zettlemoyer L., & Aghajanyan A. arXiv preprint, arXiv:2407.21770, 2024.

3. Cui Z., Yao J., Zeng L., Yang J., Liu W., & Wang X. arXiv preprint, arXiv: 2407.18054, 2024.

4. Calculate Computational Efficiency of Deep Learning Models with FLOPs and MACs, KDnuggets, June 19, 2023, https://www.kdnuggets.com/2023/06/calculate-computational-efficiency-deep-learning-models-flops-macs.html.

5. Dongarra J. Performance of various computers using standard linear equations software in a Fortran environment, Argonne National Laboratory Report ANL-80-70, Argonne, Illinois, 1979.

6. Courbariaux M., Bengio Y., & David J.-P. Binary Connect: Training Deep Neural Networks with binary weights during propagations, 2016, https://arxiv.org/abs/1511.00363.

7. Shakkouf A. and Sergeevich G.V. 36th Conference of Open Innovations Association (FRUCT), Lappeenranta, Finland, 2024, pp. 721–728, DOI: 10.23919/FRUCT64283.2024.10749876.

8. Wang L., Zhan J., Gao W., Yang K., Jiang Z., Ren R., He X., & Luo C. BOPS, Not FLOPS! A New Metric and Roofline Performance Model For Datacenter Computing, 2019, https://arxiv.org/abs/1801.09212.

9. Moosmann J., Bonazzi P., Li Y., Bian S., Mayer P., Benini L., & Magno M. Ultra-Efficient On-Device Object Detection on AI-Integrated Smart Glasses with TinyissimoYOLO, 2023, https://arxiv.org/abs/2311.01057.

10. Geiger et al. Journal of Open-Source Software, 2020, no. 5(45), pp. 1746, https://doi.org/10.21105/joss.01746.


Review

For citations:


Shakkouf A., Gromov V.S. Forward and Inverse Tasks for Inference Rate Estimation in Binary Neural Networks. Journal of Instrument Engineering. 2025;68(7):567-575. (In Russ.) https://doi.org/10.17586/0021-3454-2025-68-7-567-575

Views: 11


ISSN 0021-3454 (Print)
ISSN 2500-0381 (Online)