Next Article in Journal
Enhancing User Experience in Virtual Reality Through Optical Flow Simplification with the Help of Physiological Measurements: Pilot Study
Previous Article in Journal
Hybrid Underwater Image Enhancement via Dual Transmission Optimization and Transformer-Based Feature Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Architecture Design of a Convolutional Neural Network Accelerator for Heterogeneous Computing Based on a Fused Systolic Array

School of Information Science and Engineering, Shenyang University of Technology, Shenyang 110870, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2026, 26(2), 628; https://doi.org/10.3390/s26020628
Submission received: 20 December 2025 / Revised: 13 January 2026 / Accepted: 15 January 2026 / Published: 16 January 2026
(This article belongs to the Section Intelligent Sensors)

Abstract

Convolutional Neural Networks (CNNs) generally suffer from excessive computational overhead, high resource consumption, and complex network structures, which severely restrict the deployment on microprocessor chips. Existing related accelerators only have an energy efficiency ratio of 2.32–6.5925 GOPs/W, making it difficult to meet the low-power requirements of embedded application scenarios. To address these issues, this paper proposes a low-power and high-energy-efficiency CNN accelerator architecture based on a central processing unit (CPU) and an Application-Specific Integrated Circuit (ASIC) heterogeneous computing architecture, adopting an operator-fused systolic array algorithm with the YOLOv5n target detection network as the application benchmark. It integrates a 2D systolic array with Conv-BN fusion technology to achieve deep operator fusion of convolution, batch normalization and activation functions; optimizes the RISC-V core to reduce resource usage; and adopts a locking mechanism and a prefetching strategy for the asynchronous platform to ensure operational stability. Experiments on the Nexys Video development board show that the architecture achieves 20.6 GFLOPs of computational performance, 1.96 W of power consumption, and 10.46 GOPs/W of energy efficiency ratio, which is 58–350% higher than existing mainstream accelerators, thus demonstrating excellent potential for embedded deployment.

1. Introduction

Object Detection (OD) is a prominent research direction in computer vision and pattern recognition, aiming to accurately identify and localize multiple objects belonging to specific categories within images [1,2]. With the advancement of deep learning, the accuracy and efficiency of object detection have been significantly enhanced. Over the past decade, CNNs have found widespread application in visual tasks owing to their powerful feature extraction capabilities and efficient computational performance [3,4], thus becoming one of the mainstream methods in the field of OD. Several representative algorithms have emerged, such as Faster R-CNN [5], YOLO [6], ResNetX [7], and MobileNets [8]. Among these, the YOLO-series algorithms have achieved remarkable results across various OD tasks due to their superior performance [9,10,11].
As the complexity and performance of object detection algorithms continue to improve and model sizes gradually increase, this poses challenges for the implementation of neural network chips [12]. For instance, the YOLOv5 model series, characterized by its compact architecture and ease of deployment, has a parameter size of 46.5 MB for YOLOv5l; the later released YOLOv5x, by contrast, increases to 86.7 MB of parameters and a computational load of 205.7 GFLOPs [13]. The multi-layer feature fusion architecture and Spatial Pyramid Pooling-Fast (SPPF) operation in YOLOv5 involve extensive matrix multiplication and data interaction, which demand substantial computational units and memory resources for chip implementation to ensure operational efficiency [14]. Simultaneously, the massive model parameters result in significant time and hardware resource consumption during data movement and scheduling. These challenges restrict the application of neural networks to high-power, high-performance devices such as desktop computers or industrial control systems, thereby limiting the technological advancement of AI in edge computing and low-power scenarios [15,16].
To address the above circumstances, this paper focuses on resolving the deployment challenges of convolutional neural networks in resource-constrained scenarios. Accordingly, a heterogeneous computing convolutional neural network accelerator architecture based on the fused systolic array algorithm is designed, where the YOLOv5n network is used as the application benchmark. The main research work and contributions are as follows: (1) Operator fusion is achieved by integrating the Batch Normalization (BN) layer, Convolution (CONV) layer, and activation function. A high-performance convolutional accelerator algorithm based on the fused-operator systolic array is proposed. An attention mechanism is incorporated to mitigate the reduction in recognition accuracy induced by operator fusion, and modeling, simulations, and verification are performed in MATLAB (R2022a). (2) A low-power CPU based on the RISC-V instruction set is designed using an ICB bus. This CPU supports integration with convolutional accelerator units, pooling accelerator units, and cache modules, enabling the construction of a high-energy-efficiency CPU + ASIC heterogeneous computing platform. (3) The accelerator architecture is tested on the Digilent Nexys Video board platform (Digilent, Inc., Pullman, DC, USA) equipped with a Xilinx XCA7A35T Field Programmable Gate Array (FPGA). Test results using the BIT-Vehicle vehicle detection dataset demonstrate that operator fusion achieves significant computational speedup with minimal accuracy loss. The heterogeneous computing accelerator based on the fused systolic array achieves a throughput of 20.6 GOPS at a power consumption of only 1.96 W. Its energy efficiency significantly outperforms that of CPU and GPU deployments for convolutional neural networks, while its computational speed surpasses that of comparable accelerators validated on FPGAs.
The organization of this paper is as follows: Section 1 provides a review of the development of CNNs and the research content of this paper. Section 2 elaborates on the reasons for network selection, introduces the corresponding model, and identifies the problems to be solved. Section 3 presents a detailed description of the principles and schemes for accelerator design. Section 4 demonstrates the verification schemes and results of the accelerator. Finally, Section 5 offers a summary and discussion of the full paper.

2. Related Work

2.1. Object Detection Network

You Only Look Once (YOLO) is a widely adopted object detection algorithm characterized by high accuracy and low latency, enabling real-time object detection [17,18]. Since its debut in 2015, it has undergone multiple iterations and optimizations, with successive versions continuously improved. Its detection performance and efficiency have been significantly enhanced, providing valuable references for research and engineering applications in related fields [19,20].
At the hardware implementation level, the mainstream network models adopted include YOLOv3–Tiny [21], YOLOv4–Tiny [22], YOLOv5n [23], and YOLOv8n [24]. Their network parameters, model sizes, and release times are shown in Table 1. In comparison, YOLOv5n has the following significant advantages: First, it has a small number of model parameters and adopts an extremely lightweight design [25], which can effectively reduce the occupation of hardware computing resources and storage resources, allowing it to adapt to the resource-constrained deployment environment in integrated circuit systems. Second, the model architecture is simple, allowing it to achieve excellent real-time inference performance and a fast detection speed [26], making it suitable for integrated circuit-related application scenarios with strict requirements on real-time performance. Based on the above advantages, this study finally selects the YOLOv5n network as the core detection model to carry out relevant research work.

2.2. YOLOv5n Algorithm Model and Attention Module

The main architecture of the YOLOv5n network is illustrated in Figure 1 and is derived from the original work of the authors. It consists of fundamental operations including CONV, BN, the Sigmoid Linear Unit (SILU), Max Pooling (MaxPool), and feature concatenation (CONCAT) [27]; its input is a 640 × 640 3-channel RGB image, and the backbone implements feature extraction using Conv layers annotated as “Conv Kx,sx,px,cx”, where K, s, p, and c denote the kernel size, stride, padding and output channel number, respectively, paired with C3 modules annotated as “C3 cx”, with cx representing the output channel number, while multi-scale feature aggregation is achieved via the “SPPF K5,c1024” module. In the neck, features from different backbone levels are fused through Upsample and Concat operations, with the fused features processed by reused Conv layers and C3 modules; the detection head employs 1 × 1 convolution layers annotated as “1*1conv” to adjust channel numbers, outputting channels as “3 × (5 + 80)”, where 3 stands for prediction boxes per grid, 5 corresponds to box coordinates and confidence, and 80 refers to object classes in the COCO dataset, to generate multi-scale detection feature maps such as 80 × 80. Additionally, general modules including Conv2d, Batch Norm2d, and SiLu, which correspond to convolution operations, batch normalization, and activation functions, respectively, are integrated to support the network’s feature learning and detection tasks.
The Convolutional Block Attention Module (CBAM) aims to overcome the limitations of traditional convolutional neural networks in feature extraction. Through comparative analyses, researchers have demonstrated significant performance improvements by integrating CBAMs into classic architectures such as ResNet and MobileNet [28]. The CBAMs comprises two submodules: the Channel Attention Module (CAM) and the Spatial Attention Module (SAM) [29]. The Channel Attention Module enhances feature map representation by learning the importance of each channel, while the Spatial Attention Module strengthens spatial feature representation by learning the importance of each spatial location. The integration of CBAMs improves model performance by enabling the network to focus more on target object recognition, thereby enhancing model interpretability. CBAMs can be incorporated into any convolutional layer to boost feature extraction capabilities, ultimately improving the model’s recognition accuracy [30]. The CBAM, which is from the original work of the authors, is illustrated in Figure 2.

2.3. Analysis of Existing Problems

The YOLOv5n model itself exhibits high computational and control complexity, requiring a large number of arithmetic and control operations to complete a single inference task. In the context of accelerator design, such repetitive control operations are typically implemented using state machine groups or similar control mechanisms; however, an overly hierarchical structure can compromise the performance and reliability of the overall design [31]. Meanwhile, BN and SILU operations involve a large number of nonlinear computational operations, which are difficult to implement directly on hardware accelerators [32]. Moreover, the adoption of hardware-friendly algorithms for these operations tends to result in losses of model accuracy. In addition, convolutional operations impose a heavy computational burden on a single inference task, while feature fusion operations within the network lead to temporal discontinuities in the output data of intermediate layers [33]. These issues collectively further exacerbate the challenges in accelerator design.
To address the aforementioned problems, current hardware accelerators targeting the YOLOv5 network mainly adopt two technical approaches: model optimization and accelerator architecture improvement [34,35]. Reference [36] enhances accelerator performance by optimizing model structures and deployment quantization strategies. Specific measures include replacing standard convolutions with depthwise separable convolutions to achieve lightweighting of the YOLOv5s model and substituting the original backbone network with a MobileNet architecture to further reduce computational complexity and the parameter count. Reference [37] proposes a heterogeneous computing architecture, where software design is employed to handle complex and resource-intensive scheduling modules, while the convolutional acceleration unit is fully controlled by the CPU. By leveraging the superior scalability of the RISC-V instruction set, this architecture constructs a set of acceleration library functions and connects the accelerator to the CPU as a coprocessor via the EAI interface. The advantages of this scheme lie in transforming the accelerator design into a coprocessor form, which features simplified programming, high code density, and significant resource savings. Meanwhile, on-chip cache units can reduce data redundancy, improve data transmission efficiency, and thereby enhance the computational speed of the accelerator at the hardware level.
Nevertheless, none of the aforementioned accelerator design schemes have fully exploited the potential for performance acceleration. For instance, the model lightweighting strategy in Reference [36] comes at the cost of sacrificing recognition accuracy and fails to fully utilize the technical advantages of ASICs in large-scale matrix operations. Reference [37] transforms the hardware-driven design paradigm into a software-assisted one; although it improves data utilization through a caching mechanism, its accelerator architecture still suffers from resource wastage issues.

3. Accelerator Architecture Design

3.1. Convolution Layer and Batch Normalization Layer Fusion

When calculating the number of parameters in the YOLOv5n model, the CONV, BN, and SILU layers must be taken into account, among which CONV and BN exhibit strong interdependencies. Therefore, computational optimization should first integrate the CONV and BN layers to reduce the computational load, thereby lowering the consumption of hardware resources. This optimization strategy is illustrated in Figure 3, which is derived from the original work of the authors.
Assuming that in a single convolution operation, the input convolution kernel parameter is denoted as ω and the input feature map parameter is denoted as x, the convolution output yconv can be calculated according to Equation (1).
y c o n v = ω x + b
We then substitute the one-pass calculation formula of the aforementioned convolutional layer into the calculation formula of the BN layer [38], and the resulting expression is shown in Equation (2).
B N γ , β x = γ ω x + b μ B σ B 2 + ϵ + β = γ ω σ B 2 + ϵ x + γ σ B 2 + ϵ b μ B + β
Equation (3) describes the normalization and scaling process of network weights, while Equation (4) defines the normalization and shifting process of bias terms [39]. Substituting both equations into Equation (2) yields Equation (5).
ω ^ = γ ω σ B 2 + ϵ
b ^ = γ σ B 2 + ϵ b μ B + β
B N γ , β x = ω ^ x + b ^
As shown in Equation (5), the computations of the CONV and BN layers can be fused into a single CONV-like operation, which is termed the CBN operator. The computational complexity of this operator is equivalent to that of a convolutional operation with a bias term, and this fusion reduces the model’s computational load to approximately 60% of that of the original model.

3.2. Fused Systolic Array Algorithm

The convolution formula is presented below [39]. Specifically, the convolution kernel traverses the input feature map in a row-wise order with a stride of 1 pixel. During the traversal, it performs multiply–accumulate (MAC) operations on the covered local data. For the structural matching of convolution operations, the depth of each convolution kernel is consistent with that of the input feature map layer, while the number of convolution kernels in each layer is equal to the depth of the output feature map layer.
Q x , y , n = d j i C i , j , d , n × P x + i , y + j , d + B n
Here, Qx,y,n denotes a pixel in the output feature map layer, Ci,j,d,n represents the parameters of the convolution kernel, and Px+i,y+j,d corresponds to the pixel coordinates in the input feature map layer. To clarify the definition of each dimension parameter, x and y are the horizontal and vertical coordinates of pixels in the feature map layer, respectively; n denotes the depth coordinate of the output feature map layer; d represents the depth coordinate of the input feature map layer; and i and j are the horizontal and vertical coordinates of the convolution kernel parameters, respectively. From the parameter correlation and computation process described above, it can be observed that convolution operations inherently exhibit parallel computing characteristics.
The systolic array enhances the parallelism of convolution operations by constructing a rectangular processing array. It consists of 224 processing elements (PEs) arranged in a 16 × 16 configuration, with serial interconnections between the elements. Each processing element comprises a multiplier, an adder, and three sets of registers—where the registers are used to store pass-through parameters and intermediate computation results. Specifically, each column of the array takes in data from a convolution kernel, while each row inputs the corresponding input feature map data for that kernel. The computation proceeds in the downward and rightward directions, and the output of each processing element in the array corresponds to a single pixel in the output feature map layer. Taking a 3 × 3 systolic array as an example, Equations (7) and (8) illustrate the computational mechanism of the array and the operating principle of a single processing element, respectively, with both formulations based on the conventional matrix multiplication operation.
P E 00 P E 01 P E 02 P E 10 P E 11 P E 12 P E 20 P E 21 P E 22 = P 00 P 01 P 02 P 10 P 11 P 12 P 20 P 21 P 22 C 00 C 01 C 02 C 10 C 11 C 12 C 20 C 21 C 22
P E 00 = P 00 × C 00 + P 01 × C 10 + P 02 × C 20
In this context, PE denotes the output of a processing element within the systolic array, P represents the horizontally input feature map pixels, and C signifies the vertically input convolution kernel parameters. The final output of each systolic array corresponds to the result of a complete convolution operation, and the consistent accumulation operations within the array facilitate streamlined processing of the convolution task.
By incorporating the BIAS parameter into Equation (8), we obtain Equation (9). Considering the flushing characteristics of the systolic array, an additional set of adders (denoted as B) is added to the right and bottom sides of the array. After completing a convolution calculation and clearing the intermediate data, the results of the CBN operator can be continuously output at the adders (B). This design ultimately defines the integrated systolic array structure, which is illustrated in Figure 4.
C B N 00 = P 00 × C 00 + P 01 × C 10 + P 02 × C 20 + B I A S 0
Building upon the aforementioned foundation, we explored the secondary operator fusion technique by combining the CBN operator with the SILU function to derive a linear iterative scheme. This approach achieves compression for approximately 80% of the computational operations during network inference, along with minimal additional hardware resource consumption.

3.3. Pooled Accelerator Design

Pooling operation replaces the value of a point in a two-dimensional image layer with the maximum value within a matrix centered on that point. Taking (m, n) as the vertex and k as the radius, a k × k-sized matrix, A, is constructed. Matrix A contains k × k elements in total; let the value of the top-left element be xi, and let the maximum value in the matrix be xmax. The result of the pooling calculation is given by Equation (10).
x i = x max
As shown in Equation (10), the pooling operation requires accessing k × k elements, which imposes substantial memory access pressure on the system. To alleviate such pressure, a data-prefetching strategy is adopted. Meanwhile, data is transmitted through a cascaded pipeline with progressive discarding of lower-order bits, thereby ensuring that no data blocking occurs in the acceleration unit. Each pooling segment is defined as shown in Equation (11), where n denotes the pooling kernel index, d represents the kernel size, and Z denotes the image width.
n , n + d   ( n N , n [ 0 , Z ] )
Next, the elements within each pooling segment are compared pairwise. Combined with the corresponding encoding and decoding operations, this process ultimately yields the output result for the current pooling kernel. Given that the pooling units can be processed in a pipelined fashion, a theoretical acceleration effect of d2 can be achieved. The architecture of the pooling acceleration unit is illustrated in Figure 5.

3.4. Accelerator Architecture

The overall architecture of the accelerator is illustrated in Figure 6. The CPU is implemented based on the RISC-V instruction set and features the following functions: (1) fixed-point multiply–accumulate operations, which satisfy the scheduling requirements of the system; (2) a QSPI interface for connecting to the FLASH device that stores parameters such as weights; (3) a JTAG interface for loading control programs to adjust system functionality; (4) the DTCM, which is a CPU-integrated memory module that buffers data generated during the computation process, thereby improving the operational flexibility of the CPU; and (5) an ICB bus, where the CPU interfaces with the ASIC side via an ICB bus to handle the transmission and reception of instructions.
The ASIC side incorporates the following functional modules: (1) ICB bus interface: It interfaces with the ICB bus of the CPU system to facilitate command interaction between heterogeneous systems. (2) CBS: It is a convolution acceleration unit based on a fusion-operator systolic array. (3) Pool: It is a pipeline-based pooling acceleration unit. (4) MEM: It is an independent on-chip RAM module integrated into the ASIC side, which serves as the internal cache of the ASIC.
The primary challenge in heterogeneous computing lies in achieving data synchronization and instruction synchronization across disparate architectures. This architecture addresses this challenge via unified addressing and centralized control mechanisms. Specifically, the ASIC and CPU share a single main memory device (DTCM) with a pre-partitioned address space. During the scheduling phase, the CPU only operates on addresses within the designated data space, thus ensuring consistent data interaction across all architectures. Compared with the traditional data migration scheme, this approach reduces redundant memory accesses and thereby improves overall system performance. Centralized control refers to the strategy where the CPU acts as the master controller to manage the ASIC, which ensures the stable operation of the entire system. Critical feedback data from the ASIC is stored in a dedicated region of the MEM for subsequent retrieval by the CPU, which effectively resolves the cross-clock-domain and instruction conflict issues between heterogeneous architectures.

4. Experimental Results and Discussion

4.1. Experimental Environment

The accelerator verification platform designed in this paper is implemented on the Digilent Nexys Video development board (Digilent, Inc., Pullman, DC, USA), which is configured with the Xilinx XCA7A35T FPGA chip and operates at a system clock frequency of 100 MHz. The host system is a PC equipped with an Intel Core i5-13600K CPU and an NVIDIA GeForce RTX 3070 GPU, whose main functions include conducting simulation verification of the accelerator and performing training tasks for the network model.

4.2. Verification Plan

The construction of the accelerator verification platform is illustrated in Figure 7, where the chip designed in this study is marked by the purple box. Since the implementation scheme of this system only supports pre-trained model versions and does not have embedded training capabilities, the DDR simulation method is adopted to emulate external flash devices for inputting parameters such as weights in order to simplify the test platform and reduce testing costs. Meanwhile, an HDMI display device is configured to visually present the output results, thereby enabling an intuitive observation of whether the model outputs are consistent with the expected results.

4.3. Performance Evaluation of Operator-Fused Network Models

First, the base model was tested and validated on 889 images from the validation set of BIT-Vehicle, an open-source dataset for vehicle-category classification and detection. The accuracy (P), goodness of fit (R), mean Average Precision at 0.5 (mAP@0.5), and mean Average Precision across IoU thresholds of 0.5 to 0.95 (mAP@0.5–0.95) of the model for each vehicle category were obtained. The validation set consists of 177 SUVs, 545 sedans, 77 microbuses, 38 minivans, 88 trucks, and 39 buses, with the test results presented in Table 2. Subsequently, the operator-fused model was tested on the same validation set of BIT-Vehicle, and the corresponding results are shown in Table 3.
The results indicate that after operator fusion, the model’s accuracy remains nearly unchanged, while its size is reduced to one-third of the original. This not only verifies the effectiveness of the proposed algorithm design but also confirms its adaptability for hardware deployment—an essential requirement in integrated circuit applications.

4.4. Performance Evaluation

The resource utilized by the proposed acceleration unit, when deployed on the Nexys Video board platform, is illustrated in Table 4. As can be seen from the data, Block Random Access Memory (BRAM) accounts for the highest utilization rate of 73.84%. By contrast, the flip-flops (FFs), input/output (IO) pins, and global clock buffers (BUFGs) only occupy 2.48%, 3.16%, and 3.13% of the total available resources, respectively. In addition, the utilization rates of look-up tables (LUTs), LUT-based random access memory (LUTRAM), and digital signal processing (DSP) blocks are 8.20%, 6.65%, and 16.62%, which demonstrates a rational allocation of logic computing and dedicated signal processing resources.
To further evaluate the computational performance of the convolutional acceleration unit, a comparative analysis was conducted between its performance and resource consumption and those of other existing FPGA acceleration designs; the comparison results are presented in Table 5.
Even when accounting for parameter discrepancies across implementation schemes, including FPGA device selection, synthesis software configurations, and data type specifications, the proposed design maintains substantial advantages over existing methods. First, it features enhanced functional completeness: integrating convolutional and pooling units addresses the lack of dedicated pooling accelerators in all reference designs, obviating the need to offload pooling operations to external units. Second, CPU collaboration support enables flexible task scheduling by adapting to end-to-end CNN inference requirements. In terms of resource utilization, the design exhibits excellent LUT efficiency with a consumption of only 5.9% of the maximum in the references, boosting applicability to resource-constrained edge FPGA platforms. Moreover, it achieves the highest energy efficiency ratio of 10.46 GOPs/W, significantly outperforming the references. Notably, while maintaining float32 precision for inference accuracy, it delivers a competitive throughput of 20.6 GOPs, realizing a precision–throughput balance suitable for precision-sensitive edge tasks.
However, this design also has certain limitations; its consumption of BRAM is higher than that of the schemes in Reference [41] and Reference [43], which is a design trade-off made for dedicated memory allocation and may restrict the deployment of large-scale models. In addition, the design has limitations in device compatibility, and it requires re-optimization if migrated to high-end FPGA platforms.

5. Conclusions

CNN models are evolving toward larger scales and more complex structures. However, this trend poses enormous challenges to the low-power requirements of resource-constrained embedded devices, making it difficult to deploy large-scale CNN models. Existing acceleration solutions either fail to balance computational performance and power consumption or adopt overly complex hardware designs that are incompatible with embedded systems. To address this issue, the CNN accelerator architecture proposed in this paper adopts a heterogeneous computing framework integrating operator fusion and resource optimization strategies, which can effectively alleviate the contradiction between the complexity of CNN models and the resource constraints of embedded devices; maintain high computational efficiency and inference accuracy while ensuring low power consumption; and meet the practical needs of embedded platforms for high-performance and low-power CNN inference.
This paper proposes an FPGA accelerator for the YOLO object detection network. By adopting a heterogeneous computing architecture, the accelerator’s performance is maximized. The final experimental results demonstrate that this solution achieves a computational performance of 20.6 GFLOPs on the Nexys Video board platform, with a power consumption of only 1.96 W, exhibiting improved energy efficiency compared to other FPGA-based accelerator designs. The proposed accelerator architecture achieves approximately 56 frames per second (FPS) during YOLOv5 model inference. Its energy efficiency surpasses both GPU implementations and other large-scale FPGA-based accelerator designs, while its computational efficiency outperforms that of comparable accelerators. This research can be extended to develop specialized computing architectures for computer vision, enabling high-performance and energy-efficient processing of computer vision tasks, thus demonstrating significant potential for practical application.
This advanced technology has broad applicability in numerous edge computing scenarios, all of which have stringent requirements for real-time inference and low power consumption. Specifically, in intelligent security systems, it can support the high-speed processing of surveillance video streams to achieve real-time target detection and behavior analysis, enabling timely responses to potential security threats without relying on cloud computing resources; in the field of autonomous driving, the accelerator allows on-board embedded devices to efficiently run CNN models for environmental perception tasks such as lane detection, pedestrian recognition, and obstacle avoidance, providing reliable computing support for safe driving while meeting the strict power budget of on-board electronic systems; in addition, in wearable devices, its low-power characteristics make it suitable for integration into lightweight hardware such as smart bracelets and smart glasses, realizing functions like real-time health data analysis and gesture recognition without frequent charging, thereby enhancing the user experience.
Nonetheless, the current design still exhibits certain limitations: the throughput remains below that of comparable products, leaving room for further improvement, and the utilization ratio of Block Random Access Memory resources on the FPGA is relatively high. Future work will focus on optimizing hardware resource utilization and expanding support for additional neural network models.

Author Contributions

Conceptualization, Y.Z., J.R. and Z.M.; methodology, Y.Z., J.R., Z.M. and Y.C.; validation, Y.Z., J.R. and Z.M.; resources, Y.Z., J.R., B.L. and M.L.; writing—original draft preparation, Y.Z., Z.M. and Y.C.; writing—review and editing, Y.Z., J.R. and Z.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by Key Program of National Natural Science Foundation of China (No. 62531017); in part by General Program of National Natural Science Foundation of China (No. 62371315); in part by General Program of National Natural Science Foundation of China (No. 62571348); and in part by Liaoning Provincial Natural Science Foundation (No. 2025-MS-117).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
CPUCentral Processing Unit
ASICApplication-Specific Integrated Circuit
ODObject Detection
SPPFSpatial Pyramid Pooling-Fast
BNBatch Normalization
CONVConvolution
FPGAField-Programmable Gate Array
YOLOYou Only Look Once
ICIntegrated Circuit
SILUSigmoid Linear Unit
MaxPoolMax Pooling
CONCATConcatenation
CBAMConvolutional Block Attention Module
CAMChannel Attention Module
SAMSpatial Attention Module
DSPsDigital Signal Processors
MACMultiply–Accumulate
PEProcessing Element
BRAMBlock Random Access Memory
FFsFlip-Flops
IOInput/Output
BUFGsGlobal Clock Buffers
LUTsLook-Up Tables
LUTRAMLUT-based random access memory
FPSFrames Per Second

References

  1. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object detection in 20 years: A survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  2. Sai, B.N.K.; Sasikala, T. Object Detection and Count of Objects in Image using Tensor Flow Object Detection API. In Proceedings of the 2019 International Conference on Smart Systems and Inventive Technology (ICSSIT), Tirunelveli, India, 27–29 November 2019; pp. 542–546. [Google Scholar] [CrossRef]
  3. Li, T.Y.; Zhang, F.; Wang, S.; Cao, W.; Chen, L. FPGA-Based Unified Accelerator for Convolutional Neural Network and Vision Transformer. J. Electron. Inf. Technol. 2024, 46, 2663–2672. [Google Scholar] [CrossRef]
  4. Yu, Z.Q.; Liu, H.Y.; He, L.Y.; Yang, L.J.; Liu, B. Three Dimensional Reconstruction Method of Pipeline Defects Based on CNN-Transformer. Mod. Inf. Technol. 2024, 8, 48–55. [Google Scholar] [CrossRef]
  5. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the International Conference on Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  6. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  7. Sze, V.; Chen, Y.-H.; Yang, T.-J.; Emer, J. Efficient Processing of Deep Neural Networks: A Tutorial and Survey. Proc. IEEE 2017, 105, 2295–2329. [Google Scholar] [CrossRef]
  8. Chen, H.-Y.; Su, C.-Y. An Enhanced Hybrid MobileNet. In Proceedings of the 9th International Conference on Awareness Science and Technology (iCAST), Fukuoka, Japan, 19–21 September 2018; pp. 308–312. [Google Scholar]
  9. Wang, L.; Bai, J.; Li, W.; Jiang, J. Research Progress of YOLO Series Target Detection Algorithms. Comput. Eng. Appl. 2023, 59, 15–29. [Google Scholar]
  10. Mi, Z.; Lian, Z. Review of YOLO Methods for Universal Object Detection. Comput. Eng. Appl. 2024, 60, 38–54. [Google Scholar]
  11. Nazir, A.; Wani, M.A. You Only Look Once–Object Detection Models: A Review. In Proceedings of the 2023 10th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 15–17 March 2023; pp. 1088–1095. [Google Scholar]
  12. Zhang, Y.; Zhang, R.; Zhou, H.; Ji, Q.; Yu, Z.; Huang, T. Research status and development trends of vision foundation models. J. Image Graph. 2025, 30, 1–24. [Google Scholar] [CrossRef]
  13. Jocher, G.; Stoken, A.; Borovec, J. YOLOv5: v3.1—Bug Fixes and Performance Improvements. 2020. Available online: https://zenodo.org/record/4154370 (accessed on 1 April 2023).
  14. Dong, X.; Li, S.; Zhang, J. YOLOV5s object detection based on Sim SPPF hybrid pooling. Optoelectron. Lett. 2024, 20, 367–371. [Google Scholar] [CrossRef]
  15. Huyan, L.; Bai, Y.; Li, Y.; Jiang, D.; Zhang, Y.; Zhou, Q.; Wei, J.; Liu, J.; Zhang, Y.; Cui, T. A Lightweight Object Detection Framework for Remote Sensing Images. Remote Sens. 2021, 13, 683. [Google Scholar] [CrossRef]
  16. Qin, Y.; Lu, C.X.; Liu, Y.L.; Wu, Q.X.; Zhang, Q.; Zhang, X.B. Algorithm for the fast detection of aircraft targets on the basis of optical remote sensing for embedded platforms. Natl. Remote Sens. Bull. 2025, 29, 762–775. [Google Scholar] [CrossRef]
  17. Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo algorithm developments. Procedia Comput. Sci. 2022, 199, 1066–1073. [Google Scholar] [CrossRef]
  18. Hussain, M. YOLO-v1 to YOLO-v8, the rise of YOLO and its complementary nature toward digital manufacturing and industrial defect detection. Machines 2023, 11, 677. [Google Scholar] [CrossRef]
  19. Zhong, J.; Cheng, Q.; Hu, X.; Liu, Z. YOLO Adaptive Developments in Complex Natural Environments for Tiny Object Detection. Electronics 2024, 13, 2525. [Google Scholar] [CrossRef]
  20. Dabas, A.; Narwal, E. YOLO evolution: A comprehensive review and bibliometric analysis of object detection advancements. Int. J. Signal Imaging Syst. Eng. 2024, 13, 133–156. [Google Scholar] [CrossRef]
  21. Adarsh, P.; Rathi, P.; Kumar, M. YOLO v3-Tiny: Object Detection and Recognition using one stage improved model. In Proceedings of the 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 6–7 March 2020; pp. 687–694. [Google Scholar] [CrossRef]
  22. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  23. Miao, J.; Zhao, Y.; Gao, G.; Wen, Y. Fire Detection Algorithm Based on Improved YOLOv5. In Proceedings of the 2021 International Conference on Control, Automation and Information Sciences (ICCAIS), Xi’an, China, 14–17 October 2021; pp. 776–781. [Google Scholar] [CrossRef]
  24. Lou, H.; Duan, X.; Guo, J.; Liu, H.; Gu, J.; Bi, L.; Chen, H. DC-YOLOv8: Small-size object detection algorithm based on camera sensor. Electronics 2023, 12, 2323. [Google Scholar] [CrossRef]
  25. Xu, H.; Li, B.; Zhong, F. Light-YOLOv5: A Lightweight Algorithm for Improved YOLOv5 in Complex Fire Scenarios. Appl. Sci. 2022, 12, 12312. [Google Scholar] [CrossRef]
  26. Wang, J.; Bai, Z.; Zhang, X.; Qiu, Y. A Lightweight Remote Sensing Aircraft Object Detection Network Based on Improved YOLOv5n. Remote Sens. 2024, 16, 857. [Google Scholar] [CrossRef]
  27. Ultralytics. YOLOv5. 2022. Available online: https://github.com/ultralytics/yolov5 (accessed on 19 December 2025).
  28. Qin, J.; Zhang, S.; Wang, Y.; Yang, F.; Zhong, X.; Lu, W. Improved skeleton-based activity recognition using convolutional block attention module. Comput. Electr. Eng. 2024, 116, 109231. [Google Scholar] [CrossRef]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  30. Zhang, N.-N.; Zhang, X.; Bai, T.-C.; Shang, P.; Wang, W.-H.; Li, L. Identification Method of Cotton Leaf Pests and Diseases in Natural Environment Based on CBAM-YOLO v7. Trans. Chin. Soc. Agric. Mach. 2023, 54, 239–244. [Google Scholar]
  31. Alur, R.; Yannakakis, M. Model checking of hierarchical state machines. ACM Trans. Program. Lang. Syst. 2001, 23, 273–303. [Google Scholar] [CrossRef]
  32. Perryman, N.; Wilson, C.; George, A. Evaluation of Xilinx Versal Architecture for Next-Gen Edge Computing in Space. In Proceedings of the 2023 IEEE Aerospace Conference, Big Sky, MT, USA, 4–11 March 2023; pp. 1–11. [Google Scholar]
  33. Waeijen, L.; Sioutas, S.; Peemen, M.; Lindwer, M.; Corporaal, H. ConvFusion: A model for layer fusion in convolutional neural networks. IEEE Access 2021, 9, 168245–168267. [Google Scholar] [CrossRef]
  34. Heng, C.W.; Uttraphan, C.; Choon, C.C.; Ching, K.B. Optimizing FPGA-based YOLO series accelerators: A survey of techniques. Neurocomputing 2025, 650, 130874. [Google Scholar] [CrossRef]
  35. Yan, Z.; Zhang, B.; Wang, D. An FPGA-Based YOLOv5 Accelerator for Real-Time Industrial Vision Applications. Micromachines 2024, 15, 1164. [Google Scholar] [CrossRef]
  36. Wu, K. Research on FPGA-Based Heterogeneous Accelerator for Object Detection Algorithm. Master’s Thesis, Tiangong University, Tianjin, China, 2022. [Google Scholar]
  37. Wu, D.; Liu, Y.; Tao, C. A Universal Accelerated Coprocessor for Object Detection Based on RISC-V. Electronics 2023, 12, 475. [Google Scholar] [CrossRef]
  38. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  39. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-Based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  40. Zhang, C.; Li, P.; Sun, G.; Guan, Y.; Xiao, B.; Cong, J. Optimizing FPGA-Based Accelerator Design for Deep Convolutional Neural Networks. In Proceedings of the 2015 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays (FPGA), Monterey, CA, USA, 22–24 February 2015; pp. 161–170. [Google Scholar]
  41. Shen, X. Design and FPGA Verification of Convolutional Neural Network Acceleration System. Master’s Thesis, Southeast University, Nanjing, China, 2022. [Google Scholar]
  42. Kasem, K.; Kumar, A.; Bayoumi, M. Low-Power Convolutional Neural Network Accelerator on FPGA. In Proceedings of the 2023 IEEE 5th International Conference on Artificial Intelligence Circuits and Systems (AICAS), Hangzhou, China, 11–13 June 2023; pp. 1–5. [Google Scholar]
  43. Yu, H.; Li, S. A Higher Performance Accelerator for Resource-Limited FPGA to Deploy Deeper Object Detection Networks. In Proceedings of the International Conference on Anti-Counterfeiting, Security and Identification (ASID), Xiamen, China, 2–4 December 2022; pp. 143–147. [Google Scholar]
Figure 1. Operational diagram of YOLOv5n. The color-coded blocks represent different functional modules: pink for the standard Convolutional module (including Conv2d, Batch Normalization, and SiLu activation); yellow for the C3 feature extraction module; purple for the Concatenation operation (feature fusion); green for the Upsampling layer; and light blue for the SPPF module.
Figure 1. Operational diagram of YOLOv5n. The color-coded blocks represent different functional modules: pink for the standard Convolutional module (including Conv2d, Batch Normalization, and SiLu activation); yellow for the C3 feature extraction module; purple for the Concatenation operation (feature fusion); green for the Upsampling layer; and light blue for the SPPF module.
Sensors 26 00628 g001
Figure 2. Convolutional Block Attention Module. The diagram illustrates the sequential application of attention modules. Light blue represents the intermediate input feature map; green denotes the Channel Attention Module (CAM) which focuses on ‘what’ is meaningful; purple denotes the Spatial Attention Module (SAM) which focuses on ‘where’ is informative; and dark blue represents the final refined feature map.
Figure 2. Convolutional Block Attention Module. The diagram illustrates the sequential application of attention modules. Light blue represents the intermediate input feature map; green denotes the Channel Attention Module (CAM) which focuses on ‘what’ is meaningful; purple denotes the Spatial Attention Module (SAM) which focuses on ‘where’ is informative; and dark blue represents the final refined feature map.
Sensors 26 00628 g002
Figure 3. Schematic diagram of CONV-layer and BN-layer integration. The light blue blocks denote the individual components (CONV, BN, and SILU) in the original CBS module during the training phase. The yellow blocks represent the fused components in the modified CBS module, where BN parameters are mathematically integrated into the convolution layer (CONV with Bias) to enhance inference efficiency.
Figure 3. Schematic diagram of CONV-layer and BN-layer integration. The light blue blocks denote the individual components (CONV, BN, and SILU) in the original CBS module during the training phase. The yellow blocks represent the fused components in the modified CBS module, where BN parameters are mathematically integrated into the convolution layer (CONV with Bias) to enhance inference efficiency.
Sensors 26 00628 g003
Figure 4. Fused systolic array architecture. The green dashed box illustrates the micro-architecture of a PE, featuring a Multiplier (blue), a local Buffer (orange), and an Adder (pink) for quantification. The purple region denotes the overall SA layout, where yellow blocks represent the grid-mapped PE nodes interconnected for data flow (Pic and Conv).
Figure 4. Fused systolic array architecture. The green dashed box illustrates the micro-architecture of a PE, featuring a Multiplier (blue), a local Buffer (orange), and an Adder (pink) for quantification. The purple region denotes the overall SA layout, where yellow blocks represent the grid-mapped PE nodes interconnected for data flow (Pic and Conv).
Sensors 26 00628 g004
Figure 5. Flow-through batch acceleration unit. Different colors represent the functional stages of the pooling data pipeline: green blocks for input pooling data, purple and yellow blocks for intermediate comparison results at different stages, and the red block for the final output result. The light blue circles indicate the comparison logic units.
Figure 5. Flow-through batch acceleration unit. Different colors represent the functional stages of the pooling data pipeline: green blocks for input pooling data, purple and yellow blocks for intermediate comparison results at different stages, and the red block for the final output result. The light blue circles indicate the comparison logic units.
Sensors 26 00628 g005
Figure 6. Accelerator architecture. The components are color-coded based on their functional roles: bright yellow for the RISC-V CPU, while dark yellow represents the sub-modules including MEM, POOL, and CBS. The green block denotes the DTCM, orange signifies peripheral interfaces (JTAG and QSPI), and the blue and red blocks indicate the ICB and AXI interfaces, respectively. The purple background identifies the ASIC domain.
Figure 6. Accelerator architecture. The components are color-coded based on their functional roles: bright yellow for the RISC-V CPU, while dark yellow represents the sub-modules including MEM, POOL, and CBS. The green block denotes the DTCM, orange signifies peripheral interfaces (JTAG and QSPI), and the blue and red blocks indicate the ICB and AXI interfaces, respectively. The purple background identifies the ASIC domain.
Sensors 26 00628 g006
Figure 7. Schematic diagram of the verification platform.
Figure 7. Schematic diagram of the verification platform.
Sensors 26 00628 g007
Table 1. Network information.
Table 1. Network information.
NetworkParameters (M)Model Sizes (MB)Release Times
YOLOv3–Tiny8.733.72018
YOLOv4–Tiny6.0623.12020
YOLOv5n1.93.82021
YOLOv8n3.26.22023
Table 2. YOLOv5n model detection results.
Table 2. YOLOv5n model detection results.
Vehicle CategoryP/%R/%mAP@0.5mAP@0.5–0.95
SUV0.9290.9270.9680.918
Sedan0.9830.9610.9870.931
Microbus0.930.9090.9550.911
Minivan0.9430.8640.9490.866
Truck0.9730.9890.9940.984
Bus0.96610.9850.892
Table 3. Model detection performance in this paper.
Table 3. Model detection performance in this paper.
Vehicle CategoryP/%R/%mAP@0.5mAP@0.5–0.95
SUV0.8720.9130.9870.833
Sedan0.9590.9540.9770.905
Microbus0.8880.7530.8920.806
Minivan0.8120.680.7910.678
Truck0.9180.9550.9640.85
Bus0.91310.9680.892
Table 4. FPGA resource usage.
Table 4. FPGA resource usage.
ResourceUtilizationAvailableUtilization/%
LUT10,975133,8008.20
LUTRAM307246,2006.65
FF6670269,2002.48
BRAM269.5036573.84
DSP12374016.62
IO92853.16
BUFG1323.13
Table 5. Comparison results of relevant FPGA accelerators.
Table 5. Comparison results of relevant FPGA accelerators.
ParameterReference [40]Reference [41]Reference [42]Reference [43]This Paper
FunctionalConvolution Accelerator
Pooling
Accelerator
××××
CPU×××
Resource
Utilization
LUT186,25119,03132,17834,99510,975
BRAM102425NA102269.5
DSP2240148125197123
FPGA DeviceVirtex-VX485tPYNQ-Z2Altera 10 GXZC702Artix-7 200T
Operating Frequency (MHZ)100100120100100
Data Accuracyfloat32int16int8flaot32float32
Throughput
(GOPs)
61.62342.575.8320.6
Energy Efficiency Ratio (GOPs/W)3.315.596.59252.3210.46
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zong, Y.; Ma, Z.; Ren, J.; Cao, Y.; Li, M.; Liu, B. Architecture Design of a Convolutional Neural Network Accelerator for Heterogeneous Computing Based on a Fused Systolic Array. Sensors 2026, 26, 628. https://doi.org/10.3390/s26020628

AMA Style

Zong Y, Ma Z, Ren J, Cao Y, Li M, Liu B. Architecture Design of a Convolutional Neural Network Accelerator for Heterogeneous Computing Based on a Fused Systolic Array. Sensors. 2026; 26(2):628. https://doi.org/10.3390/s26020628

Chicago/Turabian Style

Zong, Yang, Zhenhao Ma, Jian Ren, Yu Cao, Meng Li, and Bin Liu. 2026. "Architecture Design of a Convolutional Neural Network Accelerator for Heterogeneous Computing Based on a Fused Systolic Array" Sensors 26, no. 2: 628. https://doi.org/10.3390/s26020628

APA Style

Zong, Y., Ma, Z., Ren, J., Cao, Y., Li, M., & Liu, B. (2026). Architecture Design of a Convolutional Neural Network Accelerator for Heterogeneous Computing Based on a Fused Systolic Array. Sensors, 26(2), 628. https://doi.org/10.3390/s26020628

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop