Next Article in Journal
Multi-Sensor Platform for Predictive Air Quality Monitoring
Previous Article in Journal
Multisensor Data Fusion in IoT Environments in Dempster–Shafer Theory Setting: An Improved Evidence Distance-Based Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved YOLOv7 Network Model for Gangue Selection Robot for Gangue and Foreign Matter Detection in Coal

1
School of Mechanical Engineering, Tiangong University, Tianjin 300387, China
2
Tianjin Photoelectric Detection Technology and System Key Laboratory, Tiangong University, Tianjin 300387, China
3
School of Electronics and Information Engineering, Tiangong University, Tianjin 300387, China
4
Center for Engineering Internship and Training, Tiangong University, Tianjin 300387, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(11), 5140; https://doi.org/10.3390/s23115140
Submission received: 7 April 2023 / Revised: 25 May 2023 / Accepted: 25 May 2023 / Published: 28 May 2023
(This article belongs to the Section Industrial Sensors)

Abstract

:
Coal production often involves a substantial presence of gangue and foreign matter, which not only impacts the thermal properties of coal and but also leads to damage to transportation equipment. Selection robots for gangue removal have garnered attention in research. However, existing methods suffer from limitations, including slow selection speed and low recognition accuracy. To address these issues, this study proposes an improved method for detecting gangue and foreign matter in coal, utilizing a gangue selection robot with an enhanced YOLOv7 network model. The proposed approach entails the collection of coal, gangue, and foreign matter images using an industrial camera, which are then utilized to create an image dataset. The method involves reducing the number of convolution layers of the backbone, adding a small size detection layer to the head to enhance the small target detection, introducing a contextual transformer networks (COTN) module, employing a distance intersection over union (DIoU) loss border regression loss function to calculate the overlap between predicted and real frames, and incorporating a dual path attention mechanism. These enhancements culminate in the development of a novel YOLOv71 + COTN network model. Subsequently, the YOLOv71 + COTN network model was trained and evaluated using the prepared dataset. Experimental results demonstrated the superior performance of the proposed method compared to the original YOLOv7 network model. Specifically, the method exhibits a 3.97% increase in precision, a 4.4% increase in recall, and a 4.5% increase in mAP0.5. Additionally, the method reduced GPU memory consumption during runtime, enabling fast and accurate detection of gangue and foreign matter.

1. Introduction

In recent years, the integration of robots in gangue and foreign matter screening has gained significant momentum as a notable development trend. Researchers have shown considerable interest in the visual information-based gangue robot selection method [1,2,3]. The efficacy of a gangue selection robot, which relies on visual information, is contingent upon precise gangue and foreign object recognition, as well as the speed of the manipulator. Currently, gangue selection robots face challenges, including slow selection speed and limited recognition accuracy. Attaining high-speed and accurate target recognition is a pivotal factor in enabling the automatic screening of a visual information-based gangue selection robot [4].
In the field of target recognition, researchers are progressively transitioning from relying on physical differences to exploiting faster and more intuitive visual distinctions. The integration of machine learning techniques, especially deep learning, has brought about a revolutionary change in traditional image classification recognition [5]. These methods demonstrate high accuracy in controlled laboratory settings. However, they suffer from limitations due to the manual extraction of image features inherent in traditional machine learning approaches, resulting in models with limited generalization capabilities. Additionally, the recognition speed of conventional machine learning methods is insufficient to meet the practical production requirements of coal mines.
To overcome the limitations of slow speed, low efficiency, fluctuating accuracy, and limited applicability associated with traditional image processing-based classification and recognition methods, recent research has delved into advanced techniques. Gao et al. [6] tackled these challenges by training and evaluating collected images using the U-net network, resulting in successful localization and classification of coal gangue. Li et al. [7] introduced a hierarchical coal gangue detection framework based on deep learning models. enabling the identification of multiple targets within a single image. They progressively refined their modules, transitioning from a simple LeNet structure to an Inception structure [8,9,10], and eventually to a ResNet structure [11,12]. The depth of the network modules and the number of corresponding layers increased throughout the research, starting from the 8 convolutional layers of the LeNet structure and culminating in the deep network structure of ResNetSt-269, which comprised 269 layers [13].
The continuous optimization of network model designs has resulted in significant improvements in recognition accuracy and speed. These advancements have not only enhanced the models’ recognition accuracy and generalization ability [14], but they have also addressed various challenges in image classification, image segmentation [15], and super-resolution [16]. Notable contributions include the development of target detection algorithms, such as R-CNN [17], fast R-CNN [18], R-CNN [19], as well as the SSD series [20,21] and YOLO series [22,23,24,25]. Attention mechanisms, such as SENet [26], CBAM [27], and ECA [28,29], have also played a significant role in further enhancing performance.
As network models continue to improve in speed and accuracy, the size of network parameters is also increasing. This presents challenges in practical applications, particularly in the context of the coal gangue recognition classification network based on Inception module migration learning, where the large memory requirements, reaching gigabytes, impede the embedding process. To tackle this issue, Google introduced the MobileNet network model [30,31]. By utilizing depth-separable convolution in place conventional convolution operations, the overall network parameters are reduced. However, concerns still exist regarding lower recognition accuracy.
This study addresses the issues of slow gangue selection speed and low recognition accuracy in gangue selection machine methods. To overcome these challenges, an improved gangue and foreign matter recognition network model is proposed. Anchor rods and I-beams were selected as the representative foreign matter. The YOLOv7 model structure served as the foundation for this approach. An industrial camera was utilized to capture raw images of coal blocks, gangue blocks, anchor rods, and I-beams. Data augmentation techniques were employed to enhance the dataset’s representativeness and alignment with coal mine environments. The improved model incorporated a new contextual transformer networks (COTN) module, which replaced the convolution in ResNet, resulting in the new YOLOv71 + COTN model. Additionally, heavy parameter processing was applied to the output section of the new model. These enhancements significantly improved network model performance, including enhanced output feature expressiveness, recognition speed, and accuracy, while minimizing the increase in model parameter volume and complexity.
In this study, the distance intersection over union (DIoU) loss function was chosen for border regression to effectively handle the overlap between the predicted frame and the ground truth frame. The utilization of the DIoU loss function offered several advantages, including improved robustness, faster convergence speed, and enhanced detection performance. The prediction frame was represented as (x1, x2, x3, x4), while the ground truth frame was represented as (x1’, x2’, x3’, x4’). By employing the DIoU loss, the predicted frame was less affected by the size of the ground truth bounding box, resulting in improved robustness of the model. Unlike l2 loss, DIoU loss directly minimized the distance between the two bounding boxes, leading to faster convergence and better detection performance. Furthermore, DIoU loss function was scale-invariant, further contributing to its effectiveness.
Prior to training the model on the dataset, the input sample images underwent a normalization process. This step played a vital role in enhancing the accuracy and efficiency of gangue and foreign matter image detection, thereby endowing the new model with robust multi-scale image processing capabilities. As a result, the new model demonstrated rapid detection and identification of gangue and foreign matter within coal samples. Furthermore, even in scenarios where these objects are covered with coal dust, they can still be reliably identified based on their distinctive physical characteristics. The new and improved model achieved these capabilities while reducing the number of parameters and the complexity of the network structure, leading to enhanced speed and accuracy in object identification. These advancements offer the benefits of low detection costs, high accuracy, and reliability.

2. Related Works

The presence of gangue and foreign matter in coal production has a substantial impact on the heating value of coal and can cause damage to transportation equipment, such as belt conveyors. Thus, it is essential to detect and effectively remove gangue and foreign matter to ensure reliable, efficient, and environmentally friendly sorting processes [32]. The development of automatic systems for accurate and efficient detection and removal of gangue and foreign matter holds paramount practical significance in the coal industry.
Traditional methods for gangue selection in coal processing include manual gangue selection, jigging gangue selection, heavy media gangue selection, and ray gangue selection [33]. Manual gangue separation is characterized by high labor intensity and low efficiency. Jigging gangue selection offers a simple process, ease of operation, and good processing capacity. However, it struggles with smaller gangue blocks. Heavy medium gangue exhibits high efficiency and wide applicability to different gangue types, but it suffers from a complex process system, significant equipment wear, and high cost. Ray gangue selection involves X-ray and γ-ray techniques, providing high efficiency and reliability. Nevertheless, environmental concerns arise with this method, which makes it less aligned with the principles of green development.
The YOLOv7 network is composed of three main parts [34]: the input, backbone, and head, as shown in Figure 1. In contrast to YOLOv5, YOLOv7 combines the neck and head layers into a head layer, while retaining the functionality of each part similar YOLOv5. The backbone is responsible for extracting features, and the head is utilized for making predictions. The network process in YOLOv7 begins with the preprocessing of the input image into a standardized RGB format with dimensions of 640 × 640. The processed image is then fed into the backbone network. According to the outputs of the three layers within the backbone network, feature maps of different sizes are generated and subsequently passed through the head layer. Following the RepVGG block and Conv, target detection takes place, ultimately yielding the final output.
The backbone part of the YOLOv7 network model comprises four CBSs (i.e., convolution, batch normalization, and Sigmoid weighted liner unit (SiLU) activation) followed by an enhanced light-weight aggregation network (ELAN). By combining the outputs of three MaxPooling (MP) layers with ELAN, feature maps of sizes 80 × 80 × 512, 40 × 40 × 1024, and 20 × 20 × 1024 are obtained. Each MP layer consists of 5 layers, while ELAN is composed of 8 layers, resulting in a total of 51 layers in the backbone section. The head section of YOLOv7 adopts a structure similar to pafpn observed in YOLOv4 and YOLOv5. Initially, the final output of the 32-fold downsampled feature map C5 from the backbone is obtained. Subsequently, spatial pyramid pooling with convolutional-spatial path (SPPCSP) is applied to reduce the number of channels from 1024 to 512.

3. Methods

3.1. YOLOv7 Network Model Improvement

The YOLOv7 backbone feature extraction network is a CNN-based network known for its translation invariance and localization capabilities. However, it lacks the ability to model globally and over long distances. To overcome this limitation, the transformer framework, widely used in natural language processing, is introduced to construct the CNN + transformer architecture, forming the COTN module. By integrating the transformer framework, the target detection capabilities are significantly enhanced, especially for detecting small gangue blocks, coal blocks, and densely packed objects on the conveyor. This adaptation is vital to effectively handle the presence of a large number of coal blocks on the conveyor belt. Furthermore, insights from other advancements in the YOLO model are incorporated to further improve the overall performance [35,36,37].
The introduced transformer framework, implemented in this study, incorporates a novel spatial modeling mechanism based on dot product self-attention. This framework leverages recursive gated convolution and enables higher-order spatial interactions through gated convolution and recursive design. Consequently, it offers a high degree of flexibility and customizability. YOLOv7 model’s substantial number of parameters and computational resources enable the compatibility of the CNN + transformer architecture with various convolutional variants. Furthermore, it allows for the extension of second-order interactions in self-attention to arbitrary orders without incurring additional computational overhead.
The improvements to the YOLOv7 model are depicted in Figure 2, visually illustrating the enhancements. Specifically, the initial segment of the backbone enhanced performance achieved by reducing the number of convolutional layers by half. This reduction leads to a decrease in the overall number of network layers and the spatial size of the model, resulting in a shorter model running time. Consequently, these modifications give rise to the YOLOv71 network structure. Table 1 presents a comprehensive comparison of the number of layers and running time between the two model structures.
To enhance the learning ability of small target detection, a small-sized detection layer is introduced in the head section. The inclusion of the COTN module further enhances the capabilities of the model. The DIoU loss border regression loss function is employed to calculate the overlap between the predicted frame and the ground truth frame, enabling the identification of gangue and foreign objects based on this overlap. Additionally, a dual-path attention mechanism is incorporated to improve recognition accuracy.
In this study, it was observed that approximately 70% of the gangue particles had a size greater than 50 mm, indicating a prevalence of large gangue. The remaining 30% of the gangue particles had a size less than 50 mm, representing the category of small gangue. The coal mine environment poses challenges in detecting smaller objects during target detection. To account for the variations in particle sizes encountered in real-world scenarios, the gangue used in the experiment encompassed both large and small gangue.
This study presents the process of gangue and foreign matter detection in coal using an improved YOLOv7 network model for a gangue selection robot, as shown in Figure 3. An industrial camera was used to capture images of coal, gangue, and foreign matter. These images underwent classification, annotation, and augmentation to generate a dataset. The dataset was then utilized to train and test the improved YOLOv7 network model, enabling the determination of optimal model weights. Subsequently, the obtained model weights were applied to the improved YOLOv7 network model to facilitate the detection and identification of gangue and foreign objects in coal on a belt conveyor.

3.2. Re-Parameterization

Two convolutional layers were incorporated into the batch normalization (BN) layer in the head section. This reparameterization of the three components is illustrated in Figure 4.
The convergence equation for Conv and BN can be expressed as follows:
N e w   C o n v = r w x + b m v + β = r w v x + r ( b m ) v + β
where w denotes the convolutional bias, b denotes the convolutional bias, r and β denote the parameters that can be learned in BN, m denotes the input mean in BN, and v denotes the input standard deviation in BN.
Ŵ = r w v
b = r ( b m ) v + β
Equations (2) and (3) are combined and incorporated into Equation (1), resulting in the derivation of the new fusion Equation (4).
N e w   C o n v = ŵ x + b
The YOLOv7 network model structure consists of 415 layers, contributing to a significant number of network layers. However, this extensive layer count led to prolonged training and recognition time, creating challenges in achieving fast recognition of foreign matter in coal on a rapidly moving belt conveyor. To tackle this issue, a reparameterized convolution process was introduced before image output. This process enhanced the model’s running speed, resulting in improved recognition speed while maintaining consistent model performance.

3.3. CONT Module

COTN utilized the transformer framework to replace the convolution in ResNet, serving as the backbone of the network. This replacement involves the utilization of a 1 × 1 convolution, enabling the seamless integration of contextual information mining and self-attentive learning within a unified architecture. Through the enhancement of self-attention, COTN enables efficient learning of contextual information, leading to improved expressiveness of the output features.
While the transformer framework demonstrates strong global modeling capability for long-distance interactions, it primarily calculates the attention matrix based on the interaction between query and key, neglecting the connection between adjacent keys. To overcome this limitation, a 3 × 3 convolution was applied to the key to model static contextual information. This convolution operation, illustrated in Figure 5, captures localized information. The key was combined with the modeled query and context information using the COTN module. Following this, two consecutive 1 × 1 convolutions were employed to generate dynamic contexts through self-attention. Finally, the static and dynamic context information was fused together to produce the output.

3.4. Margin Regression Loss Function

In the target detection process, the target bounding box was commonly represented by four variables (i.e., x, y, w, h). In this study, the predicted frame was denoted as (x1, x2, x3, x4) and the ground true frame was denoted as (x1’, x2’, x3’, x4’). Notably, the intersection over union (IoU) loss treats both large and small bounding boxes equally. When dealing with images of varying resolutions (i.e., different bounding box sizes), the prediction box obtained by l2 loss is more influenced by the size of the ground true bounding box. Conversely, the prediction box obtained with IoU loss is less affected by the ground true bounding box’s size, leading to improved robustness. Compared to l2 loss, IoU loss directly measures the distance between two bounding boxes, resulting in faster convergence speed and enhanced detection performance. Additionally, IoU loss demonstrates scale invariance, meaning it performs consistently well for both large and small bounding boxes.
This natural normalized loss function enhances the model’s ability to handle multiscale images. IoU value is 1 when the prediction perfectly matches the ground truth, and 0 when there is no overlap between the prediction and the ground truth. As IoU loss tends to approach positive infinity when IoU is 0, it decreases monotonically from positive infinity to 0 as IoU increases within the interval [0, 1]. IoU loss can be viewed as a form of cross-entropy loss of that quantifies the dissimilarity between the predicted and ground truth bounding boxes. Compared with l2 loss, which measures the Euclidean distance between bounding boxes, IoU loss directly captures the spatial overlap between the two bounding boxes, leading to faster convergence and improved detection performance. Additionally, IoU loss demonstrates scale invariance and remains effective for both large and small bounding boxes.
l2 loss border regression loss function can be calculated as follows:
l 2   loss = P r e d i c t i o n T r u t h 2 2
IoU loss border regression loss function can be calculated as follows:
I o U   l o s s = I n I n t e r s e c t i o n ( P r e d i c t i o n , T r u t h ) U n i o n ( P r e d i c t i o n , T r u t h )
where Prediction denotes the prediction box for the detection target, and Truth denotes the detection target truth box.
In situations where the prediction frame and the target frame do not overlap, the IoU loss function IoU(A,B) = 0, which does not capture the spatial distance between the two frames A and B. Consequently, the IoU loss function lacks differentiability and fails to effectively optimize cases where the frames do not intersect. Additionally, assuming fixed sizes for the prediction box A and target box B, the IoU value remains unchanged regardless of the specific intersection pattern between the boxes. Thus, IoU values alone do not provide detailed insights into the nature of the intersection. To overcome these limitations, the DIoU loss function was employed as a border regression loss. DIoU loss considers the distance between the prediction and target boxes, offering a more comprehensive optimization measure.
Figure 6 illustrates the data calculations for the predicted and ground truth boxes in relation to each detection target. The red boxes represent the ground truth boxes, while the green boxes correspond to the predicted boxes generated by the model. In this study, anchor rods and I-beams are selected as representative foreign matter. When Prediction and Truth do not intersect, the IoU value was 0. However, this value does not capture the spatial distance between the two frames, rendering the loss function non-differentiable. As a result, IoU loss cannot effectively optimize scenarios where the boxes do not intersect. Additionally, when assuming fixed sizes for both Prediction and Truth, their IoU values remain the same regardless of the specific intersection pattern between the boxes. Consequently, IoU values fail to provide information about the actual characteristics of the intersection. To overcome these limitations, DIoU loss is introduced as a border regression loss function, as depicted in Figure 7.
DIoU loss can be calculated as follows:
L = 1 I o U + R ( B , B g t )
R ( B , B g t ) = b b g t 2 2 c 2
where B denotes Prediction box, Bgt denotes Truth box, R(B, Bgt) denotes the added penalty term, and c denotes the square of the diagonal length of the minimum enclosing box c. The DIoU loss introduces the direct Euclidean distance between the two boxes as a penalty term, leading to a faster convergence rate compared to generalized intersection over union (GIoU) loss. Furthermore, DIoU loss takes into account the relative proportions of the rectangular boxes in the penalty term, which helps resolve the issue of box intersection between Prediction and Truth.

3.5. Attention Mechanism

The attention mechanism in artificial neural networks is inspired by the information acquisition behavior of the human brain. When humans gather information through their senses, the presence of numerous stimuli can create distractions, making it challenging to focus on the desired information. To solve this problem, the brain employs specialized processing units that facilitate effective processing and monitoring of relevant information. In the context of artificial neural networks, the attention mechanism aims to replicate this behavior by assigning specific weights to different targets during the feature extraction process. This allows the neural network to prioritize and emphasize important targets or regions of interest while suppressing or disregarding irrelevant or uninformative regions.
The attention mechanism, as shown in Figure 8, involves the output of a green square region denoted as D (tx, ty, t1). In this representation, the central coordinates are represented by tx and ty, while half of the side lengths of the square region are represented by t1. The upper left coordinates are given by (tx − t1) and (ty − t1), and the lower right coordinates are given by (tx + t1) and (ty + t1). This square region serves as the core region that captures image category features under the attention mechanism. The attention extraction network is composed of two main parts. The first part, denoted as ‘a’, represents the features extracted by the attention mechanism. This component focuses on capturing important information within the region of interest. The second part, denoted as ‘c’, represents the features extracted through the convolution operations. This component is responsible for capturing features at different scales and resolutions within the image.

4. Experiments and Analysis of Results

4.1. Configuration of Experimental Environment

The models in this study were trained and tested using PyTorch open-source framework renowned for its flexibility and versatility. The experiments were conducted on a Windows server running Windows Server 2019 with Windows 10 as the 64-bit operating system. The central processing unit (CPU) employed was an Intel(R) Xeon(R) Platinum 8255C CPU @ 2.50 GHz 2.49 GHz. To expedite the training and testing processes, a graphics processing unit (GPU) was utilized. Specifically, NVIDIA Tesla T4 (8G) was chosen for its computational prowess. The deep learning environment was established using Conda, with the following specifications: Python version 3.9.0, torch version 1.10.0, torchvision version 0.11.0, Torchaudio version 0.10.0, and CUDA version 10.2. As shown in Table 2, these hardware and software configurations were thoughtfully selected to ensure optimal performance and compatibility throughout the training and testing phases of the models.
To comprehensively analyze the performance of YOLOv71 + COTN network model, two additional modules were introduced: simulated attention mechanism (simAM) and background object training (BOT). These modules played a crucial role in facilitating a longitudinal comparison during both the model training and testing phases. To provide a cross-sectional comparison, the dataset was utilized for training and testing Yolov5 model. The pre-training parameter settings configurations for each model are detailed in Table 3, which helps to establish a solid foundation for the subsequent experiments and evaluations.

4.2. Dataset Creation

In the testing phase, a variety of targets were selected, including coal blocks, gangue blocks, anchor rods, and I-beams. These targets were deliberately chosen to encompass a wide range of physical characteristics. Gangue blocks closely resemble coal blocks in their physical form, while anchor rods exhibit a slender shape and I-beams possess a flat surface. Furthermore, both anchor rods and I-beams are primarily composed of iron elements. To acquire the necessary images for testing, an industrial camera was employed. Each detection target was photographed, and the dataset was subsequently augmented using various techniques such as rotating, mirroring, panning, brightness adjusting, and Gaussian blurring. The efficacy of these augmentation techniques can be observed in Figure 9. The objective of this dataset augmentation process was to more accurately simulate the real-world conditions of each detection target within a coal mine environment and to enhance the dataset’s representativeness.
The targets in the images were manually annotated using the label open-source annotation tool. The dataset was organized into four categories: bolt, coal, gangue, and I-beam. Each target was assigned a numerical label: 0 for anchor, 1 for coal block, 2 for gangue block, and 3 for I-beam. The annotation process involved carefully marking the targets within the images to ensure accurate labeling based on human visual perception. The annotation results were saved in visual object classes (VOC) format, with the corresponding XML files stored in a predetermined folder. Great care was taken to label all visible targets while avoiding the inclusion of ambiguous ones. This approach was employed to prevent unannotated targets from being mistaken as negative samples, thus preserving the algorithm’s ability to effectively distinguish between positive and negative samples.

4.3. Analysis of Experimental Results

When the network structure of the YOLOv7 network model is fixed, the perceptual area of the network is predetermined. The resolution of the input image plays a critical role in determining the proportion of the perceptual area within the image. Higher-resolution images result in a lower percentage of the perceptual area, which reduces the effectiveness of capturing the local information for predicting objects at different scales. Consequently, this can lead to a decrease in detection accuracy. The variance in input image size has a significant impact on the model’s detection performance. The underlying network part often generates feature maps that are 10 times smaller than the original image. As a result, features related to small targets may not be adequately captured by the detection network, particularly for relatively small gangue blocks on the belt conveyor. To mitigate these challenges, the input image pairs in the experiments were resized to a standardized size of 640 × 640, as indicated in Table 4. It is important to note that the resized dataset images did not exceed the size of the original images in the dataset. This standardization of image size enhances the robustness of the detection model to variations in target size, to a certain extent.
The number of parameters in a network model refers to the total size, in bytes, of each network layer. It quantifies the amount of video memory occupied by the model. In this study, the spatial complexity of the model is evaluated based on the number of parameters, which provides insights into the model’s size and memory requirements. On the other hand, the time complexity of the model is evaluated using the amount of computation required. The computation volume measures the duration of the model’s detection process and is expressed as the number of floating-point operations per second. Additionally, the GPU runtime memory represents the amount of server space occupied by the model when running. It reflects the memory usage during the model’s execution on the GPU and is a crucial consideration for ensuring efficient utilization of server resources.
The YOLOv7 network model, derived from the YOLOv5 network model, exhibits a reduction in both network layers and parameters compared to the original YOLOv7 network model. Interestingly, despite having fewer layers and parameters, the YOLOv5 network model occupies an additional 3.93 G of space when compared to the YOLOv7 network model. Specifically, the YOLOv71 network model boasts 22 fewer layers and 1,701,083 fewer parameters. Furthermore, the memory usage by GPU is reduced by 0.15 G. To further analyze the operating parameters, additional modules such as simAM, BOT, and COTN were incorporated into the models. Among these modules, the COTN module demonstrated superior performance compared to the simAM and BOT modules.
Table 5 presents the detection results for each detection target in the new improved model, including accuracy, recall, and average accuracy mean. Upon analyzing the results in Table 4, it is evident that anchor rods achieve higher identification accuracy compared to coal and gangue targets. This difference in performance can be attributed to the similarity in shape and appearance between gangue and coal, posing challenges for accurate discrimination. Additionally, I-beam, characterized by a larger surface area is larger and a tendency to accumulate coal dust, visually resembles coal, thereby impeding precise recognition. Consequently, the recognition performance for gangue and I-beams is relatively poorer when compared to other targets.
The YOLOv71 + simAM network model demonstrates improved recognition specifically for coal when compared to the YOLOv71 network model. However, the overall detection results for each target in the YOLOv71 network model are lower. On the other hand, the YOLOv71 + BOT network model exhibits enhanced recognition capabilities specifically for anchor rods following the model’s improvements. Notably, the YOLOv71 + COTN network model demonstrates improvements in terms of accuracy, recall, and average accuracy for each detection target when compared to the YOLOv71 model.
Box metric represents the mean value of the loss function, where a smaller value indicates more accurate detection, Objectiveness metric represents the mean value of the target detection loss, with a smaller value indicating more accurate detection of targets. The classification metric represents the mean value of the classification loss, where a smaller value indicates a more accurate classification. The comparison results presented in Figure 10 illustrate that YOLOv71 + COTN network model achieves lower detection values on the training set compared to the other models. The data curve exhibits a relatively flat trend, with a general decrease in values. These findings indicate that the YOLOv71 + COTN model outperforms the other three models in terms of detection accuracy.
The precision metric represents the accuracy of positive predictions, while the recall metric represents the rate of correctly identified positive instances. The true metric represents the accuracy of the positive metric, representing the number of positive samples recalled, which describes the number of positive samples correctly recalled by the classifier. It provides insights into how well the classifier identifies positive instances. The comparison results depicted in Figure 11a,b show that YOLOv71 + COTN network model exhibits higher accuracy and recall values compared to the other models during training. These results suggest that YOLOv71 + COTN model achieves better recognition performance, with improved accuracy in identifying positive instances and a higher recall rate for true positive examples.
mAP0.5 metric represents the mean accuracy values at a threshold of 0.5. mAP0.5:0.75 metric represents the mean accuracy values for thresholds ranging from 0.5 to 0.75, with an interval of 0.05. Upon comparing the information presented in Figure 11c,d, it is evident that the YOLOv71 + COTN network model achieves the highest mean accuracy for the detection of gangue and foreign matters at the specified threshold.
The simAM, BOT, and COTN modules were integrated into the YOLOv71 network model to conduct ablation experiments and evaluate their impact on performance. The test results for each network model are shown in Table 6. It is evident that the network model with the additional module exhibits notable improvements in accuracy and recall compared to the YOLOv7 network model. Among the three modules, the COTN module is the most effective, achieving an impressive identification accuracy of 91.3%.
Figure 12 provides a visual representation of the detection results obtained by each network model for the same target. The analysis of the model output images reveals a clear pattern: As the network model undergoes continuous improvement, there is a noticeable enhancement in the accuracy and reliability of target detection. Of particular note is the performance of the new YOLOv71 + COTN model, which achieves the highest confirmation rate for the target, reaching an impressive value of 90% when the target is clearly visible.
Figure 13 represents the model’s performance in identifying targets during the testing phase. Each target is assigned a numerical label: 0 for anchor, 1 for coal, 2 for gangue, and 3 for I-beam. Figure 13 reveals that both YOLOv7 and YOLOv71 network model exhibit recognition errors in the randomly outputted detection images. Specifically, the YOLOv7 network model has two instances of recognition errors, while the YOLOv71 network model has one instance of recognition error. However, with the incorporation of the new module, the network model demonstrates a notable absence of recognition errors.

5. Conclusions

This study presents a novel approach for detecting gangue and foreign matter in coal using an improved YOLOv7 network model. The improved YOLOv7 network model is specifically tailored to achieve accurate and reliable detection of gangue and foreign matter in coal samples. To enhance the performance of YOLOv7 network model, several key improvements were introduced. Firstly, the number of convolutional layers in the backbone was halved, resulting in a more streamlined and efficient detection process. This reduction in convolutional layers not only accelerated the detection speed but also enhanced the overall detection efficiency for gangue and foreign matter. Furthermore, small-size detection layers were incorporated into the head. This addition aimed to enhance the model’s ability to detect and classify small targets more effectively. By specifically focusing on small-sized objects, the model could improve its accuracy in identifying and localizing these challenging targets. Additionally, the COTN module further enhanced the model’s detection accuracy. This module leverages innovative techniques to refine the model’s feature representation, ultimately improving its ability to discriminate between different classes of gangue and foreign matter. To calculate the overlap between the predicted and real frames, DIoU loss border regression loss function was employed. This loss function provided a comprehensive measure of the intersection between the predicted and ground truth bounding boxes. By considering both the spatial distance and proportional characteristics of the boxes, the model could effectively identify gangue and foreign matter based on this calculated overlap. To improve recognition accuracy, a dual-path attention mechanism was integrated into the model. This mechanism allowed the model to selectively focus on relevant features while suppressing irrelevant or distracting information. By effectively attending to important regions of the input, the model achieved enhanced recognition accuracy and robustness.
This research proposes an improved the YOLOv7 network model specifically designed for the identification of gangue and foreign matter in coal, with a focus on its application in gangue sorting robots. The proposed model, YOLOv71 + COTN, was trained and evaluated using a dedicated dataset to ensure accurate identification of gangue and foreign matter. The experimental results revealed significant improvements compared to YOLOv71 + COTN network model approach. This proposed network model demonstrated notable enhancements across various performance metrics. Precision was improved by 3.97%, recall was increased by 4.4%, and mAP0.5 was improved by 4.5%, compared to the YOLOv7 network model approach. By reducing the number of parameters, the model’s efficiency was optimized, resulting in reduced memory requirements on GPU during operation. This reduction in memory consumption contributes to improved gangue selection speed and recognition accuracy, further enhancing the overall performance of the model. As a result, the YOLOv71 + COTN network model is well-suited for integration in belt conveyor gangue sorting robots.

Author Contributions

Conceptualization, D.Y. and C.M.; methodology, D.Y. and X.L.; software, Y.L.; validation, Y.W., Y.Z., and D.Y.; investigation, D.Y.; resources, D.Y.; data curation, D.Y. and Y.W.; writing—original draft preparation, D.Y.; writing—review and editing, D.Y. and C.M.; supervision, C.M.; project administration, C.M.; funding acquisition, C.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China under Grant No. NSFC51274150, and the Key R&D Programme Science and Technology Support Projects of Tianjin under No. 18YFZCGX00930.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the writing author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, P.; Tian, H.; Cao, X.; Qiao, X.; Gong, L.; Duan, X.; Qiu, Y.; Su, Y. Pick–and–Place Trajectory Planning and Robust Adaptive Fuzzy Tracking Control for Cable–Based Gangue–Sorting Robots with Model Uncertainties and External Disturbances. Machines 2022, 10, 714. [Google Scholar] [CrossRef]
  2. Wang, P.; Ma, H.; Zhang, Y.; Cao, X.; Wu, X.; Wei, X.; Zhou, W. Trajectory Planning for Coal Gangue Sorting Robot Tracking Fast-Mass Target under Multiple Constraints. Sensors 2023, 23, 4412. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, P.; Ma, H.; Cao, X.; Zhang, X.; Duan, X.; Nie, Z. Minimum Dynamic Cable Tension Workspace Generation Techniques and Cable Tension Sensitivity Analysis Methods for Cable-Suspended Gangue-Sorting Robots. Machines 2023, 11, 338. [Google Scholar] [CrossRef]
  4. Ma, H.; Wei, X.; Wang, P.; Zhang, Y.; Cao, X.; Zhou, W. Multi-Arm Global Cooperative Coal Gangue Sorting Method Based on Improved Hungarian Algorithm. Sensors 2022, 22, 7987. [Google Scholar] [CrossRef] [PubMed]
  5. Yan, P.; Sun, Q.; Yin, N.; Hua, L.; Shang, S.; Zhang, C. Detection of coal and gangue based on improved YOLOv5. 1 which embedded scSE module. Measurement 2022, 188, 110530. [Google Scholar] [CrossRef]
  6. Gao, R.; Sun, Z.; Li, W.; Pei, L.; Xiao, L. Automatic coal and gangue segmentation using u-net based fully convolutional networks. Energies 2020, 13, 829. [Google Scholar] [CrossRef]
  7. Li, D.; Zhang, Z.; Xu, Z.; Meng, G.; Li, Z.; Chen, S. An image-based hierarchical deep learning framework for coal and gangue detection. IEEE Access 2019, 7, 184686–184699. [Google Scholar] [CrossRef]
  8. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  9. Szegedy, C.; Vanhoucke, V.; Loffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  10. Szegedy, C.; Loffe, S.; Vanhoucke, V.; Alemi, A. Inception—v4, inception—Resnet and the impact of residual connections on learning. In Proceedings of the National Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2016; pp. 4278–4284. [Google Scholar]
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  12. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part IV 14. Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 630–645. [Google Scholar]
  13. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Lin, H.; Zhang, Z.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. Resnest: Split-attention networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 2736–2746. [Google Scholar]
  14. Wu, X.; Sahoo, D.; Hoi, S. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef]
  15. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  16. Dong, C.; Loy, C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. In IEEE Transactions on Pattern Analysis and Machine Intelligence; IEEE: Piscataway, NJ, USA, 2016; Volume 38, pp. 295–307. [Google Scholar]
  17. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  18. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  19. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In IEEE Transactions on Pattern Analysis and Machine Intelligence; IEEE: Piscataway, NJ, USA, 2017; Volume 39, pp. 1137–1149. [Google Scholar]
  20. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  21. Fu, C.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A. Dssd: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  22. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  23. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  24. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  25. Bochkovskiy, A.; Wang, C.; Liao, H. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  26. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar]
  28. Woo, S.; Park, J.; Lee, J.; Kweon, I. CBAM: Convolutional block attention module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Munich, Germany, 2018; pp. 3–19. [Google Scholar]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  30. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  31. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  32. Dai, S.; Zheng, X.; Wang, X.; Finkelman, R.B.; Jiang, Y.; Ren, D.; Yan, X.; Zhou, Y. Stone coal in China: A review. Int. Geol. Rev. 2018, 60, 736–753. [Google Scholar] [CrossRef]
  33. Zhao, M.; Liu, H.; Liu, C.; Li, X.; Li, F.; Yang, X.; Ma, Q. Spatial Effect Analysis of Coal and Gangue Recognition Detector Based on Natural Gamma Ray Method. Nat. Resour. Res. 2022, 31, 953–969. [Google Scholar] [CrossRef]
  34. Wang, C.; Bochkovskiy, A.; Liao, H. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  35. Zhang, J.; Zhang, J.; Zhou, K.; Zhang, Y.; Chen, H.; Yan, X. An Improved YOLOv5-Based Underwater Object-Detection Framework. Sensors 2023, 23, 3693. [Google Scholar] [CrossRef] [PubMed]
  36. Cao, X.; Zhang, Y.; Lang, S.; Gong, Y. Swin-Transformer-Based YOLOv5 for Small-Object Detection in Remote Sensing Images. Sensors 2023, 23, 3634. [Google Scholar] [CrossRef] [PubMed]
  37. Zhang, S.; Xie, M. Real-Time Recognition and Localization Based on Improved YOLOv5s for Robot’s Picking Clustered Fruits of Chilies. Sensors 2023, 23, 3408. [Google Scholar] [CrossRef] [PubMed]
Figure 1. YOLOv7 model structure.
Figure 1. YOLOv7 model structure.
Sensors 23 05140 g001
Figure 2. YOLOv71 + COTN model structure.
Figure 2. YOLOv71 + COTN model structure.
Sensors 23 05140 g002
Figure 3. Testing methodology process.
Figure 3. Testing methodology process.
Sensors 23 05140 g003
Figure 4. Re-parameterization components.
Figure 4. Re-parameterization components.
Sensors 23 05140 g004
Figure 5. CONT module.
Figure 5. CONT module.
Sensors 23 05140 g005
Figure 6. Data calculation for predicted and true boxes for each detection target. (a) Coal; (b) gangue; (c) bolt; (d) I-bean.
Figure 6. Data calculation for predicted and true boxes for each detection target. (a) Coal; (b) gangue; (c) bolt; (d) I-bean.
Sensors 23 05140 g006
Figure 7. DIoU loss diagram (example of gangue).
Figure 7. DIoU loss diagram (example of gangue).
Sensors 23 05140 g007
Figure 8. Feature extraction map (example of gangue).
Figure 8. Feature extraction map (example of gangue).
Sensors 23 05140 g008
Figure 9. Dataset augmentation process.
Figure 9. Dataset augmentation process.
Sensors 23 05140 g009
Figure 10. (a) is a plot of the results of each network model train box, (b) is a plot of the results of each network model train objectness, (c) is a plot of the results of each network model train classification.
Figure 10. (a) is a plot of the results of each network model train box, (b) is a plot of the results of each network model train objectness, (c) is a plot of the results of each network model train classification.
Sensors 23 05140 g010
Figure 11. (a) is the accuracy of each network model run, (b) is the recall of each network model run, (c) is the mAP0.5 of each network model run, (d) is the mAP0.5:0.75 of each network model run.
Figure 11. (a) is the accuracy of each network model run, (b) is the recall of each network model run, (c) is the mAP0.5 of each network model run, (d) is the mAP0.5:0.75 of each network model run.
Sensors 23 05140 g011
Figure 12. Detection of each model for the same target. From top to bottom, (a) is coal, (b) is gangue, (c) is bolt, (d) is I-beam. From left to right, they are Yolov7 network model, YOLOv71 network model, YOLOv71 + simAM network model, YOLOv71 + BOT network model, and YOLOv71 + COTN network model.
Figure 12. Detection of each model for the same target. From top to bottom, (a) is coal, (b) is gangue, (c) is bolt, (d) is I-beam. From left to right, they are Yolov7 network model, YOLOv71 network model, YOLOv71 + simAM network model, YOLOv71 + BOT network model, and YOLOv71 + COTN network model.
Sensors 23 05140 g012
Figure 13. Test results for each model. From top to bottom, they are the Yolov7 model, the YOLOv71 model, the YOLOv71 + simAM model, the YOLOv71 + BotNet model, and the YOLOv71 + COTN model.
Figure 13. Test results for each model. From top to bottom, they are the Yolov7 model, the YOLOv71 model, the YOLOv71 + simAM model, the YOLOv71 + BotNet model, and the YOLOv71 + COTN model.
Sensors 23 05140 g013
Table 1. Yolov7 network model parameters and Yolov71 network model parameters.
Table 1. Yolov7 network model parameters and Yolov71 network model parameters.
ModelLayersTime (h)
YOLOv741548.5
YOLOv7139345.2
Table 2. Model runtime environment configuration.
Table 2. Model runtime environment configuration.
DesignationVersion
GPUNVIDIA Tesla T4(8G)
CUDA10.2
Python3.9.0
Torch1.10.0
Torchvision0.11.0
Torchaudio0.10.0
Table 3. Model runtime parameters and module configuration.
Table 3. Model runtime parameters and module configuration.
ModelEpochWeight (M)Rate of LearningModule
YOLOv5100014.40.01
YOLOv71000720.01
YOLOv711000720.01
YOLOv71 + simAM1000720.01simAM
YOLOv71 + BOT1000720.01BOT
YOLOv71 + COTN1000720.01COTN
Table 4. Operating parameters for each model.
Table 4. Operating parameters for each model.
ModelLayersParametersSize GFLOPS (G)Gpu_mem (G)
YOLOv51917,260,490640 × 64016.97.13
YOLOv741537,207,344640 × 640105.13.2
YOLOv7139335,506,261640 × 64075.53.05
YOLOv71 + simAM34328,967,472640 × 64035.33.11
YOLOv71 + BOT2815,971,720640 × 64014.81.87
YOLOv71 + COTN3066,636,424640 × 64014.31.26
Table 5. Test results for each test target.
Table 5. Test results for each test target.
ModelObjectPrecisionRecallAP (0.5)AP (0.5:0.75)
YOLOv71Bolt90%87.8%90.7%85.7%
Coal88.7%88.9%89.7%83.7%
Gangue85.8%85.8%85.7%80.8%
I-beam84.8%85.1%84.5%79.8%
YOLOv71 + simAMBolt89.3%87.5%89.2%84.3%
Coal89.6%88.9%89.7%84.4%
Gangue87.4%84.6%84.6%80.3%
I-beam86.4%84.3%84.2%79.3%
YOLOv71 + BOTBolt90%90%89.5%84.9%
Coal89.7%87.9%88.6%83.8%
Gangue87.5%85.4%85.7%81.6%
I-beam86.6%85.6%86.9%80.6%
YOLOv71 + COTNBolt92%92%91.6%86.4%
Coal91.8%91.7%91.1%85.7%
Gangue90.1%88.1%89.3%83.7%
I-beam86.9%87.2%87.5%82.7%
Table 6. Test results for each network model.
Table 6. Test results for each network model.
ModelsimAMBOTCOTNPrecisionRecallMap0.5Map0.5:0.75
YOLOv7 87%86.2%87.4%81.7%
YOLOv71 87.33%86.9%87.65%82.5%
YOLOv71 + simAM 88.18%86.33%86.93%82.08%
YOLOv71 + BOT 88.45%87.23%87.68%82.73%
YOLOv71 + COTN 91.3%90.6%91.9%85.4%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, D.; Miao, C.; Li, X.; Liu, Y.; Wang, Y.; Zheng, Y. Improved YOLOv7 Network Model for Gangue Selection Robot for Gangue and Foreign Matter Detection in Coal. Sensors 2023, 23, 5140. https://doi.org/10.3390/s23115140

AMA Style

Yang D, Miao C, Li X, Liu Y, Wang Y, Zheng Y. Improved YOLOv7 Network Model for Gangue Selection Robot for Gangue and Foreign Matter Detection in Coal. Sensors. 2023; 23(11):5140. https://doi.org/10.3390/s23115140

Chicago/Turabian Style

Yang, Dengjie, Changyun Miao, Xianguo Li, Yi Liu, Yimin Wang, and Yao Zheng. 2023. "Improved YOLOv7 Network Model for Gangue Selection Robot for Gangue and Foreign Matter Detection in Coal" Sensors 23, no. 11: 5140. https://doi.org/10.3390/s23115140

APA Style

Yang, D., Miao, C., Li, X., Liu, Y., Wang, Y., & Zheng, Y. (2023). Improved YOLOv7 Network Model for Gangue Selection Robot for Gangue and Foreign Matter Detection in Coal. Sensors, 23(11), 5140. https://doi.org/10.3390/s23115140

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop