You are currently viewing a new version of our website. To view the old version click .
Agriculture
  • Article
  • Open Access

20 September 2024

A High-Accuracy Contour Segmentation and Reconstruction of a Dense Cluster of Mushrooms Based on Improved SOLOv2

,
and
1
School of Intelligent Manufacturing and Control Engineering, Shanghai Polytechnic University, Shanghai 201209, China
2
College of Mechanical and Electronic Engineering, Shandong Agricultural University, Taian 271018, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Intelligent Agricultural Machinery and Robots: Embracing Technological Advancements for a Sustainable and Highly Efficient Agricultural Future

Abstract

This study addresses challenges related to imprecise edge segmentation and low center point accuracy, particularly when mushrooms are heavily occluded or deformed within dense clusters. A high-precision mushroom contour segmentation algorithm is proposed that builds upon the improved SOLOv2, along with a contour reconstruction method using instance segmentation masks. The enhanced segmentation algorithm, PR-SOLOv2, incorporates the PointRend module during the up-sampling stage, introducing fine features and enhancing segmentation details. This addresses the difficulty of accurately segmenting densely overlapping mushrooms. Furthermore, a contour reconstruction method based on the PR-SOLOv2 instance segmentation mask is presented. This approach accurately segments mushrooms, extracts individual mushroom masks and their contour data, and classifies reconstruction contours based on average curvature and length. Regular contours are fitted using least-squares ellipses, while irregular ones are reconstructed by extracting the longest sub-contour from the original irregular contour based on its corners. Experimental results demonstrate strong generalization and superior performance in contour segmentation and reconstruction, particularly for densely clustered mushrooms in complex environments. The proposed approach achieves a 93.04% segmentation accuracy and a 98.13% successful segmentation rate, surpassing Mask RCNN and YOLACT by approximately 10%. The center point positioning accuracy of mushrooms is 0.3%. This method better meets the high positioning requirements for efficient and non-destructive picking of densely clustered mushrooms.

1. Introduction

Mushrooms, recognized as one of the edible fungi with the most extensive cultivated areas and found in some of the world’s largest cultivation countries, exhibit a significant output. In China alone, they boast an annual production of nearly 2 million tonnes. Currently, the mainstream production mode of mushrooms in developed countries is factory production, and most of the processes have realized mechanization and automation. Only picking still relies on manual labor [1,2]. Due to the decline in the rural labor force, manual mushroom picking is experiencing a significant shortage of labor, posing a serious obstacle to the industry’s continued growth. Scholars have researched the autonomous picking technology of mushrooms and achieved phased development. The relevant research has developed from the exploration stage [3,4], limited to the laboratory under ideal conditions, to the field experiment test stage [5,6] in the actual factory mushroom house. However, due to the heterogeneity of the realistic growing environment of mushroom houses and the randomness of fruiting body growth, it is challenging to achieve efficient and nondestructive picking, especially for densely overlapping mushrooms [7]. Clusters of mushrooms has become the main reason that prevents the practical application of mushroom havesting robots. Therefore, identifying and locating densely overlapping mushrooms with high precision is one of the keys to realizing more efficient, nondestructive mushroom picking.
Mushrooms are a kind of natural high-cluster crop. It is effortless for fruiting bodies to grow together densely and overlappingly. Due to the limited space for dense growth, the mushroom shape can easily vary from round to oval or irregular, or vertical to inclined. Many mushrooms are seriously overlapping, which makes it difficult to segment and reconstruct the mushroom contour with high accuracy. For example, since densely growing mushrooms are easily squeezed and stacked, the contour segmentation between them will become complex, and it is not easy to obtain accurate edge contours of mushrooms, as shown in Figure 1, Mushrooms 1, 2, 3, and 4. For another example, densely growing mushrooms are prone to appearing heavily obscured, such as mushrooms 9 and 10 in the figure. Their shapes are no longer round or elliptical but are missing and irregular. It is often difficult to restore the contour based on these missing or irregular shapes to the actual contour, such as contour 9. The deviation between the generated and accurate contours is significant, resulting in a considerable variation between the identified and actual mushroom center points. In addition, a few small mushrooms are easily overlooked, such as Mushroom 5, 6, 7, and 8. Therefore, to achieve efficient and lossless picking of densely overlapping mushrooms, it is necessary to study further and improve the segmentation and contour reconstruction algorithms of densely overlapping mushrooms to obtain high-precision positioning.
Figure 1. Main problems in segmentation and localization of densely overlapping mushrooms.
The existing system faces several challenges in accurately segmenting and recon-structing mushroom contours:
Dense Overlapping Growth: Generally, mushrooms grow in clusters with densely overlapping caps. Conventional segmentation methods encounter difficulties in distinguishing individual mushrooms within these clusters, resulting in inaccurate boundary delineation.
Deformation and Irregular Shapes: Due to mutual adhesion, extrusion, or varying growth conditions, mushrooms can exhibit irregular shapes and deformities. Most of the methods, fitting with circles only, are not very accurate in fitting the contours of irregular shapes and deformities.
Incomplete and Occluded Contours: When mushrooms are partially covered by other mushrooms, the segmentation process may yield incomplete or occluded contours, making it challenging to obtain an accurate representation of each mushroom’s shape.
Variable Lighting and Impurities: Non-uniform lighting conditions and the presence of surface impurities, such as soil and lesions, can adversely affect the quality of mushroom images, leading to additional segmentation difficulties.
To address these problems, the proposed work makes significant contributions:
(1)
The PR-SOLOv2 instance segmentation model is introduced that leverages PointRend to improve segmentation accuracy. This model effectively handles dense overlapping mushrooms and enhances boundary accuracy, achieving a segmentation accuracy of 93.04%.
(2)
The work proposes a novel contour reconstruction method based on the curvature and length of mask contours obtained through instance segmentation. This approach enables the reconstruction of mushroom contours with high precision.
(3)
The work introduces a least-squares ellipse fitting technique for mushrooms with regular shapes. For irregularly shaped or occluded mushrooms, a corner-based segmentation method is applied to accurately reconstruct contours. This classification and fitting approach significantly improves the quality of contour reconstruction.
(4)
The proposed work demonstrates a successful segmentation rate of 98.92%. This high success rate shows the effectiveness of the proposed methods.
The manuscript is structured as follows: Section 2 reviews existing research and methodologies related to mushroom segmentation and contour reconstruction. Section 3 elucidates mushroom instance segmentation dataset construction. Section 4 introduces the PR-SOLOv2 instance segmentation model. The contour construction of mushrooms using the proposed approach is presented in Section 5. It explains the classification and fitting approach based on curvature and length. Section 6 presents the experimental results of the proposed methods and compares them to other segmentation algorithms, highlighting the accuracy and effectiveness of PR-SOLOv2. Section 7 summarizes the major contributions of the research, emphasizing the improvements in mushroom segmentation and contour reconstruction.

3. Mushroom Instance Segmentation Dataset Construction

This paper takes mushrooms as the research object and collects images of mushrooms growing in the mushroom house of Shanghai Lianzhong Edible Fungus Professional Cooperative to construct the dataset. Mushrooms are raised in factory production environments and standard culture racks. The culture racks are 31 m long, 1440 mm wide, and 6 storeys high, with the storeys spaced 600 mm apart. The temperature, humidity, carbon dioxide concentration, and light in the mushroom room are intelligently controlled by an intelligent control system under conditions suitable for the growth of mushrooms, as shown in Figure 2. Three MV-GE300C-T GigE color industrial cameras, with a focal length of 4 mm, a pixel depth of 10 bits of colour and image resolution of 1280*720 is used, were installed transversely on the picking robot to traverse each layer of mushrooms and capture mushroom images.
Figure 2. Field image of mushroom image acquisition.
Mushrooms quickly form dense clusters, especially the mushrooms in the early picking stage, whose growth is particularly lush, resulting in mutual extrusion, shielding, staggering, and other phenomena among mushrooms. In addition, the fruiting bodies of mushrooms may appear on the surface of scales during the growth process or the browning phenomenon due to poor anti-browning ability, thus affecting the shape of mushrooms. Therefore, the construction of the data set should fully consider these differences in mushroom growth shape and posture. In this paper, a total of 480 images of eight typical forms of mushrooms were collected, and 60 shots were gathered for each category, as shown in Figure 3.
Figure 3. Datasets of different forms of mushrooms. (a) With scales on caps; (b) With brown spots on caps; (c) With soil on caps; (d) Serious interleaving, clumping, and adhesion; (e) Large differences in size and height; (f) Adhesion and extrusion deformation between mushrooms; (g) Heavily tilted to expose its stalk; (h) With mechanical damage on caps.
This paper also considers the images under different light source conditions when constructing the dataset. Specific composition (Figure 4): 160 images with uniform light and high brightness collected under the top light source, 160 pictures with uneven light under the side light source, and 160 images with low brightness under the dark light source.
Figure 4. Datasets for different light sources: (a) Top light source; (b) Side light source; (c) Dark light source.
The images in this paper are labeled with labelme. To ensure that the labeled polygons try to fit the real contours of the mushrooms, the number of basic labeling points for each mushroom example is stipulated, following the definition of a COCO dataset for s, m, l to divide the mushroom model: s-type mushrooms at least 10 labeling points, m-type at least 20 labeling points, and l-type mushrooms at least 25 points, to improve the degree of labeling and fit to the real contours of the mushrooms. After labeling all the initial images to obtain the initial data set, we used data enhancement to process the images to improve the generalization ability and robustness of the model. Specific image data enhancement methods included: image brightness change, horizontal flipping of the image, vertical mirroring, and random rotation. Finally, 1920 mushroom images obtained after performing the image enhancement process were randomly divided into training and validation sets in a ratio of 7:3.

4. Research on High-Accuracy Mushroom Segmentation

The SOLOv2 algorithm, shown in Figure 5, was used to segment the mushrooms first.
Figure 5. SOLOv2 network structure.
The segmentation results are as follows: for the mushrooms with a general aggregation degree and good growth form, the segmentation effect is good. The recognition rate of the mushrooms in Figure 6a is high, especially since some tiny mushrooms are segmented. The mask boundary fits well with the contour of the actual mushroom. However, the segmentation effect will be affected for mushrooms with a high aggregation degree and complex growth morphology, resulting in a rough mask and inaccurate segmentation of mushroom edges, as shown in Figure 6b. Especially in the densely overlapping area of mushrooms, the positioning accuracy of the mushroom picking center will be affected. Therefore, SOLOv2 needs to be improved to obtain more precise and accurate segmentation of mask edges.
Figure 6. Result of SOLOv2 segmented mushrooms: (a) Ordinary aggregation; (b) Complex polymerization.
The PointRend neural network can optimize object edge segmentation, treat segmentation of an image as a rendering problem, and perform the prediction of image segmentation based on an adaptive point selection iterative subdivision algorithm [58]. The structure of the network is shown in Figure 7. A lightweight prediction head generates a rough prediction mask, and bilinear interpolation is utilized for upsampling the rough prediction mask to obtain a prediction mask with twice the resolution. N points distributed on the edge are selected, and point-by-point features are constructed on the points chosen by combining fine-grained features and coarse prediction features. Finally, a simple multi-layer perceptron is used to predict these points. All points share the weight of the multi-layer perceptron. The multi-layer perceptron can be trained by standard task-specific segmental losses and finally output the classification label of each point. This is iterated until the mask at the desired resolution is upsampled. Therefore, this paper uses the PointRend network structure, which can solve the edge segmentation accuracy that is not high enough to improve the SOLOv2 network. In SOLOv2, to obtain the semantic information of the segmentation target, multiple convolutions and pooling operations on the original image will reduce the image resolution. Then, if the image with high semantic and low key is directly upsampled to be consistent with the size of the original image, information in specific areas will be lost, resulting in rough segmentation results. In addition, the method of predicting each pixel in a regular grid and inferring the instance category to which it belongs has the conditions of over-sampling and under-sampling. In particular, too many sampling points in low-frequency areas belonging to the same object in the image leads to over-sampling, and the under-sampling of such high-frequency regions close to the object boundary in the picture makes the segmented border too smooth and unreal. This will inevitably lead to rough segmentation results, especially at the edge of the segmentation object.
Figure 7. PointRend network structure.
Therefore, this paper combines the SOLOv2 network structure with the PointRend neural network module to build an improved SOLOv2 algorithm, which is called the PR-SOLOv2 algorithm in this paper. Its network structure is shown in Figure 8. After the Fully Convolution Network (FCN) of SOLOv2’s feature extraction backbone network, the PointRend module is introduced before the mask prediction branch to improve the image up-sampling method. To solve the information loss in local areas and render fine segmentation details, the delicate features of the second layer of the FCN network are introduced, combined with the rough features of the fourth layer, and sent to the taught PointRend module.
Figure 8. PR-SOLOv2 network structure.
The top-N feature pixels are extracted adaptively from the feature map of low spatial resolution and the corresponding fine feature map of high spatial resolution. The multi-layer perceptron is used to iterate and optimize each point continuously, and the instance is connected to the segmentation prediction branch to predict the instance mask. By this method, low-level and high-level features are combined layer by layer. The output image is iteratively rendered from coarse–fine to coarse–fine to solve the boundary inauthenticity caused by the over-sampling of the smooth region and under-sampling of the boundary region, which can improve the misjudgment of boundary pixels caused by the original network structure and improve the accuracy of the image segmentation edge.
The model’s total loss function L is composed of the loss L cate generated by instance classification and the loss L Mask generated by mask prediction. λ is the coefficient that balances the two losses, and the calculation is shown in Equation (1).
L = L cate + λ L Mask
L Mask = 1 N pos 1 { P i , j * > 0 } d Mask , i , j M
d Mask = 1 D p , q = 1 2 x , y p x , y q x , y x , y p x , y 2 + x , y q x , y 2
where L c a t e is the Focal Loss function for classification [59] for the calculation of L M a s k (Equation (2)). S is the number of grid units divided into the input image. i and j respectively represent the subscripts of indexes from left to right and from top to bottom on M × M grids on the input image, traversing all grids on the picture; Npos is the number of positive samples; P * represents the instance category. 1 { P i , j * > 0 } is the indicator function; if P i , j * > 0 , the indicator function value is 1, otherwise it is 0; d M a s k is the Dice Loss function, and its definition is shown in Equation (3), where D is the Dice coefficient, which is used to measure the similarity of two sets. p and q represent the predicted Dice Loss mask and the natural mask, respectively, and p x , y and q x , y are the pixel values of the point x , y of the predicted mask and the natural mask on the image, respectively.

5. Contour Reconstruction of Mushroom

In the process of mushroom growth, due to its natural cluster growth characteristics, it is common for mushrooms to overlap and extrude each other, leading to the pileus’s deformation. The PR-SOLOv2 instance segmentation model can more accurately separate them from the complex situations of adhesion between mushrooms, inclination, soil shielding on the surface, and uneven light, and obtain their masks. However, for mushrooms that are occluded or deformed due to aggregation growth, the mask obtained according to the image is only part of the shape of the mushroom. If the mask center point is directly extracted from the mushroom center point, there will be a significant error.
Therefore, given the above non-uniform and irregular mushroom growth shapes, especially the severe deformation and shape loss of densely overlapping mushroom contours, to better restore the actual mushroom contour and improve the center point positioning accuracy, this paper proposes a mushroom contour classification reconstruction method based on PR-SOLOv2 segmentation mask contour curvature and length, as shown in Figure 9.
Figure 9. Flow chart of high-precision contour reconstruction based on PR-SOLOv2.
In this method, firstly, the PR-SOLOv2 algorithm is used to segment mushrooms and obtain their masks. Then, the mushroom mask contour based on the mask is extracted. Next, mushrooms are classified into regular and irregular shapes according to the concave defect length of the extracted contour. A mushroom with a regular shape is fitted with the least square ellipse. In contrast, a mushroom with an irregular shape and difficult contour fitting, seriously occluded or squeezed, is reconstructed with the longest contour extraction and classification reconstruction method based on corner segmentation.

5.1. Reconstruction of Regular Contours

The contour is considered regular with no or only shallow concave defects.
Firstly, the convex hull method is used to find out if there are any concave defects in the contour. Contours without concave defects are considered regular contours. In addition, although some mushrooms have concave defects, the chord length ( I A r c ) and arch height ( h A r c ) of the concave contour are relatively small compared to the entire mushroom length (r L c o n t o u r ), so such concave defects can be ignored. Mushroom contours that meet the above conditions can still be considered regular contours. Due to the varying sizes of mushrooms, to better determine the impact of concave contours on the fitting of mushrooms of different sizes, mushrooms are divided into three types based on the contour length ( L c o n t o u r ): small, medium, and large. Different boundary values are used to determine mushrooms of different sizes. A mushroom with a contour length in the range [0,90) is defined as a small mushroom, [90,180) as a medium mushroom, and above 180 as a large mushroom. When I A r c is less than 46 or h A r c is less than 10 for a small mushroom, a concave contour can be ignored; when I A r c is less than 69 or hArc is less than 30 for a medium mushroom, a concave contour can be ignored; when I A r c is less than 89 or h A r c is less than 40 for a large mushroom, a concave contour can be ignored. After the above judgment, if the concavities of the mushroom contour can be ignored, then this mushroom is defined as having a regular mushroom contour. Otherwise, it has an irregular contour.
For mushrooms with regular contours, they can be fitted by the least-squares ellipse method: Take the n contour points above it as P i x i , y i i = 1,2 , , n , perform fitting operation according to the least-squares ellipse fitting objective function, calculate the ellipse position parameters and shape parameters, and reconstruct the contour of the bisporus mushroom with a smooth contour and regular shape.

5.2. Reconstruction of Irregular Contours

After segmenting the irregular contour of the overlapping mushroom, the obtained contour contains the contour fragment of the overlapping part of the other mushrooms covered. Suppose it is not processed and allowed to participate in the fitting together with the effective contour, such as the least-squares ellipse fitting. In that case, the fitted contour will differ from the actual contour. And the overlapped part cannot be well restored to the actual contour.
Given the above situation, this paper proposes the longest contour extraction and classification reconstruction approach relying on corner segmentation. The specific process is as follows:
(1)
Detect corner points at abrupt changes in contour shape. If the concave contours cannot be ignored, define its start and end point as corner points;
(2)
By corner coordinates, the coordinate line between adjacent corner coordinates is regarded as a sub-contour segment, and the whole mushroom contour is divided into N sub-contour segments;
(3)
Calculate the length of each sub-contour segment and choose the longest sub-contour segment;
(4)
Based on the longest sub-contour, the mushroom contour is classified and fitted according to the difference in arch curvature and length of the longest sub-contour.
(5)
Calculate the arch curvature C of the contour (Equation (4)).
C = 1 R
R = 0.5 * z 2 h A r c + h A r c
z = x n x 1 2 + y n y 1 2
h A r c = max ( i S i * 2 z )
S i = p i p i z p i v i p i m i , i = 1 , n
where R is the arcuate curvature radius of the longest sub-contour, vi and mi are, respectively, the length of both sides of the inscribed triangle formed by any point i (i = 1,2,…,n) on the longest sub-contour and its two upper endpoints, I A r c is the chord length of the longest sub-contour and is also the bottom length of all inscribed triangles in the longest sub-contour, p i is the half circumference of the inscribed triangle, S i is the area of the inscribed triangle, h A r c is the height of the longest arch in the inscribed triangle on the longest sub-contour.
For the longest sub-contour with an arch curvature exceeding 0.6 and a length surpassing 30, the least-squares ellipse is employed to fit the contour. However, when dealing with the longest sub-contour having an arch curvature less than 0.6 or a length below 100, using the least-squares ellipse fitting method would result in a significant deviation of the fitting contour and its center point from the actual values. Consequently, the minimum-distance circle-fitting method is applied to reconstruct the mushroom contour in these two scenarios. N points x j , y j on the contour are selected, and the circle parameters are determined based on the sum of the absolute values of the distances between the data points and the circle, as depicted in Equation (9):
f = x j x c 2 + y j y c 2 r
where x c and y c are the center point of the fitted circle, r is the radius of the fitted circle, f is the minimum value of x c , y c and r are the best-fit parameters. For irregular contours, the longest sub-contour is extracted based on the corner segmentation proposed above, and then the target contours of overlapping and irregular mushrooms are reconstructed separately by least-squares ellipse fitting and minimum-distance circle-fitting according to the different situations of the bow curvature and length of the longest sub-contour.

6. Experimental Results and Discussion

6.1. Segmentation Experiment and Discussion Based on PR-SOLOv2

6.1.1. Model Training

To validate the results and advantages of the enhanced algorithm, this article uses Mask RCNN, YOLACT, SOLOv2, and PR-SOLOv2 image segmentation networks for training based on the mushroom datasets mentioned above to compare and verify their segmentation effects.
The model training environment was built under the Windows 10 operating system and a 1×GeForceRTX3090 NVIDIA graphics card from the United States, and the running memory was 24 GB. Framework Detectron2 of Facebook under the running architecture CUDA10.1 and CUDNN7.6.5 acceleration libraries of Python3.7, PyTorch1.6, and GPU were configured as the running framework of the model.
Training parameters were configured with a training batch size of 4. The neural network optimization algorithm employed stochastic gradient descent (SGD) to progress in the direction of the maximum slope at the current location. The initial learning rate was set at 0.01 and the momentum size at 0.9, and 5000 iterations of iterative training were carried out. In addition, the mesh size S of SOLOv2 and PR-SOLOv2 algorithms was (12,16,24,36,40).

6.1.2. Model Evaluation Metrics

The instance segmentation algorithm is essentially a problem of pixel classification and category recognition. Its evaluation metrics include Average Precision (AP), Intersection over Union (IOU), forecast time, and other common metrics.
AP represents the area value of the curve enclosed by the coordinate axis, and the precision–recall (PR) curve is plotted with the recall rate as the horizontal coordinate and the accuracy rate as the vertical coordinate. IOU denotes the intersection area of the target mask predicted by the model and the actual annotation box, along with the ratio of their combination. A preset judgment threshold is utilized, and when the IOU value exceeds this threshold, the model’s prediction for the target is deemed relatively accurate.
The mushroom segmentation task necessitates predicting categories at the pixel level. Therefore, evaluation indicators such as the IOU threshold and the corresponding AP value are adopted. These metrics help distinguish the intersection ratio between the predicted segment and the ground truth. The average accuracies A P 50 and A P 75 , obtained using the commonly used decision thresholds of 0.5 and 0.75 in the example segmentation task, are used to evaluate the segmentation accuracy of the model under different confidence levels.

6.1.3. Instance Segmentation Results and Discussion

(1)
Experiment and result of dense overlapping mushroom segmentation
For the mushrooms that grow densely and form mutual adhesion or even mutual extrusion, the segmentation effect of different instance segmentation methods is shown in Figure 10. The segmentation result of Mask RCNN shows inaccurate segmentation of mushroom masks and missing identification of some mushrooms, as shown in Figure 10b. No. 1, 2, 3, and 4 mushrooms gather and grow, stick together, squeeze, and deform each other. There is no height difference between mushrooms, which leads to their indistinct boundary. No. 6 and No. 8 mushrooms were not identified because they were severely obscured by their neighbors, with only a small amount of fungus cap exposed and uneven light exposure. Due to the limited growth space of No. 5 and No. 7 mushrooms, the fruiting body can only grow slightly inclined, exposing the stalk. It is difficult to distinguish between the mushroom cap to be identified and the stem. Mask RCNN cannot handle these complex instance segmentation cases.
Figure 10. Segmentation results of densely overlapping mushrooms: (a) Original; (b) Mask RCNN; (c) YOLACT; (d) PR-SOLOv2.
The segmentation effect of the YOLACT algorithm is shown in Figure 10c. There were some cases in which the stalk was misidentified as a mushroom, some mushrooms were not identified, and the shape of the generated mushroom mask was very different from the actual mushroom shape. The stalk of mushroom No. 14, marked in yellow box in the picture, is misidentified as a mushroom; No. 5, 7, and 9 mushrooms that grow slightly at an angle and expose their stalks are missed; No. 6 and No. 8 mushrooms were seriously blocked by the large mushrooms growing around them, resulting in a small imaging area and uneven illumination and not being recognized. The shape of the segmentation mask generated by No. 2, 10, 11, 12, and 13 mushrooms is rectangular, greatly different from the actual mushroom shape. Like Mask RCNN, the YOLACT algorithm cannot solve the complex and challenging high-precision segmentation problems mentioned above.
In contrast is the PR-SOLOV2 algorithm for dense adhesion and extrusion deformation. It can be accurately segmented in complex segmentation situations, such as non-obvious features of the fungus cap and stalk, as shown in Figure 10d. The PR-SOLOv2 can accurately segment No. 1, 2, 3, and 4 mushrooms that grow densely and extrude each other without blurred edges. It can also separate No. 6 and No. 8 mushrooms that can be heavily obscured and not easily identified. In addition, an excellent segmentation effect was also shown on No. 5, 7, and 9 mushrooms with no obvious characteristics of fungus cap and stalk, and no fungus stalk was misidentified as fungus cap. The mask shapes generated on No. 10, 11, 12 and 14 mushrooms will not appear rectangular like the YOLACT but will generate masks that conform to the actual shape of the mushrooms.
(2)
Segmentation of a tilting mushroom
For the tilted mushroom, the segmentation problems of the Mask RCNN algorithm and the YOLACT algorithm are similar, as shown in Figure 11b,c. No. 1 and 2 mushrooms grow at a full tilt to the soil surface, exposing most of the stalks. The stalks look like tiny mushrooms. The growth posture of No. 3, 4, and 5 mushrooms was slightly inclined, revealing a small amount of stem. With inaccurate segmentation results, the mask RCNN and YOLACT algorithms also divided part of the stalk into the cap. No. 6 mushroom grows together with other mushrooms and hides on the stem of another mushroom. When the Mask RCNN and YOLACT algorithm segment the mushroom, in this case, part of the stalk is also divided into the fungus cap to obtain an unclear mask. The PR-SOLOv2 as shown in Figure 11d can correctly segment the stem with similar characteristics to the mushroom cap due to oblique growth, which increases the difficulty of identifying the mushroom. Including the correct segmentation of No. 1 and No. 2 of this oblique growthwhich is difficult to accurately identify. It can also accurately divide mushrooms No. 3, 4, 5, and 6, whose stalk and cap characteristics are difficult to distinguish, and get their precise edge.
Figure 11. Segmentation result of tilting mushrooms: (a) Original; (b) Mask RCNN; (c) YOLACT; (d) PR-SOLOv2.
(3)
Segmentation result of tiny mushrooms
The identification effect of small mushrooms is shown in Figure 12. Mask RCNN and YOLACT have similar products in small mushroom recognition, as shown in Figure 12b,c. No. 1, 3, 4, 7, 8, 9, 11, 12, 13, 16, 17, and 19 mushrooms are not exposed to uniform light, the individual mushroom is small, and the characteristics of the fungus cap are not prominent. Therefore, it is easy to be mistaken for mycelium and miss recognition. Individual mushrooms No. 2, 5, 6, 10, 14, 15, 18, 20, and 21 are slightly larger, belonging to the small mushroom and large mushroom adhesion with a certain height difference. It is easy to identify the small mushroom in the stalk of the large mushroom, which also missed identification. However, the effect of the PR-SOLOv2 algorithm is shown in Figure 12d; it can correctly segment the mushrooms with smaller individuals and fewer features, whose characteristics are similar to mycelia. In addition, it can also accurately segment small mushrooms that are adhered to large mushrooms and have different heights and similar features.
Figure 12. Segmentation result of tiny mushrooms: (a) Original; (b) Mask RCNN; (c) YOLACT; (d) PR-SOLOv2.
(4)
Comparison of segmentation effects
The effect of instance segmentation on mushroom segmentation is tabulated in Table 2. The average segmentation accuracy of the Mask RCNN on the dense mushroom image segmentation task is 83.510%. The AP of YOLACT is 80.743%, while that of SOLOv2 is more than 90% (90.279%). The PR-SOLOv2 proposed in this paper has the highest AP among the four segmentation methods in dense mushroom segmentation, reaching 93.037%, which is about 10% higher than Mask RCNN and YOLACT, and about 3% higher than that of SOLOv2. When the IOU threshold is set to 50%, the AP performance of PR-SOLOv2 in the mushroom image segmentation task is as high as 99.056%. In addition, the average segmentation time is 0.39 s, which is still better than the Mask RCNN and YOLACT algorithms.
Table 2. Effect of instance segmentation on mushroom segmentation.
(5)
Comparison of segmentation effect under different light sources
In order to compare the segmentation effect of the model under the three light sources, this paper selects 60 pictures for experiment; the pictures are all from the camera used in the experiment; each picture has the same resolution and is divided into three groups according to the light sources. The overall results of the experiment are as follows: 651 examples of top light source images, 638 successful segmentations, segmentation success rate of 97.81%; 794 examples of side light source images, 772 successful segmentations, segmentation success rate of 97.23%; 1102 examples of dark light source images, 1088 successful segmentations, segmentation success rate of 98.73%, as shown in Figure 13.
Figure 13. Segmentation results under different light sources.
The result shows that the segmentation accuracy of the side light source is the lowest, lower than that of the top light source, and lower than that of the dark light source. This paper concludes that since the side light source produces shadows on the mushroom surface when irradiating the mushroom, the light and dark demarcation line will affect the extraction of the mushroom edge features to a certain extent, so it will affect the final segmentation success rate. Based on the results of this experiment, considering that in the natural state, most of the stacked mushroom racks are illuminated by lights with side light sources, the direct photographs will affect the subsequent recognition and contour reconstruction. Therefore, in this paper, the camera is equipped with a parallel light source to assist illumination when taking pictures, to eliminate the influence of the side light source on the subsequent recognition and segmentation of mushrooms, and to improve the accuracy of recognition and contour reconstruction.
(6)
Further validation of PR-SOLOv2 algorithm
There are 260 mushroom sample images of different mushroom rooms and layers collected in the actual production environment from two plants, totaling 9362 mushrooms. The mask RCNN, YOLACT, SOLOV2, and PR-SOLOv2 algorithms were used to segment the sample images. The experimental outcomes are shown in Table 3. It is observed that the proposed approach achieves a segmentation success rate of 98.13%, which is significantly higher than those of the other existing works.
Table 3. Mushroom segmentation success rate of different algorithms.
There are two main reasons for the unsuccessful segmentation: first, multiple large mushrooms are blocked at the same time, resulting in too serious small mushrooms; Second, a small mushroom with a large area of dark lines on the surface. In the marking of the dataset, the two cases are not common, resulting in poor learning effect of the segmentation features of the model for these two cases. In the future, the segmentation efficiency of the model can be improved by adding the marking of special cases.

6.2. Experimental Results and Discussion of Contour Reconstruction

The mask contour obtained by segmentation mask based on the PR-SOLOv2 algorithm proposed in this paper is adopted. The process and effect of the classification and reconstruction method according to the curvature and arc length of the mask contour are shown in Figure 14: (a) the fruit-body mask obtained by PR-SOLOv2 segmentation network processing proposed in this paper; (b) the extracted mask contour. Among them, No. 1, 5, and 8 are mushrooms whose average curvatures are 0.01317, 0.01367, and 0.02313, respectively, and corresponding lengths are 444, 427, and 246 pixels, respectively. They are all mushrooms whose average curvature is less than 0.103 and whose contour length is more than 100. No. 2, 3, 4, 6 and 7 are partially occluded fruiting bodies, whose average curvatures are 0.02378, 0.01787, 0.01648, 0.02087, and 0.02252, respectively, and corresponding contour lengths are 291, 257,335, 239,and 316 pixels, respectively, belonging to non-smooth occluded contours. Among them, the arch curvature of the largest contour fragment of No. 2, No.4, and No. 7 mushrooms is more significant than 0.9. The least-squares ellipse fitting was used. The No. 3 and No. 6 mushrooms were extruded to produce a large deformation, and their contour shape was irregular. However, the arch curvature of their longest contour was greater than 0.9, so the ellipse fitting was also adopted. The figure illustrates that both the regular mushroom and the severely blocked mushroom exhibit a well-matched contour when employing the contour reconstruction method proposed in this paper. The reconstruction effect is satisfactory.
Figure 14. Reconstruction effect of overlapping extruded mushrooms: (a) PR-SOLOv2 mask; (b) Extract mask edges; (c) Contour reconstruction.
More reconstruction effects of the mushroom cluster neutron entity are shown in Figure 15. In general, the contour reconstruction method in this paper can fit the contour of the densely overlapped fruit body well. For example, the surface damage of No. 1, 7, 8, 12, and 26 mushrooms and the surface dirt of No. 3 mushrooms will affect the image segmentation effect and lead to a poor subsequent reconstruction effect. The SOLOv2-PR algorithm proposed in this paper can accurately segment the mushroom mask and extract the edge of the mask. On this basis, the mushroom contour reconstruction can have a good effect. No. 6, 10, 13, 17, 18, 19, 20, 21, 22, 26, 23, 27, 28, and 29 mushrooms are mainly adhered to a mushroom around them or are partially blocked by a single mushroom next to them. Although these mushrooms are blocked or squeezed, the deformation of the mushrooms is not severe, and the extracted mushroom outline is roughly smooth and regular. The contour reconstruction method proposed in this paper can achieve a better effect than mushroom contour reconstruction. In addition, for the mushrooms squeezed by two surrounding mushrooms, such as No. 4, 5, 9, 12, 15, 14, 16, 25, 31, 32, 34, and 33, the contour edge deformation is large, and the shape is irregular, so it is not easy to reconstruct the contour. The mushroom contour reconstruction method realized in this paper can restore the mushroom’s true contour.
Figure 15. Mushroom contour reconstruction effect.
However, in cases of extremely dense growth, a majority of the fruiting bodies, like mushrooms numbered 2, 11, 15, 24, 30, and 35, are either blocked or severely deformed by the presence of multiple surrounding mushrooms. The PR-SOLOv2 algorithm is employed to segment the mask contour of these severely deformed mushrooms and subsequently reconstruct their contours. There may be some deviation between the fitted contour and the actual contour. However, there will be missed recognition for the mushroom fruiting bodies smaller than 30 pixels, such as the mushrooms in the area marked by the yellow rectangle box. However, these are tiny mushrooms with fresh buds, so the missing identification does not affect the picking or monitoring.

6.3. Verification of Center Coordinates

6.3.1. Center Point Ground Truth Determination Method

In order to find the ground true of center point of each fruiting body in the mushroom image, we use manual marking to determine the center point of each fruiting body, as shown in the Figure 16. The center point of the daughter body with regular contour is directly obtained by using circle annotation in labelme, and the daughter body with irregular contour is obtained by using polygon annotation in labelme, as close as possible to its true contour, as shown in Figure 16b. After all mushrooms are marked, the center point of the child body marked with a circle is directly output as its benchmark, and the center point of the child body marked with a polygon is obtained by ellipse fitting output, as shown in Figure 16c. The coordinates of all the central green ‘x’ in the final output image are used as the reference point of each sub-entity center.
Figure 16. Benchmarking process for determining mushroom centers: (a) Original image; (b) Marking image; (c) Fitting contour map.

6.3.2. Calculation of Fitting Accuracy

In order to quantify the contour reconstruction accuracy, the center point positioning accuracy of the contour is used to characterize it. Firstly, the deviation from the center datum is compared from the actual fitting effect graph, as shown in Figure 17: the center point of the actual contour using manual labeling, as shown in Figure 17a; and the center point of the contour reconstructed by PR-SOLOv2, as shown in Figure 17b; for comparison, the comparison effect is shown in Figure 17c. It can be seen that the contour reconstructed by PR-SOLOv2 fits better with the actual contour manually labeled, and the center point deviation is very small.
Figure 17. Comparative effect of fitting: (a) Manually labeled contours; (b) Reconstructed contours by this paper’s method; (c) Comparison result.
In order to calculate the deviation more quantitatively, the center point positioning accuracy is calculated using the 2D coordinate deviation rate (CDR) [24] as depicted in Equation (10).
E = c j c i W + r j r i H × 100 %
where rj and cj are the row and column coordinates of the center point of the mushroom contour fitted by manual marking, ri and ci are the row and column coordinates of the center point of the mushroom obtained by the research method in this paper, and W and H are the width and height of the mushroom image, respectively.
The contour reconstruction comparison results of three more dense onsite mushroom images (169 mushrooms in total) are provided to further verify the effectiveness of the contour reconstruction method in this paper, as shown in Figure 18. The corresponding CDR comparison of the four images is shown in Figure 19.
Figure 18. Comparison effect of contour reconstruction by this paper’s method and manually labeled contours: (a) Manually labeled contours; (b) Reconstructed contours by this paper’s method; (c) Comparison result.
Figure 19. CDR plots of the four pictures.
As shown in Table 4, the average CDR between the centroid of the contour fitted by the method in this paper and the actual contour is 0.30%, which shows that the centroid positioning accuracy is higher, with better results.
Table 4. Mean deviation statistics for the average CDR coordinates of the four pictures.

7. Conclusions

This paper presents a high-precision segmentation and contour reconstruction method for densely overlapping mushrooms based on an improved SOLOv2 model (PR-SOLOv2), aiming to improve the segmentation accuracy and center point positioning accuracy of highly dense mushrooms. The experimental results show significant improvement. The PR-SOLOv2 model improves segmentation accuracy to 93.04% and segmentation rate by 98.13%, about 10% higher than Mask RCNN and YOLACT. Meanwhile, the center positioning error is as low as 0.3%, which proved that the proposed contour reconstruction method in this study could accurately segment dense, overlapped, extruded, and tilting fruiting bodies, effectively reconstruct mushroom contours that are severely occluded and compressed, and achieve accurate center point positions. It is not only suitable for the high-precision identification and positioning of the fruiting bodies of various edible fungi but also those of other densely overlapping spherical-like fruits. However, there are still some limitations. For example, for mushrooms with too a high tilt angle, occasionally it will be mistaken to identify the stalk of the mushroom as a fungus cap. When the fruiting body of the mushroom is severely obstructed, the reconstructed mushroom contour may have some relatively larger deviations from the actual mushroom contour, which needs to be further improved. In the future, we will continue to improve the structure of the segmentation model, and even try some new occlusion segmentation algorithms such as amodal instance segmentation, to further enhance the segmentation effect and reduce processing time, and optimize the reconstruction fitness of the contour reconstruction algorithm for contours covered by multiple positions or severely obstructed, to further enhance center point position accuracy.

Author Contributions

S.Y.: Writing—original draft, writing—review and editing, conceptualization, methodology, investigation, data curation, formal analysis, software, validation, production of related equipment, funding acquisition. J.Z.: writing—original draft, writing—review and editing, data curation, validation, software. J.Y.: investigation, writing-review and editing, conceptualization. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shanghai Agriculture Science and Technology Innovation Program, Grant No. I2023003; Shanghai Science and Technology Innovation Action Plan-Agriculture, Grant No. 21N21900600; Major Scientific and Technological Innovation Project of Shandong Province, Grant No. 2022CXGC010609.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Anyone can access our data by sending an email to szyang@sspu.edu.cn.

Acknowledgments

We acknowledge the Shanghai Lianzhong edible fungus professional cooperative for the use of their mushrooms and facilities.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Diego, C.Z.; Pardo-Giménez, A. Edib SOLOv2: Dynamic and fast instance segmentation le and medicinal mushrooms (technology and applications); Current overview of mushroom production in the world. In Edible and Medicinal Mushrooms: Technology and Applications; Wiley: Hoboken, NJ, USA, 2017; pp. 5–13. [Google Scholar]
  2. Zhang, J.; Cai, W.; Huang, C. Chinese Edible Fungus Cultivation, 1st ed.; China Agriculture Press: Beijing, China, 2020; Chapter 1. [Google Scholar]
  3. Reed, J.N.; Miles, S.J.; Butler, J.; Baldwin, M.; Noble, R. Automatic mushroom harvester development. J. Agric. Eng. Res. 2001, 78, 15–23. [Google Scholar] [CrossRef]
  4. Rowley, J.H. Developing Flexible Automation for Mushroom Harvesting (Agaricus bisporus): Innovation Report. Ph.D. Thesis, University of Warwick, Warwick, UK, 2009. [Google Scholar]
  5. Yang, C.H.; Xiong, L.Y.; Wang, Z.; Wang, Y.; Shi, G.; Kuremot, T.; Yang, Y. Integrated detection of citrus fruits and branches using a convolutional neural network. Comput. Electron. Agric. 2020, 174, 105469. [Google Scholar] [CrossRef]
  6. Yang, S.; Jia, B.; Yu, T.; Yuan, J. Research on multiobjective optimization algorithm for cooperative harvesting trajectory optimization of an intelligent multiarm straw-rotting fungus harvesting robot. J. Agric. 2022, 12, 986. [Google Scholar] [CrossRef]
  7. Noble, R.; Reed, J.N.; Miles, S.; Jackson, A.F.; Butler, J. Influence of mushroom strains and population density on the performance of a robotic harvester. J. Agric. Eng. Res. 1997, 68, 215–222. [Google Scholar] [CrossRef]
  8. Slaughter, D.C.; Harrell, R.C. Color vision in robotic fruit harvesting. Trans. ASAE 1987, 30, 1144–1148. [Google Scholar] [CrossRef]
  9. Gené-Mola, J.; Gregorio, E.; Guevara, J.; Auat, F.; Sanz-Cortiella, R.; Escolà, A.; Rosell-Polo, J.R. Fruit detection in an apple orchard using a mobile terrestrial laser scanner. Biosyst. Eng. 2019, 187, 171–184. [Google Scholar] [CrossRef]
  10. Lin, G.; Tang, Y.; Zou, X.; Li, J.; Xiong, J. In-field citrus detection and localisation based on RGB-D image analysis. Biosyst. Eng. 2019, 186, 34–44. [Google Scholar] [CrossRef]
  11. Magalhães, S.A.; Moreira, A.P.; Santos, F.N.D.; Dias, J. Active Perception Fruit Harvesting Robots—A Systematic Review. J. Intell. Robot. Syst. 2022, 105, 1–22. [Google Scholar] [CrossRef]
  12. Montoya-Cavero, L.E.; de León Torres, R.D.; Gómez-Espinosa, A.; Cabello, J.A.E. Vision systems for harvesting robots: Produce detection and localization. Comput. Electron. Agric. 2021, 192, 106562. [Google Scholar] [CrossRef]
  13. Yan, B.; Fan, P.; Lei, X.; Liu, Z.; Yang, F. A real-time apple targets detection method for picking robot based on improved YOLOv5. Remote Sens. 2021, 13, 1619. [Google Scholar] [CrossRef]
  14. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  15. Onishi, Y.; Yoshida, T.; Kurita, H.; Fukao, T.; Arihara, H.; Iwai, A. An automated fruit harvesting robot by using deep learning. ROBOMECH J. 2019, 6, 1–8. [Google Scholar] [CrossRef]
  16. Zhou, C.; Hu, J.; Xu, Z.; Yue, J.; Ye, H.; Yang, G. A novel greenhouse-based system for the detection and plumpness assessment of strawberry using an improved deep learning technique. Front. Plant Sci. 2020, 11, 559. [Google Scholar] [CrossRef] [PubMed]
  17. Ganesh, P.; Volle, K.; Burks, T.F.; Mehta, S.S. Deep orange: Mask R-CNN based orange detection and segmentation. IFAC PapersOnLine 2019, 52, 70–75. [Google Scholar] [CrossRef]
  18. Ghiani, L.; Sassu, A.; Palumbo, F.; Mercenaro, L.; Gambella, F. In-field automatic detection of grape bunches under a totally uncontrolled environment. Sensors 2021, 21, 3908. [Google Scholar] [CrossRef]
  19. Yu, Y.; Zhang, K.; Yang, L.; Zhang, D. Fruit detection for strawberry harvesting robot in non-structural environment based on Mask-RCNN. Comput. Electron. Agric. 2019, 163, 104846. [Google Scholar] [CrossRef]
  20. Zhang, K.; Lammers, K.; Chu, P.; Li, Z.; Lu, R. System design and control of an apple harvesting robot. Mechatronics 2021, 79, 102644. [Google Scholar] [CrossRef]
  21. Tillett, R.D.; Batchelor, B.G. An algorithm for locating mushrooms in a growing bed. Comput. Electron. Agric. 1991, 6, 191–200. [Google Scholar] [CrossRef]
  22. Yu, G.H.; Luo, J.M.; Zhao, Y. Region marking technique based on sequential scan and segmentation method of mushroom images. Trans. Chin. Soc. Agric. 2006, 22, 4. [Google Scholar]
  23. Yang, Y.Q.; Ye, M.; Lu, Y.H.; Ren, S.G. Localization algorithm based on corner density detection for overlapping mushroom image. Comput. Syst. Appl. 2018, 24, 119–125. [Google Scholar]
  24. Yang, S.; Ni, B.; Du, W.; Yu, T. Research on an improved segmentation recognition algorithm of overlapping Agaricus bisporus. Sensors 2022, 22, 3946. [Google Scholar] [CrossRef] [PubMed]
  25. Yin, H.; Yi, W.; Hu, D. Computer vision and machine learning applied in the mushroom industry: A critical review. Comput. Electron. Agric. 2022, 198, 107015. [Google Scholar] [CrossRef]
  26. Mukherjee, A.; Sarkar, T.; Chatterjee, K.; Lahiri, D.; Nag, M.; Rebezov, M.; Lorenzo, J.M. Development of artificial vision system for quality assessment of oyster mushrooms. Food Anal. Methods 2022, 15, 1663–1676. [Google Scholar] [CrossRef]
  27. Wang, Z.J.; Zhang, L.L.; Li, S.F.; Zhang, Q.Y.; Fu, Q.Q.; Liu, S.J. Identification and classification of eatable fungi based on machine learning algorithm. J. Fuyang Teach. Coll. Nat. Sci. 2021, 38, 7. [Google Scholar]
  28. Zhou, J.; Ding, W.; Zhu, X.; Cao, J.; Niu, X. Evaluation on formation rate of Pleurotus eryngii primordium under different humidity conditions by computer vision. J. Zhejiang Univ. Sci. Agric. Life Sci. 2017, 43, 262–272. [Google Scholar] [CrossRef]
  29. Huang, B.R.; Chen, J.Y.; Lai, X.Y.; Chen, G.W.; Huang, K.W. Application of Deep Learning for Mushrooms Cultivation. In Proceedings of the 2021 International Conference on Technologies and Applications of Artificial Intelligence (TAAI), Taipei, Taiwan, 21 January 2021; pp. 250–253. [Google Scholar]
  30. Huang, M.; Jiang, X.; He, L.; Choi, D.; Pecchia, J.; Li, Y. Development of a robotic harvesting mechanism for button mushrooms. Trans. ASABE 2017, 64, 565–575. [Google Scholar] [CrossRef]
  31. Ketwongsa, W.; Boonlue, S.; Kokaew, U. A new deep learning model for the classification of poisonous and edible mushrooms based on improved AlexNet convolutional neural network. J. Appl. Sci. 2022, 12, 3409. [Google Scholar] [CrossRef]
  32. Sevilla, W.H.; Hernandez, R.M.; Ligayo, M.A.D.; Costa, M.T.; Quismundo, A.Q. Machine vision recognition system of edible and poisonous mushrooms using a small training set-based deep transfer learning. In Proceedings of the 2022 International Conference on Decision Aid Sciences and Applications (DASA), Chiangrai, Thailand, 23–25 March 2022; pp. 1701–1705. [Google Scholar]
  33. Zha, L.; Gong, Z.J.; Shao, S.; Li, Z.P.; Yu, C.X.; Yang, H.L. An anchor-free network for detection of Volvariella volvacea growth status. Acta Edulis Fungi 2022, 29, 8. [Google Scholar]
  34. Lee, C.H.; Choi, D.; Pecchia, J.; He, L.; Heinemann, P. Development of a mushroom harvesting assistance system using computer vision. In Proceedings of the 2019 ASABE Annual International Meeting, American Society of Agricultural and Biological Engineers (ASABE), Boston, MA, USA, 7–10 July 2019; pp. 1–10. [Google Scholar]
  35. Rong, J.; Wang, P.; Yang, Q.; Huang, F. A field-tested harvesting robot for oyster mushroom in greenhouse. J. Agron. 2021, 11, 1210. [Google Scholar] [CrossRef]
  36. Zheng, Y.Y.; Zheng, L.X. Research on Flammulina velutipes head detection algorithm based on deep learning. Mod. Comput. 2020, 30, 23–29. [Google Scholar]
  37. Lu, C.P.; Liaw, J.J. A novel image measurement algorithm for common mushroom caps based on convolutional neural network. Comput. Electron. Agric. 2020, 171, 105336. [Google Scholar] [CrossRef]
  38. Masoudian, A. Computer Vision Algorithms for an Automated Harvester. Ph.D. Thesis, The University of Western Ontario, London, ON, Canada, 2013. [Google Scholar]
  39. Sert, E.; Okumus, I.T. Segmentation of mushroom and cap width measurement using modified K-means clustering algorithm. Adv. Electr. Electron. Eng. 2014, 12, 354–360. [Google Scholar] [CrossRef]
  40. Soomro, S.; Munir, A.; Choi, K.N. Hybrid two-stage active contour method with region and edge information for intensity inhomogeneous image segmentation. PLoS ONE 2018, 13, e0191827. [Google Scholar] [CrossRef] [PubMed]
  41. Pchitskaya, E.; Bezprozvanny, I. Dendritic spines shape analysis—Classification or clusterization? Perspective. Front. Synaptic Neurosci. 2020, 12, 31. [Google Scholar] [CrossRef]
  42. Ji, J.; Sun, J.; Zhao, K.; Jin, X.; Ma, H.; Zhu, X. Measuring the cap diameter of white button mushrooms (Agaricus bisporus) by using depth image processing. Appl. Eng. 2021, 37, 623–633. [Google Scholar] [CrossRef]
  43. Arjun, A.D.; Chakraborty, S.K.; Mahanti, N.K.; Kotwaliwale, N. Non-destructive assessment of quality parameters of white button mushrooms (Agaricus bisporus) using image processing techniques. J. Food Sci. Technol. 2022, 59, 2047–2059. [Google Scholar] [CrossRef]
  44. Baisa, N.L.; Al-Diri, B. Mushrooms Detection, Localization and 3D Pose Estimation using RGB-D Sensor for Robotic-picking Applications. arXiv 2022, arXiv:2201.02837. [Google Scholar]
  45. Zhao, K.; Zhang, M.; Ji, J.; Sun, J.; Ma, H. Whiteness measurement of Agaricus bisporus based on image processing and color calibration model. Food Meas. 2023, 17, 2152–2161. [Google Scholar] [CrossRef]
  46. Retsinas, G.; Efthymiou, N.; Anagnostopoulou, D.; Maragos, P. Mushroom detection and three dimensional pose estimation from multi-view point clouds. Sensors 2023, 23, 3576. [Google Scholar] [CrossRef]
  47. Zhang, J.; Li, C.; Rahaman, M.M.; Yao, Y.; Ma, P.; Zhang, J.; Zhao, X.; Jiang, T.; Grzegorzek, M. A comprehensive survey with quantitative comparison of image analysis methods for microorganism biovolume measurements. Arch. Comput. Methods Eng. 2023, 30, 639–673. [Google Scholar] [CrossRef]
  48. Chen, C.; Wang, F.; Cai, Y.; Yi, S.; Zhang, B. An Improved YOLOv5s-Based Agaricus bisporus Detection Algorithm. J. Agron. 2023, 13, 1871. [Google Scholar] [CrossRef]
  49. Chen, C.; Yi, S.; Mao, J.; Wang, F.; Zhang, B.; Du, F. A Novel Segmentation Recognition Algorithm of Agaricus bisporus Based on Morphology and Iterative Marker-Controlled Watershed Transform. J. Agron. 2023, 13, 347. [Google Scholar] [CrossRef]
  50. Yin, H.; Xu, J.; Wang, Y.; Hu, D.; Yi, W. A Novel Method of Situ Measurement Algorithm for Oudemansiella raphanipies Caps Based on YOLO v4 and Distance Filtering. Agronomy 2022, 13, 134. [Google Scholar] [CrossRef]
  51. Yin, H.; Yang, S.; Cheng, W.; Wei, Q.; Wang, Y.; Xu, Y. AC R-CNN: Pixelwise Instance Segmentation Model for Agrocybe cylindracea Cap. Agronomy 2023, 14, 77. [Google Scholar] [CrossRef]
  52. Wang, T.; Liew, J.H.; Li, Y.; Chen, Y.; Feng, J. SODAR: Segmenting objects by dynamically aggregating neighboring mask representations. arXiv 2022, arXiv:2202.07402. [Google Scholar]
  53. Wang, X.; Kong, T.; Shen, C.; Jiang, Y.; Li, L. SOLO: Segmenting objects by locations. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 649–665. [Google Scholar]
  54. Wang, X.; Zhang, R.; Kong, T.; Li, L.; Shen, C. SOLOv2: Dynamic and fast instance segmentation. Adv. Neural Inf. Process. Syst. 2020, 33, 17721–17732. [Google Scholar]
  55. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 12 June 2015; pp. 3431–3440. [Google Scholar]
  56. Li, T.Y.; Jiang, W.W.; Xing, J.T.; Xu, Z.; Zhang, L.; Tian, F. Intrusion detection of offshore area perimeter in oilfield operation site. Comput. Syst. Appl. 2022, 31, 236–241. [Google Scholar]
  57. Ji, J.T.; Liu, X.H.; Zhao, K.X. Automatic rumen filling scoring method for dairy cows based on SOLOv2 and cavity feature of point cloud. Trans. Chin. Soc. Agric. Eng. 2022, 38, 186–197. [Google Scholar]
  58. Kirillov, A.; Wu, Y.; He, K.; Girshick, R. PointRend: Image segmentation as rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13 June 2020; pp. 9799–9808. [Google Scholar]
  59. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22 October 2017; pp. 2980–2988. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.