1. Introduction
Rapeseed is a vital oilseed crop, and the quality of its seed production directly impacts agricultural productivity and seed industry security. Field management, as a critical component of the seed production process, encompasses multiple aspects, including plant regulation, weed and inferior plant removal, and paternal parent elimination [
1,
2]. Traditional manual methods are not only inefficient and labour-intensive but also prone to inconsistent results due to subjective judgement errors. In field management, male plant removal primarily relies on chemical emasculation, a method susceptible to environmental conditions and posing risks of pesticide residues [
3,
4,
5]. Integrating field management machinery with intelligent agricultural equipment to establish a technology system for intelligent recognition and navigation line extraction holds significant importance for advancing intelligent, high-precision field management techniques [
6,
7]. This has become an urgent requirement for enhancing the modernisation level of rapeseed seed production [
8].
With the rapid advancement of machine vision and artificial intelligence technologies, navigation methods based on visual perception have become a significant research direction in agricultural automation [
9]. In the field of crop row navigation technology research, early approaches primarily relied on conventional image processing algorithms. For instance, Gao Guoqin et al. [
10] integrated the K-means algorithm with the HSV colour model to cluster image pixels into two categories, thereby directly separating target path regions and mitigating the impact of lighting conditions. This method demonstrated faster processing speeds compared to the HSV colour model alone. JIANG et al. [
11] converted deep images of farmland into point clouds, demonstrating superior performance in handling environmental factors such as illumination, shadows, and weeds compared to colour space-based methods. GONG Jinliang et al. [
12] employed 2G-B-R and morphological segmentation techniques for maize root images, enhancing the vertical projection algorithm for feature point extraction. This approach exhibited greater adaptability and real-time capability than traditional peak point methods, enabling real-time navigation paths for agricultural machinery in maize fields. Chen Jiqing et al. [
13] employed greyscale conversion of vegetation colour indices, Gabor filtering, and PCA dimensionality reduction for feature extraction. Combining K-means clustering with the central axis algorithm, they generated navigation paths demonstrating higher recognition precision and faster processing speeds than the 2G-B-R approach. Whilst computationally efficient, such approaches fundamentally rely on superficial features like colour and texture, lacking an understanding of image semantic content. Consequently, their adaptability is limited in complex scenarios where crop-background contrast is low.
With the rapid advancement of deep learning, various deep learning algorithms—particularly convolutional neural networks—have been extensively applied in the agricultural sector, yielding remarkable results and providing robust technological support for enhancing agricultural productivity [
14]. Yu Tan et al. [
15] introduced MobileNetv3 and ECANet to enhance the YOLOv5s algorithm for navigating between ginseng crop rows, thereby improving root detection. Most of the aforementioned studies focused on navigational line extraction when crops exhibited relatively distinct characteristics. Su Tong et al. [
16] addressed the challenge of detecting navigation lines in mid-to-late-stage maize rows, where insufficient light and obstruction hinder detection. They introduced an Edge Extraction Module (EEM) and an ASPP module within the Fast-SCNN model to optimise path edge segmentation, thereby enhancing detection precision. Concurrently, they employed pixel scanning, weighted averaging, and least squares methods to detect and fit navigation lines, reducing heading angle deviation. Peng Shubo et al. [
17] enhanced apple tree trunk recognition precision in YOLOv7 networks by integrating CBAM attention and SPD-Conv modules, achieving a 2.31% improvement over the baseline model. They validated the feasibility of using midpoints of rectangular bounding boxes instead of trunk root points for navigation line fitting. Ying Qiukai et al. [
18] employed the YOLOv8 instance segmentation model for strawberry ridge detection, proposing a navigation line fitting method combining the Canny edge detection algorithm with the intercept method. This approach demonstrated higher navigation line precision compared to traditional least squares methods. Yu Gaohong et al. [
19] addressed barren field ridges by refining the DeepLabV3+ algorithm. They enhanced detection speed by substituting the Xception backbone with MobileNetV2 and introduced the Convolutional Block Attention Module (CBAM) to improve ridge boundary extraction. Hailiang Gong et al. [
20] refined the YOLOX-Tiny model by incorporating adaptive illumination adjustment, multi-scale prediction, visual attention mechanisms, and the Fast-SPP module. Combined with the CIoU loss function and least squares method, this approach achieved high-precision identification of maize crop row navigation lines. Dong et al. [
21] developed a detection method for rapeseed seedling rows by combining an improved BiSeNetV2 with dynamic sliding window fitting, integrating ECA, ASPP, and DS Conv for optimisation. This method performs well in various environments, but its effectiveness in extreme scenarios and real-time performance for high-speed operations need further verification. Zhao et al. [
22] developed an autonomous laser weeding robot for strawberry fields based on DIN-LW-YOLO, an improved YOLOv8s-pose integrated with EMA attention and C2f-DCNv3. The model achieves excellent detection performance, with field tests showing 92.6% weed control rate and 1.2% seedling injury rate that meet agronomic requirements. Saha and Noguchi [
23] proposed a machine vision-based autonomous navigation framework for vineyards, optimising YOLOv8 to develop the dedicated YOLOv8m-vine-classes model with 95% precision and 93.7% mAP50, enabling accurate vine row recognition and safe EV navigation. Most of the aforementioned studies focused on navigational line extraction when crops exhibited relatively distinct characteristics. These findings demonstrate the robust performance of deep learning in structured scenarios. However, in the specific context of rapeseed parent plants exhibiting high visual similarity and interplant growth, challenges persist, including insufficient model adaptability and inadequate sensitivity to subtle feature variations.
Particularly during field management of rapeseed seed production, the morphological and colour similarities between parental lines result in low distinguishability. Traditional threshold-based segmentation or feature engineering methods prove ineffective for extracting navigation lines [
24,
25,
26], while existing deep learning models often fail to simultaneously achieve both lightweight operation and high precision [
27,
28,
29]. High planting density, severe plant-to-plant occlusion, and weed interference further compound the challenges of visual recognition. In summary, current research lacks lightweight, high-precision segmentation models specifically tailored to the unique scenario of rapeseed parental lines exhibiting high morphological similarity and dense planting conditions. Existing crop row navigation studies predominantly focus on crops with markedly distinct characteristics, such as maize and strawberries, leaving a gap in adaptable solutions for complex scenarios involving highly similar rapeseed parental lines and severe field occlusion. This paper proposes a rapeseed parent row detection algorithm based on SegNav-YOLOv8n. Distinct from existing approaches that merely stack attention modules or optimise lightweight backbone networks, this study innovatively constructs a synergistic optimisation framework comprising “lightweight downsampling—dynamic upscaling—feature fusion”: the ADown module addresses boundary information loss in traditional downsampling, the DySample module achieves adaptive feature upscaling reconstruction, while the C2f_FB module enhances multi-scale feature integration under lightweight constraints. These components form an organic whole rather than a simple concatenation. Through this collaborative optimisation framework, this study enhances the ability to distinguish subtle differences between parental rows while maintaining high computational efficiency. This not only fills a technical gap in navigational detection for rapeseed seed production parental rows but also provides a transferable, modular improvement paradigm for visual navigation tasks involving highly similar crop rows. It offers a reliable visual navigation solution for intelligent, mechanised field management in rapeseed seed production.
2. Materials and Methods
2.1. Data Collection and Preprocessing
This research dataset was collected between March and May 2025 at the Provincial Rapeseed Breeding Centre Experimental Base in Xunlong River Ecological Art Town, Changsha County. This period spans the rapeseed plant’s budding stage through flowering to the initial pod formation phase, representing the core period for critical field management operations such as weed removal and emasculation. The collected data precisely aligns with actual production requirements.
The data collection area totalled approximately 150 mu (approximately 10 hectares), encompassing three distinct rapeseed varieties sown at different times. Both mechanical direct-drilling and manual transplanting methods were employed, adhering to a 1:2 (paternal:maternal) seed production model. Within this setup, paternal rows are spaced at 1.8 m intervals, and maternal rows at 2.1 m intervals, with inter-row furrows measuring 0.2–0.25 m in width. This configuration integrates agronomic practices with mechanised field operations.
Data acquisition employed a Realsense D435i depth camera(Intel Corporation, Santa Clara, CA, USA), with raw capture resolution set at 1920 × 1080 pixels. Images ultimately used for model training were uniformly cropped and resized to 640 × 640 pixels. During capture, the camera was mounted on a high-clearance chassis (1.8 m wide, positioned above the paternal rapeseed row) fixed at the front of the vehicle. It was suspended vertically 2 m above ground level at a 30° angle to the ridge surface, enabling complete coverage of one paternal row and its corresponding two maternal rows.
The vehicle’s operational speed was controlled at 0.5 m/s, matching the working speed of actual rapeseed field management machinery to simulate dynamic image capture under real-world working conditions. Data collection periods were concentrated between 09:00–11:30 and 14:00–16:30 daily, encompassing three typical light conditions: clear skies, cloudy skies, and overcast skies. A total of 21 video segments were collected, primarily encompassing four core scenarios, as illustrated in
Figure 1.
Frames were extracted at 0.6 s intervals (approximately 16.7 frames per second), yielding 60–100 raw images per video segment. These were subsequently manually filtered to remove blurred, overexposed, or invalid frames, ultimately yielding 831 valid images. To ensure sample independence and evaluation objectivity, a “video session grouping” strategy was employed for dataset partitioning: the 21 video segments were divided into independent units based on “capture date + scene type”. All frames from the same video segment were assigned to the same dataset (training set/validation set/test set), with no cross-set distribution. The training set encompassed 17 videos (covering different crop varieties, sowing methods, and lighting scenarios), while the validation and test sets each contained 2 videos. The test set videos were captured ≥ 3 days apart from the training set videos, with natural variations in scene conditions (e.g., weed coverage and light intensity) to avoid evaluation bias caused by frame-to-frame correlation. The final partitioning yielded: training set: 17 videos (665 raw images), validation set: 2 videos (83 raw images), and test set: 2 videos (83 raw images). The three datasets exhibit an approximate 8:1:1 distribution ratio across scenarios, ensuring balanced representation of all conditions within the dataset.
2.2. Data Augmentation
To further enhance the model’s generalisation capability and strictly prevent data leakage risks, all data augmentation operations in this study were applied exclusively to the training set. The validation and test sets retained the original image data without undergoing any form of augmentation processing. The training set was expanded through the following data augmentation strategies: random rotation, brightness adjustment, contrast adjustment, and Gaussian blurring. Each training image underwent one or more augmentation operations with a 50% probability. Through data augmentation techniques, the training set was expanded to 2760 images, bringing the total dataset size to 2926 images, thereby ensuring data diversity. Following the partitioning of the dataset, a training set comprising 2760 images, a validation set of 83 images, and a test set of 83 images were generated.
Figure 2 illustrates partial results of the data augmentation process.
2.3. Data Annotation
The annotation of the dataset was completed using the LabelMe semantic segmentation annotation tool, with the annotation process strictly adhering to the annotation specifications for instance segmentation tasks: annotators manually traced the mask regions pixel-by-pixel for each rapeseed male parent target within each image, precisely delineating the effective coverage area of the rapeseed male parent while excluding background interference factors such as soil, weeds, and field debris. Upon completion, the tool automatically generated YOLO-format text annotation files containing key information including mask coordinates, target category labels, and original image parameters. This ensured precise correspondence between mask regions and rapeseed male parent targets, providing high-quality, reliable supervised learning data for subsequent segmentation model training.
Figure 3 illustrates the data annotation process using LabelMe.
3. Navigation Line Detection Method
3.1. Improved Rapeseed Row Extraction Model
Addressing the challenges of cross-growth between parental rows and lines in rapeseed seed production fields, severe weed interference, and the inherent limitations of traditional YOLOv8-Seg in capturing parental row features and preserving boundary details, this study implements targeted enhancements based on YOLOv8-Seg. Unlike existing single-point enhancement strategies for YOLO models, this research establishes a comprehensive feature optimisation paradigm spanning core processes from backbone feature extraction to Neck layer feature fusion. First, the conventional downsampling convolutional layer in the backbone network is replaced with the lightweight ADown module. Its “average pooling + dual-branch parallel” architecture minimises computational overhead while maximally preserving critical boundary information of parent rows, addressing the loss of fine features caused by traditional downsampling. Secondly, the conventional upsampling module in the Neck layer is replaced with the dynamic upsampler DySample. This generates upsampling parameters adaptively based on input features, superseding fixed interpolation rules to enhance edge feature reconstruction precision. Concurrently, the C2f module preceding the detection head in the Neck section is replaced with the lightweight partial convolution C2f_FB. This achieves streamlining through feature channel compression and branch-parallel computation, enhancing feature extraction efficiency while reducing parameter count. This enhancement strategy represents not a simple substitution of existing modules but rather achieves the synergistic objectives of “minimising feature loss—refining feature expression—optimising computational efficiency” through complementary functionalities across modules. The improved network architecture is illustrated in
Figure 4.
The proposed collaborative optimisation framework comprises three modules: ADown, DySample, and C2f_FB. This is not a simple amalgamation of existing lightweight components, but rather a methodologically designed approach tailored to address three core constraints identified during rapeseed parent row navigation. These modules form a progressive collaborative mechanism of “feature fidelity—edge enhancement—computational efficiency optimisation”, achieving a fundamental leap from segmentation model refinement to enhanced navigation system performance.
The three enhancement modules proposed herein (ADown, DySample, and C2f_FB) are not independently optimised but constitute a collaborative system addressing key challenges in rapeseed parent row detection. 1. ADown Module: A lightweight downsampling scheme designed under boundary fidelity constraints for navigation. To resolve the loss of rapeseed parent row boundaries caused by traditional downsampling—which leads to navigation line fitting deviations exceeding tolerance thresholds—an “average pooling + dual-branch parallel” architecture is employed. This design reduces parameter count by 30% while increasing boundary feature retention to 92%, directly addressing navigation failures caused by boundary blurring. The design explicitly focuses on “boundary feature integrity” under navigation constraints, rather than merely optimising segmentation metrics. 2. DySample Module: A dynamic oversampling design developed under navigation edge precision constraints. To resolve local navigation line shifts caused by blurred rapeseed parent row edge features due to fixed oversampling, this module dynamically generates oversampling parameters based on input features. This approach elevates edge feature reconstruction precision to 89% while reducing local navigation line deviation by 40%, establishing a direct correlation between “segmentation edge precision” and “navigation alignment precision”. 3. C2f_FB Module: A lightweight feature fusion solution developed under real-time navigation constraints. To overcome computational limitations in agriculturally embedded devices, a local convolution mechanism reduces computational load by 40% while maintaining multi-scale feature fusion efficiency. This mechanism stabilises the model frame rate above 90 FPS, fulfilling “real-time requirements” under navigation constraints. The three modules synergistically form a methodological framework: “navigation performance constraints—customised module design—end-to-end performance optimisation”.
3.1.1. Lightweight Downsampling Convolution Module ADown
The core design objective of this module is to address the challenge of “navigation deviation exceeding tolerance limits due to boundary information loss” in rapeseed parent row navigation, whilst balancing lightweight implementation with the integrity of boundary features. Due to the intermingled growth of rapeseed parental lines in the field and significant weed interference, the boundary details of rapeseed paternal rows are crucial for navigation precision. However, traditional stride-2 convolution downsampling tends to lose such critical information, while real-time navigation imposes stringent computational demands on the model. These issues directly impact the precision and stability of navigation path extraction. To address this, the present study employs the ADown lightweight downsampling module, whose scene adaptability and core advantages are as follows. Centred on an architecture of “average pooling and dual-branch parallel processing”, this module achieves spatial dimension reduction and background noise smoothing through preliminary downsampling via AvgPool2d, while concurrently preserving the overall contour trends of crops to effectively counteract weed interference. The dual-branch design splits features along the channel dimension: a 3 × 3 convolution captures local boundary details, while MaxPool2d combined with a 1 × 1 convolution preserves global structural features. This avoids the boundary information loss inherent in traditional downsampling, enhancing paternal row segmentation precision in complex environments. Concurrently, parallel processing and lightweight design balance precision with computational efficiency, reducing parameters by approximately 30% to accommodate real-time field computing demands. The fusion of local and global features further enhances the model’s adaptability to varying parental growth states, significantly improving row extraction robustness and providing critical support for precise navigation in rapeseed fields. The ADown network architecture is illustrated in
Figure 5.
3.1.2. Lightweight Dynamic Oversampling DySample
The core design objective of this module is to address the challenge of “localised deviation in navigation lines caused by blurred edge features” during rapeseed parent row navigation, thereby achieving precision and adaptability in feature reconstruction. Traditional sampling methods, reliant on fixed parameters, often result in blurred edge features along the paternal rows of rapeseed. The precision of these edge features directly impacts the quality of crop row segmentation and subsequent navigation line fitting, rendering them ill-suited to the complex demands of field scenarios. To address this, this study introduces the DySample lightweight dynamic upsampling module. Its core innovation lies in dynamically generating upsampling parameters from input features, replacing fixed interpolation rules. Through dynamic parameter modulation and feature-adaptive interpolation, it achieves high-precision feature reconstruction, effectively preserving subtle edge features of rapeseed parent rows and enhancing target-background discrimination. Concurrently, this module combines lightweight advantages with high efficiency, meeting real-time field processing demands while ensuring enhanced segmentation precision. It is therefore suitable for replacing traditional upsampling modules in rapeseed crop row extraction tasks. The DySample network architecture is illustrated in
Figure 6.
3.1.3. Lightweight Partial Convolutional C2f_FB
The core design objective of this module is to address the challenge of insufficient real-time performance due to computational constraints in rapeseed parental row navigation, achieving a balance between feature fusion efficiency and lightweight implementation. To enhance the detection head’s efficiency in extracting characteristics from rapeseed paternal rows while controlling computational complexity to meet the processing demands of real-time field navigation systems, this study replaces the original C2f module with the lightweight C2f_FB partial convolutional module. Building upon the C2f architecture, this module incorporates partial convolution mechanisms. By implementing feature channel compression, branch parallel processing, and partial-region convolution operations, it reduces redundant computations to achieve lightweight optimisation. Compared to the original C2f module, it achieves approximately 40% lower computational load and 35% fewer parameters, without significantly compromising its ability to extract key features such as rapeseed parent row contours and texture of rapeseed paternal rows. This effectively enhances inference speed on embedded devices, perfectly meeting the dual demands for efficiency and lightweight design in real-time navigation scenarios within rapeseed fields. The C2f_FB network architecture is illustrated in
Figure 7.
3.2. Navigation Line Fitting Method
Mainstream navigation line fitting algorithms include the least squares method [
30] and the Hough transform [
31], both of which are applicable to the navigation line fitting task in this study. However, the least squares method is only suitable for fitting ideal linear crop rows and assumes observation errors are unidirectional (occurring solely along the
y-axis), which does not align with the actual situation where rapeseed parent rows may exhibit curvature due to uneven sowing or wind disturbance. The Hough transform demonstrates greater robustness to noise but offers lower curve fitting precision and consumes substantial computational resources, compromising real-time navigation performance. In contrast, polynomial fitting offers greater flexibility in describing non-linear crop row distributions, accommodating the slight curvature of rapeseed parent rows in the field.
Addressing phenomena such as interlaced growth of paternal and maternal rows in rapeseed seed production fields, localised gaps caused by uneven sowing of paternal rows, and inconsistent growth patterns, this paper proposes a polynomial fitting method based on an improved YOLOv8-seg model to extract navigation lines. The application of polynomial fitting in this study rests upon three core assumptions: (1) The spatial distribution of rapeseed parental rows in the field conforms to low-order polynomial curve characteristics (a quadratic polynomial was employed herein), with the curvature of rows and plants remaining within a narrow range that does not impede agricultural machinery navigation; (2) edge feature points extracted from rapeseed parent row segmentation masks possess high reliability, with noise points (e.g., weed interference) remaining within manageable limits; and (3) fitting errors between observation points and polynomial curves follow a normal distribution, enabling optimal fitting results through minimising the sum of squared orthogonal distances between observation points and the curve.
Principle of polynomial fitting: Given a series of observation points (x
i, y
i) (i = 1, 2, 3…N), the fitting formula is defined as
y = ax
2 + b
x + c (quadratic polynomial). The objective function is the sum of the squared orthogonal distances from all observation points to the quadratic polynomial curve, expressed as follows:
The denominator originates from the formula for the orthogonal distance from a point to a conic section: for the conic section y = ax2 + bx + c, the slope of the tangent line at the point (, ax2 + bx + c) is . The denominator is derived from this slope, ensuring the fitting process minimises the true orthogonal distance (perpendicular to the curve itself), rather than the projected distance perpendicular to the X-axis.
When crop rows in the field approximate a straight line (curvature ≈ 0), the quadratic term coefficient a approaches 0, the fitting formula degenerates to , and the objective function is simplified to: . This simplified function is entirely equivalent to linear orthogonal fitting (minimising the distance perpendicular to the line); hence, linear fitting constitutes a special case of quadratic polynomial fitting. It can accommodate crop rows exhibiting different geometric configurations, such as straight lines or mildly curved patterns. When f attains its minimum value, parameters a, b, and c represent the optimal fitting parameters, achieving a quadratic polynomial orthogonal fit for the observation points.
This method suppresses the impact of random noise through global data fitting, generating a smooth navigation line while effectively addressing curvature issues arising from non-standard planting practices, thereby exhibiting enhanced robustness. The key steps of the polynomial fitting algorithm for obtaining navigation lines are as follows:
Binarisation processing: The original image of the rapeseed paternal row (
Figure 8a) undergoes instance segmentation via the SegNav-YOLOv8n model to obtain segmentation results (
Figure 8b). This is subsequently binarised to produce the final binarised image (
Figure 8c).
Edge extraction of rapeseed parent row: applying the Canny edge detection algorithm to the masked region of the target ridge surface extracts its contour, yielding high-precision ridge boundary information (as shown in
Figure 9).
Navigation line fitting: Establish a coordinate system based on the image dimensions. Perform quadratic polynomial fitting on the boundary points extracted via Canny edge detection to generate navigation lines. Plot the fitted lines onto the image; the fitting results are shown in
Figure 9b.
5. Conclusions
This study addresses the field management requirements for rapeseed seed production by proposing a SegNav-YOLOv8n-based navigation line detection method for rapeseed parental rows. Significant improvements were achieved through model structure optimisation and enhanced navigation line fitting strategies. Key findings are as follows: Using YOLOv8n-seg as the baseline, the innovative integration of three core modules—ADown, DySample, and C2f_FB—constructs a segmentation model combining high performance with lightweight characteristics, establishing a distinct technical approach from existing crop row detection models. Unlike existing studies focusing on single-dimensional improvements in either lightweight design or precision, this research achieves a three-dimensional equilibrium of “precision—speed—lightweight design” through coordinated module optimisation: The ADown module reduces parameter count by 30% while preserving critical boundary information of parent rows, effectively suppressing weed interference and subsampling information loss, thereby resolving the traditional subsampling dilemma of balancing efficiency and detail. The DySample module enhances target-background discrimination through dynamic upsampling, improving edge feature reconstruction precision and overcoming the limitations of fixed interpolation upsampling in complex scenarios. The C2f_FB module reduces computational load by 40% and parameter count by 35%, balancing feature extraction efficiency with lightweight requirements. This overcomes the drawbacks of excessive parameters and computational demands in traditional feature fusion modules. Comparative experiments demonstrate that SegNav-YOLOv8n achieves an average precision of 99.2% and a mean average precision of 84.5%, outperforming baseline models and approaching YOLOv9c-seg. Inference frame rate reaches 90.21 frames per second, matching lightweight models while significantly outperforming high-parameter YOLOv9 variants. This model possesses 2.6 million parameters, a model size of 5.5 MB, and a computational load of 11.4 G, achieving the lowest resource consumption among all tested models. These lightweight characteristics lay the foundation for its embedded development and deployment in agricultural machinery. Subsequent research will focus on conducting specialised embedded deployment testing to validate its practical applicability in real-world agricultural machinery scenarios.
The segmented rapeseed parent row masks from SegNav-YOLOv8n were binarised, followed by edge contour extraction using the Canny edge detection algorithm. Finally, a navigational line was fitted using polynomial regression. Navigation line error analysis was conducted on 100 images extracted from the test set. The maximum lateral deviation was recorded based on the horizontal pixel distance between the fitted navigation line and the manually observed navigation line. The results showed deviations within 7.45 pixels (≈19.37 cm), with an average deviation of 3.35 pixels (≈8.71 cm), demonstrating the high precision of the navigation line extraction method employed herein. This research achieved favourable results in detecting rapeseed paternal rows, though certain aspects warrant refinement. For instance, model performance diminishes under extreme low-light conditions or heavy occlusion. Practical field applications must also account for variations arising from different plots, varieties, and cultivation standards. Subsequent work will focus on multimodal information fusion and embedded model deployment design to further enhance the system’s robustness and practicality in complex agricultural environments, developing models with stronger environmental adaptability.
The core innovation of this study extends beyond resolving navigation line detection for rapeseed seed production paternal rows. Crucially, it proposes a “navigation-constraint-driven lightweight segmentation-fitting integrated approach”, achieving substantial progress in three key areas: 1. Transition from scene description to failure quantification: Complex field scenarios such as high-density planting of rapeseed paternal lines and weed interference were converted into quantifiable navigation failure criteria using an 11.5 pixel threshold (corresponding to ±30 cm in actual fields). This provides a reference for navigation research on densely planted, visually similar crop rows. 2. Method optimisation from module replacement to constraint design: Establishing a framework of “navigation performance requirements—Customised Module Development—End-to-End Collaborative Optimisation” framework. Through the synergistic operation of the ADown, DySample, and C2f_FB modules, it was validated that optimisation of lightweight segmentation models must precisely align with core navigation system requirements (e.g., boundary integrity and real-time response speed), rather than merely pursuing numerical improvements in segmentation metrics. This offers new insights for model design in agricultural machinery embedded navigation systems. 3. Providing guidance for navigation in similar crop scenarios: The proposed lightweight segmentation–polynomial fitting integrated approach demonstrates strong transferability for row navigation scenarios in crops like wheat, which share dense planting patterns and similar growth characteristics with rapeseed. This offers valuable reference for navigation technology development in such crops. Leveraging the advantages of modular design, this approach holds promise for adapting to the morphological characteristics of diverse crops. It offers a technically replicable and practically valuable solution for intelligent field management of densely planted crops within smart agriculture.