Impacts of Effects of Deﬁcit Irrigation Strategy on Water Use Efﬁciency and Yield in Cotton under Different Irrigation Systems

: Irrigated cotton ( Gossypium hirsutum L.) growers in the Murray-Darling Basin (MDB) of Australia, are challenged by limited water availability. This modelling-study aimed to determine if deﬁcit irrigation (DI) practices can potentially improve water use efﬁciency (WUE) for furrow irrigation (FI), overhead sprinkler irrigation (OSI) and subsurface drip irrigation (SDI) systems. We validated the Agricultural Production System sIMulator (APSIM) against observed cotton lint yield and crop biomass accumulation for different management practices. The model achieved concordance correlation coefﬁcients of 0.93 and 0.82 against observed cotton crop biomass accumulation and lint yields, respectively. The model was then applied to evaluate the impacts of different levels of DI on lint yield, WUE across cotton growing locations in the MDB (Goondiwindi, Moree, Narrabri, and Warren), during the period from 1977 to 2017. The different levels of DI for the FI system were no irrigation, full irrigation (TF) and irrigated one out of four, one out of three, one out of two, two out of three and two out of four TF events. For the OSI and SDI systems, DI levels were no irrigation, TF, 20% of TF, 40% of TF, 60% of TF and 80% of TF. Lint yield was maximised under the OSI and SDI systems for most locations by applying 80% of TF. However; modelling identiﬁed that WUE was maximised at 60% of full irrigation for OSI and SDI systems. These results suggest there are signiﬁcant gains in agronomic performance to be gained through the application of DI practices with these systems. For FI, DI had no beneﬁt in terms of increasing yield, while DI showed marginal gains in terms of WUE in some situations. This result is due to the greater exposure to periodic water deﬁcit stress that occurred when DI practices were applied by an FI system. The results suggest that in the northern MDB, water savings could be realised for cotton production under both OSI and SDI systems if DI were adopted to a limited extent, depending on location and irrigation system.


Introduction
Approximately 90% of the Australian cotton (Gossypium hirsutum L.) crop is grown in the Murray-Darling Basin (MDB), particularly in northern New South Wales (NSW) and southern Queensland (QLD) in both irrigated and dryland cropping systems [1][2][3]. Approximately 83% of cotton grown within the MDB is irrigated, with the remaining 27% being dryland [4][5][6]. Irrigation is used to mitigate the risk of inadequate rainfall, particularly during the vegetative stages of cotton growth [7]. Water availability is one of the most important determinants of the total area of cotton grown under irrigation in the MDB [8,9]. During severe droughts, such as occurred during 2003-2004 and 2007-2008, water availability for irrigation was significantly reduced, which affected total cotton production within the MDB [1,10,11].
In the MDB, 80% of irrigated cotton is grown using furrow irrigation (FI), 14% using overhead sprinkler irrigation (OSI) and 6% using subsurface drip irrigation (SDI) [4,[12][13][14][15]. Furrow irrigation systems have significantly lower capital and energy costs, ease of operation and possibly management costs compared to the other systems [15]. The OSI and SDI systems allow cotton crops to be grown with relatively less irrigation water, which can result in higher water use efficiency (WUE) and higher marginal water use efficiency (MWUE) [16,17].
Limited water availability and the cost of water have resulted in researchers seeking to improve WUE through modifications to the farming system or better in crop management practices such as deficit irrigation (DI) [18]. Deficit irrigation is defined as the practice of applying irrigation less water than what is required to fully replace water used by the crop. It exposes the crop to periodic water deficit while enhancing WUE and MWUE [3,[19][20][21]. The aim of DI practices is to maintain soil water at a level that does not significantly limit crop yield while not completely filling the soil profile [19]. A DI strategy has the advantage over other limited irrigation strategies as it still ensures water is applied over the full growing cycle of the crop and hence minimises significant water deficit stress at critical growth stages will still reduce water use overall. Whilst the concept of DI practices for cotton was proposed in the 1970s [20], to date, it is not a common practice in the MDB. A number of studies have shown that DI practices increase yield and WUE [21,22] while other studies showed no improvements or decreases in these indicators [19,23,24]. Properly managed, DI can sustain profits while reducing irrigation water usage during periods of reduced irrigation water availability [25]. This can be demonstrated in situations where water is saved by not applying additional irrigation water when it only results in a marginal increase in yield [19,24].
Biophysical modelling can be used to identify optimum management practices (including water application strategies) and the impact of seasonal variation on crop yields [26,27]. The Agricultural Production Systems sIMulator (APSIM) is a biophysical crop model used for exploring complex crop, resource and management interactions within cropping systems [28,29]. This model operates on a daily time step integrating the supply and demand of resources (water, nutrients, light and heat) to predict crop growth, development and yield processes. It has a pedigree in simulating cropping systems and the interactions between farm resources (soil properties, water, nutrients, and weather) and crop management practices in Australia [28,29]. The model includes modules for a range of crops (including cotton), soils, climate, irrigation management, and crop management practices [28,30,31] making it an ideal tool to investigate complex farming systems questions such as irrigation practices. Using APSIM, we tested the proposition that the application of DI can decrease water use without a significant decrease in lint yield, leading to an increase in WUE and MWUE. The objective of this research was to model different levels of DI under three different irrigation systems across four locations within the northern MDB to determine the effects on lint yields, WUE, and MWUE.

APSIM Cotton Modelling
The APSIM model (Version 7.9) [28] with the cotton module [32] was used to simulate cotton crop biomass accumulation and lint yield. Input parameters for each simulation contained soil information, a description of crop management, including irrigation, and meteorological inputs. To ensure confidence in the model outputs APSIM was validated for cotton crop biomass accumulation and lint yield against observed irrigated field experiments [1] and regional irrigated and dryland cotton production data [33], prior to modelling the impact of DI across irrigation systems and locations in the MDB.

Validation Data
To validate the APSIM cotton model for its response to DI, data presented in figures describing cotton crop biomass accumulation under contrasting irrigation regimes, from Cammarano et.al. [1], were digitised. There were 139 data points from two growing seasons (2007-2008 and 2009-2010) (Table 1). Irrigation treatments occurred when soil moisture was depleted to 50% of plant available water capacity (50% of PAWC), 60% of PAWC, 70% of PAWC, and 85% of PAWC. Soil parameters were sourced from data provided in Cammarano et al. [1]. Daily meteorological patch point data were sourced from the Scientific Information for Land Owners (SILO) database (https://www.longpaddock.qld. gov.au/silo/) [34]. Crop management was set to reflect the management described with each treatment given in Cammarano et al. [1]. The data used to evaluate predictions of lint yield by APSIM across a range of locations and management practices were sourced from annual reports of the Cotton Seed Distributors cultivar evaluations [33]. This data set covered 27 locations ( Figure 1) and 111 individual cotton crops over seven years (from 2009 to 2015). They included crop management information and lint yield. For the simulations of lint yield, soil parameters were sourced directly from the APSoil database for each of the 27 locations. Crop management was set to reflect the management described for each location within the reports CSD [33].

Statistical Analysis of Validation Simulation Results
In the first instance, modelled results were visually compared to observed data for accumulated crop biomass from Cammarano et al. [1] and lint yield from CSD [33]. To assess the accuracy with which APSIM predicted cotton crop biomass and lint yield, several statistics were also calculated. The first was R 2 (coefficient of determination), which measures the proximity of data to the regression line [35]. The second statistic was the mean bias as an assessment of the difference between the mean of observed values and the mean of modelled values [35]. Third, model efficiency was calculated as an assessment of the predictive power of the model to explain the variation between observed and modelled values [35]. Fourth, the root mean square error (RMSE) was calculated as an assessment of the estimate of the amount of error between modelled values and observed values [36]. Finally, the concordance correlation coefficient was calculated as an assessment of the precision and accuracy of the model [35].

Simulation of Deficit Irrigation (DI) Practices
To assess the impacts of DI on cotton crops, we ran a series of single season simulations for four locations from the north to the lower north areas of the MDB (Goondiwindi, Moree, Narrabri, and Warren; Table 2), spanning 41 years from 1977 to 2017. Three cotton irrigation systems were represented. These were furrow irrigation (FI), overhead sprinkler irrigation (OSI), and subsurface drip irrigation (SDI) systems. Management parameters for the DI simulations were sourced from a range of industry publications [37][38][39][40]. Sowing depth was consistent across all four locations at 65 mm [37]. Crop management reflected current industry best practice within the MDB. The cultivar used at each location was Sicot 71BRF [38,41]. For Goondiwindi, Moree, Narrabri, and Warren, sowing dates were 23 October, 15 October, 11 October, and 13 October [39,40] respectively, and plant densities were 9.5, 9, 10.5, and 10 plants/m 2 , respectively [37]. For all locations, nitrogen fertiliser as urea was applied at planting (100 N kg/ha), with 50 N kg/ha applied on 15 December, and 25 January [42,43]. Daily meteorological data from 1977 to 2017 for each location were obtained from the SILO database as Patched Point Datasets [44]. Soil properties, including plant available water capacity (PAWC), were obtained from the APSoil database [45] and are provided in Table 2.
Different DI practices were used for FI, OSI, and SDI systems reflecting the different capabilities of each irrigation system. Suitable DI practices for the three irrigation systems were obtained from interviews with two irrigation researchers: Professor Steven Raine, 2018, face-to-face, 17 October and Associate Professor Joe Foley, 2019, face-to-face, 3 January. The furrow irrigation (FI) system was simulated by applying comparatively large volumes of water infrequently. The refill point of the FI system under full irrigation treatment (TF) was a soil water deficit of 100 mm. For the FI system, DI was simulated by skipping specific irrigation events (Table 3) with treatments coded and subsequently referred to as T1 through to T5. The OSI and SDI systems were simulated by applying frequent applications of small volumes of water. The trigger points for an irrigation application under the TF treatment were soil water deficits of 50 mm and 20 mm for the OSI and SDI systems, respectively. Deficit irrigation treatments levels were 20%, 40%, 60%, and 80% of TF application volumes (Table 3) and were coded and subsequently referred to in a similar fashion. Table 2. Summary of soil and climate details for each location used in the DI simulations. Soil water properties are drained upper limit, (DUL; mm); soil water capacity at 1500 kPa drained lower limit (DLL; mm); and plant available water capacity (PAWC; mm). Soil properties are the totals from the surface to 1.8 m soil depth and monthly climatic averages are for the 41 years from 1977 to 2017.

Locations
Lat where, Y represents cotton-lint yield (kg/ha), T W represents total water applied (in-crop rainfall plus irrigation; mm/ha). Marginal water use efficiency (MWUE) was calculated as the difference between irrigated yield and dryland yield relative to irrigation water applied. The MWUE was calculated as [19] where Y i represents lint yield (kg/ha) with irrigation, Y d represents lint yield (kg/ha) without irrigation and W I represents water irrigation applied (mm). For each irrigation treatment, the coefficient of variation (CV) was used as a measurement of relative variability, comparing the degree of difference relative to the mean [47].

Model Validation
Validating the APSIM cotton model against biomass data given by Cammarano et al. [1] showed that the model was capable of predicting cotton biomass accumulation under a range of deficit irrigation practices (Figure 2). The summary statistics comparing modelled biomass to observed biomass were R 2 : 0.68, mean bias: 1476, model efficiency:  The APSIM model accurately predicted lint yield across the range of different environments, years, and crop management practices (Figure 3). The model slightly over-predicted yield in the lower yield potential simulations and slightly under-predicted lint yield in the higher yield potential simulations. The summary statistics comparing modelled yield to observed yield were R 2 : 0.86, mean bias: 32, model efficiency: 0.99, root mean square error: 280 and concordance correlation coefficient: 0.93. These summary statistics indicate that the model has an acceptable level of accuracy in representing cotton lint yield processes across a wide range of environments.

The Effects of Deficit Irrigation Practices on Lint Yield
For FI, the TF treatment achieved the greatest lint yield across all locations (Figure 4) indicating that for this irrigation system, where irrigation applications are infrequent, increasing water availability drives cotton yields. For this treatment, the median lint yields at Goondiwindi, Moree, Narrabri and Warren were 2427, 2667, 2800 and 2089 kg/ha, respectively; the average water applied was 7.35, 8.93, 8.47 and 6.30 ML/ha, respectively. As the amount of irrigation water applied increased, the CV of lint yield tended to decrease as the influence from rainfall variability decreased. The lowest CVs for Goondiwindi, Moree, Narrabri and Warren were 36% (for the TF treatment), 33% (TF), 27% (T5 and TF), and 27% (T5), respectively. For the OSI system, across all study locations, the greatest median lint yield was achieved with 80% of TF treatment. This indicates that the practice of frequent applications of water, even at deficit levels, limits the exposure of the crop to yield limiting water deficits. For this treatment, the median lint yields at Goondiwindi, Moree, Narrabri and Warren were 3374, 3334, 3231 and 3179 kg/ha, respectively while the average water applied was 7.95, 9.35, 8.80 and 7.22 ML/ha, respectively. The 80% TF treatment saved, compared to the TF treatment, 1.33, 1.40, 1.20 and 1.38 ML/ha of water for Goondiwindi, Moree, Narrabri and Warren, respectively. As the amount of water applied increased under OSI, the CV of lint yield tended to decrease due to a reduced reliance on rainfall to meet water demands. The lowest CVs for Goondiwindi, Moree, Narrabri and Warren were 26% (at 80% of TF), 24% (at 80% of TF and TF), 23% (at 60% and 80% of TF) and 20% (at 80% of TF), respectively. For the SDI system, the greatest median lint yield was achieved with the 80% of TF treatment at Goondiwindi and Narrabri, and 60% of TF treatment at Moree and Warren. This indicates that higher rainfall environments are able to compensate for the greater potential for exposure to water deficit with more restrictive DI practices applied with this irrigation system. For these treatments, the median lint yields at Goondiwindi, Moree, Narrabri, and Warren were 3568, 3470, 3296 and 3279 kg/ha, respectively; the average water applied was 8.50, 7.85, 9.38 and 5.90 ML/ha, respectively. The best performing treatments in terms of lint yield resulted in water savings of 1.28, 3.10, 1.03 and 2.45 ML/ha compared to the TF treatments for Goondiwindi, Moree, Narrabri and Warren, respectively. The lowest CVs for Goondiwindi, Moree, Narrabri and Warren were 20% (60% of TF), 18% (60% of TF), 34% (80% of TF) and 16% (60% of TF), respectively.

The Effects of Deficit Irrigation Practices on Water Use Efficiency
For the FI system, the greatest median WUE was achieved with the TF treatment at Goondiwindi, Moree and Warren, and T5 treatment at Narrabri ( Figure 5) indicating the importance of limiting the exposure to water deficit stress on the agronomic performance of a FI system. For these treatments, the median WUE was 3.0, 2.9, 3.5, and 3.5 kg lint/mm, at Goondiwindi, Moree, Narrabri and Warren, respectively; the average total water applied was 7.34, 8.97, 7.57 and 6.42 ML/ha, respectively. As the amount of water applied increased, the CV of WUE tended to decrease as the influence of rainfall variability on the yield component of the WUE equation decreased. The lowest CVs for Goondiwindi, Moree, Narrabri and Warren were 31% (T5 and TF), 25% (TF), 23% (TF), and 19% (TF), respectively. For the OSI system, the greatest median WUE was achieved with the 80% of TF treatment at Goondiwindi and Warren, and 60% of TF treatment at Moree and Narrabri. This was driven by a relatively greater decrease in the T w component of the WUE equation compared to the yield component. For these treatments, the median WUEs at Goondiwindi, Moree, Narrabri and Warren were 4.0, 4.0, 4.2, and 4.3 kg lint/mm, respectively; the average total water applied was 7.69, 7.57, 7.37, and 7.21 ML/ha, respectively. The lowest CVs for Goondiwindi, Moree, Narrabri and Warren were 22% (60% of TF), 20% (80% of TF and TF), 21% (60% and 80% of TF), and 16% (80% of TF), respectively. For the SDI system, the greatest median WUEs were achieved with 40% of TF treatment at Goondiwindi, Moree and Narrabri and 60% of TF treatment at Warren. For these treatments, the median WUEs at Goondiwindi, Moree, Narrabri, and Warren were 5.2, 5.2, 5.4 and 5.6 kg lint/mm, respectively; with the average total water applied 5.74, 6.47, 5.96 and 5.97 ML/ha, respectively. The lowest CVs for Goondiwindi, Moree, Narrabri and Warren were 20% (60% of TF), 18% (40% and 60% of TF), 22% (40% of TF), and 20% (80% of TF), respectively.

The Effects of Deficit Irrigation Practices on Marginal Water Use Efficiency
For the FI system, the greatest median MWUE was achieved with the T5 treatment at Goondiwindi and Narrabri, and the TF treatment at Moree and Warren ( Figure 6) due to this system requiring large volumes of irrigation water to maximise its performance metrics. For these treatments, the median MWUEs at Goondiwindi, Moree, Narrabri and Warren were 4.5, 3.6, 5.2 and 4.6 kg lint/mm irrigation, respectively; and the average irrigation water applied was 3.16, 5.25, 3.97 and 3.82 ML/ha, respectively. As the amount of irrigation water increased, the CV of MWUEs tended to decrease as variance in the irrigated yields decreased. The lowest CV for Goondiwindi, Moree, Narrabri and Warren were 40% (T5), 58% (T4), 30% (TF), and 29% (TF), respectively. For the OSI system, the greatest median MWUE was achieved with the 80% of TF treatment at Goondiwindi and Warren, and 60% of TF treatment at Moree and Narrabri due to the greater relative reduction in irrigation water applied compared the relative reduction in yield. For these treatments, the median MWUEs at Goondiwindi, Moree, Narrabri and Warren were 6.2, 5.9, 6.1, and 5.7 kg/mm, respectively. These values were achieved with an average irrigation water use of 4.14, 3.95, 3.66, and 4.51 ML/ha, respectively. The lowest CVs for MWUE at Goondiwindi, Moree, Narrabri and Warren were 29% (60% of TF), 29% (TF), 42% (80% of TF), and 15% (80% of TF), respectively. For the SDI system, the greatest average MWUE was achieved with the 40% of TF treatment across all locations indicting that this system of the three investigated is potentially the best adapted to deficit irrigation practices. For this treatment, the median MWUEs at Goondiwindi, Moree, Narrabri, and Warren were 10.1, 9.1, 9.5 and 9.1 kg/mm, respectively; with an average irrigation water use of 2.22, 2.76, 2.32 and 2.09 ML/ha, respectively. The lowest CVs for Goondiwindi, Moree, Narrabri and Warren were 25% (at 40% of TF), 25% (at 40% of TF), 43% (at 40% of TF), and 18% (at 60% of TF), respectively.

Discussion
Before applying APSIM to model DI practices, we validated the model for its ability to model deficit irrigation (DI) through comparisons to observed crop biomass ( Figure 2) and lint yield (Figure 3) from both experimental field data and industry production data across the MDB. We showed that the APSIM crop model was able to accurately predict crop biomass and lint yield. This is in alignment with previous reports. Carberry et al. [30] previously showed that the APSIM model accurately simulated a range of different crops, including cotton grown in Dalby, QLD., Australia. Earlier versions of the APSIM cotton model (APSIM-OZCOT) accurately predicted lint yield in two locations in Xinjiang, northwest China [27]. The robustness of the APSIM's ability to represent croton growth and yield processes is testament to the suitability for modelling cotton crops and the impacts of management practices across the MDB. We, therefore, proceeded to use APSIM cotton to investigate the use of DI practices under three irrigation systems for four locations within the northern MDB.
Our results showed that modelled lint yield, WUE and MWUE under the FI system were lower than for the SDI (c.a. 27%, 8%, and 10% lower respectively for the TF treatment) and OSI (c.a. 25%, 8%, and 7% lower respectively for the TF treatment) systems (Figures 4-6). This finding through modelling is supported by Lowien and Gall [17] who reported that field data from both OSI and SDI systems demonstrated greater yields (4% increase) and WUE (7% increase) compared to FI systems. Darouich et al. [16] also ascertained that irrigation by SDI systems produced approximately 42% higher lint yields than FI systems for the same amount of water used. Our modelling with APSIM suggests that the infrequent application of irrigations under FI systems results in periodic water deficit stress in the cotton crop, which reduces plant growth and leads to a reduction in lint yield.
Furthermore, our results indicated no beneficial effects of DI for FI systems in terms of lint yield. The findings of Wen [22] from field experiments in Southwest Texas, USA support our modelling in that there was no benefit for lint yield from DI practices under the FI system. However, there may be some small improvements in WUE and MWUE with DI under FI. In Goondiwindi and Narrabri, the T5 treatment led to the best outcomes for each of these measure (Figures 4-6). There is likely an environmental influence on the suitability of deficit irrigation with a FI system with the practice potentially being better suited for environments with greater rainfall than those investigated in this study. Kaman et al. [48] reported that, when using DI in field experiments in Turkey, cotton yield was reduced, but WUE improved under a FI system. Their final conclusion aligns with our modelling results where the greatest lint yield will be achieved under a FI system using the TF treatment. This result is also supported by other researchers [48][49][50]. We conclude that the application of DI practices within a FI system results in lower lint yield because of crop stress due to the periodic wet and drying cycles. For the FI system, the results do not support our hypothesis that DI practices do not decrease lint yield in the MDB. However, in some higher rainfall situations, our hypotheses that DI achieved improvements with FI in terms of WUE and MWUE, is supported. Some deficit irrigation practices had positive effects on predicted lint yield, WUE, and MWUE under the OSI system as the impact on yield was either positive or when negative was relatively minor compared to the reduction in T w and W I . Our results indicated that, under an OSI system, lint yields, WUE and MWUE were all optimised with either 60% or 80% of the TF treatments (Figures 4-6). The 80% of TF treatment compared to the TF treatment, saved between 1.20 and 1.40 ML/ha of irrigation water across the four study locations with minimal negative impact on yield. Study location had an effect on the actual optimisation point. Evans and Sadler [51] reported that both yield and WUE increased significantly when using deficit irrigation practices under an OSI system. Our modelling with APSIM and field observations by Evans and Sadler [51] support our hypothesis that, under an OSI system, the use of DI practices can increase lint yields, WUE and MWUE. For the SDI system, DI practices had a positive effect on lint yield, WUE, and MWUE for all study locations, but there is some variation in the best DI treatment amongst locations. These differences may be dues to variation in average annual precipitation rates, especially lower summer rainfalls at Goondiwindi and Warren. Our results show that, under an SDI system, lint yields will be maximised by the use of 60% to 80% of the TF treatment DI practice. However, WUE and MWUE will be optimised with more restrictive applications of water, around 40% or 60% of TF (Figures 4-6). The 40% of TF treatment compared to the TF treatment, saved between 2.09 and 2.76 ML/ha of irrigation water across the four study locations indicating a potential for DI practices to reduce water use when applied with an SDI system. The difference between the treatments that optimised yield and the treatments that maximised WUE and MWUE for the SDI system but not for the OSI and FI system suggests that there are system design aspects that demine if it is possible to optimise both yield and efficiency metrics through DI practices. Field experiments investigating DI for cotton grown with SDI systems identified that 80% of TF treatment achieved the greatest cotton yield [52]. Our modelling with APSIM and the findings of Sampathkumar et al. [52] from field research support our hypothesis that, under SDI system, the use of DI can increase lint yields and improve both WUE and MWUE by applying the optimum application of water.
Our results also demonstrate the effect of irrigation system choice on overall yields, WUE and MWUE. Yields, WUE and MWUE under the FI system with full irrigation were lower than yields under the OSI or SDI treatments with equivalent volumes of water application. Clearly, the increasing the utilisation OSI and DSI irrigation systems across the industry will lead to increased yields and improved WUE. This presents a significant opportunity to improve the productivity and WUE of the cotton industry, as 80% of cotton production occurs under furrow irrigation [4,[12][13][14][15].
While this paper has investigated DI practices from a yield, WUE and MWUE perspective, financial returns will be a key driver of adopting DI practices on farms. If costs associated with purchasing and delivery of water are low, then a practice of applying 80% of TF treatment should be the most financially beneficial practice under OSI and SDI systems, excluding the initial start-up/capital costs which need to be further examined. However, if water is limited, costly and/or pumping costs are high, then a practice of applying 60% of TF might be the most financially beneficial practice for these systems. Alternatively, water saved could be used to irrigate more land, which would increase total cotton production and potentially whole farm financial returns.

Conclusions
Differences in lint yield and WUE, induced by the use of DI practices under three irrigation systems of FI, OSI and FI have been modelled with APSIM across four Australian cotton growing locations, using the APSIM cotton model. Our results showed that FI systems result in lower yields, WUE and MWUE than both OSI and SDI systems in all locations; moreover, DI is only a useful strategy for OSI and SDI systems to further improve these efficiency indicators. Therefore, growers should consider adoption OSI or SDI with DI practices. These results suggest it might be possible to grow more cotton with the same amount of water and increase WUE, with the use of DI practices and OSI and SDI systems. In terms of CV, the optimal DI practices resulted in lower yield variation for these two systems across all locations suggesting that DI practices may also reduce inter-annual yield variance. Water saving may be realised as reduced water consumption for a given level of lint production, or it may allow growers to increase the irrigated area, potentially increasing total cotton yields. However, given the overwhelming existing FI infrastructure across the industry, consideration of start-up/capital replacement costs and other costs should be considered in future research.