Open Access
How to translate text using browser tools
14 October 2022 New Directions in Weed Management and Research Using 3D Imaging
April M. Dobbs, Daniel Ginn, Søren Kelstrup Skovsen, Muthukumar V. Bagavathiannan, Steven B. Mirsky, Chris S. Reberg-Horton, Ramon G. Leon
Author Affiliations +
Abstract

Recent innovations in 3D imaging technology have created unprecedented potential for better understanding weed responses to management tactics. Although traditional 2D imaging methods for mapping weed populations can be limited in the field by factors such as shadows and tissue overlap, 3D imaging mitigates these challenges by using depth data to create accurate plant models. Three-dimensional imaging can be used to generate spatiotemporal maps of weed populations in the field and target weeds for site-specific weed management, including automated precision weed control. This technology will also help growers monitor cover crop performance for weed suppression and detect late-season weed escapes for timely control, thereby reducing seedbank persistence and slowing the evolution of herbicide resistance. In addition to its many applications in weed management, 3D imaging offers weed researchers new tools for understanding spatial and temporal heterogeneity in weed responses to integrated weed management tactics, including weed–crop competition and weed community dynamics. This technology will provide simple and low-cost tools for growers and researchers alike to better understand weed responses in diverse agronomic contexts, which will aid in reducing herbicide use, mitigating herbicide-resistance evolution, and improving environmental health.

Introduction

Recent innovations in remote sensing and imaging technology have created unprecedented opportunities for growers to make data-driven decisions about integrated weed management (IWM). The use of digital imaging to map and monitor weeds allows growers to use precision weed control tactics at a field scale while accounting for spatial and temporal variation in weed populations (Andújar et al. 2016; Comba et al. 2019). Until recently, most remote sensing for weed detection and mapping has used 2D approaches such as normalized difference vegetation index (NDVI) and hyperspectral imaging (Andújar et al. 2018; Comba et al. 2019; Smith et al. 2018). However, 2D techniques lack depth information and cannot accurately estimate plant volume or biomass (size), nor can they detect lower layers of the plant canopy (Paturkar et al. 2020; Smith et al. 2018). Moreover, variations in light levels, overlapping leaves and stems, and shadows can make 2D techniques ineffective for distinguishing individual plants in the field (Smith et al. 2018; Zhang et al. 2016).

To mitigate these challenges, 3D imaging techniques are being developed that include the added dimension of depth (Armean et al. 2021). Cost-effective and efficient methods such as structure-from-motion (SfM) and stereo vision are being used to reconstruct 3D models of plants and plant canopies based on many images taken in succession. By incorporating depth data to create models of plant canopies in the field, 3D imaging offers researchers and growers a powerful tool for mapping spatial heterogeneity of weeds in response to management tactics (Armean et al. 2021; Zhang et al. 2016). In this article, we discuss applications of 3D photogrammetric imaging in weed management (e.g., weed detection and mapping for targeted removal), as well as weed research (e.g., modeling of weed–crop competition to predict yield loss). In addition, we discuss applications of 3D imaging for weed management in orchards and grasslands to illustrate the breadth of potential uses of this technology.

Table 1.

Comparison of photogrammetric techniques and light detection and ranging (LIDAR) for 3D imaging.

img-z2-2_641.gif

Figure 1.

Use of images taken from different angles to create a 3D reconstruction in structure-from-motion (SfM; top) vs. stereo-vision photogrammetry (bottom).

img-z2-3_641.jpg

Overview of 3D Imaging Techniques

Three-dimensional imaging uses sensor data obtained to generate models of objects based on point clouds, which are large data sets of points in 3D space (Comba et al. 2019). These can be created using laser-based techniques such as light detection and ranging (LIDAR) or by photogrammetric techniques using computer vision algorithms (Andújar et al. 2013; Comba et al. 2019; Paturkar et al. 2020; Table 1). LIDAR, which has been used in agriculture and forestry, emits laser pulses directed toward an object (e.g., a plant), which are reflected back to the sensor to obtain depth information (Andújar et al. 2013). Although LIDAR is extremely accurate and can be used to scan wide areas in the field, it requires sophisticated equipment and is costly (Andújar et al. 2013; Paturkar et al. 2020).

In contrast, photogrammetric techniques such as SfM and stereo vision are low-cost and efficient methods for creating 3D point clouds of objects (Armean et al. 2021; Paturkar et al. 2020; Table 1). In SfM, a single camera takes many 2D images of the object (e.g., a plant) from different viewpoints of known distances with a specified amount of overlap between images (Armean et al. 2021; Figure 1). A 3D point cloud is generated based on features that are matched from different images (i.e., views) of the canopy. The point cloud is a 3D model of the canopy, which can be used to infer canopy structure and biomass based on depth data (Armean et al. 2021; Paturkar et al. 2020; Figures 2 and 3). In stereo vision, a binocular camera takes two images of the plant at once and uses triangulation to create a depth map (Teng et al., 2021). The accuracy of both SfM and stereo vision can be affected by wind conditions in the field, which cause movement and overlap in the canopy (Dandrifosse et al. 2020; Paturkar et al., 2020). However, given their relative simplicity and low cost, photogrammetric techniques have numerous applications in weed management and weed research.

Figure 2.

Red, green, and blue (RGB) image of soybeans and weeds (A) and corresponding 3D point cloud reconstruction (B). Lower panels show point cloud reconstructions from different angles, including a top view (C), top view offset 45° from vertical (D), front view (E), under canopy and offset 45° (F), directly under canopy (G), facing canopy from behind (H), facing canopy offset 45° right (I), side view (J), and facing canopy offset 45° left (K).

img-z3-1_641.jpg

Figure 3.

Data pipeline for calculating canopy height and estimating biomass in the field using red, green, and blue (RGB) images and depth data.

img-z3-3_641.jpg

Applications of 3D Imaging in Weed Management

Weed Detection and Mapping

Timely postemergence weed control is critical in crop production (López-Granados et al. 2016). Applying herbicides when weeds are not at the appropriate phenological stage or uniformly without considering spatial variability over time can interfere with weed control (López-Granados et al. 2016; Swanton et al. 1999). For example, if weed control measures are applied too late in the season to control early-emerging weeds, this can result in weed escapes, increased soil weed seedbank, and evolution of herbicide resistance (Neve et al. 2010). Although weeds often occur in spatially aggregated patches of varying size, shape, density, and species composition, weed control tactics have traditionally been applied uniformly in fields due to the difficulty of monitoring populations in situ and in real time (Thornton et al. 1990).

Photogrammetric 3D imaging allows growers to create real-time maps of weeds in the field, rather than monitor weedy patches manually. This can be accomplished early in the season for timely weed control when crops and weeds are growing at different rates (Andújar et al. 2016; Li and Tang 2018; Piron et al. 2011). For example, unmanned aerial vehicles (UAVs) as well as ground vehicles equipped with red, green, and blue (RGB) cameras have been successfully used for implementing management tactics based on within-field variability of weed populations (Castaldi et al. 2017; López-Granados et al. 2016; Wu et al. 2020). By detecting and quantifying weedy patches in real time, growers can make decisions about weed control based on spatial distribution of species types (e.g., broadleaves and grasses) or individual species in the field (Andújar et al. 2016; Wu et al. 2020). Over time, this is likely to lower production costs by reducing labor and herbicide use.

Preventing seed production in late-season weed escapes is crucial for mitigating herbicide resistance (Kutugata et al. 2021). Allowing late-season weed seed production increases the probability of herbicide-resistant mutants, particularly in species with prolific seed production. However, growers often ignore weed escapes, because escapes rarely reduce crop yield (Bagavathiannan and Norsworthy 2012). Nonetheless, detecting and managing late-season weed escapes and minimizing replenishment of the soil seedbank is important for reducing long-term weed persistence (Kutugata et al. 2021). Besides detecting individual plants, RGB-based imaging has the potential to estimate seed output of late-season weed escapes, which can direct precision weed management tactics to reduce the weed seedbank (Kutugata et al. 2021).

Automated Weed Control

The ability to detect and map weed populations in real time using 3D imaging has numerous applications in automated weed control. In addition to UAVs, ground-based robots equipped with RGB cameras have been used for automated weed removal through targeted herbicide spraying or mechanical in-row removal (Castaldi et al. 2017; Li and Tang 2018; Wu et al. 2020). Automated weeding systems allow growers added flexibility for weed control in both conventional and organic production. For example, lightweight robots can be used in the field when it is too wet for conventional equipment, allowing for more timely weed control. These systems can also dramatically reduce the overall cost of weed control, particularly in organic production or in specialty crops where there are fewer herbicide options (Fennimore et al. 2017; Lowenberg-DeBoer et al. 2020).

Three-dimensional imaging can improve the accuracy of automated systems in differentiating weeds by size class in real time. Unlike 2D approaches, which are less likely to differentiate species and capture size differences in field conditions, 3D imaging can be used to detect in-row canopy structure and characteristics of individual weed species at different stages (Wu et al. 2020). This is critical for weed control because of the relatively short window of time during which weed removal by herbicides and cultivation is effective (Kudsk and Streibig 2003; López-Granados et al. 2016). For example, postemergence herbicide labels typically refer to weed size as a critical factor for application timing, outside of which weed escapes are more likely to occur. Using automated systems with 3D imaging, rather than 2D or visual assessments, will enable growers to apply site-specific weed control at the optimum timing, as well as monitor troublesome weed patches in subsequent years.

Monitoring Weed Suppression with Cover Crops

Cover crops offer a promising strategy for weed management, because they compete with weeds for resources and inhibit weed seed germination and seedling growth (Teasdale et al. 1998). However, cover crop performance can be variable in the field, leading to weed escapes. Therefore, monitoring cover crop growth (i.e., biomass) is a crucial factor for predicting weed suppression spatially in the field (Mirsky et al. 2013). Three-dimensional imaging techniques are being developed to measure cover crop performance based on biomass and canopy structure (Cooper et al. 2017; Roth and Streit 2018). By mapping cover crop biomass throughout the season or at least at the moment of termination, growers will be able to better quantify and map spatial heterogeneity in the field, particularly areas where the cover crop is underperforming (i.e., low biomass) and where occurrence of weed escapes is most likely.

In cover crop–based reduced-tillage production systems, 3D imaging can be used to map weed escapes for targeted spraying or mechanical removal. Late-season weed escapes (i.e., weeds that survived early-season weed control or emerged later in the season) are the largest contributors to the soil seedbank. Therefore, detecting them is crucial for preventing replenishment of the seedbank, especially for weeds that pose a high risk for evolving resistance to herbicides (Bagavathiannan and Norsworthy 2012). By using 3D imaging to map areas where cover crops are underperforming, which have higher risk for weed escapes, growers can better prepare and conduct site-specific control actions to prevent the production of new weed seed. Furthermore, detecting weed escapes using 3D imaging will help with selecting the best control or removal tool based on weed size. Finally, monitoring cover crop growth and uniformity can help growers predict crop yield and biomass production and determine the optimal timing of cover crop termination as part of their weed management programs.

Applications of 3D Imaging in Weed Research

Three-dimensional imaging offers weed researchers a potentially transformative tool for understanding spatial variability in weed responses to management tactics. Because weeds are highly heterogeneous in the field, creating spatiotemporal models of weed responses can improve our understanding of the underlying patterns in this variability. Rather than measuring weed responses by point estimates in the field, researchers can use 3D imaging to capture the spatial heterogeneity of weed species, density, and biomass in real time. This will help advance our understanding of weed responses to IWM tactics, such as weed–crop competition and predicted yield loss and weed community dynamics.

Modeling Weed–Crop Competition and Predicting Yield Loss

Accurately modeling weed interference in crops as soon as possible after crop emergence is crucial for predicting yield loss. Early-season weeds affect crop yield much more than weeds that emerge after a critical period (typically a few weeks), as they compete with the crop for a longer period and can cause problems if allowed to produce seed (Kropff and Spitters 1990; Veeranampalayam Sivakumar et al. 2020). Modeling the interactions between weed density, duration of weed competition, and crop yield can help estimate yield loss as a function of weed competition over continuous time, rather than discrete time points (Weaver et al. 1992). This can provide more accurate and well-informed weed thresholds for growers to assess the optimal timing of weed control tactics (Swanton et al. 2015; Weaver et al. 1992).

Many empirical models have been developed for quantifying crop loss based on weed density and relative emergence timing with respect to the crop (Kropff and Spitters 1990). However, because weeds typically emerge in successive flushes, descriptive models based on weed density do not provide a complete picture for predicting yield loss, because weeds of different sizes exhibit different competitive ability (Kropff and Spitters 1990). Another approach has been to create dynamic simulations of competition for light and water by measuring physiological processes. Although these ecophysiological models are helpful at providing information for descriptive models of competition, they can be difficult to derive and are overly detailed for practical management applications (Kropff and Spitters 1990). Competition models that account for plant height and relative leaf cover can provide a more accurate prediction of crop yield loss. Early quantification of relative leaf area of crops and weeds has been shown to be an accurate way to predict crop yield loss and account for differences in weed emergence (Kropff and Spitters 1990). Although predictions of yield loss must be made soon after crop emergence, the competitiveness of weeds and crops largely depends on their share of leaf area in the canopy when the weed–crop canopy closes (Kropff 1988; Spitters and Aerts 1983). Therefore, modeling weed–crop competition requires quantifying the changes in relative leaf area of weeds versus crops between the time of crop emergence and canopy closure (Kropff and Spitters 1990).

Three-dimensional imaging will allow researchers to model weed–crop competition by estimating vegetation cover, particularly leaf area index. Modeling the weed–crop canopy, including each plant's height, density, and leaf area in space as a function of time, will allow researchers to study changes in canopy structure and compare physiological and morphological changes in response to competition for light, especially before canopy closure. These approaches will be able to provide real-time nondestructive sampling of biomass and leaf cover at any time point, rather than having to harvest and weigh plants at set times. Improved weed–crop competition models will also be useful for growers for monitoring weed interference and predicting yield loss in specific crop cultivars. For example, UAV-derived SfM 3D imaging has been used to compare weed competitiveness in wheat (Triticum aestivum L.) cultivars by measuring morphological and physiological traits associated with early vigor (Aharon et al. 2020). RGB imaging has also been used to compare weed competitiveness in legumes (Travlos et al. 2017).

Weed Community Dynamics

In addition to weed–crop competition, 3D imaging can help researchers better characterize weed community dynamics. Weed community structure can be influenced by environmental factors (e.g., soil type), disturbance events (e.g., tillage), colonization, and competition interactions (Cléments et al. 1994). Understanding changes in weed community structure is crucial for making IWM decisions in order to maintain weed populations below economic thresholds (Swanton and Murphy 1996). Therefore, it is crucial to understand how weed community structure changes in response to environmental factors and management tactics in the short and long term, and how these fluctuations influence succession of weed species and biodiversity (Swanton and Murphy 1996).

Three-dimensional imaging can be used to create descriptive models of weed community dynamics. By using data for leaf texture and density, and differences in plant height as proxies for community diversity and structure, 3D techniques can more accurately describe weed communities in real time. This information can be used to model community changes over many time points in response to IWM tactics or environmental factors. For example, one area of concern in conservation tillage is potential weed interference by perennials, grasses, and wind-borne species (Swanton and Murphy 1996). Three-dimensional quantification of differences in growth habit, plant height, and canopy structure and complexity can be used to describe the spatial and temporal variability of weed species, particularly in perennial systems. By monitoring weed community dynamics over time, researchers can predict changes in community structure and develop appropriate management strategies (Swanton and Murphy 1996).

Applications of 3D Imaging in Other Agricultural Systems

Three-dimensional imaging has the potential to improve weed management in a wide range of agricultural systems besides field crops (Table 1). In this section, we will discuss applications of 3D imaging for weed management in two contrasting systems: orchards and grasslands.

Weed Management in Orchards

Weed management in orchards is crucial for healthy tree growth and fruit yield, as well as soil and water quality. Orchard weeds can compete aggressively with trees for resources, provide habitat for pests, and shade and stunt young trees (Riczu et al. 2015). Poorly controlled orchard weeds can hamper operations of irrigation equipment and automated sprayers and interfere with the accurate placement of water, fertilizers, and pesticides (Brunharo et al. 2020; Liu et al. 2022). Therefore, intensive weed control is necessary in orchards both during establishment and throughout their life span. Although sequential herbicides are widely used in orchards, as are mowing and mulching, variable distribution of weeds necessitates precision control tactics (Brunharo et al. 2020; Rehman et al. 2019; Zeng et al. 2020). Three-dimensional imaging has much potential for improving precision weed control in orchards by accounting for canopy architectural features such as canopy density and depth. For example, LIDAR has been used in orchards to identify weed species, estimate weed cover, and create detailed canopy density maps to inform precision spraying equipment (Riczu et al. 2015; Zeng et al. 2020). However, machine vision using RGB-D cameras is also being developed to map orchards based on canopy height and volume. These systems are being used to guide sensor-driven intelligent spraying of herbicides and pesticides (Rehman et al. 2019; Tagarakis et al. 2022; Zeng et al. 2020).

Weed Management in Grasslands/Rangelands

Timely detection of weeds is crucial for managing grasslands, as weeds can diminish grazing capacity. However, weeds in grasslands tend to be patchy, making monitoring difficult. Therefore, it is important for farmers to have efficient methods for mapping grassland weeds to make decisions about grazing and weed control (Schellberg et al. 2008; Yuba et al. 2020). In a recent study by Yuba et al. (2020), UAV-based RGB imagery with SfM was used to generate 3D point clouds to map patches of the harmful weed fountain grass [Pennisetum alopecuroides (L.) Spreng.] in pastures. This technology has also been used to detect and map weeds in rice (Oryza sativa L.) fields (Kawamura et al. 2020). Despite the relatively few studies using 3D imaging for weeds in grasslands, this technology has considerable potential benefits. For example, 3D imaging can be used to monitor grassland health. Traditionally, “clip and weigh” methods are used to estimate biomass, as it is a key indicator for grassland health and is correlated with canopy height (Lussem et al. 2020). However, these methods are labor-intensive and do not account for variability due to grazing, species composition, and environmental factors (Bareth and Schellberg 2018; Schellberg et al. 2008). Recently, LIDAR has been used to estimate grassland species using 3D point clouds and canopy height and volume data (Paturkar et al. 2020; Schulze-Brüninghoff et al. 2019). Additionally, UAV-based SfM and stereo imaging have been used to quantify canopy height variation, which can be used as a surrogate factor for aboveground biomass (Bareth and Schellberg 2018; Cooper et al. 2017). In a study by Lussem et al. (2020), UAV-based SfM and multiview stereopsis were shown to predict forage biomass based on height with results comparable to ground-truth measurements. Although this technology is still largely untested in grasslands, continuing development of UAV-based imaging holds promise for opening new paths for sustainable weed control and monitoring in grassland systems (Cooper et al. 2017).

Figure 4.

Three-dimensional point cloud reconstructions of soybean (A, top view; B, front view) and cereal rye (Secale cereale L.) (C, top view; D, front view). Note the voids in the soybean point cloud (B) caused by dense canopy cover. Such voids are largely absent in cereal rye (D) due to a more even canopy with greater light penetration.

img-z6-1_641.jpg

Caveats

Several important caveats must be considered when using 3D imaging in weed management and research. First, canopy structure can impact efficiency and accuracy of 3D reconstructions. When the canopy is too dense for light to penetrate the upper layers, the lower layers might not be fully detected by the camera and might be absent from the point cloud. For example, in soybean [Glycine max (L.) Merr.] that have reached canopy closure, only the upper layers are included in the 3D reconstruction (Figure 4). In plants with a sparser canopy, more light penetration allows for the lower layers to be captured in the 3D reconstruction (Figure 4). Second, environmental factors can affect the efficiency and accuracy of 3D reconstructions. Leaf movement due to wind is perhaps one of the most common ways in which leaf shapes are obscured and images blurred, producing noise (Armean et al. 2021; Paturkar et al. 2020). Also, overlapping leaves and shadows can interfere with identification of plant features (Armean et al. 2021; Paturkar et al. 2020). Third, morphological features such as leaf shape and leaf orientation can interfere with 3D reconstruction (Andújar et al. 2018; Armean et al. 2021). For example, it can be more difficult to construct monocotyledonous plants with respect to dicotyledonous plants due to the elongated and thinner leaf morphology (Andújar et al. 2018). In addition, leaves that are oriented more vertically are not always correctly identified, because there is not enough information to distinguish them from stems and branches (Armean et al. 2021). Continued technological developments will likely improve the accuracy of 3D modeling in light of these challenges. It must be highlighted that despite these limitations, 3D imaging provides a much more informative description of vegetation architecture than RGB images alone. As image processing and mapping become increasingly efficient, these techniques will offer a simple yet powerful tool that is widely accessible to growers.

Future Research and Conclusions

We are at a turning point in weed management where state-of-the-art 3D imaging techniques are making it possible to create spatiotemporal maps of weeds using relatively low-cost equipment. In contrast to traditional 2D imaging techniques, 3D imaging captures depth and canopy structure, allowing more accurate estimates of plant height and volume at the species level. This technology offers growers and researchers alike unprecedented opportunities for real-time, cost-effective mapping and monitoring of weeds and their impact on crop growth and development. By capturing spatial heterogeneity of weeds in the field, 3D imaging will enable growers to create accurate weed population maps for timely site-specific weed control, in addition to improved targeted weed removal and spraying. Moreover, 3D imaging will enable growers to classify individual weeds based on size, which will allow for more targeted weed control tactics that reduce labor and use of herbicides. In addition, 3D imaging will improve our understanding of weed responses to IWM tactics and predict yield losses resulting from weed–crop competition. As 3D imaging technology continues to improve, it will provide new opportunities for growers and researchers with tools to better understand the complexities of weed biology in the field and develop better-informed tactics for weed control. These tools hold great potential to aid in the collective effort of reducing agricultural inputs and improving environmental health.

Acknowledgments.

We thank Paula Ramos for her technical advice about the use of SfM for weed canopy characterization. This research was funded by the U.S. Department of Agriculture–National Resources Conservation Service Conservation Innovation Grant (no. NR213A750013G031), USDA Hatch Project NC02653, USDA-NIFA grants (nos. 2019-68012-29818 and 2021-51300-34899). The authors declare no competing interests.

References

1.

Aharon S, Peleg Z, Argaman E, Ben-David R, Lati RN (2020) Image-based high-throughput phenotyping of cereals early vigor and weed-competitiveness traits. Remote Sens 12:3877 Google Scholar

2.

Andújar D, Calle M, Fernández-Quintanilla C, Ribeiro A, Dorado J (2018) Three-dimensional modeling of weed plants using low-cost photogrammetry. Sensors 18:1077 Google Scholar

3.

Andújar D, Dorado J, Fernández-Quintanilla C, Ribeiro A (2016) An approach to the use of depth cameras for weed volume estimation. Sensors 16:972 Google Scholar

4.

Andújar D, Escolà A, Rosell-Polo JR, Fernández-Quintanilla C, Dorado J (2013) Potential of a terrestrial LiDAR-based system to characterize weed vegetation in maize crops. Comput Electron Agric 92:11–15 Google Scholar

5.

Armean ZM, Low T, Hancock N (2021) Automatic leaf segmentation and overlapping leaf separation using stereo vision. Array 12:100099 Google Scholar

6.

Bagavathiannan MV, Norsworthy JK (2012) Late-season seed production in arable weed communities: management implications. Weed Sci 60:325–334 Google Scholar

7.

Bareth G, Schellberg J (2018) Replacing manual rising plate meter measurements with low-cost UAV-derived sward height data in grasslands for spatial monitoring. PFG—J Photogramm Remote Sens Geoinf Sci 86:157–168 Google Scholar

8.

Brunharo CACG, Watkins S, Hanson BD (2020) Season-long weed control with sequential herbicide programs in California tree nut crops. Weed Technol 34:834–842 Google Scholar

9.

Castaldi F, Pelosi F, Pascucci S, Casa R (2017) Assessing the potential of images from unmanned aerial vehicles (UAV) to support herbicide patch spraying in maize. Precis Agric 18:76–94 Google Scholar

10.

Cléments DR, Weise SF, Swanton CJ (1994) Integrated weed management and weed species diversity. Phytoprotection 75:1–18 Google Scholar

11.

Comba L, Biglia A, Aimonino DR, Barge P, Tortia C, Gay P (2019) 2D and 3D data fusion for crop monitoring in precision agriculture. Pages 62–67 in2019 IEEE International Workshop on Metrology for Agriculture and Forestry (MetroAgriFor). New York: IEEE, Inc. Google Scholar

12.

Cooper SD, Roy DP, Schaaf CB, Paynter I (2017) Examination of the potential of terrestrial laser scanning and structure-from-motion photogrammetry for rapid nondestructive field measurement of grass biomass. Remote Sens 9:531 Google Scholar

13.

Dandrifosse S, Bouvry A, Leemans V, Dumont B, Mercatoris B (2020) Imaging wheat canopy through stereo vision: overcoming the challenges of the laboratory to field transition morphological features extraction. Front Plant Sci 11:96 Google Scholar

14.

Fennimore S, Slaughter DC, Siemens MC, Leon RG, Saber MN (2017) Technology for automation of weed control in specialty crops. Weed Technol 30:823–837 Google Scholar

15.

Kawamura K, Asai H, Yasuda T, Soisouvanh P, Phongchanmixay S (2020) Discriminating crops/weeds in an upland rice field from UAV images with the SLIC-RF algorithm. Plant Prod Sci 24:198–215 Google Scholar

16.

Kropff MJ (1988) Modelling the effects of weeds on crop production. Weed Res 28:465–471 Google Scholar

17.

Kropff MJ, Spitters CJT (1990) A simple model of crop loss by weed competition from early observations on relative leaf area of the weeds. Weed Res 31: 97–105 Google Scholar

18.

Kudsk P, Streibig JC (2003) Herbicides—a two-edged sword. Weed Res 43: 90–102 Google Scholar

19.

Kutugata M, Hu C, Sapkota B, Bagavathiannan M (2021) Seed rain potential in late-season weed escapes can be estimated using remote sensing. Weed Sci 69:653–659 Google Scholar

20.

Li J, Tang L (2018) Crop recognition under weedy conditions based on 3D imaging for robotic weed control. J Field Robot 35:596–611 Google Scholar

21.

Liu S, Wang X, Li S, Chen X, Zhang X (2022) Obstacle avoidance for orchard vehicle trinocular vision system based on coupling of geometric constraint and virtual force field method. Expert Syst Appl 190:116216 Google Scholar

22.

López-Granados F, Torres-Sánchez J, Serrano-Pérez A, de Castro AI, Mesas-Carrascosa FJ, Pena JM (2016) Early season weed mapping in sunflower using UAV technology: variability of herbicide treatment maps against weed thresholds. Precis Agric 17:183–199 Google Scholar

23.

Lowenberg-DeBoer J, Huang IY, Grigoriadis V, Blackmore S (2020) Economics of robots and automation in field crop production. Precis Agric 21:278–299 Google Scholar

24.

Lussem U, Schellberg J, Bareth G (2020) Monitoring forage mass with low-cost UAV data: case study at the Rengen grassland experiment. PFG—J Photogramm Remote Sens Geoinf Sci 88:407–422. Google Scholar

25.

Mirsky SB, Ryan MR, Teasdale JR, Curran WS, Reberg-Horton CS, Spargo JT, Wells MS, Keene CL, Moyer JW (2013) Overcoming weed management challenges in cover crop-based organic rotational no-till soybean production in the eastern United States. Weed Technol 27:193–203 Google Scholar

26.

Neve P, Norsworthy JK, Smith KL, Zelaya IA (2010) Modeling evolution and management of glyphosate resistance in Amaranthus palmeri. Weed Res 51:99–112 Google Scholar

27.

Paturkar A, Gupta GS, Bailey D (2020) Non-destructive and cost-effective 3D plant growth monitoring system in outdoor conditions. Multimed Tools Appl 79:34955–34971 Google Scholar

28.

Piron A, van der Heijden F, Destain MF (2011) Weed detection in 3D images. Precis Agric 12:607–622 Google Scholar

29.

Rehman TU, Zaman QU, Chang YK, Schumann AW (2019) Development and field evaluation of a machine vision based in-season weed detection system for wild blueberry. Comput Electron Agric 162:1–13 Google Scholar

30.

Riczu P, Nagy A, Lehoczky, Tamás J (2015) Precision weed detection using terrestrial laser scanning techniques. Commun Soil Sci Plant Anal 46(S1):309–316 Google Scholar

31.

Roth L, Streit B (2018) Predicting cover crop biomass by lightweight UAS-based RGB and NIR photography: an applied photogrammetric approach. Precis Agric 19:93–114 Google Scholar

32.

Schellberg J, Hill MJ, Gerhards R, Rothmund M, Braun M (2008) Precision agriculture on grasslands: applications, perspectives and constraints. Eur J Agron 29:59–71 Google Scholar

33.

Schulze-Brüninghoff D, Hensgen F, Wachendorf M, Astor T (2019) Methods for LiDAR-based estimation of extensive grassland biomass. Comput Electron Agric 156:693–699 Google Scholar

34.

Smith LN, Zhang W, Hansen MF, Hales IJ, Smith ML (2018) Innovative 3D and 2D machine vision methods for analysis of plants and crops in the field. Comput Ind 97:122–131 Google Scholar

35.

Spitters CJT, Aerts R (1983) Simulation of competition for light and water in crop-weed associations. Asp Appl Biol 4:467–483 Google Scholar

36.

Swanton CJ, Murphy SD (1996) Weed science beyond the weeds: the role of integrated weed management (IWM) in agroecosystem health. Weed Sci 44:437–445 Google Scholar

37.

Swanton CJ, Nkoa R, Blackshaw RE (2015) Experimental methods for crop–weed competition studies. Weed Sci 63:2–11 Google Scholar

38.

Swanton CJ, Weaver S, Cowan P, Acker RV, Deen W, Shreshta A (1999) Weed thresholds: theory and applicability. J Crop Prod 2:9–29 Google Scholar

39.

Tagarakis AC, Filippou E, Kalaitzidis D, Benos L, Busato P, Bochtis D (2022) Proposing UGV and UAV Systems for 3D mapping of orchard environments. Sensors 22:1571 Google Scholar

40.

Teasdale JR, Hatfield JL, Buhler DD, Stewart BA (1998) Cover crops, smother plants, and weed management. Integr Weed Soil Manag 247:270 Google Scholar

41.

Teng X, Zhou G, Wu Y, Huang C, Dong W, Xu S (2021) Three-dimensional reconstruction method of rapeseed plants in the whole growth period using RGB-D camera. Sensors 21:4628 Google Scholar

42.

Thornton PK, Fawcett RH, Dent JB, Perkins TJ (1990) Spatial weed distribution and economic thresholds for weed control. Crop Prot 9: 337–342 Google Scholar

43.

Travlos I, Mikroulis A, Anastasiou E, Fountas S, Bilalis D, Tsiropoulos Z, Balafoutis A (2017) The use of RGB cameras in defining crop development in legumes. Adv Anim Vet Sci 8:224–228 Google Scholar

44.

Veeranampalayam Sivakumar AN, Li J, Scott S, Psota E, Jhala A, Luck JD, Shi Y (2020) Comparison of object detection and patch-based classification deep learning models on mid- to late-season weed detection in UAV imagery. Remote Sens 12:2136 Google Scholar

45.

Weaver SE, Kropff MJ, Groeneveld RMW (1992) Use of ecophysiological models for crop-weed interference: the critical period of weed interference. Weed Sci 40:302–307 Google Scholar

46.

Wu X, Aravecchia S, Lottes P, Stachniss C, Pradalier C (2020) Robotic weed control using automated weed and crop classification. J Field Robot 37:322–340 Google Scholar

47.

Yuba N, Kawamura K, Yasuda T, Lim J, Yoshitoshi R, Watanabe N, Kurokawa Y, Maeda T (2020) Discriminating Pennisetum alopecuroides plants in a grazed pasture from unmanned aerial vehicles using object-based image analysis and random forest classifier. Grassl Sci 67:73–82 Google Scholar

48.

Zeng L, Feng J, He L (2020) Semantic segmentation of sparse 3D point cloud based on geometrical features for trellis-structured apple orchard. Biosyst Eng 196:45–55 Google Scholar

49.

Zhang Y, Teng P, Shimizu Y, Hosoi F, Omasa K (2016) Estimating 3D leaf and stem shape of nursery paprika plants by a novel multi-camera photography system. Sensors 16:874 Google Scholar
© The Author(s), 2022. Published by Cambridge University Press on behalf of the Weed Science Society of America.
April M. Dobbs, Daniel Ginn, Søren Kelstrup Skovsen, Muthukumar V. Bagavathiannan, Steven B. Mirsky, Chris S. Reberg-Horton, and Ramon G. Leon "New Directions in Weed Management and Research Using 3D Imaging," Weed Science 70(6), 641-647, (14 October 2022). https://doi.org/10.1017/wsc.2022.56
Received: 30 July 2022; Accepted: 27 September 2022; Published: 14 October 2022
KEYWORDS
integrated weed management
point cloud
remote sensing
stereo vision
structure-from-motion
Back to Top