1. Introduction
Anthropogenic interaction with vegetative presence is germane for both physical and aesthetic health [
1,
2]. Within this context, people primarily gain awareness of the environment through visual perception [
3], and the photographic camera is an instrument that can be considered to extend this fundamental human sense [
4,
5,
6]. Modern cameras have evolved to offer robust optical detection, available in consumer-grade products such as the GoPro Hero 12 (GoPro Corporation, SanMateo, CA, USA) or Sony Alpha 9 (Sony Corporation New York, NY, USA) among many others. This benefit is largely attributed to the considerable and sustained investments in academic and commercial optical research which drive advancement for camera products at leading manufacturers [
7] such as Canon (Canon Corporation, Melville, NY, USA) and Nikon (Nikon Corporation, Melville, NY, USA). As a result, optoelectronic camera technology provides one of the most sophisticated consumer-available sensor products [
8,
9,
10,
11].
Current digital photographic camera technology enables efficient capture of gridded spectral reflectance data in the form of a 2-D perspective map that accurately represents real-world physical features. The subjects of photography may encompass vegetation, where Red, Green, and Blue color (RGB) foliar reflectance is linked to both subjective visual quality and the underlying productive photosystem pigmentation [
12,
13,
14]. Therefore, optical cameras can be expected to provide wide utility in the measurement of plant canopy status and change. They can detect characteristics such as the area of green biomass cover and can resolve subtle color differences presented by plant tissues [
15,
16,
17]. However, the acquisition of scientific data is more widely utilized [
18,
19,
20,
21] when it is not only of adequate descriptive quality, but also inexpensive and easy to produce, with understandable controls that involve generic parts.
Cameras have been used in various applications, such as to measure industrial parts [
22], assess soil moisture [
23], for plant phenotyping [
24,
25,
26,
27,
28] to measure disease [
29], predation effects [
30], senescence [
31], salt tolerance [
32], drought response [
33], percent green area of cover [
34,
35,
36,
37,
38], to make fertilizer recommendations [
39], measure grass-legume overyield [
40], fruit analysis [
41], and linking phenotype to genotype [
42,
43,
44,
45]. However, premier research methods often involve expensive sensing equipment (
https://terraref.org/), complex experimental facilities (
https://www.plantphenomics.org.au/), and robust data corrections that use specialized software and advanced mathematical calculations [
46,
47,
48,
49]. Improvement in the interpretation of multi-environment data with biological understanding is warranted to effectively scale plant phenotyping across its many applications [
50]. The straightforward use of the consumer camera in plant phenotyping plays a role in support of this goal.
Imagery used for plant phenotyping is a common practice because it can increase data precision, speed, and consistency over traditional manual plant assessments [
51]. Additionally, it can offer increased spatial and statistical power over single electronic point measurement collections [
52,
53]. Innovators in high-throughput plant phenotyping often use sophisticated engineering and sensing technology to generate complex datasets that require advanced handling, corrections, and analysis [
54,
55,
56]. This approach is effective in addressing research questions, but it usually has a limited scope of application and may necessitate a team of experts or expensive equipment acquisitions [
57,
58,
59,
60]. Support for high throughput plant phenotyping could be enhanced by a simpler approach that utilizes a consumer-grade camera with limited controls. The inherent complexity of current camera technology offers a promising approach to data generation. It can effectively address research questions, albeit with some trade-offs because it operates at a smaller scale and with reduced throughput. However, it also comes with reduced costs and lower technical requirements potentially making it useful across many applications. The modality of simple image-based phenotyping is most suitable for students, commercial managers, or professional researchers conducting experiments of limited size. The use of modern cameras with simple controls can enable detection of plant cover area and color qualities, and although outside the scope of this paper, it opens opportunity for further spatial analysis using shape parameters.
Grasses constitute a substantial component of the Earth’s plant community [
61,
62,
63,
64], and turfgrasses are the primary vegetative cover in many places humans interact [
65,
66]. They cover approximately 2% of the continental United States [
67], or for instance 30% of metropolitan Albuquerque New Mexico [
68]. Assessment of turfgrass canopy is an essential aspect of the relationship between humans and plants to effectively gain ecosystem services through sustainable resource utilization. The visual quality ratings of turfgrass is a traditional procedure that follows standardized guidelines (
https://www.ntep.org/, [
69]). Although there is a constraint for a single expert observer to increase accuracy and minimize data variance, human visual quality ratings of turfgrass (VQ) as a practice [
69,
70] is useful because it includes broad subjective information apparent to the complexity of human sensory perception and the cognitive expertise of the expert. Separately, NDVI is a well-regarded remotely sensed normalized difference ratio measure of the Red (RED) and Near-Infrared (NIR) spectral reflectance which detects presence of green biomass [
71,
72,
73,
74,
75]. NDVI has been successful in many applications to resolve plant cover [
76,
77,
78], plant nitrogen status [
79], plant vigor [
80,
81], and other attributes of vegetation, including chlorophyll content [
82,
83], influence of drip irrigation [
84] and deficit irrigation [
85]. Consequently, VQ and NDVI are both typical assessors of turfgrass that can be used in different environmental conditions and management scenarios [
86].
Imaging is useful in turfgrass research to detect the area of green cover and changes in plant establishment over time [
87]. This is often achieved by counting green classified pixels and calculating a percentage of green presence [
88,
89,
90,
91]. The process is enabled by compute and can be partially automated using programming scripts or recorded macros [
35]. Once data is digitized in silico, many additional and more advanced opportunities for processing and analysis become available [
92,
93,
94,
95,
96,
97]. However, there is an initial need to connect and contextualize image-based data with separate and reliable measures of the plant phenotype. Vegetation metrics calculated from a single set of image data are automatically dependent to some extent, and without a reference they do not provide a measure of alignment with the underlying biological phenomena. Therefore, isolated analysis of an individual plant image collection may not account for biases induced by the camera, photographic exposure, environmental variations, or computational process. However, once verification of a specific camera profile response to the plant phenotype trait is obtained, subsequent image data may hold increased independent value and no longer require related attribution. This idea extends to the image correction process implemented post-collection and includes the image file format. Understanding how a specific lossy file compression like JPEG (Joint Photographic Experts Group), image geometric lens distortion correction (LC), and color correction (CC) may affect phenotypic detection ability relative to various image-based calculations would ensure experimental control and improve discovery potential.
The purpose of this research is to evaluate whether a modern consumer-grade camera when used with simple image controls, could provide data representative of the information derived from conventional phenotyping metrics such as visual quality (VQ), biomass productivity, water use, and independently measured multi-spectral reflectance. The study also examines to what extent two typical image file formats and two common corrections may affect the results. Authors employ the premise that expert-assessed turfgrass VQ and NDVI can be reference measurements that enable data comparison and evaluation of image-based turfgrass metrics [
98,
99,
100,
101].
2. Materials and Methods
An economical (<
$1000) Nikon N1 aw1 DSLR action camera with 11 to 27.5 mm f 3.5 to 5.6 NIKKOR lens and custom-made LED lightbox were used to investigate 72 custom PVC lysimeters (15.2 cm diameter × 30.5 cm depth) growing ‘TifTuf’ bermudagrass (Cynodon dactylon × C. transvaalensis Burt Davy) in two separate greenhouses during the autumn season of 2023, at the U.S. Aird-Land Agricultural Research Center (ALARC) in Maricopa, Arizona, USA. The lysimeters were divided into two runs and treated with four mowing heights (2.5, 5.0, 7.5, and 10.0 cm), three water application levels [(100 × ETa (actual evapotranspiration), 65 × ETa, and 30 × ETa)], and evaluated for visual quality, clipping productivity (biomass), and water use of the 100 × ETa treatment (gravimetric loss). The primary experiment is described in detail in Hejl et al. [
102] along with major results. Information regarding the growth setup and environmental measurement are provided in
Appendix A. The camera image exposure was controlled by locking manual settings as described below. Image file format, lens distortion, and color correction were tested. The Holland Scientific RapidScan CS-45 product (Holland Scientific, Lincoln, NE, USA), was used, to acquire RED, Red-Edge (RE), and NIR spectral components at 670, 730, and 780 nm wave lengths, respectively. The active reflectance values were collected concatenate with images to provide a camera-independent measurement of the plant's optical signature. Collection of the three spectral bands allowed the calculation of NDVI and NDRE, where both are related, however, NDVI is more common and typical to assess canopy cover but can saturate at higher leaf area index values, while NDRE is considered more sensitive to plant nitrogen and chlorophyll status and incorporates information from the spectral inflection point along the RE region [
103,
104,
105]. Tables of calculated values were managed using MS Excel Version 2401 (Microsoft Corporation Redmond, WA, USA), and statistical analysis was conducted in JMP 15.2.0 (SAS Institute, Cary, NC, USA) software, but a permutation ANOVA technique [
106] was implemented in the R programming language (R Core Team 2021) as aovperm from the permuco package.
The custom lightbox design was based on a modification of the well-regarded 2001 approach of Richardson and Karcher [
107], and the 2016 Zhang et al. design [
108] that incorporates LED lights, each of which employs a portable photographic studio for digital image-based phenotyping of turfgrass. The rudimentary prototype lightbox used in this paper measures 57 (L) × 40 (W) × 74 (H) cm. It was constructed from 0.5 cm thick repurposed cardboard, assembled with tape, and equipped with an internal horizontal image backdrop that fit flush around the 16 cm diameter lysimeter top (182.4 cm2). The camera was positioned with lens protruding into a 24 (L) × 24 (W) × 14 (H) cm upper box containing LEDs. The camera lens was 27 cm distant from the lysimeter and pointed downward from the top of the box through an 11 (L) × 20 (W) × 6 (H) cm polystyrene camera cradle with a 7 cm hole that was 6 cm deep to fit the form of the lens. White printer poster paper was used as a generic, hue-neutral light reflector to cover the inside surfaces of the box. The box was wide enough so that the camera view of the lysimeter top and background would not extend to its sides; therefore, only the lysimeter top and white backdrop were in camera view. For each image sample, the box and one lysimeter were placed on a 60 (L) × 55.5 (W) cm light-colored paper poster board, which was positioned on the greenhouse bench to provide a consistent material foundation.
Photographic illumination was provided by an inexpensive generic 12-volt light emitting diode (LED) strip (Desvorry branded) consisting of 300 individual 6000 K white color lights. The light strip was affixed to the top of the box inside the perimeter around and just above the camera lens. Many LED lights were used for illumination to support light diffusion and reduce of shadows. Change in lighting due to the duration of LED activation was detected, as measured by a pyranometer positioned at the height of the lysimeter top inside the lightbox. The overly inexpensive light source decreased in intensity from 16 to 8 watts m 2 over 360 seconds, described by equation 1:
(coefficient of determination, R2 = 0.95), with a steady state of 8 watts m-2 after 360 seconds (standard deviation, STD = 0.537). Therefore, the LEDs were only activated immediately before each lysimeter image sample and then turned off to mitigate fluctuations in brightness. The approach reduced measured illumination variance, resulting in an average of 16 watts m 2 (STD = 0.482) across twelve test runs. Each test run consisted of 7 to 15 seconds of illumination with 60 seconds of de-energized time in between. This scenario was representative of the typical timing for lysimeter data collection of just over one minute per lysimeter.
The camera was set to “manual” acquisition mode with a “manual” exposure, no exposure bias, no flash, and “pattern” focus metering. “Standard color” was selected with no “Active D lighting”. The camera exposure was set to 100 ISO with an aperture of F-Stop 3.5 and a shutter speed of 1/320 second. An 11 mm focal length (equivalent to 29 for a 35 mm crop factor) was used, and the adjustable zoom lens was taped in position. A custom white balance was sampled and recorded on the camera by using the white poster paper placed inside the box and illuminated by the LEDs. The camera viewfinder displayed a circular graphic that was about the same size as the lysimeter top when visualized on the screen. This feature was utilized to verify straightness of camera angle and support geometric consistency between images. The camera display helped verify the time each image was taken and ensured that no settings were accidentally changed between image collections. A custom remote trigger was used to manually activate automatic focus and acquire images. High-resolution 4608 × 3072-pixel JPEG and 4548 × 3012-pixel Nikon NEF (RAW) image file formats were obtained (approximately 5.9 and 12.3 MB in size respectively). This equated to a measurement resolution close to 11 pixels per mm. Three 24-bit sRGB image replicates and one RapidScan composite reflectance measurement were taken consecutively for each lysimeter sample. Average values obtained from the imagery were used for comparison with the individual VQ and NDVI lysimeter point measurements. Lysimeters were imaged approximately every other week for eight weeks. Image processing was conducted on a Dell Latitude 5420 laptop with an i7-1185G7 1.8 GHz CPU and 16 GB of RAM (Dell Technologies Inc., Round Rock, TX, USA) running Microsoft Windows 11 Pro 22H2 64-bit OS (Microsoft Corporation, Redmond, WA, USA). This computer represents a typical consumer-grade product, with the Python 3.12 software (Python Software Foundation, Beaverton, OR, USA) that is commonly used in lower-division academic settings. Jupyter Lab and Notebook software (Project Jupyter, New York, NY, USA) were run using an Anaconda 2.5.0 installation (Anaconda Inc., Austin, TX, USA) and included the Pandas, Numpy, Scipy, OpenCV, PIL, Scikit-Image, Matplotlib, and Seaborn modules. When RAW imagery was used in the Python process, files were initially converted to TIFF (Tagged Image File Format) without compression using Irfan Skiljan’s IrFanView 4.59 (IrFanView, Lower Austria, Austria), which resulted in a file size of 40.1 MB. All imagery was converted to an 8-bit memory depth when processing in Python to reduce computational overhead. An 8-bit image can still display up to 16.7 million colors using 0-255 value tuples. It took approximately 60 seconds to process a single image file on the small laptop computer, regardless of the file type.
Figure 1.
The lightbox is shown in greenhouse #1 (panel a, left side) with the camera installed on top. The remote trigger with switch and the 12-volt power supply with 7.5 Ah SLA battery and wires are visible on the left and bottom left side. The lightbox diagram (panel b, right side) illustrates the placement of LED lights and demonstrates how a lysimeter would be inserted into the box and photographed against the white background.
Figure 1.
The lightbox is shown in greenhouse #1 (panel a, left side) with the camera installed on top. The remote trigger with switch and the 12-volt power supply with 7.5 Ah SLA battery and wires are visible on the left and bottom left side. The lightbox diagram (panel b, right side) illustrates the placement of LED lights and demonstrates how a lysimeter would be inserted into the box and photographed against the white background.
Color correction was performed before application of lens correction using the PerfectColor program from Evan Lund of Evens Oddities (
https://www.evenlund.com/). This program was freely available and allowed batch correction of images using the 2014 standard X-Rite Color Checker Classic (24) Legacy panel. It used lookup tables (LUT) generated from a camera profile reference image of the color checker panel that was captured inside the lightbox. Color checker reference values were provided by Babel Color (
https://babelcolor.com/), and RawTherapee was used to sample and compare color values using its Lockable Color Picker tool. Image color values were also measured using ImageJ 1.54c for reference and comparison [
109]. ImageJ is an open-source Java program credited to Wayne Rasband and contributors, in association with the National Institutes of Health, USA. The software offers powerful image processing and analysis functions.
Lens correction was performed after color correction using Python Discorpy 1.6.0, which can correct radial and perspective distortion by utilizing a grid pattern reference image. Graph paper was photographed inside the lightbox to provide the input pattern. The software calculated polynomial correction coefficients specific to the camera setup. Correction results showed linear grid uniformity of graph paper control images upon manual review. Corrections were batch applied to the experiment sample images. Separately, lens vignette correction was performed in RawTherapee using the settings of 48 strength and 45 radius for JPEG images, and 45 strength and 45 radius for TIFF images.
Color classification areas were determined based on pixel values and a binary mask to segment regions of interest in the original image. The selected classification areas were analyzed as a fraction of the total image area. They represented a manually assessed range for the total living plant material cover, the percentage of yellow senescent or chlorotic plant material, and a generic green color plant material. Details are provided below and in
Table 1. The bespoke living plant material cover segment was used to calculate the image discrete color qualities within that area segment. Individual color scalar components were generated across several color spaces. Several basic calculations were conducted to evaluate vegetation indices from existing literature, and custom relationships were also formulated based on linear correlation with NDVI and green cover area. Due to the large number of different calculated terms evaluated (280), only the most relevant metrics that best correlated with VQ and NDVI were selected to report in this paper. Although NDVI was collected at the same time as image metrics, VQ was not. There was an average difference of 2.2 days (STD = 0.92) between the assessment of VQ and the collection of image metrics.
The RGB color space is common in imagery [
111], where the R term indicates the intensity of Red, the G term intensity of Green, and the B term intensity of Blue colors. The three terms added together indicate a brightness value. The Hue-Saturation-Value (HSV) color space can be derived from RGB values. It separates the color components into a single angle of Hue (H) ranging from 1 to 360 (or another value range of different graduation). Saturation (S) indicates the linear lightness of the Hue color, and Value (V) indicates the linear brightness. Saturation and Value are typically measured on a scale of 1-100 (or 1 to 255 for 8-bit imagery). A Value of 0 represents black, while a Saturation of 0 indicates gray without any color. Percent living plant cover classification segmentation (%C) was conducted using the HSV color space. Where Hue (on a scale of 1 to 180) ranged from ≥ 27 to ≤ 90, Saturation (on a scale of 0 to 255) ranged from ≥ 60 to ≤ 255, and all brightness Values were included (on a scale of 0 to 255). The generic green fraction (%G) included Hue ≥ 30 and ≤ 60. Green fractional cover was previously reported by Casadesus to support wheat research [
112]. Yellow area classification (%Y) involved Hue ≥ 16 to ≤ 26. Both %G and %Y used the same Saturation and brightness Value designations as %C (
Table 1).
The Python-based image analysis used in this paper was modeled on the approach exemplified in the TurfAnalyzer (TA) software. TA is a freely available software tool developed by Douglas Karcher in collaboration with Pennington Seed, Inc., and Nexgen Turf Research LLC. [
113] based on the progressions of Karcher and Richardson [
35,
110] who originally used SigmaScan macros to measure turfgrass. It enables analysis of JPEG images for percent green cover and dark green color index (DGCI) and has been used in many experiments [
114,
115,
116,
117,
118,
119,
120]. The default color values for TA were included in the Python analysis. These values were determined by converting the full Hue range from 1 to 360 and the Saturation and brightness Value ranges from 1 to 100 in TA to a Hue range of 1 to 180 and Saturation and brightness Value ranges from 0 to 255 for the OpenCV commands, respectively. To better understand the compute process and verify data quality, calculated image values such as mean cover area classifications and DGCI were compared between the two software outputs and found to be equal. However, the more intricate Python approach was selected in this research due to the desire for enhanced process control, the capability to conduct full-resolution computations on TIFF format files, the option to export individual pixel values if wanted, the ability to export additional statistical values for each image, the need for additional calculations involving customized relationships between metrics, and the option to visualize and export individual values, histograms, or index calculations as separate charts.
1 A comprehensive range of Hue angles was selected for the fraction of living cover area (%C) classification segment, to include lower quality yellow-green living plant material, through the range of green color to higher-quality green living plant material and into any blue-green material (although no blue-green plant material was present). This broad range was chosen to be inclusive of all the living plant material. Selection of the generic green color area (%G) centered around hue angle 45 and was intended to select only healthy vegetation with high chlorophyll and nitrogen contents. The fraction of senesced yellow plant material (%Y) selection included color values below those of the %C, but also excluded the shades of Red and brown colors which indicated soil (the threshold was positioned between the color overlap of brown leaves and brown soil). The three classified color fractions included all brightness Values but used a Saturation threshold of about 24% to ensure that selected pixels contained a color intensity which excluded the white background and bright reflection points. These classification parameters were derived from an iterative process of manual selection and evaluation detailed in the discussion section.
Figure 2.
An example lysimeter uncorrected image and three masked views. Experiment treatment 30% water and 5.0 cm mow height is shown in an image taken on 10/26/2023 (Week 2) with associated 0.61 NDVI and 7.0 VQ (panel a, upper left), 97.8% of the lysimeter area covered in live green material (%C) segment (panel b, upper right), resulting in 0.280 DGCI, 0.400 HSVi and 7.010 COMB2 calculation values, with 31.1% yellow (%Y) plant cover (panel c, lower left), and 59.0% green (%G) cover fraction (panel d, lower right).
Figure 2.
An example lysimeter uncorrected image and three masked views. Experiment treatment 30% water and 5.0 cm mow height is shown in an image taken on 10/26/2023 (Week 2) with associated 0.61 NDVI and 7.0 VQ (panel a, upper left), 97.8% of the lysimeter area covered in live green material (%C) segment (panel b, upper right), resulting in 0.280 DGCI, 0.400 HSVi and 7.010 COMB2 calculation values, with 31.1% yellow (%Y) plant cover (panel c, lower left), and 59.0% green (%G) cover fraction (panel d, lower right).
2 The upper left (panel a) displays the original 2-D view image sample for a 5 cm mow height and low water treatment. The hue-neutral white background is adjacent to the lysimeter perimeter, with grass leaves extending outward. The lower left (panel b) depicts the HSV-based primary color segmentation of all living plant cover area. This comprehensive range of yellow-green to blue-green %C classification accurately includes all illuminated living plant material and was used as the basis for calculating color qualities in subsequent calculations. The upper right (panel c) shows the %Y fractional area of senescence with plant relevant Hue angles below the %C parameter but above soil. The lower right (panel d) shows the %G cover segmentation of healthy plant material.
This paper evaluated the RGB-based COMB2 calculation from [
121], where three vegetation indices were combined for a model intended to be more sensitive to vegetation changes and less affected by soil background. It uses the Excess Green index, which is designed to enhance detection of green vegetation, equation 2:
from [
122], where r, g, and b values are normalized Red, Green, and Blue, the Color Index of Vegetation Extraction index designed to enhance green color for vegetation segmentation, equation 3:
from [
123], and the Vegetation Extraction Index designed to emphasize green reflectance, equation 4:
from [
124], where r, g, and b values are normalized Red, Green, and Blue, and the alpha value of 0.667 was taken from [
125] as a generic camera factor. These indices are combined to create equation 5:
Color spaces beyond the standard RGB and HSV were used. CIE 1976 Lab* (CIELAB) values were calculated with OpenCV, and individual color and brightness scalars were evaluated separately, as well as in relation to each other using various equations. The CIELAB coordinate system was adopted by the International Commission on Illumination (CIE) in 1976 [
126] as a color space with increased uniformity relative to human visual perception [
127,
128]. The CIELAB color space is useful in many applications [
129,
130,
131,
132,
133] because its colors use the standard observer model for device independency. The CIELAB color space achieves increased human perceptual uniformity because equal Euclidean distances in the color space corresponds to equal perceived color differences. The CIELAB L* term represents lightness, but essentially, the a* term represents a green-to-red color dimension axis, and the b* term represents a blue-to-yellow color dimension axis. This paper uses the color opponent axis ratio of b* to a* (BA), intended to emphasize plant health indications of green and yellow pigment status, and authors are not aware of other phenotyping papers using this ratio although Wanono [
134] investigated the ratio of a* to b* to detect leaf nitrogen. A higher BA ratio could signify chlorosis, senescence, or water stress, while a lower value could indicate presence of chlorophyll, nitrogen or increased water use efficiency.
An additional image-based equation was created to compare with standard DGCI. The unconventional Hue difference illumination ratio, or HSVi, equation 6:
This equation utilizes the HSV color space, where the simple ratio of Saturation and Value was subtracted from Hue, and the result was shifted and offset to scale roughly into a range of 0 to 1. By considering the saturation-to-value ratio we capture a relative color intensity, and so as brightness increases the saturation impact reduces. A higher ratio indicates more vivid color relative to brightness. This nuanced HSVi approach modulates Hue angle as a function of color purity and attenuates the positive effect of Hue using lighting to discriminate minor plant features. Therefore, context is important for interpretation, because a larger HSVi result may indicate healthy but less dense vegetation with higher reflectance, while lower values may indicate stressed or the presence of very dark green vegetation.
Calculated metrics were summarized in tables and related to the experimental class structure. To test the effects of experimental treatments, response variables were assessed using a robust permutational multivariate analysis of variance (permANOVA). Non-parametric permutation ANOVA models are useful in environmental analysis to better handle complex data [
135,
136,
137]. Neither the VQ nor the NDVI data in this paper use camera technology. Therefore, VQ data reported previously and NDVI measurements were chosen as independent references to compare with the image-based data [
36,
138,
139]. The authors are unaware of other published works detailing the sensitivity of file format and image corrections (JPEG, TIFF, LC, and CC) on phenotypic image calculations for plant phenotyping.
Image-based metrics were calculated and compared with reference measurements to test the effects of a file format with standard JPEG compression (JPG), NEF (RAW) to lossless TIFF conversion (TIF), and image adjustments for geometric lens correction (LC), and color correction (CC). Pearson’s correlation coefficient (R) and coefficient of determination (R2) values were calculated, and the image metrics that best corresponded to each reference measurement were selected for reporting. A sensitivity analysis of calculated metrics was conducted to assess the impact of file format and image corrections. Significant differences were tested and a comparison of treatment effects on image metrics, VQ, and NDVI was performed using non-parametric Kruskal-Wallis median test and Dunn’s ranked test with JPEG as control. In a similar approach, the coefficient of variation was evaluated to investigate the effects of file format and image correction on replicate images. An NDVI time series overview for the original experiment is presented for the first time in this paper, and time series data for two image-based color metrics are offered to show comparative differences in behavior over the course of the experiment.
4. Discussion
To assist researchers, students, and managers, the authors propose a practical approach that includes accessible camera controls for collecting relevant image-based plant phenotyping data. Locked manual photographic exposure settings that balance a lower ISO, moderately higher F-Stop, and lastly a higher shutter speed should be selected based on the camera’s response to the overall lighting intensity of the target environment. Next, it is important to measure and set a customized white balance for the specific spectral intensity of the ambient illumination, rather than relying on a pre-set color temperature. Consider that a RAW file format allows post-processing of the total camera-generated data. Yet a lossless TIFF file format will retain all the image pixel color values. On the other hand, while the JPEG format is lossy, high-resolution imagery could mitigate the effects of compression loss and may therefore be acceptable for plant phenotyping functions. Although image data corrections are important to consider [
141], they can be challenging to implement, and their effects could be difficult to quantify and report. Therefore, it may be acceptable to rely solely on modern camera technology with simple exposure controls and achieve sufficient image data quality to perform basic plant phenotyping for advancement of scientific knowledge. However, in some instances, such as with the calculation of COMB2 in this paper, image corrections induced significant effects strong enough to change result interpretations. To overcome possible gaps in knowledge, transparent standardized protocols and detailed metadata is needed that includes cross or independent validations, and robust statistical analysis utilizing sample replicates or controls.
Utilizing a simple camera-based plant phenotyping approach, the highest R
2 values were observed in the relationship between VQ with %Y area classification, and NDVI with %G area classification (
Table 2). Traditionally, the evaluation of turfgrass has focused only on green color areas. However, our findings reveal that presence of yellow, which demonstrated congruence with human perception of quality and significant treatment effects, offers insight to assess the relative presence of green biomass in turfgrass. Casadesus [
142] introduced a “Green” and “Greener” segmentation in wheat, incorporating some yellow into the “Green” classification. By introducing a dedicated yellow cover class, such as the %Y metric employed in our study, the yellow stems and senesced plant material is captured and quantified more comprehensively. A more complete classification system that includes the readily observable visual cue of yellow may allow for greater precision in evaluating the composition and health of turfgrass which could improve phenotyping analyses.
Further research is warranted to clarify the COMB2 term in relation to plant biological status because the detected treatment effects significance with image corrections (
Table 4) suggests a utility of this color calculation to support understanding of plant phenotypic traits, but COMB2 behaved differently than traditional terms like NDVI (
Figure 5 and
Figure 3). Although the COMB2 calculation may provide some phenotypic detection capability, it does not predict water use. Similarly, although DGCI correlated strongest with mg clipping production (
Table 2), the relationship was not robust enough to explain most of the variation in reference measurements. The BA ratio correlated most with NDRE and RE reflectance, while the novel term HSVi correlated most with the NIR spectral reflectance component (
Table 2). Therefore, calculation of BA or HSVi may be useful to roughly estimate respective RE or NIR spectral reflectance when only color camera data is available. The BA ratio may serve to emphasize yellowness relative to redness and provide an indication of chlorophyll content or plant health, but authors did not find reference to this ratio in plant phenotyping literature so additional research is warranted. HSVi showed stronger correlation with NDVI than DGCI (
Table 2). Therefore, variations in this illumination modulated Hue angle determined by HSVi may help detect pigmentation status or tissue health across the biomass area or over time. However, because authors are not aware of other research using the HSVi image calculation for plant phenotyping, additional research is needed to determine biological significance.
This paper demonstrates how image acquisition using consumer cameras with simple controls enables not only standard calculations of traditional metrics like the green area or DGCI, but also facilitates additional calculations like bespoke yellow area, HSVi or BA which may be valuable in the description of plant phenotypes. Although it could be possible to avoid image corrections and still achieve acceptable results when simple exposure controls are used, the application of appropriate image corrections before calculation of image metrics will likely generate the most useful information. However, the varied and mixed correction effects results in this paper suggest further research is necessary to precisely identify the best methods to maximize corrections utility in image-based plant phenotyping.
Though the simple image-based approach in this paper used to identify basic cover and color features was successful, limitations persist, and several improvements can be made. First, to enhance color accuracy and consistency in subsequent investigations, the illumination source will be upgraded to a professional-grade LED option with a high color rendering index [
143,
144], such as exemplified in the Waveform Lighting Absolute Series
TM product line. Second, considering that the camera selected was released in 2013, a more modern camera may incorporate improved technology and increase the quality of the base data. Third, although no errors were detected, adjusting the camera lens aperture 3.5 F-Stop to a higher setting of around f/8 to f/11 would elongate focal depth of field, decrease lens vignette artifacts, optimize lens sharpness, and reduce chromatic aberration without inducing major diffraction. Fourth, improved computational pipelines capable of handling RAW-based imagery [
145] are anticipated to become more widely available and easier to use in the future. While uncompressed TIFF format imagery was processed in the Python pipeline used in this study, larger files are more challenging to manage, and there is less software support for uncompressed file formats compared to the common JPEG form. Yet handling all the image information generated by the camera exposure is expected to improve data quality. Likewise, advancements in color correction [
146,
147,
148] software are expected to become more accessible over time. The authors plan to obtain a license and use the Adobe Lightroom commercial software (and they may use the Python color-checker-detection library) to make color corrections in future experiments. A more robust color correction procedure could enhance data quality. Given the high image resolution capabilities of modern cameras, JEPG compression of high-resolution imagery may be acceptable for proximal measurement [
149,
150,
151,
152]. Although camera and lighting limitations are expected to introduce some level of error [
153], a careful assessment is necessary to determine whether inherent camera-induced representational errors are acceptable. The value of corrections should be weighed against their implementation costs and the potential introduction of new error sources. Refinement of the corrections methodology is essential. This study used a sensitivity approach to compare the effects of eight individual image types on six image metric correlations with eight reference measurements, including VQ and NDVI. The study indicates that %Y, BA, and HSVi have the potential to differentiate the effects of experimental treatments in addition to the more traditional %G and COMB2. However, it also highlights challenges in quantifying the effects of image corrections across diverse metrics.
The effectiveness of color corrections and other image adjustments to enhance data quality for a specific plant phenotyping collection is ambiguous. While corrections are sometimes made, evidence of improvement may be lacking [
154,
155]. There are well-developed theoretical best practices for corrections [
156,
157,
158,
159,
160,
161], but additional guidance could be provided to quantify the effects of image data adjustments under various real-world conditions. This would help understanding and determination of which imagery correction procedures are necessary in a particular circumstance. For instance, Chopin demonstrated that their camera profile for proximal field imagery was better fitted using a least squares quadratic transformation rather than linear [
162]. Despite the challenges posed by a diverse range of camera products, application spaces, and vegetative variability, clear standards for operations specific to various applications would be of benefit to users. The provision of clearer guidelines and empirical evidence to describe the effects of image adjustments would allow practitioners to make more informed decisions, ultimately enhancing the quality and reliability of plant phenotyping data.
The findings in this paper show that the image corrections applied resulted in inconsistent outcomes. While correlations with reference measurements generally improved with image corrections, improvements were often minor and sometimes negative. Only %Y showed no significant effect from the TIFF file format or the image corrections, while CC did cause large significant differences in COMB2, BA, DGCI and HSVi (
Table 3). The mean coefficient of variation for replicates was significantly affected by corrections for %G, DGCI, HSVi and COMB2 (
Table 5). The optimal combination of file format and image corrections improved R
2 values for five out of the six image metrics but resulted in less than 2% improvement. However, COMB2 was improved by 31% (
Table 2). This suggests that the application of all standard image corrections may not automatically improve plant phenotyping results, and yet lack of best file format and correction implementation could lead to oversight of crucial relationships. An improved statistical evaluation using a larger and more diverse set of images and reference measurements, including the observation of additional plants beyond TifTuf over a longer period is needed to gain deeper understanding of the image adjustments effects and inform best practices. Furthermore, the inclusion of more robust phenotyping equipment to provide thermal, hyperspectral, florescence and structural information would better contextualize the potential of the traditional camera.
Dynamic selection is an easy way to determine and verify color-based cover classification areas. These areas can be parametrically refined using gradient-based selection masks, included in TurfAnalyzer software and common in commercial image editing applications like Adobe Photoshop. By manipulating the criterion control, users can update a selection on the display of an evaluation image and visually assess whether the proper regions of interest are captured. This dynamic process enables users to determine if a particular image metric is responding to an intended color or feature while ensuring that unwanted colors and features are excluded. A similar computational process can be developed with Python scripts and was used for this paper. Given the variability in illumination conditions and camera-specific factors such as sensor characteristics and color processing algorithms, it is recommended to perform iterative visual assessment and adjustment of segmentation parameters to ensure accurate identification and selection of target chromatic values in the image data. This process certifies that correct regions are selected based on the predefined parameter set. Related, the export of a validation image, which consists of the intersection between the original image pixels and the selection mask (including its inversion) enables a zoomed-in view for more detailed inspection of feature transition edges. Detailed scrutiny, perhaps to the limit of the pixel resolution, can reveal whether the target feature area has been fully segmented by the classification parameters employed. TurfAnalyzer software outputs image files that display mask intersection results. However, these images are of reduced resolution and do not depict the original pixels themselves. The Python approach used in this paper enabled the extraction of full-resolution pixel selections output to individual image files that could be closely evaluated.
As part of the exploration process, consideration was given to the Python-based plant phenotyping research software PlantCV (
https://plantcv.danforthcenter.org/). Despite its provisions for color correction and many analysis functions, its design did not align with the generic approach of this investigation. Traditional PlantCV operation requires a specific color correction panel to be present in each image, which can pose a geometric obstacle inside a compact lightbox. Although having a dedicated correction reference panel in each image provides for superior batch correction compared to using only one reference image for an entire set, the production and processing of dedicated correction panel views increases complexity, reduces the usable image analysis view space, and may not overcome errors due to inconsistent illumination across a single image. Professional software, such as Adobe Photoshop and Lightroom, are often used to color-correct imagery [
163,
164,
165,
166]. However, these tools were not utilized in this study absent their commercial license. Therefore, although color correction is advantageous for comparing imagery taken at different times, in different illumination scenarios, and from different cameras, the benefit of color correction must be weighed against the additional complexity it introduces. Moreover, there is a possibility that color correction can introduce errors that may be difficult to quantify. Thus, the decision to implement color correction should be made judiciously, considering its potential positive impact and the trade-offs involved.
Achieving geometric regularity in images through lens distortion correction is a common enhancement, but it may be challenging to implement. An improved lens correction process could aid higher-quality data results in plant phenotyping [
167]. Although the Lensfun database of lens artifacts was available for the Nikon lens used, authors found the correction to be insufficient even after a second manual adjustment was included. This may be due to minor manufacturing imperfections unique to the specific lens used. While RawTherapee (Creative Commons 4.0), GIMP (Crunchbase Company, Charlotte, NC, USA), and Pablo d’ Angelo’s Hugin software (
https://hugin.sourceforge.io/) were employed for lens correction, the Python Discorpy package emerged as the most effective solution, offering well-documented tutorials. However, further refinements may still be possible. For example, nominal changes were observed for Hue angles (0.002%) between JPG and JPG LC, while more substantial alterations occurred for Saturation (decreased by 3.8%) and Value (increased by 11.6%). This equated to an increase in RGB (Red by 12.2%, Green by 11.6%, and Blue by 14.7%), therefore LC brightened as well as geometrically corrected images. Consequently, the accuracy of the corrections, a potential new source of error, the added challenge in making corrections for each camera setup, and a subsequent benefit from LC must be balanced together. A straightforward LC result verification is needed to better understand the extent to which LC may improve or degrade a particular imagery set. Separately, the related lens vignette was easy to correct using the RawTherapee software. The resultant brightness gradient from image edge to lysimeter edge across the white background could be sampled to verify basic brightness uniformity. Noting that the artifact effect was negligible for our dataset, other phenotyping images may require additional steps to account for vignetting. Despite challenges, image corrections are expected to improve the accuracy and reliability of resultant data. This calls for continued exploration and refinement of correction methodologies across the diverse imaging setups used in plant phenotyping.
Although a lightbox with many LEDs can provide consistent illumination to facilitate the capture of extensive gridded optical data [
168,
169,
170], its usage imposes major limitations on sample throughput. Inserting each individual lysimeter into the lightbox for imaging was considerably slower (approximately 60 seconds) than the VQ assessment (approximately 3 seconds) or NDVI measurement (approximately 15 seconds). Although a robotic high-throughput plant phenotyping greenhouse can address this restriction [
171,
172,
173,
174,
175,
176], its adoption comes at substantial monetary and complexity cost, albeit more generic versions are being developed [
177]. Some experiments have imaged multiple greenhouse plants outside the lightbox omitting robotics [
154], but the ambient illumination is likely more varied across each image. Even when a color correction panel is present within each image collected, inconsistent lighting across the entire image can compromise the accuracy of color correction. Hence, improved methods may entail consistent illumination [
178] and enable plants to be imaged without the need for relocation.
Optical cameras only measure in the RGB color spectrum (approximately 420 to 700 nm). Beyond this range lies spectral information that is significant for researchers but surpasses the detection capacity of standard cameras. Consequently, the optical camera alone cannot detect the complete spectral reflection of vegetation. Another limitation of the camera sensor is the 2-D representation it generates. Although photogrammetric or structure-from-motion techniques can be used to amalgamate multiple images into 3-D point clouds, incorporating a second view perpendicular to the nadir would provide more plant information and allow for a basic 2.5-D model with just two images, and reduced compute requirement.
Incorporating an up-looking optical diffuser with filters that transmit the RGB frequencies of the camera detector, similar to the design of some UAS multi-spectral cameras [
179] like the MicaSense Altum or RedEdge cameras (AgEagle Aerial Systems Inc., Wichita, KS, USA), which are designed specific to their spectral sensitivities, would enable a solar cosine correction [
180] frame by frame onboard the camera. This may diminish the value of color correction post-processing and greatly expedite the information process while increasing data quality.
The NDVI measurement used in this paper was derived from a unique active optical sensor technology. Unlike passive sensors, this technology allows measurements in various lighting conditions. Although the instrument used provides a set of raw values and basic statistics from its high-frequency sampling, it only offers a single data point per measurement. An NDVI camera would provide additional spatial information through its pixel grid; however, considerations related to ambient light would come into play. The authors are unaware of any active NDVI camera product.
Micrometeorological measurements of the growing environment are important in plant phenotyping. At a minimum, measurement of photosynthetic radiation, along with sensible and latent heat, provides a context for plant growth (
Appendix A). Even minor discrepancies between the expected environmental values, such as air temperature, and the actual values measured at the precise location of the plants, can accumulate over a growing season and impact results. There can also be short-duration events, such as loss of greenhouse controls or abrupt outside weather changes, which can affect plant growth inside a greenhouse. These events may go unnoticed without continuous environmental measurements at sufficient temporal resolution. In this study, greenhouse environments were measured at a frequency of 1Hz, ensuring a comprehensive characterization of the growing conditions. However, the plant images were only captured every two weeks. Increasing the frequency of image capture would facilitate the association of temporal growth indicators present in the imagery data with the measured environment. This would lead to a more comprehensive understanding of environmental effects on observed biological results.