Precision agriculture has been an application darling of remote sensing since the 1980s. With each new optical satellite constellation has come renewed focus on developing new applications and pushing existing methods for soil management, weed detection, yield measurement, crop mapping, vegetation health and many other areas.
It hasn’t just been the domain of optical satellites, however. Aerial photography has played a major role in advancing operational precision agriculture applications – the United States Department of Agriculture has its own National Agriculture Imagery Program (NAIP) dedicated to acquiring aerial imagery during the agricultural growing seasons in the continental U.S. And now, unmanned aerial vehicle (UAV) data is beginning to take root, piquing the interest of an industry that has come to rely heavily on the field intelligence that geospatial imagery, particularly high resolution data, can provide.
As the image precision has gotten better so too has the need for advanced research and powerful geospatial tools such as object-based image analysis (OBIA) technology to learn best how to harvest such data-rich information into efficient, effective and reliable tools of the precision agriculture trade.
“We are collecting an unimaginable amount of high-resolution imagery that is packed with valuable information, but we still need equally powerful tools to handle this volume of data and its complexity to discover how best to use it,” says Ovidiu Csillik, a Romanian postdoctoral research associate at the Carnegie Institution for Science in Stanford. “In agriculture, for example, the ability to automatically identify and delineate individual trees from high resolution optical imagery can help growers better manage their orchards and plantations. The question is always how do you do that?”
Answering those types of questions has been Csillik’s primary focus for the past seven years. Using a host of sources like optical satellite imagery in combination with OBIA software, he has been both developing and testing automated approaches to agricultural classifications and mapping and environmental applications to determine the feasibility of using these methods operationally.
As a visiting research scholar at UC Berkeley’s Department of Environmental Science, Policy, and Management in 2017, Csillik had the opportunity to build on his satellite-imagery based OBIA studies and develop a whole new application: automatically identifying and mapping individual citrus trees using UAV imagery, a convolutional neural network (CNN) deep-learning (DL) algorithm, and OBIA processing.
“Deep-learning algorithms are gaining a resurgence in big, geospatial data analysis because they can be trained to accurately classify large-scale datasets,” says Csillik. “Although never done before, I wanted to test the feasibility of integrating a CNN into OBIA software to automatically identify multiage citrus trees from UAV imagery. Being able to automate delineating and mapping trees would give orchard managers an efficient and precise crop management tool.”
Using the Lindcove Research and Extension Center (LREC) as his testbed, he hit paydirt with his first-ever test. A successful example of how UAV data and OBIA technology can positively impact the precision agriculture market, Csillik’s method is sowing the seeds for a viable new approach to citrus tree management, starting with managers at LREC.
The Golden Fruit
The University of California Division of Agriculture and Natural Resources’ LREC has deep roots in California’s citrus industry. Spanning from the San Joaquin Valley – where 75 per-cent of the state’s citrus is grown – to the Sierra Nevada foothills, the 175-acre LREC has been a critical link between cutting-edge citrus research and growers since 1959. More than 500 million citrus trees in California have been produced from disease-tested budwood that originated from Lindcove.
In California, citrus is gold. Valued at more than $2 billion annually, the citrus industry is a bedrock of the state’s agriculture economy and the lifeblood for more than 3,000 growers who farm 320,000 acres of citrus.
It’s a good bet that the root of any given tree on one of those farms can be traced back to the LREC, which today manages nearly 600 tree crop varieties – mostly citrus species – of mixed ages and sizes that are nurtured by soils and climate that are aligned with the commercial citrus growing in the San Joaquin Valley.
Maintaining such an agricultural research and development gem has spurred the LREC’s managers to continually adopt methods and tools that will ensure the crops are available to the hundreds of scientists and educators from all over the world who come for research. One recent innovation was producing a reference tree database of the location and attribute data of 2,912 individual trees using 1-meter NAIP imagery from 2012. But in 2017, they began to wonder how UAV imagery could benefit them.
At the same time, Csillik and his research supervisor Dr. Maggi Kelly, a professor in UC Berkeley’s Environmental Science, Policy and Management department were interested in how artificial intelligence and Trimble’s eCognition OBIA technology could be used to map the trees accurately. eCognition is an OBIA information extraction technology that employs user-defined processing workflows called rule sets, which automatically detect and classify specified objects and map them.
“Drone imagery, with its ability to provide sub-meter imagery on demand, has the potential to revolutionize precision agriculture workflows,” says Csillik. “eCognition can not only handle the complexity and large data volumes of UAV, it offers unlimited opportunities to use its algorithms to analyze spatial data and identify any object. This study would give us the chance to acquire UAV data as a proof-of-concept alternative to NAIP imagery, and to test the feasibility of using a CNN-OBIA method to automatically analyze the UAV imagery and accurately identify and map individual citrus trees.”
Citrus at the Center
In January 2017, UAV imagery was acquired over the entire 175-acre LREC site. Carrying a Parrot Sequoia multi-spectral camera with average along- and across-track overlaps of 75 percent and 75 percent, the aerial system flew at an altitude of 104 m (340 feet) and used SRTM terrain data to ensure consistent altitudes across the captured area. In two flights, the UAV captured 4,574 multispectral images, which were then photogrammetrically processed to produce a 4-band orthoimagery mosaic with a ground sample distance of 12.8 cm. The red and near infra-red bands were also used to create a Normalized Difference Vegetation Index (NDVI) image. Both datasets were used as source data for eCognition.
With the image layers prepared, Csillik could focus on developing the software’s rule set workhorse to process and classify the data. Since they used intensively managed citrus trees at the LREC, there were high diversities of tree characteristics – pruned, unpruned, different ages, different geometries, touching crowns – so he needed a rule set that would enable eCognition to delineate the trees under these varied conditions. His approach would enhance the software’s inherent analysis and object-detection aptitude by adding in eCognition’s new deep-learning, CNN algorithms.
“eCognition is often the tool of choice for identifying and mapping vegetation because it’s so good at extracting certain features and accurately mapping them,” says Csillik. “Adding a CNN model into the software would allow it go deeper into understanding what is a tree and locate it in the image to produce a more accurate result. I can also add this deeper level of analysis without having to write any code, saving me significant time.”
The first step of the ruleset involved training the CNN model with three classes–trees, bare soil and weeds–with 4,000 training samples per class. To create the samples, he input the four-band orthomosaic into eCognition, along with the tree-location points from the LREC’s existing trees database, and instructed the software to break up the mosaic into 40 by 40 pixel samples around those points. Sample collection was limited to the northern half of the LREC, leaving the southern half available for validation. Since the classes of the 12,000 samples were known, they could be used to teach the network to differentiate between classes by learning the specific features of each class. A quick study, the CNN completed its training in 13 minutes.
“The CNN mimics how the brain’s neural cortex identifies certain objects,” says Csillik. “For example, the brain recognizes a car because it has four wheels, windows, doors, etc. By showing the same image samples to the CNN thousands of times, it learns what a tree looks like in an image, even if the landscape is more complicated, or the tree is bigger or smaller, or pruned or unpruned.”
Now he was ready to run the second phase of the eCognition rule set: testing the newly trained CNN to identify and delineate individual citrus trees. Focusing on the southern half of the study site, the CNN used each 40 by 40 pixel region to obtain a likelihood that the image patch contains a citrus tree. Moving the analysis region like a sliding-window across the UAV-based orthomosaic, it searched for objects that match the tree features. When it identified a tree in an image, it located it, designated a high value to it and then moved on. In total, it took eCognition’s CNN two minutes to analyze the orthomosaic and produce a probability “heat map,” a grayscale image where higher (brighter) values correspond to likely tree locations.
The software then further processed the heat map to refine the tree objects and extract the tree center. Adding in the NDVI layer, Csillik used a Simple Linear Iterative Clustering algorithm to segment the heat map and the NDVI layer into another newly integrated feature called superpixels. With this process, eCognition constrains the segmentation towards circular objects in the heat map and constrains it to differentiate between green and bare soil in the NDVI layer, enabling it to remove multiple crown-detection errors and refine its tree delineation, improving the detection and location of single citrus trees. This task took less than 10 seconds.
In total, eCognition analyzed, identified and delineated 3,105 individual trees in 30 minutes.
To validate how well the approach performed, Csillik compared the CNN/OBIA-generated results with the LREC’s existing trees database. Using common evaluation statistics, he calculated True Positives (perfect tree identification), False Positives (tree is incorrectly identified; a commission error) and False Negatives (a tree is missed; an omission error). Based on those calculations, the overall accuracy of his debut approach and final classification was 96.2 percent–an unexpected “nice” surprise.
“Having worked with eCognition for over seven years, I was confident that the integrated CNN would perform well,” says Csillik. “But it was a nice surprise to see how well it performed. With 96 percent accuracy, that level of automated precision could provide the ability to identify and delineate individual trees using automated methods as an alternative to manual delineation. Applying this process to high-resolution UAV imagery would offer a much faster, precise and repeatable method for long-term crop management for growers and orchard managers like those at the LREC.”
Those LREC managers are indeed already intrigued by the idea. After seeing Csillik and his team’s classified map of citrus trees, they initiated plans to apply the approach to other study areas, explore how transferable it is, and determine how easily they can update their tree maps over time.
Csillik will be the first to say that more research on pairing UAV imagery with CNN-OBIA technology is needed to fully realize how this approach can better the precision agriculture industry. But he is optimistic, both about this initial result and future ones to come.
“I think drone imagery analysis is going to be a big part of precision agriculture going for-ward,” says Csillik. “Since it offers imagery on demand, you can plan the time of day, season, spatial resolution, and spectral resolution you need. Analyzing that data with the speed, intelligence and adaptability of deeper-learning OBIA not only provides a processing environment that can manage significant data volumes it can readily classify and map field assets down to the individual tree. That’s precision.”
For California citrus growers and researchers, that’s a golden opportunity.