Sick crops? These Indian subsistence farmers know just what to do: Pull out their smartphones and take their picture.
The farmers then upload the images with GPS locations to a cloud-based artificial intelligence (AI) app named Plantix. The app identifies the crop type in the image and spits out a diagnosis of a disease, pest or nutrient deficiency.
Plantix also aids farmers by recommending targeted biological or chemical treatments for ailing plants, reducing the volume of agrochemicals in groundwater and waterways that can result from overuse or incorrect application of herbicides and pesticides.
“Nearly every household in India has a smartphone, and many want to see how Plantix works,” says Srikanth Rupavatharam, a digital agriculture scientist with the International Crops Research Institute for the Semi-Arid Tropics (ICRISAT) in Hyderabad, India, which is collaborating with Plantix’s developer to adapt the tool for Telugu and Hindi languages.
Better machines and fewer workers—we’ve seen that in every other industry. Now it’s our turn.
Karen Marie Lewis, tree-fruit horticulturist, Washington State University
For common crop diseases in India, Rupavatharam says, the app has an identification success rate of more than 90 per cent.
Plantix is based on deep learning, one of today’s most powerful AI tools. Deep learning involves neural networks — digital imitations of the human brain’s system of neurons and synapses.
Deep learning models are trained to look for certain patterns in giant datasets. They also go beyond basic pattern recognition to devise their own rules as they go, deciding how best to perform their jobs.
Deep learning isn’t a new idea; it’s been around for decades. But over the past five years, deep neural networks have advanced very rapidly because of super-fast processing hardware and huge volumes of data from the internet to train models.
Deep learning already reaches into our daily lives: Facebook, for instance, uses deep learning to identify and tag the faces of our friends in posted photos. Now ag-tech companies are testing deep learning as a tool to boost productivity and reduce environmental impacts while helping farmers carry out tasks that people now perform.
As in every other industry, AI advances could lead to greater efficiencies but also job losses.
Opportunity and challenge
The world must produce at least 60 per cent more food relative to 2006 to feed an estimated global population of 9 billion by 2050, according to United Nations figures. But most arable land is already cultivated, so farmers must find ways to grow more food on the acres they already cultivate.
To improve farmers’ efficiency, researchers are experimenting with deep learning models that help guide robots and drones in monitoring fields for pests and predators, tracking crop and livestock diseases, and many other tasks. Ag-tech companies are testing deep learning in robots to identify the “faces” of fruit and pick them more rapidly. Someday, a farmer in a developed country could remain in an office all day remotely deploying AI-driven machinery.
As exciting as such applications are, they are not without challenges. Deep learning has proven successful in highly structured, predictable environments such as those involved in medical care. Farms, though, are far messier places.
“Agriculture is a little behind in adoption of advanced digital tools,” says Shriram Ramanathan, who heads the big data analytics practice at Lux Research, a technical innovation consulting firm.
“Deep learning models have improved a long way, but once you go outdoors, there are so many factors that can affect their accuracy.” Lighting conditions, shadows, weather, dust and other factors outdoors can change by the hour. A plant can look quite different at dawn, noon, or evening. If poor-quality images are fed into a deep learning model, the algorithm is more likely to misidentify the plant or its ailment.
Relatively few small and medium-size U.S. farms are experimenting with new digital tools, while larger farms with more capital are willing to take on more risk, says Ramanathan. “If you are a farmer on a very tight budget with a low-margin crop such as corn, there’s no time to run a proof of concept with a digital technology.”
That said, two small companies, California-based Abundant Robotics and Israel-based FFRobotics, are racing to develop the world’s first apple-picking robot guided by deep learning. Mounted on tractors, these machines use cameras to recognize individual apples and robotic arms and other devices to gently pick them from trees.
Abundant Robotics field-tested its apple-picking machine in Washington state orchards in 2017. FFRobotics tested a machine in Israel over the past three years, and plans to test in Washington state in 2018.
Part of the rationale behind the rush is political. Apple growers face financial disaster if fruit isn’t picked on time. Nearly one-fourth of Washington’s seasonal agricultural workers arrive on H-2A (temporary agricultural worker) guest visas, mostly from Latin America. Growers worry that future federal policies will reduce their seasonal workforce, motivating interest in robotic replacements, says Karen Marie Lewis, a Washington State University tree-fruit horticulturist.
“Better machines and fewer workers—we’ve seen that in every other industry,” Lewis says. “Now it’s our turn.”
Weeds vs. cotton
Some major agricultural companies and high-tech subsidiaries are betting on deep learning. John Deere Labs, which opened in 2017, spent US$305 million in September to purchase Blue River Technology, a startup based in Sunnyvale, California.
Blue River has developed a tractor-mounted device called See & Spray that deploys two color cameras, computer vision and a deep-learning algorithm to detect weeds and cotton plants in milliseconds and spray herbicide just on the weeds.
To train its algorithm, Blue River Technology fed it accurately labeled, high-resolution images of cotton plants and weeds at various times of day. “The algorithm has to be very robust to account for different field conditions,” says Lee Redden, the company’s co-founder and chief technology officer.
The more images that the model “sees,” the more accurate it becomes in recognizing plants. The model’s first computational layers detect simple elements in plant images such as edges and corners.
The next layers can bring together corners and edges to form features such as leaf margins. Each successive layer of the model builds on the knowledge of the previous layers. The software becomes increasingly expert at recognizing features that the programmers instruct it to look for. When the model misidentifies a plant, the programmers flag the result and adjust the model.
This story was written by John H. Tibbetts for Ensia.com and was published with permission. Read the full story.