The Prometheus League
Breaking News and Updates
- Abolition Of Work
- Ai
- Alt-right
- Alternative Medicine
- Antifa
- Artificial General Intelligence
- Artificial Intelligence
- Artificial Super Intelligence
- Ascension
- Astronomy
- Atheism
- Atheist
- Atlas Shrugged
- Automation
- Ayn Rand
- Bahamas
- Bankruptcy
- Basic Income Guarantee
- Big Tech
- Bitcoin
- Black Lives Matter
- Blackjack
- Boca Chica Texas
- Brexit
- Caribbean
- Casino
- Casino Affiliate
- Cbd Oil
- Censorship
- Cf
- Chess Engines
- Childfree
- Cloning
- Cloud Computing
- Conscious Evolution
- Corona Virus
- Cosmic Heaven
- Covid-19
- Cryonics
- Cryptocurrency
- Cyberpunk
- Darwinism
- Democrat
- Designer Babies
- DNA
- Donald Trump
- Eczema
- Elon Musk
- Entheogens
- Ethical Egoism
- Eugenic Concepts
- Eugenics
- Euthanasia
- Evolution
- Extropian
- Extropianism
- Extropy
- Fake News
- Federalism
- Federalist
- Fifth Amendment
- Fifth Amendment
- Financial Independence
- First Amendment
- Fiscal Freedom
- Food Supplements
- Fourth Amendment
- Fourth Amendment
- Free Speech
- Freedom
- Freedom of Speech
- Futurism
- Futurist
- Gambling
- Gene Medicine
- Genetic Engineering
- Genome
- Germ Warfare
- Golden Rule
- Government Oppression
- Hedonism
- High Seas
- History
- Hubble Telescope
- Human Genetic Engineering
- Human Genetics
- Human Immortality
- Human Longevity
- Illuminati
- Immortality
- Immortality Medicine
- Intentional Communities
- Jacinda Ardern
- Jitsi
- Jordan Peterson
- Las Vegas
- Liberal
- Libertarian
- Libertarianism
- Liberty
- Life Extension
- Macau
- Marie Byrd Land
- Mars
- Mars Colonization
- Mars Colony
- Memetics
- Micronations
- Mind Uploading
- Minerva Reefs
- Modern Satanism
- Moon Colonization
- Nanotech
- National Vanguard
- NATO
- Neo-eugenics
- Neurohacking
- Neurotechnology
- New Utopia
- New Zealand
- Nihilism
- Nootropics
- NSA
- Oceania
- Offshore
- Olympics
- Online Casino
- Online Gambling
- Pantheism
- Personal Empowerment
- Poker
- Political Correctness
- Politically Incorrect
- Polygamy
- Populism
- Post Human
- Post Humanism
- Posthuman
- Posthumanism
- Private Islands
- Progress
- Proud Boys
- Psoriasis
- Psychedelics
- Putin
- Quantum Computing
- Quantum Physics
- Rationalism
- Republican
- Resource Based Economy
- Robotics
- Rockall
- Ron Paul
- Roulette
- Russia
- Sealand
- Seasteading
- Second Amendment
- Second Amendment
- Seychelles
- Singularitarianism
- Singularity
- Socio-economic Collapse
- Space Exploration
- Space Station
- Space Travel
- Spacex
- Sports Betting
- Sportsbook
- Superintelligence
- Survivalism
- Talmud
- Technology
- Teilhard De Charden
- Terraforming Mars
- The Singularity
- Tms
- Tor Browser
- Trance
- Transhuman
- Transhuman News
- Transhumanism
- Transhumanist
- Transtopian
- Transtopianism
- Ukraine
- Uncategorized
- Vaping
- Victimless Crimes
- Virtual Reality
- Wage Slavery
- War On Drugs
- Waveland
- Ww3
- Yahoo
- Zeitgeist Movement
-
Prometheism
-
Forbidden Fruit
-
The Evolutionary Perspective
Category Archives: Ai
Farmers are using AI to spot pests and catch diseases and many believe its the future of agriculture – INSIDER
Posted: November 9, 2019 at 8:42 am
In Leones, Argentina,a drone with a special camera flies low over 150 acres of wheat. It's able to check each stalk, one-by-one, spottingthe beginnings of a fungal infection that could potentially threaten this year's crop.
The flying robot is powered by computer vision: a kind of artificial intelligence being developed by start-ups around the world, and deployed by farmers looking for solutions that will help them grow food on an increasingly unpredictable planet.
Many food producers are struggling to manage threats to their crop like disease and pests, made worse by climate change, monocropping, and widespread pesticide use.
Catching things early is key.
Taranis, a company that works with farms on four continents, flies high-definition cameras above fields to provides "the eyes."
Machine learning a kind of artificial intelligence that's trained on huge data sets and then learns on its own is the "brains."
"I think that today, to increase yields in our lots, it's essential to have a technology that allows us to take decisions immediately," said Ernesto Agero, the producer on San Francisco Farm in Argentina.
The algorithm teaches itself to flag something as small as an individual insect, long before humans would usually identify the problem.
Similar technology is at work in Norway's fisheries, where stereoscopic cameras are a new weapon in the battle against sea lice, a pest that plagues farmers to the tune of hundreds of millions of dollars.
The Norwegian government is considering making this technology, developed by a start-up called Aquabyte, a standard tool for farms across the country.
Farmers annotated images to create the initial data set. Over time, the algorithm has continued to sharpen its skills with the goal of finding every individual louse.
But deploying computer vision is expensive, and for many it's still out of reach.
Bigger industrial farms tried using computer vision to identify and remove sick pigs at the outset of an African swine fever epidemic that is sweeping China, according The New York Times.
But half of China's farms are small-scale operations like this one, where that wasn't an option.
Chinese pig farmer Fan Chengyou lost everything.
"When the fever came, 398 pigs were buried alive," Chengyou said. "I really don't want to raise pigs anymore."
China the world's biggest pork producing country is expected to lose half its herd this year.
For many farmers in the world's major growing regions, 2019 was devastating.
Record flooding all along the Mississippi River Valley the breadbasket of the United States meant that many farmers couldn't plant anything at all this season.
And while computer vision can't stop extreme weather, it isat the heart of a growing trend that may eventually offer an alternative, sheltered from the elements.
Root AI enlists computer vision to teach its robots to pick fruit. Root AI
"Indoor growing powered by artificial intelligence is the future," said Josh Lessing, co-founder and CEO of Root AI, a research company that develops robots to assist in-door farmers.
Computer vision has taught a fruit-picking robot named Virgo to figure out which tomatoes are ripe, and how to pick them gently, so that a hot house can harvest just the tomatoes that are ready, and let the rest keep growing.
The Boston-based start-up is installing them at a handful commercial greenhouses in Canada starting in 2020.
80 Acres Farms, another pioneer in indoor growing, opened what it says is the world's first fully-automated indoor growing facility just last year.
The company, based in Cincinnati, currently has seven facilities in the United States, and plans to expand internationally over the next six months. Artificial intelligence monitors every step of the growing process.
"We can tell when a leaf is developing and if there are any nutrient deficiencies, necrosis, whatever might be happening to the leaf," said 80 Acres Farms, CEO, Mike Zelkind. "We can identify pest issues, we can identify a whole variety of things with vision systems today that we can also process."
Because the lettuce and vine crops are grown under colored LED lights, technicians can even manage photosynthesis
Thanks to the benefits of indoor-farming practices, Zelkind says 80 Acres Farms' crops grow faster and have the potential to be more nutrient-dense.
Humans need more than salad to survive, though. Experts say indoor farms will need to expand to a more diverse range to provide a comprehensive option for growing food, but the advances being made in this space are significant.
AI-powered indoor agriculture is attracting a whole new breed of farmer.
New techie farmers are ambitious, but they are also realistic about what it takes to make AI work.
Ryan Pierce comes from a cloud computing background, but decided to jump into indoor growing, despite little to no experience in agriculture. Now, Pierce works for Fresh Impact Farms, an indoor farm in Arlington, VA.
"It's really sexy to talk about AI and machine learning, but a lot of people don't realize is the sheer amount of data points that you actually need for it to be worthwhile," Pierce said.
There is a ways to go before artificial intelligence can truly solve the issues facing agriculture today and in the future.
Many AI projects are still in beta, and some have proven too good to be true.
Still, the appetite is high for finding solutions at the intersection of data, dirt and the robots that are learning to help us grow food.
AI for agriculture is valued at $600 million, and expected to reach $2.6 billion by 2025.
Go here to see the original:
Posted in Ai
Comments Off on Farmers are using AI to spot pests and catch diseases and many believe its the future of agriculture – INSIDER
OpenAI has published the text-generating AI it said was too dangerous to share – The Verge
Posted: at 8:42 am
The research lab OpenAI has released the full version of a text-generating AI system that experts warned could be used for malicious purposes.
The institute originally announced the system, GPT-2, in February this year, but withheld the full version of the program out of fear it would be used to spread fake news, spam, and disinformation. Since then its released smaller, less complex versions of GPT-2 and studied their reception. Others also replicated the work. In a blog post this week, OpenAI now says its seen no strong evidence of misuse and has released the model in full.
GPT-2 is part of a new breed of text-generation systems that have impressed experts with their ability to generate coherent text from minimal prompts. The system was trained on eight million text documents scraped from the web and responds to text snippets supplied by users. Feed it a fake headline, for example, and it will write a news story; give it the first line of a poem and itll supply a whole verse.
Its tricky to convey exactly how good GPT-2s output is, but the model frequently produces eerily cogent writing that can often give the appearance of intelligence (though thats not to say what GPT-2 is doing involves anything wed recognize as cognition). Play around with the system long enough, though, and its limitations become clear. It particularly suffers with the challenge of long-term coherence; for example, using the names and attributes of characters consistently in a story, or sticking to a single subject in a news article.
The best way to get a feel for GPT-2s abilities is to try it out yourself. You can access a web version at TalkToTransformer.com and enter your own prompts. (A transformer is a component of machine learning architecture used to create GPT-2 and its fellows.)
Apart from the raw capabilities of GPT-2, the models release is notable as part of an ongoing debate about the responsibility of AI researchers to mitigate harm caused by their work. Experts have pointed out that easy access to cutting-edge AI tools can enable malicious actors; a dynamic weve seen with the use of deepfakes to generate revenge porn, for example. OpenAI limited the release of its model because of this concern.
However, not everyone applauded the labs approach. Many experts criticized the decision, saying it limited the amount of research others could do to mitigate the models harms, and that it created unnecessary hype about the dangers of artificial intelligence.
The words too dangerous were casually thrown out here without a lot of thought or experimentation, researcher Delip Rao told The Verge back in February. I dont think [OpenAI] spent enough time proving it was actually dangerous.
In its announcement of the full model this week, OpenAI noted that GPT-2 could be misused, citing third-party research stating the system could help generate synthetic propaganda for extreme ideological positions. But it also admitted that its fears that the system would be used to pump out a high-volume of coherent spam, overwhelming online information systems like social media, have not yet come to pass.
The lab also noted that its own researchers had created automatic systems that could spot GPT-2s output with ~95% accuracy, but that this figure was not high enough for standalone detection and means any system used to automatically spot fake text would need to be paired with human judges. This, though, is not particularly unusual for such moderation tasks, which often rely on humans in the loop to spot fake images and videos.
OpenAI says it will continue to watch how GPT-2 is used by the community and public, and will further develop its policies on the responsible publication of AI research.
Excerpt from:
OpenAI has published the text-generating AI it said was too dangerous to share - The Verge
Posted in Ai
Comments Off on OpenAI has published the text-generating AI it said was too dangerous to share – The Verge
Goldman Sachs, Nationwide, and More Highlight Benefits of AI at H2O World New York – Forbes
Posted: at 8:42 am
Water can flow, or it can crash. Water is formless, shapeless. It becomes the form of whatever container you put it in. This famous idea from actor and philosopher Bruce Lee is a part of the mission behind H2O.ai (H2O). At H2O World New York on October 22nd, H2O.ai celebrated its vision to make every company an AI company with its growing community of makers and builders.
H2O.ai has developed enterprise AI and machine learning platforms to give companies the ability to easily access and leverage data throughout their organizations. Its open source platform, H2O, allows the entire open source ecosystem to contribute to the development of AI, and contributed to the development of newer platforms such as H2O Driverless AI, an automated machine learning tool that takes some of the most difficult workflows and achieve the highest predictive accuracy by making them easier to interpret, and H2O Q, its newest platform, which helps business users and aspiring AI companies make AI apps.
While many companies attempt to embrace AI, they often lack the resources, especially when recruiting top data scientists from tech giants like Google, Facebook, and Microsoft, whose compensation packages can exceed $1M.
Aakriti Srikanth, Forbes 30 Under 30 honoree and Co-founder at Ahura AI
I see a lot of potential in H2O.ai to accelerate machine learning and to improve the workforce, said Aakriti Srikanth, Co-founder at Ahura AI and a Forbes 30 Under 30.
There is a huge talent shortage for data scientists in many companies, said Sri Ambati, CEO and Founder in his opening keynote. H2O.ai solves this problem by giving teams easier, simpler and cheaper AI platforms that implement machine learning algorithms."
H2O World New York featured sessions from many of the H2O.ai employees (or makers), as well as representatives from some of the top companies in financials services, insurance, and more, including Goldman Sachs, Nationwide Insurance, Disney, and Discover Financial Services, among others, all of whom have adopted the companys technology, including open source H2O and Driverless AI, to further their AI journey.
A core requirement of any company focused on AI is putting together a strong team.
Wieyan Zhao, Director of Data Science at Nationwide Insurance said, What matters is the skills that you have, and are you good at, for the things we want you to do...we have people coming from more than ten countries, speaking more than fifteen types of languages; the diversity gives you perspective when you come into the room and try to solve the problem.
During his H2O World session, A Decade of Data Science: The Nationwide Journey, Zhao explained how Nationwides core modeling team has a 10-to-1 ratio of data scientists to operations, with a diverse background and skill gives that helps them bring unique perspectives during peer and model reviews for each project.
Part of building a strong team, comes from having technology that lends itself to collaboration and brings technical expertise where an organization may be lacking it internally. To that end, H2O.ai boasts of attracting a large cohort of Kaggle Grandmasters, the data science equivalent of a Chess Grandmaster, to lead the way in building algorithms and new capabilities that data scientists can use in their day to day work. H2O.ai currently employs 13 Kaggle Grandmasters, roughly about 10 percent of those that exist globally. This gives H2O.ai customers the expertise necessary in the platform.
One of the things were doing in addition to building Driverless AI itself is building a platform that allows data scientists and engineers to collaborate and to store and share models and then deploy them, said Tom Kraljevic, VP of Engineering at H2O.ai.
"H2O.ai just makes modeling so much faster. Not only to produce data products, but we also created a platform for our customers, said ADP Principal Data Scientist, Xiaojing Wang.
The H2O.ai team onstage at H2O World New York
Customer love was a constant theme throughout H2O World, and Ambati compared the AI ecosystem to a team sport. At H2O.ai, this means uniting internal team members and external partners with a shared mission that challenges the status quo and creates a strong community that will build incredible things with AI. In fact, H2O.ai takes customer love very seriously, turning two of their largest customers, Goldman Sachs and Wells Fargo into investors in their recent funding rounds. When customers become investors, that is the true test of advocacy, said Ambati.
Goldman Sachs Managing Director, Charles Elkan, gave a keynote at H2O World where he spoke about The Promise and the Perils of deploying machine learning in the realm of finance. He gave the example of a healthcare company that was able to utilize machine learning to analyze patient responses and follow up with a database of clarifying questions, allowing a physician to review these answers and apply his or her experience and understanding on a deeper level, greatly increasing the physicians productivity.
"In order to be able to create something awesome like a great data science team, we need to be able to take those chances organizationally with people so as to produce those great outcomes we have been looking for, said Krish Swamy, Senior VP of Data and Enterprise Analytics at Wells Fargo.
See the original post:
Goldman Sachs, Nationwide, and More Highlight Benefits of AI at H2O World New York - Forbes
Posted in Ai
Comments Off on Goldman Sachs, Nationwide, and More Highlight Benefits of AI at H2O World New York – Forbes
The how and why of AI: enabling the intelligent network – Ericsson
Posted: at 8:42 am
Its easy to throw in a buzz word like AI these days to get some attention. But thats not my only intention with this piece. My aim is to offer some valuable insights into how AI can make a real difference in the world of radio access networks (RANs), looking at the way it works as well as the multiple benefits it brings.
Let me elaborate
From a technology perspective, AI in the RAN is about letting the hardware adjust its decision-making capability and allowing it to learn from patterns. Learning can improve the baseband functionality, which otherwise acts based on hard-coded settings. Of course, the pattern observations and execution of AI and machine learning (ML) require a certain amount of computing capacity. They also depend on software algorithms for decision making, as well as one or several data sources to use for pattern analysis.
While device types vary to some extent, 70 percent of all those used today are smartphones*. Their ability to differentiate depends on the relative level of their capabilities. Smartphones range from feature phones that support voice and text only, to full-on high-end 5G-capable devices.
Within the high-end smartphone segment, traffic behavior patterns for data services mostly shift between bursty and continuous. But on the networks side, the necessary capacity should always be available, regardless of what kind of app you are using at any given moment whether its chat, video streaming or gaming, for instance.
Smartphones perform measurements all the time, without most users being aware of it. These measurements are necessary to manage the radio link and the mobility, and to control how much output power each device needs to use. The network collects the measurement data in order to decide on the best connection for the device. Smartphones also carry key information about their network capabilities, which they conveniently report to the network. For instance, not all smartphones have 5G, but for those that do, the node can prepare a tailored network connection for each particular user.
Neighbor cells also report to each other on the status of capabilities, connected users and current load. This information can also be taken into consideration.
Ultimately, the scale of the benefits that AI can provide is determined by the hardware in place and the location of the boards. The hardware components of a mobile network today have to meet huge requirements in terms of managing the mobility of thousands of users in each cell. Not only that: they must also make sure that no connections are dropped and that the service is responding at all times.
Of course, routing and more central functions are rather executed from the core network components. So the node base stations do not have to carry full responsibility for the effectiveness of the entire network on their shoulders. But real-time mobility functions are located at the edge of the network, on the node.
Todays node often houses GSM, WCDMA, LTE and NR on a single site not always on one baseband, but such installations are soon to become commonplace as well.
Applying ML to software functionalities boosts the strength of the network significantly, since many network functions can benefit from the same algorithms. But the advantage of this comes at a cost, with some computing power being seized by the AI technology.
An Ericsson baseband will however run ML in parallel with regular node traffic without reducing the capacity of the baseband. Thats because our AI engineers have optimized the algorithms so that they can analyze huge amounts of data in real time, enabling instant traffic prediction. All this is facilitated by Ericssons many-core architecture, which is the software platform design of choice that all RAN Compute products are based on.
The reality is, service providers expect full-steam performance from their legacy products, even when new network capabilities are added and Ericsson is aware of this. Service providers also like to minimize opex and they incur significant operational costs when site visits need to be carried out. Ericsson is aware of this as well, which is why our ML features are integrated with our software releases, which can be applied on a remote basis without the need for any site visits at all.
We have reaped the benefits of AI in many areas of our lives from movie offerings being handed to us on a plate, to the voice and face recognition apps in our smartphones to the optical scanning features of our credit cards. You can look at your phone with your eyes wide open and unlock it, you can register your credit card details using the camera in your device In all such cases, AI simplifies the use of our devices, automating the steps that would otherwise have to be carried out in a repetitive, manual fashion. Imagine the hassle!
On the mobile network side, the use of AI is similar but not quite the same. While the initial use cases have been about automation, they are also about improving network coverage and the user experience by anticipating the needs of devices.
One practical example of this is that the measurements that smartphones carry out which were mentioned previously can be reduced significantly. By shortlisting the top neighbor cells at every node, the device will get an equally shortened to-do list of frequencies to listen to. This means that instead of numerous background measurements being performed by the device, battery power is conserved to do other fun stuff with.
For the service provider, one main benefit of implementing AI will be the reduction in opex, as fewer node configurations need to be added manually. But even more importantly, their spectrum assets can be used more efficiency, and spectrum is a valuable resource that they tend to have to pay for dearly.
All in all, AI for radio access networks is a sound investment. Ericssons software will improve coverage and spectrum use, and boost throughput. Then service providers can sit back, relax and let the machines do the work.
Ericsson Mobility Report, June 2019
Learn more about Ericssons AI-powered radio access networks
Join the webinar on December 3 at 15.00 CET
Learn more about artificial intelligence in the October 31 episode of the Voice of 5Gpodcast
Excerpt from:
The how and why of AI: enabling the intelligent network - Ericsson
Posted in Ai
Comments Off on The how and why of AI: enabling the intelligent network – Ericsson
Artificial Intelligence Can Be Biased. Here’s What You Should Know. – FRONTLINE
Posted: at 8:42 am
Artificial intelligence has already started to shape our lives in ubiquitous and occasionally invisible ways. In its new documentary, In The Age of AI, FRONTLINE examines the promise and peril this technology. AI systemsare being deployed by hiring managers,courts, law enforcement, and hospitals sometimes without the knowledge of the people being screened. And while these systems were initially lauded for being more objective than humans, its fast becoming clear that the algorithms harbor bias, too.
Its an issue Joy Buolamwini, a graduate researcher at the Massachusetts Institute of Technology, knows about firsthand. She founded the Algorithmic Justice League to draw attention to the issue, and earlier this year she testified at a congressional hearing on the impact of facial recognition technology on civil rights.
One of the major issues with algorithmic bias is you may not know its happening, Buolamwini told FRONTLINE. We spoke to her about how she encountered algorithmic bias, about her research, and what she thinks the public needs to know.
This interview has been edited for length and clarity.
On her first encounter with algorithmic bias.
The first time I had issues with facial detection technology was actually when I was an undergraduate at Georgia Tech, and I was working on a robot. The idea with this robot was to see if I could get it to play peek-a-boo with me. And peek-a-boo doesnt really work if your robot cant see you, and my robot couldnt see me. To get my project done, I borrowed my roommates face. She was lighter skinned than I was. That was my first time really using facial analysis technology and seeing that it didnt work for me the same way it worked for other people.
I went on to do many things and became a graduate student at MIT and I started working on projects that used facial analysis technology, face detection. So one project I did was something called the Aspire Mirror. You look into a mirror, a camera detects your face and then a lion can appear on you, or you can be somebody youre inspired by
[I]t wasnt detecting my face consistently, so I got frustrated. So what do you do when you get frustrated with your program? You debug. I started trying to figure out ways to make it work. I actually drew a face on my hand, and the system detected the face on my palm. And I was like, Wait, wait, wait, if its detecting the face I just drew on my palm, then anythings a possibility now. So I looked around my office and the white mask was there. So I was like, Theres no way! But why not?
I pick up the white mask, and I put it on and its instantaneous when I put on that white mask, and I mean just the symbolism of it was not lost to me. This is ridiculous that the system can detect this white mask that is not a real person, but cannot necessarily detect my face. So this is really when I started thinking, Okay, lets a dig a bit deeper with whats going on with these systems.
On digging a bit deeper into facial analysis technology.
Here was a question: Do these systems perform differently on various faces? There was already a 2012 reportthat actually came out from an FBI facial analysis expert showing that facial recognition systems in particular worked better on white faces than black faces. They didnt work as well on youthful faces. And they didnt work as well on women as compared to men. This was 2012, and why I keep bringing that up is this was before the deep learning revolution
Now we had a different approach that was supposed to be working much better. My question was, given these new approaches to facial analysis and facial recognition, are there still biases? Because what Im experiencing, what my friends are experiencing and what Im reading about with reports that say, Oh, weve solved face recognition, or Were 97% accurate from benchmarks those reports were not lining up to my reality.
What I focused on specifically was gender classification. I wanted to choose something that I thought would be straightforward to explain, not that gender is straightforward its highly complex. But insomuch as we were seeing binary gender classification, I thought that would be a place to start. By this time my weekend hobby was literally running my face through facial analysis and seeing what would happen. So some wouldnt detect my face and others would label me male. And I do not identify as male. This is what led down that corridor.
On finding the gold standard benchmarks were not representative.
When I ran this test, the first issue that I ran into which gave me some more insight with the issue were talking about algorithmic bias was that our measures for how well these systems perform were not representative of the world. Weve supposedly done well on gold standard benchmarks. So I started looking at the benchmarks. These are essentially the data sets we use to analyze how well were doing as a research community or as an industry on specific AI tasks. So facial recognition is one of these tasks that people are benchmarked on all the time.
What I started to see was something I call power shadows when either the inequalities or imbalances that we have in the world become embedded in our data.
The thing is, we often times dont question the status quo or the benchmark. This is the benchmark, why would I question it? But sometimes the gold standard turns out to be pyrite. And that is what was happening in this case. When I went to look at the research on the breakdown of various facial analysis systems, what I found was one of the leading gold standards, labeled Faces in the Wild, was over 70% male and 80% white. This is when I started looking into more and more data sets and seeing that you had massive skews. Sometimes you had massive skews because you were using celebrities. I mean, celebrities dont necessarily look like the rest of the world. What I started to see was something I call power shadows when either the inequalities or imbalances that we have in the world become embedded in our data.
All this to say, the measures that we had for determining progress with facial analysis technology were misleading because they werent representative of people at least the U.S. in that case. We didnt have data sets that were actually reflective of the world, so for my thesis at MIT, I created what I call the Pilot Parliaments Benchmark. I went to UN womens websites, I got a list of the top 10 nations in the world by their representation of women in parliament. So I chose European countries and African nations to try to get a spread on opposite ends of skin types, lighter skin and darker skin. After I ran into this issue that the benchmarks were misleading, I needed to make the benchmark.
On what her research found.
Then finally, I could get to the research question. So I wanted to know how accurate are they at this reduced task of binary gender classification which is not at all inclusive when it comes to guessing the gender of the face? And it turned out that there were major gaps. This was surprising because these were commercially sold products. You know how the story goes. It turns out, the systems work better on male-labeled faces than female-labeled faces, they work better on lighter faces than darker-skinned faces.
But one thing we did for this study, which I would stress for anybody whos thinking about doing research in algorithmic bias or concerned with algorithmic bias and AI harms, is we did an intersectional analysis. We didnt just look at skin type. We didnt just look at gender. We looked at the intersection. And the inspiration for this was from Kimberl Crenshaw, a legal scholar who in 1989 introduced the term of intersectionality. What would happen with the analysis is if you did it in aggregate just based on race, or if you did it in aggregate based on just gender, you might find based on those axes that there isnt substantial evidence of discrimination. But if you did it at the intersection, you would find there was a difference. And so I started looking at the research studies around facial analysis technologies and facial recognition technologies and I saw that usually we just have aggregate results just one number for accuracy. People are just optimizing for that overall accuracy, which means we dont get a sense of the various ways in which the system performs for different types of people. Its the differences in the performance, the accuracy disparities that I was fascinated by, but not just on a single axis but also on the intersection. So when we did the intersectional breakdown oooh, it was crazy.
We werent doing anything to try to trick the system. It was an optimistic test. This is why I was very surprised, because even with this optimistic test, in the worst-case scenario for the darkest-skinned women, you actually had error rates as high as 47% on a binary classification task.
I shared the results with the companies and I got a variety of responses.But I think the overall response, at least with the first study, was there was an acknowledgement of an issue with algorithmic bias.
On how AI is already affecting peoples lives.
Theres a paper that just came out from Science which is devastating, showing risk assessment algorithms used in health care actually have racial bias against black patients. Were talking about health care where the whole point is to try to optimize the benefit and what they were seeing was because they used how much money is spent on an individual as a proxy for how sick they were, it turned out it was not a good proxy because black patients who were actually sick were being said to be not as sick as they were.
When these systems fail, they fail most the people who are already marginalized, the people who are already vulnerable.
You also have AIs that are determining the kind of ads people see. And so there have been studies that show you can have discriminatory ad targeting. Or you can have a situation where you have an ad for CEO and the system over time learns to present that CEO ad to mainly men. You were saying, how do you know if youve encountered bias the thing is you might never know if youve encountered the bias. Something that might happen to other people you see phenotypic fails with passport renewals. So you have a report from a New Zealand man of Asian descent being told that his eyes are closed and he needs to upload another photo. Meanwhile, his eyes are not closed. You have, in the UK, a black man being told his mouth is open. His mouth was not open. You have these systems that are seeping into every day.
You have AI systems that are meant to verify if youre who you say you are. And so one way that can happen is with ride share apps. Uber, for example, will ping drivers to have them verify their ID. Theres actually a report from trans drivers who were saying that they were being repeatedly [asked] to submit their IDs because they were not matching. They were being either kicked out of the system or having to stop the car, test it again, which means youre not getting the same level of economic opportunity.
When these systems fail, they fail most the people who are already marginalized, the people who are already vulnerable. And so when we think about algorithmic bias, we really have to be thinking about algorithmic harm. Thats not to say we dont also have the risk of mass surveillance, which impacts everybody. We also have to think about whos going to be encountering the criminal justice system more often because of racial policing practices and injustices.
On what the public needs to know about algorithmic bias.
Theres no requirement for meaningful transparency, so these systems can easily be rolled out without our ever knowing. So one thing I wish people would do more of and something that companies also ought to do more of is having transparency so that you even know that an AI system was used in the first place. You just might never get the callback. You just might pay the higher price. You would never actually know. What I want the public to know is AI systems are being used in hidden ways that we should demand are made public.
The other thing I want the public to have is actually choice affirmative consent. Not only should I know if an AI system is being used, but lets say it makes the wrong decision or something that I contest. Theres no path to due process thats mandated right now. So if something goes wrong, what do you do?
Sometimes Ill hear, at least in the research community, efforts to de-bias AI or eradicate algorithmic bias. And its a tempting notion, lets just get rid of the bias and make the systems more fair, more inclusive, some ideal. And I always ask, but have we gotten rid of the humans? Because even if you create some system you believe is somehow more objective, its being used by humans at the end of the day. I dont think we can ever reach a true state of something being unbiased, because there are always priorities. This is something I call the coded gaze. The coded gaze is a reflection of the priorities, the preferences and also the prejudices of those who are shaping technology. This is not to say we cant do our best to try to create systems that dont produce harmful outcomes. Im not saying that at all. What I am saying is we also have to accept the fact that being human were going to miss something. Were not going to get it all right.
What I want the public to know is AI systems are being used in hidden ways that we should demand are made public.
Instead of thinking about Oh, were going to get rid of bias, what we can think about is bias mitigation knowing that we have flaws, knowing that our data has flaws, understanding that even systems we try to perfect to the best of our abilities are going to be used in the real world with all of its problems.
Before we get to the point where its having major harms with real world consequences, there need to be processes in place to check through different types of bias that could happen. So, for example, AI [systems]now have algorithmic risk assessments that they have as a process of really thinking through what the societal impact of the system are in its design and development stages before you get to the deployment. Those kinds of approaches, I believe, are extremely helpful, because then we can be proactive instead of reacting to the latest headline and playing bias whack-a-mole.
On proposals for oversight and regulation.
You have a proposal for an Algorithmic Accountability Act, this is a nationwide push that would actually require assessing systems for their social impact. And I think thats really important. We have something with the Algorithmic Justice League thats called the Safe Face Pledge, which outlines actionable steps companies can take to mitigate harms of AI systems.
I absolutely think regulation needs to be the first and foremost tool, but alongside regulation providing not just the critique of whats wrong with the system, but also steps that people can take to do better. Sometimes the step to take to do better is to commit to not developing a particular kind of technology or particular use case for technology. So with facial analysis systems, one of our banned uses is any situation where lethal force can be used. So it would mean were not supporting facial recognition on police body cameras. Or facial recognition on lethal autonomous weapons.
And I think the most important thing about the Safe Face Pledge that Ive seen is one, the conversations that Ive had with different vendors, where whether or not they adopt it actually going through those steps and thinking about their process and changes they can make in the process I believe has made internal shifts that likely would not hit the headlines. Because people would rather quietly make certain kinds of changes. The other thing is making it where the commitments have to be part of your business processes. Not a scouts honor pledge, just trust us. If you are committed to actually making this agreement, it means you have to change your terms of service and your business contracts to reflect what these commitments are.
On what should be done to fix the problem.
One, I think, demand transparency and ask questions. Ask questions if youre using a platform, if youre going to a job interview. Is AI being used? The other thing I do think is supporting legislative moves.
When I started talking about this, I think in 2016, it was such a foreign concept in the conversations that I would have. And now, today, I cant go online without seeing some kind of news article or story about a biased AI system of some shape or form. I absolutely think there has been an increase in public awareness, whether through books like Cathy ONeils Weapons of Math Destruction. Theres a great new book out by Dr. Ruha Benjamin Race After Technology.
People know its an issue and so Im excited about that. Has there been enough done? Absolutely not. Because people are just now waking up to the fact that theres a problem. Awareness is good, and then that awareness needs to lead to action. That is the phase were in. Companies have a role to play, governments have a role to play and individuals have a role to play.
When you see the bans in San Francisco [of facial recognition technology by the citys agencies] what you saw was a very powerful counter-narrative. What we were hearing was that this technology is inevitable, theres nothing you can do. When you hear theres nothing you can do, you stop trying. But what was extremely encouraging to me with the San Francisco ban and then you have Somerville that came from the folks who are in Boston people have a voice and people have a choice. This technology is not inherently inevitable. We have to look at it and say: What are the benefits and what are the harms? If the harms are too great, we can put restrictions and we can put limitations. And this is necessary. I do look to those examples and they give me hope.
The rest is here:
Artificial Intelligence Can Be Biased. Here's What You Should Know. - FRONTLINE
Posted in Ai
Comments Off on Artificial Intelligence Can Be Biased. Here’s What You Should Know. – FRONTLINE
The AI hiring industry is under scrutinybut it’ll be hard to fix – MIT Technology Review
Posted: at 8:42 am
The Electronic Privacy Information Center (EPIC) has asked the Federal Trade Commission to investigate HireVue, an AI tool that helps companies figure out which workers to hire.
Whats HireVue? HireVue is one of a growing number of artificial intelligence tools that companies use to assess job applicants. The algorithm analyzes video interviews, using everything from word choice to facial movements to figure out an employability score that is compared against that of other applicants. More than 100 companies have already used it on over a million applicants, according to the Washington Post.
Whats the problem? Its hard to predict which workers will be successful from things like facial expressions. Worse, critics worry that the algorithm is trained on limited data and so will be more likely to mark traditional applicants (white, male) as more employable. As a result, applicants who deviate from the traditionalincluding people dont speak English as a native language or who are disabledare likely to get lower scores, experts say. Plus, it encourages applicants to game the system by interviewing in a way that they know HireVue will like.
Whats next? AI hiring tools are not well regulated, and addressing the problem will be hard for a few reasons.
Most companies wont release their data or explain how the algorithms work, so its very difficult to prove any bias. Thats part of the reason there have been no major lawsuits so far. The EPIC complaint, which suggests that HireVues promise violates the FTCs rules against unfair and deceptive practices, is a start. But its not clear if anything will happen. The FTC has received the complaint but has not said whether it will pursue it.
Other attempts to prevent bias are well-meaning but limited. Earlier this year, Illinois lawmakers passed a law that requires employers to at least tell job seekers that theyll be using these algorithms, and to get their consent. But thats not very useful. Many people are likely to consent simply because they dont want to lose the opportunity.
Finally, just like AI in health or AI in the courtroom, artificial intelligence in hiring will re-create societys biases, which is a complicated problem. Regulators will need to figure out how much responsibility companies should be expected to shoulder in avoiding the mistakes of a prejudiced society.
More here:
The AI hiring industry is under scrutinybut it'll be hard to fix - MIT Technology Review
Posted in Ai
Comments Off on The AI hiring industry is under scrutinybut it’ll be hard to fix – MIT Technology Review
AI in drug development: the FDA needs to set standards – STAT
Posted: at 8:42 am
Artificial intelligence has become a crucial part of our technological infrastructure and the brain underlying many consumer devices. In less than a decade, machine learning algorithms based on deep neural networks evolved from recognizing cats in videos to enabling your smartphone to perform real-time translation between 27 different languages. This progress has sparked the use of AI in drug discovery and development.
Artificial intelligence can improve efficiency and outcomes in drug development across therapeutic areas. For example, companies are developing AI technologies that hold the promise of preventing serious adverse events in clinical trials by identifying high-risk individuals before they enroll. Clinical trials could be made more efficient by using artificial intelligence to incorporate other data sources, such as historical control arms or real-world data. AI technologies could also be used to magnify therapeutic responses by identifying biomarkers that enable precise targeting of patient subpopulations in complex indications.
Innovation in each of these areas would provide substantial benefits to those who volunteer to take part in trials, not to mention downstream benefits to the ultimate users of new medicines.
advertisement
Misapplication of these technologies, however, can have unintended harmful consequences. To see how a good idea can turn bad, just look at whats happened with social media since the rise of algorithms. Misinformation spreads faster than the truth, and our leaders are scrambling to protect our political systems.
Could artificial intelligence and machine learning similarly disrupt our ability to identify safe and effective therapies?
Even well-intentioned researchers can develop machine learning algorithms that exacerbate bias. For example, many datasets used in medicine are derived from mostly white, North American and European populations. If a researcher applies machine learning to one of these datasets and discovers a biomarker to predict response to a therapy, there is no guarantee the biomarker will work well, if at all, in a more diverse population. If such a biomarker was used to define the approved indication for a drug, that drug could end up having very different effects in different racial groups simply because it is filtered through the biased lens of a poorly constructed algorithm.
Concerns about bias and generalizability apply to most data-driven decisions, including those obtained using more traditional statistical methods. But the machine learning algorithms that enable innovations in drug development are more complex than traditional statistical models. They need larger datasets, more sophisticated software, and more powerful computers. All of that makes it more difficult, and more important, to thoroughly evaluate the performance of machine learning algorithms.
Companies operating at the intersection of drug development and technology need standards to ensure that artificial intelligence tools function as intended.
The FDA has already issued several proposals around the regulation of AI products, and it now has an opportunity to build on these efforts. The Center for Devices and Radiological Health has reviewed and cleared a number of devices that use AI. The center has also released a proposed framework, Artificial Intelligence and Machine Learning in Software as a Medical Device. These proposals, though, dont necessarily apply to AI-based tools used as part of the drug development process. As a result, biopharmaceutical and technology companies arent sure how these tools fit into current regulatory frameworks.
Im the founder and CEO of a company that uses artificial intelligence to streamline clinical trials and make them more efficient. You might expect me to counsel the FDA to back off on creating hurdles for companies that want to apply artificial intelligence to drug development. Not so. In a presentation to the FDA on Thursday, Ill argue that the agency should play an important role in ensuring that AI-based drug development tools meet appropriate standards.
The FDA has an opportunity to ease regulatory uncertainty by proposing a framework that guides how sponsors can use AI tools within drug development programs. By engaging with industry to develop a workable regulatory framework, the FDA can balance the opportunity for artificial intelligence to provide significant public health benefits with its mission to protect public health by ensuring that these new technologies are reliable. At the same time, the FDA could create a pathway for formal qualification of AI-based drug-development tools to ensure that these tools are sufficiently vetted.
In addition, it could encourage the exploratory use of AI-based technologies in drug development that would allow sponsors and regulators to better understand their advantages and disadvantages through use of new regulatory pathways, such as the Complex Innovative Trial Designs Pilot Program.
These concrete actions would open the door to innovative approaches to clinical trials that will make drug development more efficient and so help deliver new treatments to patients who need them as quickly as possible.
Charles K. Fisher, Ph.D., is the founder and CEO of San Francisco-based Unlearn.AI, Inc.
See the original post here:
AI in drug development: the FDA needs to set standards - STAT
Posted in Ai
Comments Off on AI in drug development: the FDA needs to set standards – STAT
Report: The Government and Tech Need to Cooperate on AI – WIRED
Posted: at 8:42 am
Americas national security depends on the government getting access to the artificial intelligence breakthroughs made by the technology industry.
So says a report submitted to Congress on Monday by the National Security Commission on Artificial Intelligence. The group, which includes executives from Google, Microsoft, Oracle, and Amazon, says the Pentagon and intelligence agencies need a better relationship with Silicon Valley to stay ahead of China.
AI adoption for national security is imperative, said Eric Schmidt, chair of the commission and formerly CEO of Google, at a news briefing Monday. The private sector and government officials need to build a shared sense of responsibility.
Mondays report says the US leads the world in both military might and AI technology. It predicts that AI can enhance US national security in numerous ways, for example by making cybersecurity systems, aerial surveillance, and submarine warfare less constrained by human labor and reaction times.
But the commission also unspools a litany of reasons that US dominance on the world stage and in AI may not last, noting that China is projected to overtake the US in R&D spending within 10 years, while US federal research spending as a percentage of GDP has returned to pre-Sputnik levels and should be increased significantly.
Robert Work, vice chair of the commission and previously deputy secretary of defense under Obama and Trump, continued the Cold War comparisons in Mondays news briefing. We've never faced a high-tech authoritarian competitor before, he said. The Soviet Union could compete with us in niche capabilities like nuclear weapons and space, but in the broad sense they were a technological inferior.
Created by Congress in August 2018 to offer recommendations on how the US should use AI in national security and defense, the NSCAI has strong tech industry representation. In addition to Schmidt, the 15-member commission includes Safra Katz, CEO of Oracle, Andy Jassy, the head of Amazons cloud business, and top AI executives from Microsoft and Google. Other members are from NASA, academia, the US Army, and the CIA's tech investment fund.
Mondays report says staying ahead of China depends in part on the US government getting more access to AI advances taking place inside tech companieslike those several of the commissioners work for. The document describes the Pentagon as struggling to access the best AI technology on the commercial market.
The Department of Defense has in recent years set up a series of programs aimed at forging closer relationships with Silicon Valley companies large and small. Mondays report suggests that pressure to find new ways to deepen relations will continue to grow, says William Carter, deputy director of the technology policy program at the Center for Strategic and International Studies. The report clearly articulates that DOD continuing to do business the way it always has and expecting the world to go along with it is not going to work, he says.
The commission wont send its final recommendations to Congress until late next year, but Monday's interim report says the US government should invest more in AI research and training, curtail inappropriate Chinese access to US exports and university research, and mull the ethical implications of AI-enhanced national security apparatus.
So far, attempts to draw tech companies into more national security contracts have had mixed results.
Employee protests forced Google to promise not to renew its piece of a Pentagon program, Project Maven, created to show how tech companies could help military AI projects. Microsoft has also faced internal protests over contracts with the Army and Immigration and Customs Enforcement.
Yet Microsoft CEO Satya Nadella and his Amazon counterpart Jeff Bezos have issued full-throated statements in support of the idea of taking national security contracts. Last month, Microsoft won a $10 billion Pentagon cloud-computing contract motivated in part by a desire to improve the departments AI capabilities. Deals like that could become more common if the commission proves to be influential.
Read the rest here:
Report: The Government and Tech Need to Cooperate on AI - WIRED
Posted in Ai
Comments Off on Report: The Government and Tech Need to Cooperate on AI – WIRED
These American workers are the most afraid of A.I. taking their jobs – CNBC
Posted: at 8:42 am
The Terminator movie franchise is back, and the idea that robots and artificial intelligence are coming for us specifically, our jobs is a big part of the present. But the majority of the working population remains unafraid of a T-800 stealing their employment.
Only a little over one-quarter (27%) of all workers say they are worried that the job they have now will be eliminated within the next five years as a result of new technology, robots or artificial intelligence, according to the quarterly CNBC/SurveyMonkey Workplace Happiness survey.
Nevertheless, the survey results show it may be only a matter of time: Fears about automation and jobs run higher among the youngest workers.
The survey found that 37% of workers between the ages of 18 and 24 are worried about new technology eliminating their jobs. That's nearly 10% higher than any other demographic.
Dan Schawbel, research director of Future Workplace and author of "Back to Human," said one reason for the age-based fear gap is because technology, like AI, is becoming normalized.
"They are starting to see the value of [AI] and how it's impacting their personal and professional lives," Schawbel said. "We're using AI without even thinking about it. It's a part of our lives. If you are talking to Siri or Alexa, that's AI."
Laura Wronski, senior research scientist at SurveyMonkey, said, "As digital natives, [18- to 24-year-old workers] understand the potential of technology to have a positive impact. But with 30 or 40 years left in the workforce, they likely envision vast potential changes in the nature of work over the course of their lifetime."
The survey also revealed a link between income and fear, with 34% of workers making $50,000 or under afraid of losing their jobs due to technology; that goes down to 16% among workers making between $100,000 and $150,000, and 13% for workers making $150,000 or more.
In some industries where technology already has played a highly disruptive role, worker fears of automation also run higher than the average: Workers in automotives, business support and logistics, advertising and marketing, and retail are proportionately more worried about new technology replacing their jobs than those in other industries.
Forty-two percent of workers in the business support and logistics industry have above-average concerns about new technology eliminating their jobs. Schawbel said that fear stems from the fact that the industry is already seeing it happen. Self-driving trucks already are threatening the jobs of truck drivers, and it is causing massive panic in the profession, he said.
"There is a fear, with some research to back it up, that it's going to be hard to retrain and retool truck drivers to take on other jobs," Schawbel said. "You know with a truck driver you can just eliminate the truck driver, whereas with professionals doing finance or accounting, certain tasks that they do can be automated, but they have a little more flexibility to do other tasks that could be more valuable."
Elmer Guardado, a 22-year-old account coordinator at Buie & Co. Public Relations, fits two demographics that are more likely to worry about new technology replacing them: he is young, and he is in the advertising and marketing industry. But he remains convinced that human skills will set him apart from the automated competition.
"It's not something I'm actively worried about," Guardado said. "Because I know there are so many parts of my job that require a level of nuance that technology won't be able to replace anytime soon."
Guardado says that his communication skills are a valuable asset that he brings to the workplace that a computer can't compete with quite yet. But he also understands why his peers may be more afraid than other age groups.
"I think older generations maybe process this potential fear in a more abstract way," Guardado said. "Whereas 18- 24-year-olds see it firsthand, right? We actively dealt with it growing up and saw technology consistently skyrocket throughout our entire lifetime."
The survey found a fairly optimistic view on the future of AI, with nearly half of workers (48%) saying the quest to advance the field of artificial intelligence is "important." Only 23% called it "dangerous."
They remain more worried about their own kind: 60% of workers said that human intelligence is a greater threat to humanity than artificial intelligence. Sixty-five percent of survey respondents said computer programs will always reflect the biases of the people who designed them.
Read the rest here:
These American workers are the most afraid of A.I. taking their jobs - CNBC
Posted in Ai
Comments Off on These American workers are the most afraid of A.I. taking their jobs – CNBC
How Nvidia (NVDA) and AI Can Help Farmers Fight Weeds And Invasive Plants – Nasdaq
Posted: at 8:42 am
Agricultural fields are no less than a battlefield. Irrespective of terrain, geography and type, crops have to compete against scores of different weeds, species of hungry insects, nematodes and a broad array of diseases. Weeds, or invasive plants, aggressively compete for soil nutrients, light and water, posing a serious threat to agricultural production and biodiversity. Weeds directly and indirectly result in tremendous losses to the farm sector, which convert to billions each year worldwide.
To combat these challenges, the farm sector is looking at Artificial Intelligence (AI) based solutions. Heres a look at two such initiatives powered by NVIDIA Corporation (NVDA).
Invasive Plants
The damage wrought by plant pests and diseases can reach up to 40% of global crop yields each year as perestimatesby the Food and Agricultural Organization of the United States. Among the pests, weeds are considered an important bioticconstraintto food production. The competition for survival between weeds and crops reduces agricultural output both qualitatively and quantitatively.
It isestimatedthat the annual cost of weeds to Australian agriculture is $4 billion through yield losses and product contamination. The Weed Science Society of America (WSSA)reportsthat on an annual basis, potential loss in value for corn is $27 billion and for soybean it is $16 billion based on data from 2007 to 2013. In India, an assessment by the Directorate of Weed Researchshowsthat India loses crop worth $11 billion every year to weeds.
One of the most common ways to control weed is to spray the entire field with herbicides. This method involves significant cost, wastage, health problems and environmental pollution. While the real cost of weeds to the environment is difficult to calculate, it is expected that the cost would be similar to, if not greater than, that estimated for agricultural industries, according to anoteby the department of environment of Australia.
Enter AI
Today, advanced technologies are being increasingly applied to a number of industries and sectors, agriculture being one of them. One such technique is that of precision farming, which allows for farmers to reduce their use of chemical inputs, machinery and water for irrigation by using information about the soil, temperature, humidity, seeds, farm equipment, livestock, fertilizers, terrain, crops sown, and water usage, among other things.A growing number of companies and start-ups are creating AI-based agricultural solutions.
Cameras, sensors and AI on the fields allow farmers to manage their fields better and use pesticides more precisely. Blue River Technology's See & Spray uses computer vision and AI to detect, identify, and make management decisions about every single plant in the field. In 2017, Blue River Technology wasacquiredby Deere & Company (DE). Today the See & Spray, which is a 40-feet wide machine covering 12 rows of crops, is pulled by Deere tractors and is powered by Nvidia.
The machine uses around 30 mounted cameras to capture photos of plants every 50 milliseconds and these are processed through its on-board25 Jetson AGX Xaviersupercomputing modules. As the tractor continues to move, the Jetson Xavier modules running Blue Rivers image recognition algorithms make super quick decisions based on the image inputs received on whether it is a weed or crop plant. See & Spray machine has been able to achieve good success by using less than 1/10th the herbicide of typical weed control.
Further, a researchpaperpublished in 2018 by M Dian Bah, Adel Hafiane and Raphael Canals proposed a novel fully automatic learning method using convolutional neuronal networks (CNNs) with an unsupervised training dataset collection for weed detection from UAV images. Drone images of beet, bean and spinach crops were used for the study. The researchers used a cluster ofNVIDIA Quadro GPUsto train the neural networks. The researcherssaythat, using NVIDIA Quadro GPUs shrunk training time from one week on a high-end CPU down to a few hours. The study archived a precision of 93%, 81% and 69% for beet, spinach and bean, respectively.
While these initiatives are working on precision-based use of any chemical product in the fields, neural networks can be trained to detect infected areas in plants using images. One such study is being done on the detection of symptoms of disease in grape leaves. Early detection can play an important factor in preventing a serious disease and stop an epidemic spread in vineyards.
The use of technology can help in solving multiple problems faced by farmers, saving valuable resource and reduce the damage done to the environment. Thestatementby FOA chief, the future of agriculture is not input-intensive but technology-intensive aptly sums up the role that technology and technology providers will play in the farm sector.
The views and opinions expressed herein are the views and opinions of the author and do not necessarily reflect those of Nasdaq, Inc.
Here is the original post:
How Nvidia (NVDA) and AI Can Help Farmers Fight Weeds And Invasive Plants - Nasdaq
Posted in Ai
Comments Off on How Nvidia (NVDA) and AI Can Help Farmers Fight Weeds And Invasive Plants – Nasdaq