This year marks the 78th anniversary of the end of the deadliest war in history and the beginning of the longest period in modern times without great-power war. Because World War I had been followed just two decades later by World War II, the specter of World War III, fought with weapons that had become so destructive they could theoretically threaten all of humankind, hung over the decades of the Cold War that followed. When the United States atomic destruction of Hiroshima and Nagasaki compelled Japans immediate unconditional surrender, no one thought it conceivable that the world would see a de facto moratorium on the use of nuclear weapons for the next seven decades. It seemed even more improbable that almost eight decades later, there would be just nine nuclear weapons states. The leadership demonstrated by the United States over these decades in avoiding nuclear war, slowing nuclear proliferation, and shaping an international order that provided decades of great-power peace will go down in history as one of Americas most significant achievements.
Today, as the world confronts the unique challenges posed by another unprecedented and in some ways even more terrifying technologyartificial intelligenceit is not surprising that many have been looking to history for instruction. Will machines with superhuman capabilities threaten humanitys status as master of the universe? Will AI undermine nations monopoly on the means of mass violence? Will AI enable individuals or small groups to produce viruses capable of killing on a scale that was previously the preserve of great powers? Could AI erode the nuclear deterrents that have been a pillar of todays world order?
At this point, no one can answer these questions with confidence. But as we have explored these issues for the last two years with a group of technology leaders at the forefront of the AI revolution, we have concluded that the prospects that the unconstrained advance of AI will create catastrophic consequences for the United States and the world are so compelling that leaders in governments must act now. Even though neither they nor anyone else can know what the future holds, enough is understood to begin making hard choices and taking actions todayrecognizing that these will be subject to repeated revision as more is discovered.
As leaders make these choices, lessons learned in the nuclear era can inform their decisions. Even adversaries racing to develop and deploy an unprecedented technology that could kill hundreds of millions of people nonetheless discovered islands of shared interests. As duopolists, both the United States and the Soviet Union had an interest in preventing the rapid spread of this technology to other states that could threaten them. Both Washington and Moscow recognized that if nuclear technology fell into the hands of rogue actors or terrorists within their own borders, it could be used to threaten them, and so each developed robust security systems for their own arsenals. But since each could also be threatened if rogue actors in their adversarys society acquired nuclear weapons, both found it in their interests to discuss this risk with each other and describe the practices and technologies they developed to ensure this did not happen. Once the arsenals of their nuclear weapons reached a level at which neither could attack the other without triggering a response that would destroy itself, they discovered the paradoxical stability of mutual assured destruction (MAD). As this ugly reality was internalized, each power learned to limit itself and found ways to persuade its adversary to constrain its initiatives in order to avoid confrontations that could lead to a war. Indeed, leaders of both the U.S. and the Soviet government came to realize that avoiding a nuclear war of which their nation would be the first victim was a cardinal responsibility.
The challenges presented by AI today are not simply a second chapter of the nuclear age. History is not a cookbook with recipes that can be followed to produce a souffl. The differences between AI and nuclear weapons are at least as significant as the similarities. Properly understood and adapted, however, lessons learned in shaping an international order that has produced nearly eight decades without great-power war offer the best guidance available for leaders confronting AI today.
At this moment, there are just two AI superpowers: the United States and China are the only countries with the talent, research institutes, and mass computing capacity required to train the most sophisticated AI models. This offers them a narrow window of opportunity to create guidelines to prevent the most dangerous advances and applications of AI. U.S. President Joe Biden and Chinese President Xi Jinping should seize this opportunity by holding a summitperhaps immediately after the Asia-Pacific Economic Cooperations meeting in San Francisco in Novemberwhere they could hold extended, direct, face-to-face discussions on what they should see as one of the most consequential issues confronting them today.
After atomic bombs devastated Japanese cities in 1945, the scientists who had opened Pandoras atomic box saw what they had created and recoiled in horror. Robert Oppenheimer, the principal scientist of the Manhattan Project, recalled a line from the Bhagavad Gita: Now I am become Death, the destroyer of worlds. Oppenheimer became such an ardent advocate of radical measures to control the bomb that he was stripped of his security clearance. The Russell-Einstein Manifestosigned in 1955 by 11 leading scientists including not just Bertrand Russell and Albert Einstein but also Linus Pauling and Max Bornwarned of the frightening powers of nuclear weapons and implored world leaders never to use them.
Although U.S. President Harry Truman never expressed second thoughts about his decision, neither he nor members of his national security team had a viable view of how this awesome technology could be integrated into the postwar international order. Should the United States attempt to maintain its monopoly position as the sole atomic power? Was that even feasible? In pursuit of the objective, could the United States share its technology with the Soviet Union? Did survival in a world with this weapon require leaders to invent some authority superior to national governments? Henry Stimson, Trumans secretary of war (who had just helped achieve victory over both Germany and Japan), proposed that the United States share its monopoly of the atomic bomb with Soviet leader Joseph Stalin and British Prime Minister Winston Churchill to create a great-power condominium that would prevent the spread of nuclear weapons. Truman created a committee, chaired by U.S. Undersecretary of State Dean Acheson, to develop a strategy for pursuing Stimsons proposal.
Acheson essentially agreed with Stimson: the only way to prevent a nuclear arms race ending in catastrophic war would be to create an international authority that would be the sole possessor of atomic weapons. This would require the United States to share its nuclear secrets with the Soviet Union and other members of the UN Security Council, transfer its nuclear weapons to a new UN atomic development authority, and forbid all nations from developing weapons or constructing their own capability to produce weapons-grade nuclear material. In 1946, Truman sent the financier and presidential adviser Bernard Baruch to the UN to negotiate an agreement to implement Achesons plan. But the proposal was categorically rejected by Andrei Gromyko, the Soviet representative to the UN.
Three years later, when the Soviet Union succeeded in its crash effort to build its own bomb, the United States and the Soviet Union entered what people were starting to call the Cold War: a competition by all means short of bombs and bullets. A central feature of this competition was the drive for nuclear superiority. At their heights, the two superpowers nuclear arsenals included more than 60,000 weapons, some of them warheads with more explosive power than all the weapons that had been used in all the wars in recorded history. Experts debated whether an all-out nuclear war would mean the end of every living soul on earth.
Over the decades, Washington and Moscow have spent trillions of dollars on their nuclear arsenals. The current annual budget for the U.S. nuclear enterprise exceeds $50 billion. In the early decades of this race, both the United States and the Soviet Union made previously unimaginable leaps forward in the hope of obtaining a decisive advantage. Increases in weapons explosive power required the creation of new metrics: from kilotons (equivalent to the energy released by 1,000 tons of TNT) for the original fission weapons to megatons (equivalent to that released by one million tons) for hydrogen fusion bombs. The two sides invented intercontinental missiles capable of delivering warheads to targets on the other side of the planet in 30 minutes, satellites circling the globe at a height of hundreds of miles with cameras that could identify the coordinates of targets within inches, and defenses that could in essence hit a bullet with a bullet. Some observers seriously imagined defenses that would render nuclear weapons, as President Ronald Reagan put it, impotent and obsolete.
In attempting to shape these developments, strategists developed a conceptual arsenal that distinguished between first and second strikes. They clarified the essential requirements for a reliable retaliatory response. And they developed the nuclear triadsubmarines, bombers, and land-based missilesto ensure that if an adversary were to discover one vulnerability, other components of the arsenal would remain available for a devastating response. Perception of risks of accidental or unauthorized launches of weapons spurred the invention of permissive action linkselectronic locks embedded in nuclear weapons that prevented them from being activated without the right nuclear launch codes. Redundancies were designed to protect against technological breakthroughs that might jeopardize command-and-control systems, which motivated the invention of a computer network that evolved into the Internet. As the strategist Herman Kahn famously put it, they were thinking about the unthinkable.
At the core of nuclear strategy was the concept of deterrence: preventing an adversary from attacking by threatening costs out of proportion to any conceivable benefit. Successful deterrence, it came to be understood, required not just capability but also credibility. The potential victims needed not only the means to respond decisively but also the will. Strategists refined this basic idea further with concepts such as extended deterrence, which sought to employ a political mechanisma pledge of protection via allianceto persuade key states not to build their own arsenals.
By 1962, when U.S. President John F. Kennedy confronted Soviet leader Nikita Khrushchev over nuclear-tipped missiles that the Soviets had placed in Cuba, the U.S. intelligence community estimated that even if Kennedy launched a successful first strike, the Soviet retaliatory response with their existing capabilities might kill 62 million Americans. By 1969, when Richard Nixon became president, the United States needed to rethink its approach. One of us, Kissinger, later described the challenge: Our defense strategies formed in the period of our superiority had to be reexamined in the harsh light of the new realities. . . . No bellicose rhetoric could obscure the fact that existing nuclear stockpiles were enough to destroy mankind. . . . There could be no higher duty than to prevent the catastrophe of nuclear war.
To make this condition vivid, strategists had created the ironic acronym MAD, the essence of which was summarized by Reagans oft-repeated one-liner: A nuclear war cannot be wonand must therefore never be fought. Operationally, MAD meant mutual assured vulnerability. While both the United States and the Soviet Union sought to escape this condition, they eventually recognized that they were unable to do so and had to fundamentally reconceptualize their relationship. In 1955, Churchill had noted the supreme irony in which safety will be the sturdy child of terror, and survival the twin brother of annihilation. Without denying differences in values or compromising vital national interests, deadly rivals had to develop strategies to defeat their adversary by every means possible except all-out war.
One pillar of these strategies was a series of both tacit and explicit constraints now known as arms control. Even before MAD, when each superpower was doing everything it could to achieve superiority, they discovered areas of shared interests. To reduce the risk of mistakes, the United States and the Soviet Union agreed in informal discussions not to interfere with the others surveillance of their territory. To protect their citizens from radioactive fallout, they banned atmospheric testing. To avoid crisis instabilitywhen one side feels the need to strike first in the belief that the other side is about tothey agreed in the 1972 Anti-Ballistic Missile Treaty to limit missile defenses. In the Intermediate-Range Nuclear Forces Treaty, signed in 1987, Reagan and Soviet leader Mikhail Gorbachev agreed to eliminate intermediate-range nuclear forces. The Strategic Arms Limitation Talks, which resulted in treaties signed in 1972 and 1979, limited increases in missile launchers, and later, the Strategic Arms Reduction Treaty (START), signed in 1991, and the New START, signed in 2010, reduced their numbers. Perhaps most consequentially, the United States and the Soviet Union concluded that the spread of nuclear weapons to other states posed a threat to both of them and ultimately risked nuclear anarchy. They brought about what is now known as the nonproliferation regime, the centerpiece of which is the 1968 Nuclear Nonproliferation Treaty, through which 186 countries today have pledged to refrain from developing their own nuclear arsenals.
In current proposals about ways to contain AI, one can hear many echoes of this past. The billionaire Elon Musks demand for a six-month pause on AI development, the AI researcher Eliezer Yudkowskys proposal to abolish AI, and the psychologist Gary Marcuss demand that AI be controlled by a global governmental body essentially repeat proposals from the nuclear era that failed. The reason is that each would require leading states to subordinate their own sovereignty. Never in history has one great power fearing that a competitor might apply a new technology to threaten its survival and security forgone developing that technology for itself. Even close U.S. allies such as the United Kingdom and France opted to develop their own national nuclear capabilities in addition to relying on the U.S. nuclear umbrella.
To adapt lessons from nuclear history to address the current challenge, it is essential to recognize the salient differences between AI and nuclear weapons. First, whereas governments led the development of nuclear technology, private entrepreneurs, technologists, and companies are driving advances in AI. Scientists working for Microsoft, Google, Amazon, Meta, OpenAI, and a handful of smaller startups are far ahead of any analogous effort in government. Furthermore, these companies are now locked in a gladiatorial struggle among themselves that is unquestionably driving innovation, but at a cost. As these private actors make tradeoffs between risks and rewards, national interests are certain to be underweighted.
Second, AI is digital. Nuclear weapons were difficult to produce, requiring a complex infrastructure to accomplish everything from enriching uranium to designing nuclear weapons. The products were physical objects and thus countable. Where it was feasible to verify what the adversary was doing, constraints emerged. AI represents a distinctly different challenge. Its major evolutions occur in the minds of human beings. Its applicability evolves in laboratories, and its deployment is difficult to observe. Nuclear weapons are tangible; the essence of artificial intelligence is conceptual.
A screen showing Chinese and U.S. flags, Beijing, July 2023
Third, AI is advancing and spreading at a speed that makes lengthy negotiations impossible. Arms control developed over decades. Restraints for AI need to occur before AI is built into the security structure of each societythat is, before machines begin to set their own objectives, which some experts now say is likely to occur in the next five years. The timing demands first a national, then an international, discussion and analysis, as well as a new dynamic in the relationship between government and the private sector.
Fortunately, the major companies that have developed generative AI and made the United States the leading AI superpower recognize that they have responsibilities not just to their shareholders but also to the country and humanity at large. Many have already developed their own guidelines for assessing risk before deployment, reducing bias in training data, and restricting dangerous uses of their models. Others are exploring ways to circumscribe training and impose know your customer requirements for cloud computing providers. A significant step in the right direction was the initiative the Biden administration announced in July that brought leaders of seven major AI companies to the White House for a joint pledge to establish guidelines to ensure safety, security, and trust.
As one of us, Kissinger, has pointed out in The Age of AI, it is an urgent imperative to create a systematic study of the long-range implications of AIs evolving, often spectacular inventions and applications. Even while the United States is more divided than it has been since the Civil War, the magnitude of the risks posed by the unconstrained advance of AI demands that leaders in both government and business act now. Each of the companies with the mass computing capability to train new AI models and each company or research group developing new models should create a group to analyze the human and geopolitical implications of its commercial AI operations.
The challenge is bipartisan and requires a unified response. The president and Congress should in that spirit establish a national commission consisting of distinguished nonpartisan former leaders in the private sector, Congress, the military, and the intelligence community. The commission should propose more specific mandatory safeguards. These should include requirements to assess continuously the mass computing capabilities needed to train AI models such as GPT-4 and that before companies release a new model, they stress test it for extreme risks. Although the task of developing rules will be demanding, the commission would have a model in the National Security Commission on Artificial Intelligence. Its recommendations, released in 2021, provided impetus and direction for the initiatives that the U.S. military and U.S. intelligence agencies are undertaking in the AI rivalry with China.
Even at this early stage, while the United States is still creating its own framework for governing AI at home, it is not too early to begin serious conversations with the worlds only other AI superpower. Chinas national champions in the tech sectorBaidu (the countrys top search engine), ByteDance (the creator of TikTok), Tencent (the maker of WeChat), and Alibaba (the leader in e-commerce)are building proprietary Chinese-language analogues of ChatGPT, although the Chinese political system has posed particular difficulties for AI. While China still lags in the technology to make advanced semiconductors, it possesses the essentials to power ahead in the immediate future.
Biden and Xi should thus meet in the near future for a private conversation about AI arms control. Novembers Asia-PacificEconomic Cooperation meeting in San Francisco offers that opportunity. Each leader should discuss how he personally assesses the risks posed by AI, what his country is doing to prevent applications that pose catastrophic risks, and how his country is ensuring that domestic companies are not exporting risks. To inform the next round of their discussions, they should create an advisory group consisting of U.S. and Chinese AI scientists and others who have reflected on the implications of these developments. This approach would be modeled on existing Track II diplomacy in other fields, where groups are composed of individuals chosen for their judgment and fairness although not formally endorsed by their government. From our discussions with key scientists in both governments, we are confident that this can be a highly productive discussion.
U.S. and Chinese discussions and actions on this agenda will form only part of the emerging global conversation on AI, including the AI Safety Summit, which the United Kingdom will host in November, and the ongoing dialogue at the UN. Since every country will be seeking to employ AI to enhance the lives of its citizens while ensuring the safety of its own society, in the longer run, a global AI order will be required. Work on it should begin with national efforts to prevent the most dangerous and potentially catastrophic consequences of AI. These initiatives should be complemented by dialogue between scientists of various countries engaged in developing large AI models and members of the national commissions such as the one proposed here. Formal governmental negotiations, initially among countries with advanced AI programs, should seek to establish an international framework, along with an international agency comparable to the International Atomic Energy Agency.
If Biden, Xi, and other world leaders act now to face the challenges posed by AI as squarely as their predecessors did in addressing nuclear threats in earlier decades, will they be as successful? Looking at the larger canvas of history and growing polarization today, it is difficult to be optimistic. Nonetheless, the incandescent fact that we have now marked 78 years of peace among the nuclear powers should serve to inspire everyone to master the revolutionary, inescapable challenges of our AI future.
Loading... Please enable JavaScript for this site to function properly.
Originally posted here:
Henry Kissinger: The Path to AI Arms Control - Foreign Affairs Magazine
- AI File Extension - Open . AI Files - FileInfo [Last Updated On: June 14th, 2016] [Originally Added On: June 14th, 2016]
- Ai | Define Ai at Dictionary.com [Last Updated On: June 16th, 2016] [Originally Added On: June 16th, 2016]
- ai - Wiktionary [Last Updated On: June 22nd, 2016] [Originally Added On: June 22nd, 2016]
- Adobe Illustrator Artwork - Wikipedia, the free encyclopedia [Last Updated On: June 25th, 2016] [Originally Added On: June 25th, 2016]
- AI File - What is it and how do I open it? [Last Updated On: June 29th, 2016] [Originally Added On: June 29th, 2016]
- Ai - Definition and Meaning, Bible Dictionary [Last Updated On: July 25th, 2016] [Originally Added On: July 25th, 2016]
- ai - Dizionario italiano-inglese WordReference [Last Updated On: July 25th, 2016] [Originally Added On: July 25th, 2016]
- Bible Map: Ai [Last Updated On: August 30th, 2016] [Originally Added On: August 30th, 2016]
- Ai dictionary definition | ai defined - YourDictionary [Last Updated On: August 30th, 2016] [Originally Added On: August 30th, 2016]
- Ai (poet) - Wikipedia, the free encyclopedia [Last Updated On: August 30th, 2016] [Originally Added On: August 30th, 2016]
- AI file extension - Open, view and convert .ai files [Last Updated On: August 30th, 2016] [Originally Added On: August 30th, 2016]
- History of artificial intelligence - Wikipedia, the free ... [Last Updated On: August 30th, 2016] [Originally Added On: August 30th, 2016]
- Artificial intelligence (video games) - Wikipedia, the free ... [Last Updated On: August 30th, 2016] [Originally Added On: August 30th, 2016]
- North Carolina Chapter of the Appraisal Institute [Last Updated On: September 8th, 2016] [Originally Added On: September 8th, 2016]
- Ai Weiwei - Wikipedia, the free encyclopedia [Last Updated On: September 11th, 2016] [Originally Added On: September 11th, 2016]
- Adobe Illustrator Artwork - Wikipedia [Last Updated On: November 17th, 2016] [Originally Added On: November 17th, 2016]
- 5 everyday products and services ripe for AI domination - VentureBeat [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- Realdoll builds artificially intelligent sex robots with programmable personalities - Fox News [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- ZeroStack Launches AI Suite for Self-Driving Clouds - Yahoo Finance [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- AI and the Ghost in the Machine - Hackaday [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- Why Google, Ideo, And IBM Are Betting On AI To Make Us Better Storytellers - Fast Company [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- Roses are red, violets are blue. Thanks to this AI, someone'll fuck you. - The Next Web [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- Wearable AI Detects Tone Of Conversation To Make It Navigable (And Nicer) For All - Forbes [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- Who Leads On AI: The CIO Or The CDO? - Forbes [Last Updated On: February 6th, 2017] [Originally Added On: February 6th, 2017]
- AI For Matching Images With Spoken Word Gets A Boost From MIT - Fast Company [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- Teach undergrads ethics to ensure future AI is safe compsci boffins - The Register [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- AI is here to save your career, not destroy it - VentureBeat [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- A Heroic AI Will Let You Spy on Your Lawmakers' Every Word - WIRED [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- With a $16M Series A, Chorus.ai listens to your sales calls to help your team close deals - TechCrunch [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- Microsoft AI's next leap forward: Helping you play video games - CNET [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- Samsung Galaxy S8's Bixby AI could beat Google Assistant on this front - CNET [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- 3 common jobs AI will augment or displace - VentureBeat [Last Updated On: February 7th, 2017] [Originally Added On: February 7th, 2017]
- Stephen Hawking and Elon Musk endorse new AI code - Irish Times [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- SumUp co-founders are back with bookkeeping AI startup Zeitgold - TechCrunch [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- Five Trends Business-Oriented AI Will Inspire - Forbes [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- AI Systems Are Learning to Communicate With Humans - Futurism [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- Pinterest uses AI and your camera to recommend pins - Engadget [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- Chinese Firms Racing to the Front of the AI Revolution - TOP500 News [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- Real life CSI: Google's new AI system unscrambles pixelated faces - The Guardian [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- AI could transform the way governments deliver public services - The Guardian [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- Amazon Is Humiliating Google & Apple In The AI Wars - Forbes [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- What's Still Missing From The AI Revolution - Co.Design (blog) [Last Updated On: February 9th, 2017] [Originally Added On: February 9th, 2017]
- Legaltech 2017: Announcements, AI, And The Future Of Law - Above the Law [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- Can AI make Facebook more inclusive? - Christian Science Monitor [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- How a poker-playing AI could help prevent your next bout of the flu - ExtremeTech [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- Dynatrace Drives Digital Innovation With AI Virtual Assistant - Forbes [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- AI and the end of truth - VentureBeat [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- Taser bought two computer vision AI companies - Engadget [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- Google's DeepMind pits AI against AI to see if they fight or cooperate - The Verge [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- The Coming AI Wars - Huffington Post [Last Updated On: February 10th, 2017] [Originally Added On: February 10th, 2017]
- Is President Trump a model for AI? - CIO [Last Updated On: February 11th, 2017] [Originally Added On: February 11th, 2017]
- Who will have the AI edge? - Bulletin of the Atomic Scientists [Last Updated On: February 11th, 2017] [Originally Added On: February 11th, 2017]
- How an AI took down four world-class poker pros - Engadget [Last Updated On: February 11th, 2017] [Originally Added On: February 11th, 2017]
- We Need a Plan for When AI Becomes Smarter Than Us - Futurism [Last Updated On: February 11th, 2017] [Originally Added On: February 11th, 2017]
- See how old Amazon's AI thinks you are - The Verge [Last Updated On: February 11th, 2017] [Originally Added On: February 11th, 2017]
- Ford to invest $1 billion in autonomous vehicle tech firm Argo AI - Reuters [Last Updated On: February 11th, 2017] [Originally Added On: February 11th, 2017]
- Zero One: Are You Ready for AI? - MSPmentor [Last Updated On: February 11th, 2017] [Originally Added On: February 11th, 2017]
- Ford bets $1B on Argo AI: Why Silicon Valley and Detroit are teaming up - Christian Science Monitor [Last Updated On: February 12th, 2017] [Originally Added On: February 12th, 2017]
- Google Test Of AI's Killer Instinct Shows We Should Be Very Careful - Gizmodo [Last Updated On: February 12th, 2017] [Originally Added On: February 12th, 2017]
- Google's New AI Has Learned to Become "Highly Aggressive" in Stressful Situations - ScienceAlert [Last Updated On: February 13th, 2017] [Originally Added On: February 13th, 2017]
- An artificially intelligent pathologist bags India's biggest funding in healthcare AI - Tech in Asia [Last Updated On: February 13th, 2017] [Originally Added On: February 13th, 2017]
- Ford pledges $1bn for AI start-up - BBC News [Last Updated On: February 13th, 2017] [Originally Added On: February 13th, 2017]
- Dyson opens new Singapore tech center with focus on R&D in AI and software - TechCrunch [Last Updated On: February 13th, 2017] [Originally Added On: February 13th, 2017]
- How to Keep Your AI From Turning Into a Racist Monster - WIRED [Last Updated On: February 13th, 2017] [Originally Added On: February 13th, 2017]
- How Chinese Internet Giant Baidu Uses AI And Machine Learning - Forbes [Last Updated On: February 13th, 2017] [Originally Added On: February 13th, 2017]
- Humans engage AI in translation competition - The Stack [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- Watch Drive.ai's self-driving car handle California city streets on a ... - TechCrunch [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- Cryptographers Dismiss AI, Quantum Computing Threats - Threatpost [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- Is AI making credit scores better, or more confusing? - American Banker [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- AI and Robotics Trends: Experts Predict - Datamation [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- IoT And AI: Improving Customer Satisfaction - Forbes [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- AI's Factions Get Feisty. But Really, They're All on the Same Team - WIRED [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- Elon Musk: Humans must become cyborgs to avoid AI domination - The Independent [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- Facebook Push Into Video Allows Time To Catch Up On AI Applications - Investor's Business Daily [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- Defining AI, Machine Learning, and Deep Learning - insideHPC [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- AI Predicts Autism From Infant Brain Scans - IEEE Spectrum [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- The Rise of AI Makes Emotional Intelligence More Important - Harvard Business Review [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- Google's AI Learns Betrayal and "Aggressive" Actions Pay Off - Big Think [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- AI faces hype, skepticism at RSA cybersecurity show - PCWorld [Last Updated On: February 15th, 2017] [Originally Added On: February 15th, 2017]
- New AI Can Write and Rewrite Its Own Code to Increase Its Intelligence - Futurism [Last Updated On: February 17th, 2017] [Originally Added On: February 17th, 2017]