NATO ups the ante on disruptive tech, artificial intelligence – DefenseNews.com

STUTTGART, Germany NATO has officially kicked off two new efforts meant to help the alliance invest in critical next-generation technologies and avoid capability gaps between its member nations.

For months, officials have set the ground stage to launch a new Defense Innovator Accelerator nicknamed DIANA and establish an innovation fund to support private companies developing dual-use technologies. Both of those measures were formally agreed upon during NATOs meeting of defense ministers last month in Brussels, said Secretary-General Jens Stoltenberg.

Allies signed the agreement to establish the NATO Innovation Fund and launch DIANA on Oct. 22, the final day of the two-day conference, Stoltenberg said in a media briefing that day.

He expects the fund to invest 1 billion (U.S. $1.16 billion) into companies and academic partners working on emerging and disruptive technologies.

New technologies are reshaping our world and our security, Stoltenberg said. NATOs new innovation fund will ensure allies do not miss out on the latest technology and capabilities that will be critical to our security.

We need to ensure that allies are able to operate the different technologies seamlessly, between their forces, and with each other, he added.

Seventeen allied countries agreed to help launch the innovation fund. They include: Belgium, the Czech Republic, Estonia, Germany, Greece, Hungary, Italy, Latvia, Lithuania, Luxembourg, the Netherlands, Poland, Portugal, Romania, Slovakia, Slovenia, and the United Kingdom.

NATO will develop a minimum level of funding that will be required by every participating nation, and that level is being decided by those initial 17 allies, said David van Weel, assistant secretary-general for emerging security challenges.

Soldiers from NATO member France attend the cyber defense exercise DEFNET 2021 on March 18, 2021, in Rennes, western France. Alliance members collectively have pledged to boost their focus on new and disruptive technologies, including the areas of cyber and artificial intelligence. (Damien Meyer/AFP via Getty Images)

He noted that there are a variety of reasons as to why the initial supporters stepped up, while the remaining 13 member nations did not. But he expects that more countries will sign up to participate in the fund before the alliances 2022 summit, he said during an Oct. 27 media roundtable.

The bus hasnt left the station to join the fund, and we expect more to join up, he said.

Recommendations for NATO to launch such a venture capital fund, and a technology accelerator outfit reminiscent of the U.S. Defense Advanced Research Projects Agency (DARPA), were included in a 2020 report by NATOs advisory group on emerging and disruptive technologies.

The alliance agreed to launch the DIANA accelerator at NATOs annual summit, held last June in Brussels. Both the accelerator outfit and the innovation fund will have headquarters based in both North America and Europe, and several nations have already offered to host the facilities.

The plan is for a separate company to run the day-to-day operations of the innovation fund, but that partner has yet to be selected, van Weel said. It is going to be professional venture capitalists that are going to run this fund that could either be an existing company, or we would recruit an experienced general partner to run this, he added.

The offices are expected to be in place next year, and both DIANA and the fund are scheduled to be fully in effect by NATOs next summit, June 29-30 in Madrid, per the alliance.

Meanwhile, the allies also agreed on NATOs first-ever artificial intelligence strategy, which has been in the works since early 2021. It will set standards for responsible use of artificial intelligence, in accordance with international law, outline how we will accelerate the adoption of artificial intelligence in what we do, set out how we will protect this technology, and address the threats posed by the use of artificial intelligence by adversaries, Stoltenberg said.

NATO released a summary of the strategy on Oct. 22, and it includes four sections: Principles of responsible use of artificial intelligence in defense; ensuring the safe and responsible use of allied AI; minimizing interference in allied AI; and standards.

It also lays out the six principles of AI use that member-nations should follow. They include: lawfulness; responsibility and accountability; explainability and traceability; reliability; governability; and bias mitigation.

The nascent DIANA outfit will host specialized AI test centers that will help NATO ensure standards are being kept as member-nations develop new platforms and systems and encourage interoperability, van Weel noted. That way, NATO creates a common ecosystem where all allies have access to the same levels of AI, he said.

NATO will also form a data and artificial intelligence review board with representatives from all member-nations, to ensure the operationalization of the AI strategy, he added. The principles are all great, but they only mean something if were able to actually translate that into how the technology is being developed, and then used.

NATO eventually plans to develop strategies for tackling each of the seven key emerging and disruptive technology (EDT) categories, van Weel told Defense News earlier this year. Having that strategy in place would allow the partnership to begin implementing AI capabilities into military requirements, and ensure interoperability for NATO-based and allied systems, he said at the time.

Member-nations also agreed to a new policy that treats data as a strategic asset, and sets a framework for both NATO headquarter-generated data and national data to be exploited across the alliance in a responsible fashion, van Weel said. The data and AI review board will serve as a quasi Chief Data Officer that ensures the alliances data, wherever it originates from, is stored securely and adheres to the principles agreed to by NATOs members.

This is step one to create a trust basis for allies to make them actually want them to share data, knowing that it is stored in a secure place, [and] that we have principles of responsible use, van Weel said.

It remains to be seen how each country will contribute to the innovation fund or the tech accelerator, but at least one ally already has some ideas.

Estonia has built up experience working with startups, and has invested heavily in cybersecurity technologies since the Baltic nation faced a wave of cyber attacks. That instance led to the creation of the NATO Cooperative Cyber Defence Centre of Excellence in Tallinn.

That center could play a key role in the alliances EDT efforts, particularly related to technologies like AI that will require a basis in cyber, said Tuuli Vors, counsellor to the Estonian delegation to NATO.

With cyber, we build so many different technological areas or sectors, she said in an interview with Defense News in Brussels. Having the cyber defense center in Tallinn can be used for the benefit of this initiative, or for the allies in a general way.

We have this right mindset, we are flexible, she said. I think its one of the key competencies, to bring together the private sector with the government and the civil sector.

We all know that these technological developments and the real breaks, these are in the private sector, she noted. So therefore, we need to bring them on board [in a] more effective way.

At last months ministerial allies also agreed on a specific set of capability targets to achieve jointly, Stoltenberg told reporters in Brussels. That set includes thousands of targets, heavier forces and more high-end capabilities.

Very few of us can have the whole spectrum of capabilities and defense systems, he said. One of the really important tasks of NATO ... is our ability to coordinate and agree to capability targets, so we can support and help each other as allies.

Each of the allies spend varying amounts of money on their defense budgets, but each also has expertise that can be shared, Vors said. The innovation fund and DIANA can help provide more effective collaboration among these nations, she added.

We have expertise in autonomous systems or cyber defense, we can share it to somewhere where its lacking, and we can have from them CBRN [chemical, biological, radioactive and nuclear] defense technology, she said. So its making this network.

Joe Gould in Brussels contributed to this report.

Vivienne Machi is a reporter based in Stuttgart, Germany, contributing to Defense News' European coverage. She previously reported for National Defense Magazine, Defense Daily, Via Satellite, Foreign Policy and the Dayton Daily News. She was named the Defence Media Awards' best young defense journalist in 2020.

See the article here:
NATO ups the ante on disruptive tech, artificial intelligence - DefenseNews.com

Humans in the loop: It takes people to ensure artificial intelligence success – ZDNet

When it comes to artificial intelligence, don't try to go it alone. IT departments, no matter how skilled and ready, can only go so far past proofs of concept. It takes people -- from all corners of the enterprise and working collaboratively -- to deliver AI success,

Industry experts say that AI initiatives need everyone across the enterprise on board. "A copious amount of training data and elastic compute power are not the cornerstones for successful AI implementations," saysSreedhar Bhagavatheeswaran, global head of Mindtree Consulting.

That cornerstone of AI success is people -- not only people with AI skills, but also those from all disciplines, from marketing to supply chain management. In recent years -- and especially over the past year, as the need for automated or unattended processes accelerated, "enterprises learned that they must get stakeholder buy-in, with a true champion for AI within the organization's leadership team," saysDan Simion, VP of AI and analytics at Capgemini Americas.

A concerted AI development effort also needs "strong governance, internal marketing within the company, and proper training to fuel further adoption of the AI initiatives across the business' functional areas," he adds. The key is being able to showcase the valuable insights being generated by these models,

In efforts to make AI pervasive, "enterprises are now conscious of critical factors such as identifying the right journeys and use cases where AI intervention can make a business impact, operationalizing AI by establishing an AI operations and governance mechanisms, and blending the right proportion of data engineering and AI talent," says Bhagavatheeswaran.

The catch, of course, is many of these efforts get undermined by organizational politics or simple inertia. AI seems glamorous and promising, but acceptance and adoption take time. "Companies should plan for the time and effort needed to conduct training sessions, and continuously reinforce the use and benefits of the AI system over the traditional methods," advises Nitin Aggarwal, vice president of data analytics at The Smart Cube. "Sharing and celebrating small and frequent wins is a proven catalyst."

AI also needs to have a friendly face, rather than perceptions of robots, software or otherwise, taking the reins of the company. "Make the end-user interface business-friendly and intuitive," Aggarwal suggests. "The lower the burden on the end-user to understand the insights in terms of 'so what,' the higher the chances of them actually using the system." If possible, he advises having an MLOps team on hand "to ensure the deployed solutions continue to work as expected."

To date, the areas of the business having the most success with AI "are those with direct connections to customer interactions -- such as marketing and sales," says Simion. "These areas are constantly looking to drive revenue, and are more open to innovative new methods and tactics to improve efficiencies, which AI offers." Aggarwal agrees, noting that areas seeing the most initial success with AI include "marketing mix optimization, pricing, and promotions ROI improvement, demand forecasting, CRM and hyper-personalization." Lately, however, AI's power has also been turned on areas such as supply chain risk management, he adds.

AI is more than technology -- it's new ways of thinking about problems and opportunities. Everyone needs to have access to this powerful new tool, Simion urges. "Make sure everyone across the enterprise is using the same technology stack, so each functional area can have access to the same lessons and insights. Consistency of the technology and the value it can bring is what makes the most difference."

AI adoption also hinges on perceptions that it is fair and accurate, making fighting AI bias is another challenge proponents need to address head-on. Start with the data, Aggarwal states. "As AI algorithms learn from data, make a conscious effort for collecting and feeding richer data, that is corrected for bias and is fairly representative of all classes," he advises.

In most cases, "when you deploy AI models into production at scale, you have automatic tools to monitor the results in real-time," says Simion. "When the AI models are outside of their pre-set boundaries and limits, human intervention is necessary. This is done to ensure AI is performing as expected to drive efficiencies for the business, and it also is done to ensure any issues with AI bias or trust are caught and corrected."

It's critical that humans be kept in the loop, says Aggarwal. "Sometimes human decision making alongside the algorithm is helpful to understand different responses and identify any inherent errors or biases. Human judgment can bring in more awareness, context, understanding, and research ability to guide fair decision making. Debiasing should be looked at as an ongoing commitment."

As part of this, companies may benefit by establishing an "AI governance council that reviews not only the business results influenced by their AI initiatives but is also responsible for explaining the results of specific use cases when needed," says Bhagavatheeswaran.

IT leaders and staff need to receive more training and awareness to alleviate AI bias as well. "It also ties into how staff performance is evaluated and how incentives are aligned," says Aggarwal. "If creating the most accurate AI system is the key result area for a data scientist, chances are that you will get a highly accurate system but one, which may not be the most responsible. Similarly, for all staff, an important training should be on where to look for and how to detect biases in AI, and then reward teams who are able to find and recognize flaws."

Read the rest here:
Humans in the loop: It takes people to ensure artificial intelligence success - ZDNet

Yeah, were spooked: AI starting to have big real-world impact, says expert – The Guardian

A scientist who wrote a leading textbook on artificial intelligence has said experts are spooked by their own success in the field, comparing the advance of AI to the development of the atom bomb.

Prof Stuart Russell, the founder of the Center for Human-Compatible Artificial Intelligence at the University of California, Berkeley, said most experts believed that machines more intelligent than humans would be developed this century, and he called for international treaties to regulate the development of the technology.

The AI community has not yet adjusted to the fact that we are now starting to have a really big impact in the real world, he told the Guardian. That simply wasnt the case for most of the history of the field we were just in the lab, developing things, trying to get stuff to work, mostly failing to get stuff to work. So the question of real-world impact was just not germane at all. And we have to grow up very quickly to catch up.

Artificial intelligence underpins many aspects of modern life, from search engines to banking, and advances in image recognition and machine translation are among the key developments in recent years.

Russell who in 1995 co-authored the seminal book Artificial Intelligence: A Modern Approach, and who will be giving this years BBC Reith lectures entitled Living with Artificial Intelligence, which begin on Monday says urgent work is needed to make sure humans remain in control as superintelligent AI is developed.

AI has been designed with one particular methodology and sort of general approach. And were not careful enough to use that kind of system in complicated real-world settings, he said.

For example, asking AI to cure cancer as quickly as possible could be dangerous. It would probably find ways of inducing tumours in the whole human population, so that it could run millions of experiments in parallel, using all of us as guinea pigs, said Russell. And thats because thats the solution to the objective we gave it; we just forgot to specify that you cant use humans as guinea pigs and you cant use up the whole GDP of the world to run your experiments and you cant do this and you cant do that.

Russell said there was still a big gap between the AI of today and that depicted in films such as Ex Machina, but a future with machines that are more intelligent than humans was on the cards.

I think numbers range from 10 years for the most optimistic to a few hundred years, said Russell. But almost all AI researchers would say its going to happen in this century.

One concern is that a machine would not need to be more intelligent than humans in all things to pose a serious risk. Its something thats unfolding now, he said. If you look at social media and the algorithms that choose what people read and watch, they have a huge amount of control over our cognitive input.

The upshot, he said, is that the algorithms manipulate the user, brainwashing them so that their behaviour becomes more predictable when it comes to what they chose to engage with, boosting click-based revenue.

Have AI researchers become spooked by their own success? Yeah, I think we are increasingly spooked, Russell said.

It reminds me a little bit of what happened in physics where the physicists knew that atomic energy existed, they could measure the masses of different atoms, and they could figure out how much energy could be released if you could do the conversion between different types of atoms, he said, noting that the experts always stressed the idea was theoretical. And then it happened and they werent ready for it.

The use of AI in military applications such as small anti-personnel weapons is of particular concern, he said. Those are the ones that are very easily scalable, meaning you could put a million of them in a single truck and you could open the back and off they go and wipe out a whole city, said Russell.

Russell believes the future for AI lies in developing machines that know the true objective is uncertain, as are our preferences, meaning they must check in with humans rather like a butler on any decision. But the idea is complex, not least because different people have different and sometimes conflicting preferences, and those preferences are not fixed.

Russell called for measures including a code of conduct for researchers, legislation and treaties to ensure the safety of AI systems in use, and training of researchers to ensure AI is not susceptible to problems such as racial bias. He said EU legislation that would ban impersonation of humans by machines should be adopted around the world.

Russell said he hoped the Reith lectures would emphasise that there is a choice about what the future holds. Its really important for the public to be involved in those choices, because its the public who will benefit or not, he said.

But there was another message, too. Progress in AI is something that will take a while to happen, but it doesnt make it science fiction, he said.

Read the original:
Yeah, were spooked: AI starting to have big real-world impact, says expert - The Guardian

The Cultural Benefits of Artificial Intelligence in the Enterprise – MIT Sloan

Organization-Level Cultural Benefits

The Culture-Use-Effectiveness dynamic is different at the organizational level than it is at the team level. Figure 5 shows the C-U-E dynamic at the organizational level: Organizational culture can improve AI adoption, which in turn improves organizational effectiveness, which in turn improves organizational culture.

Improving each component can lead to a virtuous cycle of cultural improvement throughout the enterprise.

At PepsiCo, executives view AI as a strategic capability. They also acknowledge that making full use of that capability goes hand in hand with strengthening the companys culture, says Colin Lenaghan, global senior vice president, net revenue management, for the food and beverage multinational. PepsiCo is very much an organization and a culture that learns by doing, he explains. We view AI as a very strategic capability that helps us solve strategic problems. We are making quite an investment in bringing literacy of advanced analytics across the broader community. We are starting to elevate that literacy among senior management. This is clearly something that has to be driven from the top. It needs cultural change. Over time, we intend to strengthen our AI capability and hopefully the culture at the same time. Pervasive AI literacy enables communication through a shared language.

A shared language improves communication about (and the identification of) new opportunities. At Levi Strauss & Co., Paul Pallath, the clothing companys global technology head of data, analytics, and AI, agrees that broad-based adoption of AI demands culture change across the organization. We need to change the overall culture of the organization, and that depends on getting our people to think in terms of AI, he says. If you dont start thinking in that direction, youre not going to ask the right questions that can eventually be solved with AI. Thinking in terms of AI such as asking what solutions might be possible with AI or whether AI could be applied in a particular situation unveils new opportunities. Collective thinking in terms of AI depends on a shared language.

We need to change the overall culture of the organization, and that depends on getting our people to think in terms of AI.

Changing the culture to make full use of AI across the enterprise is both necessary and difficult, says Chris Couch, senior vice president and CTO at Cooper Standard, which provides components and systems for diverse transportation and industrial markets. Good companies are going to develop people in all functions, whether its finance, purchasing, manufacturing you name it that have some sense about where AI tools can be applied. Bad ones wont, he explains. While AI will continue to be something special that only certain experts use, theres going to be a democratization in the next 10 years. Its one of those things that is not easy to prepare for, but we have to prepare for it. Otherwise, were going to get displaced. When the organization depends on AI literacy, those who lack literacy add discord.

Using AI doesnt merely help with effectiveness at the team level (such as by improving efficiency and decision quality); managers can also use AI to improve an organizations competitiveness. For instance, innovating new processes with AI appears to enhance a companys ability to compete with both existing and new rivals. We compared respondents who said they are using AI primarily to innovate existing processes with those who agreed that their company is using AI primarily to explore new ways of creating value. (See Figure 6.) Respondents who agreed that they are using AI primarily to explore new ways of creating value were 2.5 times more likely to agree that AI is helping their company defend against competitors and 2.7 times more likely to agree that AI is helping their company capture opportunities in adjacent industries. Exploration with AI is correlated to a greater extent with improved competitiveness than exploitation with AI.

Organizations that report greater competitiveness from AI are focused on creating new value with AI.

Organizations can use AI to accelerate these innovation processes for existing processes. Moderna rapidly developed a widely used COVID-19 vaccine with the help of AI. Johnson says Moderna focuses on having a smaller company thats very agile and can move fast. And we see AI as a key enabler for that. The hope is that it helps us to compete in ways that other companies cant. That is certainly the intention here.

Moderna began automating work that had previously been done by humans, including testing the design sequence of messenger RNA (mRNA) used in vaccines that protect against infectious diseases. One of the big bottlenecks was having this mRNA for the scientist to run testing, Johnson says. So we put in place a ton of robotic automation, and a lot of digital systems and process automation and AI algorithms as well. And we went from maybe about 30 mRNAs manually produced in a given month to a capacity of about a thousand in a monthlong period, without using significantly more resources and with much better consistency in quality. As a result, employees at Moderna can evaluate many more options for innovation than before; the companys rapid development of the COVID-19 vaccine was due, in part, to using AI to rapidly test mRNA design sequences. Using AI accelerated innovation, increasing the companys ability to compete with much larger companies.

But speed is far from the only potential benefit of AI. Amit Shah, president of floral and gift retailer 1-800-Flowers, observes, If you think about what differentiates modern organizations, it is not just the ability to adopt technologies thats become a table stake but the ability to out-solve competitors in facing deep problems.

When I think about AI, Shah continues, I think about our competitiveness on that frontier. Five years down the road, I think every new employee that starts out in any company of consequence will have an AI toolkit, like we used to get the Excel toolkit, to both solve problems better and communicate that better to clients, to colleagues, or to any stakeholder. Being a company of consequence in the future may require all employees to work with AI to out-solve competitors with new ways of creating value.

Improving organizational effectiveness is not itself an end goal. After all, organizations can become more effective at the wrong activities: They can achieve misguided objectives, reinforce outdated values, or compete against irrelevant organizations. When CBSs Subramanyam asked her AI team to assess whether executives had the right assumptions about what factors lead to a successful TV show, she was using AI to reassess what being effective means in her organization. Using AI can help a company not only achieve effective outcomes, but also change assumptions about what counts as an effective outcome.

Many executives revealed that their AI implementations were helping them develop or refine strategic assumptions and improve how they measure performance. These changes often lead to shifts in their KPIs. Indeed, our survey found that 64% of the organizations that use AI extensively or in some parts of their processes and offerings adjust their KPIs after using AI. As Pernod Ricards Calloch says, We are planning to monitor new KPIs because AI is helping us measure performance more precisely. For example, one algorithm helps us measure the performance of each marketing campaign in isolation, whereas before, campaigns were running on various media at the same time, and it was impossible to isolate the contribution of each media component. Our ability to isolate and better measure a campaigns performance allows our marketers to be more performance-focused and to make better decisions.

KLM, for example, used AI to develop a new measure to help make complex financial and operational trade-offs involving crew scheduling and passenger delays. Rather than optimizing for on-time performance, Stomph says, we quantified what it takes not to deliver as promised across different departments. That required us to quantify things that you cannot find in your P&L. The measure looks at the cost of various situations, such as a two-hour delay to a crew members schedule if that person is switched from a flight landing at 2 p.m. to one landing at 4 p.m. Whats the price of this? he asks. If you want to run an optimization across different departments, you need to create a single currency to unify all of these players. And the single currency we created was nonperformance cost. The single currency enabled everyone to make decisions based on the same criteria instead of relying on individual judgments with uncoordinated decision-making criteria.

KLMs nonperformance measurement led to changes in a cascade of decisions, including when to swap out crew members. What I find most intriguing about the solutions we have, Stomph says, is even if you will never use the tool, that process of bringing these teams together has been very valuable from a financial and a morale point of view.

Another way that AI implementations can help organizations revise assumptions about effective outcomes is to enable workers to outperform existing KPIs so consistently and so thoroughly that new KPIs are called for. People see that they are outpacing the KPIs that they agreed upon because of AI/ML, Levi Strausss Pallath says. Based on how AI/ML is delivering value to the enterprise, the goalpost keeps shifting.

New success measures become necessary when AI-based solutions make possible new performance benchmarks, obsolesce legacy KPIs, and/or reveal new drivers of performance. Changes in KPIs often accompany shifts in organizational behavior. Indeed, organizations that revise their KPIs because of how they use AI are more likely to see improvements in collaboration than organizations that dont make AI-driven adjustments to their KPIs. Sixty-six percent of respondents who agreed that their KPIs have changed because of AI also saw improvements in team-level collaboration.

Achieving these cultural benefits, particularly at the organizational level, can require considerable change. As Pernod Ricards Calloch describes it, Some processes get changed in a significant way because the data and the processing of the data through AI give us more certainty about some of the elements. You can make quicker decisions live, during a meeting. You can iterate more frequently. And you dont have to wait six months for the return on investment of a campaign to adapt the new wave or to scale it. In fact, you can have more elements. So yes, its significantly changing processes of decision-making. Using AI can accelerate the quality and pace of organizational life extensively, requiring considerable change.

But our research suggests that even when organizations make substantial changes associated with AI, culture does not suffer quite the opposite, in fact. For example, implementing AI is associated with better morale in general. But when combined with business process change, the effects are even more pronounced: The greater (in both number and extent) the change, the greater the improvements in morale. To wit, 57% of organizations that made few changes in business processes reported an increase in morale, while 66% of organizations that made many changes reported an increase in morale. (See Figure 7.) The more that an organization uses AI, the more opportunities there are for cultural benefit.

Morale improves the more processes change.

A strong culture helps encourage AI adoption, and adopting AI can strengthen organizational culture. This cyclical relationship can build through numerous individual process improvements to enhance the overall organizational culture. Zeighami says that when he introduced AI at H&M, he wanted to avoid the common practice of making one part of your organization become very good at that, and then the rest are still lagging behind.

Its almost like putting a tire on a car, he explains. You dont screw one bolt really hard and then do the next one. You just do every bolt a little bit and then tighten everything up. And I think that has been a really good approach for us. Zeighami deployed AI for many company processes, including fashion forecasting, demand forecasting, and price management, along with more personalized customer-facing initiatives. Its been a very vast approach, he observes. Not going too deep, but a little bit in every area to enhance and elevate and change the mindset for everybody so we can become data-led, AI-led, going forward. And we have seen a lot of interesting results. In some areas we even see that working with the AI product has changed peoples way of working with other stuff, because theres a proximity impact on the business. Once an organization introduces AI widely, it can come back and improve not only individual processes but the interfaces between those processes, strengthening the organization as a whole.

Through repeated application and managerial attention, the virtuous cycle between organizational culture and AI use can result in a more cohesive organization, consistently reflecting its desired strategic values. As a result, responsible AI adoption transcends legitimate issues around minimizing bias (in product design, promotion, and customer service) and manipulation (of customers, pricing, and other business practices). Instead, AI becomes a managerial tool to align microbehavior with broader goals, including societal purpose, equity, and inclusivity.

For example, JoAnn Stonier, chief data officer at Mastercard, reports that the financial services corporation launched a data responsibility initiative in 2018 that involved privacy and security issues and included working hard on our ethical AI process. Many of her workplace conversations about AI, she adds, center on minimization of bias as well as how we build an inclusive future. But the conversations dont stop there, she says. The events of this past year have taught us that we need to pay attention to how we are designing products for society and that our data sets are really important. What are we feeding into the machines, and how do we design our algorithmic processes, and what is it going to learn from us?

We understand that data sets are going to have all sorts of bias in them, she continues. I think we can begin to design a better future, but it means being very mindful of whats inherent in the data set. Whats there and whats missing? These discussions help articulate values around which the organization can align, she says. The whole firm is really getting behind this idea of developing a broad-based playbook so that everybody in the organization understands how to think about inclusive concepts.

Pervasive change is complex. As founding director of the Notre Dame-IBM Technology Ethics Lab, Elizabeth Renieris is acutely aware of the complexities of these conversations and how they continue to evolve. The ethics conversation in the past couple of years started out with the lens very much on the technology, she says. Its been turned around and focused on whos building it and whos at the table those are the really important questions.

The value of ethics, she adds, is, rather than looking at the narrow particulars and tweaking around the edges of the specific technology or implementation, to step back and have that conversation about values to ask, What are our values, and how do those values align with what it is that were working on from a technology standpoint? Stepping back may cause discomfort. But through these conversations, AI can have a profound effect on organizational culture.

Read more:
The Cultural Benefits of Artificial Intelligence in the Enterprise - MIT Sloan

Artificial intelligence is getting better at writing, and universities should worry about plagiarism – The Conversation CA

The dramatic rise of online learning during the COVID-19 pandemic has spotlit concerns about the role of technology in exam surveillance and also in student cheating.

Some universities have reported more cheating during the pandemic, and such concerns are unfolding in a climate where technologies that allow for the automation of writing continue to improve.

Over the past two years, the ability of artificial intelligence to generate writing has leapt forward significantly, particularly with the development of whats known as the language generator GPT-3. With this, companies such as Google, Microsoft and NVIDIA can now produce human-like text.

AI-generated writing has raised the stakes of how universities and schools will gauge what constitutes academic misconduct, such as plagiarism. As scholars with an interest in academic integrity and the intersections of work, society and educators labour, we believe that educators and parents should be, at the very least, paying close attention to these significant developments.

The use of technology in academic writing is already widespread. For example, many universities already use text-based plagiarism detectors like Turnitin, while students might use Grammarly, a cloud-based writing assistant. Examples of writing support include automatic text generation, extraction, prediction, mining, form-filling, paraphrasing, translation and transcription.

Read more: In an AI world we need to teach students how to work with robot writers

Advancements in AI technology have led to new tools, products and services being offered to writers to improve content and efficiency. As these improve, soon entire articles or essays might be generated and written entirely by artificial intelligence. In schools, the implications of such developments will undoubtedly shape the future of learning, writing and teaching.

Research has revealed that concerns over academic misconduct are already widespread across institutions higher education in Canada and internationally.

In Canada, there is little data regarding the rates of misconduct. Research published in 2006 based on data from mostly undergraduate students at 11 higher education institutions found 53 per cent reported having engaged in one or more instances of serious cheating on written work, which was defined as copying material without footnoting, copying material almost word for word, submitting work done by someone else, fabricating or falsifying a bibliography, submitting a paper they either bought or got from someone else for free.

Academic misconduct is in all likelihood under-reported across Canadian higher education institutions.

There are different types of violations of academic integrity, including plagiarism, contract cheating (where students hire other people to write their papers) and exam cheating, among others.

Unfortunately, with technology, students can use their ingenuity and entrepreneurialism to cheat. These concerns are also applicable to faculty members, academics and writers in other fields, bringing new concerns surrounding academic integrity and AI such as:

We are asking these questions in our own research, and we know that in the face of all this, educators will be required to consider how writing can be effectively assessed or evaluated as these technologies improve.

At the moment, little guidance, policy or oversight is available regarding technology, AI and academic integrity for teachers and educational leaders.

Over the past year, COVID-19 has pushed more students towards online learning a sphere where teachers may become less familiar with their own students and thus, potentially, their writing.

While it remains impossible to predict the future of these technologies and their implications in education, we can attempt to discern some of the larger trends and trajectories that will impact teaching, learning and research.

A key concern moving forward is the apparent movement towards the increased automation of education where educational technology companies offer commodities such as writing tools as proposed solutions for the various problems within education.

An example of this is automated assessment of student work, such as automated grading of student writing. Numerous commercial products already exist for automated grading, though the ethics of these technologies are yet to be fully explored by scholars and educators.

Read more: Online exam monitoring can invade privacy and erode trust at universities

Overall, the traditional landscape surrounding academic integrity and authorship is being rapidly reshaped by technological developments. Such technological developments also spark concerns about a shift of professional control away from educators and ever-increasing new expectations of digital literacy in precarious working environments.

Read more: Precarious employment in education impacts workers, families and students

These complexities, concerns and questions will require further thought and discussion. Educational stakeholders at all levels will being required to respond and rethink definitions as well as values surrounding plagiarism, originality, academic ethics and academic labour in the very near future.

The authors would like to sincerely thank Ryan Morrison, from George Brown College, who provided significant expertise, advice and assistance with the development of this article.

See original here:
Artificial intelligence is getting better at writing, and universities should worry about plagiarism - The Conversation CA

A Tale Of Two Jurisdictions: Sufficiency Of Disclosure For Artificial Intelligence (AI) Patents In The US And The EPO – Intellectual Property – United…

PatentNext Summary: In order to prepareapplications for filing in multiple jurisdictions, practitionersshould be cognizant of claiming styles in the various jurisdictionsthat they expect to file AI-related patent applications in, anddraft claims accordingly. For example, different jurisdictions,such as the U.S. and EPO, have different legal tests that canresult in different styles for claiming artificialintelligence(AI)-related inventions.

In this article, we will compare two applications, one in theU.S. and the other in the EPO, that have the same or similarclaims. Both applications claim priority to the same PCTApplication (PCT/AT2006/000457) (the "'427 PCTApplication"), which is published as PCT Pub. No.WO/2007/053868.

As we shall see, despite the application having the same orsimilar claims, prosecution of the applications in the twojurisdictions nonetheless resulted in different outcomes, with theU.S. application prosecuted to allowance and the EPO applicationending in rejection.

****

Pertinent to our discussion is an overview of AI. A briefdescription of AI follows before analysis of the AI-related claimsat issue.

Artificial Intelligence (AI) is fundamentally a data-driventechnology that takes unique datasets as input to train AI computermodels. Once trained, an AI computer model may take new data asinput to predict, classify, or otherwise output results for use ina variety of applications.

Machine learning, arguably the most widely used AI technique,may be described as a process that uses data and algorithms totrain (or teach) computer models, which usually involves thetraining of weights of the model. Training typically involvescalculating and updating mathematical weights (i.e., numeralvalues) of a model based on input that can comprise hundreds,thousands, millions, etc. sets of data. The trained model allowsthe computer to make decisions without the need for explicit orrule-based programming.

In particular, machine learning algorithms build a model ontraining data to identify and extract patterns from the data andtherefore acquire (or learn) unique knowledge that can be appliedto new data sets.

For more information, see Artificial Intelligence & the IntellectualProperty Landscape

AI inventions are fundamentally software-related inventions. Inthe U.S., as a practical rule, software-related patents shoulddisclose an algorithm by which the software-related invention isachieved. An algorithm provides support for a software-relatedpatent pursuant to 35 U.S.C. 112(a) including (1) byproviding sufficiency of disclosure for the patent's"written description" and (2) by "enabling" oneof ordinary skill in the art (e.g., a computer engineer or computerprogrammer) to make or use the related software-related inventionwithout "undue experimentation." Without such support, apatent claim can be held invalid. For more information regardinggeneral aspects of the sufficiency of disclosure in the U.S. forsoftware-related inventions, see Why including an "Algorithm" isImportant for Software Patents (Part 2)

U.S. Patent 8,920,327 (the "'327 Patent") issuedfrom the '457 PCT Application. The ''327 Patent is anexample of an AI patent that did not experiencesufficiency issues in the U.S. The below provides an overview ofthe '327 Patent.

The '327 Patent is titled "Method for DeterminingCardiac Output" and includes a single independent claimregarding a method for cardiac output from an arterial bloodpressure curve. The method is implemented via a cardiac device, asillustrated in Figure 1 (copied below):

Fig. 1 illustrates device 1 for implementing the invention ofthe '327 patent, where measuring device 2 measures theperipheral blood pressure curve, and where related measurement datais fed into device 1 via line 3, and stored and evaluated there.The device further comprises optical display device 4, input panel5, and keys 6 for inputting and displaying information.

The claimed method includes an AI aspect, i.e., namely the useof "an artificial neural network having weightingvalues that are determined by learning."

Claim 1 is copied below (with the AI aspectbolded):

1. A method for determiningcardiac output from an arterial blood pressure curve measured at aperipheral region, comprising the steps of:

measuring the arterial bloodpressure curve at the peripheral region; arithmeticallytransforming the measured arterial blood pressure curve to anequivalent aortic pressure; and

calculating the cardiac outputfrom the equivalent aortic pressure,

wherein the arithmetictransformation of the arterial blood pressure curve measured at theperipheral region into the equivalent aortic pressure is performedby the aid of an artificial neural networkhaving weighting values that are determined bylearning.

Figure 3 of the '327 patent (copied below) is a schematicillustration of the artificial neural network, as recited in claim1.

The specification of the '327 patent describes that"FIG. 3 illustrates the structure of the neural network...,and it is apparent that the neural network ... is comprised ofthree layers 14, 15, 16." The specification discloses that asupervised learning algorithm is used to train the weights of themodel, e.g., "[t]he weights and the bias for the latter twolayers 15 and 16 are determined by supervised learning."

The input data fed to the supervised learning algorithm to trainthe AI model includes "associated blood pressure curve pairsactually determined by measurements in the periphery or in theaorta, respectively, are used." The measurements used for theinput data may come "from patients of different ages, sexes,constitutional types, health conditions and the like."

No issues with respect to sufficiency were raised during theprosecution of the application in the U.S. that was issued as the'327 patent.

More generally, issues of sufficiency in the U.S. typicallyarise in litigation, and result in expert testimony, i.e., "abattle of the experts," where expert witnesses (e.g.,typically university professors or industry consultants) fromopposing sides opine on the knowledge of a person of ordinary skillin the art and sufficiency of disclosure in view of thatperson.

The EPO has developed its own, yet similar, stance on AI-relatedinvention when compared with the U.S. Nonetheless, outcomes ofprosecution can be different. The below provides a cursory overviewof developments in the EPO with respect to AI-related inventionsand analyzes the treatment of an EPO application as filed based onthe PCT Application '457 (which is the same PCT Application asfor the '327 patent discussed above).

Generally, artificial intelligence inventions may be patented inthe European Patent Office (EPO). For example, in its Guidelinesfor Examination, the EPO defines AI and machine learning as"based on computational models and algorithms forclassification, clustering, regression and dimensionalityreduction, such as neural networks, genetic algorithms, supportvector machines, k-means, kernel regression and discriminantanalysis." Section 3.3.1 (Artificial intelligence and machinelearning).

As such, the EPO dubs AI and machine learning as "per se ofan abstract mathematical nature," irrespective of whether suchmodels may be trained with training data. Id. Thus, simplyclaiming a machine learning model (e.g., such as a "neuralnetwork") does not, alone, necessarily imply the use of a"technical means" in accordance with EPO law.

Nonetheless, the Guidelines for Examination at the EPO recognizethat the use of an AI model, when claimed as a whole with theadditional subject matter, may demonstrate a sufficient technicalcharacter. Id. As an example, the Guidelines forExamination at the EPO states that "the use of a neuralnetwork in a heart-monitoring apparatus for the purpose ofidentifying irregular heartbeats makes a technicalcontribution." Id. As a further example, the EPOGuidelines for Examination further states that "[t]heclassification of digital images, videos, audio or speech signalsbased on low-level features (e.g. edges or pixel attributes forimages) are further typical technical applications ofclassification algorithms." Id.

In a decision in 2020, the EPO Board of Appeals rejected amachine learning-based patent application that claimed an"artificial neural network" because the patentspecification failed to sufficiently disclose how the artificialneural network was trained. See T0161/18 (Equivalent aortic pressure / ARCSEIBERSDORF). The application in question claimed priority to thePCT Application '457, which is the same parent application asthe '327 patent, as discussed above.

The claims were the same or similar as to those in the U.S.,where the claims-at-issue directed to determining cardiac outputfrom an arterial blood pressure curve measured at a periphery, andrecited, in part (with respect to AI), that the "bloodpressure curve measured on the periphery is converted into theequivalent aortic pressure with the help of anartificial neural network, the weighting values ??ofwhich are determined bylearning."

Claim 1 is reproduced below (in English based on a machinetranslation of the original opinion German):

1. A method for determining thecardiac output from an arterial blood pressure curve measured atthe periphery, in which the blood pressure curve measured at theperiphery is mathematically transformed to the equivalent aorticpressure and the cardiac output is calculated from the equivalentaortic pressure, characterized in that the transformation of theblood pressure curve measured on the periphery is converted intothe equivalent aortic pressure with the help of anartificial neural network, the weighting values ??ofwhich are determined by learning.

The Board analyzed the claim in view of the specificationpursuant to Article 83 EP (Sufficient disclosure). As described bythe Board, Article 83 EPC requires that the invention be disclosedin the European patent application so clearly and completely that aperson skilled in the art can carry it out. For this, thedisclosure of the invention in the application must enable theperson skilled in the art to reproduce the technical teachinginherent in the claimed invention on the basis of his generalspecialist knowledge.

The Board then turned to the specification to determine whetherit disclosed enough support to meet these requirements in view ofthe claimed "artificial neural network." However, thespecification was found lacking because it failed to"disclose which input data aresuitable for training the artificial neural network according tothe invention, or at least one data set suitable for solving thetechnical problem at hand."

Instead, the Board found that the specification "merelyreveals that the input data should cover a broad spectrum ofpatients of different ages, genders, constitution types, healthstatus and the like."

Therefore, the Board found that the training of the artificialneural network could therefore not be reworked by the personskilled in the art, and the person skilled in the art can thereforenot carry out the invention.

Because of these deficiencies, the Board found that thespecification failed to provide sufficient disclosure pursuant toArticle 83 EPC.

For similar reasons, the Board further found that the claimedsubject matter lacked an "inventive step" pursuant toArticle 56 EPC. Specifically, the Board found that the claimed"artificial neural network" was not adapted for thespecific, claimed application because the specification failed todisclose how the artificial neural network was trained, andspecifically failed to disclose weight values that resulted fromsuch training. For this reason, the claimed "artificial neuralnetwork" could not be distinguished from the cited prior art,which resulted in failure to demonstrate the requisite inventivestep.

As the Board described:

In the present case, the claimedneural network is therefore not adapted for the specific, claimedapplication. In the opinion of the Chamber, there is therefore onlyan unspecified adaptation of the weight values, which is in thenature of every artificial neural network. The board is thereforenot convinced that the claimed effect will be achieved in theclaimed method over the entire range claimed. This effect cannot,therefore, be taken into account in the assessment of inventivestep in the sense of an improvement over the prior art.

Accordingly, at least with respect to patent applications filedin the EPO, and where an AI or machine learning model is to bedistinguished from the prior art, then a patent applicant may wantto include an example training data set, example trained weights,or at least sufficiently describe the input used to train the modelon a specific, claimed application or end-use. For example, atleast one example of data can be provided (or claimed) to show theinputs used to train specific weights, which may allow for theclaim to have sufficient disclosure, and, at the same time allowthe claim to cover a spectrum of AI models trained with aparticular set of data.

For the time being, such disclosure for an EPO case could beconsidered as additional when compared with the sufficiency ofdisclosure in the U.S. However, it is to be understood that theU.S. Patent office has also indicated the importance of includingtraining data or specific species of data used to train a model inits example guidance. See How to Patent an Artificial Intelligence (AI)Invention: Guidance from the U.S. Patent Office (USPTO). In anyevent, while there have been few court cases on AI-relatedinventions in the U.S. (see How the Courts treat Artificial Intelligence (AI)Patent Inventions: Through the Years since Alice), future casesmay indicate whether the U.S. will trend towards the EPO'sdecision in T0161/18 with respect to the sufficiency ofdisclosure.

The content of this article is intended to provide a generalguide to the subject matter. Specialist advice should be soughtabout your specific circumstances.

Read the rest here:
A Tale Of Two Jurisdictions: Sufficiency Of Disclosure For Artificial Intelligence (AI) Patents In The US And The EPO - Intellectual Property - United...

US/EU Initiative Spotlights Cooperation, Differing Approaches To Regulation Of Artificial Intelligence Systems – Privacy – Worldwide – Mondaq News…

To print this article, all you need is to be registered or login on Mondaq.com.

In late September 2021, representatives from the U.S. and theEuropean Union met to coordinate objectives related to the U.S.-EUTrade and Technology Council, and high on the Council's agendawere the societal implications of the use of artificialintelligence systems and technologies ("AISystems"). The Council's public statements on AISystems affirmed its "willingness and intention to develop andimplement trustworthy AI" and a "commitment to ahuman-centric approach that reinforces shared democraticvalues," while acknowledging concerns that authoritarianregimes may develop and use AI Systems to curtail human rights,suppress free speech, and enforce surveillance systems. Given theincreasing focus on the development and use of AI Systems from bothusers and investors, it is becoming imperative for companies totrack policy and regulatory developments regarding AI on both sidesof the Atlantic.

At the heart of the debate over the appropriate regulatorystrategy is a growing concern over algorithmic bias thenotion that the algorithm powering the AI Systems in question hasbias "baked in" that will manifest in its results.Examples of this issue abound job applicant systemsfavoring certain candidates over others, or facial recognitionsystems treating African Americans differently than Caucasians,etc. These concerns have been amplified over the last 18 months associal justice movements have highlighted the real-worldimplications of algorithmic bias.

In response, some prominent tech industry players have postedposition statements on their public-facing websites regarding theiruse of AI Systems and other machine learning practices. Thesestatements typically address issues such as bias, fairness, anddisparate impact stemming from the use of AI Systems, but often arenot binding or enforceable in any way. As a result, these publicstatements have not quelled the debate around regulating AISystems; rather, they highlight the disparate regulatory regimesand business needs that these companies must navigate.

When the EU's General Data Protection Regulation("GDPR") came into force in 2018, itprovided prescriptive guidance regarding the treatment of automateddecision-making practices or profiling. Specifically, Article 22 isgenerally understood to implicate technology involving AI Systems.Under that provision, EU data subjects have the right not to besubject to decisions based solely on automated processing (andwithout human intervention) which may produce legal effects for theindividual. In addition to Article 22, data processing principlesin the GDPR, such as data minimization and purpose limitationpractices, are applicable to the expansive data collectionpractices inherent in many AI Systems.

Consistent with the approach enacted in GDPR, recently proposedEU legislation regarding AI Systems favors tasking businesses,rather than users, with compliance responsibilities. The EU'sArtificial Intelligence Act (the "Draft AI Regulation"),released by the EU Commission in April 2021, would requirecompanies (and users) who use AI Systems as part of their businesspractices in the EU to limit the harmful impact of AI. If enacted,the Draft AI Regulation would be one of the first legal frameworksfor AI designed to "guarantee the safety and fundamentalrights of people and businesses, while strengthening AI uptake,investment and innovation across the EU." The Draft AIRegulation adopts a risk-based approach, categorizing AISystems as unacceptable risk, high risk, and minimal risk. Much ofthe focus and discussion with respect to the Draft AI Regulationhas concerned (i) what types of AI Systems are consideredhigh-risk, and (ii) the resulting obligations on such systems.Under the current version of the proposal, activities that would beconsidered "high-risk" include employee recruiting andcredit scoring, and the obligations for high-risk AI Systems wouldinclude maintaining technical documentation and logs, establishinga risk management system and appropriate human oversight measures,and requiring incident reporting with respect to AI Systemmalfunctioning.

While AI Systems have previously been subject to guidelines fromgovernmental entities and industry groups, the Draft AI Regulationwill be the most comprehensive AI Systems law in Europe, if not theworld. In addition to the substantive requirements previewed above,it proposes establishing an EU AI board to facilitateimplementation of the law, allowing Member State regulators toenforce the law, and authorizing fines up to 6% of acompany's annual worldwide turnover. The draft law will likelybe subject to a period of discussion and revision with thepotential for a transition period, meaning that companies that dobusiness in Europe or target EU data subjects will have a few yearsto prepare.

Unlike the EU, the U.S. lacks comprehensive federal privacylegislation and has no law or regulation as specifically tailoredto AI activities. Enforcement of violations of privacy practices,including data collection and processing practices through AISystems, primarily originates from Section 5 of the Federal TradeCommission ("FTC") Act, which prohibitsunfair or deceptive acts or practices. In April 2020, the FTCissued guidance regarding the use of AI Systems designed to promotefairness and equity. Specifically, the guidance directed that theuse of AI tools should be "transparent, explainable, fair, andempirically sound, while fostering accountability." The changein administration has not changed the FTC's focus on AIsystems. First, public statements from then-FTC Acting ChairRebecca Slaughter in February 2021 cited algorithms that result inbias or discrimination, or AI-generated consumer harms, as a keyfocus of the agency. Then, the FTC addressed potential bias in AISystems on its website in April 2021 and signaled that unlessbusinesses adopt a transparency approach, test for discriminatoryoutcomes, and are truthful about data use, FTC enforcement actionsmay result.

At the state level, recently enacted privacy laws in California,Colorado and Virginia will enable consumers in those states toopt-out of the use of their personal information in the context of"profiling," defined as a form of automated processingperformed on personal information to evaluate, analyze, or predictaspects related to individuals. While AI Systems are notspecifically addressed, the three new state laws require datacontrollers (or equivalent) to conduct data protection impactassessments to determine whether processing risks associated withprofiling may result in unfair or disparate impact to consumers. Inall three cases, yet-to-be promulgated implementing regulations mayprovide businesses (and consumers) with additional guidanceregarding operationalizing automated decision-making requests upuntil the laws' effective dates (January 2023 for Virginia andCalifornia, July 2023 for Colorado).

Proliferating use of AI Systems has dramatically increased thescale, scope, and frequency of processing of personal information,which has led to an accompanying increase in regulatory scrutiny toensure that harms to individuals are minimized. Businesses thatutilize AI Systems should adopt a comprehensive governance approachto comply with both the complimentary and divergent aspects of theU.S. and EU approaches to the protection of individual rights.Although laws governing the use of AI Systems remain in flux onboth sides of the Atlantic, businesses that utilize AI in theirbusiness practices should consider asking themselves the followingquestions:

The content of this article is intended to provide a generalguide to the subject matter. Specialist advice should be soughtabout your specific circumstances.

See the original post:
US/EU Initiative Spotlights Cooperation, Differing Approaches To Regulation Of Artificial Intelligence Systems - Privacy - Worldwide - Mondaq News...

This Robotic Combat Vehicle Will Use Artificial Intelligence to Save Soldiers – The National Interest

The Army is planning to armthe newten-ton Robotic Combat Vehicle-Medium(RCV-M)withthirty-millimeterchain gun cannons, anti-tank missiles and remotely operated guns. This will enable it to conductdirect-attack missionsso thatsoldierswont have to beinthe lineofenemy fire.

Armyanddefenseindustryofficials havestatedthat humans willcontinuetomakedecisions about the use of lethal force.However, thisdoes notrestrictthem frompushing the envelope of autonomy and potentially enabling unmanned systems to process greater volumes of information and perform a widerrange of functionswithout needing human intervention.

We are looking at using unmanned vehicles to expand the network and expand the line-of-sight so we can push these robots out as far aspossible,so soldiers do not have to do that, Maj. Gen. Ross Coffman,the directorof theNext-Generation Combat Vehicle Cross-Functional TeamforArmy Futures Command, told reporters at the 2021 Association of the United States Army Annual Symposium.

TheRCV-Mwill becapable of usinga wide range of weapons, sensors and technologies.AJuly2021Congressional Research Servicereport on the vehicleshows that it willcarryJavelin anti-tank missiles and theXM813 Bushmaster chain gunas well as smoke obscuration measures, the report states. These may includeamphibious kits, electronic warfare (EW) modules, counter Unmanned Aerial System (UAS)systems, and nuclear, radiological, biological, and chemical sensors,according tothe report.

Lethal direct fire missions, such as using a Javelin or Bushmaster Chain Gun, will be closely monitored byArmy and defense industry officials.Also, the RCV-Mcouldhostnonlethal defensive interceptorsand use them to deterincoming munitionsorautonomous launch and recovery of surveillance drones.Additionally, it will be able toaccommodate a wide range of payloads and potential hardware configurationslike clearing minefields.For several years,Army Futures Command has been experimenting withusingrobotic vehiclestoclear minefields or breach obstaclesthat might preventarmored columnsfrom conducting a mission.Thiswill allowsoldiers to operate at a safe standoff distance while robotic vehiclestake major risks.

What we learned is based on their mobility, their excellent mobility and their autonomous behaviors, we can actually have them move on a separate axis of advance and link up with the humans on the objective, Coffman told the National Interest.So,they can autonomously move without humans, link up with the humans, transfer back control, and then execute the mission. This gives the enemy multiple dilemmas.

Kris Osborn is the defense editor for theNational Interest. Osborn previously served at the Pentagon as a Highly Qualified Expert with the Office of the Assistant Secretary of the ArmyAcquisition, Logistics & Technology. Osborn has also worked as an anchor and on-air military specialist at national TV networks. He has appeared as a guest military expert on Fox News, MSNBC, The Military Channel, and The History Channel. He also has aMasters Degreein Comparative Literature from Columbia University.

Image: Reuters

Go here to see the original:
This Robotic Combat Vehicle Will Use Artificial Intelligence to Save Soldiers - The National Interest

Artificial Intelligence (AI) in Manufacturing Market Worth $13.96 billion by 2028 Exclusive Report by Meticulous Research – Yahoo Finance

Artificial Intelligence in Manufacturing Market By Component, Technology (ML, NLP, Computer Vision), Application (Predictive Maintenance Quality Management, Supply Chain, Production Planning), Industry Vertical, & Geography - Global Forecast to 2028

Redding, California, Nov. 02, 2021 (GLOBE NEWSWIRE) -- According to a new market research report titled, AI in Manufacturing Market By Component, Technology (ML, NLP, Computer Vision), Application (Predictive Maintenance Quality Management, Supply Chain, Production Planning), Industry Vertical, and Geography Global Forecast to 2028, published by Meticulous Research, the artificial intelligence (AI) in manufacturing market is expected to grow at a CAGR of 38.6% during the forecast period to reach $13.96 billion by 2028.

Download Free Sample Report Now @ https://www.meticulousresearch.com/download-sample-report/cp_id=4983

The rising popularity of artificial intelligence in manufacturing industry for optimizing logistics & supply chains, enhancing production outcomes, advancing process effectiveness, reducing costs and downtime in production lines while delivering finished products to consumers are expected to drive the growth of the AI in manufacturing market. Additionally, the advent of Industrial 4.0, the increasing volume of large complex data, and the rising adoption of industrial IoT further contribute to market growth.

However, the lack of infrastructure and high procurement and operating costs are expected to restrain the growth of this market to a certain extent.

Impact of COVID-19 on AI in Manufacturing Market

The COVID-19 pandemic outbreak created serious challenges to the worlds economy and for industry verticals. The SARS-CoV-2, the virus responsible for the global COVID-19 pandemic, started showing its distressing collision on most profitable businesses across the globe, leading to a remote workforce, ensuring peoples health & safety, and business application integrity. The impact of the COVID-19 outbreak has varied by each industry sector's level of resilience. Additionally, the lockdowns imposed to contain the pandemic resulted in severe losses to businesses. Manufacturers across the globe faced grave challenges, such as diminished demand, production, and revenues, as the COVID-19 pandemic intensified in 2020. The automobile, semiconductors & electronics, and heavy metal & machinery manufacturing industries witnessed raw material shortages, with manufacturers temporarily closing down factories or minimizing production.

Story continues

Speak to our Analysts to Understand the Impact of COVID-19 on Your Business: https://www.meticulousresearch.com/speak-to-analyst/cp_id=4983

According to the United Nations Conference on Trade and Development (UNCTAD), the COVID-19 pandemic is expected to reduce the global FDI by around 515% due to the temporary shutdown of the manufacturing sector. A survey conducted by the National Association of Manufacturers (NAM) stated that around 78% of manufacturers anticipated a financial impact, and 35.5% faced supply chain disruptions due to COVID-19. These factors led manufacturing companies to deprioritize their digital transformation strategies, including equipping their production units with AI.

Consequently, the AI in manufacturing market witnessed a sharp decline in 2020. Thus, manufacturing industries require considerable productive time and assistance from local governments to get back on track and overcome the COVID-19 crisis. Several governments plan to launch favorable initiatives, such as incentive programs promoting investments in the private sector, tax exemptions, and lowering corporate interest rates. For instance, in 2021, Cisco Systems, Inc. (U.S.) launched a collaborative framework under Ciscos Country Digital Acceleration (CDA) program to accelerate digitization and support inclusive pandemic recovery across South Korea. Such developments and initiatives are exhibiting positive impacts on the growth of the market. Based on geography, the EU countries were affected the most by the COVID-19 pandemic, followed by the U.S. On the other hand, China is gradually recovering from the pandemic, with positive developments in the supply chain industry.

Several organizations post-COVID-19 pandemic might strategize to downsize by cutting business lines considered as non-critical. Many leading AI in manufacturing players are eying this crisis as a new opportunity for restructuring and revisiting their existing strategies with advanced product portfolios. AI technology providers for manufacturing industries are focused on new applications and delivery models to create smart automation technologies, digitization, and advanced AI applications. For instance, in 2021, Nvidia Corporation (U.S.) partnered with Google Cloud (U.S.) to create the industrys first AI-on-5G Lab. This partnership helped accelerate the creation of smart cities, smart factories, and other advanced 5G and AI applications. Also, in 2021, General Electric Company (U.S.) partnered with the Global Manufacturing and Industrialization Summit (GMIS) (UAE) to explore the role of digitization, lean manufacturing, and workplace safety. Such developments and initiatives are expected to help manufacturing companies recover faster and reduce dependencies on physical process handling.

Hence, despite the pandemic affecting the AI in manufacturing market, it still holds considerable potential to bounce back with the gradual recovery of the manufacturing sector.

The AI in manufacturing market is segmented based on component (hardware [processors, memory solutions, and networking solutions], software [AI platforms and AI solutions], service [deployment & integration, support & maintenance]), technology (machine learning, natural language processing, computer vision, speech & voice recognition, context-aware computing), application (predictive maintenance & machinery inspection, quality management, supply chain optimization, industrial robot, production planning, material handling, field services, safety planning, cybersecurity, energy management), industry verticals (automotive, semiconductors & electronics, heavy metals & machine manufacturing, energy & power, aerospace & defense, medical devices, pharmaceuticals, and FMCG), and region. The study also evaluates industry competitors and analyses the market at the regional and country levels.

Based on component, the hardware segment is estimated to account for the largest share of the AI in manufacturing market in 2021. The large market share of this segment is primarily driven by the increasing demand for robust and cost-effective devices, including servers, storage, and networking devices. However, the software segment is slated to grow at the fastest CAGR during the forecast period due to the high adoption of cloud-based technologies and the increasing demand for AI platforms to streamline processes and operations.

Based on technology, the machine learning segment is estimated to account for the largest share of the AI in manufacturing market in 2021. The large market share of this segment is primarily driven by the rising need for identifying, monitoring, and analyzing the critical system variables during the manufacturing process, growing demand for predictive maintenance & machinery inspection, and the increase in unstructured data generated by the manufacturing industry. However, the natural language processing segment is slated to grow at the fastest CAGR during the forecast period due to the need to strengthen interactions with search engines by allowing queries to be assessed faster in an efficient manner and the growing demand for cloud-based NLP solutions to reduce overall costs, facilitate smart environments, and enhance scalability.

Based on application, the predictive maintenance & machinery inspection segment is estimated to account for the largest share and witness the fastest CAGR of the AI in manufacturing market in 2021. This segment's large market share and high growth rate are primarily driven by the increasing demand to reduce costs related to operating heavy equipment, growing demand for equipment uptime & availability, reducing maintenance planning time, improving production capacity, and real-time reporting of manufacturing issues in industries.

Quick Buy AI in Manufacturing Market Research Report: https://www.meticulousresearch.com/Checkout/49201841

Based on industry vertical, the automotive industry is estimated to account for the largest share of the overall AI in manufacturing market in 2021. The large market share of this segment is primarily driven by the rising adoption of advanced AI automotive solutions for fault detection & isolation, quality management, smart manufacturing, production monitoring, and the need for predictive maintenance & machinery inspection solutions.

However, the medical devices manufacturing sector is slated to grow at the fastest CAGR during the forecast period due to the outbreak of the COVID-19 pandemic and the rising focus on preventive medical equipment maintenance to reduce unplanned downtime, enhance production quality control, and improve operational productivity.

Based on geography, Asia-Pacific is estimated to account for the largest share and witness the fastest CAGR of the AI in manufacturing market in 2021. This regions large market share and high growth rate are primarily attributed to the presence of major AI in manufacturing players along with several emerging startups in the region, increasing investments by technology leaders, and increasing digitization along with the strong presence of automobile and electronics and semiconductor companies and their focus on developing advanced solutions to optimize manufacturing operations and processes in the region.

The report also includes an extensive assessment of the key strategic developments adopted by the leading market participants in the industry over the past four years. The AI in manufacturing market has witnessed various strategies in recent years, such as partnerships & agreements. These strategies enabled companies to broaden their product portfolios, advance capabilities of existing products, and gain cost leadership in the AI in manufacturing market. For instance, in 2021, SAP SE (Germany) partnered with Google Cloud (U.S.) to augment existing business systems with Google Cloud capabilities in Artificial Intelligence (AI) and Machine Learning (ML). Also, SAP SE partnered with Plataine Ltd. (U.S.) to integrate IIoT and AI-based software for digital manufacturing. This partnership enabled customers to benefit from a holistic smart factory solution that extends across production operations. In 2021, Robert Bosch (Germany) collaborated with Capgemini SE (France) for intelligent manufacturing, digitization, and sustainability of their production plants.

The AI in manufacturing market is fragmented in nature. The major players operating in this market include Alphabet, Inc. (U.S.), IBM Corporation (U.S.), Intel Corporation (U.S.), Microsoft Corporation (U.S.), Nvidia Corporation (U.S.), Oracle Corporation (U.S.), Amazon Web Services, Inc. (U.S.), Siemens AG (Germany), General Electric Company (U.S.), SAP SE (Germany), Robert Bosch GmbH (Germany), Cisco Systems, Inc. (U.S.), Rockwell Automation, Inc. (U.S.), Advanced Micro Devices, Inc. (U.S.), and Sight Machine Inc. (U.S.) among others.

To gain more insights into the market with a detailed table of content and figures, click here: https://www.meticulousresearch.com/product/artificial-intelligence-in-manufacturing-market-4983

Scope of the Report:

AI in Manufacturing Market, by Component

Processors

Memory Solutions

Networking Solutions

Deployment & Integration

Support & Maintenance

AI in Manufacturing Market, by Technology

AI in Manufacturing Market, by Application

Predictive Maintenance & Machinery Inspection

Quality Management

Supply Chain Optimization

Industrial Robot/Robotics & Factory Automation

Production Planning

Material Handling

Field Services

Safety Planning

Cybersecurity

Energy management

AI in Manufacturing Market, by Industry Vertical

AI in Manufacturing Market, by Geography

North America

Europe

Germany

U.K.

France

Italy

Spain

Netherlands

Russia

Ireland

Turkey

Rest of Europe

Asia-Pacific

Japan

China

India

South Korea

Australia & New Zealand

Thailand

Indonesia

Taiwan

Vietnam

Rest of Asia-Pacific

Latin America

Mexico

Brazil

Rest of Latin America

Middle East and Africa

Download Free Sample Report Now @ https://www.meticulousresearch.com/download-sample-report/cp_id=4983

Amidst this crisis, Meticulous Research is continuously assessing the impact of COVID-19 pandemic on various sub-markets and enables global organizations to strategize for the post-COVID-19 world and sustain their growth. Let us know if you would like to assess the impact of COVID-19 on any industry here- https://www.meticulousresearch.com/custom-researchRelated Reports:

Artificial Intelligence in Retail Market by Product, Application (Predictive Merchandizing, Programmatic Advertising), Technology (Machine Learning, Natural Language Processing), Deployment (Cloud, On-Premises), and Geography - Global Forecast to 2027

https://www.meticulousresearch.com/product/artificial-intelligence-in-retail-market-4979

Healthcare Artificial Intelligence Market by Product and Services (Software, Services), Technology (Machine Learning, NLP), Application (Medical Imaging, Precision Medicine, Patient Management), End User (Hospitals, Patients) - Global Forecast to 2027

https://www.meticulousresearch.com/product/healthcare-artificial-intelligence-market-4937

Automotive Artificial Intelligence (AI) Market by Component (Hardware, Software), Technology (Machine Learning, Computer Vision), Process (Signal Recognition, Image Recognition) and Application (Semi-Autonomous Driving) - Global Forecast to 2027

https://www.meticulousresearch.com/product/automotive-artificial-intelligence-market-4996

Artificial Intelligence in Supply Chain Market by Component (Platforms, Solutions) Technology (Machine Learning, Computer Vision, Natural Language Processing), Application (Warehouse, Fleet, Inventory Management), and by End User - Global Forecast to 2027

https://www.meticulousresearch.com/product/artificial-intelligence-ai-in-supply-chain-market-5064

Artificial Intelligence (AI) in Cybersecurity Market by Technology (ML, NLP), Security (Endpoint, Cloud, Network), Application (DLP, UTM, Encryption, IAM, Antivirus, IDP), Industry (Retail, Government, Automotive, BFSI, IT, Healthcare, Education), Geography - Global Forecast to 2027

https://www.meticulousresearch.com/product/artificial-intelligence-in-cybersecurity-market-5101

About Meticulous Research

Meticulous Research was founded in 2010 and incorporated as Meticulous Market Research Pvt. Ltd. in 2013 as a private limited company under the Companies Act, 1956. Since its incorporation, the company has become the leading provider of premium market intelligence in North America, Europe, Asia-Pacific, Latin America, and the Middle East & Africa.

The name of our company defines our services, strengths, and values. Since the inception, we have only thrived to research, analyze, and present the critical market data with great attention to details. With the meticulous primary and secondary research techniques, we have built strong capabilities in data collection, interpretation, and analysis of data including qualitative and quantitative research with the finest team of analysts. We design our meticulously analyzed intelligent and value-driven syndicate market research reports, custom studies, quick turnaround research, and consulting solutions to address business challenges of sustainable growth.

Contact:Mr. Khushal BombeMeticulous Market Research Inc.1267 Willis St, Ste 200 Redding, California, 96001, U.S.USA: +1-646-781-8004Europe : +44-203-868-8738APAC: +91 744-7780008Email- sales@meticulousresearch.com Visit Our Website: https://www.meticulousresearch.com/Connect with us on LinkedIn- https://www.linkedin.com/company/meticulous-researchContent Source: https://www.meticulousresearch.com/pressrelease/294/artificial-intelligence-in-manufacturing-market-2028

See the article here:
Artificial Intelligence (AI) in Manufacturing Market Worth $13.96 billion by 2028 Exclusive Report by Meticulous Research - Yahoo Finance

Artificial Intelligence, Automation and The Future of Corporate Finance – PRNewswire

NASHVILLE, Tenn., Nov. 1, 2021 /PRNewswire/ --Algorithms rule the world or, at least, the world is headed that way. How can you prepare your company and its financial underpinnings not only to survive but also thrive under this new big data paradigm? In his new book, Deep Finance: Corporate Finance in the Information Age, author Glenn Hopper provides a clear guide for finance professionals and non-technologists who aspire to digitally transform their companies into modern, data-driven organizations streamlined for success and profitability.

Hopper, who comes to this subject armed with a unique background in finance and technology, contends that the finance department is perfectly placed to lead the digital revolution bringing companies of all sizes into a new era of efficiency while future-proofing the role of chief financial officer.

Deep Financeis written for a wide audience, ranging from those who don't know AI from A/R to those who are already working with data to drive business decisions. The book illuminates the path toward digital transformation with instructions on how finance professionals can elevate their leadership and become champions for data science.

InDeep Finance, readers will:

"In this Age of AI, every function in every company has to go through its own digital transformation to enable their organizations to succeed.Glenn Hopper provides an essential roadmap to accounting and finance executives on how to embrace analytics and AI as core tools for modern finance. This book should be a required reading for every general manager."

Karim R. Lakhani | Co-Author of Competing in the Age of AICo-Director of Laboratory for Innovation Science at Harvard and Co-Chair of Harvard Business Analytics Program

A former Navy journalist, filmmaker, and business founder, Hopper has spent the past two decades helping startups transition into going concerns, operate at scale, and prepare for funding and/or acquisition. He is passionate about transforming the role of CFO from a historical reporter and bookkeeper to a forward-looking strategist who is integral to a company's future. He has served as a finance leader in a variety of industries, including telecommunications, retail, Internet, and legal technology. He has a master's degree in finance with a graduate certificate in business analytics from Harvard University, and an MBA from Regis University.

Deep Financeis distributed by Simon & Schuster and will be available November 16, 2021, in eBook and print versions at Amazon, Barnes and Noble, and other online booksellers.

Contact:Glenn Hopper 615.756.7354[emailprotected]

SOURCE Glenn Hopper

More:
Artificial Intelligence, Automation and The Future of Corporate Finance - PRNewswire