Daily Archives: September 27, 2021

We are sleepwalking into AI-augmented work – VentureBeat

Posted: September 27, 2021 at 5:36 pm

The Transform Technology Summits start October 13th with Low-Code/No Code: Enabling Enterprise Agility. Register now!

A recent New York Times article concludes that new AI-powered automation tools such as Codex for software developers will not eliminate jobs but simply be a welcome aid to augment programmer productivity. This is consistent with the argument were increasingly hearing that people and AI have different strengths and there will be appropriate roles for each.

As discussed in a Harvard Business Review story: AI-based machines are fast, more accurate, and consistently rational, but they arent intuitive, emotional, or culturally sensitive. The belief is that AI plus humans is something of a centaur, greater than either one operating alone.

This idea of humans plus AI producing better outcomes has become a tenant of faith in technology. Everyone talks about humans being freed up to perform higher-level functions, but no one seems to know just what those high-level functions are, how they translate into real work and jobs, or the number of people needed to perform them.

A corollary of this augmented-workforce narrative is that not only will AI-augmented work enable people to pursue a higher level of abstract thinking, it will according to some also lift all of society to a higher standard of living. This is certainly an optimistic vision, and we can hope for that. However, this could also be a story imbued with magical thinking, with the true end-game being fully automated work.

Dont get me wrong; there is some evidence to support the view that AI will help us work rather than take our jobs. For example, AI lab DeepMind is designing new chess systems for the two intelligences to work in tandem with humans rather than opposed to them.

And Kai-Fu Lee, the Oracle of AI, also buys into this promise. In his new book, AI 2041: Ten Visions for our Future, he argues that repetitive tasks from stacking shelves to crunching data will be done by machines, freeing workers for more creative tasks. Forrester Research has likewise articulated that AI deployment enables people to better use their creative skills.

But, of course, some people are more creative than others, meaning that not everyone would benefit from AI-augmented work to the same degree. Which in turn reinforces a concern that AI-fueled automation, even in its augmented work capacity, could widen already existing income disparities.

One problem with the AI-augmented workforce promise is that it tells us AI will only take on the repetitive work we dont want to do. But not all work being outsourced to AI is routine or boring.

Look no further than the role of the semiconductor chip architect. This is a highly sophisticated profession, an advanced application of electrical engineering in arguably one of the most complex industries. If ever there was a job that might be thought of as immune from AI, this would have been a strong candidate. Yet recent advances from Google and Synopsys (among others using reinforcement learning neural network software) have shown the ability to do in hours what often required a team of engineers months to achieve.

One ever-faithful tech watcher still argued that the algorithms will optimize and accelerate time-intensive parts of the design process so that designers can focus on making crucial calls that require higher-level decision making.

More than likely, the current perception of work augmented by AI is a reflection on the current state of the technology and not an accurate view of the future when automation will be far more advanced. We first saw the potential of neural networks a decade ago, for example, and it took several years until that technology was developed to the point where it had practical advantages for consumers and business. Fueled in part by the pandemic, AI tech is now being widely implemented. Even massage therapists should take note, as a robot masseuse can now deliver a deep tissue massage. Yet, these are still early days for AI.

Caption: EMMA from AiTreat, a robot that uses artificial intelligence to deliver massages. Source: CNN

AI advances are being led by improvements in both hardware and software. The hardware side is driven by Moores Law, the idea that semiconductors improve by roughly 2x the number of transistors producing roughly equivalent performance and power efficiency gains every couple of years (and similarly drive down the costs of computing). This principle has been credited with all manner of electronic advances over the last several decades. As noted in a recent IEEE Spectrum article: The impact of Moores Law on modern life cant be overstated. We cant take a plane ride, make a call, or even turn on our dishwashers without encountering its effects. Without it, we would not have found the Higgs boson or created theInternet. Or have a supercomputer in your purse or pocket.

There are reasons to think that Moores-Law driven improvements in computing are nearing an end. But advanced engineering, ranging from chiplets to 3D chip packaging promise to keep the gains coming, at least for a while. These and other semiconductor design improvements have led one chip manufacturer to promise a 1000x performance improvement by 2025!

The expected improvements in AI software may be equally impressive. GPT-3, the third iteration of Generative Pre-trained Transformer from OpenAI, is a neural networkmodel consisting of 175 billion parameters. The system has proven capable of generatingcoherent prosefrom a text prompt. This is what it was designed to do, but it turns out that it can also generate other forms of text as well, including computer code and can also generate images. Moreover, while the belief is that AI will help people to be more creative, it could be that it is already capable of creativity on its own.

At its launch in May 2020, GPT-3 was the largest neural network ever introduced, and it remains among the largestdenseneural nets, exceeded only by Wu Dao 2.0 in China. (At 1.75 trillion parameters, Wu Dao 2.0 is another GPT-like language model and probably the most powerful neural network yet created.)

Some expectations are for GPT-4 to also grow and contain up to a trillion parameters. However, OpenAI CEO Sam Altman has said that it will not be larger than GPT-3 but will be far more efficient through enhanced data algorithms and fine tuning. Altman also alluded to a future GPT-5. The point being that neural networks have a long way to run in size and sophistication. We are indeed in the midst of an age of AI acceleration.

In the new book,Rule of the Robots: How Artificial Intelligence Will Transform Everything, author Martin Ford notes that nearly every technology startup is now, to some degree, investing in AI, and companies large and small in other industries are beginning to deploy the technology. The pace of innovation will only continue to accelerate as capital continues to pour into AI development. Clearly, whatever we are seeing now in the way of AI-powered automation, including the belief that AI will help us work rather than take our jobs, is but an early stage for whatever is still to come. As for what is coming, that remains the realm of speculative fiction.

In Burn In: A Novel of the Real Robotic Revolution, a Yale-educated lawyer is among those impacted when his firm replaced 80% of the legal staff with machine learning software. This could happen in the near future. The remaining 20% were indeed augmented by the AI, but the 80% had to find other work. In his case, he winds up doing gig work as an online personal assistant to the wealthy. Currently, startup company Yohana is working to realize a variation of this vision. The company is initially offering a blend of human and AI services, starting with a living, breathing assistantthat draws on data to tackle the to-do lists of subscribers. It will be telling to see if these assistants will be like the secretaries of yore, but wielding AI, or if they will be displaced cognitive workers.

The AI-driven transition to a largely automated world will take time, perhaps a few decades. This will bring many changes, with some being highly disruptive. Adjustments will not be easy. It is tempting to think that ultimately this will enrich the quality of human life. After all, as Aristotle said: When looms weave by themselves, mans slavery will end.But embracing the AI augmented work concept as currently articulated could blind us to the potential risks of job loss.Kate Crawford, a scholar focused on the social and political implications of technology, believes AI is the most profound story of our time and a lot of people are sleepwalking into it.

We all need to have a clear-eyed understanding of the growing potential for disruption and to prepare as best we can, largely by acquiring those skills most likely to be needed in the coming era. Companies need to do their part in providing skills training, and retraining will increasingly need to be a near continuous process as the pace of technology change accelerates. Government needs to develop public policies that direct the market forces driving automation towards positive outcomes for all, even while preparing for a growing social safety net that could include universal basic income.

Gary Grossman is the Senior VP of Technology Practice atEdelmanand Global Lead of the Edelman AI Center of Excellence.

Excerpt from:

We are sleepwalking into AI-augmented work - VentureBeat

Posted in Ai | Comments Off on We are sleepwalking into AI-augmented work – VentureBeat

Feeding The Cognitive Enterprise: Nestl Pushes AI, Predictive Maintenance and Robotics – IoT World Today

Posted: at 5:36 pm

Carolina Pinart, the lead for Nestls AI program, said intelligence was driving progress toward the corporates three key strategic objectives.

Day One of Informa Techs AI Summit in London saw Nestls global product director for new generation technologies provide an in-depth look at AIs potential to transform the multinational corporation.

Carolina Pinart, the lead for Nestls AI program, said intelligence was driving progress toward the corporates three key strategic objectives: enhanced efficiency, digitized operations and sustainability. The latter target is with a view of achieving Nestles goal of net zero carbon emissions by 2050.

If we look into Nestls progress last year, we made really solid progress on our structural savings program across all areas of manufacturing, procurement and administration, with massive opportunities for automating and AI, said Pinart.

Pinart stressed that all three initiatives had enabled Nestl to leverage AI, predictive maintenance and robotics to support factory automation and customization on the assembly line.

Nestl is also striving to expand the flow, accessibility and utility of real-time data as it is collected from operational technology networks, both in supply chain management and procurement operations.

These efforts support our drive to enhance consumer and customer centricity, which is a paramount objective for our company, as well as manufacturing, flexibility, agility and also transparency and traceability across the supply chain, Pinart explained.

Since 2019, Nestl has sought to transform its business into a cognitive enterprise, with unbiased machine learning tools that can be extensively used to help automate, respond, react and decide business outcomes.

Through machine learning, the mission statement set out by the company underscores its commitment to derive more value for its employees and customers, as well as society and its shareholders.

By 2025, Nestl aims to be fully data-driven and cognitive in its approach. It is currently working to define corporate directives for AI as part of its global program, which also covers topics such as ethics, organization, technology, education and communication.

Without the global program, we might get there eventually but certainly not at the speed that we want.

The global program runs together with multi-year strategic initiatives to automate operations and improve the experience of its workforce, through data, analytics and other technologies.

These programs leverage AI and machine learning where appropriate as a toolbox to solve business challenges, Pinart explained.

Pinart believes deployment of AI in Nestles global supply chain must start with a strategic analysis of the intended outcomes, followed by work to identify real use-cases where AI can be prioritized, deployed and scaled.

Also vital for Nestles implementation strategy was identifying the foundational data for AI to be trained on as well as other tech capabilities.

Pinart said: We are working on creating data assets and having the right platforms in place to really support the strategy and use-cases.

[Another] question is around organization and talent that varies depending on what were trying to do with AI. We dont have a single operating model but the question is, what is the right operating model to run all of this.

Finally governance, AI raises concerns on people. So both consumers and business leaders are employees as well, right? So how do we address those concerns.

View post:

Feeding The Cognitive Enterprise: Nestl Pushes AI, Predictive Maintenance and Robotics - IoT World Today

Posted in Ai | Comments Off on Feeding The Cognitive Enterprise: Nestl Pushes AI, Predictive Maintenance and Robotics – IoT World Today

AI Risk Management Framework Is Among Emerging Federal Initiatives on AI – JD Supra

Posted: at 5:36 pm

Artificial intelligence (AI) has drawn significant policy interest for quite some time now, and a federal approach is taking shape. A recent flurry of federal activity on the AI front has emanated in large part from the U.S. Department of Commerce (Commerce) including a move towards development of a risk management framework. The framework in particular may greatly influence how companies and organizations approach AI-related risks, including avoiding bias and promoting accuracy, privacy, and security.

Developing an AI Risk Management Framework

In the National Defense Authorization Act for Fiscal Year 2021 (2021 NDAA), Congress directed the National Institute of Standards and Technology (NIST) which falls under Commerce to develop a voluntary risk management framework for trustworthy [AI] systems. Following this directive, in late July, NIST issued a Request for Information (RFI) seeking input to help inform the development of what it refers to as the AI Risk Management Framework (AI RMF).

As NIST explained, the AI RMF is intended to help designers, developers, users, and evaluators of AI systems better manage risks across the AI lifecycle and aims to foster the development of innovative approaches to address characteristics of trustworthiness, including accuracy and mitigation of harmful bias. According to NIST, the development of the AI RMF will involve several iterations to encourage robust and continuing engagement and collaboration with interested stakeholders and will include open, public workshops, along with other forms of outreach and feedback.

In the RFI, NIST proposed that the AI RMF should feature eight key attributes. Specifically, NIST explained that the AI RMF should:

This directive as well as NISTs RFI outlining the initial attributes and goals of the AI RMF aligns with NISTs extensive work on similar frameworks for cybersecurity and privacy.

The RFI is the beginning of the collaborative development process at NIST with respect to the AI RMF. In the immediate near term, NIST is hosting a public workshop on October 19-21, 2021 to share feedback received in response to the RFI and discuss plans and goals for the AI RMF.

The National Artificial Intelligence Advisory Committee

On September 8, Secretary of Commerce Gina Raimondo announced that Commerce would be establishing a National Artificial Intelligence Advisory Committee (NAIAC) pursuant to a separate directive from the 2021 NDAA. The NAIAC, which will also include a subcommittee focused on the use of AI in law enforcement, will provide recommendations to the President on a broad range of AI-related topics, including the United States competitiveness in the AI space, the state of science related to AI, AI workforce issues, and opportunities for international AI cooperation.

NIST has already issued a Federal Register notice formally calling for nominations to the NAIAC. As this notice explains, the NAIAC will consist of at least nine members representing academia, industry, nonprofits, and federal laboratories who will be tasked with submitting a report to the President and Congress within one year, and then once every three years thereafter. Nominations for the NAIAC are due October 25, 2021.

Looking Ahead

These recent developments coming out of Commerce are by no means the only federal AI actions that have been taken since President Biden assumed office. For example, the Biden Administration launched AI.gov in May in order to introduce more Americans to the governments activities in the AI space. Similarly, in June, the White House Office of Science and Technology Policy and the National Science Foundation established a National Artificial Intelligence Research Resource Task Force that will help to expand access to educational tools and resources in order to spur AI innovation.

Given the responsibilities that have been assigned to it and the actions that it has already taken, Commerce and NIST in particular appears poised to be central to the federal governments approach. As such, stakeholders should consider engaging with NIST in its AI workstreams, including through participation in the upcoming workshop.

Link:

AI Risk Management Framework Is Among Emerging Federal Initiatives on AI - JD Supra

Posted in Ai | Comments Off on AI Risk Management Framework Is Among Emerging Federal Initiatives on AI – JD Supra

Vendor Management Is The New Customer Management, And AI Is Transforming The Sector Already – Forbes

Posted: at 5:36 pm

Woman at work in bright, open office workplace

Technology, and especially artificial intelligence (AI), is changing the way business gets done. In virtually every market worldwide, automation, business intelligence, and new systems are synthesizing operations, enabling teams to work smarter. While it may feel that new tech is being plugged in at an impressive pace, there are some late adopters, one of which is vendor relationship management.

Third-party vendors are essential for companies of any size. Manufacturers, suppliers and service providers are used for everything from product development to cybersecurity. At an enterprise level, the number of vendors can reach the tens of thousands. As is often the case in systems that have been around for a decade or more, interoperability is virtually non-existent, data is hard to find and efficiency is a pipe dream.

Sourcing and vendor management are more complex than ever, as companies move into emerging fields and leverage third-party suppliers to provide what cannot be internally produced. Vendors play long-term and essential roles in businesses. Many procurement departments have long held to an us versus them mentality, but as the number of necessary suppliers grows, so does the need for collaboration.

The type of vendor management in most business contexts today is sufficient for managing a handful of vendors with some success. But the demands of this are changing, and in larger companies, an entire vendor ecosystem requires more sophisticated management. In this context, the sheer volume of contacts, contracts, products, tasks and more far exceeds existing practices. Fortunately, data is what AI handles best, and new technology is being developed to get vendor relationship management up to speed.

CIOs, chief procurement leaders and financial officers regularly report on how difficult it is to understand the vendor relationship from a contract perspective, to understand how much tech a company actually owned versus used, to keep track of all of the support personnel and even to find the vendor data that lived in their ecosystem. Most of this has been and is still managed through email or on spreadsheets.

For years, client and customer communication have been managed via client relationship management (CRM) software, while vendor relationships have been left to outmoded methods. Solutions have been patchwork, piecemeal and largely ineffective. Even at the highest enterprise level, in companies with extensive technology, a problem persists: it is difficult to manage vendor relationships and assets in one place.

Brandon Card

A first reason is that there are disparate data sets, and enterprise businesses are notoriously siloed. Vendor information may reside in the legal department, the IT department and be tied to product records. A second reason this challenge has not been solved sooner is that every company hires vendors on different terms, and often each vendor on different terms. The compound challenges of these two issues alone have kept companies from making changes, but now that AI is more accessible, a smart solution is possible.

Industry leader Brandon Card is a well-known expert who has worked in development and sales for many companies, including IBM and Microsoft. He launched an application called Terzo, the first Vendor Relationship Management (VRM) system. The idea is to create a platform where vendors can now be managed in the way contacts have been managed by marketers.

CRMs already pull data from different sources, integrate with existing systems, have visibility across departments and centralize and visualize data in a meaningful way. CRMs have a long history of success in managing contexts, and the data obtained in them can trigger alerts, renewal notices and even alert marketers to churn risk. No successful company manages client relationships in any other way, and AI has already amplified the capacity of CRMs, making it easier to filter, sort and interpret data.

The same technology is at work in the emerging area of VRMs. Vendor management can be handled as data from various sources is added to the system, including contract dates, contract details, vendor supplier contact information, vendor communication records and more. Any procurement department in any company will attest to how transformational this process will be, alleviating the manual burden of record-keeping and retrieval, and leveraging the power of AI to manage these crucial relationships more effectively.

Enterprise resource planning (ERP) systems automate business processes, and are often the primary way businesses manage day-to-day operations. While these systems accomplish their purpose, they continue to be an inadequate way to manage vendors. Card explains, Across all of the systems in place today, customers rely upon ERP systems to manage all of their suppliers' data. ERP systems are a black box and cant solve vendor challenges for today, which is why people resort to spreadsheets. The problem is that, ten years ago, these companies were doing everything in-house with their own data centers.

As company offerings get more niched, and products become more technically complex, in-house solutions will not suffice. Businesses have grown and strategic vendors deliver cost-saving solutions. In most companies, entire departments cant exist without third-party vendors. AI provides the first glimmer of hope for a digital platform capable of helping businesses manage all of vendor relationships and the stakeholders that manage those relationships. These transactional relationships have now become critical partnerships, and because of AI, the technology has finally caught up to the need.

Vendor relationships directly impact a companys success. These critical partnerships require collaboration, and a vendor hub built on AI may provide that solution. It is a new way of working that complements existing systems. Easily integrated with a companys tech stack, this solution enriches data and user experiences and uncovers insights faster.

Contract data provides the single source of truth, from which AI can then extract all products, services and legal terms. Companies understand what they own and how much theyre paying for it. AI speeds up time to visibility and gets insights five to six times faster.

This is a new way of treating strategic suppliers as business partners, and empowering stakeholders with tools to more effectively connect, communicate, monitor and achieve results. As technology continues to improve and upgrade business operations, solutions like these make it possible for companies to do more.

Go here to see the original:

Vendor Management Is The New Customer Management, And AI Is Transforming The Sector Already - Forbes

Posted in Ai | Comments Off on Vendor Management Is The New Customer Management, And AI Is Transforming The Sector Already – Forbes

Clearview AI drops subpoenas of its critics – POLITICO

Posted: at 5:36 pm

A big about-face: The Clearview decision represents a major win for civil society groups and researchers focused on bringing transparency and accountability to dealings between the government and the private sector. Those groups had argued that the legal mandates were a tactic to dissuade advocacy groups from further investigating the company or others like it, and could undermine press freedom by targeting journalists and their sources.

A controversial technology: Facial recognition tools have long been under scrutiny by privacy and civil rights advocates whove raised alarm about biases with the technology generally and the disproportionate harm it can cause to marginalized communities.

After 2019 research from Open The Government shed light on police use of this surveillance including Clearviews tech and later prompted reporting on the startup, the software firm now finds itself tangled in multidistrict consumer privacy litigation in Illinois.

The debate over the subpoenas: The subpoenas included demands for communications with journalists about Clearview and its leaders, as well as information theyd uncovered about the company and its founders in public records requests.

Clearview's attorney Andrew J. Lichtman said previously that the subpoenas were served as part of its defense in the litigation in Illinois. But that case does not appear to involve Open The Government or the other parties subpoenaed, and Clearview would not explain how the groups' correspondence with journalists would be in any way relevant.

See the article here:

Clearview AI drops subpoenas of its critics - POLITICO

Posted in Ai | Comments Off on Clearview AI drops subpoenas of its critics – POLITICO

Inspur Comes Out on Top with Superior AI Performance in MLPerf Inference V1.1 – Maryville Daily Times

Posted: at 5:36 pm

SAN JOSE, Calif.--(BUSINESS WIRE)--Sep 27, 2021--

Recently, MLCommons, a well-known open engineering consortium, released the results of MLPerf Inference V1.1, the leading AI benchmark suite. In the very competitive Closed Division, Inspur ranked first in 15 out of 30 tasks, making it the most successful vendor at the event.

Inspur Results in MLPerf TM Inference V1.1

Vendor

Division

System

Model

Accuracy

Score

Units

Inspur

Data CenterClosed

NF5688M6

3D-UNet

Offline, 99%

498.03

Samples/s

NF5688M6

3D-UNet

Offline, 99.9%

498.03

Samples/s

NF5488A5

DLRM

Offline, 99%

2607910

Samples/s

NF5688M6

DLRM

Server, 99%

2608410

Queries/s

NF5488A5

DLRM

Offline, 99.9%

2607910

Samples/s

NF5688M6

DLRM

Server, 99.9%

2608410

Queries/s

EdgeClosed

NE5260M5

3D-UNet

Offline, 99%

93.49

Samples/s

NE5260M5

3D-UNet

Offline, 99.9%

93.49

Samples/s

NE5260M5

Bert

Offline, 99%

5914.13

Samples/s

NF5688M6

Bert

SingleStream, 99%

1.54

Latency (ms)

NF5688M6

ResNet50

SingleStream, 99%

0.43

Latency (ms)

NE5260M5

RNNT

Offline, 99%

24446.9

Samples/s

NF5688M6

RNNT

SingleStream, 99%

18.5

Latency (ms)

NF5688M6

SSD-ResNet34

SingleStream, 99%

1.67

Latency (ms)

NF5488A5

SSD-MobileNet

SingleStream, 99%

0.25

Latency (ms)

Developed by Turing Award winner David Patterson and leading academic institutions, MLPerf is the leading industry benchmark for AI performance. Founded in 2020 and based on MLPerf benchmarks, MLCommons is an open non-profit engineering consortium dedicated to advancing standards and metrics for machine learning and AI performance. Inspur is a founding member of MLCommons, along with over 50 other leading organizations and companies from across the AI landscape.

In the MLPerf Inference V1.1 benchmark test, the Closed Division included two categories Data Center (16 tasks) and Edge (14 tasks). Under the Data Center category, six models were covered, including Image Classification (ResNet50), Medical Image Segmentation (3D-UNet), Object Detection (SSD-ResNet34), Speech Recognition (RNN-T), Natural Language Processing (BERT), and Recommendation (DLRM). A high accuracy mode (99.9%) was set for BERT, DLRM and 3D-UNET. Every model task evaluated the performance in both Server and Offline scenarios with the exception 3D-UNET, which was only evaluated in the Offline scenario. For the Edge category, the Recommendation (DLRM) model was removed and the Object Detection (SSD-MobileNet) model was added. A high accuracy mode (99.9%) was set for 3D-UNET. All models were tested for both Offline and Single Stream inference.

In the extremely competitive Closed Division, in which mainstream vendors were competing, the use of the same models and optimizers was required by all participants. Doing so provided the ability to easily evaluate and compare AI computing system performance among various vendors. Nineteen vendors including Nvidia, Intel, Inspur, Qualcomm, Alibaba, Dell, and HPE participated in the Closed Division. A total of 1,130 results were submitted, including 710 for the Data Center category, and 420 for the Edge category.

Full-Stack AI Capabilities Ramp up Performance

Inspur achieved excellent results in this MLPerf competition with its three AI servers NF5488A5, NF5688M6, and NE5260M5.

Inspur ranked first in 15 tasks covering all AI models, including Medical Image Recognition, Natural Language Processing, Image Classification, Speech Recognition, Recommendation, as well as Object Detection (SSD-ResNet34 and SSD-MobileNet). The results showcase that from Cloud to Edge, Inspur is ahead of the Industry in nearly all aspects. Inspur was able to make huge strides in performance in various tasks under the Data Center category compared to previous MLPerf events despite no changes to its server configuration. Its model performance results in Image Classification (ResNet50) and Speech Recognition (RNN-T) increased by 4.75% and 3.83% compared to the V1.0 competition just six months ago.

The outstanding performance of Inspur's AI servers in the MLPerf Benchmark Test can be credited to Inspur's exceptional system design and full-stack optimization in AI computing systems. Through precise calibration and optimization, CPU and GPU performance as well as the data communication between CPUs and GPUs were able to reach the highest levels for AI inference. Additionally, by enhancing the round-robin scheduling for multiple GPUs based on GPU topology, the performance of a single GPU or multiple GPUs can be increased nearly linearly.

Inspur NF5488A5 was the only AI server in this MLPerf competition to support eight 500W A100 GPUs with liquid cooling technology, which significantly boosted AI computing performance. Among mainstream high-end AI servers with 8 NVIDIA A100 SXM4 GPUs, Inspur's servers came out on top in all 16 tasks in the Closed Division under the Data Center category.

As a leading AI computing company, Inspur is committed to the R&D and innovation of AI computing, including both resource-based and algorithm platforms. It also works with other leading AI enterprises to promote the industrialization of AI and the development of AI-driven industries through its Meta-Brain technology ecosystem.

To view the complete results of MLPerf Inference v1.1, please visit:

Inspur Information is a leading provider of data center infrastructure, cloud computing, and AI solutions. It is the worlds 2 nd largest server manufacturer. Through engineering and innovation, Inspur delivers cutting-edge computing hardware design and extensive product offerings to address important technology arenas like open computing, cloud data center, AI, and deep learning. Performance-optimized and purpose-built, our world-class solutions empower customers to tackle specific workloads and real-world challenges. To learn more, please go to https://www.inspursystems.com/.

KEYWORD: UNITED STATES NORTH AMERICA CALIFORNIA

Read the original:

Inspur Comes Out on Top with Superior AI Performance in MLPerf Inference V1.1 - Maryville Daily Times

Posted in Ai | Comments Off on Inspur Comes Out on Top with Superior AI Performance in MLPerf Inference V1.1 – Maryville Daily Times

Coevolving informatics, AI and brain-computer interfacing – Open Access Government

Posted: at 5:36 pm

Do you ever think about spatial barriers?

In 1997, a chess computer named Deep Blue became the first in the world to beat a reigning chess champion. The machine, complex and high-speed, went toe-to-toe with Garry Kasparov. The AI processed information at great speed, overcoming spatial barriers with processing power.

Now, an AI like Deep Blue is less startling to humankind than it was initially. But coevolving informatics continue to reshape the frontiers of what information can do. Associate Professor Girard examines how information can re-shape our understanding of gender equality, or evaluate the resources available to the people of Afghanistan, especially when it comes to the urban-rural divide.

When it comes back down to the kind of information a brain or AI can process, there are new ways for coevolving informatics to grow. For instance, because electromagnetic signals lack mass and travel at the speed of light, because they dont have the pathway and energy constraints of other types of information. Humans are a living testament to the power of evolved information use, as we have curated a cognitive capacity beyond that of all other living creatures. If we put those understandings together, where can we go?

According to research by Florida International University, the future could mean instant interfacing of brains with cloud-based AI enabling a kind of global access that is not limited to any individual capacity for memory. This would give an intensive advantage to groups with access to that globalised information.

But how could this future work? What kind of problems and considerations would arise?

To find out more about the imminent potential of cloud-based AI and the human brain together, read more about the work of Florida International University here.

Editor's Recommended Articles

Read the rest here:

Coevolving informatics, AI and brain-computer interfacing - Open Access Government

Posted in Ai | Comments Off on Coevolving informatics, AI and brain-computer interfacing – Open Access Government

Lone Star College-University Park Taps AI to Inspire Curiosity in Online Discussions – inForney.com

Posted: at 5:36 pm

HOUSTON, Sep. 27, 2021 /PRNewswire-PRWeb/ --Lone Star College-University Park today announced a new initiative that will enable faculty to cultivate more enriching, engaging class discussions utilizing artificial intelligence (AI). In partnership with the inquiry-based discussion platform Packback, LSC-University Park will expand faculty and student access to a platform shown to promote critical thinking, build a sense of community in class, and help improve student success.

"The power of a good class discussion is that it empowers students to engage with the curriculum beyond what's in their textbooksand help them to see that they're not alone in the questions they have or the observations they make," said Dr. Kathy Cecil-Sanchez, vice president of instruction at LSC-University Park. "That's the kind of critical thinking that Packback's platform helps to inspire, while also allowing faculty to foster deeper and more substantive engagement with their students."

Backed by entrepreneur Mark Cuban, Packback's unique platform assigns a "curiosity score" to students' discussion posts based on a range of factors, including the open-endedness of questions and the way that students reference sources. The tool also provides feedback on grammar, spelling, and offers students guidance on academic research processes like citing sources and paraphrasing. All these elements of the Packback platform are designed to spark students' intrinsic motivation, as well as to help instructors focus on providing more substantive guidance and support to students, rather than spending time on more repetitive tasks like correcting grammar.

"In my classes, Packback has been a platform that provides equity of access to the material while also pushing students to ensure that their analysis is based on facts," said Cassandra Khatri, LSC-University Park professor of political science. "Packback requires students cite the work they provide in their responses, which gets at the larger problem in political science of the gray area between analysis and takes. I really enjoy seeing my students practicing critical thinking skills as they deal with the tough issues of the day."

"I appreciate Packback's AI guidance on the writing and research, so I can focus more on the heart of the discussion topics that arise, added Jennifer Ross, LSC-University Park professor of political science. "Through Packback discussions I have seen students go from being passive learners who just accept what they read and hear, to actively taking an interest in current events and thinking about them critically."

The partnership comes in the wake of a recent research finding that inquiry-based discussion tools can spark deeper engagement and also boost academic performance. A study conducted by Packback in 2019-2020 with 10 higher education institutions, including LSC-University Park, found that students in classes that use Packback received more A, B, or C grades and fewer D, F and W grades than the control group. In the same study, students cited sources approximately 2.5 times as often in the treatment group versus the control group.

"Our work with colleges around the country is rooted in the belief that better approaches to discussion can help students build the skills that help them succeed both in and out of the classroom," said Kasey Gandham, co-founder of Packback. "As an institution that shares that vision, LSC-University Park has been an invaluable partner over the past few years, and we look forward to continuing to learn from its creative and student-centered approach in the years to come."

Packback's proprietary AI and machine learning technology provides inquiry-based online discussion to over 3,500 instructors and over 900,000 students, who have posted 22 million questions and responses to date.

Media Contact

Ben Watsky, Packback, +1 (203) 907-8930, watsky@whiteboardadvisors.com

SOURCE Packback

Continue reading here:

Lone Star College-University Park Taps AI to Inspire Curiosity in Online Discussions - inForney.com

Posted in Ai | Comments Off on Lone Star College-University Park Taps AI to Inspire Curiosity in Online Discussions – inForney.com

Part 1: Quality Control in Delivery Only Ghost Kitchens Utilizing AI and Computer – MarketScale

Posted: at 5:36 pm

There is a need to have a safe, predictable, and stable food supply chain, Jain said. Internet of Things, artificial intelligence, and data analytics can play a big role in creating these safe, predictable, and sustainable food supply chains.

On this first part of a two-part episode of To the Edge and Beyond by Intel with Voice of B2B Daniel Litwin is Maria Meow, APAC Hospitality Vertical Marketing Manager for Intels Internet of Things Group, and Ankur Jain, Founder, and CEO of UdyogYantra Technologies, a company focused on bringing Industry 4.0 and its associated ecosystem of technologies to various industries, including the Restaurant industry.

When it comes to IoT, Meow has been at the intersection for over ten years at Intel. One thing shes noticed is that as technology has evolved, so has knowledge. In her role she is regularly relying on IoT to help improve food operations and to help ensure the trust and quality of food, which addresses a lot of consumer concerns.

With population is expected to increase globally by 10 billion in 2050, this creates a demand for creative food solutions, according to Jain. There is a need to have a safe, predictable, and stable food supply chain, Jain said. Internet of Things, artificial intelligence, and data analytics can play a big role in creating these safe, predictable, and sustainable food supply chains.

Intel technology is at the forefront of these IoT innovations, powering AI and analytics solutions like UdyogYantras to address these concerns for both consumers and restaurant brands.

The wheels of ghost kitchens were already in motion before COVID-19. But, their viability increased during the pandemic, as restaurants pivoted to meet consumer demand for BOPIS and third-party delivery. Now that things have settled a bit, restaurants are still sticking with some or all of the models developed during the pandemic.

Digital transformations have been accelerated ever since the pandemic began and have pivoted to delivery, pick up and drive through, Meow said. Digitally transformed brands will reap the benefits.

View original post here:

Part 1: Quality Control in Delivery Only Ghost Kitchens Utilizing AI and Computer - MarketScale

Posted in Ai | Comments Off on Part 1: Quality Control in Delivery Only Ghost Kitchens Utilizing AI and Computer – MarketScale

AI adoption in the ETF industry begins to grow – Financial Times

Posted: at 5:36 pm

The growing appreciation that human stockpickers struggle to outperform their benchmark indices has helped fuel a massive surge in assets held by passively managed exchange traded funds. Now some companies are hoping to show that artificial intelligence can finally give them an edge.

The technology is fast-evolving but at least two fund managers, EquBot and Qraft Technologies, running dedicated AI-powered ETFs are claiming early success, even though some of their AI models decisions might have required strong nerves to implement.

For example, the team at Qraft, which offers four AI-powered ETFs, listed on NYSE Arca, witnessed its technology build a weighting of 14.7 per cent in Tesla in its Qraft AI-Enhanced US Large Cap Momentum ETF (AMOM) in August last year, but when it rebalanced a month later on September 1 it sold it all.

The ETF began buying Tesla again in November, amassing a stake of 7.6 per cent by January this year, but in the February rebalancing it sold the entire holding once again. In each case when it sold it anticipated a sharp decline in the price of Tesla and was able to profit from the subsequent rise when it bought back in.

Alpha [excess returns above the market] is getting harder and harder to find, said Francis Oh, managing director and head of AI ETFs at Qraft, pointing out that humans can become emotionally attached to certain stocks, impeding their portfolio returns. Our model has no human bias.

Academic research has certainly shown that humans tend to be reluctant to crystallise losses, while conversely they feel driven to realise gains sometimes too early.

However, it is arguable whether the AI systems used by Qraft and EquBot can really be said to eliminate human bias because both are supported by large teams of data scientists who are constantly enhancing their models EquBot has teamed up with IBM Watson and Qraft has its own dedicated team in South Korea.

The machine only has historical data. It sees opportunities according to the rules it has been programmed for, Greg Davies, head of behavioural science at consultancy Oxford Risk, pointed out.

Chris Natividad, chief investment officer at EquBot, agreed: The system only knows what it knows, and its historical, he said, adding that in addition to humans deciding what new information the self-learning system should be given, data scientists also needed to check the outcomes so we can explain it to investors.

Both of EquBots ETFs, the AI Powered Equity ETF (AIEQ) and the AI Powered International Equity ETF (AIIQ) have outperformed their respective benchmarks since inception.

Similar successes have been notched up by Qrafts suite of AI powered vehicles. But the outperformance narrative is less clear when compared to the plain vanilla SPDR S&P 500 ETF (SPY) so far this year.

AMOM and Qrafts other funds, the Qraft AI-Enhanced Large Cap ETF (QRFT), the Qraft AI-Enhanced US High Dividend ETF (HDIV) and the Qraft AI-Enhanced US Next Value ETF delivered returns of between 15.3 per cent and 20.8 per cent during the eight months to the end of August. While respectable, none of them quite matched the 21.6 per cent return of SPY over the same period.

AIEQ also just undershot SPY, notching up a 21.3 per cent gain in the eight months to the end of August, while AIIQ delivered only 12.2 per cent.

Despite the unremarkable returns this year, Oh and Natividad remain convinced that their models have much to offer.

The velocity, variety and volume of data is exploding, said Natividad, adding that bringing in new data sources was a bit like adding more pixels to an online image. You get a clearer picture. He said asset managers and index providers were embroiled in an arms race for data.

Oh said value and momentum factors were becoming so short lived and fragmented that AI systems helped to find opportunities.

EquBot scours news, social media, industry and analyst reports and financial statements to build predictive models. It also looks at things such as job posts. Qraft also uses a variety of so-called structured and unstructured data sources to drive its models.

However, despite the promise of the technology, assets under management for the ETFs at both companies remain modest. AIEQ and AIIQ have less than $200m in assets between them, although Natividad said a partnership with HSBC which was using the two EquBot indices meant that there was $1.4bn tracking the strategies.

Qrafts ETFs have attracted less than $70m across all four vehicles, although Oh said the business model was once again focused on B2B advisory asset allocation modelling.

Rony Abboud, ETF analyst at TrackInsight, said investors probably wanted to know more. He emphasised the importance of due diligence, adding: The more data points used, the higher probability of having an error. So where do they get their data from and how accurate is it?

Despite misgivings, adoption of AI techniques in the investment world is increasing. Natural language processing is certainly a growing area, said Emerald Yau, head of equity index product management for the Apac region at FTSE Russell, which has made its first foray into NLP-powered offerings with its launch of a suite of innovation-themed indices.

However, Oxford Risks Davies warned that while algorithms were good at finding arbitrage opportunities they could not deal with ambiguity.

The problem with the investing world is the rules are not static, he said, adding that humans still retained an edge. If humans learn one thing in one context they can transfer it to another.

Visit our ETF Hub for investor news and education, market updates and analysis and easy-to-use tools to help you select the right ETFs.

See the original post here:

AI adoption in the ETF industry begins to grow - Financial Times

Posted in Ai | Comments Off on AI adoption in the ETF industry begins to grow – Financial Times