Onix To Help Organizations Uncover the Power of Machine Learning-Driven Search With Amazon Kendra – PRNewswire

LAKEWOOD, Ohio, May 14, 2020 /PRNewswire/ --Onix is proud to participate in the launch of Amazon Kendra, a highly accurate and easy to use enterprise search service powered by machine learning from Amazon Web Services (AWS).

Amazon Kendra delivers powerful natural language search capabilities to customer websites and applications so their end users can more easily find the information they need. When users ask a question, Amazon Kendra uses finely tuned machine learning algorithms to understand the context and return the most relevant results, whether that be a precise answer or an entire document.

"Search capabilities have evolved over the years. Users now expect the same experience they get from the semantic and natural language search engines and conversational interfaces they use in their personal lives," notes Onix President and CEO Tim Needles. "Powered by machine learning and natural language understanding, Amazon Kendra improves employee productivity by up to 25%. With more accurate enterprise search, Amazon Kendra opens new opportunities for keyword-based on-premises and SaaS search users to migrate to the cloud and avoid contract lock-ins."

Onix has been a leader in the enterprise search space since 2002. The company provides 1:1 consulting, planning, and deployment of search solutions for hundreds of clients with a team that includes 10 certified deployment engineers. Onix has won six prestigious awards for enterprise search and boasts a 98% Customer Satisfaction Rating.

About Onix

As a leading cloud solutions provider, Onix elevates customers with consulting services for cloud infrastructure, collaboration, devices, enterprise search and geospatial technology. Onix uses its ever-evolving expertise to achieve clients' strategic cloud computing goals.

Onix backs its strategic planning and deployment with incomparable ongoing service, training and support. It also offers its own suite of standalone products to solve specific business challenges, including OnSpend, a cloud billing and budget management software solution.

Headquartered in Lakewood, Ohio, Onix serves its customers with virtual teams in major metro areas, including Atlanta, Austin, San Francisco, Boston, Chicago and New York. Onix also has Canadian offices in Toronto, Montreal and Ottawa. Learn more at http://www.onixnet.com.

Contact: Robin Suttell Onix 216-801-4984 [emailprotected]

SOURCE Onix

Home

Read the rest here:
Onix To Help Organizations Uncover the Power of Machine Learning-Driven Search With Amazon Kendra - PRNewswire

Northern Trust rolls out machine learning tech for FX management solutions – GlobalCustodian.com

Northern Trust has deployed machine learning models within its FX currency management solutions business, designed to enable greater oversight of thoughts of daily data points.

The solution has been developed in partnership with Lumint, an outsourced FX execution services provider, and will help buy-side firms reduce risk throughout the currency management lifecycle.

The technology utilised by the Robotic Oversight System (ROSY) for Northern Trust systematically cans newly arriving, anonymised data to identify anomalies across multi-dimensional data sets. It is also built on machine learning models developed byLumintusing a cloud platform that allows for highly efficient data processing.

In a data-intensive business, ROSY acts like an additional member of the team working around the clock to find and flag anomalies. The use of machine learning to detect data outliers enables us to provide increasingly robust and intuitive solutions to enhance our oversight and risk management, which can be particularly important in volatilemarkets, saidAndyLemon, head of currency management, Northern Trust.

Northern Trust announced astrategic partnership withLumintin 2018todeliver currency management services with portfolio, share class and lookthrough hedging solutions alongside transparency and analytics tools.

Northern Trusts deployment of ROSY amplifies the scalability of its already highly automated currency hedging operation; especially for the more sophisticated products such as look-through hedging offered to its global clients, addedAlexDunegan, CEO, Lumint.

The solution is the latest rollout of machine learning technology by Northern Trust, asthe bankcontinues to leverage new technologies across its businesses.

In August last year, thecustodiandeveloped a new pricing enginewithin its securities lending businessby utilising machine learning and advanced statistical technology.By using a hybrid cloud platform for highly efficient processing of data, Northern Trust will leverage a new algorithm that identifies strategic market data points from multiple asset classes and regions to project the demand for equities in the securities lending market.

The Chicago-based global custodian is underway rolling out new capabilities using robotic processing automation (RPA) and cognitive artificial intelligence (AI) within a framework called its Fund Accounting Optimisation Lab, in a bid to reduce manual entries and repetitive tasks when producing daily net asset value (NAV) for funds.

View post:
Northern Trust rolls out machine learning tech for FX management solutions - GlobalCustodian.com

How is Walmart Express Delivery Nailing that 2-Hour Window? Machine Learning – Retail Info Systems News

Walmart provided more details on its new Express two-hour delivery service, piloted last month and on its way to nearly 2,000 stores.

As agility has become the key to success within a retail landscape extraordinarily disrupted by the spread of COVID-19, the company said it tested, released and scaled the initiative in just over two weeks.

As we continue to add new machine learning-driven capabilities like this in the future, as well as the corresponding customer experiences, well be able to iterate and scale quickly by leveraging the flexible technology platforms weve developed, Janey Whiteside, Walmart chief customer officer and Suresh Kumar, global technology officer and chief development officer, wrote in a company blog post.

The contactless delivery service employs machine learning to fulfill orders from nearly 2,000 stores, fulfilled by 74,000 personal shoppers. Developed by the companys in-house global technology team, the system accounts for such variables as order quantity, staffing levels, the types of delivery vehicles available, and estimated route length between a store and home.

See also: How the Coronavirus Will Shape Retail Over the Next 35 Years

It also pulls in weather data to account for delivery speeds, and Whiteside and Kumar said its consistently refining its estimates for future orders.

Consumers must pay an additional $10, on top of any other delivery charges, to take advantage of the service.

Separately. Walmartannounced it's paying out another $390 million in cash bonuses to its U.S. hourly associates as a way to recognize their efforts during the spread of COVID-19.

Full-time associates employed as of June 5 will receive $300 while part-time and temporary associates will receive $150, paid out on June 25. Associates in stores, clubs, supply chain and offices, drivers, and assistant managers in stores and clubs are all included.

Walmart and Sams Club associates continue to do remarkable work, and its important we reward and appreciate them, said John Furner, president and CEO of Walmart U.S., in a statement. All across the country, theyre providing Americans with the food, medicine and supplies they need, while going above and beyond the normal scope of their jobs diligently sanitizing their facilities, making customers and members feel safe and welcome, and handling difficult situations with professionalism and grace.

The retailer has committed more than $935 million in bonuses for associates so far this year.

See also: Walmart Expands No-Contact Transactions During COVID-19

See original here:
How is Walmart Express Delivery Nailing that 2-Hour Window? Machine Learning - Retail Info Systems News

Reality Of Metrics: Is Machine Learning Success Overhyped? – Analytics India Magazine

In one of the most revealing research papers written recent times, the researchers from Cornell Tech and Facebook AI quash the hype around the success of machine learning. They opine and even demonstrate that the trend appears to be overstated. In other words, the so-called cutting edge research or benchmark work perform similarly to one another even if they are a decade apart. In other words, the authors believe that metric learning algorithms have not made spectacular progress.

In this work, the authors try to demonstrate the significance of assessing algorithms more diligently and how few practices can help reflect ML success in reality.

Over the past decade, deep convolutional networks have made tremendous progress. Their application in computer vision is almost everywhere; from classification to segmentation to object detection and even generative models. But is the metric evaluation carried out to track this progress has been leakproof? Are the techniques employed werent affected by the improvement in deep learning methods?

The goal of metric learning is to map data to an embedding space, where similar data are close together, and the rest are far apart. So, the authors begin with the notion that the deep networks have had a similar effect on metric learning. And, the combination of the two is known as deep metric learning.

The authors then examined flaws in the current research papers, including the problem of unfair comparisons and the weaknesses of commonly used accuracy metrics. They then propose a training and evaluation protocol that addresses these flaws and then run experiments on a variety of loss functions.

For instance, one benchmark paper in 2017, wrote the authors, used ResNet50, and then claimed huge performance gains. But the competing methods used GoogleNet, which has significantly lower initial accuracies. Therefore, the authors conclude that much of the performance gain likely came from the choice of network architecture, and not their proposed method. Practices such as these can put ML on headlines, but when we look at how much of these state-of-the-art models are really deployed, the reality is not that impressive.

The authors underline the importance of keeping the parameters constant if one has to prove that a certain new algorithm outperforms its contemporaries.

To carry out the evaluations, the authors introduce settings that cover the following:

As shown in the above plot, the trends, in reality, arent that far from the previous related works and this indicates that those who claim a dramatic improvement might not have been fair in their evaluation.

If a paper attempts to explain the performance gains of its proposed method, and it turns out that those performance gains are non-existent, then their explanation must be invalid as well.

The results show that when hyperparameters are properly tuned via cross-validation, most methods perform similarly to one another. This work, believe the authors, will lead to more investigation into the relationship between hyperparameters and datasets, and the factors related to particular dataset/architecture combinations.

According to the authors, this work exposes the following:

The authors conclude that if proper machine learning practices are followed, then the results of metric learning papers will better reflect reality, and can lead to better works in most impactful domains like self-supervised learning.

comments

More here:
Reality Of Metrics: Is Machine Learning Success Overhyped? - Analytics India Magazine

Federated Learning Fuses AI and Privacy and It Could Transform Healthcare – Built In

Its an understatement to say that doctors are swamped right now. At the beginning of April, coronavirus patients had filled New York emergency rooms so thoroughly that doctors across specialties,including dermatologists and orthopedists, had to help out.

Short-term, doctors need reliable, proven technology, like N95 masks. Longer-term, though, machine learning algorithms could help doctors treat patients. These algorithms can function as hyper-specialized doctors assistants, performing key technical tasks like scanning an MRI for signs of brain cancer, or flagging pathology slides that show breast cancer has metastasized to the lymph nodes.

One day, an algorithm could check CT scans for the lung lesions and abnormalities that indicate coronavirus.

Thats a model that could be trained, Mona G.Flores, MD, global head of medical AIat NVIDIA, told Built In.

At least, it could be trained in theory. Training an algorithm fit for a clinical setting, though, requires a large, diverse dataset. Thats hard to achieve in practice, especially when it comes to medical imaging. In the U.S.,HIPAA regulations make it very difficult for hospitals to share patient scans, even anonymized ones; privacy is a top priority at medical institutions.

More on AI and PrivacyDifferential Privacy Injects Noise Into Data Sets. Heres How It Works.

Thats not to say trained algorithms havent made it into clinical settings. A handful have passed muster with the U.S. Food and Drug Administration, according to Dr. Spyridon Bakas, a professor at University of Pennsylvanias Center for Biomedical Imaging Computing and Analytics.

In radiology, for instance, algorithms help some doctors track tumor size and progression, along with things that cannot be seen with the naked eye, Dr. Bakas told Built In like where the tumor will recur, and when.

If algorithms could train on data without puncturing its HIPAA-mandated privacy, though, machine learning could have a much bigger impact on healthcare.

And thats actually possible, thanks to a new algorithm training technique: federated learning.

Federated learning is a way of training machine learning algorithms on private, fragmented data, stored on a variety of servers and devices. Instead of pooling their data, participants all train the same algorithm on their separate data. Then they pool their trained algorithm parameters not their data on a central server, which aggregates all their contributions into a new, composite algorithm. When these steps are repeated, models across institutions converge.

Federated learning is a way of training machine learning algorithms on private, fragmented data, stored on a variety of servers and devices. Instead of pooling their data, participating institutions all train the same algorithm on their in-house, proprietary data. Then they pool their trained algorithm parameters not their data on a central server, which aggregates all their contributions into a new, composite algorithm. This composite gets shipped back to each participating institution for more training, and then shipped back to the central server for more aggregation.

Eventually, all the individual institutions algorithms converge on an optimal, trained algorithm, more generally applicable than any one institutions would have been and nearly identical to the model that would have arisen from training the algorithm on pooled data.

In December of 2019, at a radiology conference in Chicago, NVIDIA unveiled a new feature for Clara SDK. This software development kit, created expressly for the healthcare field, helps medical institutions make and deploy machine learning models with a set of tools and libraries and examples, Dr. Flores said.

The new tool was Clara Federated Learning infrastructure that allowed medical institutions to collaborate on machine learning projects without sharing patient data.

NVIDIAs not the only tech company embracing federated learning. Another medical AI company, Owkin, has rolled out a software stack for federated learning called Owkin Connect, which integrates with NVIDIAs Clara. Meanwhile, at least two general-purpose federated learning frameworks have rolled out recently, too: Googles TensorFlow Federated and the open-source PySyft.

The concept of federated learning, though, dates back to years earlier. Like many innovations, it was born at Google.

In 2017, Google researchers published a paper on a new technique they hoped could improve search suggestions on Gboard, the digital keyboard on Android phones. It was the first paper on federated learning.

In a blog post, Google AI research scientists Brendan McMahan and Daniel Ramage explained the very first federated learning use case like this:

When Gboard shows a suggested query, your phone locally stores information about the current context and whether you clicked the suggestion. Federated Learning processes that history on-device to suggest improvements to the next iteration of Gboards query suggestion model.

In other words, by blending edge computing and machine learning, federated learning offered a way to constantly improve the global query suggestion model without tracking users every move in a central database. In other words, it allowed Google to streamline its data collection processan essential given the Android OS more than 2 billion active users.

Thats just one of many potential applications, though. Dr. Bakas saw potential applications in medical imaging. This should come as no surprise: Dr. Bakas was the lead organizer of the BraTS challenge.

Since 2012, the BraTS challenge an annual data science competition has asked competitors to train algorithms to spot signs of brain tumors, specifically gliomas, on MRIs. All the competing teams use the same benchmark dataset to train, validate and test their algorithms.

In 2018, that dataset consisted of about 2,000 MRIs from roughly 500 patients, pulled from ten different medical institutions, Dr. Bakas said.

Now, this is a tiny fraction of the MRIs in the world relevant to the BraTS contest; about 20,000 people per year get diagnosed with gliomas in the U.S. alone. But obtaining medical images for a competition data set is tricky. For one, it requires the patients consent. For another, it requires approval from the contributing hospitals internal review board, which involves proving the competition serves the greater good.

The BraTS challenge is just one of many data science challenges that navigate labyrinthine bureaucracy to compile datasets of medical images.

Major companies rely on these datasets, too; theyre more robust than what even Google could easily amass on its own. Googles LYNA, a machine learning algorithm that can pinpoint signs of metastatic breast cancer in the lymph nodes, first made headlines by parsing the images from the 2016 ISBI Camelyon challenges dataset more than 10 percent more accurately than the contests original winner. NVIDIA, meanwhile, sent a team to the 2018 BraTS challenge and won.

[A]n accurate algorithm alone is insufficient to improve pathologists workflow or improve outcomes for breast cancer patients.

Even challenge-winning algorithms, though or the algorithms that beat the winning algorithms arent ready for clinical use. Googles LYNA remains in development. Despite 2018 headlines touting it asbetter than humans in detecting advanced breast cancer, it still needs more testing.

[A]n accurate algorithm alone is insufficient to improve pathologists workflow or improve outcomes for breast cancer patients, Google researchers Martin Stumpe and Craig Mermel wrote on the Google AI blog.

For one thing, it was trained to read one slide per patient but in a real clinical setting, doctors look at multiple slides per patient.

For another, accuracy in a challenge context doesnt always mean real-world accuracy. Challenge datasets are small, and biased by the fact that every patient consented to share their data. Before clinical use, even a stellar algorithm may need to train on more data.

Like, much more data.

More on Data ScienceCoronavirus Charts Are Everywhere. But Are They Good?

Federated learning, Dr. Bakas saw, could allow powerful algorithms access to massive stores of data. But how well did it work? In other words, could federated learning train an algorithm as accurate as one trained on pooled data? In 2018, he and a team of researchers from Intel publisheda paper on exactly that.

No one before has attempted to apply federated learning in medicine, he said.

He and his co-authors trained an off-the-shelf, basic algorithm on BraTS 2018 MRI images using four different techniques. One was traditional machine learning, using pooled data; another was federated learning; the other two techniques were alternate collaborative learning techniques that, like federated learning, involved training an algorithm on a fragmented dataset.

We were not married to federated learning, Dr. Bakas said.

It emerged as a clear success story in their research, though the best technique for melding AI with HIPAA-mandated data privacy. In terms of accuracy, the algorithm trained via federated learning was second only to the algorithm trained on conventional, pooled data. (The difference was subtle, too; the federated learning algorithm was 99 percent as accurate as the traditional one.) Federated learning also made all the different institutions algorithms converge more neatly on an optimal model than other collaborative learning techniques.

Once Dr. Bakas and his coauthors validated the concept of federated learning, a team of NVIDIA researchers elaborated on it further, Dr. Bakas explained. Their focus was fusing it with even more ironclad privacy technology. Though federated learning never involves pooling patient data, it does involve pooling algorithms trained on patient data and hackers could, hypothetically, reconstruct the original data from the trained algorithms.

NVIDIA found a way to prevent this with a blend of encryption and differential privacy. The reinvented model aggregation process involves transferring only partial weights... so that people cannot reconstruct the data, Dr. Flores said.

Its worth noting that NVIDIAs paper, like the one Dr. Bakas co-authored, relied on the BraTS 2018 dataset. This was largely a matter of practicality, but the link between data science competitions and federated learning could grow more substantive.

In the long-term, Dr. Bakas sees data science competitions facilitating algorithmic development; thanks to common data sets and performance metrics, these contests help identify top-tier machine learning algorithms. The winners can then progress to federated learning projects and train on much bigger data sets.

In other words, federated learning projects wont replace data science competitions. Instead, they will function as a kind of major league for competition-winning algorithms to play in and theyll improve the odds of useful algorithms making it into clinical settings.

The end goal is really to reach to the clinic, Dr. Bakas said, to help the radiologist [and] to help the clinician do their work more efficiently.

Short answer: a lot. Federated learning is still a new approach to machine learning Clara FL, lets remember, debuted less than six months agoand researchers continue to work out the kinks.

So far, NVIDIAs team has learned that clear, shared data protocols play a key role in federated learning projects.

You have to make sure that the data to each of the sites is labeled in the same fashion, Dr. Flores said, so that you're comparing apples to apples.

Open questions remain, though. For instance when a central server aggregates a group of trained algorithms, how should it do that? Its not as straightforward as taking a mathematical average, because each institutions dataset is different in terms of size, underlying population demographics and other factors.

Which ones do you give more weight to than others? Dr.Flores said. There are many different ways of aggregating the data That's something that we are still researching.

Federated learning has major potential, though, especially in Europe, where privacy regulations have already tightened due to the General Data Protection Regulation. The law, which went into effect back in 2018, is the self-proclaimed toughest privacy and security law in the world so stringent, Dr. Bakas noted, that it would prevent hospitals from contributing patient data to the BraTS challenge, even if the individual patients consented.

So far, the U.S. hasnt cracked down quite as heavily on privacy as the EU has, but federated learning could still transform industries where privacy is paramount. Already, banks can train machine learning models to recognize signs of fraud, using in-house data; however, if each bank has its own model, it will benefit big banks and leave small banks vulnerable.

While individual banks may like this outcome, it is less than ideal for solving the social issue of money laundering, writes B Capital venture capitalist Mike Fernandez.

Federated learning could even the playing field, allowing banks of all sizes to contribute to a global fraud detection model trained on more data than any one bank could amass, all while maintaining their clients privacy.

Federated learning could apply to other industries, too. As browsers like Mozilla and Google Chrome phase out third-party cookies, federated learning of cohorts could become a way of targeting digital ads to groups of like-minded users, while still keeping individual browser histories private. Federated learning could also allow self-driving cars to share the locations of potholes and other road hazards without sharing, say, their exact current location.

One thing Dr. Bakas doesnt see federated learning doing, even in the distant future: automating away doctors. Instead, he sees it freeing up doctors to do what they do best, whether thats connecting with patients or treating novel and complex ailments with innovative treatments. Doctors have already dreamed up creative approaches to the coronavirus, like using massage mattresses for pregnant women to boost patients oxygen levels.

They just dont really excel at scanning medical imaging and diagnosing common, well-documented ailments, like gliomas or metastatic breast cancer.

They can identify something that is already flaring up on a scan, Dr. Bakas said, but there are some ambiguous areas that radiologists are uncertain about.

Machine learning algorithms, too, often make mistakes about these areas. At first. But over time, they can learn to make fewer, spotting patterns in positive cases invisible to the human eye.

This is why they complement doctors so powerfully they can see routine medical protocols in a fresh, robotic way. That may sound like an oxymoron, but its not necessarily one anymore.

More on Data ScienceThe Dos and Donts of Database Design, According to Experts

Go here to read the rest:
Federated Learning Fuses AI and Privacy and It Could Transform Healthcare - Built In

What Is Differential Deep Learning? Through The Lens Of Trading – Analytics India Magazine

The explosion of the internet, in conjunction with the success of neural networks, brought the world of finance closer to more exotic approaches. Deep learning today is one such technique that is being widely adopted to cut down losses and generate profits.

When gut instincts do not do the job, mathematical methods come into play. Differential equations, for instance, can be used to represent a dynamic model. The approximation of pricing functions is a persistent challenge in quantitative finance. By the early 1980s, researchers were already experimenting with Taylor Expansions for stochastic volatility models.

For example, if company A wants to buy a commodity say oil in future from company B but is unsure of the future prices. So company A wants to make a deal with B that no matter what the price of oil is in the future, B should sell it to A for a price according to their contract.

In the world of finance, this is a watered-down version of derivatives trading. Derivatives are the securities made on underlying assets. In the above case, company A predicts a rise in price, and company B predicts a fall in price. Both these companies are making a bet on future prices and agree upon a price that cuts down their losses or can even bring profits (if A sells after price rise). So how do these companies arrive at a certain price or how do they predict the future price?

Taking the same example of derivatives trading, the researchers at Danske Bank of Denmark, have explored the implications of differential deep learning.

Deep learning offers the much needed analytic speeds, which are necessary for an approximation of volatile markets. Machine learning tools can take up the high dimensionality (many parameters) trait of a market and help resolve the computational bottlenecks.

Differential machine learning is an extension of supervised learning, where ML models are trained on differentials of labels to inputs.

In the context of financial derivatives and risk management, pathwise differentials are popularly

computed with automatic adjoint differentiation (AAD). AAD is an algorithm to calculate derivative sensitivities, very quickly. Nothing more, nothing less. AAD is also known in the field of machine learning under the name back-propagation or simply backprop.

Differential machine learning, combined with AAD, wrote the authors, provides extremely effective pricing and risk approximations. They say that fast pricing analytics can be produced and can effectively compute risk management metrics and even simulate hedge strategies.

This work compares differential machine learning to data augmentation in computer vision, where multiple labelled images are produced from a single one, by cropping, zooming, rotating or recolouring.

Data augmentation not only extends the training set but also encourages the machine learning model to learn important invariances (features that stay the same). Similarly, derivatives labels not only increase the amount of information in the training set but also encourage the model to learn the shape of the pricing function. Derivatives from feedforward networks form another neural network, efficiently computing risk sensitivities in the context of pricing approximation. Since the adjoints form a second network, one can use them for training as well as expect significant performance gain.

Risk sensitivities converge considerably slower than values and often remain blatantly wrong, even with hundreds of thousands of examples. We resolve these problems by training ML models on datasets augmented with differentials of labels with respect to the following inputs:

This simple idea, assert the authors, along with the adequate training algorithm, will allow ML models to learn accurate approximations even from small datasets, making machine learning viable in the context of trading.

Differential machine learning learns better from data alone, the vast amount of information contained in the differentials playing a similar role, and often more effective, to manual adjustments from contextual information.

The researchers posit that the unreasonable effectiveness of differential ML is applicable in situations where high-quality first-order derivatives with training inputs are available and in complex computational tasks such as the pricing and risk approximation of complex derivatives trading.

Differentials inject meaningful additional information, eventually resulting in better results with smaller datasets. Learning effectively from small datasets is critical in the context of regulations, where the pricing approximation must be learned quickly, and the expense of a large training set cannot be afforded.

The results from the experiments by Danske banks researchers show that learning the correct shape from differentials is crucial to the performance of regression models, including neural networks.

Know more about differential deep learning here.

comments

Originally posted here:
What Is Differential Deep Learning? Through The Lens Of Trading - Analytics India Magazine

Key Ways Machine Learning is Being Used in Software Testing – Techiexpert.com – TechiExpert.com

Listen : Audio version of this article

The impact that the software development industry has on the world is unparalleled to any other industry, because obviously software is one of the cores of modern-day society, no matter what industry or niche youre looking at. This is only going to continue as time goes on.

However, with the rise of services like AI, Big Data, and machine learning, its going to be interesting to see how technologies like machine learning are going to be effect industries like the software development industry, in particular, the software testing part of the process.

Today, were going to explore exactly why and how machine learning is being used and is going to be used in software testing practices, and what benefits this is going to provide to the services and procedures. Lets get into it.

When you manually test software, there are plenty of problems that currently affect the process. For one, software testing is time-consuming and expensive, and productivity is low. You also need a specialist software tester to make sure that everything is handled properly. This, of course, invites the risk of human error, which in some cases, could be incredibly costly.

On the other hand, when you train a machine to test software, once the machine has learned what its supposed to be doing, its incredibly fast at testing and will do what a human tester can do in a fraction of the time. This saves not only time, but also the money that would be spent on a software tester, shares Nick Denning, a business writer.

With a manual tester, youll have someone sitting in your office or remotely and testing your app or program. Theyll go through your software and sample different features and will test it. In bigger applications, you may have a group so you can test multiple users at once.

However, with machine learning testing, you can test up to 1,000+ of instances at the same time, meaning you can test network and user strain of multiple users of your software, or just try lots of different situations to see if bugs appear.

When youre manual testing your software, whether youre doing it yourself, or youre using a dedicated software tester, one key problem is that your tester might not be able to pick up glitches that theyre not used too. This can cause glitches to slip through the gaps and end up in your final product.

When youre using machine learning technologies, these are tools that are designed, by their very nature, to be as accurate as possible. Every single time to run them, they are going to go out of their way to deliver the results youve trained them to find. This is true whether youre purchasing machine learning software or training your own, explains Michael Taylor, a tech writer.

As we mentioned above, software testing is a slow process when its carried out by a person or a small team. It can take weeks, or even months, in extreme cases, depending on the size of the project, and this can cost a huge amount of your budget. When machine learning is involved, you only need one technology to carry out all the tasks from start to finish.

This can save you so much time and will prevent you from having to carry out mundane tasks checking data logs and trying to find areas of code that are experiencing errors. This, of course, means you can automate a ton of steps in your testing procedures.

Since machine learning applications learn every single time they run, once theyve found an error and addressed it, they can then learn this error has been dealt with, and can simultaneously run thousands of tests to make sure that nothing else was affected, all with this information still in mind.

This delivers more accurate results, more actionable data, and faster testing times that can help you get your software project to its final stages quicker than ever.

See the article here:
Key Ways Machine Learning is Being Used in Software Testing - Techiexpert.com - TechiExpert.com

The Librarians of the Future Will Be AI Archivists – Popular Mechanics

Library of Congress/Newspaper Navigator

In July 1848, L'illustration, a French weekly, printed the first photo to appear alongside a story. It depicted Parisian barricades set up during the city's June Days uprising. Nearly two centuries later, photojournalism has bestowed libraries with legions of archival pictures that tell stories of our past. But without a methodical approach to curate them, these historical images could get lost in endless mounds of data.

That's why the Library of Congress in Washington, D.C. is undergoing an experiment. Researchers are using specialized algorithms to extract historic images from newspapers. While digital scans can already compile photos, these algorithms can also analyze, catalog, and archive them. That's resulted in 16 million newspaper pages' worth of images that archivists can sift through with a simple search.

The Bridgeman Art Library

Ben Lee, innovator-in-residence at the Library of Congress, and a graduate student studying computer science at the University of Washington, is spearheading what's called Newspaper Navigator. His dataset comes from an existing project called Chronicling America, which compiles digital newspaper pages between 1789 and 1963.

He noticed that the library had already embarked on a crowdsourcing journey to turn some of those newspaper pages into a searchable database, with a focus on content relating to World War I. Volunteers could mark up and transcribe the digital newspaper pagessomething that computers aren't always so great at. In effect, what they had built was a perfect set of training data for a machine learning algorithm that could automate all of that grueling, laborious work.

"Volunteers were asked to draw the bounding boxes such that they included things like titles and captions, and so then the system would...identify that text," Lee tells Popular Mechanics. "I thought, let's try to see how we can use some emerging computer science tools to augment our abilities and how we use collections."

In total, it took about 19 days' worth of processing time for the system to sift through all 16,358,041 newspaper pages. Of those, the system only failed to process 383 pages.

Newspaper Navigator/ArXiv

Newspaper Navigator builds upon the same technology that engineers used to create Google Books. It's called optical character recognition, or OCR for short, and it's a class of machine learning algorithms that can translate images of typed or handwritten symbols, like words on a scanned magazine page, into digital, machine-readable text.

At Popular Mechanics, we have an archive of almost all of our magazines on Google Books, dating back to January 1905. Because Google has used OCR to optimize those digital scans, it's simple to go through and search our entire archive for mentions of, say, "spies," to get a result like this:

Popular Mechanics

But images are something else entirely.

Using deep learning, Lee built an object detection model that could isolate seven different types of content: photographs, illustrations, maps, comics, editorial cartoons, headlines, and advertisements. So if you want to find photos specifically of soldiers in trenches, you might search "trenches" in Newspaper Navigator and get results instantly.

Before, you'd have to sift through potentially thousands of pages' worth of data. This breakthrough will be extremely empowering for archivists, and Lee has open-sourced all of the code that he used to build his deep-learning model.

"Our hope is actually that people who have collections of newspapers...might be able to use the the code that I'm releasing, or do their own version of this at different scales," Lee says. One day your local library could use this sort of technology to help digitize and archive the history of your local community.

Newspaper Navigator/ArXiV

This is not to say that the system is perfect. "There definitely are cases in which the system will especially miscategorize say, an illustration as a cartoon or something like that," Lee says. But he has accounted for these false positives through confidence scores that highlight the likelihood that a given piece of media is a cartoon or a photograph.

"One of my goals is to use this project...to highlight some of the issues around algorithmic bias."

Lee also says that, even despite his best efforts, these kinds of systems will always encode some human bias. But to reduce any heavy-handedness, Lee tried to focus on emphasizing the classes of imagescartoon versus advertisementrather than what's actually shown in the images themselves. Lee believes this should reduce the instances of the system attempting to make judgement calls about the dataset. That should be left up to the curator, he says.

"I think a lot of these questions are very very important ones to consider and one of my goals is to use this project as an opportunity to highlight some of the issues around algorithmic bias," Lee says. "It's easy to assume that machine learning solves all the problemsthat's a fantasybut in the this project, I think it's a real opportunity to emphasize that we need to be careful how we use these tools."

The rest is here:
The Librarians of the Future Will Be AI Archivists - Popular Mechanics

Preventing procurement fraud in local government with the help of machine learning – Open Access Government

Half of the worlds organisations experienced economic crime between 2016 and 2018, according to PwCs Global Economic Crime and Fraud survey. Government organisations are among these and by no means immune from crime. And yet, the perpetrators are not always remote cyber criminals. The most worrying news is that half of fraud is carried out by agents within the organisation. For many rogue employees, their method is procurement fraud.

Though businesses may often misjudge the cost of fraud, the lack of a definable victim aside from the UK taxpayer means that the figures for fraud which hits public services are unclear. Figures for procurement fraud specifically are even more questionable, though estimates suggest that local councils quashed malicious efforts to steal over 300 million through procurement fraud in 2017/18. The amount undetected by anti-fraud investigators may dwarf that.

The importance of public sector funds not being lost to fraud or wasted has been crucially highlighted in recent weeks by the COVID-19 virus outbreak. These funds, always in high demand, are even more precious in emergency situations like were facing at the moment.

Transparency is a buzzword within the public sector. Citizens are looking for clarity on the value the public sector provides to taxpayers and are demanding more when it comes to services. It is increasingly clear to cash-strapped governments that preventing and detecting fraud is crucial for achieving goals when faced with the alternative of raising taxes which is never popular with the general public. The imperative is to safeguard taxpayers funds and for the public sector to do everything in its power to ensure that these funds are spent on crucial services.

Local government is a particular risk area for procurement fraud. Local governments, including city management,spend a lot of money particularly because many now outsource significant amounts of service provision. They may also lack expertise in contracting and commissioning, and may, therefore, be an easy target for fraudsters. The procurement process is an obvious target.

Procurement fraud can occur at any stage of the procurement lifecycle, which makes it extremely complex to detect and prevent. Analysis suggests that for government organisations, procurement fraud is most likely to occur at the payments processing stage, although vendor selection and bids are also vulnerable stages.

There are a number of ways in which procurement fraud can occur. Some involve collusion between employees and contractors, and others involve external fraudsters taking advantage of a vulnerability in the system. Organisations can also make themselves more vulnerable to fraud by not ensuring that employees follow proper procedures for procurement. One possible problem, for example, is dividing invoices up into smaller chunks to avoid particular thresholds. This is usually done in all innocence as a way to make procurement simpler, but it also leaves the organisation open to abuse because the proper checks are not made.

But if procurement fraud is on the rise, so too is counter-fraud work. Governments around the world have strategies and are monitoring the situation carefully. Many have increased the checks put on procurement processes and have also provided more information to employees and potential contractors about how to spot fraud and potential fraud.

There is growing understanding that rules-based systems are not enough to stop fraud: they may help to detect it after the event, but they are unlikely to prevent it, even in combination with systems to reduce opportunity. Analytics-based systems, however, can both improve detection of fraud, and also start to predict it. They are often based onartificial intelligence(AI), which learns from previous cases, and can then detect patterns that may be associated with fraud, or process breaches that may be a problem.

Detecting anomalies, however, is just one step in the process of preventing fraud. Its only an indicator, and all indicators can do is to indicate. In fraud detection, indicators like anomalies highlight an area for further investigation. Then its over to the fraud, audit and compliance teams to take a look.

Traditional fraud detection has often taken months to complete. Time-consuming audits could detect fraud, but these could begin months after the event, and may only occur once a year. Fraud detection systems based on analytics can spot fraud in a fraction of the time, flagging anomalies to investigation squads in real-time. The actions of those teams can then halt fraud in its tracks, before it takes place, or provide rapid evidence on the perpetrator. Public organisations that put these new technologies in place can rest assured that, with machine learning, fraud detection is not only smart, efficient and speedy, but a frightening prospect for those participating in procurement fraud.

Editor's Recommended Articles

Read this article:
Preventing procurement fraud in local government with the help of machine learning - Open Access Government

Deep Learning: An Overview in Scientific Applications – Analytics Insight

Over the last few years, deep learning has seen a huge uptake in popularity in businesses and scientific applications as well. It is defined as a subset of artificial intelligence that leverages computer algorithms to generate autonomous learning from data and information. Deep learning is prevalent across many scientific disciplines, from high-energy particle physics and weather and climate modeling to precision medicine and more. The technology has come a long way, when scientists developed a computer model in the 1940s that was organized in interconnected layers, like neurons in the human brain.

Deep learning signifies substantial progress in the ability of neural networks to automatically create problemsolving features and capture highly complex data distributions. Deep neural networks are now the state-of-the-art machine learning models across diverse areas, including image analysis and natural language processing, among others, and extensively deployed in academia and industry.

Developments in this technology have a vast potential for scientific applications and medical imaging, medical data analysis, and diagnostics. In scientific settings, data analysis is understanding as recognizing the underlying mechanisms that give rise to patterns in the data. When this is the goal, dimensionality reduction, and clustering are simple and unsupervised but highly effective techniques to divulge concealed properties in the data.

In areport, titled A Survey of Deep Learning for Scientific Discovery, where former Google CEO Eric Schmidt and Google AI researcher Maithra Raghu have put together a comprehensive overview on deep learning techniques and their application to scientific research. According to their guide, deep learning algorithms have been very effective in the processing of visual data. They also describe convolutional neural networks (CNNs) as the most eminent family of neural networks and very constructive in working with any kind of image data.

In scientific contexts, one of the best applications of CNNs is medical imaging analysis. Human experts such as radiologists and physicians have mostly performed the medical image interpretation. However, owing to large variations in pathology and potential fatigue of human experts, researchers now have started capitalizing on computer-assisted interventions. Already, many deep learning algorithms are in use to analyze CT scans and x-rays and assist in the diagnosis of diseases. Recently, in the time of crisis caused by COVID-19, scientists have started using CNNs to find out symptoms of the virus in chest x-rays.

Deep learning algorithms are also effective is natural language processing. It deals with building computational algorithms to automatically assess and represent human language. Today, NLP-based systems have enabled a various number of applications, and are useful to train machines to perform complex natural language-related tasks like machine translation and dialogue generation.

Moreover, deep learning models originally inspired by biological neural networks, which encompasses artificial neurons, or nodes, connected to a web of other nodes through edges, allowing these artificial neurons to collect and send information to each other.

Read the original here:
Deep Learning: An Overview in Scientific Applications - Analytics Insight