news and analysis for omnichannel retailers – Retail Technology Innovation Hub

Machine learning algorithms will learn patterns from the past data and predict trends and best price. These algorithms can predict the best price, discount price and promotional price based on competition, macroeconomic variables, seasonality etc.

To find out the correct pricing in real-time retailers follow the following steps:

Gather input data

In order to build a machine learning algorithm, retailers collect various data points from the customers. These are:

Transactional data

This includes the sales history of each customer and the products, which they have bought in the past.

Product description

The brands, product category, style, photos and the selling price of the previously sold products are collected. Past promotions and campaigns are also analysed to find the effect of price changes on each category.

Customer details

Demographic details and customer feedback are gathered.

Competition and inventory

Retailers also try to find the data regarding the price of products sold by their competitors and supply chain and inventory data.

Depending on the set of key performance indicators defined by the retailers, the relevant data is filtered.

For every industry, pricing would involve different goals and constraints. In terms of the dynamic nature, the retail industry can be compared to the casino industry where machine learning is involved inonline live dealer casino games too.

Like casinos, retail also has the target of profit maximisation and retention of customer loyalty. Each of these goals and constraints can be fed to a machine learning algorithm to generate dynamic prices of products.

Visit link:
news and analysis for omnichannel retailers - Retail Technology Innovation Hub

This know-it-all AI learns by reading the entire web nonstop – MIT Technology Review

This is a problem if we want AIs to be trustworthy. Thats why Diffbot takes a different approach. It is building an AI that reads every page on the entire public web, in multiple languages, and extracts as many facts from those pages as it can.

Like GPT-3, Diffbots system learns by vacuuming up vast amounts of human-written text found online. But instead of using that data to train a language model, Diffbot turns what it reads into a series of three-part factoids that relate one thing to another: subject, verb, object.

Pointed at my bio, for example, Diffbot learns that Will Douglas Heaven is a journalist; Will Douglas Heaven works at MIT Technology Review; MIT Technology Review is a media company; and so on. Each of these factoids gets joined up with billions of others in a sprawling, interconnected network of facts. This is known as a knowledge graph.

Knowledge graphs are not new. They have been around for decades, and were a fundamental concept in early AI research. But constructing and maintaining knowledge graphs has typically been done by hand, which is hard. This also stopped Tim Berners-Lee from realizing what he called the semantic web, which would have included information for machines as well as humans, so that bots could book our flights, do our shopping, or give smarter answers to questions than search engines.

A few years ago, Google started using knowledge graphs too. Search for Katy Perry and you will get a box next to the main search results telling you that Katy Perry is an American singer-songwriter with music available on YouTube, Spotify, and Deezer. You can see at a glance that she is married to Orlando Bloom, shes 35 and worth $125 million, and so on. Instead of giving you a list of links to pages about Katy Perry, Google gives you a set of facts about her drawn from its knowledge graph.

But Google only does this for its most popular search terms. Diffbot wants to do it for everything. By fully automating the construction process, Diffbot has been able to build what may be the largest knowledge graph ever.

Alongside Google and Microsoft, it is one of only three US companies that crawl the entire public web. It definitely makes sense to crawl the web, says Victoria Lin, a research scientist at Salesforce who works on natural-language processing and knowledge representation. A lot of human effort can otherwise go into making a large knowledge base. Heiko Paulheim at the University of Mannheim in Germany agrees: Automation is the only way to build large-scale knowledge graphs.

To collect its facts, Diffbots AI reads the web as a human wouldbut much faster. Using a super-charged version of the Chrome browser, the AI views the raw pixels of a web page and uses image-recognition algorithms to categorize the page as one of 20 different types, including video, image, article, event, and discussion thread. It then identifies key elements on the page, such as headline, author, product description, or price, and uses NLP to extract facts from any text.

Every three-part factoid gets added to the knowledge graph. Diffbot extracts facts from pages written in any language, which means that it can answer queries about Katy Perry, say, using facts taken from articles in Chinese or Arabic even if they do not contain the term Katy Perry.

Browsing the web like a human lets the AI see the same facts that we see. It also means it has had to learn to navigate the web like us. The AI must scroll down, switch between tabs, and click away pop-ups. The AI has to play the web like a video game just to experience the pages, says Tung.

Diffbot crawls the web nonstop and rebuilds its knowledge graph every four to five days. According to Tung, the AI adds 100 million to 150 million entities each month as new people pop up online, companies are created, and products are launched. It uses more machine-learning algorithms to fuse new facts with old, creating new connections or overwriting out-of-date ones. Diffbot has to add new hardware to its data center as the knowledge graph grows.

Researchers can access Diffbots knowledge graph for free. But Diffbot also has around 400 paying customers. The search engine DuckDuckGo uses it to generate its own Google-like boxes. Snapchat uses it to extract highlights from news pages. The popular wedding-planner app Zola uses it to help people make wedding lists, pulling in images and prices. NASDAQ, which provides information about the stock market, uses it for financial research.

Adidas and Nike even use it to search the web for counterfeit shoes. A search engine will return a long list of sites that mention Nike trainers. But Diffbot lets these companies look for sites that are actually selling their shoes, rather just talking about them.

For now, these companies must interact with Diffbot using code. But Tung plans to add a natural-language interface. Ultimately, he wants to build what he calls a universal factoid question answering system: an AI that could answer almost anything you asked it, with sources to back up its response.

Tung and Lin agree that this kind of AI cannot be built with language models alone. But better yet would be to combine the technologies, using a language model like GPT-3 to craft a human-like front end for a know-it-all bot.

Still, even an AI that has its facts straight is not necessarily smart. Were not trying to define what intelligence is, or anything like that, says Tung. Were just trying to build something useful.

Go here to see the original:
This know-it-all AI learns by reading the entire web nonstop - MIT Technology Review

Python Is About to Get the Squeeze – Built In

Python was released in the 1990s as a general-purpose programming language. Despite its clean syntax, the exposure Python got in its first decade wasntencouraging, and itdidnt really find inroads into the developers workspace. Perl was the first choice scripting language and Java had established itself as the go-to in the object-oriented programming arena. Of course, any language takes time to mature and only gets adopted when its better suited to a task than the existing tools.

For Python, that time first arrived during the early 2000s when people started realizing it has an easier learning curve than Perl and offers interoperability with other languages. This realization led to a larger number of developers incorporating Python into their applications. The emergence of Django eventually led to the doom of Perl, and Python started gaining more momentum. Still, it wasnt even close in popularity to Java and JavaScript, both of which were newer than Python.

Fast forward to the present, and Python has trumped Java to become the second-most-popular language according to the StackOverflow Developer Survey 2019. It was also the fastest-growing programming language of the previous decade. Pythons rise in popularity has a lot to do with the emergence of big data in the 2010s as well asdevelopments in machine learning and artificial intelligence. Businesses urgently required a language for quick development with low barriers of entry that could help manage large-scale data and scientific computing tasks. Python was well-suited to all these challenges.

Besides having those factors in its favor, Python was an interpreted language with dynamic typing support. More importantly, it had the backing of Google, whod invested in Python for Tensorflow, which led to its emergence as the preferred language for data analysis, visualization, and machine learning.

Yet, despite the growing demand for machine learning and AI at the turn of this decade, Python wont stay around for long. Like every programming language, it has its own set of weaknesses. Those weaknesses make it vulnerable to replacement by languages more suited to the common tasks businesses ask of them. Despite the presence of R, the emergence of newer languages such as Swift, Julia, and Rust actually poses a bigger threat to the current king of data science.

Rust is still trying to catch up with the machine learning community, and so I believe Swift and Julia are the languages that will dethrone Python and eventually rule data science. Lets see why odds are against Python.

All good things come at a cost, and Pythons dynamically typed nature is no exception. It hampers developers, especially when running the code in production. Dynamic typing that makes it easy to write code quickly without defining types increases the risk of running into runtime issues, especially when the codebase size increases. Bugs that a compiler would easily figure out could go unidentified in Python, causing hindrances in production and ultimately slowing the development process in large-scale applications.

Worse, unlike compiled code, Pythons interpreter analyzes every line of code at execution time. This leads to an overhead that causes a significantly slower performance when compared to other languages.

Julia allows you to avoid some of these problems. Despite being dynamically typed, it has a just-in-time compiler. The JIT compiler either generates the machine code right before its executed or uses previously stored, cached compilations, which makes it as performant as statically typed languages. More importantly, it has a key feature known as multiple dispatch thatis like function overloading of OOPs, albeit at runtime. The power of multiple dispatch lies in its ability to handle different argument types without the need to create separate function names or nested if statements. This helps in writing compact code, which is a big win in numeric computations since unlike Python, you can easily scale solutions to deal with all types of arguments.

Even better, Swift is a statically typed language and is highly optimized due to its LLVM (Low-Level Virtual Machine) compiler. The LLVM makes it possible to quickly compile into assembly code, making Swift super-efficient and almost as fast as C. Also, Swift boasts better memory safety and management tools known as Automatic Reference Counting. Unlike garbage collectors, ARC is a lot more deterministic as it reclaims memory whenever the reference count hits zero.

As compiled languages that offer type annotations, Swift and Julia are a lot faster and robust for development than Python. That alone might be enough to recommend them over the older language, but there are other factors to consider as well.

If slowness was not the most obvious drawback of Python, the language also has limitations with parallel computing.

In short, Python uses GIL (Global Interpreter Lock), which prevents multiple threads from executing at the same time in order to boost the performance of single threads. This process is a big hindrance because it means that developers cannot use multiple CPU cores for intensive computing.

I agree withthe commonplace notion that were currently doing finewhen leveraging Pythons interoperability with C/C++ libraries like Tensorflow and PyTorch. But a Python wrapper doesnt solve all debugging issues. Ultimately, when inspecting the underlying low-level code, were falling back on C and C++. Essentially, we cant leverage the strengths of Python at the low level,which puts it out of the picture.

This factor will soon play a decisive role in the fall of Python and rise of Julia and Swift. Julia is a language exclusively designed to address the shortcomings of Python. It primarily offers three features: coroutines (asynchronous tasks), multi-threading, and distributed computing all of which only show the immense possibilities for concurrent and parallel programming. This structure makes Julia capable of performing scientific computations and solving big data problems at a far greater speed than Python.

On the other hand, Swift possesses all the tools required for developing mobile apps and has no problems with parallel computing.

Despite the disadvantages it has with respect to speed, multi-threading, and type-safety, Python still has a huge ecosystem that boasts an enormous set of libraries and packages. Understandably, Swift and Julia are still infants in the field of machine learning and possess only a limited number of libraries. Yet, their interoperability with Python more than compensates for the lack of library support in Julia and Swift.

Julia not only lets programmers use Python code (and vice-versa), but also supports interoperability with C, R, Java, and almost every major programming language. This versatility would certainly give the language a good boost and increase its chances of a quick adoption among data scientists.

Swift, on the other hand, provides interoperability with Python with the PythonKit library.The biggest selling point for Swift (which has an Apple origin) is a strong support its been getting from Google, who fully backed Python decades ago. See how the tables have turned!

Also, the fact that the creator of Swift, Chris Lattner, is now working on Googles AI brain team just shows that Swift is being seriously groomed as Pythons replacement in the machine learning field.The Tensorflow team investing in Swift with their S4TF project only further proves that the language isnt merely regarded as a wrapper over Python. Instead Swift, thanks to its differential programming support and ability to work at a low level like C, will potentially be used to replace the underlying deep learning tools.

As the size of data continues to increase, Pythons Achilles heel will be soon found out. Gone are the days when ease of use and ability to write code quickly mattered. Speed and parallel computing are the name of the game and Python, which is a more general-purpose language, will no longer solve that problem. Inevitably, it will fade away while Julia and Swift seem like the candidates to take over the reins.

Its important to note that Python as a programming language wont disappear any time soon. Instead, itll only take a backseat in data science as the languages that are more specifically designed for deep learning will rule.

View post:
Python Is About to Get the Squeeze - Built In

Experts at Quantzig Analyze the Correlations and Functionalities of Data Lakes, Analytics, and Machine Learning in Their Recent Article – Business…

LONDON--(BUSINESS WIRE)--Given the hype around data management, it is close to impossible to dismiss the use of data lakes as a small step forward in an already accustomed technology realm. Notably, its not the implementation of new technology that matters now, as much as what it enables organizations to do using data. By making data easily accessible for everyone within an organization, data lakes are slowly turning out to be the fundamental driving forces behind innovation and disruption across industries.

To gain a competitive edge, organizations must act on data-driven insights. Book a FREE Demo to learn how we can help you leverage and act on these insights!

Moreover, as data grows and diversifies, many organizations are finding that traditional methods of data management are now becoming outdated. Quantzig, in its recent article, aims to understand the performance implications, common characteristics, and critical capabilities that need to be considered while building and maintaining data lakes as a common source for all business-critical data.

The ability of data lakes to uncover hidden data correlations in data sets can massively impact any part of the business. Request a FREE proposal to learn more about our ability to solve complex business problems, present analysis findings and communication of status, issues, and market risks in an easy to comprehend format.

Nows the time for companies to invest in data lakes and analytics to avoid being left behind in the race to success, says an analytics expert at Quantzig.

Why Quantzig?

120 +

1500

550+

15 +

Global clients includingFortune 500 companies

Comprehensiveprojects

Data scientists andanalytics experts

Years ofexperience

Detailed information about Quantzigs data management capabilities can be accessed at https://bit.ly/2XWvzzp

As organizations are now actively building data lakes and investing in analytics platforms, they need to reconsider several critical capabilities, including:

Real-time Data Movement: Data lakes enable businesses to import and store data in different formats and from various sources in real-time. Adopting such a method for data storage and analysis enables businesses to scale to data of any size while saving time of defining data structures, schema, and transformations.

Data Storage: By storing data in data lakes, businesses gain the ability to crawl, process, and index the data to better understand the data being stored. Apart from real-time transactional data, it enables businesses to store relational data and data from line of business applications, and non-relational data like mobile apps, IoT devices, and social media.

At Quantzig, we have a cross-functional team that comprises of researchers, analytics experts, and data scientists who assist our clients in implementing solutions that help devise a digital strategy that can move the needle. Speak to our analytics experts right away!

Analytics Platforms: Data lakes empower data scientists, data developers, and business analysts to access data with their choice of analytic tools and frameworks. This also enables businesses to run analytics without the need to move your data to a separate analytics system.

Machine Learning: Data lakes enable organizations to generate reports and insights that aid decision making using historical data modeling and machine learning techniques to forecast outcomes and prescribe actions to achieve the desired result.

In the analytical process of decoding unstructured data to extract actionable insight, a well-defined data lake strategy is proven to produce relevant insights promptly. Also, when the movement of data is frictionless and trustworthy, companies can discover and act upon immediate opportunities for driving growth and efficiency.

Follow us on LinkedIn and Twitter to keep abreast of the emerging trends in data management.

About Quantzig

Quantzig is a global analytics and advisory firm with offices in the US, UK, Canada, China, and India. For more than 15 years, we have assisted our clients across the globe with end-to-end data modeling capabilities to leverage analytics for prudent decision making. Today, our firm consists of 120+ clients, including 55 Fortune 500 companies. For more information on our engagement policies and pricing plans, visit: https://www.quantzig.com/request-for-proposal

Here is the original post:
Experts at Quantzig Analyze the Correlations and Functionalities of Data Lakes, Analytics, and Machine Learning in Their Recent Article - Business...

How Fidelis Leverages Machine Learning to Combat Threats Hiding in your Network – Security Boulevard

Many threats lurk in your network, hiding in external (north-south) or internal (east-west) traffic. So this is where we come in. We leverage machine learning capabilities and advanced analytics to detect the threats hiding in your network traffic.

To begin, threats hiding in external (north-south) traffic are attempting to do three things:

However, the malware activities that leave a footprint ininternal (east-west) network traffic are attempting:

To start, anomaly detection using network traffic has a long history.Traditionally, it has been done for network performance monitoring and diagnostics. There are three main challenges in adapting this approach for threat detection. First, building representative baseline models for normal or benign network activities. Second, preventing a deluge of false alarms. And third, interpreting anomalies as threat related activities to enable response.

The Fidelis Network Detection and Response (NDR) Anomaly Detection addresses the first two challenges using two strategies. Number one, it casts a wide net by analyzing network behavior using five different contexts. These are External, Internal, Application Protocols, Data Movement, and Events detected using rules and signatures.

To continue, for each context, it learns up to five different families of models to learn high fidelity baseline models. For example, for the External Traffic context, we have a family of models that focus on outbound geo-location. So within this family, we have individual baseline models for different countries or groups of countries.

Together, these five contexts and their model families capturewhat is normal baseline behavioron an enterprise network. Because of that, we are able to correlate anomalies from different models to identify high confidence detections. Then, we provide an interpretation of our anomaly detections for analysts. So, we map them to the MITRE ATT&CK TTPs to enable a response.

Inanexternal context, we focus on properties of external or north-south traffic that is independent of the application protocol. Using Unsupervised Machine Learning, statistical anomaly detection, and advanced analytics, we flag three types of suspicious activities that involve internal assets controlled by an enterprise:

With all of this, these models provide protection against threats mapped by the MITRE ATT&CK framework to the Initial Access tactics. In particular, Drive-by Compromise (T1189), and Data Exfiltration, plus the techniques related to Exfiltration Over Alternative Protocol (T1048), Exfiltration Over Web Service (T1567), and Automated Exfiltration (T1020).

Many organizations also deploy external-facing services hosted in a demilitarized zone (DMZ) that is open to the Internet. Fidelis NDR has anomaly models targeted at DMZ services. This can detect an increase in traffic to DMZ servers or traffic originating from a new location. Such anomalies often indicate that an enterprise might be the target of a new threat vector, campaign, or adversary.

In an internal context, we focus on internal traffic patterns along three dimensions. This includes who is talking to whom (I.e. connection patterns between assets), remote access and login behavior patterns, and volume of traffic exchanged between assets. Specifically, we flag five different types of suspicious activities.

Lateral Movement (TA0008)

Fidelis Network Detection and Response (NDR)uses a combination of these machine learning capabilities and advanced analytics to detect suspicious activities on an enterprise network. In a previous blog on Using Machine Learning for Threat Detection, our CTO Anubhav Arora talked about the advantages of using Machine Learning to detect patterns of cyber-attacks hiding in large amount of network traffic data. He defined the different approaches based on Supervised and Unsupervised Machine Learning algorithms. We also released a webinarhosted by SANS where we discuss this topic in more detail.

The Fidelis NDR Anomaly Detection framework involves five contexts. They include External, Internal, Application Protocol, Data Movement, and Events detected using rules and signatures. As mentioned earlier, these contexts capture what is normal baseline behavior on the network, which then helps detect any anomalies.

You can subscribe to our Threat Geek blog to receive the upcoming blogs in this series on Unsupervised Machine Learning to detect network activities. Our Data Science Manager will delve into Application Protocol and Data Movement contexts, the models and threats associated with them, and more. Contact usif you have any questions and want to learn more about our NDR solution.

See original here:
How Fidelis Leverages Machine Learning to Combat Threats Hiding in your Network - Security Boulevard

mHealth app using machine learning to increase physical activity in diabetes and depression: clinical trial protocol for the DIAMANTE Study – DocWire…

This article was originally published here

BMJ Open. 2020 Aug 20;10(8):e034723. doi: 10.1136/bmjopen-2019-034723.

ABSTRACT

INTRODUCTION: Depression and diabetes are highly disabling diseases with a high prevalence and high rate of comorbidity, particularly in low-income ethnic minority patients. Though comorbidity increases the risk of adverse outcomes and mortality, most clinical interventions target these diseases separately. Increasing physical activity might be effective to simultaneously lower depressive symptoms and improve glycaemic control. Self-management apps are a cost-effective, scalable and easy access treatment to increase physical activity. However, cutting-edge technological applications often do not reach vulnerable populations and are not tailored to an individuals behaviour and characteristics. Tailoring of interventions using machine learning methods likely increases the effectiveness of the intervention.

METHODS AND ANALYSIS: In a three-arm randomised controlled trial, we will examine the effect of a text-messaging smartphone application to encourage physical activity in low-income ethnic minority patients with comorbid diabetes and depression. The adaptive intervention group receives messages chosen from different messaging banks by a reinforcement learning algorithm. The uniform random intervention group receives the same messages, but chosen from the messaging banks with equal probabilities. The control group receives a weekly mood message. We aim to recruit 276 adults from primary care clinics aged 18-75 years who have been diagnosed with current diabetes and show elevated depressive symptoms (Patient Health Questionnaire depression scale-8 (PHQ-8) >5). We will compare passively collected daily step counts, self-report PHQ-8 and most recent haemoglobin A1c from medical records at baseline and at intervention completion at 6-month follow-up.

ETHICS AND DISSEMINATION: The Institutional Review Board at the University of California San Francisco approved this study (IRB: 17-22608). We plan to submit manuscripts describing our user-designed methods and testing of the adaptive learning algorithm and will submit the results of the trial for publication in peer-reviewed journals and presentations at (inter)-national scientific meetings.

TRIAL REGISTRATION NUMBER: NCT03490253; pre-results.

PMID:32819981 | DOI:10.1136/bmjopen-2019-034723

Original post:
mHealth app using machine learning to increase physical activity in diabetes and depression: clinical trial protocol for the DIAMANTE Study - DocWire...

Chatbots Are Machine Learning Their Way To Human Language – Forbes

Moveworks founding team from left to right Vaibhav Nivargi, CTO; Bhavin Shah, CEO; Varun Singh, VP ... [+] of Product; Jiang Chen, VP of Machine Learning.

Computers and humans have never spoken the same language. Over and above speech recognition, we also need computers to understand the semantics of written human language. We need this capability because we are building the Artificial Intelligence (AI)-powered chatbots that now form the intelligence layers in Robot Process Automation (RPA) systems and beyond.

Known formally as Natural Language Understanding (NLU), early attempts (as recently as the 1980s) to give computers the ability to interpret human text were comically terrible. This was a huge frustration to both the developers attempting to make these systems work and the users exposed to these systems.

Computers are brilliant at long division, but really bad at knowing the difference between whether humans are referring to football divisions, parliamentary division lobbies or indeed long division for mathematics. This is because mathematics is formulaic, universal and unchanging, but human language is ambiguous, contextual and dynamic.

As a result, comprehending a typical sentence requires the unprogrammable quality of common sense or so we thought.

But in just the last few years, software developers in the field of Natural Language Understanding (NLU) have made several decades worth of progress in overcoming that obstacle, reducing the language barrier between people and AI by solving semantics with mathematics.

Such progress has stemmed in no small part from giant leaps forward in NLU models, including the landmark BERT framework and offshoots like DistilBERT, RoBERTa and ALBERT. Powered by hundreds of these models, modern NLU software is able to deconstruct complex sentences to distill their essential meaning, said Vaibhav Nivargi, CTO and co-founder of Moveworks.

Moveworks software combines AI with Natural Language Processing (NLP) to understand and interpret user requests, challenges and problems before then using a further degree of AI to help deliver the appropriate actions to satisfy the users needs.

Nivargi explains that crucially here we can also now build chatbots that use Machine Learning (ML) to go a step further: autonomously addressing users requests and troubleshooting questions written in natural language. So not only can AI now communicate with employees on their terms, it can even automate many of the routine tasks that make work feel like work - thanks to this newfound capacity for reading comprehension.

Nivargi provides an illustrative example of an IT support request, which we can break down and analyze. Bhavin is a new company employee and a user is asking the chatbot how he can be added to the organizations marketing group to access its information pool and data. The request is as follows (graphic shown below at end):

Howdo [sic] I add Bhavin to the marketing group.

In large part due to the typing/spelling mistake at the start (instead of how do, the user has typed howdo) we have an immediate problem. As recently as two years ago, there was not a single application in the world capable of understanding (and then resolving) the infinite variety of similar requests to this that employees pose to their IT teams.

Of course, we could program an application to trigger the right automated workflow when it receives this exact request. But needless to say, that approach doesnt scale at all. Hard problems demand hard solutions. So here, any solution worth its salt must tackle the fundamental challenges of natural language, which is ambiguous, contextual and dynamic, said Nivargi.

A single word can have many possible meanings; for instance, the word run has about 645 different definitions. Add in the inevitable human error like the typo in this request of the phrase how do and we can see that breaking down a single sentence becomes quite daunting, quite quickly. Moveworks Nivargi explains that the initial step, therefore, is to use machine learning to identify syntactic structures that can help us rectify spelling or grammatical errors.

But, he says, to disambiguate what the employee wants, we also need to consider the context surrounding their request, including that employees department, location and role, as well as other relevant entities. A key technique in doing so is meta learning, which entails analyzing so-called metadata (information about information).

By probabilistically weighing the fact that Alex (another employee) and Bhavin are located in North America, Machine Learning models can fuzzy select the marketingna@company.abc email group, without Alex having to have specified his or her exact name. In this way, we can potentially get Alexs help and get him/her involved in the workflow at hand, said Nivargi.

As TechTarget explains here, Fuzzy logic is an approach to computing based on degrees of truth rather than the usual "true or false" (1 or 0) Boolean logic on which the modern computer is based.

Human service desk agents already factor in context by drawing on their experience, so the secret for an AI chatbot is to mimic this intuition with mathematical models.

Finally lets remember that language in particular the language used in the enterprise is dynamic. New words and expressions arise every month, while the IT systems and applications at a given company shift even more often. To deal with so much change, an effective chatbot must be rooted in advanced Machine Learning, since it needs to constantly retrain itself based on real-time information.

Despite the complexity under the hood, however, the number one criteria for a successful chatbot is a seamless user experience. Nivargi says that what his firm has learned when developing NLU technologies is that all employees care about is getting their requests resolved, instantly, via natural conversations on a messaging tool.

As we stand at the turn of the decade, we humans are arguably still not 100% comfortable with chatbot interactions. Theyre still too automated, too often non-intuitive and (perhaps unsurprisingly) too to machine-like. Technologies like these show that we've started to build chatbots with semantic intuitive intelligence, but there is still work to do. When we get to a point where technology can navigate the peculiarities and idiosyncrasies of human language.... then, just then, we may start to enjoy talking to robots.

Addressing requests written in natural language requires the combination of hundreds of machine ... [+] learning models. In this case, the Moveworks chatbot determines that Alex wants to add Bhavin to the email group for marketing.

Go here to read the rest:
Chatbots Are Machine Learning Their Way To Human Language - Forbes

Explainable AI: From the peak of inflated expectations to the pitfalls of interpreting machine learning models – ZDNet

Machine learning and artificial intelligence are helping automate an ever-increasing array of tasks, with ever-increasing accuracy. They are supported by the growing volume of data used to feed them, and the growing sophistication in algorithms.

The flip side of more complex algorithms, however, is less interpretability. In many cases, the ability to retrace and explain outcomes reached by machine learning models (ML) is crucial, as:

"Trust models based on responsible authorities are being replaced by algorithmic trust models to ensure privacy and security of data, source of assets and identity of individuals and things. Algorithmic trust helps to ensure that organizations will not be exposed to the risk and costs of losing the trust of their customers, employees and partners. Emerging technologies tied to algorithmic trust include secure access service edge, differential privacy, authenticated provenance, bring your own identity, responsible AI and explainable AI."

The above quote is taken from Gartner's newly released 2020 Hype Cycle for Emerging Technologies. In it, explainable AI is placed at the peak of inflated expectations. In other words, we have reached peak hype for explainable AI. To put that into perspective, a recap may be useful.

As experts such as Gary Marcus point out, AI is probably not what you think it is. Many people today conflate AI with machine learning. While machine learning has made strides in recent years, it's not the only type of AI we have. Rule-based, symbolic AI has been around for years, and it has always been explainable.

Incidentally, that kind of AI, in the form of "Ontologies and Graphs" is also included in the same Gartner Hype Cycle, albeit in a different phase -- the trough of disillusionment. Incidentally, again, that's conflating.Ontologies are part of AI, while graphs, not necessarily.

That said: If you are interested in getting a better understanding of the state of the art in explainable AI machine learning, reading Christoph Molnar's book is a good place to start. Molnar is a data scientist and Ph.D. candidate in interpretable machine learning. Molnar has written the bookInterpretable Machine Learning: A Guide for Making Black Box Models Explainable, in which he elaborates on the issue and examines methods for achieving explainability.

Gartner's Hype Cycle for Emerging Technologies, 2020. Explainable AI, meaning interpretable machine learning, is at the peak of inflated expectations. Ontologies, a part of symbolic AI which is explainable, is in the trough of disillusionment

Recently, Molnar and a group of researchers attempted to addresses ML practitioners by raising awareness of pitfalls and pointing out solutions for correct model interpretation, as well as ML researchers by discussing open issues for further research. Their work was published as a research paper, titledPitfalls to Avoid when Interpreting Machine Learning Models, by the ICML 2020 Workshop XXAI: Extending Explainable AI Beyond Deep Models and Classifiers.

Similar to Molnar's book, the paper is thorough. Admittedly, however, it's also more involved. Yet, Molnar has striven to make it more approachable by means of visualization, using what he dubs "poorly drawn comics" to highlight each pitfall. As with Molnar's book on interpretable machine learning, we summarize findings here, while encouraging readers to dive in for themselves.

The paper mainly focuses on the pitfalls of global interpretation techniques when the full functional relationship underlying the data is to be analyzed. Discussion of "local" interpretation methods, where individual predictions are to be explained, is out of scope. For a reference on global vs. local interpretations, you can refer to Molnar's book as previously covered on ZDNet.

Authors note that ML models usually contain non-linear effects and higher-order interactions. As interpretations are based on simplifying assumptions, the associated conclusions are only valid if we have checked that the assumptions underlying our simplifications are not substantially violated.

In classical statistics this process is called "model diagnostics," and the research claims that a similar process is necessary for interpretable ML (IML) based techniques. The research identifies and describes pitfalls to avoid when interpreting ML models, reviews (partial) solutions for practitioners, and discusses open issues that require further research.

Under- or overfitting models will result in misleading interpretations regarding true feature effects and importance scores, as the model does not match the underlying data generating process well. Evaluation of training data should not be used for ML models due to the danger of overfitting. We have to resort to out-of-sample validation such as cross-validation procedures.

Formally, IML methods are designed to interpret the model instead of drawing inferences about the data generating process. In practice, however, the latter is the goal of the analysis, not the former. If a model approximates the data generating process well enough, its interpretation should reveal insights into the underlying process. Interpretations can only be as good as their underlying models. It is crucial to properly evaluate models using training and test splits -- ideally using a resampling scheme.

Flexible models should be part of the model selection process so that the true data-generating function is more likely to be discovered. This is important, as the Bayes error for most practical situations is unknown, and we cannot make absolute statements about whether a model already fits the data optimally.

Using opaque, complex ML models when an interpretable model would have been sufficient (i.e., having similar performance) is considered a common mistake. Starting with simple, interpretable models and gradually increasing complexity in a controlled, step-wise manner, where predictive performance is carefully measured and compared is recommended.

Measures of model complexity allow us to quantify the trade-off between complexity and performance and to automatically optimize for multiple objectives beyond performance. Some steps toward quantifying model complexity have been made. However, further research is required as there is no single perfect definition of interpretability but rather multiple, depending on the context.

This pitfall is further analyzed in three sub-categories: Interpretation with extrapolation, confusing correlation with dependence, and misunderstanding conditional interpretation.

Interpretation with Extrapolation refers to producing artificial data points that are used for model predictions with perturbations. These are aggregated to produce global interpretations. But if features are dependent, perturbation approaches produce unrealistic data points. In addition, even if features are independent, using an equidistant grid can produce unrealistic values for the feature of interest. Both issues can result in misleading interpretations.

Before applying interpretation methods, practitioners should check for dependencies between features in the data (e.g., via descriptive statistics or measures of dependence). When it is unavoidable to include dependent features in the model, which is usually the case in ML scenarios, additional information regarding the strength and shape of the dependence structure should be provided.

Confusing correlation with dependence is a typical error. The Pearson correlation coefficient (PCC) is a measure used to track dependency among ML features. But features with PCC close to zero can still be dependent and cause misleading model interpretations. While independence between two features implies that the PCC is zero, the converse is generally false.

Any type of dependence between features can have a strong impact on the interpretation of the results of IML methods. Thus, knowledge about (possibly non-linear) dependencies between features is crucial. Low-dimensional data can be visualized to detect dependence. For high-dimensional data, several other measures of dependence in addition to PCC can be used.

Misunderstanding conditional interpretation. Conditional variants to estimate feature effects and importance scores require a different interpretation. While conditional variants for feature effects avoid model extrapolations, these methods answer a different question. Interpretation methods that perturb features independently of others also yield an unconditional interpretation.

Conditional variants do not replace values independently of other features, but in such a way that they conform to the conditional distribution. This changes the interpretation as the effects of all dependent features become entangled. The safest option would be to remove dependent features, but this is usually infeasible in practice.

When features are highly dependent and conditional effects and importance scores are used, the practitioner has to be aware of the distinct interpretation. Currently, no approach allows us to simultaneously avoid model extrapolations and to allow a conditional interpretation of effects and importance scores for dependent features.

Global interpretation methods can produce misleading interpretations when features interact. Many interpretation methods cannot separate interactions from main effects. Most methods that identify and visualize interactions are not able to identify higher-order interactions and interactions of dependent features.

There are some methods to deal with this, but further research is still warranted. Furthermore, solutions lack in automatic detection and ranking of all interactions of a model as well as specifying the type of modeled interaction.

Due to the variance in the estimation process, interpretations of ML models can become misleading. When sampling techniques are used to approximate expected values, estimates vary, depending on the data used for the estimation. Furthermore, the obtained ML model is also a random variable, as it is generated on randomly sampled data and the inducing algorithm might contain stochastic components as well.

Hence, themodel variance has to be taken into account. The true effect of a feature may be flat, but purely by chance, especially on smaller data, an effect might algorithmically be detected. This effect could cancel out once averaged over multiple model fits. The researchers note the uncertainty in feature effect methods has not been studied in detail.

It's a steep fall to the peak of inflated expectations to the trough of disillusionment. Getting things done for interpretable machine learning takes expertise and concerted effort.

Simultaneously testing the importance of multiple features will result in false-positive interpretations if the multiple comparisons problem (MCP) is ignored. MCP is well known in significance tests for linear models and similarly exists in testing for feature importance in ML.

For example, when simultaneously testing the importance of 50 features, even if all features are unimportant, the probability of observing that at least one feature is significantly important is 0.923. Multiple comparisons will even be more problematic, the higher dimensional a dataset is. Since MCP is well known in statistics, the authors refer practitioners to existing overviews and discussions of alternative adjustment methods.

Practitioners are often interested in causal insights into the underlying data-generating mechanisms, which IML methods, in general, do not provide. Common causal questions include the identification of causes and effects, predicting the effects of interventions, and answering counterfactual questions. In the search for answers, researchers can be tempted to interpret the result of IML methods from a causal perspective.

However, a causal interpretation of predictive models is often not possible. Standard supervised ML models are not designed to model causal relationships but to merely exploit associations. A model may, therefore, rely on the causes and effects of the target variable as well as on variables that help to reconstruct unobserved influences.

Consequently, the question of whether a variable is relevant to a predictive model does not directly indicate whether a variable is a cause, an effect, or does not stand in any causal relation to the target variable.

As the researchers note, the challenge of causal discovery and inference remains an open key issue in the field of machine learning. Careful research is required to make explicit under which assumptions what insight about the underlying data generating mechanism can be gained by interpreting a machine learning model

Molnar et. al. offer an involved review of the pitfalls of global model-agnostic interpretation techniques for ML. Although as they note their list is far from complete, they cover common ones that pose a particularly high risk.

They aim to encourage a more cautious approach when interpreting ML models in practice, to point practitioners to already (partially) available solutions, and to stimulate further research.

Contrasting this highly involved and detailed groundwork to high-level hype and trends on explainable AI may be instructive.

View post:
Explainable AI: From the peak of inflated expectations to the pitfalls of interpreting machine learning models - ZDNet

Practical NLP: The perfect guide for executives and machine learning practitioners – TechTalks

Practical Natural Language Processing provides in-depth coverage of NLP with Python machine learning libraries and beyond.

This article is part ofAI education, a series of posts that review and explore educational content on data science and machine learning. (In partnership withPaperspace)

By many accounts, linguistics is one of the most complicated functions of the human mind. Likewise, natural language processing (NLP) is one of the most complicated subfields of artificial intelligence. Most books on AI, including educational books on machine learning, provide an introduction to natural language processing. But the field of NLP is so vast that covering all its aspects would require several separate books.

When I picked up Practical Natural Language Processing: A Comprehensive Guide to Building Real-World NLP Systems, what I expected was a book that covered Python machine learning for NLP in depth. Though the book didnt exactly turn out to be what I had in mind, it provided the exact kind of coverage that the field misses in the craze and hype that surrounds deep learning today.

The best way to describe Practical Natural Language Processing is a zoomed-out view of the NLP landscape, a close-up of the NLP process, and plenty of practical tips and guidelines to avoid making mistakes in one of the most important fields of AI.

What you take away from Practical Natural Language Processing depends on two things: Your previous background in mathematics and Python machine learning, and your involvement in the field. I recommend this book to two types of readers:

Ironically, these two audiences are at two ends of the NLP spectrum. On the one hand, you have hardcore Python machine learning coders while on the other, you have people whose daily routine doesnt involve coding.

But the authors of Practical Natural Language Processing, who have extensive experience implementing NLP in different fields, have managed to write a book that will provide value to both audiences. Veteran coders can go through the entire book, the accompanying Jupyter Notebooks, and the many references the authors provide. Executives, on the other hand, can skip over the code and read the high-level overview that each chapter provides before digging into the technical Python coding.

I would not recommend this book to novice Python machine learning coders. If you dont have experience with numpy, pandas, matplotlib, scikit-learn, tensorflow, and keras libraries, then this book probably isnt for you. I suggest you go through a book or course on data science and another on Python machine learning before picking up Practical Natural Language Processing.

Anyone who has done machine learning knows that the development cycle of ML applications is different from the classic, rule-based software development lifecycle. But many people mistakenly think that the NLP development pipeline is identical to the data gathering, modeling, testing cycle of any machine learning application. There are some similarities, but there are also many nuances that are specific to NLP.

Some of the most valuable parts of Practical Natural Language Processing are the overview of the NLP development pipeline for different applications. The book brilliantly gives a high-level view of natural language processing that is detached from machine learning and deep learning.

Youll get to know a lot of the challenges involved in gathering, cleaning, and preparing data for NLP applications. Youll also learn about important NLP disciplines such as information extraction, name-entity recognition, temporal information extraction, and more.

One of the key challenges of NLP is that data tends to be very application-specific. Recent advances in deep learning have enabled the development of very large language models that can adapt to different tasks without further tuning. But in a lot of applications and settings, the use of expensive deep learning language models is still not feasible.

Practical Natural Language Processing shows how you can start with small and simple NLP applications and gradually scale them and transition to more complex and automated AI models as you gather more data on your problem domain.

As you go deeper into the book, youll get to know the specific development cycle for different NLP applications such as chatbots, search engines, text summarization, recommender systems, machine translation, ecommerce, and more.

At the end of the book, youll get a review of the end-to-end NLP process and some of the key questions that can determine which path and technology to choose when starting a new application. There are also important guidelines on storing and deploying models and problems youll need to solve in real-world NLP applications, such as reproducing results, which is a great challenge in machine learning in general.

These high-level discussions make Practical Natural Language Processing a valuable read to both developers, team leaders, and executives at tech companies.

But the book also contains a lot of coding examples, which Ill get to next.

Although a large part of Practical Natural Language Processing is about using Python machine learning libraries for language tasks, the book has much more to offer. Interestingly, the most important takeaway is that you dont need machine learning for every single task. In fact, large deep learningbased language models should be your last resort in most cases. This is a recurring theme across the book andin my opiniona very important one, given all the excitement surrounding the use of larger and larger neural networks for NLP tasks.

As the authors will show you, in many cases, simple solutions such as regular expressions will provide better results. At the very least, the simpler rule-based systems will provide you with an interim solution to gather the data required for the more complex AI models.

With that out of the way, the book does go deep on Python machine learning and deep learning for natural language processing. Practical Natural Language Processing provides in-depth coverage of many critical concepts such as word embeddings, bag of words, ngrams, and TF-IDF.

Aside from the popular machine learning and NLP libraries, the book also introduces and explore Python NLP libraries that basic machine learning books dont cover, such as spacy and genism. Theres also a good deal on using other Python tools to better assess the performance of language models, such as t-SNE for visualizing embeddings and LIME for dealing with AI explainability issues.

As you go into the details of each technique and its associated libraries, the authors continue to provide some key tips, such as how to decide between general-purpose embeddings and hand-crafted features.

The book also introduces you to Googles BERT (but you have to bring your own pytorch skills).

I had mixed feelings about the code samples and how theyre presented in the book. On the one hand, there are plenty of valuable material in the Jupyter Notebooks. However, the code tends to get a bit buggy due to bad file addresses in some of the notebooks. (You also need to spend a great deal of time downloading embedding and models from other sources, which is inevitable anyway.)

In the book, snippets are presented very scantly with only the very important parts highlighted. But in many cases, those parts have been selected from the middle of a sequence, which means you wont make sense of it unless you also open the original code file and read what has come before (to their credit, the authors have taken great care to comment the code comprehensively.)

Theres also a lot of we leave it as an exercise for the reader holes, which give some great directions for personal research but would not be appealing to less experienced developers.

Another great thing about Practical Natural Language Processing is the introduction of a roster of tools and application programming interfaces (API) that let you get started with language tasks without much coding.

The point is, you dont need to reinvent the wheel, and theres a likely chance that theres already a tool out there that can boost your initial efforts to integrate NLP into your applications.

Throughout the book, youll get to use tools such as DialogFlow, Elasticsearch, and Semantic3 for different NLP applications. Youll also see how APIs such as Bing can abstract language tasks (if you have the financial means to rent them).

Youll also get an idea of how these different pieces can be integrated into other NLP applications or gradually transitioned to your own custom-made language models.

Familiarity with these tools will also be very useful for product managers who must decide on which direction to take with the development of their applications given their time, budget, and talent constraints.

Practical Natural Language Processing is a must-read for anyone who wants to become seriously involved in NLP. Whether youre a c-level executive or a hands-on coder, this book is for you.

However, dont pick up this book if you just want to learn the basics of NLP. There are plenty of good Python machine learning books and courses that will introduce you to the basics. Once you feel comfortable with Python machine learning and the basics of natural language processing, this will be your go-to book for taking the next step.

See more here:
Practical NLP: The perfect guide for executives and machine learning practitioners - TechTalks

Machine Learning Market 2020 by Solution (Software,services);Deployment Type(Cloud/MLaaS, On-Premise); and End-User (BFSI, Risk Management,Predictive…

Machine Learning market is expected to grow to US$ 39.98 billion by 2025 from US$ 1.29 billion in 2016. The sales of Machine Learnings is largely influenced by numerous economic and environmental factors and the global economy plays a key role in the development of machine learning market. In todays competitive environment, machine learning technology has become an important part in many applications of the BFSI ecosystem, from approving loans, to managing assets, to assessing risks. BFSI Institutions face a dynamic & challenging environment with superior competition from specialized Fin-Tech enterprises, increasing regulatory supplies and pressure on interest margins in a low interest rate market. All of this at a time when consumer behavior is transforming and traditional banking practices and models are no longer adequate to achieve the increasing consumer demands. Machine learning enables them to pioneer in the dynamic business landscape and attain profitable and sustained growth.

Get Sample [emailprotected]: https://www.premiummarketinsights.com/sample/TIP00003837

Key Players:

After studying key companies, the report focuses on the startups contributing towards the growth of the market. Possible mergers and acquisitions among the startups and key organizations are identified by the reports authors in the study. Most companies in the Machine Learning Market are currently engaged in adopting new technologies, strategies, product developments, expansions, and long-term contracts to maintain their dominance in the global market

Analysis tools such as SWOT analysis and Porters five force model have been inculcated in order to present a perfect in-depth knowledge about Machine Learning Market. Ample graphs, tables, charts are added to help have an accurate understanding of this market. The Machine Learning Market is also been analyzed in terms of value chain analysis and regulatory analysis.

Avail Discount on this [emailprotected]: https://www.premiummarketinsights.com/discount/TIP00003837

In-depth qualitative analyses include identification and investigation of the following aspects:

Market Structure

Growth Drivers

Restraints and Challenges

Emerging Product Trends & Market Opportunities

Risk Assessment for Investing in Global Market

Critical Success Factors (CSFs)

The competitive landscape of the market has been examined on the basis of market share analysis of key players. Detailed market data about these factors is estimated to help vendors take strategic decisions that can strengthen their positions in the market and lead to more effective and larger stake in the global Machine Learning Market. Pricing and cost teardown analysis for products and service offerings of key players has also been undertaken for the study.

Table of Contents:

1 Executive Summary

2 Preface

3 Machine Learning Market Overview

4 Market Trend Analysis

5 Global Machine Learning Market Segmentation

6 Market Effect Factors Analysis

7 Market Competition by Manufacturers

8 Key Developments

9 Company Profiling

About Premium market insights:

Premiummarketinsights.comis a one stop shop of market research reports and solutions to various companies across the globe. We help our clients in their decision support system by helping them choose most relevant and cost effective research reports and solutions from various publishers. We provide best in class customer service and our customer support team is always available to help you on your research queries.

Sameer Joshi Call: US: +1-646-491-9876, Apac: +912067274191Email: [emailprotected]

Original post:
Machine Learning Market 2020 by Solution (Software,services);Deployment Type(Cloud/MLaaS, On-Premise); and End-User (BFSI, Risk Management,Predictive...