Veteran Brian Kolfage Arrested in Build the Wall Scam – Connecting Vets

Former Trump administration strategist Steve Bannon and Iraq war veteranBrian Kolfage have been arrested and indicted in Manhattan along with colleaguesAndrew Badolato andTimothy Shea in relation to their We Build the Wall Gofundme campaign which raised over 25 million dollars. The campaign has since been deleted by Gofundme with Kolfage claiming censorship by the online platform.

Acting U.S. Attorney Audrey Strauss said of the case, "As alleged, the defendants defrauded hundreds of thousands of donors, capitalizing on their interest in funding a border wall to raise millions of dollars, under the false pretense that all of that money would be spent on construction.While repeatedly assuring donors that Brian Kolfage, the founder and public face of We Build the Wall, would not be paid a cent, the defendants secretly schemed to pass hundreds of thousands of dollars to Kolfage, which he used to fund his lavish lifestyle."

Authorities allege that the four men made fake invoices and conspired to steal at least some of the funds for their own use. "This case should serve as a warning to other fraudsters that no one is above the law, not even a disabled war veteran or a millionaire political strategist,"Inspector-in-Charge Philip R. Bartlett said in a press release. They allege that Kolfage took $350,000 for hispersonal use while Bannon claimed over a million dollars through a non-profit he ran to cover his personal expenses.

They have been charged with conspiracy to commit wire fraud and conspiracy to commit money laundering, and are facing up to 20 years in prison.

This isn't the first time Kolfage has been in trouble as it has been previously reported that he has, "a history of scamming would-be donors. Kolfage previously led a crowdfunding charge that claimed to raise money for veterans at military hospitals, but a spokesperson for the medical facilitiestold Buzzfeed Newsthey have no record of Kolfage donating a penny."

As Connecting Vets reported at the time, donors to the build the wall campaign were supposed to be refunded as per GoFundMe policy but Kolfage got creative and gave donors the option to instead give to his non-profit We Build the Wall, Inc.

"As everyone knows, President Trump has no involvement in this project and felt it was only being done in order to showboat, and perhaps raise funds," White House spokeswomanKayleigh McEnany said upon news of the arrests. "President Trump has not been involved with Steve Bannon since the campaign and the early part of the Administration, and he does not know the people involved with this project," she said.

Want to get more connected to the stories and resources Connecting Vets has to offer?Click hereto sign up for our weekly newsletter.

Reach Jack Murphy:jack@connectingvets.comor@JackMurphyRGR.

The rest is here:

Veteran Brian Kolfage Arrested in Build the Wall Scam - Connecting Vets

How COVID-19 gives cover to press crackdowns the world over – The Detroit News

Mae Anderson, Associated Press Published 1:33 p.m. ET Aug. 19, 2020 | Updated 1:34 p.m. ET Aug. 19, 2020

Governments around the world are taking advantage of the coronavirus pandemic to justify or to divert attention from crackdowns on press freedom.

Media tycoon Jimmy Lai wasarrested in Hong Kongearlier in August as police enforced a new national security law. In June, journalist Maria Ressa was convicted of cyber libel" in the Philippines. In Egypt, at least 12 journalists have been arrested this year under laws against spreading misinformation" related to the coronavirus.

Rappler CEO and Executive Editor Maria Ressa.(Photo: Aaron Favila, AP)

In some cases, regimes have moved to curb alleged misinformation about the coronavirus pandemic that doesnt align with official proclamations about its spread or severity. In others, the pandemic serves as a distraction by directing national attention away from these incidents.

Egypt, for instance, has been jailing young journalists such asNora Younis, editor-in-chief of the al-Manassa news agency, who according to the International Press Institute was arrested on June 24. In Russia,the AP foundat least nine cases against ordinary Russians accused of spreading untrue information on social media and via messenger apps, with at least three of them receiving significant fines.

The IPI has beentracking media freedom violationssince the pandemic began. Such repression includes arrests and charges, restrictions to access to information, censorship, excessive fake news regulation, and physical attack.

Incomplete figures make it difficult to say whether such crackdowns are on the rise. At least 17 countries, including Hungary, Russia, the Philippines and Vietnam, have enacted new laws ostensibly intended to fight misinformation about the coronavirus, according to an IPI tally. In reality, those measures have actually served as pretexts to fine or jail journalists who are critical of the government, the organization said.

In Hungary, for example, Prime Minister Viktor Orban passed a coronavirus law that could mean up to five years in prison for false information. Russia can fine people up to $25,000 or imprison them for five years if they're deemed to have spread false information about the virus. Media outlets can be fined up to $127,000, according to the IPI.

The Committee to Protect Journalists has tracked 163violations of press freedomrelated to the coronavirus this year as of July 29. The group says its data is not comprehensive. The IPI hastracked 421 violationsrelated to the virus, including arrests, censorship, excessive fake news" regulation and physical or verbal attacks.

We see an ongoing crackdown on the press that is compounded by the coronavirus," said Courtney Radsch, CPJs advocacy director.

Hong Kong media tycoon Jimmy Lai.(Photo: Vincent Yu, AP)

Even incidents unrelated to alleged pandemic misinformation can escape broader notice amid the flood of coronavirus news. Jimmy Lai'sarrest in Hong Kong, for instance, shortly followed enactment of a new national security law that gives China more power to squash dissent in Hong Kong. Lai operates Apple Daily, a feisty pro-democracy tabloid that often criticizes Chinas Communist Party-led government.

The libel convictions ofRessaand another journalist were also unrelated to COVID-19. But Radsch said the pandemic can serve as a distraction for such cases that might otherwise have gotten more international attention.

Theres just much less attention being paid to a lot of this since people are just caught up in other news," she said. It's difficult to break through the morass to raise concerns and public concerns about crackdowns."

That's been exacerbated by the absence of a robust response from the U.S. under President Donald Trump, experts said.

In the age before Trump, clearly the United States would be the one advocating for press freedom and independent media worldwide," said David Kaye, a law professor at the University of California, Irvine, and a former UN special rapporteur on freedom of expression. Trump routinely refers to the mainstream press as fake news.

While the Trump administration sanctioned Chinese officials, including Hong Kong leader Carrie Lam, over Lai's arrest, its traditional rhetoric in support of the free press has fallen short. We dont see the robust condemnation that we would expect from the U.S. over press freedom crackdowns or deaths of journalists in custody, Radsch said. The administration also could have done more for Ressa, she said, as the journalist holds American as well as Filipino citizenship.

We have not seen a robust call at the highest level for charges to be dropped," she said. It's not what we expect."

The U.S. does still intervene on occasion. For example, U.S. negotiators have been active negotiations overAustin Tice, a Houston journalist and veteran held in Syria. But that is a rare exception.

Kaye said increasing media repression is a direct consequence of a global rise in authoritarian government.

Authoritarians and populists of the last several years have been elected into office," he said. Theres been pressure on independent media, that hasnt changed, and that has been happening in parallel prior to and into COVID."

The pandemic has added a new vector toward repression," he said. There is existing repression that's continued, and COVID-oriented repression thats new."

Read or Share this story: https://www.detroitnews.com/story/news/world/2020/08/19/virus-outbreak-press-crackdown/5607418002/

Excerpt from:

How COVID-19 gives cover to press crackdowns the world over - The Detroit News

Explained: Who is Edward Snowden? – The Indian Express

By: Explained Desk | New Delhi | Updated: August 18, 2020 12:19:11 pmEdward Snowden, former National Security Agency contractor, speaks during a virtual conversation at the South By Southwest (SXSW) Interactive Festival in Austin, Texas, US, March 10, 2014. (Bloomberg Photo: David Paul Morris)

Last Saturday (August 15), US President Donald Trump said he was considering a pardon for Edward Snowden, former employee of the Central Intelligence Agency (CIA) and former National Security Agency (NSA) contractor, who exposed a surveillance programme under which the US government was collecting data on millions of people.

Trump has previously called Snowden a traitor and a spy who should be executed. In 2013, Trump tweeted, ObamaCare is a disaster and Snowden is a spy who should be executed-but if he could reveal Obamas records, I might become a major fan.

Since he was charged in 2013, Snowden has been in exile in Russia. A pardon could mean he can finally return to the US.

In 2013, The Guardian broke the news that the NSA was collecting phone records of millions of Americans from telecom service provider Verizon. It was further revealed that the intelligence agency was tapping servers of Facebook, Google and Microsoft to track Americans online activities.

Subsequently, The Guardian revealed its source of information, and named Snowden as the whistle-blower who leaked information on these surveillance programmes.

Express Explainedis now onTelegram. Clickhere to join our channel (@ieexplained)and stay updated with the latest

The documents leaked by Snowden showed the NSA and its counterpart in the UK, the Government Communications Headquarters (GCHQ), had found ways to bypass the encryption offered to consumers by various companies on the Internet.

The NSAs ability to decipher the data of millions of Americans was one of its closest-guarded secrets. Knowledge of this was restricted to those who were a part of a highly classified programme, called Bullrun.

The surveillance programme was not only limited to ordinary American citizens, but also foreign leaders such as German Chancellor Angela Merkel.

In 2013, Snowden was charged with theft of US government property and unauthorised communication of national defence information, in violating of the 1917 Espionage Act, and providing classified information to The Guardian and The Washington Post.

The leaks had triggered a debate on surveillance and privacy. While critics accused Snowden of treason, his supporters, including privacy activists, lauded him for releasing the documents.

In 2019, a lawsuit was filed against him by the US for publishing a book, titled Permanent Record, which was in violation of the non-disclosure agreements he had signed with the NSA and CIA. The lawsuit alleged Snowden published the book without submitting it to agencies for pre-publication review, in violation of his agreements with the two agencies.

According to a report in The New York Times, the data intelligence agencies were able to access included sensitive information like trade secrets and medical records, and automatically secures the e-mails, Web searches, Internet chats and phone calls of Americans and others around the world, the documents show.

In an interview Snowden gave to The Guardian in 2014, he said through the arrangement between the NSA and private Internet companies, such as Facebook, the agency was able to get copies of ones Facebook messages, Skype conversations and Gmail inboxes.

The Indian Express is now on Telegram. Click here to join our channel (@indianexpress) and stay updated with the latest headlines

For all the latest Explained News, download Indian Express App.

The Indian Express (P) Ltd

View original post here:
Explained: Who is Edward Snowden? - The Indian Express

Trump and Snowdens War on Proper Authority – The Bulwark

President Trump, who exhibits unrestrained contempt for proper authority, including his own, recently declared that he may soon grant a pardon for Edward Snowden, the former National Security Agency contractor who has lived in Moscow since 2013 after leaking information on the nations most secretive spy agencies. This would be the culmination of a years-long relationship that Trump and his entourage has cultivated with Snowden (and Julian Assange) since it became clear that these forces shared Trumps unalloyed hostility for the American order and all its works.

The prospect of a presidential pardon for Snowden, whose revelations about the National Security Agencys foreign and domestic surveillance techniques were disclosed at his personal whim rather than through democratic audit, is a befitting offering from a chief executive whose flagrant criminality and contempt for the democratic process will be his most enduring legacy.

The controversy over Snowden calls to mind an episode from revolutionary America in which Alexander Hamilton laid down an important standard about the appropriate balance between justice and principle. In May 1775, when a nasty mob gathered at Kings College to tar and feather Myles Cooper, the colleges loyalist president, Hamilton interceded to lecture the crowd on the disgrace such unwarranted violence would bring on the cause of liberty. Thanks to this effort to hold the agitated crowd at bay, Cooper was able to escape to the safety of a British warship, but Hamilton was nonetheless shaken by the experience. He wrote to John Jay soon afterwards, I am always more or less alarmed at every thing which is done of mere will and pleasure, without any proper authority.

It seems never to have occurred to Snowden that his shabby conscience does not and never did constitute a proper authority. He abrogated to himself the responsibilities of governance under the dubious ruse of exposing unconstitutional methods, and his media allies assumed the same responsibilities under the equally dubious ruse of investigative journalism. It has been widely suggested that Snowdens stolen secretsand his unscrupulous acolytes in media who have lustily distributed them on multiple continentsrepresent a valiant defense of individual liberties against mighty and minatory states. This narrative could not be further from the truth.

The apologists for and beneficiaries of Vladimir Putins regime have suggested that Americans civil liberties are held in contempt by their own government. And yet it is the United States government that has seen fit to obtain such a sturdy legal footing before venturing to gather user information from Verizon and other private companies that so offended Snowdens delicate sensibilities. Far from being a violation of the Fourth Amendment guarantee against unreasonable searches and seizures, the collection of sensitive information by the NSAexpressly permitted by Congress and extensively reviewed by the Foreign Intelligence Surveillance Courthas been fully consistent with longstanding judicial precedent.

In United States v. Choate, for instance, the courts decided that the Postal Service may record mail cover, i.e., whats written on the outside of an envelopethe addresses of sender and receiver. As Charles Krauthammer argued at the time of Snowdens revelations, the NSAs recording of U.S. phone data merely extends the logic of this ruling to telephone records. The program is not a mass, roving tap, as it has been widely characterized, because it only records the numbers dialed and time stampsthe outside of the envelope, as it were. The content of the conversation, however, is like the letter inside the envelope. It may not be opened without a court order.

In Smith v. Maryland, the Supreme Court addressed this method of surveillance more directly. It held that the warrantless state installation of a pen register that collected telephone information, known as metadata, could not be construed as an unreasonable search by the standards of the Fourth Amendment. This was decided on the grounds that people have no expectation of privacy in records that plainly belong to another party.

Once the case of government trampling civil liberties is demonstrated to be an obnoxious fiction, the critics tend to fall back on challenging the notion that its surveillance measures even foil the countrys enemies. After all, we are knowingly assured, the enemy already understood the danger of placing calls. This argument has gained a surprising amount of traction given the obvious flaw in its reasoning. A moments thought should tell us that if we, an informed citizenry, didnt know about the details of such operations, our enemies probably didnt know either.

This argument calls to mind the deceitful habit cultivated by detractors of Americas post-September 11 regime of enhanced interrogation. They frequently insisted that torture was not merely immoral; it was ineffective. At times it seemed as if they could not mount a case against it if it were proven to yield benefits to American security. The suggestion that Snowden ever exposed wrongdoing is at least tenable, if not compelling. What strains credulity is that his imprudent action has not done real harm to American security. The ability to hinder terrorists communications is a vital American advantage in a war where the asymmetric benefits so often reside on the other side.

At this point one simply has to reiterate the initial questions: To justify Snowdens wanton disregard for democratic procedure, what laws were ever shown to have been broken? What liberties were violated? This is what Snowdens defendersor, if you like, defenders of the act of bringing these classified materials to lighthave never adequately explained. Indeed, this is what they have doggedly refused to answer. And for good reason. In the vast labyrinth of the American security state, there are numerous possibilities of illegal activity being brought to light.

And yet even at this late date, not a shred of evidence has been produced suggesting chronic abuse, let alone law-breaking, by the NSA. Even if there were such evidence, there is an established and credible process in place to reform government intelligence gathering techniques without arbitrarily leaking classified informationa process that Snowden shows no sign of having availed himself.

Anyone who recalls this sordid affair will not fail to notice the cavalier contempt for authority that propelled Snowden to splash his ill-gotten goods onto the world stage has been replicated every day of the Trump administration. (At least once, this happened literally, when Trump disclosed high-level U.S. secrets to the Russian foreign minister in the Oval Office.) It is this exercise in mere will and pleasure that most urgently needs to be ruled out of courtthe court of public opinion. If the president decides to grant him a pardon anyway, in open defiance of the spirit of his executive authority, it will not be the least of reasons for the American public to turn him out of office this November.

Visit link:
Trump and Snowdens War on Proper Authority - The Bulwark

Chatbots Are Machine Learning Their Way To Human Language – Forbes

Moveworks founding team from left to right Vaibhav Nivargi, CTO; Bhavin Shah, CEO; Varun Singh, VP ... [+] of Product; Jiang Chen, VP of Machine Learning.

Computers and humans have never spoken the same language. Over and above speech recognition, we also need computers to understand the semantics of written human language. We need this capability because we are building the Artificial Intelligence (AI)-powered chatbots that now form the intelligence layers in Robot Process Automation (RPA) systems and beyond.

Known formally as Natural Language Understanding (NLU), early attempts (as recently as the 1980s) to give computers the ability to interpret human text were comically terrible. This was a huge frustration to both the developers attempting to make these systems work and the users exposed to these systems.

Computers are brilliant at long division, but really bad at knowing the difference between whether humans are referring to football divisions, parliamentary division lobbies or indeed long division for mathematics. This is because mathematics is formulaic, universal and unchanging, but human language is ambiguous, contextual and dynamic.

As a result, comprehending a typical sentence requires the unprogrammable quality of common sense or so we thought.

But in just the last few years, software developers in the field of Natural Language Understanding (NLU) have made several decades worth of progress in overcoming that obstacle, reducing the language barrier between people and AI by solving semantics with mathematics.

Such progress has stemmed in no small part from giant leaps forward in NLU models, including the landmark BERT framework and offshoots like DistilBERT, RoBERTa and ALBERT. Powered by hundreds of these models, modern NLU software is able to deconstruct complex sentences to distill their essential meaning, said Vaibhav Nivargi, CTO and co-founder of Moveworks.

Moveworks software combines AI with Natural Language Processing (NLP) to understand and interpret user requests, challenges and problems before then using a further degree of AI to help deliver the appropriate actions to satisfy the users needs.

Nivargi explains that crucially here we can also now build chatbots that use Machine Learning (ML) to go a step further: autonomously addressing users requests and troubleshooting questions written in natural language. So not only can AI now communicate with employees on their terms, it can even automate many of the routine tasks that make work feel like work - thanks to this newfound capacity for reading comprehension.

Nivargi provides an illustrative example of an IT support request, which we can break down and analyze. Bhavin is a new company employee and a user is asking the chatbot how he can be added to the organizations marketing group to access its information pool and data. The request is as follows (graphic shown below at end):

Howdo [sic] I add Bhavin to the marketing group.

In large part due to the typing/spelling mistake at the start (instead of how do, the user has typed howdo) we have an immediate problem. As recently as two years ago, there was not a single application in the world capable of understanding (and then resolving) the infinite variety of similar requests to this that employees pose to their IT teams.

Of course, we could program an application to trigger the right automated workflow when it receives this exact request. But needless to say, that approach doesnt scale at all. Hard problems demand hard solutions. So here, any solution worth its salt must tackle the fundamental challenges of natural language, which is ambiguous, contextual and dynamic, said Nivargi.

A single word can have many possible meanings; for instance, the word run has about 645 different definitions. Add in the inevitable human error like the typo in this request of the phrase how do and we can see that breaking down a single sentence becomes quite daunting, quite quickly. Moveworks Nivargi explains that the initial step, therefore, is to use machine learning to identify syntactic structures that can help us rectify spelling or grammatical errors.

But, he says, to disambiguate what the employee wants, we also need to consider the context surrounding their request, including that employees department, location and role, as well as other relevant entities. A key technique in doing so is meta learning, which entails analyzing so-called metadata (information about information).

By probabilistically weighing the fact that Alex (another employee) and Bhavin are located in North America, Machine Learning models can fuzzy select the marketingna@company.abc email group, without Alex having to have specified his or her exact name. In this way, we can potentially get Alexs help and get him/her involved in the workflow at hand, said Nivargi.

As TechTarget explains here, Fuzzy logic is an approach to computing based on degrees of truth rather than the usual "true or false" (1 or 0) Boolean logic on which the modern computer is based.

Human service desk agents already factor in context by drawing on their experience, so the secret for an AI chatbot is to mimic this intuition with mathematical models.

Finally lets remember that language in particular the language used in the enterprise is dynamic. New words and expressions arise every month, while the IT systems and applications at a given company shift even more often. To deal with so much change, an effective chatbot must be rooted in advanced Machine Learning, since it needs to constantly retrain itself based on real-time information.

Despite the complexity under the hood, however, the number one criteria for a successful chatbot is a seamless user experience. Nivargi says that what his firm has learned when developing NLU technologies is that all employees care about is getting their requests resolved, instantly, via natural conversations on a messaging tool.

As we stand at the turn of the decade, we humans are arguably still not 100% comfortable with chatbot interactions. Theyre still too automated, too often non-intuitive and (perhaps unsurprisingly) too to machine-like. Technologies like these show that we've started to build chatbots with semantic intuitive intelligence, but there is still work to do. When we get to a point where technology can navigate the peculiarities and idiosyncrasies of human language.... then, just then, we may start to enjoy talking to robots.

Addressing requests written in natural language requires the combination of hundreds of machine ... [+] learning models. In this case, the Moveworks chatbot determines that Alex wants to add Bhavin to the email group for marketing.

Go here to read the rest:
Chatbots Are Machine Learning Their Way To Human Language - Forbes

Explainable AI: From the peak of inflated expectations to the pitfalls of interpreting machine learning models – ZDNet

Machine learning and artificial intelligence are helping automate an ever-increasing array of tasks, with ever-increasing accuracy. They are supported by the growing volume of data used to feed them, and the growing sophistication in algorithms.

The flip side of more complex algorithms, however, is less interpretability. In many cases, the ability to retrace and explain outcomes reached by machine learning models (ML) is crucial, as:

"Trust models based on responsible authorities are being replaced by algorithmic trust models to ensure privacy and security of data, source of assets and identity of individuals and things. Algorithmic trust helps to ensure that organizations will not be exposed to the risk and costs of losing the trust of their customers, employees and partners. Emerging technologies tied to algorithmic trust include secure access service edge, differential privacy, authenticated provenance, bring your own identity, responsible AI and explainable AI."

The above quote is taken from Gartner's newly released 2020 Hype Cycle for Emerging Technologies. In it, explainable AI is placed at the peak of inflated expectations. In other words, we have reached peak hype for explainable AI. To put that into perspective, a recap may be useful.

As experts such as Gary Marcus point out, AI is probably not what you think it is. Many people today conflate AI with machine learning. While machine learning has made strides in recent years, it's not the only type of AI we have. Rule-based, symbolic AI has been around for years, and it has always been explainable.

Incidentally, that kind of AI, in the form of "Ontologies and Graphs" is also included in the same Gartner Hype Cycle, albeit in a different phase -- the trough of disillusionment. Incidentally, again, that's conflating.Ontologies are part of AI, while graphs, not necessarily.

That said: If you are interested in getting a better understanding of the state of the art in explainable AI machine learning, reading Christoph Molnar's book is a good place to start. Molnar is a data scientist and Ph.D. candidate in interpretable machine learning. Molnar has written the bookInterpretable Machine Learning: A Guide for Making Black Box Models Explainable, in which he elaborates on the issue and examines methods for achieving explainability.

Gartner's Hype Cycle for Emerging Technologies, 2020. Explainable AI, meaning interpretable machine learning, is at the peak of inflated expectations. Ontologies, a part of symbolic AI which is explainable, is in the trough of disillusionment

Recently, Molnar and a group of researchers attempted to addresses ML practitioners by raising awareness of pitfalls and pointing out solutions for correct model interpretation, as well as ML researchers by discussing open issues for further research. Their work was published as a research paper, titledPitfalls to Avoid when Interpreting Machine Learning Models, by the ICML 2020 Workshop XXAI: Extending Explainable AI Beyond Deep Models and Classifiers.

Similar to Molnar's book, the paper is thorough. Admittedly, however, it's also more involved. Yet, Molnar has striven to make it more approachable by means of visualization, using what he dubs "poorly drawn comics" to highlight each pitfall. As with Molnar's book on interpretable machine learning, we summarize findings here, while encouraging readers to dive in for themselves.

The paper mainly focuses on the pitfalls of global interpretation techniques when the full functional relationship underlying the data is to be analyzed. Discussion of "local" interpretation methods, where individual predictions are to be explained, is out of scope. For a reference on global vs. local interpretations, you can refer to Molnar's book as previously covered on ZDNet.

Authors note that ML models usually contain non-linear effects and higher-order interactions. As interpretations are based on simplifying assumptions, the associated conclusions are only valid if we have checked that the assumptions underlying our simplifications are not substantially violated.

In classical statistics this process is called "model diagnostics," and the research claims that a similar process is necessary for interpretable ML (IML) based techniques. The research identifies and describes pitfalls to avoid when interpreting ML models, reviews (partial) solutions for practitioners, and discusses open issues that require further research.

Under- or overfitting models will result in misleading interpretations regarding true feature effects and importance scores, as the model does not match the underlying data generating process well. Evaluation of training data should not be used for ML models due to the danger of overfitting. We have to resort to out-of-sample validation such as cross-validation procedures.

Formally, IML methods are designed to interpret the model instead of drawing inferences about the data generating process. In practice, however, the latter is the goal of the analysis, not the former. If a model approximates the data generating process well enough, its interpretation should reveal insights into the underlying process. Interpretations can only be as good as their underlying models. It is crucial to properly evaluate models using training and test splits -- ideally using a resampling scheme.

Flexible models should be part of the model selection process so that the true data-generating function is more likely to be discovered. This is important, as the Bayes error for most practical situations is unknown, and we cannot make absolute statements about whether a model already fits the data optimally.

Using opaque, complex ML models when an interpretable model would have been sufficient (i.e., having similar performance) is considered a common mistake. Starting with simple, interpretable models and gradually increasing complexity in a controlled, step-wise manner, where predictive performance is carefully measured and compared is recommended.

Measures of model complexity allow us to quantify the trade-off between complexity and performance and to automatically optimize for multiple objectives beyond performance. Some steps toward quantifying model complexity have been made. However, further research is required as there is no single perfect definition of interpretability but rather multiple, depending on the context.

This pitfall is further analyzed in three sub-categories: Interpretation with extrapolation, confusing correlation with dependence, and misunderstanding conditional interpretation.

Interpretation with Extrapolation refers to producing artificial data points that are used for model predictions with perturbations. These are aggregated to produce global interpretations. But if features are dependent, perturbation approaches produce unrealistic data points. In addition, even if features are independent, using an equidistant grid can produce unrealistic values for the feature of interest. Both issues can result in misleading interpretations.

Before applying interpretation methods, practitioners should check for dependencies between features in the data (e.g., via descriptive statistics or measures of dependence). When it is unavoidable to include dependent features in the model, which is usually the case in ML scenarios, additional information regarding the strength and shape of the dependence structure should be provided.

Confusing correlation with dependence is a typical error. The Pearson correlation coefficient (PCC) is a measure used to track dependency among ML features. But features with PCC close to zero can still be dependent and cause misleading model interpretations. While independence between two features implies that the PCC is zero, the converse is generally false.

Any type of dependence between features can have a strong impact on the interpretation of the results of IML methods. Thus, knowledge about (possibly non-linear) dependencies between features is crucial. Low-dimensional data can be visualized to detect dependence. For high-dimensional data, several other measures of dependence in addition to PCC can be used.

Misunderstanding conditional interpretation. Conditional variants to estimate feature effects and importance scores require a different interpretation. While conditional variants for feature effects avoid model extrapolations, these methods answer a different question. Interpretation methods that perturb features independently of others also yield an unconditional interpretation.

Conditional variants do not replace values independently of other features, but in such a way that they conform to the conditional distribution. This changes the interpretation as the effects of all dependent features become entangled. The safest option would be to remove dependent features, but this is usually infeasible in practice.

When features are highly dependent and conditional effects and importance scores are used, the practitioner has to be aware of the distinct interpretation. Currently, no approach allows us to simultaneously avoid model extrapolations and to allow a conditional interpretation of effects and importance scores for dependent features.

Global interpretation methods can produce misleading interpretations when features interact. Many interpretation methods cannot separate interactions from main effects. Most methods that identify and visualize interactions are not able to identify higher-order interactions and interactions of dependent features.

There are some methods to deal with this, but further research is still warranted. Furthermore, solutions lack in automatic detection and ranking of all interactions of a model as well as specifying the type of modeled interaction.

Due to the variance in the estimation process, interpretations of ML models can become misleading. When sampling techniques are used to approximate expected values, estimates vary, depending on the data used for the estimation. Furthermore, the obtained ML model is also a random variable, as it is generated on randomly sampled data and the inducing algorithm might contain stochastic components as well.

Hence, themodel variance has to be taken into account. The true effect of a feature may be flat, but purely by chance, especially on smaller data, an effect might algorithmically be detected. This effect could cancel out once averaged over multiple model fits. The researchers note the uncertainty in feature effect methods has not been studied in detail.

It's a steep fall to the peak of inflated expectations to the trough of disillusionment. Getting things done for interpretable machine learning takes expertise and concerted effort.

Simultaneously testing the importance of multiple features will result in false-positive interpretations if the multiple comparisons problem (MCP) is ignored. MCP is well known in significance tests for linear models and similarly exists in testing for feature importance in ML.

For example, when simultaneously testing the importance of 50 features, even if all features are unimportant, the probability of observing that at least one feature is significantly important is 0.923. Multiple comparisons will even be more problematic, the higher dimensional a dataset is. Since MCP is well known in statistics, the authors refer practitioners to existing overviews and discussions of alternative adjustment methods.

Practitioners are often interested in causal insights into the underlying data-generating mechanisms, which IML methods, in general, do not provide. Common causal questions include the identification of causes and effects, predicting the effects of interventions, and answering counterfactual questions. In the search for answers, researchers can be tempted to interpret the result of IML methods from a causal perspective.

However, a causal interpretation of predictive models is often not possible. Standard supervised ML models are not designed to model causal relationships but to merely exploit associations. A model may, therefore, rely on the causes and effects of the target variable as well as on variables that help to reconstruct unobserved influences.

Consequently, the question of whether a variable is relevant to a predictive model does not directly indicate whether a variable is a cause, an effect, or does not stand in any causal relation to the target variable.

As the researchers note, the challenge of causal discovery and inference remains an open key issue in the field of machine learning. Careful research is required to make explicit under which assumptions what insight about the underlying data generating mechanism can be gained by interpreting a machine learning model

Molnar et. al. offer an involved review of the pitfalls of global model-agnostic interpretation techniques for ML. Although as they note their list is far from complete, they cover common ones that pose a particularly high risk.

They aim to encourage a more cautious approach when interpreting ML models in practice, to point practitioners to already (partially) available solutions, and to stimulate further research.

Contrasting this highly involved and detailed groundwork to high-level hype and trends on explainable AI may be instructive.

View post:
Explainable AI: From the peak of inflated expectations to the pitfalls of interpreting machine learning models - ZDNet

Practical NLP: The perfect guide for executives and machine learning practitioners – TechTalks

Practical Natural Language Processing provides in-depth coverage of NLP with Python machine learning libraries and beyond.

This article is part ofAI education, a series of posts that review and explore educational content on data science and machine learning. (In partnership withPaperspace)

By many accounts, linguistics is one of the most complicated functions of the human mind. Likewise, natural language processing (NLP) is one of the most complicated subfields of artificial intelligence. Most books on AI, including educational books on machine learning, provide an introduction to natural language processing. But the field of NLP is so vast that covering all its aspects would require several separate books.

When I picked up Practical Natural Language Processing: A Comprehensive Guide to Building Real-World NLP Systems, what I expected was a book that covered Python machine learning for NLP in depth. Though the book didnt exactly turn out to be what I had in mind, it provided the exact kind of coverage that the field misses in the craze and hype that surrounds deep learning today.

The best way to describe Practical Natural Language Processing is a zoomed-out view of the NLP landscape, a close-up of the NLP process, and plenty of practical tips and guidelines to avoid making mistakes in one of the most important fields of AI.

What you take away from Practical Natural Language Processing depends on two things: Your previous background in mathematics and Python machine learning, and your involvement in the field. I recommend this book to two types of readers:

Ironically, these two audiences are at two ends of the NLP spectrum. On the one hand, you have hardcore Python machine learning coders while on the other, you have people whose daily routine doesnt involve coding.

But the authors of Practical Natural Language Processing, who have extensive experience implementing NLP in different fields, have managed to write a book that will provide value to both audiences. Veteran coders can go through the entire book, the accompanying Jupyter Notebooks, and the many references the authors provide. Executives, on the other hand, can skip over the code and read the high-level overview that each chapter provides before digging into the technical Python coding.

I would not recommend this book to novice Python machine learning coders. If you dont have experience with numpy, pandas, matplotlib, scikit-learn, tensorflow, and keras libraries, then this book probably isnt for you. I suggest you go through a book or course on data science and another on Python machine learning before picking up Practical Natural Language Processing.

Anyone who has done machine learning knows that the development cycle of ML applications is different from the classic, rule-based software development lifecycle. But many people mistakenly think that the NLP development pipeline is identical to the data gathering, modeling, testing cycle of any machine learning application. There are some similarities, but there are also many nuances that are specific to NLP.

Some of the most valuable parts of Practical Natural Language Processing are the overview of the NLP development pipeline for different applications. The book brilliantly gives a high-level view of natural language processing that is detached from machine learning and deep learning.

Youll get to know a lot of the challenges involved in gathering, cleaning, and preparing data for NLP applications. Youll also learn about important NLP disciplines such as information extraction, name-entity recognition, temporal information extraction, and more.

One of the key challenges of NLP is that data tends to be very application-specific. Recent advances in deep learning have enabled the development of very large language models that can adapt to different tasks without further tuning. But in a lot of applications and settings, the use of expensive deep learning language models is still not feasible.

Practical Natural Language Processing shows how you can start with small and simple NLP applications and gradually scale them and transition to more complex and automated AI models as you gather more data on your problem domain.

As you go deeper into the book, youll get to know the specific development cycle for different NLP applications such as chatbots, search engines, text summarization, recommender systems, machine translation, ecommerce, and more.

At the end of the book, youll get a review of the end-to-end NLP process and some of the key questions that can determine which path and technology to choose when starting a new application. There are also important guidelines on storing and deploying models and problems youll need to solve in real-world NLP applications, such as reproducing results, which is a great challenge in machine learning in general.

These high-level discussions make Practical Natural Language Processing a valuable read to both developers, team leaders, and executives at tech companies.

But the book also contains a lot of coding examples, which Ill get to next.

Although a large part of Practical Natural Language Processing is about using Python machine learning libraries for language tasks, the book has much more to offer. Interestingly, the most important takeaway is that you dont need machine learning for every single task. In fact, large deep learningbased language models should be your last resort in most cases. This is a recurring theme across the book andin my opiniona very important one, given all the excitement surrounding the use of larger and larger neural networks for NLP tasks.

As the authors will show you, in many cases, simple solutions such as regular expressions will provide better results. At the very least, the simpler rule-based systems will provide you with an interim solution to gather the data required for the more complex AI models.

With that out of the way, the book does go deep on Python machine learning and deep learning for natural language processing. Practical Natural Language Processing provides in-depth coverage of many critical concepts such as word embeddings, bag of words, ngrams, and TF-IDF.

Aside from the popular machine learning and NLP libraries, the book also introduces and explore Python NLP libraries that basic machine learning books dont cover, such as spacy and genism. Theres also a good deal on using other Python tools to better assess the performance of language models, such as t-SNE for visualizing embeddings and LIME for dealing with AI explainability issues.

As you go into the details of each technique and its associated libraries, the authors continue to provide some key tips, such as how to decide between general-purpose embeddings and hand-crafted features.

The book also introduces you to Googles BERT (but you have to bring your own pytorch skills).

I had mixed feelings about the code samples and how theyre presented in the book. On the one hand, there are plenty of valuable material in the Jupyter Notebooks. However, the code tends to get a bit buggy due to bad file addresses in some of the notebooks. (You also need to spend a great deal of time downloading embedding and models from other sources, which is inevitable anyway.)

In the book, snippets are presented very scantly with only the very important parts highlighted. But in many cases, those parts have been selected from the middle of a sequence, which means you wont make sense of it unless you also open the original code file and read what has come before (to their credit, the authors have taken great care to comment the code comprehensively.)

Theres also a lot of we leave it as an exercise for the reader holes, which give some great directions for personal research but would not be appealing to less experienced developers.

Another great thing about Practical Natural Language Processing is the introduction of a roster of tools and application programming interfaces (API) that let you get started with language tasks without much coding.

The point is, you dont need to reinvent the wheel, and theres a likely chance that theres already a tool out there that can boost your initial efforts to integrate NLP into your applications.

Throughout the book, youll get to use tools such as DialogFlow, Elasticsearch, and Semantic3 for different NLP applications. Youll also see how APIs such as Bing can abstract language tasks (if you have the financial means to rent them).

Youll also get an idea of how these different pieces can be integrated into other NLP applications or gradually transitioned to your own custom-made language models.

Familiarity with these tools will also be very useful for product managers who must decide on which direction to take with the development of their applications given their time, budget, and talent constraints.

Practical Natural Language Processing is a must-read for anyone who wants to become seriously involved in NLP. Whether youre a c-level executive or a hands-on coder, this book is for you.

However, dont pick up this book if you just want to learn the basics of NLP. There are plenty of good Python machine learning books and courses that will introduce you to the basics. Once you feel comfortable with Python machine learning and the basics of natural language processing, this will be your go-to book for taking the next step.

See more here:
Practical NLP: The perfect guide for executives and machine learning practitioners - TechTalks

Machine Learning Market 2020 by Solution (Software,services);Deployment Type(Cloud/MLaaS, On-Premise); and End-User (BFSI, Risk Management,Predictive…

Machine Learning market is expected to grow to US$ 39.98 billion by 2025 from US$ 1.29 billion in 2016. The sales of Machine Learnings is largely influenced by numerous economic and environmental factors and the global economy plays a key role in the development of machine learning market. In todays competitive environment, machine learning technology has become an important part in many applications of the BFSI ecosystem, from approving loans, to managing assets, to assessing risks. BFSI Institutions face a dynamic & challenging environment with superior competition from specialized Fin-Tech enterprises, increasing regulatory supplies and pressure on interest margins in a low interest rate market. All of this at a time when consumer behavior is transforming and traditional banking practices and models are no longer adequate to achieve the increasing consumer demands. Machine learning enables them to pioneer in the dynamic business landscape and attain profitable and sustained growth.

Get Sample [emailprotected]: https://www.premiummarketinsights.com/sample/TIP00003837

Key Players:

After studying key companies, the report focuses on the startups contributing towards the growth of the market. Possible mergers and acquisitions among the startups and key organizations are identified by the reports authors in the study. Most companies in the Machine Learning Market are currently engaged in adopting new technologies, strategies, product developments, expansions, and long-term contracts to maintain their dominance in the global market

Analysis tools such as SWOT analysis and Porters five force model have been inculcated in order to present a perfect in-depth knowledge about Machine Learning Market. Ample graphs, tables, charts are added to help have an accurate understanding of this market. The Machine Learning Market is also been analyzed in terms of value chain analysis and regulatory analysis.

Avail Discount on this [emailprotected]: https://www.premiummarketinsights.com/discount/TIP00003837

In-depth qualitative analyses include identification and investigation of the following aspects:

Market Structure

Growth Drivers

Restraints and Challenges

Emerging Product Trends & Market Opportunities

Risk Assessment for Investing in Global Market

Critical Success Factors (CSFs)

The competitive landscape of the market has been examined on the basis of market share analysis of key players. Detailed market data about these factors is estimated to help vendors take strategic decisions that can strengthen their positions in the market and lead to more effective and larger stake in the global Machine Learning Market. Pricing and cost teardown analysis for products and service offerings of key players has also been undertaken for the study.

Table of Contents:

1 Executive Summary

2 Preface

3 Machine Learning Market Overview

4 Market Trend Analysis

5 Global Machine Learning Market Segmentation

6 Market Effect Factors Analysis

7 Market Competition by Manufacturers

8 Key Developments

9 Company Profiling

About Premium market insights:

Premiummarketinsights.comis a one stop shop of market research reports and solutions to various companies across the globe. We help our clients in their decision support system by helping them choose most relevant and cost effective research reports and solutions from various publishers. We provide best in class customer service and our customer support team is always available to help you on your research queries.

Sameer Joshi Call: US: +1-646-491-9876, Apac: +912067274191Email: [emailprotected]

Original post:
Machine Learning Market 2020 by Solution (Software,services);Deployment Type(Cloud/MLaaS, On-Premise); and End-User (BFSI, Risk Management,Predictive...

Informatica Acquires GreenBay Technologies to Advance AI and Machine Learning Capabilities – PRNewswire

REDWOOD CITY, Calif., Aug. 18, 2020 /PRNewswire/ --Informatica, the enterprise cloud data management leader, today announced it has acquired GreenBay Technologies Inc. to accelerate its innovation in AI and machine learning data management technology. The acquisition will strengthen the core capabilities of Informatica's AI-powered CLAIRE engine across its Intelligent Data Platform, empowering businesses to more easily identify, access, and derive insights from organizational data to make informed business decisions.

"We continue to invest and innovate in order to empower enterprises in the shift to the next phase of their digital transformations," said Amit Walia, CEO of Informatica. "GreenBay Technologies is instrumental in delivering on our vision of Data 4.0, by strengthening our ability to deliver AI and machine learning in a cloud-first, cloud-native environment. This acquisition gives us a competitive advantage that will further enable our customers to unleash the power of data to increase productivity with enhanced intelligence and automation."

Core to the GreenBay acquisition are three distinct and advanced capabilities in entity matching, schema matching, and metadata knowledge graphs that will be integrated across Informatica's product portfolio. These technologies will accelerate Informatica's roadmap across Master Data Management, Data Integration, Data Catalog, Data Quality, Data Governance, and Data Privacy.

GreenBay Technologies' AI and machine learning capabilities will be embedded in the CLAIRE engine for a more complete and accurate, 360-degree view and understanding of business, with innovative matching techniques of master data of customers, products, suppliers, and other domains. With the acquisition, GreenBay Technologies will accelerate Informatica's vision for self-integrating systems that automatically infer and link target schemas to source data, enhance capabilities to infer data lineage and relationships, auto-generate and apply data quality rules based on concept schema matching, and increase accuracy of identifying sensitive data across the enterprise data landscape.

GreenBay Technologies was co-founded by Dr. AnHai Doan, University of Wisconsin Madison's Vilas Distinguished Achievement Professor, together with his Ph.D. students, Yash Govind and Derek Paulsen. Dr. Doan oversees multiple data management research projects at the University of Wisconsin's Department of Computer Science and is the co-author of "Principles of Data Integration," a leading textbook in the field, and was among the first to apply machine learning to data integration in 2001. Doan's pioneering work in the area of data integration has received multiple awards, including the prestigious ACM Doctoral Dissertation Award and the Alfred P. Sloan Research Fellowship. Dr. Doan and Informatica have a long history collaborating in the use of AI and machine learning in data management. In 2019, Informatica became the sole investor in GreenBay Technologies, which also has ties to the University of Wisconsin (UW) at Madison and the Wisconsin Alumni Research Foundation (WARF), one of the first and most successful technology transfer offices in the nation focused on advancing transformative discoveries to the marketplace.

"What started as a collaborative project with Informatica's R&D will now help thousands of Informatica customers better manage and utilize their data and solve complex problems at the pace of digital transformation," said Dr. Doan. "GreenBay Technologies will provide Informatica customers with AI and ML innovations for more complete 360 views of the business, self-integrating systems, and more automated data quality and governance tasks."

The GreenBay acquisition is an important part of Informatica's collaboration with academic and research institutions globally to further its vision of AI-powered data management including most recently in Europe with The ADAPT Research Center, a world leader in Natural Language Processing (NLP), in Dublin.

About InformaticaInformatica is the only proven Enterprise Cloud Data Management leader that accelerates data-driven digital transformation. Informatica enables companies to fuel innovation, become more agile, and realize new growth opportunities, resulting in intelligent market disruptions. Over the last 25 years, Informatica has helped more than 9,000 customers unleash the power of data. For more information, call +1 650-385-5000 (1-800-653-3871 in the U.S.), or visit http://www.informatica.com. Connect with Informatica on LinkedIn, Twitter, and Facebook.

Informatica and CLAIRE aretrademarks or registered trademarks of Informatica in the United States and in jurisdictions throughout the world. All other company and product names may be trade names or trademarks of their respective owners.

The information provided herein is subject to change without notice. In addition, the development, release, and timing of any product or functionality described today remain at the sole discretion of Informatica and should not be relied upon in making a purchasing decision, nor as a representation, warranty, or commitment to deliver specific products or functionality in the future.

SOURCE Informatica

http://www.informatica.com

More:
Informatica Acquires GreenBay Technologies to Advance AI and Machine Learning Capabilities - PRNewswire

J&J’s Janssen Partners with BioSymetrics for Predicting Onset and Severity of COVID-19 Using Machine Learning – HospiMedica

Johnson & Johnsons (New Brunswick, NJ, USA) Janssen Pharmaceuticals Inc. (Beerse, Belgium) has entered into a collaboration with BioSymetrics Inc. (New York, NY, USA) and Sema4 (Stamford, CT, USA) that will focus on predicting the onset and severity of COVID-19 among different populations using machine learning.

As part of the collaboration, the parties will use BioSymetrics' Contingent-AI engine across several projects to characterize high-risk populations, measure and predict disease progression based on biological risk factors and treatment course, and identify markers for clinical phenotype and severity of disease. BioSymetrics, a biomedical artificial intelligence company that provides clinical insights and helps researchers develop drugs with greater speed and precision, has developed the platform based on a patent pending AI iteration framework that can be used in conjunction with clinical research to predict target mechanism, identify lead compounds, or provide clinical insights. The collaboration will operate across several projects with a goal of enabling a vaccine and course of treatment against SARS-CoV-2.

"We've been working on deploying AI in the clinical setting for several years," said Anthony Iacovone, Co-Founder and Chairman of BioSymetrics. "We've demonstrated that machine learning can bring speed and precision to helping identify at risk patient populations, predict disease outcomes, and build better treatments, but the pandemic has now pushed biomedical AI technology to the fore front of innovative necessity."

"There is dramatic heterogeneity within the COVID-19 patient groups and a spectrum of disease risk that must be interpreted probabilistically something of which I believe this collaboration will drive through innovation and combined expertise," added Eric Schadt, Founder and CEO of Sema4, a patient-centered health intelligence company.

Related Links:BioSymetrics Inc.Sema4Johnson & JohnsonJanssen Pharmaceuticals Inc.

Read more here:
J&J's Janssen Partners with BioSymetrics for Predicting Onset and Severity of COVID-19 Using Machine Learning - HospiMedica