What the open source community can teach the suddenly remote workforce – Security Boulevard

Productive remote teamwork is possible. Just ask the open source community, who has been doing it for years. Here are some top tips for working remotely.

By now we are all familiar with the, uh, challenges (thats the printable word) of uprooting millions of workers from their offices so they can work more safely from home.

Remote workall of a sudden with no time to plan for itis disruptive. Its unfamiliar. Its stressful. Its distracting, especially for those with school-aged children who are not at school. As one frazzled parent put it, I now have two more full-time jobs. Im a principal and a teacher.

And in the tech sector, software developers who have been working together, collaborating in open office environments, are suddenly isolated. Sure, there are virtual connections, but they are not the same as being in the same room.

That doesnt mean development has to crash and burn, though. There is a template available for overcoming that challenge. The open source software sector has been working remotely since well, since open source became a thing.

In most cases, participants have never met. They dont know each other. They might never see or speak with one another. They are likely in different parts of the country, or different countries, many on different sides of the world. Frequently they dont even speak the same language.

Yet they work together, in many cases with astonishing efficiency, and they produce products of superb quality. The Linux operating system, for example, started as an open source project and remains open source to this day. Open source software is part of virtually every application, network, and system in operation today. It often represents the majoritysometimes more than 90%of the code in a codebase.

So yes, productive remote teamwork is possible.

Thats not to say that open source development is an exact parallel to the corporate world. Open source is a community, not a company. Those who participate are essentially volunteers, not paid employees. There is a hierarchy, but the supervisor is generally more along the lines of community leader than boss.

But conventional development teams and their managers in need of new ideas for working remotely can still learn plenty of things from the remote operation of the open source world. There are even books about itone of the most popular is The Art of Community by Jono Bacon, former community manager for Ubuntu.

Tim Mackey, principal strategist at the Synopsys Cybersecurity Research Center (CyRC), knows about the remote operation of open source communities as well. While he works for a company, he has been a community leader, and still is a community member, for open source projects. He has worked remotely and managed remote teams for the bulk of his career.

So he knows from experience that remote doesnt have to mean disconnected. It just takes some awareness, effort, and cooperation. He described some of the ways open source communities mitigate the absence of physical human contact:

It sounds like the tech version of the real estate mantra Location, location, location, which describes the most important factor in buying a house.

But that is because communication is the foundation for everything else. Of course, working remotely cant be exactly like the physical office environment, where, as Mackey puts it, If someone is working on something that relates to what you are working on, you will know because up will pop a head when you say, I really dont understand why this is doing this.

But it can come close, as long as teams arent too largeideally fewer than 10 people.

You could actually have everyone put their phone on a Skype call. The phone is just sitting in the corner, and it doesnt have any other purpose than to serve as the proxy for the office, he said. There are many ways to solve the problem. You just need to find the pieces that are missing.

Resiliency flows from communicationwhat Mackey says is completely and transparently communicating all of the issues regarding a project.

As is the case with open source projects, resiliency is the result when there is nobody who is magically special who needs to know extra stuff, he said. Anyone at any point in time can know everything. That level of egalitarianism really starts to increase the engagement.

It also means there is no single point of failure, which is a mandatory element of resiliency. If somebody gets sick, goes on vacation, or gets a different job, it doesnt hamstring the rest of the team, because one person isnt carrying all the institutional knowledge. Everybody is.

You dont have to worry about one person having all the magic knowledge and then you are massively disrupted when that one person has to deal with some personal issue or, for that matter, wins a Powerball ticket, Mackey said.

It gives you flexibility. Everybody is going to have some aspect of their life that is going to be variable. Some people want to ski, some people want to surf, some people dont like the cold, some people love the cold.

Emotionally intelligent feedback, which also flows from good communication, can be much trickier among a team working remotely, since emails and texts usually lack tone. Facial expressions, speech volume, and other physical cues present in face-to-face communication can bring a lot of helpful nuance to comments that might seem harsh or even accusatory in writing.

Not to mention that cultural and language barriers can be easier to overcome face-to-face than in writing. If a recipient from another country who speaks another language gets a note and puts it into something like Google Translate, the results can be unpredictable.

If you know multiple languages and have tried that, then you know Google Translate is sometimes really good and sometimes it is absolutely atrocious, Mackey said. He noted that in the open source world or any remote situation, he makes a point of using very precise language when he writes comments.

If you have ever been in a situation where somebody has complained about the tone of your writing, that is exactly the type of scenario that successful open source teams figure out pretty early how to overcome, he said. In some countries, tone is such a key component of their written language that they might miss what you meant more often than you would prefer.

A breakdown in the emotional element of feedback can be a huge kick in morale, Mackey said.

Every team and every project needs a process to govern how things get done. But remember that the whole point of the process is to help things get done. As Bacon says in his book, processes are only useful when they are a means to an end.

Or as Mackey puts it, It really boils down to making certain that everyone knows what it is, why it is, and to a certain extent knows that they can raise their hand and say, But you do realize youre not doing this, right?

And if the shift from the office to working remotely means certain things cant get done, then the process needs a revision.

A perfect example, Mackey said, is a scenario where IT and legal have imposed a security policy that says, To protect our source code, you can only commit code on this special network that will never be accessible outside of the company.

While that might have made perfect sense before, it doesnt work when nobody is at the office. Process needs to be a living entity. You cant just fall back on, But this is the way we have always done it, he said.

One reason is that someone on the team might come up with a workaround just to get their job done, but such a workaround amounts to shadow IT, since it is outside security policy.

Does that mean Ive created a situation where, in trying to do the work Ive been assigned, I have now circumvented every process in place because it wasnt designed for the reality of everybody working from home? he asked.

It is clearly much better to figure out a way to maintain the security of source code without making it impossible for a remote development team to do its work.

All of which, once again, comes down to communication, communication, communication.

Of course, it will take some getting used to. There will likely be some bumps in the road. But if the open source community can do it, organizations can too.

Like what youre reading? Subscribe to the blog!

See more here:
What the open source community can teach the suddenly remote workforce - Security Boulevard

Open source made the cloud in its image – ITworld

The cloud was built for running open source, Matt Wilson once told me, which is why open source [has] worked so well in the cloud.

While true, theres something more fundamental that open source offers the cloud. As one observer put it, The whole intellectual foundation of open interfaces and combinatorial single-purpose tools is pretty well ingrained in cloud. That approach is distinctly open source, which in turn owes much to the Unix mentality that early projects like Linux embraced.

Hence, the next time you pull together different components to build an application on Microsoft Azure, Google Cloud, AWS, or another cloud, realize that the reason you can do this is because the open source ethos permeates the cloud.

Open source has become so commonplace today that we are apt to forget its origins. While it would be an overstatement to suggest that Unix is wholly responsible for what open source became, many of the open source pioneers came from a Unix background, and it shows.

Heres a summary of the Unix philosophy by Doug McIlroy, the creator of Unix pipes:

Sound familiar? From this ideological parentage its not hard to see where open source gets its preference for modularity, transparency, composability, etc. Its also not much of a stretch to see where the open source-centric clouds are picking up their approach to microservices.

In turn, the different clouds have all converged on similar design principles. As Wilson notes, the composable pieces ethos of open source is a property of open systems, and a general Unix philosophy that [is] carried forward in the foundational building blocks of cloud as we know it.

Cloud is impossible without the economics of free and open source software, but cloud is arguably even more impossible at least, in the way we experience it today without the freedoms and design principles offered by open source. Erica Brescia makes this point perfectly.

Importantly, were now in a hyper-growth development phase for the cloud, with different companies with different agendas combining to open source incredibly complex, powerful, and cloud-native software to tackle everything from machine learning to network management. As Jono Bacon notes,

Open source created the model for collaborative technology development in a competitive landscape.

The cloud manifested as the most pressing need to unite this competitive landscape together.

This led to a rich tapestry of communities sharing best practices and approaches.

This rich tapestry of communities sharing owes its existence to open source. Clouds may provide the platforms where open source increasingly lives and grows, but the animating force behind the clouds is open source. Given the pressing problems all around us, were going to need both cloud and communities each driven by open source to help tackle them.

This story, "Open source made the cloud in its image" was originally published by InfoWorld.

Read this article:
Open source made the cloud in its image - ITworld

How Edge Is Different From Cloud And Not – The Next Platform

As the dominant supplier of commercial-grade open source infrastructure software, Red Hat sets the pace and it is not a surprise that IBM was willing to shell out an incredible $34 billion to acquire the company. It is no surprise, then, that Red Hat has its eyes on the edge, that amorphous and potentially substantial collection of distributed computing systems that everyone is figuring out how to chase.

To get a sense of what Red Hat thinks about the edge, we sat down with Joe Fernandes, vice president and general manager of core cloud platforms at what amounts to the future for IBMs software business. Fernandes has been running Red Hats cloud business for nearly a decade, starting with CloudForms and moving through the evolution of OpenShift from a proprietary (but open source) platform to one that has become the main distribution of the Kubernetes cloud controller by enterprises. Meaning those who cant or wont roll their own open source software products.

Timothy Prickett Morgan: Is the edge different, or is it just a variation on the cloud theme?

Joe Fernandes: For Red Hat, the edge is really an extension of our core strategy, which is open hybrid cloud and which is around providing a consistent operating environment for applications that extends from the datacenter across multiple public clouds and now out at the edge. Linux is definitely the foundation of that, and Linux for us is of course Red Hat Enterprise Linux, which we see running in all footprints.

It is not just about trying to get into the core datacenter. Its about trying to deal with the growing opportunity at the edge, and I think its not just important for Red Hat. Look at what Amazon is doing with Outposts, what Microsoft is doing with Azure Stack, and what and Google is doing with Anthos, trying to put out cloud appliances for on premises use. This hybrid cloud is as strategic for any of them as it is for any of us.

TPM: What is your projection for how much compute is on the edge and how much is in the datacenter? If you added up all of the clock cycles, how is it going to balance out?

Joe Fernandes: It is very workload driven. Generally, the advice we always give to clients is that you should always centralize what you can because at the core is where you have the most capacity in terms of infrastructure, the most capacity in terms of your SREs and your ops teams, and so forth. As you start distributing out to the edge, then you are in constrained environments and you are also not going to have humans out there managing things. So centralize what you can and distribute what you must, right.

That being said, specific workloads do need to be distributed. They need to be closer to the sources of data that they are operating upon. We see alignment of the trends around AI and machine learning with the trends around edge, and thats where we see some of the biggest demand. That makes sense because people want to process data close to where it is being generated and they cant they cant incur either the cost or the latency of sending that data back to their datacenter or even the public cloud regions.

And it is not specific to one vertical. Its certainly important for service providers and 5G deployments, but its also important for auto companies doing autonomous vehicles, where those vehicles are essentially data generating machines on wheels that need to have made quick decisions that are as tell.

TPM: As far as I can tell, cars are just portable entertainment units. The only profit anybody gets from a car is all the extra entertainment stuff we add. The rest of the price covers commissions for dealers and the bill of materials for the parts in the car.

Joe Fernandes: At last years Red Hat Summit, we had both BMW and Volkswagen talking about their autonomous vehicle programs, and this year we received an award from Ford Motor Company, who also has major initiatives around autonomous driving as well as electrification. Theyll be speaking at this years Red Hat Summit. Another edge vertical is retail, allowing companies to make decisions in stores to the extent that they still have physical locations.

TPM: I didnt give much thought to the Amazon store where it has something ridiculous like 1,700 cameras and you walk in, you grab stuff, you walk out, it watches everything you do and it takes your money electronically. This is looking pretty attractive this week is my guess. And I thought it was kind of a bizarre two months ago, not shopping as I know and enjoy it. And I know were not going to have a pandemic for the rest of our lives, but this could be the way we do things in the future. My guess is that people are going to be less inclined to do all kinds of things that seem very normal only one or two months ago.

Joe Fernandes: Exactly. The other interesting vertical for edge is financial services, which has branch offices and remote offices. The oil and gas industry is interested in edge deployments close to where they are doing exploration and drilling, and the US Department of Defense is also thinking about remote battlefield and control of ships and planes and tanks.

The thing that those environments have in common is Linux. People arent running these edge platforms on Windows Servers, and they are not using mainframes or Unix systems. It is obviously all Linux and it puts a premium on performance and security, on which Red Hat has obviously made its mark with RHEL. People are interested in driving on open systems anyway, and moving to containers and Kubernetes, and Linux is the foundation of this.

TPM: Are containers a given for edge at this point? I think they are, except where bare metal is required.

Joe Fernandes: I dont think that containers are a prerequisite. But certainly, just like the rest of the Linux deployments, it is going in the direction of containers. The reason is portability, having that same environment to package and deploy and manage at the edge as you do in the datacenter and in the cloud. Bare metal containers can run directly on Linux; you dont need to have a virtualization layer in between.

TPM: Well, when I say bare metal, I mean not even a container. Its Linux. Thats it.

Joe Fernandes: I think that that distinction between bare metal Linux versus bare metal Linux containers is more around do what those packaged as container images, or as something like RPMs or Debian and you need orchestration, do you need orchestrated containers. Right. And again, thats very workload specific. We certainly see folks asking us about environments that are really small, that you might not do orchestration because youre not running more than a single container or a small number of containers. In that case, its just Linux on metal.

TPM: OK, but you didnt answer my question yet, and that is really my fault, not yours. So, to circle back: How much compute is at the edge and how much is on premises or in the cloud? Do you think it will be 50/50? Whats your guess?

Joe Fernandes: I dont think itll be 50/50 for some time. I think in the range of 10 percent to 20 percent in the next couple of years is possible, and I would put that at 10 percent or less because there is just a ton of applications running in core datacenter and a ton running out in the public cloud. People are still making that shift to cloud.

But again, itll be very industry specific. I think the adoption of edge compute using analytics and AI/ML is still now just taking off. For the auto makers doing autonomous vehicles, there is no other choice. It is a datacenter on wheels that needs to make life and death decisions on where to turn and when to brake, and in that market, the aggregate edge compute will be the majority at these companies pretty darn quick. You will see edge compute adoption go to 50 percent or more in some very specific areas, but if you took the entire population of IT, its probably still going to be in the single digits.

TPM: Does edge require a different implementation of Linux, say a cut-down version? Do you need a JEOS-type thing like we used to have in the early days of server virtualization? Do you need a special, easier, more distributed version of OpenShift for Kubernetes? Whats different?

Joe Fernandes: With Linux, the valuable thing is the hardware compatibility that RHEL provides. But we certainly see demand for Linux on different footprints. So, for example, RHEL on Arm devices or RHEL with GPU enablement.

When it comes to OpenShift, obviously Kubernetes is a distributed system, where the cluster is the computer, while Linux is focused on individual servers. What we are seeing is demand for smaller clusters, with OpenShift enabled on three node clusters. Three node clusters, which is sort of the minimum to have a highly available control plane because etcd, which is core to Kubernetes, requires three nodes for quorum. But in that situation, we may put the control plane and the applications run on the same three machines, whereas in a larger setup, you would have a three-node OpenShift control plane and then at least two separate machines running your actual containers so that you have HA for the apps. Obviously those application clusters will grow to tens or even hundreds of nodes. But at the edge, the premium is on size and power, so three nodes might be as much space as youre going to get in the rack out at the edge.

TPM: Either that or you might end up having put your control plane on a bunch of embedded microcontroller type systems and compacting that part down.

Joe Fernandes: Actually, we see a kind of the progression. So there are standard clusters as small as you can get them. So maybe its control plane with one or two nodes. And then the next step weve moved into is a control plane and app nodes are the same three machines. And then you get into what Id call distributed nodes, where you might have a control plane shared across five or ten or twenty edge locations that are running applications and talk back to a shared control plane. You have to worry about connectivity to the control plane.

TPM: If you lose the control plane or your connectivity to it, all it should mean is that you cant change the configuration of the compute cluster at the edge.

Joe Fernandes: Not exactly, because Kubernetes is a declarative system, so it thinks that needs to start up containers on another node or start a new node. In a case where you might have intermittent connectivity, we need to meet to make it more tolerant so it doesnt actually start that process unless it doesnt reconnect for some amount of time. And then the next step beyond that is clusters that have two nodes or a single node, and at that point the control plane, if it exists, is not HA, so youre focusing on high availability some other way.

TPM: You can do virtual machines on a slightly beefier server and have software resilience, but you have the potential of having a hardware resilience issue.

Joe Fernandes: Maybe their resiliency is between edge locations.

TPM: What happens with OpenStack at this point, if anything? AT&T obviously has been widely deploying OpenStack at the edge, with tens of thousands of baby datacenters planned, all linked by and controlled by OpenStack. Is this going to be something like use OpenShift where you can, use OpenStack where you must?

Joe Fernandes: We certainly see Red Hat OpenStack deployed at the edge. Theres an architecture that we put out called the distributed compute node architecture, which customers are adopting. It is relevant that customers virtualized application workloads and also want an open solution, and so I think you will continue to see Red Hat OpenStack at the edge and you continue to see vSphere at the edge, too.

For example, in telco, OpenStack has a big footprint where companies have been creating virtualized network functions, or VNFs, for a number of years and that has driven a lot of our business for OpenStack in telco because a lot of the companies we work with, like Verizon and others, they wanted an open platform to deploy VNFs.

TPM: These telcos are not going to suddenly just decide, to hell with it, and containerize all this and get rid of VMs and server virtualization?

Joe Fernandes: Its not going to be an either/or, but we now see a new wave of containerized network functions, or CNFs, particularly around like the 5G deployment. So the telcos are coming around to containers, but like every other vertical, they dont all switch overnight. Just because Kubernetes has been out for five years now doesnt mean the VMs are gone.

TPM: Is the overhead for containers a lot less than VMs? It must be, and that must be a huge motivator.

Joe Fernandes: Remember that the overhead of a VM includes the operating system that runs inside the guest and the overhead of a container, where you are not virtualizing that hardware, you are virtualizing just the process. You can make a container as small as the process that it runs. And for a VM, you can only make it as small as the operating system.

TPM: We wouldnt have done all this VM stuff if we could have just figured out containers to start with.

Joe Fernandes: You know, Red Hat Summit is coming up in a few weeks and we will be providing an update on KubeVirt, which allows Kubernetes to manage standard virtual machines along with containers. In the past year or more, we have been talking about it strictly in terms of what we are doing in the community to enable it. But it has not been something that we can sell and support. This is the year, its ready for primetime and that presents an opportunity to have a converged management plane. You could have Kubernetes directly on bare metal, managing both container workloads and VM workloads, and also manage the transition as more of those workloads move from VMs to containers. You wont have to switch environments or have that additional layer and so forth.

TPM: And I fully expect people to do that. Ive got nothing against OpenStack. Five years ago, when we started The Next Platform, it was not obvious if the future control plane and management and compute metaphor would be Mesos or OpenStack or Kubernetes. And for a while there, Mesos looked like it was certainly better than OpenStack because of some of the mixed workload capabilities and the fact that it could run Kubernetes better than OpenStack could. But if you can get KubeVirt to work and it provides Kubernetes essentially the same functionality that you get for OpenStack in terms of managing the VMs, then I think were done. It is emotional for me to just put a nail in the coffin like that.

Joe Fernandes: The question is: Is it going to put a nail not just in OpenStack, but in VMware, too.

TPM: VMware is an impressive legacy environment in the enterprise, and it generates more than $8 billion in sales for Dell. There is a lot of inertia with legacy environments I mean, there are still System z mainframes out there doing a lot of useful work and providing value to IT organizations and their businesses. I have seen so many legacy environments in my life, but this may be the last big one I see this decade.

Joe Fernandes: You have covered vSphere 7.0 and Project Pacific and look at the contrast in strategy. Were taking Kubernetes and trying to apply it to standard VM workloads as a cloud native environment. What VMware has done is take Kubernetes and wrap it back around the vSphere stack to keep people on the old environment that theyve been on for the last decade.

Read the original post:
How Edge Is Different From Cloud And Not - The Next Platform

No New Normal: Building the Commons – Resilience

Author and archdruid John Michael Greer talks about catabolic collapse, not as the guns & ammo, post-apocalyptic-yet-still-powered-by-capitalism scenariofavored in the media, but as an ongoing process of societal disintegration. Looking at our mainstream institutions, economics and beliefs, its clear that weve been collapsing for a while. A pandemic punctuates the catabolic curve with an eye-popping shock set against systemic processes bedrocked as background, never foreground.

The etymology of apocalypse points to anunveiling, dropping illusion and findingrevelation. As our global production systems and social institutions (eg. healthcare, education) are suddenly overwhelmed, their basic unsuitability is exposed. Just weeks ago so mighty, economies now sputter when faced with this latest adversity, and this sudden spike in the process of collapse portends a larger undertaking in ecological and social entropy. As Covid-19 takes its human toll worldwide, weve begun to see the best and worst of humanity in its choice of loyalties, whether to human life or to economic systems, and the power struggles in finding the right balance (if such a thing exists). Its another opportunity to consider, what is inherent in us as people, and what is the product of our systems? Growing up in systems preaching that greed is good, that the only social responsibility of businesses is to increase profits, or that there is no alternative, its no surprise that the worst reactions to the crisis are marked by individualism, paranoia and accumulation.

Image by Sam Wallman and Miroslav Sandev

Natural systems are rebounding because pollution and emissions are down, but its impossible to fist-pump about this while people are suffering, dying, or working beyond capacity to save lives. In fact, its a good time to question the very validity of work: which services are essential, how to use our free time. What solutionscan the market offerto the health crisis, to overcrowded hospitals, to breaks in supply lines of essential goods and services? To those unable to meet their rent, mortgage or future expenses? Some claim ourglobal, industrialized model is to blame for the virus, others cry that the cure is worse than the disease, that the economic effects of quarantining will create more destruction than the virus itself.

These predictions are not endemic to economic science, but to a history of accumulatory, command and control dynamics which, via longstanding institutions including patriarchy and colonialism, have found their apex incapitalist realism: the widespread sense that not only iscapitalismthe only viable political and economic system, but also that it is now impossible even to imagine a coherent alternative to it. Short a few weeks of predatory feeding, the growth-based model shows its weakness against the apocalypse. Another veil is lifting.

What else can we see? Whatwill the world look likewhenever this is over (and how will we know when it is)?

Could this be the herald of another political economybased on abundance, not scarcity and greed? We can help nature to restore itself, cut down emissions, our consumption of mass manufactured and designed-to-break-down crap. We can radically curtail speculative ventures and fictitious commodities. Slash inequality from the bottom up, spend our time away frombullshit jobstoreimagine the world. Use this free time to reconnect, cherish our aliveness, break out of containment, care for each other, grieve what weve lost and celebrate what we still have.

We do have the frameworks, we have been creating this capacity for quite a while. Here I refer tothe Commons. Simply put, the Commons are living systems to meet shared needs. As old as humanity itself and as new as the latest trends in decentralized technology, the Commons are best understood as a verb, not a noun; more action than static. A commons needs three elements:

Examples includecooperatively managed forests,water distribution irrigation systems,social currencies,Free/Libre and Open-Source Software,self organized urban spaces,distributed manufacturing networksandso much more. As George Monbiotdescribes, the most inspiring and effective reactions to the Covid-19 crisis are not coming from markets or states, but from the Commons. Often invisibilized, the practices of the commons offer fairer economic and human frameworks to meet our needs, especially in challenging times.

From localized yetglobally connected systems of productionthat can rapidly respond to urgent needs without depending on massive global chains, to ways to organize the workforce intorestorative and purpose-oriented clusters of peoplewho take care of each other. This new economy will need a new politics and a more emancipated relation to the state:we have tried it and succeeded. What new worlds (many worlds are possible) can we glimpse under this lifted veil?

We Must Reimagine Everything was originally published in Spanish by Miguel Brieva Clismn. It was translated to English by Guerrilla Translation

Heres a question: did you already know about these potentials? Are we still having this conversation among ourselves, or have these terrible circumstances gifted us with anopportunity for (apocalyptic) clarity? The normal is collapsing, while our weirdness looks saner than ever before.

Timothy Leary famously called for us to find the others. I think that the others areallof us, and this may be the moment where more of us can recognise that. A few years ago, we createdan accessible, easy to use platformto share the potential of the Commons with everyone. Today its more relevant than ever. The projects we work on (Commons TransitionandDisCO) are based on two simple precepts:

This is why we strive to create accessible and relatable frameworks for people to find the commoner within themselves. But we need to grow out of our bubbles, algorithmically predetermined or not; we need to rewild our message beyond the people who already know. Movements like Degrowth, Open Source software and hardware, anti-austerity, Social Solidarity Economy, Ecofeminism, Buen Vivir we are all learning from each other. We must continue to humbly and patiently pass the knowledge on, listen to more voices and experiences, and keep widening the circle to include everyone, until thereare no others.

Please share this article with anyone who may benefit from these crazy ideas that suddenly dont look so crazy anymore.Start a conversationwith people who, aghast at the rapid collapse and lack of reliable systemic support, are eager for new ideas, solutions, and hope. The greatest enclosure of the commons is that of the mind: our capacity to imagine better worlds, to be kinder to each other and to the Earth. This will not be an easy or straightforward process. We need to hold each other through the loss and pain. We need to keep finding the others among all of us, until there are no more.

Teaser photo credit: Photograph by Antonio Marn Segovia, CC by 2.0

Continued here:
No New Normal: Building the Commons - Resilience

FBI expects a rise in scams involving cryptocurrency related to the COVID-19 pandemic – The Highland County Press

Fraudsters are leveraging increased fear and uncertainty during the COVID-19 pandemic to steal your money and launder it through the complex cryptocurrency ecosystem.

People of all ages, including the elderly, are being victimized by criminals through cryptocurrency-related fraud schemes. Developments in cryptocurrency technology and an increasing number of businesses accepting it as payment have driven the growing popularity and accessibility of cryptocurrency.

There are not only numerous virtual asset service providers online but also thousands of cryptocurrency kiosks located throughout the world which are exploited by criminals to facilitate their schemes. Many traditional financial crimes and money laundering schemes are now orchestrated via cryptocurrencies.

The FBI advises you to be on the lookout for an increase in the following cryptocurrency fraud schemes related to COVID-19:

Blackmail Attempts. Threatening emails or letters in which scammers claim to have access to your personal information or knowledge of your dirty secrets and demand payment in Bitcoin to prevent release of this information have been circulating for years. With the advent of COVID-19, there is a new twist on this scam. The correspondence claims that the writer will both release your information and infect you and/or your family with coronavirus unless payment is sent to a Bitcoin wallet.

Work from Home Scams. Scammers, posing as employers, may ask you to accept a donation of funds into your own bank account and to deposit them into a crypto kiosk. The so-called donation is likely money stolen from others. Your acceptance and transfer of the stolen money is considered illegal money mule activity and potentially unlicensed money transmission.

Paying for Non-Existent Treatments or Equipment. Scammers have been known to lure customers from trusted e-commerce sites offering products that claim to prevent COVID-19 onto unrelated and unregulated messaging sites to accept payment in cryptocurrencies for products that do not actually exist.

Investment Scams. Criminals often pitch fraudulent investments in a new and developing cryptocurrency, such as an initial coin offering (ICO) or other investment vehicle to take a victims money. These scams typically involve scenarios that seem too good to be true offering large monetary returns for a short-term, small investment. The reality is that scammers steal the investment money for personal use and utilize the complexities of cryptocurrency to hide the true destination of the stolen funds.

Although there are legitimate charities, investment platforms, and e-commerce sites that accept payment in cryptocurrency, pressure to use a virtual currency should be considered a significant red flag.

By remembering the following tips regarding finances and cryptocurrency, you can better protect yourself from fraud:

Verify that a vendor/charity is legitimate and accepts cryptocurrency before sending payments/donations.

Conduct extensive research on potential investment opportunities.

Do not use your personal bank accounts for work-from-home business-related activity or provide your bank account information to someone who is not named on the account.

Contact law enforcement before paying out blackmail and/or extortion attempts and before converting your money into cryptocurrency to pay them.

The FBIs Criminal Investigative Division has an entire team dedicated to preventing and combating cryptocurrency money laundering and frauds. If you believe you are the victim of a fraud, or if you want to report suspicious activity, pleasecontact your local field officeor visit the FBIs Internet Crime Complaint Center atic3.gov.

See the article here:
FBI expects a rise in scams involving cryptocurrency related to the COVID-19 pandemic - The Highland County Press

TOP-20 Cryptocurrency and Bitcoin (BTC) Wallets in 2020 – U.Today

A cryptocurrency wallet refers to a tool created to store, manage andtransfer crypto token funds. Which style of crypto wallet serves you best?

Firstly, the best crypto wallet has very little in common with wallets, in the traditional sense: no assets are kept directly insidethe Bitcoin (BTC) wallet, neither in their classic form nor asbinary code.

A cryptocurrency wallet (crypto wallet, Bitcoin (BTC) wallet) represents a special class of computational frameworks or physical device able to authorize blockchain-based transactions. In a nutshell, it may send, receive, and operate the digital assets. The best crypto wallet reliably secures keypairs to protect cryptoinvestments.

People obtain software crypto wallets by activating an account (web client, cloud storage) or through the installation of the crypto wallet software on a desktop or portablecomputer. Also, the best crypto wallet could be developed manually, but thisrequires professional skills in Python coding.

Holders also create Hard wallets by setting up a digital gadget (hardware)and printing a keypair on paper or by engraving it on a metal plate.

When acrypto wallet is active, the user has its public key. The blockchain uses this to identify the unique address of thetoken holder. Everyone who has a public key can authorize a crypto remittance to a certain address with the sum and the name of the token required. The confidentiality of the private key is of crucial importance,the unveiling of which would put the users riches at risk.

Hardware crypto wallets constitute a high-end class of electronic gadgets that consistof a core processor, touchpad and display. It defends the integrity of keypairs and authorizesblockchain transactions. This gadget needs a USB or Bluetooth connection to your computer in order to process operations.

A software crypto wallet refers to a computational environment utilized to operate your tokens. Typically, it represents an in-browser cloud interface or downloadable program. The most secure crypto wallet develops the keypair and allows the token holder to store it in extreme confidentiality. Typically, the best crypto wallet in thisclass allows traders to conduct multi-chain token conversion.

Hardware crypto wallet

Software crypto wallet

What is it?

Portable gadget

Computer program

Is it secure?

Extremely secure

May vary from low to very high

How much does it cost?

Starts at $49

Typically, you can download onefree of charge

Is it too sophisticated for non-CS users?

No

Some software wallets have very sophisticated UX/UI

Mostly, it'sdesigned for...?

Hodlers with low diversityin their portfolio and a big sum of savings

Daily and weekly traders with a well-diversified portfolio and a medium amount of savings

The best software cryptocurrency wallets have intuitive UI/UX, a high level of security and powerful functionality.

Blockchain.com cross-chain wallet is a veteran of this market as it was proudly presented in 2011. As of today, it has been downloaded 47 million times. This giant operates $200 billion in USD equivalent. It gainedpopularity as aBitcoin (BTC) wallet, but the communities of Ethereum (ETH), Bitcoin Cash (BCH) and Stellar Lumens (XLM) can alsoenjoy its functionality. Also, it has block explorers for Bitcoin (BTC) and Ethereum (ETH) blockchains.

Coinbase is the second mogul of our industry. It was establishedin 2012 in the U.S. by Brian Armstrong and Fred Ehrsam. It pioneered the sphere of direct credit card deposits. Now it supports the crypto king as well as altcoins: Ethereum (ETH), XRP, Bitcoin Cash (BCH), Litecoin (LTC), Eos (EOS), Tezos (XTZ) and its native stablecoin USD Coin (USDC). Furthermore, it delivers processing services to corporate entities.

Opera portable computercrypto wallet is a pioneering native browser app for iOS and Android devices. The flagship currency, Ethereum (ETH) and Ethereum-fam ERC tokens are in its toolkit. This solution allows Opera to support online payouts in cryptocurrency, to browse Ether-based dApps and to send money between wallets.

TheVexel team deliversa one-stop-shop ecosystem for crypto remittances fueled by fiat gateway. Besides theflagship coin, Ethereum (ETH), Bitcoin Cash (BCH), Dash (DASH) and Litecoin (LTC) are welcome. Visa and Mastercard transfers are supported with reasonable fees (1.2%). Users can operate their funds through both mobile and web apps as well as with a Telegram bot. Unique Vexels (anonymous cryptocurrency vouchers) allow merchants to build a Blockchain-as-a-Service-powered payment system for their B2B and B2C entities.

Try Prizm, a one-size-fits-all multi-blockchain software wallet with a web client, crypto wallet app for Apple, Windows and Android OS. Prizm processes the flagship currency and nine major altcoins including itsnative PRIZM (PZM). Prizm wallet is characterized by an attractive interface: it allows users to launch numerous accounts in a single click, to generate a remittance QR-code. The Prizm team created a cutting-edge security toolkit: Face ID, Touch ID, seed phrases.

Guarda Wallet is one of themost secure crypto wallets: 45 high-demandprotocols and tons of tokens are on board. It works as a multi-platform crypto wallet app andGoogle Chrome plug-in. A full range of operations with Bitcoin (BTC), Ethereum (ETH), Monero (XMR), Litecoin (LTC) and many other types of assets are on Guarda Wallet'sservices. Guarda Wallet integrates the Ledger hardware wallet (see below - U.Today) and pioneered multi-sig authorization of Bitcoin (BTC) transactions.

Wallet

Date Launched

Interfaces

Custodial / non-custodial

Killer feature

Blockchain.com

2011

Web client, mobile crypto wallet.

Custodial

Native BTC and ETH blockchain explorers

Coinbase.com

2012

Web client, mobile crypto wallet

Custodial

Operations for corporate clients

Opera

2018

Browser wallet

Custodial

First browser with a built-in wallet

Vexel

2019

Web client, mobile crypto wallet, Telegram bot, desktop apps

Non-custodial

Telegram bot, anonymous crypto vouchers

Prizm

2017

Web client, mobile crypto wallet, desktop apps

Non-custodial

Paramining passive income program

Guarda Wallet

2017

Web client, mobile crypto wallet, desktop apps

Non-custodial

Multi-signature authorization of BTC transactions

The best hardware crypto wallets ensure the most secure crypto wallet safety for meaningful crypto savings. If this class of electronic instruments seems attractive for your strategy, look through our detailed guide on hardware wallets.

Ledger Nano X is a high-end gadget with 100+ applications on-boarded that accesses a wide rangeof currencies. One of the most secure crypto wallets in this market, Ledger Nano X supports 1250+ ERC-20 tokens and all of the behemoths you may need - digital gold, the second blockchain, XRP, Bitcoin Cash (BCH), as well as Stellar Lumens (XLM) and EOS.

Trezor Model T was announced by Satoshi Labs in 2018. Equipped with a high-res color touchscreen display, its interface is strikingly different from all competitors. This device is the perfect storage for Bitcoin (BTC), Litecoin (LTC), Dash (DASH), Ethereum (ETH), Bitcoin Cash (BCH), XRP, Monero (XMR) and Cardano (ADA).

Trezor ONE crypto hardware wallet represents the initial stage of Trezor devices announced six years ago. Bitcoin (BTC), Litecoin (LTC), Dash (DASH), Ethereum (ETH), Bitcoin Cash (BCH) and a plethora of Ethereum-based ERC-20 altcoins can be easily accessed by its users.

Ledger Nano S crypto hardware wallet tracks its history back to June 2016. Ledger Nano S works with Bitcoin (BTC) and all of the most-demanded altcoins: Ethereum (ETH), XRP, Bitcoin Cash (BCH), Litecoin (LTC), Dash (DASH). It merges the performance of 3 - 7 applications to process the assets of top-demanded blockchains simultaneously.

Simply put, the best hardware cryptocurrency wallets are those from theindustry leaders, Ledger and Trezor. Ledger Nano X andTrezor Model T are first-choice tools for the old hands while their smaller brothers, Ledger Nano S and Trezor ONE offer more common solutions for the average Joe.

And to close this list off,here is thedigest of thetop crypto wallets U.Today preparedone year ago:

More here:
TOP-20 Cryptocurrency and Bitcoin (BTC) Wallets in 2020 - U.Today

Cryptocurrency Day Trading – Everything You Need to Know as a Beginner – Crypto Daily

As you already know, the cryptocurrency market is a very busy one; hence, there are frequent changes due to its volatility.

For this reason, it might be difficult for beginners to penetrate as they have a lot of strategies to learn. Among the various cryptocurrency trading strategies, day trading happens to be the most common. You may have heard about the term before without knowing what it means.

In this post, we will show you everything you need to know as a beginner to cryptocurrency day trading. But before we proceed, lets answer the basic question what is day trading.

From the words day and trading, you should have a clue about what the term implies. Cryptocurrency day trading involves buying cryptocurrencies and selling them off on the same trading day.

This trade strategy is not unique to the cryptocurrency market alone. It exists in all other markets, such as forex, stocks, commodity, and bond markets. However, it is different in the cryptocurrency market because it is always live; in fact, the cryptocurrency does not close.

Crypto day traders approach the market in various ways; hence, there are different types of day traders. Of the lot, the most common ones are Technical Analysts and the Speculators.

Technical analysts are traders that trade based on current happenings in the market. They dont get moved by the ups and downs displayed outside. Their interest is placed on financial charts, most notably from the past, as it helps them to predict future cryptocurrency prices.

In essence, most of the professional cryptocurrency day traders are technical analysts; moreover, it requires a high level of practical experience. At the same time, it is not an easy venture. Understanding cryptocurrency trading charts - bars, lines, candlesticks - requires excellent technical skills and cryptocurrency knowledge.

Basically, spectator day traders are observers of the cryptocurrency market. Spectator day traders are concerned about happenings outside of the market. Conversely, several external forces affect the cryptocurrency market, such as political and economic activities. Spectators follow the news and trade based on the information they get.

The method of day trading to follow depends entirely on the trader. But as a beginner, here are some steps you should follow before you start cryptocurrency day trading.

Depending for how long you want to stay in your trades, you should select a cryptocurrency exchange or a crypto broker. Focus on trading fees, minimum deposit, availability of coins you want to trade and other import aspects. If you are based in the UK, you can go for a popular exchange like Binance or Kraken or a cryptocurrency broker like IQ Option which is too available in the UK. Crypto brokers are for traders who want to stay in a trade for a couple of minutes or hour tops. Crypto exchanges for traders who want to stay in their trades for a longer time period.

Conversely, there are centralized and decentralized exchanges. Both are ideal, but if you prefer staying anonymous on the net, decentralized exchanges are the perfect option. In fact, you have to identify yourself in most of the centralized exchanges. Nevertheless, you can utilize more than one cryptocurrency exchange for day trading. Mostly, day traders utilize 3 to 4 platforms. Its tolerable so long as you can efficiently manage them.

Before starting any day trade, ensure that you have a plan to follow. Your plan should include the coins you wish to trade, trading amounts, chart pattern setups, and more. As a result, this would require some research.

As a beginner, your stop-loss limit and limit sell order is an integral part of your day trading plan. These determine when you exit a trade at a low or high point. They are set automatically.

Also, cryptocurrency day traders utilize crypto trading bots in their plans. This can be worthwhile but should be done with caution.

Whether you choose to be a technical analyst or a spectator, it is recommended that you carry out market analysis. For day trading, you must understand the trading charts even if not at 100%.

The trading time range is a short one, so you should act quickly on price trends and changes as they happen in real-time.

This is what every trade is all about - making profits. You need to pay attention to the ideal time for cashing out and making profits. The trade value mustnt reach your limit sell order before you get the profits.

In cryptocurrency day trading, traders often aim at making small profits via each trade. At the end of the day, the accumulated profits of each trade become the main profit. Remember the exchange fees you have to pay for each trade. Making profits from your cryptocurrency day trading also involves determining loss. Encountering small losses on each trade can result in a significant decline at the end of the trading day.

As a beginner to cryptocurrency day trading, you must spend time studying and understanding the daily changes in the crypto market. Not to mention, you need to get the experience, and this requires time.

However, there is no specific way of cryptocurrency day trading. Besides, crypto day traders make use of various trading strategies, and you can come up with your individualized plan. Nonetheless, you can get started by applying the steps mentioned above.

Originally posted here:
Cryptocurrency Day Trading - Everything You Need to Know as a Beginner - Crypto Daily

BitHull Cryptocurrency Miners on Promotional Offer, Business Insider – Business Insider

COPENHAGEN,DENMARK EQS Newswire April 9, 2020BH Miner and BH Miners Box, two recently launched cryptocurrency miners from BitHull S.A (www.BitHull.com), are now available for a lucrative promotional offer. Anyone purchasing three of these extraordinarily profitable FPGA miners is now entitled to receive one more miner absolutely free of cost. This promotional offer will remain active until April 30.

BH Miner is the basic product from BitHull S.A that has been designed specifically for the newbies looking to try their hand in crypto mining. BH Miners Box, on other hand, is a combination of six BH Miner with serious profit making potential. Both these miners can be used for mining Bitcoin, Litecoin, Ethereum, and Monero.

Both Products can generate Return of Investment within one month and they are delivered pre-configured to customers, so is not required any level of knowledge to start mining and generate profits within one month.

We are pleased to announce that anyone purchasing three or more miners from us will now receive one miner for no additional cost whatsoever. This promotion is specifically for the mining enthusiasts looking to build an extremely profitable home based mining business, said Matias Milet, Vice President of BitHull S.A.

For more details, please visithttps://www.bithull.com/

BitHull S.A is a technology company dedicated to developing next-generation hardware for cryptocurrency mining. The company is run by a team of experts with a track record of delivering world-class tech components such as FPGA chips to numerous industry heavyweights.

Follow this link:
BitHull Cryptocurrency Miners on Promotional Offer, Business Insider - Business Insider

Some countries in the Middle East are using artificial intelligence to fight the coronavirus pandemic – CNBC

View of an empty street amid the COVID-19 pandemic in Doha, capital of Qatar, on April 13, 2020.

Nikku | Xinhua News Agency via Getty

Countries in the Gulf Cooperation Council are stepping up their use of artificial intelligence tools to halt the spread of the coronavirus pandemic.

Governments throughout the GCC a group of countries in the Middle East that includesBahrain, Saudi Arabia, Qatar, Oman, Kuwait and the United Arab Emirates have enacted some of world's strictest measures, including suspending passenger flights and imposing curfews on citizens to put brakes on the number of new cases of Covid-19 that currently totalover 2 million (2,064,115)globally, according to Johns Hopkins University data.

But countries aren't restricting their efforts to simply imploring their residents to stay locked in and shutting down all but the most essential of businesses.

They are increasingly deploying sophisticated technology to ensure that movement is limited and social distancing is in place through the use of speed cameras, drones and robots.

By applying location-based contact tracing, governments can monitor those who have tested positive for coronavirus, and try to limit their exposure to the population.

AI's ability to crunch large amounts of data has allowed governments worldwide to collect information to try and stop the pandemic. Contact-tracing has allowedHong Kong, China and Singapore to monitor cases.

While governments and companies grapple with what could be a controversial violation ofprivacy issues, many countries have found it to be the key to lifting lockdown measures.

In Bahrain, an application called 'BeAware' allows residents to track proximity to someone with Covid-19. The application uses location data to alert individuals in the event they approach an active case.

"BeAware registration is mandatory for those in quarantine, while non-quarantined cases may choose to register," Mohammed Ali AlQaed, chief executive of Information & eGovernment Authority in Bahrain told CNBC.

Bahrain has reported1,671cases according to Hopkins data, and was one of the first to begin easing restrictions, allowing some stores and malls to reopen.

AI can also help businesses work more efficiently throughout the pandemic.

Majed M. Al Tahan, co-founder & MD of Danube Online told CNBC the Saudi-based hypermarket and supermarket chain is using AI to minimize delivery time.

Using 'aisle-mapping' technology, packers can locate items in an online customer's order, which are tracked around stores using an app.

Saudi Arabia extended its curfew indefinitely on Sunday and the country remains in total lockdown. Saudi Arabia has reported the highest number of cases in the GCC5,862 on Thursday,according to Hopkins data.

A Qatari Government communications spokesperson told CNBC the government is working with the Qatar Computing Research Institute on a diagnostic monitoring app,connected to a ministry of health database that uses computing and geolocation services to help diagnose and track Covid-19 cases. According to Hopkins data,Qatar has reported3,711cases of coronavirus to date.

In the United Arab Emirates, the government is using AI to limit the movement of Dubai residents, the UAE's most densely-populated emirate and home to 3.3 million people.

Dubai police are monitoring permits required by residents leaving their homes in the region's business hub.

Dubai Police use a program called 'Oyoon' which, through a network of cameras in the city uses facial, voice and license plate recognition. The information is fed through a large database and the computer can cross-reference and analyze the data to determine, in this instance, if aresidentis employedin a vital sector or in possession of a valid permit.

The United Arab Emirates has reported5,365cases of coronavirus, according to Hopkins.

UAE-based healthcare startupNabta Healthwill use AI to provide risk and symptom assessments for Covid-19. Co-founder Sophie Smith told CNBC that advanced technologies such as AI, applied machine learning and blockchain could help alleviate the effects of future pandemics.

"When the dust settles, people will look at this pandemic and say 'we are only as strong as our lowest common denominator, and that's people with underlying health conditions,'"Smith said.Nabta Health uses AI to diagnose those very conditions.

View original post here:
Some countries in the Middle East are using artificial intelligence to fight the coronavirus pandemic - CNBC

How The White House Guidance For Regulation Of Artificial Intelligence Invites Overregulation – Forbes

Excessive top-down federal funding and governance of scientific and technology research will be increasingly incompatible with a future of lightly regulated science and technology specifically, and with limited government generally.

Neither political party takes that view though. In a rule-of-experts, send-the tax-dollars-home environment, America risks becoming vulnerable to industrial policy and market socialist mechanisms as frontier technologies become more complex.

Addressing infrastructure and other broad initiatives a year ago in his February 5, 2019, State of the Union address, for example, president Donald Trump called for legislation including investments in the cutting edge industries of the future and proclaimed, This is not an option, this is a necessity.

AI, Artificial Intelligence concept,3d rendering,conceptual image.

Along with such spending having thick strings attached and accompanying regulatory effects that propagate, it is not proper for the sciences nor practical applications of them to proceed walled off from one another in the arbitrary legislative appropriations and regulatory environments that prevail in Washiington.

Artificial intelligence in particular serves as a case study or warning. Emblematic was Executive Order 13859 of February 11, 2019 on Maintaining American Leadership on Artificial Intelligence and the establishment of the AI Initiative, which were followed by the March 19, 2019 launching of the federal hub AI.gov (now whitehouse.gov/ai).

Executive orders are not law, but they can influence policy, and this one promotes sustained investment in AI R&D in collaboration with industry, academia, and other doings.

E.O 13859 also calls for federal collection of data among other centrally coordinated moves. Actions shall be implemented by agencies that conduct foundational AI R&D, develop and deploy applications of AI technologies, provide educational grants, and regulate and provide guidance for applications of AI technologies.

Whew. This federalization is concerning on its own, but it occurs in an environment in which much AI research at the federal level happens under the auspices of the Department of Defense.

Bet you didnt know that the Pentagon, on the very day after Trumps 2019 AI executive order, released its own AI strategy, subtitled Harnessing AI to Advance Our Security and Prosperity, describing use, plans, and ethical standards in deployment. There are now new promises by DoD to adopt rules for how it develops and uses AI.

But where, indeed, is the only spot where a definition of AI is codified in federal statute? In the John S. McCain National Defense Authorization Act for Fiscal Year 2019.

When it comes to robotics and military, the concern is that Isaac Asimovs famous Laws of Robotics (devised to forbid the harm of humans) are programmed out, not in. This is a part of what makes fusion of government and private AI deployment problematic. Where a tech titans one-time motto had been Dont Be Evil, a fitting admonition now for the technology sector as a whole is:

Dont Be Government.

The most recent development is the White House Office of Management and Budgets 2020 Guidance for Regulation of Artificial Intelligence Applications, directed at heads of federal executive branch agencies. In fulfillment of Trump E.O. 13859 and building upon it, the January 2020 document at first blush strikes the right tone, aiming at engaging the public and forbearance, limiting regulatory overreach, eliminating duplication and redundancy across agencies, improving access to government-data and models, recognizing that a one-size regulatory shoe does not fit all, using performance based objectives rather than rigid rules, and in particular, avoiding over-precaution. For example, the guidance on p. 2 instructs:

Agencies must avoid a precautionary approach that holds AI systems to such an impossibly high standard that society cannot enjoy their benefits.

The OMBs Request for Comments on the Guidance at one point seems to adopt the same reasoned laissez-faire stance: OMB guidance on these matters seeks to support the U.S. approach to free-market capitalism, federalism, and good regulatory practices (GRPs).

Michael Kratsios, Chief Technology Officer of the United States, called the Guidance the first-of-its-kind set of regulatory principles to govern AI development in the private sector to address the challenging technical and ethical questions that AI can create.

But make no mistake, the new AI guidance constitutes a set of regulatory principles, especially as they will be interpreted by less market-oriented administrations that later assume the helm.

The Guidance states:

When considering regulations or policies related to AI applications, agencies should continue to promote advancements in technology and innovation, while protecting American technology, economic and national security, privacy, civil liberties, and other American values, including the principles of freedom, human rights, the rule of law.

The guidance mentions American values five times, without recognizing the degree of incompatibility of the top-down administrative state form of governance that now prevails, as distinct from Article I lawmaking, with those values.

Nor is there sufficient appreciation of the extent to which the regulatory bureaucracy can hold conflicting visions of rule of law. Todays administrative state has its own set of value pursuits and visions, of what are costs and what are benefits, and the sources of each. As such, the administrations AI Guidance contains elements that can be exploited by creative agencies seeking to expand once the ostensibly less-regulatory Trump administration has left the state.

The AI Guidance correctly states: The deployment of AI holds the promise to improve safety, fairness, welfare, transparency, and other social goals, and Americas maintenance of its status as a global leader in AI development is vital to preserving our economic and national security.

But on the other hand, the Guidance (p. 3) says AI applications could pose risks to privacy, individual rights, autonomy, and civil liberties that must be carefully assessed and appropriately addressed.

Well thats interesting. Governments, as post-9/11 and more recent surveillance history shows not the institution of orderly, competitive free enterprise are the primary threat to these very values; so opening the door too far to agencies misidentifies sources of values problems, and lays bedrock for counterproductive and harmful regulation.

Unfortunately, agencies wanting to be granted the legitimacy necessary to throw their weight around on the new and exciting AI playground have been needlessly invited to do so by the Guidance.

For example, in evaluating benefits and costs of regulatory alternatives, agencies are to (p. 12) evaluate impacts to equity, human dignity, fairness, potential distributive impacts, privacy and civil liberties, and personal freedom.

These bureau-speak formulations and directives plainly favor agency governmental proclivities moreso than they defer to the competitive process and non-governmental resolutions of the inevitable difficult issues that will naturally arise from the proliferation of AI.

Unless externally restrained, a regulatory bureaucracys inclination is to answer the question, Is there call for regulation? in the affirmative. The Guidance invites agencies (p. 11) to consider whether a change in regulatory policy is needed due to the adoption of AI applications in an already regulated industry, or due to the development of substantially new industries facilitated by AI.

Why would the Trump adiministration open this Pandoras Box? As a wholly blank canvas, this approach to AI policy will prove an irresistable unleashing of the bureaus. Trumps regulatory reduction Task Forces notwithstanding, there exists no permanent Office of No anchored at any agency to vigorously resist to top-down discretion and reject the more appealing heavy Washington influence they are invited to proffer.

The unfortunate iron law that industry generally prefers regulation that advances its interests and walls out competition will prove true of AI regulation specifically: Companies cannot just build new technology and let market forces decide how it will be used, said one prominent CEO in January 2020.

Companies may dislike like the kind or regulation that makes them ask Mother-may-I? before they take a risky step. But on the other hand, established playersespeccially given the head start of the government contracting and military presence in AIwill appreciate federal approaches that just so happen to forestall those nettlesome upstarts with a different idea, even when those new ideas advance safety or accountability.

Here are a few additional concerns with federal AI Guidance at this stage.

Too frequently there occurs misdiagnosis and denial regarding the root source government itself of frontier technologies risks. The OMB guidance (p. 6) calls on agencies to encourage the consideration of safety and security issues throughout the AI design, development, deployment, and operation process. But the government is more prone to undermine security-enhancing encryption used in private sector applications, for example. And, especially given the heavy government collaborative role sought, to indemnify winner companies when things go wrong and thereby mangle risk-management mechanisms like insurance and containment in AI ecosystems.

Since the administrtions AI proclamations belong in the regulatory rather than deregulatory camp, it is good that strong AI (the potentially sentient, self-improvingversion) is ostensibly not addressed (exempted) by the Guidance. Fortunately, the Guidance acknowledges that (p. 11) current technical challenges in creating interpretable AI can make it difficult for agencies to ensure a level of transparency necessary for humans to understand the decision-making of AI applications. Indeed, agencies cannot do this; no one can; it is the very nature of black box machine learning. But it is a sure bet that agencies would seize this authority anyway, made apparent in some of the bullets above.

The AI guidance appears in a policy climate in which Republicans and Democrats alike seek major government funding of science generally, an environment replete with proposals that have marinated in the regulatory, administrative state frameworks up to and including a manufacturing czar, and quasi -military terminlogy such that energy security gets equated with national security. AIis vulnerable to all this. Internationally, governments are moving toward regulation of AI; and the U.S., by these new actions, has demonstrated readiness to do so as well.

This state of affairs is not particularly the fault of well meaning policymakers within the White House, but results from the fact that there exists no audience or consituency for keeping governments hands out of complex, competitive free enterprise generally. The disruptions purportedly to be caused by AI create irresistable magnets for the opportunistic and cynical to pursue regulation.

Unfortunately in part due to Trumps order and related/derivative guidance yet to come, we can predict that future administrations and legislators will expand government alliances with a subset of private sector winners, perhaps even a sort of cartelization. The legitimization of this concept at the top by an ostensibly deregulation-oreinted president will make it harder for our decendents to achieve regulatory liberalization and maintain any separation of technology and state in future complex undertakings, many of which will be AI-driven.

In similar vein and illustrative of the concerns raised here, the establishment of a Space Force, enacted in the National Defense Authorization Act of 2020, presents the same lock-in of a top-down federal managerialism of private sector undertakings, given that commercial space activities have hardly taken root beyond NASA contractors and partners. Making the (AI-driven) force asixth branch of the armed forceswill inevitably alter freedoms and private commercial space activities, heavily influencing technology investment and evolution in a sector that barely exists yet. The Space Force move had already been preceded by a presidential directive on space traffic management complete with tracking, cataloging, and data sharing with government. It is worth remembering that most debris in space used to justify calls for regulation is there thanks to the NASA legacy, not private entrepreneurs who would have needed to ponder property rights in sub-orbital and orbital space in a different way. Even though normalizing commercial space activies for a diverse portfolio of actors and approaches is not compatible with heavy regulation, the role of competitive discipline may yet be improperly overlooked or squelched.

So the AI Guidance is by no means making an appearance in a policy vacuum, which is not altogether encouraging. In similar vein, an October 2019 executive order established a new Presidents Council of Advisors on Science and Technology to strengthen .... the ties that connect government, industry, and academia. This project entails collaborative partnerships across the American science and technology enterprise, which includes an unmatched constellation of public and private educational institutions, research laboratories, corporations, and foundations, [by which] the United States can usher extraordinary new technologies into homes, hospitals, and highways across the world. Even this appeared in the wake of E.O. 13,885 on Establishing the National Quantum Initiative Advisory Committee, aimed at implementing the 2018 National Quantum Initiative Act in its purpose of supporting research, development, demonstration, and application of quantum information science and technology.

While big science need not entail big government; the alignment of forces implies that it likely will. There is, however, no commandment to regulate frontier sectors via the same administrative state model that has dominated policy in recent decades, and policymakers are at a fork in the road that will affect the evolution of business and enterprise. On matters of safety, economics and jobs, the government need not steer while the market merely rows.

(This article is based on my comments to OMB on its Request for Comments on the Guidance for Regulation of Artificial Intelligence Applications.)

View original post here:
How The White House Guidance For Regulation Of Artificial Intelligence Invites Overregulation - Forbes