Daily Archives: June 15, 2024

Top 6 examples of humanoid robots – TechTarget

Posted: June 15, 2024 at 7:51 pm

Humanoids are a fusion of AI and robotics. They typically have a body structure similar to humans, often sport skin and eyes, and are equipped with sensors and cameras to recognize human faces, respond to voice commands and engage in conversations.

Also embedded with the technology to mimic human traits, humanoids can learn and adapt in real time. The most recent versions of these robots may even exhibit a wide spectrum of human emotions and move and talk like people.

Besides captivating human imagination, these anthropomorphic creations also serve as groundbreaking tools across various industries. According to a Goldman Sachs report, the global market for humanoid robots could reach $38 billion by 2035, underscoring their importance across numerous industries.

Various sources, such as Interesting Engineering, Business Today and Built In, identify the following as the top examples of humanoid robots:

Sophia is an emotionally intelligent, AI-powered social robot that a team of AI experts and David Hanson of the Hong Kong-based company Hanson Robotics developed. It was activated on February 14, 2016, and unlike previous models of humanoids, Sophia can imitate human expressions and engage in conversations.

Sophia is a service robot developed to fulfill specific roles such as caring for the elderly, serving customers, engaging with kids and handling crowds at events. Sophia's exceptional natural language processing skills, fueled by AI and neural networks, enable it to maintain eye contact, answer questions, converse and synchronize body language with its voice. Sophia is also skilled at reading the emotions and body language of humans. Sophia has been featured at numerous events and conferences, such as the Consumer Electronics Show (CES) 2019 and is scheduled to appear at the Global AI Show and Global Blockchain Show in December 2024.

Interesting facts about Sophia: Sophia's look is an ideal fusion of science fiction and historical elegance and was inspired by the Hollywood actress Audrey Hepburn, Amanda Hudson (the wife of Hanson) and the ancient Egyptian queen Nefertiti.

In 2019, Sophia displayed the ability to create drawings, including portraits. Notably, a non-fungible token (NFT) self-portrait created by Sophia sold for nearly $700,000 at an auction in Hong Kong, China, in 2021.

Developed by an American robotics design company, Boston Dynamics, and funded by the Defense Advanced Research Project (DARPA), Atlas made its public debut on July 11, 2013. Measuring 5 feet tall and weighing 196 pounds, the first iteration of Atlas relied on a robust and intricate hydraulics system, that enhanced its agility. Capable of backflips and bending down, this robot was designed to undertake hazardous tasks in search and rescue missions. Atlas also aids in real-world applications, such as industrial automation tasks, and mobile manipulation that involve the integration of navigation and interaction with the environment, such as welding, screwing and quality control.

In April 2024, Boston Dynamics revealed intentions to replace the hydraulic Atlas with an electric version to boost its strength and provide a wider range of motion.

Interesting facts about Atlas: The retired hydraulics version of Atlas was the most agile humanoid around. It effortlessly lifted and transported items such as boxes and crates. However, its signature moves were running, jumping and performing backflips.

Ameca's designer and vendor, Engineered Arts claims that Ameca is the world's most advanced humanoid robot. Originally conceived as a foundation for advancing robotics technologies in human-robot interaction and as a development platform for testing AI and machine learning systems, this humanoid incorporates embedded microphones, binocular eye-mounted cameras, a chest camera and facial recognition software for engaging with the public.

Ameca was developed at Engineered Arts' base in Falmouth, Cornwall, UK, in 2021. It quickly captured the spotlight on X (formerly known as Twitter) and TikTok before its debut demonstration at CES 2022, where it attracted vast coverage from various media outlets.

Interesting facts about Ameca: Since Ameca has cameras in each of its eyes, it can recognize and track faces, identify objects and respond appropriately when a hand is placed in front of its face. It also has humanlike shoulder motions and can extend its hand to the side of its head.

Geminoid DK is a teleoperated android boasting a metallic skeleton covered in silicone skin and complemented by human and artificial hair. When it debuted in 2011, the world was taken aback by its lifelike appearance and facial expressions.

The Geminoid DK also shares an uncanny resemblance with its creator, the Danish professor Henrik Scharfe of Aalborg University, who collaborated on the project along with Japanese engineer Hiroshi Ishiguro, his team at Advanced Telecommunication Institute International, and Sanrio Group's robot manufacturer Kokoro.

Geminoid DK's goal is to study human-robot interactions, especially how people respond to robotic representations of real humans.

Interesting facts about Geminoid DK: Geminoid-DK can establish eye contact, exhibit various expressions and perform involuntary muscle and breathing movements. It's also the first humanoid robot to sport a beard, which, along with other facial hair, was manually implanted and trimmed using Henrik Scharfe's personal trimmer.

Nadine is a gynoid social robot, also known as a fembot, that was created in 2013. It was modeled after Professor Nadia Magnenat Thalmann, one of Nadine's creators and a visiting professor at Nanyang Technological University's Institute (NTU). Japanese firm Kokoro developed Nadine's hardware, while Thalmann's team at NTU crafted the software and articulated the robot's hands to achieve natural grasping.

Nadine was designed to interact with humans in social settings, displaying empathy, answering queries and remembering conversations. Nadine is equipped with 3D depth cameras and microphones to ensure seamless operation.

Interesting facts about Nadine: Nadine is full of personality, returns greetings, makes eye contact and interacts with arm movements. It assists individuals with special needs by reading stories and helping with other communication tasks. Additionally, Nadine has served as an office receptionist or a personal coach.

Pepper was developed by SoftBank Robotics and made its debut in 2014. This advanced and commercially available social humanoid robot stands at approximately 4 feet tall and features a tablet display on its chest for enabling interactions with users.

Pepper was created to serve various functions and industries. For example, it has served as a companion in various settings such as homes, schools, hospitality, healthcare and retail. It is equipped with several cameras, touch sensors and microphones that enable it to engage with humans through speech, touch and emotion recognition.

Interesting facts about Pepper: Pepper's voice can be adjusted depending on preferences. Pepper utilizes tactile sensors in its hands that enable it to perform human actions such as gently picking up and setting down objects. Pepper uses these sensors during activities such as playing games or engaging in social interactions. These sensors are also present in Pepper's head to perceive touch and interactions.

Kinza Yasar is a technical writer for WhatIs with a degree in computer networking.

See the original post here:

Top 6 examples of humanoid robots - TechTarget

Posted in Robotics | Comments Off on Top 6 examples of humanoid robots – TechTarget

Quantum computers are like kaleidoscopes why unusual metaphors help illustrate science and technology – The Conversation

Posted: at 7:50 pm

Quantum computing is like Forrest Gumps box of chocolates: You never know what youre gonna get. Quantum phenomena the behavior of matter and energy at the atomic and subatomic levels are not definite, one thing or another. They are opaque clouds of possibility or, more precisely, probabilities. When someone observes a quantum system, it loses its quantum-ness and collapses into a definite state.

Quantum phenomena are mysterious and often counterintuitive. This makes quantum computing difficult to understand. People naturally reach for the familiar to attempt to explain the unfamiliar, and for quantum computing this usually means using traditional binary computing as a metaphor. But explaining quantum computing this way leads to major conceptual confusion, because at a base level the two are entirely different animals.

This problem highlights the often mistaken belief that common metaphors are more useful than exotic ones when explaining new technologies. Sometimes the opposite approach is more useful. The freshness of the metaphor should match the novelty of the discovery.

The uniqueness of quantum computers calls for an unusual metaphor. As a communications researcher who studies technology, I believe that quantum computers can be better understood as kaleidoscopes.

The gap between understanding classical and quantum computers is a wide chasm. Classical computers store and process information via transistors, which are electronic devices that take binary, deterministic states: one or zero, yes or no. Quantum computers, in contrast, handle information probabilistically at the atomic and subatomic levels.

Classical computers use the flow of electricity to sequentially open and close gates to record or manipulate information. Information flows through circuits, triggering actions through a series of switches that record information as ones and zeros. Using binary math, bits are the foundation of all things digital, from the apps on your phone to the account records at your bank and the Wi-Fi signals bouncing around your home.

In contrast, quantum computers use changes in the quantum states of atoms, ions, electrons or photons. Quantum computers link, or entangle, multiple quantum particles so that changes to one affect all the others. They then introduce interference patterns, like multiple stones tossed into a pond at the same time. Some waves combine to create higher peaks, while some waves and troughs combine to cancel each other out. Carefully calibrated interference patterns guide the quantum computer toward the solution of a problem.

The term bit is a metaphor. The word suggests that during calculations, a computer can break up large values into tiny ones bits of information which electronic devices such as transistors can more easily process.

Using metaphors like this has a cost, though. They are not perfect. Metaphors are incomplete comparisons that transfer knowledge from something people know well to something they are working to understand. The bit metaphor ignores that the binary method does not deal with many types of different bits at once, as common sense might suggest. Instead, all bits are the same.

The smallest unit of a quantum computer is called the quantum bit, or qubit. But transferring the bit metaphor to quantum computing is even less adequate than using it for classical computing. Transferring a metaphor from one use to another blunts its effect.

The prevalent explanation of quantum computing is that while classical computers can store or process only a zero or one in a transistor or other computational unit, quantum computers supposedly store and handle both zero and one and other values in between at the same time through the process of superposition.

Superposition, however, does not store one or zero or any other number simultaneously. There is only an expectation that the values might be zero or one at the end of the computation. This quantum probability is the polar opposite of the binary method of storing information.

Driven by quantum sciences uncertainty principle, the probability that a qubit stores a one or zero is like Schroedingers cat, which can be either dead or alive, depending on when you observe it. But the two different values do not exist simultaneously during superposition. They exist only as probabilities, and an observer cannot determine when or how frequently those values existed before the observation ended the superposition.

Leaving behind these challenges to using traditional binary computing metaphors means embracing new metaphors to explain quantum computing.

The kaleidoscope metaphor is particularly apt to explain quantum processes. Kaleidoscopes can create infinitely diverse yet orderly patterns using a limited number of colored glass beads, mirror-dividing walls and light. Rotating the kaleidoscope enhances the effect, generating an infinitely variable spectacle of fleeting colors and shapes.

The shapes not only change but cant be reversed. If you turn the kaleidoscope in the opposite direction, the imagery will generally remain the same, but the exact composition of each shape or even their structures will vary as the beads randomly mingle with each other. In other words, while the beads, light and mirrors could replicate some patterns shown before, these are never absolutely the same.

Using the kaleidoscope metaphor, the solution a quantum computer provides the final pattern depends on when you stop the computing process. Quantum computing isnt about guessing the state of any given particle but using mathematical models of how the interaction among many particles in various states creates patterns, called quantum correlations.

Each final pattern is the answer to a problem posed to the quantum computer, and what you get in a quantum computing operation is a probability that a certain configuration will result.

Metaphors make the unknown manageable, approachable and discoverable. Approximating the meaning of a surprising object or phenomenon by extending an existing metaphor is a method that is as old as calling the edge of an ax its bit and its flat end its butt. The two metaphors take something we understand from everyday life very well, applying it to a technology that needs a specialized explanation of what it does. Calling the cutting edge of an ax a bit suggestively indicates what it does, adding the nuance that it changes the object it is applied to. When an ax shapes or splits a piece of wood, it takes a bite from it.

Metaphors, however, do much more than provide convenient labels and explanations of new processes. The words people use to describe new concepts change over time, expanding and taking on a life of their own.

When encountering dramatically different ideas, technologies or scientific phenomena, its important to use fresh and striking terms as windows to open the mind and increase understanding. Scientists and engineers seeking to explain new concepts would do well to seek out originality and master metaphors in other words, to think about words the way poets do.

Originally posted here:

Quantum computers are like kaleidoscopes why unusual metaphors help illustrate science and technology - The Conversation

Posted in Quantum Computing | Comments Off on Quantum computers are like kaleidoscopes why unusual metaphors help illustrate science and technology – The Conversation

Quantum control’s role in scaling quantum computing – McKinsey

Posted: at 7:50 pm

June 14, 2024by Henning Soller and Niko Mohr with Elisa Becker-Foss, Kamalika Dutta, Martina Gschwendtner, Mena Issler, and Ming Xu

Quantum computing can leverage the states of entangled qubits1 to solve problems that classical computing cannot currently solve and to substantially improve existing solutions. These qubits, which are typically constructed from photons, atoms, or ions, can only be manipulated using specially engineered signals with precisely controlled energy that is barely above that of a vacuum and that changes within nanoseconds. This control system for qubits, referred to as quantum control, is a critical enabler of quantum computing because it ensures quantum algorithms perform with optimal efficiency and effectiveness.

While the performance and scaling limitations of current quantum control systems preclude large-scale quantum computing, several promising technological innovations may soon offer scalable control solutions.

A modern quantum computer comprises various hardware and software components, including quantum control components that require extensive space and span meters. In quantum systems, qubits interact with the environment, causing decoherence and decay of the encoded quantum information. Quantum gates (building blocks of quantum circuits) cannot be implemented perfectly at the physical system level, resulting in accumulated noise. Noise leads to decoherence, which lowers qubits superposition and entanglement properties. Quantum control minimizes the quantum noisefor example, thermal fluctuations and electromagnetic interferencecaused by the interaction between the quantum hardware and its surroundings. Quantum control also addresses noise by improving the physical isolation of qubits, using precise control techniques, and implementing quantum error correction codes. Control electronics use signals from the classical world to provide instructions for qubits, while readout electronics measure qubit states and transmit that information back to the classical world. Thus, the control layer in a quantum technology stack is often referred to as the interface between the quantum and classical worlds.

Components of the control layer include the following:

A superconducting- or spin qubitbased computer, for example, includes physical components such as quantum chips, cryogenics (cooling electronics), and control and readout electronics.

Quantum computing requires precise control of qubits and manipulation of physical systems. This control is achieved via signals generated by microwaves, lasers, and optical fields or other techniques that support the underlying qubit type. A tailored quantum control system is needed to achieve optimal algorithm performance.

In the context of a quantum computing stack, control typically refers to the hardware and software system that connects to the qubits the application software uses to solve real-world problems such as optimization and simulation (Exhibit 1).

At the top of the stack, software layers translate real-world problems into executable instructions for manipulating qubits. The software layer typically includes middleware (such as a quantum transpiler2) and control software comprising low-level system software that provides compilation, instrument control, signal generation, qubit calibration, and dynamical error suppression.3 Below the software layer is the hardware layer, where high-speed electronics and physical components work together to send signals to and read signals from qubits and to protect qubits from noise. This is the layer where quantum control instructions are executed.

Quantum control hardware systems are highly specialized to accommodate the intricacies of qubits. Control hardware interfaces directly with qubits, generating and reading out extremely weak and rapidly changing electromagnetic signals that interact with qubits. To keep qubits functioning for as long as possible, control hardware systems must be capable of adapting in real time to stabilize the qubit state (feedback calibration) and correct qubits from decaying to a completely decoherent state4 (quantum error correction).

Although all based on similar fundamental principles of quantum control, quantum control hardware can differ widely depending on the qubit technology with which it is designed to be used (Exhibit 2).

For example, photonic qubits operate at optical frequencies (similar to fiber internet), while superconducting qubits operate at microwave frequencies (similar to a fifth-generation network). Different types of hardware using laser technology or electronic circuits are needed to generate, manipulate, and transmit signals to and from these different qubit types. Additional hardware may be needed to provide environmental control. Cryostats, for example, cool superconducting qubits to keep them in a working state, and ion trap devices are used in trapped-ion qubit systems to confine ions using electromagnetic fields.

Quantum control is critical to enable fault-tolerant quantum computingquantum computing in which as many errors as possible are prevented or suppressed. But realizing this capability on a large scale will require substantial innovation. Existing control systems are designed for a small number of qubits (1 to 1,000) and rely on customized calibration and dedicated resources for each qubit. A fault-tolerant quantum computer, on the other hand, needs to control 100,000 to 1,000,000 qubits simultaneously. Consequently, a transformative approach to quantum control design is essential.

Specifically, to achieve fault-tolerant quantum computing on a large scale, there must be advances to address issues with current state-of-the-art quantum control system performance and scalability, as detailed below.

Equipping quantum systems to perform at large scales will require the following:

The limitations that physical space poses and the cost to power current quantum computing systems restrict the number of qubits that can be controlled with existing architecture, thus hindering large-scale computing.

Challenges to overcoming these restrictions include the following:

Several technologies show promise for scaling quantum control, although many are still in early-research or prototyping stages (Exhibit 3).

Multiplexing could help reduce costs and prevent overheating. The cryogenic complementary metal-oxide-semiconductor (cryo-CMOS) approach also helps mitigate overheating; it is the most widely used approach across industries because it is currently the most straightforward way to add control lines, and it works well in a small-scale R&D setup. However, cryo-CMOS is close to reaching the maximum number of control lines, creating form factor and efficiency challenges to scaling. Even with improvements, the number of control lines would only be reduced by a few orders of magnitude, which is not sufficient for scaling to millions of qubits. Another option to address overheating is single-flux quantum technology, while optical links for microwave qubits can increase efficiency in interconnections as well as connect qubits between cryostats.

Whether weighing options to supply quantum controls solutions or to invest in or integrate quantum technologies into companies in other sectors, leaders can better position their organizations for success by starting with a well-informed and strategically focused plan.

The first strategic decision leaders in the quantum control sector must make is whether to buy or build their solutions. While various levels of quantum control solutions can be sourced from vendors, few companies specialize in control, and full-stack solutions for quantum computing are largely unavailable. The prevailing expertise is that vendors can offer considerable advantages in jump-starting quantum computing operations, especially those with complex and large-scale systems. Nevertheless, a lack of industrial standardization means that switching between quantum control vendors could result in additional costs down the road. Consequently, many leading quantum computing players opt to build their own quantum control.

Ideally, business leaders also determine early on which parts of the quantum tech stack to focus their research capacities on and how to benchmark their technology. To develop capabilities and excel in quantum control, it is important to establish KPIs that are tailored to measure how effectively quantum control systems perform to achieve specific goals, such as improved qubit fidelity.5 This allows for the continuous optimization and refinement of quantum control techniques to improve overall system performance and scalability.

Quantum control is key to creating business value. Thus, the maturity and scalability of control solutions are the chief considerations for leaders exploring business development related to quantum computing, quantum solutions integration, and quantum technologies investment. In addition to scalability (the key criterion for control solutions), leaders will need to consider and address the other control technology challenges noted previously. And as control technologies mature from innovations to large-scale solutions, establishing metrics for benchmarking them will be essential to assess, for example, ease of integration, cost effectiveness, error-suppression effectiveness, software offerings, and the possibility of standardizing across qubit technologies.

Finally, given the shortage of quantum talent, recruiting and developing the highly specialized capabilities needed for each layer of the quantum stack is a top priority to ensure quantum control systems are properly developed and maintained.

Henning Soller is a partner in McKinseys Frankfurt office, and Niko Mohr is a partner in the Dsseldorf office. Elisa Becker-Foss is a consultant in the New York office, Kamalika Dutta is a consultant in the Berlin office, Martina Gschwendtner is a consultant in the Munich office, Mena Issler is an associate partner in the Bay Area office, and Ming Xu is a consultant in the Stamford office.

1 Entangled qubits are qubits that remain in a correlated state in which changes to one affect the other, even if they are separated by long distances. This property can enable massive performance boosts in information processing. 2 A quantum transpiler converts code from one quantum language to another while preserving and optimizing functionality to make algorithms and circuits portable between systems and devices. 3 Dynamical error suppression is one approach to suppressing quantum error and involves the periodic application of control pulse sequences to negate noise. 4 A qubit in a decoherent state is losing encoded quantum information (superposition and entanglement properties). 5 Qubit fidelity is a measure of the accuracy of a qubits state or the difference between its current state and the desired state.

Original post:

Quantum control's role in scaling quantum computing - McKinsey

Posted in Quantum Computing | Comments Off on Quantum control’s role in scaling quantum computing – McKinsey

Quantum Computing and AI: A Perfect Match? – InformationWeek

Posted: at 7:50 pm

It's a marriage that could only happen in cyberspace -- quantum computing and artificial intelligence.

Quantum AI is a burgeoning computer science sector, dedicated to exploring the potential synergy that exists between quantum computing and AI, says Gushu Li, a professor at the University of Pennsylvania School of Engineering and Applied Science, in an email interview. "It seeks to apply principles from quantum mechanics to enhance AI algorithms." A growing number of researchers now believe that AI models developed with quantum computing will soon outpace classical computing AI development.

Quantum AI creates an intersection between quantum computing and artificial intelligence, observes Romn Ors, chief scientific officer at quantum computing software development firm Multiverse Computing, via email. He notes that quantum computing has the potential to take AI to entirely new levels of performance. "For instance, it's possible to develop quantum neural networks that teach a quantum computer to detect anomalies, do image recognition, and other tasks." Ors adds that it's also possible to improve traditional AI methods by using quantum-inspired approaches to dramatically reduce the development and training costs of large language models (LLMs).

Related:Demystifying Quantum Computing: Separating Fact from Fiction

Combining the quantum physics properties of superposition and entanglement, which can perform limitless processes simultaneously with machine learning and AI, and suddenly it's possible to do more than ever imagined, says Tom Patterson, emerging technology security lead at business advisory firm Accenture, via email. "Unfortunately, that includes being used by adversaries to crack our encryption and develop new and insidious ways to separate us from our information, valuables, and anything else we hold dear."

Still, Patterson is generally optimistic. Like ChatGPT, he expects quantum AI to arrive gradually, and then all at once. "While full use of an AI-relevant quantum computer remains years away, the benefits of thinking about AI with quantum information science capabilities are exciting and important today," he states. "The opportunities are here and now, and the future is brighter than ever with quantum AI."

For his part, Li believes that quantum AI's biggest initial impact will be in four specific areas:

Drug Discovery: Simulating molecules to design new drugs and materials with superior properties.

Financial Modeling: Optimizing complex financial portfolios and uncovering hidden trends in the market.

Related:Cybersecurity's Future: Facing Post-Quantum Cryptography Peril

Materials Science: Developing new materials with specific properties for applications like superconductors or ultra-efficient solar cells.

Logistics and Optimization: Finding the most efficient routes for transportation and optimizing complex supply chains.

Quantum AI is already here, but it's a silent revolution, Ors says. "The first applications of quantum AI are finding commercial value, such as those related to LLMs, as well as in image recognition and prediction systems," he states. More quantum AI applications will become available as quantum computers grow more powerful. "It's expected that in two-to-three years there will be a broad range of industrial applications of quantum AI."

Yet the road ahead may be rocky, Li warns. "It's well known that quantum hardware suffers from noise that can destroy computation," he says. "Quantum error correction promises a potential solution, but that technology isn't yet available."

Meanwhile, while quantum AI algorithms are being developed, classical computing competitors are achieving new AI successes. "While progress is being made, it's prudent to acknowledge that the integration of quantum computing with AI is a complex endeavor that will unfold gradually," Li says.

Related:What Is the Future of AI-Driven Employee Monitoring?

Patterson notes that many of the most promising quantum AI breakthroughs aren't arriving from university and corporate research teams, but from various regional developer and support communities that closely mirror natural ecosystems. "Regions that have decided that quantum and AI are too big and too important to leave to one group or another have organized around providing everything progress demands -- from investment to science to academics to entrepreneurs, growth engines, and tier-one buyers," he says. "These regional ecosystems are where the magic happens with quantum AI."

GenAI and quantum computing are mind-blowing advances in computing technology, says Guy Harrison, enterprise architect at cybersecurity technology company OneSpan, in a recent email interview. "AI is a sophisticated software layer that emulates the very capabilities of human intelligence, while quantum computing is assembling the very building blocks of the universe to create a computing substrate," he explains. "We're pushing computing both into the realm of the mind and the realm of the sub-atomic."

The transition to quantum AI won't be optional, Ors warns, since current AI is fundamentally flawed due to excessive energy costs. New models and methods will be needed to lower energy demands and to make AI feasible in the long term. "Early adopters of quantum AI will get a competitive advantage and will survive, as opposed to those that do not adopt or adopt it too late."

See the original post:

Quantum Computing and AI: A Perfect Match? - InformationWeek

Posted in Quantum Computing | Comments Off on Quantum Computing and AI: A Perfect Match? – InformationWeek

Better Qubits: Quantum Breakthroughs Powered by Silicon Carbide – SciTechDaily

Posted: at 7:50 pm

By U.S. Department of Energy June 14, 2024

Artists representation of the formation pathway of vacancy complexes for spin-based qubits in the silicon carbide host lattice and to the right the associated energy landscape. Credit: University of Chicago

Quantum computers, leveraging the unique properties of qubits, outperform classical systems by simultaneously existing in multiple states. Focused research on silicon carbide aims to optimize qubits for scalable application, with studies revealing new methods to control and enhance their performance. This could lead to breakthroughs in large-scale quantum computing and sensor technologies.

While conventional computers use classical bits for calculations, quantum computers use quantum bits, or qubits, instead. While classical bits can have the values 0 or 1, qubits can exist in a mix of probabilities of both values at the same time. This makes quantum computing extremely powerful for problems conventional computers arent good at solving. To build large-scale quantum computers, researchers need to understand how to create and control materials that are suitable for industrial-scale manufacturing.

Semiconductors are very promising qubit materials. Semiconductors already make up the computer chips in cell phones, computers, medical equipment, and other applications. Certain types of atomic-scale defects, called vacancies, in the semiconductor silicon carbide (SiC) show promise as qubits. However, scientists have a limited understanding of how to generate and control these defects. By using a combination of atomic-level simulations, researchers were able to track how these vacancies form and behave.

Quantum computing could revolutionize our ability to answer challenging questions. Existing small scale quantum computers have given a glimpse of the technologys power. To build and deploy large-scale quantum computers, researchers need to know how to control qubits made of materials that make technical and economic sense for industry.

The research identified the stability and molecular pathways to create the desired vacancies for qubits and determine their electronic properties.

These advances will help the design and fabrication of spin-based qubits with atomic precision in semiconductor materials, ultimately accelerating the development of next-generation large-scale quantum computers and quantum sensors.

The next technological revolution in quantum information science requires researchers to deploy large-scale quantum computers that ideally can operate at room temperature. The realization and control of qubits in industrially relevant materials is key to achieving this goal.

In the work reported here, researchers studied qubits built from vacancies in silicon carbide (SiC) using various theoretical methods. Until now, researchers knew little about how to control and engineer the selective formation process for the vacancies. The involved barrier energies for vacancy migration and combination pose the most difficult challenges for theory and simulations.

In this study, a combination of state-of-the-art materials simulations and neural-network-based sampling technique led researchers at the Department of Energys (DOE) Midwest Center for Computational Materials (MICCoM) to discover the atomistic generation mechanism of qubits from spin defects in a wide-bandgap semiconductor.

The team showed the generation mechanism of qubits in SiC, a promising semiconductor with long qubit coherence times and all-optical spin initialization and read-out capabilities.

MICCoM is one of the DOE Computational Materials Sciences centers across the country that develops open-source, advanced software tools to help the scientific community model, simulate, and predict the fundamental properties and behavior of functional materials. The researchers involved in this study are from Argonne National Laboratory and the University of Chicago.

Reference: Stability and molecular pathways to the formation of spin defects in silicon carbide by Elizabeth M. Y. Lee, Alvin Yu, Juan J. de Pablo and Giulia Galli, 3 November 2021,Nature Communications. DOI: 10.1038/s41467-021-26419-0

This work was supported by the Department of Energy (DOE) Office of Science, Office of Basic Energy Sciences, Materials Sciences and Engineering Division and is part of the Basic Energy Sciences Computational Materials Sciences Program in Theoretical Condensed Matter Physics. The computationally demanding simulations used several high-performance computing resources: Bebop in Argonne National Laboratorys Laboratory Computing Resource Center; the Argonne Leadership Computing Facility (ALCF), a DOE Office of Science user facility; and the University of Chicagos Research Computing Center. The team was awarded access to ALCF computing resources through DOEs Innovative and Novel Computational Impact on Theory and Experiment (INCITE) program. Additional support was provided by NIH.

Read more from the original source:

Better Qubits: Quantum Breakthroughs Powered by Silicon Carbide - SciTechDaily

Posted in Quantum Computing | Comments Off on Better Qubits: Quantum Breakthroughs Powered by Silicon Carbide – SciTechDaily

Quantum, AI Combine to Transform Energy Generation, AI Summit London – AI Business

Posted: at 7:50 pm

The electrical grid is very complicated. Nobody thinks about it ever until it doesn't work. But it is critical infrastructure that runs minute-to-minute energy being consumed now was generated milliseconds ago, somewhere far away, instantaneously shot through power lines and delivered.

This gets more complicated when locally generated sustainable energy joins the mix, pushing it beyond the capabilities of classical computing solutions. Home energy supplier E.ON is trialing quantum computer solutions to manage this future grid.

Speaking at the AI Summit London, E.ON chief quantum scientist Corey OMeara explained the challenges presented by future decentralized grids.

The way grids are changing now is, if buildings have solar panels on the roofs, you want to use that renewable energy yourself, or you might want to inject that back into the grid to power your neighbor's house, he said.

This decentralized energy production and peer-to-peer energy-sharing model presents a massive overhead for an aging grid that was never meant to be digital. E.ON is working on solving this renewable energy integration optimization problem using quantum computing.

E.ON also uses AI extensively and some functions could in the future be enhanced using quantum computing. An important example is AI-driven predictive maintenance for power plants.

Related:Unilever's Alberto Prado on Quantum Computing's Future, Impact on Emerging Tech

Power plants are complex objects that have thousands of sensors that measure and monitor factors such as temperatures and pressures and store the data in the cloud. We have AI solutions to analyze them to make sure that they're functioning correctly, said OMeara.

We published a paper where we invented a novel anomaly detection algorithm using quantum computing as a subroutine. We used it with our gas turbine data as well as academic benchmark data sets from the computer science field and found that the quantum-augmented solution did perform better but only for certain metrics.

E.ON plans to develop this trial into an integrated quantum software solution that could run on today's noisy, intermediate-scale quantum computers rather than waiting for next-generation fully error-corrected devices.

See the article here:

Quantum, AI Combine to Transform Energy Generation, AI Summit London - AI Business

Posted in Quantum Computing | Comments Off on Quantum, AI Combine to Transform Energy Generation, AI Summit London – AI Business

Diraq is Partnering with GlobalFoundries for Integrated CMOS/Qubits Logic on the Same Chip and Achieves Record Spin Qubit Gate Fidelities at 1 Kelvin…

Posted: at 7:50 pm

Diraq is Partnering with GlobalFoundries for Integrated CMOS/Qubits Logic on the Same Chip and Achieves Record Spin Qubit Gate Fidelities at 1 Kelvin for Test Chips Fabricated at Imec.  Quantum Computing Report

The rest is here:

Diraq is Partnering with GlobalFoundries for Integrated CMOS/Qubits Logic on the Same Chip and Achieves Record Spin Qubit Gate Fidelities at 1 Kelvin...

Posted in Quantum Computing | Comments Off on Diraq is Partnering with GlobalFoundries for Integrated CMOS/Qubits Logic on the Same Chip and Achieves Record Spin Qubit Gate Fidelities at 1 Kelvin…

New technique could help build quantum computers of the future – EurekAlert

Posted: at 7:50 pm

image:

Kaushalya Jhuria in the lab testing the electronics from the experimental setup used to make qubits in silicon.

Credit: Thor Swift/Berkeley Lab

Quantum computers have the potential to solve complex problems in human health, drug discovery, and artificial intelligence millions of times faster than some of the worlds fastest supercomputers. A network of quantum computers could advance these discoveries even faster. But before that can happen, the computer industry will need a reliable way to string together billions of qubits or quantum bits with atomic precision.

Connecting qubits, however, has been challenging for the research community. Some methods form qubits by placing an entire silicon wafer in a rapid annealing oven at very high temperatures. With these methods, qubits randomly form from defects (also known as color centers or quantum emitters) in silicons crystal lattice. And without knowing exactly where qubits are located in a material, a quantum computer of connected qubits will be difficult to realize.

But now, getting qubits to connect may soon be possible. A research team led by Lawrence Berkeley National Laboratory (Berkeley Lab) says that they are the first to use a femtosecond laser to create and annihilate qubits on demand, and with precision, by doping silicon with hydrogen.

The advance could enable quantum computers that use programmable optical qubits or spin-photon qubits to connect quantum nodes across a remote network. It could also advance a quantum internet that is not only more secure but could also transmit more data than current optical-fiber information technologies.

To make a scalable quantum architecture or network, we need qubits that can reliably form on-demand, at desired locations, so that we know where the qubit is located in a material. And that's why our approach is critical, said Kaushalya Jhuria, a postdoctoral scholar in Berkeley Labs Accelerator Technology & Applied Physics (ATAP) Division. She is the first author on a new study that describes the technique in the journal Nature Communications. Because once we know where a specific qubit is sitting, we can determine how to connect this qubit with other components in the system and make a quantum network.

This could carve out a potential new pathway for industry to overcome challenges in qubit fabrication and quality control, said principal investigator Thomas Schenkel, head of the Fusion Science & Ion Beam Technology Program in Berkeley Labs ATAP Division. His group will host the first cohort of students from the University of Hawaii in June as part of a DOE Fusion Energy Sciences-funded RENEW project on workforce development where students will be immersed in color center/qubit science and technology.

Forming qubits in silicon with programmable control

The new method uses a gas environment to form programmable defects called color centers in silicon. These color centers are candidates for special telecommunications qubits or spin photon qubits. The method also uses an ultrafast femtosecond laser to anneal silicon with pinpoint precision where those qubits should precisely form. A femtosecond laser delivers very short pulses of energy within a quadrillionth of a second to a focused target the size of a speck of dust.

Spin photon qubits emit photons that can carry information encoded in electron spin across long distances ideal properties to support a secure quantum network. Qubits are the smallest components of a quantum information system that encodes data in three different states: 1, 0, or a superposition that is everything between 1 and 0.

With help from Boubacar Kant, a faculty scientist in Berkeley Labs Materials Sciences Division and professor of electrical engineering and computer sciences (EECS) at UC Berkeley, the team used a near-infrared detector to characterize the resulting color centers by probing their optical (photoluminescence) signals.

What they uncovered surprised them: a quantum emitter called the Ci center. Owing to its simple structure, stability at room temperature, and promising spin properties, the Ci center is an interesting spin photon qubit candidate that emits photons in the telecom band. We knew from the literature that Ci can be formed in silicon, but we didnt expect to actually make this new spin photon qubit candidate with our approach, Jhuria said.

The researchers learned that processing silicon with a low femtosecond laser intensity in the presence of hydrogen helped to create the Ci color centers. Further experiments showed that increasing the laser intensity can increase the mobility of hydrogen, which passivates undesirable color centers without damaging the silicon lattice, Schenkel explained.

A theoretical analysis performed by Liang Tan, staff scientist in Berkeley Labs Molecular Foundry, shows that the brightness of the Ci color center is boosted by several orders of magnitude in the presence of hydrogen, confirming their observations from laboratory experiments.

The femtosecond laser pulses can kick out hydrogen atoms or bring them back, allowing the programmable formation of desired optical qubits in precise locations, Jhuria said.

The team plans to use the technique to integrate optical qubits in quantum devices such as reflective cavities and waveguides, and to discover new spin photon qubit candidates with properties optimized for selected applications.

Now that we can reliably make color centers, we want to get different qubits to talk to each other which is an embodiment of quantum entanglement and see which ones perform the best. This is just the beginning, said Jhuria.

The ability to form qubits at programmable locations in a material like silicon that is available at scale is an exciting step towards practical quantum networking and computing, said Cameron Geddes, Director of the ATAP Division.

Theoretical analysis for the study was performed at the Department of EnergysNational Energy Research Scientific Computing Center (NERSC) at Berkeley Lab with support from the NERSC QIS@Perlmutterprogram.

The Molecular Foundry and NERSC are DOE Office of Science user facilities at Berkeley Lab.

This work was supported by the DOE Office of Fusion Energy Sciences.

###

Lawrence Berkeley National Laboratory (Berkeley Lab) is committed to delivering solutions for humankind through research in clean energy, a healthy planet, and discovery science. Founded in 1931 on the belief that the biggest problems are best addressed by teams, Berkeley Lab and its scientists have been recognized with 16 Nobel Prizes. Researchers from around the world rely on the Labs world-class scientific facilities for their own pioneering research. Berkeley Lab is a multiprogram national laboratory managed by the University of California for the U.S. Department of Energys Office of Science.

DOEs Office of Science is the single largest supporter of basic research in the physical sciences in the United States, and is working to address some of the most pressing challenges of our time. For more information, please visitenergy.gov/science.

Nature Communications

Experimental study

Not applicable

Programmable quantum emitter formation in silicon

27-May-2024

Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.

Read this article:

New technique could help build quantum computers of the future - EurekAlert

Posted in Quantum Computing | Comments Off on New technique could help build quantum computers of the future – EurekAlert

The quiet return of eugenics – The Spectator

Posted: at 7:49 pm

Here follows a non-exhaustive list of my genetic flaws. I am short-sighted, more so as I age. I have bunions, dodgy knees and even dodgier shoulders. I have asthma. My skin blisters easily. My hair started going grey when I was in my late teens. I have zero talent for foreign languages, running or music. I am prone to nightmares, as well as to depression and anxiety.

Relatively mild flaws, as they go. But still, these arent traits Im eager to pass on. Our three-year-old already shows a tendency for nightmares that sometimes makes me wince with guilt. Not that its my fault, of course. We dont get to choose which of our genes we pass on. Every conception is a roll of the dice.

The technology will be described with euphemisms such as genetic enhancement or genetic health

But soon that will no longer be true. In fact, its already not quite true, at least for those who have the means and determination to load the dice. Emerging technology is about to present parents with a set of ethical questions that make the usual kinds of debates breast milk or formula? Nanny or daycare? seem trivial. We have always had the power (more or less) to control our childrens nurture. Before long perhaps in just a few years any parent who can afford to will have control over the minutest details of a childs nature too.

The crucial change set to turn our lives upside-down is called preimplantation genetic testing for polygenic disorders (PGT-P), hereafter polygenic screening. Testing a foetus or embryo for some conditions is now a routine part of the modern pregnancy experience. Prenatal Downs Syndrome tests, for instance, are so widespread that in some Scandinavian countries almost 100 per cent of women choose to abort a foetus diagnosed with the condition, or if using IVF not implant the affected embryo. The result is a visible change to these populations: there are simply no more people with Downs to be seen on the streets of Iceland and Denmark.

Until now, these prenatal tests have been available only for some conditions. Whats revolutionary about polygenic screening is that it allows parents to take a batch of embryos conceived through IVF, have a report compiled for each one, based on their genetic risk factors, and then use these reports to decide which embryo to implant.

Such reports give a very full picture of the adult that embryo could become, including their vulnerability to an enormous number of diseases heart disease, diabetes, cancer and their likely physical and psychological traits: height, hair colour, athletic ability, conscientiousness, altruism, intelligence.

The list is long, and ethically fraught. Polygenic screening permits parents to choose the very best children, according to their own preferences, almost entirely removing the role of luck in the normal genetic lottery.

Should we welcome a new kind of commercial product that will allow some people mostly rich ones to have healthier, happier and cleverer children? And should you the reader seek out such a product for yourself? Should I?

Its a live question. Im currently pregnant with our second baby conceived the old fashioned way but we want to have more children. And I know enough people in the world of biotech to gain access to this novel service, which is not, at present, advertised as a single package, but must be procured via at least two different companies: one for the IVF, one for the polygenic screening. The screening itself is expensive, but not prohibitively so probably in the region of 7,000-12,000, which is less than a year of full-time daycare in London. Equally expensive, and far more physically onerous for the mother, is the IVF process, which my husband and I would otherwise have no reason to pursue.

But think of whats on offer: the opportunity to offer your children the best possible chance in life. Why would the kind of upper-middle-class parents who think nothing of spending vast sums on their childrens education not opt for polygenic screening? My bet is that they will, and soon.

If the word eugenics has sprung to mind while reading this, youre not alone. What were talking about here can best be understood as a new kind of eugenics one likely to be quite different from the first eugenics movement that emerged in Britain at the end of the 19th century.

For one thing, the new eugenics will be far more scientifically sophisticated. The earliest eugenicists did not know that the gene was the basic unit of heredity, since the term was not coined until 1909. They talked instead of gemmules and pangens. And they assumed that some traits such as homosexuality were far more heritable than they really are. The first eugenicists made many factual errors, as well as moral ones.

But while eugenics may be a dirty word in the 21st-century West, the fundamental claim behind the first eugenics movement nevertheless remains true. Victorian and Edwardian scientists were correct to notice that our genetic inheritance affects often to a large degree not only our physical but also our psychological characteristics. It is therefore possible to manipulate the characteristics of a population by encouraging or discouraging the reproduction of some genes which historically meant, in practice, the reproduction of some people.

Its one thing to deplore eugenics on ideological, political, moral grounds, as the evolutionary biologist Richard Dawkins tweeted in 2020. Its quite another to conclude that it wouldnt work in practice. Of course it would. It works for cows, horses, pigs, dogs & roses. Why on earth wouldnt it work for humans? Facts ignore ideology.

This is racist trash, Richard, replied Dan Hicks, professor of archaeology at Oxford, putting ideology before facts, and highlighting the key contemporary objection to the use of the word eugenics (if not, as we shall see later, the actual practice of it).

Critics such as Hicks are wrong to suggest that eugenics, as Dawkins phrased it, wouldnt work. Evolution can occur quite quickly, given the right conditions. Mutations that provide protection against malaria have been strongly selected for over the past few centuries in parts of the world where the disease is endemic. In the West, the invention of the caesarean section in the 18th century removed the selection pressure that historically restricted the size of the human head, meaning that heads are growing larger. These are not examples of artificial selection that is, humans consciously selecting particular features but they nevertheless demonstrate the speed and power of the evolutionary process upon which eugenics relies.

Which is not to say, of course, that it all comes down to genes. On the ancient question of nature or nurture?, by far the most defensible scientific answer is both. But while it is rare to come across anyone who insists that environment plays no role whatsoever in the development of physical and psychological traits, it is common to find people on the left who reject the role of nature altogether, insisting that humans are born as blank slates.

There are two motivations behind this ideological stance. The first is utopian: if you support any kind of revolutionary political project egalitarianism, say then you need to believe humans could, given the right conditions, radically overhaul their instinctive behaviour and desires. For revolutionaries, the idea of a fixed human nature presents a hateful obstacle to their political ambitions (good ideology; wrong species as E.O. Wilson said of communism, a political system that he declared might work well for ants, but is reliably disastrous for Homo sapiens).

The second motivation comes from a well-meaning urge to reject eugenics on moral grounds. The horror most modern people feel when they hear the word is justified by the atrocities associated with the first movement. The extermination programmes of the Nazis, for example, were directly inspired by the eugenics movement of the Anglosphere, not least the programmes permitted by American eugenic legislation that saw more than 64,000 individuals forcibly sterilised between 1907 and 1963, disproportionately African-American and indigenous women. The procedure was known in medical slang as a Mississippi appendectomy.

The instinct to condemn as racist trash even a partial defence of the science of eugenics is rooted in the recognition that this science has, within living memory, been used to justify many evil deeds. And this is a question that any defender of the new eugenics must provide an answer to: is that process inevitable? Does a widespread belief that some genes are better or worse than others lead to the widespread conclusion that some people are better or worse than others? And does this conclusion always lead to some very dark places?

We are about to find out. The new eugenics will shortly be with us, although it will not describe itself as such. It will be described with euphemisms such as genetic enhancement or genetic health.

Unlike the first eugenics movement, which attempted to harness the power of the state to determine who should and should not be encouraged (or forbidden) to reproduce, the new version will not concern itself especially with government policy. Rather, it will mostly take the form of private individuals quietly opting for new commercial services like polygenic screening and, in the future, more radical biotech. These individuals will typically spend large sums of money on these services because they will have reached the conclusion that socially desirable traits such as intelligence and beauty are heavily influenced by genetics.

Some countries may well subsidise polygenic screening. Israel already offers its citizens free IVF services, and China has recently announced its intention to do the same. Laws that permit or incentivise the use of these biotech services can accurately be described as eugenic laws, albeit not ones written with the intention of manipulating the gene pool at scale.

My prediction is that the new eugenics will be just as popular as the first which is to say, very. What is often forgotten about the first eugenics movement is how extraordinarily influential it was in its day, particularly among the self-defined progressive upper-middle classes of Britain and America.

The best contemporary comparison is perhaps the environmentalist movement, which has also achieved rapid mainstreaming within a few decades. Like environmentalism, eugenics was endorsed by the most prestigious scientific associations and journals. Like environmentalism, it found passionate advocates among celebrities and the socially conscious middle classes. It wasnt popular only among Wasp conservatives. Black progressives Kelly Miller and W.E.B. Dubois were eugenicists, for example, as were some of the leading socialists of the day. For the Fabian reformer Sidney Webb, the first eugenics movement combined perfectly with his famous injunction to Interfere! Interfere! Interfere! Moulding a healthier and more intelligent population was regarded as not just a virtuous cause, but a duty.

One technology the first eugenicists made use of was abortion. Margaret Sanger, the founder of Planned Parenthood, was a prominent eugenicist, as was Marie Stopes, her British counterpart, who gave her name to Marie Stopes International (MSI), one of the worlds foremost providers of abortion services to this day. So great was Stopess eugenics fervour that in 1947 she forbade her son from marrying a beautiful heiress because the woman was short-sighted. After he went ahead anyway, Stopes cut him out of her will.

When criticism of eugenics came, it was mainly from Catholics, in part because most eugenicists vigorously endorsed the use of both birth control and abortion to further their goals. G.K. Chesterton was perhaps the best-known opponent of the movement. He once wrote a comic story about a woman (strongly reminiscent of Stopes) who breaks off an engagement after her fianc falls off his bicycle, since this revealed his genetic feebleness. He condemned eugenics as a thing no more to be bargained about than poisoning.

But such objections were rare before the second world war. And even after Nazi atrocities were made known, it took some decades for the word eugenics to fall entirely out of favour (the American Eugenics Society did not change its name until 1973).

Yet while the term is now stigmatised, plenty of eugenic laws and practices remain popular. In a recent essay titled Youre Probably a Eugenicist, the evolutionary psychologist Diana Fleischman points out that the efforts of the non-profit organisation Dor Yeshorim to reduce the incidence of Tay-Sachs disease and cystic fibrosis in Jewish families could accurately be described as eugenicist. The practice of aborting foetuses likely to be affected by Downs Syndrome is also eugenicist. Laws forbidding sibling or cousin marriage are definitely eugenicist, in that they are motivated by a desire to reduce the incidence of disease caused by inbreeding. And whats more, as Fleischman writes: Gay men and lesbian women in the US often use gamete donors from egg and sperm banks to have kids in a process that is transparently eugenicOrganisations that recruit egg and sperm donors dont just recruit for fertility, they also screen for mental and physical health, height, education and criminal history because thats what their clients want and expect.

It is common to find people on the left who reject the role of nature altogether

The current bien-pensant position on eugenics is to talk the talk on opposing it, without walking the walk. The increasing availability of polygenic screening is likely to make that dissonance more obvious.

Jonathan Anomaly is one of the few philosophers thinking seriously about the ethical implications. In his 2020 book, Creating Future People, he explored the many practical and moral problems that might result from widespread use of polygenic screening, including the risk of what evolutionary biologists call speciation: that is, a group becoming so genetically distinct from the rest of its species that the two populations are no longer genetically similar enough to interbreed. Strange as this may sound, the run-away use of polygenic screening by an elite group could result in just such an outcome. The social and political differences between the two human species would then become so enormous that the fracturing of polities would be likely, with genetically enhanced people eventually forming their own nation states that exclude the non-enhanced.

Personally, I share the nervousness that many feel in response to the unnaturalness of polygenic screening. But it is worth remembering how unnatural our modern lives already are not least the artificially low levels of child mortality we now enjoy. For most of our species history, something in the region of 40-50 per cent of children would die before their 15th birthdays. Now, the rate globally is at about 4 per cent, and much lower in the rich world.

This is an enormous blessing. It also ensures that people who in other eras would have died as children perhaps including me, as a fairly sickly asthmatic are now able to pass on the genes that make them vulnerable to premature disease and death. This so-called crumbling genome problem means that without the use of genetic enhancement technology of some kind, we will become steadily more genetically sick as a species: childhood cancers will become more common, our immune systems will become weaker and we will become steadily more reliant on modern medical technology to allow us to weather threats. If for any reason those medical systems fail, its game over.

The only natural way out of this quandary is to return to historically normal child mortality rates a possibility that strikes terror into my parental heart. The truth is that any parent grateful for unnatural technologies such as vaccines and effective treatments for childhood cancer should also be open to the prospect of using other unnatural technologies such as polygenic screening.

Parents have historically moved heaven and earth to protect the health and happiness of their children. We should expect those of the future to do the same. And very soon they will have another tool at their disposal: a radical and potentially dangerous tool, but one that any parent with the means to acquire it will almost certainly be unable to resist.

More:

The quiet return of eugenics - The Spectator

Posted in Eugenics | Comments Off on The quiet return of eugenics – The Spectator

Ep 2 | The Monster Masterminds of Eugenics | The Beck Story – iHeart

Posted: at 7:49 pm

This is the story of Madison Grant, a wealthy conservationist, influential progressive, and one of the fathers of eugenics in the U.S. His 1916 book, The Passing of the Great Race, prompted a flattering letter from Adolf Hitler about the book, which Hitler called his bible. Progressives regarded eugenics as cutting-edge science, but in the hands of experts, this science led to unspeakable horrors, the consequences of which are still felt today.

Sponsors

Jase

History teaches us to have at least a little bit of skepticism for what the so-called experts tell us at any given time.In fact, its best to have as many of the decisions that get made about you and your family in your own hands, and not in theirs.For instance, what about life-saving medications?If theres an emergency and you or a member of your family desperately needs a medication that just isnt available right now, what are you going to do?How do you even prepare for something like that?

You should get a Jase case.

Its a personalized emergency kit that contains essential antibiotics and medications that treat the most common and deadly bacterial infections.It provides five life-saving antibiotics for emergency use.All you have to do is fill out a simple form online, and youll have it in case you need it.There are add-on options too, like EpiPens and Ivermectin.

Jase Medical encourages you to take your familys health into your own hands.Go to https://jasemedical.com and enter code Beck at checkout for a discount on your order.

Relief Factor

I hope its obvious to you by now that when it comes to decisions that directly affect you and the people you love, sometimes its better not to leave things in the hands of the experts.

Years ago, when I was dealing with almost daily, horrible pain in my hands, I sought out the advice of a lot of experts.And do you know what that got me?Medical bills, and no easing of my burden.

That was before I started taking Relief Factor, and things changed for me forever.I got rid of the pain.I got my hope back, my joy backI basically got my whole life back.

And you know what?Im no expert, either, but dont take my word for it even at that.If youre living with aches and pains, see for yourself how Relief Factora daily, drug-free supplementcould help you feel and live better every day.Join the over 1 million people whove turned to Relief Factor and you could start feeling better in 3 weeks or less.Visit https://www.relieffactor.com or call 800-4-RELIEF and save on your first order.

Learn more about your ad choices. Visit megaphone.fm/adchoices

More:

Ep 2 | The Monster Masterminds of Eugenics | The Beck Story - iHeart

Posted in Eugenics | Comments Off on Ep 2 | The Monster Masterminds of Eugenics | The Beck Story – iHeart