БЛОГ

Archive for the ‘supercomputing’ category: Page 55

Mar 14, 2019

Why modern enterprises need to adopt cognitive computing for faster business growth in a digital economy

Posted by in categories: business, economics, robotics/AI, supercomputing

Cognitive computing (CC) technology revolves around making computers adept at mimicking the processes of the human brain, which is basically making them more intelligent. Even though the phrase cognitive computing is used synonymously with AI, the term is closely associated with IBM’s cognitive computer system, Watson. IBM Watson is a supercomputer that leverages AI-based disruptive technologies like machine learning (ML), real-time analysis, natural language processing, etc. to augment decision making and deliver superior outcomes.

Read more

Mar 7, 2019

Physicists Used Supercomputers to Map the Bone-Crushing Pressures Hiding Inside Protons

Posted by in categories: physics, supercomputing

If you shrank yourself down and entered a proton, you’d experience among the most intense pressures found anywhere in the universe.

Read more

Mar 5, 2019

Scientists use machine learning to identify high-performing solar materials

Posted by in categories: engineering, robotics/AI, solar power, supercomputing, sustainability

Finding the best light-harvesting chemicals for use in solar cells can feel like searching for a needle in a haystack. Over the years, researchers have developed and tested thousands of different dyes and pigments to see how they absorb sunlight and convert it to electricity. Sorting through all of them requires an innovative approach.

Now, thanks to a study that combines the power of supercomputing with and experimental methods, researchers at the U.S. Department of Energy’s (DOE) Argonne National Laboratory and the University of Cambridge in England have developed a novel “design to device” approach to identify promising materials for dye-sensitized solar cells (DSSCs). DSSCs can be manufactured with low-cost, scalable techniques, allowing them to reach competitive performance-to-price ratios.

The team, led by Argonne materials scientist Jacqueline Cole, who is also head of the Molecular Engineering group at the University of Cambridge’s Cavendish Laboratory, used the Theta supercomputer at the Argonne Leadership Computing Facility (ALCF) to pinpoint five high-performing, low-cost dye materials from a pool of nearly 10,000 candidates for fabrication and device testing. The ALCF is a DOE Office of Science User Facility.

Continue reading “Scientists use machine learning to identify high-performing solar materials” »

Feb 18, 2019

How Do Particles Escape Black Holes? Supercomputers May Have the Answer

Posted by in categories: cosmology, particle physics, supercomputing

The gravitational pull of a black hole is so strong that nothing, not even light, can escape once it gets too close. However, there is one way to escape a black hole — but only if you’re a subatomic particle.

As black holes gobble up the matter in their surroundings, they also spit out powerful jets of hot plasma containing electrons and positrons, the antimatter equivalent of electrons. Just before those lucky incoming particles reach the event horizon, or the point of no return, they begin to accelerate. Moving at close to the speed of light, these particles ricochet off the event horizon and get hurled outward along the black hole’s axis of rotation.

Known as relativistic jets, these enormous and powerful streams of particles emit light that we can see with telescopes. Although astronomers have observed the jets for decades, no one knows exactly how the escaping particles get all that energy. In a new study, researchers with Lawrence Berkeley National Laboratory (LBNL) in California shed new light on the process. [The Strangest Black Holes in the Universe].

Read more

Feb 1, 2019

Exploring the dark side of the genome

Posted by in categories: biotech/medical, genetics, supercomputing

Dr. Rico explained: “When we compare human genomes from different people, we see that they are way more different than we initially expected when the Human Genome Project was declared to be ”completed” in 2003. One of the main contributions to these differences are the so called Copy Number Variable (CNV) regions. CNV regions are in different copy number depending on each individual, and their variability can be greater in some human populations than others. The number of copies of CNV regions can contribute to both normal phenotypic variability in the populations and susceptibility to certain diseases.


Research has shown a direct relationship between mutations in introns and variability in human populations.

One of the greatest challenges of genomics is to reveal what role the ”dark side” of the human genome plays: those regions where it has not yet been possible to find specific functions. The role that introns play within that immense part of the genome is especially mysterious. The introns, which represent almost half the size of the human genome, are constitutive parts of genes that alternate with regions that code for proteins, called exons.

Continue reading “Exploring the dark side of the genome” »

Jan 29, 2019

Quantum Computing Research at NASA

Posted by in categories: quantum physics, supercomputing

A participation in the annual Supercomputing conference taking place in Salt Lake City, UT, USA from November 14–17, 2016.

Read more

Jan 26, 2019

Big Pharma’s Drug Studies Are Getting a NASA-Style Makeover

Posted by in categories: biotech/medical, supercomputing, surveillance

Trying to streamline an operation that spends more than $5 billion a year on developing new drugs, Novartis dispatched teams to jetmaker Boeing Co. and Swissgrid AG, a power company, to observe how they use technology-laden crisis centers to prevent failures and blackouts. That led to the design of something that looks like the pharma version of NASA’s Mission Control: a global surveillance hub where supercomputers map and chart Novartis’s network of 500 drug studies in 70 countries, trying to predict potential problems on a minute-by-minute basis.


A third of development costs comes from clinical trials. Novartis wants to make them cheaper and faster.

Read more

Jan 19, 2019

Why it is dangerous to build ever larger big bang machines

Posted by in categories: alien life, astronomy, cosmology, energy, engineering, ethics, existential risks, general relativity, governance, gravity, innovation, law, nuclear energy, nuclear weapons, particle physics, philosophy, physics, policy, quantum physics, science, scientific freedom, security, singularity, space travel, supercomputing, theory, time travel

CERN has revealed plans for a gigantic successor of the giant atom smasher LHC, the biggest machine ever built. Particle physicists will never stop to ask for ever larger big bang machines. But where are the limits for the ordinary society concerning costs and existential risks?

CERN boffins are already conducting a mega experiment at the LHC, a 27km circular particle collider, at the cost of several billion Euros to study conditions of matter as it existed fractions of a second after the big bang and to find the smallest particle possible – but the question is how could they ever know? Now, they pretend to be a little bit upset because they could not find any particles beyond the standard model, which means something they would not expect. To achieve that, particle physicists would like to build an even larger “Future Circular Collider” (FCC) near Geneva, where CERN enjoys extraterritorial status, with a ring of 100km – for about 24 billion Euros.

Experts point out that this research could be as limitless as the universe itself. The UK’s former Chief Scientific Advisor, Prof Sir David King told BBC: “We have to draw a line somewhere otherwise we end up with a collider that is so large that it goes around the equator. And if it doesn’t end there perhaps there will be a request for one that goes to the Moon and back.”

“There is always going to be more deep physics to be conducted with larger and larger colliders. My question is to what extent will the knowledge that we already have be extended to benefit humanity?”

Continue reading “Why it is dangerous to build ever larger big bang machines” »

Jan 10, 2019

Scientists Discover That Our Brains Can Process the World in 11 Dimensions

Posted by in categories: neuroscience, supercomputing

Neuroscientists have used a classic branch of maths in a totally new way to peer into the structure of our brains. What they’ve discovered is that the brain is full of multi-dimensional geometrical structures operating in as many as 11 dimensions.

We’re used to thinking of the world from a 3D perspective, so this may sound a bit tricky, but the results of this new study could be the next major step in understanding the fabric of the human brain – the most complex structure we know of.

This latest brain model was produced by a team of researchers from the Blue Brain Project, a Swiss research initiative devoted to building a supercomputer-powered reconstruction of the human brain.

Continue reading “Scientists Discover That Our Brains Can Process the World in 11 Dimensions” »

Jan 2, 2019

Cosmic Ray Showers Crash Supercomputers. Here’s What to Do About It

Posted by in categories: military, particle physics, space, supercomputing

The Cray-1 supercomputer, the world’s fastest back in the 1970s, does not look like a supercomputer. It looks like a mod version of that carnival ride The Round Up, the one where you stand, strapped in, as it dizzies you up. It’s surrounded by a padded bench that conceals its power supplies, like a cake donut, if the hole was capable of providing insights about nuclear weapons.

After Seymour Cray first built this computer, he gave Los Alamos National Laboratory a six-month free trial. But during that half-year, a funny thing happened: The computer experienced 152 unattributable memory errors. Later, researchers would learn that cosmic-ray neutrons can slam into processor parts, corrupting their data. The higher you are, and the bigger your computers, the more significant a problem this is. And Los Alamos—7,300 feet up and home to some of the world’s swankiest processors—is a prime target.

The world has changed a lot since then, and so have computers. But space has not. And so Los Alamos has had to adapt—having its engineers account for space particles in its hard- and software. “This is not really a problem we’re having,” explains Nathan DeBardeleben of the High Performance Computing Design group. “It’s a problem we’re keeping at bay.”

Continue reading “Cosmic Ray Showers Crash Supercomputers. Here’s What to Do About It” »

Page 55 of 82First5253545556575859Last