Perovskites power up the solar industry

Tsutomu Miyasaka was on a mission to build a better solar cell. It was the early 2000s, and the Japanese scientist wanted to replace the delicate molecules that he was using to capture sunlight with a sturdier, more effective option.

So when a student told him about an unfamiliar material with unusual properties, Miyasaka had to try it. The material was “very strange,” he says, but he was always keen on testing anything that might respond to light.
Other scientists were running electricity through the material, called a perovskite, to generate light. Miyasaka, at Toin University of Yokohama in Japan, wanted to know if the material could also do the opposite: soak up sunlight and convert it into electricity. To his surprise, the idea worked. When he and his team replaced the light-sensitive components of a solar cell with a very thin layer of the perovskite, the illuminated cell pumped out a little bit of electric current.

The result, reported in 2009 in the Journal of the American Chemical Society, piqued the interest of other scientists, too. The perovskite’s properties made it (and others in the perovskite family) well-suited to efficiently generate energy from sunlight. Perhaps, some scientists thought, this perovskite might someday be able to outperform silicon, the light-absorbing material used in more than 90 percent of solar cells around the world.
Initial excitement quickly translated into promising early results. An important metric for any solar cell is how efficient it is — that is, how much of the sunlight that strikes its surface actually gets converted to electricity. By that standard, perovskite solar cells have shone, increasing in efficiency faster than any previous solar cell material in history. The meager 3.8 percent efficiency reported by Miyasaka’s team in 2009 is up to 22 percent this year. Today, the material is almost on par with silicon, which scientists have been tinkering with for more than 60 years to bring to a similar efficiency level.
“People are very excited because [perovskite’s] efficiency number has climbed so fast. It really feels like this is the thing to be working on right now,” says Jao van de Lagemaat, a chemist at the National Renewable Energy Laboratory in Golden, Colo.

Now, perovskite solar cells are at something of a crossroads. Lab studies have proved their potential: They are cheaper and easier to fabricate than time-tested silicon solar cells. Though perovskites are unlikely to completely replace silicon, the newer materials could piggyback onto existing silicon cells to create extra-effective cells. Perovskites could also harness solar energy in new applications where traditional silicon cells fall flat — as light-absorbing coatings on windows, for instance, or as solar panels that work on cloudy days or even absorb ambient sunlight indoors.

Whether perovskites can make that leap, though, depends on current research efforts to fix some drawbacks. Their tendency to degrade under heat and humidity, for example, is not a great characteristic for a product meant to spend hours in the sun. So scientists are trying to boost stability without killing efficiency.

“There are challenges, but I think we’re well on our way to getting this stuff stable enough,” says Henry Snaith, a physicist at the University of Oxford. Finding a niche for perovskites in an industry so dominated by silicon, however, requires thinking about solar energy in creative ways.

Leaping electrons
Perovskites flew under the radar for years before becoming solar stars. The first known perovskite was a mineral, calcium titanate, or CaTiO3, discovered in the 19th century. In more recent years, perovskites have expanded to a class of compounds with a similar structure and chemical recipe — a 1:1:3 ingredient ratio — that can be tweaked with different elements to make different “flavors.”

But the perovskites being studied for the light-absorbing layer of solar cells are mostly lab creations. Many are lead halide perovskites, which combine a lead ion and three ions of iodine or a related element, such as bromine, with a third type of ion (usually something like methylammonium). Those ingredients link together to form perovskites’ hallmark cagelike pyramid-on-pyramid structure. Swapping out different ingredients (replacing lead with tin, for instance) can yield many kinds of perovskites, all with slightly different chemical properties but the same basic crystal structure.

Perovskites owe their solar skills to the way their electrons interact with light. When sunlight shines on a solar panel, photons — tiny packets of light energy — bombard the panel’s surface like a barrage of bullets and get absorbed. When a photon is absorbed into the solar cell, it can share some of its energy with a negatively charged electron. Electrons are attracted to the positively charged nucleus of an atom. But a photon can give an electron enough energy to escape that pull, much like a video game character getting a power-up to jump a motorbike across a ravine. As the energized electron leaps away, it leaves behind a positively charged hole. A separate layer of the solar cell collects the electrons, ferrying them off as electric current.

The amount of energy needed to kick an electron over the ravine is different for every material. And not all photon power-ups are created equal. Sunlight contains low-energy photons (infrared light) and high-energy photons (sunburn-causing ultraviolet radiation), as well as all of the visible light in between.

Photons with too little energy “will just sail right on through” the light-catching layer and never get absorbed, says Daniel Friedman, a photovoltaic researcher at the National Renewable Energy Lab. Only a photon that comes in with energy higher than the amount needed to power up an electron will get absorbed. But any excess energy a photon carries beyond what’s needed to boost up an electron gets lost as heat. The more heat lost, the more inefficient the cell.
Because the photons in sunlight vary so much in energy, no solar cell will ever be able to capture and optimally use every photon that comes its way. So you pick a material, like silicon, that’s a good compromise — one that catches a decent number of photons but doesn’t waste too much energy as heat, Friedman says.

Although it has dominated the solar cell industry, silicon can’t fully use the energy from higher-energy photons; the material’s solar conversion efficiency tops out at around 30 percent in theory and has hit 20-some percent in practice. Perovskites could do better. The electrons inside perovskite crystals require a bit more energy to dislodge. So when higher-energy photons come into the solar cell, they devote more of their energy to dislodging electrons and generating electric current, and waste less as heat. Plus, by changing the ingredients and their ratios in a perovskite, scientists can adjust the photons it catches. Using different types of perovskites across multiple layers could allow solar cells to more effectively absorb a broader range of photons.

Perovskites have a second efficiency perk. When a photon excites an electron inside a material and leaves behind a positively charged hole, there’s a tendency for the electron to slide right back into a hole. This recombination, as it’s known, is inefficient — an electron that could have fed an electric current instead just stays put.

In perovskites, though, excited electrons usually migrate quite far from their holes, Snaith and others have found by testing many varieties of the material. That boosts the chances the electrons will make it out of the perovskite layer without landing back in a hole.

“It’s a very rare property,” Miyasaka says. It makes for an efficient sunlight absorber.

Some properties of perovskites also make them easier than silicon to turn into solar cells. Making a conventional silicon solar cell requires many steps, all done in just the right order at just the right temperature — something like baking a fragile soufflé. The crystals of silicon have to be perfect, because even small defects in the material can hurt its efficiency. The need for such precision makes silicon solar cells more expensive to produce.

Perovskites are more like brownies from a box — simpler, less finicky. “You can make it in an office, basically,” says materials scientist Robert Chang of Northwestern University in Evanston, Ill. He’s exaggerating, but only a little. Perovskites are made by essentially mixing a bunch of ingredients together and depositing them on a surface in a thin, even film. And while making crystalline silicon requires temperatures up to 2000° Celsius, perovskite crystals form at easier-to-reach temperatures — lower than 200°.

Seeking stability
In many ways, perovskites have become even more promising solar cell materials over time, as scientists have uncovered exciting new properties and finessed the materials’ use. But no material is perfect. So now, scientists are searching for ways to overcome perovskites’ real-world limitations. The most pressing issue is their instability, van de Lagemaat says. The high efficiency levels reported from labs often last only days or hours before the materials break down.

Tackling stability is a less flashy problem than chasing efficiency records, van de Lagemaat points out, which is perhaps why it’s only now getting attention. Stability isn’t a single number that you can flaunt, like an efficiency value. It’s also a bit harder to define, especially since how long a solar cell lasts depends on environmental conditions like humidity and precipitation levels, which vary by location.

Encapsulating the cell with water-resistant coatings is one strategy, but some scientists want to bake stability into the material itself. To do that, they’re experimenting with different perovskite designs. For instance, solar cells containing stacks of flat, graphenelike sheets of perovskites seem to hold up better than solar cells with the standard three-dimensional crystal and its interwoven layers.

In these 2-D perovskites, some of the methylammonium ions are replaced by something larger, like butylammonium. Swapping in the bigger ion forces the crystal to form in sheets just nanometers thick, which stack on top of each other like pages in a book, says chemist Aditya Mohite of Los Alamos National Laboratory in New Mexico. The butylammonium ion, which naturally repels water, forms spacer layers between the 2-D sheets and stops water from permeating into the crystal.
Getting the 2-D layers to line up just right has proved tricky, Mohite says. But by precisely controlling the way the layers form, he and colleagues created a solar cell that runs at 12.5 percent efficiency while standing up to light and humidity longer than a similar 3-D model, the team reported in 2016 in Nature. Although it was protected with a layer of glass, the 3-D perovskite solar cell lost performance rapidly, within a few days, while the 2-D perovskite withered only slightly. (After three months, the 2-D version was still working almost as well as it had been at the beginning.)

Despite the seemingly complex structure of the 2-D perovskites, they are no more complicated to make than their 3-D counterparts, says Mercouri Kanatzidis, a chemist at Northwestern and a collaborator on the 2-D perovskite project. With the right ingredients, he says, “they form on their own.”

His goal now is to boost the efficiency of 2-D perovskite cells, which don’t yet match up to their 3-D counterparts. And he’s testing different water-repelling ions to reach an ideal stability without sacrificing efficiency.

Other scientists have mixed 2-D and 3-D perovskites to create an ultra-long-lasting cell — at least by perovskite standards. A solar panel made of these cells ran at only 11 percent efficiency, but held up for 10,000 hours of illumination, or more than a year, according to research published in June in Nature Communications. And, importantly, that efficiency was maintained over an area of about 50 square centimeters, more on par with real-world conditions than the teeny-tiny cells made in most research labs.

A place for perovskites?
With boosts to their stability, perovskite solar cells are getting closer to commercial reality. And scientists are assessing where the light-capturing material might actually make its mark.

Some fans have pitted perovskites head-to-head with silicon, suggesting the newbie could one day replace the time-tested material. But a total takeover probably isn’t a realistic goal, says Sarah Kurtz, codirector of the National Center for Photovoltaics at the National Renewable Energy Lab.

“People have been saying for decades that silicon can’t get lower in cost to meet our needs,” Kurtz says. But, she points out, the price of solar energy from silicon-based panels has dropped far lower than people originally expected. There are a lot of silicon solar panels out there, and a lot of commercial manufacturing plants already set up to deal with silicon. That’s a barrier to a new technology, no matter how great it is. Other silicon alternatives face the same limitation. “Historically, silicon has always been dominant,” Kurtz says.
For Snaith, that’s not a problem. He cofounded Oxford Photo-voltaics Limited, one of the first companies trying to commercialize perovskite solar cells. His team is developing a solar cell with a perovskite layer over a standard silicon cell to make a super-efficient double-decker cell. That way, Snaith says, the team can capitalize on the massive amount of machinery already set up to build commercial silicon solar cells.
A perovskite layer on top of silicon would absorb higher-energy photons and turn them into electricity. Lower-energy photons that couldn’t excite the perovskite’s electrons would pass through to the silicon layer, where they could still generate current. By combining multiple materials in this way, it’s possible to catch more photons, making a more efficient cell.

That idea isn’t new, Snaith points out: For years, scientists have been layering various solar cell materials in this way. But these double-decker cells have traditionally been expensive and complicated to make, limiting their applications. Perovskites’ ease of fabrication could change the game. Snaith’s team is seeing some improvement already, bumping the efficiency of a silicon solar cell from 10 to 23.6 percent by adding a perovskite layer, for example. The team reported that result online in February in Nature Energy.

Rather than compete with silicon solar panels for space on sunny rooftops and in open fields, perovskites could also bring solar energy to totally new venues.

“I don’t think it’s smart for perovskites to compete with silicon,” Miyasaka says. Perovskites excel in other areas. “There’s a whole world of applications where silicon can’t be applied.”

Silicon solar cells don’t work as well on rainy or cloudy days, or indoors, where light is less direct, he says. Perovskites shine in these situations. And while traditional silicon solar cells are opaque, very thin films of perovskites could be printed onto glass to make sunlight-capturing windows. That could be a way to bring solar power to new places, turning glassy skyscrapers into serious power sources, for example. Perovskites could even be printed on flexible plastics to make solar-powered coatings that charge cell phones.

That printing process is getting closer to reality: Scientists at the University of Toronto recently reported a way to make all layers of a perovskite solar cell at temperatures below 150° — including the light-absorbing perovskite layer, but also the background workhorse layers that carry the electrons away and funnel them into current. That could streamline and simplify the production process, making mass newspaper-style printing of perovskite solar cells more doable.

Printing perovskite solar cells on glass is also an area of interest for Oxford Photovoltaics, Snaith says. The company’s ultimate target is to build a perovskite cell that will last 25 years, as long as a traditional silicon cell.

Moon had a magnetic field for at least a billion years longer than thought

The moon had a magnetic field for at least 2 billion years, or maybe longer.

Analysis of a relatively young rock collected by Apollo astronauts reveals the moon had a weak magnetic field until 1 billion to 2.5 billion years ago, at least a billion years later than previous data showed. Extending this lifetime offers insights into how small bodies generate magnetic fields, researchers report August 9 in Science Advances. The result may also suggest how life could survive on tiny planets or moons.
“A magnetic field protects the atmosphere of a planet or moon, and the atmosphere protects the surface,” says study coauthor Sonia Tikoo, a planetary scientist at Rutgers University in New Brunswick, N.J. Together, the two protect the potential habitability of the planet or moon, possibly those far beyond our solar system.

The moon does not currently have a global magnetic field. Whether one ever existed was a question debated for decades (SN: 12/17/11, p. 17). On Earth, molten rock sloshes around the outer core of the planet over time, causing electrically conductive fluid moving inside to form a magnetic field. This setup is called a dynamo. At 1 percent of Earth’s mass, the moon would have cooled too quickly to generate a long-lived roiling interior.
Magnetized rocks brought back by Apollo astronauts, however, revealed that the moon must have had some magnetizing force. The rocks suggested that the magnetic field was strong at least 4.25 billion years ago, early on in the moon’s history, but then dwindled and maybe even got cut off about 3.1 billion years ago.
Tikoo and colleagues analyzed fragments of a lunar rock collected along the southern rim of the moon’s Dune Crater during the Apollo 15 mission in 1971. The team determined the rock was 1 billion to 2.5 billion years old and found it was magnetized. The finding suggests the moon had a magnetic field, albeit a weak one, when the rock formed, the researchers conclude.
A drop in the magnetic field strength suggests the dynamo driving it was generated in two distinct ways, Tikoo says. Early on, Earth and the moon would have sat much closer together, allowing Earth’s gravity to tug on and spin the rocky exterior of the moon. That outer layer would have dragged against the liquid interior, generating friction and a very strong magnetic field (SN Online: 12/4/14).

Then slowly, starting about 3.5 billion years ago, the moon moved away from Earth, weakening the dynamo. But by that point, the moon would have started to cool, causing less dense, hotter material in the core to rise and denser, cooler material to sink, as in Earth’s core. This roiling of material would have sustained a weak field that lasted for at least a billion years, until the moon’s interior cooled, causing the dynamo to die completely, the team suggests.

The two-pronged explanation for the moon’s dynamo is “an entirely plausible idea,” says planetary scientist Ian Garrick-Bethell of the University of California, Santa Cruz. But researchers are just starting to create computer simulations of the strength of magnetic fields to understand how such weaker fields might arise. So it is hard to say exactly what generated the lunar dynamo, he says.

If the idea is correct, it may mean other small planets and moons could have similarly weak, long-lived magnetic fields. Having such an enduring shield could protect those bodies from harmful radiation, boosting the chances for life to survive.

Here are the paths of the next 15 total solar eclipses

August’s total solar eclipse won’t be the last time the moon cloaks the sun’s light. From now to 2040, for example, skywatchers around the globe can witness 15 such events.

Their predicted paths aren’t random scribbles. Solar eclipses occur in what’s called a Saros cycle — a period that lasts about 18 years, 11 days and eight hours, and is governed by the moon’s orbit. (Lunar eclipses follow a Saros cycle, too, which the Chaldeans first noticed probably around 500 B.C.)

Two total solar eclipses separated by that 18-years-and-change period are almost twins — compare this year’s eclipse with the Sept. 2, 2035 eclipse, for example. They take place at roughly the same time of year, at roughly the same latitude and with the moon at about the same distance from Earth. But those extra eight hours, during which the Earth has rotated an additional third of the way on its axis, shift the eclipse path to a different part of the planet.
This cycle repeats over time, creating a family of eclipses called a Saros series. A series lasts 12 to 15 centuries and includes about 70 or more eclipses. The solar eclipses of 2019 and 2037 belong to a different Saros series, so their paths too are shifted mimics. Their tracks differ in shape from 2017’s, because the moon is at a different place in its orbit when it passes between the Earth and the sun. Paths are wider at the poles because the moon’s shadow is hitting the Earth’s surface at a steep angle.

Predicting and mapping past and future eclipses allows scientists “to examine the patterns of eclipse cycles, the most prominent of which is the Saros,” says astrophysicist Fred Espenak, who is retired from NASA’s Goddard Spaceflight Center in Greenbelt, Md.

He would know. Espenak and his colleague Jean Meeus, a retired Belgian astronomer, have mapped solar eclipse paths from 2000 B.C. to A.D. 3000. For archaeologists and historians peering backward, the maps help match up accounts of long-ago eclipses with actual paths. For eclipse chasers peering forward, the data are an itinerary.

“I got interested in figuring out how to calculate eclipse paths for my own use, for planning … expeditions,” says Espenak, who was 18 when he witnessed his first total solar eclipse. It was in 1970, and he secured permission to drive the family car from southern New York to North Carolina to see it. Since then, Espenak, nicknamed “Mr. Eclipse,” has been to every continent, including Antarctica, for a total eclipse of the sun.

“It’s such a dramatic, spectacular, beautiful event,” he says. “You only get a few brief minutes, typically, of totality before it ends. After it’s over, you’re craving to see it again.”

Rumors swirl that LIGO snagged gravitational waves from a neutron star collision

Speculation is running rampant about potential new discoveries of gravitational waves, just as the latest search wound down August 25.

Publicly available logs from astronomical observatories indicate that several telescopes have been zeroing in on one particular region of the sky, potentially in response to a detection of ripples in spacetime by the Advanced Laser Interferometer Gravitational-Wave Observatory, LIGO. These records have raised hopes that, for the first time, scientists may have glimpsed electromagnetic radiation — light — produced in tandem with gravitational waves. That light would allow scientists to glean more information about the waves’ source. Several tweets from astronomers reporting rumors of a new LIGO detection have fanned the flames of anticipation and amplified hopes that the source may be a cosmic convulsion unlike any LIGO has seen before.
“There is a lot of excitement,” says astrophysicist Rosalba Perna of Stony Brook University in New York, who is not involved with the LIGO collaboration. “We are all very anxious to actually see the announcement.”

An Aug. 25 post on the LIGO collaboration’s website announced the end of the current round of data taking, which began November 30, 2016. Virgo, a gravitational wave detector in Italy, had joined forces with LIGO’s two on August 1 (SN Online: 8/1/17). The three detectors will now undergo upgrades to improve their sensitivity. The update noted that “some promising gravitational-wave candidates have been identified in data from both LIGO and Virgo during our preliminary analysis, and we have shared what we currently know with astronomical observing partners.”

When LIGO detects gravitational waves, the collaboration alerts astronomers to the approximate location the waves seemed to originate from. The hope is that a telescope could pick up light from the aftermath of the cosmic catastrophe that created the gravitational waves — although no light has been found in previous detections.

LIGO previously detected three sets of gravitational waves from merging black holes (SN: 6/24/17, p. 6). Black hole coalescences aren’t expected to generate light that could be spotted by telescopes, but another prime candidate could: a smashup between two remnants of stars known as neutron stars. Scientists have been eagerly awaiting LIGO’s first detections of such mergers, which are suspected to be the sites where the universe’s heaviest elements are formed. An observation of a neutron star crash also could provide information about the ultradense material that makes up neutron stars.
Since mid-August, seemingly in response to a LIGO alert, several telescopes have observed a section of sky around the galaxy NGC 4993, located 134 million light-years away in the constellation Hydra. The Hubble Space Telescope has made at least three sets of observations in that vicinity, including one on August 22 seeking “observations of the first electromagnetic counterparts to gravitational wave sources.”

Likewise, the Chandra X-ray Observatory targeted the same region of sky on August 19. And records from the Gemini Observatory’s telescope in Chile indicate several potentially related observations, including one referencing “an exceptional LIGO/Virgo event.”

“I think it’s very, very likely that LIGO has seen something,” says astrophysicist David Radice of Princeton University, who is not affiliated with LIGO. But, he says, he doesn’t know whether its source has been confirmed as merging neutron stars.

LIGO scientists haven’t commented directly on the veracity of the rumor. “We have some substantial work to do before we will be able to share with confidence any quantitative results. We are working as fast as we can,” LIGO spokesperson David Shoemaker of MIT wrote in an e-mail.

Tabby’s star is probably just dusty, and still not an alien megastructure

Alien megastructures are out. The unusual fading of an oddball star is more likely caused by either clouds of dust or an abnormal cycle of brightening and dimming, two new papers suggest.

Huan Meng of the University of Arizona in Tucson and his colleagues suggest that KIC 8462852, known as Tabby’s star, is dimming thanks to an orbiting cloud of fine dust particles. The team observed the star with the infrared Spitzer and ultraviolet Swift space telescopes from October 2015 to December 2016 — the first observations in multiple wavelengths of light. They found that the star is dimming faster in short blue wavelengths than longer infrared ones, suggesting smaller particles.
“That almost absolutely ruled out the alien megastructure scenario, unless it’s an alien microstructure,” Meng says.

Tabby’s star is most famous for suddenly dropping in brightness by up to 22 percent over the course of a few days (SN Online: 2/2/16). Later observations suggested the star is also fading by about 4 percent per year (SN: 9/17/16, p. 12), which Meng’s team confirmed in a paper posted online August 24 at arXiv.org.

Joshua Simon of the Observatories of the Carnegie Institution for Science in Pasadena, Calif., found a similar dimming in data on Tabby’s star from the All Sky Automated Survey going back to 2006. Simon and colleagues also found for the first time that the star grew brighter in 2014, and possibly in 2006, they reported in a paper August 25 at arXiv.org.

“That’s fascinating,” says astrophysicist Tabetha Boyajian of Louisiana State University in Baton Rouge. She first reported the star’s flickers in 2015 (the star is nicknamed for her) and is a coauthor on Meng’s paper. “We always speculated that it would brighten sometime. It can’t just get fainter all the time — otherwise it would disappear. This shows that it does brighten.”

The brightening could be due to a magnetic cycle like the sun’s, Simon suggests. But no known cycle makes a star brighten and dim by quite so much, so the star would still be odd.
Brian Metzger of Columbia University previously suggested that a ripped-up planet falling in pieces into the star could explain both the long-term and short-term dimming. He thinks that model still works, although it needs some tweaks.

“This adds some intrigue to what’s going on, but I don’t think it really changes the landscape,” says Metzger, who was not involved in the new studies. And newer observations could complicate things further: The star went through another bout of dimming between May and July. “I’m waiting to see the papers analyzing this recent event,” Metzger says.

50 years ago, West Germany embraced nuclear power

West German power companies have decided to go ahead with two nuclear power station projects…. Compared with the U.S. and Britain, Germany has been relatively backward in the application of nuclear energy…. The slow German start is only partly the result of restrictions placed upon German nuclear research after the war. — Science News, September 16, 1967

Update
Both East and West Germany embraced nuclear power until antinuclear protests in the 1970s gathered steam. In 1998, the unified German government began a nuclear phaseout, which Chancellor Angela Merkel halted in 2009. The 2011 Fukushima nuclear disaster in Japan caused a rapid reversal. Germany closed eight of its nuclear plants immediately, and announced that all nuclear power in the country would go dark by 2022 (SN Online: 6/1/11). A pivot to renewable energy — wind, solar, hydropower and biomass — produced 188 billion kilowatt-hours of electricity in 2016, nearly 32 percent of German electricity usage.

Hundreds of dietary supplements are tainted with potentially harmful drugs

From 2007 to 2016, the U.S. Food and Drug Administration flagged nearly 800 over-the-counter dietary supplements as tainted with potentially harmful pharmaceutical drugs, a study shows. Fewer than half of those products were recalled by their makers, scientists found.

Researchers analyzed the FDA’s public database of tainted supplements, identifying both the type of contaminating ingredients they contained and how the products were marketed. Most of these supplements, which are allowed to contain only dietary ingredients, included drugs such as steroids, the active ingredient in Viagra and a weight loss drug banned from the U.S. market eight years ago. The products had been marketed primarily for sexual enhancement, weight loss or muscle building, scientists report online October 12 in JAMA Network Open.

More than half of American adults have reported taking dietary supplements, such as vitamins, minerals and other specialty products. More than 85,000 supplements are estimated to be available in the United States, and the FDA says it cannot test all of them.
No No’s
These pharmaceutical ingredients are not permitted in dietary supplements, but were found to be contaminating supplements.

Sildenafil
What it is: A medication that dilates blood vessels in the penis, and is the active ingredient in Viagra
Health issue: Can lower blood pressure to levels that are unsafe for people taking medications for diabetes, high blood pressure or high cholesterol
Supplement type: Sexual enhancement
Sibutramine
What it is: An appetite suppressant removed from the U.S. market in 2010
Health issue: Increased risk of heart attack or stroke
Supplement type: Weight loss
Phenolphthalein
What it is: A laxative removed from the U.S. market in 1999
Health issue: Potential carcinogen
Supplement type: Weight loss
Anabolic steroids
What they are: Chemicals related to the male sex hormone testosterone
Health issue: Associated with liver injury, kidney damage, heart attack and stroke
Supplement type: Muscle building
Aromatase inhibitors
What they are: A class of drugs that lower estrogen levels, and are used to treat breast cancer
Health issue: Associated with decreased bone growth, infertility, liver dysfunction
Supplement type: Muscle building
These supplements aren’t subject to the same regulations, testing and approval process that are required for pharmaceutical drugs. But if the FDA identifies tainted supplements after they’re on the market, the agency can issue public warnings or suggest the company voluntarily remove the product.

Whether that approach is effective raises questions, though, says general internist Pieter Cohen of Cambridge Health Alliance in Cambridge, Mass., who was not involved in the new work. Voluntary recalls don’t necessarily mean a product is completely removed from shelves or that consumers become aware and stop using a product, Cohen’s research has found.

And only 360 of the 776 supplements flagged as tainted from 2007 to 2016 were recalled, the study found. “What really jumped out at me,” Cohen says, is that “when the FDA detects drugs in supplements, more than half the time the product isn’t even recalled.”

Supplement use does carry health risks. A 2015 study estimated that 23,000 emergency room visits each year are due to health problems related to dietary supplements. Of those, about 2,100 patients are hospitalized annually, commonly for symptoms related to heart trouble.
In 2013, 20 percent of drug-induced liver injury cases recorded in the Drug-Induced Liver Injury Network registry were caused by dietary supplements. That’s up from 7 percent in 2004. Liver damage can be fatal or require a liver transplant. A 2013 report by the U.S. Centers for Disease Control and Prevention on 29 cases of liver injury found that 24 of those patients reported using a dietary supplement for weight loss.

“The law allows companies to advertise supplements as if they’re good for your health, even if there’s no evidence in humans that that’s the case,” Cohen says. He began studying dietary supplements after noting that his patients developed health problems, including panic attacks, chest pain and kidney failure, related to weight-loss supplements. One patient was suspended from his job when his urine tested positive for amphetamine; a chemical derivative of the drug was found in the weight-loss pills that he was taking.

Cohen’s recommendation? Avoid supplements “that promise you anything.”

Virtual avatars learned cartwheels and other stunts from videos of people

Animated characters can learn from online tutorials, too.

A new computer program teaches virtual avatars new skills, such as dances, acrobatic stunts and martial art moves, from YouTube videos. This kind of system, described in the November ACM Transactions on Graphics, could render more physically coordinated characters for movies and video games, or serve as a virtual training ground for robots.

“I was really impressed” by the program, says Daniel Holden, a machine-learning researcher at Ubisoft La Forge in Montreal not involved in the work. Rendering accurate, natural-looking movements based on everyday video clips “has always been a goal for researchers in this field.”
Animated characters typically have learned full-body motions by studying motion capture data, collected by a camera that tracks special markers attached to actors’ bodies. But this technique requires special equipment and often works only indoors.

The new program leverages a type of computer code known as an artificial neural network, which roughly mimics how the human brain processes information. Trained on about 100,000 images of people in various poses, the program first estimates an actor’s pose in each frame of a video clip. Then, it teaches a virtual avatar to re-create the actor’s motion using reinforcement learning, giving the character a virtual “reward” when it matches the video actor’s pose in a frame.

Computer scientist Jason Peng and colleagues at the University of California, Berkeley, fed YouTube videos into the system to teach characters to do somersaults, backflips, vaulting and other stunts.
Even characters such as animated Atlas robots with bodies drastically different from those of their human video teachers mastered these motions (SN: 12/13/14, p. 16). Characters could also perform under conditions not seen in the training video, like cartwheeling while being pelted with blocks or moving across terrain riddled with holes.
The work, also reported October 8 at arXiv.org, is a step “toward making motion capture easier, cheaper and more accessible,” Holden says. Videos could be used to render virtual versions of outdoor activities, since motion capture is difficult to do outdoors, or to create lifelike avatars of large animals that would be difficult to stick with motion capture markers.

This kind of program may also someday be used to teach robots new skills, Peng says. An animated version of a robot could master skills in a virtual environment before that learned computer code powered a machine in the physical world.

These animated characters still struggle with nimble dance steps, such as the “Gangnam Style” jig, and learn from short clips featuring only a single person. David Jacobs, a computer scientist at the University of Maryland in College Park not involved in the work, looks forward to future virtual avatars that can reenact longer, more complex actions, such as pairs of people dancing or soccer teams playing a game.

“That’s probably a much harder problem, because [each] person’s not as clearly visible, but it would be really cool,” Jacobs says. “This is only the beginning.”

Physicists finally calculated where the proton’s mass comes from

A proton’s mass is more than just the sum of its parts. And now scientists know just what accounts for the subatomic particle’s heft.

Protons are made up of even smaller particles called quarks, so you might expect that simply adding up the quarks’ masses should give you the proton’s mass. However, that sum is much too small to explain the proton’s bulk. And new, detailed calculations show that only 9 percent of the proton’s heft comes from the mass of constituent quarks. The rest of the proton’s mass comes from complicated effects occurring inside the particle, researchers report in the Nov. 23 Physical Review Letters.

Quarks get their masses from a process connected to the Higgs boson, an elementary particle first detected in 2012 (SN: 7/28/12, p. 5). But “the quark masses are tiny,” says study coauthor and theoretical physicist Keh-Fei Liu of the University of Kentucky in Lexington. So, for protons, the Higgs explanation falls short.

Instead, most of the proton’s 938 million electron volts of mass is due to complexities of quantum chromodynamics, or QCD, the theory which accounts for the churning of particles within the proton. Making calculations with QCD is extremely difficult, so to study the proton’s properties theoretically, scientists rely on a technique called lattice QCD, in which space and time are broken up into a grid, upon which the quarks reside.
Using this technique, physicists had previously calculated the proton’s mass (SN: 12/20/08, p. 13). But scientists hadn’t divvied up where that mass comes from until now, says theoretical physicist André Walker-Loud of Lawrence Berkeley National Laboratory in California. “It’s exciting because it’s a sign that … we’ve really hit this new era” in which lattice QCD can be used to better understand nuclear physics.

In addition to the 9 percent of the proton’s mass that comes from quarks’ heft, 32 percent comes from the energy of the quarks zipping around inside the proton, Liu and colleagues found. (That’s because energy and mass are two sides of the same coin, thanks to Einstein’s famous equation, E=mc2.) Other occupants of the proton, massless particles called gluons that help hold quarks together, contribute another 36 percent via their energy.

The remaining 23 percent arises due to quantum effects that occur when quarks and gluons interact in complicated ways within the proton. Those interactions cause QCD to flout a principle called scale invariance. In scale invariant theories, stretching or shrinking space and time makes no difference to the theories’ results. Massive particles provide the theory with a scale, so when QCD defies scale invariance, protons also gain mass.

The results of the study aren’t surprising, says theoretical physicist Andreas Kronfeld of Fermilab in Batavia, Ill. Scientists have long suspected that the proton’s mass was made up in this way. But, he says, “this kind of calculation replaces a belief with scientific knowledge.”

A ‘fire wolf’ fish could expand what we know about one unusual deep-sea ecosystem

Off the Pacific coast of Costa Rica sits a deep-sea chimera of an ecosystem. Jacó Scar is a methane seep, where the gas escapes from sediment into the seawater, but the seep isn’t cold like the others found before it. Instead, geochemical activity gives the Scar lukewarm water that enables organisms from both traditionally colder seeps and scalding hot hydrothermal vents to call it home.

One resident of the Scar is a newly identified species of small, purplish fish called an eelpout, described for the first time on January 19 in Zootaxa. This fish is the first vertebrate species found at the Scar and could help scientists understand how the unique ecosystem developed.
Jacó Scar was discovered during exploration of a known field of methane seeps off the Costa Rican coast and named for the nearby town of Jacó. It is “a really diverse place” with many different organisms living in various microhabitats, says Lisa Levin, a marine ecologist at Scripps Institution of Oceanography in La Jolla, Calif.

Levin was on one of the first expeditions to the Scar but wasn’t involved in the new study. She recalls the team finding and collecting one of the fish during this early excursion, but the researchers didn’t recognize it as a new species.

Several more specimens were snagged during later submersible dives. Charlotte Seid, an invertebrate biologist at Scripps who is working on a checklist of organisms found at the Costa Rican seeps, brought the fishy finds to ichthyologist Ben Frable, also of Scripps, for formal identification.

Frable says he knew the fish was an eelpout. They look exactly as one would expect based on their name: like frowning eels, though they aren’t true eels. But he was having trouble determining what type. Eelpouts are a diverse family of fish comprised of nearly 300 species that can be found all over the world at various ocean depths.

Because the physical differences between species can be subtle, they are “kind of a tricky group” to identify, Frable says. “I just was not really getting anywhere.” So the team turned to eelpout expert Peter Rask Møller of the Natural History Museum of Denmark in Copenhagen, sending him X-rays, pictures and eventually one of the fish specimens.

Møller narrowed the enigmatic eelpout to the genus Pyrolycus, meaning “fire wolf.” Turns out, the tool, called a dichotomous key, that Frable had been using to identify the specimens was outdated, made before Pyrolycus was described in 2002. “I did not know that genus existed,” Frable says.

Because the other two known Pyrolycus species live far away in the western Pacific and have different physical features, the team dubbed the mystery fish P. jaco — a new species.

The first eelpouts most likely evolved in cold waters, Frable says, but many have since made their home in the scalding waters of hydrothermal vents. Of the 24 known fish species that live only at hydrothermal vents, “13 of them are eelpouts,” Frable says.
The new finding raises questions about how the known Pyrolycus species came to live so far apart. It may have to do with the fact that methane seeps are more common than previously thought on the ocean floor, and if some are lukewarm like Jacó Scar, the new species could have used them as refuges while moving east.

And by comparing P. jaco to its vent-living relatives, researchers may be able to figure out how it adapted to live in the tepid waters of the Scar — which may provide clues to how other species living there did too.

The eelpout is part of a medley of other species that form Jacó Scar’s composite ecosystem, along with, for example, clams typically found at cold seeps and bacteria found at hydrothermal vents. Jacó Scar is a “mixing bowl” of species found in other parts of the world, Seid says. Figuring out how this eclectic bunch interacts “is part of the fun.”