Mooching roommates are an ancient problem. Certain species of beetles evolved to live with and leech off social insects such as ants and termites as long ago as the mid-Cretaceous, two new beetle fossils suggest. The finds date the behavior, called social parasitism, to almost 50 million years earlier than previously thought.
Ants and termites are eusocial — they live in communal groups, sharing labor and collectively raising their young. The freeloading beetles turn that social nature to their advantage. They snack on their hosts’ larvae and use their tunnels for protection, while giving nothing in return.
Previous fossils have suggested that this social parasitism has been going on for about 52 million years. But the new finds push that date way back. The specimens, preserved in 99-million-year-old Burmese amber, would have evolved relatively shortly after eusociality is thought to have popped up.
One beetle, Mesosymbion compactus, was reported in Nature Communications in December 2016. A different group of researchers described the other, Cretotrichopsenius burmiticus, in Current Biology on April 13. Both species have shielded heads and teardrop-shaped bodies, similar to modern termite-mound trespassers. Those adaptations aren’t just for looks. Like a roommate who’s found his leftovers filched one too many times, termites frequently turn against their pilfering housemates.
Smarty-pants have 40 new reasons to thank their parents for their powerful brains. By sifting through the genetics of nearly 80,000 people, researchers have uncovered 40 genes that may make certain people smarter. That brings the total number of suspected “intelligence genes” to 52.
Combined, these genetic attributes explain only a very small amount of overall smarts, or lack thereof, researchers write online May 22 in Nature Genetics. But studying these genes, many of which play roles in brain cell development, may ultimately help scientists understand how intelligence is built into brains. Historically, intelligence research has been mired in controversy, says neuroscientist Richard Haier of the University of California, Irvine. Scientists disagreed on whether intelligence could actually be measured and if so, whether genes had anything at all to do with the trait, as opposed to education and other life experiences. But now “we are so many light-years beyond that, as you can see from studies like this,” says Haier. “This is very exciting and very positive news.”
The results were possible only because of the gigantic number of people studied, says study coauthor Danielle Posthuma, a geneticist at VU University Amsterdam. She and colleagues combined data from 13 earlier studies on intelligence, some published and some unpublished. Posthuma and her team looked for links between intelligence scores, measured in different ways in the studies, and variations held in the genetic instruction books of 78,308 children and adults. Called a genome-wide association study or GWAS, the method looks for signs that certain quirks in people’s genomes are related to a trait.
This technique pointed out particular versions of 22 genes, half of which were not previously known to have a role in intellectual ability. A different technique identified 30 more intelligence genes, only one of which had been previously found. Many of the 40 genes newly linked to intelligence are thought to help with brain cell development. The SHANK3 gene, for instance, helps nerve cells connect to partners.
Together, the genetic variants identified in the GWAS account for only about 5 percent of individual differences in intelligence, the authors estimate. That means that the results, if confirmed, would explain only a very small part of why some people are more intelligent than others. The gene versions identified in the paper are “accounting for so little of the variance that they’re not telling us much of anything,” says differential developmental psychologist Wendy Johnson of the University of Edinburgh.
Still, knowing more about the genetics of intelligence might ultimately point out ways to enhance the trait, an upgrade that would help people at both the high and low ends of the curve, Haier says. “If we understand what goes wrong in the brain, we might be able to intervene,” he says. “Wouldn’t it be nice if we were all just a little bit smarter?”
Posthuma, however, sees many roadblocks. Beyond ethical and technical concerns, basic brain biology is incredibly intricate. Single genes have many jobs. So changing one gene might have many unanticipated effects. Before scientists could change genes to increase intelligence, they’d need to know everything about the entire process, Posthuma says. Tweaking genetics to boost intelligence “would be very tricky.”
Julius Caesar could have stayed home on March 15, 44 B.C. But mocking the soothsayer who had predicted his death, the emperor rode in his litter to Rome’s Forum. There he met the iron daggers of 60 senators.
As he lay in a pool of blood, he may have gasped a final incrimination to his protégé Brutus: You too, my son? Or maybe not. But he certainly would have breathed a dying breath, a final exhalation of some 25 sextillion gas molecules. And it’s entirely possible that you just breathed in one of them. In fact, calculating the probability of a particle of Caesar’s dying breath appearing in any given liter of air (the volume of a deep breath) has become a classic exercise for chemistry and physics students. If you make a few assumptions about the mixing of gases and the lifetimes of molecules in the atmosphere, it turns out that, on average, one molecule of “Caesar air” — or any other historic liter of air, for that matter — appears in each breath you take.
Author Sam Kean begins his book Caesar’s Last Breath with this exercise, noting that “we can’t escape the air of those around us.” It’s all recycled, and every day we breathe in a bit of our, and Earth’s, history. “The story of Earth,” he writes, “is the story of its gases.”
Kean, author of a best seller about the periodic table, The Disappearing Spoon, then tells that story. As he did in his fascinating portraits of the elements, Kean profiles individual gases such as nitrogen and oxygen primarily through the scientists and entrepreneurs who discovered or sought to harness them. These are quirky men (and they are mostly men) — every bit as obsessed, greedy and brilliant as one could hope for in a page-turner.
Along with lesser-known backstories of textbook heroes such as James Watt, Antoine-Laurent Lavoisier and Albert Einstein (who was surprisingly obsessed with building a better refrigerator), Kean clearly delights in weaving in the unexpected. In the discussion of helium, we learn about Joseph-Michel Montgolfier, the papermaker who was inspired to build the first hot-air balloon as he watched his wife’s pantaloons billowing suggestively above a fire. And in a chapter on the radioactive elements carried in nuclear fallout, there’s Pig 311, a sow that survived a nuclear test blast only to be used as propaganda for the weapons’ supposed safety.
Along the way, Kean threads in the history of Earth’s atmosphere in a surprisingly compelling narrative of geologic history. He steps aside from Lavoisier’s work on life-giving oxygen, for example, to describe the Great Oxygenation Event, which infused the atmosphere a couple billion years ago with a gas that, at the time, was toxic to most living things. The explanations of science here and throughout the book are written clearly and at a level that should be understandable with a high school education. And while they’re straightforward, the explanations have enough depth to be satisfying; by the end of the book, you realize you’ve learned quite a bit. Even those who rarely read science will enjoy the drama — death, for instance, plays a big role in these stories. Over and over, we learn, men have taken gases’ powers too lightly, or wielded their own power too cruelly, and paid the price. Fritz Haber, for instance, could have died a hero for finding a way to make fertilizer from the nitrogen in air. Instead, he died broke and loathed for his World War I work on gas warfare.
Then there was Harry Truman — not that Truman, but the one who refused to leave his home when scientists warned of an impending volcanic eruption. Truman contended that officials were “lying like horses trot” right up until Mount St. Helens blew searing gases that erased him from the mountainside.
The links between these stories can seem at first as ephemeral as the gases, but together they tell the story of the birth of the atmosphere and humans’ history in it. In the end, like Caesar’s breath, it all comes full circle.
Books about the history of science, like many other histories, must contend with the realization that others have come before. Their tales have already been told. So such a book is worth reading, or buying, only if it offers something more than the same old stories.
In this case, The Oxford Illustrated History of Science offers most obviously an excellent set of illustrations and photographs from science’s past, from various ancient Egyptian papyruses to the Hubble Space Telescope’s ultradeep view of distant galaxies. Some of the images will be familiar to science fans; many others are obscure but apt; nearly all help illustrate various aspects of science’s history. And yet the pictures, while many may be worth more than 10,000 words, are still just complements to the text. Oxford attempts a novel organization for recounting the story of science: a sometimes hard-to-follow mix of chronological and topical. The first section, “Seeking Origins,” has six chapters that cover ancient Mediterranean science, science in ancient China, medieval science (one chapter for the Islamic world and Europe, one for China), plus the scientific revolution and science in the Enlightenment. The second section, “Doing Science,” shifts to experimenting, fieldwork, biology, cosmology, theory and science communication. Each chapter has a different author, which has the plus of bringing distinct expertise to each subject matter but the minus of vast divergence in readability and caliber of content. Some chapters (see “Exploring Nature,” on field science) are wordy, repetitive and lack scientific substance. Others (“Mapping the Universe”) are compelling, engaging and richly informative. A particularly disappointing chapter on biology (“The Meaning of Life”) focuses on 19th century evolution, with only a few paragraphs for the life science of the 20th and 21st centuries. That chapter closes with an odd, antiscientific tone lamenting the “huge numbers of people … addicted to antidepressants” and complaining that modern biology (and neuroscience) “threatens to undermine traditional values of moral responsibility.”
Some of the book’s strongest chapters are the earliest, especially those that cover aspects of science often missing in other histories, such as science in China. Who knew that the ancient Chinese had their own set of ancient elements — not the Greeks’ air, earth, water and fire, but rather wood, fire, water, soil and metal?
With the book’s second-half emphasis on how science was done rather than what science found out, the history that emerges is sometimes disjointed and out of order. Discussions of the modern view of the universe, which hinges on Einstein’s general theory of relativity, appear before the chapter on theory, where relativity is mentioned. In fact, both relativity and quantum theory are treated superficially in that chapter, as examples of the work of theorists rather than the components of a second scientific revolution. No doubt lack of space prevented deeper treatment of science from the last century. Nevertheless the book’s merits outweigh its weaknesses. For an accessible account of the story of pre-20th century science, it’s informative and enjoyable. For more recent science, you can at least look at the pictures.
This year’s Atlantic hurricane season has already proven to be active and deadly. Powerful hurricanes such as Harvey, Irma and Maria are also providing a testing ground for new tools that scientists hope will save lives by improving forecasts in various ways, from narrowing a storm’s future path to capturing swift changes in the intensity of storm winds.
Some of the tools that debuted this year — such as the GOES-16 satellite — are already winning praise from scientists. Others, such as a new microsatellite system aiming to improve measurements of hurricane intensity and a highly anticipated new computer simulation that forecasts hurricane paths and intensities, are still in the calibration phase. As these tools get an unprecedented workout thanks to an unusually ferocious series of storms, scientists may know in a few months whether hurricane forecasting is about to undergo a sea change.
The National Oceanic and Atmospheric Administration’s GOES-16 satellite is perhaps the clearest success story of this hurricane season so far. Public perceptions of hurricane forecasts tend to focus on uncertainty and conflicting predictions. But in the big picture, hurricane models adeptly forecasted Irma’s ultimate path to the Florida Keys nearly a week before it arrived there, says Brian Tang, an atmospheric scientist at the University at Albany in New York. “I found that remarkable,” he says. “Ten or so years ago that wouldn’t have been possible.”
One reason for this is GOES-16, which launched late last year and will become fully operational in November. The satellite offers images at four times the resolution of previous satellites. “It’s giving unparalleled details about the hurricanes,” Tang says, including data on wind speeds and water temperatures delivered every minute that are then fed into models.
GOES-16’s crystal-clear images also give forecasters a better picture of the winds swirling around a storm’s central eye. But more data from this crucial region is needed to improve predictions of just how strong a hurricane might get. Scientists continue to struggle to predict rapid changes in hurricane intensity, Tang says. He notes how Hurricane Harvey, for example, strengthened suddenly to become a Category 4 storm right before it made landfall in Texas, offering emergency managers little time to issue warnings. “That’s the sort of thing that keeps forecasters up at night,” he says. In December, NASA launched a system of eight suitcase-sized microsatellites called the Cyclone Global Navigation Satellite System, or CYGNSS, into orbit. The satellites measure surface winds near the inner core of a hurricane, such as between the eyewall and the most intense bands of rain, at least a couple of times a day. Those regions have previously been invisible to satellites, measured only by hurricane-hunter airplanes darting through the storm.
“Improving forecasts of rapid intensification, like what occurred with Harvey on August 25, is exactly what CYGNSS is intended to do,” says Christopher Ruf, an atmospheric scientist at the University of Michigan in Ann Arbor and the lead scientist for CYGNSS. Results from CYGNSS measurements of both Harvey and Irma look very promising, he says. While the data are not being used to inform any forecasts this year, the measurements are now being calibrated and compared with hurricane-hunter flight data. The team will give the first detailed results from the hurricane season at the annual meeting of the American Geophysical Union in December. Meanwhile, NOAA has also been testing a new hurricane forecast model this year. The U.S. forecasting community is still somewhat reeling from its embarrassing showing during 2012’s Hurricane Sandy, which the National Weather Service had predicted would go out to sea while a European meteorological center predicted, correctly, that it would squarely hit New York City. In the wake of that event, Congress authorized $48 million to improve U.S. weather forecasting, and in 2014 NOAA held a competition to select a new weather prediction tool to improve its forecasts.
The clear winner was an algorithm developed by Shian-Jiann Lin and colleagues at NOAA’s Geophysical Fluid Dynamics Laboratory in Princeton, N.J. In May, NOAA announced that it would test the new model this hurricane season, running it alongside the more established operational models to see how it stacks up. Known as FV3 (short for Finite-Volume Cubed-Sphere Dynamical Core), the model divides the atmosphere into a 3-D grid of boxes and simulates climate conditions within the boxes, which may be as large as 4 kilometers across or as small as 1 kilometer across. Unlike existing models, FV3 can also re-create vertical air currents that move between boxes, such as the updrafts that are a key element of hurricanes as well as tornadoes and thunderstorms.
But FV3’s performance so far this year hasn’t been a slam dunk. FV3 did a far better job at simulating the intensity of Harvey than the other two leading models, but it lagged behind the European model in determining the hurricane’s path, Lin says. As for Irma, the European model outperformed the others on both counts. Still, Lin says he is confident that FV3 is on the right track in terms of its improvement. That’s good because pressure to work out the kinks may ramp up rapidly. Although NOAA originally stated that FV3 would be operational in 2019, “I hear some hints that it could be next year,” he says.
Lin adds that a good model alone isn’t enough to get a successful forecast; the data that go into a model are ultimately crucial to its success. “In our discipline, we call that ‘garbage in, garbage out,’” he says. With GOES-16 and CYGNSS nearly online, scientists are looking forward to even better hurricane models thanks to even better data.
WASHINGTON, D.C. — Helper cells in the brain just got tagged with a new job — forming traumatic memories.
When rats experience trauma, cells in the hippocampus — an area important for learning — produce signals for inflammation, helping to create a potent memory. But most of those signals aren’t coming from the nerve cells, researchers reported November 15 at the Society for Neuroscience meeting.
Instead, more than 90 percent of a key inflammation protein comes from astrocytes. This role in memory formation adds to the repertoire of these starburst-shaped cells, once believed to be responsible for only providing food and support to more important brain cells (SN Online: 8/4/15). The work could provide new insight into how the brain creates negative memories that contribute to post-traumatic stress disorder, said Meghan Jones, a neuroscientist at the University of North Carolina at Chapel Hill.
Jones and her colleagues gave rats a short series of foot shocks painful enough to “make you curse,” she said. A week after that harrowing experience, rats confronted with a milder shock remained jumpy. In some rats, Jones and her colleagues inhibited astrocyte activity during the original trauma, which prevented the cells from releasing the inflammation protein. Those rats kept their cool in the face of the milder shock.
These preliminary results show that neurons get a lot of help in creating painful memories. Studies like these are “changing how we think about the circuitry that’s involved in depression and post-traumatic stress disorder,” says neuroscientist Georgia Hodes of Virginia Tech in Blacksburg. “Everyone’s been focused on what neurons are doing. [This is] showing an important effect of cells we thought of as only being supportive.”
Sperm from stressed-out dads can carry that stress from one generation to another. “But one question that really hasn’t been addressed is, ‘How do dad’s experiences actually change his germ cell?’” Jennifer Chan, a neuroendocrinologist at the University of Pennsylvania, said November 13 in Washington, D.C., at the annual meeting of the Society for Neuroscience.
Now, from a study in mice, Chan and her colleagues have some answers, and even hints at ways to stop this stress inheritance. The researchers focused on the part of the male reproductive tract called the caput epididymis, a place where sperm cells mature. Getting rid of a stress-hormone sensor there called the glucocorticoid receptor stopped the transmission of stress, the researchers found. When faced with an alarming predator odor, offspring of chronically stressed mice dads overproduce the stress hormone corticosterone. But mice dads that lacked this receptor in the epididymis had offspring with normal hormonal responses.
Earlier work has shown that epididymis cells release small packets filled with RNA that can fuse to sperm and change their genetic payload. Experiments on cells in dishes revealed that chronic exposure to corticosterone changed the RNA in these vesicles. The results offer an explanation of how stress can change sperm: By activating the glucocorticoid receptor, stress tweaks the RNA in epididymis vesicles. Then, those vesicles deliver their altered contents to sperm, passing stress to the next generation.
Similar vesicles are present in human seminal fluid, even after ejaculation. Chan and colleagues are testing whether humans carry similar signs of stress in these RNA-loaded vesicles by studying college students’ semen samples. Exam schedules will be used as a stress indicator, she said.
Immune reactions against proteins commonly used as molecular scissors might make CRISPR/Cas9 gene editing ineffective in people, a new study suggests.
About 79 percent of 34 blood donors tested had antibodies against the Cas9 protein from Staphylococcus aureus bacteria, Stanford University researchers report January 5 at bioRxiv.org. About 65 percent of donors had antibodies against the Cas9 protein from Streptococcus pyogenes.
Nearly half of 13 blood donors also had T cells that seek and destroy cells that make S. aureus Cas9 protein. The researchers did not detect any T cells that attack S. pyogenes Cas9, but the methods used to detect the cells may not be sensitive enough to find them, says study coauthor Kenneth Weinberg. Cas9 is the DNA-cutting enzyme that enables researchers to make precise edits in genes. Antibodies and T cells against the protein could cause the immune system to attack cells carrying it, making gene therapy ineffective.
The immune reactions may be a technical glitch that researchers will need to work around, but probably aren’t a safety concern as long as cells are edited in lab dishes rather than in the body, says Weinberg, a stem cell biologist and immunologist.
“We think we need to address this now … as we move toward clinical trials,” he says, but “this is probably going to turn out to be more of a hiccup than a brick wall.”
LOS ANGELES — In the search for new superconductors, scientists are leaving no stone — and no meteorite — unturned. A team of physicists has now found the unusual materials, famous for their ability to conduct electricity without resistance, within two space rocks.
The discovery implies that small amounts of superconducting materials might be relatively common in meteorites, James Wampler of the University of California, San Diego, said March 6 at a meeting of the American Physical Society. While the superconducting materials found weren’t new to science, additional interplanetary interlopers might harbor new, more technologically appealing varieties of superconductors, the researchers suggest. Superconductors could potentially beget new, energy-saving technologies, but they have one fatal flaw: They require very cold temperatures to function, making them impractical for most uses. So scientists are on the hunt for new types of superconductors that work at room temperature (SN: 12/26/15, p. 25). If found, such a substance could lead to dramatic improvements in power transmission, computing and high-speed magnetically levitated trains, among other things.
Space rocks are a good avenue to explore in the search for new, exotic materials, says Wampler. “Meteorites are formed under these really unique, really extreme conditions,” such as high temperatures and pressures.
What makes the meteorite superconductors special, the researchers say, is that they occurred naturally, instead of being fabricated in a lab, as most known superconductors are. In fact, says physicist Ivan Schuller, also of University of California, San Diego, these are the highest temperature naturally occurring superconductors known — although they still have to be superchilled to about 5 kelvins (–268.15° C) to work. They are also the first known to have formed extraterrestrially.
“At this point, it’s a novelty,” says chemist Robert Cava of Princeton University. Although Cava is skeptical that scrutinizing meteorites will lead to new, useful superconductors, he says, it’s “kinda cool” that superconductors show up in meteorites. Wampler, Schuller and colleagues bombarded bits of powdered meteorite with microwaves and looked for changes in how those waves were absorbed as the temperature changed. The sensitive technique can pick out minute traces of superconducting material within a sample.
Analysis of powdered scrapings from more than a dozen meteorites showed that two meteorites contained superconducting material. However, the superconductors found within the meteorites were run-of-the-mill varieties, made from alloys of metals including indium, tin and lead, which are already known to superconduct.
“The idea is, try to look for something that is very unusual,” such as a room temperature superconductor, says Schuller, who led the research. So far, that hope hasn’t been realized — but that hasn’t deterred the search for something more exotic. For a previous study, Wampler, Schuller and colleagues scanned 65 tiny micrometeorites, but found no superconductors at all.
Since parts of space are colder than 5 kelvins, some meteorites may even contain materials that were once superconducting in their chilly natural habitat. That’s an interesting idea, Wampler says, although it’s too early to say whether that possibility might have any astronomical implications for how the objects behave out in space.
Bit by bit, Jupiter is revealing its deepest, darkest secrets.
The latest findings are in from the Juno spacecraft. And they unveil the roots of the planet’s storms, what lies beneath the opaque atmosphere and a striking geometric layout of cyclones parked around the gas giant’s north and south poles.
“We’re at the beginning of dissecting Jupiter,” says Juno mission leader Scott Bolton of the Southwest Research Institute in San Antonio. And the picture that’s emerging — still just a sketch — topples many preconceived notions. The results appear in four papers in the March 8 Nature. Juno has been orbiting Jupiter since July 4, 2016, on a mission to map the planet’s interior (SN: 6/25/16, p. 16). The probe loops around once every 53 days, traveling on an elongated orbit that takes the spacecraft from pole to pole and as close as about 4,000 kilometers above the cloud tops.
As it plows through Jupiter’s gravity field, Juno speeds up and slows down in response to shifting masses inside the planet. By measuring these minute accelerations and decelerations, scientists can calculate subtle variations in Jupiter’s gravity and deduce how its mass is distributed. That lets researchers build up a three-dimensional map of the planet’s internal structure. At the same time, Juno snaps pictures in visible and infrared light. While other probes have extensively photographed much of the planet, Juno is the first to get an intimate look at the north and south poles.
“The whole thing is really intriguing, especially when you compare [Jupiter] to other giant planets,” says Imke de Pater, a planetary scientist at the University of California, Berkeley. “They are all unique, it looks like.” Check out these four surprising new things we’ve learned that make Jupiter one of a kind:
Rings of cyclones Parked at each pole is a cyclone several thousand kilometers wide. That part isn’t surprising. But each of those cyclones is encircled by a polygonal arrangement of similarly sized storms — eight in the north and five in the south. The patterns have persisted throughout Juno’s visit.
“We don’t really understand why that would happen, and why they would collect up there in such a geometric fashion,” Bolton says. “That’s pretty amazing that nature is capable of something like that.”
More than skin deep Researchers have long debated whether the photogenic bands of clouds that wrap around Jupiter have deep roots or just skim the top of the atmosphere. Juno’s new look shows that the bands penetrate roughly 3,000 kilometers below the cloud tops. That’s 30 times as thick as the bulk of Earth’s atmosphere. While just a tiny fraction of Jupiter’s diameter, that’s deeper than previously thought, Bolton says.
Weighty weather Within those 3,000 kilometers lies what passes for an atmosphere on Jupiter. It’s the stage on which Jupiter’s turbulent weather plays out. The atmosphere alone is about three times as massive as our planet, or 1 percent of Jupiter’s entire mass, researchers estimate.
Stuck together Below the atmosphere, Jupiter is fluid. But unlike most fluids, the planet rotates as if it’s a solid mass. Like kids playing crack-the-whip, atoms of hydrogen and helium figuratively link arms and spin around the planet in unison, scientists report. Earlier results from Juno also indicate there’s no solid core lurking beneath this fluid (SN: 6/24/17, p. 14), so anyone dropped into the planet can expect a terribly long fall.
Many of these results are preliminary, and it’s unclear what it all means for how Jupiter operates. But what’s been learned so far, Bolton says, “is quite different than anybody anticipated.”