Migrating north may trigger immediate health declines among Mexicans (Demography)

Mexican-Square_351x351_45
Photo credit: Ticiana Jardim Marini, Woodrow Wilson School

By B. Rose Huber, Woodrow Wilson School of Public and International Affairs

Mexican immigrants who relocate to the United States often face barriers like poorly paying jobs, crowded housing and family separation. Such obstacles – including the migration process itself – may be detrimental to the health of Mexican immigrants, especially those who have recently moved.

A study led by Princeton University’s Woodrow Wilson School of Public and International Affairs finds that Mexican immigrants who relocate to the United States are more likely to experience declines in health within a short time period compared with other Mexicans.

While past studies have attempted to examine the consequences of immigration for a person’s health, few have had adequate data to compare recent Mexican immigrants, those who moved years ago and individuals who never left Mexico. Published in the journal Demography, the Princeton-led study is one of the first to examine self-reported health at two stages among these groups.

“Our study demonstrates that declines in health appear quickly after migrants’ arrival in the United States,” said Noreen Goldman, lead author and professor of demography and public affairs at the Wilson School and faculty associate at the Wilson School’s Office of Population Research (OPR). “Overall, we find that recent Mexican migrants are more likely to experience rapid changes in health, both good and bad, than the other groups. The deteriorations in health within a year or two of migration far outweigh the improvements.”

For the study, the researchers used data from the Mexican Family Life Survey, a longitudinal survey containing demographic and health information on nearly 20,000 Mexicans who were 20 years or older at the time of the first interview in 2002. Follow-up interviews took place in 2005-06 with individuals who stayed in Mexico as well as with those who moved to the United States between 2002 and 2005. Goldman and her collaborators based their analysis on a sample of 14,257 adults, excluding those who didn’t report health conditions at the follow-up interview.

In order to assess whether migrants from Mexico to the United States experienced changes in their health after they moved, the researchers used two health assessments: self-rated health (compared to someone of the same age and sex) at each of the two interviews and perceived change in health at the second interview. The latter measure was based on the following question: “Comparing your health to a year ago, would you say your health is much better, better, the same, worse or much worse?” Goldman and her collaborators narrowed the original five response categories to three: better, worse or the same. Changes in health for Mexicans who migrated between 2002 to 2005 were compared with those of migrants from earlier time periods and with people who remained in Mexico.

The researchers also took health measures at the first wave into account: obesity, anemia, hypertension – which were all determined by at-home visits by trained health workers – and hospitalization within the past year. They also controlled for socioeconomic factors – years of schooling and household spending. Additionally, they included data from 136 municipalities in Mexico (as past research has found that migration decisions can differ based on place of origin.)

Using statistical models, the researchers analyzed changes in health status. The two health measures revealed that recent migrants to the United States were more apt to experience both improvements and declines in their health than either earlier migrants or non-migrants. However, the overall net change was a substantial deterioration in the health of recent migrants relative to the other groups. The health of recent migrants was about 60 percent more likely to have worsened within a one- or two-year period than that of those who never left Mexico.

“The speed of the health decline for recent migrants suggests that the process of border crossing for both documented and undocumented immigrants combined with the physical and psychological costs of finding work, crowded housing, limited access to health care in the United States and isolation from family members can result in rapid deterioration of immigrants’ physical and mental wellbeing,” said Goldman.

“Immigrants are often assumed to be resilient and in good health because they have not yet adopted unhealthy American behaviors like poor diet and a sedentary lifestyle,” said co-author Anne Pebley from the California Center for Population Research at the University of California, Los Angeles. “But these results suggest that the image of the ‘healthy migrant’ is an illusion – at least for many recent immigrants.”

“These results demonstrate the high personal costs that many immigrants are willing to pay for a chance to improve their lives,” said Goldman. “From a humanitarian standpoint, the health declines underscore the need for public health, social service and immigration agencies to provide basic services for physical and psychological health to recent migrants.”

Given the limitations of the dataset, Goldman and her collaborators could not provide a more nuanced analysis regarding the causes of the changes in health status, but, with the availability of the third wave of data (collected between 2009-12), many of these questions can be later addressed.

In addition to Goldman and Pebley, study researchers include Chang Chung from OPR; Mathew Creighton from the University of Massachusetts; Graciela Teruel from the Universidad Iberoamericana; and Luis Rubalcava from the Centro de Análisis y Medición del Bienestar Social.

Support for this project from the Eunice Kennedy Shriver National Institute of Child Health and Human Development (R01HD051764, R24HD047879, R03HD040906, and R01HD047522) and from the Sector Research Fund for Social Development of the National Council for Science and Technology of Mexico.

Read the abstract.

Goldman N, Pebley AR, Creighton MJ, Teruel GM, Rubalcava LN, Chung C. 2014. The Consequences of Migration to the United States for Short-Term Changes in the Health of Mexican Immigrants. Demography. 2014 May 1 (Epub ahead of print).

Public interest in climate change unshaken by scandal, but unstirred by science (Environ. Res. Lett.)

Public interest in climate change
Princeton University and University of Oxford researchers found that negative media reports seem to have only a passing effect on public opinion, but that positive stories don’t appear to possess much staying power, either. Measured by how often people worldwide scour the Internet for information related to climate change, overall public interest in the topic has steadily waned since 2007. To gauge public interest, the researchers used Google Trends to document the Internet search-engine activity for “global warming” (blue line) and “climate change” (red line) from 2004 to 2013. They examined activity both globally (top) and in the United States (bottom). The numbers on the left indicate how often people looked up each term based on its percentage of the maximum search volume at any given point in time. (Image courtesy of William Anderegg)

By Morgan Kelly, Office of Communications

The good news for any passionate supporter of climate-change science is that negative media reports seem to have only a passing effect on public opinion, according to Princeton University and University of Oxford researchers. The bad news is that positive stories don’t appear to possess much staying power, either. This dynamic suggests that climate scientists should reexamine how to effectively and more regularly engage the public, the researchers write.

Measured by how often people worldwide scour the Internet for information related to climate change, overall public interest in the topic has steadily waned since 2007, according to a report in the journal Environmental Research Letters. Yet, the downturn in public interest does not seem tied to any particular negative publicity regarding climate-change science, which is what the researchers primarily wanted to gauge.

First author William Anderegg, a postdoctoral research associate in the Princeton Environmental Institute who studies communication and climate change, and Gregory Goldsmith, a postdoctoral researcher at Oxford’s Environmental Change Institute, specifically looked into the effect on public interest and opinion of two widely reported, almost simultaneous events.

The first involved the November 2009 hacking of emails from the Climate Research Unit at the University of East Anglia in the United Kingdom, which has been a preeminent source of data confirming human-driven climate change. Known as “climategate,” this event was initially trumpeted as proving that dissenting scientific views related to climate change have been maliciously quashed. Thorough investigations later declared that no misconduct took place.

The second event was the revelation in late 2009 that an error in the 2007 Fourth Assessment Report of the Intergovernmental Panel on Climate Change (IPCC) — an organization under the auspices of the United Nations that periodically evaluates the science and impacts of climate change — overestimated how quickly glaciers in the Himalayas would melt.

To first get a general sense of public interest in climate change, Anderegg and Goldsmith combed the freely available database Google Trends for “global warming,” “climate change” and all related terms that people around the world searched for between 2004 and 2013. The researchers documented search trends in English, Chinese and Spanish, which are the top three languages on the Internet. Google Trends receives more than 80 percent of the world’s Internet search-engine activity, and it is increasingly called upon for research in economics, political science and public health.

Internet searches related to climate change began to climb following the 2006 release of the documentary “An Inconvenient Truth” starring former vice president Al Gore, and continued its ascent with the release of the IPCC’s fourth report, the researchers found.

Anderegg and Goldsmith specifically viewed searches for “climategate” between Nov. 1 and Dec. 31, 2009. They found that the search trend had a six-day “half-life,” meaning that search frequency dropped by 50 percent every six days. After 22 days, the number of searches for climategate was a mere 10 percent of its peak. Information about climategate was most sought in the United States, Canada and Australia, while the cities with the most searchers were Toronto, London and Washington, D.C.

Searches for the phrase "global warming hoax" correlate with conservative political leanings
The researchers found that searchers for the phrase “global warming hoax” and related terms correlate in the United States with Republican or conservative political leanings. They compared the prevalence of searches for “global warming hoax” with the Cook Partisan Voting Index — which gauges how far toward Republicans or Democrats a congressional district leans — for 34 US states (above). They found that the more Republican/conservative the state (bottom measurement), the more frequently people in that state looked up related terms. The bottom graph shows how often a state votes Democrat (low numbers) versus Republican (high numbers). The numbers on the left indicate how often people looked up “global warming hoax” based on its percentage of the maximum search volume at any given point in time. (Image courtesy of William Anderegg)

The researchers tracked the popularity of the term “global warming hoax” to gauge the overall negative effect of climategate and the IPCC error on how the public perceives climate change. They found that searches for the term were actually higher the year before the events than during the year afterward.

“The search volume quickly returns to the same level as before the incident,” Goldsmith said. “This suggests no long-term change in the level of climate-change skepticism.

We found that intense media coverage of an event such as ‘climategate’ was followed by bursts of public interest, but these bursts were short-lived.”

All of this is to say that moments of great consternation for climate scientists seem to barely register in the public consciousness, Anderegg said. The study notes that independent polling data also indicate that these events had very little effect on American public opinion. “There’s a lot of handwringing among scientists, and a belief that these events permanently damaged public trust. What these results suggest is that that’s just not true,” Anderegg said.

While that’s good in a sense, Anderegg said, his and Goldsmith’s results also suggest that climate change as a whole does not top the list of gripping public topics. For instance, he said, climategate had the same Internet half-life as the public fallout from pro-golfer Tiger Woods’ extramarital affair, which happened around the same time (but received far more searches).

A public with little interest in climate change is unlikely to push for policies that actually address the problem, Anderegg said. He and Goldsmith suggest communicating in terms familiar to the public rather than to scientists. For example, their findings suggest that most people still identify with the term “global warming” instead of “climate change,” though the shift toward embracing the more scientific term is clear.

“If public interest in climate change is falling, it may be more difficult to muster public concern to address climate change,” Anderegg said. “This long-term trend of declining interest is worrying and something I hope we can address soon.”

One outcome of the research might be to shift scientists’ focus away from battling short-lived, so-called scandals, said Michael Oppenheimer, Princeton’s Albert G. Milbank Professor of Geosciences and International Affairs. The study should remind climate scientists that every little misstep or controversy does not make or break the public’s confidence in their work, he said. Oppenheimer, who was not involved in the study, is a long-time participant in the IPCC and an author of the Fifth Assessment Report being released this year in sections.

“This is an important study because it puts scientists’ concerns about climate skepticism in perspective,” Oppenheimer said. “While scientists should maintain the aspirational goal of their work being error-free, they should be less distracted by concerns that a few missteps will seriously influence attitudes in the general public, which by-and-large has never heard of these episodes.”

Read the article.

Anderegg, William R. L., Gregory R. Goldsmith. 2014. Public interest in climate change over the past decade and the effects of the ‘climategate’ media event. Environmental Research Letters 9 054005. doi:10.1088/1748-9326/9/5/054005 Article published online May 20, 2014.

Unlocking the potential of bacterial gene clusters to discover new antibiotics (Proc. Natl. Acad. Sci.)

High-throughput screening for the discovery of small molecules that activate silent bacterial gene clusters
High-throughput screening for the discovery of small molecules that activate silent bacterial gene clusters. Image courtesy of Mohammad Seyedsayamdost.

by Tien Nguyen, Department of Chemistry

Resistance to antibiotics has been steadily rising and poses a serious threat to the stronghold of existing treatments. Now, a method from Mohammad Seyedsayamdost, an assistant professor of chemistry at Princeton University, may open the door to the discovery of a host of potential drug candidates.

The vast majority of anti-infectives on the market today are bacterial natural products, made by biosynthetic gene clusters. Genome sequencing of bacteria has revealed that these active gene clusters are outnumbered approximately ten times by so-called silent gene clusters.

“Turning these clusters on would really expand our available chemical space to search for new antibiotic or otherwise therapeutically useful molecules,” Seyedsayamdost said.

In an article published last week in the journal Proceedings of the National Academy of Sciences, Seyedsayamdost reported a strategy to quickly screen whole libraries of compounds to find elicitors, small molecules that can turn on a specific gene cluster. He used a genetic reporter that fluoresces or generates a color when the gene cluster is activated to easily identify positive hits. Using this method, two silent gene clusters were successfully activated and a new metabolite was discovered.

Application of this work promises to uncover new bacterial natural products and provide insights into the regulatory networks that control silent gene clusters.

Read the abstract.

Seyedsayamdost, M. R. “High-throughput platform for the discovery of elicitors of silent bacterial gene clusters.” Proc. Natl. Acad. Sci. 2014, Early edition.

Why a bacterium got its curve — and why biologists should know (Nature Communications)

Art by Laura Ancona
Princeton University researchers found that the banana-like curve of the bacteria Caulobacter crescentus provides stability and helps them flourish as a group in the moving water they experience in nature. The findings suggest a new way of studying the evolution of bacteria that emphasizes using naturalistic settings. The illustration shows how C. crescentus divides asymmetrically into a “stalked” mother cell that anchors to a bacterium’s home surface, and an upper unattached portion that forms a new, juvenile cell known as a “swarmer.” Swarmer cells later morph into stalked cells and lay down roots nearby. They repeat the life cycle with their own swarmer cell and the bacterial colony grows. The Princeton researchers found that in moving water, curvature points the swarmer cell toward the surface to which it needs to attach. This ensures that the bacteria’s next generation does not stray too far from its progenitors. (Image by Laura Ancona)

By Morgan Kelly, Office of Communications

Drawing from his engineering background, Princeton University researcher Alexandre Persat had a notion as to why the bacteria Caulobacter crescentus are curved — a hunch that now could lead to a new way of studying the evolution of bacteria, according to research published in the journal Nature Communications.

Commonly used in labs to study cell division, C. crescentus naturally take on a banana-like curve, but they also can undergo a mutation in which they grow to be perfectly straight. The problem was that in a laboratory there was no apparent functional difference between the two shapes. So a question among biologists was, why would nature bother?

Then Persat, who is a postdoctoral researcher in the group of Associate Professor of Molecular Biology Zemer Gitai, considered that the bacteria dwell in large groups attached to surfaces in lakes, ponds and streams. That means that their curvature could be an adaptation that allows C. crescentus to better develop in the water currents the organisms experience in nature.

In the new paper, first author Persat, corresponding author Gitai and Howard Stone, Princeton’s Donald R. Dixon ’69 and Elizabeth W. Dixon Professor of Mechanical and Aerospace Engineering, report that curvature does more than just help C. crescentus hold their ground in moving fluid. The researchers monitored C. crescentus growth on surfaces in flow and found that the bacteria’s arched anatomy is crucial to flourishing as a group.

“It didn’t take a long time to figure out how flow brought out the advantages of curvature,” Persat said. “The obvious thing to me as someone with a fluid-dynamics background was that this shape had something to do with fluid flow.”

The findings emphasize the need to study bacteria in a naturalistic setting, said Gitai, whose group focuses on how bacterial shapes are genetically determined. While a petri dish generally suffices for this line of study, the functionality of bacterial genes and anatomy can be elusive in most lab settings, he said. For instance, he said, 80 percent of the genes in C. crescentus are seemingly disposable — but they might not be in nature.

“We now see there can be benefits to bacterial shapes that are only seen in a growth environment that is close to the bacteria’s natural environment,” Gitai said.

“For C. crescentus, the ecology was telling us there is an advantage to being curved, but nothing we previously did in the lab could detect what that was,” he said. “We need to not only think of the chemical environment of the bacteria — we also need to think of the physical environment. I think of this research as opening a whole new axis of studying bacteria.”

While most bacteria grow and divide as two identical “daughter” cells, C. crescentus divides asymmetrically. A “stalked” mother cell anchors to a bacterium’s home surface while the upper unattached portion forms a new, juvenile version of the stalked cell known as a “swarmer” cell. The swarmer cells later morph into stalked cells then eventually detach before laying down roots nearby. They repeat the life cycle with their own swarmer cell and the bacterial colony grows.

The Princeton researchers found that in moving water, curvature points the swarmer cell toward the surface to which it needs to attach. This ensures that the bacteria’s next generation does not stray too far from its progenitors, as well as from the nutrients that prompted cell division in the first place, Gitai said. On the other hand, the upper cells of straight bacteria — which are comparatively higher from the ground — are more likely to be carried far away as they are to stay near home.

But the advantage of curvature only goes so far. The researchers found that when the water current was too strong, both curved and straight bacteria were pressed flat against the surface, eliminating the curved cells’ colonization advantage.

These findings put some interesting boundaries on what is known about C. crescentus, starting with the upper limits of the current in which the organism can thrive, Gitai said. He and Persat also plan to pursue whether the bacteria are able to straighten out and cast offspring downstream when the home colony faces a decline in available nutrients.

At the same time, understanding why C. crescentus got its curve helps in figuring out the evolution of other bacteria, he said. Close relatives of the bacteria, for example, are not curved — could it have to do with the severity of their natural environment, such as the powerful turbulence of an ocean? Harmful bacteria such as Vibrio cholerae, strains of which cause cholera, are curved, though the reason is unclear. It’s possible this shape could be related to the organism’s environment in a way that might help treat those infected by it, Gitai said.

Whatever the reason for a specific bacteria’s shape, the Princeton research shows that exploring the influence of its natural habitat could be worthwhile, Gitai said.

“It was clear with C. crescentus that we needed to try something different,” Gitai said. “People didn’t really think of flow as a major driver of this bacteria’s evolution. That really is a new idea.”

Read the article..

Persat, Alexandre, Howard A. Stone, Zemer Gitai. 2014. The curved shape of Caulobacter crescentus enhances surface colonization in flow. Nature Communications. Article published online May 8, 2014. DOI: 10.1038/ncomms4824

The work was supported by the Gordon and Betty Moore Foundation (grant no. GBMF 2550.02), the National Science Foundation (grant no. CBET-1234500), and the National Institutes of Health Director’s New Investigator Innovator Award (grant no. 1DP2OD004389).

Too many chefs: Smaller groups exhibit more accurate decision-making (Proceedings of the Royal Society B)

Flock behavior
Smaller groups actually tend to make more accurate decisions, according to a new study from Princeton University Professor Iain Couzin and graduate student Albert Kao. (Photo credit: Gabriel Miller)

By Morgan Kelly, Office of Communications

The trope that the likelihood of an accurate group decision increases with the abundance of brains involved might not hold up when a collective faces a variety of factors — as often happens in life and nature. Instead, Princeton University researchers report that smaller groups actually tend to make more accurate decisions, while larger assemblies may become excessively focused on only certain pieces of information.

The findings present a significant caveat to what is known about collective intelligence, or the “wisdom of crowds,” wherein individual observations — even if imperfect — coalesce into a single, accurate group decision. A classic example of crowd wisdom is English statistician Sir Francis Galton’s 1907 observation of a contest in which villagers attempted to guess the weight of an ox. Although not one of the 787 estimates was correct, the average of the guessed weights was a mere one-pound short of the animal’s recorded heft. Along those lines, the consensus has been that group decisions are enhanced as more individuals have input.

But collective decision-making has rarely been tested under complex, “realistic” circumstances where information comes from multiple sources, the Princeton researchers report in the journal Proceedings of the Royal Society B. In these scenarios, crowd wisdom peaks early then becomes less accurate as more individuals become involved, explained senior author Iain Couzin, a professor of ecology and evolutionary biology.

“This is an extension of the wisdom-of-crowds theory that allows us to relax the assumption that being in big groups is always the best way to make a decision,” Couzin said.

“It’s a starting point that opens up the possibility of capturing collective decision-making in a more realistic environment,” he said. “When we do see small groups of animals or organisms making decisions they are not necessarily compromising accuracy. They might actually do worse if more individuals were involved. I think that’s the new insight.”

Couzin and first author Albert Kao, a graduate student of ecology and evolutionary biology in Couzin’s group, created a theoretical model in which a “group” had to decide between two potential food sources. The group’s decision accuracy was determined by how well individuals could use two types of information: One that was known to all members of the group — known as correlated information — and another that was perceived by only some individuals, or uncorrelated information. The researchers found that the communal ability to pool both pieces of information into a correct, or accurate, decision was highest in a band of five to 20. After that, the accurate decision increasingly eluded the expanding group.

At work, Kao said, was the dynamic between correlated and uncorrelated cues. With more individuals, that which is known by all members comes to dominate the decision-making process. The uncorrelated information gets drowned out, even if individuals within the group are still well aware of it.

In smaller groups, on the other hand, the lesser-known cues nonetheless earn as much consideration as the more common information. This is due to the more random nature of small groups, which is known as “noise” and typically seen as an unwelcome distraction. Couzin and Kao, however, found that noise is surprisingly advantageous in these smaller arrangements.

“It’s surprising that noise can enhance the collective decision,” Kao said. “The typical assumption is that the larger the group, the greater the collective intelligence.

“We found that if you increase group size, you see the wisdom-of-crowds benefit, but if the group gets too large there is an over-reliance on high-correlation information,” he said. “You would find yourself in a situation where the group uses that information to the point that it dominates the group’s decision-making.”

None of this is to suggest that large groups would benefit from axing members, Couzin said. The size threshold he and Kao found corresponds with the number of individuals making the decisions, not the size of the group overall. The researchers cite numerous studies — including many from Couzin’s lab — showing that decisions in animal groups such as schools of fish can often fall to a select few members. Thusly, these organisms can exhibit highly coordinated movements despite vast numbers of individuals. (Such hierarchies could help animals realize a dual benefit of efficient decision-making and defense via strength-in-numbers, Kao said.)

“What’s important is the number of individuals making the decision,” Couzin said. “Just looking at group size per se is not necessarily relevant. It depends on the number of individuals making the decision.”

Read the abstract.

Kao, Albert B., Iain D. Couzin. 2014. Decision accuracy in complex environments is often maximized by small group sizes. Proceedings of the Royal Society B. Article published online April 23, 2014. DOI: 10.1098/rspb.2013.3305

This work was supported by a National Science Foundation Graduate Research Fellowship, National Science Foundation Doctoral Dissertation Improvement (grant no. 1210029), the National Science Foundation (grant no. PHY-0848755), the Office of Naval Research Award (no. N00014-09-1-1074), the Human Frontier Science Project (grant no. RGP0065/2012), the Army Research Office (grant no. W911NG-11-1-0385), and an NSF EAGER grant (no. IOS-1251585).

Study resolves controversy over nitrogen’s ocean “exit strategies” (Science)

By Catherine Zandonella, Office of the Dean for Research

Seawater collection device
Princeton University graduate student Andrew Babbin (left) prepares a seawater collection device known as a rosette. The team used samples of seawater to determine how nitrogen is removed from the oceans. (Research photos courtesy of A. Babbin)

A decades-long debate over how nitrogen is removed from the ocean may now be settled by new findings from researchers at Princeton University and their collaborators at the University of Washington.

The debate centers on how nitrogen — one of the most important food sources for ocean life and a controller of atmospheric carbon dioxide — becomes converted to a form that can exit the ocean and return to the atmosphere where it is reused in the global nitrogen cycle.

Researchers have argued over which of two nitrogen-removal mechanisms, denitrification and anammox, is most important in the oceans. The question is not just a scientific curiosity, but has real world applications because one mechanism contributes more greenhouse gases to the atmosphere than the other.

Bess Ward
Bess Ward, Princeton’s William J. Sinclair Professor of Geosciences (Photo courtesy of Georgette Chalker)

“Nitrogen controls much of the productivity of the ocean,” said Andrew Babbin, first author of the study and a graduate student who works with Bess Ward, Princeton’s William J. Sinclair Professor of Geosciences. “Understanding nitrogen cycling is crucial to understanding the productivity of the oceans as well as the global climate,” he said.

In the new study, the researchers found that both of these nitrogen “exit strategies” are at work in the oceans, with denitrification mopping up about 70 percent of the nitrogen and anammox disposing of the rest.

The researchers also found that this 70-30 ratio could shift in response to changes in the quantity and quality of the nitrogen in need of removal. The study was published online this week in the journal Science.

The two other members of the research team were Richard Keil and Allan Devol, both professors at University of Washington’s School of Oceanography.

Research vessel
The researchers collected the samples in 2012 in the ocean off Baja California. Click on the image to read blog posts from a similar expedition that took place the following year.

Essential for the Earth’s life and climate, nitrogen is an element that cycles between soils and the atmosphere and between the atmosphere and the ocean. Bacteria near the surface help shuttle nitrogen into the ocean food chain by converting or “fixing” atmospheric nitrogen into forms that phytoplankton can use.

Without this fixed nitrogen, phytoplankton could not absorb carbon dioxide from the air, a feat which is helping to check today’s rising carbon dioxide levels in the atmosphere. When these tiny marine algae die or are consumed by predators, their biomass sinks to the ocean interior where it becomes food for other types of bacteria.

Test tubes
Researchers added specific amounts and types of nitrogen and organic compounds to test tubes containing seawater, and then noted whether denitrification or anammox occurred.

Until about 20 years ago, most scientists thought that denitrification, carried out by some of these bacteria, was the primary way that fixed nitrogen was recycled back to nitrogen gas. The second process, known as anaerobic ammonia oxidation, or anammox, was discovered by Dutch researchers studying how nitrogen is removed in sewage treatment plants.

Both processes occur in regions of the ocean that are naturally low in oxygen, or anoxic, due to local lack of water circulation and intense phytoplankton productivity overlying these regions. Within the world’s ocean, such regions occur only in the Arabian Sea, and off the coasts of Peru and Mexico.

In these anoxic environments, anaerobic bacteria feast on the decaying phytoplankton, and in the process cause the denitrification of nitrate into nitrogen gas, which cannot be used as a nutrient by most phytoplankton. During this process, ammonium is also produced, although marine geochemists had never been able to detect the ammonium that they knew must be there.

That riddle was solved in the early 2000s by the discovery of the anammox reaction in the marine environment, in which anaerobic bacteria feed on ammonium and convert it to nitrogen gas.

Glove bag
Graduate student Andrew Babbin filling incubation vials under an anoxic atmosphere in a transparent container called a “glove bag.” (Photo courtesy of Andrew Babbin)

But another riddle soon appeared: the anammox rates that Dutch and German teams of researchers measured in the oceans appeared to account for the entire nitrogen loss, leaving no role for denitrification.

Then in 2009, Ward’s team published a study in the journal Nature showing that denitrification was still a major actor in returning nitrogen to the air, at least in the Arabian Sea. The paper further fueled the controversy.

Back at Princeton, Ward suspected that both processes were necessary, with denitrification churning out the ammonium that anammox then converted to nitrogen gas.

To settle the issue, Ward and Babbin decided to look at exactly what was going on in anoxic ocean water when bacteria were given nitrogen and other nutrients to chew on.

They collected water samples from an anoxic region in the ocean south of Baja California and brought test tubes of the water into an on-ship laboratory. Working inside a sturdy, flexible “glove bag” to keep air from contaminating the low-oxygen water, Babbin added specific amounts and types of nitrogen and organic compounds to each test tube, and then noted whether denitrification or anammox occurred.

“We conducted a suite of experiments in which we added different types of organic matter, with variable ammonium content, to see if the ratio between denitrification and anammox would change,” said Babbin. “We found that not only did increased ammonia favor anammox as predicted, but that the precise proportions of nitrogen loss matched exactly as predicted based on the ammonium content.”

The explanation of why, in past experiments, some researchers found mostly denitrification while others found only anammox comes down to a sort-of “bloom and bust” cycle of phytoplankton life, explained Ward.

“If you have a big plankton bloom, then when those organisms die, a large amount of organic matter will sink and be degraded,” she said, “but we scientists are not always there to measure this. In other words, if you aren’t there on the day lunch is delivered, you won’t measure these processes.”

The researchers also linked the rates of nitrogen loss with the supply of organic material that drives the rates: more organic material equates to more nitrogen loss, so the quantity of the material matters too, Babbin said.

The two pathways have distinct metabolisms that turn out to be important in global climate change, he said.  “Denitrification produces carbon dioxide and both produces and consumes nitrous oxide, which is another major greenhouse gas and an ozone depletion agent,” he said. “Anammox, however, consumes carbon dioxide and has no known nitrous oxide byproduct. The balance between the two therefore has a significant impact on the production and consumption of greenhouse gases in the ocean.”

The research was funded by National Science Foundation grant OCE-1029951.

Read the abstract.

Andrew R. Babbin, Richard G. Keil, Allan H. Devol, and Bess B. Ward. Organic Matter Stoichiometry, Flux, and Oxygen Control Nitrogen Loss in the Ocean. Science. Published Online April 10 2014. DOI: 10.1126/science.1248364

 

A promising concept on the path to fusion energy (IEEE Transactions on Plasma Science)

by John Greenwald, Princeton Plasma Physics Laboratory

QUASAR stellerator design
QUASAR stellerator design (Source: PPPL)

Completion of a promising experimental facility at the U.S. Department of Energy’s Princeton Plasma Laboratory (PPPL) could advance the development of fusion as a clean and abundant source of energy for generating electricity, according to a PPPL paper published this month in the journal IEEE Transactions on Plasma Science.

The facility, called the Quasi-Axisymmetric Stellarator Research (QUASAR) experiment, represents the first of a new class of fusion reactors based on the innovative theory of quasi-axisymmetry, which makes it possible to design a magnetic bottle that combines the advantages of the stellarator with the more widely used tokamak design. Experiments in QUASAR would test this theory. Construction of QUASAR — originally known as the National Compact Stellarator Experiment — was begun in 2004 and halted in 2008 when costs exceeded projections after some 80 percent of the machine’s major components had been built or procured.

“This type of facility must have a place on the roadmap to fusion,” said physicist George “Hutch” Neilson, the head of the Advanced Projects Department at PPPL.

Both stellarators and tokamaks use magnetic fields to control the hot, charged plasma gas that fuels fusion reactions. While tokamaks put electric current into the plasma to complete the magnetic confinement and hold the gas together, stellarators don’t require such a current to keep the plasma bottled up. Stellarators rely instead on twisting — or 3D —magnetic fields to contain the plasma in a controlled “steady state.”

Stellarator plasmas thus run little risk of disrupting — or falling apart — as can happen in tokamaks if the internal current abruptly shuts off. Developing systems to suppress or mitigate such disruptions is a challenge that builders of tokamaks like ITER, the international fusion experiment under construction in France, must face.

Stellarators had been the main line of fusion development in the 1950s and early 1960s before taking a back seat to tokamaks, whose symmetrical, doughnut-shaped magnetic field geometry produced good plasma confinement and proved easier to create. But breakthroughs in computing and physics understanding have revitalized interest in the twisty, cruller-shaped stellarator design and made it the subject of major experiments in Japan and Germany.

PPPL developed the QUASAR facility with both stellarators and tokamaks in mind. Tokamaks produce magnetic fields and a plasma shape that are the same all the way around the axis of the machine — a feature known as “axisymmetry.” QUASAR is symmetrical too, but in a different way. While QUASAR was designed to produce a twisting and curving magnetic field, the strength of that field varies gently as in a tokamak — hence the name “quasi-symmetry” (QS) for the design.  This property of the field strength was to produce plasma confinement properties identical to those of tokamaks.

“If the predicted near-equivalence in the confinement physics can be validated experimentally,” Neilson said, “then the development of the QS line may be able to continue as essentially a ‘3D tokamak.’”

Such development would test whether a QUASAR-like design could be a candidate for a demonstration — or DEMO —fusion facility that would pave the way for construction of a commercial fusion reactor that would generate electricity for the power grid.

Read the paper.

George Neilson, David Gates, Philip Heitzenroeder, Joshua Breslau, Stewart Prager, Timothy Stevenson, Peter Titus, Michael Williams, and Michael Zarnstorff. Next Steps in Quasi-Axisymmetric Stellarator Research IEEE Transactions on Plasma Science, vol. 42, No. 3, March 2014.

The research was supported by the U.S. Department of Energy under contract DE-AC02 09CH11466. Princeton University manages PPPL, which is part of the national laboratory system funded by the U.S. Department of Energy through the Office of Science.

A more potent greenhouse gas than CO2, methane emissions will leap as Earth warms (Nature)

Freshwater wetlands can release methane, a potent greenhouse gas, as the planet warms. (Image source: RGBstock.com)
Freshwater wetlands can release methane, a potent greenhouse gas, as the planet warms. (Image source: RGBstock.com)

By Morgan Kelly, Office of Communications

While carbon dioxide is typically painted as the bad boy of greenhouse gases, methane is roughly 30 times more potent as a heat-trapping gas. New research in the journal Nature indicates that for each degree that the Earth’s temperature rises, the amount of methane entering the atmosphere from microorganisms dwelling in lake sediment and freshwater wetlands — the primary sources of the gas — will increase several times. As temperatures rise, the relative increase of methane emissions will outpace that of carbon dioxide from these sources, the researchers report.

The findings condense the complex and varied process by which methane — currently the third most prevalent greenhouse gas after carbon dioxide and water vapor — enters the atmosphere into a measurement scientists can use, explained co-author Cristian Gudasz, a visiting postdoctoral research associate in Princeton’s Department of Ecology and Evolutionary Biology. In freshwater systems, methane is produced as microorganisms digest organic matter, a process known as “methanogenesis.” This process hinges on a slew of temperature, chemical, physical and ecological factors that can bedevil scientists working to model how the Earth’s systems will contribute, and respond, to a hotter future.

The researchers’ findings suggest that methane emissions from freshwater systems will likely rise with the global temperature, Gudasz said. But to not know the extent of methane contribution from such a widely dispersed ecosystem that includes lakes, swamps, marshes and rice paddies leaves a glaring hole in climate projections.

“The freshwater systems we talk about in our paper are an important component to the climate system,” Gudasz said. “There is more and more evidence that they have a contribution to the methane emissions. Methane produced from natural or manmade freshwater systems will increase with temperature.”

To provide a simple and accurate way for climate modelers to account for methanogenesis, Gudasz and his co-authors analyzed nearly 1,600 measurements of temperature and methane emissions from 127 freshwater ecosystems across the globe.

New research in the journal Nature found that for each degree that the Earth's temperature rises, the amount of methane entering the atmosphere from microorganisms dwelling in freshwater wetlands — a primary source of the gas — will increase several times. The researchers analyzed nearly 1,600 measurements of temperature and methane emissions from 127 freshwater ecosystems across the globe (above), including lakes, swamps, marshes and rice paddies. The size of each point corresponds with the average rate of methane emissions in milligrams per square meter, per day, during the course of the study. The smallest points indicate less than one milligram per square meter, while the largest-sized point represents more than three milligrams. (Image courtesy of Cristian Gudasz)
New research in the journal Nature found that for each degree that the Earth’s temperature rises, the amount of methane entering the atmosphere from microorganisms dwelling in freshwater wetlands — a primary source of the gas — will increase several times. The researchers analyzed nearly 1,600 measurements of temperature and methane emissions from 127 freshwater ecosystems across the globe (above), including lakes, swamps, marshes and rice paddies. The size of each point corresponds with the average rate of methane emissions in milligrams per square meter, per day, during the course of the study. The smallest points indicate less than one milligram per square meter, while the largest-sized point represents more than three milligrams. (Image courtesy of Cristian Gudasz)

The researchers found that a common effect emerged from those studies: freshwater methane generation very much thrives on high temperatures. Methane emissions at 0 degrees Celsius would rise 57 times higher when the temperature reached 30 degrees Celsius, the researchers report. For those inclined to model it, the researchers’ results translated to a temperature dependence of 0.96 electron volts (eV), an indication of the temperature-sensitivity of the methane-emitting ecosystems.

“We all want to make predictions about greenhouse gas emissions and their impact on global warming,” Gudasz said. “Looking across these scales and constraining them as we have in this paper will allow us to make better predictions.”

Read the abstract.

Yvon-Durocher, Gabriel, Andrew P. Allen, David Bastviken, Ralf Conrad, Cristian Gudasz, Annick St-Pierre, Nguyen Thanh-Duc, Paul A. del Giorgio. 2014. Methane fluxes show consistent temperature dependence across microbial to ecosystem scales. Nature. Article published online before print: March 19, 2014. DOI: 10.1038/nature13164 and in the March 27, 2014 print edition.

It slices, it dices, and it protects the body from harm (Science)

By Catherine Zandonella, Office of the Dean for Research

RNase L enzyme structure
Researchers at Princeton have deciphered the 3D structure of RNase L, an enzyme that slices through RNA and is a first responder in the innate immune system. The structure contains two subunits, represented in red as two parts of a pair of scissors. Illustration by Sneha Rath, Inset courtesy of Science.

An essential weapon in the body’s fight against infection has come into sharper view. Researchers at Princeton University have discovered the 3D structure of an enzyme that cuts to ribbons the genetic material of viruses and helps defend against bacteria.

The discovery of the structure of this enzyme, a first-responder in the body’s “innate immune system,” could enable new strategies for fighting infectious agents and possibly prostate cancer and obesity. The work was published Feb. 27 in the journal Science.

Until now, the research community has lacked a structural model of the human form of this enzyme, known as RNase L, said Alexei Korennykh, an assistant professor of molecular biology and leader of the team that made the discovery.

“Now that we have the human RNase L structure, we can begin to understand the effects of carcinogenic mutations in the RNase L gene. For example, families with hereditary prostate cancers often carry genetic mutations in the region, or locus, encoding RNase L,” Korennykh said. The connection is so strong that the RNase L locus also goes by the name “hereditary prostate cancer 1.” The newly found structure reveals the positions of these mutations and explains why some of these mutations could be detrimental, perhaps leading to cancer, Korennykh said. RNase L is also essential for insulin function and has been implicated in obesity.

The Princeton team’s work has also led to new insights on the enzyme’s function.

The enzyme is an important player in the innate immune system, a rapid and broad response to invaders that includes the production of a molecule called interferon. Interferon relays distress signals from infected cells to neighboring healthy cells, thereby activating RNase L to turn on its ability to slice through RNA, a type of genetic material that is similar to DNA. The result is new cells armed for destruction of the foreign RNA.

The 3D structure uncovered by Korennykh and his team consists of two nearly identical subunits called protomers. The researchers found that one protomer finds and attaches to the RNA, while the other protomer snips it.

The initial protomer latches onto one of the four “letters” that make up the RNA code, in particular, the “U,” which stands for a component of RNA called uridine. The other protomer “counts” RNA letters starting from the U, skips exactly one letter, then cuts the RNA.

Although the enzyme can slice any RNA, even that of the body’s own cells, it only does so when activated by interferon.

“We were surprised to find that the two protomers were identical but have different roles, one binding and one slicing,” Korennykh said. “Enzymes usually have distinct sites that bind the substrate and catalyze reactions. In the case of RNase L, it appears that the same exact protein surface can do both binding and catalysis. One RNase L subunit randomly adopts a binding role, whereas the other identical subunit has no other choice but to do catalysis.”

To discover the enzyme’s structure, the researchers first created a crystal of the RNase L enzyme. The main challenge was finding the right combination of chemical treatments that would force the enzyme to crystallize without destroying it.

Korennykh groupAfter much trial and error and with the help of an automated system, postdoctoral research associate Jesse Donovan and graduate student Yuchen Han succeeded in making the crystals.

Next, the crystals were bombarded with powerful X-rays, which diffract when they hit the atoms in the crystal and form patterns indicative of the crystal’s structure. The diffraction patterns revealed how the atoms of RNase L are arranged in 3D space.

At the same time Sneha Rath, a graduate student in Korennykh’s laboratory, worked on understanding the RNA cleavage mechanism of RNase L using synthetic RNA fragments. Rath’s results matched the structural findings of Han and Donovan, and the two pieces of data ultimately revealed how RNase L cleaves its RNA targets.

Han, Donovan and Rath contributed equally to the paper and are listed as co-first authors.

Finally, senior research specialist Gena Whitney and graduate student Alisha Chitrakar conducted additional studies of RNase L in human cells, confirming the 3D structure.

Now that the human structure has been solved, researchers can explore ways to either enhance or dampen RNase L activity for medical and therapeutic uses, Korennykh said.

“This work illustrates the wonderful usefulness of doing both crystallography and careful kinetic and enzymatic studies at the same time,” said Peter Walter, professor of biochemistry and biophysics at the University of California-San Francisco School of Medicine. “Crystallography gives a static picture which becomes vastly enhanced by studies of the kinetics.”

Support for the work was provided by Princeton University.

Read the abstract.

Han, Yuchen, Jesse Donovan, Sneha Rath, Gena Whitney, Alisha Chitrakar, and Alexei Korennykh. Structure of Human RNase L Reveals the Basis for Regulated RNA Decay in the IFN Response Science 1249845. Published online 27 February 2014 [DOI:10.1126/science.1249845]

Now in 3D: Video of virus-sized particle trying to enter cell (Nature Nanotechnology)

Video of virus trying to enter cell
3D movie (below) of virus-like nanoparticle trying to gain entry to a cell

By Catherine Zandonella, Office of the Dean for Research

Tiny and swift, viruses are hard to capture on video. Now researchers at Princeton University have achieved an unprecedented look at a virus-like particle as it tries to break into and infect a cell. The technique they developed could help scientists learn more about how to deliver drugs via nanoparticles — which are about the same size as viruses — as well as how to prevent viral infection from occurring.

The video reveals a virus-like particle zipping around in a rapid, erratic manner until it encounters a cell, bounces and skids along the surface, and either lifts off again or, in much less time than it takes to blink an eye, slips into the cell’s interior. The work was published in Nature Nanotechnology.

Video caption: ‘Kiss and run’ on the cell surface. This 3D movie shows actual footage of a virus-like particle (red dot) approaching a cell (green with reddish brown nucleus), as captured by Princeton University researchers Kevin Welcher and Haw Yang. The color of the particle represents its speed, with red indicating rapid movement and blue indicating slower movement. The virus-like particle lands on the surface of the cell, appears to try to enter it, then takes off again. Source: Nature Nanotechnology.

“The challenge in imaging these events is that viruses and nanoparticles are small and fast, while cells are relatively large and immobile,” said Kevin Welsher, a postdoctoral researcher in Princeton’s Department of Chemistry and first author on the study. “That has made it very hard to capture these interactions.”

The problem can be compared to shooting video of a hummingbird as it roams around a vast garden, said Haw Yang, associate professor of chemistry and Welsher’s adviser. Focus the camera on the fast-moving hummingbird, and the background will be blurred. Focus on the background, and the bird will be blurred.

The researchers solved the problem by using two cameras, one that locked onto the virus-like nanoparticle and followed it faithfully, and another that filmed the cell and surrounding environment.

Putting the two images together yielded a level of detail about the movement of nano-sized particles that has never before been achieved, Yang said. Prior to this work, he said, the only way to see small objects at a similar resolution was to use a technique called electron microscopy, which requires killing the cell.

“What Kevin has done that is really different is that he can capture a three-dimensional view of a virus-sized particle attacking a living cell, whereas electron microscopy is in two-dimensions and on dead cells,” Yang said. “This gives us a completely new level of understanding.”

In addition to simply viewing the particle’s antics, the researchers can use the technique to map the contours of the cell surface, which is bumpy with proteins that push up from beneath the surface. By following the particle’s movement along the surface of the cell, the researchers were able to map the protrusions, just as a blind person might use his or her fingers to construct an image of a person’s face.

“Following the motion of the particle allowed us to trace very fine structures with a precision of about 10 nanometers, which typically is only available with an electron microscope,” Welsher said. (A nanometer is one billionth of a meter and roughly 1000 times smaller than the width of a human hair.) He added that measuring changes in the speed of the particle allowed the researchers to infer the viscosity of the extracellular environment just above the cell surface.

The technology has potential benefits for both drug discovery and basic scientific discovery, Yang said.  “We believe this will impact the study of how nanoparticles can deliver medicines to cells, potentially leading to some new lines of defense in antiviral therapies,” he said. “For basic research, there are a number of questions that can now be explored, such as how a cell surface receptor interacts with a viral particle or with a drug.”

Welsher added that such basic research could lead to new strategies for keeping viruses from entering cells in the first place.

“If we understand what is happening to the virus before it gets to your cells,” said Welsher, “then we can think about ways to prevent infection altogether. It is like deflecting missiles before they get there rather than trying to control the damage once you’ve been hit.”

To create the virus-like particle, the researchers coated a miniscule polystyrene ball with quantum dots, which are semiconductor bits that emit light and allow the camera to find the particle. Next, the particle was studded with protein segments known as Tat peptides, derived from the HIV-1 virus, which help the particle find the cell. The width of the final particle was about 100 nanometers.

The researchers then let loose the particles into a dish containing skin cells known as fibroblasts. One camera followed the particle while a second imaging system took pictures of the cell using a technique called laser scanning microscopy, which involves taking multiple images, each in a slightly different focal plane, and combining them to make a three-dimensional picture.

The research was supported by the US Department of Energy (DE-SC0006838) and by the Eric and Wendy Schmidt Transformative Technology Fund at Princeton University.

Read the abstract.

Kevin Welsher and Haw Yang. 2014. Multi-resolution 3D visualization of the early stages of cellular uptake of peptide-coated nanoparticles. Nature nanotechnology. Published online: 23 February 2014 | DOI: 10.1038/NNANO.2014.12