A new study finds that climate has been the dominant controller of wildfire activity in the Sierra Nevada region of the past 1,400 years, suggesting that future climate change is poised to make fires worse.
PROVIDENCE, R.I. [Brown University] – In the wake of recent wildfires that have ravaged northern and central California, a new study finds that the severity of fire activity in the Sierra Nevada region has been sensitive to changes in climate over the past 1,400 years. The findings, published in Environmental Research Letters, suggest that future climate change is likely to drive increased fire activity in the Sierras.
“Our data show that climate has been the main driver of fire on a regional scale,” said Richard Vachula, a Ph.D. student in Brown University’s Department of Earth, Environmental and Planetary Sciences and the study’s lead author. “We find that warm and dry conditions promote fire, which in light of climate model predictions suggests that future fires may be more extensive than we have observed in the last century.”
Vachula’s co-authors on the study were James Russell and Yongsong Huang, both professors at Brown and research fellows in the Institute at Brown for Environment and Society.
The findings are based upon sediment cores taken from the bottom of Swamp Lake, located toward the northern edge of Yosemite National Park. As fires have burned in the region over the years, charcoal from smoke plumes falls into the lake, eventually sinking to the bottom and embedding itself in sediment. By tracking changes in charcoal content with depth, Vachula and his colleagues created a sequential record of how much area had burned around the lake over the past 1,400 years up to 2007, when the cores were taken.
The sizes of the charcoal particles enabled the team to track both local and regional fire activity. Smaller particles, which can travel farther in the air, record regional fire in a roughly 90-mile radius of the lake. Larger particles, which don’t carry as far, represent local activity within about 15 miles.
By synchronizing the charcoal data with existing climate records measuring temperature and rainfall, the team could measure correlations between climate and fire over time. The researchers also included data aimed at measuring the effects of direct human forest management, either by indigenous populations or the U.S. government.
The study found that the regional trends track closely with climate trends, with warmer and dryer periods correlating to periods of increased fire activity. Fire activity peaked during what’s known as the Medieval Climate Anomaly, a period of unusual heat and drought lasting from about 950 to 1,250 C.E.
The effect of human activity was largely confined to the local scale for most of the study period, the research found. The team used archeological records to track the size of Miwok Native American populations over time. The Miwok, whose dietary staple was acorns, used fire to control underbrush in order to promote the growth of acorn-producing oaks. The study showed that when the Miwok population was large, burning increased at the local level even when climate was adverse to burning. At the regional level, however, the influence of climate remained dominant.
The only time human activity seemed to supplant climate as a regional influence on fire was from 1900 to 1970, when the National Parks Service systematically suppressed fire throughout Yosemite. That intervention seemed to reduce fire even when climate conditions favored increased activity.
Taken together, the findings underscore the importance of climate in driving fire activity in the Sierras, the researchers say.
“Our results are significant because they clarify the relative roles of human management and climate as controls of fire regimes in California, and they highlight the need to prepare for future fire activity,” Vachula said.