The IPCC Report: Key Findings and Radical Implications

The IPCC Report: Key Findings and Radical Implications

This article originally appeared in Climate & Capitalism.

Editor’s note: DGR has always argued that civilizations are inherently destructive and environmental destruction and degradation has been ongoing for millenia. Climate change is only another concequence of this inherently destructive way of life. This is why technical solutions will never work. What we need to do to save the planet is 1. immediately stop destroying it, and 2. restore what we already have destroyed. This logic is easy to understand if your loyalty lies with the planet and all life on it, but it seems very hard to understand if your loyalty lies with this destructive and addictive way of life.


By Brian Tokar

Beyond the headlines: what climate science now shows about Earth’s future. Can we act in time?

The UN-sponsored Intergovernmental Panel on Climate Change (IPCC) recently released its latest comprehensive report on the state of the earth’s climate. The much-anticipated report dominated the headlines for a few days in early August, then quickly disappeared amidst the latest news from Afghanistan, the fourth wave of Covid-19 infections in the US, and all the latest political rumblings. The report is vast and comprehensive in its scope, and is worthy of more focused attention outside of specialist scientific circles than it has received thus far.

The report affirms much of what we already knew about the state of the global climate, but does so with considerably more clarity and precision than earlier reports. It removes several elements of uncertainty from the climate picture, including some that have wrongly served to reassure powerful interests and the wider public that things may not be as bad as we thought. The IPCC’s latest conclusions reinforce and significantly strengthen all the most urgent warnings that have emerged from the past 30 to 40 years of climate science. It deserves to be understood much more fully than most media outlets have let on, both for what it says, and also what it doesn’t say about the future of the climate and its prospects for the integrity of all life on earth.

Click image to download report. (PDF, 248MB)

First some background. Since 1990, the IPCC has released a series of comprehensive assessments of the state of the earth’s climate, typically every 5–6 years. The reports have hundreds of authors, run for many hundreds of pages (this one has over 3000), and represent the international scientific consensus that has emerged from the period since the prior report. Instead of releasing a comprehensive report in 2019, as originally scheduled, the IPCC followed a mandate from the UN to issue three special reports: on the implications of warming above 1.5 degrees (all temperatures here are in Celsius except where otherwise noted), and on the particular implications of climate change for the earth’s lands and oceans. Thus the sixth comprehensive Assessment Report (dubbed AR6) is being released during 2021–22 instead of two years prior.

Also the report released last week only presents the work of the first IPCC working group (WGI), focused on the physical science of climate change. The other two reports, on climate impacts (including implications for health, agriculture, forests, biodiversity, etc.) and on climate mitigation — including proposed policy measures — are scheduled for release next February and March, respectively. While the basic science report typically receives far more press coverage, the second report on climate impacts and vulnerabilities is often the most revealing, describing in detail how both ecosystems and human communities will experience the impacts of climate changes.

In many respects, the new document represents a qualitative improvement over the previous Assessment Reports, both in terms of the precision and reliability of the data and also the clarity of its presentation. There are countless detailed charts and infographics, each illuminating the latest findings on a particular aspect of current climate science in impressive detail. There is also a new Interactive Atlas (freely available at interactive-atlas.ipcc.ch), which allows any viewer to produce their own maps and charts of various climate phenomena, based on a vast array of data sources and climate models.

If there is a key take-home message, it is that climate science has vastly improved over the past decade in terms of its precision and the degree of confidence in its predictions. Many uncertainties that underlay past reports appear to have been successfully addressed, for example how a once-limited understanding of the behavior and dynamics of clouds were a major source of uncertainty in global climate models. Not only have the mathematical models improved, but we now have more than thirty years of detailed measurements of every aspect of the global climate that enable scientists to test the accuracy of their models, and also to substitute direct observations for several aspects that once relied heavily upon modeling studies. So we have access to better models, and are also less fully reliant upon them.

Second, scientists’ understanding of historic and prehistoric climate trends have also vastly improved. While the IPCC’s third report in 2001 made headlines for featuring the now-famous “hockey stick” graph, showing how average temperatures had been relatively stable for a thousand years before starting to spike rapidly in the past few decades, the current report highlights the relative stability of the climate system over many thousands of years. Decades of detailed studies of the carbon contents of polar ice cores, lake and ocean sediments and other geologically stable features have raised scientists’ confidence in the stark contrast between current climate extremes and a couple of million years of relative climate stability.

The long-term cycle of ice ages, for example, reflects shifts of about 50 to 100 parts per million (ppm) in atmospheric carbon dioxide concentrations, compared to a current concentration (approximately 410 ppm) that is well over 150 ppm higher than the million-year average. We need to look back to the last interglacial era (125,000 years ago) to find an extended period of high average temperatures comparable to what we are experiencing now, and current carbon dioxide concentrations in the atmosphere are believed to be higher than any time in at least two million years.

With these overarching issues in mind, it is time to summarize some of the report’s most distinctive findings and then reflect upon their implications.

First, the question of “climate sensitivity” has been one of the more contentious ones in climate science. It is a measure of how much warming would result from a doubling of atmospheric CO2 from preindustrial levels, i.e. from 280 ppm to 560 ppm. Early estimates were all over the map, giving policymakers the wiggle room to suggest it is reasonable to reduce emissions more slowly or wait for newer technologies — from better batteries to carbon capture and even nuclear fusion — to come along. This report greatly narrows the scope of that debate, with a “best estimate” that doubling CO2 will produce approximately 3 degrees of warming — far too high to avoid extremely dire consequences for all of life on earth.

Climate sensitivity is very likely (more than 90% confidence) between 2.0–4.5 degrees and likely (2/3 confidence) between 2.5 and 4 degrees. Of the five main future scenarios explored in the report, only those where global greenhouse gas emissions reach their peak before 2050 will avoid that disastrous milestone. If emissions continue increasing at rates comparable to the past few decades, we’ll reach doubled CO2 by 2100; if emissions accelerate, it could happen in just a few decades, vastly compounding the climate disruptions the world is already experiencing.

A second key question is, how fast do temperatures rise with increasing emissions? Is it a direct, linear relationship, or might temperature rises begin to level off any time in the foreseeable future? The report demonstrates that the effect remains linear, at least up to the level of 2 degrees warming, and quantifies the effect with high confidence. Of course there are important deviations from this number (1.65 degrees per thousand gigatons of carbon): the poles heat up substantially more quickly than other regions, the air over continental land masses heats up faster than over the oceans, and temperatures are warming almost twice as fast during cold seasons than warm seasons, accelerating the loss of arctic ice and other problems.

Of course more extreme events remain far less predictable, except that their frequency will continue to increase with rising temperatures. For example the triple digit (Fahrenheit) temperatures that swept the Pacific Northwest of the US and southwestern Canada this summer have been described as a once in 50,000 years event in “normal” times and no one excludes the possibility that they will happen again in the near future. So-called “compound” events, for example the combination of high temperatures and dry, windy conditions that favor the spread of wildfires, are the least predictable events of all.

The central conclusion from the overall linear increase in temperatures relative to emissions is that nothing short of a complete cessation of CO2 and other greenhouse gas emissions will significantly stabilize the climate, and there is also a time delay of at least several decades after emissions cease before the climate can begin to stabilize.

Third, estimates of likely sea level rise, in both the near- and longer-terms, are far more reliable than they were a few years ago. Global sea levels rose an average of 20 centimeters during the 20th century, and will continue to rise throughout this century under all possible climate scenarios — about a foot higher than today if emissions begin to fall rapidly, nearly 2 feet if emissions continue rising at present rates, and 2.5 feet if emissions rise faster. These, of course, are the most cautious scientific estimates. By 2150 the estimated range is 2–4.5 feet, and more extreme scenarios where sea levels rise from 6 to 15 feet “cannot be ruled out due to deep uncertainty in ice sheet processes.”

With glacial melting expected to continue for decades or centuries under all scenarios, sea levels will “remain elevated for thousands of years,” potentially reaching a height of between 8 and 60 feet above present levels. The last time global temperatures were comparable to today’s for several centuries (125,000 years ago), sea levels were probably 15 to 30 feet higher than they are today. When they were last 2.5 to 4 degrees higher than preindustrial temperatures — roughly 3 million years ago — sea levels may have been up to 60 feet higher than today. Again these are all cautious estimates, based on the available data and subject to stringent statistical validation. For residents of vulnerable coastal regions around the world, and especially Pacific Island dwellers who are already forced to abandon their drinking water wells due to high infiltrations of sea water, it is far from just a theoretical problem.

Also, for the first time, the new report contains detailed projections for the unfolding of various climate-related phenomena in every region of the world. There is an entire chapter devoted to regionally-specific effects, and much attention to the ways in which climate disruptions play out differently in different locations. “Current climate in all regions is already distinct from the climate of the early or mid-20th century,” the report states, and many regional differences are expected to become more pronounced over time. While every place on earth is getting hotter, there are charts showing how different regions will become consistently wetter or dryer, or various combinations of both, with many regions, including eastern North America, anticipated to experience increasingly extreme precipitation events.

There are also more specific discussions of potential changes in monsoon patterns, as well as particular impacts on biodiversity hotspots, cities, deserts, tropical forests, and other places with distinctive characteristics in common. Various drought-related phenomena are addressed in more specific terms, with separate projections for meteorological drought (lack of rainfall), hydrological drought (declining water tables) and agricultural/ecological drought (loss of soil moisture). It can be expected that all these impacts will be discussed in greater detail in the upcoming report on climate impacts that is due in February.

There are numerous other important observations, many of which directly counter past attempts to minimize the consequences of future climate impacts. For those who want to see the world focus more fully on emissions unrelated to fossil fuel use, the report points out that between 64 and 86 percent of carbon emissions are directly related to fossil fuel combustion, with estimates approaching 100 percent lying well within the statistical margin of error. Thus there is no way to begin to reverse climate disruptions without an end to burning fossil fuels. There are also more detailed projections of the impacts of shorter-lived climate forcers, such as methane (highly potent, but short-lived compared to CO2), sulfur dioxide (which counteracts climate warming) and black carbon (now seen as a substantially less significant factor than before).

To those who assume the vast majority of emissions will continue to be absorbed by the world’s land masses and oceans, buffering the effects on the future atmosphere, the report explains how with rising emissions, a steadily higher proportion of the CO2 remains in the atmosphere, rising from only 30 to 35 percent under low emissions scenarios, up to 56 percent with emissions continuing to increase at present rates and doubling to 62 percent if emissions begin to rise more rapidly. So we will likely see a declining capacity for the land and oceans to absorb a large share of excess carbon dioxide.

The report is also more skeptical than in the past toward geoengineering schemes based on various proposed technological interventions to absorb more solar radiation. The report anticipates a high likelihood of “substantial residual or overcompensating climate change at the regional scales and seasonal time scales” resulting from any interventions designed to shield us from climate warming without reducing emissions, as well as the certainty that ocean acidification and other non-climate consequences of excess carbon dioxide would inevitably continue. There will likely be substantially more discussion of these scenarios in the third report of this IPCC cycle, which is due in March.

In advance of the upcoming international climate conference in Glasgow, Scotland this November, several countries have pledged to increase their voluntary climate commitments under the 2015 Paris Agreement, with some countries now aiming to achieve a peak in climate-altering emissions by mid-century. However this only approaches the middle range of the IPCC’s latest projections. The scenario based on a 2050 emissions peak is right in the middle of the report’s range of predictions, and shows the world surpassing the important threshold of 1.5 degrees of average warming in the early 2030s, exceeding 2 degrees by mid-century, and reaching an average temperature increase between 2.1 and 3.5 degrees (approximately 4–6 degrees Fahrenheit) between 2080 and 2100, nearly two and a half times the current global average temperature rise of 1.1 degrees since preindustrial times.

We will learn much more about the impacts of this scenario in the upcoming February report, but the dire consequences of future warming have been described in numerous published reports in recent years, including an especially disturbing very recent paper reporting signs that the Atlantic circulation (AMOC), which is the main source of warm air for all of northern Europe, is already showing signs of collapse. If carbon emissions continue to increase at current rates, we are looking at a best estimate of a 3.6 degree rise before the end of this century, with a likely range reaching well above 4 degrees — often viewed as a rough threshold for a complete collapse of the climate system.

There are two lower-emissions scenarios in the report, the lowest of which keeps the temperature rise by the century’s end under 1.5 degrees (after exceeding it briefly), but a quick analysis from MIT’s Technology Review points out that this scenario relies mainly on highly speculative “negative emissions” technologies, especially carbon capture and storage, and a shift toward the massive-scale use of biomass (i.e. crops and trees) for energy. We know that a more widespread use of “energy crops” would consume vast areas of the earth’s landmass, and that the regrowing of trees that are cut down to burn for energy would take many decades to absorb the initial carbon release– a scenario the earth clearly cannot afford.

The lower-emissions scenarios also accept the prevailing rhetoric of “net-zero,” assuming that more widespread carbon-sequestering methods like protecting forests can serve to compensate for still-rising emissions. We know that many if not most carbon offset schemes to date have been an absolute failure, with Indigenous peoples often driven from their traditional lands in the name of “forest protection,” only to see rates of commercial logging increase rapidly in immediately surrounding areas.

It is increasingly doubtful that genuine long-term climate solutions can be found without a thorough transformation of social and economic systems. It is true that the cost of renewable energy has fallen dramatically in the past decade, which is a good thing, and that leading auto manufacturers are aiming to switch to electric vehicle production over the coming decade. But commercial investments in renewable energy have leveled off over the same time period, especially in the richer countries, and continue to favor only the largest-scale projects that begin to meet capitalist standards of profitability. Fossil fuel production has, of course, led to exaggerated standards of profitability in the energy sector over more than 150 years, and most renewable projects fall far short.

We will likely see more solar and wind power, a faster tightening of fuel efficiency standards for the auto industry and subsidies for electric charging stations in the US, but nothing like the massive reinvestment in community-scaled renewables and public transportation that is needed. Not even the landmark Biden-Sanders budget reconciliation plan that is under consideration in in the US Congress, with all its necessary and helpful climate measures, addresses the full magnitude of changes that are needed to halt emissions by midcentury. While some obstructionists in Congress appear to be stepping back from the overt climate denial that has increasingly driven Republican politics in recent years, they have not backed away from claims that it is economically unacceptable to end climate-altering pollution.

Internationally, the current debate over reducing carbon pollution (so called “climate mitigation”) also falls far short of addressing the full magnitude of the problem, and generally evades the question of who is mainly responsible. While the US and other wealthy countries have produced an overwhelming share of historic carbon pollution since the dawn of the industrial era, there is an added dimension to the problem that is most often overlooked, and which I reviewed in some detail in my Introduction to a recent book (co-edited with Tamar Gilbertson), Climate Justice and Community Renewal (Routledge 2020). A 2015 study from Thomas Piketty’s research group in Paris revealed that inequalities within countries have risen to account for half of the global distribution of greenhouse gas emissions, and several other studies confirm this.

Researchers at Oxfam have been studying this issue for some years, and their most recent report concluded that the wealthiest ten percent of the global population are responsible for 49 percent of individual emissions. The richest one percent emits 175 times more carbon per person on average than the poorest ten percent. Another pair of independent research groups have released periodic Carbon Majors Reports and interactive graphics profiling around a hundred global companies that are specifically responsible for almost two-thirds of all greenhouse gases since the mid-19th century, including just fifty companies — both private and state-owned ones — that are responsible for half of all today’s industrial emissions (See climateaccountability.org). So while the world’s most vulnerable peoples are disproportionately impacted by droughts, floods, violent storms and rising sea levels, the responsibility falls squarely upon the world’s wealthiest.

When the current IPCC report was first released, the UN Secretary General described it as a “code red for humanity,” and called for decisive action. Greta Thunberg described it as a “wake-up call,” and urged listeners to hold the people in power accountable. Whether that can happen quickly enough to stave off some of the worst consequences will be a function of the strength of our social movements, and also our willingness to address the full scope of social transformations that are now essential for humanity and all of life on earth to continue to thrive.


Brian Tokar is the co-editor (with Tamra Gilbertson) of Climate Justice and Community Renewal: Resistance and Grassroots Solutions. He is a lecturer in Environmental Studies at the University of Vermont and a long-term faculty and board member of the Vermont-based Institute for Social Ecology.

World On Track For “Worst Case” IPCC Climate Change Scenario

World On Track For “Worst Case” IPCC Climate Change Scenario

This article, originally posted by the Woods Hole Research Centre on August 3rd 2020, states that the “Worst case” for CO2 emissions scenario is actually the best match for assessing the climate risk, impact by 2050. 


The RCP 8.5 CO2 emissions pathway, long considered a “worst case scenario” by the international science community, is the most appropriate for conducting assessments of climate change impacts by 2050, according to a new article published today in the Proceedings of the National Academy of Sciences. The work was authored by Woods Hole Research Center (WHRC) Risk Program Director Dr. Christopher Schwalm, Dr. Spencer Glendon, a Senior Fellow at WHRC and founder of Probable Futures, and by WHRC President Dr. Philip Duffy.

Long dismissed as alarmist or misleading, the paper argues that is actually the closest approximation of both historical emissions and anticipated outcomes of current global climate policies, tracking within 1% of actual emissions.  “Not only are the emissions consistent with RCP 8.5 in close agreement with historical total cumulative CO2 emissions (within 1%), but RCP8.5 is also the best match out to mid-century under current and stated policies with still highly plausible levels of CO2 emissions in 2100,” the authors wrote. “Not using RCP8.5 to describe the previous 15 years assumes a level of mitigation that did not occur, thereby skewing subsequent assessments by lessening the severity of warming and associated physical climate risk.

Four scenarios known as Representative Concentration Pathways (RCPs) were developed in 2005 for the most recent Intergovernmental Panel on Climate Change Assessment Report (AR5). The RCP scenarios are used in global climate models, and include historical greenhouse gas emissions until 2005, and projected emissions subsequently. RCP 8.5 assumes the greatest fossil fuel use, and a resulting additional 8.5 watts per square meter of radiative forcing by 2100. The commentary also emphasizes that while there are signs of progress on bending the global emissions curve and that our emissions picture may change significantly by 2100, focusing on the unknowable, distant future may distort the current debate on these issues. “For purposes of informing societal decisions, shorter time horizons are highly relevant, and it is important to have scenarios which are useful on those horizons. Looking at mid-century and sooner, RCP8.5 is clearly the most useful choice,” they wrote.The article also notes that RCP 8.5 would not be significantly impacted by the COVID-19 pandemic, adding that “we note that the usefulness of RCP 8.5 is not changed due to the ongoing COVID-19 pandemic. Assuming pandemic restrictions remain in place until the end of 2020 would entail a reduction in emissions of -4.7 Gt CO2. This represents less than 1% of total cumulative CO2 emissions since 2005 for all RCPs and observations.”
“Given the agreement of 2005-2020 historical and RCP8.5 total CO2 emissions and the congruence between current policies and RCP8.5 emission levels to mid-century, RCP8.5 has continued utility, both as an instrument to explore mean outcomes as well as risk,”
they concluded. “Indeed, if RCP8.5 did not exist, we’d have to create it.”


You can access the original article here:

https://whrc.org/worst-case-co2-emissions-scenario-is-best-match-for-assessing-climate-risk-impact-by-2050/

Featured image: Efbrazil / CC BY-SA (https://creativecommons.org/licenses/by-sa/4.0)

Sea levels rising 60% faster than projected by IPCC

By Institute of Physics

Sea-levels are rising 60 per cent faster than the Intergovernmental Panel on Climate Change’s (IPCC) central projections, new research suggests.

While temperature rises appear to be consistent with the projections made in the IPCC’s fourth assessment report (AR4), satellite measurements show that sea-levels are actually rising at a rate of 3.2 mm a year compared to the best estimate of 2 mm a year in the report.

These findings, which have been published today, 28 November, in IOP Publishing’s journal Environmental Research Letters, are timely as delegates from 190 countries descend on Doha, Qatar, for the United Nation’s 18th Climate Change Conference this week.

The researchers, from the Potsdam Institute for Climate Impact Research, Tempo Analytics and Laboratoire d’Etudes en Géophysique et Océanographie Spatiales, state that the findings are important for keeping track of how well past projections match the accumulating observational data, especially as projections made by the IPCC are increasingly being used in decision making.

The study involved an analysis of global temperatures and sea-level data over the past two decades, comparing them both to projections made in the IPCC’s third and fourth assessment reports.

Results were obtained by taking averages from the five available global land and ocean temperature series.

After removing the three known phenomena that cause short-term variability in global temperatures – solar variations, volcanic aerosols and El Nino/Southern Oscillation – the researchers found that the overall warming trend at the moment is 0.16°C per decade, which closely follows the IPCC’s projections.

Satellite measurements of sea-levels showed a different picture, however, with current rates of increase being 60 per cent faster than the IPCC’s AR4 projections.

Satellites measure sea-level rise by bouncing radar waves back off the sea surface and are much more accurate than tide gauges as they have near-global coverage; tide gauges only sample along the coast. Tide gauges also include variability that has nothing to do with changes in global sea level, but rather with how the water moves around in the oceans, such as under the influence of wind.

The study also shows that it is very unlikely that the increased rate is down to internal variability in our climate system and also shows that non-climatic components of sea-level rise, such as water storage in reservoirs and groundwater extraction, do not have an effect on the comparisons made.

Lead author of the study, Stefan Rahmstorf, said: “This study shows once again that the IPCC is far from alarmist, but in fact has under-estimated the problem of climate change. That applies not just for sea-level rise, but also to extreme events and the Arctic sea-ice loss.”

Beyond Reforestation, Let’s Try Proforestation

Beyond Reforestation, Let’s Try Proforestation

Editor’s note: The International Day for Biodiversity was celebrated on May 22, which commemorates the adoption of the Convention on Biological Diversity, a global treaty. What lessons have we learned from undoing past harms and conserving biodiversity for our planet’s future?

Global efforts to restore forests are gathering pace, driven by promises of combating climate change, conserving biodiversity and improving livelihoods. Yet a recent review published in Nature Reviews Biodiversity warns that the biodiversity gains from these initiatives are often overstated — and sometimes absent altogether.

Restoration has typically prioritized utilitarian goals such as timber production, carbon sequestration or erosion control. This bias is reflected in the widespread use of monoculture plantations or low-diversity agroforests. Nearly half the forest commitments in the Bonn Challenge to restore degraded and deforested landscapes consist of commercial plantations of exotic species, a trend that risks undermining biodiversity rather than enhancing it.

Scientific evidence shows that restoring biodiversity requires more than planting trees. Methods like natural regeneration — allowing forests to recover on their own — can often yield superior biodiversity outcomes, though they face social and economic barriers. By contrast, planting a few fast-growing species may sequester carbon quickly but offers little for threatened plants and animals.

Biodiversity recovery is influenced by many factors: the intensity of prior land use, the surrounding landscape and the species chosen for restoration. Recovery is slow, often measured in decades, and tends to lag for rare and specialist species. Alarmingly, most projects stop monitoring after just a few years, long before ecosystems stabilize.

Scientists underline that while proforestation, reforestation and forest rewilding can contribute to curbing climate change and biodiversity loss, they have their limits and must be combined with deep carbon emissions cuts and conservation of existing forests and wilderness.


By Sruthi Gurudev / Mongabay

Edward Faison, an ecologist at the Highstead Foundation, stood quietly in a patch of forest that stretched for miles in all directions. Above him, the needles from white pine trees swayed — common in the Adirondack Forest Preserve in northern New York state. He stepped past downed wood and big, broken snags, observing how the forest functioned with minimal interference.

“These forests have been essentially unmanaged for over 125 years. To see them continue to thrive and accumulate carbon, recover from natural disturbances and develop complexity without our help reveal just how resilient these systems are,” Faison says.

Protected from logging in 1894 by an act of the New York Legislature, the Adirondack Forest Preserve (AFP) is a model of natural forest growth, or letting forests simply “get on with it.” The largest trees, white pines (Pinus strobus), are more than a century old and stretch more than 150 feet tall and are 4-5 feet in diameter.

The AFP, the largest wilderness preserve in the eastern United States, is a prime example of what researchers have come to call “proforestation.” Coined in 2019 by Tufts University professor William Moomaw and Trinity College professor of applied science Susan Masino, the term proforestation describes the process of allowing existing forests to continue growing without human interference until they achieve their full ecological potential for carbon sequestration and biological diversity.

Proforestation is considered a natural climate solution, i.e., a strategy to steward the Earth’s vegetation to increase the removal of carbon dioxide (CO2) from the atmosphere. According to Faison, a forest naturally develops greater complexity over time, with a diversity of tree sizes and heights as well as large standing dead trees and downed logs. This complexity provides habitat for various animals, plants and fungi, which make the forest more resilient to disturbances associated with climate change.

Proforestation is distinct from reforestation, which can involve planting new trees in deforested areas to restore them (or allowing deforested areas to naturally regenerate). It is also different from afforestation, which is the process of planting new forests in previously unforested areas. Proforestation’s merit lies in inaction: simply leaving old forests undisturbed, allowing for continuous growth to maximize carbon accumulation over time. As forests mature and trees grow larger, they sequester greater amounts of carbon.

“The largest 1% diameter trees in a mature multiage forest hold half the carbon,” according to Moomaw. “It’s the existing forests that we have that are doing the work.” Existing forests remove almost 30% of CO2 from the atmosphere that humans put in every year from burning fossil fuels.

Older is better

In Mohawk Trail State Forest in Massachusetts, Moomaw studied the tallest grove of white pine trees in New England, aged between 150 and 200 years, observing how the trees grew. When comparing them with younger trees of the same type growing under similar conditions, he found that “the amount of carbon added by these trees between 100 and 150 years of age is greater than the amount added between zero and 50.”

In addition to carbon storage capabilities, old forests are pivotal in controlling regional and global water cycles through a process called evapotranspiration, by which water is transferred from the land to the atmosphere. Due to deeper and more complex root systems as well as larger canopies and leaves, old forests capture more water and release it as vapor into the atmosphere.

“Old forests have the genetic competence to do this work,” Masino says. “It’s not done by meadows. It’s not done by grassy areas. It’s not done as effectively by forests that have been cut or planted. It’s these ancient systems that have the complexity to bring water to themselves. And in doing that, they’re bringing it to the rest of the landscape. Once you start cutting the landscape, you’re drying it out.”

Masino, who also has a joint appointment in neuroscience and psychology at Trinity College, emphasizes the importance of designating natural areas appropriately and allowing more room for proforestation.

“It’s urgent to decide where we intend to prioritize natural processes, where we are doing research, and what areas we are dedicating for our resource needs,” she says. “Nature needs room to breathe. We can’t leave everything open to manipulation and extraction. It’s deadly.”

She says that planting trees on streets, on campuses or in parks is good for temperature regulation, flood protection and creating habitat, but these trees don’t grow up in a web of life. Planting trees in a forest, too, can risk disrupting the dynamic complexity of evolved and evolving genetic knowledge.

Wildlife dependent on old growth

Over on the West Coast, University of Oregon professor emeritus Beverly Law has studied forests for decades. She describes watching three logging trucks, each with a giant log from an old, single tree strapped to the back, passing in a procession while waiting at an intersection on her bike, a frequent occurrence on her way to work at the university in the late 1980s.

“There are plant and animal species that rely on these old forests for their survival. You take away the forest, and they’re gone,” Law says. “It’s important to have diverse genetics in the forest. Some of them will be more genetically able to withstand climate change than others. You don’t know which ones they will be. That is why genetic diversity within species is important.”

Mature forests are crucial to the survival of certain critically endangered animals that rely on the connected canopies or the soil-rich forest floor. Preserving the biodiversity of the Pacific Northwest, which hosts forests more than a thousand years old, is especially dire. According to a 2022 paper published in Environmental Chemistry Letters, old growth forests retain a number of species from both the top and bottom of the food chain, such as the Olympic salamander (Rhyacotriton olympicus), the Del Norte salamander (Plethodon elongatus) and the two species of tailed frog (Ascaphidae). Losing them forever could kick off a cascade effect and result in severe consequences for the environment.

The spotted owl (Strix occidentalis), too, depends on old-growth forests in the Pacific Northwest, requiring the specific environment for roosting and nesting, and remains a central figure in forest management debates.

Such hulking ancient trees are the eyes of the woods, having stood through changing years and the changing climate.

“Ten to 12% of old-growth forests are left [in the US], and it’s insane that people are still trying to cut them down,” Law says. “They are the only survivors of American handiwork. Is it man’s dominion over the forest? We should have reverence, considering they’re all that’s left.”

Banner image: Pine cone of a white pine (Pinus strobus). Image by Denis Lifanov via Flickr (CC BY-NC-SA 2.0).

Energy Transition: Never Was and Never Will Be One

Energy Transition: Never Was and Never Will Be One

Editor’s note: “Energy is, of course, fundamental to both human existence and the functioning of capitalism. It is central to production, as well as the heating and lighting systems that most people take for granted, and the energy sector is by far the single largest producer of greenhouse emissions.” A transition to 100% electrical energy will never happen. The percentage of electrical energy is 20%, of which 3% are “renewable”. Those figures have never been higher in well over 50 years. Also everywhere in the world, the development of “renewables” has and remains propped up by government support.

From a distance, the Ivanpah solar plant looks like a shimmering lake in the Mojave Desert(a death trap for migratory). Up close, it’s a vast alien-like installation of hundreds of thousand of mirrors pointed at three towers, each taller than the Statue of Liberty. When this plant opened near the California-Nevada border in early 2014, it was pitched as the future of solar power. Just over a decade later, it’s closing. Ivanpah now stands as a huge, shiny monument to wasted tax dollars and environmental damage — campaign groups long criticized the plant for its impact on desert wildlife.

“It was a monstrosity combining huge costs, huge subsidies, huge environmental damage, and justifications hugely spurious. It never achieved its advertised electricity production goals even remotely, even as the excuses flowed like wine, as did the taxpayer bailouts.

And now, despite all the subventions, it is shutting down about 15 years early as a monument to green fantasies financed with Other People’s Money, inflicted upon electricity ratepayers in California denied options to escape the madness engendered by climate fundamentalism.”

Instead of forcing coal and oil into obsolescence, we’re merely adding more energy to the system — filling the gap with “renewables” while still burning record amounts of fossil fuels. This is the real danger of the “energy abundance” mindset: it assumes that a limitless supply of “clean” energy will eventually render fossil fuels obsolete. In reality, “renewable” energies are not replacing fossil fuels, but supplementing them, contributing to a continued pattern of broad energy consumption.


 

Historian Jean-Baptiste Fressoz: ‘Forget the energy transition: there never was one and there never will be one’

At first glance, no one is waiting for a historian to play down the idea of an energy transition. Certainly not at a time of environmental headwinds. But above all, Fressoz wants to correct historical falsehoods and reveal uncomfortable truths. ‘Despite all the technological innovation of the 20th century, the use of all raw materials has increased. The world now burns more wood and coal than ever before.’

In his latest book, More and more and more, the historian of science, technology and environment explains why there has never been an energy transition, and instead describes the modern world in all its voracious reality. The term “transition” that has come into circulation has little to do with the rapid, radical upheaval of the fossil economy needed to meet climate targets.

In France, Jean-Baptiste Fressoz has been provoking the energy and climate debate for some time. He denounces the obsession with technological solutions to climate change and advocates a reduction in the use of materials and energy.

The cover of the French edition of your book says ‘the energy transition is not going to happen’. Why do you so strongly oppose this narrative?

We are reducing the carbon intensity of the economy, but that is not a transition. You hear very often that we just need to organise ‘a new industrial revolution’, most recently by US climate envoy John Kerry. You cannot take this kind of historical analogy seriously, this is really stupid.

The idea of an energy transition is actually a very bizarre form of future thinking, as if we would transform from one energy system to another over a 30-year period and stop emitting CO2. If it were to come across as credible, it is because we do not understand the history of energy.

But don’t we have historic precedents? Didn’t we transform from a rural economy that ran on wood to an industrial society with coal as the big driver?

This is an example of the many misconceptions of the history of energy. In the 19th century, Britain used more wood annually just to shore up the shafts of coal mines than the British economy consumed as fuel during the 18th century.

Of course it is true that coal was very important for the new industrial economy in 1900, but you cannot imagine that as if one energy source replaced the other. Without wood, there would be no coal, and therefore no steel and no railways either. So different energy sources, materials and technologies are highly interdependent and everything expands together.

So I guess you won’t agree either with the claim that oil replaced coal in the last century?

Again, oil became very important, but this is not a transition. Because what do you use oil for? To drive a car. Look at Ford’s first car of the 1930s. While it ran on fuel, it was made of steel, requiring 7 tonnes of coal. That’s more than the car would consume in oil over its lifetime! Today it is no different: if you buy a car from China, it still requires about three tonnes of coal.

You should also take into account the infrastructure of highways and bridges, the world’s biggest consumers of steel and cement, and that is just as dependent on coal. Oil drilling rigs and pipelines also use large amounts of steel. So behind the technology of a car is both oil and a lot of coal.

You suggest looking at energy and the climate problem without the idea of ‘transition’. How? 

Focus on material flows. Then you see that despite all the technological innovation of the 20th century, the use of all raw materials has increased (excluding wool and asbestos). So modernisation is not about ‘the new’ replacing ‘the old’, or competition between energy sources, but about continuous growth and interconnection. I call it ‘symbiotic expansion’.

How do you apply this idea of symbiotic expansion of all materials to the current debate about the energy transition?

The energy transition is a slogan but no scientific concept. It derives its legitimacy from a false representation of history. Industrial revolutions are certainly not energy transitions, they are a massive expansion of all kinds of raw materials and energy sources.

Moreover, the word energy transition has its main origins in political debates in the 1970s following the oil crisis. But in these, it was not about the environment or climate, but only about energy autonomy or independence from other countries.

Scientifically, it is a scandal to then apply this concept to the much more complex climate problem. So when we seek solutions to the climate crisis and want to reduce CO2 emissions, it is better not to talk about a transition. It is better to look at the development of raw materials in absolute terms and to understand their intertwinedness. This will also restrain us from overestimating the importance of technology and innovation .

Didn’t technological innovation bring about major changes?

Numerous new technologies did appear and sometimes they rendered the previous ones obsolete, but that is not linked to the evolution of raw materials. Take lighting, for example. Petroleum lamps were in mass use around 1900, before being replaced by electric light bulbs. Yet today we use far more oil for artificial lighting than we did then: to light the headlights of the many millions of cars.

So despite impressive technological advances, the central issue for ecological problems remains: raw materials, which never became obsolete. We speak frivolously about technological solutions to climate problems, and you can see this in the reports of the IPCC’s Working Group 3.

Don’t you trust the IPCC as the highest scientific authority on climate?

Let me be clear, I certainly trust the climate scientists of groups 1 and 2 of the IPCC, but I am highly critical of the third working group that assesses options for the mitigation of the climate crisis. They are obsessed with technology. There are also good elements in their work, but in their latest report they constantly refer to new technologies that do not yet exist or are overvalued, such as hydrogen, CCS and bioenergy (BECCS).

The influence of the fossil industry is also striking. All this is problematic and goes back to the history of this institution. The US has been pushing to ‘play the technology card’ from the beginning in 1992. Essentially, this is a delaying tactic that keeps attention away from issues like decreasing energy use, which is not in the interest of big emitters like the US.

What mitigation scenarios do exist that do not rely excessively on technology? 

As late as 2022, the IPCC’s Working Group 3 report wrote about ‘sufficiency’, the simple concept of reducing emissions by consuming less. I’m astonished that there is so little research on this. Yet it is one of the central questions we should be asking, rather than hoping for some distant technology that will solve everything in the future.

Economists tell what is acceptable to power because it is the only way to be heard and to be influential, it is as simple as that. That is why the debate in the mainstream media is limited to: ‘the energy transition is happening, but it must be speeded up’.

The transition narrative is the ideology of 21st century capitalism. It suits big companies and investors very well. It makes them part of the solution and even a beacon of hope, even though they are in part responsible for the climate crisis. Yet it is remarkable that experts and scientists go along with this greenwashing.

Do you take hope from the lawsuits against fossil giants like Shell and Exxon? 

Of course Exxon has a huge responsibility and they have been clearly dishonest, but I think it is too simplistic to look at them as the only bad guys.  Those companies simultaneously satisfy a demand from a lot of other industries that are dependent on oil, like the meat industry or aviation. More or less the whole economy depends on fossil fuels, but we don’t talk as much about them.

That’s why it is inevitable to become serious about an absolute reduction in material and energy use, and that is only possible with degrowth and a circular economy. That is a logical conclusion of my story, without being an expert on this topic.

Degrowth is not an easy political message. How can it become more accepted?

I do not offer ‘solutions’ in my book since I don’t believe in green utopias. It is clear that many areas of the economy won’t be fully decarbonized before 2050, such as cement, steel, plastics and also agriculture. We have to recognise this and it means that we simply won’t meet the climate targets.

Once you realise this, the main issue becomes: what to do with the CO2 that we are still going to emit? Which emissions are really necessary and what is their social utility?  As soon as economists do a lot more research into this, we can have this debate and make political choices. Yet another skyscraper in New York or a water supply network in a city in the Global South?