The true price of nuclear power

Maple leaf

Several times this blog has discussed whether climate change is making nuclear power a more acceptable option (1, 2, 3). One element of the debate that bears consideration is the legacy of contamination at sites that form part of the nuclear fuel cycle: from uranium mines to post-reactor fuel processing facilities. The Rocky Flats Plant in the United States is an especially sobering example.

Insiders at the plant started “tipping” the FBI about the unsafe conditions sometime in 1988. Late that year the FBI started clandestinely flying light aircraft over the area and noticed that the incinerator was apparently being used late into the night. After several months of collecting evidence both from workers and by direct measurement, they informed the DOE on June 6, 1989 that they wanted to meet about a potential terrorist threat. When the DOE officers arrived, they were served with papers. Simultaneously, the FBI raided the facilities and ordered everyone out. They found numerous violations of federal anti-pollution laws including massive contamination of water and soil, though none of the original charges that led to the raid were substantiated.

In 1992, Rockwell was charged with minor environmental crimes and paid an $18.5 million fine.

Accidents and contamination have been a feature of facilities handling nuclear materials worldwide. Of course, this does not suffice to show that nuclear energy is a bad option. Coal mines certainly produce more than their share of industrial accidents and environmental contamination.

The trickiest thing, when it comes to evaluating the viability of nuclear power, is disentangling exactly what sort of governmental subsidies do, have, and will exist. These subsidies are both direct (paid straight to operators) and more indirect (soft loans for construction, funding for research and development). They also include guarantees that the nuclear industry is only responsible for a set amount of money in the result of a catastrophic accident, as well as the implicit cost that any contamination that corporations cannot be legally forced to correct after the fact will either fester or be fixed at taxpayer expense. Plenty of sources claim to have a comprehensive reckoning of these costs and risks, but the various analyses seem to be both contradictory and self-serving.

Before states make comprehensive plans to embrace or reject nuclear power as a climate change mitigation option, some kind of extensive, comprehensive, and impartial study of the caliber of the Stern Review would be wise.

Carbon pricing and local food

Ottawa Hostel Outdoor Club

I have been hearing a lot about food miles lately. While it is good for people to be aware of the productive processes that support them, I also have issues with the shape of the local foods debate. Just because something is produced closer to where you live does not mean it is more ecologically friendly or less climate harming. To take an extreme case: people living in very cold regions may find that it is far more environmentally sound to import food from warmer places than to grow it in greenhouses nearby. My focus here is on greenhouse gas emissions, but similar arguments can be made regarding water use, pesticides, etc.

What this debate really demonstrates is the lack of proper carbon pricing in the market. If CO2 emissions were included in the price of tomatoes or bananas, producers could choose to base production in whichever location is most efficient when carbon emissions (along with other factors) are taken into account. Until proper carbon pricing exists, there is justification for the intelligent application of the food miles concept. That said, I think the energy of environmentally aware people is much better spent advocating carbon taxes or cap-and-trade schemes than on encouraging people to spent their time buying local zucchini, rather than whatever sort is at the supermarket. By all means, attend farmers’ markets if you like them, but I think it is deluded to think they can make a meaningful contribution to reducing human emissions to 5,000 Mt of CO2 equivalent.

I can already feel the dissenting opinions coming on, for this post…

Jeffersonian trivia

Little known facts:

  1. Former American President Thomas Jefferson was an avid amateur palaeontologist.
  2. In an attempt to mock him, his political opponents gave him the nickname “Mr. Mammoth” during the 1808 election.
  3. He is credited with the discovery of an enormous ground sloth, larger than an elephant, that inhabited North America during the late Pleistocene.
  4. The creature now bears his name: Megalonyx jeffersonii.

These and many other entertaining facts come from the marvellous recent book The World Without Us, which has leapt to the top of my reading pile. I will post a full review when I finish it.

Sputnik at 50

Bridge on the Rideau Canal

Even the Google logo has been altered to commemorate the 50th anniversary of the launch of Sputnik 1: the first artificial satellite. As someone who spends a very considerable amount of time thinking about how things are going to be in 2050 and 2100, it is remarkable to reflect upon both how different the world is from that of 1957 and how similar it is. The big changes that occurred have often been in areas that few if any people would have anticipated the importance of back then. Areas of great enthusiasm, such as nuclear power and space exploration, have only progressed incrementally since the 1950s and 60s.

I mentioned one Sputnik-related irony in a paper published back in 2005:

At the end of August, 1955, the Central Committee of the Communist Party approved the Soviet satellite program that would lead to Sputnik and authorized the construction of the Baikonour Cosmodrome. This facility, the largest of three Soviet launch sites that would eventually built, was the launching place of Sputnik I (and subsequent Sputniks), and the launch site for all Soviet manned missions…

This former stretch of Kazakhstani desert was also, fatefully, the place to which Nikifor Nikitin was exiled by the Czar in1830 for “making seditious speeches about flying to the moon.” He might have taken cold comfort in the fact that in 1955, the Central Committee gave control of the site to the new Soviet ‘Permanent Commission for Interplanetary Travel.’

For all the drama, it remains unclear to me that manned spaceflight serves any useful scientific or practical purpose at this point in time (see previous). In that sense, perhaps Sputnik – rather than John Glenn – was the true template for humanity’s future involvement in space: an 83.6kg ball of metal with a radio transmitter.

PS. My thesis mentions one somewhat surprising connection between Sputnik and climatic science:

A fortuitous bit of funding produced one of the most famous graphs in the climate change literature: the one tracking CO2 concentrations at Mauna Loa in Hawaii. Examining it closely, a gap can be seen in 1957, where David Keeling’s funding for the project ran out. The Soviet launch of Sputnik I on 4 October 1957 led to a marked concern in the United States that American science and technology had fallen behind. One result of the subsequent surge in funding was the resumption of the CO2 recording program, which continues to the present day.

This graph is the jagged, upward-sloping line that Al Gore devotes so much attention to near the beginning of An Inconvenient Truth.

Vermont’s regulatory victory

Well known as a progressive place, Vermont seems to have recently struck a notable blow in the fight to develop regulatory structures to address climate change. A heated court case had developed between car manufacturers and the state government about whether the latter could impose tough emission limits on cars and light trucks. William Sessions, a federal judge, found in favour of the state’s right to do so. You can read the entire judgment here: PDF, Google Cache.

Among the arguments brought forward by the auto makers (and rejected by Sessions) were that the regulations were unconstitutional, impossible to meet with existing technology, economically disastrous, ineffective, and anti-consumer. The case also involved a reasonably complex jurisdictional issue regarding California’s special exemptions to set environmental policy more broadly than other states.

There do seem to be a suspicious number of cases where industries have followed this trajectory in relation to new regulations: saying that they are unnecessary, saying they would be financially ruinous, then quietly adapting to them with little fuss once they come into force. The phase-out of CFCs in response to the Montreal Protocol is an excellent example. This trend is explicitly recognized in the ruling:

Policy-makers have used the regulatory process to prompt automakers to develop and employ new, state-of-the-art technologies, more often than not over the industry’s objections. The introduction of catalytic converters in the 1970s is just one example. In each case the industry responded with technological advancements designed to meet the challenges…

On this issue, the automotive industry bears the burden of proving the regulations are beyond their ability to meet…

In light of the public statements of industry representatives, history of compliance with previous technological challenges, and the state of the record, the Court remains unconvinced automakers cannot meet the challenges of Vermont and California’s GHG regulations.

The fact that Chinese cars have to meet better emission standards than American ones strongly suggests that the objections of industry are bogus. Given the price inelasticity of demand for gasoline (people keep buying about the same amount when the price goes up), regulating fuel efficiency and emissions does seem like an efficient way to reduce GHG emissions in the transport sector.

The folly of Apollo redux

In an earlier post, I discussed the wastefulness of manned spaceflight. In particular, plans to return to the Moon or go to Mars cannot be justified in any sensible cost-benefit analysis. The cost is high, and the main benefit seems to be national prestige. Human spaceflight is essentially defended in a circular way: we need to undertake it so that we can learn how human beings function in space.

A post on Gristmill captures it well:

Let me be clear. There is a 0 percent chance that this Moon base or anything like it will ever be built, for the following reason: the moon missions in the ’60s and early ’70s cost something like $100 billion in today’s dollars. There is no way that setting up a semipermanent lunar base will be anything other than many times more expensive. That would put the total cost at one to a few trillion dollars.

Assuming that this taxpayer money needs to be lavished on big aerospace firms like Lockheed anyhow, it would be much better spent on satellites for the study of our planet (Some comprehensive temperature data for Antarctica, perhaps? Some RADAR analysis of the Greenland icecap? Some salaries for people studying climatic feedbacks?) or on robotic missions to objects of interest in the solar system.

HCFC phaseout

While international negotiations on climate change don’t seem to be going anywhere at the moment, some further tightening has been agreed within the regime that combats substances that deplete the ozone layer (the Vienna Convention and Montreal Protocol). The parties have decided to speed up the elimination of hydrochlorofluorocarbons (HCFCs), which were permitted as temporary substitutes for the chlorofluorocarbons (CFCs) that destroy ozone most energetically.

The BBC reports that:

The US administration says the new deal will be twice as effective as the Kyoto Protocol in controlling greenhouse gas emissions.

This seems quite implausible to me. HFCs, PFCs, and SF6 collectively contribute about 1% of anthropogenic warming. As such, their complete elimination would have a fairlylimited effect. In addition, the Vienna Convention process always envisioned their elimination, so there is nothing substantially new about this announcement, other than the timing. An agreement for eliminating HCFCs has been in place since 1992:

1996 – production freeze
2004 – 35% reduction
2010 – 65% reduction
2015 – 90% reduction
2020 – 99.5% reduction
2030 – elimination

While it does seem that this timeline isn’t being followed, it remains to be seen whether this new announcement will have any effect on that.

The Kyoto Protocol targets a six different greenhouse gases, most importantly the carbon dioxide that constitutes 77% of anthropogenic climate change. If it had succeeded at reducing emissions among Annex I signatories by 5.2%, as planned, it would have been both a significant contribution and an important starting point.

None of this is to say that we shouldn’t welcome the HCFC phaseout. If nothing else, it should help with the recovery of the ozone layer. We just need to be cautious about accepting statements like the one quoted.

Oryx and Crake

Fire truck valves

Margaret Atwood‘s novel, which was short-listed for the Booker Prize, portrays a future characterized by the massive expansion of human capabilities in genetic engineering and biotechnology. As such, it bears some resemblance to Neal Stephenson‘s The Diamond Age, which ponders what massive advances in material science could do, and posits similar stratification by class. Of course, biotechnology is an area more likely to raise ethical hackles and engage with the intuitions people have about what constitutes the ethical use of science.

Atwood does her best to provoke many such thoughts: bringing up food ethics, that of corporations, reproductive ethics, and survivor ethics (the last time period depicted is essentially post-apocalyptic). The degree to which this is brought about by a combination of simple greed, logic limited by one’s own circumstances, and unintended consequences certainly has a plausible feel to it.

The book is well constructed and compelling, obviously the work of someone who is an experienced storyteller. From a technical angle, it is also more plausible than most science fiction. It is difficult to identify any element that is highly likely to be impossible for humanity to ever do, if desired. That, of course, contributes to the chilling effect, as the consequences for some such actions unfold.

All in all, I don’t think the book has a straightforwardly anti-technological bent. It is more a cautionary tale about what can occur in the absence of moral consideration and concomitant regulation. Given how the regulation of biotechnology is such a contemporary issue (stem cells, hybrid embryos, genetic discrimination, etc), Atwood has written something that speaks to some of the more important ethical discussions occurring today.

I recommend the book without reservation, with the warning that readers may find themselves disturbed by how possible it all seems.

Types of goods

In economic theory, most things you can buy are ‘normal goods.’ This means that, as the price rises, people buy less of them. Conversely, people buy more as the price falls. This is all quite self-explanatory but it is interesting to note that there are other types of goods that operate in different ways.

The most common example may be inferior goods. The richer people get, the less they spend on inferior goods. This includes most kinds of discount items: once people can afford something better, they make the switch. Inferior goods reflect this property both at the micro level (an individual gets a big raise and buys less cheap IKEA furniture) and at a macro level (the mean income in a state rises and demand for low-cost gruel falls). Long distance bus trips are a classic example of an inferior good, as anyone who has spent more than twelve hours in a smelly, noisy coach can easily understand.

A somewhat perverse counterpoint to inferior goods can be found in Veblen goods. Named after the economist Thorstein Veblen, these are products for which the demand actually rises as the price does. This is essentially on account of their exclusivity. People buy Velben goods (such as Rolls Royce cars and $50,000 cell phones) precisely to demonstrate that they can. Of course, this makes them a godsend for those hoping to part status conscious rich suckers from some of their wealth.

A final possibility, which may not actually exist, is a Giffen good. To qualify, the good needs to be inferior (in the sense described above), there must be a lack of close substitutes, and the good must comprise a significant share of the purchaser’s budget. With these goods, price rises also lead to people buying more, though for a rather different reason. People who have become too poor to buy a better option fall back on a worse option. The failure of economists to find any well-defended empirical examples suggests that this kind of good may exist only in the minds of academics.

Both Giffen goods and Veblen goods exist because of possible characteristics of the buyer, rather than of the good itself. Whereas Giffen goods are easy to reconcile with ‘rationality’ as understood by economists, Velben goods do so only when they are viewed as inputs in the manufacture of the commodity actually sought: such as social status or prestige.

People wanting to read even more about goods and economic theory can look into the distinction between rivalrous and non-rivalrous goods and excludable and non-excludable goods. The two ideas together define public goods and common property goods, the existence of which make even the most hard-nosed economist recognize the efficiency of governmental action to regulate markets.

Hired guns

I heard a lot fair amount about mercenaries when I was at Oxford, but this is the most interesting thing to happen in relation to them in decades. The degree to which war has been privatized would probably shock Eisenhower.

What remains to be seen is the degree to which the United States will respect the sovereignty of the democratic government that all the entire second Iraq war was meant to create.