Book Review: Smart Power: Climate Change, the Smart Grid, & the
Future of Electric Utilities: Anniversary Edition – by Peter Fox-Penner (Island
Press, 2010, 2014)
This is an excellent book that delves into the changing
utility models taking shape with new sources of distributed energy, often
renewable, and smart grid technologies that can make the electric grid more
efficient and more responsive, yet still keep it stable and reliable.
The new edition begins with four long forwards, the first by
retired Duke Energy CEO and chairman Jim Rogers. He credits the “pressures of
climate change” and the “capabilities of emerging technologies” with forcing
the utility industry to begin to change from their outdated business models and
regulatory structures. Trends that have been occurring will continue: less
coal, more natural gas, probably not much change in nuclear, more wind, more
solar, more distributed resources, more battery storage, and more renewables-based
peak load shifting. Rogers echoes Dr. Fox-Penner’s two main conclusions:
changes are needed in utility business models and utility regulations.
The next forward is by Daniel Esty, a professor of
environmental law and policy at Yale. He notes the unexpected rise of shale gas
and oil through fracking and horizontal drilling which brought vast new and
abundant gas supply onto the market, quickly making gas cheaper than oil and
coal which had not been the case before. Now gas is the cheapest and the lowest
emissions fossil fuel, and this will likely be the case for at least the next
decade or two, he notes. He mentions state incentives, including making private
capital available for solar startups and standardizing energy service company
(ESCOs) contracts. His state, Connecticut, has implemented these projects to
make renewable energy cheaper. They have also begun buildout of a system of
microgrids powered with small gas turbines, fuel cells, and biogas from
anaerobic digesters. These are forms of distributed generation (capable of
offgrid “islanding”) in addition to most renewable sources, especially if they
have battery backup. He also notes the need for a change in utility business
models with other sources of revenue besides selling energy. One is selling
energy services where energy efficiency is commodified. Another is managing
demand fluctuation (demand response services). Others are managing microgrids
and providing backup power for distributed resources.
The third forward is by Daniel Dobbeni, a principal in
several companies. He notes that utilities and power sources are affected by
events and policies (Fukushima in Japan, carbon markets and laws in Europe,
etc.) He notes that the effects of renewables on the grid (need for backup/peak
generation and equipment to balance supply and demand of loads) has been
largely ignored until now. He notes the potential for stranded assets in the
form of transmission lines that could come about as new technologies replace
them. He notes that the focus of energy policy in Europe shifted from
competitiveness to sustainability, culminating with the 2007 renewable energy
directive, with new infrastructure from 2013 onward focused on connecting
Europe as a whole. Europe has adopted a common market model for the power
industry, unlike other areas such as in the U.S. Power ‘flexibility,’ such as
demand side management, is required in the increasingly distributed energy
systems. The European model is to be watched and studied so better models can
follow it.
The final forward is by Lyndon Rive, cofounder and executive
officer of SolarCity. He focuses on the potential for economic catastrophe due
to climate change. Making the grid cleaner and more efficient should be vastly
accelerated, he notes (as any solar executive should/would). His focus is to do
away with the old utility business models faster to save money since the
transition to renewables is inevitable and doing otherwise would be wasteful.
He sees midday “overgeneration” from solar more as an opportunity than as a
problem, one that might be used for water desalinization and electrification of
transportation. He says the old U.S. utility monopoly model of “sole
source/cost plus,” where competition is suppressed and profits are guaranteed,
is economically undesirable and unsustainable. NASA migrated away from such a
model quite successfully and so should U.S. utilities, he says.
Fox-Penner starts out with an account of the first electric
revolution. Electric machines and electromagnets in foundries led to drastic
increases in work output in the late 1800’s. He talks about the industry
developed by Insull, one of Thomas Edison’s staff who became the CEO of
Commonwealth Edison. Power was aggregated into what we know as the “grid” and
power was sold to take advantage of economies of scale. The model was monopoly with
the more power consumed the cheaper the cost per unit of power for the
consumer. He also noted that regulation of the industry would provide both
stability and protections. It was Insull’s vision that influenced the investor
owned utility (IOU) model that has been the standard in the U.S until recently.
The second electric revolution is in progress. Two of the
main motives are the need to reduce greenhouse gas emissions and the need for
energy security. Renewable energy, energy efficiency, and replacing high carbon
fossil fuels (coal) with low carbon fossil fuels (gas) are the three main ways
of reducing carbon emissions in the electric sector. Securing domestic energy
supply (oil, gas, coal, nuclear, and renewables) helps reduce our dependence on
foreign suppliers, particularly on OPEC countries and oligarchies like Russia
who use energy as political leverage. Electric transport via EVs can also help
us with both. Such needed changes, says Fox-Penner, will be both challenging
and expensive for utilities and their rate-payers. Implementation of the
so-called “smart grid” with software embedded technologies to optimize
efficiency, balance local grids, and buy and sell energy at different rates,
will happen more and more over the next few decades. Less coal, more gas, and
more renewables will be on the grid. More battery and other energy storage
projects will be added. Grid management will be complexified and yet it will be
more efficient overall. Fox-Penner mentions three objectives: “creating a
decentralized control paradigm, re-tooling the system for low-carbon supplies,
and finding a business model that promotes much more efficiency. The
consumption model where utilities profit strictly from sales of power is not
sustainable. Electric power demand in the U.S. has been flat for a while and is
expected to stay there. This is due to efficiency increases and reduction of
waste usage on both sides of the meter. One question is how investor owned
utilities can remain viable – can remain profitable while meeting all these
challenges. New business models are required. Some utilities have and will be
resistant while others have or are adopting new economic paradigms.
Electric deregulation is the next issue. In 1990 the
consensus was that the electric utility industry should follow airlines,
natural gas suppliers, telephone companies, and trucking firms in allowing
markets rather than regulators to set prices. Deregulation was oversold, says
Fox-Penner, and executed poorly, resulting in the California energy crisis. Early
problems have been fixed with better oversight and better market designs but
there are still problems. Twenty-three states originally implemented
deregulation but eight of those have suspended it or scaled it back. Analysis
indicates some regulation is required for successful power markets.
The three vertical stages of power production from an
engineering perspective are: generation, transmission, and distribution. When a
company owns all three of these aspects it is said to be vertically integrated. Only some of the U.S. power industry is
vertically integrated. Wholesale, or “bulk” power trading between generators
and distributors is subject to pricing set by the Federal Energy Regulatory
Commission (FERC). This includes control of high-voltage transmission systems.
The authority to build transmission systems resides in the individual states
through their public service commissions (PSCs). The PSCs set rates for
transporting power over these systems. Residential, small-commercial, and
large-commercial pricing is at different rates. There are also generators owned
by federal, state, and local governments which are subject to less wholesale
and transmission regulation as are government and customer owned (co-ops)
distribution systems. All the government and customer owned utilities are
thought to be less likely to charge unfair prices so they are less regulated. In
a deregulated market it is the market itself that sets prices for the wholesale
market – transmission and distribution are still regulated by FERC or PSCs. In 1994 FERC was permitted to create an “open
access” system in which any generator could use anyone else’s transmission
system (first-come, first-served) to deliver power to a state-regulated system.
FERC also began allowing some generators to sell wholesale power to other
utilities, not end users, at deregulated rates. In order to develop fair
competition in all power markets there are three requirements: 1) enough
competing power generators (deconcentration), 2) a transmission system large
enough to accommodate all generators, and 3) “open access” rules. If markets
can allow buyers to react to price changes then they will be more functional.
Power prices tend to vary hour to hour. Unfortunately, there are very few
places where all these conditions occur currently but some regional power
markets are testing new dynamic pricing, or ‘time-of-use’ models.
Deregulation in the states that adopted it was given as a
consumer choice so that consumers may choose a provider or choose not to
participate in the deregulated market. This is likely due to the previous
deregulation of telephone service where consumers were forced to choose a
provider. Since deregulated prices were expected to drop the non-deregulated
(provider of last resort, POLR) prices, the new prices were set 10% lower and
frozen for five years. However, in many areas deregulated providers could not
compete with POLR rates. At the time the cost of fuels rose and deregulated
providers had to raise their rates but POLRs could not which made POLRs cheaper
than deregulated sellers, a situation opposite of what was intended. This was
not the case everywhere and some deregulated providers were/are able to offer lower
prices. In May of 2000 hourly electric prices in California began spiking.
Energy was in short supply, especially natural gas. This caused intentional
blackouts and rolling blackouts for consumers. Prices up to 10 times the
previous averages continued in the northwest. The FERC stepped in to provide
caps on prices and other measures. By July 2001, the crisis was over. It cost
Californians $33 billion more than the previous year’s cost! This gave
deregulation a bad reputation. By 2006-2008 when the POLR 10% off rates expired
they went up considerably, further bashing the customer advantage claims for a
deregulated power industry. Fox-Penner suggests that the relative failures and
turbulent history of deregulation in the power industry have galvanized
resistance to further “changes,” some of which are necessary to adapt to
changing energy sources and business paradigms. Deregulation was only adopted
by some states so it further confuses and adds to the already significant
variability of generation owner-types and regulation models. The U.S. power
industry as a result is quite heterogeneous which makes adopting new
regulations and business models and getting new projects going difficult,
especially as smart grid technology advances.
Next he goes through some early (2005) dynamic
pricing/time-of-use pricing experiments where energy consumers could alter
their daily power use to take advantage of the best prices, through smart grid
software embedded technology. The results were excellent in keeping power use
down, with the possible application of decreasing power spikes that require
extra generation usually provided through idling and ready power plants
(peaking plants) built for demand peaks.
He explains how second-by-second balancing of power supply
and demand on the grid happens with system operators constantly monitoring it
at a control center called the ‘balancing authority.’ System operators
typically have ‘reserve capacity’ in the form of demand “peaking” plants.
Typically for every 100 MW of baseload capacity there is 15 MW of reserve
capacity, so 15%. About 5 MW, or 5%, will be ready and idling at any given time,
burning fuel and producing emissions.
Another feature of power use assessment is its one-way
nature. The old “dumb” meters measure how much power we use over the course of
a month, the billing period. Pricing may vary by season but that is about it.
New smart meters keep track of hourly power use and may charge according to
times, the price per kilowatt-hour varying with established demand periods. The
technology is capable of responding to “dynamic pricing” which may change
minute-by-minute and communicating back and forth through high-speed internet.
In the past utilities encouraged more power use by dropping prices as power use
increased through the month, so-called “declining block rates.” Dumb meters are
like charging for groceries by weight rather than by item as one utility
executive puts it.
Fox-Penner gives a kind of average scenario where the cost
for a utility to generate one kWh varies from 2-3 cents in the dead of night to
8-20 cents in high demand periods from hot or cold days when old and
inefficient plants in reserve are turned on. Another potential feature of smart
meters is that they can work with smart appliances to respond directly to price
changes or simply programmed to run at lower demand/lower price times of the
day or night. Smart meters are also more useful for integrating renewables and
other distributed sources and battery storage. Thus the end users can
self-balance their own power supply and demand. During high demand times if
enough users reduce their power usage with smart appliances then there may be
no need to turn on expensive reserve capacity. Fox-Penner says that while
system operators do employ demand response software that they will still
balance the grid manually for a long time to come.
The smart grid offers three basic advantages: 1) greater
customer control over energy use and costs, 2) enabling of local small-scale
power production, and 3) a more reliable and secure grid. However, he notes,
there are complications. Smart meter technology that regulates the downstream
distribution end of the grid is different from high-voltage transmission coming
from the upstream generation end where many power plants feed in. Hourly
dynamic pricing and trading has long been a feature at this
generation-transmission end as has been smart computerized control and
switching. Even so, he notes, new software technologies can still improve that
end. Such technologies will allow better “situational awareness’ of grid issues
and disruptions than previously. Smart grid technology will make it easier to
incorporate storage. Having energy storage available means the grid can respond
to demand and balance the grid instantaneously and at low cost with the stored
energy. Utility-scale storage as well as storage associated with distributed
resources on the other end can both be utilized and eventually optimized.
However, cost of implementation, battery life, and other issues keep it from
being implemented widely. Local distributed generation allows two-way flow of
the grid which can be advantageous to the local generators and to the grid as a
whole. The smart grid can provide self-activating tools to increase grid
reliability and prevent blackouts.
Next he goes into real-time pricing, or time-of-use (TOU)
rates, a feature likely to happen more and more in power markets as smart grid
tech advances. In TOU pricing the utility may set the daily variable rates in
advance. In ‘critical peak pricing’ (CPP) the utility may warn users say a day
in advance of expected high demand and high prices so they can plan
accordingly. In real-time pricing (RTP) new rates may be set every hour. Such
variable rates are designed to be neutral to the utilities (prices are shifted
so that the utility makes the same amount of money no matter how much power is
used) in terms of power sales but beneficial to consumers. Dynamic pricing
advocate Ahmad Faruqui found that typical TOU rates can reduce demand peaks by
about 5% and CPP can reduce them by up to 20%. Fox-Penner notes that if this
could be sustained as a national average then about 200 medium-sized power
plants could be eliminated! So too can their carbon emissions and pollution.
This can happen in a fully developed smart grid automatically by programming
devices to respond to expected and short-term price changes. Kurt Yeager
formerly of the Electric Power Research Institute calls it ‘prices to devices.’
Smart thermostats are now more widely available and many other devices are
being outfitted with these ‘enabling technologies.’ Customers can save over a
hundred dollars a year while helping the utilities save money (by not building
and using more reserve capacity and associated grid) and help a little to mitigate
climate change and pollution. By not having to turn on expensive reserve
capacity the overall pricing can be reduced as well which benefits not only
those provided the ‘demand response’ by reducing their usage, but all users on
the system. It takes about 5% of demand drop through demand response to lower
prices for all so these benefits should be quite achievable as the tech
advances. Those who provide the demand response by lowering their usage through
smart tech will of course save much more than the other users. Avoiding the
building of more reserve capacity and associated grid is called avoided capacity cost as the reserve
capacity is now provided by built-in demand response. This may easily become
the largest benefit of DR as it is advanced. Another pricing mechanism is increasing, or inclining block rates. This is simply raising per kWh prices with
excess use per customer. This has long been done and was first implemented by
Insull – no smart tech required.
The barriers and resistance to DR and dynamic pricing come
from the utilities who have to pay for and implement the smart tech. Changing tens
of millions of meters from dumb to smart, or installing ‘advanced metering
infrastructure’ (AMI) is quite expensive. However, it does allow utilities to
eliminate meter readers as meters can now be read electronically. As smart grid
tech advances the cost scenarios will become more familiar to utilities and
regulators so business cases can be more predictable. Some users are not able
to change their use patterns much such as those who do their most business at
times of high demand so they argue that dynamic pricing could hurt them but of
course the overall drop due to DR would mitigate this somewhat. However,
research indicates that only a few percent of users would end up with higher
bills due to well-designed dynamic pricing. Right now we should be around 50%
give or take of smart meter implementation, or AMI, in the U.S. The shift from
flat power rates to variable ones will save customers quite a bit of money and
a little carbon as well.
Next we come to the regulatory realm. The main goals of
utility regulation are to keep utilities from making excessive profits or from
incurring excessive losses. Regulators and utility executives face significant
uncertainty regarding the results of smart grid investments and it will take
time to evaluate them. The benefits are large, mostly external to utilities,
but not easy to measure quantitatively. Large upfront capital investments are
required. Regulators will have a lot to consider. Both DG and DR can add
significantly to ‘avoided capital costs’ for utilities, costs which are much
greater than energy saved costs. However, figuring avoided costs is not easy.
Congress dictated in the 1980’s that states must determine avoided costs every
few years and there are many different methods of determining them. Part of the
difficulty is that the grid is highly connected and determining avoided costs requires
making detailed and extensive hypothetic cases. Wholesale prices are also
regional now as different regional markets sell electricity at different rates.
Transmission can be considered the difference in costs from the two markets it
connects. This is also how natural gas is regionally priced. It is considered a
reasonable approximation of value. However, there are requirements: knowledge
of daily price fluctuations, knowing the value of scheduled power, requiring DG
and DR to keep safety margins of power, and planning for blackouts. Fluctuating
hourly prices in wholesale markets benefit small DG and renewables generators,
say DG advocates. At press time, 42 states had net metering so that distributed
generators, including rooftop solar generators, could sell their power to the
grid at the same rate as they buy it. In terms of avoided costs it is actually
best to install DG or DR in the middle between generation and end-user distribution.
The author mentions Amory Lovins work on DG and his 2002 book – Small is Profitable – as a monumental
text on DG. He catalogued many other side benefits to DG. However, most are
nearly impossible to measure. For instance, having DGs would make terrorist
attacks less disruptive.
More downstream distributed energy sources also mean that
the distribution ends of the grid will need to be reengineered and that will
cost and also raise local issues. Fox-Penner gives a scenario where a new solar
DG generator will temporarily enjoy selling energy for high cost with those
around them also having to pay that high cost until the utility upgrades the
system in that area after which the DG will profit less and those around them
will be relieved. Such seemingly unfair changing pricing may cause disputes and
resentment. Such occurrences have happened on the high voltage end, slowing
development.
Two-way communication between grid operators and customers
and their devices will be necessary for full smart grid implementation and software
platform standards will have to be developed. This is happening now. This is a requirement
for “plug and play interoperability.” Non-profits like the Institute for
Electrical and Electronics Engineers (IEEE or I-triple-E) will likely write up
the standards. Agreeing on standards has been no easy matter in the past and
there are many standards to consider in expanding the smart grid.
Cyber-security is another issue in smart grids. With many different languages
and protocols in ‘supervisory control and data acquisition (SCADA) systems, the
smart grid can be vulnerable to hackers. Russian and Chinese hackers are said
to be already mapping U.S. power systems.
The next topic addressed is total electricity sales, or
power usage. This has remained flat for a while in the U.S. and is projected to
stay that way. This is due to vastly increasing efficiency from all parts of
the system. LED lights are an example from the user end. Utilities,
particularly IOUs, have long planning times and significant investments to
build generation. If planned power plants are built and later found to be not
needed there is a huge sunk cost. If they are not built and there are
blackouts, the consequences can be equally severe. Lower sales make it harder
for the utilities to invest in low carbon energy and the smart grid. There is
no incentive for them. An earlier example is nuclear plant cost overruns in the
1980’s which caused electric rates to skyrocket and bond defaults.
Renewables expansion will require more transmission lines,
particularly from areas with rich wind resources that in the U.S. are typically
remote from populations. People have resisted large transmission expansions in
scenic areas. Transmission regulation is a mix of states, FERC, and regional
transmission organizations. Different types of lines have different purposes and
abilities. He compares AC and DC high-voltage lines. DC is unidirectional and
AC is two-way, or bidirectional. DC lines lose less energy as they traverse
long distances (in one direction only) so they can be more economical in
certain situations. DC lines move hydroelectric power from the
Oregon-California border to Los Angeles and from Canada into New England. They
are better for moving power under water and can be used for offshore wind.
Apparently, there are only about 15-20 of these DC lines currently in the U.S.
Another new transmission technology is superconducting cables which can handle
3-4 times as much current as regular wires. Grid expansion planning is
generally difficult in the deregulated markets due to jurisdictions and rules.
There are often debates about who should pay for new lines, especially those
traversing multiple states and power market regions. New lines will be needed,
especially with utility-scale renewable power plants – which will be required
to help meet carbon reduction goals and state RPSs. Unfortunately, the best
wind, geothermal, and solar resources in the U.S. tend to be in areas with poor
transmission infrastructure in addition to being far away from population
centers where the power is needed. The bottom line is that a low-carbon future
requires more transmission per kWh than our current high-carbon grid. A new
transmission superhighway is in the planning with large power volume lines,
both AC and DC proposed. Even so, a “supergrid” does raise some reliability and
security issues if large amounts of power are coming from single sources. DERs
can provide downstream reliability enhancements but upstream disruptions would
still be problematic.
Availability, cost, and reliability are the three main
issues with low-carbon energy sources. Natural gas peaking plants used to back-up
renewables are the best cost effective means for doing that. Combined cycle gas
turbines (CCGT) are the most efficient power plants with energy conversion efficiency
over 60%. Their cost per kWh is also the lowest of all power sources. They emit
far less carbon and pollutants than coal. Thus coal-to-gas switching is happening on a
big scale and is by far the main reason U.S. greenhouse gas emissions have
dropped. The big risk for these plants is high gas prices. Gas prices will
likely remain quite low but gas needs to be available to new plants in order
for more switching to occur so pipeline expansions are required. Many have been
delayed recently due to public opposition. Low-carbon coal plants that use
gasification and/or carbon sequestration are very few and far between and are
still after many years not widely deployed. I think CCGT gas plants with some
carbon sequestration will make these plants nearly as clean as renewables and will
be a better deal than any coal plants. Carbon capture and sequestration (CCS)
is simply quite expensive and will only be used in a world with carbon prices.
It may be better to switch to gas then incorporate some sequestration with gas
exhausts. Due to the scattered nature of power plants large networks of CO2 pipelines
would be required which people tend not to like and CO2 leaks high enough in
certain areas could be poisonous.
20% of U.S. power comes from nuclear plants and most are set
to be retired over the next 40-50 years. Will more nuclear replace them? Maybe,
maybe not. Costs, safety, security, waste storage, and decommissioning costs
are the major issues.
Wind power will continue to grow, mostly onshore but some
offshore as well. Currently wind makes up about 4.5% of U.S. electricity. That
could double in the next decade. The availability and cost of transmission
limits wind, especially from its best sources in the Great Plains where
transmission is scarce. Variability and so reliability is also an issue. Wind
is most available when least needed. Storage of excess wind power would be
ideal but storage technologies are currently far too expensive even if there
are many projects operating and some mandates. When these “grid integration”
issues are added the economics of wind power drop significantly.
PV solar has similar availability problems but is
significantly more unpredictable than wind due to clouds and there is far less
sun in winter. It does have the advantage that peak generation is close to high
electric demand times. Even so, without demand or storage for the excess supply
during peak generation times there will be grid integration issues with massive
solar deployment. Solar economics, rooftop, utility scale, or solar thermal can’t
even come close to the economics of CCGT gas, even with significant direct
subsidies. Concentrated solar power (CSP), also called thermal solar, has been
built in the U.S. southwest but the economics and the performance have been
poor so far. It also tends to be in areas that require transmission upgrades.
Biomass power comes from four sources: wood waste from paper
and furniture makers, forestry residue, agricultural residue, and methane from
landfills and anaerobic digesters. The attribution of biomass as “net-zero
carbon,” while technically close to
being true (it is carbon lean rather than carbon zero) , should be caveated
with the fact that the carbon is entering the atmosphere much faster than it
would have naturally. This is also true of biogas from anaerobic digestion.
Geothermal and hydroelectric power will see some growth but
overall the places it can be developed are small and there are some
environmental risks with both. Hydrokinetic power from ocean tides and waves is
also in this category of making small contributions to the energy picture.
California defines DG as sources less than 20MW that connect
to the local distribution grid rather than to the high-voltage grid. The four sources are combined heat and power (CHP,
also called cogeneration) – which are usually small gas microturbines; wind; small
PV solar plants; and fuel cells. CHP utilizes waste heat to heat buildings.
Factories and residential complexes utilize these 5-20 MW sources around the
world, with the heat often being delivered through steam tunnels. Utilities
have not been cooperative with CHP developments since it decreases their sales
but more CHP means less carbon emissions so there should be no opposition.
Small home-sized CHP gas microturbines are being used more and more by property
developers. They are currently higher cost than grid power but are expected to
drop as the technology advances. Siting energy using equipment is a challenge
in these systems. They can also be used stand-alone as offgrid islands in many
cases. Small-scale wind is simply too expensive although there is some development.
The same is true of fuel cells. Costs could come down in the far future but now
they are not at all economic. The costs of DG are often better than they appear
due to the avoided costs they provide to the grid utilities which are hard to
measure and often depend on siting and local grid circumstances. CHP is the
cheapest current form of DG. “Observable” costs of DG (without accounting for costs
avoided) are still 2-3 times as conventional large-scale power sources. Regulation
and subsidies help but not enough. Currently there are 76-85 GW of CHP plants
at 3300 sites but this could be expanded to an additional 80-100 GW by 2030. He
doesn’t predict the DG revolution will truly begin until about 2030 and suggests
most of it will be CHP, small and large. He gives two possible scenarios for
future power: ‘small-scale wins’ which favors DG and ‘traditional triumphs’
which favors utility-scale centralized power projects on a more traditional
grid. Either way the smart grid will be built.
Fox-Penner shows comparison charts where gas is the cheapest
of all sources as long as gas costs are less than $6 per MCF. Current projections
put gas prices at $3-4.50 through 2020 so gas is by far our least expensive
power source. Solar PV remains the most expensive energy source even with tax
credits. Even with a price on carbon, gas would still be the least expensive
and solar the most expensive. Only if the carbon price were above about $50 per
ton would it change. Nuclear and carbon sequestration require long lead times
and by 2030, he says, we will know if they will be a big part of the picture or
not.
Problems with scenarios of complete renewables are many and
include increasing costs per kWh as more renewables are added to the grid. Even
in the distant future mixed energy sources will likely be used including gas,
hydro, and nuclear – all which can provide baseload power. Thus most power
providers now and in the future have diverse power source portfolios, used to
satisfy demand, reliability, low carbon mandates and pollution mandates such as
the Clean Power Plan. The shift toward renewables and smaller sources is indeed
inevitable but how much and how fast are the questions.
Next he explores in detail new power provider business
models. First is finding a way to commodify and incentivize energy efficiency
(EE) so that providers can profit from it as well as power sales. EE is a straightforward
investment for someone who buys power but not so for those who sell power. EE
also offers the best potential source of carbon emissions and pollution reductions,
so these investments are also emissions reduction investments which makes them
doubly valuable. Thus carbon pricing would further incentivize them as well. He
goes through the barriers to EE adoption: 1) information, 2) capital availability,
3) transaction costs, and 4) inaccurate prices. First accurate information must
be obtained about comparative efficiencies of energy sources and technologies
big and small on each system. This is true on utility-scale as well as
home-scale end-user efficiencies. People especially need to understand efficiency
and the potential benefits before they invest in it. Capital availability is
simply that. Efficiency is an upfront all-capital cost where 100% must be paid
out before savings are realized. Utilities have many projects like new plants,
transmission build-out, and smart grid expansion – so EE must compete for
capital. Transaction costs refer to the disruption effects of contractors
adding the new measures which may cause occupancy delays in new housing and building
projects. Inaccurate price signals refer to the common situation that the
builder will not be the one paying the bills but will be the one choosing the energy
systems and appliances so has no incentive to look at potential energy use
costs. One help to some of these problems is when free energy audits and sometimes
installation advice are provided by organizations or utility companies. Appliance
efficiency standards have saved massive amounts of energy and have saved
consumers massive amounts of money. However, the construction and real estate
industries as well as appliance manufacturers routinely oppose them because
they increase their costs and change construction and manufacturing practices.
Building efficiency codes and appliance efficiency standards are surefire ways
to save energy and emissions and need to be further pursued. Utility efficiency
programs can also be effective – offering free energy audits, low-interest
loans, rebates, and free technical assistance. State government incentives and financing
have also worked to enable EE. Private sector EE has also been fairly successful
through energy service companies (ESCOs) utilizing ‘shared savings’ business
models whereby the ESCO pays for the EE upgrades and recoups their money and a
certain amount of profit over time while the customer enjoys cheaper energy
costs that drop even more when the ESCO is paid off. This is a great model for
companies since the ESCO does all the work but often they don’t do the full EE
upgrade but only partial ones with rapid paybacks except in government
installations where the government mandates full slow payback EE upgrades which
are better in the long run. To mandate utility EE there is the Energy
Efficiency resource Standard (EERS) whereby the utility or seller of electric
power has to meet a certain percentage of sales growth through energy savings
through efficiency upgrades. Thus they are mandated to save a certain amount of
energy in addition to selling it. About half of U.S. states have some version
of EERS. Utilities can often raise capital at lower interest rates – they are
credit worthy because they are providing an essential service. This gives them
an advantage in investing in EE by loaning to their customers. They are an
ideal low hassle financing entity. On the downside utilities face ‘divided
incentives’ – selling energy vs. saving energy. Government control of EE
programs also has downsides: changing administrations changing the rules and complaints
about excessive government loans (as in the DOE paying for weatherizing
upgrades) since EE is capital intensive. Further government control could
involve tax increases which are unpopular. He gives the three approaches as 1)
requiring utilities to control EE against their own financial interests (as is
mostly done today), 2) letting the government do it, 3) changing regulation and
utility business model to give them an incentive to implement EE. He notes that
none of these are done (enough) today. He thinks utilities can do it best and
can find more opportunities for EE if they had an incentive to do so.
Finally, he gets to business models. First he describes the
two “triads” (referring to generation, transmission, and distribution) - business
models in place today. Then he describes what he thinks will be the two possible
models that will dominate the future. The first model of today is the
traditional vertically integrated utility that owns generation, transmission,
and distribution. This is the integrated regulated public/cooperative structure
of 36 states. The second model is the de-integrated, deregulated generation and
regulated grid which occurs in 14 states. The FERC regulated transmission is
open-access – any company can put energy onto it if there is capacity. Vertically
integrated companies have probably been the most economically successful but
that was before the advent of the smart grid and the necessity to reduce
emissions. After deregulation and de-integration in the U.S. and Europe the
percentage of vertically integrated companies dropped but after a while they
came back up again with many re-integrating for the economic advantages. Will
integration with its economic efficiencies survive the smart grid? Regulation
arose due to the belief that power companies were natural monopolies with
economies of scale. These days only transmission and distribution are
considered natural monopolies. Generation and retailing are not considered so. Competition
in generation could lead to cost benefits for consumers. Thus it was generation
that was deregulated in the 1990’s. With deregulation the owners of
transmission and distribution would be required to utilize the lowest cost
generation whether they owned it or not or they would not be allowed even to
own generation. Many now say that as generation becomes less centralized and
more distributed there will be more benefits to deregulated competition. Theoretically
they should work even better with dynamic pricing. Fox-Penner goes into detail
comparing these models. Smart-phone controlled home-energy management
responding to dynamic price signals will be possible when the smart grid is
fully employed and will likely be a major feature of demand response and peak
shaving. Such innovations will likely be easier if regulated distributors have
less control and there is more deregulated competition. Competition and dynamic
pricing are a good match but traditional utilities may be able to manage it as
well.
Now we come to his two (and a half) business models for the
future of utilities. First is the Smart
Integrator (SI) model which is “a utility that operates a regulated smart
grid offering independent power and other services at market prices.”
Decoupling sales to profits can remove disincentives. The second model is the Energy Services Utility (ESU), which “is
vertically integrated, regulated, and must have strong EE incentives built into
its regulatory structure to offset its regulated profit motive.” He notes that these
two models are really not much different at all than the models available today
– just that sales and profits are decoupled in the SI and efficiency is set as
a core mission in the ESU. The other differences are that the ESU may own some
or most of its power supply and is regulated and the regulators set prices
while in the SI the market sets prices. The half model (of the two and a half)
is a Smart Integrator in which distributed generators (DG) are owned mainly by
communities rather than mainly by individuals. Mid-scale generation may be
owned by individuals, businesses, DG management companies, utilities, or
communities. Community energy systems (CES) may become a public power model not
unlike municipal power companies owning mid-scale DG and becoming the CES.
The mission of the Smart Integrator is to deliver power with
superb reliability and maintain mostly downstream wires and assets – but it
does not own or sell the power. The SI will work mainly as a distribution company
(distco) controlling the two-way flow of electrons to and from DG in response
to prices and so must have an ‘open architecture’ format to let more sources in
when needed and turn them off when not needed. The whole two-way system will be
managed with software, sending price signals and switching sources on and off
in real-time. Hourly spot trading prices will have to be determined, but this
mechanism of management can get both expensive and complex. Germany is
currently having to deal with this as millions of new DG resources that send
and receive energy must be managed. Geographical situations where advantages of
local DG providers get too strong may have to be regulated in some way. Determining
the value of DG and DR services in avoiding expansion costs is not easy nor is how
to reward them for their services. Software and software platform evaluation,
costs, and agreement are currently big issues and these problems will have to
be worked out. Information management will become more important as more DG is
added to systems. Software needs to be queried and analyzed in order to provide
decision support. More expertise in IT and regulatory economics will be needed
in addition to electrical engineering. Regulated utilities set rates according
to ‘cost of service’ or rate of return plus an agreed upon profit. This is the
sales component by which rates are set for customers every few years. These
fixed costs per kWh set in such a way do not encourage EE. The sales incentive
exists even for the SI. ‘Decoupling’ is the current solution to mitigating
sales incentives and replacing them with energy savings incentives where the SI
or other utility is paid for saving energy. This is a short-term fix, he says.
Investors will see it as sales declining which is not what Wall Street likes to
see! This is especially the case for companies that own generation. If SIs just
deliver power the drop in sales won’t affect their profit. Another issue is who
provides customer service – will it be the SI itself in a business-to customer
(B2C) or a third party software vender-type provider in a business-to-business
(B2B) format?
The Energy Service Utility (ESU) will differ from the SI in
two main respects. “The ESU will not necessarily have an incentive to cooperate
with local generators who want to connect and sell power into its smart system.”
It may view them as competition. Second is the disincentive to help customers
reduce their power use due to the ESU’s ownership of generation. These two
issues will require regulatory interventions. Letting in local generators, or
open-access, and incentivizing such access will require regulation. The only
difference here between the SI and the ESU is that the ESU owns generation and
other generators would be seen as competitors. Another issue is that utilities
often will go to the bare minimum to comply with any EE mandates rather than
beyond. EE is considered a public goal and should be pursued as much as
possible as such so it needs to be incentivized in such a way as it will indeed
be pursued as much as possible rather than bare minimum. If utilities get to
keep higher percentages of the value of energy they help their customers save
then they will be encouraged to save more – simple as that. This has been
successful with PG & E in California. Thus the public goal of saving energy
is aligned with the business interests of the utility. Even so, the sales
profits from PG & E and others still far outweigh (by nearly 10 times) the
efficiency profits. Next he considers Duke Energy’s Jim Rogers’ “Save-a Watt” program
where home efficiency improvements would become automatically provided by the
utility, the costs avoided by deferring new plants would be part of the utility’s
profit, and the heavily-regulated planning and approval cycles would be
eliminated. The problem is that regulators were/are averse to ceding control of
money allocation and investment to the utilities. Rogers’ plan was seen as too
lucrative for the utility. It was approved in four states after some tweaking
and lowering the profit a bit. Rogers plan called for the possibility of charging
customers not for kWhs but for heat, light, and other units – what he called “value
billing.” Utility executive Ralph Izzo calls a similar idea “universal access
to energy efficiency.” For an ESU to be able to pitch its ability to investors
to paradoxically create more value by selling less energy it will have to be
able to sell its energy services in such a way as to out-compete an SI and size
and site its investments in new generation and transmission in a more precise
way. It will also have to manage effectively the Smart Grid.
Defining utilities missions as selling energy services
rather than energy is not new – Edison first sold light, not power. While Amory
Lovins and others promoted selling energy services back in the 1980’s it was
impractical then because the services could not be measured. Now with IT
technologies, dynamic pricing, and automation we can measure them much better.
The idea can be extended to other realms – instead of buying products we can
rent them while those we rent from take them back to be recycled. Many products
are like this. Software-as-service models are standard in some industries.
Much of the regulatory hurdles to overcome will revolve
around how to measure the benefit of investments, how to allocate system costs,
and how to blend markets and regulations. Changes in regulation, the smart
grid, more decentralized distributed generation, more demand response provided
by distributed generation, more focus on decarbonization, and more regional,
state, and grid-to-grid cooperation will be the requirements of the future as
smart power arrives in full. He notes that public utility commissions have in
the past been unduly criticized for some decisions and that has affected their
propensity to innovate and take risks with new technology which may be required
in the new environment. The California energy crisis and the perceived failure
of deregulation has also had negative effects on experimenting with new models
and technology integration. State commissions, he says, need their independence
restored, and less government oversight. He advocates for regulator training
programs and accreditation for commissioners. They need to be well-educated on
current trends and technologies. Basically, the public utility commissions need
an upgrade, he says. Decentralized DG
will include more and more community-owned sources, co-operatives, and
community municipal utility-owned sources. Their ESU formats will be slightly
different than those of IOUs.
Fox-Penner suggests that the SI and ESU models will likely
both occur and if one is revealed as better, then that one will win out but it
may also be that both will remain. He calls for better national policy on EE
and to better inform state regulatory policies – but other than that no new federal
laws are needed.
In the Afterword he gives four pillars to the new power
paradigm: adequacy, reliability, universally affordable service, and rapid
decarbonization. These changes will be paid for by customers and financed by
investors. There is the grid and the increasingly important “grid edge” which
is the decentralized distribution and microgrid technologies that will grow. So
far, he says, Germany, California, New York, and Hawaii have been at the forefront
of these changes to new models. These changes have not been without problems.
Traditional utilities in Germany have been hit with revenue losses and low spot
prices. When this has happened the U.S. the utilities have requested higher
rates for all customers. While they do see some saving from DG it has not been
enough to offset revenue losses. Such revenue losses will cause distribution
companies to become Smart Integrators while government and customer-owned power
companies have moved toward the Energy Services Utility model.
This book is essential reading for anyone wanting to
understand the dynamic state of the current and future power industry. New
developments happen frequently. Battery and other energy storage is now
entering the picture more and more, both at home-scale and utility-scale. There
are state battles occurring with net metering and feed-in tariff rules. There
are new renewables mandates and some that have been scaled back. Gas continues
to replace coal. Wind continues to grow. Future nuclear and CCS-endowed coal
continued to be uncertain but may happen in the 2020’s. More microgrids,
community energy, and dynamic pricing experiments are happening. Regional
carbon pricing mechanisms have been established and are functional. EVs are set
to take-off in the 2020’s and 2030’s. Efficiency improvements continue at all
levels. I enjoyed this one as energy is a big interest of mine.
No comments:
Post a Comment